text
stringlengths
6
128k
# A Trust-Centric Privacy-Preserving Blockchain for Dynamic Spectrum Management in IoT Networks Jingwei Ye, Xin Kang, , Ying-Chang Liang, , and Sumei Sun J. Ye, X. Kang and Y.-C. Liang are with the Center for Intelligent Networking and Communications (CINC), University of Electronic Science and Technology of China (UESTC), Chengdu 611731, China (e-mail<EMAIL_ADDRESS><EMAIL_ADDRESS>liangyc@ieee.org).S. Sun is with the Institute for Infocomm Research, Agency for Science, Technology and Research, Singapore (e-mail: sunsm@i2r.a-star.edu.sg). ###### Abstract In this paper, we propose a trust-centric privacy-preserving blockchain for dynamic spectrum access in IoT networks. To be specific, we propose a trust evaluation mechanism to evaluate the trustworthiness of sensing nodes and design a Proof-of-Trust (PoT) consensus mechanism to build a scalable blockchain with high transaction-per-second (TPS). Moreover, a privacy protection scheme is proposed to protect sensors’ real-time geolocatioin information when they upload sensing data to the blockchain. Two smart contracts are designed to make the whole procedure (spectrum sensing, spectrum auction, and spectrum allocation) run automatically. Simulation results demonstrate the expected computation cost of the PoT consensus algorithm for reliable sensing nodes is low, and the cooperative sensing performance is improved with the help of trust value evaluation mechanism. In addition, incentivization and security are also analyzed, which show that our design not only can encourage nodes’ participation, but also resist to many kinds of attacks which are frequently encountered in trust-based blockchain systems. ###### Index Terms: Dynamic spectrum access, blockchain, trust model, consensus algorithm, cooperative spectrum sensing. ## I Introduction Recent years have witnessed the exponential growth of mobile data traffic. The rapid development of new wireless applications such as autonomous vehicles, remote healthcare, augmented and virtual reality will continue driving the demand for radio spectrum[1]. However, the current static spectrum management method has resulted in under-utilization of spectrum resources[2]. To this end, dynamic spectrum access (DSA) has been proposed to allow secondary users (SUs) sense and then use the idle spectrum bands. Since single SU’s sensing capability is usually limited, cooperative sensing[3] is proposed to fuse the sensing results of multiple SUs to improve the sensing accuracy. However, the traditional usage of a centre to collect and fuse sensing results faces the risk of single point of failure and the increasing management and regulatory costs. Moreover, SUs usually need to upload their sensing data to the fusion center at the risk of leakage of their privacy, such as identity and location. As an emerging technology, blockchain has attracted attention from both academia and industry. By using blockchain, a decentralized resource management system can be established without the requirement of trustworthy central management agencies[4, 5, 6, 7, 8]. Moreover, smart contracts which are implemented on a blockchain can be used to replace traditional central management agencies and facilitate the cooperation of users [9]. Recently, researchers have investigated how to apply the blockchain technology to DSA. [10] summarizes the types of blockchains applicable in different spectrum sharing scenarios, and the possible advantages and disadvantages of applying blockchain technology to DSA. With the help of cryptocurrency issued by the blockchain and smart contracts, flexible and automatic spectrum trading markets are made possible for spectrum sellers and buyers. [11, 12, 13, 14, 15]. However, the integration of blockchain and DSA still faces many challenges. Firstly, although the data recorded on a blockchain is prevented from being tampered, the quality or value of it cannot be guaranteed. Especially, in a public blockchain, a malicious user can easily join the blockchain and record their data in the blockchain. The data from such a malicious user or an unreliable user is valueless or even harmful to cooperation tasks based on the data, e.g., cooperative sensing. Therefore, there is compelling need to evaluate the quality of the data from each user. Secondly, though the communication among Ethereum external accounts can be protected by encryption and decryption with their public and private key pairs, a contract account is not equipped with a key pair hence such protection method is not applicable to smart contracts. As a result, sensing results uploaded to the smart contract cannot be encrypted, and the sensitive information in the sensing results (such as real-time geolocatioin), can thus be accessed by malicious users. Thirdly, traditional consensus algorithms like Proof of Work (PoW) used in public blockchain introduce much computation overhead and thus make the transaction processing speed very low. As a result, it is inefficient to directly use existing consensus algorithms (such as PoW) for spectrum management, especially considering the limited computation capabilities of IoT devices. In this paper, we consider the DSA for an IoT network where IoT devices opportunistically access the licensed spectrum bands through cooperative sensing. Blockchain is used as a platform for dynamic spectrum management. Specifically, key component designs for such a blockchain-enabled DSA system, including the trust evaluation mechanism, privacy protection, consensus algorithm and smart contracts, are studied. The main contributions of this paper are summarized as follows. * • We propose a trust evaluation mechanism to evaluate the trustworthiness of sensing nodes which participate in the collaborative sensing. We show that the proposed trust evaluation mechanism is effective in incentivizing sensing nodes to be honest, and thus improving the spectrum sensing result. * • We propose a strong privacy protection mechanism by combining ring signature with the commit-and-reveal scheme for providing sensing nodes’ privacy such as their real-time location. The ring signature is used to protect sensing nodes’ identities when they upload sensing data. The commit-and-reveal scheme is proposed to figure out the connection between sensing data packets and its originated sensing node after the fusion of sensing results, so that the trust value of each sensing node is updated and the incentive tokens are distributed. * • We propose a new consensus algorithm named as Proof-of-Trust (PoT) by connecting the mining difficulty with a miner’s trust value. We show that the proposed PoT can greatly reduce the computation cost of honest nodes and enhance the scalability of current blockchian-based DSA system. * • We design the system protocol and two smart contracts to make the whole DSA procedure, including spectrum sensing, spectrum auction, and spectrum allocation, run automatically. In addition, we implement and verify a prototype of our proposed protocol and smart contracts using Solidity[16] on the Ethereum test net. The rest of the paper is organized as follows. In Section II, we introduce the basic concept of blockchain and smart contract. In Section III, we describe the works related to the application of blockchain to DSA. In Section IV, we introduce our system model. In Section V, we describe the design of our proposed protocol including the block structure, the trust evaluation mechanism, the consensus algorithm and the privacy protection scheme. In section VI, the design of smart contract and workflow are introduced. In Section VII, we discuss the performance of the trust evaluation mechanism, and analyze the security of the system. Finally, Section VIII concludes this paper. ## II Preliminaries ### II-A Blockchain Blockchain is a growing chain of data blocks with a specific data structure constructed using cryptographic algorithms. It uses hash pointers to connect the data blocks to form an entangled chain. In this way, the integrity of the data can be protected. Blockchain uses a measurable and verifiable mechanism to reach a consensus among nodes in the network on the generation of new blocks. Such a mechanism is usually referred to as the consensus algorithm. On the premise that the data block entanglement is guaranteed, the structure and the consensus algorithm of the blockchain can be flexibly designed based on the demand of application scenarios. Blockchain can be roughly classified into three types, depending on degree of openness, which are public blockchain, private blockchain, and consortium blockchain. Public blockchain is designed to enable and to record the transactions in a public network, which means that any node can freely join or leave the blockchain network without authorization. Bitcoin, the most famous digital cryptocurrency, is generated and managed by a public blockchain. Consortium blockchain is a permissioned blockchain, which is open to authorized members of external institutions in limited roles and functions. The authorization of the node to join are all determined by an authorized organization. The nodes in a private blockchain network trust each other. Thus, private chains can simplify operations in data authentication and the consensus algorithm to improve the efficiency. ### II-B Smart Contract In the 1990s, Nick Szabo first proposed the concept of smart contracts[17]. It was defined as computerized transaction protocols that execute terms of a contract. In the field of cryptocurrency, a smart contract is defined as an application or a program that runs on a blockchain. Normally, it contains a set of digital agreements with specific rules. These rules are predefined in the form of computer source codes, and all network nodes will copy and execute these computer source codes independently. Smart contracts are highly customizable and can be flexibly designed to provide different services and solutions. ### II-C Ring Signature Ring signatures were first introduced in 2001[18] which follow a ring-like structure of the signature algorithm. It is a type of digital signature that can be performed by any member of a group of users without the agreement of others. Therefore, the only “open” information is that a message signed with a ring signature is endorsed by someone in a particular group of people. Ring signatures are deliberately designed so that it is computationally infeasible to determine which of the group members’ keys was used to produce the signature. Ring signatures are similar to group signatures but differ in two key ways: firstly, there is no group administrator to revoke the anonymity of an individual signature; secondly, any set of users can be used as a group by the signer without additional setup. ## III Related work Blockchain was first proposed as a secure decentralized ledger for recording spectrum transaction data and analyzing overall spectrum usage in dynamic spectrum management. In [19], the authors proposed to use the blockchain as a distributed database to securely store the transactions in dynamic spectrum sharing, and to use the digital currency issued by the blockchain as the currency for spectrum auctions. In [15], a smart contract running on Ethereum was proposed as a platform to provide spectrum sensing services. In [13], a blockchain-based spectrum trading and sharing scheme was proposed for Mobile Network Operators (MNOs) to lease their spectrum to secondary Unmanned Aerial Vehicle (UAV). In [20], spectrum sharing contracts deployed on a permissioned blockchain platform are constructed for multi-operators spectrum trading. In [14], the authors proposeds to use the smart contract to securely store and then autonomously implement the spectrum leasing agreements, thus to avoid the interference from the disordered spectrum access. However, these works did not make any improvement to the existing blockchain. Thus, the TPS of the blockchain is low, limiting the performance of the blockchain-based DSA platform. Thus, in [21], the authors proposed a consensus algorithm named as "blockchain-KM protocol" by modifying the PoW, to speed up the transaction processing for the license-free spectrum sharing. In [22], a novel consensus algorithm named as proof-of-strategy was proposed, where the best strategy to allocate the spectrum bands in the whole network is the proof of authority to publish a new block. Such a consensus algorithm can not only regulate the transaction generation, but also optimize the spectrum allocation. In [23], by modifying the PoW, the authors proposed a consensus algorithm named as Proof-of-Device, which selects a device, represented by its unique identifier, to publish a new block in a way like lottery. It thus eliminates the heavy computation cost in PoW. On the other hand, blockchain system is an open system, hence the reliability of data source needs to be guaranteed. In [24], the data uploaded by the sensors are compared with the data collected by the trustworthy validator to determine the credibility degree of the sensor, in this way, the system can evaluate the credibility of nodes for better performance. However, this method is a semi-centralized method which needs to deploy trustworthy validators. Different from the above works, we not only delve into the design of block structure and consensus algorithm, intending to overcome the scalability problem of the blockchain technology, but also design a decentralized trust evaluation mechanism and a privacy protection mechanism for the openness and transparency of blockchain system while protecting the data’s confidentiality and the sensors’ privacy. To the best of our knowledge, this is the first work of joint trust evaluation and privacy protection mechanism in cooperative sensing in a blockchain-enabled DSA system. ## IV System Model, challenges and design guidelines In this section, we introduce the dynamic spectrum sharing scenario in the future IoT network, and list the difficulties and challenges in applying the blockchain to this scenario. Finally, we give the guidelines on how to design a blockchain-based DSA system. ### IV-A Blockchain-enabled DSA System Model In this paper, we consider the DSA in a two-tier cognitive radio network consisting of licensed/primary users (PUs) and unlicensed/secondary users (SUs). In this network, PUs are inactive users who may not use their spectrum all the time so SUs can detect these spectrum holes and use them. As illustrated in Fig.2, SUs can be a variety of heterogeneous IoT devices, which form a mesh network. SUs access the spectrum bands of a PU in an opportunistic manner based on cooperative sensing, and a blockchain is used as the platform of the cooperative spectrum sensing decision and the spectrum allocation. Figure 1: System model Basically, as shown in Fig. 3, there are four phases in our blockchain-enabled DSA system. 1. 1. Individual Sensing: The sensors perform spectrum sensing to detect the state of the PU’s spectrum band. 2. 2. Sensing Fusion: Each of the sensor broadcasts its sensing data which mainly consists of the sensing result and the geolocation where the sensing is performed. These data can be raw data or quantized bits, which will be fed to a smart contract implemented upon a blockchain. Then, the smart contract provides the final sensing result according to predefined fusion rules. 3. 3. Spectrum Allocation: When the final sensing result indicates that PU is inactive, it should be decided which spectrum requester can access the spectrum. Spectrum auction, as a common and fair way for spectrum allocation, is chosen to achieve that. In the spectrum auction, the digital cryptocurrency, which is issued by the blockchain through incentive mechanisms for spectrum sensing and mining, is commonly considered. 4. 4. Spectrum Access: The SU who obtains the access right through spectrum auction accesses the idle spectrum for data transmission. Compared with traditional cooperative spectrum sensing and spectrum allocation methods, blockchain-enabled DSA has the following benefits: * • Decentralized: Blockchain-based dynamic spectrum access does not require the deployment of trusted central nodes which avoids the single point of failure. * • Transparency: The blockchain ledger can record the whole process of DSA. * • Automatic: Using smart contract instead of traditional contracts, we can achieve automatic spectrum management and on-chain payment settlement. * • Flexibility: For spectrum trading on the blockchain platform, diversified spectrum trading rules can be dynamically enforced by adjusting parameters of smart contracts. Compared with the current fixed paper-based contracts signed between mobile users and operators, this method has higher flexibility. Figure 2: Blockchain based DSA ### IV-B Difficulties and Challenges Traditional cooperative spectrum sensing is divided into two modes: centralized mode and distributed mode. Centralized collaborative spectrum sensing requires the deployment of a central fusion node to process the data collected from sensors which usually contain sensors’ geolocation information. Therefore, this mode not only requires redundant regulation fees (usually charged to maintain the operation of the central node), but also faces the risk of data leakage when the central node is attacked. While the distributed mode often suffers from the fake reporting issue caused by malicious nodes and the iterative-based malicious node detection algorithms [25] introduce extra computation cost. Therefore, how to implement the cooperative sensing in a secure and effective way is a challenge. The decentralized nature of the blockchain system is promising to realize distributed cooperative sensing and encourage all sensors to share spectrum sensing results. However, there are still some challenges when applying blockchain into spectrum management. Firstly, IoT devices usually lack sufficient computing and storage resources to maintain and store the whole blockchain ledger. Secondly, a malicious node can easily join a public blockchain network and threaten the reliability of the cooperative spectrum sensing. Moreover, the sensing-related data recorded on blockchain or collected by smart contracts is usually in an unencrypted format which may leak the private information of IoT devices contained in their sensing data packets. ### IV-C Design Guidelines Considering above challenges, our discussion on system design mainly includes the following four parts. * • _Trust Evaluation_ : Specially, in a public blockchain network, anyone can join or exit the system at will, and one user can create multiple accounts in the system. Therefore, if there is no suitable node evaluation mechanism, malicious users can launch sybil attacks easily. This is because, malicious nodes can keep on applying for accounts and continue launching attacks. Therefore, the system should be able to evaluate the trust values for nodes, and identity the malicious nodes based on the their trust values. * • _Lightweight Consensus Algorithm_ : Due to the limitation of the computing power and storage resources of IoT devices, maintaining a normal public blockchain system is a heavy burden. To reduce the computing cost, it is necessary to invent a new consensus algorithm with low computational and storage complexity. On the other hand, as time goes by, the length of the ledger keeps on increasing, resulting in increased node storage. To reduce the storage cost, methods such as edge storage can alleviate storage pressure on the blockchain nodes but may introduce additional budget. Thus, a method that can directly reduce the storage cost of the blockchain is preferred. * • _Privacy Protection Mechanism_ : Smart contracts does not have public-and- private key pair and thus can not communicate with external accounts in an encrypted manner. Therefore, when a smart contract is used as the data fusion center of cooperative spectrum sensing, a privacy protection mechanism is needed to ensure that the private information of the sensors will not be leaked during the interaction between sensors and the smart contract. ## V Blockchain Design, Trust Evaluation and Privacy Protection In this section, to improve the accuracy of cooperative sensing, we first introduce the block structure of our blockchain. We then propose the trust evaluation mechanism to evaluate the trustworthiness of each SU on spectrum sensing. After that, a high-efficiency consensus algorithm is proposed for our blockchain. Finally, a privacy protection mechanism is proposed for protecting participants’ privacy. TABLE I: List of notations Notations | Descriptions ---|--- $N_{i,w}(n)$ | Accumulated number of wrong sensing results of $sensor_{i}$ at round $n$ $N_{i,r}(n)$ | Accumulated number of right sensing results of $sensor_{i}$ at round $n$ $P$ | Length of window for calculating $N_{i,w}(n)$ $\mathcal{D}^{i}_{n}$ | Mining difficulty of node $i$ at block $n$ $\mathcal{H(\cdot)}$ | Hash operation $R_{sleep}$ | Number of inactive rounds since last active round $RND_{i}$ | Random number picked by $sensor_{i}$ $SR$ | Sensing result of unknown but legal sensors $SR_{i}$ | Sensing result of $sensor_{i}$ $N_{1}$ | Number of required sensors in Cooperative Sensing Contract ($CSC$) $N_{2}$ | Number of required bidders in Spectrum Auction Contract ($SAC$) $TV_{thr}$ | Trust value threshold set in $CSC$ $T_{ddl}$ | Sensing deadline in $SAC$ $d_{s}$ | Deposit for sensing $d_{a}$ | Deposit for auction $T_{self-d}$ | Time for $SAC$ to self-destruct $TS$ | Timestamp $loc$ | Location of unknown but legal sensors ### V-A Block Structure The block structure should be tailor-designed for the blockchain-based DSA system. Some of the unnecessary components in the block header can be removed to reduce the overheads on the block transmission and storage and other necessary information, such as the trust value of the block generator, needs to be recorded in the block header to prevent possible attacks in the DSA system. * • Block header: As shown in Fig. 3, The block header mainly contains (i) the merkle root of transactions; (ii) the merkle root of account states; (iii) the hash value of the header of the previous block, as the hash pointer of the chain structure; (iv) the trust value of the miner who mines this block; (v) the signature of the miner who mines this block; (vi) a Nonce, which is the solution of Hash puzzle of the consensus algorithm; (vii) the timestamp when the block is mined. * • Block body: The block body mainly stores transaction generated during spectrum allocation using merkle tree. Unlike the unspent transaction output (UTXO) [26] model used in Bitcoin, the system based on the account model usually contains an account tree to record the account states such as the balance and the trust value of each account. Merkle Patricia Trie (MPT) [27] which is a commonly adopted data structure, is used here to store the account states. Figure 3: Block structure ### V-B Trust Evaluation Mechanism Although the blockchain can guarantee that the data recorded in a block will not be tampered, it cannot guarantee that the data source is trustworthy. Especially, in a public blockchain, a malicious node can upload misleading data which may interfere the cooperation sensing based on such data. In such cases, the performance of the cooperative sensing will be degraded, and the honest and reliable sensors will be discouraged from participating in cooperative sensing. To this end, we design a trust evaluation mechanism to evaluate the credibility of every sensing node in the blockchain. Instead of using a central authority to record and maintain every account’s trust value, we propose to include the trust value as an attribute of the blockchain account, which is recorded in the block and maintained by all the miners. In the following, we will discuss the design of trust value updating method. Intuitively, the trust value of a sensing node should be adjusted based on its performance on the current and historical cooperative sensing. In detail, if it plays a positive role in the cooperative sensing, its trust value should be increased, and vice versa. Here, we define that the effect of a sensor is positive when its sensing result is consistent with the cooperative sensing result, and is negative when its sensing result is inconsistent with the cooperative sensing result, or when the sensing node does not participate in the cooperative sensing. Denote the number of times when the sensing result of the i-th node is consistent with the cooperative sensing result as $N_{i,r}$; the number of times when the sensing result of a node is inconsistent with the cooperative sensing result as $N_{i,w}$. Inspired by [28], we propose a new model to calculate trust value of $sensor_{i}$ denoted by $TV_{i}^{(1)}(n)$ in sensing round $n$, where $n$ is also the number of sensing tasks published since this system was created. $\displaystyle TV_{i}^{(1)}(n)=e^{-\rho N_{i,w}(n)}\left(1-e^{-\eta N_{i,r}(n)}\right),$ (1) where $\rho>0$ and $\eta>0$ are coefficient that determine how fast the trust value changes with respect to $N_{i,w}$ and $N_{i,r}$ , respectively. Moreover, we let $N_{i,w}(n)$ decay by $\frac{1}{P}$ every time when the node participates in sensing, and we only consider the latest $P$ sensing rounds of each node. In this way, the effect of a wrong sensing result on the trust value will gradually be degraded over time. As a result, a node that unintentionally submitting an inconsistent result will be gradually forgotten. Mathematically, $N_{i,w}(n)$ can be defined as $N_{i,w}(n)=\sum\limits_{m=n-P}^{n}r_{i}(m)\left(1-\frac{n-m}{P}\right),$ (2) where $r_{i}(m)=1$ if $sensor_{i}$ broadcasts the wrong sensing result in round $m$, and $r_{i}(m)=0$ otherwise. Besides, we denote the number of rounds when a sensing node is inactive as $R_{sleep}$, and we add a function $f(R_{sleep})$ which models the negative impact of being inactive on the trust value. It is designed to satisfy the following criteria. * • ${\forall}R_{sleep}\in[1,+\infty)$, $f(R_{sleep})\in(0,1]$, * • $\frac{\partial f_{dcv}}{\partial R_{sleep}}<0$, The first criterion is for normalization; the second criterion is to ensure that the negative effect increases as $R_{sleep}$ increases. Moreover, as $R_{sleep}$ increases, the downward trend of the function is first gentle, and then becomes severe. This is designed to punish the peers that does not participate in the sensing process. The punishment is light when the peer does not participate only a few times, but becomes severe when the node always does not participate in the sensing. $f(R_{sleep})$ will reach a certain low value when $R_{sleep}$ is big enough. For example, the trust value of a node will reduce to $r_{1}$ ($r_{1}<1$) of its original value for $k_{1}$ consecutive non-participation, and to $r_{2}$ ($r_{2}<r_{1}<1$) for $k_{2}$ consecutive non-participation. Here, a piecewise function consisting of multiple linear functions is utilized as our $f(R_{sleep})$. $f(R_{sleep})=\left\\{\begin{array}[]{lcl}\frac{k_{1}-R_{sleep}}{k_{1}}\cdot(1-r_{1})+r_{1}&&{0\leq R_{sleep}\leq k_{1}}\\\ \\\ \frac{R_{sleep}-k_{2}}{k_{1}-k_{2}}\cdot(r_{1}-r_{2})+r_{2}&&{k_{1}\leq R_{sleep}\leq k_{2}}\\\ \\\ r_{2}&&{k_{2}<R_{sleep}}\\\ \end{array}\right.$ (3) These parameters can be set flexibly. If the actual deployment requires stricter penalties, then all these parameters should be set smaller. By adding the attenuation function, the model in (1) can be modified as: $TV^{(2)}_{i}(n)=\begin{cases}TV_{i}^{(2)}(n-1)+\Delta TV_{i}\cdot f(R_{sleep}),&\mbox{if $\Delta TV_{i}>0$}\\\ \\\ TV_{i}^{(1)}(n),&\mbox{if $\Delta TV_{i}<0$}\\\ \\\ TV_{i}^{(2)}(n-1)\cdot f(R_{sleep}),&\mbox{if inactive}\end{cases}$ (4) where $\Delta TV_{i}$ = $TV_{i}^{(1)}(n)-TV_{i}^{(2)}(n-1)$ is defined as the original increment of trust value, which can be either positive or negative. The design criterion of (4) is explained as follows. We consider three cases. First of all, when the sensor senses correctly, the original increment of trust value is positive, i.e., $\Delta TV_{i}>0$. Such increment is first degraded by the negative effect of sleeps in the previous sensing activities, i.e., $f(R_{sleep})$, and then added to the trust value. Secondly, when the sensor gives the wrong sensing result, the original increment of trust value is negative, i.e., $\Delta TV_{i}<0$. In this case, we choose not to consider the negative effect of the sleep in previous sensing rounds, in order to encourage the sensors which unintentionally derive the wrong sensing result in this time. Finally, for sensors who do not participate in sensing, their trust value will not remain unchanged but gradually decrease as their sleep time increases. This design is to ensure that sensors can only maintain its high trust value by actively participating in spectrum sensing. ### V-C Consensus Algorithm, Forking and Block Compression Based on the last block of the block chain, nodes in the blockchain network will compete for publishing the next block, and only one block will be accepted as the next valid block. The traditional consensus algorithm for public blockchain like PoW is computation-intensive, which may be inapplicable in IoT networks, where IoT nodes are usually with limited computational capabilities. One reason why the PoW is designed to be computation-insensitive is that it assumes there is nearly no trust among nodes in a blockchain network. Nevertheless, with the trust value mechanism proposed in this paper, we can evaluate the credibility of a node on spectrum sensing in the blockchain network. Therefore, based on the trust value, we can optimize our design of the blockchain consensus algorithm, forking solution and blockchain compression. Consensus Algorithm: Intuitively, the nodes with higher trust value on sensing are more likely to be honest in publishing a new block. Therefore, we can reduce the difficulty of such nodes to mine a new block. In this way, the computation consumption for reliable nodes will be decreased, and a block will thus be more quickly mined and published. Based on PoW and trust value, we propose a Proof of Trust (PoT) consensus algorithm. Similar to the Proof-of-Stake (PoS), PoT assigns different nodes different mining difficulties. To be specific, for successful mining, the miner with the higher trust value is required to find a hash value with fewer leading zeros and vice versa. Mathematically, assuming that the difficulty of node $i$ at block $n$ is denoted as $\mathcal{D}^{i}_{n}$, which can be denoted as $\mathcal{D}^{i}_{n}=\beta_{n}\cdot\left(1-sin\left(\frac{\pi}{2}\cdot TV_{i}\right)\right),$ (5) where $\beta_{n}$ denotes the base difficulty of block $n$. $TV_{i}$ is the trust value of node $i$. The initial difficulty is denoted as $\beta_{0}$. $\beta_{0}$ can be determined by evaluating the computing power of actual IoT devices. We will discuss this problem in the simulation section. A node with $TV_{i}=0.8$ has about $0.049\cdot\beta_{n}$ difficulty, which is 17 times less than a node with $TV_{i}=0.1$ whose difficulty is about $0.844\cdot\beta_{n}$. We denote the timestamp of block $n$ as $T_{n}$, the ideal time interval of two consecutive blocks is $T_{0}$. Then,we have $q=\left\lfloor\frac{T_{n-1}-T_{n-2}}{T_{0}}\right\rfloor,$ (6) where $q$ is defined as adaptive adjustment factor of base difficulty, and $g$ is defined as the difficulty adjustment granularity, which can be written as $g=\left\lfloor\frac{\beta_{n-1}}{128}\right\rfloor,$ (7) The updating of $\beta_{n}$ can be denoted as $\beta_{n}=\beta_{n-1}-q\cdot g,$ (8) Forking Solution: As the mining speed increases, there may be multiple blocks mined at nearly the same time. Because of the communication latency, the block that is first mined might not be the first to be received by all the nodes in the blockchain network. Instead, the block that is first received and recognized by different nodes might be different. In this case, the blockchain forking occurs. It is harmful and thus needs to be solved. In Algorithm 1, we propose a trust value based forking solution. It first compares the trust values in the head of blocks, and it then validates the block with the highest trust value. If the trust values of multiple blocks are the same, the block whose timestamp is earlier is selected as the valid block. If the timestamp is the same, it compares the hash value of the blocks, and then selects the block with the smallest hash value as the valid block. Since the probability of hash collision is negligible, Algorithm 1 can select one valid block eventually. Blockchain Compression: Since IoT devices are usually with limited storage space, each time when the blockchain grows by $L$ blocks, compression will be performed. For the compression of blockchain, in the existing literature, the RSA accumulator as a data structure, which functions similarly to that of a Merkle tree, can be used to compress the blockchain [29]. Another approach to compress blockchain is to use the chameleon hash function to replace the traditional hash function of the blockchain [30]. Here, using the trust value, we propose a compression method called as Trust-based Compression (TBC). The node with the highest current trust value will be authorized to compress the blockchain. To be specific, such a node will first extract the account tree in the last block as the body of the new block, calculate a hash value for the body, and finally combine the obtained nonce, miner’s signature, and timestamp to form the block header. The obtained block is the new genesis block. Since each node stores the original blockchain, it is easy to check whether the account status has been tampered by the selected node during the compression process. After a node verifies the new genius block, it will clear the original blockchain. Algorithm 1 Block Selection Input: the block to be compared: $Block_{i}$, $Block_{j}$; Output: The winner block; if $TV_{i}<TV_{j}$ then $Block_{i}$ wins. else if $TV_{i}>TV_{j}$ then $Block_{j}$ wins. else if $Timestamp_{i}<Timestamp_{j}$ then $Block_{i}$ wins. else if $Timestamp_{i}>Timestamp_{j}$ then $Block_{j}$ wins. else if $\mathcal{H}(Block_{i})<\mathcal{H}(Block_{j})$ then $Block_{i}$ wins. else $Block_{j}$ wins. end if end if end if ### V-D Privacy Protection Mechanism The location where a sensing node senses the spectrum bands of interest is useful for the fusion centre to cluster the sensing nodes and to improve the cooperative sensing accuracy [31]. Therefore, alongside the sensing result, the location is needed to be uploaded by a sensing node. However, it is difficult to protect the location information of a sensing node from being leaked when a smart contract is used as the fusion centre. This is because a smart contract account is not equipped with a key pair which can be used by sensors to encrypt their upload data packet. The privacy protection issue in this case is to hide the source of a sensing packet, i.e., from which sensing node the sensing packet comes. However, the sensing node cannot be allowed to be totally anonymous when submitting the sensing packet because this will make the cooperative sensing system vulnerable to malicious attacks. To this end, we propose the use of the ring signature [18] by each sensing node to hide the source of a sensing packet in a group of valid sensing nodes. Also, the smart contract as the fusion centre can identify the validity of each received sensing packet. Moreover, since sensing packets are unencrypted, fusing these sensing results can be carried out directly and automatically in the smart contract. In the following, we illustrate the procedures of one sensor, denoted by $Sensor_{s}$, in generating the ring signature. * 1. $Sensor_{s}$ selects $n-1$ legal sensors to form a group and collect their public keys. The format of a sensing data packet, denoted as msg, is given as follows $msg=\\{\mathcal{H}(msgID),SR,time,location\\},$ (9) where $\mathcal{H}(msgID)$ is used in the update of the trust value and $SR$ is the sensing result denoted by one bit. * 2. $Sensor_{s}$ uses one-way hash function to compute $k=\mathcal{H}(msg)$, which is a symmetric key of the symmetric encryption function $E_{k}$. * 3. $Sensor_{s}$ generates a random value for each of other members in the group where it belongs to. Specifically, the random number $x_{i}$ is first generated for the $i$-th node in the group. It then calculates corresponding $y_{i}$=$g_{i}(x_{i})$ using corresponding public key, where the function $g_{i}(\cdot)$ is the encryption function encrypted with the public key $pk_{i}$. * 4. $Sensor_{s}$ finds the solution to the ring equation (10) and gets the undetermined parameter $y_{s}$, where $v$ is a random value chosen by $Sensor_{s}$. Then, $Sensor_{s}$ calculates $x_{s}$ using its own privacy key: $x_{s}$=$g_{s}^{-1}(y_{s})$. To find solution, a private key of a sensor in this group is needed , anyone who is not in the group cannot generate a legal ring signature. $\begin{split}C_{k,v}(y_{1},y_{2},...,y_{n})&=E_{k}(y_{n}\oplus E_{k}(y_{n-1}\oplus E_{k}(...\\\ &\oplus E_{k}(y_{1}\oplus v)...)))=v,\end{split}$ (10) Finally, a valid ring signature, denoted as $Ring_{sig}$, can be generated by $Sensor_{s}$. $Ring_{sig}=(pk_{1},pk_{2},...,pk_{n},v,x_{1},x_{2},...,x_{n}),$ (11) The signature verification by the smart contract as the fusion centre involves the following three steps. * 1. Calculate all $y_{i}$ using corresponding public key $pk_{i}$ and $x_{i}$; * 2. Calculate the symmetric key $k=\mathcal{H}(msg)$; * 3. Verify that if $C_{k,v}(y_{1},y_{2},...,y_{n})=v$ holds. If the verification succeeds, the smart contract recognizes that $msg$ is sent by a valid sensing node from the group $\\{Sensor_{1},Sensor_{2},...,Sensor_{n}\\}$. In this way, we can cut the connection between the data packet and its corresponding owner. However, for the trust value update and payments assignment process, it is necessary to know the mapping of the sensing result and its corresponding sensor. To this end, we propose a two stage commitment scheme as follows. This scheme consists of two stages including Commit Stage and Reveal Stage. As shown in Fig. 4, at the commit stage, each sensor participating in the cooperative sensing needs to submit the $msg$ whose privacy is protected by the ring signature, and the field “hash of $msgID$”, i.e. $\mathcal{H}(msgID)$, is used as a commitment. $msgID$ is an identifier which is the input of hash function, and can be any message such as "I am User 3" or "I like apples". At the reveal stage, each sensor directly uploads the unhashed $msgID$ for miners to verify the consistency. This checking mechanism is effective since hash function is preimage resistance and second-preimage resistance [32], i.e., given a hash output $\mathcal{R}$=$\mathcal{H}(msgID)$, it is difficulty to find the input $msgID$ or another input $msgID^{\prime}$ such that $\mathcal{H}(msgID)$ = $\mathcal{R}$ or $\mathcal{H}(msgID^{\prime})=\mathcal{R}$. Figure 4: Two-Stage commitment scheme. ## VI Smart Contracts and DSA Protocol Design Smart contracts are computer codes that run on the blockchain platform, which are automatically executed when the predefined conditions are satisfied. Moreover, smart contracts are not controlled by any third party. Therefore, we propose to use smart contracts to realize the automated operation of cooperative spectrum sensing and spectrum auction on blockchain. In this section, we will give the designs of the corresponding smart contracts, and then propose the DSA protocol based on these smart contracts. ### VI-A Smart Contract Design We first discuss the design of our smart contracts for cooperative sensing and spectrum auction. In the following, we describe the parameters and functions in the two smart contracts, respectively. _Cooperative Sensing Contract (CSC)_ : The parameters in CSC that need to be predefined are $T_{ddl}$, $d_{s}$, $N_{1}$ and $TV_{thr}$, which are introduced as follows. * • $T_{ddl}$: The deadline for a sensor to send its sensing packet to CSC; * • $d_{s}$: The deposit that a sensor needs to pay before participating in sensing, which is used to be a guarantee for acting honestly. The deposit can be withdrawn only if the sensor uploads the sensing result which is consistent to the cooperative sensing result.. * • $N_{1}$: The maximum number of sensors needed in cooperative sensing. If there are more than $N_{1}$ sensors who apply to participate in cooperative sensing, those with $N_{1}$ highest trust value will be selected. * • $TV_{thr}$: the minimum trust value of a sensor is needed to participate in cooperative sensing. The functions in CSC are introduced as follows. * • $SensorRegister(\cdot)$: The inputs to this function are address $Addr$, deposit $dpt$ and trust value $TV$ of the node who invokes this function. This function will decide whether this node can register successfully by checking its trust value. * • $CheckRegisterQuality(\cdot)$: This function is used to check node’s quality of participating in sensing by checking the parameters passed into it, i.e., node’s trust value $TV_{i}$ and deposit $deposit_{i}$. For new users, participating in sensing is the only way to improve their trust value. This function offers users a way to convert their tokens to additional trust value, which enables the users whose trust value is below threshold to participate in sensing by paying more deposit. * • $Fusion(\cdot)$: The input to this function is $msgList$, which is a list consisting of all the sensing results from the registered sensors. This function outputs the final sensing result according to the pre-defined fusion rule. * • $UploadSensingData(\cdot)$: The inputs to this function are $msg$ and $RSIG$ , which are the sensing packets in (9) and the ring signature of the node, respectively. This function is invoked by legal sensors to upload their sensing result. The pseudocode of the CSC is summarized in Algorithm 2. _Spectrum Auction Contract (SAC)_ : The parameters in SAC are as follows. * • $CSC_{id}$: The identity of corresponding CSC, which is used to identify the connection between SAC and CSC since every SAC is associated with a particular CSC. * • $N_{2}$: The maximum number of the bidders, which is used to prevent too many nodes from sending bidding message. * • $T_{self-d}$: The time when this SAC will conduct the self-destruct operation to release memory. * • $d_{a}$: The amount of tokens that the bidders need to deposit before the final sensing result is released. The functions in SAC are introduced as follows. * • $BidderRegister(\cdot)$: Bidders invoke this function to register in SAC. * • $Commit(\cdot)$: The input to this function is $bldBid$ which denotes the commitment of $bidder_{i}$. Note that since the account balance is transparent, nodes can identify others’ bid by checking their balance[33]. Therefore, everyone makes several commits in order to prevent others from inferring your bid based on changes of their account balance. * • $Reveal(\cdot)$: The inputs to this function include $Dps$, $Bools$ and $RNDs$. $Dps$ is the list of deposit ($Dp$) the bidders make; $Bools$ is the list of boolean values which indicate whether these bids are valid or not; $RNDs$ are random values to make the hash of $\\{Dp$, $Bool$, $RND\\}$ hard to guess. A $Commit$ is the hash of these three parameters, i.e., $Commit$= $\mathcal{H}$$\\{$$Dp$, $Bool$, $RND\\}$. $bidsList[msg.sender]$ is the list of the invoker’s commits. By comparing the reveal messages and the commits, this function can identify which bids are revealed correctly. Valid bids are added together as the bidder’s total bid, invalid but correctly revealed bids are allowed to be withdrawn, bids that are not correctly revealed will not be returned. * • $Win(\cdot)$: The input to this function is $bidderList$, which is the list of all valid bids. Here we adopt the second-price sealed-bid auction for better economic profit [34]. Accordingly, this function selects the bidder with the highest bid as the winner, who only needs to pay the second highest bid. * • $EndOfAuction(\cdot)$: SAC will execute self-destruct operation at the preset time $T_{self-d}$. The pseudocode of SAC is shown in Algorithm 3. Algorithm 2 Cooperative Sensing Contract 1:function init($TV_{thr}$, $d_{s}$, $N_{1}$) 2: require(msg.sender==ContractOwner); 3: initialize $TV_{thr}$, $d_{s}$, $N_{1}$; 4:end function 5: 6:function SensorRegister($Addr,dpt,TV$) 7: $sensor_{num}\leftarrow 0$; 8: if checkRegisterQuality($TV_{i}$, $addr_{i}$) then 9: $sensorMap\leftarrow sensor_{i}$; 10: $sensor_{num}\leftarrow sensor_{num}+1$; 11: EMIT event("Registration success!"); 12: else 13: EMIT event("Registration failed, please checkout the trust value and deposit."); 14: end if 15: if $sensor_{num}>N_{1}$ then 16: Select top $N_{1}$ sensors according to trust value; 17: end if 18:end function 19: 20:function Fusion($msgList$) 21: Decision fusion in majority rule; 22: return Cooperative sensing result; 23:end function 24: 25:function checkRegisterQuality($TV_{i}$,$deposit_{i}$) 26: if $deposit_{i}>d_{s}$ then 27: $TV^{{}^{\prime}}_{i}$ $\leftarrow$ $convert(deposit_{i}-d_{s})$ 28: if ($TV_{i}+TV^{{}^{\prime}}_{i})>TV_{thr}$ then 29: if $totalNum<N_{1}$ or $TV_{i}>lowestTV$ then 30: Register Successful; 31: end if 32: else 33: Register Failed; 34: end if 35: end if 36:end function 37: 38:function UploadSensingData($msg$, $RSIG$) 39: if $RSIG$ is legal then 40: $msgList\leftarrow msg$; 41: EMIT event("Upload successfully!"); 42: else 43: EMIT event("Upload failed, illegal sensor!"); 44: end if 45:end function Algorithm 3 Sealed Spectrum Auction Contract 1:$Bid\leftarrow{bldBid_{i},msg.value}$; 2:$bidsList\leftarrow mapping(address=>Bid[])$; 3:$refund\leftarrow 0$ 4: 5:function Commit($bldBid$) 6: $bidsList[msg.sender].push(bldBid);$ 7:end function 8: 9:function Reveal($Dps,Bools,RNDs$) 10: $len\leftarrow bidsList[msg.sender].length$; 11: while $len>0$ do 12: $Commit\leftarrow\mathcal{H}\\{Dps[len],Bools[len],RNDs[len]\\}$; 13: if $Commit!=bidsList[msg.sender][len]$ then 14: EMIT event("Illegal Reveal!"); 15: Continue; 16: else if $Commit==bidsList[msg.sender][len]$ and $Bools[len]==true$ then 17: $refund\leftarrow refund+Dps[len]$; 18: else if $Commit==bidsList[msg.sender][len]$ and $Bools[len]==false$ then 19: withdraw invalid but reveal correctly bid; 20: end if 21: $len\leftarrow len-1$; 22: end while 23:end function 24: 25:function Win($bidderList$) 26: return Bidder with the highest bid; 27:end function 28: 29:function EndofAuction($T_{self-d}$) 30: Execute self-destruct operation; 31:end function ### VI-B The DSA Protocol In this part, we illustrate our proposed blockchain-based DSA protocol. To make it clearer, we elaborate this protocol in Fig. 5. * • Phase 1 (_Spectrum Sensing Request_): To ask for the channel state, SUs need to send a request message to Task Issuer (TI) through the control channel. TI is played by the node whose trust value is currently the highest. The role of TI is set to prevent multiple contracts from appearing in the network, which results in users participate in different contracts. TI will then creates and deploys the CSC and the corresponding SAC in the blockchain. * • Phase 2 (_Smart Contracts and Nodes Register_): The CSC and SAC are instantiated when TI issues the transaction $CSC=\\{CSC_{id}\mid T_{ddl}\mid N_{1}\mid TV_{thr}\mid Fusion(\cdot)\mid d_{s}\\}$ (12) with the signature $Sig_{TI}(STC)$, and the transaction $SAC=\\{CSC_{id}\mid SAC_{id}\mid N_{2}\mid T_{self-d}\mid Win(\cdot)\mid d_{a}\\}$ (13) with the signature $Sig_{TI}$=$(AC)$. To register to CSC, a sensor issues the transaction $Deposit_{CSC_{id}}=\\{pk_{i}\mid TV_{i}\mid CSC_{id}\mid d_{s}\\}$ (14) with the signature $Sig_{sk_{i}}(Deposit_{CSC_{id}})$ to make deposits. Similarly, a SU who intends to participate in SAC makes its deposit by issuing $Bty_{SAC_{id}}=\\{pk_{i}\mid TV_{i}\mid SAC_{id}\mid d_{a}\mid\mathcal{H}(bid_{i},RND_{i})\\}$ (15) with the signature $Sig_{sk_{i}}(Bty_{SAC_{id}})$. The $\mathcal{H}(bid_{i},RND_{i})$ denotes the bidding commitment this SU makes for the sealed auction. * • Phase 3 (_Players selection_): Players (including sensors and bidder) are selected based on $SensorRegister(\cdot)$ and $BidderRegister(\cdot)$, respectively. Finally, an event will be triggered to provide an appropriate notification to every eventually selected nodes. * • Phase 4 (_Sensing phase_): Sensors upload the sensing data by issuing the transaction $msg_{tx}=\\{CSC_{id}\mid\mathcal{H}(msgID)\mid TS\mid SR\mid loc\\}$ (16) with the ring signature generated by (11) before $T_{ddl}$, and $TS$ denotes timestamp, $SR$ denotes the sensing result. In this process, we use the ring signature to protect the SU’s identity. Meanwhile, these sensors need to make commitments mentioned in Section V-D by sending the transaction $Commit_{i}=\\{CSC_{id}\mid\mathcal{H}(SR,RND_{i},msgID)\mid pk_{i})\ $ (17) with $Sig_{sk_{i}}(Commit_{i})$. After the sensing deadline, data fusion is done, and the final sensing result will be published in the blockchain network. * • Phase 5 (_Auction if necessary_): If the cooperative sensing result indicates that the corresponding spectrum bands are idle, the SAC will be invoked. Users who have been selected will reveal their bids for the spectrum by sending bids and random numbers. Then the winner will be published on the blockchain. * • Phase 6 (_Trust value updating_): In this phase, the trust value of each node is updated using equation (4) by miners in the network. Figure 5: The workflow of our proposed blockchain-based DSA. ## VII Implementation and Performance Analysis ### VII-A Implementation of Smart Contract We implement the proposed smart contracts using Solidity[16] in the Ethereum Virtual Machine (EVM). The smart contracts are compiled and deployed in the Remix IDE[35] which is used for testing. After that, the nodes in the blockchain network can send transactions to deploy the smart contracts and invoke functions in these contracts. In the cooperative sensing phase, five sensors are considered and the information of their accounts in the blockchain network are listed in table II. In the following, we describe the implementation procedures in detail. 1. 1. _Smart Contract Preparation_ : The smart contract is created and deployed by the node with the highest trust value, i.e., the fifth account in table II. The relevant parameters and functions which need to be predefined are set as follows: $N_{1}$ is set as 3, the trust value threshold $TV_{thr}$ is set as 900 where we magnify the trust values by 100 times and make they as integers, since fixed point numbers are not fully supported by Solidity yet. $d_{s}$ is set as 100 wei, 2. 2. _Smart Contract Deployment_ : Fig.6 shows the details about the deployed smart contract. The field “from" is the account address of TI. The field “decoded input" shows all of our preset parameters. 3. 3. _Sensor Registration_ : All the five sensors send their registration request to the smart contract. Then, three sensors are selected by the function $SensorRegister(\cdot)$. 4. 4. _Sensing Results Fusion_ : The function $Fusion(\cdot)$, where the majority rule is selected, is used to fuse the sensing results from three legal sensors. If the spectrum is detected as idle, the auction phase initiates. In this phase, we consider two bidders: each bidder has one true bid and one false bid. The related information is listed in Table III. The first bidder has two bids: one is a true bid with 100 wei, the other one is a false bid with 200 wei. The second bidder also has two bids: one is true with 150 wei, the other is false with 300 wei. TABLE II: Related information of sensor accounts Account addresses | Trust value | Sensing Result ---|---|--- 0x5B38Da6a701c568545dCfcB03FcB875f56beddC4 | 0.91 | 0 0xAb8483F64d9C6d1EcF9b849Ae677dD3315835cb2 | 0.92 | 1 0x4B20993Bc481177ec7E8f571ceCaE8A9e22C02db | 0.87 | 1 0x78731D3Ca6b7E34aC0F824c42a7cC18A495cabaB | 0.93 | 0 0x5B38Da6a701c568545dCfcB03FcB875f56beddC4 | 0.94 | 1 Figure 6: Initialization of Smart Contract 1. 1. _Commit_ : Bidders invoke $Commit(\cdot)$ function to make commits. Fig. 7 shows the log of commit made by bidder 2. There are address information and bid information about this commit, thus others can know that bidder 2 make a bid with 150 wei, but they can not identify whether this is a real commit or not. 2. 2. _Reveal_ : After the commit phase, bidders invoke the function $Reveal(\cdot)$ to reveal its commits. The field “decoded input” in Fig. 8 shows the whole information about this commit. The smart contract will automatically verify the information uploaded in the two phase and the final bidding result can be accessed by all nodes in the system. TABLE III: Related Information of Bidders Bidder address | Bid | isReal | Commit ---|---|---|--- 0xAb84…5835cb2 | 100wei | yes | 0x591a291ad67…1b62c96a2092 200wei | no | 0x4871a30b671…4d08cfaa8bc1 0x4B20…2C02db | 150wei | yes | 0x4e6a24e35c7be…23643d68efa 300wei | no | 0x0c98edecac…74d24d971ec88 Figure 7: The commit made by bidder 2 Figure 8: The reveal made by bidder 2. ### VII-B Performance Analysis of The Proposed PoT Consensus Algorithm In this part, we evaluate the performance of the proposed PoT consensus algorithm. We first discuss how to set the initial mining difficulty at the beginning when every node’s trust value is 0. We simulate the mining process in our personal computer as a reference to explain how to select a suitable mining difficulty. We use a given string to represent the transactions in a block in the blockchain network. The result is shown in Fig. 9. The running time increases exponentially when the number of leading zeros is larger than 20. Supposing that the hash rate of the IoT devices is similar to that of our personal computer, the interval between two blocks $T_{0}$ is about 1 second, then the mining difficulty should be less than 18 leading zeros, the $\beta_{0}$ is about $2^{18}=262144$. It’s worth mentioning that even under the same mining difficulty and experimental environment, the mining time may be different. For example, given the mining difficulty where 32 leading zeros are needed in the target hash, it sometimes take tens of seconds, and sometimes can take more than an hour. In order to eliminate the impact of luck, we convert the mining difficulty into the expected mining cost and evaluate the expected mining cost instead of running time. Theoretically, assuming that the output of each hash calculation is unpredictable, each time when we increase the leading zero by one, the success probability of each hash trial will be halved. The trust value is associated with the mining difficulty. Therefore, we first convert the trust value into the number of leading zeros, and then the number of leading zeros is converted into the expected mining cost proportionally. Simulation is conducted for 1000 time slots, and the average expected mining cost of every type of nodes is calculated in Fig.11. To be specific, we simulate a network with 20 nodes and 4 types of nodes are considered. Rnode means a node with high performance and always act honestly; UAnode means a node with high performance and act honestly, however it participates in sensing infrequently. OOnode is a node who conducts malicious behavior periodically. We assume that OOnode will act maliciously after every two normal sensing rounds. The Lnode is a node who randomly provides binary sensing result without sensing. There are 12 Rnode with probability of detection $p_{d}$ = 0.90 and probability of false alarm $p_{f}$ = 0.15 [36]; 3 OOnode whose $p_{d}$ = 0.90 and $p_{f}$ = 0.15; 3 Lnode with $p_{d}$ = 0.5 and $p_{f}$ = 0.5; and 2 UAnode with $p_{d}$ = 0.90 and $p_{f}$ = 0.15. Besides, it is assumed that they only have a 50% chance to sign up in CSS each sensing round. In Fig.11, it can be observed that the reliable node performs best with respect to the average expected mining cost per timeslot, which is about one third of the other three types of nodes. The expected mining cost of a on-off node is a little smaller than Lnode and UAnode but still cost three times than Rnode. This proves the effectiveness of our proposed consensus algorithm. Figure 9: Running time (second) with increasing number of leading zeros Figure 10: Curve of trust value Figure 11: Expected mining consumption of different nodes ### VII-C Performance of Cooperative Sensing In this part, we evaluate the performance of cooperative spectrum sensing with the help of our proposed trust evaluation mechanism. In traditional blockchain network, there is no trust among nodes hence anyone can record the sensing information into blockchain, including nodes with poor performance or malicious behavior. However, due to the introduction of the trust value mechanism, the cooperative sensing contract can exclude bad nodes according to nodes’ trust value, thus the system’s sensing performance can be improved. For comparison, we consider 3 kinds of selection schemes to select the candidate nodes for the cooperative sensing. (i) random selection scheme; (ii) select according to register time; (iii) select according to sensor’s trust value. We simulate a 20 nodes network with the same setup in VII-B. It can be seen from Fig.12 that when the number of needed sensors $N_{1}$ is small, $p_{d}$ and $p_{f}$ of cooperative sensing in the first two schemes which do not take advantage of trust value perform worse than that of the last selection scheme. This indicates that our proposed trust value mechanism can effectively improve the cooperative sensing performance of the system. Moreover, in the last selection scheme, $p_{d}$ is close to $1$ when $N_{1}$ is about one fourth of all network nodes. The fewer nodes involved, the smaller the total monetary reward that the network needs to pay to sensors, and the smaller the economic burden for spectrum buyers. As $N_{1}$ grows, the performance of different schemes will gradually close, since most of the network nodes (including bad nodes) will participate in cooperative sensing. (a) $P_{d}$ of cooperative sensing (b) $P_{f}$ of cooperative sensing Figure 12: Performance of cooperative sensing under three selection schemes ### VII-D Incentive Mechanism Analysis Since both sensing and mining will consume SU’s computing power, an incentive mechanism is needed to reward nodes for the work they have done. The users who upload a sensing result that is consistent to the final cooperative sensing result will be rewarded with $R_{s}$ tokens; The users who successfully mine will be rewarded with $R_{m}$ tokens. The tokens can be used in auction to bid for spectrum resources. In implementation, the tokens rewarded for accurate spectrum sensing and successful mining, i.e., $R_{s}$ and $R_{m}$, need to be designed based on the evaluation of the computation consumption of spectrum sensing and mining. This incentive mechanism not only encourages the nodes to participate in spectrum sensing, but also encourage them to behave honestly and accurately in spectrum sensing. This is because, on the one hand, an honest and accurate sensing node is more likely to derive a sensing result that is consistent to the final cooperative sensing result and obtain tokens rewarded for spectrum sensing. On the other hand, with the proposed consensus algorithm, the node with a higher trust value is easier to mine successfully so that they have a higher probability of obtaining tokens for mining. On the contrary, the dishonest nodes will be discouraged since they are less likely to obtain rewards since they need to spend more computing resources on mining. ### VII-E Security Analysis * • _Distributed Denial of Service (DDoS) Attack_ : The DDoS attack here means that malicious users try to make the sensing service unavailable to other users. Our system is resist to this attack since we adopt the deposit mechanism. Thus, under our scheme, the cost of launching large-scale DDoS attack is very high since the attackers need to obtain lots of tokens in order to launch the attack. * • _Spoofing Attack_ : Spoofing attack means someone tries to masquerade others to create forged transactions. Secure Elliptic Curve Digital Signature Algorithm (ECDSA)[37] used in our blockchain can prevent this attack on the premise that attack does not have the user’s private key. * • _Free-riding Attack (lazy node)_ : The Free-riding attack here means lazy users may directly copy others’ sensing results at the phase of uploading sensing data. Firstly, there is no motivation for sensors to submit the sensing result before sensing deadline. Thus, when the lazy nodes get the sensing result, it is difficult for them to repack the sensing result and submit before the deadline. Secondly, even if a few sensors submit sensing results in advance, the connection between a user’s identity and sensing data is cut by the ring signature, thus the lazy users cannot determine the owner of the sensing data, and thus the credibility of the data cannot be guaranteed. Thus, it is no better than submit a sensing result randomly. * • _On-off Attack_ : On-off attack means that a node performs malicious behaviors periodically. If a trust management mechanism satisfy the condition that the dropping rate of trust value is larger than its increasing rate, it is considered to be resist to the on-off attack [28]. However, the model in [28] may cause the trust value drop a lot even the misbehavior is unintentional, since the increasing rate is much smaller than dropping rate. In our proposed model, as show in Fig.13, the unintentional mistakes made by sensors can be compensated by making right sensing decision. When the parameters $\tau$ and $\eta$ satisfy $\rho\textgreater\frac{\eta}{1-e^{-\eta}}-\eta$, our proposed trust value management mechanism is considered to be resist to the on-off attack, the proof is given in Appendix A. (a) Existing model (b) Proposed model Figure 13: Trust value of different types of nodes under two model ## VIII Conclusion In this paper, we have proposed a blockchain-based dynamic spectrum sharing protocol. This protocol mainly consists of three parts: the first part is the trust value management mechanism, which is designed to evaluate the credibility of nodes; the second part is the PoT consensus algorithm, which make the mining difficulty for malicious nodes greatly increased. The combination of node’s trust value and their mining difficulty can motivate nodes to be more willing to behave honestly; the third part is the privacy protection mechanism, in which we combine the ring signature and the commit- reveal scheme to solve the problem of privacy issue in the process of cooperative spectrum sensing. Finally, we implemented the prototype of our proposed smart contracts and analyzed the performance of PoT consensus algorithm and the improvement in cooperative sensing. Security analysis of the system show that our framework can resist many kinds of attacks which are frequently encountered in trust-based blockchain systems. ## Appendix A Proof of the resistance of on-off attack ###### Proposition 1. The proposed trust model given in (1) is resistant to on–off attack when $\rho\textgreater\frac{\eta}{1-e^{-\eta}}-\eta$. ###### Proof. Let $f(N_{r},N_{w})=e^{-\rho\cdot N_{w}}\cdot\left(1-e^{-\eta\cdot N_{r}}\right)$ denotes the trust value update function in (1). Because when $N_{r}=0$, $f(N_{r},N_{w})$ can only increase, this situation is not discussed here. At any other points $(N_{r},N_{w})$ in this function, the decreasing rate should be larger than the increasing rate, that is: $\left\lVert\frac{\partial f}{\partial N_{w}}\right\rVert>\left\lVert\frac{\partial f}{\partial N_{r}}\right\rVert,$ (18) where $\left\lVert\frac{\partial f}{\partial N_{w}}\right\|=\rho\cdot e^{-\rho\cdot N_{w}}\cdot\left(1-e^{-\eta\cdot N_{r}}\right)$ (19) and $\left\|\frac{\partial f}{\partial N_{r}}\right\|=\eta\cdot e^{-\rho\cdot N_{w}}\cdot\left(e^{-\eta\cdot N_{r}}\right),$ (20) then we have $\left\|\frac{\partial f}{\partial N_{w}}\right\|>\left\|\frac{\partial f}{\partial N_{r}}\right\|\Leftrightarrow\frac{\rho}{\eta}>\frac{e^{-\eta\cdot N_{r}}}{1-e^{-\eta\cdot N_{r}}}$ (21) which can also be denoted as: $\left\|\frac{\partial f}{\partial N_{w}}\right\|>\left\|\frac{\partial f}{\partial N_{r}}\right\|\Leftrightarrow\frac{\rho}{\eta}>\frac{1}{1-e^{-\eta\cdot N_{r}}}-1$ (22) When $N_{r}\geq 1$, $\frac{1}{1-e^{-\eta\cdot N_{r}}}-1\leq\frac{1}{1-e^{-\eta}}-1,$ (23) Therefore, as long as it is satisfied $\frac{\rho}{\eta}>\frac{1}{1-e^{-\eta}}-1$, the above inequality holds. ∎ ## References * [1] G. Forecast, “Cisco visual networking index: global mobile data traffic forecast update, 2017–2022,” _Update_ , vol. 2017, p. 2022, 2019. * [2] P. Kolodzy and I. Avoidance, “Spectrum policy task force,” _Federal Commun. Comm., Washington, DC, Rep. ET Docket_ , vol. 40, no. 4, pp. 147–158, 2002\. * [3] I. F. Akyildiz, B. F. Lo, and R. Balakrishnan, “Cooperative spectrum sensing in cognitive radio networks: A survey,” _Phy. Commun._ , vol. 4, no. 1, pp. 40–62, 2011. * [4] J. Wang, M. Li, Y. He, H. Li, K. Xiao, and C. Wang, “A blockchain based privacy-preserving incentive mechanism in crowdsensing applications,” _IEEE Access_ , vol. 6, pp. 17 545–17 556, 2018. * [5] L. Xiao, Y. Ding, D. Jiang, J. Huang, D. Wang, J. Li, and H. Vincent Poor, “A reinforcement learning and blockchain-based trust mechanism for edge networks,” _IEEE Transactions on Communications_ , vol. 68, no. 9, pp. 5460–5470, 2020. * [6] G. Liu, H. Dong, Z. Yan, X. Zhou, and S. Shimizu, “B4SDC: A blockchain system for security data collection in MANETs,” _IEEE Transactions on Big Data_ , pp. 1–1, 2020. * [7] W. Feng and Z. Yan, “MCS-Chain: Decentralized and trustworthy mobile crowdsourcing based on blockchain,” _Future Generation Computer Systems_ , vol. 95, pp. 649–666, 2019. * [8] M. Li, J. Weng, A. Yang, W. Lu, Y. Zhang, L. Hou, J. Liu, Y. Xiang, and R. H. Deng, “CrowdBC: A blockchain-based decentralized framework for crowdsourcing,” _IEEE Transactions on Parallel and Distributed Systems_ , vol. 30, no. 6, pp. 1251–1266, 2019. * [9] M. Yang, T. Zhu, K. Liang, W. Zhou, and R. H. Deng, “A blockchain-based location privacy-preserving crowdsensing system,” _Future Generation Computer Systems_ , vol. 94, pp. 408 – 418, 2019. [Online]. Available: http://www.sciencedirect.com/science/article/pii/S0167739X18320909 * [10] M. B. H. Weiss, K. Werbach, D. C. Sicker, and C. E. C. Bastidas, “On the application of blockchains to spectrum management,” _IEEE Transactions on Cognitive Communications and Networking_ , vol. 5, no. 2, pp. 193–205, 2019. * [11] E. D. Pascale, J. McMenamy, I. Macaluso, and L. Doyle, “Smart contract slas for dense small-cell-as-a-service,” _CoRR_ , vol. abs/1703.04502, 2017. [Online]. Available: http://arxiv.org/abs/1703.04502 * [12] G. Qiao, S. Leng, H. Chai, A. Asadi, and Y. Zhang, “Blockchain empowered resource trading in mobile edge computing and networks,” in _ICC 2019 - 2019 IEEE International Conference on Communications (ICC)_ , 2019, pp. 1–6. * [13] J. Qiu, D. Grace, G. Ding, J. Yao, and Q. Wu, “Blockchain-based secure spectrum trading for unmanned-aerial-vehicle-assisted cellular networks: An operator’s perspective,” _IEEE Internet of Things Journal_ , vol. 7, no. 1, pp. 451–466, 2020. * [14] T. Ariyarathna, P. Harankahadeniya, S. Isthikar, N. Pathirana, H. M. N. D. Bandara, and A. Madanayake, “Dynamic spectrum access via smart contracts on blockchain,” in _2019 IEEE Wireless Communications and Networking Conference (WCNC)_ , 2019, pp. 1–6. * [15] S. Bayhan, A. Zubow, and A. Wolisz, “Spass: Spectrum sensing as a service via smart contracts,” in _2018 IEEE International Symposium on Dynamic Spectrum Access Networks (DySPAN)_ , 2018, pp. 1–10. * [16] Solidity webpage, https://docs.soliditylang.org/en/v0.7.5/. * [17] N. Szabo, “Formalizing and securing relationships on public networks,” _First Monday_ , vol. 2, no. 9, Sep. 1997. [Online]. Available: https://firstmonday.org/ojs/index.php/fm/article/view/548 * [18] R. L. Rivest, A. Shamir, and Y. Tauman, “How to leak a secret,” in _Advances in Cryptology — ASIACRYPT 2001_ , C. Boyd, Ed. Berlin, Heidelberg: Springer Berlin Heidelberg, 2001, pp. 552–565. * [19] K. Kotobi and S. G. Bilen, “Secure blockchains for dynamic spectrum access: A decentralized database in moving cognitive radio networks enhances security and user access,” _IEEE Vehicular Technology Magazine_ , vol. 13, no. 1, pp. 32–39, 2018. * [20] S. Zheng, T. Han, Y. Jiang, and X. Ge, “Smart contract-based spectrum sharing transactions for multi-operators wireless communication networks,” _IEEE Access_ , vol. 8, pp. 88 547–88 557, 2020. * [21] X. Fan and Y. Huo, “Blockchain based dynamic spectrum access of non-real-time data in cyber-physical-social systems,” _IEEE Access_ , vol. 8, pp. 64 486–64 498, 2020. * [22] H. Zhang, S. Leng, and H. Chai, “A blockchain enhanced dynamic spectrum sharing model based on proof-of-strategy,” in _ICC 2020 - 2020 IEEE International Conference on Communications (ICC)_ , 2020, pp. 1–6. * [23] X. Ling, J. Wang, T. Bouchoucha, B. C. Levy, and Z. Ding, “Blockchain radio access network (b-ran): Towards decentralized secure radio access paradigm,” _IEEE Access_ , vol. 7, pp. 9714–9723, 2019. * [24] M. A. A. Careem and A. Dutta, “Sensechain: Blockchain based reputation system for distributed spectrum enforcement,” in _2019 IEEE International Symposium on Dynamic Spectrum Access Networks (DySPAN)_ , 2019, pp. 1–10. * [25] A. Vosoughi, J. R. Cavallaro, and A. Marshall, “Trust-aware consensus-inspired distributed cooperative spectrum sensing for cognitive radio ad hoc networks,” _IEEE Transactions on Cognitive Communications and Networking_ , vol. 2, no. 1, pp. 24–37, 2016. * [26] Vitalik Buterin, Thoughts on UTXOs, 2016, http://timmurphy.org/2009/07/22/line-spacing-in-latex-documents/. * [27] G. Wood, “Ethereum: A secure decentralised generalised transaction ledger,” pp. 1–32, 2014. * [28] X. Kang and Y. Wu, “A trust-based pollution attack prevention scheme in peer-to-peer streaming networks,” _Computer Networks_ , vol. 72, pp. 62 – 73, 2014. [Online]. Available: http://www.sciencedirect.com/science/article/pii/S1389128614002667 * [29] D. Boneh, B. Bünz, and B. Fisch, “Batching techniques for accumulators with applications to iops and stateless blockchains,” in _Annual International Cryptology Conference_. Springer, 2019, pp. 561–586. * [30] G. Ateniese, B. Magri, D. Venturi, and E. Andrade, “Redactable blockchain – or – rewriting history in bitcoin and friends,” in _2017 IEEE European Symposium on Security and Privacy (EuroS P)_ , 2017, pp. 111–126. * [31] S. Li, H. Zhu, Z. Gao, X. Guan, Kai Xing, and X. Shen, “Location privacy preservation in collaborative spectrum sensing,” in _2012 Proceedings IEEE INFOCOM_ , 2012, pp. 729–737. * [32] P. Rogaway and T. Shrimpton, “Cryptographic hash-function basics: Definitions, implications, and separations for preimage resistance, second-preimage resistance, and collision resistance,” in _Fast Software Encryption_ , B. Roy and W. Meier, Eds. Berlin, Heidelberg: Springer Berlin Heidelberg, 2004, pp. 371–388. * [33] Solidity webpage, https://docs.soliditylang.org/en/v0.7.5/. * [34] Y.-C. Liang, _Blockchain for Dynamic Spectrum Management_. Singapore: Springer Singapore, 2020, pp. 121–146. [Online]. Available: https://doi.org/10.1007/978-981-15-0776-2_5 * [35] Remix-Ethereum IDE, https://remix-ide.readthedocs.io/en/latest/. * [36] X. Kang, Y.-C. Liang, H. K. Garg, and L. Zhang, “Sensing-based spectrum sharing in cognitive radio networks,” _IEEE Transactions on Vehicular Technology_ , vol. 58, no. 8, pp. 4649–4654, 2009. * [37] D. Johnson, A. Menezes, and S. Vanstone, “The elliptic curve digital signature algorithm (ecdsa),” _Int. J. Inf. Sec._ , vol. 1, pp. 36–63, 08 2001.
.gifpng.pngconvert gif:#1 png: .gif # The DAQ system of the 12,000 Channel CMS High Granularity Calorimeter Prototype B. Acar G. Adamov C. Adloff S. Afanasiev N. Akchurin B. Akgün M. Alhusseini J. Alison G. Altopp M. Alyari S. An S. Anagul I. Andreev M. Andrews P. Aspell I. A. Atakisi O. Bach A. Baden G. Bakas A. Bakshi P. Bargassa D. Barney E. Becheva P. Behera A. Belloni T. Bergauer M. Besancon S. Bhattacharya S. Bhattacharya D. Bhowmik P. Bloch A. Bodek M. Bonanomi A. Bonnemaison S. Bonomally J. Borg F. Bouyjou D. Braga J. Brashear E. Brondolin P. Bryant J. Bueghly B. Bilki B. Burkle A. Butler- Nalin S. Callier D. Calvet X. Cao B. Caraway S. Caregari L. Ceard Y. C. Cekmecelioglu G. Cerminara N. Charitonidis R. Chatterjee Y. M. Chen Z. Chen K. y. Cheng S. Chernichenko H. Cheung C. H. Chien S. Choudhury D. Čoko G. Collura F. Couderc I. Dumanoglu D. Dannheim P. Dauncey A. David G. Davies E. Day P. DeBarbaro F. De Guio C. de La Taille M. De Silva P. Debbins E. Delagnes J. M. Deltoro G. Derylo P.G. Dias de Almeida D. Diaz P. Dinaucourt J. Dittmann M. Dragicevic S. Dugad V. Dutta S. Dutta J. Eckdahl T. K. Edberg M. El Berni S. C. Eno Yu. Ershov P. Everaerts S. Extier F. Fahim C. Fallon B. A. Fontana Santos Alves E. Frahm G. Franzoni J. Freeman T. French E. Gurpinar Guler Y. Guler M. Gagnan P. Gandhi S. Ganjour A. Garcia-Bellido Z. Gecse Y. Geerebaert H. Gerwig O. Gevin W. Gilbert A. Gilbert K. Gill C. Gingu S. Gninenko A. Golunov I. Golutvin T. Gonzalez N. Gorbounov L. Gouskos Y. Gu F. Guilloux E. Gülmez M. Hammer A. Harilal K. Hatakeyama A. Heering V. Hegde U. Heintz V. Hinger N. Hinton J. Hirschauer J. Hoff W. S. Hou C. Isik J. Incandela S. Jain H. R. Jheng U. Joshi O. Kara V. Kachanov A. Kalinin R. Kameshwar A. Kaminskiy A. Karneyeu O. Kaya M. Kaya A. Khukhunaishvili S. Kim K. Koetz T. Kolberg A. Kristić M. Krohn K. Krüger N. Kulagin S. Kulis S. Kunori C. M. Kuo V. Kuryatkov S. Kyre O. K. Köseyan Y. Lai K. Lamichhane G. Landsberg J. Langford M. Y. Lee A. Levin A. Li B. Li J.-H. Li H. Liao D. Lincoln L. Linssen R. Lipton Y. Liu A. Lobanov R. S. Lu I. Lysova A. M. Magnan F. Magniette A. A. Maier A. Malakhov I. Mandjavize M. Mannelli J. Mans A. Marchioro A. Martelli P. Masterson B. Meng T. Mengke A. Mestvirishvili I. Mirza S. Moccia I. Morrissey T. Mudholkar J. Musić I. Musienko S. Nabili A. Nagar A. Nikitenko D. Noonan M. Noy K. Nurdan C. Ochando B. Odegard N. Odell Y. Onel W. Ortez J. Ozegović L. Pacheco Rodriguez E. Paganis D. Pagenkopf V. Palladino S. Pandey F. Pantaleo C. Papageorgakis I. Papakrivopoulos J. Parshook N. Pastika M. Paulini P. Paulitsch T. Peltola R. Pereira Gomes H. Perkins P. Petiot F. Pitters F. Pitters H. Prosper M. Prvan I. Puljak T. Quast R. Quinn M. Quinnan K. Rapacz L. Raux G. Reichenbach M. Reinecke M. Revering A. Rodriguez T. Romanteau A. Rose M. Rovere A. Roy P. Rubinov R. Rusack A. E. Simsek U. Sozbilir O. M. Sahin A. Sanchez R. Saradhy T. Sarkar M. A. Sarkisla J. B. Sauvan I. Schmidt M. Schmitt E. Scott C. Seez F. Sefkow S. Sharma I. Shein A. Shenai R. Shukla E. Sicking P. Sieberer Y. Sirois V. Smirnov E. Spencer A. Steen J. Strait T. Strebler N. Strobbe J. W. Su E. Sukhov L. Sun M. Sun C. Syal B. Tali U. G. Tok A. Kayis Topaksu C. L. Tan I. Tastan T. Tatli R. Thaus S. Tekten D. Thienpont T. Pierre-Emile E. Tiras M. Titov D. Tlisov J. Troska Z. Tsamalaidze G. Tsipolitis A. Tsirou N. Tyurin S. Undleeb D. Urbanski V. Ustinov A. Uzunian M. van de Klundert J. Varela M. Velasco M. Vicente Barreto Pinto P. M. da Silva T. Virdee R. Vizinho de Oliveira J. Voelker E. Voirin Z. Wang X. Wang F. Wang M. Wayne S. N. Webb M. Weinberg A. Whitbeck D. White R. Wickwire J. S. Wilson H. Y. Wu L. Wu C. H Yeh R. Yohay G. B. Yu S. S. Yu D. Yu F. Yumiceva A. Zacharopoulou N. Zamiatin A. Zarubin S. Zenz H. Zhang J. Zhang ###### Abstract The CMS experiment at the CERN LHC will be upgraded to accommodate the 5-fold increase in the instantaneous luminosity expected at the High-Luminosity LHC (HL-LHC) [1]. Concomitant with this increase will be an increase in the number of interactions in each bunch crossing and a significant increase in the total ionising dose and fluence. One part of this upgrade is the replacement of the current endcap calorimeters with a high granularity sampling calorimeter equipped with silicon sensors, designed to manage the high collision rates [2]. As part of the development of this calorimeter, a series of beam tests have been conducted with different sampling configurations using prototype segmented silicon detectors. In the most recent of these tests, conducted in late 2018 at the CERN SPS, the performance of a prototype calorimeter equipped with ${\approx}12,000\rm{~{}channels}$ of silicon sensors was studied with beams of high-energy electrons, pions and muons. This paper describes the custom-built scalable data acquisition system that was built with readily available FPGA mezzanines and low-cost Raspberry PI computers. ## 1 Introduction The HL-LHC at CERN is planned to operate with an instantaneous luminosity of $5\times 10^{34}$ cm-2s-1 or higher, delivering up to ten times more integrated luminosity than is expected in the current LHC programme. This increase poses significant challenges in the design and operation of the detectors at the HL-LHC. In particular, in the forward direction the absorbed dose will be up as much as 2 MGy and the fluence in the innermost region is expected to reach $10^{16}$ neq/cm2, which is an unprecedented level in high energy collider experiments. Additionally there will be ${\approx}140$ proton- proton interactions occurring (pile up) in every bunch crossing, which happens at a rate of 40 MHz. This considerably complicates the reconstruction of events. To contend with these conditions, the CMS Collaboration is planning a series of upgrades to some of the existing detector components, and replacing others with new detectors designed specifically to mitigate the effect of the high pile up [1]. As part of this upgrade programme the current electromagnetic and hadronic calorimeters in the endcaps will be replaced with a new calorimeter, known as the ‘High-Granularity Calorimeter’ (HGCAL) [2]. This new sampling calorimeter (CE) will be sub-divided into two sections, the electromagnetic (CE-E) and the hadronic (CE-H), in an arrangement shown in Figure 1. The CE-E will be equipped with silicon sensors, while the CE-H will be equipped with both silicon sensors and scintillator tiles read out directly with SiPMs. In the CE-H the silicon sensors will be at small radii, close to the beam where the radiation levels are highest, and scintillator tiles at the larger radii. The absorber of CE-E will be a mixture of lead, copper and sintered copper-tungsten, while in CE-H the absorber plates will be stainless steel. The hexagonal silicon sensors will be subdivided into hexagonal cells with areas of ${\approx}1.1\,\rm{cm}^{2}$ or 0.5 cm2, with the sensors with smaller cells placed at small radii. The full calorimeter will be operated at -30∘C to reduce the dark current in the silicon sensors and in the SiPMs. There will be 28 sampling layers in CE-E and 22 in CE-H. This high-transverse granularity, combined with the high longitudinal segmentation of the calorimeter has been selected, within the constraints of cost and available space, to optimise the identification and measurement of hadronic and electromagnetic showers in the presence of the high pile up. Figure 1: Schematic view of the CMS high granularity Endcap Calorimeter. The basic detector unit is a silicon module. A module consists of a silicon sensor, glued to a baseplate on one side and to a printed circuit board (PCB) for the readout on the opposite side. The individual cells of the sensor are connected electrically to a readout ASIC on the PCB with wire-bonds that pass through holes in the PCB. The ASIC amplifies and digitises the analogue signals and transmits them to the off-detector electronics on receipt of an external trigger signal. By 2018 more than 100 prototype silicon modules have been produced using 6-inch hexagonal silicon sensors subdivided into cells with areas of ${\approx}$1.1 cm2. A module is shown in Fig. 2 (left). Further details of the construction and assembly of the silicon modules used in these tests may be found in [3]. The modules were assembled into a prototype of HGCAL that was tested with beams of electrons, pions and muons at the CERN SPS. The ASIC used to read out of the signals from the silicon cells was the Skiroc2-CMS [4], a custom ASIC designed by the OMEGA group at Ecole Polytechnique. The function of this 64-channel ASIC was to measure both the amplitude of the signal and its time of arrival, so not only the energy response can be measured, but the timing performance can be characterised. This Skiroc2-CMS ASIC has many of the features of the HGCAL front-end readout ASIC in development. In the latest beam test in late 2018 the HGCAL prototype was equipped with 94-hexagonal silicon modules arranged into a 26 radiation length electromagnetic section and 5 nuclear interaction length hadronic section. Behind the prototype calorimeter we placed the Analog Hadronic Calorimeter (AHCAL) prototype, developed by the CALICE Collaboration [5]. This calorimeter is a scintillator-based sampling calorimeter, similar in design to the proposed design of the HGCAL [2], but with much finer longitudinal segmentation. In the final test at the CERN SPS, data were taken with beams of muons, charged hadrons and electrons with energies ranging from 20 to 300 GeV at the H2 beam line of the CERN-SPS over a period of two weeks in October 2018. The data acquisition (DAQ) system for the beam tests needed to be flexible and scalable to control and read out the increasing number of prototype silicon modules as they became available. It was designed with readily available FPGA mezzanines and low-cost Raspberry PIs, and scaled up to work with ${\approx}$12,000 channels for the final test. This paper describes the DAQ system and is structured as follows: the overall architecture of the system is described in section 2; the data format and the back-end DAQ components are described in section 3; the DAQ software is explained in section 4; in section 5 the detector systems used for system synchronisation is discussed and the operational experience is discussed in section 6. ## 2 DAQ System Architecture Each hexagonal cell of the silicon sensor was connected to the 64-channel Skiroc2-CMS ASIC. Each channel of this ASIC had a low noise pre-amplifier followed by high- and low-gain shapers, with a shaping time of 40 ns, and time-over-threshold (ToT) and time-of-arrival (ToA) circuits. Both of the shapers had analogue-to-digital converters (ADCs) that sampled the signal every 25 ns. It also had a circuit to measure the ToA of large amplitude signals (${}>3\rm{fC}$) with a precision of 50 ps. Further details of the design can be found in [4]. To simplify routing of the signals in a very dense board, four Skiroc2-CMS ASICs were used to readout the 128 channels of each silicon sensor, leaving half of the channels unused. The PCB also had a MAX®10 field programmable gate array (FPGA) to control the readout of the module. It received the clock, trigger and busy signals from the off-detector electronics, aggregated the data from the Skiroc2-CMS ASICs and transmitted it to the off-detector electronics. Figure 2 (left) is a photograph of a prototype module, with the Skiroc2-CMS ASICs marked with rectangles, and the MAX®10 FPGA, on the top left, is indicated with a white circle. Figure 2: (left) A prototype module used in beam tests. The Skiroc2-CMS ASICs are marked with white rectangles. The MAX®10 FPGA is marked with a white circle. The red grounding wire (on the left side) and, red and violet bias wires (on the right side) are soldered on the PCB. The micro HDMI (uHDMI) cable is connected on the top. (right) An interposer board used in beam tests. The HDMI cable and bias voltage wires, connecting the interposer board to the prototype module, are connected on the top. The HDMI and RG174 cables, connecting the interposer board to the DAQ board, are connected on the bottom. Each of the prototype silicon modules was connected to the off-detector DAQ boards through an interposer board. These boards regulated the 5 V output from the DAQ boards to the 3.3 V needed by the prototype modules via HDMI cables. They also filtered and transmitted the bias voltage coming from the DAQ boards through RG174 cables to wires soldered to the prototype modules. A photograph of an interposer board is shown in Figure 2 (right). The off-detector electronics consisted of a set of custom 9U readout boards, each of which could receive data from up to eight silicon modules. All the the readout boards were controlled by a single custom 9U synchronisation board, the ‘sync board’. The sync board distributed the clock, trigger and busy signals to all the readout boards. The readout boards communicated with the data acquisition computer through Ethernet, with a 100 Mbit/s output of the readout boards connected to a Gigabit Ethernet switch, from which data were sent to the DAQ computer for processing. In the tests with 94 silicon modules, one sync board and fourteen readout boards were mounted in two custom air- cooled crates. The crate that was equipped with one sync board and seven readout boards is shown in Figure 3. Figure 3: A crate, used in HGCAL beam tests, populated with one sync board and seven readout boards. The sync board (in the leftmost slot) distributed the clock, trigger and busy signals to the readout boards via HDMI cables, not shown. The readout boards were used to send control data to the silicon modules, to receive data from them, and to transmit the data to the acquisition computer through an Ethernet switch (on the right side of the readout boards). The readout and sync boards were powered by the power modules located under the DAQ crate. Figure 4 shows a schematic view of the inter-connectivity of the DAQ system. The readout boards were connected to the prototype silicon modules by HDMI cables both between the readout boards and the interposer boards and between the interposer boards and the modules. The trigger signal was formed from a coincidence of signals from two scintillation counters located upstream of the calorimeter. The 40 MHz system clock, generated on the sync board and the trigger signals were transmitted to the readout boards with HDMI cables. Since the beam was not synchronised with the clock, the trigger and the clock were asynchronous. The bias voltage was distributed to the silicon sensors through the readout boards with separate RG174 cables. Figure 4: Schematic view of the DAQ system used in HGCAL beam tests. ## 3 Data format and back-end DAQ components ### 3.1 Skiroc2-CMS ASIC and prototype module data format In operation, the analog signals were stored every 25 ns in a switched- capacitor array (SCA) with a depth of 13 cells. When a trigger was received, the updating of the SCA was halted and the two values for the ToA were stored, one referenced to the next falling edge of the 40 MHz clock, and the other to the next rising clock edge. Two values of the ToT were also kept, one with a fast ramp time-to-digital converter and another with a slow ramp. [bitwidth=1.6em]16 []5 & 0,1,2,3,4,15 1924 $\times$ 16-bit integers []5$\times$64 channels 111 110 110 11$H_{A}$ 1212Low gain ADC (SCA0) []5$\times$64 channels 111 110 110 11$H_{A}$ 1212High gain ADC (SCA0) []5 1616…$\times$13 SCA cells []5$\times$64 channels 111 110 110 11$H_{A}$ 1212Low gain ADC (SCA12) []5$\times$64 channels 111 110 110 11$H_{A}$ 1212High gain ADC (SCA12) []5$\times$64 channels 111 110 110 11$H_{A}$ 1212ToA (stop falling clk) []5$\times$64 channels 111 110 110 11$H_{A}$ 1212ToA (stop rising clk) []5$\times$64 channels 111 110 110 11$H_{T}$ 1212ToT (fast ramp) []5$\times$64 channels 111 110 110 11$H_{T}$ 1212ToT (slow ramp) []5 110 110 110 1313Roll position (13-bit) []5 110 110 1414Global timestamp MSB (14-bit) []5 110 110 110 1212Global timestamp LSB (12-bit) 110 []5 111 111 110 110 110 110 110 110 88Chip ID (8-bit) Figure 5: Data format of the Skiroc2-CMS ASIC. $H_{A}$ is the hit bit for ToA and is set to ’1’ when ToA is fired. Similarly, $H_{T}$ is the hit bit for ToT and is set to ’1’ when ToT is fired. The 13-bit roll position is used to reorder the SCA cells in time. Figure 5 shows the data format of the Skiroc2-CMS ASIC. When a trigger was received, by the MAX®10 FPGA of the hexaboard the four Skiroc2-CMS ASICs converted the data in analogue memory to the digital data format. These data were then read out by the MAX®10 FPGA and packaged as shown in Figure 6. The bits $b_{Ai}$ belong to the ASIC "i" which followed the Skiroc2-CMS data format shown in Fig. 5. For every event 30784 bytes of data were transmitted from each hexaboard. These data were then gathered, via the HDMI-uHDMI cables, by the back-end readout boards. [bitwidth=3.0em]8 0,3,4,7 44HEADER & 441 bit per ASIC 1924 $\times$ 16 bytes 11 0 0 0 $b_{A0}$ $b_{A1}$ $b_{A2}$ $b_{A3}$ 88… 11 0 0 0 $b_{A0}$ $b_{A1}$ $b_{A2}$ $b_{A3}$ Figure 6: Hexaboard data format. The bits $b_{Ai}$ belong to the ASIC "i" which followed the Skiroc2-CMS data format described by Figure 5. ### 3.2 Back-end DAQ electronics The DAQ system was designed to be easily scalable to provide a readout for different numbers of silicon modules. To minimise costs, readily available commercial components were used. Additionally, optical receiver modules (oRMs) [6], recovered from the CMS level-1 trigger system, when it was upgraded with faster electronics, were used. Each oRM was equipped with a Kintex-7 FPGA, 4.8 Mbits of block RAM, two 6.6 Gbit/s bi-directional serial ports, and a 128 Mbit FLASH memory for configuration. The connection from the Kintex-7 FPGA to the gigabit Ethernet switch was made with SFP to RJ-45 adapters. #### 3.2.1 Readout board In the final beam test in October 2018, there were 94 silicon modules that were read out with 14 readout boards mounted in two racks, controlled by a sync board. The readout boards performed the following tasks: * • Loading firmware on the Max®10 FPGAs and module initialisation and reset. * • Generating and distributing control signals for the prototype silicon modules. * • Accumulating the data received from the prototype silicon modules. * • Distributing the clock, busy and trigger signals. * • Distributing the low voltage power and the bias voltage for the prototype silicon modules. The readout board, shown in Figure 7, was a custom PCB equipped with five oRMs and a Raspberry Pi. Each board had eight HDMI ports on the front panel connected to the silicon modules through the interposers and one HDMI port on the back panel for connection to the sync board. On the front of each board there were eight RG174 connectors that were used to distribute, after filtering, the bias voltage through the interposers to each detector module. Figure 7: The readout board used to send control data to the silicon modules, to receive data from them and to transmit it to the data acquisition computer. It readout data from up to eight silicon modules via HDMI connectors. It also supplied the bias voltage for up to eight prototype silicon modules via standard RG174 connectors. It was equipped with one control and four data oRMs and one Raspberry Pi. The overall readout cycle was controlled via helper processes running on the Raspberry Pi, which communicated with each oRM through the SPI bus. Each Pi was connected to the central DAQ server through its Ethernet port, from which it also received ‘Start’, ‘Stop’, and other commands. A single readout board was equipped with five oRMs: one control (CTL) oRM, and four DATA oRMs. The DATA oRMs were responsible for reading the data from up to two silicon modules, while the CTL oRM received data from the DATA oRMs and transferred it to the central server. It also managed the communication with the sync board. The firmware installed on the CTL oRM’s FPGA included the IPBus firmware [7, 8] for this purpose. The IPBus IP and MAC addresses of the oRM were set by the Raspberry Pi, as well as other parameters used by the CTL oRM as it combined the four streams of data. Before a run started, the helper processes on the Raspberry Pis first configured the ASICs on the prototype silicon modules, and data collection was initiated. The Trigger signal was broadcast from the sync board to the readout boards, from where it was forwarded on to the modules. On the readout boards, the helper processes running on the Raspberry Pis after receiving the trigger signal, prompted the ASICs to initiate data transmission. The data from the ASICs were then sent unprocessed, via the MAX®10 FPGA, to the DATA oRMs, where it was then merged into a single data stream by the CTL oRM. The 4-bit headers of Figure 6 were then dropped and 32-bit integers were built with the data from up to eight modules, corresponding to 32 ASICs. These 32-bit integers were written to a FIFO to be readout by the central server using the IPBus protocol over a gigabit Ethernet link. When ready, a flag was set inside the CTL’s RAM to indicate that the data were ready for transfer. In Fig. 8 the output data format of the CTL oRM FIFO is shown. Once the data had been fully read out by the server, the helper processes on the Pis reset the ASICs, and sent a start acquisition signal. The CTL oRM then sent a ‘ReadoutDone’ signal to the sync board, indicating the boards had finished their cycles and were ready to receive the next trigger. The firmware block diagram of the readout board is shown in Figure 9. [bitwidth=2.0em,bitformatting=16 0-15 30784 $\times$ 32-bit integers 1 $b_{0,0}$ $b_{0,1}$ $b_{0,2}$ $b_{0,3}$ 88… 1$b_{7,0}$ $b_{7,1}$ $b_{7,2}$ $b_{7,3}$ 1616… 1 $b_{0,0}$ $b_{0,1}$ $b_{0,2}$ $b_{0,3}$ 88… 1$b_{7,0}$ $b_{7,1}$ $b_{7,2}$ $b_{7,3}$ Figure 8: Data format of the CTL oRM FIFO readout by the central server using the IPBus protocol. The bits $b_{i,j}$ correspond to the data of ASIC "j" of module "i". Figure 9: The firmware block diagram of the readout board. #### 3.2.2 Sync board The function of the sync board, shown in Figure 10, was the distribution of the common signals to the readout boards and to synchronise the flow of data from the readout boards. The sync board generated the 40 MHz system clock on a small mezzanine card mounted at the rear of the board. On the same mezzanine there were four 50 $\Omega$ RG174 connectors. Two were for the Trigger and a Veto signal inputs and two for the Clock and a copy of the Trigger signal outputs. The Veto signal was not used in these tests. Processing on the sync boards was handled by a Raspberry Pi as well as a central (SYNC) oRM. One sync board could control up to 15 readout boards through 15 HDMI ports mounted on the front. An extra HDMI port was mounted on the front to allow for the possibility of daisy-chaining two or more sync boards together when more than 15 readout boards are to be readout. Figure 10: The sync board was used to control up to 15 readout boards. It received the Veto and the Trigger signals and distributed control signals to up to 15 readout boards. On top of the 15 HDMI ports for readout board control there was one extra port for connection to another sync board for daisy- chaining. It was equipped with a Raspberry Pi and an Kintex-7 FPGA for control and communication. It was also equipped with a mezzanine card for clock generation and receiving external signals. At the start of a readout cycle, the sync board waited for an asynchronous External Trigger signal. Then this signal was synchronized with the on-board 40 MHz clock and sent to the readout boards to be distributed to the silicon modules. The sync board then waited for a ‘ReadoutDone’ signal from each readout board. Once this signal was received from all the readout boards, the sync board made itself ready to process the next available trigger. The firmware block diagram of the sync board is shown in Figure 11. Figure 11: The firmware block diagram of the sync board. ## 4 Data acquisition software The DAQ software selected for these tests was based on the EUDAQ [9] framework. This framework, written in C++, was developed specifically for small-to-medium scale systems with significantly less overhead than frameworks used in large scale experiments, like XDAQ [10]. Additionally, in separate earlier tests of the AHCAL prototype, the EUDAQ system had already been used successfully. The EUDAQ framework was designed to be modular and portable. It was structured so that software for the readout of specific detector components was kept separate and distinct from the core processes. For this each detector component that produced data had a ‘Producer’ process running. The functions performed by the Producer was to initialise, configure, issue stops and starts to the component, and to collect the data and forward it to the core process. ### 4.1 CMS-CE EUDAQ Producer A Readout Producer was developed to read out the data from one or more readout boards in parallel. During the combined beam test of October 2018, the DAQ had seven of these Readout Producers, each connected to two readout boards. The $\mu$TCA Hardware Access Library ($\mu$HAL) was used to read the IPBus UDP transactions from the readout boards. The sequence of operations of the Readout Producer were as follows: 1. 1. Wait until each $\mu$HAL interface is notified that a trigger occurred (by checking an IPBus register of the CTL oRM boards). 2. 2. Read out the FIFO of the CTL oRM board from each readout board and fill raw data containers. The data format of this FIFO is described in Section 3.2.1. A time-stamp – the number of 40 MHz clock cycles in a 64-bit integers since the last configuration – is read out with the data. 3. 3. ‘ReadoutDone’ signal is sent from CTL oRM to Sync oRM. 4. 4. Create an event block containing the raw data from each readout board, the time-stamp of the readout board and the event ID. 5. 5. Forward the event block to the EUDAQ data collector. 6. 6. Increment the event ID. 7. 7. Return to step 1 and wait for the next trigger. Once the readout was complete, data from each of the event blocks from each of the Readout Producers were combined with data from the Producers connected to other detector components to form a complete event data block. #### 4.1.1 EUDAQ online data monitoring Part of the EUDAQ framework were tools to monitor the data collection. Online analyses were developed to monitor in real time the stability of the pedestal values, noise and occupancies of each of the silicon sensor channels. #### 4.1.2 Data unpacking and first analysis steps The first step of the data analysis consisted of unpacking the CMS CE EUDAQ events. For this purpose a C++ library embedded in the EUDAQ framework has been developed. An initial data quality check was performed before unpacking the raw data by comparing the difference in time-stamp with the time-stamp of the previous event as a check of the synchronisation for all the readout boards. During the October data taking only a few runs had events with a synchronization failure. After this test the data were unpacked and the data from each ASIC were sorted into tables of 16-bit integers with the structure shown in Figure 5 and stored in a ROOT [11] file. As zero-suppression was not used for simplicity, these tables contained the data from every channel of the ASICs, including those not connected to a detector channel. The data for each cell in an event contained data and pointer information from the 13 SCA cells for both the high- and the low-gain slow shapers, the ToA and the ToT measurements. The pointer was the address of first SCA cell data for the event, which allowed the ordering of the SCA data in time. This was required for the data reconstruction since the trigger was asynchronous with the 40 MHz clock. The data analysis workflow, which was developed in the CMSSW framework [12], and transformed the tables of 16-bit integers into a collection of calibarated hits for data analysis used the ROOT files as input. ## 5 System synchronization ### 5.1 Beam-characterization detectors The tests of HGCAL and AHCAL prototypes with particle beams in the H2 area at CERN had been complemented by the readout of various beam-characterization detectors. Four delay wire chambers (DWC) [13] measured the trajectory and impact of the particles in the beam, two scintillator detectors served as external trigger source and two micro-channel plates (MCP) had been used to provide fast signals for reference timing measurement of the incident particles [14]. For this purpose, two 16-channel CAEN v1290N TDCs and one v1742 digitiser were integrated into the HGCAL prototype DAQ. From each DWC four signals were separately discriminated at a threshold of -30 mV and fed as inputs to the TDC. Since the binning of the time-stamp digitization should be less than 1 ns corresponding to an optimal resolution of the position measurement of 200$~{}\mu$m [13], a binning of 25 ps had been chosen. For proper event synchronization, the trigger for all CAEN modules stemmed from the duplicated TTL trigger signal issued by the synchronization board. After conversion to NIM, it was copied three times and fed into each module individually. After receiving a trigger, events were built, were labeled with trigger time-stamps and subsequently stored in a local buffer. Two dedicated EUDAQ producers [9] had been developed. They ran on a separate computer and communicated to these modules through optical link and VMEbus. These producers polled the event data from the buffers at a (configurable) frequency of 500 Hz (HGCAL DAQ rate $\approx 40~{}\text{Hz}$), converted the raw data into the EUDAQ format and sent it to the main DAQ for storage and online data monitoring. ### 5.2 AHCAL The front-end electronics of the AHCAL prototype were designed for low power operation under a specific ILC accelerator timing with a less than 1 ms long spill followed by 199 ms idle time. The operation of the Spiroc ASIC [15] was therefore split in 3 phases: 1) acquisition phase, where self-triggered events were stored into up to 16 analog memory columns; 2) conversion phase, where up to 16 events were sequentially digitized by internal ADC; 3) readout phase, where the digitized data was read out. Detailed timing was described in [16]. For beam test purposes, the acquisition phase length was extended to 16 ms and the readout phase varied typically between 2 to 20 ms, depending on the hit occupancy. Any external trigger from the sync board stopped the acquisition for immediate readout, as shown in Fig 12. Figure 12: The time diagram of AHCAL acquisition and readout phases. Due to the ‘self-trigger’ design of the ASIC, the AHCAL did not require an external trigger for data taking. All the hits (including noise hits) were read out and referenced by a number of bunch-crossing clock cycles (4 us period, called BXID) from the start of the acquisition phase. In order to assign an external trigger to the hits in the AHCAL, the DAQ internally samples the external trigger number (with a time-stamp, 48-bit counter with a 25 ns resolution) and the time-stamp of the start of the relevant acquisition phase. The trigger was assigned to one of the self-triggered events in the acquisition cycle. Assignment was based on the startup time from the start of the acquisition phase to the beginning of the first BXID and the additional delay due to the length of the trigger cable. Figure 13: The organization of self-triggered particles in AHCAL BXIDs, showing the BXID ambiguity for an example of split events due to the time of arrival with respect to the BXID clock. The collation of events according to the BXID might have however led to an existence of incomplete events for particles, that arrived close to the BXID counter value switching in the ASICs. An example for such a split event is shown in Figure 13 (particle no. 2). Several factors contribute to the BXID ambiguity: time walk of the signals, clock skew due to board and ASIC location, clock tree distribution through the FPGAs and clock jitter. The internal ASIC TDC [15] had also a region of non-linearity around the BXID change. Therefore, particles arriving close to the BXID change, within 10 ns, needed to be excluded from the data analysis. The AHCAL provided two means of synchronization with the HGCAL data: the trigger number and the trigger time-stamp, which used the 40 MHz clock from the HGCAL sync board. Both pieces of information were accessible in the data file. ## 6 Collected Data The DAQ was operated at a readout rate of 40 Hz. At this rate the amount of data readout for a typical 5 second spill of the SPS for the HGCAL prototype was approx 300 MB, as each of the 94 silicon modules sent 16 KB of data per event. In the last run, over a period of two weeks, six million events were collected with beams of charged hadrons, electrons, and muons with momenta from 20 to 300 GeV/c, with different detector configurations. Figure 14 shows accumulated events for different detectors over the beam-test campaign. Figure 14: The accumulated events for different detectors during beam-test campaign in October 2018. ## 7 Summary In the upgrade of the CMS detector for when the HL-LHC is operational, the two endcap calorimeters will be replaced with high granularity sampling calorimeters equipped with silicon sensors. As part of the development of this calorimeter, a series of beam tests have been conducted with different sampling configurations using prototype segmented silicon detectors readout with a low-cost custom scalable data acquisition system. The software framework used for the run control and data collection was the portable modular EUDAQ framework. In the most recent of the tests conducted in late 2018 at the CERN SPS in 2018, the performance of a prototype calorimeter equipped with ${\approx}12,000\rm{~{}channels}$ of silicon sensors, in conjunction with the CALICE prototype analogue hadron calorimeter, was studied with beams of high-energy electrons, pions and muons, with six million events collected over a two week period. ## Acknowledgments We thank the technical and administrative staffs at CERN and at other CMS institutes for their contributions to the success of the CMS effort. We acknowledge the enduring support provided by the following funding agencies: BMBWF and FWF (Austria); CERN; CAS, MoST, and NSFC (China); MSES and CSF (Croatia); CEA and CNRS/IN2P3 (France); SRNSF (Georgia); BMBF, DFG, and HGF (Germany); GSRT (Greece); DAE and DST (India); MES (Latvia); MOE and UM (Malaysia); MOS (Montenegro); PAEC (Pakistan); FCT (Portugal); JINR (Dubna); MON, RosAtom, RAS, RFBR, and NRC KI (Russia); MST (Taipei); ThEPCenter, IPST, STAR, and NSTDA (Thailand); TUBITAK and TENMAK (Turkey); STFC (United Kingdom); DOE (USA). ## References * [1] D. Contardo, M. Klute, J. Mans, L. Silvestris, and J. Butler, “Technical Proposal for the Phase-II Upgrade of the CMS Detector,” 2015. * [2] C. Collaboration, “The Phase-2 Upgrade of the CMS Endcap Calorimeter,” Tech. Rep. CERN-LHCC-2017-023. CMS-TDR-019, CERN, Geneva, Nov 2017. Technical Design Report of the endcap calorimeter for the Phase-2 upgrade of the CMS experiment, in view of the HL-LHC run. * [3] A. Steen, B. Akgun, et al., “Construction, commissioning and response equalisation of cms ce prototype silicon modules,” In preparation for JINST. * [4] J. Borg, S. Callier, D. Coko, F. Dulucq, C. de La Taille, L. Raux, T. Sculac, and D. Thienpont, “SKIROC2-CMS an ASIC for testing CMS HGCAL,” JINST, vol. 12, no. 02, p. C02019, 2017. * [5] F. Sefkow et al., “A highly granular SiPM-on-tile calorimeter prototype,” J. Phys.: Conf. Ser., vol. 1162, p. 012012, 2019. * [6] M. Baber et al., “Development and testing of an upgrade to the CMS level-1 calorimeter trigger,” JINST, vol. 9, no. 01, pp. C01006–C01006, 2014. * [7] C. G. Larrea, K. Harder, D. Newbold, D. Sankey, A. Rose, A. Thea, and T. Williams, “IPbus: a flexible ethernet-based control system for xTCA hardware,” JINST, vol. 10, no. 02, pp. C02019–C02019, 2015. * [8] J. Mans and E. Frahm, “Status report on a MicroTCA card for HCAL trigger and readout at SLHC,” JINST, vol. 5, no. 12, pp. C12027–C12027, 2010\. * [9] E. D. Team, “EUDAQ User Manual - Last update on October 2016 for EUDAQ version 1.7,” 2016. * [10] V. B. et al., “Using xdaq in application scenarios of the cms experiment,” FERMILAB-CONF-03-293, CHEP-2003-MOGT008 (2003). * [11] “https://root.cern.ch/guides/reference-guide,” * [12] “https://twiki.cern.ch/twiki/bin/view/cmspublic/workbook,” * [13] J. Spanggaard, “Delay Wire Chambers - A Users Guide,” SL-Note-98-023, 1998\. * [14] N. Akchurin et al., “First beam tests of prototype silicon modules for the CMS High Granularity Endcap Calorimeter,” JINST, vol. 13, p. P10023, 2018. * [15] S. Lorenzo, S. Callier, J. Fleury, F. Dulucq, C. de La Taille, G. Martin Chassard, L. Raux, and N. Seguin-Moreau, “SPIROC: design and performances of a dedicated very front-end electronics for an ILC Analog Hadronic CALorimeter (AHCAL) prototype with SiPM read-out,” Journal of Instrumentation, vol. 8, p. C01027, 01 2013. * [16] J. Kvasnicka, “Data acquisition system for the CALICE AHCAL calorimeter,” Journal of Instrumentation, vol. 12, no. 03, p. C03043, 2017.
# Causal dynamics of null horizons under linear perturbations Peter K S Dunsby<EMAIL_ADDRESS>Cosmology and Gravity Group, Department of Mathematics and Applied Mathematics, University of Cape Town, Rondebosch 7701, Cape Town, South Africa South African Astronomical Observatory, Observatory 7925, Cape Town, South Africa Centre for Space Research, North-West University, Potchefstroom 2520, South Africa Seoktae Koh <EMAIL_ADDRESS>Department of Science Education, Jeju National University, Jeju, 63243, South Korea Abbas M Sherif<EMAIL_ADDRESS>Department of Science Education, Jeju National University, Jeju, 63243, South Korea ###### Abstract We study the causal dynamics of an embedded null horizon foliated by marginally outer trapped surfaces (MOTS) for a locally rotationally symmetric background spacetime subjected to linear perturbations. We introduce a simple procedure which characterizes the transition of the causal character of the null horizon. We apply our characterization scheme to non-dissipative perturbations of the Schwarzschild and spatially homogeneous backgrounds. For the latter, a linear equation of state was imposed. Assuming a harmonic decomposition of the linearized field equations, we clarify the variables of a formal solution to the linearized system that determine how the null horizon evolves. For both classes of backgrounds, the shear and vorticity 2-vectors are essential to the characterization, and their roles are made precise. Finally, we discuss aspects of the relationship between the characterizing conditions. Various properties related to the self-adjointness of the MOTS stability operator are extensively discussed. ## I Introduction Background and motivation: The discovery of supermassive black holes as well as the detection of gravitational waves emitted by inspiralling black holes have ignited renewed interest in black hole physics. A key concept in understanding the dynamics of black holes is the notion of a trapped surface, those on which both ingoing and outgoing null rays are converging, introduced in the proof of the 1965 singularity theorem by Roger Penrose [1]. Marginally outer trapped surfaces (MOTS) are the marginal case of trapped surfaces, closed surfaces on which the outgoing null ray is neither converging nor diverging. Slight iterations to the definition of MOTS can be found in the literature from Wald’s marginally trapped surface [2] (note the “outer” here has been dropped) which rather imposes that both null rays are non-diverging, to Hayward’s marginal surface [3] which is in fact what we call a MOTS here, but in Hayward’s case the choice is which of the null ray that is neither converging nor diverging is a not fixed. (See the references [4, 5] for excellent reviews and additional details. The notion of a MOTS will also be rigorously defined in Section III). MOTS foliate 3-dimensional hypersurfaces known as marginally outer trapped tubes (MOTTs) [8], and where the causal character is fixed at all points of a MOTT it will be referred to by the common name horizon, which under certain conditions bound trapped regions containing trapped surfaces. Dynamical and null horizons are well known cases of horizons (see for example [9]). The timelike MOTTs are usually referred to simply as membranes as they, by definition, cannot enclose trapped surfaces. MOTS have also been fundamental to understanding the formation of black holes due to gravitational collapse, as well as their dynamics and evolution. MOTS admit a notion of stability [6, 7] (consequently leading to a notion of stability of black holes). This is a way to determine whether one can deform a MOTS $\mathcal{S}$ to another $\mathcal{S^{\prime}}$ along a unit direction normal to $\mathcal{S}$ that points along the slice in which $\mathcal{S}$ is embedded. This stability condition is given in terms of conditions on the principal eigenvalue of a second order linear elliptic operator. It has been shown that under physically reasonable conditions, strict stability, where the principal eigenvalue is positive, a MOTS will evolve to foliate a DH [6] (the same conclusion follows in the case that the operator has no zero eigenvalue). Given a horizon embedded in a background spacetime which is subjected to a perturbation, one expects the perturbation to in general affect the dynamics, and hence causal character of the horizon. Perturbation of horizons have been examined variously in different contexts using different approaches. For example, in [10] linear perturbations of a null horizon by a gravitational field was studied and it was established that such perturbations do not affect the causal character of the horizon. This feature of the invariance of the causal character, as have already been mentioned, will not be the case in general, and this was neatly discussed in [11], where generic first and higher order perturbations of non-expanding horizons were considered. A more explicit example was considered by Pilkington et al. in [12]. Embedding a stationary metric in the Weyl solution is a way of smoothly distorting stationary blackholes (see for example [13]). The Weyl potentials tune the behavior of horizons and are referred to as distortion potentials. The authors demonstrated that while for small distortions the horizon is apparently ‘well behaved’, for large distortions, even the MOTS structure of the horizon is not assured, let alone the causal character. Objectives: Our aim here is to introduce an approach which characterizes the evolution of MOTS in a linearly perturbed locally rotationally symmetric (LRS) background spacetime [14, 15]. More precisely we consider the following problem: How does a null horizon foliated by MOTS, embedded in a LRS background solution, evolve when the background is subjected to linear perturbations? We will employ the 1+1+2 covariant formalism [16, 17], which has been used in several works to study MOTS and their evolution in background LRS spacetimes [18, 19, 20, 21]. After formulating the characterization scheme, we will adapt our approach to cases of some well known background solutions. Specifically, we consider the null event horizon in the Schwarzschild background and null horizons in hypersurface orthogonal and spatially homogeneous background solutions. As will be seen, at least in these examples to which our scheme will be applied, the shear and vorticity 2-vectors appear to consistently play the pivotal roles in determining the causal behavior of the null horizons under perturbations. (The roles of these 2-vectors will be made more precise in the text). We also aim to consider aspects of the self-adjointness of the stability operator. As it will be seen, the 1-form that characterizes the self- adjointness of the operator is defined in terms of the shear and vorticity 2-vectors as well, and so it is quite natural to consider this in relation to the evolution of the horizons. Structure: Section II of the paper gives a quick overview of the covariant formalism that is used, and the procedure for linearizing a LRS background solution, which provides the mapping between a background solution and a perturbed one, is outlined. The process to harmonically decompose first order scalars, 2-vectors and 2-tensor, in this formalism is briefly reviewed following the standard literature. In Section III, we introduce an approach to characterize MOTS evolution in the perturbed solution. We start by briefly introducing the notion of a MOTS, which is well proliferated in the literature but nonetheless necessary for the flow of the paper. We then go on to introduce the characterization scheme where it becomes transparent how the required gauge choice forces us to consider a null horizon in the background. That is, our scheme does not work for a non-null horizon. Section IV applies the characterization scheme to static and spatially homogeneous LRS backgrounds. In both cases the evolution of the null horizon is characterized, with the case of the latter an equations of state being imposed and a partial characterization provided. In Section V, we discuss some simple implications of the results of Section IV for the self-adjointness of the MOTS stability. Several additional properties of the MOTS and aspects of the stability operator are discussed. We conclude with discussion of results in Section VI and discuss potential avenues for future research. ## II Preliminaries In this section we briefly introduce the 1+1+2 covariant formalism. We then describe the general prescription for the perturbative scheme that have been used extensively over the last few years, to carry out perturbations of background LRS solutions. ### II.1 Covariant decomposition There is by now a well established trove of literature on the 1+1+2 covariant formalism, with several references providing details of the steps in the decomposition. A reader interested in more details is referred to the following standard texts [16, 17] (Also see the references [22, 23, 24] for some recent corrections to the full set of equations). As such, the introduction here would be very brief, only necessary enough for the work carried out in this paper. In the well known and powerful 1+3 approach, a unit tangent direction - call this $u^{a}$ \- along which timelike observers flow, is what threads the spacetime. The field equations are decomposed along $u^{a}$ plus some constraint equations. If there exists some unit spacelike vector $n^{a}$ obeying $u_{a}n^{a}=0$, the 3-space from the previous split can be further decomposed along this direction as a $1+2$ product manifold. This, in addition to the evolution equations along $u^{a}$, introduces a set of propagation equations along $n^{a}$, as well as some constraint equations. The splitting naturally introduces new derivative along $n^{a}$ and on the 2-space (generally referred to as the sheet): * • Along $u^{a}$: $\dot{\psi}^{a\cdots b}_{\ \ \ \ c\cdots d}=u^{f}\nabla_{f}\psi^{a\cdots b}_{\ \ \ \ c\cdots d}$; * • Along $n^{a}$: $\hat{\psi}^{a\cdots b}_{\ \ \ \ c\cdots d}=n^{f}\nabla_{f}\psi^{a\cdots b}_{\ \ \ \ c\cdots d}$; * • Along the sheet: $\delta_{f}\psi^{a\cdots b}_{\ \ \ \ c\cdots d}=N^{\bar{c}}_{\ c}\cdots N^{\bar{d}}_{\ d}N^{a}_{\ \bar{a}}\cdots N^{b}_{\ \bar{b}}N^{e}_{\ f}\nabla_{e}\psi^{\bar{a}\cdots\bar{b}}_{\ \ \ \ \bar{c}\cdots\bar{d}}$, for any tensor $\psi^{a\cdots b}_{\ \ \ \ c\cdots d}$, where $N^{ab}=g^{ab}+u^{a}u^{b}-n^{a}n^{b}$ projects vectors and tensors to the sheet (it is the metric induced from the splitting, on the sheet). A spacetime vector $\psi^{a}$ can accordingly be decomposed as $\psi^{a}=\Psi u^{a}+\bar{\Psi}n^{a}+\Psi^{a}$ ($\Psi^{a}$ is the sheet component of $\psi^{a}$), and a fully projected tensor ${}^{3}\psi_{ab}$ on the 3-space decomposes as ${}^{3}\psi_{ab}=\Psi\left(n_{a}n_{b}-\frac{1}{2}N_{ab}\right)+\Psi_{(a}n_{b)}+\Psi_{ab},$ (1) with $\Psi=-^{3}\psi_{ab}N^{ab}$, $\Psi_{a}=N_{a}^{\ b}\ {}^{3}\psi_{bc}n^{c}$ and $\displaystyle\Psi_{ab}=\left(N_{(a}^{\ c}N_{b)}^{\ d}-\frac{1}{2}N_{ab}N^{cd}\right)\ ^{3}\psi_{cd},$ (2) is the symmetric, fully projected and trace-free part of ${}^{3}\psi_{ab}$. The above can be seen as the 1+1+2 analogue of the 1+3 orthogonally projected symmetric trace-free part of a 4-tensor: $\displaystyle\psi^{\langle ab\rangle}=\left(h^{(a}_{\ c}h^{b)}_{\ d}-\frac{1}{3}h^{ab}h_{cd}\right)\psi^{cd},$ (3) where $h_{ab}=g_{ab}+u_{a}u_{b}$ which introduces the derivative operator $\displaystyle D_{f}\psi^{a\cdots b}_{\ \ \ \ c\cdots d}=h^{\bar{c}}_{\ c}\cdots h^{\bar{d}}_{\ d}h^{a}_{\ \bar{a}}\cdots h^{b}_{\ \bar{b}}h^{e}_{\ f}\nabla_{e}\psi^{\bar{a}\cdots\bar{b}}_{\ \ \ \ \bar{c}\cdots\bar{d}}.$ So, for example the shear, electric and magnetic Weyl 3-tensors decompose respectively as $\displaystyle\sigma_{ab}$ $\displaystyle=\sigma\left(n_{a}n_{b}-\frac{1}{2}N_{ab}\right)+\Sigma_{(a}n_{b)}+\Sigma_{ab},$ (4) $\displaystyle E_{ab}$ $\displaystyle=\mathcal{E}\left(n_{a}n_{b}-\frac{1}{2}N_{ab}\right)+\mathcal{E}_{(a}n_{b)}+\mathcal{E}_{ab},$ (5) $\displaystyle H_{ab}$ $\displaystyle=\mathcal{H}\left(n_{a}n_{b}-\frac{1}{2}N_{ab}\right)+\mathcal{H}_{(a}n_{b)}+\mathcal{H}_{ab}.$ (6) Similarly, gradients of scalars decompose along these directions: $\displaystyle\nabla_{a}\psi=-\dot{\psi}u_{a}+\hat{\psi}n_{a}+\delta_{a}\psi.$ (7) The energy momentum tensor takes the covariant form $\displaystyle T_{ab}=\rho u_{a}u_{b}+ph_{ab}+2q_{(a}u_{b)}+\pi_{ab}.$ Here, $\rho=T_{ab}u^{a}u^{b}$ is the (local) energy density, $3p=T_{ab}h^{ab}$ is the (isotropic) pressure, $q_{a}=T_{bc}h^{b}_{\ a}u^{c}$ is the heat flux vector ($q_{a}=Qn_{a}+Q_{a}$), and $\pi_{ab}$, which can be decomposed using (1) (with $\Pi=T_{ab}n^{a}n^{b}-p$), captures the degree of anisotropy. From the Einstein field equations with cosmological constant $\displaystyle R_{ab}-\frac{1}{2}Rg_{ab}+\Lambda g_{ab}=T_{ab},$ (8) where $R_{ab}$ and $R$ are respectively the spacetime Ricci tensor and scalar curvature, one can write down the Ricci tensor of the spacetime in the covariant form $\displaystyle R_{ab}$ $\displaystyle=g_{1}u_{a}u_{b}+g_{2}n_{a}n_{b}+2\left(Qu_{(a}+\Pi_{(a}\right)n_{b)}$ $\displaystyle+\left(g_{2}-\frac{3}{2}\Pi\right)N_{ab}+2Q_{(a}u_{b)}+\Pi_{ab},$ where $\displaystyle g_{1}=\frac{1}{2}\left(\rho+3p-2\Lambda\right)\quad\mbox{and}\quad g_{2}=\frac{1}{2}\left(\rho-p+2\Lambda+2\Pi\right).$ The covariant derivatives of the unit vector fields can be written down in terms of the kinematic (geometrical) variables: $\displaystyle\nabla_{a}u_{b}$ $\displaystyle=-u_{a}\left(\mathcal{A}e_{b}+\mathcal{A}_{b}\right)+\left(\frac{1}{3}\theta+\sigma\right)n_{a}n_{b}$ $\displaystyle+\frac{1}{2}\left(\frac{2}{3}\theta-\sigma\right)N_{ab}+\Omega\varepsilon_{ab}+\Sigma_{ab}$ $\displaystyle+2\left(n_{(a}\Sigma_{b)}+n_{[a}\varepsilon_{b]c}\Omega^{c}\right),$ (9a) $\displaystyle\nabla_{a}n_{b}$ $\displaystyle=-u_{a}\left(\mathcal{A}u_{b}+\alpha_{b}\right)+\left(\frac{1}{3}\theta+\sigma\right)n_{a}u_{b}$ $\displaystyle+\frac{1}{2}\phi N_{ab}+\xi\varepsilon_{ab}+\zeta_{ab}+n_{a}a_{b}$ $\displaystyle+\left(\Sigma_{a}-\varepsilon_{ac}\Omega^{c}\right)u_{b}.$ (9b) $\mathcal{A}=n_{a}\dot{u}^{a}$ is acceleration, $\theta=h^{ab}\bar{D}_{a}u_{b}$ is expansion of $u^{a}$, $\phi$ is expansion of $n^{a}$ (called sheet expansion), $\sigma=\sigma_{ab}n^{a}n^{b}$ is the shear, $2\Omega=\varepsilon^{ab}\nabla_{a}u_{b}$ and $2\xi=\varepsilon^{ab}\nabla_{a}n_{b}$ are the respective twists of $u^{a}$ and $n^{a}$ ($\Omega$ is usually referred to as vorticity or rotation, and $\varepsilon_{ab}$ is the 2-dimensional alternating tensor); $\mathcal{A}^{a}$ is the sheet component of $\dot{u}_{a}$; $\alpha_{a}$ is the sheet component of $\dot{n}_{a}$; $a_{a}=\hat{n}_{a}$ is the acceleration of $n^{a}$ and; $\zeta_{ab}=\delta_{\\{a}n_{b\\}}$ is the shear of $n^{a}$. The field equations then takes a covariant form as a set of evolution and propagation equations of the covariant scalars, vectors and tensors, along with some constraint equations. These are obtained from the field equations themselves, and the Ricci and Bianchi identities for the unit fields $u^{a}$ and $n^{a}$. We do not need all of these equations here and will not enlist them. The few equations that will be needed for the current work will be introduced accordingly. ### II.2 Mapping the background and ‘true’ spacetimes Perturbative schemes to study LRS cosmologies have been recently implemented using the 1+1+2 formalism. They have largely been applied to the LRS II class of spacetimes, with a recent consideration being a generalization which introduces dissipation of the fluid via the non-vanishing of the anisotropic stress on the background (see for example these recent works [22, 23]). In this section, we begin by first introducing LRS spacetimes in context of the 1+1+2 formulation, and we briefly discuss the linearization procedure (more details can be found in the references [16, 17]), up to where it is relevant to this paper. Locally rotationally symmetric (LRS) spacetimes admit a continuous isotropy group at each point, as well as a preferred spatial direction (see the references [14, 15] for more details). The symmetry of these spacetimes allows one to write the metric in local coordinates $(t,\mathcal{R},y,z)$ as (see [15] for example) $\displaystyle ds^{2}$ $\displaystyle=-A_{1}^{2}dt^{2}+A_{2}^{2}dr^{2}+A_{3}^{2}dy^{2}$ $\displaystyle+\left(\left(DA_{3}\right)^{2}+\left(A_{2}h\right)^{2}-\left(A_{1}g\right)^{2}\right)dz^{2}$ $\displaystyle+2\left(A_{1}^{2}gdt-A_{2}^{2}hdr\right)dz,$ (10) where $A_{i}=A_{i}\left(t,r\right)$, and $g=g(y)$, $h=h(y)$. $D=D(y,k)$ where $k$ is a constant and specified $D$: For $k=-1$, $D=\sinh y$; for $k=0$, $D=y$ and; for $k=1$, $D=\sin y$. In the limiting case that $g=0=h$, we recover the well studied LRS II class of spacetimes, which generalizes spherically symmetric solutions to the Einstein field equations. This class includes FLRW models, LTB spacetime model, Schwardschild solution, Oppenheimer–Snyder dust model, etc. The covariant set $\displaystyle\mathcal{D}:\equiv\\{{\mathcal{A},\theta,\sigma,\phi,\xi,\Omega,\rho,p,Q,\Pi,\mathcal{E},\mathcal{H}\\}},$ (11) specifies a LRS spacetime. In general, the mapping between the background and perturbed spacetime generally depends on the problem of physical interest. And for the specific background under consideration, the linearization procedure may or may not be particularly cumbersome. Essentially, the problem reduces to determining which quantities vanish in the background and which do not. Here, we outline the procedure for the perturbation. (A detailed discussion on this is found in [17], formulated differently but unchanging underlying concept.) * • We first define the background spacetime: start with a background LRS spacetime $M$, and let the collection of scalars $\displaystyle\mathcal{D}_{1}:\equiv\\{{\varphi^{i}\\}}_{i\in I},$ (12) for some indexing set $I$, specify $M$, i.e. $\mathcal{D}_{1}\subset\mathcal{D}$. (Note that all 2-vector and 2-tensor quantities, those defined on the sheet, vanish in the background). * • Define the set $\displaystyle\mathcal{D}_{2}:\equiv\mathcal{D}\setminus\mathcal{D}_{1},$ (13) of those covariant scalars in $\mathcal{D}$ that vanish in the background. * • Further introduce the set of 2-vectors and 2-tensors: $\displaystyle\mathcal{D}_{3}:\equiv\\{{\Psi_{a}^{j},\Psi_{ab}^{w}\\}}_{j\in J,w\in W},$ (14) with $J$ and $W$ being indexing sets, which also vanish in the background. * • The full set of gauge invariant quantities is then $\displaystyle\mathcal{D}_{2}\cup\mathcal{D}_{3},$ (15) which follows from the Stewart-Walker Lemma [25]. One can then use the Ricci and Bianchi identities for $u^{a}$ and the direction of anisotropy $n^{a}$ to obtain the linearized equations for the set $\bar{D}:\equiv\mathcal{D}_{1}\cup\mathcal{D}_{2}\cup\mathcal{D}_{3}$. The linearized system is generally not gauge invariant. The evolution and/or propagation equations of the background quantities may contain terms that do not vanish in the background. This is resolved by introducing $\delta$ gradients of the background quantities whose evolution and propagation equations replace those of the background scalars. * • We note that the ‘type’ of perturbation being carried out will further simplify the collection $\mathcal{D}_{2}\cup\mathcal{D}_{3}$. For example, one may consider an irrotational or a shear-free perturbation, eliminating all shear and vorticity quantities. There is also a degree of freedom in choosing the frame as $n^{a}$ can be freely chosen. This choice will depend on the particular physical consideration at hand. One common choice when dealing with black hole perturbation is to set the sheet component of the acceleration to zero [16] which corresponds to a hovering observer. (At some point in this work we will make this choice of frame to transparently demonstrate our characterization scheme.) Other choices include the aligning of $n^{a}$ along the vorticity or the electric Weyl scalar. That is, one makes the respective frame choices $\Omega^{a}=0$ and $\mathcal{E}^{a}=0$. * • In general the linearized equations are quite messy and seldom tractible, analytically. This generally results from the presence of the angular ‘$\delta$’ derivative. One approach is to decompose first order variables into harmonics, which converts the linearized equations into ordinary differential equations which in some cases can be solved exactly. In this work we do not seek to solve the linearized system. Rather, we suppose that the harmonically decomposed linearized system can be solved exactly for an appropriate basis. Then, we specify the roles that certain components of the even (electric) vector harmonics (to be defined shortly) play in the evolution of a null horizon embedded in a LRS background which is subject to a perturbation. ### II.3 Harmonic decomposition We now present a quick overview of the harmonic decomposition procedure adapted to the formalism. Here we again follow the general setup in the references [16, 17]. One starts by introducing dimensionless harmonic functions that satisfy $\displaystyle\delta^{2}\mathcal{Q}=-\frac{\bar{k}^{2}}{r^{2}}\mathcal{Q},\quad\hat{\mathcal{Q}}=\dot{\mathcal{Q}}=0,\quad\quad\quad\quad\mbox{($0\leq\bar{k}^{2}$)}$ (16) on any LRS background, i.e. they are eigenfunctions of the sheet Laplacian. The function $r$ is covariantly specified by the relations $\displaystyle\dot{r}=\frac{r}{2}\left(\frac{2}{3}\theta-\sigma\right),\quad\hat{r}=\frac{r}{2}\phi,\quad\delta_{a}r=0.$ (17) One may decompose any first order scalar $\Psi$ as $\displaystyle\psi^{(1)}=\sum\psi_{S}^{(\bar{k})}\mathcal{Q}^{(\bar{k})},$ (18) usually written simply as $\psi^{(1)}=\psi_{S}\mathcal{Q}$, with the sum implicit over $\bar{k}$, and the $S$ subscript signalling a scalar decomposition. The 2-vectors are also expanded in harmonics via the following defined basis $\displaystyle\mathcal{Q}^{\bar{k}}_{a}$ $\displaystyle=r\delta_{a}\mathcal{Q}^{\bar{k}}\implies N^{b}_{\ a}\hat{\mathcal{Q}}_{b}=0=N^{b}_{\ a}\dot{\mathcal{Q}}_{b},$ $\displaystyle\delta^{2}\mathcal{Q}_{a}$ $\displaystyle=\frac{1}{r^{2}}\left(1-\bar{k}^{2}\right)\mathcal{Q}_{a};$ (19a) $\displaystyle\bar{\mathcal{Q}}^{\bar{k}}_{a}$ $\displaystyle=r\varepsilon_{ab}\delta^{b}\bar{\mathcal{Q}}^{\bar{k}}\implies N^{b}_{\ a}\hat{\bar{\mathcal{Q}}}_{b}=0=N^{b}_{\ a}\dot{\bar{\mathcal{Q}}}_{b},$ $\displaystyle\delta^{2}\bar{\mathcal{Q}}_{a}$ $\displaystyle=\frac{1}{r^{2}}\left(1-\bar{k}^{2}\right)\bar{\mathcal{Q}}_{a},$ (19b) respectively named even (electric) and odd (magnetic) vector harmonics. (For all that are to follow, there should be no confusion that is to arise if we drop the superscript $\bar{k}$ in relevant expressions.) Notice that the harmonic basis vectors $\mathcal{Q}_{a}$ and $\bar{\mathcal{Q}}_{a}$ can be obtained from one another, up to sign reversal, by applying $\varepsilon_{ab}$: $\bar{\mathcal{Q}}_{a}=\varepsilon_{ab}\mathcal{Q}^{b}\iff\mathcal{Q}_{a}=-\varepsilon_{ab}\bar{\mathcal{Q}}^{b}$. For this reason, $\varepsilon_{ab}$ is understood as a parity operator, in the sense of the harmonics introduced here. A quick check also shows that $\displaystyle\delta_{a}\bar{\mathcal{Q}}^{a}$ $\displaystyle=0=\varepsilon_{ab}\delta^{a}\mathcal{Q}^{b},$ $\displaystyle\delta_{a}\mathcal{Q}^{a}$ $\displaystyle=-\frac{\bar{k}^{2}}{r}\mathcal{Q}=-\varepsilon_{ab}\delta^{a}\bar{\mathcal{Q}}^{b}.$ Since for each $\bar{k}$, $\mathcal{Q}_{a}\bar{\mathcal{Q}}^{a}=0$, first order vectors ($2$-vectors) can always be decomposed as $\displaystyle\Psi_{a}=\sum\Psi_{V}^{(\bar{k})}\mathcal{Q}_{a}+\bar{\Psi}_{V}^{(\bar{k})}\bar{\mathcal{Q}}_{a}=\Psi_{V}\mathcal{Q}_{a}+\bar{\Psi}_{V}\bar{\mathcal{Q}}_{a},$ (21) again, where the sum over $\bar{k}$ is implicit, with the $V$ signalling a vector decomposition. We note the following important point: for any first order 2-vector $\Psi^{a}$, a quick check gives its sheet divergence as $\displaystyle\delta_{a}\Psi^{a}=-\frac{\bar{k}^{2}}{r}\Psi_{V}Q.$ (22) Thus, $\Psi^{a}$ is gradient if and only if $\Psi_{V}$ vanishes. For a $2$-tensor $\Psi_{ab}$, the even and odd parity tensor harmonics are defined respectively as $\displaystyle\mathcal{Q}_{ab}$ $\displaystyle=r^{2}\delta_{\\{a}\delta_{b\\}}\mathcal{Q}\implies\hat{\mathcal{Q}}_{ab}=0=\dot{\mathcal{Q}}_{ab},$ (23a) $\displaystyle\bar{\mathcal{Q}}_{ab}$ $\displaystyle=r^{2}\varepsilon_{c\\{a}\delta^{c}\delta_{b\\}}\mathcal{Q}\implies\hat{\bar{\mathcal{Q}}}_{ab}=0=\dot{\bar{\mathcal{Q}}}_{ab}.$ (23b) Again, we see that $\varepsilon_{ab}$ is a parity operator here: $\bar{\mathcal{Q}}_{ab}=\varepsilon_{c\\{a}\mathcal{Q}_{b\\}}^{\ \ c}\iff\mathcal{Q}_{ab}=-\varepsilon_{c\\{a}\bar{\mathcal{Q}}_{b\\}}^{\ \ c}$. Indeed, for each $\bar{k}$ $\mathcal{Q}_{ab}\bar{\mathcal{Q}}^{ab}=0$ so that any $\Psi_{ab}$ can be decomposed as $\displaystyle\Psi_{ab}=\Psi_{T}\mathcal{Q}_{ab}+\bar{\Psi}_{T}\bar{\mathcal{Q}}_{ab},$ with the $T$ signalling a tensor decomposition. ## III Characterizing MOTS evolution We introduce a characterization scheme for the evolution of MOTS orthogonal to $u^{a}$ and $n^{a}$. We are interested in how MOTS in constant time slices in linearized LRS background geometries evolve. We begin with a quick definition of MOTS, after which we introduce the characterization. As will be seen, the particular problem we are trying to address, or more precisely, how our objective is formulated, is imposed by the required gauge invariance of the evolution characterizing function. ### III.1 MOTS For the surfaces under consideration, the following must hold in order for ‘2-space’ to be a genuine surface rather than simply a collection of tangent planes (see the reference [17] for discussion): $\displaystyle\xi=\Omega=a^{a}$ $\displaystyle=0;$ $\displaystyle\Sigma^{a}+\varepsilon^{ab}\Omega_{b}-\alpha^{a}$ $\displaystyle=0.$ (24) (The left hand side of the second line is known as the Greenberg vector.) In particular, these conditions ensure that the Lie bracket of $u^{a}$ and $n^{a}$ vanishes, i.e. $u^{a}$ and $n^{a}$ are surface forming, and that the operator $\delta_{a}$ is a ‘true’ covariant derivative for the surface [17]. Consider a 2-surface $S$ in spacelike constant $t$ slices, with unit spacelike normal $n^{a}$. Outgoing and ingoing null normals are respectively $\displaystyle k^{a}=u^{a}+n^{a};\quad l^{a}=\frac{1}{2}\left(u^{a}-n^{a}\right),$ normalized so that $l_{a}k^{a}=-1$. There is a degree of freedom to rescale the null vectors as $k^{a}\rightarrow fk^{a},l^{a}\rightarrow f^{-1}l^{a}$, for $f>0$. But as will be noted shortly, the MOTS character relevant here is invariant under such scaling. The induced metric on $S$ is $\displaystyle\mathcal{F}_{ab}=g_{ab}+2k_{(a}l_{b)},$ (25) and the ingoing and outgoing expansions are, respectively, $\displaystyle\chi=\mathcal{F}^{ab}\nabla_{a}k_{b};\qquad\bar{\chi}=\mathcal{F}^{ab}\nabla_{a}l_{b}.$ (For our current purposes, $\mathcal{F}_{ab}$ here is just $N_{ab}$.) By MOTS, we mean that on $S$, one has $\chi=0$. In the case that the sign of $\bar{\chi}$ is constrained as $\bar{\chi}<0$ on $S$, MOTS is abbreviated to MTS. (A MTS is generally considered to be associated to horizons that enclose black holes in dynamical spacetimes.) One sees that, under the rescaling, the expansion becomes $\chi_{f}=f\chi$ (similarly, $\bar{\chi}_{f}=f^{-1}\bar{\chi}$). Thus, it is clear that any sign conditions on $\chi$ and $\bar{\chi}$ is seen to be invariant under the freedom to scale the null vectors. ### III.2 MOTS evolution Because of the MOTS condition $\chi=0$, we know that the gradient $\nabla_{a}\chi$ is orthogonal to the MOTS. Hence the norm, which we notate by $z$, specifies the causal character, pointwise: $\displaystyle z>0\quad\mbox{(Timelike)};\quad z<0\quad\mbox{(Spacelike)};\ z=0\quad\mbox{(Null)}.$ Let us decompose the expansion $\chi$ into zeroth (background) and first order scalars: $\displaystyle\chi=\chi_{0}+\chi_{1},$ with the subscripts $\ast_{0}$ and $\ast_{1}$ denoting, respectively, zeroth and first orders. (In general this will be the case as this is a linear combination of the covariant scalars, some of which may or may not vanish in the background.) We expand the norm $z$ as $\displaystyle z$ $\displaystyle=\left(-\dot{\chi}_{0}^{2}+\hat{\chi}_{0}^{2}\right)+\left(-\dot{\chi}_{1}^{2}+\hat{\chi}_{1}^{2}\right)+2\left(-\dot{\chi}_{0}\dot{\chi}_{1}+\hat{\chi}_{0}\hat{\chi}_{1}\right)$ $\displaystyle+\left(\delta_{a}\chi_{0}\delta^{a}\chi_{0}+\delta_{a}\chi_{1}\delta^{a}\chi_{1}+2\delta_{a}\chi_{0}\delta^{a}\chi_{1}\right).$ (26) As we neglect the products of first order quantities, the second and fourth parenthesized terms of (III.2) vanish and we have $\displaystyle z=\left(-\dot{\chi}_{0}^{2}+\hat{\chi}_{0}^{2}\right)+2\left(-\dot{\chi}_{0}\dot{\chi}_{1}+\hat{\chi}_{0}\hat{\chi}_{1}\right).$ (27) Indeed, it is clear that depending on the particular background solution one begins with, the above expression will simplify as will later be observed. Of course there will be some cases where even extending to the nonlinear regime, i.e. in cases of nonlinear perturbations, the characterization can proceed using the form of $z$ (27), for example when $z$ is optical, i.e. when $\delta_{a}z\delta^{a}z$ vanishes on the MOTS. Let us consider some cases where this is true for a particular frame choice or a class of backgrounds: The commutation relations between the ‘$\delta_{a}$’, and the dot and hat derivatives, acting on a scalar $\psi$, are given respectively as (we set $\Omega=\xi=a_{a}=0$) $\displaystyle\delta_{a}\dot{\psi}-N^{\ b}_{a}\left(\delta_{b}\psi\right)^{\cdot}$ $\displaystyle=-\mathcal{A}_{a}\dot{\psi}+\left(\Sigma_{a}-\varepsilon_{ab}\Omega^{b}+\alpha_{a}\right)\hat{\psi}$ $\displaystyle+\frac{1}{2}\left(\frac{2}{3}\theta-\sigma\right)\delta_{a}\psi,$ (28) $\displaystyle\delta_{a}\hat{\psi}-N^{\ b}_{a}\widehat{\left(\delta_{b}\psi\right)}$ $\displaystyle=-\left(\Sigma_{a}-\varepsilon_{ab}\Omega^{b}\right)\dot{\psi}+\frac{1}{2}\phi\delta_{a}\psi.$ (29) Now notice that the function $r$, introduced in the harmonic decomposition obeys $\displaystyle\dot{r}+\hat{r}=\frac{r}{2}\chi.$ And upon using the equations (28) and (29) with $\psi=r$ we find that $\displaystyle\delta_{a}\chi=\left(\mathcal{A}_{a}+\alpha_{a}\right)\phi.$ (30) So, a frame choice $\mathcal{A}_{a}=\alpha_{a}=0$, $\delta_{a}\chi=0$. Also, on any background with a vanishing sheet expansion, it is true that $\delta_{a}\chi=0$ on the MOTS. Thus, analyzing MOTS in perturbed spatially homogeneous or static LRS backgrounds, the form of the function $z$ suffices even in nonlinear regimes, although the linearized equations is now more formidable. However, we will keep the forthcoming applications modest and restrict ourselves to linear perturbations. As is easily seen from (27), the scalar $z$ will generally not be gauge invariant: Indeed, the second parenthesized terms do vanish in the background due to the presence of $\chi_{1}$ and derivatives thereof, but the first parenthesized ones do not. Characterization of the evolution of the MOTS in the perturbed spacetime obviously requires a gauge invariant $z$. In fact, $z$ is gauge invariant if and only if the combination of terms in the first parenthesis vanishes. It indeed follows that, were one to start with a null horizon $\bar{H}$, foliated by MOTS, in the background, $z$ is gauge invariant with respect to $\bar{H}$ since the vanishing of the sum in the first parenthesis characterizes the nullity of $\bar{H}$ in the first place. For this reason, we can more precisely frame what is being done in this work: Consider an exact LRS background spacetime $M$ and let $\bar{H}$ be a null horizon in $M$, foliated by MOTS. Let us suppose that $M$ is subject to a perturbation of linear order. How do we characterize the evolution of the horizon $\bar{H}$ under the perturbation? Now we emphasize again that we are not interested in solving linearized system. It is assumed that an appropriate gauge (and possibly frame) choice has been made, so that one has a mapping between the background and the perturbed spacetimes. Independent of the gauge choice, Then, by simply imposing the vanishing condition $\displaystyle-\dot{\chi}_{0}^{2}+\hat{\chi}_{0}^{2}=0,$ the function $z$ will characterize the evolution of a null horizon from the background. The gauge choice that maps the background and perturbed spacetimes of course determines the simplification of $z$ as it is constructed from elements of the linearized equations. Now, to ensure a consistent sign of $z$ along a horizon foliated by the MOTS under consideration, one requires $\delta_{a}z=0$. Explicitly, this is the requirement that $\displaystyle\dot{\chi}_{0}\delta_{a}\dot{\chi}-\hat{\chi}_{0}\delta_{a}\hat{\chi}=0.$ (31) Again, once a background is specified, the expression (31) simplifies as will be seen in Section IV. * Remark 1. We point out that if one relaxes the $\delta_{a}z=0$ condition, one allows for a range of possibilities, for example, horizons with different portions exhibiting varying causal characters. This is beyond the scope of the current work and will be deferred for a subsequent paper. Let us now compute the forms of the outgoing and ingoing null expansion scalars: $\displaystyle\chi=\frac{2}{3}\theta-\sigma+\phi,\qquad\bar{\chi}=\frac{1}{2}\left(\frac{2}{3}\theta-\sigma-\phi\right).$ (32) Thus, a MOTS is specified by the covariant triple $(\theta,\sigma,\phi)$, and for any particular background, this splits into the background and first order parts. If our interest is in MOTS that are MTS, then the requirement that $\bar{\chi}<0$ implies that necessarily $\phi>0$. For our purposes here, however, we will not impose such restriction and will instead consider the more general ‘MOTS’. We note that the form of $\chi$, first equation of (32), is form invariant with respect to the background and so the gauge problem is automatically fixed. This is a consequence of the particular class of MOTS we are considering for this work. For a different class of MOTS, this correspondence may be broken and some additional condition may have to be imposed to fix the gauge, in general. However, up to linear order for null horizons as is considered here (or more generally nonexpanding horizons) $\chi$ vanishes on the perturbed horizon as was established in [11]. * Remark 2. Before proceeding, we point out that caution is to be exercised when considering the vanishing of the first parenthesized terms of (III.2). For example if one is to consider a static background, then obviously, $\dot{\chi}^{2}_{0}$ should vanish. However, $\hat{\chi}^{2}_{0}$ does not vanish on a null horizon in the background. So, the terms should be taken together. In fact, its easily seen that the first parenthesized terms of (III.2) – we denote this as $F$ – factors as $\displaystyle F=-\mathcal{L}_{k}\chi\mathcal{L}_{l}\chi.$ (33) The vanishing condition in the background, assuming the null energy condition is then $\mathcal{L}_{k}\chi=0$. So, one has to take care of the function $F$, so that once a background is specified, $F$ vanishes on a null horizon in the background. The evolution and propagation equations for $\phi$ and $(2/3)\theta-\sigma$, up to linear order, are [16] $\displaystyle\dot{\phi}$ $\displaystyle=\left(\frac{2}{3}\theta-\sigma\right)\left(\mathcal{A}-\frac{1}{2}\phi\right)+Q+\delta_{a}\alpha^{a},$ (34a) $\displaystyle\hat{\phi}$ $\displaystyle=-\frac{1}{2}\phi^{2}+\left(\frac{2}{3}\theta-\sigma\right)\left(\frac{1}{3}\theta+\sigma\right)-\mathcal{E}-\frac{1}{2}\Pi$ $\displaystyle-\frac{2}{3}\left(\rho+\Lambda\right)+\delta_{a}a^{a},$ (34b) $\displaystyle\frac{2}{3}\dot{\theta}-\dot{\sigma}$ $\displaystyle=\mathcal{A}\phi-\frac{1}{2}\left(\frac{2}{3}\theta-\sigma\right)^{2}+\mathcal{E}-\frac{1}{2}\Pi$ $\displaystyle-\frac{1}{3}\left(\rho+3p-2\Lambda\right)+\delta_{a}\mathcal{A}^{a},$ (34c) $\displaystyle\ \frac{2}{3}\hat{\theta}-\hat{\sigma}$ $\displaystyle=\frac{3}{2}\phi\sigma+Q+\delta_{a}\left(\Sigma^{a}+\varepsilon^{ab}\Omega_{b}\right).$ (34d) Equation (34d) is obtained by projecting along $n^{a}$ the shear divergence equation [26] $\displaystyle D^{b}\sigma_{ba}-\frac{2}{3}D_{a}\theta+\varepsilon_{abc}\left(D^{b}+2\dot{u}^{b}\right)\omega^{c}+q_{a}=0,$ where $\displaystyle\omega^{a}=\Omega n^{a}+\Omega^{a},\quad q^{a}=Qn^{a}+Q^{a}.$ Equations (34a), (34b), and (34c) are obtained, respectively, by applying $u^{a}N^{bc}$, $n^{a}N^{bc}$, and $-n^{a}u^{b}u^{c}$ to the Ricci identities for $n^{a}$. We can write down the evolution of the null expansion along the null rays as $\displaystyle\mathcal{L}_{k}\chi$ $\displaystyle=\chi\left(\mathcal{A}-\frac{1}{2}\phi+\frac{3}{2}\sigma\right)-\left(\rho+p+\Pi\right)$ $\displaystyle+2Q+\delta_{a}\left(\alpha^{a}+\mathcal{A}^{a}+a^{a}+\Sigma^{a}+\varepsilon^{ab}\Omega_{b}\right),$ (35) $\displaystyle\mathcal{L}_{l}\chi$ $\displaystyle=-\frac{1}{2}\chi\left(\chi-\frac{1}{2}\phi+\frac{3}{2}\sigma\right)+\frac{1}{3}\left(\rho-3p\right)+2\mathcal{E}$ $\displaystyle+\delta_{a}\left(\alpha^{a}+\mathcal{A}a-a^{a}-\Sigma^{a}-\varepsilon^{ab}\Omega_{b}\right).$ (36) Note that $\chi$ vanishes on the horizon and hence the above are simplified there. The scalar $F$ now expands, up to linear order on the horizon, to the form $\displaystyle F$ $\displaystyle=\left(\frac{1}{3}\left(\rho-3p\right)+2\mathcal{E}\right)$ $\displaystyle\times\left(\left(\rho+p+\Pi\right)-2Q-\delta_{a}Z^{a}\right)$ (37) where we have defined $\displaystyle Z^{a}=\alpha^{a}+\mathcal{A}^{a}+a^{a}+\Sigma^{a}+\varepsilon^{ab}\Omega_{b}.$ (38) Depending on the particular class of background, and how one chooses the $\\{u^{a}-n^{a}\\}$ frame by appropriately fixing the $n^{a}$ direction, the scalar $F$ can be considerably simplified as will be seen in the next section. #### The future outer trapping condition A MOTS $\mathcal{S}$ on which the ingoing expansion $\bar{\chi}$ is strictly negative and $\mathcal{L}_{l}\chi<0$ is known as a future outer trapped surface (FOTS). A horizon foliated by FOTS is a future outer trapping horizon (FOTH). These were introduced by Hayward in [3] and have since be extensively studied [27, 28].) In the dynamical and isolated cases, respectively, assuming the null energy condition, one has that $\mathcal{L}_{k}\chi<0$ and $\mathcal{L}_{k}\chi=0$. In this case we have a dynamical FOTH and an isolated FOTH. Crucially, the FOTH condition immediately implies the presence of a black hole as it ensures the existence of trapped surfaces just to the inside of the horizon. Thus, the conditions characterizing a FOTH for the dynamical and isolated cases are $\bar{\chi}<0$, (31), plus $\mathcal{L}_{k}\chi<0$ and $\mathcal{L}_{k}\chi=0$, respectively, $\displaystyle z$ $\displaystyle<0,\dot{\chi}<0,\hat{\chi}<0,$ (39) $\displaystyle z$ $\displaystyle=0,\dot{\chi}<0,\hat{\chi}>0.$ (40) Henceforth, when specifying to a FOTH the condition $\bar{\chi}<0$ will be assumed. In the next section, we consider some backgrounds as particular examples, to apply the horizon characterization scheme developed in this section to horizons in the corresponding perturbed spacetime. ## IV Applications In this section we consider some particular background LRS solutions. Specifically, we consider static and spatially homogeneous LRS backgrounds, with restriction to the class II. We will also briefly comment on the case of a background with both $t$ and $r$ dependence (see Remark 3). ### IV.1 Static background Let us take as an example a non-dissipative linear perturbation of the Schwarzschild background with metric (we set the mass to unity here, which does not, in any serious way, affect the calculations) $\displaystyle ds^{2}=-\left(1-\frac{2}{r}\right)dt^{2}+\left(1-\frac{2}{r}\right)^{-1}dr^{2}+r^{2}d\varsigma^{2},$ (41) where $d\varsigma^{2}$ is the metric on the unit 2-sphere. We know the above is time-symmetric, i.e. $\theta=0=\sigma$. Hence, we have the zeroth and first orders decomposition of the null expansion $\chi$ as $\displaystyle\chi_{0}=\phi\qquad\mbox{and}\qquad\chi_{1}=\frac{2}{3}\theta-\sigma.$ Therefore, $\dot{\chi}_{0}$ and $\dot{\chi}_{1}$ are first order quantities so that $\dot{\chi}_{0}^{2}=\dot{\chi}_{0}\dot{\chi}_{1}=0$. The only non- vanishing background covariant scalars are $\mathcal{A},\phi,\mathcal{E}$, given in coordinate forms as $\displaystyle\mathcal{A}=\frac{1}{r^{2}}\left(\sqrt{1-\frac{2}{r}}\right)^{-1};\quad\phi=\frac{2}{r}\sqrt{1-\frac{2}{r}};\quad\mathcal{E}=-\frac{2}{r^{3}}.$ In this background the MOTS is located at $\phi=0\iff r=2$. Now, the norm $z$ becomes $\displaystyle z$ $\displaystyle=F+2\hat{\chi}_{0}\hat{\chi}_{1}$ $\displaystyle=-4\mathcal{E}\delta_{a}\left(Z^{a}+\Sigma^{a}+\varepsilon^{ab}\Omega_{b}\right).$ (42) Additionally one also requires, from (31), that (we have now set $a^{a}=0$ on the horizon as it should be) $\displaystyle 0=\hat{\chi}_{0}\delta_{a}\hat{\chi}=\delta_{a}\mathcal{E}+\delta_{a}\left[\delta_{b}\left(\Sigma^{b}+\varepsilon^{bc}\Omega_{c}\right)\right].$ (43) Since $\mathcal{E}<0$, coupled with (43), the $r=2$ horizon in the Schwarzschild background evolves into an isolated (resp. dynamical) horizon provided that $\displaystyle\delta_{a}\left(Z^{a}+\Sigma^{a}+\varepsilon^{ab}\Omega_{b}\right)=\mbox{({resp.} $>$)}0.$ (44) Separately, we examine the two cases of (44). #### IV.1.1 Null case We start with the null criterion, with the vanishing of (44), in harmonics, yielding $\displaystyle 0=-\frac{\bar{k}^{2}}{r^{2}}\left[\mathcal{A}_{V}+2\left(\Sigma_{V}-\bar{\Omega}_{V}\right)\right],$ (45) which should hold for all $\bar{k}$. By defining $Y_{a}=\delta_{a}\mathcal{E}$ and decomposing the equation (43) into harmonics, we write it down as the pair of equations $\displaystyle Y_{V}=-\frac{\bar{k}^{2}}{r^{2}}\left(\Sigma_{V}-\bar{\Omega}_{V}\right);\quad\bar{Y}_{V}=0.$ (46) Upon substituting the first equation of (46) into (45) gives $\displaystyle 0=-\frac{\bar{k}^{2}}{r^{2}}\mathcal{A}_{V}+2Y_{V}.$ (47) Evoking the vanishing of the Greenberg vector, (46) is $r^{2}Y_{V}+\bar{k}^{2}\alpha_{V}=0$, so that (47) reduces to the condition $\displaystyle\mathcal{A}_{V}+2\alpha_{V}=0.$ (48) This is the required condition that preserves the null character of the null horizon $r=2$ under linear perturbation. Were one to fix the frame by setting $\mathcal{A}^{a}=0$ (so that $n^{a}$ points along $\dot{u}^{a}$), we will have that the condition (48) now becomes $\alpha_{V}=0$. Otherwise, the horizon does not retain its null character. On the other hand, by (48), the vanishing condition $\Sigma_{V}-\bar{\Omega}_{V}=0$ would lead to the nullity condition fixing the frame: $\mathcal{A}_{V}=0$. FOTH: Let us consider the FOTH condition, $\dot{\chi}<0$ and $\hat{\chi}>0$. If one fixes the frame by setting $\mathcal{A}^{a}=0$, the FOTH condition trivially follows: Explicitly, the FOTH condition is the pair $\displaystyle\mathcal{E}$ $\displaystyle<-\delta_{a}\alpha^{a},$ (49) $\displaystyle\mathcal{E}$ $\displaystyle<\delta_{a}a^{a}+\delta_{a}\alpha^{a},$ (50) where we have used the fact that the Greenberg vector must vanish. So, again, how the black hole evolves under linear perturbation depends crucially on how the frame is choosen, as one would expect. Again, $a^{a}$ must vanish for our MOTS to be a genuine surface. As the specification of the frame choice $\mathcal{A}^{a}=0$ gives $\alpha_{V}=0$, the FOTH condition is simply the requirement that $\mathcal{E}<0$, which always holds. #### IV.1.2 Dynamical case Let us now consider the case where the $r=2m$ null horizon possibly evolves to acquire a different character. (Again for simplicity let us make the same frame choice here by setting $\mathcal{A}^{a}=0$.) Then, with all considerations of the vanishing Greenberg vector, as well as the necessity of the vanishing of $a^{a}$, the characterizing norm $z$ takes the form $\displaystyle z=-12\frac{\bar{k}^{2}}{r}\mathcal{E}\left(\Sigma_{V}-\bar{\Omega}_{V}\right).$ (51) We note that from the field equations the vectors $\Sigma_{a}$ and $\Omega_{a}$ do not evolve independently [16, 17], and so the presence of one ensures the non-vanishing of the other. Hence, we see that as long as there is a nonvanishing shear, $z\neq 0$ for $\bar{k}^{2}\neq 0$, and if there is no shear the horizon stays null as was seen in the previous case. We do expect that under the perturbation and in the presence of a nonzero shear the Schwarzschild event horizon becomes spacelike. This then imposes that $\displaystyle\Sigma_{V}-\bar{\Omega}_{V}<0,$ (52) thereby imposing a constraint on the solution to the harmonically decomposed linearized system, when restricted to the horizon. We summarize the above discussion as follows: Let $\tilde{M}$ be the spacetime obtained by linearly perturbing a Schwarzschild background. Then the $r=2$ horizon is null in $\tilde{M}$ if and only if the shear and vorticity 2-vectors obey $\Sigma_{V}-\bar{\Omega}_{V}=0$, but otherwise spacelike in $\tilde{M}$ with $\Sigma_{V}-\bar{\Omega}_{V}<0$. Note: A nondissipative perturbation is assumed here. The above conclusion drawn therefore may or may not be true in the presence of a nonzero $Q$ and $\Pi$, depending on their relative magnitudes. ### IV.2 Spatially homogeneous background We consider a hypersurface orthogonal LRS II solution, which is spatially homogeneous, with metric $\displaystyle ds^{2}=-dt^{2}+\bar{a}^{2}_{1}(t)dr^{2}+\bar{a}^{2}_{2}(t)d\bar{\varsigma}^{2}.$ (53) (The 2-surface with metric $d\bar{\varsigma}^{2}$ is allowed different geometries including a toroidal one.) In this case we have that $\mathcal{A}=\phi=0$, with the non-vanishing background kinematic quantities being the expansion and shear: $\displaystyle\theta=\frac{\bar{a}_{1t}}{\bar{a}_{1}}+2\frac{\bar{a}_{2t}}{\bar{a}_{2}};\qquad\sigma=\frac{2}{3}\left(\frac{\bar{a}_{1t}}{\bar{a}_{1}}-\frac{\bar{a}_{2t}}{\bar{a}_{2}}\right),$ (54) where the underscore ‘$t$’ indicates partial differentiation. Hence, $\displaystyle\chi_{0}=\frac{2}{3}\theta-\sigma\qquad\mbox{and}\qquad\chi_{1}=\phi.$ Indeed, $\dot{\chi}_{1}$ and $\hat{\chi}_{0}$ are first order quantities so that $\hat{\chi}_{0}^{2}=\hat{\chi}_{0}\hat{\chi}_{1}=0$. For simplicity, we specify to a conformally flat perturbation, and set $\Lambda=0$. We further restrict to the nondissipative case. This then gives the norm $z$ as (we recall that this is always evaluated at $\chi=0$) $\displaystyle z$ $\displaystyle=F-2\dot{\chi}_{0}\dot{\chi}_{1}$ $\displaystyle=\frac{1}{3}\left(\rho-3p\right)\left[\left(\rho+p\right)-\delta_{a}Z^{a}\right]-\frac{4}{3}\rho\delta_{a}\alpha^{a}.$ (55) Also, the vanishing condition (31) becomes (again evaluating at $\chi=0$) $\displaystyle 0=\dot{\chi}_{0}\delta_{a}\dot{\chi}$ $\displaystyle=-\frac{1}{3}\left(\rho+3p\right)\delta_{a}\left(\delta_{b}\mathcal{A}^{b}+\delta_{b}\alpha^{b}\right)$ $\displaystyle=\frac{\bar{k}^{2}}{3r^{2}}\left(\rho+3p\right)\left(\mathcal{A}_{V}+\alpha_{V}\right)\mathcal{Q}_{a}.$ (56) So, either we fix the equation of state $\rho+3p=0$ or $\mathcal{A}_{V}+\alpha_{V}=0$. Now, for the former case this would require that $\rho$ (and consequently $p$) vanishes in the background to ensure the gauge invariance of $z$, in which case the horizon is always null. Let us consider the latter case, and consider the particular frame choice $\mathcal{A}^{a}=0$. This would impose that $\alpha_{V}=0$. (Note again that this condition is compatible with $\Sigma_{V}-\bar{\Omega}_{V}=0$ by virtue of the vanishing of the Greenberg vector.) Therefore, the characterizing function $z$ reduces to $\displaystyle z=\frac{1}{3}\left(\rho-3p\right)\left(\rho+p\right).$ (57) Hence, assuming the weak energy condition, the character of a null horizon, evolving under linear perturbation, behaves according to the sign of $\rho-3p$: the horizon stays null, or becomes spacelike or timelike provided $\rho=3p,\rho<3p$ or $\rho>3p$. So, for example, with a linear equation of state (EoS) $p=w\rho$, the above conditions are the respective requirements $w=0,w>0$ and $w<0$. Of course, a choice different from $\mathcal{A}^{a}=0$ may be used to fix the frame. In this case, we will have $z$ assuming the form $\displaystyle z=\frac{1}{3}\left[\left(\rho-3p\right)\left(\rho+p\right)+\frac{\bar{k}^{2}}{r}\left(5\rho-3p\right)\alpha_{V}\right].$ (58) We see that even with a linear EoS of the form $p=w\rho$, the signature of $z$ above acquires some complexity an carries a crucial dependence on sign of the component $\alpha_{V}$ (Note that $\alpha_{V}=\Sigma_{V}-\bar{\Omega}_{V}$). We may however still comment on some possible characters of the perturbed horizon (we are assuming a positive $\rho$). For example, it is not difficult to see that for the range $\displaystyle-1\leq w\leq\frac{1}{3},$ (59) $\alpha_{V}>0$ precludes a spacelike or null character, allowing for only a timelike character. This indeed differs from the timelike characterization of the background as the extremes $w=-1$ and $w=1/3$ allow only for the null character. (See for example the references [29, 19].) If we look outside the bound (59) in the case $\alpha_{V}>0$, for $w<-1$, the horizon character will depend on the magnitude of $\alpha_{V}$ at each $r$ in relation to each $\bar{k}^{2}$. (In the background, this is quite straightforward as outside the bound (59), the horizon is spacelike. Again, see the references [29, 19].) This will also be the case for the bound $\displaystyle\frac{1}{3}<w<\frac{5}{3},$ (60) but for $w\geq 5/3$, only the timelike character is possible. Now consider the case for $\alpha_{V}<0$. Then, a timelike or null character is ruled out within the bound (60), allowing for only a spacelike character. For the bounds $\displaystyle-1<w<\frac{1}{3}\qquad\mbox{and}\qquad w>\frac{5}{3},$ (61) the character again depends on the magnitude of $\alpha_{V}$ at each $r$ in relation to each $\bar{k}^{2}$. At the extreme $w=1/3$ the horizon is spacelike and timelike for $\displaystyle\frac{1}{3}<w\leq\frac{5}{3},$ (62) And for values $w<-1$, a timelike or null character is prohibited. (Notice that at the critical point $w=5/3$ the horizon signature is insensitive to the sign of $\alpha_{V}$.) We summarize the above results as follows: Let $\tilde{M}$ be the spacetime obtained by linearly perturbing a spatially homogeneous background $M$ with EoS $p=w\rho$, and let $\mathcal{T}$ be a null horizon foliated by MOTS in $M$. Then for $\Sigma_{V}-\bar{\Omega}_{V}=0$, $\mathcal{T}$ is null, spacelike or timelike in $\tilde{M}$ provided $w=0,w>0$ or $w<0$, respectively. Otherwise, for $\Sigma_{V}-\bar{\Omega}_{V}\neq 0$, see Table 1. (Note that $\alpha_{V}=\Sigma_{V}-\bar{\Omega}_{V}$ on the MOTS.) Null horizon character under linear perturbation --- $\alpha_{V}$ | $w$ | Character | $(-1\leq w\leq\frac{1}{3})\cup(w\geq\frac{5}{3})$ | Timelike $\alpha_{V}>0$ | | | $(\frac{1}{3}<w<\frac{5}{3})\cup(w<-1)$ | – | $(-1<w<\frac{1}{3})\cup(w>\frac{5}{3})$ | – $\alpha_{V}<0$ | $\frac{1}{3}<w\leq\frac{5}{3}$ | Timelike | $(w\leq-1)\cup(w=\frac{1}{3})$ | Spacelike Table 1: The behavior of a null horizon evolving under linear perturbation. Assuming a linear equation of state $p=w\rho$, the above table shows the character of the horizon for different ranges of the equation of state parameter $w$, and the relationship with the sheet vector $\alpha^{a}$ decomposed harmonically. The ‘$-$’ indicates a non-trivial relationship which is dependent on the magnitudes of $\alpha_{V}$ and $\rho$ as well as the value of the nonnegative integer $\bar{k}^{2}$. Let it be emphasized that the linear choice of the EoS is for demonstrable purposes for our particular application and there is no restriction here. Any EoS is in principle allowable with (possible) caveat being the the analysis acquiring a more complicated character. We will now conclude this section with the following remark: * Remark 3. In a general inhomogeneous case, $\chi_{1}=0$ and we see that the associated norm is simplified: $\displaystyle z=F,$ so that the characterization proceeds from (III.2). However, the consideration of the $\delta_{a}z=0$ criterion is now significantly more complicated. Relaxing this condition of course then gives a simple pointwise characterization via (III.2). If one does not impose $\delta_{a}z=0$ a priori, this allows for the possibility of varying character across the horizon. This is clearly a more involving case and this is worth considering for a subsequent work. However, as was just mentioned, while generally these cases are expected to pose problems, we see that in a general inhomogeneous background the condition provides a certain analytic simplicity, at least point-wise. ## V On the MOTS stabiliity operator As have been seen from the previous section, the 2-vectors associated to the shear and vorticity play a crucial role in how a null horizon in a LRS background spacetime evolves under a linear perturbation. This particular combination of the shear and vorticity evolves simultaneously according to the field equations and have no independent evolution or propagation equations as was mentioned in the previous section. This relationship has implications for the self-adjointness of the MOTS stability operator as these same 2-vectors also control the self-adjointness of the operator. We will begin by introducing the MOTS stability operator after which we provide some commentary on this relationship. We will then delve into several properties of the operator in context of our current considerations as it relates to self- adjointness. The evolution, propagation and sheet equations that will be utilized here have been obtained in [16, 17] (with corrections in [22, 23] as stated in Section III). So, especially the reference [16] is implicit wherever these equations appear. ### V.1 The MOTS stability operator Given a MOTS $\mathcal{S}$ and a normal foliation by a collection $\mathcal{S}_{v}$, with the subscript ‘$v$’ labelling the foliation, one may define the tangent vector to curves generating the deformation as $\displaystyle\partial_{v}=\bar{\psi}n^{a},$ for some smooth function $\bar{\psi}$ on $\mathcal{S}$. Then, stability of $\mathcal{S}$ is captured via the variation of the expansion $\chi$ along $\partial_{v}$ (several discussions of this can be found in the original papers [6, 7], as well as, for example, the works [30, 31] and references therein): $\displaystyle\bar{\delta}_{\bar{\psi}n}\chi=L_{\mathcal{S}}\bar{\psi},$ where the operator $L_{\mathcal{S}}$ is $\displaystyle L_{\mathcal{S}}=-\delta^{2}+2\tilde{s}^{a}\delta_{a}+\bar{F},$ (63) with the definition $\displaystyle\bar{F}=\frac{1}{2}\mathcal{R}_{\mathcal{S}}-\tilde{s}_{a}\tilde{s}^{a}+\delta_{a}s^{a}-G_{ab}k^{a}l^{b},$ (64) (It is important to note that the scalar $\bar{\psi}$ completely characterizes the variation of $\mathcal{S}$.) The one-form $\tilde{s}_{a}=-N^{c}_{\ a}l_{b}\nabla_{c}k^{b}$ is the projection of the torsion of $k^{a}$ onto $\mathcal{S}$, $G_{ab}$ is the Einstein tensor and $\mathcal{R}_{\mathcal{S}}$ is the scalar curvature of $\mathcal{S}$. $\mathcal{S}$ is said to be strictly or marginally stable if there exists a nonnegative (and not identically zero) $\bar{\psi}$ obeying respectively $L_{\mathcal{S}}\bar{\psi}>0$ or $L_{\mathcal{S}}\bar{\psi}=0$, and unstable otherwise. For a 1+1+2 decomposed spacetime, $\tilde{s}_{a}$ is explicitly $\displaystyle\tilde{s}_{a}=\Sigma_{a}-\varepsilon_{ab}\Omega^{b},$ (65) And in the harmonic basis, $\displaystyle\tilde{s}_{a}=\underbrace{\left(\Sigma_{V}+\bar{\Omega}_{V}\right)}_{\text{$\tilde{s}_{V}$}}\mathcal{Q}_{a}+\underbrace{\left(\bar{\Sigma}_{V}-\Omega_{V}\right)}_{\text{$\bar{\tilde{s}}_{V}$}}\bar{\mathcal{Q}}_{a}.$ (66) In general the operator $L_{\mathcal{S}}$ is not self-adjoint, which results from the presence of the linear term $\tilde{s}^{a}\delta_{a}$ so that $L_{\mathcal{S}}$ is allowed to have complex eigenvalues. The imaginary part encodes information about rotation through the Komar angular momentum, obtained by integrating over the MOTS the inner product of $\tilde{s}_{a}$ and an axial Killing field. Still, the principal eigenvalue, which we denote as $\bar{\lambda}$, is always real [6]. The stability of a MOTS $\mathcal{S}$ then reduces to the sign of $\bar{\lambda}$: $\mathcal{S}$ is strictly or marginally stable if $\bar{\lambda}>0$ or $\bar{\lambda}=0$, and unstable otherwise. Here, we are interested in the implications of the characterization conditions obtained for the backgrounds in the previous sections, for the self- adjointness of the stability operator. Crucially, the dependence on the sign of $\alpha_{V}$ (or equivalently $\Sigma_{V}-\bar{\Omega}_{V}$). Now, it is known that the self-adjointness of the MOTS stability operator is guaranteed by the one-form $\tilde{s}_{a}$ being a gradient [31]. Particular, if the 1-form $\tilde{s}_{a}$ is gradient, then the operator $L_{\mathcal{S}}$ is similar to a self-adjoint operator. (Of course, the case of vanishing $\tilde{s}_{a}$ is trivial, and so we consider this possibility $\tilde{s}_{a}\neq 0$.) This self-adjointness in the case that $\tilde{s}_{a}$ is gradient can be seen here in a rather trivial way. To see this, let us consider the action of the operator $L_{\mathcal{S}}$ on a first order scalar $\psi^{(1)}=\psi_{S}\mathcal{Q}$: $\displaystyle L_{\mathcal{S}}y^{(1)}=\left[\left(-\delta^{2}+\bar{F}\right)\mathcal{Q}+\frac{1}{r}\tilde{s}_{V}\right]\psi_{S}.$ (67) (The normalization $\mathcal{Q}_{a}\mathcal{Q}^{a}=1$ has been assumed for simplicity.) Since if the $\tilde{s}_{a}$ is gradient it has a vanishing divergence, it is then clear that this requires the vanishing of $\tilde{s}_{V}$, i.e. $\displaystyle\Sigma_{V}+\bar{\Omega}_{V}=0.$ (68) Thus, the eigenvalue problem for the MOTS stability operator simply reduces to the following eigenvalue problem for $\mathcal{Q}$: $\displaystyle-\delta^{2}\mathcal{Q}+F\mathcal{Q}=\lambda\mathcal{Q}.$ (69) That is, the problem reduces to the eigenvalue problem for the self-adjoint operator $L^{\prime}_{\mathcal{S}}\mathcal{Q}=\left(-\delta^{2}+F\right)\mathcal{Q}=\lambda\mathcal{Q}$ so that $\displaystyle\lambda_{\bar{k}^{2}}=\frac{\bar{k}^{2}}{r}+\bar{F},$ (70) i.e. the eigenvalues are parametrized by $\bar{k}^{2}$. Thus, strict stability is always guaranteed for $\displaystyle\bar{F}=\frac{1}{2}\left(\mathcal{R}_{\mathcal{S}}-R\right)-\bar{\tilde{s}}_{V}^{2}-\left(p+Q\right)+2\Lambda>0.$ (71) (recall that $R$ is the four dimensional scalar curvature.) It will later be seen that $\bar{\tilde{s}}_{V}$ is necessarily zero on the MOTS in the case that $\tilde{s}_{V}=0$, i.e. $\tilde{s}_{a}$ is a gradient. So, in this case for $p+Q\leq 0$ and $\Lambda\geq 0$, whenever $\mathcal{R}_{\mathcal{S}}>R$, the operator has no negative eigenvalue. The principal eigenvalue and the MOTS is strictly stable. This is indeed the case, say, for vacuum solutions undergoing a shear-free and a nondissipative perturbation where one will have $\tilde{s}_{V}$ vanishing. On the other hand, consider an irradiating ($Q=0$) spacetime with a positive scalar curvature, a vanishing $\Lambda$ and a nonnegative pressure. Then, for a sufficiently large $p$ it is possible that $\bar{F}<0$, so that the principal eigenvalue is always negative and hence the MOTS will be unstable. * Remark 4. Actually, in the linear regime any perturbation to the MOTS scalar curvature $\mathcal{R}_{\mathcal{S}}$ is encoded in first order scalars. Particularly, if the scalars $\mathcal{E},\Pi$, and $\Lambda$ (if $\Lambda$ is included as a perturbation variable) are nonvanishing in the background, then $\mathcal{R}_{\mathcal{S}}$ is not perturbed. However, nonlinear effects are present under perturbation of the background as $\displaystyle\mathcal{R}_{\mathcal{S}}=\ ^{0}\mathcal{R}_{\mathcal{S}}+\mathcal{Z},$ (72) where $\mathcal{Z}=\Sigma_{ab}\Sigma^{ab}-\zeta_{ab}\zeta^{ab}$ is just twice the inner product of the projected shears of $k^{a}$ and $l^{a}$. So, it is clear that once nonlinear effects are factored in any analysis, we do expect the behavior of the eigenvalues to be modified. Now, the commutation relations of the dot and hat derivatives for an arbitrary 2-vector $\Psi_{a}$ (imposing the vanishing of the Greenberg vector and taken up to linear order) $\displaystyle\hat{\dot{\Psi}}_{a}-\dot{\hat{\Psi}}_{a}=-\mathcal{A}\dot{\Psi}_{a}+\left(\frac{1}{3}\theta+\sigma\right)\hat{\Psi}_{a}+\mathcal{H}\varepsilon_{ab}\Psi^{b}.$ (73) Harmonically decomposing the above relation and applying to the one-form $\tilde{s}_{a}$ while setting the component $\tilde{s}_{V}=0$ says that $\displaystyle\hat{\dot{\bar{\tilde{s}}}}_{V}-\dot{\hat{\bar{\tilde{s}}}}_{V}$ $\displaystyle=-\mathcal{A}\dot{\bar{\tilde{s}}}_{V}+\left(\frac{1}{3}\theta+\sigma\right)\hat{\bar{\tilde{s}}}_{V},$ (74) $\displaystyle\mathcal{H}\bar{\tilde{s}}_{V}$ $\displaystyle=0.$ (75) The above is meant to suggest that for a magnetized ($\mathcal{H}\neq 0$) linear perturbation $\tilde{M}$ of a LRS background spacetime $M$, if the 1-form $\tilde{s}_{a}$ is gradient in which case the MOTS stability operator is self-adjoint (remember this is for the class of MOTS under consideration), then $\tilde{s}_{a}$ necessarily vanishes on the MOTS. But as it will be seen shortly, the quantities $\bar{\tilde{s}}_{V}$ and $\mathcal{H}$ are proportional so that in the linear regime on any MOTS it is true that both $\bar{\tilde{s}}_{V}$ and $\mathcal{H}$ vanish. More specifically, $\mathcal{H}$ must be first order. ### V.2 Comments on relations to the characterization of null horizons Let us now specify our discussions to the backgrounds of the previous section. #### V.2.1 Schwarzschild background To begin with, consider the case of the Schwarzschild background. We will establish the following statement (C1) and the corresponding implication (C2): 1. (C1). For a linear perturbation $\tilde{M}$ of a Schwarzschild background, the 1-form $\tilde{s}_{a}$ is gradient (in which case the MOTS stability operator is self-adjoint) if and only if $\tilde{s}_{a}$ is identically zero; $\implies$ 2. (C2). For a linear perturbation $\tilde{M}$ of a Schwarzschild background, if the one-form $\tilde{s}_{a}\neq 0$ and the vorticity 2-vector has a nonzero contribution from the odd sector, the $r=2$ null horizon necessarily evolves to a spacelike character. Note: As was mentioned in the previous section, away from the null case the horizon can only evolve to the spacelike case with the even parity component of $\Sigma_{a}+\varepsilon_{ab}\Omega^{a}$ being negative. (A timelike character would imply that timelike curves do enter the trapped region, which is of course ruled out.) Thus, the second statement above, (C2), follows from the fact that $\tilde{s}_{V},\bar{\Omega}_{V}\neq 0$ ensures that $\Sigma_{V}-\bar{\Omega}_{V}\neq 0$. Let us apply the tensor $\varepsilon^{ab}n^{c}$ to the Ricci identities for the unit field $n^{a}$ to get $\displaystyle\delta_{a}\left(\Omega^{a}+\varepsilon^{ab}\Sigma_{b}\right)=\left(2\mathcal{A}-\phi\right)\Omega+\mathcal{H},$ (76) and upon noting $\Omega$ should be zero on the MOTS, the above reduces to $\displaystyle-\frac{\bar{k}^{2}}{r}\bar{\tilde{s}}_{V}\mathcal{Q}=\mathcal{H},$ (77) so that $\mathcal{H}$ is a first order quantity that vanishes on the background, i.e. $\mathcal{H}$ is decomposed into the harmonic $\mathcal{Q}$ basis. Then, by (75) it follows that $\mathcal{H}=\bar{\tilde{s}}_{V}=0$. It therefore follows that $\tilde{s}_{V}=0\iff\tilde{s}_{a}=0$, i.e. $\tilde{s}_{a}$ is gradient if and only if $\tilde{s}_{a}$ vanishes identically. Finally, the linearized evolution and propagation equations of $\mathcal{H}$ are respectively given by (these are respectively the $u^{a}$ and $n^{a}$ components of the magnetic Weyl divergence equation [26]) $\displaystyle\dot{\mathcal{H}}$ $\displaystyle=-3\mathcal{E}\xi-\varepsilon_{ab}\delta^{a}\mathcal{E}^{b},$ (78) $\displaystyle\hat{\mathcal{H}}$ $\displaystyle=-3\mathcal{E}\Omega-3\phi\mathcal{H}-\delta_{a}\mathcal{H}^{a},$ (79) which translates to, by the vanishing of $\mathcal{H}$ as well as noting that $\Omega=\xi=0$ on the MOTS, $\displaystyle\mathcal{H}_{V}=0;\quad\bar{\mathcal{E}}_{V}=0.$ (80) This is to say that, necessarily, on the MOTS one requires that $\mathcal{H}^{a}$ is solenoidal (divergence-free) and that $\mathcal{E}^{a}$ has no odd parity contribution on the MOTS. While there appears to be freedom in the quantities $\mathcal{E}_{V}$ and $\bar{\mathcal{H}}_{V}$, they are actually constrained by a coupling of the sheet expansion $\phi$ and the 1-form $\tilde{s}_{a}$ (see the discussions of the next subsection). For example, for a vanishing $\tilde{s}_{a}$ in the case of a nondissipative perturbation, it follows that one requires (again, see the next subsection for more details on this) $\displaystyle\mathcal{E}_{V}+\bar{\mathcal{H}}_{V}=0.$ Of course, we do not rule out self-adjointness of the operator in the case of a spacelike character. This is in principle possible depending on the nature of the perturbation. This case will however impose that $\Sigma_{V}<0$. On the other hand the null character does not necessarily imply the self- adjointness of the operator. The self-adjointness due to vanishing of $\tilde{s}_{a}$, in this case, would imply (and is implied by) $\Sigma_{V}=\bar{\Omega}_{V}=0$. #### V.2.2 Hypersurface orthogonal and spatially homogeneous background We now consider the spatially homogeneous case of the previous section, and stick with a nondissipative perturbation. In this case the equations corresponding to (78), (79) and (76) are respectively $\displaystyle\dot{\mathcal{H}}$ $\displaystyle=-\frac{3}{2}\left(\frac{2}{3}\theta-\sigma\right)\mathcal{H}-3\mathcal{E}\xi-\varepsilon_{ab}\delta^{a}\mathcal{E}^{b},$ (81) $\displaystyle\hat{\mathcal{H}}$ $\displaystyle=-3\left(\mathcal{E}+\rho+p\right)\Omega-3\phi\mathcal{H}-\delta_{a}\mathcal{H}^{a},$ (82) $\displaystyle\mathcal{H}$ $\displaystyle=\delta_{a}\left(\Omega^{a}+\varepsilon^{ab}\Sigma_{b}\right)+3\xi\sigma.$ (83) The vanishing of $\Omega,\xi=0$ and $\mathcal{H}$ then implies the result $\mathcal{H}_{V},\bar{\mathcal{E}}_{V},\bar{\tilde{s}}_{V}=0$ follows as in the Schwarzschild case. For this reason, the statement (C1) of the previous subsection also holds here, although for obvious reasons the statement (C2) does not hold. * Remark 5. We do expect these three vanishing properties to be generic to the MOTS we are considering in the sense that they are independent of the particular background one works with. It will be quite interesting to see how much these vanishing conditions further constrains the MOTS in full generality. This is however beyond the scope of the current work. In the self-adjoint case with a vanishing $\tilde{s}_{a}$, obviously either one of the components $\Sigma_{V}$ or $\bar{\Omega}_{V}$, paired with the parameter ranges suffice for the characterization of the horizon. On the other hand, if one is to impose the condition that $|\Sigma_{V}|\neq|\bar{\Omega}_{V}|$, this of course ensures that $\tilde{s}_{a}$ is nonvanishing, in which case one may then check which conditions this further impose on the horizon character. However, this does not in by itself rule out self-adjointness of the operator. In principle, if one can expand a nonvanishing $\tilde{s}_{V}$ in the harmonic $\mathcal{Q}$ basis, then the operator can still be brought to a self-adjoint form. In the forthcoming subsection we will see a particular class of perturbations where this may be possible. ### V.3 More on the MOTS and stability operator We obtain some additional results required to be satisfied on the MOTS, given properties derived from the one-form $\tilde{s}_{a}$. As we shall see, this will provide some interesting insights into properties of the 1-form $\tilde{s}_{a}$. The magnetic Weyl tensor is written down as a constraint to the field equation [26]: $\displaystyle H_{ab}=\varepsilon_{cd\langle a}D^{c}\sigma_{b\rangle}^{\ d}-2\dot{u}_{\langle a}\omega_{b\rangle}-D_{\langle a}\omega_{b\rangle}.$ (84) One may interpret the above as the magnetic Weyl tensor measuring the degree of distortion of vorticity. Taking the $n^{a}$ component of (84) produces the constraint equation, up to linear order, $\displaystyle-\delta_{a}\left(\frac{2}{3}\theta-\sigma\right)$ $\displaystyle=-2\delta^{b}\Sigma_{ab}-\phi\tilde{s}_{a}$ $\displaystyle-2\varepsilon_{ab}\left[\mathcal{H}^{b}+\left(\delta^{b}+2\mathcal{A}^{b}\right)\Omega\right]$ $\displaystyle+\xi\left(\Omega_{a}-3\varepsilon_{ab}\Sigma^{b}\right).$ (85) By setting $\Omega=\xi=0$ we have $\displaystyle-\delta_{a}\left(\frac{2}{3}\theta-\sigma\right)=-2\left(\delta^{b}\Sigma_{ab}+\varepsilon_{ab}\mathcal{H}^{b}\right)-\phi\tilde{s}_{a}.$ (86) Also, fully projecting the Ricci identities for $n^{a}$ taken up to linear order and imposing the MOTS condition and the vanishing of $\xi$ and $\Omega$ as required we obtain $\displaystyle-\delta_{a}\phi=2\left(\mathcal{E}_{a}-\delta^{b}\zeta_{ab}\right)-\Pi_{a}-\phi\tilde{s}_{a},$ (87) and upon combining with (86) and (87) we have $\displaystyle-\cancel{\delta}_{a}\chi=2\left[\mathcal{E}_{a}-\delta^{b}\left(\Sigma_{ab}+\zeta_{ab}\right)-\varepsilon_{ab}\mathcal{H}^{b}\right]-\phi\tilde{s}_{a},$ (88) where we have written $\cancel{\delta}_{a}=\delta_{a}-\tilde{s}_{a}$. One may then use the Codazzi equations to establish the following $\displaystyle- Q_{a}+Z_{ab}k^{b}=2\left(\mathcal{E}_{a}-\varepsilon_{ab}\mathcal{H}^{b}\right)-\phi\tilde{s}_{a},$ (89) where $Z_{ab}=N^{\ c}_{a}C_{cbde}k^{d}l^{e}$, with $C_{abcd}$ being the spacetime Weyl tensor. We will now utilize a curvature property of the MOTS to establish that the second term on the left hand side of (89) is of nonlinear order on the MOTS. With this we will have established a relationship between the Weyl and heat flux 2-vectors and the rotation 1-form. The curvature 2-form on the MOTS, i.e. the curvature of $\tilde{s}_{a}$ – which we denote by $\tilde{\Omega}_{ab}$ – contracted along the null direction $k^{a}$ can be shown to obey the relation $\displaystyle k^{b}\tilde{\Omega}_{ba}=-N^{\ c}_{a}Z_{cb}k^{b},$ so that $\displaystyle k^{b}\tilde{\Omega}_{ba}=-\left[2\left(\mathcal{E}_{a}-\varepsilon_{ab}\mathcal{H}^{b}\right)-\phi\tilde{s}_{a}+Q_{a}\right].$ (90) Explicitly, the left hand side of (90) can be reduced to $\displaystyle k^{b}\tilde{\Omega}_{ba}$ $\displaystyle=-\frac{1}{2}\Sigma_{b}\delta_{a}k^{b}$ $\displaystyle=-\frac{1}{2}\left(\Sigma_{ab}+\zeta_{ab}\right)\Sigma^{b},$ which we ignore as it is of second order. Thus, $\displaystyle\phi\tilde{s}_{a}=Q_{a}+2\left(\mathcal{E}_{a}-\varepsilon_{ab}\mathcal{H}^{b}\right).$ (91) Furthermore, imposing the MOTS condition on (29) we have $\displaystyle\cancel{\delta}_{a}\phi=0,$ so that by (87) it follows $\displaystyle\Pi_{a}=2\left(\mathcal{E}_{a}-\delta^{b}\zeta_{ab}\right).$ (92) Clearly, away from the case of a vanishing $\phi$, the following statement is concluded: The MOTS stability operator in a conformally flat and nondissipative (this imposes that $\zeta_{ab}$ is divergence-free) linearly perturbed inhomogeneous LRS background is self-adjoint. So for example, the above will be true for a nondissipative perturbation of a Lemaitre-Tolman-Bondi type solutions. Notice that if the tensor $\delta_{a}n_{b}$ is pure trace, dissipation on the MOTS is generated entirely by contributions from the Weyl tensor. On the other hand, without restricting to the nondissipative case, using the vanishing of $\bar{\mathcal{E}}_{V},\mathcal{H}_{V}$ and $\bar{\tilde{s}}_{V}$ we have the following pair of equations from (91). $\displaystyle\bar{Q}_{V}$ $\displaystyle=0,$ (93) $\displaystyle\phi\tilde{s}_{V}$ $\displaystyle=Q_{V}+2\left(\mathcal{E}_{V}+\bar{\mathcal{H}}_{V}\right).$ (94) In any case, we can draw some immediate conclusions from both (91) and (92) about the generation of rotation on a horizon: * • In a conformally flat linearly perturbed inhomogeneous LRS background, rotation of a horizon foliated by MOTS can be entirely sourced by the heat flux 2-vector along the MOTS. This then implies that self-adjointness of the MOTS stability operator in this particular case may be obtained by switching off any heat flux contribution along the MOTS. * • In a nondissipative linearly perturbed inhomogeneous LRS background, rotation of a horizon foliated by MOTS can be sourced by the 2-vector components of the Weyl curvature tensor. This is to say that in this case the contributing sources are tidal forces acting along the MOTS. * Remark 6. We remark here that the form of (94) suggests that it is possible, in principle, to have a nonvanishing $\tilde{s}_{V}$ – i.e. a nonvanishing $\tilde{s}_{a}$ – even in the case of a conformally flat perturbation with a vanishing heat flux and vanishing $\phi$. Thus, in the first bulleted point above we were quite cautious with semantics when we used “may be” when switching off the heat flux contribution along the MOTS to obtain self- adjointness of the stability operator. Only when $\phi$ is nonzero in the perturbed spacetime that this is true, at least in our current context. Finally, let us describe a simple case where $\tilde{s}_{V}$ is non-vanishing but the stability operator may take a self-adjoint form. More specifically, we may be able to express $\tilde{s}_{V}$ in terms of the $\mathcal{Q}$ basis. We shall consider a static background, i.e. the shear and expansion scalars vanish on the background. We note that for any 2-tensor $\Psi_{ab}$ one has as the divergence $\displaystyle\delta^{b}\Psi_{ab}=\frac{1}{r}\left(\bar{k}^{2}-2\right)\left(-\bar{\Psi}_{T}\mathcal{Q}_{a}+\Psi_{T}\bar{\mathcal{Q}}_{a}\right).$ (95) Thus, one can harmonically decompose (86) to obtain the pair of constraint equations $\displaystyle 0$ $\displaystyle=\frac{1}{r}\left(\bar{k}^{2}-2\right)\Sigma_{T}+\mathcal{H}_{V},$ (96) $\displaystyle 0$ $\displaystyle=\frac{1}{r}\left(\bar{k}^{2}-2\right)\bar{\Sigma}_{T}+\bar{\mathcal{H}}_{V}$ $\displaystyle+\frac{1}{2r}\left(\frac{2}{3}\theta_{S}-\sigma_{S}\right)-\phi\tilde{s}_{V}.$ (97) Indeed, from the first equation (96) it follows that the tensorial part $\Sigma_{T}$, for $\bar{k}^{2}\neq 2$, vanishes by virtue of the vanishing of the quantity $\mathcal{H}_{V}$. Furthermore, by imposing the MOTS condition $\chi=0$, the second equation (97) can be recast as $\displaystyle 0=\frac{1}{r}\left(\bar{k}^{2}-2\right)\bar{\Sigma}_{T}+\bar{\mathcal{H}}_{V}+\phi\left(\frac{1}{2r\mathcal{Q}}-\tilde{s}_{V}\right).$ (98) Therefore, as long as the MOTS is nonminimal in the perturbed spacetime (recall that by nonminimal it is meant that $\phi$ and $(2/3)\theta-\sigma$ cannot vanish simultaneously on the MOTS), for a perturbation with no contributions from the odd vector harmonics part of $\mathcal{H}_{a}$ and the shear 2-tensor $\Sigma_{ab}$, the 1-form component $\tilde{s}_{V}$ can effectively be cast as $\displaystyle\tilde{s}_{V}=\left(\frac{1}{2r\mathcal{Q}^{2}}\right)\mathcal{Q}.$ (99) (Of course $\tilde{s}_{V}$ is nonzero and so $\tilde{s}_{a}$ is not a gradient.) This then allows us to write down the eigenvalues of the operator as $\displaystyle\lambda_{\bar{k}^{2}}=\frac{\bar{k}^{2}}{r}+\frac{1}{2r^{2}\mathcal{Q}^{2}}+\bar{F}.$ (100) Consequently the positivity of the modification term means that $\bar{F}>0$ again ensures nonnegativity of the eigenvalues. ## VI Summary and outlook Summary – In this work we have considered the behavior of a null horizon foliated by MOTS in a LRS background spacetime subjected to linear perturbations. The problem is fixed by the gauge invariant requirement of the characterizing function. More precisely, we set out to characterize the evolution of a MOTS in a linearized spacetime admitting a 1+1+2 decomposition, which lies in the slice orthogonal to the $u^{a}$. The gauge invariance of the chacterizing function introduced is possible if and only if a certain function vanishes in the background. This vanishing condition is equivalent to the condition for a horizon in the background to be null, thus phrasing the problem in terms of the evolutionary behavior of a null horizon in a LRS background subjected to linear perturbations. In other words – and emphasis is placed on this – this characterization fails in the case of a non-null horizon in the background. While this work has been dedicated to the case of linear perturbations, it was demonstrated that the characterizing function is form invariant even when extending to non-linear regimes, for a certain frame choice or provided that the perturbed spacetime has a vanishing sheet expansion. We focused the application of our characterization to non-dissipative perturbations. As was described in the introduction, perturbing a seed metric generally affects the dynamics of an embedded horizon, and that this is the case of the null event horizon in the Schwarzschild spacetime. The presence of shear ensures a spacelike transition of the null horizon. We have assumed a harmonic decomposition of first order scalars $\Psi$ and 2-vectors $\Psi_{a}$, where the notations $\Psi_{V}$ and $\bar{\Psi}_{V}$ are used respectively for the components of $\Psi_{a}$ along the even and odd vector harmonics. It is established here that this spacelike transition requires that a particular relationship between the shear and the vorticity 2-vectors holds, specifically, $\Sigma_{V}-\bar{\Omega}_{V}<0$. In order for the causal preservation of the horizon character, $\Sigma_{V}-\bar{\Omega}_{V}=0$. The result was obtained for the particular consideration of a non-dissipative perturbation where the scalars, tensors and vector quantities associated to the heat flux and anisotropic pressure are vanish. We also considered the case of a spatially homogeneous background solution and wrote down the form of the horizon characterizing function. In the Schwarzschild case, any transition from a null character is necessarily spacelike as the horizon encloses a black hole (note that we are assuming the MOTS condition is preserved). However, for the spatially homogeneous case null horizons do not enclose black holes, and therefore the causal character could transition to a timelike one under the perturbation. Imposing a linear equation of state, for a particular frame choice, i.e., fixing the spatial unit vector to align with the acceleration of the temporal unit vector (which coincides with the vanishing of the difference between the 2-vector components, i.e., $\Sigma_{V}-\bar{\Omega}_{V}=0$), the sign of the equation of state parameter $w$ alone characterizes the evolution: null if $w=0$, spacelike if $w>0$, and timelike if $w<0$. Without this frame choice, the evolution of the horizon is a bit more involved. We nonetheless present a partial characterization of the horizon dynamics in yhje presence of perturbations. In particular, where the difference between these vector components is strictly negative or strictly positive, we find the ranges of the equation of state parameter where the induced metric on the horizon has absolute sign, i.e., the causal character is timelike or spacelike. However, there are ranges of the parameter for which the causal character is determined by a complex relationship between the 2-vectors, the energy density and the eigenvalues of MOTS Laplacian, which we could not analytically determine. Finally, we considered the relationship between the shear and vorticity 2-vectors which play a crucial role in the characterization of the horizon, to the MOTS stability operator. These 2-vectors also specify the 1-form $\tilde{s}_{a}$ which controls the self-adjointness of the operator, and points to an obvious connection to the characterization we have introduced. We established that the component $\bar{\tilde{s}}_{V}$ vanishes for our particular consideration in this work, and that $\tilde{s}_{V}=\Sigma_{V}+\bar{\Omega}_{V}$. This implied that for our consideration the 1-form $\tilde{s}_{a}$ is a gradient (so that the MOTS stability operator is self-adjoint) if and only if $\tilde{s}_{a}$ vanishes identically. As a consequence, in the case of perturbation of a Schwarzschild background, it follows that when the stability operator is not self-adjoint and $\bar{\Omega}_{V}\neq 0$, the event horizon necessarily transitions to a spacelike character. In the case of a spatially homogeneous background, in the self-adjoint case with a vanishing $\tilde{s}_{a}$, it is clear that only one of the components $\Sigma_{V}$ or $\bar{\Omega}_{V}$ is needed for the characterization. Several additional results were also obtained, restricting the form of 2-tensors and 2-vectors on the MOTS. For example, it was established that (and this appears independent of the choice of background) the components of the even and odd vector harmonics, respectively, of the magnetic and electric Weyl 2-vectors necessarily vanish on the MOTS. In fact, it was demonstrated that the bi-implication $\bar{\tilde{s}}_{V}=0\iff\mathcal{H}=0$ follows, which is due to the fact that $\mathcal{H}$ is necessarily a first order quantity. Furthermore, the component of the odd tensor harmonics of the shear 2-tensor also vanishes on the MOTS. As the 1-form $\tilde{s}_{a}$ is tied to rotation generation on the horizon, we provide a clear picture, at least in the linear regime, of the precise source of rotation on the horizon: for a non-dissipative linear perturbation of a LRS background spacetime, rotation is entirely sourced by contributions from the Weyl 2-vectors on the MOTS. If one includes dissipation, then rotation gains a contribution from the heat flux 2-vector. Actually, there could be dissipation from the anisotropic 2-vector and no contribution from the heat flux 2-vector, and it remains true that horizon rotation will be entirely sourced by Weyl contributions. It then follows that for a conformally flat and non-dissipative perturbation of an inhomogeneous LRS background spacetime, the stability operator is necessarily self-adjoint. We also presented a class of perturbations for which the 1-form $\tilde{s}_{a}$ is non-vanishing but one may nonetheless expand the component $\bar{\tilde{s}}_{V}$ in the harmonic basis so that the MOTS stability operator can be brought to a self-adjoint form. Outlook – There are several interesting directions that could be explored. In this work, we have restricted ourselves to first order perturbations. What happens when non-linearities are incorporated into the analysis of the horizon evolution? Naturally the problem becomes a bit (or a lot) more complicated as seen from the characterizing function $z$ in (27). For example, there will be additional contributions from the non-linearities to the scalar curvature of the MOTS, so that large distortions are introduced on the MOTS. These distortions are not present at first order as was discussed earlier. Modification to the eigenvalue in the self-adjoint case will present in the form of the inclusion of dissipation and the cosmological constant $\Lambda$ (if $\Lambda$ is considered a perturbation ‘variable’). Any non-linearity introduced is an extra contribution to the principal eigenvalue through the scalar curvature of the MOTS. Another example is the following. At every stage of our analysis we have imposed the MOTS condition $\chi=0$, meaning that the built-in assumption here is that a background horizon, under perturbations of its ambient background spacetime, evolves to a horizon foliated by MOTS. However, it is quite possible that under the perturbation the horizon is no longer sliced into MOTS. This is the case of the Weyl-distorted Schwarzschild geometry discussed in the introduction, where by the tuning up the distortion parameter the $r=2$ horizon is no longer a MOTS [12]. This degeneracy may be a feature of non- linearities under the perturbation and so might not happen in the linear perturbative case considered in this work. However, as mentioned in the second point above, extending our analysis to non-linear perturbations, it would suffice as a more general consideration, not to impose the MOTS condition and analyze the non-linear equations. ## VII Acknowledgement AS and SK research is supported by the Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of education (grant numbers) (NRF-2022R1I1A1A01053784) and (NRF-2021R1A2C1005748). PKSD acknowledges support through the First Rand Bank, South Africa. AS would also like to thank the IBS Center for Theoretical Physics of the Universe, Daejeon, South Korea, for its hospitality, where a part of this work was carried out. ## References * [1] R. Penrose, Phys. Rev. D 14, 57 (1965). * [2] R. Wald, (Chicago IL: University of Chicago Press, 1984). * [3] S. A. Hayward, Phys. Rev. D 49, 6467 (1994). * [4] I. Booth, Can. J. Phys. 83, 1073 (2005). * [5] I. Booth, L. Brits, J. A. Gonzalez, and C. Van Den Broeck, Class. Quantum Grav. 23, 413 (2005). * [6] L. Andersson, M. Mars, and W. Simon Phys. Rev. Lett. 95, 111102 (2005). * [7] L. Andersson, M. Mars, and W. Simon Adv. Theor. Math. Phys. 12, 853 (2008). * [8] A. Ashtekar and G. J. Galloway, Adv. Theor. Math. Phys. 9, 1 (2005). * [9] A. Ashtekar and B. Krishnan, Living Rev. Relativ. 7, 10 (2004). * [10] K. Mikolaj, Phys. Rev. D 44, 104029 (2006). * [11] A. Ashtekar, N. Khera, M. Kolanowski, and J. Lewandowski, J. High Energ. Phys. 2022, 66 (2022). * [12] T. Pilkington, A. Melanson, J. Fitzgerald, and I. Booth, Class. Quantum Grav. 28, 125018 (2011). * [13] R. Geroch and J. Hartle, J. Math. Phys. 63, 680 (1982). * [14] G. F. R. Ellis, J. Math. Phys. 8, 1171 (1967). * [15] J. M. Stewart and G. F. R. Ellis, J. Math. Phys. 9, 1072 (1968). * [16] C. Clarkson and R. K. Barrett, Class. Quantum Grav. 20, 3855 (2003). * [17] C. Clarkson, Phys. Rev. D 76, 104034 (2007). * [18] George F. R. Ellis, R. Goswami, A. I. Hamid, and S. D. Maharaj, Phys. Rev. D 90, 084013 (2014). * [19] A. M. Sherif, R. Goswami, and S. D. Maharaj, Class. Quantum Grav. 36, 215001 (2019). * [20] A. M. Sherif, Eur. Phys. J. C 81, 1 (2021). * [21] A. M. Sherif and P. K. S. Dunsby, Class. Quantum Grav. 40, 045005 (2023). * [22] P. Semrén, Universe, 8, 406 (2022). * [23] P. Semrén and M. Bradley, Class. Quantum Grav. 39, 235003 (2022) * [24] C. Hansraj, R. Goswami, and S. D. Maharaj, Eur. Phys. J. C 83, 321 (2023). * [25] J. M. Stewart and M. Walker, Proc. R. Soc. A 431, 49 (1974). * [26] G. F. R. Ellis and H. Van Elst, NATO Adv. Study Inst. Ser. C, Math. Phys. Sci. 541, 1 (1999). * [27] I. Booth and S. Fairhurst, Phys. Rev. Lett. 92, 011102 (2004). * [28] I. Booth and S. Fairhurst, Class. Quantum Grav. 22, 4515 (2005). * [29] I. Ben-Dov, Phys. Rev. D 70, 124031 (2004). * [30] J. L. Jaramillo, Class. Quantum Grav. 32, 132001 (2015). * [31] I. Booth, G. Cox, and J. Margalef-Bentabol, ArXiv 2311, 02063v1 (2023).
# Edit Distance between Merge Trees Raghavendra Sridharamurthy, , Talha Bin Masood, Adhitya Kamakshidasan, and Vijay Natarajan R. Sridharamurthy, T.B. Masood, A. Kamakshidasan, and V. Natarajan are with the Department of Computer Science and Automation, Indian Institute of Science, Bangalore, 560012. E-mail<EMAIL_ADDRESS> ###### Abstract Topological structures such as the merge tree provide an abstract and succinct representation of scalar fields. They facilitate effective visualization and interactive exploration of feature-rich data. A merge tree captures the topology of sub-level and super-level sets in a scalar field. Estimating the similarity between merge trees is an important problem with applications to feature-directed visualization of time-varying data. We present an approach based on tree edit distance to compare merge trees. The comparison measure satisfies metric properties, it can be computed efficiently, and the cost model for the edit operations is both intuitive and captures well-known properties of merge trees. Experimental results on time-varying scalar fields, 3D cryo electron microscopy data, shape data, and various synthetic datasets show the utility of the edit distance towards a feature-driven analysis of scalar fields. ###### Index Terms: Merge tree, scalar field, distance measure, persistence, edit distance. ## 1 Introduction The study of the behavior of physical quantities over time helps in understanding underlying scientific processes. Physical quantities are either measured using imaging devices or computed via simulation. In either case, they are often modeled as scalar functions (also referred to as scalar fields). Direct analysis and visualization of such a scalar function using isosurfaces or volume rendering provides a good overview but is limited by two factors. First, increasing size of data makes storage and retrieval inefficient. Second, the analysis often requires a sweep over a large subset of the domain or range of the function even when the features of interest may be contained within a small region. These limitations are amplified when we consider time-varying scalar functions. Thus, these techniques are not well suited for feature directed analysis and visualization. Topological structures such as the _merge tree_ [1] shown in Figure 1 provide a succinct representation of the scalar function, support feature-directed visualization and exploration, and hence enable the user to quickly identify patterns and gain insights. Multiple scenarios demand a method for comparing scalar functions. For example, a distance or similarity measure between scalar functions is essential for detecting periodicity in a time-varying dataset. The matrix of distances between all pairs of time steps will display a characteristic pattern if the function is periodic. A method for comparing scalar functions is also useful for tracking features in time-varying phenomena [2], topological shape matching [3], detecting symmetry/asymmetry in scalar fields [4, 5, 6, 7, 8], or clustering [9], computing temporal summaries of large data sets, to identify features that are preserved in ensemble simulations or multi-field data, or to compare simulated data against measured data [10, 11, 12, 13]. In the above-mentioned scenarios, the similarity or dissimilarity between scalar functions is often captured by a distance measure between topological structures that represent the functions. We want such a distance measure to satisfy useful theoretical properties and be efficiently computable in order to be applicable in practice. (a) 2D scalar field (b) left: join tree and right: split tree Figure 1: Merge trees. (a) A 2D scalar field (b) A merge tree tracks the connectivity of sub-level sets (preimage of $f^{-1}(-\infty,c]$) or the super- level sets (preimage of $f^{-1}[c,\infty)$). ### 1.1 Related Work Assuming identical domains, RMS distance, Chebyschev distance, and other norms such as $L_{p},1\leq p\leq\infty$ can be used for point-to-point comparisons. However, a direct comparison of the two scalar functions may not be appropriate because of its sensitivity to noise and minor perturbations. Distance measures between various topological structures have been studied in the literature, beginning with the bottleneck distance between persistence diagrams [14]. A topological feature is often represented by a creator- destructor pair, a critical point pair in the case of scalar functions. For example, a minimum creates a 0-dimensional topological feature (connected component) in the sub-level set that is destroyed by a saddle. A persistence diagram (see Figure 3) depicts the persistence or “lifetime” of all topological features by plotting their orresponding time of creation (birth) and destruction (death) as points in $\mathbb{R}^{2}$. The bottleneck distance ($D_{B}$) between two persistence diagrams is equal to the weight of the minimum weight mapping between points of the two diagrams. The weight of a mapping is equal to the largest $L_{\infty}$ distance between a point and its image under the mapping. We say that the persistence diagram is _stable_ with respect to a distance measure if it is bounded above by the $L_{\infty}$ distance between the two scalar functions. Intuitively, we require that small perturbations to the scalar functions translate to small changes in the distance between the respective persistence diagrams. The persistence diagram is stable with respect to $D_{B}$. But, the persistence diagram is only a multiset. It does not capture the spatial configuration of critical points, which reduces its discriminative capability. Morozov et al. [15] proposed the interleaving distance between merge trees. This distance is defined by a continuous map that shifts points of one merge tree onto the other and vice- versa. The distance is equal to the smallest value of the shift such that the map satisfies certain compatibility conditions. The merge tree is stable under this distance and the distance measure is more discriminative compared to the bottleneck distance but computing it is not a tractable problem. Beketayev et al. [16] define a distance measure between merge trees that can be computed by considering all possible branch decompositions. This measure can be computed in polynomial time but provide no guarantees on stability. Distance measures have been also defined for other topological structures. The _Reeb graph_ captures the topology of both sub-level sets and super-level sets of scalar functions defined on a manifold [17]. Bauer et al. [18] imposed a metric on Reeb graphs called the functional distortion distance. They proved its stability and connections with other distances such as the bottleneck and interleaving distance. The computation depends on the Gromov-Hausdorff distance, which is proven to be NP-hard [19] to even approximate up to a constant factor for general metric graphs. Di Fabio and Landi [20] defined an edit distance for Reeb graphs on surfaces. They also proved its stability and showed connections with interleaving distance and function distortion distance but there is no polynomial time algorithm to compute the distance. Dey et al. [21] defined the persistence distortion distance to compare metric graphs, proved its stability, and described a polynomial time algorithm with asymptotic running time $O(m^{12}\log n)$ (continuous version) and $O(n^{2}m^{1.5}\log m)$ (discrete version), where $m$ is the number of edges and $n$ is the number of vertices in the larger graph. They also reported applications to shape matching. Narayanan et al. [22] defined a distance measure to compare _extremum graphs_ , whose nodes corresponds to critical points of the scalar function and arcs correspond to integral lines. The distance measure is based on the maximum weight common subgraph and they use pruning techniques to speedup the computation. While stability is not guaranteed, they present many experimental results on time-varying data to demonstrate its application to time-varying data analysis and visualization. In contrast to the rigorous definitions of distance measures introduced in the above-mentioned works, simpler but practical similarity measures have also been studied. Saikia et al. [7] introduced the extended branch decomposition graph (eBDG) that describes a hierarchical representation of all subtrees of a join/split tree and designed an efficient algorithm to compare them. They also present experimental results on time-varying data. Saikia et al. [23] studied a measure that compared histograms that are constructed together with the merge trees. As in the case of bottleneck distance, this measure ignores the structure but it can be computed efficiently and is therefore useful in practice. Saikia and Weinkauf [2] later extended this measure and demonstrated applications to feature tracking in time-varying data. Edit distances and alignment distances for trees are inspired by edit distances defined on strings. They have found various applications, such as comparing neuronal trees [24], comparing shapes [25], comparing music genre taxonomy [26], analysis of glycan structures [27], comparing RNA structures [28], and comparing plant architectures [29]. Given two strings, one is transformed into the other via a sequence of operations where each operation has a non-negative associated cost. The distance is defined as the minimum cost over all such transformations. Similar distance measure may be defined for labeled trees with edit operations like relabeling, addition, and deletion of nodes. Zhang and Shasha [30] described an algorithm to compute the tree edit distance for ordered labeled trees. Later, Zhang [31] proposed a new algorithm for constrained tree edit distance for ordered labeled trees. The computation of tree edit distance for unordered labeled trees is NP-complete [32]. However, the constrained version of the problem can be solved in polynomial time using a dynamic programming based algorithm [33]. A gap corresponds to a collection of nodes that are inserted / deleted during a sequence of edit operations. Edit distance with arbitrary gap costs were first proposed by Touzet [34], who showed that the distance computation is NP-hard. But, the distance between ordered labeled binary trees can be computed in polynomial time [35]. While tree edit distance based algorithms have been employed in many applications, they have not been well studied for comparing topological structures like merge trees except in very recent work. Riecke et al. [36] defined a hierarchy of persistence pairs and a tree edit distance based dissimilarity measure to compare hierarchies. Sridharamurthy et al. [37] adapt Xu’s algorithm [35] for computing distance between ordered labeled binary trees to the case of the general subtree gap model that preserve the merge tree structure. The general subtree gap model allows for interior nodes to be inserted / deleted while retaining the child nodes. The cost model is intuitive and they present preliminary experimental results to show its utility. However, this method has multiple shortcomings: * • The gap model is too general. In the case of merge trees, we require a constrained version that considers gaps as persistence pairs in order to preserve the structural integrity of the tree. These pairs depend on function values. So, the constraints are ad hoc, difficult to express directly and to incorporate into the dynamic programming based algorithm that is used to compute the measure. * • The above-mentioned pairs are not stable under perturbations to the scalar function. * • Merge trees constructed on real world data are not necessarily binary trees. * • Absence of a natural left-to-right ordering of children of a node in the merge tree. The algorithm requires such an ordering, random or canonical orderings lead to instabilities. * • The running time of the algorithm is approximately $O(n^{5})$, where $n$ is the number of nodes in the tree. This is very slow for practical applications. In summary, existing work either propose a rigorous definition of distance with theoretical guarantees but without practical value (with very few exceptions) or describe a similarity / dissimilarity measure with practical applications but without theoretical analysis. Further, existing methods do not provide natural support for a fine grained analysis of similar/dissimilar regions. ### 1.2 Contributions In this paper, we propose a tree edit distance based approach to compare merge trees. The distance measure is an adaptation of the constrained unordered tree edit distance [33], but is a significant modification that caters to merge trees and alleviates the shortcomings of the measure proposed by Sridharamurthy et al. [37]. Individual edits correspond to topological features. The edit operations may be subsequently studied for a fine grained analysis. The paper makes the following key contributions: 1. 1. An intuitive and mathematically sound cost model for the individual edit operations. 2. 2. A proof that the distance measure is a metric under the proposed cost model. 3. 3. A computational solution to handle instabilities. 4. 4. Experiments to demonstrate the practical value of the distance measure using various applications — 2D time-varying data analysis by detecting periodicity, summarization to support visualization of 3D time-varying data, detection of symmetry and asymmetry in scalar fields, study of topological effects of subsampling and smoothing, and shape matching. In addition, we describe a comprehensive set of validation experiments that are designed to help understand the properties of the measure. ## 2 Background In this section, we introduce necessary definitions and background on merge trees, list some desirable properties of distance measures, and describe three edit operations on a merge tree that define a tree edit distance. ### 2.1 Merge tree Figure 2: Persistence pairs in the join (left) and split (right) trees. A merge tree [1] captures the connectivity of sub-level sets (_join tree_) or super-level sets (_split tree_) of a scalar function $f:\mathbb{X}\longrightarrow\mathbb{R}$ defined on a manifold domain $\mathbb{X}$, see Figure 1. A value $c$ in the range of $f$ is called an _isovalue_. Given an isovalue, an _isocontour_ is defined as the collection of all points $x\in\mathbb{X}$ such that $f(x)=c$. Nodes of join trees consist of minima $M=\\{m_{i}\\}$, saddles $S=\\{s_{j}\\}$, and the global maximum. In theory, the structure of a join tree is simple. Excluding the global maximum, which is the root of the tree, every node has either 0 (minimum) or 2 children (saddle). All minima are paired with saddles based on the notion of topological persistence [38] except for one which is paired to the lone global maximum. Each such pair $(m,s)$ represents a topological feature and its persistence is defined as $pers(m)=pers(s)=f(s)-f(m)$. In practice, saddles may have more than two children. We discuss how to handle them in Section 4.4. A split tree is defined likewise. It contains a set of maxima and saddles together with the global minimum. Figure 2 shows the persistence pairing for the trees from Figure 1. Several fast algorithms have been developed for computing merge trees (join or split) for piecewise linear functions defined on simply connected domains [1, 39, 40]. Figure 3: A 1D scalar function (left) and the persistence diagram of the function (right). Each birth-death pair $(b_{i},d_{i})$ is a feature of the scalar function and its persistence is defined as $d_{i}-b_{i}$. Each pair is represented as a point in $\mathbb{R}^{2}$. ### 2.2 Distance measures Designing distance measures is a well studied problem and has several applications in data analysis, visualization, pattern recognition, data mining, and machine learning. A distance measure $D:\mathbb{X}\times\mathbb{X}\longrightarrow\mathbb{R}$ on a domain $\mathbb{X}$ satisfies the metric properties: 1. 1. _Non-negativity_ : $D(x,y)\geq 0$ 2. 2. _Identity of indiscernibles_ : $D(x,y)=0$ iff $x=y$ 3. 3. _Symmetry_ : $D(x,y)=D(y,x)$ 4. 4. _Triangle inequality_ : $D(x,z)\leq D(x,y)+D(y,z)$ When metric properties such as triangle inequality are relaxed, we get a _dissimilarity measure_ rather than a distance measure. Further, while comparing scalar fields or topological structures constructed based on scalar fields, it is desirable that the distance measure satisfies two additional properties – stability and discrimination. In the following discussion, we will use $D$ to refer to the distance between topological structures that represent the functions. Given two scalar functions $f,g$, 1. 1. _Stability_ : $D(f,g)\leq\|f-g\|_{\infty}$ 2. 2. _Discrimination_ : $D_{B}(f,g)\leq D(f,g)$. Intuitively, stability requires that if the functions are not too “different” in terms of the $L_{\infty}$ norm of the difference between the functions then the distance measure between the topological structures representing the scalar functions should also be small. Discrimination, on the other hand, requires that however “small” the difference between two functions, it should be captured by the distance measure. Specifically, the distance measure equals $0$ should imply that the functions are equal. Since the bottleneck distance $D_{B}$ between persistence diagrams of $f$ and $g$ was among the first measures defined between topological structures, we typically state this property in terms of how the distance $D$ is related to bottleneck distance. Figure 4 shows an example where $D_{B}=0$ for a pair of functions that are not equal, which implies that $D_{B}$ is not discriminative enough. One reason for the low discriminative power is that the persistence diagram, and hence $D_{B}$, does not incorporate the connectivity between critical points as in the merge tree. We wish to design a distance measure that satisfies the following property: $\displaystyle D_{B}(f,g)\leq D(f,g)\leq\|f-g\|_{\infty}$ (1) From the computational perspective, $D$ should be computable either exactly or within a constant factor of approximation in polynomial time in order for it to be useful in a practical application. Figure 4: The discriminative power of the bottleneck distance $D_{B}$ is low. Two scalar functions (blue and red) and the corresponding persistence diagrams and merge trees. Even though the scalar functions are different, $D_{B}$ is not able to capture the difference because the persistence diagrams are equal. A distance measure that considers the structure of the merge tree would discriminate the two scalar functions. ### 2.3 Tree edit distance Tree edit distances have been studied extensively in the past few decades [41]. All these measures typically employ a set of edit operations with associated costs and try to minimize the total cost over the set of all edit operations. Let $T$ be a rooted tree with node set $V$ and edge set $E$. For a node $v\in V$, $deg(v)$ is the number of children of $v$, and $parent(v)$ is its parent in the tree. The maximum degree of a node in the tree is denoted as $deg(T)$. We denote an empty tree by $\theta$. Since we are interested in labeled trees, let $\Sigma$ be the set of labels, and $\lambda\notin\Sigma$ denote the null or empty character, which corresponds to a gap. In the following discussion, we use notations and definitions from Zhang [33]. Edit operations. The edit operations differ based on the gap model. For this discussion we consider edit operations that modify the tree, one node at a time. Xu [35] gives a detailed discussion of general gaps where edits modify multiple nodes. We consider a total of three edit operations as shown in Figure 5. (a) delete (b) insert (c) relabel Figure 5: Three different tree edit operations. Each edit affects only one node in the tree. The null character $\lambda$ corresponds to a gap. 1. 1. relabel: A relabel $a\longrightarrow b$ corresponds to an operation where the label $a\in\Sigma$ of a node is changed to a label $b\in\Sigma$. 2. 2. delete: A delete operation $a\longrightarrow\lambda$ removes a node $n$ with label $a\in\Sigma$ and all the children of $n$ are made the children of $parent(n)$. 3. 3. insert: An insert operation $\lambda\longrightarrow b$ inserts a node $n$ with label $b\in\Sigma$ as a child of another node $m$ by moving all the children of $m$ to children of $n$. We define a cost function $\gamma$ that assigns a non-negative real number to each edit operation of the form $a\longrightarrow b$. It is useful if the cost function $\gamma$ satisfies metric properties i.e. $\forall a,b,c\in\Sigma\cup\\{\lambda\\}$ 1. 1. $\gamma(a\longrightarrow b)\geq 0,\ \gamma(a\longrightarrow a)=0$ 2. 2. $\gamma(a\longrightarrow b)=\gamma(b\longrightarrow a)$ 3. 3. $\gamma(a\longrightarrow c)\leq\gamma(a\longrightarrow b)+\gamma(b\longrightarrow c)$ In particular, Zhang [33] proved that if $\gamma$ is a metric then the edit distance is also a metric, else it will be merely a dissimilarity measure. Given a tree $T_{1}$, we can apply a sequence of edit operations to transform it into another tree $T_{2}$. If $S=s_{1},s_{2},\ldots,s_{k}$ is a sequence of edit operations, where each $s_{i}$ is an edit, we can extend the cost function to $S$ by defining $\gamma(S)=\Sigma_{i=1}^{|S|}\gamma(s_{i})$. Edit distance. Formally, the distance between two trees $T_{1},T_{2}$ is defined as $\displaystyle D_{e}(T_{1},T_{2})=\min_{S}\\{\gamma(S)\\}$ (2) where $S$ is an edit operation sequence from $T_{1}$ to $T_{2}$. ## 3 Edit Distance Mappings Computing the edit distance between merge trees is a minimization problem with a huge search space. In order to understand this search space and how it affects the computation, we first define some edit distance mappings – unconstrained, constrained, and restricted – and their properties as described by Zhang [33]. We refer the reader to the supplementary material for additional description and illustrations of the mappings. ### 3.1 Unconstrained edit distance mapping The sequence of edit operations performed to transform $T_{1}$ into $T_{2}$ determines a mapping between the two trees. For convenience, we order the nodes of both the trees. This ordering does not affect the distance. Let $t_{1}$ and $t_{2}$ denote the ordering of nodes in $T_{1}$ and $T_{2}$, respectively, and $t_{1}[i]$ represents the $i$th node in the ordering. Let $M_{e}$ denote a collection of ordered integer pairs $(i,j)$. A triple $(M_{e},T_{1},T_{2})$ defines the _edit distance mapping_ from $T_{1}$ to $T_{2}$, where each pair $(i_{1},j_{1}),(i_{2},j_{2})\in M_{e}$ satisfies the following properties: * • $i_{1}=i_{2}$ iff $j_{1}=j_{2}$ (one-to-one) * • $t_{1}[i_{1}]$ is an ancestor of $t_{1}[i_{2}]$ iff $t_{2}[j_{1}]$ is an ancestor of $t_{2}[j_{2}]$ (ancestor ordering). The cost of transforming $T_{1}$ into $T_{2}$ can be expressed through the mapping as $\displaystyle\gamma(M_{e})$ $\displaystyle=\sum\limits_{(i,j)\in M_{e}}\gamma(t_{1}[i]\longrightarrow t_{2}[j])$ (3) $\displaystyle+\sum\limits_{\\{i|\nexists j,(i,j)\in M_{e}\\}}\gamma(t_{1}[i]\longrightarrow\lambda)$ $\displaystyle+\sum\limits_{\\{j|\nexists i,(i,j)\in M_{e}\\}}\gamma(\lambda\longrightarrow t_{2}[j])$ Given a sequence of edit operations $S$ that transforms $T_{1}$ into $T_{2}$, there exists a mapping $M_{e}$ such that $\gamma(M_{e})\leq\gamma(S)$. Conversely, given an edit distance mapping $M_{e}$ ,there exists a sequence of edit operations $S$ such that $\gamma(S)=\gamma(M_{e})$. Using the above, it can be shown that $\displaystyle D_{e}(T_{1},T_{2})=\min_{M_{e}}\\{\gamma(M_{e})\\}$ (4) where $(M_{e},T_{1},T_{2})$ defines the _edit distance mapping_ from $T_{1}$ to $T_{2}$. Zhang et al. [32] showed that computing $D_{e}(T_{1},T_{2})$ is NP-complete even when the trees are binary and $|\Sigma|=2$. ### 3.2 Constrained and restricted mappings Adding constraints to the edit distance mapping brings it within the computationally tractable realm. The main constraint imposed is that disjoint subtrees are mapped to disjoint subtrees. Let $T[i]$ denote the subtree rooted at the node with label $i$ and $F[i]$ denote the unordered forest obtained by deleting the node $t[i]$ from $T[i]$. A node $t_{1}[i]$ is a _proper ancestor_ of $t_{1}[j]$ if $t_{1}[i]$ lies on the path from the root to $t_{1}[j]$ and $t_{1}[i]\neq t_{1}[j]$. The triple $(M_{c},T_{1},T_{2})$ is called a _constrained edit distance mapping_ if, * • $(M_{c},T_{1},T_{2})$ is an edit distance mapping, and * • Given three pairs $(i_{1},j_{1}),(i_{2},j_{2}),(i_{3},j_{3})\in M_{c}$, the _least common ancestor_ $lca(t_{1}[i_{1}],t_{1}[i_{2}])$ is a proper ancestor of $t_{1}[i_{3}]$ iff $lca(t_{2}[j_{1}],t_{2}[j_{2}])$ is a proper ancestor of $t_{2}[j_{3}]$. The constrained edit distance mappings can be composed. Given two constrained edit distance mappings $M_{c_{1}}$ from $T_{1}$ to $T_{2}$ and $M_{c_{2}}$ from $T_{2}$ to $T_{3}$, $M_{c_{2}}\circ M_{c_{1}}$ is a constrained edit distance mapping between $T_{1}$ and $T_{3}$. Also, $\displaystyle\gamma(M_{c_{2}}\circ M_{c_{1}})\leq\gamma(M_{c_{1}})+\gamma(M_{c_{2}})$ (5) which can be proven using the triangle inequality imposed on the edit operation costs. This leads to the definition of constrained edit distance $\displaystyle D_{c}(T_{1},T_{2})=\min_{M_{c}}\\{\gamma(M_{c})\\}$ (6) $D_{c}$ also satisfies metric properties. Both $M_{e}$ and $M_{c}$ deal with mapping between unordered trees. Similar mappings work for forests. We define a _restricted mapping_ $M_{r}(i,j)$ between $F_{1}[i]$ and $F_{2}[j]$ as follows: * • $M_{r}(i,j)$ corresponds to a constrained edit distance mapping between $F_{1}[i]$ and $F_{2}[j]$. * • Given two pairs $(i_{1},j_{1}),(i_{2},j_{2})\in M_{c}$, $t_{1}[l_{1}]$ and $t_{1}[l_{2}]$ belong to a common tree in $F_{1}[i]$ if and only if $t_{2}[j_{1}]$ and $t_{2}[j_{2}]$ belong to a common tree in $F_{2}[i]$. Essentially, nodes within different trees of $F_{1}$ are mapped to nodes lying in different trees of $F_{2}$. ### 3.3 Constrained edit distance We recall the properties of $D_{c}$. Let $t_{1}[i_{1}],t_{1}[i_{2}],\ldots,t_{1}[i_{n_{i}}]$ be the children of $t_{1}[i]$ and $t_{2}[j_{1}],t_{2}[j_{2}],\ldots,t_{2}[j_{n_{j}}]$ be the children of $t_{2}[j]$. Further, let $\theta$ denote the empty tree. Then, $\displaystyle D_{c}(\theta,\theta)$ $\displaystyle=0,$ (7) $\displaystyle D_{c}(F_{1}[i],\theta)$ $\displaystyle=\sum\limits_{k=1}^{n_{i}}D_{c}(T_{1}[i_{k}],\theta),$ (8) $\displaystyle D_{c}(T_{1}[i],\theta)$ $\displaystyle=D_{c}(F_{1}[i],\theta)+\gamma(t_{1}[i]\longrightarrow\lambda),$ (9) $\displaystyle D_{c}(\theta,F_{2}[j])$ $\displaystyle=\sum\limits_{k=1}^{n_{j}}D_{c}(\theta,T_{2}[j_{k}]),$ (10) $\displaystyle D_{c}(\theta,T_{2}[j])$ $\displaystyle=D_{c}(\theta,F_{2}[j])+\gamma(\lambda\longrightarrow t_{2}[j]),$ (11) $\displaystyle\scriptsize D_{c}(T_{1}[i],T_{2}[j])$ $\displaystyle=\scriptsize\min\begin{cases}D_{c}(\theta,T_{2}[j])+\min\limits_{1\leq t\leq n_{j}}\\{D_{c}(T_{1}[i],T_{2}[j_{t}])-D_{c}(\theta,T_{2}[j_{t}])\\},\\\ D_{c}(T_{1}[i],\theta)+\min\limits_{1\leq s\leq n_{i}}\\{D_{c}(T_{1}[i_{s}],T_{2}[j])-D_{c}(T_{1}[i_{s}],\theta)\\},\\\ D_{c}(F_{1}[i],F_{2}[j])+\gamma(t_{1}[i]\longrightarrow t_{2}[j]).\end{cases}$ (12) If the cost is not a metric, we need to include one additional case, namely $D_{c}(F_{1}[i],F_{2}[j])+\gamma(t_{1}[i]\longrightarrow\lambda)+\gamma(\lambda\longrightarrow t_{2}[j])$. The distance between two forests is given by $\displaystyle\scriptsize D_{c}(F_{1}[i],F_{2}[j])$ $\displaystyle=\scriptsize\min\begin{cases}D_{c}(\theta,F_{2}[j])+\min\limits_{1\leq t\leq n_{j}}\\{D_{c}(F_{1}[i],F_{2}[j_{t}])-D_{c}(\theta,F_{2}[j_{t}])\\},\\\ D_{c}(F_{1}[i],\theta)+\min\limits_{1\leq s\leq n_{i}}\\{D_{c}(F_{1}[i_{s}],F_{2}[j])-D_{c}(F_{1}[i_{s}],\theta)\\},\\\ \min\limits_{M_{r}(i,j)}\gamma(M_{r}(i,j)).\end{cases}$ (13) The minimum restricted mapping may be computed by constructing a weighted bipartite graph in such a way that the cost of the minimum weight maximum matching $MM(i,j)$ is exactly the same as the cost of the minimum restricted mapping $M_{r}(i,j)$, $\displaystyle\min\limits_{M_{r}(i,j)}\gamma(M_{r}(i,j))=\min\limits_{MM(i,j)}\gamma(MM(i,j))$ (14) ### 3.4 Algorithm Zhang described an algorithm for computing the tree edit distance for labeled unordered trees [33]. It is a dynamic programming based algorithm that follows from the properties discussed in Section 3.3. The pseudo code is presented in the supplementary material (Section 2). The entry $D(T_{1}[m],T_{2}[n])$ in the table with $m=|T_{1}|$ and $n=|T_{2}|$ corresponds to the final result. The algorithm computes the distance in $O(|T_{1}|\times|T_{2}|\times(deg(T_{1})+deg(T_{2}))\times log_{2}(deg(T_{1})+deg(T_{2})))$ time in the worst case. ## 4 Tree Edit Distance We now describe a new tree edit distance that is appropriate for comparing merge trees, discuss its properties, and an algorithm for computing the distance measure. ### 4.1 Comparing merge trees Our proposed measure is based on a variant of tree edit distance that applies to unordered general trees as opposed to ordered binary trees. This variant is appropriate because * • Merge trees are unordered trees. * • Merge trees are not binary in general. * • Persistence pairs represent topological features. So, it is natural that the edit operations are defined in terms of persistent pairs. * • The pairs do not fit into any subtree gap model that has been studied in the literature. Consider the properties of edit distance mapping mentioned in Section 3.1, but now in the context of merge trees. The one-to-one property is applicable but ancestor ordering might not hold in all cases. Small perturbations in the function value may result in swaps similar to rotations in AVL or red-black trees [42, Chapter 13], which violate the ancestor ordering. Such violations also result in instabilities i.e., cause significant fluctuations in the distance (see Section 4.4). Computing the edit distance with the ancestor order preserving mappings is already infeasible. Removing that constraint will make the computation more difficult. We introduce a stability parameter to ensure that ancestor order preserving mappings are identified in practically all cases. More details on this computational solution to handling instabilities can be found in Section 4.4. This solution does discard some mappings and may lead us away from the optimum solution. But, the stabilization ensures that the mapping remains meaningful and helps reduce the search space thereby making the problem tractable. To summarize, $D_{c}$ between unordered trees with suitable modifications seems to be a good candidate for comparing merge trees. In this section, we describe one such distance measure and demonstrate its use in the following section. The additional constraint of mapping disjoint subtrees to disjoint subtrees may seem limiting. Also, $D_{e}(T_{1},T_{2})\leq D_{c}(T_{1},T_{2})$, which implies that the constrained edit distance may not be optimal in many cases. But, we observe that, in practice, it is not as limiting and gives good results in many applications. ### 4.2 Cost model The edit distance mapping $M_{e}$ and the constrained edit distance mapping $M_{c}$ need to be suitably modified so that they are applicable for comparing merge trees. We begin by considering the edit operations as applicable to merge trees together with appropriate cost models. The literature on tree edit distances study generic trees and hence do not describe particular cost models. The following discussions focus on join trees but all results hold for split trees also. Tree edit operations on the join tree need to preserve the structural integrity of the join tree. This reduces the number of operations, say insertions and deletions. Consider a min-saddle pair $(m_{2},s_{1})$ in Figure 6. If $s_{1}$ is deleted, then it’s pair $m_{2}$ should also be deleted, and vice-versa. After deletion, $m_{1}$ is adjacent to $s_{2}$. But deletion of $s_{1}$ does not necessarily require that the entire subtree rooted at $s_{1}$ be deleted. In fact, deleting the entire subtree may not result in a valid join tree as illustrated in Figure 6. In this particular illustration, we consider the pairing imposed by the persistence. But, in general, we may consider other pairings based on say volume, hyper-volume, etc. Figure 6: Permitted and forbidden edit operations. (left) A gap is introduced by removing a persistence pair. (right) An edit operation that is permitted for generic trees but is invalid for a join tree. Nodes and arcs are repositioned to improve the tree layout. Gaps in the join tree can be represented as a collection of min-saddle pairs. In Figure 6, we can transform the first tree into the last tree by deleting the pairs $\\{(m_{2},s_{1}),(m_{3},s_{2})\\}$. We propose two cost models that capture the preservation of topological features and are applicable for join trees. Consider nodes $p\in T_{1}$ and $q\in T_{2}$. Then $p$ and $q$ are creators or destroyers of topological features in $T_{1}$ and $T_{2}$, respectively. Let the birth and death times of these features be $(b_{p},d_{p})$ and $(b_{q},d_{q})$, respectively. These birth-death pairs correspond to points in the persistence diagrams. Alternatively, they are represented as closed intervals $[b_{p},d_{p}]$ and $[b_{q},d_{q}]$ in a persistence barcode. #### 4.2.1 $L_{\infty}$ cost $C_{W}$ $\displaystyle\gamma(p\longrightarrow q)$ $\displaystyle=\min\begin{cases}\max(|b_{q}-b_{p}|,|d_{q}-d_{p}|),\\\ \frac{(|d_{p}-b_{p}|+|d_{q}-b_{q}|)}{2}\end{cases}$ (15) $\displaystyle\gamma(p\longrightarrow\lambda)$ $\displaystyle=\frac{|d_{p}-b_{p}|}{2}$ (16) $\displaystyle\gamma(\lambda\longrightarrow q)$ $\displaystyle=\frac{|d_{q}-b_{q}|}{2}$ (17) This cost model is based on the bottleneck and Wasserstein distances. Note that the insert / delete cost is based on the $L_{\infty}$-distance of the points $p$ (or $q$) from the diagonal in the persistence diagram. The relabel cost is the minimum of the $L_{\infty}$-distance between the points $p$ and $q$ and the sum of the $L_{\infty}$-distance from the points $p$ (or $q$) to the diagonal. This corresponds to the scenario where transforming $p$ to $q$ ($p\longrightarrow q$) by deleting $p$ and inserting $q$ ($p\longrightarrow\lambda$ and $\lambda\longrightarrow q$) has a lower cost in some cases. Figure 7 shows how these costs can be derived from the persistence diagram when there is no overlap in the barcodes and when there is overlap between the barcodes. #### 4.2.2 Overhang cost $C_{O}$ $\displaystyle\gamma(p\longrightarrow q)$ $\displaystyle=\min\begin{cases}|b_{q}-b_{p}|+|d_{q}-d_{p}|,\\\ |d_{p}-b_{p}|+|d_{q}-b_{q}|\end{cases}$ (18) $\displaystyle\gamma(s\longrightarrow\lambda)$ $\displaystyle=|d_{p}-b_{p}|$ (19) $\displaystyle\gamma(\lambda\longrightarrow t)$ $\displaystyle=|d_{q}-b_{q}|$ (20) This cost model is based on the overlap of the barcodes or the intervals. We consider the lengths of the overhang or the non-overlapping section to determine the costs. Consider $p\longrightarrow\lambda$, the interval corresponding to $p$ is given by $[b_{p},d_{p}]$ with length $|d_{p}-b_{p}|$ and the interval corresponding to $\lambda$ is $\emptyset$ with length $0$. Since there is no overlap, the cost is $|d_{p}-b_{p}|+0=|d_{p}-b_{p}|$. The cost of $\lambda\longrightarrow q$ can be derived similarly. Let us now consider the cost of $p\longrightarrow q$. If there is an overlap, we discard the overlap and obtain $|b_{q}-b_{p}|+|d_{q}-d_{p}|$. If there is no overlap then the cost is equal to $|d_{p}-b_{p}|+|d_{q}-b_{q}|$. The minimum of the two expressions is the relabel cost. The barcodes are shown in the fourth quadrant of the persistence diagrams in Figure 7. Figure 7: Illustration of the cost models when there is no overlap in the barcodes (left) and when there is overlap (right). We distinguish between birth and death events in the barcode by using different glyphs for the start and the end of the intervals. ### 4.3 Metric properties Metric property enables us to study the space of all the trees, compute the mean, and also compose transformations between merge trees. From Sections 2.3 and 3.2, we know that if the cost model satisfies the metric property then the distance measure is also a metric. We now prove the metric properties for our cost model. The overhang cost is similar to symmetric difference, which is a well-known metric [43]. We now prove that the $L_{\infty}$ cost $C_{W}$ is a metric. #### 4.3.1 $C_{W}$ is a metric We show that the cost $C_{W}$ is equal to the Wasserstein distance between two corresponding persistence diagrams. Let $N$ denote the set of all nodes in the merge trees and $\lambda$ denote a node corresponding to the null character. We define a mapping $\mathcal{M}:N\cup\\{\lambda\\}\longrightarrow Dgm$, where $Dgm$ is set of all persistence diagrams, as follows: 1. 1. $\forall p\in N,\ \mathcal{M}(p)=\\{(b_{p},d_{p})\\}\cup\\{(x,x),x\geq 0\\}$, 2. 2. $\mathcal{M}(\lambda)=\\{(x,x),x\geq 0\\}$. Define the distance on the set $N\cup\\{\lambda\\}$ as the Wasserstein distance of the first order _i.e._ , given $p,q\in N\cup\\{\lambda\\}$ $d(p,q)=W_{1}(\mathcal{M}(p),\mathcal{M}(q))$ Now, the cost $C_{W}$ can be rewritten as $\displaystyle\gamma(p\longrightarrow q)$ $\displaystyle=W_{1}(\mathcal{M}(p),\mathcal{M}(q))$ (21) $\displaystyle\gamma(p\longrightarrow\lambda)$ $\displaystyle=W_{1}(\mathcal{M}(p),\mathcal{M}(\lambda))$ (22) $\displaystyle\gamma(\lambda\longrightarrow q)$ $\displaystyle=W_{1}(\mathcal{M}(\lambda),\mathcal{M}(q)).$ (23) Since the Wasserstein distance $W_{1}(\cdot,\cdot)$ between persistence diagrams is known to be a metric [44, Chapter 6], $C_{W}$ is also a metric. However, this proof of the metric property is for general distributions. We have an alternative proof for merge trees from first principles with the aim to better understand the cost. #### 4.3.2 $C_{W}$ is a metric : proof from first principles Non-negativity and symmetry follows by definition because $C_{W}$ is based on sum, max, min of absolute values. Figure 8: The cost of edit operations can be reformulated as the weight of a minimum weight maximum matching in a bipartite graph. Bipartite graph for a relabel operation $p\longrightarrow q$ (left) and a delete operation (right). To prove the triangle inequality, we first reformulate the cost of the edit operations as the weight of a minimum weight maximum matching. The matching is defined in a bipartite graph. Nodes of the bipartite graph consists of the merge tree nodes together with an equal number of copies of $\lambda$. We collect the nodes of the graph to construct sets of the form $P=\\{p,\lambda\\}$ and a special multiset $\Lambda=\\{\lambda,\lambda\\}$, see Figure 8. All pairs of nodes from different multisets are connected by an edge. The edge weight $c$ is given by the $L_{\infty}$ distance between the corresponding points in the persistence diagram: $\displaystyle c_{pq}=L_{\infty}(p,q)$ $\displaystyle=\max(|b_{q}-b_{p}|,|d_{q}-d_{p}|)$ (24) $\displaystyle c_{p\lambda}=L_{\infty}(p,\lambda)$ $\displaystyle=\frac{|d_{p}-b_{p}|}{2}$ (25) $\displaystyle c_{\lambda q}=L_{\infty}(\lambda,q)$ $\displaystyle=\frac{|d_{q}-b_{q}|}{2}$ (26) $\displaystyle c_{\lambda\lambda}=L_{\infty}(\lambda,\lambda)$ $\displaystyle=0$ (27) The cost of the edit operations is equal to the cost of the minimum weight maximum matching $MM$ in this bipartite graph. In Figure 8, one of the two matchings will determine the cost of the edit operation. $\displaystyle\gamma(p\longrightarrow q)=MM(P,Q)$ $\displaystyle=\min\begin{cases}c_{pq}+c_{\lambda\lambda},\\\ c_{p\lambda}+c_{\lambda q}\end{cases}$ (28) $\displaystyle\gamma(p\longrightarrow\lambda)=MM(P,\Lambda)$ $\displaystyle=\min\begin{cases}c_{p\lambda}+c_{\lambda\lambda},\\\ c_{\lambda\lambda}+c_{p\lambda}\end{cases}$ (29) $\displaystyle\gamma(\lambda\longrightarrow q)=MM(\Lambda,Q)$ $\displaystyle=\min\begin{cases}c_{\lambda\lambda}+c_{\lambda q},\\\ c_{\lambda q}+c_{\lambda\lambda}\end{cases}$ (30) Figure 9: The cost function satisfies triangle inequality. (left) The blue or the red matching may be the minimum weight maximum matching that corresponds to the cost of the edit operation. (right) The matching between $P$ and $R$ is a composition of matching between $P,Q$ and $Q,R$. The inequality can be proved via case analysis by considering all possible compositions. Consider three multisets as shown in Figure 9. Using the above construction, we prove triangle inequality by considering the two cases, namely when the minimum weight matching is equal to either the red or blue matching. Case red: $MM(P,R)$ is given by the red matching. The cost of the relabel $\gamma(p\longrightarrow r)=c_{pr}+c_{\lambda\lambda}=c_{pr}$. Two different paths from $p$ lead to $r$, $p\longrightarrow q\longrightarrow r$ and $p\longrightarrow\lambda\longrightarrow r$. Consider the first path, $\displaystyle\gamma(p\longrightarrow r)=c_{pr}=L_{\infty}(p,r)$ $\displaystyle\leq L_{\infty}(p,q)+L_{\infty}(q,r)$ (31) $\displaystyle=c_{pq}+c_{qr}$ (32) $\displaystyle=\gamma(p\longrightarrow q)+\gamma(q\longrightarrow r)$ (33) Now, let us consider the second path. If $c_{pr}\leq c_{p\lambda}+c_{\lambda r}$ then $c_{pr}\leq c_{p\lambda}+c_{\lambda q}+c_{q\lambda}+c_{\lambda r}$ and we are done. Else, we have two sub-cases $\displaystyle c_{pr}$ $\displaystyle>c_{p\lambda}+c_{\lambda q}+c_{q\lambda}+c_{\lambda r}>c_{p\lambda}+c_{\lambda r},\mbox{ or}$ (34) $\displaystyle c_{pr}$ $\displaystyle>c_{p\lambda}+c_{\lambda r}\mbox{ but }c_{pr}<c_{p\lambda}+c_{\lambda q}+c_{q\lambda}+c_{\lambda r}.$ (35) In both sub-cases, we have a matching with weight $MM^{\prime}(P,R)=c_{p\lambda}+c_{\lambda r}\leq MM(P,R)=c_{pr}+c_{\lambda\lambda}$, which contradicts our assumption. The cost $\gamma(\lambda\longrightarrow\lambda)=c_{\lambda\lambda}=0$. Both paths via $Q$, $\lambda\longrightarrow\lambda\longrightarrow\lambda$ and $\lambda\longrightarrow q\longrightarrow\lambda$, should necessarily have a non-zero total cost. So, the inequality holds trivially. Case blue: $MM(P,R)$ is given by the blue matching. The cost of the relabel is equal to the sum $c_{p\lambda}+c_{\lambda r}$. We consider the two weights individually. Two paths from $p\in P$ lead to $\lambda\in R$ via a node in $Q$, $p\longrightarrow\lambda\longrightarrow\lambda$ and $p\longrightarrow q\longrightarrow\lambda$. Similarly, two paths from $\lambda\in P$ lead to $r\in R$, $\lambda\longrightarrow\lambda\longrightarrow r$ and $\lambda\longrightarrow q\longrightarrow r$. In both cases, triangle inequality holds trivially for the first path via $Q$, $c_{p\lambda}\leq c_{p\lambda}+c_{\lambda\lambda}$ and $c_{\lambda r}\leq c_{\lambda\lambda}+c_{\lambda r}$. We need to show that the inequality holds for the second paths as well. From the persistence diagram, we observe that $L_{\infty}(p,\lambda)\leq L_{\infty}(p,q)+L_{\infty}(q,\lambda)$ for all $q$, even when $q$ lies on the perpendicular from $p$ onto the diagonal. So, $c_{p\lambda}\leq c_{pq}+c_{q\lambda}$. A similar argument can be used to show that $c_{\lambda r}\leq c_{\lambda q}+c_{qr}$. The red and blue cases together imply that the cost $C_{W}$ satisfies the triangle inequality and is therefore a metric. It follows that the tree edit distance measure $D$ is also a metric. ### 4.4 Handling instabilities Saikia et al. [7] discuss two kinds of instabilities, _vertical_ and _horizontal_ , that affect branch decompositions and hence the distance measures. Figure 10 illustrates how horizontal instability can occur. In our case, the horizontal stability has a more drastic effect on the measure because Figure 10: Illustrating instabilities. Since the difference in the function values between $s_{1}$ and $s_{2}$ is small, a slight perturbation leads to a change in the structure of the tree, which affects the distance measure. * • It changes the persistence pairing, which in turn affects the cost. * • It also changes the subtrees thereby affecting the matching found by the algorithm. We employ a strategy similar to the one used for branch decompositions by Thomas and Natarajan [4] and apply it to merge trees. We introduce a stability parameter $\varepsilon$ and use it to determine how to merge simple saddles into a multi-saddle where instabilities occur. We merge the saddles in a bottom-up manner as follows. Begin from the lower saddle $s_{l}$ that is further from the root and merge it into a higher saddle $s_{h}$ that is nearer to the root if the function difference $|f(s_{h})-f(s_{l})|<\varepsilon$. Repeat this process until none of the saddles satisfy the merging condition. In the implementation, the multi-saddle is represented by the saddle with the highest persistence. We compute the distance between the stabilized trees. Optionally, a fixed value may be added to the final distance to incorporate the cost incurred due to the stabilization. In Section 5.2, we experimentally analyze how varying the stability parameter $\varepsilon$ affects the distance measure. ### 4.5 Algorithm We adapt Zhang’s algorithm [33] (See supplementary material, Section 2) with the edit costs discussed in Section 4.2 to compute the tree edit distance between merge trees. The input to this algorithm is a pair of merge trees that are stabilized using the strategy described in Section 4.4. ### 4.6 Implementation The computation proceeds in a bottom up manner. Distances for the subtrees are computed and stored in a table. These are next used for computing distances between subtrees at higher levels of the merge trees. This proof of concept implementation does not include code and memory optimizations for efficiently computing and storing the dynamic programming tables. We use the simple Kuhn- Munkres algorithm [45] for computing $MM(i,j)$. We still observe reasonable running times for most of the data sets as reported in the individual experiments in the following section. ## 5 Experiments and case studies We demonstrate the utility of the tree edit distance measure by applying it to analyze time-varying data, to study symmetry in scalar fields, for summarizing data, and for shape matching. We use the Recon library [46] to compute merge trees, the algorithm described in Section 4.5 to compute the tree edit distance between the merge trees, and Paraview [47] together with the Topology ToolKit TTK [48] to generate renderings of the merge trees together with the scalar fields. We uniformly use the $L_{\infty}$ cost $C_{W}$ (4.2.1). All experiments were performed on a machine with an Intel Xeon CPU with $8$ cores running at $2.0$ GHz and $16$ GB main memory. ### 5.1 Understanding the distance measure (a) Three scalar field $f_{1},f_{2},f_{3}$ (b) Merge tree driven segmentation for each field. (c) Mapping determined by $W_{1}$ for ($f_{1},f_{2}$) and ($f_{2},f_{3}$) (d) Mapping determined by $D$ for ($f_{1},f_{2}$) and ($f_{2},f_{3}$) Figure 11: Comparing mappings established by tree edit distance measure $D$ and Wasserstein distance $W_{1}$. 11(a) Three scalar functions $f_{1},f_{2},f_{3}$ in the synthetic data set. 11(b) Regions corresponding to the maxima and arcs incident on them in the merge trees of $f_{1},f_{2},f_{3}$. Each region is assigned a unique color. 11(c) Mapping determined by $W_{1}$ between ($f_{1},f_{2}$) and between ($f_{2},f_{3}$). 11(d) Mapping determined by the tree edit distance $D$ between ($f_{1},f_{2}$) and between ($f_{2},f_{3}$). Merge tree nodes and their corresponding spatial regions have the same color. We construct three synthetic datasets to understand the difference between the tree edit distance $D$ and other well known distances between topological structures. The scalar functions $f_{1},f_{2},f_{3}$ are sums of gaussians whose extrema are fixed in space. The scalar values change in a controlled manner for the three functions so that the values at the extrema increase / decrease monotonically as we step from $f_{1}$ to $f_{2}$ to $f_{3}$. We compute the tree edit distances together with the corresponding mapping for each pair $(f_{1},f_{2})$ and $(f_{2},f_{3})$. Figure 11 shows the three scalar functions $f_{1},f_{2},f_{3}$. We observe in Figure 11(d) that $D$ establishes intuitively correct mappings. The mappings also preserve the tree hierarchy. On the other hand, the Wasserstein distance $W_{1}$ (Figure 11(c), left) maps the brown regions to the null character. The tree edit distance prefers the relabel over a sequence of delete-insert operations. The reason $W_{1}$ does not find the correspondence between the two brown nodes is because their birth-death intervals do not overlap. As a result, these nodes are mapped to the null character _i.e._ , inserted or deleted. The intervals corresponding to the two nodes in question are $[1.17,1.39]$ in $f_{1}$ and $[1.06,1.08]$ in $f_{2}$. We also see from Figure 11(c) that $W_{1}$ maps the brown region to the magenta region, thereby mapping nodes that lie within different subtrees. This also causes a pair of nodes being mapped to the null character. The tree edit distance $D$ is constrained to map disjoint subtrees to disjoint subtrees and establishes a better mapping. To summarize, $D$ in general establishes mappings that are better than $D_{B}$ and $W_{1}$ because it is aware of the structure of the merge tree and preserves the hierarchy captured in the tree. ### 5.2 Comparison with other distance measures $0$$50$$100$$150$$200$$0$$0.5$$1$TED $D$$W_{1}$$D_{B}$$0$$50$$100$$150$$200$$0$$0.5$$1$$\varepsilon$ = 0%$\varepsilon$ = 1.5%$\varepsilon$ = 5%$0$$50$$100$$150$$200$$0$$0.5$$1$$\varepsilon$ = 5%$\varepsilon$ = 10%$\varepsilon$ = 20%$0$$50$$100$$150$$200$$0$$0.5$$1$Timesteps$\varepsilon$ = 20%$\varepsilon$ = 50%$\varepsilon$ = 100%$W_{1}$ Figure 12: Comparing distance measures on the von Kármán vortex street dataset. (top) Plot of distance measures between the first time step and others when stability parameter $\varepsilon$ is set to $0$ and comparison with the Wasserstein distance $W_{1}$ and bottleneck distance $D_{B}$. (rows 2-4) Effect of stabilization parameter $\varepsilon=0,1.5,5,10,20,50,100\%$ and comparison with Wasserstein distance $W_{1}$. Results are shown in three plots to reduce clutter. We compare the proposed tree edit distance measure $D$ with existing measures such as bottleneck distance $D_{B}$ and Wasserstein distance $W_{1}$ via computational experiments on the 2D Bénard-von Kármán vortex street dataset [49]. Figure 13(a) shows a few time steps of the data, which represents flow around a cylinder. The dataset contains the velocity magnitude on a $400\times 50$ grid over $1001$ time steps. Each split tree contains approximately $55-65$ nodes. We calculate $D$ and plot it together with the bottleneck and Wasserstein distance, see top row of Figure 12. The tree edit distance $D$ is always greater than $W_{1}$ and $D_{B}$. Indeed, $D$ is likely to be more discriminative than $W_{1}$ and $D_{B}$ because it incorporates the structure of the merge tree in addition to the persistence pairs. We also compute and plot $D$ for increasing values of the stability parameter $\varepsilon$. The values of $\varepsilon$ are reported as a percentage of the maximum persistence of the particular dataset. While there are some anomalies for small values of $\varepsilon$, in general we observe in Figure 12 that with increase in $\varepsilon$, $D$ tends towards $W_{1}$. For a high enough value of $\varepsilon$, $D$ becomes almost equal to $W_{1}$. The reason for this behavior is that the bottleneck/Wasserstein distance does not consider the structure of the trees. Increasing the stability parameter transforms the tree to become more like a bush. Finally, all the nodes become children of the root thereby simplifying and eliminating the tree structure. Varying $\varepsilon$ from $0-5\%$ results in a decrease of up to $25$ nodes in the split tree. Further increasing $\varepsilon$ led to an additional reduction by only $1-2$ nodes. We observe this trend in the distance plots also. ### 5.3 Periodicity in time-varying data (a) Three time steps from the flow around a cylinder simulation. (b) Distance matrix highlights the periodicity. Figure 13: 13(a) Time step 0 (top), 37 (middle) and 74 (bottom) of the von Kármán vortex street dataset. The split tree and critical points are overlaid. 13(b) A truncated version of the DM showing the tree edit distance measure between all pairs of time steps. Blue bands indicate periodicity with time period 74-75. A half period of 37, corresponding to the alternating nature of vortex shedding, is also visible. Earlier studies of the Bénard-von Kármán vortex street dataset have successfully identified periodicity in the dataset. Narayanan et al. [22] detect both a half period of 38 and the full period of 75. We also aim to identify periodicity. Towards this, we compare the split tree of time step $1$ with the remaining $1000$ time steps of the dataset. We plot the tree edit distance for time steps $1-220$, see top plot of Figure 12. We rerun the experiment and compare all $1000$ time steps with all other time steps. The distances are stored in a distance matrix (DM). Each split tree contains approximately $55-65$ nodes. The distances were computed in parallel using 12 threads and took approximately $25$ minutes. A truncated version is shown in Figure 13(b) for clarity. From Figure 13(b), we can also observe a periodicity of 37, which matches with the results reported by Narayanan et al. [22]. The tree edit distance was computed in this experiment without stabilization. ### 5.4 Topological effects of subsampling and smoothing (a) $f_{1}$ (b) subsampled $f_{1}$ (c) smoothened $f_{1}$ (d) $f_{2}$ (e) subsampled $f_{2}$ (f) smoothened $f_{2}$ (g) DM for $f_{1}$, original and subsampled (h) DM for $f_{2}$, original and subsampled (i) DM for $f_{2}$, $\varepsilon=0.5\%$ (j) DM for $f_{1}$, original and smoothened (k) DM for $f_{2}$, original and smoothened (l) DM for $f_{2}$, $\varepsilon=0.5\%$ Figure 14: Measuring the effect of subsampling and smoothing. 14(a),14(d) Two synthetic functions sampled over a $300\times 300$ grid. 14(b),14(e) Subsampled down to $30\times 30$ over $9$ iterations. 14(c),14(f) Smoothed in $9$ iterations. 14(g)-14(i) DMs showing distance between all pairs of subsampled datasets without and with stabilization. 14(j)-14(l) DMs showing distances for all pairs of smoothed datasets. Row and column indices correspond to the iteration number, $0$ corresponds to the original, $9$ corresponds to the lowest resolution/extreme smoothing. Red indicates high and blue indicates low values. Colormaps for $f_{1}$ and $f_{2}$ are not on the same scale. The size of datasets are ever increasing and this mandates the use of subsampling and/or smoothing of the data as a preprocessing step. The aim of this preprocessing is to reduce the data size while ensuring a limited effect on geometric accuracy. However, the effect on the topological features of the scalar field is often not quantified. We want to observe how the tree edit distance measure captures these topological effects. We consider two synthetically generated datasets of size $300\times 300$ (iteration $0$), see Figures 14(a), 14(d). The data is downsampled over $9$ iterations to a $30\times 30$ grid by reducing the number of samples in each dimension by $30$ within each iteration. We also apply $9$ iterations of laplacian smoothing on both $300\times 300$ datasets. Next, we compare all merge trees corresponding to the subsampled and smoothed datasets pairwise. The distance matrix (DM) for the function $f_{1}$ indicates that the distances are monotonic, which conforms to the expected behavior. But we see a different pattern in the case of function $f_{2}$. A small stabilization applied on $f_{2}$ with $\varepsilon=0.5\%$ results in distance matrices that conform to the expected behavior. This indicates that the stabilization may indeed be required, particularly when the scalar functions contain flat regions and multi-saddles. In both datasets, we notice that the distances between the lowest resolution ($30\times 30$) dataset and others is relatively high. We identified two reasons for the high values. First, the number of critical points reduces significantly between iterations $8$ and $9$. For example, in the case of $f_{2}$, it goes down from $66-70$ in earlier iterations to $58$ in iteration $9$. Second, the function value at the critical points in the lowest resolution dataset are also different. Hence, the relabel costs increase significantly, up to a factor of 1.5 in some cases. ### 5.5 Detecting symmetry / asymmetry Identifying symmetric or repeating patterns in scalar fields enables feature- directed visualization. For example, it supports applications such as symmetry-aware transfer function design for volume rendering, anomaly detection, and query-driven exploration. A distance measure is central to any method for identifying symmetry. Consider the synthetic dataset in Figure 15 that contains six regions corresponding to six subtrees of the merge tree. Four regions colored green in Figure 15(b) are symmetric copies. The remaining two regions, colored orange and magenta, are slightly perturbed to cause asymmetry. We compute the tree edit distance measure to compare each subtree corresponding to a region with other subtrees. The measure clearly distinguishes between symmetric and asymmetric regions as can be seen from the distance matrix (DM) in Figure 15(c). These results are consistent with the premise upon which the data is generated. We present additional case studies that demonstrate the applicability of the tree edit distance measure to symmetry identification. EMDB111https://www.ebi.ac.uk/pdbe/emdb/ contains 3D electron microscopy density data of macromolecules, subcellular structures, and viruses. Some of these structures contain symmetric subunits. We study two structures, EMDB 1654, and 1897, see Figure 16. First, we compute the split tree for each structure. We then use a semi-automated method to extract sub-trees corresponding to significant features from the merge tree based on user specified persistence and minimum scalar value thresholds. Next, we compute the tree edit distance measure between sub-trees corresponding to these regions of interest. We observe two distinct groups from the DMs. The tree edit distance measure clearly identifies two groups with $4$ and $8$ regions each in EMDB 1654, see Figure 16(a). Similarly, it identifies two groups containing $3$ and $6$ symmetric regions each in EMDB 1897, see Figure 16(b). (a) Synthetic field (b) Segmentation (c) Distance matrix Figure 15: Identifying symmetry and asymmetry. 15(a) Sum of 2D gaussians. 15(b) The DM indicates presence of a symmetric group containing 4 regions. Two regions are correctly identified as being different from the rest. 15(c) DM between various subtrees of the merge tree. (a) EMDB 1654 (b) EMDB 1897 Figure 16: Detecting groups of symmetric regions in EMDB datasets. (centre) DM showing tree edit distance between various pairs of subtrees of the merge tree. Low values are mapped to blue and high values to red. The DM indicates the presence of two distinct groups. All regions within a group are symmetric copies of each other. (left, right) Volume rendering where one region from each symmetric group is highlighted. ### 5.6 Shape matching Shape matching involves comparing geometric shapes and finding similarity between them. A good distance measure helps quantify this notion of similarity more concretely. The TOSCA non-rigid world dataset222http://tosca.cs.technion.ac.il/book/resources_data.html contains a set of different shapes, see Figure 17. The shapes are in different poses and the project aims to develop methods to identify similarity between shapes in a pose invariant manner. We compute the average geodesic distance field [3] on the surface mesh. This field is well studied in the literature and is known to be a good shape descriptor. We apply a persistence simplification threshold of 1% on the merge trees both to remove topological noise and to reduce the number of nodes. Next, we compute the tree edit distance measure between all pairs of shapes. It takes around $15$ seconds to generate the distance matrix with the same setup used for the periodicity experiment. Figure 18 shows the distance matrix. Each collection of shape appears as a blue block irrespective of variations in pose. We also observe higher values for a pair of shapes that are different. Note the blue blocks away from the diagonal. They correspond to Michael vs Victoria, David vs Victoria, David vs Michael, and David vs Gorilla. These pairs have similar shapes, which is more apparent in a few poses. Not all poses are shown in Figure 17. Figure 17: Collection of shapes from the TOSCA non-rigid world dataset. The average geodesic distance field [3] is computed on the surface. Each shape is available in multiple poses (number of poses mentioned within parenthesis), only one pose is shown here. Figure 18: Tree edit distance matrix for all pairs of shapes from the TOSCA non-rigid world dataset. Blocks of low values (blue) correspond to similar shapes but in different poses. ### 5.7 Data Summarization Exploring large scientific data, particularly time-varying data, and identifying patterns of interest is often time consuming even with good visualization tools. Well designed abstract representations provide good overviews of the data and direct the user to features of interest. Abstractions such as the merge trees present a summary of spatial features. Temporal summaries enable effective visualization of time-varying data. Central to the design of a temporal summary is a good distance measure that can distinguish between periods of significant activity and inactive time periods. In this experiment, we consider the 3D Bénard-von Kármán vortex street dataset. The velocity magnitude is available as a scalar field on a $192\times 64\times 48$ grid over $102$ time steps [50]. Figure 19 shows volume renderings and isosurfaces for a few time steps. Topological features of the velocity magnitude scalar field are represented using the split tree. Each split tree had approximately $180-200$ nodes. We compute the tree edit distance between all pairs of time steps. It takes around $4$ seconds to generate the distance matrix (DM) with the same setup used for the periodicity experiment. Figure 19: The 3D Bénard-von Kármán vortex street dataset. (top) Volume rendering of the velocity magnitude field for time steps $15,35,58,91,98$ ordered left to right. (bottom) Isosurfaces at isovalue $0.7$ extracted for the above time steps. Figure 20: Tree edit distance matrix for all time steps of the 3D Bénard-von Kármán vortex street dataset. Columns corresponding to time steps $15,35,58,91,98$ are highlighted. Patterns that help in generating a temporal summary are highlighted using black and green boxes. The DM shown in Figure 20 contains multiple patterns. A fluid dynamics expert helped study and interpret the results. The distance between time steps $2-28$ are small because the flow does not contain any vortices and the features do not change. The top left blue block in the matrix corresponds to this time period. This is followed by the period when new vortex structures are formed (small block highlighted in green that corresponds to time steps $29-39$). Next, the vortices exhibit shedding, which is shown by the repeating patterns present in the larger green block in the matrix (time steps $40-85$). Finally, the vortices are significantly distorted, which is captured by the high values of distance in the bottom right block. Thus we can use the patterns that emerge in the distance matrix to distinguish between different types of behavior and summarize the scientific phenomena using these patterns. ## 6 Conclusions We described a distance measure between two scalar fields that compares their merge trees. The distance measure is defined as the minimum cost of a set of restricted edit operations that transforms one tree into another. The edit operations and the associated costs are both intuitive and mathematically sound. The measure satisfies metric properties, can be efficiently computed, and is useful in practice. We study the properties of the measure and demonstrate its application to data analysis and visualization using various computational experiments. In future work, we plan to develop a theoretical analysis of the stability properties of the measure. Developing a comparative visualization framework based on the tree edit distance measure is also an interesting problem with potential applications to time-varying data and multifield data visualization. ## Acknowledgments This work is supported by the Department of Science and Technology, India (DST/SJF/ETA-02/2015-16), and Joint Advanced Technology Programme, Indian Institute of Science (JATP/RG/PROJ/2015/16), and the Robert Bosch Centre for Cyber Physical Systems, Indian Institute of Science. We thank Shrisha Rao for discussions on the data summarization experiment. ## References * [1] H. Carr, J. Snoeyink, and U. Axen, “Computing contour trees in all dimensions,” _Computational Geometry_ , vol. 24, no. 2, pp. 75–94, 2003\. * [2] H. Saikia and T. Weinkauf, “Global feature tracking and similarity estimation in time-dependent scalar fields,” in _Computer Graphics Forum_ , vol. 36, no. 3, 2017, pp. 1–11. * [3] M. Hilaga, Y. Shinagawa, T. Kohmura, and T. L. Kunii, “Topology matching for fully automatic similarity estimation of 3d shapes,” in _Proceedings of the 28th annual conference on Computer graphics and interactive techniques_. ACM, 2001, pp. 203–212. * [4] D. M. Thomas and V. Natarajan, “Symmetry in scalar field topology,” _IEEE Trans. Vis. Comp. Graphics_ , vol. 17, no. 12, pp. 2035–2044, 2011\. * [5] ——, “Detecting symmetry in scalar fields using augmented extremum graphs,” _IEEE Trans. Vis. Comp. Graphics_ , vol. 19, no. 12, pp. 2663–2672, 2013. * [6] T. B. Masood, D. M. Thomas, and V. Natarajan, “Scalar field visualization via extraction of symmetric structures,” _The Visual Computer_ , vol. 29, no. 6-8, pp. 761–771, 2013. * [7] H. Saikia, H. P. Seidel, and T. Weinkauf, “Extended branch decomposition graphs: Structural comparison of scalar data,” _Computer Graphics Forum_ , vol. 33, no. 3, pp. 41–50, 2014. * [8] D. M. Thomas and V. Natarajan, “Multiscale symmetry detection in scalar fields by clustering contours,” _IEEE Trans. Vis. Comp. Graphics_ , vol. 20, no. 12, pp. 2427–2436, 2014. * [9] P. Oesterling, C. Heine, G. H. Weber, D. Morozov, and G. Scheuermann, “Computing and Visualizing Time-Varying Merge Trees for High-Dimensional Data,” in _TopoinVis IV_. Springer, 2017, pp. 87–101. * [10] N. Sauber, H. Theisel, and H.-P. Seidel, “Multifield-graphs: An approach to visualizing correlations in multifield scalar data,” _IEEE Trans. Vis. Comp. Graphics_ , vol. 12, no. 5, pp. 917–924, 2006. * [11] S. Nagaraj, V. Natarajan, and R. S. Nanjundiah, “A gradient-based comparison measure for visual analysis of multifield data,” in _Computer Graphics Forum_ , vol. 30, no. 3, 2011, pp. 1101–1110. * [12] I. Demir, C. Dick, and R. Westermann, “Multi-charts for comparative 3d ensemble visualization,” _IEEE Trans. Vis. Comp. Graphics_ , vol. 20, no. 12, pp. 2694–2703, 2014. * [13] S. Dutta, J. Woodring, H.-W. Shen, J.-P. Chen, and J. Ahrens, “Homogeneity guided probabilistic data summaries for analysis and visualization of large-scale data sets,” in _PacificVis_ , 2017, pp. 111–120. * [14] D. Cohen-Steiner, H. Edelsbrunner, and J. Harer, “Stability of persistence diagrams,” _Disc. Comp. Geom._ , vol. 37, no. 1, pp. 103–120, 2007. * [15] D. Morozov, K. Beketayev, and G. Weber, “Interleaving distance between merge trees,” _Discrete and Computational Geometry_ , vol. 49, no. 52, pp. 22–45, 2013. * [16] K. Beketayev, D. Yeliussizov, D. Morozov, G. H. Weber, and B. Hamann, “Measuring the distance between merge trees,” in _TopoinVis III_. Springer, 2014, pp. 151–165. * [17] G. Reeb, “Sur les points singuliers d’une forme de Pfaff completement intégrable ou d’une fonction numérique,” _CR Acad. Sci. Paris_ , vol. 222, no. 847-849, p. 2, 1946. * [18] U. Bauer, X. Ge, and Y. Wang, “Measuring distance between Reeb graphs,” in _Proc. 13th Symp. Comp. Geom_. ACM, 2014, pp. 464–474. * [19] P. K. Agarwal, K. Fox, A. Nath, A. Sidiropoulos, and Y. Wang, “Computing the Gromov-Hausdorff distance for metric trees,” in _Int. Symp. Alg. Comp._ Springer, 2015, pp. 529–540. * [20] B. Di Fabio and C. Landi, “The edit distance for Reeb graphs of surfaces,” _Disc. Comp. Geom._ , vol. 55, no. 2, pp. 423–461, 2016. * [21] T. Dey, D. Shi, and Y. Wang, “Comparing Graphs via Persistence Distortion,” _Proc. 31rd Annu. Sympos. Comput. Geom._ , pp. 491–506, 2015. * [22] V. Narayanan, D. M. Thomas, and V. Natarajan, “Distance between extremum graphs,” in _PacificVis_ , 2015, pp. 263–270. * [23] H. Saikia, H.-P. Seidel, and T. Weinkauf, “Fast similarity search in scalar fields using merging histograms,” in _TopoInVis 2015_ , 2015, pp. 1–14. * [24] T. A. Gillette, P. Hosseini, and G. A. Ascoli, “Topological characterization of neuronal arbor morphology via sequence representation: II - global alignment,” _BMC Bioinformatics_ , vol. 16, no. 1, pp. 209–226, 2015. * [25] P. Klein, S. Tirthapura, D. Sharvit, and B. Kimia, “A tree-edit-distance algorithm for comparing simple, closed shapes,” in _Proc. 11th annual ACM-SIAM symposium on Discrete algorithms_ , 2000, pp. 696–704. * [26] M. McVicar, B. Sach, C. Mesnage, J. Lijffijt, E. Spyropoulou, and T. De Bie, “SuMoTED: An intuitive edit distance between rooted unordered uniquely-labelled trees,” _Pattern Recognition Letters_ , vol. 79, pp. 52–59, 2016. * [27] D. Fukagawa, T. Tamura, A. Takasu, E. Tomita, and T. Akutsu, “A clique-based method for the edit distance between unordered trees and its application to analysis of glycan structures,” _BMC bioinformatics_ , vol. 12, no. Suppl 1, p. S13, 2011. * [28] S. Schirmer and R. Giegerich, “Forest alignment with affine gaps and anchors, applied in RNA structure comparison,” _Theoretical Computer Science_ , vol. 483, pp. 51–67, 2013. * [29] P. Ferraro and C. Godin, “A distance measure between plant architectures,” _Annals of Forest Science_ , vol. 57, no. 5, pp. 445–461, 2000. * [30] K. Zhang and D. Shasha, “Simple Fast Algorithms for the Editing Distance between Trees and Related Problems,” _SIAM Journal on Computing_ , vol. 18, no. 6, pp. 1245–1262, 1989. * [31] K. Zhang, “Algorithms for the Constrained Editing Distance between Ordered Labeled Trees and related Problems,” _Pattern Recognition_ , vol. 28, no. 3, pp. 463–474, 1995. * [32] K. Zhang, R. Statman, and D. Shasha, “On the editing distance between unordered labeled trees,” _Information Processing Letters_ , vol. 42, no. 3, pp. 133–139, 1992. * [33] K. Zhang, “A Constrained Edit Distance Between Unordered Labeled Trees,” _Algorithmica_ , vol. 15, pp. 205–222, 1996. * [34] H. Touzet, “Tree edit distance with gaps,” _Information Processing Letters_ , vol. 85, no. 3, pp. 123–129, 2003. * [35] H. Xu, “An algorithm for comparing similarity between two trees,” _arXiv preprint arXiv:1508.03381_ , 2015. * [36] B. Rieck, H. Leitte, and F. Sadlo, “Hierarchies and ranks for persistence pairs,” in _TopoInVis 2017_ , 2017. * [37] R. Sridharamurthy, A. Kamakshidasan, and V. Natarajan, “Edit distances for comparing merge trees,” in _IEEE SciVis Posters_ , 2017. * [38] H. Edelsbrunner, D. Letscher, and A. Zomorodian, “Topological persistence and simplification,” in _Foundations of Computer Science_. IEEE, 2000, pp. 454–463. * [39] D. Morozov and G. Weber, “Distributed merge trees,” in _Proc. ACM SIGPLAN Symposium on Principles and Practice of Parallel Programming_ , ser. PPoPP ’13, 2013, pp. 93–102. * [40] A. Acharya and V. Natarajan, “A parallel and memory efficient algorithm for constructing the contour tree,” in _PacificVis_ , 2015, pp. 271–278. * [41] P. Bille, “A survey on tree edit distance and related problems,” _Theoretical Computer Science_ , vol. 337, no. 1-3, pp. 217–239, 2005. * [42] T. H. Cormen, C. E. Leiserson, R. L. Rivest, and C. Stein, _Introduction to Algorithms_ , 3rd ed. MIT Press, 2009\. * [43] M. Levandowsky and D. Winter, “Distance between sets,” _Nature_ , vol. 234, no. 5323, p. 34, 1971. * [44] C. Villani, _Optimal Transport: Old and New_. Springer-Verlag Berlin Heidelberg, 2000. * [45] H. W. Kuhn, “The hungarian method for the assignment problem,” _Naval Research Logistics (NRL)_ , vol. 2, no. 1-2, pp. 83–97, 1955. * [46] H. Doraiswamy and V. Natarajan, “Computing Reeb graphs as a union of contour trees,” _IEEE Trans. Vis. Comp. Graphics_ , vol. 19, no. 2, pp. 249–262, 2013. * [47] J. Ahrens, B. Geveci, C. Law, C. Hansen, and C. Johnson, “Paraview: An end-user tool for large-data visualization,” _The Visualization Handbook_ , vol. 717, 2005. * [48] J. Tierny, G. Favelier, J. A. Levine, C. Gueunet, and M. Michaux, “The Topology ToolKit,” _IEEE Trans. Vis. Comp. Graphics_ , vol. 24, no. 1, pp. 832–842, 2018, https://topology-tool-kit.github.io/. * [49] T. Weinkauf and H. Theisel, “Streak lines as tangent curves of a derived vector field,” _IEEE Trans. Vis. Comp. Graphics_ , vol. 16, no. 6, pp. 1225–1234, 2010. * [50] W. von Funck, T. Weinkauf, H. Theisel, and H.-P. Seidel, “Smoke surfaces: An interactive flow visualization technique inspired by real-world flow experiments,” _IEEE Trans. Vis. Comp. Graphics_ , vol. 14, no. 6, pp. 1396–1403, 2008. | Raghavendra Sridharamurthy is a PhD candidate in computer science at Indian Institute of Science, Bangalore. He received BE degree in information technology from National Institute of Technology Karnataka, Surathkal and M.Sc degree in computer science from Indian Institute of Science. His research interests include scientific visualization, computational topology and its applications. ---|--- | Talha Bin Masood is a PhD candidate in computer science at Indian Institute of Science, Bangalore. He received B.Tech degree from Aligarh Muslim University and ME degree in computer science from Indian Institute of Science. His research interests include scientific visualization, computational geometry, computational topology and its applications to various scientific domains. ---|--- | Adhitya Kamakshidasan is a Junior Research Fellow at Visualization and Graphics Lab, Indian Institute of Science. He holds a B.Tech degree in computer science from Visvesvaraya National Institute of Technology, Nagpur. His research interests include fluid simulation, information visualization and cartography. ---|--- | Vijay Natarajan is an associate professor in the Department of Computer Science and Automation at the Indian Institute of Science, Bangalore. He received the Ph.D. degree in computer science from Duke University in 2004. His research interests include scientific visualization, computational topology, and geometry processing. ---|---
Higher order hesitant fuzzy Choquet integral operator and its application to multiple criteria decision making B. Farhadinia111Corresponding author. U. Aickelin† H.A. Khorshidi† Dept. Math., Quchan University of Technology, Iran. <EMAIL_ADDRESS> † Dept. Computing and Information Systems, University of Melbourne, Australia. <EMAIL_ADDRESS> <EMAIL_ADDRESS> Abstract. Generally, the criteria involved in a decision making problem are interactive or inter-dependent, and therefore aggregating them by the use of traditional operators which are based on additive measures is not logical. This verifies that we have to implement fuzzy measures for modelling the interaction phenomena among the criteria. On the other hand, based on the recent extension of hesitant fuzzy set, called higher order hesitant fuzzy set (HOHFS) which allows the membership of a given element to be defined in forms of several possible generalized types of fuzzy set, we encourage to propose the higher order hesitant fuzzy (HOHF) Choquet integral operator. This concept not only considers the importance of the higher order hesitant fuzzy arguments, but also it can reflect the correlations among those arguments. Then, a detailed discussion on the aggregation properties of the HOHF Choquet integral operator will be presented. To enhance the application of HOHF Choquet integral operator in decision making, we first assess the appropriate energy policy for the socio-economic development. Then, the efficiency of the proposed HOHF Choquet integral operator-based technique over a number of exiting techniques is further verified by employing another decision making problem associated with the technique of TODIM (an acronym in Portuguese of Interactive and Multicriteria Decision Making). Keywords: Higher order hesitant fuzzy set (HOHFS), Choquet integral, Multiple criteria decision making. ## 1 Introduction The theory of extending and generalizing fuzzy sets is very prospective tool in different research domains. One of interesting extensions is the concept of hesitant fuzzy set (HFS) which was proposed by Torra and Narukawa [28]. HFS is quite suit for the case where the decision maker has a set of possible values, rather than a margin of error (like that in intuitionistic fuzzy sets [2]) or some possibility distribution on the possible values (like that is considered in defining type-2 fuzzy sets [19]). By the way, HFS [28] and its extension which is called the generalized HFS (G-HFS) [22], have their own drawbacks because of the reason that the membership degrees of an element to a given set can be expressed only by crisp numbers or intuitionistic fuzzy sets. The evidence to this fact is that in a real and practical decision making problem where the information provided by a group of experts may not be described only by hesitant fuzzy sets (HFSs) or just by one kind of HFS extensions. A more specific situation occurs when a number of experts are intended to return their evaluations in the form of hesitant multiplicative sets, and others are interested to describe their evaluation by the help of hesitant fuzzy linguistic term sets. Therefore, it is difficult for the decision maker to provide exact hesitant multiplicative set or hesitant fuzzy linguistic term set for the membership degrees. Such a difficulty is avoided by considering the higher order HFS (HOHFS) [10] in order to describe the membership degrees. The concept of HOHFS not only encompasses fuzzy sets, intuitionistic fuzzy sets, type 2 fuzzy sets and HFSs, but also it extends the concept of generalized hesitant fuzzy set (G-HFS) [22]. It does worth to say that the topic of developing information aggregations in fuzzy multiple criteria decision making (MCDM) has been an interesting research topic [8, 9, 11, 24, 30] so far. As it is known, most of the existing fuzzy aggregation operators only consider those situations in which all the elements are independent. Moreover, in many real-world situations, we observe that the elements are usually interactive or interdependent, that is, they are correlative. The Choquet integral [7] with respect to fuzzy measures [25] is an aggregation function that has been employed successfully in MCDM problems where it determines the relative importance of decision criteria as well as their interactions [17]. There exist a growing number of studies on the Choquet integral, and it has been investigated by many researchers. Yager [41] proposed the induced Choquet ordered averaging operator, and moreover, he introduced the concept of Choquet aggregation based on the order induced aggregation. By extending the concept of Choquet integral to that for fuzzy context, Meyer and Roubens [18] applied this concept to multiple criteria decision making problems. Yu et al. [35] defined a hesitant fuzzy aggregation operator by taking the concept of Choquet integral into account, and they implemented it in modelling a hesitant-based multiple criteria decision making problem. Tan and Chen [27] and Tan [26] proposed respectively the intuitionistic and the interval-valued intuitionistic fuzzy Choquet integral operators. Bustince et al. [3] investigated a multiple criteria decision making technique by considering the concept of interval-valued Choquet integral. Wang et al. [31] investigated multiple criteria group decision making problems in which the Choquet integral aggregation operators together with interval two-tuple linguistic information play the main role. Some more studies have been given by Abdullah et al. [1], Candeloro et al. [4], Pasi et al. [20], etc. By the way, inspired by the concept of Choquet integral, we will develop here a higher order hesitant fuzzy Choquet integral operator for aggregating higher order hesitant fuzzy (HOHF) information in MCDM, and investigate different properties of this operator. As will be shown in Section 3, such a proposed concept not only is well applicable to the case where the decision makers have a hesitation among several possible memberships with uncertainties in a general form (i.e., HOHFSs), but also it models the interaction phenomena among the criteria (i.e., the Choquet integral property). The present paper is organized as follows: The concept of higher order HFS (HOHFS) as an extension of HFSs is reviewed in Section 2, and then the higher order hesitant fuzzy (HOHF) Choquet integral operator is presented together with a detailed discussion on the aggregation properties of the HOHF Choquet integral operator. In Section 3, we apply the proposed HOHF Choquet integral operator to MCDM problems involving the higher order hesitant fuzzy information, and then a comparative analysis is then given to demonstrate the effectiveness of the proposed operator. Finally, conclusion is drown in Section 4. ## 2 The HOHF Choquet integral operator Among the several extensions of the theory of HFSs, we deal in brief with the higher order hesitant fuzzy set (HOHFS) which was first introduced by Farhadinia [10]. ###### Definition 2.1. [10] Let $X$ be the universe of discourse. A generalized type of fuzzy set (G-Type FS) on $X$ is defined as $\displaystyle\widetilde{A}=\\{\langle x,\widetilde{A}(x)\rangle:x\in X\\},$ (1) where $\displaystyle\widetilde{A}:X\rightarrow\psi([0,1]).$ Here, $\psi([0,1])$ denotes a family of crisp or fuzzy sets that can be defined with in the universal set $[0,1]$. We mention here some special cases of G-Type FS which are encountered by Klir and Yuan in [15]: * • if $\psi([0,1])=[0,1]$, then the G-Type FS $\widetilde{A}$ reduces to an ordinary fuzzy set; * • if $\psi([0,1])=\varepsilon([0,1])$ denoting the set of all closed intervals, then the G-Type FS $\widetilde{A}$ reduces to an interval-valued fuzzy set; * • if $\psi([0,1])={\cal F}([0,1])$ denoting the set of all ordinary fuzzy sets, then the G-Type FS $\widetilde{A}$ reduces to a type II fuzzy set; * • if $\psi([0,1])=L$ denoting a partially ordered Lattice, then the G-Type FS $\widetilde{A}$ reduces to a lattice fuzzy set. ###### Definition 2.2. [28] Let $X$ be the universe of discourse. A hesitant fuzzy set (HFS) on $X$ is symbolized by $H=\\{\langle x,h(x)\rangle:x\in X\\},$ where $h(x)$, referred to as the hesitant fuzzy element (HFE), is a set of some values in $[0,1]$ denoting the possible membership degree of the element $x\in X$ to the set $H$. As can be seen from Definition 2.2, HFS expresses the membership degrees of an element to a given set only by several real numbers between 0 and 1, while in many real-world situations assigning exact values to the membership degrees does not describe properly the imprecise or uncertain decision information. Thus, it seems to be difficult for the decision makers to rely on HFSs for expressing uncertainty of an element. To overcome the difficulty associated with expressing uncertainty of an element to a given set, the concept of HOHFS allows the membership degrees of an element to be expressed by several possible G-Type FSs. ###### Definition 2.3. [10] Let $X$ be the universe of discourse. A higher order hesitant fuzzy set (HOHFS) on $X$ is defined in terms of a function when it is applied to $X$, it then returns a set of G-Type FSs. A HOHFS is denoted by $\displaystyle\widetilde{H}=\\{\langle x,\widetilde{h}(x)\rangle:x\in X\\},$ (2) where $\widetilde{h}(x)$, referred to as the higher order hesitant fuzzy element (HOHFE), is a set of some G-Type FSs denoting the possible membership degree of the element $x\in X$ to the set $\widetilde{H}$. In this regards, the HOHFS $\widetilde{H}$ is also represented as $\displaystyle\widetilde{H}=\\{\langle x,\\{\widetilde{h_{\widetilde{h}}}^{(1)}(x),...,\widetilde{h_{\widetilde{h}}}^{(|\widetilde{h}(x)|)}(x)\\}\rangle:x\in X\\},$ where all $\widetilde{h_{\widetilde{h}}}^{(1)}(x),...,\widetilde{h_{\widetilde{h}}}^{(|\widetilde{h}(x)|)}(x)$ are G-Type FSs on $X$, and $|\widetilde{h}(x)|$ is the number of G-type FSs in $\widetilde{h}(x)$. The noteworthy feature about Definition 2.3 is that it encompasses all the above-mentioned generalizations of ordinary fuzzy sets which have been discussed in detail by Farhadinia [10]. Having introduced HOHFEs of a HOHFS $\widetilde{H}=\\{\langle x,\widetilde{{h}}(x)\rangle:x\in X\\}$, we now turn our attention to the interpretation of HOHFS $\widetilde{H}$ as the union of all HOHFEs i.e., $\displaystyle\widetilde{H}=\bigcup_{\widetilde{{h}}(x)\in\widetilde{H}}\\{\widetilde{{h}}(x)\\}$ (3) which is fundamental in the study of HOHFS aggregation operators within the next parts of the paper. Hereafter, to simplify the notation, we use only $\widetilde{{h}}$ instead of $\widetilde{{h}}(x)$ in the following descriptions, that is, we assume that $\displaystyle\widetilde{H}=\bigcup_{\widetilde{{h}}\in\widetilde{H}}\\{\widetilde{{h}}\\}.$ ###### Definition 2.4. Let $\widetilde{H}_{1}=\bigcup_{\widetilde{{h}}_{1}\in\widetilde{H}_{1}}\\{\widetilde{{h}_{1}}\\}$ and $\widetilde{H}_{2}=\bigcup_{\widetilde{{h}}_{2}\in\widetilde{H}_{2}}\\{\widetilde{{h}_{2}}\\}$ be two HOHFSs. We give the definition of operation $\odot$ on HOHFSs based on its counterpart on HOHFEs as follows: $\displaystyle\widetilde{H}_{1}\odot\widetilde{H}_{2}=\bigcup_{\widetilde{h_{1}}\in\widetilde{H_{1}},\widetilde{h_{2}}\in\widetilde{H_{2}}}\\{\widetilde{{h_{1}}}\odot\widetilde{{h_{2}}}\\}=\bigcup_{\widetilde{h_{1}}\in\widetilde{H_{1}},\widetilde{h_{2}}\in\widetilde{H_{2}}}\\{\bigcup_{\widetilde{h_{h1}}\in\widetilde{h_{1}},\widetilde{h_{h2}}\in\widetilde{h_{2}}}\\{\widetilde{{h_{h1}}}\odot\widetilde{{h_{h2}}}\\}~{}\\},$ (4) where the last operation $\odot$ stands for the operation on G-type FSs. Note that, if there is no confusion, all the operations performing over HOHFSs, HOHFEs and G-type FSs are denoted by the same symbol. Furthermore, that property holds for G-type FSs is denoted by $\widetilde{(p_{G})}$, and that is defined on HOHFEs is symbolized by $\widetilde{(p)}$. ###### Theorem 2.5. Let $\widetilde{h}_{1}=\bigcup_{\widetilde{{h}}_{h1}\in\widetilde{h}_{1}}\\{\widetilde{{h}_{h1}}\\}$ and $\widetilde{h}_{2}=\bigcup_{\widetilde{{h}}_{h2}\in\widetilde{h}_{2}}\\{\widetilde{{h}_{h2}}\\}$ be two HOHFEs. If the operation $\odot$ on G-type FSs has the property $\widetilde{(p_{G})}$, then the operation $\odot$ on HOHFEs has the counterpart property, denoted by $\widetilde{(p)}$. Proof. The proof is obvious from Definition 2.4 Suppose that, for example, the operation $\odot$ on G-type FSs is commutative, that is, for any G-type FSs $\widetilde{h_{h1}}\in\widetilde{h_{1}}$ and $\widetilde{h_{h2}}\in\widetilde{h_{2}}$, we have $\displaystyle\widetilde{{h}_{h1}}\odot\widetilde{{h}_{h2}}=\widetilde{{h}_{h2}}\odot\widetilde{{h}_{h1}}.\quad\quad(\textrm{Property}~{}\widetilde{(p_{G})})$ Then, one can see from Definition 2.4 that $\displaystyle\widetilde{h}_{1}\odot\widetilde{h}_{2}=\bigcup_{\widetilde{h_{h1}}\in\widetilde{h_{1}},\widetilde{h_{h2}}\in\widetilde{h_{2}}}\\{\widetilde{{h_{h1}}}\odot\widetilde{{h_{h2}}}\\}=\bigcup_{\widetilde{h_{h1}}\in\widetilde{h_{1}},\widetilde{h_{h2}}\in\widetilde{h_{2}}}\\{\widetilde{{h_{h2}}}\odot\widetilde{{h_{h1}}}\\}=\widetilde{h}_{2}\odot\widetilde{h}_{1}.\quad\quad(\textrm{Property}~{}\widetilde{(p)})$ This means that the operation $\odot$ on HOHFEs is accordingly commutative. By similar reasoning, one can easily prove that the HOHFE operator inherits all the properties of the G-type FS operator. $\Box$ As a corollary of Theorem 2.5 , it can be observed that HOHFS operators inherit subsequently all operational properties of G-type FS operators. ###### Corollary 2.6. Let $\widetilde{H}_{1}=\bigcup_{\widetilde{{h}}_{1}\in\widetilde{H}_{1}}\\{\widetilde{{h}_{1}}\\}$ and $\widetilde{H}_{2}=\bigcup_{\widetilde{{h}}_{2}\in\widetilde{H}_{2}}\\{\widetilde{{h}_{2}}\\}$ be two HOHFSs. If the operation $\odot$ on HOHFEs has the property $\widetilde{(p)}$, then the operation $\odot$ on HOHFSs has the mentioned property, denoted by $\widetilde{(P)}$. Ranking fuzzy information play a key role in fuzzy decision-making procedure. This fact is very encouraging for introducing a score function of HOHFSs. To compare the HOHFSs, we define the following comparison laws: ###### Definition 2.7. Let $X=\\{x_{1},x_{2}...,x_{n}\\}$. For HOHFS $\widetilde{H}=\\{\langle x,\widetilde{{h}}(x)\rangle:x\in X\\}=\bigcup_{\widetilde{h}\in\widetilde{H}}\\{\bigcup_{\widetilde{h_{h}}\in\widetilde{h}}\\{\widetilde{{h_{h}}}\\}\\}$, the score function ${\cal S}(.)$ is defined by $\displaystyle{\cal S}(\widetilde{H})=\frac{1}{n}\sum_{i=1}^{n}S_{\widetilde{h}}(\widetilde{h}(x_{i}))=\frac{1}{n}\sum_{i=1}^{n}(~{}\frac{1}{|\widetilde{h}(x_{i})|}\sum_{\widetilde{h_{h}}\in\widetilde{h}(x_{i})}S_{\widetilde{{h_{h}}}}(\widetilde{{h_{h}}}(x_{i}))~{}),$ (5) where $S_{\widetilde{h}}(.)$ is a score function of HOHFEs, $|\widetilde{h}(x_{i})|$ is the number of G-type FSs in $\widetilde{h}(x_{i})$ and $S_{\widetilde{{h_{h}}}}(.)$ is a score function of G-type FSs. For two HOHFSs $\widetilde{H_{1}}$ and $\widetilde{H_{2}}$, if ${\cal S}(\widetilde{H}_{1})<{\cal S}(\widetilde{H}_{2})$ then $\widetilde{H_{1}}\prec\widetilde{H_{2}}$; if ${\cal S}(\widetilde{H}_{1})={\cal S}(\widetilde{H}_{2})$ then $\widetilde{H_{1}}\approx\widetilde{H_{2}}$. Similar comparison rules can be considered for HOHFEs and G-type FSs such that * • for two HOHFEs $\widetilde{h_{1}}$ and $\widetilde{h_{2}}$ if $S_{\widetilde{h}}(\widetilde{h}_{1})<S_{\widetilde{h}}(\widetilde{h}_{2})$ then $\widetilde{h_{1}}\prec\widetilde{h_{2}}$; if $S_{\widetilde{h}}(\widetilde{h}_{1})=S_{\widetilde{h}}(\widetilde{h}_{2})$ then $\widetilde{h_{1}}\approx\widetilde{h_{2}}$; * • for two G-type FSs $\widetilde{h_{h}}^{(1)}$ and $\widetilde{h_{h}}^{(2)}$ if $S_{\widetilde{h_{h}}}(\widetilde{h}_{h}^{(1)})<S_{\widetilde{h_{h}}}(\widetilde{h}_{h}^{(2)})$ then $\widetilde{h_{h}}^{(1)}\prec\widetilde{h_{h}}^{(2)}$; if $S_{\widetilde{h_{h}}}(\widetilde{h}_{h}^{(1)})=S_{\widetilde{h_{h}}}(\widetilde{h}_{h}^{(2)})$ then $\widetilde{h_{h}}^{(1)}\approx\widetilde{h_{h}}^{(2)}$. In most of the real world applications, we usually face to the elements in the universe of discourse which may have a different importance. This impulses us to consider the weight of each element $x_{i}\in X$. Assume that the weight of $x_{i}\in X$ is $\mu(\\{x_{i}\\})$, $(i=1,...,n)$ where $\mu$ is a fuzzy measure. ###### Definition 2.8. [25] A fuzzy measure on $X$ is a set function $\mu$ from the power set of $X$ into $[0,1]$, satisfying the following conditions: $\displaystyle(1)~{}\mu(\emptyset)=0,~{}\mu(X)=1;$ $\displaystyle(2)~{}\textrm{For}~{}\textrm{any}~{}A,B\subseteq X,~{}\textrm{if}~{}A\subseteq B,~{}\textrm{then}~{}\mu(A)\leq\mu(B);$ $\displaystyle(3)~{}\textrm{For}~{}\textrm{any}~{}A,B\subseteq X,~{}\mu(A\cup B)=\mu(A)+\mu(B)-\rho\mu(A)\mu(B),~{}\textrm{where}~{}A\cap B=\emptyset~{}\textrm{and}~{}\rho\in(-1,\infty).$ A fuzzy measure $\mu$ on $X$ is said to be * • Additive if $\mu(A\cup B)=\mu(A)+\mu(B)$ for all $A,B\subseteq X$ where $A\cap B=\emptyset$; * • Subadditive if $\mu(A\cup B)\leq\mu(A)+\mu(B)$ for all $A,B\subseteq X$ where $A\cap B=\emptyset$; * • Superadditive if $\mu(A\cup B)\geq\mu(A)+\mu(B)$ for all $A,B\subseteq X$ where $A\cap B=\emptyset$. Sometimes, in MCDM, we need an aggregation function to determine the relative importance of decision criteria considering their interactions. Such a need could be satisfied by implementing the Choquet integral [7] which is defined on the basis of fuzzy measures. ###### Definition 2.9. [7] Let $f$ be a positive real-valued function on $X$, and $\mu$ be a fuzzy measure on $X$. The Choquet integral of $f$ with respect to $\mu$ is defined by $\displaystyle C_{\mu}(f)=\sum_{i=1}^{n}(\mu(A_{\sigma(i)})-\mu(A_{\sigma(i-1)})f_{\sigma(i)}$ (6) where $(\sigma(1),\sigma(2),...,\sigma(n))$ is a permutation of $(1,2,...,n)$ such that $f_{\sigma(1)}\geq f_{\sigma(2)}\geq...\geq f_{\sigma(n)}$, $A_{\sigma(k)}=\\{x_{\sigma(1)},x_{\sigma(2)},...,x_{\sigma(k)}\\}$ for $k\geq 1$, and $A_{\sigma(0)}=\emptyset$. The Choquet integral $C_{\mu}(.)$ has some considerable aggregation properties, such as idempotence, compensativeness and comonotonic additivity [17]. Moreover, it generalizes the weighted arithmetic mean (WAM) and the ordered weighted averaging (OWA) [40]. In view of the Choquet integral formulation (6), we here introduce the HOHF Choquet integral: ###### Definition 2.10. Let $\mu$ be a fuzzy measure on $X$, and $\widetilde{h}(x_{i}),~{}(i=1,...,n)$ be a collection of HOHFEs on $X$. The HOHF Choquet integral of $\widetilde{h}(x_{i}),~{}(i=1,...,n)$ with respect to $\mu$ is defined by $\displaystyle HOHFEC_{\mu}(\widetilde{h}(x_{1}),...,\widetilde{h}(x_{n}))=\oplus_{i=1}^{n}[(\mu(A_{\sigma(i)})-\mu(A_{\sigma(i-1)})\widetilde{h}(x_{\sigma(i)})]\hskip 142.26378pt$ $\displaystyle\hskip 28.45274pt=\bigcup_{\widetilde{h_{h}}(x_{\sigma(1)})\in\widetilde{h}(x_{\sigma(1)}),...,\widetilde{h_{h}}(x_{\sigma(n)})\in\widetilde{h}(x_{\sigma(n)})}\\{\oplus_{i=1}^{n}[(\mu(A_{\sigma(i)})-\mu(A_{\sigma(i-1)})\widetilde{h_{h}}(x_{\sigma(i)})]\\},$ (7) where $(\sigma(1),\sigma(2),...,\sigma(n))$ is a permutation of $(1,2,...,n)$ such that $\widetilde{h}(x_{\sigma(1)})\succ\widetilde{h}(x_{\sigma(2)})\succ...\succ\widetilde{h}(x_{\sigma(n)})$, $A_{\sigma(k)}=\\{x_{\sigma(1)},x_{\sigma(2)},...,x_{\sigma(k)}\\}$ for $k\geq 1$, and $A_{\sigma(0)}=\emptyset$. ###### Proposition 2.11. If for any G-type FSs $\widetilde{h_{h}}^{(1)}$ and $\widetilde{h_{h}}^{(2)}$ of HOHFE $\widetilde{h}$, the addition $\widetilde{h_{h}}^{(1)}\oplus\widetilde{h_{h}}^{(2)}$ and the scalar multiplication $\lambda\widetilde{h_{h}}^{(1)}$ where $\lambda>0$ are also G-type FSs, then the HOHF Choquet integral of $\widetilde{h}(x_{i}),~{}(i=1,...,n)$ with respect to $\mu$ introduced in Definition 2.10 is a HOHFE. Proof. The proof is obvious by putting together the definition of the HOHF Choquet integral of $\widetilde{h}(x_{i}),~{}(i=1,...,n)$ with respect to $\mu$ introduced in Definition 2.10 and Theorem 2.5. $\Box$ ###### Example 2.12. Let $X=\\{x_{1},x_{2}\\}$, and $\widetilde{h}(x_{i}),~{}(i=1,2)$ be a collection of HOHFEs on $X$ whose G-type FSs are in the form of IFSs given by $\displaystyle\widetilde{h}(x_{1})=\\{\widetilde{h_{h}}^{(1)}(x_{1})=\langle 0.2,0.5\rangle,\widetilde{h_{h}}^{(2)}(x_{1})=\langle 0.3,0.4\rangle\\},\quad\quad\widetilde{h}(x_{2})=\\{\widetilde{h_{h}}^{(1)}(x_{2})=\langle 0.3,0.5\rangle\\},$ and a fuzzy measure $\mu:X\rightarrow[0,1]$ is given by $\displaystyle\mu(\emptyset)=0,\quad\mu(\\{x_{1}\\})=0.2,\quad\mu(\\{x_{2}\\})=0.3,\quad\mu(\\{x_{1},x_{2}\\})=0.6.$ If we take the G-type FSs score function as the hesitancy degree, i.e., $\displaystyle S_{\widetilde{h_{h}}}(\widetilde{h_{h}})=S_{\widetilde{h_{h}}}(\langle\mu_{\widetilde{{h_{h}}}},\nu_{\widetilde{{h_{h}}}}\rangle)=1-\mu_{\widetilde{{h_{h}}}}-\nu_{\widetilde{{h_{h}}}},$ then the HOHFE scores $S_{\widetilde{h}}(\widetilde{h}(x_{1}))=0.3$ and $S_{\widetilde{h}}(\widetilde{h}(x_{2}))=0.1$ indicate that $\widetilde{h}(x_{1})\succ\widetilde{h}(x_{2})$ and therefore ${\sigma(1)}=1,{\sigma(2)}=2$. Now from (7) the HOHF Choquet integral of $\widetilde{h}(x_{i}),~{}(i=1,2)$ with respect to $\mu$ is calculated as follows: $\displaystyle HOHFEC_{\mu}(\widetilde{h}(x_{1}),\widetilde{h}(x_{2}))=\bigcup_{\widetilde{h_{h}}(x_{1})\in\widetilde{h}(x_{1}),\widetilde{h_{h}}(x_{2})\in\widetilde{h}(x_{2})}\\{\oplus_{i=1}^{2}[(\mu(A_{\sigma(i)})-\mu(A_{\sigma(i-1)})\widetilde{h_{h}}(x_{\sigma(i)})]\\}$ $\displaystyle=\\{[(\mu(A_{1})-\mu(A_{0})]\widetilde{h_{h}}^{(1)}(x_{1})\oplus[(\mu(A_{2})-\mu(A_{1})]\widetilde{h_{h}}^{(1)}(x_{2}),$ $\displaystyle\hskip 113.81102pt[(\mu(A_{1})-\mu(A_{0})]\widetilde{h_{h}}^{(2)}(x_{1})\oplus[(\mu(A_{2})-\mu(A_{1})]\widetilde{h_{h}}^{(1)}(x_{2})\\}$ $\displaystyle=\\{[(\mu(\\{x_{1}\\})-\mu(\emptyset)]\widetilde{h_{h}}^{(1)}(x_{1})\oplus[(\mu(\\{x_{2}\\})-\mu(\\{x_{1}\\})]\widetilde{h_{h}}^{(1)}(x_{2}),$ $\displaystyle\hskip 113.81102pt[(\mu(\\{x_{1}\\})-\mu(\emptyset)]\widetilde{h_{h}}^{(2)}(x_{1})\oplus[(\mu(\\{x_{2}\\})-\mu(\\{x_{1}\\})]\widetilde{h_{h}}^{(1)}(x_{2})\\}$ $\displaystyle=\\{[0.2]\langle 0.2,0.5\rangle\oplus[0.6-0.2]\langle 0.3,0.5\rangle,$ $\displaystyle\hskip 113.81102pt[0.2]\langle 0.3,0.4\rangle\oplus[0.6-0.2]\langle 0.3,0.5\rangle\\}.$ By the use of the following relations (see e.g. [44]) $\displaystyle\widetilde{{h_{h1}}}\oplus\widetilde{{h_{h2}}}=\langle\mu_{\widetilde{{h_{h1}}}},\nu_{\widetilde{{h_{h1}}}}\rangle\oplus\langle\mu_{\widetilde{{h_{h2}}}},\nu_{\widetilde{{h_{h2}}}}\rangle=\langle\mu_{\widetilde{{h_{h1}}}}+\mu_{\widetilde{{h_{h2}}}}-\mu_{\widetilde{{h_{h1}}}}.\mu_{\widetilde{{h_{h2}}}},~{}~{}\nu_{\widetilde{{h_{h1}}}}.\nu_{\widetilde{{h_{h2}}}}\rangle;$ $\displaystyle\widetilde{{h_{h1}}}^{\lambda}=\langle(\mu_{\widetilde{{h_{h1}}}})^{\lambda},1-(1-\nu_{\widetilde{{h_{h1}}}})^{\lambda}\rangle,\quad\lambda>0,$ in which $\mu_{\widetilde{{h_{h2}}}}$ and $\nu_{\widetilde{{h_{h2}}}}$ stand respectively for the membership and the non-membership functions, we get $\displaystyle\widetilde{h}_{{}_{HOHFEC}}:=HOHFEC_{\mu}(\widetilde{h}(x_{1}),\widetilde{h}(x_{2}))=\\{\langle 0.1708,0.6598\rangle,\langle 0.1927,0.6310\rangle\\}.$ ###### Proposition 2.13. (Idempotency). If for any G-type FS $\widetilde{h_{h}}$ of HOHFE $\widetilde{h}$, it holds $\lambda_{1}\widetilde{h_{h}}\oplus\lambda_{2}\widetilde{h_{h}}=(\lambda_{1}+\lambda_{2})\widetilde{h_{h}}$ where $\lambda_{1},\lambda_{2}>0$, and if all HOHFEs $\widetilde{h}(x_{i}),~{}(i=1,...,n)$ are equal, that is, for all $i$, $\widetilde{h}(x_{i})=\overline{\widetilde{h}}$, then the HOHF Choquet integral of $\widetilde{h}(x_{i}),~{}(i=1,...,n)$ with respect to $\mu$ introduced in Definition 2.10 is equal to the HOHFE $\overline{\widetilde{h}}$. Proof. Suppose that for all $i,~{}(i=1,...,n)$, we have $\widetilde{h}(x_{i})=\overline{\widetilde{h}}$. From definition of the HOHF Choquet integral of $\widetilde{h}(x_{i}),~{}(i=1,...,n)$ with respect to $\mu$ introduced in Definition 2.10, we have $\displaystyle HOHFEC_{\mu}(\widetilde{h}(x_{1}),...,\widetilde{h}(x_{n}))$ $\displaystyle=$ $\displaystyle\oplus_{i=1}^{n}[(\mu(A_{\sigma(i)})-\mu(A_{\sigma(i-1)})\widetilde{h}(x_{\sigma(i)})]$ $\displaystyle=[(\mu(A_{\sigma(1)})-\mu(A_{\sigma(0)})]\overline{\widetilde{h}}\oplus...\oplus[(\mu(A_{\sigma(n)})-\mu(A_{\sigma(n-1)})]\overline{\widetilde{h}}.$ If for any G-type FS $\widetilde{h_{h}}$ of HOHFE $\widetilde{h}$, it holds $\lambda_{1}\widetilde{h_{h}}\oplus\lambda_{2}\widetilde{h_{h}}=(\lambda_{1}+\lambda_{2})\widetilde{h_{h}}$ where $\lambda_{1},\lambda_{2}>0$, then by Theorem 2.5 and the fact that $\mu(A_{\sigma(0)}=\emptyset)=0,\mu(A_{\sigma(n)}=X)=1$, one gets $\displaystyle HOHFEC_{\mu}(\widetilde{h}(x_{1}),...,\widetilde{h}(x_{n}))$ $\displaystyle=([(\mu(A_{\sigma(1)})-\mu(A_{\sigma(0)})]+...+[(\mu(A_{\sigma(n)})-\mu(A_{\sigma(n-1)})])\overline{\widetilde{h}}=\overline{\widetilde{h}}.~{}\Box$ ###### Lemma 2.14. Suppose that the score functions $S_{\widetilde{h}}$ on HOHFEs and $S_{\widetilde{{h_{h}}}}$ on G-type FSs are those introduced in Definition 2.7, that is, $\displaystyle S_{\widetilde{h}}(\widetilde{h})=\frac{1}{|\widetilde{h}|}\sum_{\widetilde{h_{h}}\in\widetilde{h}}S_{\widetilde{{h_{h}}}}(\widetilde{{h_{h}}}).$ (8) If for each HOHFE $\widetilde{h}=\bigcup_{\widetilde{h_{h}}\in\widetilde{h}}\\{\widetilde{h_{h}}\\}$, $S_{\widetilde{{h_{h}}}}$ satisfies $S_{\widetilde{{h_{h}}}}(\lambda\widetilde{{h_{h}}})=\lambda S_{\widetilde{{h_{h}}}}(\widetilde{{h_{h}}})$, for any $\lambda>0$. Then, $S_{\widetilde{h}}$ also satisfies the latter property, i.e., $\displaystyle S_{\widetilde{h}}(\lambda\widetilde{h})=\lambda S_{\widetilde{h}}(\widetilde{h}),\quad\forall\lambda>0.$ (9) Proof. From definition of $S_{\widetilde{h}}$ given by (8) and the mentioned property of $S_{\widetilde{{h_{h}}}}$, one gets for any $\lambda>0$ $\displaystyle S_{\widetilde{h}}(\lambda\widetilde{h})=\frac{1}{|\lambda\widetilde{h}|}\sum_{\widetilde{h_{h}}\in\widetilde{h}}S_{\widetilde{{h_{h}}}}(\lambda\widetilde{{h_{h}}})=\frac{1}{|\widetilde{h}|}\sum_{\widetilde{h_{h}}\in\widetilde{h}}S_{\widetilde{{h_{h}}}}(\lambda\widetilde{{h_{h}}})=\frac{1}{|\widetilde{h}|}\sum_{\widetilde{h_{h}}\in\widetilde{h}}\lambda S_{\widetilde{{h_{h}}}}(\widetilde{{h_{h}}})=\lambda S_{\widetilde{h}}(\widetilde{h}).~{}\Box$ Hereafter, the HOHFE score function $S_{\widetilde{h}}$ is said to be $\lambda$-invariance if it satisfies the property (9). ###### Lemma 2.15. Assume that the HOHFE score function $S_{\widetilde{h}}$ is $\lambda$-invariance. For any two HOHFEs $\widetilde{h}_{1}$ and $\widetilde{h}_{2}$, if $\widetilde{h}_{1}\succ\widetilde{h}_{2}$, then $\lambda\widetilde{h}_{1}\succ\lambda\widetilde{h}_{2}$, for any $\lambda>0$. Proof. As follows from Definition 2.7, we find that $\widetilde{h}_{1}\succ\widetilde{h}_{2}$ whenever $S_{\widetilde{h}}(\widetilde{h}_{1})>S_{\widetilde{h}}(\widetilde{h}_{2})$. On the other hand, with the $\lambda$-invariance property of the HOHFE score function $S_{\widetilde{h}}$ in mind, the relation $S_{\widetilde{h}}(\widetilde{h}_{1})>S_{\widetilde{h}}(\widetilde{h}_{2})$ gives rise to $S_{\widetilde{h}}(\lambda\widetilde{h}_{1})>S_{\widetilde{h}}(\lambda\widetilde{h}_{2})$ for any $\lambda>0$, which implies that $\lambda\widetilde{h}_{1}\succ\lambda\widetilde{h}_{2}$. This completes the proof. $\Box$ ###### Proposition 2.16. (Monotonicity). Let $\mu$ be a fuzzy measure on $X$, and $\widetilde{h}_{1}(x_{i})$ and $\widetilde{h}_{2}(x_{i}),~{}(i=1,...,n)$ be two collection of HOHFEs on $X$. Assume that the HOHFE score function $S_{\widetilde{h}}$ is $\lambda$-invariance, and $(\sigma(1),\sigma(2),...,\sigma(n))$ denotes a permutation of $(1,2,...,n)$ such that $\widetilde{h}_{1}(x_{\sigma(1)})\succ\widetilde{h}_{1}(x_{\sigma(2)})\succ...\succ\widetilde{h}_{1}(x_{\sigma(n)})$ and $\widetilde{h}_{2}(x_{\sigma(1)})\succ\widetilde{h}_{2}(x_{\sigma(2)})\succ...\succ\widetilde{h}_{2}(x_{\sigma(n)})$. If $\widetilde{h}_{1}(x_{\sigma(i)})\succ\widetilde{h}_{2}(x_{\sigma(i)})$ for $(i=1,...,n)$, then $\displaystyle HOHFEC_{\mu}(\widetilde{h}_{1}(x_{1}),...,\widetilde{h}_{1}(x_{n}))\succ HOHFEC_{\mu}(\widetilde{h}_{2}(x_{1}),...,\widetilde{h}_{2}(x_{n})).$ Proof. The assertion follows by applying Lemma 2.15 to Definition 2.10. $\Box$ ###### Proposition 2.17. (Boundedness). Let $\mu$ be a fuzzy measure on $X$, $\widetilde{h}(x_{i}),~{}(i=1,...,n)$ be a collection of HOHFEs on $X$, and $(\sigma(1),\sigma(2),...,\sigma(n))$ be a permutation of $(1,2,...,n)$ such that $\widetilde{h}_{1}(x_{\sigma(1)})\succ\widetilde{h}_{1}(x_{\sigma(2)})\succ...\succ\widetilde{h}_{1}(x_{\sigma(n)})$. Then, the $\lambda$-invariance property of the HOHFE score function $S_{\widetilde{h}}$ gives rise to $\displaystyle\widetilde{h}_{Max}\succ HOHFEC_{\mu}(\widetilde{h}_{1}(x_{1}),...,\widetilde{h}_{1}(x_{n}))\succ\widetilde{h}_{Min},$ where $\widetilde{h}_{Min}=\min\\{\widetilde{h}_{1}(x_{1}),...,\widetilde{h}_{1}(x_{n})\\}=\widetilde{h}_{1}(x_{\sigma(n)})$ and $\widetilde{h}_{Max}=\max\\{\widetilde{h}_{1}(x_{1}),...,\widetilde{h}_{1}(x_{n})\\}=\widetilde{h}_{1}(x_{\sigma(1)})$. Proof. The application of Proposition 2.16 gives rise to the assertion. $\Box$ ###### Proposition 2.18. If for any G-type FS $\widetilde{h_{h}}$ of HOHFE $\widetilde{h}$, it holds $(\widetilde{h_{h}}^{(1)}\oplus\widetilde{h_{h}}^{(2)})\oplus\widetilde{h_{h}}^{(3)}=\widetilde{h_{h}}^{(1)}\oplus(\widetilde{h_{h}}^{(2)}\oplus\widetilde{h_{h}}^{(3)})$ and $\lambda_{1}\widetilde{h_{h}}\oplus\lambda_{2}\widetilde{h_{h}}=(\lambda_{1}+\lambda_{2})\widetilde{h_{h}}$ where $\lambda_{1},\lambda_{2}>0$, then the HOHF Choquet integral of $\widetilde{h}(x_{i}),~{}(i=1,...,n)$ with respect to $\mu$ introduced in Definition 2.10 satisfies $\displaystyle HOHFEC_{\mu}(\kappa\widetilde{h}_{1}(x_{1})\oplus\widetilde{\overline{h}},...,\kappa\widetilde{h}_{1}(x_{n})\oplus\widetilde{\overline{h}})=\kappa HOHFEC_{\mu}(\widetilde{h}_{1}(x_{1}),...,\widetilde{h}_{1}(x_{n}))\oplus\widetilde{\overline{h}},\quad\kappa>0.$ Proof. The proof is much like that of Proposition 2.13 and therefore is omitted. $\Box$ ## 3 MCDM with the HOHF Choquet integral operator In this section, we present a new method for MCDM, in which the evaluations of the alternatives are given by HOHFEs and the interaction among the criteria are allowed. The MCDM procedure with HOHF Choquet integral operator can be described as follows: suppose that $Y=\\{y_{1},y_{2},...,y_{m}\\}$ is an alternative set of $m$ alternatives. In a MCDM problem, the decision maker is interested in choosing the best one(s) from $Y$ according to the criteria set $X=\\{x_{1},x_{2},...,x_{n}\\}$. In the HOHFE MCDM, the evaluation of each alternative on each criterion is a HOHFE. By using the score function $S_{\widetilde{h}}$ introduced in Definition 2.7, all HOHFEs associated with each alternative are re-ordered such that $\widetilde{h}(x_{\sigma(1)})\succ\widetilde{h}(x_{\sigma(2)})\succ...\succ\widetilde{h}(x_{\sigma(n)})$ where $(\sigma(1),\sigma(2),...,\sigma(n))$ is a permutation of $(1,2,...,n)$. Taking into account the correlations of the HOHFEs, the evaluations of an alternative can be aggregated to its overall evaluation by the HOHF Choquet integral operator. According to a given total order relation on aggregated HOHFEs, the decision maker can rank those overall evaluations and get the best alternative(s). ### 3.1 A practical example In what follows, we are going to demonstrate the practicality of implementing the proposed concept of HOHFE in a higher order hesitant fuzzy multi-attribute decision making problem. Energy plays an important role in socio-economic development. Thus selecting an appropriate policy for energy is critical for economic development and environment. Assume that there exist five alternatives (energy projects) ${y}_{i},(i=1,2,3,4,5)$ which are invested in accordance with four criteria: $x_{1}$: technological; $x_{2}$: environmental; $x_{3}$: socio-political; $x_{4}$: economic. A number of decision makers are invited to evaluate the performance of the above five alternatives. For an alternative under a criterion, the decision makers give their evaluations anonymously in the form of HOHFSs. In this regard, the results evaluated by the decision makers are contained in a higher order hesitant fuzzy decision matrix which is shown in Table 1. Table 1. Higher order hesitant fuzzy decision matrix | $x_{1}$ | $x_{2}$ | $x_{3}$ | $x_{4}$ ---|---|---|---|--- ${y}_{1}$ | $\\{(0.3,0.4,0.5)\\}$ | $\\{(0.4,0.5,0.6),\\{0.7,0.8,0.9\\}\\}$ | $\\{(0.5,0.7,0.7),(0.7,0.8,0.9)\\}$ | $\\{(0.2,0.3,0.4),\\{0.3,0.4,0.5\\}\\}$ ${y}_{2}$ | $\\{(0.1,0.2,0.3),(0.2,0.3,0.4)\\}$ | $\\{(0.5,0.6,0.7),\\{0.3,0.4,0.5\\}\\}$ | $\\{(0.2,0.4,0.6),(0.7,0.8,0.9)\\}$ | $\\{(0.1,0.4,0.7),(0.6,0.7,0.8)\\}$ ${y}_{3}$ | $\\{(0.1,0.2,0.3),\\{0.3,0.4,0.5\\}\\}$ | $\\{0.7,0.8,0.9\\}$ | $\\{(0.2,0.3,0.4),(0.5,0.6,0.7)\\}$ | $\\{(0.4,0.5,0.6)\\}$ ${y}_{4}$ | $\\{0.2,0.3,0.4\\}$ | $\\{(0.3,0.4,0.5),(0.2,0.4,0.6)\\}$ | $\\{(0.5,0.6,0.7),\\{0.3,0.4,0.5\\}\\}$ | $\\{(0.1,0.2,0.3),(0.3,0.4,0.5)\\}$ ${y}_{5}$ | $\\{(0.2,0.4,0.6),(0.7,08,0.9)\\}$ | $\\{0.7,0.8,0.9\\}$ | $\\{(0.3,0.4,0.5)\\}$ | $\\{(0.5,0.7,0.7),(0.7,0.8,0.9)\\}$ In Table 1, the notation $A=(a_{1},a_{2},a_{3})$ is used to describe a fuzzy event by the help of a triangular fuzzy number in which the values $a_{1}$, $a_{2}$ and $a_{3}$ denote respectively the smallest possible value, the most promising value and the largest possible value, and moreover, $\\{\widetilde{h_{\widetilde{h}}}^{(1)},\widetilde{h_{\widetilde{h}}}^{(2)},\widetilde{h_{\widetilde{h}}}^{(3)}\\}$ indicates a HFE. In this setting and as a consequence, we observe that the second and third elements of the last row of Table 1 are conceptually different. The second array of Table 1 indicates a HFE, while, the third array is used to indicate a triangular fuzzy number. Suppose that the fuzzy measures of criteria $X=\\{x_{1},x_{2},x_{3},x_{4}\\}$ are given as follows: $\displaystyle\mu(\emptyset)=0;$ $\displaystyle\mu(\\{x_{1}\\})=0.2,~{}\mu(\\{x_{2}\\})=0.4,~{}\mu(\\{x_{3}\\})=0.3,~{}\mu(\\{x_{4}\\})=0.1;$ $\displaystyle\mu(\\{x_{1},x_{2}\\})=0.4,~{}\mu(\\{x_{1},x_{3}\\})=0.2,~{}\mu(\\{x_{1},x_{4}\\})=0.1,~{}\mu(\\{x_{2},x_{3}\\})=0.3,$ $\displaystyle\mu(\\{x_{2},x_{4}\\})=0.3,~{}\mu(\\{x_{3},x_{4}\\})=0.1;$ $\displaystyle\mu(\\{x_{1},x_{2},x_{3}\\})=0.3,~{}\mu(\\{x_{1},x_{2},x_{4}\\})=0.1,~{}\mu(\\{x_{1},x_{3},x_{4}\\})=0.2,~{}\mu(\\{x_{2},x_{3},x_{4}\\})=0.2;$ $\displaystyle\mu({X})=1.$ For re-ordering HOHFEs given in Table 1, we recall the HOHFE score function $S_{\widetilde{h}}$ introduced in Definition 2.7, where $\displaystyle S_{\widetilde{h}}(\widetilde{h})=\frac{1}{|\widetilde{h}|}\sum_{\widetilde{h_{h}}\in\widetilde{h}}S_{\widetilde{{h_{h}}}}(\widetilde{{h_{h}}}).$ (10) Here, the G-type FS score function $S_{\widetilde{{h_{h}}}}$ is taken as * • (For triangular fuzzy number): $\displaystyle S_{\widetilde{{h_{h}}}}({(a_{1},a_{2},a_{3})})=\frac{1}{3}(a_{1}+a_{2}+a_{3}),$ (11) which was proposed by Lee and Li in [16]; * • (For hesitant fuzzy element): $\displaystyle S_{\widetilde{{h_{h}}}}(\\{\widetilde{h_{\widetilde{h}}}^{(1)},...,\widetilde{h_{\widetilde{h}}}^{(|\widetilde{h}|)}\\})=\frac{1}{|\widetilde{h}|}(\\{\widetilde{h_{\widetilde{h}}}^{(1)}+...+\widetilde{h_{\widetilde{h}}}^{(|\widetilde{h}|)}\\}),$ (12) Although, the corresponding score functions of HFEs and triangular fuzzy numbers may differ from each other, but in this example, both of them have a same rule. However, by using the HOHFE score function $S_{\widetilde{h}}$ given by (10), we re-arrange the HOHFEs corresponding to each criterion in descending order as follows: $\displaystyle y_{1}:\quad\widetilde{h}(x_{\sigma(1)}=x_{3})\succ\widetilde{h}(x_{\sigma(2)}=x_{2})\succ\widetilde{h}(x_{\sigma(3)}=x_{1})\succ\widetilde{h}(x_{\sigma(4)}=x_{4});$ $\displaystyle y_{2}:\quad\widetilde{h}(x_{\sigma(1)}=x_{3})\succ\widetilde{h}(x_{\sigma(2)}=x_{4})\succ\widetilde{h}(x_{\sigma(3)}=x_{2})\succ\widetilde{h}(x_{\sigma(4)}=x_{1});$ $\displaystyle y_{3}:\quad\widetilde{h}(x_{\sigma(1)}=x_{2})\succ\widetilde{h}(x_{\sigma(2)}=x_{4})\succ\widetilde{h}(x_{\sigma(3)}=x_{3})\succ\widetilde{h}(x_{\sigma(4)}=x_{1});$ $\displaystyle y_{4}:\quad\widetilde{h}(x_{\sigma(1)}=x_{3})\succ\widetilde{h}(x_{\sigma(2)}=x_{2})\succ\widetilde{h}(x_{\sigma(3)}=x_{1})\succ\widetilde{h}(x_{\sigma(4)}=x_{4});$ $\displaystyle y_{5}:\quad\widetilde{h}(x_{\sigma(1)}=x_{2})\succ\widetilde{h}(x_{\sigma(2)}=x_{4})\succ\widetilde{h}(x_{\sigma(3)}=x_{1})\succ\widetilde{h}(x_{\sigma(4)}=x_{3}).$ If we employ the HOHF Choquet integral operator, we then are able to aggregate all HOHFEs in the i-$th$ row of the HOHFE decision matrix into an overall value. For instance, $\displaystyle\widetilde{h}_{y_{1}}=HOHFEC_{\mu}(\widetilde{h}(x_{1}),...,\widetilde{h}(x_{4}))=$ $\displaystyle[(\mu(A_{\sigma(1)})-\mu(A_{\sigma(0)}))\widetilde{h}(x_{\sigma(1)})]\oplus[(\mu(A_{\sigma(2)})-\mu(A_{\sigma(1)}))\widetilde{h}(x_{\sigma(2)})]$ $\displaystyle\oplus[(\mu(A_{\sigma(3)})-\mu(A_{\sigma(2)}))\widetilde{h}(x_{\sigma(3)})]\oplus[(\mu(A_{\sigma(4)})-\mu(A_{\sigma(3)}))\widetilde{h}(x_{\sigma(4)})]$ $\displaystyle=[(\mu(\\{x_{\sigma(1)}\\})-\mu(\emptyset))\widetilde{h}(x_{\sigma(1)})]\oplus[(\mu(\\{x_{\sigma(1)},x_{\sigma(2)}\\})-\mu(\\{x_{\sigma(1)}\\}))\widetilde{h}(x_{\sigma(2)})]$ $\displaystyle\oplus[(\mu(\\{x_{\sigma(1)},x_{\sigma(2)},x_{\sigma(3)}\\})-\mu(\\{x_{\sigma(1)},x_{\sigma(2)}\\}))\widetilde{h}(x_{\sigma(3)})]\oplus[(\mu(X)-\mu(\\{x_{\sigma(1)},x_{\sigma(2)},x_{\sigma(3)}\\}))\widetilde{h}(x_{\sigma(4)})]$ $\displaystyle=[(\mu(\\{x_{3}\\})-\mu(\emptyset))\widetilde{h}(x_{3})]\oplus[(\mu(\\{x_{3},x_{2}\\})-\mu(\\{x_{3}\\}))\widetilde{h}(x_{2})]$ $\displaystyle\oplus[(\mu(\\{x_{3},x_{2},x_{1}\\})-\mu(\\{x_{3},x_{2}\\}))\widetilde{h}(x_{1})]\oplus[(\mu(X)-\mu(\\{x_{3},x_{2},x_{1}\\}))\widetilde{h}(x_{4})]$ $\displaystyle=[(0.3-0)\\{(0.5,0.7,0.7),(0.7,0.8,0.9)\\}]\oplus[(0.3-0.3)\\{(0.4,0.5,0.6),\\{0.7,0.8,0.9\\}\\}]$ $\displaystyle\oplus[(0.3-0.3)\\{(0.3,0.4,0.5)\\}]\oplus[(1-0.3)\\{(0.2,0.3,0.4),\\{0.3,0.4,0.5\\}\\}].$ Let us now apply the following arithmetic operations * • (For triangular fuzzy number [14]): $\displaystyle\lambda A=\lambda(a_{1},a_{2},a_{3})=(\lambda a_{1},\lambda a_{2},\lambda a_{3}),$ $\displaystyle A\oplus B=(a_{1},a_{2},a_{3})\oplus(b_{1},b_{2},b_{3})=(a_{1}+b_{1}-a_{1}b_{1},a_{2}+b_{2}-a_{2}b_{2},a_{3}+b_{3}-a_{3}b_{3});$ * • (For hesitant fuzzy element [11]): $\displaystyle\lambda\\{\widetilde{h_{\widetilde{h}}}^{(1)},...,\widetilde{h_{\widetilde{h}}}^{(|\widetilde{h}|)}\\}=\\{\lambda\widetilde{h_{\widetilde{h}}}^{(1)},...,\lambda\widetilde{h_{\widetilde{h}}}^{(|\widetilde{h}|)}\\},\quad\lambda\geq 0,$ $\displaystyle\\{\widetilde{h_{\widetilde{h_{1}}}}^{(1)},...,\widetilde{h_{\widetilde{h_{1}}}}^{(|\widetilde{h}|)}\\}\oplus\\{\widetilde{h_{\widetilde{h_{2}}}}^{(1)},...,\widetilde{h_{\widetilde{h_{2}}}}^{(|\widetilde{h}|)}\\}$ $\displaystyle\quad\quad=\\{\widetilde{h_{\widetilde{h_{1}}}}^{(1)}+\widetilde{h_{\widetilde{h_{2}}}}^{(1)}-\widetilde{h_{\widetilde{h_{1}}}}^{(1)}\widetilde{h_{\widetilde{h_{2}}}}^{(1)},...,\widetilde{h_{\widetilde{h_{1}}}}^{(|\widetilde{h}|)}+\widetilde{h_{\widetilde{h_{2}}}}^{(|\widetilde{h}|)}-\widetilde{h_{\widetilde{h_{1}}}}^{(|\widetilde{h}|)}\widetilde{h_{\widetilde{h_{2}}}}^{(|\widetilde{h}|)}\\},$ to the latter findings. Then, one achieves that $\displaystyle\widetilde{h}_{y_{1}}$ $\displaystyle=$ $\displaystyle[\\{(0.15,0.21,0.21),(0.21,0.24,0.27)\\}]\oplus[\\{(0.14,0.21,0.28),\\{0.21,0.28,0.35\\}\\}]$ $\displaystyle=$ $\displaystyle\\{(0.27,0.37,0.43),(0.32,0.40,0.47),\\{0.21,0.28,0.35\\}\\}.$ The HOHFE score value of $\widetilde{h}_{y_{1}}$ is then equal to $\displaystyle S_{\widetilde{h}}(\widetilde{h}_{y_{1}})=\frac{1}{|\widetilde{h}_{y_{1}}|}\sum_{\widetilde{h_{{h}_{y_{1}}}}\in\widetilde{h}_{y_{1}}}S_{\widetilde{{h_{h}}}}(\widetilde{{h_{{h}_{y_{1}}}}})=0.3456.$ Similarly, the other HOHFE score values are obtained as: $\displaystyle S_{\widetilde{h}}(\widetilde{h}_{y_{2}})=0.4915,\quad S_{\widetilde{h}}(\widetilde{h}_{y_{3}})=0.4364,\quad S_{\widetilde{h}}(\widetilde{h}_{y_{4}})=0.2739,\quad S_{\widetilde{h}}(\widetilde{h}_{y_{5}})=0.5601.$ Therefore, $\displaystyle S_{\widetilde{h}}(\widetilde{h}_{y_{5}})>S_{\widetilde{h}}(\widetilde{h}_{y_{2}})>S_{\widetilde{h}}(\widetilde{h}_{y_{3}})>S_{\widetilde{h}}(\widetilde{h}_{y_{1}})>S_{\widetilde{h}}(\widetilde{h}_{y_{4}}),$ which are led to $\displaystyle y_{5}\succ y_{2}\succ y_{3}\succ y_{1}\succ y_{4},$ and hence the most appropriate energy project is $y_{5}$. ### 3.2 Comparisons and further discussions Here, we adopt a multiple criteria group decision making problem from [21] which is originally based on extending TODIM method by encountering the Choquet integral within a multiset hesitant fuzzy environment. Peng et al. [21] considered an investment company for investing in a project. They supposed five alternatives ${y}_{1}$: car company; ${y}_{2}$: food company; ${y}_{3}$: computer company; ${y}_{4}$: arms company; and ${y}_{5}$: TV company which need to be invested. In this setting, the desired decision has to be made according to the four criteria including $x_{1}$: the environment impact which is referred to as the impact on the companys environment $x_{2}$: the risk factor including product risk and development environment risk; $x_{3}$: the growth prospects including increased profitability and returns; and $x_{4}$: social-political impact that is referred to as the governments and local residents support for company. Peng et al. [21] used multiset hesitant fuzzy sets to evaluate alternatives by two decision makers. In this contribution to have a complete picture about the performance of the proposed approach and that of Peng et al. [21], we re-state Peng et al’s [21] decision matrix in the following form of higher order hesitant fuzzy decision matrix: Table 2. Higher order hesitant fuzzy decision matrix | $x_{1}$ | $x_{2}$ | $x_{3}$ | $x_{4}$ ---|---|---|---|--- ${y}_{1}$ | $\\{(0.4,0.5,0.7)\\}$ | $\\{(0.5,0.5,0.8)\\}$ | $\\{(0.6,0.6,0.9)\\}$ | $\\{0.5,0.6\\}$ ${y}_{2}$ | $\\{(0.6,0.7,0.8)\\}$ | $\\{0.5,0.6\\}$ | $\\{(0.6,0.7,0.7)\\}$ | $\\{0.4,0.5\\}$ ${y}_{3}$ | $\\{0.6,0.8\\}$ | $\\{(0.2,0.3,0.5)\\}$ | $\\{0.6,0.6\\}$ | $\\{0.5,0.7\\}$ ${y}_{4}$ | $\\{(0.5,0.5,0.7)\\}$ | $\\{0.4,0.5\\}$ | $\\{0.8,0.9\\}$ | $\\{(0.3,0.4,0.5)\\}$ ${y}_{5}$ | $\\{0.6,0.7\\}$ | $\\{0.5,0.7\\}$ | $\\{0.7,0.8\\}$ | $\\{(0.3,0.3,0.4)\\}$ some of its arrays are in form of triangular fuzzy numbers and some of them are HFEs. Needless to say that preserving both forms of arrays appears to maintain the implication of HOHFEs. According to Peng et al’s [21] assumption, the fuzzy measures of criteria $X=\\{x_{1},x_{2},x_{3},x_{4}\\}$ are as follows: $\displaystyle\mu(\emptyset)=0;$ $\displaystyle\mu(\\{x_{1}\\})=0.40,~{}\mu(\\{x_{2}\\})=0.25,~{}\mu(\\{x_{3}\\})=0.37,~{}\mu(\\{x_{4}\\})=0.20;$ $\displaystyle\mu(\\{x_{1},x_{2}\\})=0.60,~{}\mu(\\{x_{1},x_{3}\\})=0.70,~{}\mu(\\{x_{1},x_{4}\\})=0.56,~{}\mu(\\{x_{2},x_{3}\\})=0.68,$ $\displaystyle\quad\quad\quad\mu(\\{x_{2},x_{4}\\})=0.43,~{}\mu(\\{x_{3},x_{4}\\})=0.54;$ $\displaystyle\mu(\\{x_{1},x_{2},x_{3}\\})=0.88,~{}\mu(\\{x_{1},x_{2},x_{4}\\})=0.75,~{}\mu(\\{x_{1},x_{3},x_{4}\\})=0.84,~{}\mu(\\{x_{2},x_{3},x_{4}\\})=0.73;$ $\displaystyle\mu(X)=1.$ We re-arrange the HOHFEs corresponding to each criterion in descending order by the use of HOHFE score function $S_{\widetilde{h}}$ given by (10) as follows: $\displaystyle y_{1}:\quad\widetilde{h}(x_{\sigma(1)}=x_{3})\succ\widetilde{h}(x_{\sigma(2)}=x_{2})\succ\widetilde{h}(x_{\sigma(3)}=x_{1})\succ\widetilde{h}(x_{\sigma(4)}=x_{4});$ $\displaystyle y_{2}:\quad\widetilde{h}(x_{\sigma(1)}=x_{1})\succ\widetilde{h}(x_{\sigma(2)}=x_{3})\succ\widetilde{h}(x_{\sigma(3)}=x_{2})\succ\widetilde{h}(x_{\sigma(4)}=x_{4});$ $\displaystyle y_{3}:\quad\widetilde{h}(x_{\sigma(1)}=x_{1})\succ\widetilde{h}(x_{\sigma(2)}=x_{3})\succ\widetilde{h}(x_{\sigma(3)}=x_{4})\succ\widetilde{h}(x_{\sigma(4)}=x_{2});$ $\displaystyle y_{4}:\quad\widetilde{h}(x_{\sigma(1)}=x_{3})\succ\widetilde{h}(x_{\sigma(2)}=x_{1})\succ\widetilde{h}(x_{\sigma(3)}=x_{2})\succ\widetilde{h}(x_{\sigma(4)}=x_{4});$ $\displaystyle y_{5}:\quad\widetilde{h}(x_{\sigma(1)}=x_{3})\succ\widetilde{h}(x_{\sigma(2)}=x_{1})\succ\widetilde{h}(x_{\sigma(3)}=x_{2})\succ\widetilde{h}(x_{\sigma(4)}=x_{4}).$ Now, by the help of HOHF Choquet integral operator, we aggregate all HOHFEs in the i-$th$ row of the HOHFE decision matrix into an overall value, for example, $\displaystyle\widetilde{h}_{y_{1}}=HOHFEC_{\mu}(\widetilde{h}(x_{1}),...,\widetilde{h}(x_{4}))=$ $\displaystyle[(\mu(A_{\sigma(1)})-\mu(A_{\sigma(0)}))\widetilde{h}(x_{\sigma(1)})]\oplus[(\mu(A_{\sigma(2)})-\mu(A_{\sigma(1)}))\widetilde{h}(x_{\sigma(2)})]$ $\displaystyle\oplus[(\mu(A_{\sigma(3)})-\mu(A_{\sigma(2)}))\widetilde{h}(x_{\sigma(3)})]\oplus[(\mu(A_{\sigma(4)})-\mu(A_{\sigma(3)}))\widetilde{h}(x_{\sigma(4)})]$ $\displaystyle=[(\mu(\\{x_{\sigma(1)}\\})-\mu(\emptyset))\widetilde{h}(x_{\sigma(1)})]\oplus[(\mu(\\{x_{\sigma(1)},x_{\sigma(2)}\\})-\mu(\\{x_{\sigma(1)}\\}))\widetilde{h}(x_{\sigma(2)})]$ $\displaystyle\oplus[(\mu(\\{x_{\sigma(1)},x_{\sigma(2)},x_{\sigma(3)}\\})-\mu(\\{x_{\sigma(1)},x_{\sigma(2)}\\}))\widetilde{h}(x_{\sigma(3)})]\oplus[(\mu(X)-\mu(\\{x_{\sigma(1)},x_{\sigma(2)},x_{\sigma(3)}\\}))\widetilde{h}(x_{\sigma(4)})]$ $\displaystyle=[(\mu(\\{x_{3}\\})-\mu(\emptyset))\widetilde{h}(x_{3})]\oplus[(\mu(\\{x_{3},x_{2}\\})-\mu(\\{x_{3}\\}))\widetilde{h}(x_{2})]$ $\displaystyle\oplus[(\mu(\\{x_{3},x_{2},x_{1}\\})-\mu(\\{x_{3},x_{2}\\}))\widetilde{h}(x_{1})]\oplus[(\mu(X)-\mu(\\{x_{3},x_{2},x_{1}\\}))\widetilde{h}(x_{4})]$ $\displaystyle=\\{(0.0765,0.0795,0.1205),\\{0.0100,0.0120\\}\\}.$ By applying the same procedure, we achieve $\displaystyle\widetilde{h}_{y_{2}}=\\{(0.0550,0.0642,0.0672),\\{0.0533,0.0645\\}\\},$ $\displaystyle\widetilde{h}_{y_{3}}=\\{(0.0110,0.0165,0.0275),\\{0.0650,0.0750\\}\\},$ $\displaystyle\widetilde{h}_{y_{4}}=\\{(0.0210,0.0230,0.0310),\\{0.0713,0.0830\\}\\},$ $\displaystyle\widetilde{h}_{y_{5}}=\\{(0.0060,0.0060,0.0080),\\{0.1330,0.1633\\}\\}.$ Moreover, the HOHFE score function (11) results in $\displaystyle S_{\widetilde{h}}(\widetilde{h}_{y_{1}})=0.2985,\quad S_{\widetilde{h}}(\widetilde{h}_{y_{2}})=0.3041,\quad S_{\widetilde{h}}(\widetilde{h}_{y_{3}})=0.1950,\quad S_{\widetilde{h}}(\widetilde{h}_{y_{4}})=0.2293,\quad S_{\widetilde{h}}(\widetilde{h}_{y_{5}})=0.3163.$ Therefore, $\displaystyle S_{\widetilde{h}}(\widetilde{h}_{y_{5}})>S_{\widetilde{h}}(\widetilde{h}_{y_{2}})>S_{\widetilde{h}}(\widetilde{h}_{y_{1}})>S_{\widetilde{h}}(\widetilde{h}_{y_{4}})>S_{\widetilde{h}}(\widetilde{h}_{y_{3}}).$ These findings indicate that $\displaystyle y_{5}\succ y_{2}\succ y_{1}\succ y_{4}\succ y_{3},$ and hence the most appropriate energy project is $y_{5}$. In what follows, we are going to verify the validity of proposed technique compared to the other existing techniques including Xu’s [37], Wei’s [36], Zhang et al.’s [46], Chen et al.’s [5], Xu’s [39], Farhadinia’s [10], Zhang and Wei’s [47], Zhang and Xu’s [48], Wang et al.’s [33] and Peng et al.’s [21] from a perspective of performance. Indeed, from the performance stand point of view, we intend to discuss the resulted rankings of techniques macroscopy, and not microscopy, that is, we do not intend to investigate the pros and cons of each technique. Before going more in depth with future analysis, let us first re-state here from Farhadinia [12] the algorithm being implemented to sort decision making techniques based on their outcomes. ###### Algorithm 3.1. [12] (Sorting of decision making techniques based on their outcomes) Step 1. Take into account each outcome of decision making techniques (i.e., the resulted ranking order of alternatives) as an individual preference $R_{k}$ $(k=1,...,m)$ together with the associated dominance-vector $\pi_{R_{k}}$. Step 2. Add up all the individual preference matrices component-wisely to get the collective preference matrix $\overline{R}=[\overline{r}_{ij}]_{n\times n}$. Step 3. Extract the collective preference from the matrix $\overline{R}=[\overline{r}_{ij}]_{n\times n}$ with respect to the collective majority decision rule. Step 4. Characterize the corresponding dominance-vector hesitant fuzzy sets to the individual and the collective preferences. Then, determine the importance weight of decision making techniques in accordance with the distance value of their individual preference from the collective preference using a corresponding distance measure. Now, for determining the importance weights of decision making techniques according to their distance values of their individual rankings from the collective one, we implement here Algorithm 3.1 above. Indeed, Algorithm 3.1 is allocated to assess the best outcome(s) without regarding of how the preference ordering is obtained or from what numerical values one concludes the preference ordering. Now, by considering the total ranking order $R$ on $Y=\\{y_{1},y_{2},y_{3},y_{4},y_{5}\\}$, we can correspond to each of them the dominance-vectors $\pi_{R_{k}}$ (for $k=X1,W1,...,Pro$) which are given in Table 3. Table 3. The dominance-vectors corresponding to the different techniques. Techniques | Ranking orders | Dominance-vector ---|---|--- $(e_{X1})$: Xu’s [37] technique | $R_{X1}$: $y_{4}\succ y_{5}\succ y_{2}\succ y_{1}\succ y_{3}$ | $\pi_{R_{X1}}=(2,3,1,4,5)$ $(e_{W1})$: Wei’s [36] technique | $R_{W1}$: $y_{2}\succ y_{5}\succ y_{4}\succ y_{1}\succ y_{3}$ | $\pi_{R_{W1}}=(2,5,1,3,4)$ $(e_{Z1})$: Zhang et al.’s [46] technique | $R_{Z1}$: $y_{5}\succ y_{2}\succ y_{1}\succ y_{4}\succ y_{3}$ | $\pi_{R_{Z1}}=(3,4,1,2,5)$ $(e_{C})$: Chen et al.’s [5] technique | $R_{C}$: $y_{5}\succ y_{1}\succ y_{2}\succ y_{4}\succ y_{3}$ | $\pi_{R_{C}}=(4,3,1,2,5)$ $(e_{X2})$: Xu’s [39] technique | $R_{X2}$: $y_{5}\succ y_{2}\succ y_{4}\succ y_{1}\succ y_{3}$ | $\pi_{R_{X2}}=(2,4,1,3,5)$ $(e_{F})$: Farhadinia’s [10] technique | $R_{F}$: $y_{5}\succ y_{1}\succ y_{2}\succ y_{4}\succ y_{3}$ | $\pi_{R_{F}}=(4,3,1,2,5)$ $(e_{Z2})$: Zhang and Wei’s [47] technique | $R_{Z2}$: $y_{5}\succ y_{1}\succ y_{2}\succ y_{4}\succ y_{3}$ | $\pi_{R_{Z2}}=(4,3,1,2,5)$ $(e_{Z3})$: Zhang and Xu’s [48] technique | $R_{Z3}$: $y_{5}\succ y_{1}\succ y_{2}\succ y_{3}\succ y_{4}$ | $\pi_{R_{Z3}}=(4,3,2,1,5)$ $(e_{W2})$: Wang et al.’s [33] technique | $R_{W2}$: $y_{5}\succ y_{1}\succ y_{2}\succ y_{3}\succ y_{4}$ | $\pi_{R_{W2}}=(4,3,2,1,5)$ $(e_{P})$: Peng et al.’s [21] technique | $R_{P}$: $y_{5}\succ y_{2}\succ y_{1}\succ y_{3}\succ y_{4}$ | $\pi_{R_{P}}=(3,4,2,1,5)$ $(e_{Pro})$: The proposed technique | $R_{Pro}$: $y_{5}\succ y_{2}\succ y_{1}\succ y_{4}\succ y_{3}$ | $\pi_{R_{Pro}}=(3,4,1,2,5)$ By considering all the individual preference matrices $R_{k}$ $(k=X1,W1,...,Pro)$, for instance, $R_{X1}$: $y_{4}\succ y_{5}\succ y_{2}\succ y_{1}\succ y_{3}$ which is given by $\displaystyle R_{X1}=\left(\begin{array}[]{ccccc}0&-1&1&-1&-1\\\ 1&0&1&-1&-1\\\ -1&-1&0&-1&-1\\\ 1&1&1&0&1\\\ 1&1&1&-1&0\\\ \end{array}\right),$ the collective preference $\overline{R}$ is then achieved through the aggregation process in which all the individual preference matrices are added up component-wisely. The result is as follows: $\displaystyle\overline{R}=\left(\begin{array}[]{ccccc}0&-1&11&5&-11\\\ 1&0&11&9&-9\\\ -11&-11&0&-5&-11\\\ -5&-9&5&0&-9\\\ 11&9&11&9&0\\\ \end{array}\right),$ which returns $y_{5}\succ y_{2}\succ y_{1}\succ y_{4}\succ y_{3}$ as the collective preference. Now, with the distance $\displaystyle{\mathcal{D}}({\pi_{R_{k}}},{\pi_{\overline{R}}})=\max\\{\max_{\pi_{R_{k}}}\min_{\pi_{\overline{R}}}|\pi_{R_{k}},\pi_{\overline{R}}|~{},~{}\min_{\pi_{R_{k}}}\max_{\pi_{\overline{R}}}|\pi_{R_{k}},\pi_{\overline{R}}|\\},\quad(k=X1,W1,...,Pro)$ (15) at hand, we may obtain the distance of individual preferences from the collective preference as those given in Table 4. Table 4. The distance of individual preferences from the collective preference. Techniques | Ranking orders | Distance values ---|---|--- $(e_{X1})$: Xu’s [37] technique | $R_{X1}$: $y_{4}\succ y_{5}\succ y_{2}\succ y_{1}\succ y_{3}$ | ${{\cal{D}}}(\pi_{R_{X1}},{\pi}_{\overline{R}})=4$ $(e_{W1})$: Wei’s [36] technique | $R_{W1}$: $y_{2}\succ y_{5}\succ y_{4}\succ y_{1}\succ y_{3}$ | ${{\cal{D}}}(\pi_{R_{W1}},{\pi}_{\overline{R}})=4$ $(e_{Z1})$: Zhang et al.’s [46] technique | $R_{Z1}$: $y_{5}\succ y_{2}\succ y_{1}\succ y_{4}\succ y_{3}$ | ${{\cal{D}}}(\pi_{R_{Z1}},{\pi}_{\overline{R}})=0$ $(e_{C})$: Chen et al.’s [5] technique | $R_{C}$: $y_{5}\succ y_{1}\succ y_{2}\succ y_{4}\succ y_{3}$ | ${{\cal{D}}}(\pi_{R_{C}},{\pi}_{\overline{R}})=2$ $(e_{X2})$: Xu’s [39] technique | $R_{X2}$: $y_{5}\succ y_{2}\succ y_{4}\succ y_{1}\succ y_{3}$ | ${{\cal{D}}}(\pi_{R_{X2}},{\pi}_{\overline{R}})=2$ $(e_{F})$: Farhadinia’s [10] technique | $R_{F}$: $y_{5}\succ y_{1}\succ y_{2}\succ y_{4}\succ y_{3}$ | ${{\cal{D}}}(\pi_{R_{F}},{\pi}_{\overline{R}})=2$ $(e_{Z2})$: Zhang and Wei’s [47] technique | $R_{Z2}$: $y_{5}\succ y_{1}\succ y_{2}\succ y_{4}\succ y_{3}$ | ${{\cal{D}}}(\pi_{R_{Z2}},{\pi}_{\overline{R}})=2$ $(e_{Z3})$: Zhang and Xu’s [48] technique | $R_{Z3}$: $y_{5}\succ y_{1}\succ y_{2}\succ y_{3}\succ y_{4}$ | ${{\cal{D}}}(\pi_{R_{Z3}},{\pi}_{\overline{R}})=4$ $(e_{W2})$: Wang et al.’s [33] technique | $R_{W2}$: $y_{5}\succ y_{1}\succ y_{2}\succ y_{3}\succ y_{4}$ | ${{\cal{D}}}(\pi_{R_{W2}},{\pi}_{\overline{R}})=4$ $(e_{P})$: Peng et al.’s [21] technique | $R_{P}$: $y_{5}\succ y_{2}\succ y_{1}\succ y_{3}\succ y_{4}$ | ${{\cal{D}}}(\pi_{R_{P}},{\pi}_{\overline{R}})=2$ $(e_{Pro})$: The proposed technique | $R_{Pro}$: $y_{5}\succ y_{2}\succ y_{1}\succ y_{4}\succ y_{3}$ | ${{\cal{D}}}(\pi_{R_{Pro}},{\pi}_{\overline{R}})=0$ The larger distance ${{\cal{D}}}$ indicates the more dis-similar between each individual preference and the collective preference. This implies that the larger value of distance shows lower performance of decision-making technique. In view of Table 4, we are able to sort the considered techniques which are labelled by $k=X1,W1,...,Pro$ as the following: $\displaystyle\\{(e_{Pro}),(e_{Z1})\\}>\\{(e_{C}),(e_{X2}),(e_{F}),(e_{Z2}),(e_{P})\\}>\\{(e_{X1}),(e_{W1}),(e_{Z3}),(e_{W2})\\}.$ The above results verify that among the above decision-making techniques, the proposed technique can be chosen as the one of the most suitable techniques. However, these findings indicate that although the procedure of obtaining the most appropriate alternative using different techniques may be different, but their outcomes can be used for constructing an admissible assessment which was called here as the collective preference. ## 4 Conclusion The current contribution proposed the higher order hesitant fuzzy (HOHF) Choquet integral operator that not only considers the importance of the elements, but also it can reflect the correlations among the elements. Then, we employed the HOHF Choquet integral operator to aggregate HOHFEs in a higher order hesitant fuzzy MCDM problem. Eventually, a comparative analysis of different techniques with the HOHF Choquet integral-based technique presented, and it indicates that the latter technique is enough efficient. In future work, we plan to study the other aggregation operators in the higher order hesitant fuzzy setting for handling MCDM with higher order hesitant fuzzy information. In addition, since the application potentials of HOHFS are diverse, it can be investigated in other domains such as clustering, pattern recognition, image processing, etc. ## References * [1] L. Abdullah, N. Zulkifli, H. Liao, E. Herrera-Viedma, A. Al-Barakati, An interval-valued intuitionistic fuzzy DEMATEL method combined with Choquet integral for sustainable solid waste management, Engineering Applications of Artificial Intelligence 82 (2019) 207-215. * [2] K.T. Atanassov, Intuitionistic fuzzy sets, Fuzzy Sets and Systems 20 (1986) 87-96. * [3] H. Bustince, J. Fernandez, J. Sanz, M. Galar, R. Mesiar, A. Kolesarova, Multicriteria decision making by means of interval-valued Choquet integrals, Advances in Intelligent and Soft Computing 107 (2012) 269-278. * [4] D. Candeloro, R. Mesiar, A.R. Sambucini, A special class of fuzzy measures: Choquet integral and applications, Fuzzy sets and Systems, 355 (2019) 83-99. * [5] N. Chen, Z.S. Xu, M.M. Xia, Correlation coefficients of hesitant fuzzy sets and their applications to clustering analysis, Applied Mathematical Modelling 37 (2013) 2197-2211. * [6] N. Chen, Z.S. Xu, M.M. Xia, Interval-valued hesitant preference relations and their applications to group decision making, Knowl. Based Syst. 37 (2013) 528-540. * [7] G. Choquet, Theory of capacities, Annales de lntuinstitut Fourier 5 (1953) 131-295. * [8] B. Farhadinia, A novel method of ranking hesitant fuzzy values for multiple attribute decision-making problems, Int. J. Intelligent Systems, DOI 10.1002/int.21600 (2013) 1-16. * [9] B. Farhadinia, Information measures for hesitant fuzzy sets and interval-valued hesitant fuzzy sets, Information Sciences 240 (2013) 129-144. * [10] B. Farhadinia, Distance and similarity measures for higher order hesitant fuzzy sets, Knowledge-Based Systems 55 (2014) 43-48. * [11] B. Farhadinia, Z.S. Xu, Information Measures for Hesitant Fuzzy Sets and Their Extensions, Singapore, Springer 2019\. * [12] B. Farhadinia, E. Herrera-Viedma, Sorting of decision-making methods based on their outcomes using dominance-vector hesitant fuzzy-based distance, Soft Computing 23 (2019) 1109-1121. * [13] L.C. Jang, B.M. Kil, Y.K. Kim, J.S. Kwon, Some properties of Choquet integrals of set-valued functions, Fuzzy Sets and Systems 91 (1997) 95-113. * [14] A. Kauffmann, M.M. Gupta, Introduction to fuzzy Arithmetic: Theory and Applications, Van Nostrand Reinhold, New York, 1991. * [15] G.J. Klir, B. Yuan, Fuzzy Sets and Fuzzy Logic: Theory and Applications, Prentice-Hall, Upper Saddle River, NJ, 1995. * [16] E.S. Lee, R.L. Li, Comparison of fuzzy numbers based on the probability measure of fuzzy events, Computers and Mathematics with Applications 15 (10) (1988) 887-896. * [17] J.L. Marichal, An axiomatic approach of the discrete Choquet integral as a tool to aggregate interacting criteria, IEEE Transactions on Fuzzy Systems 8 (6) (2000) 800-807. * [18] P. Meyer, M. Roubens, On the use of the Choquet integral with fuzzy numbers in multiple criteria decision support, Fuzzy Sets and Systems 157 (2006) 927-938. * [19] M. Mizumoto, K. Tanaka, Some properties of fuzzy sets of type-2, Inform. Control 31 (1976) 312-340. * [20] G. Pasi, M. Viviani, A. Carton, A Multi-Criteria Decision Making approach based on the Choquet integral for assessing the credibility of User-Generated Content, Information Sciences, 503 (2019) 574-588. * [21] J.J. Peng, J.Q. Wang, H. Zhou, X.H. Chen, A multi-criteria decision-making approach based onTODIM and Choquet integral within multiset hesitant fuzzy environment, Appl. Math. Inf. Sci. 9 (2015) 2087-2097. * [22] G. Qian, H. Wang, X. Feng, Generalized hesitant fuzzy sets and their application in decision support system, Knowledge Based Systems 37 (2013) 357-365. * [23] R.M. Rodriguez, L. Martinez, F. Herrera, Hesitant fuzzy linguistic term sets for decision making, IEEE Transactions on Systems 20 (2012) 109-119. * [24] L. Rui, Model for multiple attribute decision making with hesitant fuzzy information and their application, Int. J. Knowledge-based and Intelligent Engineering Systems 23 (2019) 181-189. * [25] M. Sugeno, Theory of fuzzy integrals and its applications, Ph.D. Dissertation, Tokyo Institute of Technology, 1974. * [26] C.Q. Tan, A multi-criteria interval-valued intuitionistic fuzzy group decision making with Choquet integral-based TOPSIS, Expert Systems with Applications 38 (2011) 3023C3033. * [27] C.Q. Tan, X.H. Chen, Intuitionistic fuzzy Choquet integral operator for multi-criteria decision making, Expert Systems with Applications 37 (2010) 149-157. * [28] V. Torra, Hesitant fuzzy sets, International Journal of Intelligent Systems 25 (2010) 529-539. * [29] Z. Wang, Convergence theorems for sequences of Choquet integral, International Journal of General Systems 26 (1997) 133-143. * [30] W. Wang, J.M. Mendel, Multiple attribute group decision making with linguistic variables and complete unknown weight information, Iranian Journal of Fuzzy Systems, 16 (2019) 145-157. * [31] J.Q. Wang, D.D. Wang, H.Y. Zhang, X.H. Chen, Multicriteria group dcision making method based on interval 2-tuple linguistic and Choquet integral aggregation operators, Soft Computing, doi:10.1007/s00500-014-1259-z, 2014. * [32] L. Wang, M. Ni, Z. Yu, L. Zhu, Power geometric operators of hesitant multiplicative fuzzy numbers and their application to multiple attribute group decision making. Math. Probl. Eng. 2014, 186502 (2014). doi:10.1155/2014/186502. * [33] J.Q. Wang, D.D. Wang, H.Y. Zhang, X.H. Chen, Multicriteria outranking approach with hesitant fuzzy sets, OR Spectrum 36 (2014) 1001-1019. * [34] Z. Wang, R. Yang, P.A. Heng, K.S. Leung, Real-valued Choquet integrals with fuzzy-valued integrand, Fuzzy Sets and Systems 157 (2006) 256-269. * [35] D.J. Yu, Y.Y. Wu, W. Zhou, Multi-criteria decision making based on Choquet integral under hesitant fuzzy environment, Journal of Computational Information Systems 7 (2011) 4506-4513. * [36] G.W. Wei, Hesitant fuzzy prioritized operators and their application to multiple attribute decision making, Knowledge-Based Systems 31 (2012) 176-182. * [37] M.M. Xia, Z.S. Xu, Hesitant fuzzy information aggregation in decision making, Int. J. Approximate Reasoning 52 (2011) 395-407. * [38] Z.S. Xu, Deviation measures of linguistic preference relations in group decision making, Omega 17 (2005) 432-445. * [39] Z.S. Xu, M.M. Xia, Distance and similarity measures for hesitant fuzzy sets, Information Sciences 181 (2011) 2128-2138. * [40] R.R. Yager, Generalized OWA aggregation operators, Fuzzy Optimization and Decision Making 3 (2004) 93-107. * [41] R.R. Yager, Induced aggregation operators, Fuzzy Sets and Systems 137 (2003) 59-69. * [42] R. Yang, Z. Wang, P.A. Heng, K.S. Leung, Fuzzy numbers and fuzzification of the Choquet integral, Fuzzy Sets and Systems 153 (2005) 95-113. * [43] J. Ye, Correlation coefficient of dual hesitant fuzzy sets and its application to multiple attribute decision making, Appl. Math. Modell. 38 (2014) 659-666. * [44] Z.S. Xu, R.R. Yager, Some geometric aggregation operators based on intuitionistic fuzzy sets, International Journal of General Systems 35 (2006) 417-433. * [45] D. Zhang, C. Guo, D. Liu, Set-valued Choquet integrals revisited, Fuzzy Sets and Systems 147 (2004) 475-485. * [46] Z.M. Zhang, C.Wang, D.Z. Tian, K. Li, Induced generalized hesitant fuzzy operators and their application to multiple attribute group decision making, Computers and Industrial Engineering 67 (2014) 116-138. * [47] N. Zhang, G.W. Wei, Extension of VIKOR method for decision making problem based on hesitant fuzzy set, Applied Mathematical Modeling 37 (2013) 4938-4947. * [48] X.L. Zhang, Z.S. Xu, The TODIM analysis approach based on novel measured functions under hesitant fuzzy environment, Knowledge-Based Systems 61 (2014) 48-58.
# Consistency of sample-based stationary points for infinite-dimensional stochastic optimization Johannes Milz H. Milton Stewart School of Industrial and Systems Engineering, Georgia Institute of Technology, Atlanta, GA 30332, USA <EMAIL_ADDRESS> (June 19, 2023) ###### Abstract We consider stochastic optimization problems with possibly nonsmooth integrands posed in Banach spaces and approximate these stochastic programs via a sample-based approaches. We establish the consistency of approximate Clarke stationary points of the sample-based approximations. Our framework is applied to risk-averse semilinear PDE-constrained optimization using the average value-at-risk and to risk-neutral bilinear PDE-constrained optimization. ###### keywords: stochastic programming, sample average approximation, optimization under uncertainty, PDE-constrained optimization, uncertainty quantification, bilinear optimal control AMS subject classifications. 65C05, 90C15, 35R60, 90C48, 90C30, 60H25, 49M41, 35Q93 ## 1 Introduction Infinite-dimensional optimization problems arise in a plethora of research fields such as dynamic programming [56], statistical estimation [26], feedback stabilization of dynamical systems [54], and optimization problems governed by partial differential equations (PDEs) [49]. PDE-constrained optimization is an active research field with a focus on modeling, analyzing and solving complex optimization problems with PDE constraints. For example, numerous applications in the field of renewable and sustainable energy yield challenging PDE- constrained optimization problems, such as wind plant layout optimization [44], wind turbine blade planform design [1], and tidal stream array optimization [28]. Parameters in PDEs may be uncertain, such as diffusion coefficients and boundary conditions, and as such can have a significant influence on the simulation output. Using parameter-dependent PDEs for decision making naturally leads to decision making under uncertainty. A common approach in optimization under uncertainty models uncertain parameters as random variables with known probability distribution. Using this stochastic programming approach, we can formulate an optimization problem having solutions performing best on average; its objective function is the expected value of a parameterized cost function, the integrand. Such optimization problems are nowadays often referred to as risk-neutral problems to distinguish them from risk-averse formulations. Since high-dimensional integrals cannot be accurately evaluated and solving PDEs requires numerical computations, we approximate the expectations using a Monte Carlo sample-based approximation, the sample average approximation (SAA) approach [45, 72, 74], and other schemes approximating the true probability distribution via weakly convergent probability distributions. In this manuscript, we demonstrate the consistency of Clarke-(C-)stationary points of the approximated problems towards the risk-neutral problem’s set of C-stationary points. Motivated by infinite-dimensional optimization problems governed by complex physical systems under uncertainty, we consider the risk-neutral composite optimization problem $\displaystyle\min_{x\in X}\,\int_{\Xi}\,F_{\xi}(x)\,\mathrm{d}\mathbb{P}(\xi)+\psi(x),$ (1.1) where $X$ is a reflexive Banach space, $\psi\colon X\to(-\infty,\infty]$ is a proper, convex, lower semicontinuous function having the so-called Kadec property, $\mathbb{P}$ is a probability measure on the complete, separable metric space $\Xi$, and $F_{\xi}$ is a locally Lipschitz continuous integrand with integrable Lipschitz constants. In this manuscript, we consider two types of approximations of the expectations in (1.1). While our main focus is on analyzing the SAA approach [45, 72], we also consider approximations of $\mathbb{P}$ via weakly converging probability measures [56, 22]. To introduce, the SAA approach, let $\xi^{1}$, $\xi^{2},\ldots$ be independent identically distributed $\Xi$-valued random elements such that each $\xi^{i}$ has distribution $\mathbb{P}$. The SAA problem with sample size $N\in\mathbb{N}$ corresponding to (1.1) is given by $\displaystyle\min_{x\in X}\,\frac{1}{N}\sum_{i=1}^{N}F_{\xi^{i}}(x)+\psi(x).$ (1.2) The principle contributions of this manuscript are twofold. 1. (i) We demonstrate the consistency of approximate C-stationary points of the SAA problem (1.2) for a class of risk-neutral nonlinear optimization problems (1.1) by adapting and extending a framework developed in [59], where the consistency of SAA optimal values and SAA solutions has been demonstrated. More specifically, our analysis relies on the identity $\displaystyle\partial_{C}F_{\xi}(x)=BM_{\xi}(x)\quad\text{for all}\quad(x,\xi)\in X\times\Xi,$ (1.3) where $\partial_{C}F_{\xi}(x)$ is Clarke’s subdifferential of $F_{\xi}$ at $x$, $V$ is a Banach space, $B\colon V\to X^{*}$ is linear and compact, and $M_{\xi}\colon X\rightrightarrows V$ is a multifunction for each $\xi\in\Xi$. Moreover, using proof techniques developed in [62], we establish consistency statements for the significantly larger class of functions $\psi$ than those considered in [59], so-called functions having the Kadec property [9]. 2. (ii) We apply our framework to risk-averse semilinear PDE-constrained optimization problems using the average value-at-risk and to steady-state elliptic bilinear optimal control problems under uncertainty. While the literature on risk- neutral and risk-averse semilinear PDE-constrained optimization is extensive [24, 48, 50, 64], bilinear PDE-constrained problems under uncertainty have not been analyzed, but serve as yet another class of nonconvex problems and arise in many applications. The consistency of SAA solutions and SAA stationary points of nonlinear possibly infinite-dimensional stochastic programs are typically based on some form of the feasible set’s compactness. In particular, the consistency results for SAA stationary points of infinite-dimensional stochastic programs established in [5, 78] require the feasible set’s compactness. However, compact sets in infinite dimensions may regarded as rare phenomena, given the fact that closed unit balls in Banach spaces are compact if and only if the spaces are finite-dimensional. This fact complicates the consistency analysis for nonconvex stochastic programs. For strongly convex risk-neutral PDE- constrained optimization, the SAA approach has been analyzed in [38, 58, 71, 60, 61] without requiring the compactness of the feasible set. Contributions to analyzing the SAA approach for optimization problems governed by nonlinear operator equations with random inputs are relatively recent. The SAA objective function’s almost sure epiconvergence and the SAA stationary points’ weak consistency is analyzed in [68] for optimization problems governed by ordinary differential equations with random inputs. Utilizing epiconvergence, performance guarantees of optimal values of PDE-constrained optimization problems under uncertainty are provided in [16] with respect to sample average approximations and control and state space discretizations, for example. The asymptotic consistency of SAA optimal values and SAA solutions to risk-averse PDE-constrained optimization problems has recently been demonstrated in [62] using (Mosco-)epiconvergence, and sample size estimates for risk-neutral semilinear PDE-constrained problems are derived in [64]. We use the compact operator $B$ in (1.3) to mathematically model problem structure typically found in PDE-constrained optimization problems. Related problem characteristics are used for establishing error estimates for finite- dimensional approximations to PDE-constrained optimization problems [13, 20, 52] and for algorithmic design and analysis in function space [33, 36, 35, 80], for example. The analysis of bilinear PDE-constrained optimization problems is generally more challenging than those of their linear counterparts, as bilinear PDEs may lack solutions for each point in the control space and bilinear control problems are generally nonconvex. Bilinear control problems can be governed by steady-state elliptic PDEs [52, 81], convection-diffusion equations [11], and advection-reaction-diffusion equations [27], for instance. A substantial body of literature is available on the analysis, algorithmic design, and applications of deterministic bilinear control problems. We refer the reader to [14, 27, 31] for the analysis of deterministic bilinear control problems. Finite element error estimates are established in [14, 23, 52, 82, 84] and algorithms are designed in [11, 27, 41, 43, 53, 81]. Bilinear control problems arise in medical image analysis [57], damping design [75], groundwater remediation system design [30], optical flow [40], PDE-constrained regression problems [65], and chemorepulsion production [31], for example. ## Outline We introduce notation and terminology in section 2. Section 3 establishes consistency of C-stationary points of optimization problems obtained through approximations of infinite-dimensional nonconvex optimization posed in reflexive Banach spaces. This basic result is used to establish consistency of SAA C-stationary points in section 4 and of C-stationary points of approximate stochastic programs defined by weakly convergent probability measures in section 5. In section 6, we apply the framework developed in section 4 to risk-averse semilinear PDE-constrained optimization using the average value- at-risk. Section 7 discusses the application of our theory to a risk-neutral bilinear PDE-constrained optimization problem. We summarize our contributions and discuss open research questions in section 8. ## 2 Notation and preliminaries Metric spaces are defined over the real numbers and equipped with their Borel sigma-algebra if not specified otherwise. For a Hilbert space $H$, we denote by $(\cdot,\cdot)_{H}$ its inner product. Let $X_{1}$ and $X_{2}$ be Banach spaces. The space of linear, bounded operators from $X_{1}$ to $X_{2}$ is denoted by $\mathscr{L}(X_{1},X_{2})$. An operator $A\in\mathscr{L}(X_{1},X_{2})$ is compact if the image $A(W)$ is precompact in $X_{2}$ for each bounded set $W\subset X_{1}$. The operator $A^{*}\in\mathscr{L}(X_{2}^{*},X_{1}^{*})$ is the adjoint operator of $A\in\mathscr{L}(X_{1},X_{2})$. We denote by $\mathrm{D}f$ the Fréchet derivative of $f$ and by $\mathrm{D}_{y}f$ or $f_{y}$ the Fréchet derivative with respect to $y$. The dual to a Banach space $\Lambda$ is $\Lambda^{*}$ and we use $\langle\cdot,\cdot\rangle_{{\Lambda}^{*}\\!,\Lambda}$ to denote the dual pairing between $\Lambda^{*}$ and $\Lambda$. Let $X_{0}\subset X$ be an open, nonempty subset of a Banach space $X$. We denote Clarke’s generalized directional derivative of a locally Lipschitz continuous mapping $f\colon X_{0}\to$ at $x\in X_{0}$ in the direction $h\in X$ by $f^{\circ}(x;h)$ and Clarke’s subdifferential at $x$ by $\partial_{C}f(x)\subset X^{*}$ [18, pp. 25 and 27]. We refer the reader to [18, p. 39] for the definition of Carke regularity of a function $f$. We use $\partial f$ to denote the usual convex subdifferential of a function $f$. We define the distance $\mathrm{dist}({v,\Upsilon})$ from $v\in\Lambda\subset X$ to $\Upsilon\subset X$ by $\displaystyle\mathrm{dist}({v,\Upsilon})=\inf_{w\in\Upsilon}\,\|v-w\|_{X}\quad\text{if}\quad\Upsilon\neq\emptyset,\quad\text{and}\quad\mathrm{dist}({v,\Upsilon})=\infty\quad\text{otherwise}.$ and the deviation $\mathbb{D}({\Lambda,\Upsilon})$ between the sets $\Lambda$ and $\Upsilon$ by $\mathbb{D}({\Lambda,\Upsilon})=\sup_{v\in\Lambda}\,\mathrm{dist}({v,\Upsilon})$. Here $\|\cdot\|_{X}$ is a norm of $X$. If $\phi\colon X\to(-\infty,\infty]$, then $\mathrm{dom}(\phi)=\\{x\in X\colon\phi(x)<\infty\\}$ denotes its domain. Following [9, p. 147] and [10, p. 306], we say that a proper function $\phi\colon X\to(-\infty,\infty]$ has the Kadec property if for each $\bar{x}\in\mathrm{dom}(\phi)$, we have $x_{k}\to\bar{x}$ whenever $\phi(x_{k})\to\phi(\bar{x})$, $(x_{k})\subset\mathrm{dom}(\psi)$, and $x_{k}\rightharpoonup\bar{x}$. Here $\to$ denotes strong convergence and $\rightharpoonup$ denotes weak convergence. For $x\in X$ and $\varepsilon\geq 0$, we denote by $\mathbb{B}_{X}(x;\varepsilon)$ the open ball about $x$ in $X$ with radius $\varepsilon$ and by $\overline{\mathbb{B}}_{X}(x;\varepsilon)$ its closure in $X$. The following notation is used in sections 7 and 6. Let $D\subset^{d}$ be a bounded domain. The space $C^{0,1}(\bar{D})$ is the set of Lipschitz continuous functions on $\bar{D}$, $L^{p}(D)$ ($1\leq p\leq\infty$) and $H^{1}(D)$, $H^{2}(D)$, and $H_{0}^{1}(D)$ are the “usual” Lebesgue and Sobolev spaces, respectively. The space $H_{0}^{1}(D)$ is equipped with the norm $\|v\|_{H_{0}^{1}(D)}=\|\nabla v\|_{L^{2}(D)^{d}}$ and the dual to $H_{0}^{1}(D)$ is denoted by $H^{-1}(D)$. Moreover, let $C_{D}\in(0,\infty)$ be Friedrichs’ constant of $D$. If $X$ is a reflexive Banach space, then we identify $(X^{*})^{*}$ with $X$ and write $(X^{*})^{*}=X$. Moreover, we identify $L^{2}(D)^{*}$ with $L^{2}(D)$ and write $L^{2}(D)^{*}=L^{2}(D)$. We denote by $\iota_{0}\colon H_{0}^{1}(D)\to L^{2}(D)$ the embedding operator of the compact embedding $H_{0}^{1}(D)\xhookrightarrow{}L^{2}(D)$ and by $\iota_{1}\colon H^{1}(D)\to L^{2}(D)$ that of $H^{1}(D)\xhookrightarrow{}L^{2}(D)$. We introduce further application- specific notation in section 7. ## 3 Consistency of C-stationary points We consider the optimization problem $\displaystyle\min_{x\in\mathrm{dom}(\psi)}\,f(x)+\psi(x)$ (3.1) and its approximations $\displaystyle\min_{x\in\mathrm{dom}(\psi)}\,f_{k}(x)+\psi(x).$ (3.2) We explicitly add $\mathrm{dom}(\psi)$ as a constraint sets in (3.1) and (3.2), as $f$ and $f_{k}$ may only be finite-valued on an open neighborhood of $\mathrm{dom}(\psi)$. We provide conditions sufficient for the asymptotic consistency of C-stationary points of the approximated problems (3.2) to those of “true” problem (3.1). Our analysis is inspired by those in [72, 74, 17, 62] and based on technical assumptions used in [59]. Using these consistency results, we establish asymptotic consistency of C-stationary points of infinite-dimensional stochastic programs in sections 4 and 5. ###### Assumption 3.1. 1. 1. The spaces $X$ and $V$ are Banach spaces, $X$ is reflexive, and $X_{0}\subset X$ is nonempty and open. 2. 2. The function $\psi\colon X\to(-\infty,\infty]$ is proper, convex, and lower semicontinuous. Moreover, $\psi$ has the Kadec property and $\mathrm{dom}(\psi)\subset X_{0}$. 3. 3. The functions $f\colon X_{0}\to$ and $f_{k}\colon X_{0}\to$, $k\in\mathbb{N}$, are locally Lipschitz continuous. 4. 4. The operator $B\in\mathscr{L}(V,X^{*})$ is compact, $M_{k}\colon\mathrm{dom}(\psi)\rightrightarrows V$ is a set-valued mapping, and $\partial_{C}f_{k}(x)=BM_{k}(x)$ for all $x\in\mathrm{dom}(\psi)$ and $k\in\mathbb{N}$. 5. 5. For each $\bar{x}\in\mathrm{dom}(\psi)$, each sequence $(x_{k})\subset\mathrm{dom}(\psi)$ with $x_{k}\rightharpoonup\bar{x}$, and each sequence $(v_{k})$ with $v_{k}\in M_{k}(x_{k})$ for all $k\in\mathbb{N}$, $\limsup_{k\to\infty}\,\|v_{k}\|_{V}<\infty$. 6. 6. For each $h\in X$, each $\bar{x}\in\mathrm{dom}(\psi)$, and each sequence $(x_{k})\subset\mathrm{dom}(\psi)$ with $x_{k}\to\bar{x}$, it holds that $\limsup_{k\to\infty}f_{k}^{\circ}(x_{k};h)\leq f^{\circ}(\bar{x};h)$. Under 3.1.1, 3.1.3, and 3.1.2, Clarke’s subdifferential $\partial_{C}f_{k}(x)$ for each $x\in X_{0}$ is nonempty, convex, bounded, and weakly∗-compact [18, Prop. 2.1.2]. Hence the set-valued mapping $M_{k}$ in 3.1.4 is nonempty- valued. If $M_{k}$ has bounded images, then 3.1.4 ensures that $\partial_{C}f_{k}(x)$ is precompact-valued. For Monte Carlo sample-based approximations of stochastic programs, we verify 3.1.6 using an epigraphical law of large numbers [3], basic properties of Clarke’s generalized directional derivative, and by imposing Clarke regularity of integrands studied in section 4. 3.1.4 is satisfied for large problem classes as demonstrated next. ###### Example 3.2. Let 3.1.1, 3.1.2, and 3.1.3 hold. If $W$ is a Banach space, $\iota\in\mathscr{L}(X,W)$ is compact, $F_{k}\colon W\to$ is locally Lipschitz continuous and Clarke regular, and $f_{k}\colon X\to$ is defined by $f_{k}(x)=F_{k}(\iota x)$, then $\partial_{C}f_{k}(x)=\iota^{*}\partial_{C}F_{k}(\iota x)$ [18, Thm. 2.3.10], and we may choose $B=\iota^{*}$ and $M_{k}(x)=\partial_{C}F_{k}(\iota x)$ to satisfy 3.1.4. Composite objective functions with compact linear operators arise in nonsmooth optimal control [32, 33, 35, 76], for example. Next we discuss 3.1.5 and 3.1.4 when $X=^{n}$. ###### Remark 3.3. Let 3.1.1, 3.1.2, and 3.1.3 hold. Let $X=^{n}$ (equipped with the Euclidean norm $\|\cdot\|_{2}$) and let $X^{*}=V$. We identify the dual to n with n. Let $B\colon^{n}\to^{n}$ be the identity mapping. Then 3.1.4 holds true with $M_{k}=\partial_{C}f_{k}$. Next we discuss 3.1.5. Let $\bar{x}\in\mathrm{dom}(\psi)$ and let $L_{k}\geq 0$ be Lipschitz constant of $f_{k}$ on $\mathbb{B}_{X}(\bar{x};\varepsilon)$, where $\varepsilon=\varepsilon(\bar{x})>0$. If $\limsup_{k\to\infty}\,L_{k}<\infty$, then 3.1.5 holds true. Let us verify this assertion. Let $(x_{k})\subset\mathrm{dom}(\psi)$ with $x_{k}\rightharpoonup\bar{x}$, and let $(v_{k})$ with $v_{k}\in M_{k}(x_{k})$ for all $k\in\mathbb{N}$. Then $x_{k}\to\bar{x}$ and there exists $K=K(\bar{x})$ such that $x_{k}\in\mathbb{B}_{X}(\bar{x};\varepsilon)$ for all $k\geq K$. Since $\partial_{C}f_{k}(x_{k})=M_{k}(x_{k})$, [18, Prop. 2.1.2] ensures $\|v_{k}\|_{2}\leq L_{k}$ for all $k\geq K$. Hence $\limsup_{k\to\infty}\,\|v_{k}\|_{2}<\infty$. Theorem 3.4 establishes consistency of approximate C-stationary points. Let $\mathcal{C}$ be the set of C-stationary points of (3.1), that is, $\mathcal{C}$ is the set of all points $x\in\mathrm{dom}(\psi)$ with $0\in\partial_{C}f(x)+\partial\psi(x)$. Let $\varepsilon\geq 0$ and let $\mathcal{C}_{k}^{\varepsilon}$ be the set of all points $x\in\mathrm{dom}(\psi)$ with $0\in\bigcup_{y\in\overline{\mathbb{B}}_{X}(x;\varepsilon)\cap\mathrm{dom}(\psi)}\partial_{C}f_{k}(y)+\partial\psi(x)$ (cf. [74, eqns. (3.2) and (4.2)]). Each point in $\mathcal{C}_{k}^{\varepsilon}$ is referred to as an approximate C-stationary point of (3.2). ###### Theorem 3.4. Let Assumption 3.1 hold and let $(\varepsilon_{k})\subset[0,\infty)$ satisfy $\varepsilon_{k}\to 0$ as $k\to\infty$. Moreover, let $\mathcal{C}_{k}^{\varepsilon_{k}}$ be nonempty for each $k\in\mathbb{N}$. 1. (i) If $(x_{k})$ is a bounded sequence with $x_{k}\in\mathcal{C}_{k}^{\varepsilon_{k}}$ for each $k\in\mathbb{N}$, then $\mathcal{C}$ is nonempty and $\mathrm{dist}({x_{k},\mathcal{C}})\to 0$ as $k\to\infty$. 2. (ii) If there exists a bounded set $\mathscr{C}\subset X$ with $\mathcal{C}_{k}^{\varepsilon_{k}}\subset\mathscr{C}$ for all sufficiently large $k\in\mathbb{N}$, then $\mathbb{D}({\mathcal{C}_{k}^{\varepsilon_{k}},\mathcal{C}})\to 0$ as $k\to\infty$. ###### Proof. 3.1.3 ensures that the Clarke subdifferentials of $f$ and $f_{k}$ are well- defined [18, Prop. 2.1.2]. (i) Let $(\mathrm{dist}({x_{k},\mathcal{C}}))_{K_{0}}$ be a subsequence of $(\mathrm{dist}({x_{k},\mathcal{C}}))\subset[0,\infty]$. Since $x_{k}\in\mathcal{\mathcal{C}}_{k}^{\varepsilon_{k}}$, there exists $y_{k}\in\mathrm{dom}(\psi)$ with $\|x_{k}-y_{k}\|_{X}\leq\varepsilon_{k}$ and $g_{k}\in\partial_{C}f_{k}(y_{k})$ with $\displaystyle\langle g_{k},x-x_{k}\rangle_{{X}^{*}\\!,X}+\psi(x)\geq\psi(x_{k})\quad\text{for all}\quad x\in X.$ (3.3) The boundedness of $(x_{k})$ implies that of $(x_{k})_{K_{0}}$. Combined with the fact that $X$ is reflexive (see 3.1.1) and that $\mathrm{dom}(\psi)$ is weakly sequentially closed (see 3.1.2 and [7, Thm. 2.23 (ii)]), we find that $(x_{k})_{K_{0}}$ has a subsequence $(x_{k})_{K_{1}}$ with $x_{k}\rightharpoonup\bar{x}\in\mathrm{dom}(\psi)$ as $K_{1}\ni k\to\infty$. Since $\|x_{k}-y_{k}\|_{X}\leq\varepsilon_{k}$ and $\varepsilon_{k}\to 0$, we find that $y_{k}\rightharpoonup\bar{x}$ as $K_{1}\ni k\to\infty$. Since $g_{k}\in\partial_{C}f_{k}(x_{k})$, 3.1.4 ensures $g_{k}=Bv_{k}$ for some $v_{k}\in M_{k}(y_{k})$. 3.1.5 implies that $(v_{K})_{K_{1}}$ is bounded. Combined with the compactness of $B$ (see 3.1.4), we obtain that $(g_{k})_{K_{1}}\subset X^{*}$ has a further subsequence $(g_{k})_{K_{2}}$ with $g_{k}\to\bar{g}$ as $K_{2}\ni k\to\infty$ [51, Thm. 8.1-5]. Hence $\langle g_{k},x-x_{k}\rangle_{{X}^{*}\\!,X}\to\langle\bar{g},x-\bar{x}\rangle_{{X}^{*}\\!,X}$ as $K_{2}\ni k\to\infty$ [7, Thm. 2.23 (iv)]. 3.1.2 implies that $\psi$ is weakly lower semicontinuous. Hence $\liminf_{K_{2}\ni k\to\infty}\,\psi(x_{k})\geq\psi(\bar{x})$. Putting together the pieces and using (3.3), we have $\displaystyle\langle\bar{g},x-\bar{x}\rangle_{{X}^{*}\\!,X}+\psi(x)\geq\psi(\bar{x})\quad\text{for all}\quad x\in X.$ (3.4) Using (3.3) once more, we obtain $\psi(\bar{x})=\lim_{K_{2}\ni k\to\infty}\langle g_{k},\bar{x}-x_{k}\rangle_{{X}^{*}\\!,X}+\psi(\bar{x})\geq\limsup_{K_{2}\ni k\to\infty}\psi(x_{k})$. Hence $\psi(x_{k})\to\psi(\bar{x})$ as $K_{2}\ni k\to\infty$. Since $\psi$ is Kadec according to 3.1.2, $x_{k}\to\bar{x}$ as $K_{2}\ni k\to\infty$. Since $\|x_{k}-y_{k}\|_{X}\leq\varepsilon_{k}$, we also have $y_{k}\to\bar{x}$ as $K_{2}\ni k\to\infty$. We define $\widetilde{x}_{k}=y_{k}$ if $k\in K_{2}$ and $\widetilde{x}_{k}=\bar{x}$ otherwise. We have $\widetilde{x}_{k}\to\bar{x}$ as $k\to\infty$. Fix $h\in X$. Since $g_{k}\in\partial_{C}f_{k}(y_{k})$, $f_{k}^{\circ}(y_{k};h)\geq\langle g_{k},h\rangle_{{X}^{*}\\!,X}$ [18, Prop. 2.1.5]. Combined with 3.1.6, we have $\displaystyle\langle\bar{g},h\rangle_{{X}^{*}\\!,X}=\lim_{K_{2}\ni k\to\infty}\,\langle g_{k},h\rangle_{{X}^{*}\\!,X}\leq\limsup_{K_{2}\ni k\to\infty}\,f_{k}^{\circ}(y_{k};h)\leq\limsup_{k\to\infty}\,f_{k}^{\circ}(\widetilde{x}_{k};h)\leq f^{\circ}(\bar{x};h).$ Hence $\bar{g}\in\partial_{C}f(\bar{x})$ [18, Prop. 2.1.5]. Now (3.4) ensures $0\in\partial_{C}f(\bar{x})+\partial\psi(\bar{x})$, that is, $\bar{x}\in\mathcal{C}$. Hence $\mathrm{dist}({x_{k},\mathcal{C}})\leq\|x_{k}-\bar{x}\|_{X}\to 0$ as $K_{2}\ni k\to 0$, and $(\mathrm{dist}({x_{k},\mathcal{C}}))$ is bounded. Our derivations also show that each subsequence of $(\mathrm{dist}({x_{k},\mathcal{C}}))$ has a further subsequence converging to zero. Hence $\mathrm{dist}({x_{k},\mathcal{C}})\to 0$ as $k\to\infty$. (ii) Since $\mathcal{C}_{k}^{\varepsilon_{k}}\subset\mathscr{C}$, and $\mathcal{C}_{k}^{\varepsilon_{k}}$ and $\mathcal{C}$ are nonempty, and $\mathscr{C}$ is bounded, we have $\mathbb{D}({\mathcal{C}_{k}^{\varepsilon_{k}},\mathcal{C}})\leq\mathbb{D}({\mathscr{C},\mathcal{C}})<\infty$ for all sufficiently large $k\in\mathbb{N}$. Hence for all sufficiently large $k\in\mathbb{N}$, there exists $x_{k}\in\mathcal{C}_{k}^{\varepsilon_{k}}$ with $\mathbb{D}({\mathcal{C}_{k}^{\varepsilon_{k}},\mathcal{C}})\leq\mathrm{dist}({x_{k},\mathcal{C}})+1/k$. Using part (i), we have $\mathrm{dist}({x_{k},\mathcal{C}})\to 0$ as $k\to\infty$. Hence $\mathbb{D}({\mathcal{C}_{k}^{\varepsilon_{k}},\mathcal{C}})\to 0$ as $k\to\infty$. ∎ ## 4 Empirical approximations via Monte Carlo sampling Let $\Xi$ be a complete, separable metric space, and let $(\Xi,\mathcal{A},\mathbb{P})$ be a complete probability space. Let $\xi^{1}$, $\xi^{2},\ldots$ be independent identically distributed $\Xi$-valued random elements defined on a complete probability space $(\Omega,\mathcal{F},P)$ such that each $\xi^{i}$ has distribution $\mathbb{P}$. We consider the risk-neutral optimization problem $\displaystyle\min_{x\in\mathrm{dom}(\psi)}\,\int_{\Xi}\,F_{\xi}(x)\,\mathrm{d}\mathbb{P}(\xi)+\psi(x)$ (4.1) and its SAA problem $\displaystyle\min_{x\in\mathrm{dom}(\psi)}\,\frac{1}{N}\sum_{i=1}^{N}F_{\xi^{i}}(x)+\psi(x).$ (4.2) We study the almost sure convergence of SAA C-stationary points of (4.2) to the C-stationary point’s set of (4.1). We analyze the consistency under conditions related to those in Assumption 3.1. ###### Assumption 4.1. 1. 1. The spaces $X$ and $V$ are separable Banach spaces, $X$ is reflexive, and $X_{0}\subset X$ is nonempty and open. 2. 2. The function $\psi\colon X\to(-\infty,\infty]$ is proper, convex, and lower semicontinuous. Moreover, $\psi$ has the Kadec property and $\mathrm{dom}(\psi)\subset X_{0}$. 3. 3. For each $\bar{x}\in\mathrm{dom}(\psi)$, there exists an open neighborhood $\mathcal{V}_{\bar{x}}\subset X_{0}$ of $\bar{x}$ and a $\mathbb{P}$-integrable random variable $L_{\bar{x}}\colon\Xi\to[0,\infty)$ such that for each $\xi\in\Xi$, $F_{\xi}\colon X_{0}\to$ is Lipschitz continuous on $\mathcal{V}_{\bar{x}}$ with Lipschitz constant $L_{\bar{x}}(\xi)$. 4. 4. For each $\xi\in\Xi$, $F_{\xi}\colon X_{0}\to$ is Clarke regular. 5. 5. For each $x\in X_{0}$, $\Xi\ni\xi\mapsto F_{\xi}(x)\in$ is measurable and $\mathbb{P}$-integrable. 6. 6. The operator $B\in\mathscr{L}(V,X^{*})$ is compact, $M_{\xi}\colon\mathrm{dom}(\psi)\rightrightarrows V$ is a set-valued map with nonempty images for each $\xi\in\Xi$, and $\partial_{C}F_{\xi}(x)=BM_{\xi}(x)$ for each $(x,\xi)\in\mathrm{dom}(\psi)\times\Xi$. 7. 7. There exists $\Omega_{0}\subset\Omega$ with $\Omega_{0}\in\mathcal{F}$ and $P(\Omega_{0})=1$ such that for each $\bar{x}\in\mathrm{dom}(\psi)$, each sequence $(x_{N})\subset\mathrm{dom}(\psi)$ with $x_{N}\rightharpoonup\bar{x}$, and each sequence $(v_{N})\subset V$, we have $\limsup_{N\to\infty}\,\|v_{N}\|_{V}<\infty$ whenever $\omega\in\Omega_{0}$ and $v_{N}\in(1/N)\sum_{i=1}^{N}M_{\xi^{i}(\omega)}(x_{N})$ for each $N\in\mathbb{N}$. 4.1.1 and 4.1.2 correspond to 3.1.1 and 3.1.2, but we impose in addition separability of $X$ and $V$. 4.1.1, 4.1.4, and 4.1.3 allow us to use Clarke’s calculus for generalized gradients of integral functions [18, sect. 2.7]. To apply the epigraphical law of large numbers derived in [3], we use 4.1.4 and 4.1.3. 4.1.6 and 4.1.7 generalize technical conditions used in [59]. ###### Lemma 4.2. Let $\zeta\colon\Xi\to[0,\infty)$ be measurable and $\mathbb{P}$-integrable, and let 4.1.1, 4.1.6, and 4.1.2 hold. If $\|m_{\xi}(x)\|_{V}\leq\zeta(\xi)$ for all $m_{\xi}(x)\in M_{\xi}(x)$, each $x\in\mathrm{dom}(\psi)$, and every $\xi\in\Xi$, then 4.1.7 holds true. ###### Proof. Since $\zeta(\xi^{1}),\zeta(\xi^{2}),\ldots$ are independent, the law of large numbers ensures that $(1/N)\sum_{i=1}^{N}\zeta(\xi^{i})$ converges almost surely to $\int_{\Xi}\zeta(\xi)\,\mathrm{d}\mathbb{P}(\xi)$ as $N\to\infty$. Hence there exists a set $\Omega_{0}\subset\Omega$ with $\Omega_{0}\in\mathcal{F}$, $P(\Omega_{0})=1$, and for each $\omega\in\Omega_{0}$, we have $(1/N)\sum_{i=1}^{N}\zeta(\xi^{i}(\omega))\to\int_{\Xi}\zeta(\xi)\,\mathrm{d}\mathbb{P}(\xi)$ as $N\to\infty$. Fix $\omega\in\Omega_{0}$. Let $\bar{x}\in\mathrm{dom}(\psi)$, let $(x_{N})\subset\mathrm{dom}(\psi)$ fulfill $x_{N}\rightharpoonup\bar{x}$, and let $(v_{N})$ satisfy $v_{N}\in(1/N)\sum_{i=1}^{N}M_{\xi^{i}(\omega)}(x_{N})$ for all $N\in\mathbb{N}$. The latter ensures the existence of $v_{N}^{i}\in M_{\xi^{i}(\omega)}(x_{N})$ with $v_{N}=(1/N)\sum_{i=1}^{N}v_{N}^{i}$. Now the lemma’s hypotheses ensure $\|v_{N}^{i}\|_{V}\leq\zeta(\xi^{i})$. Hence $\|v_{N}\|_{V}\leq(1/N)\sum_{i=1}^{N}\|v_{N}^{i}\|_{V}\leq(1/N)\sum_{i=1}^{N}\zeta(\xi^{i})$. We obtain $\limsup_{N\to\infty}\,\|v_{N}\|_{V}<\infty$. ∎ We impose a technical measurability condition on Clarke’s generalized directional derivative of $F_{\xi}$ in Theorem 4.3. The measurability of Clarke’s subdifferential as a function of the decision variables and parameters has been analyzed, for example, in [66, Lem. 4] and [83, Prop. 2.1]. Let $\mathcal{C}$ be the set of all C-stationary points of (4.1), that is, the set of all $x\in\mathrm{dom}(\psi)$ with $0\in\partial_{C}\big{[}\int_{\Xi}\,F_{\xi}(x)\,\mathrm{d}\mathbb{P}(\xi)\big{]}+\partial\psi(x)$. Let $\varepsilon\geq 0$ and for each $\omega\in\Omega$, let $\hat{\mathcal{C}}_{N}^{\varepsilon}(\omega)$ be the set of all points $x\in\mathrm{dom}(\psi)$ with $0\in\bigcup_{v\in\overline{\mathbb{B}}_{X}(x;\varepsilon)\cap\mathrm{dom}(\psi)}\partial_{C}\Big{[}\frac{1}{N}\sum_{i=1}^{N}F_{\xi^{i}(\omega)}(v)\Big{]}+\partial\psi(x);$ cf. [74, eqns. (3.2) and (4.2)]. Each element in $\hat{\mathcal{C}}_{N}^{\varepsilon}(\omega)$ is referred to as an approximate C-stationary point of (4.2). ###### Theorem 4.3. Let Assumption 4.1 hold true and let $(\varepsilon_{N})\subset[0,\infty)$ be a sequence with $\varepsilon_{N}\to 0$ as $N\to\infty$. Suppose that $\mathrm{dom}(\psi)\times X\times\Xi\ni(x,h,\xi)\mapsto F_{\xi}^{\circ}(x;h)$ is measurable, and let $\mathcal{C}$ and $\hat{\mathcal{C}}_{N}^{\varepsilon_{N}}(\omega)$ be nonempty for each $N\in\mathbb{N}$ and $\omega\in\Omega$. 1. (i) If $\Omega_{1}\subset\Omega$ with $\Omega_{1}\in\mathcal{F}$ and $P(\Omega_{1})=1$ and for each $\omega\in\Omega_{1}$, $(x_{N}(\omega))$ is a bounded sequence with $x_{N}(\omega)\in\hat{\mathcal{C}}_{N}^{\varepsilon_{N}}(\omega)$ for all sufficiently large $N\in\mathbb{N}$, then with probability one (w.p. $1$), $\mathrm{dist}({x_{N},\mathcal{C}})\to 0$ as $N\to\infty$. 2. (ii) If there exists a bounded set $\mathscr{C}\subset X$ such that w.p. $1$, $\hat{\mathcal{C}}_{N}^{\varepsilon_{N}}\subset\mathscr{C}$ for all sufficiently large $N\in\mathbb{N}$, then w.p. $1$, $\mathbb{D}({\hat{\mathcal{C}}_{N}^{\varepsilon_{N}},\mathcal{C}})\to 0$ as $N\to\infty$. ###### Proof. We establish the assertions via applications of Theorem 3.4. First, we observe that for each $\omega\in\Omega_{0}$, 4.1.7 implies 3.1.5 with $M_{k}(\cdot)$ replaced by $\hat{M}_{N}(\cdot,\omega)=(1/N)\sum_{i=1}^{N}M_{\xi^{i}(\omega)}(\cdot)$. Second, 4.1.4, 4.1.3, and 4.1.6, and the sum rule [18, Cor. 3 on p. 40] ensure for all $\omega\in\Omega$, $x\in X_{0}$, and every $N\in\mathbb{N}$, $\displaystyle\partial_{C}\Big{[}\frac{1}{N}\sum_{i=1}^{N}F_{\xi^{i}(\omega)}(x)\Big{]}=\frac{1}{N}\sum_{i=1}^{N}\partial_{C}F_{\xi^{i}(\omega)}(x)=\frac{1}{N}\sum_{i=1}^{N}BM_{\xi^{i}(\omega)}(x)=B\hat{M}_{N}(x,\omega).$ Hence for all $\omega\in\Omega$, 3.1.4 holds true with $M_{k}(\cdot)$ replaced by $\hat{M}_{N}(\cdot,\omega)$. Third, we show that for almost every $\omega\in\Omega$, 3.1.6 is satisfied with $f(\cdot)$ replaced by the expectation function $\int_{\Xi}F_{\xi}(\cdot)\,\mathrm{d}\mathbb{P}(\xi)$ and with $f_{k}(\cdot)$ replaced by the SAA objective function $(1/N)\sum_{i=1}^{N}F_{\xi^{i}(\omega)}(\cdot)$. For each $\xi\in\Xi$, $X_{0}\times X\ni(x,h)\mapsto F_{\xi}^{\circ}(x;h)$ is upper semicontinuous [18, Prop. 2.1.1]. Fix $\bar{x}\in\mathrm{dom}(\psi)$. Let $\mathcal{V}_{\bar{x}}$ be an open neighborhood and let $L_{\bar{x}}$ be a Lipschitz constant given by 4.1.3. Using [18, Prop. 2.1.1], we have $|F_{\xi}^{\circ}(x;h)|\leq L_{\bar{x}}(\xi)$ for each $(x,h,\xi)\in\mathcal{V}_{\bar{x}}\times X\times\Xi$. Moreover $(x,h,\xi)\mapsto F_{\xi}^{\circ}(x;h)$ is measurable by assumption. Combined with [3, Thm. 2.3], we find that w.p. $1$, $\mathrm{dom}(\psi)\times X\ni(x,h)\mapsto(1/N)\sum_{i=1}^{N}F_{\xi^{i}}^{\circ}(x;h)$ hypoconverges to $(x,h)\mapsto\int_{\Xi}F_{\xi}^{\circ}(x;h)\,\mathrm{d}\mathbb{P}(\xi)$.111We refer the reader to [3, p. 3] for a definition of epiconvergence of a sequence of functions $(g_{k})$ to some function $g$. A sequence of functions $(g_{k})$ hypoconverges to $g$ if (and only if) $(-g_{k})$ epiconverges to $-g$ [70, pp. 242–243]. Using 4.1.1, [18, Cor. 3 on p. 40 and Thm. 2.7.2] and Clarke regularity of $F_{\xi}$ for each $\xi\in\Xi$ (see 4.1.4), we have the identities $\int_{\Xi}F_{\xi}^{\circ}(x;h)\,\mathrm{d}\mathbb{P}(\xi)=(\int_{\Xi}F_{\xi}(\cdot)\,\mathrm{d}\mathbb{P}(\xi))^{\circ}(x;h)$ and $(1/N)\sum_{i=1}^{N}F_{\xi^{i}}^{\circ}(x;h)=\big{(}(1/N)\sum_{i=1}^{N}F_{\xi^{i}}(\cdot)\big{)}^{\circ}(x;h)$ for every $(x,h)\in X_{0}\times X$. Hence there exists $\Omega_{2}\subset\Omega$ with $\Omega_{2}\in\mathcal{F}$ and $P(\Omega_{2})=1$ such that for each $\omega\in\Omega_{2}$, all $h\in X$ and each $(x_{N})\subset\mathrm{dom}(\psi)$ with $x_{N}\to x$, we have $\limsup_{N\to\infty}\,\Big{(}\frac{1}{N}\sum_{i=1}^{N}F_{\xi^{i}(\omega)}(\cdot)\Big{)}^{\circ}(x_{N};h)\leq\Big{(}\int_{\Xi}F_{\xi}(\cdot)\,\mathrm{d}\mathbb{P}(\xi)\Big{)}^{\circ}(x;h).$ Now we establish parts (i) and (ii). (i) By assumption and construction, we have $\Omega_{0}\cap\Omega_{1}\cap\Omega_{2}\in\mathcal{F}$. Moreover this event happens w.p. $1$. Combining our derivations, Theorem 3.4 ensures for each $\omega\in\Omega_{0}\cap\Omega_{1}\cap\Omega_{2}$, we have $\mathrm{dist}({x_{N}(\omega),\mathcal{C}})\to 0$ as $N\to\infty$. (ii) Since w.p. $1$, $\hat{\mathcal{C}}_{N}^{\varepsilon_{N}}\subset\mathscr{C}$ for all sufficiently large $N\in\mathbb{N}$, there exists a set $\Omega_{3}\in\mathcal{F}$ with $P(\Omega_{3})=1$ such that for each $\omega\in\Omega_{3}$, there exists $n(\omega)\in\mathbb{N}$ with $\hat{\mathcal{C}}_{N}^{\varepsilon_{N}}(\omega)\subset\mathscr{C}$ for all $N\geq n(\omega)$. In particular $\Omega_{0}\cap\Omega_{2}\cap\Omega_{3}\in\mathcal{F}$ and $P(\Omega_{0}\cap\Omega_{2}\cap\Omega_{3})=1$. Now Theorem 3.4 ensures for each $\omega\in\Omega_{0}\cap\Omega_{2}\cap\Omega_{3}$, $\mathbb{D}({\hat{\mathcal{C}}_{N}^{\varepsilon_{N}}(\omega),\mathcal{C}})\to 0$ as $N\to\infty$. ∎ ## 5 Approximations via weakly converging probability measures Motivated by recent contributions on epiconvergent discretizations of stochastic programs [19, 22, 34, 67], we establish consistency of approximations to the stochastic program (4.1) which result from approximating $\mathbb{P}$ by a sequence of weakly convergent probability measures $(\mathbb{P}_{N})$. Our main interest in considering weakly convergent probability measures stems from those considered in [67, sect. 2] and [46, sect. 3]. Throughout the section, we assume $\Xi$ be a complete, separable metric space and $\mathbb{P}$ be a probability measure on $\Xi$. We refer the reader to [42, p. 65] for the definition of weak convergence of probability measures. For a sequence $(\mathbb{P}_{N})$ of probability measures defined on $\Xi$, we approximate (4.1) via the optimization problems $\displaystyle\min_{x\in\mathrm{dom}(\psi)}\,\int_{\Xi}\,F_{\xi}(x)\,\mathrm{d}\mathbb{P}_{N}(\xi)+\psi(x).$ (5.1) Our consistency analysis is built on the conditions in assumption 5.1 and those needed to apply [56, Cor. 3.4]. ###### Assumption 5.1. In addition to 4.1.1, 4.1.2, 4.1.6, and 4.1.5, we consider the following conditions. 1. 1. For each $N\in\mathbb{N}$, $\mathbb{P}_{N}$ is a probability measure on $\Xi$, and $(\mathbb{P}_{N})$ weakly converges to $\mathbb{P}$ as $N\to\infty$. 2. 2. For each $\xi\in\Xi$, $F_{\xi}\colon X_{0}\to$ is continuously differentiable. For every $x\in X_{0}$, $\Xi\ni\xi\mapsto F_{\xi}(x)\in$ is $\mathbb{P}_{N}$-integrable. For all $h\in X$, $X_{0}\times\Xi\ni(x,\xi)\mapsto\langle\mathrm{D}F_{\xi}(x),h\rangle_{{X}^{*}\\!,X}\in$ is upper semicontinuous. 3. 3. 4.1.3 holds true with $L_{\bar{x}}$ being continuous, $\mathbb{P}_{N}$-integrable for each $N\in\mathbb{N}$, and satisfying $\limsup_{N\to\infty}\,\int_{\Xi}L_{\bar{x}}(\xi)\,\mathrm{d}\mathbb{P}_{N}(\xi)\leq\int_{\Xi}L_{\bar{x}}(\xi)\,\mathrm{d}\mathbb{P}(\xi)$. 4. 4. For each $\bar{x}\in\mathrm{dom}(\psi)$, each sequence $(x_{N})\subset\mathrm{dom}(\psi)$ with $x_{N}\rightharpoonup\bar{x}$, and each sequence $(v_{N})$ with $v_{N}\in\int_{\Xi}M_{\xi}(x_{N})\,\mathrm{d}\mathbb{P}_{N}(\xi)$ for all $N\in\mathbb{N}$, $\limsup_{N\to\infty}\,\|v_{N}\|_{V}<\infty$. While 5.1.2 imposes continuity and differentiablity, these conditions are satisfied for certain problem classes [63]. However the continuity conditions may be relaxed in view of [2, Thms. 3.6 and 5.2]. Under 5.1.2, the continuously differentiable function $F_{\xi}\colon X_{0}\to$ is Clarke regular for each $\xi\in\Xi$ [18, Prop. 2.3.6] and $\partial_{C}F_{\xi}(x)=\\{\,\mathrm{D}F_{\xi}(x)\,\\}$ [18, Cor. on p. 32 and Prop. 2.2.4]. Using Assumption 5.1 and [24, Lem. C.3], we find that $X_{0}\ni x\mapsto\int_{\Xi}\,F_{\xi}(x)\,\mathrm{d}\mathbb{P}(\xi)$ and $X_{0}\ni x\mapsto\int_{\Xi}\,F_{\xi}(x)\,\mathrm{d}\mathbb{P}_{N}(\xi)$ are Fréchet differentiable and we can interchange derivatives and integrals. Let $\mathfrak{C}$ be the set of all $x\in\mathrm{dom}(\psi)$ with $0\in\mathrm{D}\big{[}\int_{\Xi}\,F_{\xi}(x)\,\mathrm{d}\mathbb{P}(\xi)\big{]}+\partial\psi(x)$. Let $\varepsilon\geq 0$ and for each $\omega\in\Omega$, let $\mathfrak{C}_{N}^{\varepsilon}$ be the set of all points $x\in\mathrm{dom}(\psi)$ with $0\in\bigcup_{v\in\overline{\mathbb{B}}_{X}(x;\varepsilon)\cap\mathrm{dom}(\psi)}\mathrm{D}\big{[}\int_{\Xi}\,F_{\xi}(v)\,\mathrm{d}\mathbb{P}_{N}(\xi)\big{]}+\partial\psi(x)$. ###### Theorem 5.2. Let Assumption 5.1 hold true and let $(\varepsilon_{N})\subset[0,\infty)$ be a sequence with $\varepsilon_{N}\to 0$ as $N\to\infty$. Let $\mathfrak{C}$ and $\mathfrak{C}_{N}^{\varepsilon_{N}}$ be nonempty for each $N\in\mathbb{N}$. 1. (i) If $(x_{N})$ is bounded with $x_{N}\in\mathfrak{C}_{N}^{\varepsilon_{N}}$ for each $N\in\mathbb{N}$, then $\mathrm{dist}({x_{N},\mathcal{C}})\to 0$ as $N\to\infty$. 2. (ii) If there exists a bounded set $\mathscr{C}\subset X$ such that $\mathfrak{C}_{N}^{\varepsilon_{N}}\subset\mathscr{C}$ for all sufficiently large $N\in\mathbb{N}$, then $\mathbb{D}({\mathfrak{C}_{N}^{\varepsilon_{N}},\mathfrak{C}})\to 0$ as $N\to\infty$. ###### Proof. Using [56, Cor. 3.4], we show that 3.1.6 holds true with $f(\cdot)$ replaced by $\int_{\Xi}\,F_{\xi}(\cdot)\,\mathrm{d}\mathbb{P}(\xi)$, and $f_{k}(\cdot)$ replaced by $\int_{\Xi}\,F_{\xi}(\cdot)\,\mathrm{d}\mathbb{P}_{N}(\xi)$. Fix $h\in X$, let $\bar{x}\in\mathrm{dom}(\psi)$ and let $(x_{N})\subset\mathrm{dom}(\psi)$ with $x_{N}\to\bar{x}$ be arbitrary. According to 5.1.3, there exists $\varepsilon=\varepsilon(\bar{x})>0$ such $F_{\xi}$ is Lipschitz continuous on $\mathbb{B}_{X}(\bar{x},\varepsilon)$ with a Lipschitz constant $L_{\bar{x}}(\xi)$ for each $\xi\in\Xi$, $L_{\bar{x}}$ is continuous, $\mathbb{P}$\- and $\mathbb{P}_{N}$-integrable, and $\limsup_{N\to\infty}\,\int_{\Xi}L_{\bar{x}}(\xi)\,\mathrm{d}\mathbb{P}_{N}(\xi)\leq\int_{\Xi}L_{\bar{x}}(\xi)\,\mathrm{d}\mathbb{P}(\xi)$. The mean value theorem yields $|\langle\mathrm{D}F_{\xi}(x),h\rangle_{{X}^{*}\\!,X}|\leq L_{\bar{x}}(\xi)\|h\|_{X}$ for all $x\in\mathbb{B}_{X}(\bar{x},\varepsilon)$, $h\in X$, and $\xi\in\Xi$. Combined with 5.1.1 and 5.1.2, [56, Cor. 3.4] yields $\displaystyle\limsup_{N\to\infty}\,\int_{\Xi}\langle\mathrm{D}F_{\xi}(x_{N}),h\rangle_{{X}^{*}\\!,X}\,\mathrm{d}\mathbb{P}_{N}(\xi)\leq\int_{\Xi}\langle\mathrm{D}F_{\xi}(\bar{x}),h\rangle_{{X}^{*}\\!,X}\,\mathrm{d}\mathbb{P}(\xi).$ Using the chain rule [24, Lem. C.3], we can interchange derivatives and integrals in the above equation. Hence 3.1.6 holds true. Both assertions are now implied by Theorem 3.4. ∎ ## 6 Application to risk-averse semilinear PDE-constrained optimization We show that our main result, Theorem 4.3, applies to risk-averse PDE- constrained optimization using the average value-at-risk. For $\beta\in(0,1)$, and a $\mathbb{P}$-integrable random variable $Z\colon\Xi\to$, the average value-at-risk $\mathrm{AVaR}_{\beta}$ evaluated at $Z$ is defined by $\displaystyle\mathrm{AVaR}_{\beta}[Z]=\inf_{t\in}\,\big{\\{}\,t+\tfrac{1}{1-\beta}\mathbb{E}_{\mathbb{P}}[(Z-t)_{+}]\,\big{\\}},$ (6.1) where $(s)_{+}=\max\\{0,s\\}$ for $s\in$ [73, eq. (6.23)]. We consider the risk-averse semilinear PDE-constrained problem $\displaystyle\min_{u\in U_{\text{ad}}}\,(1/2)\mathrm{AVaR}_{\beta}[\|\iota_{0}S(u,\cdot)-y_{d}\|_{L^{2}(D)}^{2}]+\wp(\|u\|_{L^{2}(D)}),$ (6.2) where $\wp\colon[0,\infty)\to[0,\infty)$, $y_{d}\in L^{2}(D)$, $U_{\text{ad}}\subset L^{2}(D)$, and for each $(u,\xi)\in L^{2}(D)\times\Xi$, $y=S(u,\xi)\in H_{0}^{1}(D)$ solves the semilinear PDE $\displaystyle(\kappa(\xi)\nabla y,\nabla v)_{L^{2}(D)^{d}}+(y^{3},v)_{L^{2}(D)}=(u,v)_{L^{2}(D)}+(b(\xi),v)_{L^{2}(D)}\>\>\text{for all}\>\>v\in H_{0}^{1}(D).$ We impose mild conditions on the semilinear PDE, the feasible set $U_{\text{ad}}$, and $\wp$. ###### Assumption 6.1 (semilinear control problem). 1. 1. $D\subset^{d}$ is a bounded, convex, polygonal/polyhedral domain with $d\in\\{2,3\\}$. 2. 2. $\kappa:\Xi\to L^{\infty}(D)$ is strongly measurable and there exists $0<\kappa_{\min}\leq\kappa_{\max}<\infty$ such that for all $\xi\in\Xi$, $\kappa_{\min}\leq\kappa(\xi)(x)\leq\kappa_{\max}$ for a.e. $x\in D$. 3. 3. $b:\Xi\to L^{2}(D)$ is measurable and there exists $b_{\max}\geq 0$ with $\|b(\xi)\|_{L^{2}(D)}\leq b_{\max}$ for all $\xi\in\Xi$. 4. 4. $U_{\text{ad}}\subset L^{2}(D)$ is closed, convex, nonempty, and there exists $r_{\text{ad}}\in(0,\infty)$ such that $\|u\|_{L^{2}(D)}\leq r_{\text{ad}}$ for all $u\in U_{\text{ad}}$. 5. 5. $\wp\colon[0,\infty)\to[0,\infty)$ is convex and strictly increasing. If not stated otherwise, let Assumption 6.1 hold throughout the section. Let us define $G\colon\times\to$ and $\widehat{J}\colon L^{2}(D)\times\Xi\to[0,\infty)$ by $\displaystyle G(s,t)=t+\tfrac{1}{1-\beta}(s-t)_{+}\quad\text{and}\quad\widehat{J}(u,\xi)=(1/2)\|\iota_{0}S(u,\xi)-y_{d}\|_{L^{2}(D)}^{2}.$ Moreover, we define $L^{2}(D)\times\times\Xi\ni(u,t,\xi)\mapsto F_{\xi}(u,t)\in[0,\infty)$ by $\displaystyle F_{\xi}(u,t)=G(\widehat{J}(u,\xi),t).$ To study consistency of SAA C-stationary points, we reformulate (6.2) equivalently as $\displaystyle\min_{(u,t)\in U_{\text{ad}}\times}\,\int_{\Xi}G(\widehat{J}(u,\xi),t)\,\mathrm{d}\mathbb{P}(\xi)+\wp(\|u\|_{L^{2}(D)}).$ (6.3) Let $\Xi$, let $\mathbb{P}$, and let $\xi^{1}$, $\xi^{2},\ldots$ be as in section 4. The SAA problem of (6.3) is given by $\displaystyle\min_{(u,t)\in U_{\text{ad}}\times}\,\frac{1}{N}\sum_{i=1}^{N}G(\widehat{J}(u,\xi^{i}),t)+\wp(\|u\|_{L^{2}(D)}).$ (6.4) For the remainder of the section, we verify Assumption 4.1. For the spaces $L^{2}(D)\times$, $V=H_{0}^{1}(D)\times$, and the open, convex, and bounded set $U_{0}=U_{\text{ad}}+\mathbb{B}_{L^{2}(D)}(0;1)$, 4.1.1 holds true. Let $\psi$ be the sum of the indicator function of $U_{\text{ad}}$ and the regularizer $\wp\circ\|\cdot\|_{L^{2}(D)}$. Since $\wp$ is convex and strictly increasing, and $L^{2}(D)$ is a Hilbert space, $\psi$ has the Kadec property; see [62, Lem. 1] and [8, Thm. 6.5]. Hence 4.1.2 holds true. For each $(u,\xi)\in L^{2}(D)\times\Xi$, the PDE solution $S(u,\xi)$ is well- defined [85, Thm. 26.A], and we have for all $(u,\xi)\in L^{2}(D)\times\Xi$, the stability estimate $\displaystyle\|S(u,\xi)\|_{H_{0}^{1}(D)}\leq(C_{D}/\kappa_{\min})\|u\|_{L^{2}(D)}+C_{D}(b_{\max}/\kappa_{\min}).$ (6.5) Defining $\displaystyle\mathfrak{c}_{S}=(C_{D}/\kappa_{\min})(r_{\text{ad}}+1)+C_{D}(b_{\max}/\kappa_{\min}),$ we obtain for all $(u,\xi)\in U_{0}\times\Xi$, the stability estimate $\displaystyle\widehat{J}(u,\xi)\leq\|\iota_{0}S(u,\xi)\|_{L^{2}(D)}^{2}+\|y_{d}\|_{L^{2}(D)}^{2}\leq C_{D}^{2}\mathfrak{c}_{S}^{2}+\|y_{d}\|_{L^{2}(D)}^{2}.$ (6.6) The adjoint approach [37, sect. 1.6.2] ensures that for each $\xi\in\Xi$, $\widehat{J}(\cdot,\xi)$ is continuously differentiable and $\displaystyle\nabla_{u}\widehat{J}(\cdot,\xi)=-\iota_{0}z(u,\xi)\quad\text{for all}\quad(u,\xi)\in L^{2}(D)\times\Xi,$ where for each $(u,\xi)\in L^{2}(D)\times\Xi$, $z=z(u,\xi)\in H_{0}^{1}(D)$ solves the adjoint equation $\displaystyle(\kappa(\xi)\nabla z,\nabla v)_{L^{2}(D)^{d}}+3(S(u,\xi)^{2}z,v)_{L^{2}(D)}=-(\iota_{0}S(u,\xi)-y_{d},v)_{L^{2}(D)}$ for all $v\in H_{0}^{1}(D)$. For all $(u,\xi)\in L^{2}(D)\times\Xi$, we have $\displaystyle\|z(u,\xi)\|_{H_{0}^{1}(D)}\leq(C_{D}/\kappa_{\min})\|\iota_{0}S(u,\xi)\|_{L^{2}(D)}+(C_{D}/\kappa_{\min})\|y_{d}\|_{L^{2}(D)}.$ Combined with the stability estimate (6.5), we have for each $(u,\xi)\in U_{0}\times\Xi$, $\displaystyle\|z(u,\xi)\|_{H_{0}^{1}(D)}\leq(C_{D}^{2}/\kappa_{\min})\mathfrak{c}_{S}+(C_{D}/\kappa_{\min})\|y_{d}\|_{L^{2}(D)}.$ (6.7) Let $\mathfrak{c}_{z}$ be the right-hand side in (6.7). Since $G$ is Lipschitz continuous and the bound $\|\nabla_{u}\widehat{J}(\cdot,\xi)\|_{L^{2}(D)}\leq C_{D}\mathfrak{c}_{z}$ is valid for all $(u,\xi)\in U_{0}\times\Xi$, 4.1.3 is satisfied. Fix $\xi\in\Xi$. Since the mapping $(u,t)\mapsto(\widehat{J}(u,\xi),t)\in^{2}$ is continuously differentiable and $g$ is finite-valued and convex, $F_{\xi}$ is Clarke regular on $L^{2}(D)\times$ and $\displaystyle\partial_{C}F_{\xi}(u,t)=(\nabla_{u}\widehat{J}(u,\xi),-1)^{*}\partial G(\widehat{J}(u,\xi),t)=\begin{bmatrix}\frac{1}{1-\beta}\partial(\widehat{J}(u,\xi)-t)_{+}\nabla_{u}\widehat{J}(u,\xi)\\\ 1-\frac{1}{1-\beta}\partial(\widehat{J}(u,\xi)-t)_{+}\end{bmatrix};$ see [18, Prop. 2.3.6 and Thm. 2.3.10]. The Clarke regularity ensures 4.1.4. Using [4, Thm. 8.2.9], we can show that $S(u,\cdot)$ is measurable for each $u\in L^{2}(D)$. Combined with (6.6), we find that 4.1.5 holds true. We define $B\colon H_{0}^{1}(D)\times\to L^{2}(D)\times$ and $M_{\xi}\colon L^{2}(D)\times\rightrightarrows H_{0}^{1}(D)\times$ by $\displaystyle B(v,s)=(-\iota_{0}v,s)\quad\text{and}\quad M_{\xi}(u,t)=\begin{bmatrix}\frac{1}{1-\beta}\partial(\widehat{J}(u,\xi)-t)_{+}z(u,\xi)\\\ 1-\frac{1}{1-\beta}\partial(\widehat{J}(u,\xi)-t)_{+}\end{bmatrix}.$ (6.8) We obtain $\displaystyle\partial_{C}F_{\xi}(u,t)=BM_{\xi}(u,t)\quad\text{for all}\quad(u,t,\xi)\in L^{2}(D)\times\times\Xi.$ Combined with the fact that $\iota_{0}$ is linear and compact [37, Thm. 1.14], we conclude that 4.1.6 is fulfilled. Next we verify 4.1.7 using Lemma 4.2. For each $s\in$, $\partial(s)_{+}\subset[0,1]$. Fix $(u,t,\xi)\in U_{\text{ad}}\times\times\Xi$. For all $m=(m_{1},m_{2})\in M_{\xi}(u,t)$, (6.8) yields $\displaystyle\|m\|_{H_{0}^{1}(D)\times}=\|m_{1}\|_{H_{0}^{1}(D)}+|m_{2}|\leq\tfrac{1}{1-\beta}\mathfrak{c}_{z}+1+\tfrac{1}{1-\beta}.$ Hence Lemma 4.2 implies 4.1.7. To summarize, we have shown that Assumption 4.1 holds true. In order to demonstrate the convergence of SAA C-stationary points of (6.4) towards those of (6.3) via Theorem 4.3, we need to show that the SAA C-stationary points are contained in a bounded, deterministic set. Let $(u_{N},t_{N})\in U_{\text{ad}}\times$ be a C-stationary point of (6.3). We have $u_{N}\in U_{\text{ad}}$, which is by assumption a bounded set. Next we show that $\displaystyle t_{N}\in[0,C_{D}^{2}\mathfrak{c}_{S}^{2}+\|y_{d}\|_{L^{2}(D)}^{2}].$ Since $(u_{N},t_{N})$ is a C-stationary point of (6.3), the second formula in (6.8) implies that $1\in\tfrac{1}{N(1-\beta)}\sum_{i=1}^{N}\partial(\widehat{J}(u_{N},\xi^{i})-t_{N})_{+}$. Combined with $\beta\in(0,1)$ and $\widehat{J}\geq 0$, we have $t_{N}\geq 0$. Let $\widehat{\mathrm{AVaR}}_{\beta,N}$ be the empirical estimate of $\mathrm{AVaR}_{\beta}$ with $\mathbb{P}$ replaced by the empirical distribution of the sample $\xi^{1},\ldots,\xi^{N}$ in (6.1). The minimization problem in (6.1) with $\mathbb{P}$ replaced by the empirical distribution is convex. Hence $t_{N}$ solves (6.1) with $\mathbb{P}$ replaced by the empirical distribution. As with $\mathrm{AVaR}_{\beta}$ [73, p. 231], the empirical average value-at-risk $\widehat{\mathrm{AVaR}}_{\beta,N}$ is a coherent risk measure. Since $(s)_{+}\geq 0$ for all $s\in$, $\beta\in(0,1)$, and coherent risk measures are monotone, translation equivariant, and positively homogeneous [73, Def. 6.4], the upper bound on $\widehat{J}$ in (6.6) yields $\displaystyle t_{N}\leq\widehat{\mathrm{AVaR}}_{\beta,N}[\widehat{J}(u_{N},\xi)]\leq C_{D}^{2}\mathfrak{c}_{S}^{2}+\|y_{d}\|_{L^{2}(D)}^{2}.$ ## 7 Application to risk-neutral bilinear PDE-constrained optimization Motivated by the deterministic bilinear control problems studied in [14, 27, 41, 52], we consider risk-neutral optimization problems governed by a bilinear elliptic PDEs with random inputs: $\displaystyle\min_{u\in U_{\text{ad}}}\,\int_{\Xi}J(S(u,\xi))\,\mathrm{d}\mathbb{P}(\xi)+(\alpha/2)\|u\|_{L^{2}(D)}^{2},$ (7.1) where $\alpha>0$, and for each $(u,\xi)\in U_{\text{ad}}\times\Xi$, $y=S(u,\xi)\in H_{0}^{1}(D)$ is the solution to $\displaystyle(\kappa(\xi)\nabla y,\nabla v)_{L^{2}(D)^{d}}+(g(\xi)uy,v)_{L^{2}(D)}=(b(\xi),v)_{L^{2}(D)}\quad\text{for all}\quad v\in H_{0}^{1}(D).$ (7.2) The following assumptions ensure existence and regularity of the solutions to (7.2), and impose conditions on the function $J$ and the set $U_{\text{ad}}$. These conditions allow us to Assumption 4.1 and apply Theorem 4.3 to study the consistency of SAA C-stationary points corresponding to (7.1). Conditions beyond those formulated in Assumption 7.1 may be required to allow for applications of Theorem 5.2. ###### Assumption 7.1 (bilinear control problem). 1. 1. $D\subset^{d}$ is a bounded, convex, polygonal/polyhedral domain with $d\in\\{2,3\\}$. 2. 2. $\kappa:\Xi\to C^{1}(\bar{D})$ is measurable and there exists $\kappa_{\min}>0$ such that $\kappa_{\min}\leq\kappa(\xi)(x)$ for all $(\xi,x)\in\Xi\times\bar{D}$. Moreover $\mathbb{E}_{\mathbb{P}}[\|\kappa\|_{C^{1}(\bar{D})}^{p}]<\infty$ for all $p\in[1,\infty)$. 3. 3. $b:\Xi\to L^{2}(D)$ and $g:\Xi\to C^{0,1}(\bar{D})$ are measurable and there exist $b_{\max}$, $g_{\max}>0$ with $\|b(\xi)\|_{L^{2}(D)}\leq b_{\max}$ and $\|g(\xi)\|_{C^{0,1}(\bar{D})}\leq g_{\max}$ for all $\xi\in\Xi$. Moreover, $g(\xi)(x)\geq 0$ for all $(\xi,x)\in\Xi\times\bar{D}$. 4. 4. $J:H_{0}^{1}(D)\to[0,\infty)$ is convex, continuously differentiable and its derivative is Lipschitz continuous with Lipschitz constant $\ell>0$. For some nondecreasing function $\varrho:[0,\infty)\to[0,\infty)$, it holds that $\displaystyle J(y)\leq\varrho(\|y\|_{H_{0}^{1}(D)})\quad\text{for all}\quad y\in H_{0}^{1}(D).$ 5. 5. $U_{\text{ad}}=\\{\,u\in L^{2}(D):\,0\leq u(x)\leq\mathfrak{u}(x)\;\text{a.e.}\;x\in D\,\\}$ with $\mathfrak{u}\in L^{\infty}(D)$ and $\mathfrak{u}(x)\geq 0$ a.e. $x\in D$. If not stated otherwise, let Assumption 7.1 hold throughout the section. We introduce notation used throughout the section. Let $C_{H_{0}^{1};L^{4}}\in(0,\infty)$ be the embedding constant of the compact embedding $H_{0}^{1}(D)\xhookrightarrow{}L^{4}(D)$ and let $C_{H^{2};L^{\infty}}>0$ be that of the continuous embedding $H^{2}(D)\xhookrightarrow{}L^{\infty}(D)$ [37, Thm. 1.14]. We define $C_{\mathfrak{u}}=\|\mathfrak{u}\|_{L^{\infty}(D)}$ and $|D|=\int_{D}1\,\mathrm{d}x$. Let $\psi$ be the sum of the indicator function of $U_{\text{ad}}$ and $(\alpha/2)\|u\|_{L^{2}(D)}^{2}$. The function $\psi$ has the Kadec property [8, Thm. 6.5], and 4.1.2 holds true. Next we provide an example function satisfying 7.1.4. Let $y_{d}\in L^{2}(D)$. We define $J:H_{0}^{1}(D)\to[0,\infty)$ by $J(y)=(1/2)\|\iota_{0}y-y_{d}\|_{L^{2}(D)}^{2}$. The function $J$ satisfies 7.1.4 with $\ell=C_{D}^{2}$ and $\varrho(t)=C_{D}^{2}t^{2}+\|y_{d}\|_{L^{2}(D)}^{2}$. We use two technical facts for our analysis. Lemma 7.2 establishes a bound on the derivatives of the objective functions fulfilling 7.1.4. ###### Lemma 7.2. If $D\subset^{d}$ is a bounded domain and 7.1.4 holds, then $\|\mathrm{D}J(y)\|_{H^{-1}(D)}^{2}\leq 2\ell J(y)$ for all $y\in H_{0}^{1}(D)$. ###### Proof. Fix $\tilde{y}$, $y\in H_{0}^{1}(D)$. Let $\nabla J(y)\in H_{0}^{1}(D)$ be the Riesz representation of $\mathrm{D}J(y)$. Using 7.1.4 and [86, Cor. 3.5.7 and Rem. 3.5.2], we have $0\leq J(\tilde{y})\leq J(y)+(\nabla J(y),\tilde{y}-y)_{H_{0}^{1}(D)}+(\ell/2)\|\tilde{y}-y\|_{H_{0}^{1}(D)}^{2}$. Minimizing over $\tilde{y}\in H_{0}^{1}(D)$ yields $0\leq J(y)-(1/(2\ell))\|\nabla J(y)\|_{H_{0}^{1}(D)}^{2}$. ∎ Using 7.1.1 and theorems on the multiplication of Sobolev functions (see [6, Thm. 7.4] and [29, Thm. 1.4.1.1]), we can show that the trilinear (pointwise multiplication) operator $\displaystyle C^{0,1}(\bar{D})\times H^{1}(D)\times H^{2}(D)\ni(g,v,w)\mapsto gvw\in H^{1}(D)$ (7.3) is continuous. Hence it is bounded [55, p. 68]. Therefore, we obtain the existence of a constant $C_{\mathrm{tri}}\in(0,\infty)$ such that for all $(g,v,w)\in C^{0,1}(\bar{D})\times H^{1}(D)\times H^{2}(D)$, $\displaystyle\|gvw\|_{H^{1}(D)}\leq C_{\mathrm{tri}}\|g\|_{C^{0,1}(\bar{D})}\|v\|_{H^{1}(D)}\|w\|_{H^{2}(D)}.$ (7.4) ### 7.1 Existence and uniqueness of PDE solutions We show that the PDE (7.2) has a unique solution $S(u,\xi)\in H_{0}^{1}(D)$ for each $\xi\in\Xi$ and $u$ in an open neighborhood of the set $U_{\text{ad}}$ in $L^{2}(D)$. This allows us to verify 4.1.3, 4.1.4, 4.1.5, and 4.1.1. The existence results established in [14, sect. 3.1] are not applicable to our setting, as we consider $L^{2}(D)$ instead of $L^{\infty}(D)$ the control space. An approach different from ours to constructing such a neighborhood is developed in [82, p. 158]. We define $\displaystyle U_{0}=\bigcup_{u_{0}\in U_{\text{ad}}}\,\mathbb{B}_{L^{2}(D)}(u_{0};\delta),\quad\text{where}\quad\delta=\tfrac{\kappa_{\min}}{2g_{\max}C_{H_{0}^{1};L^{4}}^{2}}>0.$ (7.5) We have the identity $U_{0}=U_{\text{ad}}+\mathbb{B}_{L^{2}(D)}(u_{0};\delta)$. Fix $(u,\xi)\in U_{0}\times\Xi$ and $y\in H_{0}^{1}(D)$. Since $u\in U_{0}$, there exists $u_{0}\in U_{\text{ad}}$ with $u\in\mathbb{B}_{L^{2}(D)}(u_{0};\delta)$. Combining Hölder’s inequality and the continuity of $H_{0}^{1}(D)\xhookrightarrow{}L^{4}(D)$ with the definition of $\delta$ provided in (7.5) and $\|u-u_{0}\|_{L^{2}(D)}\leq\delta$, we obtain $\displaystyle|(g(\xi)(u-u_{0})y,y)_{L^{2}(D)}|\leq C_{H_{0}^{1};L^{4}}^{2}g_{\max}\|u-u_{0}\|_{L^{2}(D)}\|y\|_{H_{0}^{1}(D)}^{2}\leq(\kappa_{\min}/2)\|y\|_{H_{0}^{1}(D)}^{2}.$ Using $u_{0}(x)\geq 0$ and $g(\xi)(x)\geq 0$ for a.e. $x\in D$, $(g(\xi)u_{0}y,y)_{L^{2}(D)}\geq 0$. It follows that $\displaystyle\begin{aligned} &(\kappa(\xi)\nabla y,\nabla y)_{L^{2}(D)^{d}}+(g(\xi)uy,y)_{L^{2}(D)}\\\ &\quad=(\kappa(\xi)\nabla y,\nabla y)_{L^{2}(D)^{d}}+(g(\xi)(u-u_{0})y,y)_{L^{2}(D)}+(g(\xi)u_{0}y,y)_{L^{2}(D)}\\\ &\quad\geq\kappa_{\min}\|y\|_{H_{0}^{1}(D)}^{2}-(\kappa_{\min}/2)\|y\|_{H_{0}^{1}(D)}^{2}=(\kappa_{\min}/2)\|y\|_{H_{0}^{1}(D)}^{2}.\end{aligned}$ (7.6) The right-hand side in (7.2) defines a continuous bilinear form. Hence the Lax–Milgram lemma ensures that the state equation (7.2) has a unique solution $S(u,\xi)\in H_{0}^{1}(D)$ and yields the stability estimate $\displaystyle\|S(u,\xi)\|_{H_{0}^{1}(D)}\leq(2/\kappa_{\min})C_{D}\|b(\xi)\|_{L^{2}(D)}\quad\text{for all}\quad(u,\xi)\in U_{0}\times\Xi.$ (7.7) We define the parameterized operator $E:H_{0}^{1}(D)\times L^{2}(D)\times\Xi\to H^{-1}(D)$ by $\displaystyle\langle E(y,u,\xi),v\rangle_{H^{-1}(D),H_{0}^{1}(D)}=(\kappa(\xi)\nabla y,\nabla v)_{L^{2}(D)^{d}}+(g(\xi)uy,v)_{L^{2}(D)}-(b(\xi),v)_{L^{2}(D)}.$ The mapping $E$ is well-defined and $E(\cdot,\cdot,\xi)$ is infinitely many times continuously differentiable. For each $(y,u,\xi)\in H_{0}^{1}(D)\times L^{2}(D)\times\Xi$ and $h$, $v\in H_{0}^{1}(D)$, we have $\displaystyle\begin{aligned} \langle E_{y}(y,u,\xi)h,v\rangle_{H^{-1}(D),H_{0}^{1}(D)}&=(\kappa(\xi)\nabla h,\nabla v)_{L^{2}(D)^{d}}+(g(\xi)uh,v)_{L^{2}(D)}.\end{aligned}$ Together with (7.6), we obtain for all $(u,\xi)\in U_{0}\times\Xi$, $\displaystyle\langle E_{y}(S(u,\xi),u,\xi)h,h\rangle_{H^{-1}(D),H_{0}^{1}(D)}\geq(\kappa_{\min}/2)\|h\|_{H_{0}^{1}(D)}^{2}\quad\text{for all}\quad h\in H_{0}^{1}(D).$ The Lax–Milgram lemma ensures that $E_{y}(S(u,\xi),u,\xi)$ has a bounded inverse and $\displaystyle\|E_{y}(S(u,\xi),u,\xi)^{-1}\|_{\mathscr{L}(H_{0}^{1}(D),H^{-1}(D))}\leq(2/\kappa_{\min})\quad\text{for all}\quad(u,\xi)\in U_{0}\times\Xi.$ (7.8) For each $y$, $v\in H_{0}^{1}(D)$ and $u\in L^{2}(D)$, we can show that $\langle E(y,u,\cdot),v\rangle_{H^{-1}(D),H_{0}^{1}(D)}$ is measurable, as 7.1.1, 7.1.2, and 7.1.3 hold. Since $H^{-1}(D)$ is separable, $E(y,u,\cdot)$ is measurable [39, Thm. 1.1.6]. Hence $S(u,\cdot)$ is measurable [4, Thm. 8.2.9]. Combined with 7.1.4, we find that $J(S(u,\cdot))$ is a random variable for each $u\in U_{0}$. ### 7.2 PDE regularity We show that the solution $S(u,\xi)$ to (7.2) is contained in $H^{2}(D)$ and derive a stability estimate for each $(u,\xi)\in U_{\text{ad}}\times\Xi$. The computations performed here form the basis of verifying 4.1.6. Fix $(u,\xi)\in U_{\text{ad}}\times\Xi$. Hence $u\in L^{\infty}(D)$. To establish an $H^{2}(D)$-stability estimate, we apply [77, Thm. 3.1] to the solution $\widetilde{y}=\widetilde{y}(u,\xi)\in H_{0}^{1}(D)$ to $\displaystyle(\kappa(\xi)\nabla\widetilde{y},\nabla v)_{L^{2}(D)^{d}}=(b(\xi),v)_{L^{2}(D)}-(g(\xi)uS(u,\xi),v)_{L^{2}(D)}\quad\text{for all}\quad v\in H_{0}^{1}(D).$ Since this equation has a unique solution $\widetilde{y}(u,\xi)$ and $S(u,\xi)$ is a solution, we have $\widetilde{y}(u,\xi)=S(u,\xi)$. Defining the random variable $\displaystyle C_{\kappa}(\xi)=\frac{\|\kappa(\xi)\|_{C^{0}(\bar{D})}\|\kappa(\xi)\|_{C^{1}(\bar{D})}^{2}}{\kappa_{\min}^{4}},$ (7.9) and applying [77, Thm. 2.1], we find that $S(u,\xi)\in H^{2}(D)$ and $\displaystyle\|S(u,\xi)\|_{H^{2}(D)}\leq C_{H^{2}}C_{\kappa}(\xi)\Big{(}\|g(\xi)uS(u,\xi)\|_{L^{2}(D)}+\|b(\xi)\|_{L^{2}(D)}\Big{)},$ where $C_{H^{2}}=C_{H^{2}}(D)>0$ is a deterministic constant. Combined with Hölder’s and Friedrichsínequalities, we obtain for all $(u,\xi)\in U_{\text{ad}}\times\Xi$, $\displaystyle\|S(u,\xi)\|_{H^{2}(D)}\leq C_{H^{2}}C_{\kappa}(\xi)\big{(}C_{D}C_{\mathfrak{u}}g_{\max}\|S(u,\xi)\|_{H_{0}^{1}(D)}+\|b(\xi)\|_{L^{2}(D)}\big{)}.$ (7.10) ### 7.3 Gradient regularity In this section, we show that 4.1.3, 4.1.4, 4.1.6, and 4.1.7 hold true. Using the construction of $U_{0}$, we can define the objective function $U_{0}\times\Xi\ni(u,\xi)\mapsto F_{\xi}(u)\in[0,\infty)$ by $\displaystyle F_{\xi}(u)=J(S(u,\xi)).$ Assumption 7.1 ensures that $F_{\xi}$ is continuously differentiable for every $\xi\in\Xi$. Hence 4.1.4 holds true [18, Prop. 2.3.6]. The adjoint approach [37, sect. 1.6.2] yields for each $(u,\xi)\in U_{0}\times\Xi$, $\displaystyle(\nabla F_{\xi}(u),s)_{L^{2}(D)}=(g(\xi)S(u,\xi)z(u,\xi),s)_{L^{2}(D)}\quad\text{for all}\quad s\in L^{2}(D),$ (7.11) where for each $(u,\xi)\in U_{0}\times\Xi$, the adjoint state $z=z(u,\xi)\in H_{0}^{1}(D)$ solves $\displaystyle(\kappa(\xi)\nabla z,\nabla v)_{L^{2}(D)^{d}}+(g(\xi)uz,v)_{L^{2}(D)}=-\langle\mathrm{D}J(S(u,\xi)),v\rangle_{H^{-1}(D),H_{0}^{1}(D)}$ (7.12) for all $v\in H_{0}^{1}(D)$. Using (7.8) and (7.12), we obtain the stability estimate $\displaystyle\begin{aligned} \|z(u,\xi)\|_{H_{0}^{1}(D)}&\leq(2/\kappa_{\min})\|\mathrm{D}J(S(u,\xi))\|_{H^{-1}(D)}\;\;\text{for all}\;\;(u,\xi)\in U_{0}\times\Xi.\end{aligned}$ Combined with the stability estimate (7.7), 7.1.4, and Lemma 7.2, we have for all $(u,\xi)\in U_{0}\times\Xi$, $\displaystyle\begin{aligned} \|z(u,\xi)\|_{H_{0}^{1}(D)}&\leq(2/\kappa_{\min})\big{(}2\ell\varrho((2/\kappa_{\min})C_{D}b_{\max})\big{)}^{1/2}.\end{aligned}$ (7.13) We show that 4.1.3 is satisfied. Using the Hölder and Friedrichs inequalities, the continuity of $H_{0}^{1}(D)\xhookrightarrow{}L^{4}(D)$, the stability estimates (7.7) and (7.13), and the gradient formula (7.11), we have $(u,\xi)\in U_{0}\times\Xi$, $\displaystyle\|\nabla F_{\xi}(u)\|_{L^{2}(D)}$ $\displaystyle\leq C_{H_{0}^{1};L^{4}}^{2}g_{\max}\|S(u,\xi)\|_{H_{0}^{1}(D)}\|z(u,\xi)\|_{H_{0}^{1}(D)}$ $\displaystyle\leq C_{H_{0}^{1};L^{4}}^{2}g_{\max}(2/\kappa_{\min})^{2}C_{D}b_{\max}\big{(}2\ell\varrho((2/\kappa_{\min})C_{D}b_{\max})\big{)}^{1/2}.$ Right-hand side in the above equation is a deterministic constant. Hence 4.1.3 holds true. The stability estimate (7.7) and 7.1.4 yield for all $(u,\xi)\in U_{0}\times\Xi$, $\displaystyle F_{\xi}(u)\leq\varrho((2/\kappa_{\min})C_{D}b_{\max}).$ (7.14) Combined with the measurability of $J(S(u,\cdot))$ for each $u\in U_{0}$, we find that 4.1.5 holds true. We verify 4.1.6 by showing that for each $(u,\xi)\in U_{\text{ad}}\times\Xi$, $\displaystyle\nabla F_{\xi}(u)=\iota_{1}m_{\xi}(u),\quad\text{where}\quad m_{\xi}(u)=g(\xi)S(u,\xi)z(u,\xi).$ (7.15) We define $M_{\xi}(u)=\\{\,m_{\xi}(u)\,\\}$, $V=H^{1}(D)$, and $B=\iota_{1}$. Fix $(u,\xi)\in U_{\text{ad}}\times\Xi$. We have $S(u,\xi)\in H^{2}(D)$ (see section 7.2 and (7.10)) and $z(u,\xi)\in H^{1}(D)$ (see (7.13)). Combined with $g(\xi)\in C^{0,1}(\bar{D})$ and (7.4), we find that $m_{\xi}(u)\in H^{1}(D)$. Hence (7.11) implies (7.15). We show that 4.1.7 holds true using Lemma 4.2. Fix $(u,\xi)\in U_{\text{ad}}\times\Xi$. Using (7.4) and (7.15), we have $\displaystyle\|m_{\xi}(u)\|_{H^{1}(D)}\leq C_{\mathrm{tri}}g_{\max}\|S(u,\xi)\|_{H^{2}(D)}\|z(u,\xi)\|_{H^{1}(D)}.$ Combined with the stability estimates (7.7), (7.10), and (7.13), and $\|y\|_{H^{1}(D)}\leq(1+C_{D})\|y\|_{H_{0}^{1}(D)}$ valid for all $y\in H_{0}^{1}(D)$, we find that $\displaystyle\|m_{\xi}(u)\|_{H^{1}(D)}$ $\displaystyle\leq C_{\mathrm{tri}}g_{\max}C_{H^{2}}C_{\kappa}(\xi)\big{(}C_{D}^{2}C_{\mathfrak{u}}g_{\max}(2/\kappa_{\min})b_{\max}+b_{\max}\big{)}$ $\displaystyle\quad\cdot(2/\kappa_{\min})C_{D}(1+C_{D})\big{(}2\ell\varrho((2/\kappa_{\min})C_{D}b_{\max})\big{)}^{1/2}.$ We define $\zeta$ as the right-hand side in the above equation. Using 7.1.2 and (7.9), we find that $\zeta$ is integrable. Hence Lemma 4.2 yields 4.1.7. ## 8 Discussion A multitude of applications in science and engineering yield PDE-constrained optimization problems under uncertainty. Motivated by such applications, we have considered infinite-dimensional stochastic optimization problems. We approximated the expectations in the stochastic programs using two approaches: the SAA approach and using probability measures weakly converging to the random element’s probability distribution. For both approximation approaches, we established asymptotic consistency statements for C-stationary points under Clarke regularity of integrands. We applied our framework to a risk-averse semilinear PDE-constrained optimization problem with the average value-at-risk as a risk measure, and to risk-neutral bilinear PDE-constrained optimization. Our consistency analysis requires integrands be Clarke regular. For many applications in PDE-constrained optimization, integrands are continuously differentiable and hence Clarke regular. For nonregular integrands, consistency analysis of SAA first-order stationary points is more demanding. Using smoothing functions of integrands, the SAA approach has been analyzed in [12, 83] as the sample size approaches infinity and smoothing parameters converge to zero. We refer the reader to [69] for recent contributions to the consistency analysis of SAA stationary points as applied to risk-neutral optimization problems with integrands that lack Clarke regularity. Our main result, Theorem 4.3, establishes the consistency of SAA C-stationary points of infinite-dimensional risk-neutral optimization problems. While applicable to a large set of applications, it may be desirable to establish consistency results for sample-based approximations of chance-constrained problems governed by PDEs with random inputs [15, 21, 79], and PDE-constrained problems under uncertainty with state constraints [25, 47]. Further future work includes extending the consistency analysis to a broader class of risk- averse PDE-constrained optimization problems. ## Acknowledgments JM is very grateful to Professor Alexander Shapiro for taking the time to address several of my questions related to the SAA approach. ## References * [1] J. Allen, E. Young, P. Bortolotti, R. King, and G. Barter, Blade planform design optimization to enhance turbine wake control, Wind Energy, 25 (2022), pp. 811–830, https://doi.org/10.1002/we.2699. * [2] Z. Artstein and R. J.-B. Wets, Stability results for stochastic programs and sensors, allowing for discontinuous objective functions, SIAM J. Optim., 4 (1994), pp. 537–550, https://doi.org/10.1137/0804030. * [3] Z. Artstein and R. J.-B. Wets, Consistency of minimizers and the SLLN for stochastic programs, J. Convex Anal., 2 (1995), pp. 1–17. * [4] J.-P. Aubin and H. Frankowska, Set-Valued Analysis, Mod. Birkhäuser Class., Springer, Boston, 2009, https://doi.org/10.1007/978-0-8176-4848-0. * [5] R. Balaji and H. Xu, Approximating stationary points of stochastic optimization problems in Banach space, J. Math. Anal. Appl., 347 (2008), pp. 333–343, https://doi.org/10.1016/j.jmaa.2008.06.015. * [6] A. Behzadan and M. Holst, Multiplication in Sobolev spaces, revisited, Ark. Mat., 59 (2021), pp. 275–306, https://doi.org/10.4310/arkiv.2021.v59.n2.a2. * [7] J. F. Bonnans and A. Shapiro, Perturbation Analysis of Optimization Problems, Springer Ser. Oper. Res., Springer, New York, 2000, https://doi.org/10.1007/978-1-4612-1394-9. * [8] J. M. Borwein and A. S. Lewis, On the convergence of moment problems, Trans. Amer. Math. Soc., 325 (1991), pp. 249–271, https://doi.org/10.2307/2001670. * [9] J. M. Borwein and A. S. Lewis, Strong rotundity and optimization, SIAM J. Optim., 4 (1994), pp. 146–158, https://doi.org/10.1137/0804008. * [10] J. M. Borwein and J. D. Vanderwerff, Convex functions: Constructions, Characterizations and Counterexamples, Encyclopedia Math. Appl. 109, Cambridge University Press, Cambridge, 2010, https://doi.org/10.1017/CBO9781139087322. * [11] A. Borzì, E.-J. Park, and M. Vallejos Lass, Multigrid optimization methods for the optimal control of convection-diffusion problems with bilinear control, J. Optim. Theory Appl., 168 (2016), pp. 510–533, https://doi.org/10.1007/s10957-015-0791-z. * [12] J. V. Burke, X. Chen, and H. Sun, The subdifferential of measurable composite max integrands and smoothing approximation, Math. Program., 181 (2020), pp. 229–264, https://doi.org/10.1007/s10107-019-01441-9. * [13] E. Casas, R. Herzog, and G. Wachsmuth, Optimality conditions and error analysis of semilinear elliptic control problems with $L^{1}$ cost functional, SIAM J. Optim., 22 (2012), pp. 795–820, https://doi.org/10.1137/110834366. * [14] E. Casas, D. Wachsmuth, and G. Wachsmuth, Second-order analysis and numerical approximation for bang-bang bilinear control problems, SIAM J. Control Optim., 56 (2018), pp. 4203–4227, https://doi.org/10.1137/17M1139953. * [15] P. Chen and O. Ghattas, Taylor approximation for chance constrained optimization problems governed by partial differential equations with high-dimensional random parameters, SIAM/ASA J. Uncertain. Quantif., 9 (2021), pp. 1381–1410, https://doi.org/10.1137/20M1381381. * [16] P. Chen and J. O. Royset, Performance bounds for PDE-constrained optimization under uncertainty, preprint, https://arxiv.org/abs/2110.10269, 2021. * [17] X. Chen, R. J.-B. Wets, and Y. Zhang, Stochastic variational inequalities: residual minimization smoothing sample average approximations, SIAM J. Optim., 22 (2012), pp. 649–673, https://doi.org/10.1137/110825248. * [18] F. H. Clarke, Optimization and Nonsmooth Analysis, Classics Appl. Math. 5, SIAM, Philadelphia, PA, 1990, https://doi.org/10.1137/1.9781611971309. * [19] H. T. H. Diem, Consistency of statistical estimators of solutions to stochastic optimization problems, J. Global Optim., 83 (2022), pp. 825–842, https://doi.org/10.1007/s10898-022-01125-3. * [20] R. S. Falk, Approximation of a class of optimal control problems with order of convergence estimates, J. Math. Anal. Appl., 44 (1973), pp. 28–47, https://doi.org/10.1016/0022-247X(73)90022-X. * [21] M. H. Farshbaf-Shaker, M. Gugat, H. Heitsch, and R. Henrion, Optimal Neumann boundary control of a vibrating string with uncertain initial data and probabilistic terminal constraints, SIAM J. Control Optim., 58 (2020), pp. 2288–2311, https://doi.org/10.1137/19M1269944. * [22] E. A. Feinberg, P. O. Kasyanov, and J. O. Royset, Epi-convergence of expectation functions under varying measures and integrands, J. Convex Anal., 30 (2023), pp. 917–936. * [23] F. Fuica and E. Otárola, A posteriori error estimates for an optimal control problem with a bilinear state equation, J. Optim. Theory Appl., 194 (2022), pp. 543–569, https://doi.org/10.1007/s10957-022-02039-6. * [24] C. Geiersbach and T. Scarinci, Stochastic proximal gradient methods for nonconvex problems in Hilbert spaces, Comput. Optim. Appl., 78 (2021), pp. 705–740, https://doi.org/10.1007/s10589-020-00259-y. * [25] C. Geiersbach and W. Wollner, Optimality conditions for convex stochastic optimization problems in Banach spaces with almost sure state constraints, SIAM J. Optim., 31 (2021), pp. 2455–2480, https://doi.org/10.1137/20M1363558. * [26] E. Giné and R. Nickl, Mathematical Foundations of Infinite-Dimensional Statistical Models, Camb. Ser. Stat. Probab. Math. 40, Cambridge University Press, New York, 2016, https://doi.org/10.1017/CBO9781107337862. * [27] R. Glowinski, Y. Song, X. Yuan, and H. Yue, Bilinear optimal control of an advection-reaction-diffusion system, SIAM Rev., 64 (2022), pp. 392–421, https://doi.org/10.1137/21M1389778. * [28] Z. L. Goss, D. S. Coles, S. C. Kramer, and M. D. Piggott, Efficient economic optimisation of large-scale tidal stream arrays, Appl. Energy, 295 (2021), p. 116975, https://doi.org/10.1016/j.apenergy.2021.116975. * [29] P. Grisvard, Elliptic Problems in Nonsmooth Domains, Classics Appl. Math. 69, SIAM, Philadelphia, PA, 2011, https://doi.org/10.1137/1.9781611972030. * [30] J. Guan and M. Aral, Optimal remediation with well locations and pumping rates selected as continuous decision variables, Journal of Hydrology, 221 (1999), pp. 20–42, https://doi.org/10.1016/S0022-1694(99)00079-7. * [31] F. Guillen-Gonzalez, E. Mallea-Zepeda, and M. A. Rodriguez-Bellido, A regularity criterion for a 3D chemo-repulsion system and its application to a bilinear optimal control problem, SIAM J. Control Optim., 58 (2020), pp. 1457–1490, https://doi.org/10.1137/18M1209891. * [32] L. Hertlein, A.-T. Rauls, M. Ulbrich, and S. Ulbrich, An inexact bundle method and subgradient computations for optimal control of deterministic and stochastic obstacle problems, in Non-Smooth and Complementarity-Based Distributed Parameter Systems: Simulation and Hierarchical Optimization, M. Hintermüller, R. Herzog, C. Kanzow, M. Ulbrich, and S. Ulbrich, eds., Internat. Ser. Numer. Math. 172, Birkhäuser, Cham, 2022, pp. 467–497, https://doi.org/10.1007/978-3-030-79393-7_19. * [33] L. Hertlein and M. Ulbrich, An inexact bundle algorithm for nonconvex nonsmooth minimization in Hilbert space, SIAM J. Control Optim., 57 (2019), pp. 3137–3165, https://doi.org/10.1137/18M1221849. * [34] C. Hess and R. Seri, Generic consistency for approximate stochastic programming and statistical problems, SIAM J. Optim., 29 (2019), pp. 290–317, https://doi.org/10.1137/17M1156769. * [35] M. Hintermüller and T. Surowiec, A bundle-free implicit programming approach for a class of elliptic MPECs in function space, Math. Program., 160 (2016), pp. 271–305, https://doi.org/10.1007/s10107-016-0983-9. * [36] M. Hintermüller and M. Ulbrich, A mesh-independence result for semismooth Newton methods, Math. Program., 101 (2004), pp. 151–184, https://doi.org/10.1007/s10107-004-0540-9. * [37] M. Hinze, R. Pinnau, M. Ulbrich, and S. Ulbrich, Optimization with PDE Constraints, Math. Model. Theory Appl. 23, Springer, Dordrecht, 2009, https://doi.org/10.1007/978-1-4020-8839-1. * [38] M. Hoffhues, W. Römisch, and T. M. Surowiec, On quantitative stability in infinite-dimensional optimization under uncertainty, Optim. Lett., 15 (2021), pp. 2733–2756, https://doi.org/10.1007/s11590-021-01707-2. * [39] T. Hytönen, J. van Neerven, M. Veraar, and L. Weis, Analysis in Banach Spaces: Martingales and Littlewood-Paley Theory, Ergeb. Math. Grenzgeb. (3) 63, Springer, Cham, 2016, https://doi.org/10.1007/978-3-319-48520-1. * [40] P. Jarde and M. Ulbrich, Existence of minimizers for optical flow based optimal control problems under mild regularity assumptions, Control Cybernet., 48 (2019), pp. 261–306. * [41] M. Kahlbacher and S. Volkwein, POD a-posteriori error based inexact SQP method for bilinear elliptic optimal control problems, ESAIM Math. Model. Numer. Anal., 46 (2012), pp. 491–511, https://doi.org/10.1051/m2an/2011061. * [42] O. Kallenberg, Foundations of Modern Probability, Probab. Appl., Springer, New York, NY, 2nd ed., 2002, https://doi.org/10.1007/978-1-4757-4015-8. * [43] Y. L. Keung and J. Zou, An efficient linear solver for nonlinear parameter identification problems, SIAM J. Sci. Comput., 22 (2000), pp. 1511–1526, https://doi.org/10.1137/S1064827598346740. * [44] R. N. King, K. Dykes, P. Graf, and P. E. Hamlington, Optimization of wind plant layouts using an adjoint approach, Wind Energy Science, 2 (2017), pp. 115–131, https://doi.org/10.5194/wes-2-115-2017. * [45] A. J. Kleywegt, A. Shapiro, and T. Homem-de Mello, The sample average approximation method for stochastic discrete optimization, SIAM J. Optim., 12 (2002), pp. 479–502, https://doi.org/10.1137/S1052623499363220. * [46] M. Koivu, Variance reduction in sample approximations of stochastic programs, Math. Program., 103 (2005), pp. 463–485, https://doi.org/10.1007/s10107-004-0557-0. * [47] D. P. Kouri, M. Staudigl, and T. M. Surowiec, A relaxation-based probabilistic approach for PDE-constrained optimization under uncertainty with pointwise state constraints, Comput. Optim. Appl., 85 (2023), pp. 441–478, https://doi.org/10.1007/s10589-023-00461-8. * [48] D. P. Kouri and T. M. Surowiec, Risk-averse PDE-constrained optimization using the conditional value-at-risk, SIAM J. Optim., 26 (2016), pp. 365–396, https://doi.org/10.1137/140954556. * [49] D. P. Kouri and T. M. Surowiec, Existence and optimality conditions for risk-averse PDE-constrained optimization, SIAM/ASA J. Uncertain. Quantif., 6 (2018), pp. 787–815, https://doi.org/10.1137/16M1086613. * [50] D. P. Kouri and T. M. Surowiec, Risk-averse optimal control of semilinear elliptic PDEs, ESAIM Control. Optim. Calc. Var., 26 (2020), https://doi.org/10.1051/cocv/2019061. * [51] E. Kreyszig, Introductory Functional Analysis with Applications, John Wiley & Sons, New York, NY, 1978. * [52] A. Kröner and B. Vexler, A priori error estimates for elliptic optimal control problems with a bilinear state equation, J. Comput. Appl. Math., 230 (2009), pp. 781–802, https://doi.org/10.1016/j.cam.2009.01.023. * [53] K. Kunisch and X.-C. Tai, Sequential and parallel splitting methods for bilinear control problems in Hilbert spaces, SIAM J. Numer. Anal., 34 (1997), pp. 91–118, https://doi.org/10.1137/S0036142993255253. * [54] K. Kunisch and D. Walter, Semiglobal optimal feedback stabilization of autonomous systems via deep neural network approximation, ESAIM Control Optim. Calc. Var., 27 (2021), https://doi.org/10.1051/cocv/2021009. * [55] S. Lang, Real and Functional Analysis, Grad. Texts in Math. 142, Springer, Berlin, 3rd ed., 1993, https://doi.org/10.1007/978-1-4612-0897-6. * [56] H.-J. Langen, Convergence of dynamic programming models, Math. Oper. Res., 6 (1981), pp. 493–512, https://doi.org/10.1287/moor.6.4.493. * [57] A. Mang, A. Gholami, C. Davatzikos, and G. Biros, PDE-constrained optimization in medical image analysis, Optim. Eng., 19 (2018), pp. 765–812, https://doi.org/10.1007/s11081-018-9390-9. * [58] M. Martin, S. Krumscheid, and F. Nobile, Complexity analysis of stochastic gradient methods for PDE-constrained optimal control problems with uncertain parameters, ESAIM Math. Model. Numer. Anal., 55 (2021), pp. 1599–1633, https://doi.org/10.1051/m2an/2021025. * [59] J. Milz, Consistency of Monte Carlo estimators for risk-neutral PDE-constrained optimization, Appl. Math. Optim., 87 (2023), https://doi.org/10.1007/s00245-023-09967-3. * [60] J. Milz, Sample average approximations of strongly convex stochastic programs in Hilbert spaces, Optim. Lett., 17 (2023), pp. 471–492, https://doi.org/10.1007/s11590-022-01888-4. * [61] J. Milz, Reliable error estimates for optimal control of linear elliptic PDEs with random inputs, preprint, https://arxiv.org/abs/2206.09160, June 2022. * [62] J. Milz and T. M. Surowiec, Asymptotic consistency for nonconvex risk-averse stochastic optimization with infinite dimensional decision spaces, preprint, https://arxiv.org/abs/2207.14756, July 2022. * [63] J. Milz and M. Ulbrich, An approximation scheme for distributionally robust PDE-constrained optimization, SIAM J. Control Optim., 60 (2022), pp. 1410–1435, https://doi.org/10.1137/20M134664X. * [64] J. Milz and M. Ulbrich, Sample size estimates for risk-neutral semilinear PDE-constrained optimization, preprint, https://arxiv.org/abs/2207.14755, July 2022. * [65] R. Nickl, S. van de Geer, and S. Wang, Convergence rates for penalized least squares estimators in PDE constrained regression problems, SIAM/ASA J. Uncertain. Quantif., 8 (2020), pp. 374–413, https://doi.org/10.1137/18M1236137. * [66] V. I. Norkin, Random Lipschitz functions, Cybern. Syst. Anal., (1986), pp. 226–233, https://doi.org/10.1007/BF01074785. * [67] T. Pennanen and M. Koivu, Epi-convergent discretizations of stochastic programs via integration quadratures, Numer. Math., 100 (2005), pp. 141–163, https://doi.org/10.1007/s00211-004-0571-4. * [68] C. Phelps, J. Royset, and Q. Gong, Optimal control of uncertain systems using sample average approximations, SIAM J. Control Optim., 54 (2016), pp. 1–29, https://doi.org/10.1137/140983161. * [69] Z. Qi, Y. Cui, Y. Liu, and J.-S. Pang, Asymptotic properties of stationary solutions of coupled nonconvex nonsmooth empirical risk minimization, Math. Oper. Res., 47 (2022), pp. 2034–2064, https://doi.org/10.1287/moor.2021.1198. * [70] R. T. Rockafellar and R. J.-B. Wets, Variational Analysis, Grundlehren Math. Wiss. 317, Springer, Berlin, 2009, https://doi.org/10.1007/978-3-642-02431-3. * [71] W. Römisch and T. M. Surowiec, Asymptotic properties of Monte Carlo methods in elliptic PDE-constrained optimization under uncertainty, preprint, https://arxiv.org/abs/2106.06347, 2021. * [72] A. Shapiro, Monte Carlo Sampling Methods, in Stochastic Programming, Handbooks in Oper. Res. Manag. Sci. 10, Elsevier, 2003, pp. 353–425, https://doi.org/10.1016/S0927-0507(03)10006-0. * [73] A. Shapiro, D. Dentcheva, and A. Ruszczyński, Lectures on Stochastic Programming: Modeling and Theory, MOS-SIAM Ser. Optim., SIAM, Philadelphia, PA, 3rd ed., 2021, https://doi.org/10.1137/1.9781611976595. * [74] A. Shapiro and H. Xu, Uniform laws of large numbers for set-valued mappings and subdifferentials of random functions, J. Math. Anal. Appl., 325 (2007), pp. 1390–1399, https://doi.org/10.1016/j.jmaa.2006.02.078. * [75] S. Stojanovic, Optimal damping control and nonlinear elliptic systems, SIAM J. Control Optim., 29 (1991), pp. 594–608, https://doi.org/10.1137/0329033. * [76] T. M. Surowiec, Numerical optimization methods for the optimal control of elliptic variational inequalities, in Frontiers in PDE-Constrained Optimization, H. Antil, D. P. Kouri, M.-D. Lacasse, and D. Ridzal, eds., IMA Vol. Math. Appl. 163, Springer, New York, NY, 2018, pp. 123–170, https://doi.org/10.1007/978-1-4939-8636-1_4. * [77] A. L. Teckentrup, R. Scheichl, M. B. Giles, and E. Ullmann, Further analysis of multilevel Monte Carlo methods for elliptic PDEs with random coefficients, Numer. Math., 125 (2013), pp. 569–600, https://doi.org/10.1007/s00211-013-0546-4. * [78] P. Terán, On consistency of stationary points of stochastic optimization problems in a Banach space, J. Math. Anal. Appl., 363 (2010), pp. 569–578, https://doi.org/10.1016/j.jmaa.2009.07.062. * [79] S. Tong, A. Subramanyam, and V. Rao, Optimization under rare chance constraints, SIAM J. Optim., 32 (2022), pp. 930–958, https://doi.org/10.1137/20M1382490. * [80] M. Ulbrich, Semismooth Newton Methods for Variational Inequalities and Constrained Optimization Problems in Function Spaces, MOS-SIAM Ser. Optim., SIAM, Philadelphia, PA, 2011, https://doi.org/10.1137/1.9781611970692. * [81] M. Vallejos and A. Borzì, Multigrid optimization methods for linear and bilinear elliptic optimal control problems, Computing, 82 (2008), pp. 31–52, https://doi.org/10.1007/s00607-008-0261-7. * [82] M. Winkler, Error estimates for the finite element approximation of bilinear boundary control problems, Comput. Optim. Appl., 76 (2020), pp. 155–199, https://doi.org/10.1007/s10589-020-00171-5. * [83] H. Xu and D. Zhang, Smooth sample average approximation of stationary points in nonsmooth stochastic optimization and applications, Math. Program., 119 (2009), pp. 371–401, https://doi.org/10.1007/s10107-008-0214-0. * [84] D. Yang, Y. Chang, and W. Liu, A priori error estimate and superconvergence analysis for an optimal control problem of bilinear type, J. Comput. Math., 26 (2008), pp. 471–487, https://www.jstor.org/stable/43693457. * [85] E. Zeidler, Nonlinear Functional Analysis and its Applications _II/B_ : Nonlinear Monotone Operators, Springer, New York, 1990, https://doi.org/10.1007/978-1-4612-0981-2. * [86] C. Zălinescu, Convex Analysis in General Vector Spaces, World Scientific, River Edge, NJ, 2002, https://doi.org/10.1142/9789812777096.
# Caught in the Act: A Metal-Rich High-Velocity Cloud in the Inner Galaxy Frances H. Cashman Space Telescope Science Institute, 3700 San Martin Drive, Baltimore, MD 21218, USA Andrew J. Fox AURA for ESA, Space Telescope Science Institute, 3700 San Martin Drive, Baltimore, MD 21218, USA Bart P. Wakker Department of Astronomy, University of Wisconsin-Madison, 475 North Charter Street, Madison, WI 53706, USA Trisha Ashley Space Telescope Science Institute, 3700 San Martin Drive, Baltimore, MD 21218, USA Derck Massa Space Science Institute, 4750 Walnut Street Suite 205 Boulder, Colorado 80301, USA Edward B. Jenkins Department of Astrophysical Sciences, Princeton University, Princeton, NJ 08544-1001, USA Dhanesh Krishnarao NSF Astronomy & Astrophysics Postdoctoral Fellow, Johns Hopkins University, 3400 N. Charles Street, Baltimore, MD 21218, USA Department of Physics, Colorado College, 14 East Cache La Poudre Street, Colorado Springs, CO 80903, USA Space Telescope Science Institute, 3700 San Martin Drive, Baltimore, MD 21218, USA Felix J. Lockman Green Bank Observatory, P.O. Box 2, Rt. 28/92, Green Bank, WV 24944, USA Robert A. Benjamin Department of Physics, University of Wisconsin- Whitewater, 800 West Main Street, Whitewater, WI 53190, USA Rongmon Bordoloi Department of Physics, North Carolina State University, 421 Riddick Hall, Raleigh, NC 27695-8202, USA Tae-Sun Kim Department of Astronomy, University of Wisconsin-Madison, 475 North Charter Street, Madison, WI 53706, USA (Received September 28, 2022) ###### Abstract We characterize the chemical and physical conditions in an outflowing high- velocity cloud in the inner Galaxy. We report a super-solar metallicity of [O/H] = $+0.36\pm 0.12$ for the high-velocity cloud at $v_{\mathrm{LSR}}$ = 125.6 km s-1 toward the star HD 156359 (catalog ) ($l$ = 328.$\degree$7, $b$ = $-$14.$\degree$5, $d$ = 9 kpc, $z$ = $-$2.3 kpc). Using archival observations from FUSE, HST STIS, and ESO FEROS we measure high-velocity absorption in H I, O I, C II, N II, Si II, Ca II, Si III, Fe III, C IV, Si IV, N V, and O VI. We measure a low H I column density of log $N$(H I) = $15.54\pm 0.05$ in the HVC from multiple unsaturated H I Lyman series lines in the FUSE data. We determine a low dust depletion level in the HVC from the relative strength of silicon, iron, and calcium absorption relative to oxygen, with [Si/O]=$-0.33\pm 0.14$, [Fe/O]=$-0.30\pm 0.20$, and [Ca/O] =$-0.56\pm 0.16$. Analysis of the high-ion absorption using collisional ionization models indicates that the hot plasma is multi-phase, with the C IV and Si IV tracing 104.9 K gas and N V and O VI tracing 105.4 K gas. The cloud’s metallicity, dust content, kinematics, and close proximity to the disk are all consistent with a Galactic wind origin. As the HD 156359 (catalog ) line of sight probes the inner Galaxy, the HVC appears to be a young cloud caught in the act of being entrained in a multi-phase Galactic outflow and driven out into the halo. Galactic Center (565) — Ultraviolet astronomy (1736) — High-velocity clouds (735) – Chemical abundances (224) — Galactic winds (572) ††journal: ApJ††facilities: FUSE, HST(STIS), ESO(FEROS)††software: astropy (Astropy Collaboration et al., 2018), Cloudy (Ferland et al., 2017), linetools (Prochaska et al., 2017), VPFIT (Carswell & Webb, 2014), CalFUSE (Dixon & Kruk, 2009), calSTIS (Dressel et al., 2007), FEROS-DRS (Kaufer et al., 1999) ## 1 Introduction Figure 1: Left panel: The location of HD 156359 (catalog ) with reference to the Fermi and eROSITA bubbles. The composite Fermi–eROSITA image from Predehl et al. (2020), where the softer X-ray emission (0.6–1 keV, in cyan) envelopes the harder component of the extended GeV emission of the Fermi bubbles (in red; adapted from Selig et al. 2015). Right panel: H I column density map from the 21 cm HI4PI survey showing the distribution of H I in the region from 100–150 km s-1 HI4PI Collaboration et al. (2016) in a section of “Complex WE” by Wakker & van Woerden (1991). The location of HD 156359 (catalog ) is marked with a white cross. The contours correspond to log $N_{\mathrm{HI}}$ = 2.08, 6.23, 10.4, 14.5, 18.7 $\times$ 1018 cm-2. The supermassive black hole, Sagittarius A∗, and surrounding regions of active star formation power an outflowing multiphase wind from the center of the Milky Way (MW). Outflows play a critical role in the baryon cycle, the cycling of gas into and out of galaxies, which helps regulate the evolution of galaxies. The MW offers us a front-row seat view to study outflows over a range of wavelengths and phases and understand their impact on galaxy evolution. Presently, the farthest-reaching known outflows in the MW are the Fermi (Su et al., 2010; Ackermann et al., 2014) and eROSITA (Bland-Hawthorn & Cohen, 2003; Predehl et al., 2020) Bubbles, which are giant gamma- and X-ray emission extending $\sim$10 kpc and 14 kpc above and below the Galactic disk, respectively (see Figure 1). The origin of the Fermi Bubbles is believed to be an energetic GC Seyfert flare event $\sim$3.5 Myr ago (Bland-Hawthorn et al., 2019). Ultraviolet (UV) absorption-line studies have found gas outflowing at high- velocity in a number of sight lines passing through or near the Fermi Bubbles (see Keeney et al. 2006; Zech et al. 2008; Fox et al. 2015; Bordoloi et al. 2017; Savage et al. 2017; Karim et al. 2018; Ashley et al. 2020, 2022). These high-velocity clouds (HVCs) consist of neutral (e.g., C I, O I), singly ionized (S II, Si II, Fe II), and highly ionized (C IV, Si IV, O VI) gas. Cold gas has also been detected in the nuclear outflow via emission in neutral hydrogen and molecules. Several hundred H I 21 cm clouds have been detected at low latitudes in the Fermi Bubbles outflowing from the Galactic Center (GC) (McClure-Griffiths et al., 2013; Di Teodoro et al., 2018; Lockman et al., 2020). Carbon monoxide (CO) has been detected in submillimeter emission in two dense molecular clouds comoving within the H I 21 cm outflow (Di Teodoro et al., 2020). More recently, molecular hydrogen was discovered in the Galactic nuclear outflow $\sim$1 kpc below the GC (Cashman et al., 2021). Outflowing gas associated with the Fermi Bubbles has also been seen in optical emission (Krishnarao et al., 2020a). Together, these observations of HVCs near the Galactic Center provide observational constraints on the properties of the MW nuclear wind. The outflowing wind interacts with gas in the disk in multiple ways. First, it can disrupt the position and velocity of the disk (Krishnarao et al., 2020b), leading to warps and perturbations. Second, it can accelerate and entrain cool gas into the halo. In some instances, this high-velocity gas may cool, lose buoyancy, and fall back onto the disk as a “Galactic fountain”, supplying new fuel for further star formation (see Shapiro & Field 1976; Bregman 1980; Kahn 1981; de Avillez 1999). In other cases, the high-velocity gas may survive being accelerated into the halo, and even escape. Cloud survival is the subject of recent theoretical work focusing on how cool gas clouds develop and grow in the hot wind (see Gronke & Oh 2020; Sparre et al. 2020). Metallicities of HVCs are important to determine because they provide direct evidence for the origin of the clouds. Fully UV-based metallicities (with the metal and H I column densities measured along the same UV sight lines) are ideal to ensure that the metal absorption and hydrogen absorption are tracing the same gas, but they are rare because of the difficulty in isolating unsaturated high-velocity H I components in UV absorption. These high-velocity H I components are often saturated or blended, even for high-order Lyman series lines, thus there is only a narrow range of $N$(H I) values for which precise measurements are possible (French et al., 2021). Therefore HVC metallicities are frequently determined from a combination of a UV measurement along an infinitesimal sight line with a H I measurement made using a much larger 21 cm beam. However, this combination introduces beam-smearing uncertainties on the metallicity. Currently, the only fully-UV-based metallicities are for two HVCs toward M5-ZNG1 (Zech et al., 2008), but this is a high-latitude direction ($b\approx 50\degree$) far from the GC. At lower latitudes, we can target UV-bright stars as background sources, but very few UV spectra of distant GC stars ($d\gtrsim 8$ kpc) exist. Therefore, it is crucial to fully analyze the few spectra available, including LS 4825 (Savage et al., 2017; Cashman et al., 2021) and HD 156359 (this paper). It is precisely these low-latitude sight lines that are likely to harbor young clouds that have only recently been entrained in the Galactic outflow. Finding an outflowing cloud at low latitude means catching it close to its origin, thus offering a rare opportunity to study the cloud before it has undergone significant mixing as it begins its journey into the halo. Observing recently entrained clouds therefore gives a snapshot of outflowing gas when it first exits the disk. In this paper we present a detailed spectroscopic analysis of an HVC detected toward an inner Galaxy sight line, HD 156359 (catalog ). By studying the properties of this cloud, we characterize the physical and chemical conditions of the gas in the inner Galaxy, in a region likely influenced by the nuclear wind. ## 2 Data ### 2.1 The HD 156359 Sight line HD 156359 (at $l,b,d$ = 328.68°, $-$14.52°, 9 kpc) is one of the best-studied GC sight lines (Sembach et al., 1991, 1995). The sight line lies in the inner Galaxy at the boundary of the southern Fermi Bubble and within the eROSITA X-ray bubble (Predehl et al., 2020), a region of enhanced X-ray emission (see Figure 1). This sight line also passes close to a complex of small HVCs dubbed “Complex WE” by Wakker & van Woerden (1991), less than 1$\degree$ from one of the densest cores. Sembach et al. (1991) classified HD 156359 (catalog ) as a O9.7 Ib–II star on the basis of stellar photospheric lines and wind profiles in its UV spectrum. The spectral type and apparent magnitude of the star yield a spectroscopic distance of 11.1$\pm$2.8 kpc. However, recent Gaia EDR3 parallax measurements (Bailer-Jones et al., 2021) place HD 156359 (catalog ) at a distance of 9.0${}^{+3.1}_{-2.1}$ kpc, which implies a $z$-distance below the plane of 2.3 kpc. We adopt the Gaia distance in our analysis. The sight line toward HD 156359 (catalog ) intersects at least three spiral arms, the Sagittarius, Scutum, and Norma arms. Spiral-arm models from Vallée (2017) give the expected velocities of the Sagittarius, Scutum, and Norma arms at approximately $-$10, $-$55, and $-$103 km s-1, respectively. Thus, absorption components observed at these velocities can be attributed to gas in (or associated with) the spiral arms. However, the spiral arms cannot explain the HVC observed at $v_{\rm LSR}$=125 km s-1. High-ion absorption toward HD 156359 (catalog ) was first observed with the International Ultraviolet Explorer (IUE; Sembach et al., 1991) and later by the Goddard High Resolution Spectrograph (GHRS; Sembach et al., 1995). The high-ion information in the FUSE and HST STIS spectra of this sight line has not been previously published – the sight line is not included in the FUSE O VI survey of Galactic disk sight lines by Bowen et al. (2008), as that survey was restricted to $|b|<10\degree$. We also include analysis of a single archival optical spectrum of HD 156359 (catalog ) taken with FEROS at the European Southern Observatory (ESO) at La Silla. Details of all observations of spectra used in this paper are described below. ### 2.2 UV Observations HD 156359 (catalog ) was observed by FUSE (Moos et al., 2000) under programs P101, S701, and U109 between 2000 April 12 and 2006 April 25 (PIs Sembach, Andersson, and Blair, respectively). The raw spectra were downloaded from the MAST FUSE archive, and the CalFUSE pipeline (v3.2.2; Dixon & Kruk 2009) was used to reduce and extract the spectra. Data from the SiC, LiF1, and LiF2 channels were used for the analysis. A detailed explanation of the data reduction, as well as refinements to the CalFUSE data reduction procedures can be found in Wakker et al. (2003) and Wakker (2006). The spectra have a signal- to-noise ratio (S/N) $\sim$12–26 per resolution element and a velocity resolution of 20 km s-1 (FWHM). The data were binned by three pixels for the fitting analysis. The FUSE wavelength coverage is $\sim$912–1180Å. A single exposure of HD 156359 was obtained with HST/STIS on 2003 March 26 under program 9434 (PI Lauroesch) using the E140M grating. The data were downloaded from the MAST HST archive and reduced using calstis (v.3.4.2, Dressel et al. 2007). The echelle orders were combined to create a single continuous spectrum, and in regions of order-overlap spectral counts were combined to increase the S/N ratio. The data have S/N $\sim$10–38 per resolution element and a FWHM velocity resolution of 6.5 km s-1, i.e. spectral resolution ($\lambda$/$\Delta\lambda$) $\sim$45,800. The STIS E140M wavelength coverage is $\sim$1160–1725Å. Wavelengths and velocities for the absorption- line features are given in the local standard of rest (LSR) reference frame, where the correction factor $v_{\mathrm{LSR}}$ $-$ $v_{\mathrm{Helio}}$ = $-0.35$ km s-1 for the direction toward HD 156359 (catalog ). ### 2.3 Optical Observations A spectrum of HD 156359 (catalog ) was captured using the Fibre-fed Extended Range Optical Spectrograph (FEROS) at the European Southern Observatory (ESO) La Silla 2.2 meter telescope on 2006 April 30 (PI: Bouret, PID: 077.D-0635). The retrieved archival product covers the wavelength range $\sim$3565–9214 Å with R = 48000 and FWHM = 6.25 km s-1. The primary data product was reduced automatically using the FEROS Data Reduction Software (DRS) pipeline version fern/1.0. In the automated reduction process, the bias is subtracted using overscan regions and bad columns are replaced by the mean values of the neighboring columns. The orders are rectified and then extracted using the standard method. Next, the extracted spectra are flat-fielded and wavelength calibrated, then rebinned to a constant dispersion of 0.03 Å. Finally, the individual orders are combined into a single 1D spectrum. As the archival reduced data are not flux calibrated, the continuum of the stellar spectrum was normalized using the linetools software package. ## 3 Measurements In this section we describe our absorption-line measurement processes, including stellar continuum fitting and procedures for measuring lines of different ionization states. Our measurements are based on the VPFIT software program (v12.2; Carswell & Webb 2014), which we used to conduct a set of Voigt-profile fits to the absorption-line profiles, with wavelengths and oscillator strengths taken from the compilations of Morton (2003) and Cashman et al. (2017). The measured lines are listed in Table 1. ### 3.1 Stellar continuum modeling Given the spectral type of HD 156359 (catalog ) (O9.7 Ib–II; Sembach et al., 1991), the stellar continuum placement requires careful consideration. In addition to estimating the stellar continuum through comparison with FUSE spectra of stars of similar spectral type, we also constructed a TLUSTY model (Lanz & Hubeny, 2003) of this type of star as a continuum placement guide. This was particularly useful in regions where the interstellar absorption was stronger. The adopted TLUSTY model has $T_{\mathrm{eff}}$ = 26000 K, log $g$ = 3.0, $v_{\mathrm{turb}}$ = 10.0 km s-1 and is rotationally broadened by 90 km s-1 in order to match the Si III 1300 Å triplets. We matched the observed UV wind signatures found in the observed FUSE and STIS spectra, where all observed spectra are adjusted for the stellar radial velocity $v_{\mathrm{rad}}=-82$ km s-1 (Gontcharov, 2006). The model is reddened by an $E(B-V)=0.09$, using the mean Fitzpatrick & Massa (2007) extinction curve (extrapolated to 912 Å), and Lyman series absorption for a foreground H I column of $N$(H I) = 6.0 $\times$ 1020 cm-2 is also included (Sembach et al., 1991). The model is then scaled by 1.25 $\times$ 10-20 in order to match observations to $\sim$10%. Finally, all spectra were binned to 0.1 Å, $\sim$15 to 30 km s-1, depending on wavelength, see Figure 2. When consulting the model as a guide for interstellar line continua, additional tweaks of about 5% were needed to match local continua. Our subsequent column density measurements account for the uncertainty inherent in the continuum-placement process. Figure 2: A comparison of a section of the FUSE FUV spectrum with a TLUSTY stellar model for the same spectral type, O9.7, which aided in the continuum placement. The FUSE SiC2a spectrum from 914–944 Å is shown in black with a stellar TLUSTY model shown in red. The TLUSTY model has $T_{\mathrm{eff}}$ = 26000 K, log $g$ = 3.0, $v_{\mathrm{turb}}$ = 10.0 km s-1, and is rotationally broadened by 90 km s-1 in order to match the observed Si III 1300 Å triplets. The model is reddened by an $E(B-V)$ = 0.09 using the mean Fitzpatrick & Massa (2007) extinction curve and Lyman line absorption for log $N$(H I) = 20.78 (Sembach et al., 1991) is also applied. The flux was scaled by 1.25$\times$10-20 to provide an overall agreement with observations to $\sim$10%. ### 3.2 H I absorption Figure 3: Velocity profiles of the neutral, low-, and intermediate-ion absorption lines detected toward HD 156359. The normalized flux is shown in black, the continuum level is in red, and the 1$\sigma$ error in the normalized flux is in blue. The vertical line at 0 km s-1 marks the region associated with the Milky Way. Each of these profiles shows a Voigt fit to the data except for O I for which we provide an AOD measurement, and for Fe II, which is a non-detection. The solid green curve is the overall Voigt profile fit to the absorption. The magenta curve is the Voigt profile fit to the HVC absorption feature near $+$125 km s-1 and the vertical dashed magenta line marks the velocity centroid of the fitted component. Figure 4: A comparison of the normalized apparent column density plots between O I 1302 in violet and Si II 1526 in green. The profiles have similar shapes although the O I 1302 profile shows a lower signal to noise. The inset panel shows the apparent column density ratio of Si II 1526 to O I 1302. A linear fit to 5 pixels in the region from $+110\lesssim v\lesssim 125$ km s-1 has a slope near zero within the margin of error, supporting the reality of the detection in both ions. Figure 5: Velocity profiles of the Ca II $\lambda\lambda$3934, 3969 and Na I $\lambda\lambda$5891, 5897 absorption lines detected toward HD 156359 in the archival FEROS spectrum. The normalized flux is shown in black, the continuum level is in red. The gray vertical line at 0 km s-1 marks the absorption region associated with the Milky Way. The green curves in the top two panels are the Voigt profile fits to the data, where the green vertical line at 123.7 km s-1 marks the location of the HVC. There is a non-detection of high-velocity Na I, although telluric H2O absorption lines are present in this region. The lower end of the y-axis begins at $+0.2$ in all panels to improve the visibility of the absorption features. The FUSE spectrum of HD 156359 (catalog ) covers almost the entire H I Lyman series, from Lyman-$\beta$ at 1025 Å down to the Lyman limit at 912 Å. However, we limited our fitting to H I lines in regions where the stellar continuum was better-defined and showed the least contamination from stellar features and/or interstellar molecular lines, which are prolific in the FUSE bandpass. We selected H I $\lambda\lambda$923, 926, 930, and 937 (see Figure 3) to derive the H I column density. We fit a continuum to each of these lines locally. In order to account for sensitivity to continuum placement, we consider a high and low continuum placement for our column density measurement across the wavelength range $\lambda$917–944. H I $\lambda$923 lies in a noisy region of the upper Lyman lines frequently associated with the Inglis-Teller effect (Inglis & Teller, 1939), in which overlapping stellar absorption lines have merged to produce the appearance of a depressed continuum. We perform a simultaneous Voigt profile fit to $\lambda\lambda$926, 930, 937 with the positions of the low-velocity absorption initially based on weak ISM lines. We derive log $N_{\mathrm{HI}}$ = 15.54 $\pm$ 0.02 $\pm$ 0.05 in the HVC at $+125.6$ km s-1, where the first error is the statistical error due to photon noise and the second is the systematic continuum-placement error. The result of these fits is shown Table 2 and in Figure 3, where we show that the simultaneous fit is also consistent with the noisier H I $\lambda$923 profile. We include a detailed explanation of our continuum placement procedures and how we considered contamination from stellar features in Appendix A. An important step in measuring H I absorption in FUSE spectra is to decontaminate H2 absorption. Fortunately, the FUSE LiF1 and LiF2 channels provide us with an opportunity to model isolated H2 lines. This allows us to account for the Milky Way’s molecular contribution to the interstellar absorption lines blended with H I in the SiC2A spectrum. We conducted an H2 decontamination and describe and illustrate this process in detail in Appendix B. ### 3.3 O I absorption After placing a smooth continuum through the O I line at 1302 Å in the STIS E140M spectrum, we noticed a small absorption feature near $+125$ km s-1 (see Figure 3). This feature spans five pixels and has a significance of 3.2$\sigma$ ($EW$/$\sigma_{EW}$=3.2). Since only one STIS exposure exists, we cannot confirm the O I detection in another dataset, and this feature is not seen in the much weaker O I line at 1039 Å in the FUSE spectrum. To explore whether this feature is real, we compared the absorption profile to another low-ion line. Figure 4 shows an apparent column density profile comparison of the O I 1302 Å line to the weak Si II 1526 Å line. We chose $\lambda$1526 because it is unsaturated, unblended, and seen in a region with high signal to noise. Although noisier, the profile of O I $\lambda$1302 is very similar to Si II $\lambda$1526 in the velocity region near $+125$ km s-1, and the ratio of their apparent column densities is flat with velocity. To confirm this similarity, the inset plot in Figure 4 shows a linear fit to the ratio over five pixels (over 2 resolution elements) with a slope equal to zero within the margin of error, as expected for a genuine O I detection. Therefore, we proceed with treating the O I $\lambda$1302 feature as real on the basis that: 1) the line is detected at 3.2$\sigma$ significance, 2) there is close kinematic consistency with Si II, and 3) the feature is centered at the same velocity as multiple other ions, e.g., H I, C II, N II, and Ca II. After applying a low and high continuum to this region, we measure log $N_{\mathrm{OI}}$ = 12.43 $\pm$ 0.08 $\pm$ 0.07, which includes the statistical error and a continuum-placement (systematic) error. This measurement is the foundation of our metallicity measurement in the HVC, which we discuss in Section 4. ### 3.4 Low and intermediate ion absorption We detected C II $\lambda$1334, Si II $\lambda\lambda$1190, 1193, 1304, 1526, N II $\lambda$1083, Si III $\lambda$1206, and Fe III $\lambda$1122 in the HVC near $+125$ km s-1 in the FUSE and STIS spectra (see Figure 3). Our absorption-line measurements of these lines are given in Table 2. Achieving a simultaneous Voigt profile fit using all Si lines is hindered by difficulties in continuum placement for the stronger lines since large spans of the variable stellar continuum are absorbed. Instead, we adopt the Voigt profile measurement log $N_{\mathrm{SiII}}$ = 12.94$\pm$0.07 for the weakest unblended line available at 1526 Å. We show that it is a good fit for $\lambda\lambda$1190, 1193, 1304 in Figure 3. Although detected, N II 1083 lies in a portion of the FUSE spectrum where the stellar continuum is highly variable over a short range in wavelength. Comparing the neighboring low-velocity H2 $J$3 1084 Å line to other $J$3 lines in regions with smooth continua reveals that the continuum must drop significantly in this region, making it difficult to estimate log $N_{\mathrm{NII}}$. For this reason, our measurement for N II is an upper limit. Fe II 1144 has a low absorption profile in the vicinity of the HVC. We applied a high, middle, and low continuum across the absorbing region of $\lambda$1144 and calculate a significance of 2.2$\sigma$ after including a continuum placement error. We find a 3$\sigma$ limiting column density of log $N_{\mathrm{FeII}}$ $\leq$ 12.99, however, given the Fe III detection in the HVC, we use the Fe III column density in metallicity calculations going forward. In addition to the high-velocity absorption we observe for Fe III $\lambda$1122, we detect an intermediate-velocity cloud (IVC) centered near $+$86 km s-1 with log $N$ = 12.98$\pm$0.13. The archival ESO FEROS optical spectrum covers the Ca II K and H and Na I D lines, see Figure 5. Telluric H2O absorption lines are present in the velocity range of the HVC, however, there is a non-detection of high-velocity Na I and we find a 3$\sigma$ limiting column density limit of log $N_{\mathrm{NaI}}<9.45$. We see absorption in Ca II across 6 pixels at 123.7 km s-1 and measure an AOD column density of log $N_{\mathrm{CaII}}$ = 10.68$\pm$0.01 for Ca II 3934. We also performed a simultaneous Voigt profile fit to Ca II $\lambda$3934, 3969 and find log $N$ = 10.63$\pm$0.10. The resulting $b$-value for this small component has a high error. However, since the log $N$ value from Voigt profile fitting agrees with the AOD measurement within the margin of error, we adopt its value. ### 3.5 High ion absorption Figure 6: Velocity profiles of the high-ion absorption lines detected toward HD 156359. The flux is shown in black and the 1$\sigma$ error in the flux is in blue. The dashed black vertical line marks 0 km s-1. The solid green curve is the overall Voigt profile fit to the absorption. The solid magenta curve is the Voigt profile fit to the HVC absorption near $+125$ km s-1 and the vertical dashed magenta line marks the velocity centroid of the fitted component. The pink and yellow curves are fits to components at intermediate velocities. We detected C IV $\lambda\lambda$1548, 1550, Si IV $\lambda\lambda$1393, 1402, and N V $\lambda\lambda$1238, 1242 absorption near $+$130 km s-1 in our STIS E140M spectrum. The C IV and Si IV profiles are complex, showing multiple distinct low and intermediate-velocity components, in addition to the HVC. The high-ion absorption features lie on well-defined continua due to their corresponding broad and smooth stellar P Cygni wind profiles, which gradually elevates their stellar continua in the region from $-100$ to $+200$ km s-1 for this star, as shown in Figure 6. The high-ion continuum fitting was also guided by fits to the GHRS data previously published by Sembach et al. (1991, 1995). We performed Voigt-profile fitting to all high ions with the initial positions of the components and $b$-values determined from the C IV 1548 Å line. The position and $b$-values were allowed to vary freely and the results for the fits are given in Table 2. We note the detection of an IVC centered near $+$78 km s-1 in both C IV and Si IV with log $N$ = $13.57\pm 0.08$ and $13.09\pm 0.10$, respectively. We observe high-velocity absorption in O VI 1031 and 1037 in the LiF1 channel of the FUSE spectrum at $v_{\mathrm{LSR}}$ = 141 km s-1. However, we only include $\lambda$1031 in our analysis, because $\lambda$1037 lies on the steep blueward side of the O VI P Cygni profile and suffers from significant blending with C II∗ 1037 and H2 $J$1 1038 Å. The wavelength region around O VI 1031 Å contains several absorption lines which can serve as contaminants, most notably HD 6–0 $R$(0) $\lambda$1031.915, Cl I $\lambda$1031.507, and several lines of molecular hydrogen including H2 6–0 $P$(3) $\lambda$1031.192 and H2 6–0 $R$(4) $\lambda$1032.351. To gauge the effect of contamination by HD 6–0 $R$(0) $\lambda$1031.915, we examined other HD lines of similar oscillator strength that are isolated from interstellar absorption, such as 5–0 $R$(0) $\lambda$1042.850, 7–0 $R$(0) $\lambda$1021.460, and 8–0 $R$(0) $\lambda$1011.461. We detect no HD molecular absorption in these lines and conclude that no subtraction of the HD 6–0 $R$(0) line at $\lambda$1031.915 from the O VI profile is necessary. For Cl I, a small amount of absorption in Cl I $\lambda$1347 is present near 0 km s-1 in the STIS spectrum and we simultaneously fit Cl I $\lambda$1031 to derive its contribution to the O VI profile. For molecular hydrogen, H2 6–0 $R$(4) $\lambda$1032.351 is the most relevant potential contaminant as it overlaps with the high-velocity component in O VI $\lambda$ 1031\. We modeled other isolated H2 $J4$ lines in the FUSE spectrum, e.g. 5–0 $R$(4) $\lambda$1044.543 and 4–0 $R$(4) $\lambda$1057.381 (see description in Appendix B). Through simultaneous fitting of those lines with H2 6–0 $R$(4) $\lambda$1032 , we account for the small amount of H2 present in the O VI high-velocity absorption. A similar procedure was followed to account for contamination from H2 6–0 $P$(3) $\lambda$1031.192 near $v_{\mathrm{LSR}}$ = $-$200 km s-1 using the isolated H2 $J3$ lines 6–0 $R$(3) $\lambda$1028.986 and 5–0 $P$(3) $\lambda$1043.503. The resulting Voigt profile fit and O VI column density are shown in Figure 6 and Table 2. ## 4 Results: The Cool Low-ion Gas The neutral and low ionization species including H I, O I, C II, N II, Si II, Ca II, Fe II, Si III, and Fe III trace the cool photoionized phase of the HVC detected at $v_{\mathrm{LSR}}$ = +125 km s-1 toward HD 156359 (catalog ). In this section we use the ratios of metal column densities to H I column densities to derive the ion abundances for each observed species in the HVC, as shown in Table 3. The elemental abundances are then derived from the ion abundances using ionization corrections derived from custom photoionization modeling. Finally a comparison of the relative abundances of different elements is used to derive the dust depletion pattern in the HVC. Table 1: Measured Absorption Lines Ion | $\lambda^{a}$ | $f$ | Dataset ---|---|---|--- | (Å) | | H I | 923.150 | 0.0022b | FUSE SiC2A | 926.226 | 0.0032b | FUSE SiC2A | 930.748 | 0.0048b | FUSE SiC2A | 937.804 | 0.0078b | FUSE SiC2A C II | 1334.532 | 0.1290c | STIS E140M C IV | 1548.202 | 0.1900d | STIS E140M | 1550.774 | 0.0948d | STIS E140M N II | 1083.990 | 0.1110c | FUSE SiC2B N V | 1238.821 | 0.1560e | STIS E140M | 1242.804 | 0.0777e | STIS E140M O I | 1302.168 | 0.0480c | STIS E140M O VI | 1031.912 | 0.1330e | FUSE LiF1 Na I | 5891.583 | 0.6500g | FEROS Si II | 1190.420 | 0.2560f | STIS E140M | 1193.280 | 0.5440f | STIS E140M | 1304.370 | 0.0910f | STIS E140M | 1526.720 | 0.1440f | STIS E140M Si III | 1206.510 | 1.6100g | STIS E140M Si IV | 1393.760 | 0.5130g | STIS E140M | 1402.770 | 0.2540g | STIS E140M Ca II | 3934.770 | 0.6490h | FEROS | 3969.590 | 0.3210h | FEROS Fe II | 1144.926 | 0.0830i | FUSE LiF2 Fe III | 1122.524 | 0.0642j | FUSE LiF2 Note. — a All wavelengths are taken from Morton (2003). b Pal’chikov (1998), c Froese Fischer & Tachiev (2004), d Yan et al. (1998), e Peach et al. (1988), f Bautista et al. (2009), g Froese Fischer et al. (2006), h Safronova & Safronova (2011) i Donnelly & Hibbert (2001), j Deb & Hibbert (2009) Table 2: HD 156359 High-velocity absorption Ion | $v_{\mathrm{LSR}}$ | $b$ | log $N$ ---|---|---|--- | (km s-1) | (km s-1) | ($N$ in cm-2) H I | 125.6 $\pm$ 0.9 | 18.9 $\pm$ 1.2 | 15.54 $\pm$ 0.05 O I | 125.0 $\pm$ 4.8 | … | 12.43 $\pm$ 0.11a Na I | … | … | $<$ 9.45b C II | 120.9 $\pm$ 0.6 | 9.8 $\pm$ 1.4 | 13.83 $\pm$ 0.10 N II | … | … | $<$ 13.98b Si II | 118.5 $\pm$ 1.0 | 7.0 $\pm$ 1.6 | 12.94 $\pm$ 0.07 Ca II | 123.7 $\pm$ 1.0 | 2.3 $\pm$ 2.7 | 10.63 $\pm$ 0.10 Fe II | … | … | $<$ 12.99b Si III | 117.6 $\pm$ 1.4 | 6.9 $\pm$ 2.8 | 12.32 $\pm$ 0.16 Fe III | 129.3 $\pm$ 2.3 | 11.0 $\pm$ 3.6 | 12.82 $\pm$ 0.14 C IV | 133.3 $\pm$ 2.3 | 8.6 $\pm$ 3.6 | 12.60 $\pm$ 0.13 Si IV | 129.5 $\pm$ 1.5 | 5.7 $\pm$ 2.9 | 11.93 $\pm$ 0.15 N V | 130.2 $\pm$ 2.7 | 12.2 $\pm$ 4.0 | 12.64 $\pm$ 0.11 O VI | 141.3 $\pm$ 1.2 | 24.0 $\pm$ 1.8 | 13.65 $\pm$ 0.03 Note. — a Except where noted, all measurements presented in this table are derived from Voigt-profile fitting, with the exception of O I, for which we derive an apparent column density from the profile in the range $+110\lesssim v\lesssim+125$ km s-1, see Figure 4. The error listed for log $N$(O I) is the combination of the statistical error and a continuum placement error in quadrature, and is $\pm$ 0.08 $\pm$ 0.07 separately. b This measurement is a non-detection and the 3$\sigma$ limiting column density is given. ### 4.1 Photoionization Modeling To model the ionization breakdown in the HVC and characterize its physical conditions, we ran a multi-dimensional grid of Cloudy (v. 17.02, Ferland et al. 2017) photoionization models for the low- and intermediate-ions, assuming they arise in the same gas phase as the H I and are photoionized by the same incident radiation field. Our model assumes a plane-parallel slab of uniform density exposed to the magnitude of the escaping UV ionizing flux of the Milky Way (Fox et al., 2005; Barger et al., 2013; Fox et al., 2014) at the location of HD 156359 (catalog ) and includes the extragalactic UV background radiation field from Khaire & Srianand (2019). We ran our models for a grid of metallicity values with [Z/H] varying from $-0.5$ to $+1.5$ in steps of 0.1 dex, each over a range of hydrogen number densities log ($n_{\mathrm{H}}$/cm-3) from $-3$ to 0, in order to explore possible metallicities and densities. We determined the best-fit log $n_{\mathrm{H}}$ for each metallicity model by matching the observed column density ratio of Si III/Si II to the model value (see the top panel of Figure 7). The Si III/Si II ratio was chosen because both ions have unsaturated detections in the HVC, and using a ratio of adjacent ions from the same element (Si) minimizes metallicity or depletion effects from different metal ions. Using the pairs of metallicity and log $n_{\mathrm{H}}$ determined from the Si III/Si II ratio, we narrow the range of metallicities by identifying which models are consistent with the observed log $N_{\mathrm{OI}}$ = 12.43 $\pm$ 0.11, as illustrated in the bottom left panel of Figure 7. We then run an even finer grid of metallicity values in increments of 0.02 dex over the narrowed metallicity and number density region to determine the range of metallicities and densities allowed by the data. We find that metallicities in the range $0.18\leq\mathrm{[O/H]}\leq 0.51$ and densities from $-1.69\leq$ log ($n_{\mathrm{H}}$/cm-3) $\leq-1.37$ are consistent with the data, giving a model best-fit of [O/H] = 0.36${}^{+0.15}_{-0.18}$ at log $n_{\mathrm{H}}$ = $-$1.53$\pm$-0.16. We repeated this procedure for C II (see bottom-right panel of Figure 7), which is expected to be weakly depleted (Jenkins, 2009), and find that the range 0.21 $\leq$ [C/H] $\leq$ 0.42 overlaps with the observed data, giving a model best-fit of [C/H] = 0.30${}^{+0.12}_{-0.09}$ at log $n_{\mathrm{H}}$ = $-$1.49${}^{+0.09}_{-0.10}$. The agreement between the oxygen-based metallicity and the carbon-based metallicity lends support to our methodology and to the robustness of the super-solar metallicity we infer. The metallicity for this cloud ranks among the highest UV-based metallicities of any HVC observed thus far, along with the HVC at $-$125 km s-1 toward M5-ZNG ($l$ = 3.$\degree$9, $b$ = $+$47.$\degree$7 at $z$ = $+$5.3 kpc) with [O/H] = $+$0.22 $\pm$ 0.10 (Zech et al., 2008). M5-ZNG was also observed with FUSE and STIS E140M spectra. Their observed metallicity is not corrected for ionization effects, but is likely higher than reported, given the measured low log $N_{\mathrm{HI}}$ = 16.50 $\pm$ 0.06 in the cloud and that oxygen shows a positive ionization correction when log $N$(H I) $<$ 18.5 (Bordoloi et al., 2017). The metallicity of the HD 156359 (catalog ) HVC is also on the high end of the range of $<$20% to 3 times solar reported by Ashley et al. (2022) in their survey of metallicities of Fermi Bubble HVCs. We note that a super-solar abundance in the inner Galaxy may not be unexpected, given the oxygen abundance gradients reported from emission line measurements in H II regions (see Wenger et al. 2019; Arellano-Córdova et al. 2021). Table 3: HVC elemental abundances and depletions Ion | [Xi/H I]a | IC(Xi)b | [X/H]c | $\delta_{\mathrm{O}}$(X)d ---|---|---|---|--- O I | 0.20$\pm$0.10 | $+$0.16${}^{+0.04}_{-0.07}$ | 0.36${}^{+0.11}_{-0.12}$ | 0 C II | 1.86$\pm$0.12 | $-$1.59$\pm$0.04 | 0.27$\pm$0.13 | $-$0.09$\pm$0.17 N II | $<$ 2.60 | $<-$1.67 | $<$ 0.93 | $<$ 0.57 Si II | 1.89$\pm$0.08 | $-$1.86$\pm$0.03 | 0.03$\pm$0.09 | $-$0.33$\pm$0.14 Fe II | $<$ 1.95 | $-$1.17${}^{+0.20}_{-0.12}$ | $<$ 0.79 | $<$ 0.43 Fe III | 1.78$\pm$0.15 | $-$1.73$\pm$0.08 | 0.06$\pm$0.17 | $-$0.30$\pm$0.20 Ca II | 0.75$\pm$0.11 | $-$0.95$\pm$0.01 | $-$0.20$\pm$0.11 | $-$0.56$\pm$0.16 Note. — a Ion abundance [Xi/H I] = log (Xi/H I)HVC $-$ log (X/H)⊙ where Xi is the observed ion of element X. These are not corrected for ionization effects. The error includes both systematic and continuum placement uncertainties. b Ionization correction IC(Xi) = [X/H]model $-$ [Xi/H I], derived from the difference between the model elemental abundance and the observed ion abundance. c Ionization-corrected elemental abundance [X/H] = [Xi/H I] + IC(Xi). Also referred to as gas-phase abundance. d $\delta_{\mathrm{O}}$(X) $\equiv$ [X/O] = [X/H] $-$ [O/H] is the depletion of element X relative to oxygen. Figure 7: Cloudy photoionization models of the HVC toward HD 156359 (catalog ) that explore a parameter space in metallicity and number density, $-0.50\leq$ [Z/H] $\leq+1.50$ and $-3\leq$ log ($n_{\mathrm{H}}$/cm-3) $\leq 0$, respectively. Top panel: Model results for log $N_{\mathrm{SiIII}}$/$N_{\mathrm{SiII}}$ for the set metallicities are shown as solid colored curves. The dashed horizontal turquoise line is the observed ratio of $-$0.62$\pm$0.17. The best-fit log $n_{\mathrm{H}}$ for each metallicity model is determined by the intersection of the model and observed ratio and is marked by vertical colored lines and circle markers. Bottom left panel: The model curves for log $N_{\mathrm{OI}}$ for each metallicity are plotted against log $n_{\mathrm{H}}$, where the values derived from log $N_{\mathrm{SiIII}}$/$N_{\mathrm{SiII}}$ are marked with vertical lines and markers. The horizontal turquoise line and bar show the observed log $N_{\mathrm{OI}}$ = 12.43 $\pm$ 0.11. The model data corresponding to the metallicity region from $+0.18\leq$[Z/H] $\leq+0.51$ and $-1.69\leq$ log $n_{\mathrm{H}}$ $\leq-1.37$ overlap with the observed data. The model best- fit values are [Z/H] = $+$0.36${}^{+0.15}_{-0.18}$ and log $n_{\mathrm{H}}$ = $-1.53\pm 0.16$. Bottom right panel: Same as for the bottom left panel, but for C II. The horizontal turquoise line and bar show observed log $N_{\mathrm{CII}}$ = 13.83 $\pm$ 0.10. The model data corresponding to the metallicity region from $+0.21\leq$ [Z/H] $\leq+0.42$ and $-1.59\leq$ log $n_{\mathrm{H}}$ $\leq-1.40$ overlap with the observed data. The model best fit values are [Z/H]= $+$0.30${}^{+0.12}_{-0.09}$ and log $n_{\mathrm{H}}$ = $-1.49^{+0.09}_{-0.10}$. Figure 8: Top panel: ionization corrections for the low ions O I, C II, N II, Si II, Fe III, and Ca II (colored curves) plotted against log $n_{\mathrm{H}}$ from our Cloudy model to the HD 156359 (catalog ) HVC. The model uses our best-fit [Z/H]=$+$0.36. The vertical line marks the best-fit log ($n_{\rm H}$/cm-3)=$-$1.53, determined from the Cloudy models for O I (see Figure 7). The ionization correction calculated at $-$1.53 is added to the observed ion abundance to determine the elemental abundance. Middle panel: comparison of the ionization-corrected abundances of the low ions detected in the HVC, where the solar abundance is plotted as a horizontal line. Bottom panel: Comparison of the dust depletion levels $\delta$(X) = [X/O] in the HVC with the depletion pattern for [X/O] measured for sight lines with the lowest and highest collective depletions ($F_{*}$=0 and $F_{*}$=1, respectively) from Jenkins (2009). A slight offset is applied in the $x$-direction of each element for distinction. Figure 9: Our final Cloudy photoionization model of the low- and intermediate ions detected in the HVC toward HD 156359 near $+125$ km s-1. The model uses the best-fit [Z/H] = $+$0.36 determined from our O I analysis. These model predictions for each ion (colored dotted curves) have been scaled by the dust depletions required to match the observed values, where the observed column densities are indicated by circle markers at the best-fit log ($n_{\mathrm{H}}$/cm-3)=$-$1.53 (black vertical line). ### 4.2 Ionization corrections In ISM abundance work, [O I/H I] is often considered a robust indicator of elemental abundance [O/H], since (1) charge-exchange reactions closely couple the ionization state of hydrogen and oxygen together Field & Steigman (1971), and (2) oxygen is only lightly depleted onto dust grains (Jenkins, 2009). However, the assumption that [O I/H I]=[O/H] breaks down when $N$(H I) is so low that the gas is optically thin or when the ionizing photon flux is extremely high (Viegas, 1995). In the HVC at +125 km s-1 toward HD 156359 (catalog ), log $N$(H I) is only 15.54$\pm$0.05, so an ionization correction must be made to account for the column densities of all observed neutral and ion stages. Our Cloudy models directly provide the ionization corrections for all our observed ion stages. We define the ionization correction as the difference between the model (true) elemental abundance and the observed ion abundance, i.e., $IC(X^{i})=[\text{X/H}]-[\text{X}^{i}/\textsc{H i}].$ (1) The resulting ion abundances, ionization corrections, and ionization-corrected elemental abundances are given in Table 3. The relationships between the ionization corrections and the hydrogen number density for [Z/H] = 0.36 are shown in the top panel of Figure 8 and the ionization-corrected abundances are shown in the middle panel of Figure 8. The positive ionization correction of $IC$(O I) = $0.16^{+0.04}_{-0.07}$ from the photoionization modeling results in a high gas-phase oxygen abundance of $+0.36$. We calculate gas-phase abundances of [C/H] = 0.27$\pm$0.13, [N/H] = $\leq$ 0.93, [Si/H] = 0.03$\pm$0.09, [Fe/H] = 0.06$\pm$0.17, and [Ca/H] = $-$0.20$\pm$0.11 (see Table 3). ### 4.3 Dust depletion effects Following convention, we define the depletion $\delta_{\mathrm{O}}$(X) of each refractory element X as the ionization-corrected abundances of that element compared to the ionization-corrected oxygen abundance, i.e., $\delta_{\mathrm{O}}(\text{X})\equiv[\text{X/O}]=[\text{X/H}]-[\text{O/H}],$ (2) where oxygen represents an undepleted (or lightly depleted) volatile element. In this formalism, a negative value of $\delta_{\mathrm{O}}$(X) means that element X is depleted relative to oxygen. This method assumes that the total (gas+dust) abundances are solar, which is often assumed to apply to the Galactic ISM (Savage & Sembach, 1996), though local ISM abundance variations cannot be ruled out in some sight lines (De Cia et al., 2021). We show $\delta_{\mathrm{O}}$(X) for the low ions in the bottom panel of Figure 8 compared to $F_{*}$, the line-of-sight depletion strength factor, for [X/O] determined for low-depletion ($F_{*}$=0) and high-depletion ($F_{*}$=1) gas from the comprehensive ISM gas-phase element depletions study of Jenkins (2009). For the HVC toward HD 156359 (catalog ), we find a low value of $\delta_{\mathrm{O}}$(C)=$-$0.09$\pm$0.17, i.e. carbon shows no significant depletion. This is consistent with the low values of [C/O] measured in Galactic ISM gas (Jenkins, 2009). For the other low ions, we find a 3$\sigma$ upper limit for the nitrogen depletion $\delta_{\mathrm{O}}$(N)$\leq+0.57$ and low depletions for silicon, iron, and calcium of $\delta_{\mathrm{O}}$(Si)=$-$0.33$\pm$0.14$,\delta_{\mathrm{O}}$(Fe)=$-$0.30$\pm$0.20, and $\delta_{\mathrm{O}}$(Ca)=$-$0.56$\pm$0.16, respectively. To summarize the final results of the CLOUDY modeling after the dust depletion levels have been derived, Figure 9 shows the model curves of the detected ions shifted by their respective depletions at [X/H] = $+$0.36 and log $n_{\mathrm{H}}$= $-$1.53. The ability of this model to match the observed column densities shows that all low-ion measurements can be explained by photoionization once dust depletion effects are accounted for. A low level of dust depletion for the HVC is consistent with the well-known Routly-Spitzer effect (RS effect), in which the amount of dust observed in high-velocity clouds decreases significantly at higher LSR velocity (see Routly & Spitzer 1952; Siluk & Silk 1974; Vallerga et al. 1993; Smoker et al. 2011; Ben Bekhti et al. 2012; Murga et al. 2015). The RS effect is typically traced by the $N$(Na I)/$N$(Ca II) ratio, which has been found to significantly decrease with increasing LSR velocity. Although the RS effect was historically interpreted as a dust depletion effect, it may also be related to ionization effects, and likely differs depending on the environment of the cloud. We measure an 3$\sigma$ upper limit of log $N$(Na I)/$N$(Ca II) $\leq$ $-$1.2 in the $+124$ km s-1 HVC toward HD 156359 (catalog ) from the optical FEROS data, which is on the low end of the observed ratios for HVCs (Smoker et al., 2011; Ben Bekhti et al., 2012; Murga et al., 2015), and is also lower than is typically observed at low velocities. In summary, the dust depletion pattern we derive in the HVC is consistent with other HVCs and with the Routly-Spitzer effect, even though the cloud has high metallicity. ## 5 Results: the Highly Ionized Gas The HVC toward HD 156359 (catalog ) shows high-ion absorption in the resonance doublets of Si IV, C IV, N V, and O VI at a velocity range of 130–140 km s-1 (Figure 6; Table 2), slightly higher than the velocity range of the low ions, which are centered near 125 km s-1. To compare the high-ion absorption between different ions, we show in Figure 10 a comparison of the normalized optical depth profiles of each high ion. These plots provide a way to inter-compare the profiles of different ions in a linear manner (e.g., Fox et al. 2003). The HVC line profiles of C IV and Si IV are well aligned in velocity, but are offset from the centroid of O VI absorption by $\approx$8 km s-1. This suggests that the O VI exists in a separate (likely hotter) gas phase. The STIS profile of N V is too noisy to draw any conclusions about line centroid. However, the GHRS N V profiles (Sembach et al., 1995) are less noisy and have a similar profile shape to the O VI seen in the FUSE data, supporting the placement of O VI and N V in the same phase. ### 5.1 Collisional Ionization Modeling Figure 10: Optical depth profiles of the high ions, where each profile is arbitrarily normalized to its peak value for the velocity region shown in order to facilitate comparison. Top panel: C IV $\lambda$1548 and Si IV $\lambda$1393 display similar line shapes in the velocity region of the HVC from $112\lesssim v\lesssim 140$ km s-1, indicated with cyan shading. Middle panel: C IV $\lambda$1548 and O VI $\lambda$1031 are compared, where the center of the HVC absorption for O VI is shifted to a higher velocity at $v\approx 141$ km s-1. Bottom panel: N V $\lambda$1238 and O VI $\lambda$1031 are compared, where the range of velocities attribute to absorption by O VI are indicated in light pink. Figure 11: Comparison of observed high-ion column-density ratios with predictions from the collisional ionization models of Gnat & Sternberg (2007). The observed ratios in the HD 156359 (catalog ) HVC are shown as dashed orange horizontal lines. The top and bottom panels show the C IV/Si IV and O VI/N V ratios, respectively. In both panels, the solid black curve is the collisional ionization equilibrium model ion fraction. The solid green and violet curves are the time-dependent isochoric and isobaric collisional ionization models for a solar metallicity, respectively. Their dashed counterparts are the same but for a metallicity twice the solar value. The vertical dashed blue line is the temperature of the gas determined by the models. The vertical blue band in the top panel represents a second time-dependent solution for a range of temperatures from log $T$ = 4.2–4.4 K. The observed high ions have column densities that are too high to be explained by the Cloudy photoionization models described in Section 4.1. Where the Si IV and C IV column densities differ by $\sim$0.7 and 1.0 dex respectively, the N V and O VI differ by orders of magnitudes. These discrepancies imply that a separate ionization mechanism is required for the high ions. We do not account for energetic photons from the radiation field intrinsic to the Fermi Bubbles, however, we explore collisional ionization as a separate mechanism for gas with temperatures of $T\gtrsim$ 105 K. We consult the collisional ionization models from Gnat & Sternberg (2007) to determine the exact range of temperatures allowed by the high-ion column densities and their ratios. While we considered the more recent models from Gnat 2017, which include photoionization effects from the extragalactic UV background, we do not adopt them because HD 156359 lies in close proximity to the radiation field of the Galactic plane which is much stronger. We considered both the collisional ionization equilibrium and non-equilibrium regimes. A single-temperature solution explaining all four high ions (Si IV, C IV, N V, O VI) in the HVC is ruled out, as no such solution exists to the observed high-ion column densities. Instead, we find that a two-phase solution is needed, with one temperature explaining the $N_{\mathrm{CIV}}$/$N_{\mathrm{SiIV}}$ ratio and another explaining $N_{\mathrm{OVI}}$/$N_{\mathrm{NV}}$. This two-phase model is consistent with the kinematic information in the UV spectra, where Si IV and C IV show very similar line profiles but O VI is offset in velocity centroid. The two-phase high-ion model is illustrated in Figure 11, which shows the observed high-ion ratios of $N_{\mathrm{CIV}}$/$N_{\mathrm{SiIV}}$ and $N_{\mathrm{OVI}}$/$N_{\mathrm{NV}}$ compared to model predictions from Gnat & Sternberg (2007) for solar ([Z/H]=0) and super-solar ([Z/H]=$+$0.3) metallicities. For the C IV/Si IV phase, we find two possible solutions for the temperature. First, the non-equilibrium isochoric (constant volume) and isobaric (constant pressure) models give a low-temperature solution at $T$ = $10^{4.2-4.4}$ K, where the lower end corresponds to the solar-metallicity isochoric model and the higher end with the super-solar isobaric model. Second, the collisional ionization equilibrium (CIE) model returns a higher temperature $T$=104.9 K. We are inclined to adopt the non-equilibrium (lower-temperature) solution, as plasma near $T$=105 K is at the peak of the cooling curve, where oxygen dominates the radiative cooling and the gas can cool faster than it recombines, reaching a non-equilibrium state. For the O VI/N V phase, a single temperature solution for the observed log $N_{\mathrm{OVI}}$/$N_{\mathrm{NV}}$=1.01 is found at $T$=105.4 K for all models (see bottom panel of Figure 11). In conclusion, we can successfully model the high-ion plasma in the HVC toward HD 156359 (catalog ) as containing collisionally-ionized gas at two temperatures: a cooler phase seen in C IV and Si IV at $T$ = $10^{4.2-4.4}$ K, and a hotter phase seen in N V and O VI at $T$=105.4 K. ## 6 Summary Using archival FUSE, HST STIS, and ESO FEROS spectra, we have analyzed the chemical composition of the HVC near $+$125 km s-1 toward HD 156359 (catalog ), a massive star lying 9 kpc away toward the Galactic Center. The sight line passes less than 1$\degree$ from one of the densest cores of a complex of small HVCs dubbed “Complex WE” by Wakker & van Woerden (1991), as shown in Figure 1. Furthermore, the sight line passes through a region of enhanced X-ray emission (the southern eROSITA Bubble; Predehl et al., 2020); this region indicates energetic feedback from the Galactic Center. Our main results are as follows. 1. 1. We determined an H I column density measurement of log $N$(H I) = $15.54\pm 0.05$ in the HVC using unsaturated Lyman series absorption lines. 2. 2. We measured a super-solar O I abundance of [O I/H I] = 0.20$\pm$0.11 in the HVC. After applying an ionization correction, we derive an oxygen abundance of [O/H] = 0.36$\pm$0.12, indicating the cloud is enriched to 2.3 times the solar level. This abundance is consistent with super solar O abundances of H II regions measured in the inner Galaxy (Wenger et al., 2019; Arellano-Córdova et al., 2021). 3. 3. The ionization-corrected carbon abundance [C/H] = 0.27$\pm$0.13 is consistent with the oxygen abundance. This indicates that carbon shows low dust depletion relative to oxygen, consistent with the pattern seen by Jenkins (2009), as shown in Figure 8. 4. 4. A low level of depletion is inferred from the silicon, iron, and calcium lines, with [Si/O] = $-$0.33 $\pm$ 0.14, [Fe/O] = $-$0.30 $\pm$ 0.20, and [Ca/O] = $-$0.56 $\pm$ 0.16. The Na I/Ca II ratio in the HVC is $\leq-1.2$, which is lower than is typically observed at low velocities, consistent with the trend known as the Routly-Spitzer effect. 5. 5. We detect high ion species C IV, Si IV, N V, and O VI near $+$130 km s-1, at slightly higher velocities than the velocity range of the lower ions. We find that an ionization mechanism separate from photoionization, such as collisional ionization, is required to explain the column densities of the high ions. We determine that a two-phase temperature solution best explains the observed $N_{\mathrm{CIV}}$/$N_{\mathrm{SiIV}}$ and $N_{\mathrm{OVI}}$/$N_{\mathrm{NV}}$ ratios, with a cooler phase seen in C IV and Si IV at $T$ = $10^{4.2-4.4}$ K, and a hotter phase seen in N V and O VI at $T$=105.4 K. The high metallicity, low depletion, complex high-ion absorption, and high positive velocity of the HVC toward HD 156359 (catalog ) are all consistent with a wind origin, in which a swept-up clump of material is being carried out from the Galactic disk into the halo. As such, this HVC may represent a freshly entrained cool clump of gas caught in the act of being accelerated into the nuclear wind. While we cannot rule out a foreground origin for the HVC, in which the cloud exists at an anomalous velocity somewhere between the Sun and the Galactic Center, we can exclude a spiral-arm explanation for the HVC on kinematic grounds, because the cloud’s $+125$ km s-1 velocity lies over 100 km s-1 away from the nearest spiral arm, the Sagittarius Arm at $\sim-10$ km s-1. Under a biconical outflow model (Fox et al. 2015; Bordoloi et al. 2017; Di Teodoro et al. 2018), the cloud’s radial velocity and its location only 2.3 kpc below the disk imply a short timescale of $\sim$5 Myr for the HVC to have reached its current position, which is much shorter than the timescales on which chemical mixing is expected to be significant (10s to 100s of Myr; Gritton et al., 2014; Heitsch et al., 2022). Our observations therefore provide a snapshot into the chemical and physical conditions prevailing in this early stage of a nuclear outflow, before chemical mixing has diluted or enriched the gas from its initial state. We gratefully acknowledge the invaluable contributions to early versions of this manuscript from the late Blair Savage, who passed away during the manuscript’s preparation. This paper would not have been possible without Blair’s foundational work on the HD 156359 (catalog ) sight line, the chemical abundances in the ISM, apparent optical depth analysis, and the inner Galaxy. We gratefully acknowledge support from the NASA Astrophysics Data Analysis Program (ADAP) under grant 80NSSC20K0435, 3D Structure of the ISM toward the Galactic Center. The FUSE data were obtained under program P101. FUSE was operated for NASA by the Department of Physics and Astronomy at the Johns Hopkins University. D.K. is supported by an NSF Astronomy and Astrophysics Postdoctoral Fellowship under award AST-2102490. We thank Isabel Rebollido for valuable conversations on the FEROS spectrograph. The FUSE and HST STIS data presented in this paper were obtained from the Mikulski Archive for Space Telescopes (MAST) at the Space Telescope Science Institute. The ESO FEROS spectrum was obtained from the ESO Archive Science Portal. The FUSE and HST STIS data presented in this paper were obtained from the Mikulski Archive for Space Telescopes (MAST) at the Space Telescope Science Institute. The specific observations analyzed can be accessed via http://dx.doi.org/10.17909/rrnk-3e58 (catalog 10.17909/rrnk-3e58). ## Appendix A Global Continuum Fitting Stellar absorption lines present a continuum-fitting challenge, particularly in the complex far-UV FUSE spectra used to derive the H I column density in the HVC under study. Our approach followed in the analysis is to fit the continuum _locally_ around each H I Lyman series line of interest, since this allows us to account for stellar absorption lines, which are present in the continuum when the radiation field encounters the HVC. However, for completeness here we consider a _global_ continuum fitting process, which fits the H I continuum over a larger range (916–944 Å) in the SiC2 channel. This approach neglects stellar absorption features but ensures the continua are continuous between adjacent lines in the Lyman series. Wavelength regions free from stellar emission and absorption lines defined the flux of the global continuum. We show a global fit to the stellar flux in the vicinity of the higher order H I Lyman series lines in Figure 12, in which areas of emission and absorption due to both stellar and interstellar features can be seen. We performed a simultaneous Voigt profile fit on the same H I lines modeled with the global continua, using the mid-, high-, and low-continuum fits shown in Figure 12. A separate local continuum was applied to the depressed region resulting from stellar line blanketing and which contains H I $\lambda$923\. Using the global continuum fit (which has a slightly higher continuum placement than the local fits) results in log $N_{\mathrm{HI}}$ = 15.69$\pm$0.03$\pm$ 0.03 for the HVC, where the first error is the statistical error due to photon noise and the second error is the systematic continuum- placement error. This corresponds to a moderate difference of $\Delta$log $N_{\mathrm{HI}}$ = 0.15 dex between the H I column densities derived by the global and local continuum methods. Figure 12: A portion of the FUSE SiC2 spectrum from 916–944 Å in which several higher order H I Lyman series lines are identified. The flux and 1$\sigma$ error in the flux are shown in black and blue, respectively. A global fit to the stellar continuum is traced by a solid red curve, with a high and low stellar continuum placement marked by dashed red lines. The orange box shows a local fit to a depressed portion of the stellar continuum created by stellar line blanketing, from $\sim$920–925 Å. ## Appendix B Decontamination of Molecular Hydrogen The FUSE LiF1 and LiF2 channels provide us with an opportunity to model isolated H2 lines. This effort is important because it enables us to account for the Milky Way’s molecular contribution to the interstellar absorption lines. In particular, contamination from Milky Way H2 affected the higher order Lyman series H I in the SiC2A channel and O VI in the LiF1 channel. We performed simultaneous single-component Voigt profile fits to multiple H2 transitions from rotational levels $J$ = 0–6, allowing both the velocity and $b$-value to vary freely for the stronger $J$ = 0–4 transitions. The $b$-values of the significantly weaker $J$ = 5,6 transitions were fixed to the $b$-value of the $J$ = 4 transition in order to deter unrealistic column densities for these fainter lines. These results of these fittings are shown in Figure 13. Figure 13: We performed Voigt profile fits to vibrationally and rotationally excited H2 lines from the FUSE LiF1 and LiF2 channels in order to account for molecular blending with H I. In each panel, the flux and 1$\sigma$ error in the flux are shown in black and blue, respectively. Voigt profile fits to the data are shown in green. Green vertical lines mark the positions of the H2 absorption lines. Top panel: LiF1 channel H2 transitions from upper vibrational levels 4 and 5, spanning rotational levels $J=0-6$. Fits to Ar I $\lambda$1048 and Fe II $\lambda$1055 are included. Bottom left panel: LiF1 channel H2 transitions from upper vibrational levels 2 and 3, including rotational levels $J=0,1,2,5,6$. Bottom right panel: LiF2 channel H2 transitions from upper vibrational levels 1 and 2, including rotational levels $J=0,1,2,5$. ## References * Ackermann et al. (2014) Ackermann, M., Albert, A., Atwood, W. B., et al. 2014, ApJ, 793, 64, doi: 10.1088/0004-637X/793/1/64 * Arellano-Córdova et al. (2021) Arellano-Córdova, K. Z., Esteban, C., García-Rojas, J., & Méndez-Delgado, J. E. 2021, MNRAS, 502, 225, doi: 10.1093/mnras/staa3903 * Ashley et al. (2022) Ashley, T., Fox, A. J., Cashman, F. H., et al. 2022, arXiv e-prints, arXiv:2207.08838. https://arxiv.org/abs/2207.08838 * Ashley et al. (2020) Ashley, T., Fox, A. J., Jenkins, E. B., et al. 2020, ApJ, 898, 128, doi: 10.3847/1538-4357/ab9ff8 * Astropy Collaboration et al. (2018) Astropy Collaboration, Price-Whelan, A. M., Sipőcz, B. M., et al. 2018, AJ, 156, 123, doi: 10.3847/1538-3881/aabc4f * Bailer-Jones et al. (2021) Bailer-Jones, C. A. L., Rybizki, J., Fouesneau, M., Demleitner, M., & Andrae, R. 2021, AJ, 161, 147, doi: 10.3847/1538-3881/abd806 * Barger et al. (2013) Barger, K. A., Haffner, L. M., & Bland-Hawthorn, J. 2013, ApJ, 771, 132, doi: 10.1088/0004-637X/771/2/132 * Bautista et al. (2009) Bautista, M. A., Quinet, P., Palmeri, P., et al. 2009, A&A, 508, 1527, doi: 10.1051/0004-6361/200913179 * Ben Bekhti et al. (2012) Ben Bekhti, N., Winkel, B., Richter, P., et al. 2012, A&A, 542, A110, doi: 10.1051/0004-6361/201118673 * Bland-Hawthorn & Cohen (2003) Bland-Hawthorn, J., & Cohen, M. 2003, ApJ, 582, 246, doi: 10.1086/344573 * Bland-Hawthorn et al. (2019) Bland-Hawthorn, J., Maloney, P. R., Sutherland, R., et al. 2019, ApJ, 886, 45, doi: 10.3847/1538-4357/ab44c8 * Bordoloi et al. (2017) Bordoloi, R., Fox, A. J., Lockman, F. J., et al. 2017, ApJ, 834, 191, doi: 10.3847/1538-4357/834/2/191 * Bowen et al. (2008) Bowen, D. V., Jenkins, E. B., Tripp, T. M., et al. 2008, ApJS, 176, 59, doi: 10.1086/524773 * Bregman (1980) Bregman, J. N. 1980, ApJ, 236, 577, doi: 10.1086/157776 * Carswell & Webb (2014) Carswell, R. F., & Webb, J. K. 2014, VPFIT: Voigt profile fitting program, Astrophysics Source Code Library, record ascl:1408.015. http://ascl.net/1408.015 * Cashman et al. (2017) Cashman, F. H., Kulkarni, V. P., Kisielius, R., Ferland, G. J., & Bogdanovich, P. 2017, ApJS, 230, 8, doi: 10.3847/1538-4365/aa6d84 * Cashman et al. (2021) Cashman, F. H., Fox, A. J., Savage, B. D., et al. 2021, ApJ, 923, L11, doi: 10.3847/2041-8213/ac3cbc * de Avillez (1999) de Avillez, M. A. 1999, in Astronomical Society of the Pacific Conference Series, Vol. 166, Stromlo Workshop on High-Velocity Clouds, ed. B. K. Gibson & M. E. Putman, 103. https://arxiv.org/abs/astro-ph/9901056 * De Cia et al. (2021) De Cia, A., Jenkins, E. B., Fox, A. J., et al. 2021, Nature, 597, 206, doi: 10.1038/s41586-021-03780-0 * Deb & Hibbert (2009) Deb, N. C., & Hibbert, A. 2009, Atomic Data and Nuclear Data Tables, 95, 184, doi: 10.1016/j.adt.2008.12.001 * Di Teodoro et al. (2020) Di Teodoro, E. M., McClure-Griffiths, N. M., Lockman, F. J., & Armillotta, L. 2020, Nature, 584, 364, doi: 10.1038/s41586-020-2595-z * Di Teodoro et al. (2018) Di Teodoro, E. M., McClure-Griffiths, N. M., Lockman, F. J., et al. 2018, ApJ, 855, 33, doi: 10.3847/1538-4357/aaad6a * Dixon & Kruk (2009) Dixon, W. V., & Kruk, J. W. 2009, in American Institute of Physics Conference Series, Vol. 1135, Future Directions in Ultraviolet Spectroscopy: A Conference Inspired by the Accomplishments of the Far Ultraviolet Spectroscopic Explorer Mission, ed. M. E. van Steenberg, G. Sonneborn, H. W. Moos, & W. P. Blair, 218–220, doi: 10.1063/1.3154053 * Donnelly & Hibbert (2001) Donnelly, M. P., & Hibbert, A. 2001, MNRAS, 321, 247, doi: 10.1046/j.1365-8711.2001.03996.x * Dressel et al. (2007) Dressel, L., Holfeltz, S., & Quijano, J. K. 2007, in STIS Data Handbook v. 5, Vol. 5, 5 * Ferland et al. (2017) Ferland, G. J., Chatzikos, M., Guzmán, F., et al. 2017, Rev. Mexicana Astron. Astrofis., 53, 385\. https://arxiv.org/abs/1705.10877 * Field & Steigman (1971) Field, G. B., & Steigman, G. 1971, ApJ, 166, 59, doi: 10.1086/150941 * Fitzpatrick & Massa (2007) Fitzpatrick, E. L., & Massa, D. 2007, ApJ, 663, 320, doi: 10.1086/518158 * Fox et al. (2003) Fox, A. J., Savage, B. D., Sembach, K. R., et al. 2003, ApJ, 582, 793, doi: 10.1086/344692 * Fox et al. (2005) Fox, A. J., Wakker, B. P., Savage, B. D., et al. 2005, ApJ, 630, 332, doi: 10.1086/431915 * Fox et al. (2014) Fox, A. J., Wakker, B. P., Barger, K. A., et al. 2014, ApJ, 787, 147, doi: 10.1088/0004-637X/787/2/147 * Fox et al. (2015) Fox, A. J., Bordoloi, R., Savage, B. D., et al. 2015, ApJ, 799, L7, doi: 10.1088/2041-8205/799/1/L7 * French et al. (2021) French, D. M., Fox, A. J., Wakker, B. P., et al. 2021, ApJ, 923, 50, doi: 10.3847/1538-4357/ac1b9f * Froese Fischer & Tachiev (2004) Froese Fischer, C., & Tachiev, G. 2004, Atomic Data and Nuclear Data Tables, 87, 1, doi: 10.1016/j.adt.2004.02.001 * Froese Fischer et al. (2006) Froese Fischer, C., Tachiev, G., & Irimia, A. 2006, Atomic Data and Nuclear Data Tables, 92, 607, doi: 10.1016/j.adt.2006.03.001 * Gnat (2017) Gnat, O. 2017, ApJS, 228, 11, doi: 10.3847/1538-4365/228/2/11 * Gnat & Sternberg (2007) Gnat, O., & Sternberg, A. 2007, ApJS, 168, 213, doi: 10.1086/509786 * Gontcharov (2006) Gontcharov, G. A. 2006, Astronomy Letters, 32, 759, doi: 10.1134/S1063773706110065 * Gritton et al. (2014) Gritton, J. A., Shelton, R. L., & Kwak, K. 2014, ApJ, 795, 99, doi: 10.1088/0004-637X/795/1/99 * Gronke & Oh (2020) Gronke, M., & Oh, S. P. 2020, MNRAS, 492, 1970, doi: 10.1093/mnras/stz3332 * Heitsch et al. (2022) Heitsch, F., Marchal, A., Miville-Deschênes, M. A., Shull, J. M., & Fox, A. J. 2022, MNRAS, 509, 4515, doi: 10.1093/mnras/stab3266 * HI4PI Collaboration et al. (2016) HI4PI Collaboration, Ben Bekhti, N., Flöer, L., et al. 2016, A&A, 594, A116, doi: 10.1051/0004-6361/201629178 * Inglis & Teller (1939) Inglis, D. R., & Teller, E. 1939, ApJ, 90, 439, doi: 10.1086/144118 * Jenkins (2009) Jenkins, E. B. 2009, ApJ, 700, 1299, doi: 10.1088/0004-637X/700/2/1299 * Kahn (1981) Kahn, F. D. 1981, in Astrophysics and Space Science Library, Vol. 91, Investigating the Universe, ed. F. D. Kahn, 1, doi: 10.1007/978-94-009-8534-6_1 * Karim et al. (2018) Karim, T., Fox, A. J., Jenkins, E. B., et al. 2018, ApJ, 860, 98, doi: 10.3847/1538-4357/aac167 * Kaufer et al. (1999) Kaufer, A., Stahl, O., Tubbesing, S., et al. 1999, The Messenger, 95, 8 * Keeney et al. (2006) Keeney, B. A., Danforth, C. W., Stocke, J. T., et al. 2006, ApJ, 646, 951, doi: 10.1086/505128 * Khaire & Srianand (2019) Khaire, V., & Srianand, R. 2019, MNRAS, 484, 4174, doi: 10.1093/mnras/stz174 * Krishnarao et al. (2020a) Krishnarao, D., Benjamin, R. A., & Haffner, L. M. 2020a, ApJ, 899, L11, doi: 10.3847/2041-8213/aba8f0 * Krishnarao et al. (2020b) —. 2020b, Science Advances, 6, 9711, doi: 10.1126/sciadv.aay9711 * Lanz & Hubeny (2003) Lanz, T., & Hubeny, I. 2003, ApJS, 146, 417, doi: 10.1086/374373 * Lockman et al. (2020) Lockman, F. J., Di Teodoro, E. M., & McClure-Griffiths, N. M. 2020, ApJ, 888, 51, doi: 10.3847/1538-4357/ab55d8 * McClure-Griffiths et al. (2013) McClure-Griffiths, N. M., Green, J. A., Hill, A. S., et al. 2013, ApJ, 770, L4, doi: 10.1088/2041-8205/770/1/L4 * Moos et al. (2000) Moos, H. W., Cash, W. C., Cowie, L. L., et al. 2000, ApJ, 538, L1, doi: 10.1086/312795 * Morton (2003) Morton, D. C. 2003, ApJS, 149, 205, doi: 10.1086/377639 * Murga et al. (2015) Murga, M., Zhu, G., Ménard, B., & Lan, T.-W. 2015, MNRAS, 452, 511, doi: 10.1093/mnras/stv1277 * Pal’chikov (1998) Pal’chikov, V. G. 1998, Phys. Scr, 57, 581, doi: 10.1088/0031-8949/57/5/006 * Peach et al. (1988) Peach, G., Saraph, H. E., & Seaton, M. J. 1988, Journal of Physics B Atomic Molecular Physics, 21, 3669, doi: 10.1088/0953-4075/21/22/006 * Predehl et al. (2020) Predehl, P., Sunyaev, R. A., Becker, W., et al. 2020, Nature, 588, 227, doi: 10.1038/s41586-020-2979-0 * Prochaska et al. (2017) Prochaska, J. X., Tejos, N., Crighton, N., et al. 2017, Linetools/Linetools: Third Minor Release, v0.3, Zenodo, doi: 10.5281/zenodo.1036773 * Routly & Spitzer (1952) Routly, P. M., & Spitzer, Lyman, J. 1952, ApJ, 115, 227, doi: 10.1086/145535 * Safronova & Safronova (2011) Safronova, M. S., & Safronova, U. I. 2011, Phys. Rev. A, 83, 012503, doi: 10.1103/PhysRevA.83.012503 * Savage & Sembach (1996) Savage, B. D., & Sembach, K. R. 1996, ARA&A, 34, 279, doi: 10.1146/annurev.astro.34.1.279 * Savage et al. (2017) Savage, B. D., Kim, T.-S., Fox, A. J., et al. 2017, ApJS, 232, 25, doi: 10.3847/1538-4365/aa8f4c * Selig et al. (2015) Selig, M., Vacca, V., Oppermann, N., & Enßlin, T. A. 2015, A&A, 581, A126, doi: 10.1051/0004-6361/201425172 * Sembach et al. (1995) Sembach, K. R., Savage, B. D., & Lu, L. 1995, ApJ, 439, 672, doi: 10.1086/175206 * Sembach et al. (1991) Sembach, K. R., Savage, B. D., & Massa, D. 1991, ApJ, 372, 81, doi: 10.1086/169956 * Shapiro & Field (1976) Shapiro, P. R., & Field, G. B. 1976, ApJ, 205, 762, doi: 10.1086/154332 * Siluk & Silk (1974) Siluk, R. S., & Silk, J. 1974, ApJ, 192, 51, doi: 10.1086/153033 * Smoker et al. (2011) Smoker, J. V., Fox, A. J., & Keenan, F. P. 2011, MNRAS, 415, 1105, doi: 10.1111/j.1365-2966.2011.18647.x * Sparre et al. (2020) Sparre, M., Pfrommer, C., & Ehlert, K. 2020, MNRAS, 499, 4261, doi: 10.1093/mnras/staa3177 * Su et al. (2010) Su, M., Slatyer, T. R., & Finkbeiner, D. P. 2010, ApJ, 724, 1044, doi: 10.1088/0004-637X/724/2/1044 * Vallée (2017) Vallée, J. P. 2017, The Astronomical Review, 13, 113, doi: 10.1080/21672857.2017.1379459 * Vallerga et al. (1993) Vallerga, J. V., Vedder, P. W., Craig, N., & Welsh, B. Y. 1993, ApJ, 411, 729, doi: 10.1086/172875 * Viegas (1995) Viegas, S. M. 1995, MNRAS, 276, 268, doi: 10.1093/mnras/276.1.268 * Wakker (2006) Wakker, B. P. 2006, ApJS, 163, 282, doi: 10.1086/500365 * Wakker & van Woerden (1991) Wakker, B. P., & van Woerden, H. 1991, A&A, 250, 509 * Wakker et al. (2003) Wakker, B. P., Savage, B. D., Sembach, K. R., et al. 2003, ApJS, 146, 1, doi: 10.1086/346230 * Wenger et al. (2019) Wenger, T. V., Balser, D. S., Anderson, L. D., & Bania, T. M. 2019, ApJ, 887, 114, doi: 10.3847/1538-4357/ab53d3 * Yan et al. (1998) Yan, Z.-C., Tambasco, M., & Drake, G. W. F. 1998, Phys. Rev. A, 57, 1652, doi: 10.1103/PhysRevA.57.1652 * Zech et al. (2008) Zech, W. F., Lehner, N., Howk, J. C., Dixon, W. V. D., & Brown, T. M. 2008, ApJ, 679, 460, doi: 10.1086/587135
# Robust Implicit Adaptive Low Rank Time-Stepping Methods for Matrix Differential Equations Daniel Appelö Department of Mathematics, Virginia Tech, Blacksburg, VA 24061 U.S.A<EMAIL_ADDRESS>Research supported by DOE Office of Advanced Scientific Computing Research under the Advanced Research in Quantum Computing program, subcontracted from award 2019-LLNL-SCW-1683, NSF DMS-2208164, and Virginia Tech. Yingda Cheng Department of Mathematics, Virginia Tech, Blacksburg, VA 24061 U.S.A<EMAIL_ADDRESS>Research supported by NSF DMS-2011838, DOE grant DE-SC0023164 and Virginia Tech. ###### Abstract In this work, we develop implicit rank-adaptive schemes for time-dependent matrix differential equations. The dynamic low rank approximation (DLRA) is a well-known technique to capture the dynamic low rank structure based on Dirac–Frenkel time-dependent variational principle. In recent years, it has attracted a lot of attention due to its wide applicability. Our schemes are inspired by the three-step procedure used in the rank adaptive version of the unconventional robust integrator (the so called BUG integrator) [2] for DLRA. First, a prediction (basis update) step is made computing the approximate column and row spaces at the next time level. Second, a Galerkin evolution step is invoked using a base implicit solve for the small core matrix. Finally, a truncation is made according to a prescribed error threshold. Since the DLRA is evolving the differential equation projected on to the tangent space of the low rank manifold, the error estimate of the BUG integrator contains the tangent projection (modeling) error which cannot be easily controlled by mesh refinement. This can cause convergence issue for equations with cross terms. To address this issue, we propose a simple modification, consisting of merging the row and column spaces from the explicit step truncation method together with the BUG spaces in the prediction step. In addition, we propose an adaptive strategy where the BUG spaces are only computed if the residual for the solution obtained from the prediction space by explicit step truncation method, is too large. We prove stability and estimate the local truncation error of the schemes under assumptions. We benchmark the schemes in several tests, such as anisotropic diffusion, solid body rotation and the combination of the two, to show robust convergence properties. ## 1 Introduction In this work, we are interested in solving linear matrix differential equations in the following form $\frac{d}{dt}X(t)=F(X(t),t),\quad X(t)\in\mathbb{R}^{m_{1}\times m_{2}},\quad X(0)=X_{0},$ (1) where $F(X(t),t)=\sum_{j=1}^{s}A_{j}X(t)B_{j}^{T}+G(t),$ and $A_{j}\in\mathbb{R}^{m_{1}\times m_{1}},B_{j}\in\mathbb{R}^{m_{2}\times m_{2}}$ are sparse or structured matrices for which a fast matrix-vector product is assumed to be known. Further, $G(t)\in\mathbb{R}^{m_{1}\times m_{2}}$ is a given function that is assumed to have a known low rank decomposition. The integer $s,$ which denotes the separation rank of $F(\cdot),$ is assumed not to be too large. We focus on implicit methods, which are particularly needed for computing solutions to stiff equations. Equation (1) arises in many applications governed by partial differential equations (PDE). For example, using the method of lines approach, we can cast numerical discretizations of two-dimensional (or even higher dimensional) linear time-dependent convection-diffusion-reaction equations in the form of (1). If we represent the unknowns on a Cartesian grid, the elements of $X(t)$ will represent the unknowns (e.g. point values of the solution) at a given time $t.$ The matrices $A_{j},B_{j}$ will, e.g., correspond to difference operators or variable coefficients, and $G(t)$ encodes the given source terms and boundary conditions. Compared to the traditional approach of numerical PDE solvers, the matrix approach does not vectorize the unknowns into a single vector at each time step, but rather we represent the unknowns in a matrix format, retaining the relative relations in each spatial dimension. In recent years, the matrix (or tensor in higher dimensions) approach received increased attention in numerical analysis because it offers a way to tame _the curse of dimensionality_ , enabling approximations to solutions of high dimensional PDEs [10, 13]. The overarching idea is to expose low rank structure in the solution manifold in order to drastically reduce storage and computation. This is the idea of low rank tensor methods for PDEs [7, 1]. In two dimensions, this approach reduces to the low rank matrix method, where it is assumed that the matrix $X(t)$ has a small rank $r.$ As is well known, a rank $r$ matrix of size $m_{1}\times m_{2}$ allows for a SVD decomposition with only $(m_{1}+m_{2}-r)r$ degrees of freedom. This is much smaller than $m_{1}m_{2}$, the degrees of freedom of the full matrix, when $r$ is small compared to $m_{1}$ and $m_{2}$. If the rank of $X(t)$ is fixed, i.e. we constrain the solution $X(t)$ to live on the rank $r$ matrix manifold $\mathcal{M}_{r}=\\{X\in\mathbb{R}^{m_{1}\times m_{2}},{\sf rank}(X)=r\\},$ then an effective numerical solution can be computed by the dynamic low rank approximation (DLRA) [15, 16]. The DLRA solves $\frac{d}{dt}X(t)=\Pi_{X(t)}F(X(t),t),$ where $\Pi_{X(t)}$ is the orthogonal projection onto the tangent space $T_{X(t)}\mathcal{M}_{r}$ of the manifold $\mathcal{M}_{r}$ at $X(t).$ The DLRA has found success in many applications, including extensions to various tensor formats [18, 17]. However, the constraint that $X(t)\in\mathcal{M}_{r}$ requires that the rank $r$ is estimated a-priori and this is in contrast with the rank adaptive schemes. A rank adaptive method adaptively choose the rank at each time step according to a prescribed error tolerance. Several approaches to rank adaptivity can be found in the literature. For example, a straightforward idea is the so-called step truncation method [4] which evolves the low rank solution for one time step by a traditional time stepping method in an ambient space of higher rank, then performs a truncation (by SVD with given tolerance). This method is intuitive and effective for explicit time stepping methods. For implicit schemes [21], implementation of step truncation methods is nontrivial and requires an effective iterative solver in adaptive low rank format [7, 1]. Another approach that has been suggested by multiple groups, and that is similar to what we propose, is to modify the DLRA method by using rank increase and decrease indicators and augment or truncate the spaces at each time step [12, 11, 24]. In particular the ideas in [24], while being specialized to time dependent Schrödinger, are similar to the approach we take here. A promising idea in [2] uses the so called unconventional (BUG) integrator [3] to achieve rank adaptivity. This approach was recently extended to higher order in time in [19]. In the BUG integrator [3], two subproblems that come from the projector splitting of DLRA [16] are solved. Then, the solutions to the two subproblems are used to update the column and row spaces of the matrix; and a Galerkin evolution step is performed in the resulting space, searching for the coefficients of a small matrix. For the rank adaptive version of the unconventional integrator [2], this translates to solving two matrix differential equation of sizes $m_{1}\times r,m_{2}\times r$ in the first step and one matrix differential equation of size $2r\times 2r$ in the Galerkin step, followed by a truncated SVD step. The ideas in [3, 2] can be implemented for implicit methods because the matrix size is fixed in each subproblems, so a standard linear solver will suffice. However, the errors of both methods in [3, 2] are subject to the tangent projection error or the so called _modeling error_ [14]. It is important to understand that this modeling error is present in many problems and can, as we will see below, result in non convergent numerical methods. We now describe a trivial but prototypical example that will make the DLRA/BUG method fail. Assume that $X(t)=U\Sigma V^{T}$ and that the column and row space of the matrix $F(X(t),t)$ are both (respectively) orthogonal to $U$ and $V$, then $\Pi_{X(t)}F(X(t),t)=0,$ which means the solution to the DLRA description will remain stationary. For example, this will happen for the numerical discretization when $X(t)$ is a rank 1 even function in both variables (say $X(t,x_{1},x_{2})=\exp(-x_{1}^{2})\exp(-x_{2}^{2})$) on a square domain with center $(0,0)$, and $F(X(t),t)=AXB^{T}$, with $A={\sf diag}(x_{1})$ and $B={\sf diag}(x_{2})$. The main contribution of this paper is to propose a simple improvement to the BUG integrator that enhances the robustness with respect to convergence. In our method, we merge the row and column spaces from the explicit step truncation method with the BUG spaces in the prediction step to alleviate the tangent projection error associated with DLRA. As the BUG spaces can be expensive to compute when implicit schemes are used, we propose an adaptive acceleration strategy based on the residual of the underlying classic scheme. The strategy is straightforward, if the residual obtained when only the step truncation spaces are used, is larger than the truncation threshold then we add in the BUG spaces. We observe that this strategy is robust in convergence and can reduce the computational time for moderately stiff problems. The rest of the paper is organized as follows. Section 2 reviews the background on low rank time integrators. In Section 3, we describe the proposed schemes and perform numerical analysis of their properties. Section 4 provides numerical experiments and in Section 5 we draw conclusions. ## 2 Background review In this section, we will review basic concepts of low rank time stepping methods for (1). In Section 2.1, we gather the notations used in the paper. Section 2.2 reviews an explicit scheme truncation method based on the forward Euler method, while Section 2.3 reviews the rank adaptive BUG scheme. ### 2.1 Notations and preliminaries In this paper, we use $\|\cdot\|$ to denote the matrix Frobenius norm. The Frobenius norm is a natural choice for matrix functions as $\|A\|$ coincides with the $L^{2}$ vector norm of the vectorized matrix ${\sf vec}(A),$ which is commonly used in the analysis of the standard ODE/PDE solvers. We also use matrix inner product defined as follows, for $A,B\in\mathbb{R}^{m_{1}\times m_{2}},$ we define $\langle A,B\rangle=\sum_{i,j=1}^{m_{1},m_{2}}A_{ij}B_{i,j}.$ Then $\langle A,A\rangle=\|A\|^{2}.$ The following property, $\langle A,USV^{T}\rangle=\langle U^{T}AV,S\rangle,$ (2) holds for any $U\in\mathbb{R}^{m_{1}\times r},S\in\mathbb{R}^{r\times s},V\in\mathbb{R}^{m_{2}\times s}$. Here we stop to caution the reader that throughout this paper a matrix $\Sigma,S,$ or $\mathcal{S}$ does not always represent the diagonal matrix holding the singular values of another matrix. The numerical solution at $t^{n}$ is denoted by $\hat{X}^{n}\in\mathbb{R}^{m_{1}\times m_{2}}.$ In particular, this is a rank $r_{n}$ matrix with singular value decomposition (SVD) $\hat{X}^{n}=U^{n}\Sigma^{n}(V^{n})^{T},$ where $U^{n}\in\mathbb{R}^{m_{1}\times r_{n}}$, $V^{n}\in\mathbb{R}^{m_{2}\times r_{n}}$ have orthogonal columns, and $\Sigma^{n}={\sf diag}(\sigma_{1},\cdots,\sigma_{r_{n}})$ is a $r_{n}\times r_{n}$ diagonal matrix with diagonal entries $\sigma_{1}\geq\cdots\geq\sigma_{r_{n}}>0.$ Here, the rank $r_{n}$ will be chosen adaptively by the numerical scheme so that $\hat{X}^{n}$ approximate the true solution $X(t^{n})$ with a prescribed accuracy. We use $\mathcal{T}_{\epsilon}$ to denote a generic matrix approximation operator with accuracy $\epsilon$ in the Frobenius norm, i.e. $\|A-\mathcal{T}_{\epsilon}(A)\|\leq\epsilon.$ A prominent example is the truncated SVD of a matrix. Namely, given a generic rank $r$ matrix $A\in\mathbb{R}^{m_{1}\times m_{2}}$ with reduced SVD: $A=U\Sigma V^{T},$ then $\mathcal{T}^{svd}_{\epsilon}(A)=U[:,1:s]\,{\sf diag}(\sigma_{1},\cdots,\sigma_{r})\,V[:,1:s]^{T},$ where we have used the standard MATLAB notation for submatrices and $r$ is chosen to be the smallest integer so that $(\sum_{j=s+1}^{r}\sigma_{j}^{2})^{1/2}\leq\epsilon.$ In this paper, we will also encounter low rank matrix sum operations. This frequently appearing operation in low rank methods can be efficiently computed by Algorithm 1. The inputs to this algorithm are the SVD representations of the matrices $X_{j}=U_{j}\Sigma_{j}(V_{j})^{T},j=1\ldots m,$ and the output is the truncated SVD of their sum $\mathcal{T}^{sum}_{\epsilon}(\sum_{j=1}^{m}X_{j})=\mathcal{U}\mathcal{S}\mathcal{V}^{T}.$ We first write $U=[U_{1},\ldots,U_{m}],$ $\Sigma={\sf diag}(\Sigma_{1},\ldots,\Sigma_{m}),$ $V=[V_{1},\ldots,V_{m}]$. Then use the column-pivoted QR decomposition, [6], denoted by $[Q,R,\Pi]={\sf qr}(A)$, which computes $QR\Pi=A.$ In this factorization the columns of $Q$ are orthogonal as usual, but the introduction of the permutation matrix $\Pi$, makes it possible to guarantee that the diagonal entries of the upper triangular matrix $R$ are strictly decreasing in magnitude. Assume that the column pivoted QR procedure applied to to $U$ and $V$ yields the factorizations $Q_{1}R_{1}\Pi_{1}$, and $Q_{2}R_{2}\Pi_{2}$. Then the truncated matrix sum $\mathcal{T}^{sum}_{\epsilon}(\sum_{j=1}^{m}X_{j})=Q_{1}\mathcal{U}\mathcal{S}(Q_{2}\mathcal{V})^{T}.$ is obtained by a truncated SVD on the small matrix $\mathcal{T}^{svd}_{\epsilon}(R_{1}\Pi_{1}\,\Sigma\,\Pi_{2}^{T}R_{2}^{T})=\mathcal{U}\mathcal{S}\mathcal{V}^{T}$. 1 Input : low rank matrices $X_{j},j=1\ldots m$ or their SVD $U_{j}\Sigma_{j}(V_{j})^{T},j=1\ldots m$ Output : truncated SVD of their sum $\mathcal{T}^{sum}_{\epsilon}(\sum_{j=1}^{m}X_{j})=\mathcal{U}\mathcal{S}\mathcal{V}^{T}$ Parameter : tolerance $\epsilon$ 2 Form $U=[U_{1},\ldots,U_{m}],\,\Sigma={\sf diag}(\Sigma_{1},\ldots,\Sigma_{m}),\,V=[V_{1},\ldots,V_{m}]$. 3Perform column pivoted QR: $[Q_{1},R_{1},\Pi_{1}]={\sf qr}(U),$ $[Q_{2},R_{2},\Pi_{2}]={\sf qr}(V).$ 4Compute the truncated SVD: $\mathcal{T}^{svd}_{\epsilon}(R_{1}\Pi_{1}\,\Sigma\,\Pi_{2}^{T}R_{2}^{T})=\mathcal{U}\mathcal{S}\mathcal{V}^{T}$. 5Form $\mathcal{U}\leftarrow Q_{1}\mathcal{U},\mathcal{V}\leftarrow Q_{2}\mathcal{V}$. Return $[\mathcal{U},\mathcal{S},\mathcal{V}]=\mathcal{T}^{sum}_{\epsilon}(\sum_{j=1}^{m}X_{j})$ Algorithm 1 Sum of low rank matrices ### 2.2 Explicit step truncation schemes The primary assumption to guarantee the efficiency of any low rank solver is that $r_{n}\ll m_{1},r_{n}\ll m_{2},$ i.e. the solution can be well approximated by a low rank matrix. When such assumptions hold, the main steps of the numerical algorithm should specify the evolution of $U^{n},\Sigma^{n},V^{n}$ based on the matrix differential equation (1). For explicit schemes, a simple approach is to numerically integrate $\hat{X}^{n}$ according to a standard time integrator to $t^{n+1}$ and then perform a truncated SVD according to the error threshold to obtain $\hat{X}^{n+1}.$ This is the so-called step truncation method. Methods of this type and their variations have been discussed in [14, 4]. The main steps of the rank adaptive forward Euler time scheme are highlighted in Algorithm 2. The first step of Algorithm 2 is a standard forward Euler step for (1) with the truncated right hand side in Line 2. This will result in a numerical solution $\hat{X}^{n+1,pre}$ with higher rank than needed. Then, the second step in Line 2 is to truncate the solution to a lower rank matrix. In both steps, we use Algorithm 1 to perform the sum of the low rank matrices involved. 1 Input : numerical solution at $t^{n}:$ rank $r_{n}$ matrix $\hat{X}^{n}$ in its SVD form $U^{n}\Sigma^{n}(V^{n})^{T}.$ Output : numerical solution at $t^{n+1}:$ rank $r_{n+1}$ matrix $\hat{X}^{n+1}$ in its SVD form $U^{n+1}\Sigma^{n+1}(V^{n+1})^{T}.$ Parameter : time step $\Delta t$, error tolerance $\epsilon_{1},\epsilon_{2}$ 2 (Evolution). $\hat{X}^{n+1,pre}=\hat{X}^{n}+\Delta t\mathcal{T}^{sum}_{\epsilon_{1}}(F(\hat{X}^{n},t^{n})).$ (Truncation). $\hat{X}^{n+1}=\mathcal{T}^{sum}_{\epsilon_{2}}(\hat{X}^{n+1,pre}).$ Algorithm 2 Forward Euler scheme $t^{n}\rightarrow t^{n+1}$ The convergence of this algorithm is well understood [21]. The local truncation error is on the order of $O(\Delta t^{2}+\epsilon_{1}\Delta t+\epsilon_{2}).$ If one choose $\epsilon_{1}=O(\Delta t),\epsilon_{2}=O(\Delta t^{2}),$ we will obtain a first order accurate solution. As for computational cost, we can see that the procedure involves QR factorization of tall matrices and SVD for a small matrix. If we assume $r_{n}=O(r),$ $s$ is small (i.e. $s=O(1)$) and $G$ has low rank (i.e. ${\sf rank}(G)=O(r)$), the computational complexity is on the order of $O(m_{1}r^{2}+m_{2}r^{2}+r^{3}).$ Algorithm 2 can be readily extended to higher order by embedding in Runge-Kutta, multistep methods, and was also considered with tangent projection in the projected Runge-Kutta schemes [14, 4, 8]. ### 2.3 Dynamic low rank approximation The DLRA modifies the equation, and solves $\frac{d}{dt}X(t)=\Pi_{X(t)}F(X(t),t),$ (3) where $\Pi_{X(t)}$ is the orthogonal projection onto the tangent space $T_{X(t)}\mathcal{M}_{r}$ of the the rank $r$ matrix manifold $\mathcal{M}_{r}=\\{X\in\mathbb{R}^{m_{1}\times m_{2}},{\sf rank}(X)=r\\}$ at $X(t).$ The DLRA is particularly suited for a fixed rank calculation, if the computation is constrained on $\mathcal{M}_{r}.$ It has found success in many applications, including extensions to various tensor formats [18, 17] and particularly in quantum mechanics [20]. We reminder the reader that if the SVD of $X\in\mathcal{M}_{r}$ is given by $U\Sigma V^{T},$ then the tangent space of $\mathcal{M}_{r}$ at $X$ is given by $T_{X}\mathcal{M}_{r}=\left\\{[U\quad U_{\perp}]\begin{bmatrix}\mathbb{R}^{r\times r}&\mathbb{R}^{r\times(m_{2}-r)}\\\ \mathbb{R}^{(m_{1}-r)\times r}&0^{(m_{1}-r)\times(m_{2}-r)}\end{bmatrix}[V\quad V_{\perp}]^{T}\right\\},$ where $U_{\perp},V_{\perp}$ are orthogonal complements of $U,V$ in $\mathbb{R}^{m_{1}},\mathbb{R}^{m_{2}},$ respectively. As we can see later in the PDE examples, typically if there are cross terms like $F(X)=AXB$ (e.g. cross derivatives or variable coefficient problem with operators acting on both left and right side of the matrix), then in general $AXB\notin T_{X}\mathcal{M}_{r}.$ In this case, the _modeling error_ [14] (i.e. the error from the tangent projection, which is $\|(I-\Pi_{X(t)})F(X,t)\|$) will be evident in the numerical approximation, see for example error estimates in [3, 2]. Below we will review the rank adaptive BUG integrator in [2] for the discretization of (3). First, two subproblems that come from the projector splitting of DLRA [16] are solved in the $K-$ and $L-$step. Then, the solutions to the two subproblems are used to update the column and row spaces of the matrix; and a Galerkin evolution step is performed in the resulting space, searching for the optimal solution with the Galerkin condition. The idea of finding the appropriate subspace and computing the solution in that space is similar to a widely used technique in numerical linear algebra called projection methods [22]. For completeness, we describe the rank adaptive BUG integrator in Algorithm 3. The original algorithm is written in a time continuous format, but here we present it using an implicit Euler discretization of the $K-,L-,S-$steps to facilitate the discussion in the remaining part of the paper. Input : numerical solution at $t^{n}:$ rank $r_{n}$ matrix $\hat{X}^{n}$ in its SVD form $U^{n}\Sigma^{n}(V^{n})^{T}.$ Output : numerical solution at $t^{n+1}:$ rank $r_{n+1}$ matrix $\hat{X}^{n+1}$ in its SVD form $U^{n+1}\Sigma^{n+1}(V^{n+1})^{T}.$ Parameter : time step $\Delta t$, error tolerance $\epsilon_{2}$ 1 (Prediction). $K-$step and $L-$step integrating from $t^{n}$ to $t^{n+1}$. 2Solve $K^{n+1}-K^{n}=\Delta tF(K^{n+1}(V^{n})^{T},t)V^{n},\quad K^{n}=U^{n}\Sigma^{n},$ to obtain $K^{n+1},$ and $[\tilde{U},\sim,\sim]={\sf qr}([U^{n},K^{n+1}]).$ 3Solve $L^{n+1}-L^{n}=\Delta tF(U^{n}(L^{n+1})^{T},t)^{T}U^{n},\quad L^{n}=V^{n}\Sigma^{n},$ to obtain $L^{n+1},$ and $[\tilde{V},\sim,\sim]={\sf qr}([V^{n},L^{n+1}]).$ 4(Galerkin Evolution). $S-$step: solve for $S^{n+1}$ from $S^{n+1}-S^{n}=\Delta t\tilde{U}^{T}F(\tilde{U}S^{n+1}\tilde{V}^{T},t)\tilde{V},\quad S^{n}=\tilde{U}^{T}U^{n}\Sigma^{n}(V^{n})^{T}\tilde{V},$ to obtain $S^{n+1}.$ 5(Truncation). $\hat{X}^{n+1}=\tilde{U}\mathcal{T}_{\epsilon_{2}}^{svd}(S^{n+1})\tilde{V}^{T}=U^{n+1}\Sigma^{n+1}(V^{n+1})^{T}.$ Algorithm 3 Rank adaptive BUG integrator using implicit Euler [2]. ## 3 Numerical method We now present the proposed numerical methods, the Merge method and its adapted version. We first describe the algorithms, and then discuss the properties and the rationale behind the design of the schemes. ### 3.1 The Merge method In the Merge method, we merge the column and row spaces from the explicit step truncation method and the spaces from the $K-$ and $L-$ steps in the BUG method. As mentioned above the method consists of three stages. Prediction: Our first ingredient in the method is to predict the column and row space using the explicit Euler scheme in combination with the BUG prediction spaces from Algorithm 3. Precisely, the predicted column and row spaces are defined as the column and row spaces of the collection of matrices $[\hat{X}^{n},\,\mathcal{T}_{\epsilon_{1}}(F(\hat{X}^{n},t^{n})),\,K^{n+1}(L^{n+1})^{T}].$ Here the truncation level $\epsilon_{1}$ can be set to zero (no truncation used) or chosen as $\epsilon_{1}=C_{1}\Delta t$. Both choices gives first order accurate solutions but depending on the form of (1) the added cost from the $\epsilon_{1}=C_{1}\Delta t$ truncation may be offset by the reduction in the dimension of the predicted row and column spaces. To complete the next step we first orthogonalize the predicted column and row spaces. That is, we use column pivoted QR to find orthogonal matrices $\tilde{U}\in\mathbb{R}^{m_{1}\times s_{1}},\tilde{V}\in\mathbb{R}^{m_{2}\times s_{2}}$ spanning these spaces. Here $s_{1},s_{2}$ is bounded from above by $2r_{n}(s+1)+r_{G},$ where $r_{G}$ is the rank of the source $G(t^{n}).$ Galerkin evolution: The Galerkin evolution step can be understood as to find an approximation $\hat{X}^{n+1,pre}\in W_{\tilde{U},\tilde{V}},$ such that $\langle\hat{X}^{n+1,pre},A\rangle=\langle\hat{X}^{n}+\Delta tF(\hat{X}^{n+1,pre},t^{n+1}),A\rangle,\qquad\forall A\in W_{\tilde{U},\tilde{V}},$ (4) where $W_{\tilde{U},\tilde{V}}=\\{A\in\mathbb{R}^{m_{1}\times m_{2}}:A=\tilde{U}\Sigma\tilde{V}^{T},\textrm{with }\Sigma\in\mathbb{R}^{s_{1}\times s_{2}}\\}$ denote all size $m_{1}\times m_{2}$ spaces with column and row spaces as $\tilde{U}$ and $\tilde{V}.$ To solve this problem we let $\hat{X}^{n+1,pre}=\tilde{U}\tilde{\Sigma}\tilde{V}^{T},A=\tilde{U}\Sigma^{*}\tilde{V}^{T}.$ Then by (2), we get $\langle\tilde{\Sigma},\Sigma^{*}\rangle=\langle\tilde{U}^{T}(\hat{X}^{n}+\Delta tF(\tilde{U}\tilde{\Sigma}\tilde{V}^{T},t^{n+1})))\tilde{V},\Sigma^{*}\rangle,\qquad\forall\Sigma^{*}\in\mathbb{R}^{s_{1}\times s_{2}}.$ This means $\displaystyle\tilde{\Sigma}$ $\displaystyle=$ $\displaystyle\tilde{U}^{T}(\hat{X}^{n}+\Delta tF(\tilde{U}\tilde{\Sigma}\tilde{V}^{T},t^{n+1})))\tilde{V}$ (5) $\displaystyle=$ $\displaystyle\tilde{U}^{T}U^{n}\Sigma^{n}(V^{n})^{T}\tilde{V}^{T}+\Delta t\sum_{j=1}^{s}\tilde{U}^{T}A_{j}\tilde{U}\tilde{\Sigma}\tilde{V}^{T}B_{j}^{T}\tilde{V}+\Delta t\tilde{U}^{T}U^{n}_{G}\Sigma^{n}_{G}(V^{n}_{G})^{T}\tilde{V}.$ The equation (5) is a linear matrix equation (generalized Sylvester equation) for the unknown $\tilde{\Sigma}.$ Since the problems we are considering have low rank solutions, we expect the dimensions of the matrix $\tilde{\Sigma}$ to be small. To solve this equation one can either use a direct solves or an iterative solver (e.g. GMRES) [23]. Truncation: The truncation step uses the truncated SVD of $\hat{X}^{n+1}=\mathcal{T}_{\epsilon_{2}}^{svd}(\tilde{U}\tilde{\Sigma}\tilde{V}^{T})=\tilde{U}\mathcal{T}_{\epsilon_{2}}^{svd}(\tilde{\Sigma})\tilde{V}^{T}.$ Here the truncation level is chosen as $\epsilon_{2}=C_{2}\Delta t^{2}$ to ensure accuracy. Algorithm 4 summarizes the Merge method. 1 Input : numerical solution at $t^{n}:$ rank $r_{n}$ matrix $\hat{X}^{n}$ in its SVD form $\hat{U}^{n}\hat{\Sigma}^{n}(\hat{V}^{n})^{T}$. Output : numerical solution at $t^{n+1}:$ rank $r_{n+1}$ matrix $\hat{X}^{n+1}$ in its SVD form $U^{n+1}\Sigma^{n+1}(V^{n+1})^{T}.$ Parameters : truncation tolerances $\epsilon_{1},\epsilon_{2}$ 2 (Merge Prediction). Compute the truncated SVD of the right hand side: $[\mathcal{U},\mathcal{S},\mathcal{V}]=\mathcal{T}^{sum}_{\epsilon_{1}}(F(\hat{X}^{n},t^{n})),$. 3Compute the BUG prediction spaces $K^{n+1},L^{n+1}$ according to Algorithm 3. 4Orthogonalize the merged spaces by column pivoted QR to get the prediction spaces $[\tilde{U},\sim,\sim]={\sf qr}([[\hat{U}^{n},\mathcal{U},K^{n+1}]),$ $[\tilde{V},\sim,\sim]={\sf qr}([\hat{V}^{n},\mathcal{V}^{n},L^{n+1}]).$ 5(Galerkin Evolution). Find $\tilde{\Sigma}$ by solving $\tilde{\Sigma}=\tilde{U}^{T}(\hat{X}^{n}+\Delta tF(\tilde{U}\tilde{\Sigma}\tilde{V}^{T},t^{n+1}))\tilde{V}.$ (Truncation). $\hat{X}^{n+1}=\hat{U}^{n+1}\hat{\Sigma}^{n+1}(\hat{V}^{n+1})^{T}=\tilde{U}\mathcal{T}_{\epsilon_{2}}^{svd}(\tilde{\Sigma})\tilde{V}^{T}.$ Algorithm 4 Merge method $t^{n}\rightarrow t^{n+1}$ We now provide estimates of the computational cost of the different parts of the computation. Here we assume $m_{1}=m_{2}=m$, $s$ terms in differential equation and rank $r$. * • An evaluation of $AU,(BV)^{T}$ costs $\mathcal{O}(mr)$ for each of $AU$, $BV$, assuming $A$ and $B$ are sparse. * • The truncation of the right hand side $\mathcal{T}^{sum}_{\epsilon_{1}}(F)$ has two components. Two QR orthogonalizations, each costs $\mathcal{O}(2(m(rs)^{2}-\frac{(rs)^{2}}{3})$ and one SVD of the core dense matrix, this costs $\mathcal{O}((rs)^{3})$. * • The $K$ and $L$ solves used to compute the BUG spaces also require the solution of matrix equations but the dimension of these matrix solves are for a $m\times r$ matrix rather than a $sr\times sr$ matrix. For these equations we have found that applying GMRES in vectorized form works well. It should be noted that this is typically the most expensive part of the solvers. * • The cost of solving for the Galerkin core matrix $\tilde{\Sigma}$ in $\tilde{\Sigma}-\Delta t\tilde{U}^{T}F(\tilde{U}\tilde{\Sigma}\tilde{V}^{T},t^{n+1})\tilde{V}=\tilde{U}^{T}\hat{X}^{n}\tilde{V},$ depends on the algorithm used. For example, if GMRES is applied on the vectorized version of this equation each evaluation of the left hand side costs $\mathcal{O}(m(rs)^{2})$ if $F$ is a general function. If $F$ is in the form described in equation (1) it is possible to pre-compute dense $rs\times rs$ matrices (at a cost of $\mathcal{O}(ms(rs)^{2})$) and bring down the cost of the evaluation of the left hand side to $\mathcal{O}(s(rs)^{3})$ per evaluation. Neglecting the cost associated with the growth of the GMRES Krylov subspace this cost would then be multiplied with the number of iterations needed to converge (which is difficult to estimate). We have found that when the governing equation has elliptic terms it also works well to reformulate the Galerkin equation into a fixed point iteration $A_{2}^{-1}(I-\tilde{U}^{T}\Delta tA_{1}\tilde{U})\tilde{\Sigma}^{k}-\tilde{\Sigma}^{k}(\Delta tB_{2}\tilde{V})^{T}\tilde{V}B^{-1}_{1}=P(\tilde{\Sigma}^{k-1},\hat{X}^{n},\tilde{U},\tilde{V}).$ In the iteration we use a dense Sylvester solve with cost $\mathcal{O}((rs)^{3})$ for each iteration. When the matrices $A_{1}$ and $B_{2}$ correspond to approximations to second derivatives in the 1 and 2 direction and $A_{2}$ and $B_{1}$ are identity or diagonal positive definite matrices we find that this iteration converges to machine precision in a handful of iterations. Finally, we would like to comment on the rank evolution of the numerical solution. We find in our numerical experiments that the low rank method generally track the rank growth of the implicit Euler scheme well, i.e. the rank of $\hat{X}^{n}$ is on par with the rank of the implicit Euler solution at $t^{n}.$ However, due to the merge in the prediction step, the predicted rank is larger than the rank of the BUG space (which is equal to $2r$). This do translate to larger computational cost in the Galerkin step compared to the BUG solver. However we note that the upper bound in the BUG space $2r$ can be insufficient to capture rapidly changing initial layers in the PDEs as pointed out in [19] (where it was observed a larger than actual rank needs to be imposed for the numerical initial solution). For the numerical experiments we performed in this paper, we find that by merging with the spaces generated by the explicit schemes, this issue seems to be addressed and we do not need to impose large artificial rank for the numerical initial condition. ### 3.2 The Merge-adapt method As will be shown in Lemma 3.3, the column and row spaces generated from the current solution $\hat{U}^{n},\hat{V}^{n}$ combined with $\mathcal{U}$ and $\mathcal{V}$ in $[\mathcal{U},\mathcal{S},\mathcal{V}]=\mathcal{T}^{sum}_{\epsilon_{1}}(F(\hat{X}^{n},t^{n})),$ will be sufficiently accurate to yield a first order accurate solution as long as the time step is sufficiently small compared to the Lipschitz constant $L$ of $\|F\|$. This time step restriction can be overcome by merging with the BUG spaces $K^{n+1},L^{n+1}$, which are computed useing an implicit solver and can handle the stiffness as shown in the Merge method in Algorithm 4. However, since the computation of $K^{n+1}$ and $L^{n+1}$ requires a linear solve, it is preferable to only add those when necessary. This motivates the design of the Merge-adapt method, which is described in Algorithm 5. In Steps 1-4 in Algorithm 5, we perform a calculation based purely on column and row spaces generated by the explicit scheme for the term $\mathcal{T}^{sum}_{\epsilon_{1}}(F(\hat{X}^{n},t^{n})).$ In Step 5, we do a residual check, and if this fails, the method will fall back to the Merge method as shown in Steps 6-8. Otherwise, we proceed to the next time step. We, heuristically, argue that for this residual check can be used to robustly maintain first order accuracy. For simplicity, assume that the differential equation we would like to solve is $\frac{d}{dt}X(t)=AX(t)B^{T}+G(t).$ Let $X^{n+1}$ be the solution obtained by the classic implicit Euler method, i.e. it solves ${\sf vec}(X^{n+1})=\underbrace{(I-\Delta t(B\otimes A))^{-1}}_{C}{\sf vec}(X^{n})+\Delta t\,{\sf vec}(G(t_{n+1})).$ The low rank solution is not, in general, expected to solve this equation but we may introduce the residual $\hat{R}^{n+1}$ so that ${\sf vec}(\hat{X}^{n+1})=\underbrace{(I-\Delta t(B\otimes A))^{-1}}_{C}{\sf vec}(\hat{X}^{n})+\Delta t\,{\sf vec}(G(t_{n+1}))+{\sf vec}(\hat{R}^{n+1}).$ Combining the equations we find ${\sf vec}(\hat{X}^{n+1})-{\sf vec}(X^{n+1})=C[{\sf vec}(\hat{X}^{n})-{\sf vec}(X^{n})]+{\sf vec}(\hat{R}^{n+1}).$ Now if we assume $\|C\|\leq 1$ and the residual check pass, then the error between the implicit Euler solution and the implicit adaptive low rank solution $e^{n}={\sf vec}(\hat{X}^{n})-{\sf vec}(X^{n})$ satisfies $\|e^{n+1}\|\leq\|e^{n}\|+C_{2}\Delta t^{2}.$ This is enough to guarantee that the low rank solution is convergent because the implicit Euler solution is first order in time. It is reasonable to assume that the condition $\|C\|\leq 1$ is satisfied for most discretizations of diffusion or advection-diffusion equations [9]. Input : numerical solution at $t^{n}:$ rank $r_{n}$ matrix $\hat{X}^{n}$ in its SVD form $\hat{U}^{n}\hat{\Sigma}^{n}(\hat{V}^{n})^{T}$. Output : numerical solution at $t^{n+1}:$ rank $r_{n+1}$ matrix $\hat{X}^{n+1}$ in its SVD form $U^{n+1}\Sigma^{n+1}(V^{n+1})^{T}.$ Parameters : truncation tolerances $\epsilon_{1},\epsilon_{2}$ 1 (Cheap Prediction). Compute a first order prediction of the column and row spaces of $\hat{X}^{n+1}$. Compute the truncated SVD of the right hand side: $[\mathcal{U},\mathcal{S},\mathcal{V}]=\mathcal{T}^{sum}_{\epsilon_{1}}(F(\hat{X}^{n},t^{n}))$. 2Orthogonalize by column pivoted QR to get the prediction spaces $[\tilde{U},\sim,\sim]={\sf qr}([[\hat{U}^{n},\mathcal{U}]),$ $[\tilde{V},\sim,\sim]={\sf qr}([\hat{V}^{n},\mathcal{V}^{n}]).$ 3(Galerkin Evolution). Find $\tilde{\Sigma}$ by solving $\tilde{\Sigma}=\tilde{U}^{T}(\hat{X}^{n}+\Delta tF(\tilde{U}\tilde{\Sigma}\tilde{V}^{T},t^{n+1}))\tilde{V}.$ 4(Truncation). $\hat{X}^{n+1}=\hat{U}^{n+1}\hat{\Sigma}^{n+1}(\hat{V}^{n+1})^{T}=\tilde{U}\mathcal{T}_{\epsilon_{2}}^{svd}(\tilde{\Sigma})\tilde{V}^{T}.$ 5(Residual Check). Compute $\hat{R}^{n+1}=\hat{X}^{n+1}-\hat{X}^{n}-\Delta tF(\hat{X}^{n+1},t^{n+1})$. If $\|\hat{R}^{n+1}\|<\epsilon_{2}$ return the solution $\hat{X}^{n+1}$. If not 6(Merge Predicion). Compute the BUG prediction spaces $K^{n+1},L^{n+1}$ according to Algorithm 3. Orthogonalize the merged spaces by column pivoted QR: $[\tilde{U},\sim,\sim]={\sf qr}([[\hat{U}^{n},\mathcal{U},K^{n+1}]),$ $[\tilde{V},\sim,\sim]={\sf qr}([\hat{V}^{n},\mathcal{V}^{n},L^{n+1}]).$ 7(Galerkin Evolution). Find $\tilde{\Sigma}$ by solving $\tilde{\Sigma}=\tilde{U}^{T}(\hat{X}^{n}+\Delta tF(\tilde{U}\tilde{\Sigma}\tilde{V}^{T},t^{n+1}))\tilde{V}.$ 8(Truncation). $\hat{X}^{n+1}=\hat{U}^{n+1}\hat{\Sigma}^{n+1}(\hat{V}^{n+1})^{T}=\tilde{U}\mathcal{T}_{\epsilon_{2}}^{svd}(\tilde{\Sigma})\tilde{V}^{T}.$ Algorithm 5 Merge-adapt method $t^{n}\rightarrow t^{n+1}$ We note that the acceleration of Algorithm 5 over Algorithm 4 is based on the assumption that residual check passes in Step 5 and the BUG spaces are not computed. If the residual check fails, extra computation is performed in Steps 3-5 which may incur more computational cost. We observe that, for moderately stiff problems with small $s$, the Merge-adapt method has computational advantages. More discussion is provided in the numerical experiment section. ### 3.3 Analysis In this section, we analyze properties of the proposed methods. #### 3.3.1 Stability First, we consider conservative or dissipative systems for which $\langle F(X,t),X\rangle\leq 0,\forall t,X.$ Then it follows that $\frac{d}{dt}\|X\|^{2}=\langle F(X,t),X\rangle\leq 0,$ that is, the energy ($L^{2}$ norm) is monotonically decreasing. ###### Theorem 3.1. If we have $\langle F(X,t),X\rangle\leq 0,\forall t,X,$ then the numerical solutions from Algorithm 4 or 5 satisfy $\|\hat{X}^{n+1}\|\leq\|\hat{X}^{n}\|.$ ###### Proof. For Algorithm 4, first note that $\|X^{n+1,pre}\|=\|\tilde{U}\tilde{\Sigma}\tilde{V}^{T}\|=\|\tilde{\Sigma}\|.$ By (5) and (2), $\displaystyle\|\tilde{\Sigma}\|^{2}=\langle\tilde{\Sigma},\tilde{\Sigma}\rangle=\langle\tilde{U}^{T}(\hat{X}^{n}+\Delta tF(\tilde{U}\tilde{\Sigma}\tilde{V}^{T},t^{n+1}))\tilde{V},\tilde{\Sigma}\rangle$ $\displaystyle=\langle\hat{X}^{n}+\Delta tF(\tilde{U}\tilde{\Sigma}\tilde{V}^{T},t^{n+1}),\tilde{U}\tilde{\Sigma}\tilde{V}^{T}\rangle$ $\displaystyle\leq\langle\hat{X}^{n},\tilde{U}\tilde{\Sigma}\tilde{V}^{T}\rangle=\langle\hat{X}^{n},X^{n+1,pre}\rangle.$ Therefore, by Cauchy-Schwarz inequality, $\|X^{n+1,pre}\|^{2}\leq\|\hat{X}^{n}\|\|X^{n+1,pre}\|,$ which gives $\|X^{n+1,pre}\|\leq\|\hat{X}^{n}\|.$ Finally, because of the property of the truncated SVD, $\|\hat{X}^{n+1}\|=\|\mathcal{T}_{\epsilon_{2}}^{svd}(X^{n+1,pre})\|\leq\|X^{n+1,pre}\|\leq\|\hat{X}^{n}\|.$ The proof for Algorithm 5 is the same and is omitted. ∎ We can easily generalize this result to semi-bounded operator [9]. Specifically, if $\langle F(X,t),X\rangle\leq\alpha\|X\|^{2},\forall\,t,X,$ then $\frac{d}{dt}\|X\|^{2}=\langle F(X,t),X\rangle\leq\alpha\|X\|^{2},$ which implies $\|X(t)\|\leq e^{\alpha t}\|X(0)\|.$ ###### Theorem 3.2 (Stability for semi-bounded operator). If we have $\langle F(X,t),X\rangle\leq\alpha\|X\|^{2},\forall\,t,X$ then the numerical solutions from Algorithm 4 or 5 satisfy $\|\hat{X}^{n+1}\|\leq e^{\alpha\Delta t}\|\hat{X}^{n}\|,\quad\|\hat{X}^{n}\|\leq e^{\alpha\,t^{n}}\|\hat{X}^{0}\|,\quad\textrm{if}\,\,\alpha\Delta t\leq 1.$ ###### Proof. The proof is similar to the proof of Theorem 3.1, so we only highlight the difference. We have $\langle\hat{X}^{n+1},\hat{X}^{n+1}\rangle\leq\langle\hat{X}^{n},\hat{X}^{n+1}\rangle+\Delta t\alpha\|\hat{X}^{n+1}\|^{2},$ which implies $\|\hat{X}^{n+1}\|\leq\frac{1}{1-\alpha\Delta t}\|\hat{X}^{n}\|,$ if $\alpha\Delta t\leq 1.$ The theorem follows using a similar arguments as in Theorem 3.1. ∎ #### 3.3.2 Convergence A convergence estimate has been shown in Theorem 2 of [2] for the rank adaptive BUG scheme. Assuming the Lipschitz continuity and boundedness of the operator $F,$ the authors show that the BUG scheme has numerical error bounded by the sum of the initial numerical error, the tangent projection error, $\epsilon_{2}$ and first order in time error. The proof is based on the time continuous version of the BUG schemes. We want to point out that for PDE applications, in general the (differential) operators are not Lipschitz bounded. Nevertheless, it is still of theoretical interests to investigate the convergence properties of the schemes under such assumptions. For simplicity, below we will estimate the local truncation error of the first order implicit schemes with column and row space spanned by the cheap prediction space, i.e. we are investigating the Steps 1-4 in Algorithm 5. Since the Merge method uses a space that is union of the cheap prediction (as shown below) and the BUG space, the local truncation error of Algorithm 4 will be also be upper bounded by the estimate below. ###### Lemma 3.3 (Local truncation error with cheap prediction space). Suppose $F$ is Lipschitz-continuous in both variables and bounded, i.e. there exists a constant $L$ such that $\|F(Y,s)-F(Z,t)\|\leq L\|Y-Z\|+L|s-t|$ and $\|F(X,t)\|\leq B$. If we denote $\mathcal{X}$ as the exact solution to (1) with initial condition $\hat{X}^{n}=U^{n}\Sigma^{n}(V^{n})^{T}$ from $t^{n}$ to $t^{n+1}$, $\hat{X}^{n+1}$ the numerical solution obtained from from Algorithm 5, Steps 1-4 with the same initial condition $\hat{X}^{n},$ then for sufficiently small $\Delta t,$ we have the following error bound $\|\mathcal{X}-\hat{X}^{n+1}\|\leq C\Delta t^{2}+2\Delta t\epsilon_{1}+\epsilon_{2}.$ (6) In (6), $C$ is a constant that only depends $L$ and $B$. ###### Proof. First we prove (6) for $\epsilon_{1}=0$. Then $\mathcal{T}^{sum}_{\epsilon_{1}=0}(F(\hat{X}^{n},t^{n}))=F(\hat{X}^{n},t^{n})$. From (5), and left multiplying by $\tilde{U}$ and right multiplying by $\tilde{V}^{T},$ we obtain $\hat{X}^{n+1,pre}=\tilde{U}\tilde{\Sigma}\tilde{V}^{T}=\tilde{U}\tilde{U}^{T}(\hat{X}^{n}+\Delta tF(\hat{X}^{n+1,pre},t^{n+1})))\tilde{V}\tilde{V}^{T}.$ Since the column space of $\hat{X}^{n}$ is a subset of the column space of $\tilde{U},$ i.e. $R(\hat{X}^{n})\subseteq R(\tilde{U}),$ we have $\tilde{U}\tilde{U}^{T}\hat{X}^{n}=\hat{X}^{n}.$ Similarly, because $R((\hat{X}^{n})^{T})\subseteq R(\tilde{V}),$ $\hat{X}^{n}\tilde{V}\tilde{V^{T}}=\hat{X}^{n}.$ Therefore, $\tilde{U}\tilde{U}^{T}\hat{X}^{n}\tilde{V}\tilde{V^{T}}=\hat{X}^{n}.$ By a similar argument, $R(F(\hat{X}^{n},t^{n}))\subseteq R(\tilde{U})$ and $R(F(\hat{X}^{n},t^{n})^{T})\subseteq R(\tilde{V})$ implies $\tilde{U}\tilde{U}^{T}F(\hat{X}^{n},t^{n})\tilde{V}\tilde{V}^{T}=F(\hat{X}^{n},t^{n}).$ This gives $\displaystyle\hat{X}^{n+1,pre}=\hat{X}^{n}+\Delta t\tilde{U}\tilde{U}^{T}F(\hat{X}^{n+1,pre},t^{n+1})\tilde{V}\tilde{V}^{T}$ $\displaystyle=\hat{X}^{n}+\Delta t\tilde{U}\tilde{U}^{T}F(\hat{X}^{n},t^{n})\tilde{V}\tilde{V}^{T}+\Delta t\tilde{U}\tilde{U}^{T}(F(\hat{X}^{n+1,pre},t^{n+1})-F(\hat{X}^{n},t^{n}))\tilde{V}\tilde{V}^{T}$ $\displaystyle=\hat{X}^{n}+\Delta tF(\hat{X}^{n},t^{n})+\Delta t\tilde{U}\tilde{U}^{T}(F(\hat{X}^{n+1,pre},t^{n+1})-F(\hat{X}^{n},t^{n}))\tilde{V}\tilde{V}^{T}.$ (7) Therefore, $\|\hat{X}^{n+1,pre}-\mathcal{X}\|\\\ =\|\hat{X}^{n}-\mathcal{X}+\Delta tF(\hat{X}^{n},t^{n})+\Delta t\tilde{U}\tilde{U}^{T}(F(\hat{X}^{n+1,pre},t^{n+1})-F(\hat{X}^{n},t^{n}))\tilde{V}\tilde{V}^{T}\|.$ (8) We have $\hat{X}^{n}-\mathcal{X}=-\Delta tF(X(t^{*}),t^{*}),$ where $t^{*}$ is a point on the interval $[t^{n},t^{n+1}].$ Therefore, $\displaystyle\|\hat{X}^{n}-\mathcal{X}+\Delta tF(\hat{X}^{n},t^{n})\|=\Delta t\|-F(X(t^{*}),t^{*})+F(\hat{X}^{n},t^{n})\|$ $\displaystyle\leq\Delta tL(\Delta t+\|X(t^{*})-\hat{X}^{n}\|)\leq\Delta t^{2}L(B+1).$ (9) On the other hand, $\displaystyle\|\tilde{U}\tilde{U}^{T}(F(\hat{X}^{n+1,pre},t^{n+1})-F(\hat{X}^{n},t^{n}))\tilde{V}\tilde{V}^{T}\|$ $\displaystyle\leq\|\tilde{U}\tilde{U}^{T}\|_{2}\|F(\hat{X}^{n+1,pre},t^{n+1})-F(\hat{X}^{n},t^{n})\|\|\tilde{V}\tilde{V}^{T}\|_{2},$ where $\|\cdot\|_{2}$ is the matrix 2-norm. Since $\|\tilde{U}\tilde{U}^{T}\|_{2}=\|\tilde{V}\tilde{V}^{T}\|_{2}=1$ and $\|F(\hat{X}^{n+1,pre},t^{n+1})-F(\hat{X}^{n},t^{n})\|\leq L(\Delta t+\|\hat{X}^{n}-\mathcal{X}\|+\|\hat{X}^{n+1,pre}-\mathcal{X}\|)\leq L(\Delta t+B\Delta t+\|\hat{X}^{n+1,pre}-\mathcal{X}\|),$ we get $\displaystyle\|\tilde{U}\tilde{U}^{T}(F(\hat{X}^{n+1,pre},t^{n+1})-F(\hat{X}^{n},t^{n}))\tilde{V}\tilde{V}^{T}\|\leq L(\Delta t+B\Delta t+\|\hat{X}^{n+1,pre}-\mathcal{X}\|).$ (10) Combining (8),(3.3.2),(10), we get $\|\hat{X}^{n+1,pre}-\mathcal{X}\|\leq\Delta t^{2}L(2B+2)+L\Delta t\|\hat{X}^{n+1,pre}-\mathcal{X}\|,$ Choosing $\Delta t$ small enough (i.e. $\Delta t\leq\frac{1}{2L}$), we proved $\displaystyle\|\hat{X}^{n+1,pre}-\mathcal{X}\|\leq 4\Delta t^{2}L(B+1).$ (11) Therefore, (6) follows by using $\|\hat{X}^{n+1,pre}-\hat{X}^{n+1}\|=\|\hat{X}^{n+1,pre}-\mathcal{T}_{\epsilon_{2}}^{svd}(\hat{X}^{n+1,pre})\|\leq\epsilon_{2}.$ The proof for (6) with $\epsilon_{1}\neq 0$ largely follows the same process. The only difference is that we no longer have $\tilde{U}\tilde{U}^{T}F(\hat{X}^{n},t^{n})\tilde{V}\tilde{V}^{T}=F(\hat{X}^{n},t^{n})$ due to the truncation in the prediction step. Instead, we have $\tilde{U}\tilde{U}^{T}\mathcal{T}_{\epsilon_{1}}(F(\hat{X}^{n},t^{n}))\tilde{V}\tilde{V}^{T}=\mathcal{T}_{\epsilon_{1}}(F(\hat{X}^{n},t^{n})).$ In this case, $\tilde{U}\tilde{U}^{T}F(\hat{X}^{n},t^{n})\tilde{V}\tilde{V}^{T}=\mathcal{T}_{\epsilon_{1}}(F(\hat{X}^{n},t^{n}))+\tilde{U}\tilde{U}^{T}(F(\hat{X}^{n},t^{n})-\mathcal{T}_{\epsilon_{1}}(F(\hat{X}^{n},t^{n})))\tilde{V}\tilde{V}^{T},$ and because $\|\tilde{U}\tilde{U}^{T}\|_{2}=\|\tilde{V}\tilde{V}^{T}\|_{2}=1,$ this leads to $\|\tilde{U}\tilde{U}^{T}F(\hat{X}^{n},t^{n})\tilde{V}\tilde{V}^{T}-F(\hat{X}^{n},t^{n})\|\leq 2\|F(\hat{X}^{n},t^{n})-\mathcal{T}_{\epsilon_{1}}(F(\hat{X}^{n},t^{n}))\|=2\epsilon_{1}.$ By a similar argument, (6) follows. The details are omitted for brevity. ∎ The numerical convergence follows from Lemma 3.3 by a standard result for the convergence of one-step method for ODE and is skipped for brevity. Now we would like to comment on the results in this Lemma to understand property of the scheme. As we can see from the results in Section 3.3.1, the schemes are unconditionally stable due to the Galerkin step. Also, in Lemma 3.3, the tangent projection error is absent because the DLRA approximation is not used. However, the time step is limited by the Lipschitz constant and also the constant in the estimate $C$ depends on the Lipschitz constant. For standard error estimates for implicit schemes for stiff problems, the concept of B-convergence [5] can remove such restrictions with the help of one-sided Lipschitz continuity. Unfortunately, such results are not available in our case. ## 4 Numerical results We now present numerical results that illustrate the features of the methods described above. In this section we will refer to the different methods in figures and tables according to the following naming convention. We will denote the Merge method by M, and the Merge-adapt method by MA. For many problems we will also compare to the classic implicit Euler discretization as IE or Implicit Euler. When computing errors we will compare to a solution computed using Matlab’s built in ODE15s or ODE45 solvers with relative and absolute tolerance set to $10^{-12}$. We always report relative errors in the Frobenius norm. We note that in all of the examples below, the BUG method does not converge since all or part of the right hand side of (1) is outside the tangent projection space. We also note that in all the examples below we use $\epsilon_{1}=0$ but that the results are very similar for $\epsilon_{1}=\Delta t$. The main PDE examples used in this section are simulations of solutions to the equation $\frac{\partial\rho}{\partial t}+r_{1}(x_{1})\frac{\partial\rho}{\partial x_{2}}+r_{2}(x_{2})\frac{\partial\rho}{\partial x_{1}}=\\\ b_{1}(x_{2})\frac{\partial}{\partial x_{1}}\left[a_{1}(x_{1})\frac{\partial\rho}{\partial x_{1}}\right]+b_{2}(x_{2})\frac{\partial^{2}\left[a_{2}(x_{1})\rho\right]}{\partial x_{1}\partial x_{2}}+a_{3}(x_{1})\frac{\partial^{2}\left[b_{3}(x_{2})\rho\right]}{\partial x_{1}\partial x_{2}}+a_{4}(x)\frac{\partial}{\partial y}\left[b_{4}(y)\frac{\partial\rho}{\partial y}\right].$ (12) Here $\rho=\rho(t,x_{1},x_{2})$ and we always use homogenous Dirichlet boundary conditions. Equation (12) is discretized on the domain $(x_{1},x_{2})\in[-1,1]^{2},$ using second order accurate finite differences. Consider the grid $(x_{1,i},x_{2,j})=(-1+ih_{1},-1+jh_{2})$, with $h_{1}=2/(m_{1}+1),$ $h_{2}=2/(m_{2}+1).$ The solution $\rho(t_{n},x_{1,i},x_{2,j})$ is then approximated by the grid function (matrix) $\hat{X}^{n}_{i,j}$ and derivatives are approximated as follow: $\frac{\partial}{\partial x_{1}}\left[a_{1}(x_{1})\frac{\partial\rho(t_{n},x_{1},x_{2})}{\partial x_{1}}\right]\Bigg{|}_{x_{1,i},x_{2,j}}\approx D_{+}^{1}\left[\frac{a_{1}(x_{1,i})+a_{1}(x_{1,i-1})}{2}\right]D_{-}^{1}\,X_{i,j}^{n},$ $\frac{\partial}{\partial x_{1}}\left[a_{2}(x_{1})\frac{\partial\rho(t_{n},x_{1},x_{1})}{\partial x_{2}}\right]\Bigg{|}_{x_{1,i},x_{2,j}}\approx D_{0}^{1}a_{1}(x_{1,i})D_{0}^{2}\,X_{i,j}^{n}.$ with the remaining two terms discretized in the same way. Here the difference operators in the 1-direction are defined $2h_{1}D_{0}^{1}w_{i,j}\equiv w_{i+1,j}-w_{i-1,j},\ \ h_{1}D_{+}^{1}w_{i,j}\equiv w_{i+1,j}-w_{i,j},\ \ h_{1}D_{-}^{1}w_{i,j}\equiv w_{i,j}-w_{i-1,j},$ with the operators in the 2-direction defined analogously. ### 4.1 Rotation with anisotropic diffusion In this experiment we consider solid body rotation, $r_{1}(x_{1})=x_{1}$ $r_{2}(x_{2})=-x_{2}$, with anisotropic diffusion. The diffusion coefficients are: $\displaystyle a_{1}(x_{1})=a_{4}(x_{1})=\sqrt{\mu}(1+0.1\sin(\pi x_{1})),$ $\displaystyle a_{2}(x_{1})=\sqrt{\mu}(0.15+0.1\sin(\pi x_{1})),$ $\displaystyle a_{3}(x_{1})=\sqrt{\mu}(0.15+0.1\cos(\pi x_{1})),$ $\displaystyle b_{1}(x_{2})=b_{4}(x_{2})=\sqrt{\mu}(1+0.1\cos(\pi x_{2})),$ $\displaystyle b_{2}(x_{2})=\sqrt{\mu}(0.15+0.1\cos(\pi x_{2})),$ $\displaystyle b_{3}(x_{2})=\sqrt{\mu}(0.15+0.1\sin(\pi x_{2})),$ and $\mu=10^{-3}$. We start from the rank-1 initial data $\rho(0,x_{1},x_{2})=e^{-\left(\frac{x_{1}}{0.3}\right)^{2}}e^{-\left(\frac{x_{2}}{0.1}\right)^{2}}$ and evolve the solution until time $t=\pi$ using a timestep $\Delta t=\pi/n_{T}$, with $40,80,160,320$. We carry out the computation on three different grids with $m_{1}=m_{2}=99,199,799$. In Figure 1 we display 10 equidistant contours between 0.1 and 0.9 of the solution. Results using the Merge method (left), the classic implicit Euler (middle) and the BUG method (right) are presented. The solutions are displayed at time $0.5\pi$ when they have rotated from being horizontal to vertical. The BUG method does not see the rotation (it is outside the tangent space) and remains stationary. The solution computed using the Merge method rotates as expected and is almost identical to the classical implicit Euler solution. Figure 1: Displayed are 10 equidistant contours between 0.1 and 0.9 of the solution computed using the Merge method (left), the classic implicit Euler (middle) and the BUG method (right). The solutions are displayed at time $0.5\pi$ when they have rotated from being horizontal to vertical. The BUG method does not see the rotation (it is outside the tangent space) and remains stationary. The solution computed using the Merge method rotates as expected and is almost identical to the classical implicit Euler solution. $n_{T}$ | M | MA | IE | $\Delta t/h$ : $\mu\Delta t/h^{2}$ ---|---|---|---|--- 40 | 1.60(-1) | 1.60(-1) F = 5 | 1.60(-1) | 3.9 : 0.2 80 | 1.01(-1) [0.65] | 1.01(-1) [0.65] F = 3 | 1.01(-1) [0.65] | 2 : 0.098 160 | 6.01(-2) [0.75] | 6.01(-2) [0.75] F = 4 | 6.01(-2) [0.75] | 0.98 : 0.049 320 | 3.36(-2) [0.83] | 3.36(-2) [0.83] F = 0 | 3.36(-2) [0.83] | 0.49 : 0.025 40 | 1.60(-1) | 1.60(-1) F = 8 | 1.60(-1) | 7.9 : 0.79 80 | 1.02(-1) [0.65] | 1.02(-1) [0.65] F = 6 | 1.02(-1) [0.65] | 3.9 : 0.39 160 | 6.07(-2) [0.75] | 6.07(-2) [0.75] F = 4 | 6.07(-2) [0.75] | 2 : 0.2 320 | 3.40(-2) [0.83] | 3.40(-2) [0.83] F = 1 | 3.40(-2) [0.83] | 0.98 : 0.098 40 | 1.61(-1) | 1.61(-1) F = 13 | 1.61(-1) | 31 : 13 80 | 1.02(-1) [0.65] | 1.02(-1) [0.65] F = 11 | 1.02(-1) [0.65] | 16 : 6.3 160 | 6.09(-2) [0.75] | 6.09(-2) [0.75] F = 11 | 6.09(-2) [0.75] | 7.9 : 3.1 320 | 3.41(-2) [0.83] | 3.41(-2) [0.83] F = 9 | 3.41(-2) [0.83] | 3.9 : 1.6 Table 1: Solid body rotation with anisotropic diffusion. Displayed are the errors (here 1.3(-1) means $1.3\cdot 10^{-1}$) for different timesteps along with estimated rates of convergence (in brackets). The numbers on the far right are the “hyperbolic and parabolic CFL numbers”. The top box is for $m_{1}=m_{2}=99$, the middle for $m_{1}=m_{2}=199$ and the bottom for $m_{1}=m_{2}=799$. The F indicates how many times the Merge-adapt needed to add the BUG spaces. In Table 1 we display the errors and rates of convergence for the Merge, Merge-adapt and the implicit Euler method. All three methods behave almost identically (the errors and rates of convergence only differ in the third or fourth digit). However, the time it takes to compute the solution is quite different. In Table 2 we display the CPU times for the different methods. The computations were performed in Matlab on a MacBook Pro with an M2 chip and 16GB RAM. Reasonable effort went into optimizing the codes but the timings should be seen as indicative rather than decisive. As can be seen the growth in CPU time as a function of number of degrees of freedom is very mild for the Merge and Merge-adapt methods in comparison to the classic implicit Euler method. We further note that the Merge-adapt method has comparable cost with the Merge method in this example because the equation involves relatively complex dynamics, which means the explicit prediction is quite costly because of the large number $s.$ $m_{1}=m_{2}$ | M time$[s]$ | MA time$[s]$ | IE time$[s]$ ---|---|---|--- 99 | 0.9 | 0.9 | 5.0 199 | 1.4 | 1.5 | 27.1 799 | 12.8 | 11.4 | 926.2 Table 2: Displayed are the CPU times in seconds for the rotation with anisotropic diffusion example. The results are all for $n_{T}=320$. As can be seen the growth in CPU time as a function of number of degrees of freedom is very mild for the Merge and Merge-adapt methods in comparison to the classic implicit Euler method. We also track the rank of the solution as a function of time. In Figure 2 we display the rank for the different methods as a function of time. Once the solution becomes more diagonal at time $\pi/4$ the rank is decreased and then increased again as the rotation continues. For the Merge and Merge-adapt method we use the truncation tolerance $\epsilon_{2}=\Delta t^{2}$ and in order to compare with the classic implicit Euler we compute the SVD and count the rank of that solution using the same threshold. We observe the ranks from the low rank methods are slightly smaller than the implicit Euler method. Figure 2: Displayed is the rank for the Merge, Merge-adapt and Implicit Euler method for the problem with rotation and anisotropic diffusion. Here “Fail for MA” is the indicator where the BUG space is needed for the Merge-adapt method. For the Implicit Euler method we constantly use the threshold $\Delta t^{2}$ when computing the rank via the truncated SVD. All the computations are done with $m_{1}=m_{2}=799$ and $n_{T}=320.$ ### 4.2 Anisotropic diffusion In this experiment we consider anisotropic diffusion without rotation ($r_{1}=r_{2}=0$). We take the diffusion coefficients to be constants with $a_{1}=a_{4}=b_{1}=b_{4}=1$ and $a_{2}=a_{3}=b_{2}=b_{3}=0.3$. Starting from the rank-1 initial data $\rho(0,x_{1},x_{2})=\sin(\pi x_{1})\sin(\pi x_{2}),$ we evolve the solution until time $t=0.5$ using a base timestep of $\Delta t=0.5/n_{T}$, with $40,80,160,320,640,1280$. We carry out the computation on three different grids with $m_{1}=m_{2}=99,399,799$. $n_{T}$ | M | MA | IE | $\Delta t/h$ : $\Delta t/h^{2}$ ---|---|---|---|--- 40 | 8.65(-2) | 8.65(-2) F = 37 | 9.31(-2) | 0.62 : 31 80 | 2.84(-2) [1.60] | 2.84(-2) [1.60] F = 70 | 4.39(-2) [1.08] | 0.31 : 16 160 | 9.94(-3) [1.51] | 9.94(-3) [1.51] F = 160 | 2.13(-2) [1.03] | 0.16 : 7.8 320 | 5.89(-3) [0.75] | 5.89(-3) [0.75] F = 296 | 1.05(-2) [1.01] | 0.078 : 3.9 640 | 3.78(-3) [0.63] | 3.78(-3) [0.63] F = 595 | 5.22(-3) [1.01] | 0.039 : 2 1280 | 2.19(-3) [0.78] | 2.19(-3) [0.78] F = 989 | 2.58(-3) [1.01] | 0.02 : 0.98 40 | 1.13(-1) | 1.13(-1) F = 40 | 9.30(-2) | 1.2 : 120 80 | 4.23(-2) [1.42] | 4.23(-2) [1.42] F = 80 | 4.38(-2) [1.08] | 0.62 : 62 160 | 1.33(-2) [1.66] | 1.33(-2) [1.66] F = 160 | 2.13(-2) [1.03] | 0.31 : 31 320 | 4.64(-3) [1.52] | 4.64(-3) [1.52] F = 320 | 1.05(-2) [1.01] | 0.16 : 16 640 | 2.88(-3) [0.69] | 2.88(-3) [0.69] F = 640 | 5.22(-3) [1.01] | 0.078 : 7.8 1280 | 1.87(-3) [0.61] | 1.87(-3) [0.61] F = 1280 | 2.58(-3) [1.01] | 0.039 : 3.9 40 | 1.03(-1) | 1.03(-1) F = 40 | 9.30(-2) | 5 : 2000 80 | 6.52(-2) [0.66] | 6.52(-2) [0.66] F = 80 | 4.38(-2) [1.08] | 2.5 : 1000 160 | 2.70(-2) [1.27] | 2.70(-2) [1.27] F = 160 | 2.13(-2) [1.03] | 1.2 : 500 320 | 9.69(-3) [1.48] | 9.69(-3) [1.48] F = 320 | 1.05(-2) [1.02] | 0.62 : 250 640 | 2.97(-3) [1.70] | 2.97(-3) [1.70] F = 640 | 5.21(-3) [1.01] | 0.31 : 1.25 1280 | 1.01(-3) [1.55] | 1.01(-3) [1.55] F = 1280 | 2.57(-3) [1.01] | 0.16 : 62 Table 3: Anisotropic diffusion. Displayed are the errors for different timesteps along with estimated rates of convergence (in brackets). The numbers on the far right are the “hyperbolic and parabolic CFL numbers”. The top box is for $m_{1}=m_{2}=99$, the middle for $m_{1}=m_{2}=199$ and the bottom for $m_{1}=m_{2}=799$. The F indicates how many times the Merge-adapt needed to add the BUG spaces. Figure 3: Displayed is the rank for the Merge, Merge-adapt and Implicit Euler method for the problem with anisotropic diffusion initial data $\sin(\pi x_{1})\sin(\pi x_{2}),$ (left) and $\sin(2\pi x_{1})\sin(2\pi x_{2})$ (right) and the solid body rotation problem (right). Here “Fail for MA” is the indicator where the BUG space is needed for the Merge-adapt method. For the Implicit Euler method we constantly use the threshold $\Delta t^{2}$ when computing the rank via the truncated SVD. All the computations are done with $m_{1}=m_{2}=799$ and $n_{T}=320.$ In Table 3 we display the errors and rates of convergence. Again, the methods are comparable although it should be noted that the rates of convergence for the classic implicit Euler method is more uniform than for the low rank methods. Curiously the error levels are smaller for the low rank method. It should also be noted that for this example the Merge-adapt method requires the addition of the BUG spaces almost every time step due to the stiffness of the problem. We also track the rank of the solution as a function of time, the results can be found in Figure 3. The rank evolution is very similar for all methods. We also consider the more oscillatory initial data $\rho(0,x_{1},x_{2})=\sin(2\pi x_{1})\sin(2\pi x_{2}),$ which we again evolve until time $t=0.5$ using a base timestep of $\Delta t=0.5/n_{T}$, with $40,80,160,320,640,1280$. Again, we carry out the computation on three different grids with $m_{1}=m_{2}=99,399,799$. This initial data decays more rapidly in time and as a consequence the residual norm becomes smaller and the Merge-adapt method does not have to add the BUG spaces as often. The errors and rates of convergence can be found in Table 4, and the rank of the solution as a function of time, the results can be found in Figure 2. It can be observed that for this example the classic implicit Euler method at first is a bit more accurate than the low rank methods but they catch up as the timestep decreases. $n_{T}$ | M | MA | IE | $\Delta t/h$ : $\Delta t/h^{2}$ ---|---|---|---|--- 40 | 4.26(-1) | 4.30(-1) F = 18 | 7.58(-2) | 0.62 : 31 80 | 2.07(-1) [1.03] | 2.08(-1) [1.04] F = 35 | 3.80(-2) [0.99] | 0.31 : 16 160 | 7.90(-2) [1.39] | 7.96(-2) [1.38] F = 61 | 1.91(-2) [0.99] | 0.16 : 7.8 320 | 2.36(-2) [1.74] | 2.42(-2) [1.71] F = 104 | 9.67(-3) [0.98] | 0.078 : 3.9 640 | 5.14(-3) [2.19] | 5.15(-3) [2.23] F = 188 | 4.97(-3) [0.95] | 0.039 : 2 1280 | 1.00(-3) [2.35] | 1.01(-3) [2.34] F = 334 | 2.68(-3) [0.89] | 0.02 : 0.98 40 | 4.87(-1) | 4.89(-1) F = 20 | 7.57(-2) | 1.2 : 125 80 | 2.62(-1) [0.89] | 2.63(-1) [0.89] F = 40 | 3.80(-2) [0.99] | 0.62 : 62 160 | 1.13(-1) [1.20] | 1.13(-1) [1.21] F = 81 | 1.90(-2) [0.99] | 0.31 : 31 320 | 3.89(-2) [1.54] | 3.89(-2) [1.54] F = 134 | 9.60(-3) [0.98] | 0.16 : 16 640 | 1.05(-2) [1.87] | 1.06(-2) [1.87] F = 277 | 4.87(-3) [0.97] | 0.078 : 7.8 1280 | 2.01(-3) [2.39] | 2.01(-3) [2.39] F = 521 | 2.52(-3) [0.95] | 0.039 : 3.9 40 | 5.23(-1) | 5.23(-1) F = 39 | 7.58(-2) | 5 : 2000 80 | 3.49(-1) [0.58] | 3.49(-1) [0.58] F = 75 | 3.80(-2) [0.99] | 2.5 : 1000 160 | 1.80(-1) [0.95] | 1.80(-1) [0.95] F = 132 | 1.91(-2) [0.99] | 1.2 : 500 320 | 7.71(-2) [1.22] | 7.71(-2) [1.22] F = 242 | 9.71(-3) [0.97] | 0.62 : 250 640 | 2.89(-2) [1.41] | 2.89(-2) [1.41] F = 479 | 5.04(-3) [0.94] | 0.31 : 125 1280 | 9.17(-3) [1.65] | 9.17(-3) [1.65] F = 1162 | 2.78(-3) [0.85] | 0.16 : 62 Table 4: Anisotropic diffusion for a higher frequency initial data. Displayed are the errors for different timesteps along with estimated rates of convergence (in brackets). The numbers on the far right are the “hyperbolic and parabolic CFL numbers”. The top box is for $m_{1}=m_{2}=99$, the middle for $m_{1}=m_{2}=199$ and the bottom for $m_{1}=m_{2}=799$. The F indicates how many times the Merge-adapt needed to add the BUG spaces. ### 4.3 Solid body rotation with isotropic diffusion $m_{1}=m_{2}$ | time MA [s] | time M [s] | $\Delta t/h$ : $\mu\Delta t/h^{2}$ ---|---|---|--- 999 | 1.9 [0.1%] | 3.1 | 0.39 : 0.020 1999 | 3.5 [0.1%] | 6.0 | 0.79 : 0.79 3999 | 7.1 [0.4%] | 15.5 | 1.57 : 0.31 7999 | 52.8 [72%] | 46.2 | 3.14 : 1.26 Table 5: CPU-times for the problem with solid body rotation with isotropic diffusion. The numbers in brackets are the percentages for how often MA has a residual that necessitates the addition of the BUG space. In this experiment we time a problem with solid body rotation and isotropic diffusion, corresponding to, $r_{1}(x_{1})=x_{1}$ $r_{2}(x_{2})=-x_{2}$, $a_{1}=b_{1}=a_{4}=b_{4}=10^{-4}$, and $a_{j}=b_{j}=0,\,j=2,\ldots,3$. We focus on the comparison of the computational cost of Merge and Merge-adapt schemes for large mesh size. We start from the rank-1 initial data $\rho(0,x_{1},x_{2})=e^{-\left(\frac{x_{1}}{0.3}\right)^{2}}e^{-\left(\frac{x_{2}}{0.1}\right)^{2}}$ and evolve the solution until time $t=0.25\pi$ using a base timestep of $\Delta t=0.25\pi/n_{T}$, with $n_{T}=1000$. We carry out the computation on very fine grids with $m_{1}=m_{2}=999,1999,3999,7999$. Here the size of the diffusion is chosen so that the ratios $\Delta t/h$ : $\mu\Delta t/h^{2}$ change from smaller than one to larger than one. We report the CPU times for the M and MA methods in Table 5. As can be seen the MA is faster than the M method when the number of timesteps where the BUG space is used is small. The more intensive use of the BUG space appears to happen when the ratio $\mu\Delta t/h^{2}$ increases above one. ### 4.4 Solid body rotation In this experiment we consider a hyperbolic problem with pure solid body rotation, corresponding to, $r_{1}(x_{1})=x_{1}$ $r_{2}(x_{2})=-x_{2}$ and $a_{j}=b_{j}=0,\,j=1,\ldots,4$. We start from the rank-1 initial data $\rho(0,x_{1},x_{2})=e^{-\left(\frac{x_{1}}{0.3}\right)^{2}}e^{-\left(\frac{x_{2}}{0.1}\right)^{2}}$ and evolve the solution until time $t=\pi$ using a base timestep of $\Delta t=\pi/n_{T}$, with $n_{T}=40,80,160,320$. We carry out the computation on three different grids with $m_{1}=m_{2}=99,199,799$. The errors are listed in Table 6. Again similar conclusions can be drawn, namely all three methods give comparable errors. The rank evolution is reported in Figure 4. In this problem, the rank evolution exhibits periodic pattern in time due to the solution structure. The rank of the low rank methods are lower than the implicit Euler method. $n_{T}$ | M | MA | IE | $\Delta t/h$ ---|---|---|---|--- 40 | 2.47(-1) | 2.47(-1) F = 11 | 2.51(-1) | 3.9 80 | 1.71(-1) [0.53] | 1.71(-1) [0.53] F = 19 | 1.73(-1) [0.53] | 2 160 | 1.10(-1) [0.63] | 1.10(-1) [0.63] F = 4 | 1.10(-1) [0.64] | 0.98 320 | 6.57(-2) [0.74] | 6.57(-2) [0.74] F = 4 | 6.60(-2) [0.74] | 0.49 40 | 2.50(-1) | 2.50(-1) F = 14 | 2.54(-1) | 7.9 80 | 1.74(-1) [0.52] | 1.74(-1) [0.52] F = 27 | 1.76(-1) [0.53] | 3.9 160 | 1.12(-1) [0.62] | 1.12(-1) [0.62] F = 32 | 1.13(-1) [0.63] | 2 320 | 6.79(-2) [0.73] | 6.79(-2) [0.73] F = 12 | 6.81(-2) [0.73] | 0.98 40 | 2.51(-1) | 2.51(-1) F = 24 | 2.55(-1) | 31 80 | 1.75(-1) [0.52] | 1.75(-1) [0.52] F = 41 | 1.77(-1) [0.52] | 16 160 | 1.13(-1) [0.62] | 1.13(-1) [0.62] F = 61 | 1.14(-1) [0.63] | 7.9 320 | 6.85(-2) [0.72] | 6.85(-2) [0.72] F = 117 | 6.88(-2) [0.73] | 3.9 Table 6: Solid body rotation. Displayed are the errors for different timesteps along with estimated rates of convergence (in brackets). The top box is for $m_{1}=m_{2}=99$, the middle for $m_{1}=m_{2}=199$ and the bottom for $m_{1}=m_{2}=799$. The F indicates how many times the Merge-adapt needed to add the BUG spaces. Figure 4: Displayed is the rank for the Merge, Merge- adapt and Implicit Euler method for the problem with solid body rotation problem. Here “Fail for MA” is the indicator where the BUG space is needed for the Merge-adapt method. For the Implicit Euler method we constantly use the threshold $\Delta t^{2}$ when computing the rank via the truncated SVD. All the computations are done with $m_{1}=m_{2}=799$ and $n_{T}=320.$ ## 5 Conclusions and future work This work prototypes a class of implicit adaptive low rank time-stepping schemes. By simply merging the explicit step truncation and the BUG spaces, the modeling error is removed and the numerical schemes achieve robust convergence upon mesh refinement. An adaptive strategy is proposed for the prediction of row and column spaces, which is computationally advantageous for moderately stiff problems. The immediate future work is the generalization to higher order in time. More importantly, the ideas can be applied to tensor differential equations, which will be investigated in the future. ## 6 Acknowledgement We would like to thank Steven R. White for pointing out the reference [24]. ## References * [1] M. Bachmayr. Low-rank tensor methods for partial differential equations. Acta Numerica, 32:1–121, 2023. * [2] G. Ceruti, J. Kusch, and C. Lubich. A rank-adaptive robust integrator for dynamical low-rank approximation. BIT Numerical Mathematics, 62(4):1149–1174, 2022. * [3] G. Ceruti and C. Lubich. An unconventional robust integrator for dynamical low-rank approximation. BIT Numerical Mathematics, 62(1):23–44, 2022. * [4] A. Dektor, A. Rodgers, and D. Venturi. Rank-adaptive tensor methods for high-dimensional nonlinear PDEs. Journal of Scientific Computing, 88(2):36, 2021. * [5] R. Frank, J. Schneid, and C. W. Ueberhuber. The concept of B-convergence. SIAM Journal on Numerical Analysis, 18(5):753–780, 1981. * [6] G. H. Golub and C. F. Van Loan. Matrix computations. JHU press, 2013. * [7] L. Grasedyck, D. Kressner, and C. Tobler. A literature survey of low-rank tensor approximation techniques. GAMM-Mitteilungen, 36(1):53–78, 2013. * [8] W. Guo and J.-M. Qiu. A low rank tensor representation of linear transport and nonlinear vlasov solutions and their associated flow maps. Journal of Computational Physics, 458:111089, 2022. * [9] B. Gustafsson, H.-O. Kreiss, and J. Oliger. Time dependent problems and difference methods, volume 2. Wiley Online Library, 2013. * [10] W. Hackbusch. Tensor spaces and numerical tensor calculus, volume 42. Springer, 2012. * [11] C. D. Hauck and S. Schnake. A predictor-corrector strategy for adaptivity in dynamical low-rank approximations. SIAM Journal on Matrix Analysis and Applications, 44(3):971–1005, 2023. * [12] M. Hochbruck, M. Neher, and S. Schrammer. Rank-adaptive dynamical low-rank integrators for first-order and second-order matrix differential equations. BIT Numerical Mathematics, 63(1):9, 2023. * [13] B. N. Khoromskij. Tensor numerical methods in scientific computing, volume 19. Walter de Gruyter GmbH & Co KG, 2018. * [14] E. Kieri and B. Vandereycken. Projection methods for dynamical low-rank approximation of high-dimensional problems. Computational Methods in Applied Mathematics, 19(1):73–92, 2019\. * [15] O. Koch and C. Lubich. Dynamical low-rank approximation. SIAM Journal on Matrix Analysis and Applications, 29(2):434–454, 2007. * [16] C. Lubich and I. V. Oseledets. A projector-splitting integrator for dynamical low-rank approximation. BIT Numerical Mathematics, 54(1):171–188, 2014. * [17] C. Lubich, I. V. Oseledets, and B. Vandereycken. Time integration of tensor trains. SIAM Journal on Numerical Analysis, 53(2):917–941, 2015. * [18] C. Lubich, B. Vandereycken, and H. Walach. Time integration of rank-constrained tucker tensors. SIAM Journal on Numerical Analysis, 56(3):1273–1290, 2018. * [19] J. Nakao, J.-M. Qiu, and L. Einkemmer. Reduced augmentation implicit low-rank (RAIL) integrators for advection-diffusion and Fokker-Planck models. 11 2023. * [20] S. Paeckel, T. Köhler, A. Swoboda, S. R. Manmana, U. Schollwöck, and C. Hubig. Time-evolution methods for matrix-product states. Annals of Physics, 411:167998, 2019. * [21] A. Rodgers and D. Venturi. Implicit integration of nonlinear evolution equations on tensor manifolds. Journal of Scientific Computing, 97(2):33, 2023. * [22] Y. Saad. Iterative methods for sparse linear systems. SIAM, 2003. * [23] V. Simoncini. Computational methods for linear matrix equations. SIAM Review, 58(3):377–441, 2016. * [24] M. Yang and S. R. White. Time-dependent variational principle with ancillary krylov subspace. Phys. Rev. B, 102:094315, Sep 2020.
University at Albany, State University of New York, USA jmcurry@albany.eduhttps://orcid.org/ 0000-0003-2504-8388Supported by NSF CCF-1850052 and NASA 80GRC020C0016 Florida State University, Tallahassee, <EMAIL_ADDRESS>by NSF grant DMS-1722995 Florida State University, Tallahassee, Floridatneedham@fsu.eduhttps://orcid.org/0000-0001-6165-3433Supported by NSF DMS-2107808 Max Planck Institute for Mathematics in the Sciences, Leipzig, <EMAIL_ADDRESS>Graz University of Technology, <EMAIL_ADDRESS>by the Austrian Science Fund (FWF): W1230 Justin Curry, Washington Mio, Tom Needham, Osman Okutan, Florian Russold[300]Mathematics of computing Algebraic topology [300]Theory of computation Computational geometry # Convergence of Leray Cosheaves for Decorated Mapper Graphs Justin Curry Washington Mio Tom Needham Osman Berat Okutan Florian Russold111Corresponding Author ###### Abstract We introduce decorated mapper graphs as a generalization of mapper graphs capable of capturing more topological information of a data set. A decorated mapper graph can be viewed as a discrete approximation of the cellular Leray cosheaf over the Reeb graph. We establish a theoretical foundation for this construction by showing that the cellular Leray cosheaf with respect to a sequence of covers converges to the actual Leray cosheaf as the resolution of the covers goes to zero. ###### keywords: Leray cosheaves, Reeb graphs, Mapper, convergence ## 1 Introduction Reeb11footnotetext: This is an abstract of a presentation given at CG:YRF 2023. It has been made public for the benefit of the community and should be considered a preprint rather than a formally reviewed paper. Thus, this work is expected to appear in a conference with formal proceedings and/or in a journal. graphs and their discrete analogs—mapper graphs—are important tools in computational topology [3, 14, 8, 2]. They are used for data visualization [12, 10], for comparing scalar fields via distances between Reeb graphs [7, 1, 6], and data skeletonization [9], among other things. Given a continuous map $f\colon X\rightarrow\mathbb{R}$, the Reeb graph $R_{f}$ summarizes the zero- dimensional connectedness of $X$. In practice, we deal with maps $f\colon P\rightarrow\mathbb{R}$ on discrete data sets $P$, where Reeb graphs are replaced by mapper graphs. A mapper graph is constructed by choosing a cover $\mathcal{U}=(U_{i})_{i\in I}$ of $f(X)$ or $f(P)$, taking the components or clustering the data points of $f^{\minus 1}(U_{i})$, collapsing each of the obtained components to a single vertex and connecting two vertices if their corresponding components have common points. In [6, 11] it is shown that Reeb graphs can be viewed as cosheaves and that the cellular Reeb cosheaf w.r.t. a cover converges to the Reeb cosheaf if the resolution of the cover goes to zero, establishing a theoretical justification for working with finite covers. A limitation of this approach is that Reeb/mapper graphs fail to capture higher-dimensional topological features. To overcome these limitations we introduce decorated mapper graphs. Figure 1: Pulling back the cover $\mathcal{U}$ of $f(X)$ along the induced map $\hat{f}$ on the Reeb graph and refining it into components yields a cover $\hat{\mathcal{U}}$ of $R_{f}$. The decorated mapper graph is the nerve complex of this cover decorated by the homology $H_{\bullet}\big{(}\pi_{f}^{\minus 1}(-)\big{)}$ of the preimages of intersections in $\hat{\mathcal{U}}$ under the quotient map $\pi_{f}$. Since $f=\hat{f}\circ\pi_{f}$, this is analogous to using components of preimages under $f$. A decorated mapper graph collects the homology (possibly inferred from a point sample) in all degrees $H_{\bullet}(-)\coloneqq\bigoplus_{n\geq 0}H_{n}(-)$ of the components of $f^{\minus 1}(U_{i})$ and $f^{\minus 1}(U_{i}\cap U_{j})$ on the corresponding vertices and edges of the mapper graph; see Figure 1. It can be viewed as a discrete approximation of the Leray cosheaf over the Reeb graph. In this abstract we show that the cellular Leray cosheaf w.r.t. a cover converges to the actual Leray cosheaf for any finite 1D cover $\mathcal{U}$ as the resolution of the cover goes to zero. ## 2 Leray Cosheaves The Leray (pre)cosheaf [4, 5, 15] parameterizes the homology of a space $X$ when viewed along a continuous map $f\colon X\rightarrow Y$. It does this by recording for each pair of open subsets $V\subseteq W\subseteq Y$ their homology and the induced map $H_{\bullet}\big{(}f^{\minus 1}(V)\big{)}\xrightarrow{H_{\bullet}(f^{\minus 1}(V)\subseteq f^{\minus 1}(W))}H_{\bullet}\big{(}f^{\minus 1}(W)\big{)}$. We further assume that $X$ is compact and $H_{\bullet}\big{(}f^{\minus 1}(V)\big{)}$ is finite- dimensional for every $V\subseteq Y$. ###### Definition 2.1 (Leray (pre)cosheaf). We define the graded Leray (pre)cosheaf $\mathcal{L}^{f}$ of a continuous map $f\colon X\rightarrow Y$ by the following assignments: For all $V\subseteq W\subseteq Y$ open $\displaystyle\mathcal{L}^{f}(V)=\bigoplus_{n\in\mathbb{N}_{0}}\mathcal{L}^{f}_{n}(V)$ $\displaystyle\coloneqq\bigoplus_{n\in\mathbb{N}_{0}}H_{n}\big{(}f^{\minus 1}(V)\big{)}$ $\displaystyle\mathcal{L}^{f}(V\subseteq W)=\bigoplus_{n\in\mathbb{N}_{0}}\mathcal{L}^{f}_{n}(V\subseteq W)$ $\displaystyle\coloneqq\bigoplus_{n\in\mathbb{N}_{0}}H_{n}\big{(}f^{\minus 1}(V)\subseteq f^{\minus 1}(W)\big{)}\hskip 2.0pt.$ In practice, we can not access the whole Leray cosheaf. We can only get a cellular version [4, 4.1.6] given by its values on a finite cover of $f(X)$. An open cover $\mathcal{U}=(U_{i})_{i\in I}$ of $f(X)$ defines a simplicial complex $\mathcal{N}_{\mathcal{U}}\coloneqq\\{\sigma=(i_{0},\ldots,i_{k})\text{}|\text{ }\mathbf{U}_{\sigma}\coloneqq U_{i_{0}}\cap\ldots\cap U_{i_{k}}\neq\emptyset\\}$, called the nerve complex of $\mathcal{U}$. We call $\mathcal{U}$ a finite 1D cover if it is finite and has a one-dimensional nerve complex and denote by $\leq$ the face-relation of $\mathcal{N}_{\mathcal{U}}$, i.e. $\sigma\leq\tau\iff\sigma\text{ is a face of }\tau$ . ###### Definition 2.2 (Cellular Leray cosheaf). We define the cellular Leray cosheaf $D_{\mathcal{U}}\mathcal{L}^{f}$ w.r.t. a cover $\mathcal{U}$ as the cosheaf on $\mathcal{N}_{\mathcal{U}}$ given by the following assignments: For all $\sigma\leq\tau\in\mathcal{N}_{\mathcal{U}}$ $\displaystyle D_{\mathcal{U}}\mathcal{L}^{f}(\sigma)$ $\displaystyle\coloneqq\mathcal{L}^{f}(U_{\sigma})$ $\displaystyle D_{\mathcal{U}}\mathcal{L}^{f}(\sigma\leq\tau)$ $\displaystyle\coloneqq\mathcal{L}^{f}(U_{\tau}\subseteq U_{\sigma})\hskip 2.0pt.$ If $\pi_{f}\colon X\rightarrow R_{f}$ is the quotient map from $X$ to the Reeb graph and $\hat{\mathcal{U}}$ a cover of $R_{f}$, then we define a decorated mapper graph as $D_{\hat{\mathcal{U}}}\mathcal{L}^{\pi_{f}}$; see Figure 1. ## 3 Convergence It is obvious that this process of discretization can lose information. This raises the question: How well is $\mathcal{L}^{f}$ represented by $D_{\mathcal{U}}\mathcal{L}^{f}$? To compare $\mathcal{L}^{f}$ and $D_{\mathcal{U}}\mathcal{L}^{f}$, we define a process of transforming $D_{\mathcal{U}}\mathcal{L}^{f}$ into a (pre)cosheaf on $Y$. We want to approximate the value of $\mathcal{L}^{f}$ on any open set $V\subseteq Y$ given only the information in $D_{\mathcal{U}}\mathcal{L}^{f}$. To this end, we define the subcomplex $K_{V}\coloneqq\\{\sigma\in\mathcal{N}_{\mathcal{U}}\text{ }|\text{ }U_{\sigma}\cap V\neq\emptyset\\}\leq\mathcal{N}_{\mathcal{U}}$, which can be viewed as a simplicial approximation of $V\subseteq\underset{\sigma\in K_{V}}{\bigcup}U_{\sigma}$; see Figure 2. Moreover, if $V\subseteq W$, we obtain an inclusion $\iota\colon K_{V}\xhookrightarrow{}K_{W}$. ###### Definition 3.1 (Continuous extension). Let $\mathcal{U}$ be a finite 1D cover of $Y$. We define the continuous extension $C_{\mathcal{U}}D_{\mathcal{U}}\mathcal{L}^{f}$ of $D_{\mathcal{U}}\mathcal{L}^{f}$ as a (pre)cosheaf on $Y$ via the following assignments: For all $V\subseteq W\subseteq Y$ open $\displaystyle C_{\mathcal{U}}D_{\mathcal{U}}\mathcal{L}^{f}(V)$ $\displaystyle\coloneqq\bigoplus_{n\in\mathbb{N}_{0}}\Big{(}H_{0}\big{(}K_{V};D_{\mathcal{U}}\mathcal{L}^{f}_{n}|_{K_{V}}\big{)}\oplus H_{1}\big{(}K_{V};D_{\mathcal{U}}\mathcal{L}^{f}_{n\minus 1}|_{K_{V}}\big{)}\Big{)}$ $\displaystyle C_{\mathcal{U}}D_{\mathcal{U}}\mathcal{L}^{f}(V\subseteq W)$ $\displaystyle\coloneqq\bigoplus_{n\in\mathbb{N}_{0}}\Big{(}H_{0}\big{(}\iota\big{)}_{n}\oplus H_{1}\big{(}\iota\big{)}_{n\minus 1}\Big{)}$ where $H_{i}\big{(}K_{V};D_{\mathcal{U}}\mathcal{L}^{f}_{n}|_{K_{V}}\big{)}$ is the degree $i$ homology [4, 6.2.2] of the cosheaf $D_{\mathcal{U}}\mathcal{L}^{f}_{n}$ restricted to $K_{V}$ and $H_{i}(\iota)_{n}$ is the map induced on cosheaf homology by $\iota\colon K_{V}\xhookrightarrow{}K_{W}$ cf. [13, A.17]. As shown in [4, 5, 15], if $\mathcal{U}$ is a finite 1D cover of $f(X)$ the continuous extension yields $C_{\mathcal{U}}D_{\mathcal{U}}\mathcal{L}^{f}(V)\cong\underset{n\in\mathbb{N}_{0}}{\bigoplus}H_{n}\Big{(}f^{\minus 1}\big{(}\underset{\sigma\in K_{V}}{\bigcup}U_{\sigma}\big{)}\Big{)}$ approximating $H_{\bullet}\big{(}f^{\minus 1}(V)\big{)}$; see Figure 2. Figure 2: The figure shows $D_{\mathcal{U}}\mathcal{L}^{f}_{0}\cong D_{\mathcal{U}}\mathcal{L}^{f}$ on $\mathcal{N}_{\mathcal{U}}$ w.r.t. two covers with different resolution as well as the restriction $D_{\mathcal{U}}\mathcal{L}^{f}_{0}|_{K_{V}}$ to $K_{V}$ (in orange). Homology is taken with coefficients in a field $k$ and unlabeled arrows represent identity maps. For the coarse cover we get $C_{\mathcal{U}}D_{\mathcal{U}}\mathcal{L}^{f}(V)\cong k\ncong H_{0}\big{(}f^{\minus 1}(V)\big{)}$ but for the finer one we get $C_{\mathcal{U}}D_{\mathcal{U}}\mathcal{L}^{f}(V)\cong k^{2}\cong H_{0}\big{(}f^{\minus 1}(V)\big{)}$. The following proposition shows that $C_{\mathcal{U}}D_{\mathcal{U}}\mathcal{L}^{f}(V\subseteq W)$ also gives us the correct induced map. ###### Proposition 3.2. Let $\mathcal{U}$ be a finite 1D cover of $f(X)$. Then, for all $V\subseteq W\subseteq Y$ open, the following diagram commutes and the horizontal arrows are isomorphisms: $\leavevmode\hbox to319.22pt{\vbox to114.94pt{\pgfpicture\makeatletter\hbox{\hskip 159.60762pt\lower-61.62509pt\hbox to0.0pt{\pgfsys@beginscope\pgfsys@invoke{ }\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\pgfsys@setlinewidth{0.4pt}\pgfsys@invoke{ }\nullfont\hbox to0.0pt{\pgfsys@beginscope\pgfsys@invoke{ }{}{}{}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{\offinterlineskip{}{}{{{}}{{}}{{}}{{}}}{{{}}}{{}{}{{ {}{}}}{ {}{}} {{}{{}}}{{}{}}{}{{}{}} { }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{-159.60762pt}{-53.31956pt}\pgfsys@invoke{ }\hbox{\vbox{\halign{\pgf@matrix@init@row\pgf@matrix@step@column{\pgf@matrix@startcell#\pgf@matrix@endcell}&#\pgf@matrix@padding&&\pgf@matrix@step@column{\pgf@matrix@startcell#\pgf@matrix@endcell}&#\pgf@matrix@padding\cr\hfil\hskip 83.00818pt\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{}{{ {}{}}}{ {}{}} {{}{{}}}{{}{}}{}{{}{}} { }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{-78.70264pt}{0.0pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{${H_{0}\big{(}K_{V};D_{\mathcal{U}}\mathcal{L}_{n}^{f}|_{K_{V}}\big{)}\oplus H_{1}\big{(}K_{V};D_{\mathcal{U}}\mathcal{L}_{n\minus 1}^{f}|_{K_{V}}\big{)}}$} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope{}}}&\hskip 83.00818pt\hfil&\hfil\hskip 95.37718pt\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{}{{ {}{}}}{ {}{}} {{}{{}}}{{}{}}{}{{}{}} { }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{-47.07167pt}{0.0pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{${H_{n}\Big{(}f^{\minus 1}\big{(}\underset{\sigma\in K_{V}}{\bigcup}U_{\sigma}\big{)}\Big{)}}$} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}&\hskip 51.37721pt\hfil\cr\vskip 48.00005pt\cr\hfil\hskip 85.45264pt\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{}{{ {}{}}}{ {}{}} {{}{{}}}{{}{}}{}{{}{}} { }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{-81.1471pt}{0.0pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{${H_{0}\big{(}K_{W};D_{\mathcal{U}}\mathcal{L}_{n}^{f}|_{K_{W}}\big{)}\oplus H_{1}\big{(}K_{W};D_{\mathcal{U}}\mathcal{L}_{n\minus 1}^{f}|_{K_{W}}\big{)}}$} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}&\hskip 85.45264pt\hfil&\hfil\hskip 96.15497pt\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{}{{ {}{}}}{ {}{}} {{}{{}}}{{}{}}{}{{}{}} { }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{-47.84946pt}{0.0pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{${H_{n}\Big{(}f^{\minus 1}\big{(}\underset{\sigma\in K_{W}}{\bigcup}U_{\sigma}\big{)}\Big{)}}$} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}&\hskip 52.155pt\hfil\cr}}}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}{{{{}}}{{}}{{}}{{}}{{}}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} {}{ {}{}{}}{}{ {}{}{}} {{{{{}}{ {}{}}{}{}{{}{}}}}}{}{{{{{}}{ {}{}}{}{}{{}{}}}}}{{}}{}{}{}{}{}{{{}{}}}{}{{}}{}{}{}{{{}{}}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@setlinewidth{0.39998pt}\pgfsys@invoke{ }{}{{ {\pgfsys@beginscope \pgfsys@setdash{}{0.0pt}\pgfsys@roundcap\pgfsys@roundjoin{} {}{}{} {}{}{} \pgfsys@moveto{-2.07988pt}{2.39986pt}\pgfsys@curveto{-1.69989pt}{0.95992pt}{-0.85313pt}{0.27998pt}{0.0pt}{0.0pt}\pgfsys@curveto{-0.85313pt}{-0.27998pt}{-1.69989pt}{-0.95992pt}{-2.07988pt}{-2.39986pt}\pgfsys@stroke\pgfsys@endscope}} }{}{}{{}}{}{}{{}}\pgfsys@moveto{9.05319pt}{26.50003pt}\pgfsys@lineto{55.47545pt}{26.50003pt}\pgfsys@stroke\pgfsys@invoke{ }{{}{{}}{}{}{{}}{{{}}}}{{}{{}}{}{}{{}}{{{}}{{{}}{\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{55.67543pt}{26.50003pt}\pgfsys@invoke{ }\pgfsys@invoke{ \lxSVG@closescope }\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}{{}}}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{}{{ {}{}}}{ {}{}} {{}{{}}}{{}{}}{}{{}{}} { }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{29.8393pt}{30.25279pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{$\scriptstyle{\cong}$} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope{}{ {}{}{}}{}{ {}{}{}} {{{{{}}{ {}{}}{}{}{{}{}}}}}{}{{{{{}}{ {}{}}{}{}{{}{}}}}}{{}}{}{}{}{}{}{{{}{}}}{}{{}}{}{}{}{{{}{}}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@setlinewidth{0.39998pt}\pgfsys@invoke{ }{}{}{}{}{{}}{}{}{{}}\pgfsys@moveto{-74.15498pt}{17.13367pt}\pgfsys@lineto{-74.15498pt}{-40.05989pt}\pgfsys@stroke\pgfsys@invoke{ }{{}{{}}{}{}{{}}{{{}}}}{{}{{}}{}{}{{}}{{{}}{{{}}{\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{0.0}{-1.0}{1.0}{0.0}{-74.15498pt}{-40.25987pt}\pgfsys@invoke{ }\pgfsys@invoke{ \lxSVG@closescope }\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}{{}}}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{}{{{}{}}}{{}{}} {{}{{}}}{{}{}}{}{{}{}} { }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{-114.5712pt}{-13.41309pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{$\scriptstyle{H_{0}(\iota)\oplus H_{1}(\iota)}$} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope{}{ {}{}{}}{}{ {}{}{}} {{{{{}}{ {}{}}{}{}{{}{}}}}}{}{{{{{}}{ {}{}}{}{}{{}{}}}}}{{}}{}{}{}{}{}{{{}{}}}{}{{}}{}{}{}{{{}{}}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@setlinewidth{0.39998pt}\pgfsys@invoke{ }{}{}{}{}{{}}{}{}{{}}\pgfsys@moveto{107.45262pt}{10.1402pt}\pgfsys@lineto{107.45262pt}{-37.05989pt}\pgfsys@stroke\pgfsys@invoke{ }{{}{{}}{}{}{{}}{{{}}}}{{}{{}}{}{}{{}}{{{}}{{{}}{\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{0.0}{-1.0}{1.0}{0.0}{107.45262pt}{-37.25987pt}\pgfsys@invoke{ }\pgfsys@invoke{ \lxSVG@closescope }\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}{{}}}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{}{{{}{}}}{{}{}} {{}{{}}}{{}{}}{}{{}{}} { }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{-17.7089pt}{-16.15977pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{$\scriptstyle{H_{n}\Big{(}f^{\minus 1}\big{(}\underset{\sigma\in K_{V}}{\bigcup}U_{\sigma}\big{)}\subseteq f^{\minus 1}\big{(}\underset{\sigma\in K_{W}}{\bigcup}U_{\sigma}\big{)}\Big{)}}$} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope{}{ {}{}{}}{}{ {}{}{}} {{{{{}}{ {}{}}{}{}{{}{}}}}}{}{{{{{}}{ {}{}}{}{}{{}{}}}}}{{}}{}{}{}{}{}{{{}{}}}{}{{}}{}{}{}{{{}{}}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@setlinewidth{0.39998pt}\pgfsys@invoke{ }{}{}{}{}{{}}{}{}{{}}\pgfsys@moveto{11.49765pt}{-50.81956pt}\pgfsys@lineto{54.69766pt}{-50.81956pt}\pgfsys@stroke\pgfsys@invoke{ }{{}{{}}{}{}{{}}{{{}}}}{{}{{}}{}{}{{}}{{{}}{{{}}{\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{54.89764pt}{-50.81956pt}\pgfsys@invoke{ }\pgfsys@invoke{ \lxSVG@closescope }\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}{{}}}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{}{{ {}{}}}{ {}{}} {{}{{}}}{{}{}}{}{{}{}} { }{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{30.67264pt}{-58.07233pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{$\scriptstyle{\cong}$} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope{}{}{}\hss}\pgfsys@discardpath\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope\hss}}\lxSVG@closescope\endpgfpicture}}\hskip 2.0pt.$ To talk about convergence, we have to define a distance on (pre)cosheaves. Assume $Y$ is a metric space and that all the involved (pre)cosheaves are constructible [4, 11.0.10]. For every $\epsilon>0$ define $V^{\epsilon}\coloneqq\\{y\in Y\text{ }|\text{ }d(y,V)<\epsilon\\}$. To a (pre)cosheaf $F$ on $Y$ we now associate $F^{\bullet}\coloneqq\\{(F^{\epsilon})_{\epsilon\geq 0}\hskip 2.0pt,\hskip 2.0pt(F_{\epsilon}^{\delta}\colon F^{\epsilon}\rightarrow F^{\delta})_{\epsilon\leq\delta}\\}$, a parameterized family of (pre)cosheaves on $Y$, by setting $F^{\epsilon}(V)\coloneqq F(V^{\epsilon})$. This allows us to define the distance of two (pre)cosheaves $F$ and $G$ as the interleaving distance of $F^{\bullet}$ and $G^{\bullet}$, i.e. $d(F,G)\coloneqq d_{I}(F^{\bullet},G^{\bullet})$. Denote by $\text{res}(\mathcal{U})\coloneqq\underset{U\in\mathcal{U}}{\text{sup}}\hskip 2.0pt\underset{x,y\in U}{\text{sup}}d(x,y)$ the resolution of an open cover $\mathcal{U}$. We are now able to show that if the resolution of the cover $\mathcal{U}$ goes to zero, $C_{\mathcal{U}}D_{\mathcal{U}}\mathcal{L}^{f}$ converges to $\mathcal{L}^{f}$. ###### Theorem 3.3. If $f\colon X\rightarrow Y$ is continuous and $\mathcal{U}$ is a finite 1D cover of $f(X)$, then $d\big{(}C_{\mathcal{U}}D_{\mathcal{U}}\mathcal{L}^{f},\mathcal{L}^{f}\big{)}\leq\text{res}(\mathcal{U})\hskip 2.0pt.$ If we only consider $\mathcal{L}^{f}_{0}$, the degree-zero part of $\mathcal{L}^{f}$, the construction in Definition 3.1 and Theorem 3.3 specializes to an abelianization of the convergence result in [11]. ## 4 Future Work Theorem 3.3 guarantees that, if we choose a cover of resolution $\leq\delta$, the cellular Leray cosheaf is a $\delta$-approximation of the continuous one. This result establishes a theoretical justification for working with finite covers. Since we have to deal with maps from finite point sets in practice, in future work we plan to investigate under what conditions we can infer the cellular Leray cosheaf of a map $f\colon X\rightarrow Y$ from a finite sample of $X$; cf. [2]. ## References * [1] Ulrich Bauer, Xiaoyin Ge, and Yusu Wang. Measuring distance between reeb graphs. In Proceedings of the Thirtieth Annual Symposium on Computational Geometry, SOCG’14, page 464–473, New York, NY, USA, 2014. Association for Computing Machinery. doi:10.1145/2582112.2582169. * [2] Adam Brown, Omer Bobrowski, Elizabeth Munch, and Bei Wang. Probabilistic convergence and stability of random mapper graphs. Journal of Applied and Computational Topology, 5(1):99–140, 2021\. doi:10.1007/s41468-020-00063-x. * [3] Hamish Carr, Jack Snoeyink, and Ulrike Axen. Computing contour trees in all dimensions. Computational Geometry, 24(2):75–94, 2003. Special Issue on the Fourth CGC Workshop on Computational Geometry. doi:https://doi.org/10.1016/S0925-7721(02)00093-7. * [4] Justin Curry. Sheaves, Cosheaves and Applications. PhD thesis, University of Pennsylvania, 2014. arXiv:1303.3255. * [5] Justin Curry, Robert Ghrist, and Vidit Nanda. Discrete morse theory for computing cellular sheaf cohomology. Foundations of Computational Mathematics, 16, 2013. doi:10.1007/s10208-015-9266-8. * [6] Vin de Silva, Elizabeth Munch, and Amit Patel. Categorified reeb graphs. Discrete & Computational Geometry, 55(4):854–906, 2016. doi:10.1007/s00454-016-9763-9. * [7] Barbara Di Fabio and Claudia Landi. The edit distance for reeb graphs of surfaces. Discrete & Computational Geometry, 55(2):423–461, 2016. doi:10.1007/s00454-016-9758-6. * [8] Herbert Edelsbrunner, John Harer, and Amit K. Patel. Reeb spaces of piecewise linear mappings. In Proceedings of the Twenty-Fourth Annual Symposium on Computational Geometry, SCG ’08, page 242–250, New York, NY, USA, 2008. Association for Computing Machinery. doi:10.1145/1377676.1377720. * [9] Xiaoyin Ge, Issam Safa, Mikhail Belkin, and Yusu Wang. Data skeletonization via reeb graphs. Advances in Neural Information Processing Systems, 24, 2011. * [10] P. Y. Lum, G. Singh, A. Lehman, T. Ishkanov, M. Vejdemo-Johansson, M. Alagappan, J. Carlsson, and G. Carlsson. Extracting insights from the shape of complex data using topology. Scientific Reports, 3(1):1236, 2013. doi:10.1038/srep01236. * [11] Elizabeth Munch and Bei Wang. Convergence between Categorical Representations of Reeb Space and Mapper. In 32nd International Symposium on Computational Geometry (SoCG 2016), volume 51 of Leibniz International Proceedings in Informatics (LIPIcs), pages 53:1–53:16. Schloss Dagstuhl–Leibniz-Zentrum fuer Informatik, 2016. doi:10.4230/LIPIcs.SoCG.2016.53. * [12] Monica Nicolau, Arnold Levine, and Gunnar Carlsson. Topology based data analysis identifies a subgroup of breast cancers with a unique mutational profile and excellent survival. Proceedings of the National Academy of Sciences of the United States of America, 108:7265–70, 04 2011. doi:10.1073/pnas.1102826108. * [13] Florian Russold. Persistent sheaf cohomology, 2022. arXiv:2204.13446. * [14] Gurjeet Singh, Facundo Memoli, and Gunnar Carlsson. Topological Methods for the Analysis of High Dimensional Data Sets and 3D Object Recognition. In Eurographics Symposium on Point-Based Graphics. The Eurographics Association, 2007. doi:10.2312/SPBG/SPBG07/091-100. * [15] Hee Rhang Yoon and Robert Ghrist. Persistence by parts: Multiscale feature detection via distributed persistent homology, 2020. arXiv:2001.01623.
TextBox # An Exploratory Study on Refactoring Documentation in Issues Handling Eman Abdullah AlOmar Rochester Institute of TechnologyRochester, New YorkUSA <EMAIL_ADDRESS>, Anthony Peruma Rochester Institute of TechnologyRochester, New YorkUSA<EMAIL_ADDRESS>, Mohamed Wiem Mkaouer Rochester Institute of TechnologyRochester, New YorkUSA<EMAIL_ADDRESS>, Christian D. Newman Rochester Institute of TechnologyRochester, New YorkUSA <EMAIL_ADDRESS>and Ali Ouni ETS Montreal, University of QuebecMontreal, QuebecCanada<EMAIL_ADDRESS> ###### Abstract. Understanding the practice of refactoring documentation is of paramount importance in academia and industry. Issue tracking systems are used by most software projects enabling developers, quality assurance, managers, and users to submit feature requests and other tasks such as bug fixing and code review. Although recent studies explored how to document refactoring in commit messages, little is known about how developers describe their refactoring needs in issues. In this study, we aim at exploring developer-reported refactoring changes in issues to better understand what developers consider to be problematic in their code and how they handle it. Our approach relies on text mining 45,477 refactoring-related issues and identifying refactoring patterns from a diverse corpus of 77 Java projects by investigating issues associated with 15,833 refactoring operations and developers’ explicit refactoring intention. Our results show that (1) developers mostly use move refactoring related terms/phrases to target refactoring-related issues; and (2) developers tend to explicitly mention the improvement of specific quality attributes and focus on duplicate code removal. We envision our findings enabling tool builders to support developers with automated documentation of refactoring changes in issues. Refactoring documentation, issues, software quality, mining software repositories ††copyright: none††conference: MSR ’22: Proceedings of the 19th International Conference on Mining Software Repositories; May 23–24, 2022; Pittsburgh, PA, USA††booktitle: MSR ’22: Proceedings of the 19th International Conference on Mining Software Repositories, May 23–24, 2022, Pittsburgh, PA, USA††price: 15.00††isbn: 978-1-4503-XXXX-X/18/06††ccs: Software Engineering Software Quality††ccs: Software Engineering Refactoring ## 1\. Introduction Code refactoring is a disciplined software engineering practice that is known as the process of changing a software system in such a way that it does not alter the external behavior of the code yet improves its internal structure (Fowler et al., 1999; AlOmar et al., 2021c). Refactoring is commonly used in different development and maintenance tasks (AlOmar et al., 2021e). It supports developers in revolving submitted issues such as feature requests (Nyamawe et al., 2019) or bug reports (Di Penta et al., 2020). Issue tracking systems are used by most contemporary software projects enabling developers, quality assurance, managers, and users to submit feature or enhancement requests, as well as other tasks such as bug fixing and code review. Previous studies have focused on recommending refactorings through the detection of refactoring opportunities, either by identifying code anti- patterns that need correction (Brito et al., 2020; Oizumi et al., 2020; Lenarduzzi et al., 2020), or by optimizing code quality metrics (Mkaouer et al., 2015; Anwer et al., 2017; Sellitto et al., [n. d.]). Yet, recent studies have shown that there is a gap between automated refactoring tools, and what developers consider to a need-to-refactor situation in code (Cedrim et al., 2016; AlOmar et al., 2019b; Fernandes et al., 2020). To bridge this gap, it is important to understand what triggers developers to refactor their code, and what do developers care about when it comes to code improvement. Such information provides insights, to software practitioners and researchers, about the developer’s perception of refactoring. This can question whether developers do care about structural metrics and code smells when refactoring their code, or if there are other factors that are of direct influence on these non-functional changes. In this paper, we focus on investigating issues that are written to express a need for refactoring. We extract patterns differentiating refactoring issues. These patterns represent what developers consider to be worth refactoring. We are also interested in the solutions, i.e., refactoring operations, being proposed as correction measures. Our investigation is driven by answering the following research questions: * • RQ1: What textual patterns do developers use to describe their refactoring needs in issues? This RQ explores the existence of refactoring documentation in issues containing refactorings. This RQ aims to identify developers’ common phrases when describing their refactoring problem/challenge. * • RQ2: What are the quality attributes developers care about when documenting in issues? In this RQ, we investigate whether developers explicitly indicate the purpose of their refactoring activity applied in issues, e.g., improving structural metrics of fixing code smells. The results of this exploratory study strengthen our understanding of what circumstances cause the need for refactorings. Using the evolution history of 77 open-source projects exhibiting a total of 45,477 refactoring commits with issues, our study reveals that developers are mainly driven by reducing complexity and increasing comprehension and performance. While various studies associate refactoring tightly with fixing code smells, the only anti-pattern that was highlighted is duplicate code. Furthermore, various studies have shown that the rename category is the most frequent in terms of refactoring operations (e.g., rename method, rename attribute, etc.) (Peruma et al., 2018; Negara et al., 2013), but our study reveals that developers tend also to discuss more complex refactorings, including refactorings dealing with extract and move code fragments, due to their impact on the design and code semantics preservation. We have also prepared a replication package of issues and their corresponding fixes (refactorings), to support the reproducibility and extension of our work (AlOmar, 2022). ## 2\. Study Design Figure 1. Overview of our experiment design. Table 1. Dataset overview of refactoring-related issues. Item Count Refactoring commits with issues 45,477 Refactoring commits with issues having keyword ‘refactor*’ in title 835 Refactoring operations associated with issues 15,833 Issue Status Item Count Closed 603 In progress 8 Open 28 Resolved 196 Issue Resolution Item Count Done 4 Fixed 780 Implemented 8 Resolved 4 Won’t fix 3 Issue Types Item Count Bug 95 Improvement 390 New Feature 2 Story 1 Sub-task 71 Task 274 Test 2 Figure 1 depicts a general overview of our experimental setup. In the following subsections, we elaborate on the activities involved in the process. We provide the dataset that we generate available in our replication package (AlOmar, 2022). ### 2.1. Source Dataset Our study utilizes the SmartSHARK MongoDB Release 2.1 dataset (Trautsch et al., 2021). This dataset contains a wide range of information for 77 open- source Java projects, such as commit history, issues, refactorings, code metrics, mailing lists, and continuous integration data, among others. All 77 Java projects are part of the Apache ecosystem and utilize GitHub as their version control repository and JIRA for issue tracking. Furthermore, SmartSHARK utilizes RefDiff (Silva and Valente, 2017) and RefactoringMiner (Tsantalis et al., 2018) to mine refactoring operations. Finally, the SmartSHARK dataset schema provides the necessary relationship attributes between data collections to join two or more related collection types. ### 2.2. Refactoring Documentation Detection To identify refactoring documentation patterns in issues, we perform a series of manual and automated activities, as follows: Step #1: Issues associated with a refactoring activity. As our study focuses on issues and refactorings, our analysis is limited to issues where one or more refactoring operations were performed as part of the issue resolution. Hence, we first extracted all different refactorings from the source dataset. Next, we identify all commits containing the refactoring operations. Finally, we extracted issues that was addressed using the identified commits. Step #2: Issues associated with developer intention about refactoring. To ensure that the selected issues are about refactoring, we focused on a subset of 835 issues that reported developers’ intention about the application of refactoring (i.e., having the keyword ‘refactor’). The choice of ‘refactor’, besides being used by all related studies, is intuitively the first term to identify ideal refactoring-related issues (Murphy-Hill et al., 2008; Kim et al., 2014; AlOmar et al., 2019a; AlOmar et al., 2021a). Finally, to reduce the occurrence of false positives, we limited our analysis to only the occurrence of the term ‘refactor’ in the title of the issue as the title is a concise description of the problem (Peruma et al., 2022; Rosen and Shihab, 2016). Step #3: Annotation of issues. When creating issues, developers use natural language to describe the issues. Hence, given the diverse nature of developers describing the problem, an automated approach to analyzing the issue text is not feasible. Therefore, we performed a manual analysis of the issue title and body to identify refactoring documentation patterns. Next, we grouped this subset of issues based on specific patterns. Further, to avoid redundancy of any pattern, we only considered one phrase if we found different patterns with the same meaning. For example, if we find patterns such as ‘simplifying the code’, ‘code simplification’, and ‘simplify code’, we add only one of these similar phrases in the list of patterns. This enables having a list of the most insightful and unique patterns, and it also helps in making more concise patterns that are usable for readers. ## 3\. Experimental Results ### 3.1. RQ1: What textual patterns do developers use to describe their refactoring needs in issues? Methodology. To identify refactoring documentation patterns, we manually inspect a subset of issues. These patterns are represented in the form of a keyword or phrase that frequently occurs in the issues associated with refactoring-related commits. Table 2. List of refactoring documentation in issues (‘*’ captures the extension of the keyword) . Patterns Add* Chang* Chang* the name Clean* up code Cleanup Code clean* Code optimization Creat* Extend* Extract* Fix* Fix* code style Improv* Improv* code quality Inlin* Introduc* Merg* Mov* Pull* up Push* down Repackag* Redesign* Reduc* Refactor* Refin* Remov* Renam* Reorganiz* Replac* Restructur* Rewrit* Simplify* code Split* Results. Our in-depth inspection of the issues results in a list of 33 refactoring documentation patterns, as shown in Table 2. Our findings show that the names of refactoring operations (e.g., ‘extract*’, ‘mov*’, ‘renam*’) occur in the top frequently occurring patterns, and these patterns are mainly linked to code elements at different levels of granularity such as classes, methods, and variables. These specific terms are well-known software refactoring operations and indicate developers’ knowledge of the catalog of refactoring operations. We also observe that the top-ranked refactoring operation-related keywords include ‘extract*’ and ‘mov*’. ‘pull up’ and ‘push down’ operations are among the least discussed refactoring operations (similar to findings in (Silva et al., 2016; Peruma et al., 2022)). Moreover, we observe the occurrences of issue-fixing specific terms such as ‘fix*’, ‘remov*’, and ‘reduc*’. Next, we examine the most common keywords that developers use when expressing refactoring documentation in issues. Figure 3 shows the top keywords used to identify refactoring documentations across the examined projects, that are ranked according to their number of occurrences. Extract 37.2% Move 24.6% Rename 21.5% Change 14.7% Inline 2.0% Figure 2. Percentage of refactorings, clustered by operation class. Table 3. Summary of refactoring patterns, clustered by refactoring related categories. Internal QA (%) | External QA (%) | Code Smell (%) ---|---|--- Complexity (0.26 %) | Readability (0.23 %) | Duplicate code (0.73 %) Design Size (0.25 %) | Performance (0.11 %) | Encapsulation (0.18 %) | Usability (0.08 %) | Dependency (0.16 %) | Extensibility (0.07 %) | Inheritance (0.08 %) | Compatibility (0.06 %) | Coupling (0.02 %) | Accuracy (0.05 %) | Abstraction (0.02 %) | Modularity (0.05 %) | | Flexibility (0.04 %) | | Understandability (0.04 %) | | Reusability (0.04 %) | | Testability (0.03 %) | | Maintainability (0.03 %) | | Manageability (0.02 %) | | Stability (0.006 %) | | Accessibility (0.01 %) | | Configurability (0.01 %) | | Robustness (0.006 %) | | Repeatability (0.006 %) | | Effectiveness (0.006 %) | To better understand the nature of refactoring documentation, we have classified the associated refactoring operations into 5 classes, namely, ‘changing’, ‘extracting’, ‘inlining’, ‘moving’, and ‘renaming’. Depicted in Figure 2, we cluster these operations associated with issues using a list of refactoring keywords defined in a previous work (AlOmar et al., 2022). The changing of the types belongs to the ‘changing’ class, whereas the extraction of classes and methods are included in the ‘extracting’ class. As for, ‘moving’, it gathers all the movement of code elements, e.g., moving methods, or pushing code elements across hierarchies. Merging-related activities are included in the ‘inlining’ class. Finally, the ‘renaming’ class contains all refactorings that rename a given code element such as a class, a package or an attribute. As shown in Figure 2, the ‘extracting’ operations are highly documented in issues across the projects, and it reached the percentage of 37.2%, higher than ‘moving’, ‘renaming’, and ‘changing’ whose percentage is respectively 24.6%, 21.5%, and 14.7%. The ‘inlining’ operations, however, is the least documented refactoring which had a ratio of only 2 %. Figure 3. Popular refactoring textual patterns in issues. ### 3.2. RQ2: What are the quality attributes developers care about when documenting in issues? Methodology. After identifying the different refactoring documentation patterns, we identify and categorize the patterns into three main categories (similar to (AlOmar et al., 2019a; AlOmar et al., 2021e; AlOmar et al., 2021b; AlOmar et al., 2021d)): (1) internal quality attributes, (2) external quality attributes, and (3) code smells. Results. Table 3 provides the list of refactoring documentation patterns, ranked based on their frequency, we identify in refactoring-related issues. We observe that developers frequently mention key internal quality attributes (such as inheritance, complexity, etc.), a wide range of external quality attributes (such as readability and performance), and code duplication code smell that might impact code quality. To improve the internal design, the system structure optimization regarding its complexity and design size seems to be the dominant focus that is consistently mentioned in issues (0.26% and 0.25%, respectively). Concerning external quality attribute-related issues, we observe the mention of refactorings to enhance nonfunctional attributes. Patterns such as ‘readability’, ‘performance’, and ‘usability’ represent the developers’ main focus, with 0.23%, 0.11%, and 0.08%, respectively. Finally, for code smell-focused refactoring issues, duplicate code represents the most popular anti-pattern developers intend to refactor (0.73%). ## 4\. Discussion Our research aims to explore refactoring documentation in issues to provide future research directions that support developers in understanding refactoring applied in issues. RQ1 indicates that developers tend to use a variety of textual patterns to document their refactorings in issues. These patterns can provide either a (1) generic description of problems developers encounter or (2) a specific refactoring operation name following Fowler’s names (Fowler et al., 1999). Although previous studies show that rename refactorings are a common type of refactoring, e.g., (Peruma et al., 2018), we notice that ‘mov*’ and ‘extract*’ are the topmost documented refactorings in issues. This can be explained by the fact that developers tend to make many design improvement decisions that include remodularizing packages by moving classes, reducing class-level coupling, and increasing cohesion by moving methods. Additionally, developers might use similar terminology when performing move-related refactoring operations, i.e., Extract/Inline/Pull-up/Push-down (AlOmar et al., 2022). As shown in Figure 2, ‘extracting’ is the most documented refactorings. An interpretation for this comes from the nature of the debugging process that may include the separation of concerns which helps in reducing the core complexity of a larger module and reduce its proneness to errors (Tsantalis and Chatzigeorgiou, 2011). In other words, developers tend to discuss more complex refactorings in issues, including refactorings from the extract and move categories, due to their impact on the design and code semantic preservation. This information can provide valuable references for refactoring documentation practice in issues. For example, whether refactoring-related issue descriptions have the relevant information is a critical indicator for reproducing refactoring-related issues. From RQ2, we observe that developers discuss quality concerns when documenting refactorings in issues that can be related to: (1) internal quality attributes, (2) external quality attributes, or (3) code smells. When analyzing these quality concerns per issue types reported in Table 1, we notice that complexity and duplicate code are mostly documented with issue type named ‘bug’, whereas the duplicate code and readability were the popular sub-categories for refactoring-related issues type named ‘improvement’. For instance, the developer discussed fixing design issues by putting common functionalities into a superclass to eliminate duplicate code, breaking up lengthier methods to make the code more readable, and avoiding nested complex data structure to reduce code complexity. Moreover, we observe that code smell is rarely documented in issues. As shown in Table 3, developers only focused on duplicate code removal. Conversely, developers tend to report a variety of external quality attributes, focusing mainly on improving readability of the code. This corroborates the finding by Palomba et al. (Palomba et al., 2017), where refactoring targeting program comprehension was mostly applied during bug fixing activities. As developers discussed functional and non-functional aspects of source code, future research can further investigate the intent as to why and how developers perform refactoring in issues. With a better understanding of this phenomenon, researchers and tool builders can support developers with automatically documenting refactorings in issues. One of the main purposes of exploring refactoring documentation in issues is to better understand how developers cope with their software decay by extracting any refactoring strategies that can be associated with removing code smells (Tsantalis et al., 2008; Bavota et al., 2013), or improving the design structural measurements (Mkaouer et al., 2014; Bavota et al., 2014). However, these techniques only analyze the changes at the source code level, and provide the operations performed, without associating it with any textual description, which may infer the rationale behind the refactoring application. Our proposal, of textual patterns, is the first step towards complementing the existing effort in detecting refactorings, by augmenting it with any description that was intended to describe the refactoring activity. As previously shown in Tables 2 and 3, developers tend to add a high-level description of their refactoring activity, and mention their intention behind refactoring (remove duplicate code, improve readability, etc.), along with mentioning the refactoring operations they apply (type migration, inline methods, etc.). Overall, the documentation of refactoring in issues is an important research direction that requires further attention. It has been known that there is a general shortage of refactoring documentation, and there is no consensus about how refactoring should be documented, which makes it subjective and developer- specific. Lack of design documentation forced developers to rely on the source code to identify design problems (Sousa et al., 2018). Moreover, the fine- grained description of refactoring can be time-consuming, as a typical description should contain an indication about the operations performed, refactored code elements, and a hint about the intention behind the refactoring. In addition, the developer specification can be ambiguous as it reflects the developer’s understanding of what has been improved in the source code, which can be different in reality, as the developer may not necessarily adequately estimate the refactoring impact on the quality improvement. ## 5\. Threats To Validity The first threat relates to the analysis of open-source Java projects. Our results may not generalize to systems written in other languages. Another potential threat to validity relates to our findings regarding counting the reported quality attributes and code smells. Due to the large number of commit messages, we have not performed a manual validation to remove false positive commit messages. Thus, this may have an impact on our findings. Finally, we constructed our dataset by extracting issues containing the term ‘refactor’ in the title. There is the possibility that we may have excluded synonymous terms/phrases. However, even though this approach reduces the number of issues in our dataset, it also decreases false-positives, and ensures that we analyze issues that are explicitly focused on refactorings. ## 6\. Conclusion & Future Work In this study, we performed an exploratory study to understand how developers document refactorings in issues. Specifically, we identify refactoring terms/phrases patterns, study possible refactoring documentation types, and determine how many refactoring terms/phrases exist in issues. Our results show that (1) developers mostly use move refactoring related terms/phrases to target refactoring-related issues; and (2) developers tend to explicitly mention the improvement of specific quality attributes and focus on duplicate code removal. We envision our findings enabling tool builders to support developers with automatically document refactoring in issues. Future work in this area includes investigating which refactoring operation is more problematic in issues. ## References * (1) * AlOmar (2022) AlOmar. 2022. _ReplicationPackage_. https://smilevo.github.io/self-affirmed-refactoring/ * AlOmar et al. (2021a) Eman Abdullah AlOmar, Hussein AlRubaye, Mohamed Wiem Mkaouer, Ali Ouni, and Marouane Kessentini. 2021a. Refactoring Practices in the Context of Modern Code Review: An Industrial Case Study at Xerox. In _2021 IEEE/ACM 43rd International Conference on Software Engineering: Software Engineering in Practice (ICSE-SEIP)_. IEEE, 348–357. * AlOmar et al. (2022) Eman Abdullah AlOmar, Jiaqian Liu, Kenneth Addo, Mohamed Wiem Mkaouer, Christian Newman, Ali Ouni, and Zhe Yu. 2022\. On the documentation of refactoring types. _Automated Software Engineering_ 29, 1 (2022), 1–40. * AlOmar et al. (2021c) Eman Abdullah AlOmar, Mohamed Wiem Mkaouer, Christian Newman, and Ali Ouni. 2021c. On preserving the behavior in software refactoring: A systematic mapping study. _Information and Software Technology_ (2021), 106675. * AlOmar et al. (2019a) Eman Abdullah AlOmar, Mohamed Wiem Mkaouer, and Ali Ouni. 2019a. Can refactoring be self-affirmed? an exploratory study on how developers document their refactoring activities in commit messages. In _International Workshop on Refactoring-accepted. IEEE_. * AlOmar et al. (2021b) Eman Abdullah AlOmar, Mohamed Wiem Mkaouer, and Ali Ouni. 2021b. Toward the automatic classification of self-affirmed refactoring. _Journal of Systems and Software_ 171 (2021), 110821. * AlOmar et al. (2019b) Eman Abdullah AlOmar, Mohamed Wiem Mkaouer, Ali Ouni, and Marouane Kessentini. 2019b. On the impact of refactoring on the relationship between quality attributes and design metrics. In _2019 ACM/IEEE International Symposium on Empirical Software Engineering and Measurement (ESEM)_. IEEE, 1–11. * AlOmar et al. (2021e) Eman Abdullah AlOmar, Anthony Peruma, Mohamed Wiem Mkaouer, Christian Newman, Ali Ouni, and Marouane Kessentini. 2021e. How we refactor and how we document it? On the use of supervised machine learning algorithms to classify refactoring documentation. _Expert Systems with Applications_ 167 (2021), 114176. * AlOmar et al. (2021d) Eman Abdullah AlOmar, Anthony Peruma, Mohamed Wiem Mkaouer, Christian D Newman, and Ali Ouni. 2021d. Behind the scenes: On the relationship between developer experience and refactoring. _Journal of Software: Evolution and Process_ (2021), e2395. * Anwer et al. (2017) Sajid Anwer, Ahmad Adbellatif, Mohammad Alshayeb, and Muhammad Shakeel Anjum. 2017. Effect of coupling on software faults: An empirical study. In _2017 International Conference on Communication, Computing and Digital Systems (C-CODE)_. IEEE, 211–215. * Bavota et al. (2013) Gabriele Bavota, Bogdan Dit, Rocco Oliveto, Massimiliano Di Penta, Denys Poshyvanyk, and Andrea De Lucia. 2013. An empirical study on the developers’ perception of software coupling. In _Proceedings of the 2013 International Conference on Software Engineering_. IEEE Press, 692–701. * Bavota et al. (2014) Gabriele Bavota, Sebastiano Panichella, Nikolaos Tsantalis, Massimiliano Di Penta, Rocco Oliveto, and Gerardo Canfora. 2014\. Recommending refactorings based on team co-maintenance patterns. In _Proceedings of the 29th ACM/IEEE international conference on Automated software engineering_. ACM, 337–342. * Brito et al. (2020) Aline Brito, Andre Hora, and Marco Tulio Valente. 2020\. Refactoring Graphs: Assessing Refactoring over Time. _arXiv preprint arXiv:2003.04666_ (2020). * Cedrim et al. (2016) Diego Cedrim, Leonardo Sousa, Alessandro Garcia, and Rohit Gheyi. 2016. Does refactoring improve software structural quality? A longitudinal study of 25 projects. In _Proceedings of the 30th Brazilian Symposium on Software Engineering_. ACM, 73–82. * Di Penta et al. (2020) Massimiliano Di Penta, Gabriele Bavota, and Fiorella Zampetti. 2020. On the relationship between refactoring actions and bugs: a differentiated replication. In _Proceedings of the 28th ACM Joint Meeting on European Software Engineering Conference and Symposium on the Foundations of Software Engineering_. 556–567. * Fernandes et al. (2020) Eduardo Fernandes, Alexander Chávez, Alessandro Garcia, Isabella Ferreira, Diego Cedrim, Leonardo Sousa, and Willian Oizumi. 2020\. Refactoring effect on internal quality attributes: What haven’t they told you yet? _Information and Software Technology_ 126 (2020), 106347. * Fowler et al. (1999) Martin Fowler, Kent Beck, John Brant, William Opdyke, and don Roberts. 1999. _Refactoring: Improving the Design of Existing Code_. Addison-Wesley Longman Publishing Co., Inc., Boston, MA, USA. http://dl.acm.org/citation.cfm?id=311424 * Kim et al. (2014) Miryung Kim, Thomas Zimmermann, and Nachiappan Nagappan. 2014\. An empirical study of refactoringchallenges and benefits at microsoft. _IEEE Transactions on Software Engineering_ 40, 7 (2014), 633–649. * Lenarduzzi et al. (2020) Valentina Lenarduzzi, Francesco Lomio, Heikki Huttunen, and Davide Taibi. 2020. Are SonarQube Rules Inducing Bugs?. In _2020 IEEE 27th International Conference on Software Analysis, Evolution and Reengineering (SANER)_. IEEE, 501–511. * Mkaouer et al. (2014) Mohamed Wiem Mkaouer, Marouane Kessentini, Slim Bechikh, Kalyanmoy Deb, and Mel Ó Cinnéide. 2014. Recommendation system for software refactoring using innovization and interactive dynamic optimization. In _Proceedings of the 29th ACM/IEEE international conference on Automated software engineering_. ACM, 331–336. * Mkaouer et al. (2015) Wiem Mkaouer, Marouane Kessentini, Adnan Shaout, Patrice Koligheu, Slim Bechikh, Kalyanmoy Deb, and Ali Ouni. 2015. Many-objective software remodularization using NSGA-III. _ACM Transactions on Software Engineering and Methodology (TOSEM)_ 24, 3 (2015), 1–45. * Murphy-Hill et al. (2008) Emerson Murphy-Hill, Andrew P Black, Danny Dig, and Chris Parnin. 2008. Gathering refactoring data: a comparison of four methods. In _Proceedings of the 2nd Workshop on Refactoring Tools_. 1–5. * Negara et al. (2013) Stas Negara, Nicholas Chen, Mohsen Vakilian, Ralph E Johnson, and Danny Dig. 2013. A comparative study of manual and automated refactorings. In _European Conference on Object-Oriented Programming_. Springer, 552–576. * Nyamawe et al. (2019) Ally S Nyamawe, Hui Liu, Nan Niu, Qasim Umer, and Zhendong Niu. 2019. Automated recommendation of software refactorings based on feature requests. In _2019 IEEE 27th International Requirements Engineering Conference (RE)_. IEEE, 187–198. * Oizumi et al. (2020) Willian Oizumi, Ana C Bibiano, Diego Cedrim, Anderson Oliveira, Leonardo Sousa, Alessandro Garcia, and Daniel Oliveira. 2020\. Recommending Composite Refactorings for Smell Removal: Heuristics and Evaluation. In _Proceedings of the 34th Brazilian Symposium on Software Engineering_. 72–81. * Palomba et al. (2017) Fabio Palomba, Andy Zaidman, Rocco Oliveto, and Andrea De Lucia. 2017. An exploratory study on the relationship between changes and refactoring. In _2017 IEEE/ACM 25th International Conference on Program Comprehension (ICPC)_. IEEE, 176–185. * Peruma et al. (2018) Anthony Peruma, Mohamed Wiem Mkaouer, Michael J Decker, and Christian D Newman. 2018. An empirical investigation of how and why developers rename identifiers. In _Proceedings of the 2nd International Workshop on Refactoring_. 26–33. * Peruma et al. (2022) Anthony Peruma, Steven Simmons, Eman Abdullah AlOmar, Christian D Newman, Mohamed Wiem Mkaouer, and Ali Ouni. 2022. How do I refactor this? An empirical study on refactoring trends and topics in Stack Overflow. _Empirical Software Engineering_ 27, 1 (2022), 1–43. * Rosen and Shihab (2016) Christoffer Rosen and Emad Shihab. 2016. What are mobile developers asking about? A large scale study using stack overflow. _Empirical Software Engineering_ 21, 3 (01 Jun 2016), 1192–1223. https://doi.org/10.1007/s10664-015-9379-3 * Sellitto et al. ([n. d.]) Giulia Sellitto, Emanuele Iannone, Zadia Codabux, Valentina Lenarduzzi, Andrea De Lucia, Fabio Palomba, and Filomena Ferrucci. [n. d.]. Toward Understanding the Impact of Refactoring on Program Comprehension. ([n. d.]). * Silva et al. (2016) Danilo Silva, Nikolaos Tsantalis, and Marco Tulio Valente. 2016\. Why We Refactor? Confessions of GitHub Contributors. In _Proceedings of the 2016 24th ACM SIGSOFT International Symposium on Foundations of Software Engineering_ (Seattle, WA, USA) _(FSE 2016)_. ACM, New York, NY, USA, 858–870. https://doi.org/10.1145/2950290.2950305 * Silva and Valente (2017) Danilo Silva and Marco Tulio Valente. 2017. Refdiff: detecting refactorings in version histories. In _2017 IEEE/ACM 14th International Conference on Mining Software Repositories (MSR)_. IEEE, 269–279. * Sousa et al. (2018) Leonardo Sousa, Anderson Oliveira, Willian Oizumi, Simone Barbosa, Alessandro Garcia, Jaejoon Lee, Marcos Kalinowski, Rafael de Mello, Baldoino Fonseca, Roberto Oliveira, et al. 2018\. Identifying design problems in the source code: A grounded theory. In _Proceedings of the 40th International Conference on Software Engineering_. 921–931. * Trautsch et al. (2021) Alexander Trautsch, Fabian Trautsch, and Steffen Herbold. 2021\. MSR Mining Challenge: The SmartSHARK Repository Data. (2021). * Tsantalis et al. (2008) Nikolaos Tsantalis, Theodoros Chaikalis, and Alexander Chatzigeorgiou. 2008. JDeodorant: Identification and removal of type-checking bad smells. In _2008 12th European Conference on Software Maintenance and Reengineering_. IEEE, 329–331. * Tsantalis and Chatzigeorgiou (2011) Nikolaos Tsantalis and Alexander Chatzigeorgiou. 2011. Identification of extract method refactoring opportunities for the decomposition of methods. _Journal of Systems and Software_ 84, 10 (2011), 1757–1782. * Tsantalis et al. (2018) Nikolaos Tsantalis, Matin Mansouri, Laleh Eshkevari, Davood Mazinanian, and Danny Dig. 2018\. Accurate and efficient refactoring detection in commit history. In _2018 IEEE/ACM 40th International Conference on Software Engineering (ICSE)_. IEEE, 483–494.
$\lim_{k\rightarrow\infty}\operatorname*{p-lim}_{n\rightarrow\infty}(\lambda^{2}/n)(Q({\bm{m}}^{k-1})-q_{\infty})\|{\bm{m}}^{k-1}\|=0.$ (123) Because $\operatorname*{p-lim}_{n\rightarrow\infty}Q({\bm{m}}^{k-1})=q_{k}$, and $q_{k}\rightarrow q_{\infty}$ as $k\rightarrow\infty$, the previous display holds, and the result follows.
# HyperDeepONet: learning operator with complex target function space using the limited resources via hypernetwork Jae Yong Lee Center for Artificial Intelligence and Natural Sciences Korea Institute for Advanced Study Seoul, 02455, Republic of Korea <EMAIL_ADDRESS> &Sung Woong Cho11footnotemark: 1 & Hyung Ju Hwang Department of Mathematics Pohang University of Science and Technology Pohang, 37673, Republic of Korea <EMAIL_ADDRESS>These authors contributed equally.corresponding author ###### Abstract Fast and accurate predictions for complex physical dynamics are a significant challenge across various applications. Real-time prediction on resource- constrained hardware is even more crucial in real-world problems. The deep operator network (DeepONet) has recently been proposed as a framework for learning nonlinear mappings between function spaces. However, the DeepONet requires many parameters and has a high computational cost when learning operators, particularly those with complex (discontinuous or non-smooth) target functions. This study proposes HyperDeepONet, which uses the expressive power of the hypernetwork to enable the learning of a complex operator with a smaller set of parameters. The DeepONet and its variant models can be thought of as a method of injecting the input function information into the target function. From this perspective, these models can be viewed as a particular case of HyperDeepONet. We analyze the complexity of DeepONet and conclude that HyperDeepONet needs relatively lower complexity to obtain the desired accuracy for operator learning. HyperDeepONet successfully learned various operators with fewer computational resources compared to other benchmarks. ## 1 Introduction Operator learning for mapping between infinite-dimensional function spaces is a challenging problem. It has been used in many applications, such as climate prediction (Kurth et al., 2022) and fluid dynamics (Guo et al., 2016). The computational efficiency of learning the mapping is still important in real- world problems. The target function of the operator can be discontinuous or sharp for complicated dynamic systems. In this case, balancing model complexity and cost for computational time is a core problem for the real-time prediction on resource-constrained hardware (Choudhary et al., 2020; Murshed et al., 2021). Many machine learning methods and deep learning-based architectures have been successfully developed to learn a nonlinear mapping from an infinite- dimensional Banach space to another. They focus on learning the solution operator of some partial differential equations (PDEs), e.g., the initial or boundary condition of PDE to the corresponding solution. Anandkumar et al. (2019) proposed an iterative neural operator scheme to learn the solution operator of PDEs. Simultaneously, Lu et al. (2019; 2021) proposed a deep operator network (DeepONet) architecture based on the universal operator approximation theorem of Chen & Chen (1995). The DeepONet consists of two networks: branch net taking an input function at fixed finite locations, and trunk net taking a query location of the output function domain. Each network provides the $p$ outputs. The two $p$-outputs are combined as a linear combination (inner- product) to approximate the underlying operator, where the branch net produces the coefficients ($p$-coefficients) and the trunk net produces the basis functions ($p$-basis) of the target function. While variant models of DeepONet have been developed to improve the vanilla DeepONet, they still have difficulty approximating the operator for a complicated target function with limited computational resources. Lanthaler et al. (2022) and Kovachki et al. (2021b) pointed out the limitation of linear approximation in DeepONet. Some operators have a slow spectral decay rate of the Kolmogorov $n$-width, which defines the error of the best possible linear approximation using an $n-$dimensional space. A large $n$ is required to learn such operators accurately, which implies that the DeepONet requires a large number of basis $p$ and network parameters for them. Hadorn (2022) investigated the behavior of DeepONet, to find what makes it challenging to detect the sharp features in the target function when the number of basis $p$ is small. They proposed a Shift-DeepONet by adding two neural networks to shift and scale the input function. Venturi & Casey (2023) also analyzed the limitation of DeepONet via singular value decomposition (SVD) and proposed a flexible DeepONet (flexDeepONet), adding a pre-net and an additional output in the branch net. Recently, to overcome the limitation of the linear approximation, Seidman et al. (2022) proposed a nonlinear manifold decoder (NOMAD) framework by using a neural network that takes the output of the branch net as the input along with the query location. Even though these methods reduce the number of basis functions, the total number of parameters in the model cannot be decreased. The trunk net still requires many parameters to learn the complex operators, especially with the complicated (discontinuous or non-smooth) target functions. In this study, we propose a new architecture, HyperDeepONet, which enables operator learning, and involves a complex target function space even with limited resources. The HyperDeepONet uses a hypernetwork, as proposed by Ha et al. (2017), which produces parameters for the target network. Wang et al. (2022) pointed out that the final inner product in DeepONet may be inefficient if the information of the input function fails to propagate through a branch net. The hypernetwork in HyperDeepONet transmits the information of the input function to each target network’s parameters. Furthermore, the expressivity of the hypernetwork reduces the neural network complexity by sharing the parameters (Galanti & Wolf, 2020). Our main contributions are as follows. * • We propose a novel HyperDeepONet using a hypernetwork to overcome the limitations of DeepONet and learn the operators with a complicated target function space. The DeepONet and its variant models are analyzed primarily in terms of expressing the target function as a neural network (Figure 4). These models can be simplified versions of our general HyperDeepONet model (Figure 5). * • We analyze the complexity of DeepONet (Theorem 2) and prove that the complexity of the HyperDeepONet is lower than that of the DeepONet. We have identified that the DeepONet should employ a large number of basis to obtain the desired accuracy, so it requires numerous parameters. For variants of DeepONet combined with nonlinear reconstructors, we also present a lower bound for the number of parameters in the target network. * • The experiments show that the HyperDeepONet facilitates learning an operator with a small number of parameters in the target network even when the target function space is complicated with discontinuity and sharpness, which the DeepONet and its variants suffer from. The HyperDeepONet learns the operator more accurately even when the total number of parameters in the overall model is the same. ## 2 Related work Many machine learning methods and deep learning-based architectures have been successfully developed to solve PDEs with several advantages. One research direction is to use the neural network directly to represent the solution of PDE (E & Yu, 2018; Sirignano & Spiliopoulos, 2018). The physics-informed neural network (PINN), introduced by Raissi et al. (2019), minimized the residual of PDEs by using automatic differentiation instead of numerical approximations. There is another approach to solve PDEs, called operator learning. Operator learning aims to learn a nonlinear mapping from an infinite-dimensional Banach space to another. Many studies utilize the convolutional neural network to parameterize the solution operator of PDEs in various applications (Guo et al., 2016; Bhatnagar et al., 2019; Khoo et al., 2021; Zhu et al., 2019; Hwang et al., 2021). The neural operator (Kovachki et al., 2021b) is proposed to approximate the nonlinear operator inspired by Green’s function. Li et al. (2021) extend the neural operator structure to the Fourier Neural Operator (FNO) to approximate the integral operator effectively using the fast Fourier transform (FFT). Kovachki et al. (2021a) proved the universality of FNO and identified the size of the network. The DeepONet (Lu et al., 2019; 2021) has also been proposed as another framework for operator learning. The DeepONet has significantly been applied to various problems, such as bubble growth dynamics (Lin et al., 2021), hypersonic flows (Mao et al., 2021), and fluid flow (Cai et al., 2021). Figure 1: Example of operator learning: the input function and the output function for the solution operator of shallow water equation Lanthaler et al. (2022) provided the universal approximation property of DeepONet. Wang et al. (2021) proposed physics-informed DeepONet by adding a residual of PDE as a loss function, and Ryck & Mishra (2022) demonstrated the generic bounds on the approximation error for it. Prasthofer et al. (2022) considered the case where the discretization grid of the input function in DeepONet changes by employing the coordinate encoder. Lu et al. (2022) compared the FNO with DeepONet in different benchmarks to demonstrate the relative performance. FNO can only infer the output function of an operator as the input function in the same grid as it needs to discretize the output function to use Fast Fourier Transform(FFT). In contrast, the DeepONet can predict from any location. Ha et al. (2017) first proposed hypernetwork, a network that creates a weight of the primary network. Because the hypernetwork can achieve weight sharing and model compression, it requires a relatively small number of parameters even as the dataset grows. Galanti & Wolf (2020) proved that a hypernetwork provides higher expressivity with low-complexity target networks. Sitzmann et al. (2020) and Klocek et al. (2019) employed this approach to restoring images with insufficient pixel observations or resolutions. de Avila Belbute-Peres et al. (2021) investigated the relationship between the coefficients of PDEs and the corresponding solutions. They combined the hypernetwork with the PINN’s residual loss. For time-dependent PDE, Pan et al. (2022) designated the time $t$ as the input of the hypernetwork so that the target network indicates the solution at $t$. von Oswald et al. (2020) devised a chunk embedding method that partitions the parameters of the target network; this is because the output dimension of the hypernetwork can be large. ## 3 Operator learning ### 3.1 Problem setting The goal of operator learning is to learn a mapping from infinite-dimensional function space to the others using a finite pair of functions. Let $\mathcal{G}:\mathcal{U}\to\mathcal{S}$ be a nonlinear operator, where $\mathcal{U}$ and $\mathcal{S}$ are compact subsets of infinite-dimensional function spaces $\mathcal{U}\subset C(\mathcal{X};\mathbb{R}^{d_{u}})$ and $\mathcal{S}\subset C(\mathcal{Y};\mathbb{R}^{d_{s}})$ with compact domains $\mathcal{X}\subset\mathbb{R}^{d_{x}}$ and $\mathcal{Y}\subset\mathbb{R}^{d_{y}}$. For simplicity, we focus on the case $d_{u}=d_{s}=1$, and all the results could be extended to a more general case for arbitrary $d_{u}$ and $d_{s}$. Suppose we have observations $\\{u_{i},\mathcal{G}(u_{i})\\}_{i=1}^{N}$ where $u_{i}\in\mathcal{U}$ and $\mathcal{G}(u_{i})\in\mathcal{S}$. We aim to find an approximation $\mathcal{G}_{\theta}:\mathcal{U}\to\mathcal{S}$ with parameter $\theta$ using the $N$ observations so that $\mathcal{G}_{\theta}\approx\mathcal{G}$. For example, in a dam break scenario, it is an important to predict the fluid flow over time according to a random initial height of the fluid. Figure 2: Diagram for the three components for operator learning. To this end, we want to find an operator $\mathcal{G}_{\theta}$, which takes an initial fluid height as an input function and produces the fluid height over time at any location as the output function (Figure 1). As explained in Lanthaler et al. (2022), the approximator $\mathcal{G}_{\theta}$ can be decomposed into the three components (Figure 2) as $\mathcal{G}_{\theta}:=\mathcal{R}\circ\mathcal{A}\circ\mathcal{E}.$ (1) First, the encoder $\mathcal{E}$ takes an input function $u$ from $\mathcal{U}$ to generate the finite-dimensional encoded data in $\mathbb{R}^{m}$. Then, the approximator $\mathcal{A}$ is an operator approximator from the encoded data in finite dimension space $\mathbb{R}^{m}$ to the other finite-dimensional space $\mathbb{R}^{p}$. Finally, the reconstructor $\mathcal{R}$ reconstructs the output function $s(y)=\mathcal{G}(u)(y)$ with $y\in\mathcal{Y}$ using the approximated data in $\mathbb{R}^{p}$. ### 3.2 DeepONet and its limitation DeepONet can be analyzed using the above three decompositions. Assume that all the input functions $u$ are evaluated at fixed locations $\\{x_{j}\\}_{j=1}^{m}\subset\mathcal{X}$; they are called ”sensor points.” DeepONet uses an encoder as the pointwise projection $\mathcal{E}(u)=(u(x_{1}),u(x_{2}),...,u(x_{m}))$ of the continuous function $u$, the so-called ”sensor values” of the input function $u$. An intuitive idea is to employ a neural network that simply concatenates these $m$ sensor values and a query point $y$ as an input to approximate the target function $\mathcal{G}(u)(y)$. DeepONet, in contrast, handles $m$ sensor values and a query point $y$ separately into two subnetworks based on the universal approximation theorem for the operator (Lu et al., 2021). See Appendix B for more details. Lu et al. (2021) use the fully connected neural network for the approximator $\mathcal{A}:\mathbb{R}^{m}\to\mathbb{R}^{p}$. They referred to the composition of these two maps as branch net $\beta:\mathcal{U}\to\mathbb{R}^{p},\beta(u):=\mathcal{A}\circ\mathcal{E}(u)$ (2) for any $u\in\mathcal{U}$. The role of branch net can be interpreted as learning the coefficient of the target function $\mathcal{G}(u)(y)$. They use one additional neural network, called trunk net $\tau$ as shown below. $\tau:\mathcal{Y}\to\mathbb{R}^{p+1},\tau(y):=\\{\tau_{k}(y)\\}_{k=0}^{p}$ (3) for any $y\in\mathcal{Y}$. The role of trunk net can be interpreted as learning an affine space $V$ that can efficiently approximate output function space $C(\mathcal{Y};\mathbb{R}^{d_{s}})$. The functions $\tau_{1}(y)$, …,$\tau_{p}(y)$ become the $p$-basis of vector space associated with $V$ and $\tau_{0}(y)$ becomes a point of $V$. By using the trunk net $\tau$, the $\tau$-induced reconstructor $\mathcal{R}$ is defined as $\mathcal{R}_{\tau}:\mathbb{R}^{p}\to C(\mathcal{Y};\mathbb{R}^{d_{s}}),\mathcal{R}_{\tau}(\beta)(y):=\tau_{0}(y)+\sum_{k=1}^{p}\beta_{k}\tau_{k}(y)$ (4) where $\beta=(\beta_{1},\beta_{2},...,\beta_{p})\in\mathbb{R}^{p}$. In DeepONet, $\tau_{0}(y)$ is restricted to be a constant $\tau_{0}\in\mathbb{R}$ that is contained in a reconstructor $\mathcal{R}$. The architecture of DeepONet is described in Figure 4 (b). Here, the $\tau$-induced reconstructor $\mathcal{R}_{\tau}$ is the linear approximation of the output function space. Because the linear approximation $\mathcal{R}$ cannot consider the elements in its orthogonal complement, a priori limitation on the best error of DeepONet is explained in Lanthaler et al. (2022) as $\left(\int_{\mathcal{U}}\int_{\mathcal{Y}}|\mathcal{G}(u)(y)-\mathcal{R}_{\tau}\circ\mathcal{A}\circ\mathcal{E}(u)(y)|^{2}dyd\mu(u)\right)^{\frac{1}{2}}\geq\sqrt{\sum_{k>p}\lambda_{k}},$ (5) where $\lambda_{1}\geq\lambda_{2}\geq...$ are the eigenvalues of the covariance operator $\Gamma_{\mathcal{G}_{\\#\mu}}$ of the push-forward measure $\mathcal{G}_{\\#\mu}$. Several studies point out that the slow decay rate of the lower bound leads to inaccurate approximation operator learning using DeepONet (Kovachki et al., 2021b; Hadorn, 2022; Lanthaler et al., 2022). For example, the solution operator of the advection PDEs (Seidman et al., 2022; Venturi & Casey, 2023) and of the Burgers’ equation (Hadorn, 2022) are difficult to approximate when we are using the DeepONet with the small number of basis $p$. Figure 3: The perspective target network parametrization for operator learning. ### 3.3 Variant models of DeepONet Several variants of DeepONet have been developed to overcome its limitation. All these models can be viewed from the perspective of parametrizing the target function as a neural network. When we think of the target network that receives $y$ as an input and generates an output $\mathcal{G}_{\theta}(u)(y)$, the DeepONet and its variant model can be distinguished by how information from the input function $u$ is injected into this target network $\mathcal{G}_{\theta}(u)$, as described in Figure 3. From this perspective, the trunk net in the DeepONet can be considered as the target network except for the final output, as shown in Figure 4 (a). The output of the branch net gives the weight between the last hidden layer and the final output. Hadorn (2022) proposed Shift-DeepONet. The main idea is that a scale net and a shift net are used to shift and scale the input query position $y$. Therefore, it can be considered that the information of input function $u$ generates the weights and bias between the input layer and the first hidden layer, as explained in Figure 4 (b). Venturi & Casey (2023) proposed FlexDeepONet, explained in Figure 4 (c). They used the additional network, pre-net, to give the bias between the input layer and the first hidden layer. Additionally, the output of the branch net also admits the additional output $\tau_{0}$ to provide more information on input function $u$ at the last inner product layer. NOMAD is recently developed by Seidman et al. (2022) to overcome the limitation of DeepONet. They devise a nonlinear output manifold using a neural network that takes the output of branch net $\\{\beta_{i}\\}_{i=1}^{p}$ and the query location $y$. As explained in Figure 4 (d), the target network receives information about the function $u$ as an additional input, similar to other conventional neural embedding methods (Park et al., 2019; Chen & Zhang, 2019; Mescheder et al., 2019). Figure 4: DeepONet and its variant models for operator learning. These methods provide information on the input function $u$ to only a part of the target network. It is a natural idea to use a hypernetwork to share the information of input function $u$ to all parameters of the target network. We propose a general model HyperDeepONet (Figure 5), which contains the vanilla DeepONet, FlexDeepONet, and Shift-DeepONet, as a special case of the HyperDeepONet. Figure 5: The proposed HyperDeepONet structure ## 4 Proposed model: HyperDeepONet ### 4.1 Architecture of HyperDeepONet The HyperDeepONet structure is described in Figure 5. The encoder $\mathcal{E}$ and the approximator $\mathcal{A}$ are used, similar to the vanilla DeepONet. The proposed structure replaces the branch net with the hypernetwork. The hypernetwork generate all parameters of the target network. More precisely, we define the hypernetwork $h$ as $h_{\theta}:\mathcal{U}\to\mathbb{R}^{p},\;h_{\theta}(u):=\mathcal{A}\circ\mathcal{E}(u)$ (6) for any $u\in\mathcal{U}$. Then, $h(u)=\Theta\in\mathbb{R}^{p}$ is a network parameter of the target network, which is used in reconstructor for the HyperDeepONet. We define the reconstructor $\mathcal{R}$ as $\mathcal{R}:\mathbb{R}^{p}\to C(\mathcal{Y};\mathbb{R}^{d_{s}}),\;\mathcal{R}(\Theta)(y):=\text{NN}(y;\Theta)$ (7) where $\Theta=[W,b]\in\mathbb{R}^{p}$, and NN denotes the target network. Two fully connected neural networks are employed for the hypernetwork and target network. Therefore, the main idea is to use the hypernetwork, which takes an input function $u$ and produces the weights of the target network. It can be thought of as a weight generator for the target network. The hypernetwork determines the all parameters of the target network containing the weights between the final hidden layer and the output layer. It implies that the structure of HyperDeepONet contains the entire structure of DeepONet. As shown in Figure 4 (b) and (c), Shift-DeepONet and FlexDeepONet can also be viewed as special cases of the HyperDeepONet, where the output of the hypernetwork determines the weights or biases of some layers of the target network. The outputs of the hypernetwork determine the biases for the first hidden layer in the target network for NOMAD in Figure 4 (d). ### 4.2 Comparison on complexity of DeepONet and HyperDeepONet In this section, we would like to clarify the complexity of the DeepONet required for the approximation $\mathcal{A}$ and reconstruction $\mathcal{R}$ based on the theory in Galanti & Wolf (2020). Furthermore, we will show that the HyperDeepONet entails a relatively lower complexity than the DeepONet using the results on the upper bound for the complexity of hypernetwork (Galanti & Wolf, 2020). #### 4.2.1 Notations and definitions Suppose that the pointwise projection values (sensor values) of the input function $u$ is given as $\mathcal{E}(u)=(u(x_{1}),u(x_{2}),...,u(x_{m}))$. For simplicity, we consider the case $\mathcal{Y}=[-1,1]^{d_{y}}$ and $\mathcal{E}(u)\in[-1,1]^{m}$. For the composition $\mathcal{R}\circ\mathcal{A}:\mathbb{R}^{m}\rightarrow C(\mathcal{Y};\mathbb{R})$, we focus on approximating the mapping $\mathcal{O}:\mathbb{R}^{m+d_{y}}\rightarrow\mathbb{R}$, which is defined as follows: $\mathcal{O}(\mathcal{E}(u),y):=(\mathcal{R}\circ\mathcal{A}(\mathcal{E}(u)))(y),\quad\text{for $y\in[-1,1]^{d_{y}},\mathcal{E}(u)\in[-1,1]^{m}$}.$ (8) The supremum norm $\|h\|_{\infty}$ is defined as $\max_{y\in\mathcal{Y}}\|h(y)\|$. Now, we introduce the Sobolev space $\mathcal{W}_{r,n}$, which is a subset of $C^{r}([-1,1]^{n};\mathbb{R})$. For $r,n\in\mathbb{N}$, $\mathcal{W}_{r,n}:=\bigg{\\{}h:[-1,1]^{n}\rightarrow\mathbb{R}\quad\Big{|}\|h\|_{r}^{s}:=\|h\|_{\infty}+\sum_{1\leq|\mathbf{k}|\leq r}\|D^{\mathbf{k}}h\|_{\infty}\leq 1\bigg{\\}},$ where $D^{\mathbf{k}}h$ denotes the partial derivative of $h$ with respect to multi-index $\mathbf{k}\in\left\\{\mathbb{N}\cup\left\\{0\right\\}\right\\}^{d_{y}}$. We assume that the mapping $\mathcal{O}$ lies in the Sobolev space $\mathcal{W}_{r,m+d_{y}}$. For the nonlinear activation $\sigma$, the class of neural network $\mathcal{F}$ represents the fully connected neural network with depth $k$ and corresponding width $(h_{1}=n,h_{2},\cdots,h_{k+1})$, where $W^{i}\in\mathbb{R}^{h_{i}}\times\mathbb{R}^{h_{i+1}}$ and $b_{i}\in\mathbb{R}^{h_{i+1}}$ denote the weights and bias of the $i$-th layer respectively. $\displaystyle\mathcal{F}:=\left\\{f:[-1,1]^{n}\rightarrow\mathbb{R}|f(y;[\mathbf{W},\mathbf{b}])=W^{k}\cdot\sigma(W^{k-1}\cdots\sigma(W^{1}\cdot y+b^{1})+b^{k-1})+b^{k}\right\\}$ Some activation functions facilitate an approximation for the Sobolev space and curtail the complexity. We will refer to these functions as universal activation functions. The formal definition can be found below, where the distance between the class of neural network $\mathcal{F}$ and the Sobolev space $\mathcal{W}_{r,n}$ is defined by $d(\mathcal{F};\mathcal{W}_{r,n}):=\sup_{\psi\in\mathcal{W}_{r,n}}\inf_{f\in\mathcal{F}}\|f-\psi\|_{\infty}$. Most well-known activation functions are universal activations that are infinitely differentiable and non-polynomial in any interval (Mhaskar, 1996). Furthermore, Hanin & Sellke (2017) state that the ReLU activation is also universal. ###### Definition 1. (Galanti & Wolf, 2020) (Universal activation). The activation function $\sigma$ is called universal if there exists a class of neural network $\mathcal{F}$ with activation function $\sigma$ such that the number of parameters of $\mathcal{F}$ is $O(\epsilon^{-n/r})$ with $d(\mathcal{F};\mathcal{W}_{r,n})\leq\epsilon$ for all $r,n\in\mathbb{N}$. We now introduce the theorem, which offers a guideline on the neural network architecture for operator learning. It suggests that if the entire architecture can be replaced with a fully connected neural network, large complexity should be required for approximating the target function. It also verifies that the lower bound for a universal activation function is a sharp bound on the number of parameters. First, we give an assumption to obtain the theorem. ###### Assumption 1. Suppose that $\mathcal{F}$ and $\mathcal{W}_{r,n}$ represent the class of neural network and the target function space to approximate, respectively. Let $\mathcal{F}^{\prime}$ be a neural network class representing a structure in which one neuron is added rather than $\mathcal{F}$. Then, the followings holds for all $\psi\in\mathcal{W}_{r,n}$ not contained in $\mathcal{F}$. $\inf_{f\in\mathcal{F}}\|f-\psi\|_{\infty}>\inf_{f\in\mathcal{F}^{\prime}}\|f-\psi\|_{\infty}.$ For $r=0$, Galanti & Wolf (2020) remark that the assumption is valid for 2-layered neural networks with respect to the $L^{2}$ norm when an activation function $\sigma$ is either a hyperbolic tangent or sigmoid function. With Assumption 1, the following theorem holds, which is a fundamental approach to identifying the complexity of DeepONet and its variants. Note that a real- valued function $g\in L^{1}(\mathbb{R})$ is called a bounded variation if its total variation $\sup_{\phi\in C_{c}^{1}(\mathbb{R}),\|\phi\|_{\infty}\leq 1}\int_{\mathbb{R}}g(x)\phi^{\prime}(x)dx$ is finite. ###### Theorem 1. (Galanti & Wolf, 2020). Suppose that $\mathcal{F}$ is a class of neural networks with a piecewise $C^{1}(\mathbb{R})$ activation function $\sigma:\mathbb{R}\rightarrow\mathbb{R}$ of which derivative $\sigma^{\prime}$ is bounded variation. If any non-constant $\psi\in\mathcal{W}_{r,n}$ does not belong to $\mathcal{F}$, then $d(\mathcal{F};W_{r,n})\leq\epsilon$ implies the number of parameters in $\mathcal{F}$ should be $\Omega(\epsilon^{-n/r})$. #### 4.2.2 Lower bound for the complexity of the DeepONet Now, we provide the minimum number of parameters in DeepONet. The following theorem presents a criterion on the DeepONet’s complexity to get the desired error. It states that the number of required parameters increases when the target functions are irregular, corresponding to a small $r$. $\mathcal{F}_{\text{DeepONet}}(\mathcal{B},\mathcal{T})$ denotes the class of function in DeepONet, induced by the class of branch net $\mathcal{B}$ and the class of trunk net $\mathcal{T}$. ###### Theorem 2. (Complexity of DeepONet) Let $\sigma:\mathbb{R}\rightarrow\mathbb{R}$ be a universal activation function in $C^{r}(\mathbb{R})$ such that $\sigma$ and $\sigma^{\prime}$ are bounded. Suppose that the class of branch net $\mathcal{B}$ has a bounded Sobolev norm (i.e., $\|\beta\|_{r}^{s}\leq l_{1},\forall\beta\in\mathcal{B}$). Suppose any non-constant $\psi\in\mathcal{W}_{r,n}$ does not belong to any class of neural network. In that case, the number of parameters in the class of trunk net $\mathcal{T}$ is $\Omega(\epsilon^{-d_{y}/r})$ when $d(\mathcal{F}_{\text{DeepONet}}(\mathcal{B},\mathcal{T});\mathcal{W}_{r,d_{y}+m})\leq\epsilon$. The core of this proof is showing that the inner product between the branch net and the trunk net could be replaced with a neural network that has a low complexity (Lemma 1). Therefore, the entire structure of DeepONet could be replaced with a neural network that receives $[\mathcal{E}(u),y]\in\mathbb{R}^{d_{y}+m}$ as input. It gives the lower bound for the number of parameters in DeepONet based on Theorem 1. The proof can be found in Appendix C.1. The analogous results holds for variant models of DeepONet. Models such as Shift-DeepONet and flexDeepONet could achieve the desired accuracy with a small number of basis. Still, there was a trade-off in which the first hidden layer of the target network required numerous units. There was no restriction on the dimension of the last hidden layer in the target network for NOMAD, which uses a fully nonlinear reconstruction. However, the first hidden layer of the target network also had to be wide enough, increasing the number of parameters. Details can be found in Appendix C.2. For the proposed HyperDeepONet, the sensor values $\mathcal{E}(u)$ determine the weight and bias of all other layers as well as the weight of the last layer of the target network. Due to the nonlinear activation functions between linear matrix multiplication, it is difficult to replace HyperDeepONet with a single neural network that receives $[\mathcal{E}(u),y]\in\mathbb{R}^{d_{y}+m}$ as input. Galanti & Wolf (2020) state that there exists a hypernetwork structure (HyperDeepONet) such that the number of parameters in the target network is $O(\epsilon^{-d_{y}/r})$. It implies that the HyperDeepONet reduces the complexity compared to all the variants of DeepONet. ## 5 Experiments In this section, we verify the effectiveness of the proposed model HyperDeepONet to learn the operators with a complicated target function space. To be more specific, we focus on operator learning problems in which the space of output function space is complicated. Each input function $u_{i}$ generates multiple triplet data points $(u_{i},y,\mathcal{G}(u)(y))$ for different values of $y$. Except for the shallow water problem, which uses 100 training function pairs and 20 test pairs, we use 1,000 training input-output function pairs and 200 test pairs for all experiments. For the toy example, we first consider the identity operator $\mathcal{G}:u_{i}\mapsto u_{i}$. The Chebyshev polynomial is used as the input (=output) for the identity operator problem. The Chebyshev polynomials of the first kind $T_{l}$ of degree 20 can be written as $u_{i}\in\\{\sum_{l=0}^{19}c_{l}T_{l}(x)|c_{l}\in[-1/4,1/4]\\}$ with random sampling $c_{l}$ from uniform distribution $U[-1/4,1/4]$. The differentiation operator $\mathcal{G}:u_{i}\mapsto\frac{d}{dx}u_{i}$ is considered for the second problem. Previous works handled the anti-derivative operator, which makes the output function smoother by averaging (Lu et al., 2019; 2022). Here, we choose the differentiation operator instead of the anti-derivative operator to focus on operator learning when the operator’s output function space is complicated. We first sample the output function $\mathcal{G}(u)$ from the above Chebyshev polynomial of degree 20. The input function is generated using the numerical method that integrates the output function. Finally, the solution operators of PDEs are considered. We deal with two problems with the complex target function in previous works (Lu et al., 2022; Hadorn, 2022). The solution operator of the advection equation is considered a mapping from the rectangle shape initial input function to the solution $w(t,x)$ at $t=0.5$, i.e., $\mathcal{G}:w(0,x)\mapsto w(0.5,x)$. We also consider the solution operator of Burgers’ equation which maps the random initial condition to the solution $w(t,x)$ at $t=1$, i.e., $\mathcal{G}:w(0,x)\mapsto w(1,x)$. The solution of the Burgers’ equation has a discontinuity in a short time, although the initial input function is smooth. For a challenging benchmark, we consider the solution operator of the shallow water equation that aims to predict the fluid height $h(t,x_{1},x_{2})$ from the initial condition $h(0,x_{1},x_{2})$, i.e., $\mathcal{G}:h(0,x_{1},x_{2})\mapsto h(t,x_{1},x_{2})$ (Figure 1). In this case, the input of the target network is three dimension ($t,x_{1},x_{2}$), which makes the solution operator complex. Detail explanation is provided in Appendix E. Model | DeepONet | Shift | Flex | NOMAD | Hyper(ours) ---|---|---|---|---|--- Identity | 0.578$\pm$0.003 | 0.777$\pm$0.018 | 0.678$\pm$0.062 | 0.578$\pm$0.020 | 0.036$\pm$0.005 Differentiation | 0.559$\pm$0.001 | 0.624$\pm$0.015 | 0.562$\pm$0.016 | 0.558$\pm$0.003 | 0.127$\pm$0.043 Table 1: The mean relative $L^{2}$ test error with standard deviation for the identity operator and the differentiation operator. The DeepONet, its variants, and the HyperDeepONet use the target network $d_{y}$-20-20-10-1 with tanh activation function. Five training trials are performed independently. Figure 6: One test data example of differentiation operator problem. Expressivity of target network. We have compared the expressivity of the small target network using different models. We focus on the identity and differentiation operators in this experiment. All models employ the small target network $d_{y}$-20-20-10-1 with the hyperbolic tangent activation function. The branch net and the additional networks (scale net, shift net, pre-net, and hypernetwork) also use the same network size as the target network for all five models. | Branch net (Hypernetwork) | Target | $\\#$Param | Rel error ---|---|---|---|--- Advection | DeepONet | $m$-256-256 | $d_{y}$-256-256-256-256-1 | 274K | 0.0046$\pm$0.0017 Hyper(ours) | $m$-70-70-70-70-70-$N_{\theta}$ | $d_{y}$-33-33-33-33-1 | 268K | 0.0048$\pm$0.0009 c-Hyper(ours) | $m$-128-128-128-128-128-$1024$ | $d_{y}$-256-256-256-256-1 | 208K | 0.0043$\pm$0.0004 Burgers | DeepONet | $m$-128-128-128-128 | $d_{y}$-128-128-128-128-1 | 115K | 0.0391$\pm$0.0040 Hyper(ours) | $m$-66-66-66-66-66-$N_{\theta}$ | $d_{y}$-20-20-20-20-1 | 114K | 0.0196$\pm$0.0044 c-Hyper(ours) | $m$-66-66-66-66-66-$512$ | $d_{y}$-128-128-128-128-1 | 115K | 0.0066$\pm$0.0009 Shallow | DeepONet | $m$-100-100-100-100 | $d_{y}$-100-100-100-100-1 | 107K | 0.0279 $\pm$ 0.0042 Hyper(ours) | $m$-30-30-30-30-$N_{\theta}$ | $d_{y}$-30-30-30-30-1 | 101K | 0.0148 $\pm$ 0.0002 Shallow w/ small param | DeepONet | $m$-20-20-10 | $d_{y}$-20-20-10-1 | 6.5K | 0.0391 $\pm$ 0.0066 Hyper(ours) | $m$ -10-10-10-$N_{\theta}$ | $d_{y}$-10-10-10-1 | 5.7K | 0.0209 $\pm$ 0.0013 Table 2: The mean relative $L^{2}$ test error with standard deviation for solution operator learning problems. $N_{\theta}$ and $\\#$Param denote the number of parameters in the target network and the number of learnable parameters, respectively. Five training trials are performed independently. Table 1 shows that the DeepONet and its variant models have high errors in learning complex operators when the small target network is used. In contrast, the HyperDeepONet has lower errors than the other models. This is consistent with the theorem in the previous section that HyperDeepONet can achieve improved approximations than the DeepONet when the complexity of the target network is the same. Figure 6 shows a prediction on the differentiation operator, which has a highly complex target function. The same trends are observed when the activation function or the number of sensor points changes (Table 5) and the number of layers in the branch net and the hypernetwork vary (Figure 11). Same number of learnable parameters. The previous experiments compare the models using the same target network structure. In this section, the comparison between the DeepONet and the HyperDeepONet is considered when using the same number of learnable parameters. We focus on the solution operators of the PDEs. For the three solution operator learning problems, we use the same hyperparameters proposed in Lu et al. (2022) and Seidman et al. (2022) for DeepONet. First, we use the smaller target network with the larger hypernetwork for the HyperDeepONet to compare the DeepONet. Note that the vanilla DeepONet is used without the output normalization or the boundary condition enforcing techniques explained in Lu et al. (2022) to focus on the primary limitation of the DeepONet. More Details are in Appendix E. Table 2 shows that the HyperDeepONet achieves a similar or better performance than the DeepONet when the two models use the same number of learnable parameters. The HyperDeepONet has a slightly higher error for advection equation problem, but this error is close to perfect operator prediction. It shows that the complexity of target network and the number of learnable parameters can be reduced to obtain the desired accuracy using the HyperDeepONet. The fourth row of Table 2 shows that HyperDeepONet is much more effective than DeepONet in approximating the solution operator of the shallow water equation when the number of parameters is limited. Figure 7 and Figure 12 show that the HyperDeepONet learns the complex target functions in fewer epochs for the desired accuracy than the DeepONet although the HyperDeepONet requires more time to train for one epoch (Table 8). Figure 7: One test data example of prediction on the advection equation (First row) and Burgers’ equation (Second row) using the DeepONet and the HyperDeepONet. Scalability. When the size of the target network for the HyperDeepONet is large, the output of the hypernetwork would be high-dimensional (Ha et al., 2017; Pawlowski et al., 2017) so that its complexity increases. In this case, the chunked HyperDeepONet (c-HyperDeepONet) can be used with a trade-off between accuracy and memory based on the chunk embedding method developed by von Oswald et al. (2020). It generates the subset of target network parameters multiple times iteratively reusing the smaller chunked hypernetwork. The c-HyperDeepONet shows a better accuracy than the DeepONet and the HyperDeepONet using an almost similar number of parameters, as shown in Table 2. However, it takes almost 2x training time and 2$\sim$30x memory usage than the HyperDeepOnet. More details on the chunked hypernetwork are in Appendix D. ## 6 Conclusion and discussion In this work, the HyperDeepONet is developed to overcome the expressivity limitation of DeepONet. The method of incorporating an additional network and a nonlinear reconstructor could not thoroughly solve this limitation. The hypernetwork, which involves multiple weights simultaneously, had a desired complexity-reducing structure based on theory and experiments. We only focused on when the hypernetwork and the target network is fully connected neural networks. In the future, the structure of the two networks can be replaced with a CNN or ResNet, as the structure of the branch net and trunk net of DeepONet can be changed to another network (Lu et al., 2022). Additionally, it seems interesting to research a simplified modulation network proposed by Mehta et al. (2021), which still has the same expressivity as HyperDeepONet. Some techniques from implicit neural representation can improve the expressivity of the target network (Sitzmann et al., 2020). Using a sine function as an activation function with preprocessing will promote the expressivity of the target network. We also leave the research on the class of activation functions satisfying the assumption except for hyperbolic tangent or sigmoid functions as a future work. #### Acknowledgments J. Y. Lee was supported by a KIAS Individual Grant (AP086901) via the Center for AI and Natural Sciences at Korea Institute for Advanced Study and by the Center for Advanced Computation at Korea Institute for Advanced Study. H. J. Hwang and S. W. Cho were supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (RS-2022-00165268) and by Institute for Information & Communications Technology Promotion (IITP) grant funded by the Korea government(MSIP) (No.2019-0-01906, Artificial Intelligence Graduate School Program (POSTECH)). ## References * Anandkumar et al. (2019) Anima Anandkumar, Kamyar Azizzadenesheli, Kaushik Bhattacharya, Nikola Kovachki, Zongyi Li, Burigede Liu, and Andrew Stuart. Neural operator: Graph kernel network for partial differential equations. In _ICLR 2020 Workshop on Integration of Deep Neural Models and Differential Equations_ , 2019. URL https://openreview.net/forum?id=fg2ZFmXFO3. * Bhatnagar et al. (2019) Saakaar Bhatnagar, Yaser Afshar, Shaowu Pan, Karthik Duraisamy, and Shailendra Kaushik. Prediction of aerodynamic flow fields using convolutional neural networks. _Comput. Mech._ , 64(2):525–545, 2019. ISSN 0178-7675. doi: 10.1007/s00466-019-01740-0. URL https://doi.org/10.1007/s00466-019-01740-0. * Cai et al. (2021) Shengze Cai, Zhicheng Wang, Lu Lu, Tamer A Zaki, and George Em Karniadakis. Deepm&mnet: Inferring the electroconvection multiphysics fields based on operator approximation by neural networks. _Journal of Computational Physics_ , 436:110296, 2021. * Chen & Chen (1995) Tianping Chen and Hong Chen. Universal approximation to nonlinear operators by neural networks with arbitrary activation functions and its application to dynamical systems. _IEEE Transactions on Neural Networks_ , 6(4):911–917, 1995. * Chen & Zhang (2019) Zhiqin Chen and Hao Zhang. Learning implicit fields for generative shape modeling. In _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition_ , pp. 5939–5948, 2019. * Choudhary et al. (2020) Tejalal Choudhary, Vipul Mishra, Anurag Goswami, and Jagannathan Sarangapani. A comprehensive survey on model compression and acceleration. _Artificial Intelligence Review_ , 53(7):5113–5155, 2020. * de Avila Belbute-Peres et al. (2021) Filipe de Avila Belbute-Peres, Yi-fan Chen, and Fei Sha. Hyperpinn: Learning parameterized differential equations with physics-informed hypernetworks. In _The Symbiosis of Deep Learning and Differential Equations_ , 2021\. * E & Yu (2018) Weinan E and Bing Yu. The deep Ritz method: a deep learning-based numerical algorithm for solving variational problems. _Commun. Math. Stat._ , 6(1):1–12, 2018. ISSN 2194-6701. doi: 10.1007/s40304-018-0127-z. URL https://doi.org/10.1007/s40304-018-0127-z. * Galanti & Wolf (2020) Tomer Galanti and Lior Wolf. On the modularity of hypernetworks. _Advances in Neural Information Processing Systems_ , 33:10409–10419, 2020. * Guo et al. (2016) Xiaoxiao Guo, Wei Li, and Francesco Iorio. Convolutional neural networks for steady flow approximation. In _Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining_ , pp. 481–490, 2016. * Ha et al. (2017) David Ha, Andrew M. Dai, and Quoc V. Le. Hypernetworks. In _International Conference on Learning Representations_ , 2017. URL https://openreview.net/forum?id=rkpACe1lx. * Hadorn (2022) Patrik Simon Hadorn. Shift-deeponet: Extending deep operator networks for discontinuous output functions. ETH Zurich, Seminar for Applied Mathematics, 2022. * Hanin & Sellke (2017) Boris Hanin and Mark Sellke. Approximating continuous functions by relu nets of minimal width. _arXiv preprint arXiv:1710.11278_ , 2017. * Hwang et al. (2021) Rakhoon Hwang, Jae Yong Lee, Jin Young Shin, and Hyung Ju Hwang. Solving pde-constrained control problems using operator learning. _arXiv preprint arXiv:2111.04941_ , 2021. * Khoo et al. (2021) Yuehaw Khoo, Jianfeng Lu, and Lexing Ying. Solving parametric PDE problems with artificial neural networks. _European J. Appl. Math._ , 32(3):421–435, 2021\. ISSN 0956-7925. doi: 10.1017/S0956792520000182. URL https://doi.org/10.1017/S0956792520000182. * Kingma & Ba (2015) Diederik P. Kingma and Jimmy Ba. Adam: A method for stochastic optimization. In Yoshua Bengio and Yann LeCun (eds.), _3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Conference Track Proceedings_ , 2015. URL http://arxiv.org/abs/1412.6980. * Klocek et al. (2019) Sylwester Klocek, Łukasz Maziarka, Maciej Wołczyk, Jacek Tabor, Jakub Nowak, and Marek Śmieja. Hypernetwork functional image representation. In _International Conference on Artificial Neural Networks_ , pp. 496–510. Springer, 2019. * Kovachki et al. (2021a) Nikola Kovachki, Samuel Lanthaler, and Siddhartha Mishra. On universal approximation and error bounds for fourier neural operators. _Journal of Machine Learning Research_ , 22:Art–No, 2021a. * Kovachki et al. (2021b) Nikola Kovachki, Zongyi Li, Burigede Liu, Kamyar Azizzadenesheli, Kaushik Bhattacharya, Andrew Stuart, and Anima Anandkumar. Neural operator: Learning maps between function spaces. _arXiv preprint arXiv:2108.08481_ , 2021b. * Kurth et al. (2022) Thorsten Kurth, Shashank Subramanian, Peter Harrington, Jaideep Pathak, Morteza Mardani, David Hall, Andrea Miele, Karthik Kashinath, and Animashree Anandkumar. Fourcastnet: Accelerating global high-resolution weather forecasting using adaptive fourier neural operators. _arXiv preprint arXiv:2208.05419_ , 2022. * Lanthaler et al. (2022) Samuel Lanthaler, Siddhartha Mishra, and George E Karniadakis. Error estimates for deeponets: A deep learning framework in infinite dimensions. _Transactions of Mathematics and Its Applications_ , 6(1):tnac001, 2022. * Li et al. (2021) Zongyi Li, Nikola Borislavov Kovachki, Kamyar Azizzadenesheli, Burigede liu, Kaushik Bhattacharya, Andrew Stuart, and Anima Anandkumar. Fourier neural operator for parametric partial differential equations. In _International Conference on Learning Representations_ , 2021. URL https://openreview.net/forum?id=c8P9NQVtmnO. * Lin et al. (2021) Chensen Lin, Zhen Li, Lu Lu, Shengze Cai, Martin Maxey, and George Em Karniadakis. Operator learning for predicting multiscale bubble growth dynamics. _The Journal of Chemical Physics_ , 154(10):104118, 2021. * Lu et al. (2019) Lu Lu, Pengzhan Jin, and George Em Karniadakis. Deeponet: Learning nonlinear operators for identifying differential equations based on the universal approximation theorem of operators. _arXiv preprint arXiv:1910.03193_ , 2019. * Lu et al. (2021) Lu Lu, Pengzhan Jin, Guofei Pang, Zhongqiang Zhang, and George Em Karniadakis. Learning nonlinear operators via DeepONet based on the universal approximation theorem of operators. _Nature Machine Intelligence_ , 3(3):218–229, 2021. * Lu et al. (2022) Lu Lu, Xuhui Meng, Shengze Cai, Zhiping Mao, Somdatta Goswami, Zhongqiang Zhang, and George Em Karniadakis. A comprehensive and fair comparison of two neural operators (with practical extensions) based on fair data. _Computer Methods in Applied Mechanics and Engineering_ , 393:114778, 2022. * Mao et al. (2021) Zhiping Mao, Lu Lu, Olaf Marxen, Tamer A. Zaki, and George Em Karniadakis. DeepM&Mnet for hypersonics: predicting the coupled flow and finite-rate chemistry behind a normal shock using neural-network approximation of operators. _J. Comput. Phys._ , 447:Paper No. 110698, 24, 2021. ISSN 0021-9991. doi: 10.1016/j.jcp.2021.110698. URL https://doi.org/10.1016/j.jcp.2021.110698. * Mehta et al. (2021) Ishit Mehta, Michaël Gharbi, Connelly Barnes, Eli Shechtman, Ravi Ramamoorthi, and Manmohan Chandraker. Modulated periodic activations for generalizable local functional representations. In _Proceedings of the IEEE/CVF International Conference on Computer Vision_ , pp. 14214–14223, 2021. * Mescheder et al. (2019) Lars Mescheder, Michael Oechsle, Michael Niemeyer, Sebastian Nowozin, and Andreas Geiger. Occupancy networks: Learning 3d reconstruction in function space. In _Proceedings of the IEEE/CVF conference on computer vision and pattern recognition_ , pp. 4460–4470, 2019. * Mhaskar (1996) Hrushikesh N Mhaskar. Neural networks for optimal approximation of smooth and analytic functions. _Neural computation_ , 8(1):164–177, 1996. * Murshed et al. (2021) MG Sarwar Murshed, Christopher Murphy, Daqing Hou, Nazar Khan, Ganesh Ananthanarayanan, and Faraz Hussain. Machine learning at the network edge: A survey. _ACM Computing Surveys (CSUR)_ , 54(8):1–37, 2021\. * Pan et al. (2022) Shaowu Pan, Steven L Brunton, and J Nathan Kutz. Neural implicit flow: a mesh-agnostic dimensionality reduction paradigm of spatio-temporal data. _arXiv preprint arXiv:2204.03216_ , 2022. * Park et al. (2019) Jeong Joon Park, Peter Florence, Julian Straub, Richard Newcombe, and Steven Lovegrove. Deepsdf: Learning continuous signed distance functions for shape representation. In _Proceedings of the IEEE/CVF conference on computer vision and pattern recognition_ , pp. 165–174, 2019. * Pawlowski et al. (2017) Nick Pawlowski, Andrew Brock, Matthew CH Lee, Martin Rajchl, and Ben Glocker. Implicit weight uncertainty in neural networks. _arXiv preprint arXiv:1711.01297_ , 2017. * Prasthofer et al. (2022) Michael Prasthofer, Tim De Ryck, and Siddhartha Mishra. Variable-input deep operator networks. _arXiv preprint arXiv:2205.11404_ , 2022. * Raissi et al. (2019) M. Raissi, P. Perdikaris, and G. E. Karniadakis. Physics-informed neural networks: a deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations. _J. Comput. Phys._ , 378:686–707, 2019. ISSN 0021-9991. doi: 10.1016/j.jcp.2018.10.045. URL https://doi.org/10.1016/j.jcp.2018.10.045. * Ryck & Mishra (2022) Tim De Ryck and Siddhartha Mishra. Generic bounds on the approximation error for physics-informed (and) operator learning. In Alice H. Oh, Alekh Agarwal, Danielle Belgrave, and Kyunghyun Cho (eds.), _Advances in Neural Information Processing Systems_ , 2022. URL https://openreview.net/forum?id=bF4eYy3LTR9. * Seidman et al. (2022) Jacob H Seidman, Georgios Kissas, Paris Perdikaris, and George J Pappas. Nomad: Nonlinear manifold decoders for operator learning. _arXiv preprint arXiv:2206.03551_ , 2022. * Sirignano & Spiliopoulos (2018) Justin Sirignano and Konstantinos Spiliopoulos. DGM: a deep learning algorithm for solving partial differential equations. _J. Comput. Phys._ , 375:1339–1364, 2018. ISSN 0021-9991. doi: 10.1016/j.jcp.2018.08.029. URL https://doi.org/10.1016/j.jcp.2018.08.029. * Sitzmann et al. (2020) Vincent Sitzmann, Julien Martel, Alexander Bergman, David Lindell, and Gordon Wetzstein. Implicit neural representations with periodic activation functions. _Advances in Neural Information Processing Systems_ , 33:7462–7473, 2020. * Takamoto et al. (2022) Makoto Takamoto, Timothy Praditia, Raphael Leiteritz, Dan MacKinlay, Francesco Alesiani, Dirk Pflüger, and Mathias Niepert. PDEBench: An extensive benchmark for scientific machine learning. In _Thirty-sixth Conference on Neural Information Processing Systems Datasets and Benchmarks Track_ , 2022. URL https://openreview.net/forum?id=dh_MkX0QfrK. * Venturi & Casey (2023) Simone Venturi and Tiernan Casey. Svd perspectives for augmenting deeponet flexibility and interpretability. _Computer Methods in Applied Mechanics and Engineering_ , 403:115718, 2023. * von Oswald et al. (2020) Johannes von Oswald, Christian Henning, Benjamin F. Grewe, and João Sacramento. Continual learning with hypernetworks. In _International Conference on Learning Representations_ , 2020. URL https://openreview.net/forum?id=SJgwNerKvB. * Wang et al. (2021) Sifan Wang, Hanwen Wang, and Paris Perdikaris. Learning the solution operator of parametric partial differential equations with physics-informed deeponets. _Science advances_ , 7(40):eabi8605, 2021. * Wang et al. (2022) Sifan Wang, Hanwen Wang, and Paris Perdikaris. Improved architectures and training algorithms for deep operator networks. _Journal of Scientific Computing_ , 92(2):1–42, 2022. * Zhu et al. (2019) Yinhao Zhu, Nicholas Zabaras, Phaedon-Stelios Koutsourelakis, and Paris Perdikaris. Physics-constrained deep learning for high-dimensional surrogate modeling and uncertainty quantification without labeled data. _J. Comput. Phys._ , 394:56–81, 2019. ISSN 0021-9991. doi: 10.1016/j.jcp.2019.05.024. URL https://doi.org/10.1016/j.jcp.2019.05.024. ## Appendix A Notations We list the main notations in Table 3 which is not concretely described in this paper. $\mathcal{U}$ | Domain of input function ---|--- $\mathcal{Y}$ | Domain of output function $d_{u}$ | Dimension of $\mathcal{U}$ $d_{y}$ | Dimension of $\mathcal{Y}$ $d_{x}$ | Dimension of the codomain of input function $d_{s}$ | Dimension of the codomain of output function $\left\\{x_{1},\cdots,x_{m}\right\\}$ | Sensor points $m$ | Number of sensor points $\mathbb{R}^{d}$ | Euclidean space of dimension $d$ $C^{r}(\mathbb{R})$ | Set of functions that has continuous $r$-th derivative. $C(\mathcal{Y};\mathbb{R}^{d_{s}})$ | Set of continuous function from $\mathcal{Y}$ to $\mathbb{R}^{d_{s}}$ $C^{r}([-1,1]^{n};\mathbb{R})$ | Set of functions from $[-1,1]^{n}$ to $\mathbb{R}$ whose $r$-th partial derivatives are continuous. $n=O(\epsilon)$ | There exists a constant $C$ such that $n\leq C\cdot\epsilon,\forall\epsilon>0$ $n=\Omega(\epsilon)$ | There exists a constant $C$ such that $n\geq C\cdot\epsilon,\forall\epsilon>0$ $n=o(\epsilon)$ | $n/\epsilon$ converges to 0 as $\epsilon$ approaches to 0. Table 3: Notation ## Appendix B Original structure of DeepONet Figure 8: The structure of DeepONet For a clear understanding of previous works, we briefly leave a description of DeepONet. In particular, we explain the structure of the unstacked DeepONet in Lu et al. (2019) which is being widely used in various experiments of the papers. Note that Figure 4(a) represents the corresponding model which is called simply DeepONet throughout this paper. The overall architecture of the model is formulated as $\mathcal{R}_{\tau}\circ\mathcal{A}_{\beta}(u(x_{1}),\cdots,u(x_{m}))(y):=\langle\beta(u(x_{1}),\cdots,u(x_{m});\theta_{\beta}),\tau(y;\theta_{\tau})\rangle,$ where $\tau$ and $\beta$ are referred to as the trunk net and the branch net respectively. Note that $R_{\tau}$ and $A_{\beta}$ denote the reconstructor and the operator in Section 3.1. For $m$ fixed observation points $(x_{1},\cdots,x_{m})\in\mathcal{X}^{m}$, the unstacked DeepONet consists of an inner product of branch Net and trunk Net, which are fully connected neural networks. For a function $u$, the branch net receives pointwise projection values $(u(x_{1}),\cdots,u(x_{m}))$ as inputs to detect which function needs to be transformed. The trunk net queries a location $y\in\mathcal{Y}$ of interest where $\mathcal{Y}$ denotes the domain of output functions. It was revealed that the stacked DeepONet, the simplified version of the unstacked DeepONet, is a universal approximator in the set of continuous functions. Therefore, the general structure also becomes a universal approximator which enables close approximation by using a sufficient number of parameters. Motivated by the property, we focus on how large complexity should be required for DeepONet and its variants to achieve the desired error. ## Appendix C On complexity of DeepONet and its variants ### C.1 Proof of Theorem 2 on DeepONet complexity The following lemma implies that the class of neural networks is sufficiently efficient to approximate the inner product. ###### Lemma 1. For the number of basis $p\in\mathbb{N}$, consider the inner product function $\pi_{p}:[-1,1]^{2p}\rightarrow\mathbb{R}$ defined by $\pi_{p}(a_{1},\cdots,a_{p},b_{1},\cdots,b_{p}):=\sum_{i=1}^{p}a_{i}b_{i}=\langle(a_{1},\cdots,a_{p}),(b_{1},\cdots,b_{p})\rangle.$ For an arbitrary positive $t$, there exists a class of neural network $\mathcal{F}$ with universal activation $\sigma:\mathbb{R}\rightarrow\mathbb{R}$ such that the number of parameters of $\mathcal{F}$ is $O(p^{1+1/t}\epsilon^{-1/t})$ with $\inf_{f\in\mathcal{F}}\|f-\pi_{p}\|_{\infty}\leq\epsilon$. ###### Proof. Suppose that $t$ is a positive integer and the Sobolev space $W_{2t,2}$ is well defined. First, we would like to approximate the product function $\pi_{1}:[-1,1]^{2}\rightarrow\mathbb{R}$ which is defined as $\pi_{1}(a,b)=ab.$ Note that partial derivatives $D^{\mathbf{k}}\pi_{1}=0$ for all multi-index $\mathbf{k}\in\left\\{\mathbb{N}\bigcup\left\\{0\right\\}\right\\}^{2}$ such that $|\mathbf{k}|\geq 2$. For a multi-index $\mathbf{k}$ with $|\mathbf{k}|=1$, $D^{\mathbf{k}}\pi_{1}$ contains only one term which is either $a$ or $b$. In this case, we can simply observe that $\sum_{|\mathbf{k}|=1}\|D^{\mathbf{k}}\pi_{1}\|_{\infty}\leq 2\cdot 1=2$ by the construction of the domain $[-1,1]^{2}$ for $\pi_{1}$. And finally, $\displaystyle\|\pi_{1}\|_{\infty}\leq\|ab\|_{\infty}\leq\|a\|_{\infty}\|b\|_{\infty}\leq 1\cdot 1=1,$ so that a function $\pi_{1}/3$ should be contained in $\mathcal{W}_{r,2}$ for any $r\in\mathbb{N}$. In particular, $\pi_{1}/3$ lies in $\mathcal{W}_{2t,2}$ so that there exists a neural network approximation $f_{nn}$ in some class of neural network $\mathcal{F}^{*}$ with an universal activation function $\sigma$ such that the number of parameters of $\mathcal{F}^{*}$ is $O((\epsilon/3p)^{-2/2t})=O(p^{1/t}\epsilon^{-1/t})$, and $\displaystyle\|\pi_{1}/3-f_{nn}\|_{\infty}\leq\epsilon/3p,$ by Definition 1. Then the neural network $3f_{nn}$ approximates the function $\pi_{1}$ by an error $\epsilon/p$ which can be constructed by adjusting the last weight values directly involved in the output layer of neural network $f_{nn}$. Finally, we construct a neural network approximation for the inner product function $\pi_{p}$. Decompose the $2p-$dimensional inner product function $\pi_{p}$ into $p$ product functions $\left\\{\text{Proj}_{i}(\pi_{p})\right\\}_{i=1}^{p}$ which are defined as $\displaystyle\text{Proj}_{i}(\pi_{p}):\mathbb{R}^{2p}\rightarrow\mathbb{R},\quad\text{Proj}_{i}(\pi_{p})(a_{1},\cdots,a_{p},b_{1},\cdots,b_{p}):=\pi_{1}(a_{i},b_{i})=a_{i}b_{i},$ for $\forall i\in\left\\{1,\cdots,p\right\\}$. Then each function $\text{Proj}_{i}(\pi_{p})$ could be approximated within an error $\epsilon/p$ by neural network $NN_{i}$ which has $O(p^{1/t}\epsilon^{-1/t})$ parameters by the above discussion. Finally, by adding the last weight $[1,1,\cdots,1]\in\mathbb{R}^{1\times p}$ which has input as the outputs of $p$ neural networks $\left\\{NN_{i}\right\\}_{i=1}^{p}$, we can construct the neural network approximation $NN$ of $\pi_{p}=\sum_{i=1}^{p}\text{Proj}_{i}(\pi_{p})$ such that the number of parameters is $O(1+p+p\cdot p^{1/t}\epsilon^{-1/t})=O(p^{1+1/t}\epsilon^{-1/t})$. Class of neural network $\mathcal{F}$, which represents the structure of $NN$, satisfies the desired property. Obviously, the statement holds for an arbitrary real $t$ which is not an integer. ∎ Now we assume that $\mathcal{O}$ (defined in Eq. (8)) lies in the Sobolev space $\mathcal{W}_{r,d_{y}+m}$. Then, we can obtain the following lemma which presents the lower bound on the number of basis $p$ in DeepONet structure. Note that we apply $L_{\infty}$-norm for the outputs of branch net and trunk net which are multi-dimensional vectors. ###### Lemma 2. Let $\sigma:\mathbb{R}\rightarrow\mathbb{R}$ be a universal activation function in $C^{r}(\mathbb{R})$ such that $\sigma^{\prime}$ is a bounded variation. Suppose that the class of branch net $\mathcal{B}$ has a bounded Sobolev norm (i.e., $\|\beta\|_{r}^{s}\leq l_{1},\forall\beta\in\mathcal{B}$). Assume that supremum norm $\|\cdot\|_{\infty}$ of the class of trunk net $\mathcal{T}$ is bounded by $l_{2}$ and the number of parameters in $\mathcal{T}$ is $o(\epsilon^{-(d_{y}+m)/r})$. If any non-constant $\psi\in\mathcal{W}_{r,d_{y}+m}$ does not belong to any class of neural network, then the number of basis $p$ in $\mathcal{T}$ is $\Omega(\epsilon^{-d_{y}/r})$ when $d(\mathcal{F}_{\text{DeepONet}}(\mathcal{B},\mathcal{T});\mathcal{W}_{r,d_{y}+m})\leq\epsilon$. ###### Proof. To prove the above lemma by contradiction, we assume the opposite of the conclusion. Suppose that there is no constant $C$ that satisfies the inequality $p\geq C(\epsilon^{-d_{y}/r})$ for $\forall\epsilon>0$. In other words, there exists a sequence of DeepONet which has $\left\\{p_{n}\right\\}_{n=1}^{\infty}$ as the number of basis with a sequence of error $\left\\{\epsilon_{n}\right\\}_{n=1}^{\infty}$ in $\mathbb{R}$ such that $\epsilon_{n}\rightarrow 0$, and satisfies $\displaystyle p_{n}\leq\frac{1}{n}\epsilon_{n}^{-d_{y}/r}\quad\left(\text{i.e.,}\quad p_{n}=o(\epsilon_{n}^{-d_{y}/r})\quad\text{with respect to }n\right),$ (9) and $\displaystyle d(\mathcal{F}_{\text{DeepONet}}(\mathcal{B}_{n},\mathcal{T}_{n});\mathcal{W}_{r,d_{y}+m})\leq\epsilon_{n},$ where $\mathcal{B}_{n}$ and $\mathcal{T}_{n}$ denote the corresponding class sequence of branch net and trunk net respectively. Then, we can choose the sequence of branch net $\left\\{\beta_{n}:\mathbb{R}^{m}\rightarrow\mathbb{R}^{p_{n}}\right\\}_{n=1}^{\infty}$ and trunk net $\left\\{\tau_{n}:\mathbb{R}^{d_{y}}\rightarrow\mathbb{R}^{p_{n}}\right\\}_{n=1}^{\infty}$ satisfying $\displaystyle\|\mathcal{O}(\mathcal{E}(u),y)-\pi_{p_{n}}(\beta_{n}(\mathcal{E}(u)),\tau_{n}(y))\|_{\infty}\leq 2\epsilon_{n},\forall[\mathcal{E}(u),y]\in[-1,1]^{d_{y}+m},\mathcal{O}\in\mathcal{W}_{r,d_{y}+m}$ for the above sequence of DeepONet by the definition of $d(\mathcal{F}_{\text{DeepONet}}(\mathcal{B}_{n},\mathcal{T}_{n});\mathcal{W}_{r,d_{y}+m})$. Now, we would like to construct neural network approximations $f_{n}$ for the branch net $\beta_{n}$. By the assumption on the boundedness of $\mathcal{B}$, the $i$-th component $[\beta_{n}]_{i}$ of $\beta_{n}$ has a Sobolev norm bounded by $l_{1}$. In other words, $\|[\beta_{n}]_{i}/l_{1}\|_{r}^{s}\leq 1$ and therefore, $[\beta_{n}]_{i}/l_{1}$ is contained in $W_{1,m}$. Since $\sigma$ is a universal activation function, we can choose a neural network approximation $[f_{n}]_{i}$ of $[\beta_{n}]_{i}$ such that the number of parameters is $O((\epsilon_{n}/l_{1})^{-m/r})=O(\epsilon_{n}^{-m/r})$, and $\displaystyle\|[f_{n}]_{i}-[\beta_{n}]_{i}\|_{\infty}\leq\epsilon_{n}/l_{1}.$ Then, $f_{n}=(l_{1}[f_{n}]_{1},l_{1}[f_{n}]_{2},\cdots,l_{1}[f_{n}]_{p_{n}})$ becomes neural network approximation of $\beta_{n}$ which has $O(p_{n}\epsilon_{n}^{-m/r})$ parameters within an error $\epsilon_{n}$. Recall the target function corresponding to $m$ observation $\mathcal{E}(u)\in[-1,1]^{m}$ by $\mathcal{O}(\mathcal{E}(u),\cdot):\mathbb{R}^{d_{y}}\rightarrow\mathbb{R}$ which is defined in Eq. (8). Then, for $\forall\mathcal{E}(u)$, we can observe the following inequalities: $\|\mathcal{O}(\mathcal{E}(u),y)-\pi_{p_{n}}(f_{n}(\mathcal{E}(u)),\tau_{n}(y))\|_{\infty}\\\ \leq\|\mathcal{O}(\mathcal{E}(u),y)-\pi_{p_{n}}\left(\beta_{n}\left(\mathcal{E}(u)\right),\tau_{n}(y)\right)\|_{\infty}+\|\pi_{p_{n}}\left(\beta_{n}\left(\mathcal{E}\left(u\right)\right)-f_{n}(\mathcal{E}(u)\right),\tau_{n}(y))\|_{\infty}\\\ \leq 2\epsilon_{n}+\epsilon_{n}\|\tau_{n}(y)\|_{\infty}\leq\epsilon_{n}(2+l_{2}),$ (10) by the assumption on the boundedness of $\mathcal{T}$. Now we would like to consider the sequence of neural network which is an approximation of inner product between $p_{n}-$dimensional vector in $[-1,1]^{p_{n}}$. Note the following inequality $\displaystyle\|f_{n}(\mathcal{E}(u))\|_{\infty}$ $\displaystyle\leq\|f_{n}(\mathcal{E}(u))-\beta_{n}(\mathcal{E}(u))\|_{\infty}+\|\beta_{n}(\mathcal{E}(u))\|_{\infty}$ $\displaystyle\leq\epsilon_{n}+\|\beta_{n}(\mathcal{E}(u))\|_{r}^{s}$ $\displaystyle\leq\epsilon_{n}+l_{1}\leq 2l_{1},$ with $\|\tau_{n}\|_{\infty}\leq l_{2}$ for large $n$. It implies that $f_{n}(\mathcal{E}(u))/2l_{1}$ and $\tau_{n}(x)/l_{2}$ lie in $[-1,1]^{p_{n}}$. By Lemma 1, there exists a class of neural network $\mathcal{H}_{n}$ such that the number of parameters is $O(p_{n}^{1+1/2d_{y}r}\epsilon_{n}^{-1/2d_{y}r})$ and, $\displaystyle\inf_{h\in\mathcal{H}_{n}}\|h-\pi_{p_{n}}\|_{\infty}\leq\epsilon_{n}$ where $\pi_{p_{n}}$ is the inner product corresponding to $p_{n}-$dimensional vector. Choose a neural network $h_{n}:[-1,1]^{2p_{n}}\rightarrow\mathbb{R}$ such that $\|h_{n}-\pi_{p_{n}}\|_{\infty}\leq 2\epsilon_{n}$. Then, by the triangular inequality, $\|\mathcal{O}(\mathcal{E}(u),y)-2l_{1}l_{2}h_{n}(f_{n}(\mathcal{E}(u))/2l_{1},\tau_{n}(y)/l_{2})\|_{\infty}\\\ \leq\|\mathcal{O}(\mathcal{E}(u),y)-\pi_{p_{n}}(f_{n}(\mathcal{E}(u)),\tau_{n}(y))\|_{\infty}\\\ +2l_{1}l_{2}\|\pi_{p_{n}}(f_{n}(\mathcal{E}(u))/2l_{1},\tau_{n}(y)/l_{2})-h_{n}(f_{n}(\mathcal{E}(u))/2l_{1},\tau_{n}(y)/l_{2})\|_{\infty}\\\ \leq\epsilon_{n}(2+l_{2})+2l_{1}l_{2}(2\epsilon_{n})=\epsilon_{n}(2+l_{2}+4l_{1}l_{2}).$ (11) Finally, we compute the number of parameters which is required to implement the function $2l_{1}l_{2}h_{n}(f_{n}(\mathcal{E}(u))/2l_{1},\tau_{n}(y)/l_{2})$. The only part that needs further consideration is scalar multiplication. Since we need one weight to multiply a constant with one real value, three scalar multiplications $\displaystyle h_{n}(f_{n}(\mathcal{E}(u))/2l_{1},\tau_{n}(y)/l_{2})\mapsto 2l_{1}l_{2}h_{n}(f_{n}(\mathcal{E}(u))/2l_{1},\tau_{n}(y)/l_{2}),$ $\displaystyle f_{n}(\mathcal{E}(u))\mapsto f_{n}(\mathcal{E}(u))/2l_{1},\quad\text{and}\quad\tau_{n}(x)\mapsto\tau_{n}(y)/l_{2},$ require $1,p_{n},p_{n}$-parameters respectively. Combining all the previous results with the size of trunk net, the total number of parameters is obtained in the form of $\displaystyle O(1+2p_{n}+p_{n}^{1+1/2d_{y}r}\epsilon_{n}^{-1/2d_{y}r}+p_{n}\epsilon_{n}^{-m/r})+o(\epsilon_{n}^{-(d_{y}+m)/r})=o(\epsilon_{n}^{-(d_{y}+m)/r}),$ since the initial assumption (9) on the number of basis gives the following inequality. $\displaystyle p_{n}^{1+1/2d_{y}r}\epsilon_{n}^{-1/2d_{y}r}+p_{n}\epsilon_{n}^{-m/r}$ $\displaystyle\leq p_{n}(p_{n}^{1/2d_{y}r}\epsilon_{n}^{-1/2d_{y}r}+\epsilon_{n}^{-m/r})$ $\displaystyle\leq\frac{1}{n}\epsilon_{n}^{-d_{y}/r}(\epsilon_{n}^{-1/2r^{2}-1/2d_{y}r}+\epsilon_{n}^{-m/r})$ $\displaystyle\leq\frac{1}{n}\epsilon_{n}^{-d_{y}/r}2\epsilon_{n}^{-m/r}=\frac{2}{n}\epsilon_{n}^{-(d_{y}+m)/r}.$ On the one hand, the sequence of function $\left\\{2l_{1}l_{2}h_{n}(f_{n}(\mathcal{E}(u))/2l_{1},\tau_{n}(y)/l_{2})\right\\}_{n=1}^{\infty}$ is an sequence of approximation for $\mathcal{O}(\mathcal{E}(u),y)$ within a corresponding sequence of error $\left\\{\epsilon_{n}(2+l_{2}+4l_{1}l_{2})\right\\}_{n=1}^{\infty}$. Denote the sequence of the class of neural networks corresponding to the sequence of the function $\left\\{2l_{1}l_{2}h_{n}(f_{n}(\mathcal{E}(u))/2l_{1},\\\ \tau_{n}(y)/l_{2})\right\\}_{n=1}^{\infty}$ by $\left\\{\mathcal{F}_{n}\right\\}_{n=1}^{\infty}$. By the assumption, Theorem 1 implies the number of parameters in $\left\\{\mathcal{F}_{n}\right\\}_{n=1}^{\infty}$ is $\Omega((\epsilon_{n}(2+l_{2}+4l_{1}l_{2}))^{-(d_{y}+m)/r})=\Omega(\epsilon_{n}^{-(d_{y}+m)/r})$. Therefore, the initial assumption (9) would result in a contradiction so the desired property is valid. ∎ Note that the assumption on the boundedness of the trunk net could be valid if we use the bounded universal activation function $\sigma$. Using the above results, we can prove our main theorem, Theorem 2. ###### Proof of Theorem 2. Denote the number of parameters in $\mathcal{T}$ by $N_{\mathcal{T}}$. Suppose that there is no constant $C$ satisfies the inequality $N_{\mathcal{T}}\geq C\epsilon^{-d_{y}/r},\forall\epsilon>0$. That is, there exists a sequence of DeepONet with the corresponding sequence of trunk net class $\left\\{\mathcal{T}_{n}\right\\}_{n=1}^{\infty}$ and sequence of error $\left\\{\epsilon_{n}\right\\}_{n=1}^{\infty}$ such that $\epsilon_{n}\rightarrow 0$, and it satisfies $\displaystyle N_{\mathcal{T}_{n}}<\frac{1}{n}\epsilon_{n}^{-d_{y}/r}\left(\text{i.e.,}N_{\mathcal{T}_{n}}=o(\epsilon_{n}^{-d_{y}/r})\text{ with respect to }n\right).$ Note that the above implies $N_{\mathcal{T}_{n}}=o(\epsilon_{n}^{-(d_{y}+m)/r})$. On the one hand, the following inequality holds where $\mathcal{B}_{n}$ denotes the corresponding class sequence of branch net. $\displaystyle d(\mathcal{F}_{\text{DeepONet}}(\mathcal{B}_{n},\mathcal{T}_{n});\mathcal{W}_{r,d_{y}+m})\leq\epsilon_{n}.$ Since $\sigma$ is bounded, $\mathcal{T}_{n}$ consists of bounded functions with respect to the supremum norm. Therefore, if we apply the Lemma 2 with respect to $n$, the number of basis $p_{n}$ should be $\Omega(\epsilon_{n}^{-d_{y}/r})$. Since $p_{n}$ is also the number of output dimensions for the class of trunk net $\mathcal{T}_{n}$, the number of parameters in $\mathcal{T}_{n}$ should be larger than $p_{n}=\Omega(\epsilon_{n}^{-d_{y}/r})$. This leads to a contradiction. ∎ Finally, we present a lower bound on the total number of parameters of DeepONet, considering the size of the branch net. Keep in mind that the proof of this theorem can be applied to other variants of DeepONet. ###### Theorem 3. (Total Complexity of DeepONet) Let $\sigma:\mathbb{R}\rightarrow\mathbb{R}$ be a universal activation function in $C^{r}(\mathbb{R})$ such that $\sigma$ and $\sigma^{\prime}$ are bounded. Suppose that the class of branch net $\mathcal{B}$ has a bounded Sobolev norm (i.e., $\|\beta\|_{r}^{s}\leq l_{1},\forall\beta\in\mathcal{B}$). If any non-constant $\psi\in\mathcal{W}_{r,n}$ does not belong to any class of neural network, then the number of parameters in DeepONet is $\Omega(\epsilon^{-(d_{y}+m)/R})$ for any $R>r$ when $d(\mathcal{F}_{\text{DeepONet}}(\mathcal{B},\mathcal{T});\mathcal{W}_{r,d_{y}+m})\leq\epsilon$. ###### Proof. For a positive $\epsilon<l_{1}$, suppose that there exists a class of branch net $\mathcal{B}$ and trunk net $\mathcal{T}$ such that $d(\mathcal{F}_{\text{DeepONet}}(\mathcal{B},\mathcal{T});\mathcal{W}_{r,d_{y}+m})\leq\epsilon$. By the boundedness of $\sigma$, there exists a constant $l_{2}$ which is the upper bound on the supremum norm $\|\cdot\|_{\infty}$ of the trunk net class $\mathcal{T}$. Let us denote the number of parameters in $\mathcal{F}_{\text{DeepONet}}$ by $N_{\mathcal{F}_{\text{DeepONet}}}$. Using the Lemma 1 to replace DeepONet’s inner products with neural networks as in the inequality (11), we can construct a class of neural network $\mathcal{F}$ such that the number of parameters $\mathcal{F}$ is $O(N_{\mathcal{F}_{\text{DeepONet}}}+p^{1+1/t}\epsilon^{-1/t})$ and, $\displaystyle d(\mathcal{F};\mathcal{W}_{r,d_{y}+m})\leq(1+l_{1}l_{2})\epsilon.$ Suppose that $N_{\mathcal{F}_{\text{DeepONet}}}=o(\epsilon^{-(d_{y}+m)/r})$. Then, by Theorem 1, $p^{1+1/t}\epsilon^{-1/t}$ should be $\Omega(\epsilon^{-(d_{y}+m)/r})$. Since $t$ can be arbitrary large, the number of basis $p$ should be $\Omega(\epsilon^{-(d_{y}+m)/R})$ for any $R>r$. ∎ ### C.2 Complexity analysis on variants of DeepONet. We would like to verify that variant models of DeepONet require numerous units in the first hidden layer of the target network. Now we denote the class of Pre-net in Shift-DeepONet and flexDeepONet by $\mathcal{P}$. The class of Shift-DeepONet and flexDeepONet will be written as $\mathcal{F}_{\text{shift- DeepONet}}(\mathcal{P},\mathcal{B},\mathcal{T})$ and $\mathcal{F}_{\text{flexDeepONet}}(\mathcal{P},\mathcal{B},\mathcal{T})$ respectively. The structure of Shift-DeepONet can be summarized as follows. Denote the width of the first hidden layer of the target network by $w$. We define the pre-net as $\rho=[\rho_{1},\rho_{2}]:\mathbb{R}^{m}\rightarrow\mathbb{R}^{w\times({d_{y}+1})}$ where $\rho_{1}:\mathbb{R}^{m}\rightarrow\mathbb{R}^{w\times d_{y}}$ and $\rho_{2}:\mathbb{R}^{m}\rightarrow\mathbb{R}^{w}$, the branch net as $\beta:\mathbb{R}^{m}\rightarrow\mathbb{R}^{p}$, and the trunk net as $\tau:\mathbb{R}^{w}\rightarrow\mathbb{R}^{p}$. The Shift-DeepONet $f_{\text{Shift-DeepONet}}(\rho,\beta,\tau)$ is defined as $\displaystyle f_{\text{Shift- DeepONet}}(\rho,\beta,\tau)(\mathcal{E}(u),y):=\pi_{p}(\beta(\mathcal{E}(u)),\tau(\Phi(\rho_{1}(\mathcal{E}(u)),y)+\rho_{2}(\mathcal{E}(u))))$ where $\Phi$ is defined in Eq. (12). We claim that it does not improve performance for the branch net to additionally output the weights on the first layer of a target network. The following lemma shows that the procedure can be replaced by a small neural network structure. ###### Lemma 3. Consider a function $\Phi:\mathbb{R}^{d_{y}(w+1)}\rightarrow\mathbb{R}^{w}$ which is defined below. $\Phi(x_{1},\cdots,x_{d_{y}},\cdots,x_{(d_{y}-1)w+1}\cdots,x_{d_{y}w},y_{1},\cdots,y_{d_{y}})\\\ :=\left(\sum_{i=1}^{d_{y}}x_{i}y_{i},\cdots,\sum_{i=1}^{d_{y}}x_{(d_{y}-1)w+i}y_{i}\right).$ (12) For any arbitrary positive $t$, there exists a class of neural network $\mathcal{F}$ with universal activation $\sigma:\mathbb{R}\rightarrow\mathbb{R}$ such that the number of parameters of $\mathcal{F}$ is $O(wd_{y}^{1+1/t}\epsilon^{-1/t})$ with $\inf_{f\in\mathcal{F}}\|f-\Phi\|_{\infty}\leq\epsilon$. ###### Proof. Using the Lemma 1, we can construct a sequence of neural network $\left\\{f_{i}\right\\}_{i=1}^{w}$ which is an $\epsilon$-approximation of the inner product with $O(d_{y}^{1+{1/t}}\epsilon^{-1/t})$ parameters. If we combine all of the $w$ approximations, we get the desired neural network. ∎ Now we present the lower bound on the number of parameters for Shift-DeepONet. We derive the following theorem with an additional assumption that the class of trunk net is Lipschitz continuous. The function $\tau:\mathbb{R}^{d_{y}}\rightarrow\mathbb{R}^{p}$ is called Lipschitz continuous if there exists a constant $C$ such that $\|\tau(y_{1})-\tau(y_{2})\|_{1}\leq C\|y_{1}-y_{2}\|_{1}.$ For the neural network $f$, the upper bound of the Lipschitz constant for $f$ could be obtained as $L^{k-1}\Pi_{i=1}^{k}\|W^{i}\|_{1}$, where $L$ is the Lipschitz constant of $\sigma$ and the norm $\|\cdot\|_{1}$ denotes the matrix norm induced by vector $1$-norms. We can impose constraints on the upper bound of the weights, which consequently enforce affine transformation $W^{i}$ to be bounded with respect to the $L^{1}$ norm. Therefore, we can guarantee the Lipschitz continuity of the entire neural network in this way. We would like to remark on the validity of the weight assumptions in the theorem since the bounded assumptions of the weight may be a possible reason for increasing the number of parameters. However, the definition of Sobolev space forces all elements to have the supremum norm $\|\cdot\|_{\infty}$ less than 1. It may be somewhat inefficient to insist on large weights for approximating functions with a limited range. ###### Theorem 4. Let $\sigma:\mathbb{R}\rightarrow\mathbb{R}$ be a universal activation function in $C^{r}(\mathbb{R})$ such that $\sigma$ and $\sigma^{\prime}$ are bounded. Suppose that the class of branch net $\mathcal{B}$ and pre-net $\mathcal{P}$ has a bounded Sobolev norm (i.e., $\|\beta\|_{r}^{s}\leq l_{1},\forall\beta\in\mathcal{B}$, and $\|\rho\|_{r}^{s}\leq l_{3},\forall\rho\in\mathcal{P}$) and any neural network in the class of trunk net $\mathcal{T}$ is Lipschitz continuous with constant $l_{2}$. If any non- constant $\psi\in\mathcal{W}_{r,n}$ does not belong to any class of neural network, then the number of parameters in $\mathcal{T}$ is $\Omega(\epsilon^{-d_{y}/r})$ when $d(\mathcal{F}_{\text{shift- DeepONet}}(\mathcal{P},\mathcal{B},\mathcal{T});\mathcal{W}_{r,d_{y}+m})\leq\epsilon$. ###### Proof. Denote the number of parameters in $\mathcal{T}$ by $N_{\mathcal{T}}$. Suppose that there exists a sequence of pre-net $\left\\{\rho_{n}\right\\}_{n=1}^{\infty}$, branch net $\left\\{\beta_{n}\right\\}_{n=1}^{\infty}$ and trunk net $\left\\{\tau_{n}\right\\}_{n=1}^{\infty}$ with the corresponding sequence of error $\left\\{\epsilon_{n}\right\\}_{n=1}^{\infty}$ such that $\epsilon_{n}\rightarrow 0$ and, $\displaystyle N_{\mathcal{T}}=o(\epsilon_{n}^{-d_{y}/r}),\quad\text{and}\quad\sup_{\psi\in\mathcal{W}_{r,d_{y}+m}}\|f_{\text{shift- DeepONet}}(\rho_{n},{\beta_{n}},\tau_{n})-\psi\|_{\infty}\leq\epsilon_{n}.$ The proof can be divided into three parts. Firstly, we come up with a neural network approximation $\rho^{NN}_{n}$ of $\rho_{n}$ of which size is $O(w\cdot\epsilon_{n}^{-m/r})$ within an error $\epsilon_{n}$. Next, construct a neural network approximation of $\Phi$ using the Lemma 3. Finally, the inner product $\pi_{p_{n}}(\beta_{n},\tau_{n})$ is replaced with a neural network as in (11) of Lemma 2. Since all techniques such as triangular inequality are consistent with the previous discussion, we will briefly explain why additional Lipschitz continuity is required for the trunk network, and omit the details. Approximating the Pre-Net of Shift DeepOnet, which is not in DeepOnet, inevitably results in an error in the input of the trunk net. We are reluctant to allow this error to change the output of the trunk net significantly. In this situation, the Lipschitz continuity provides the desired result. ∎ For $d_{y}=1$, the additional rotation is only multiplying by $1$ or $-1$. Since the weight and bias of the first layer alone can cover the scalar multiplication, flexDeepONet has the same properties as Shift-DeepONet in the above theorem. ###### Theorem 5. Consider the case $d_{y}=1$. Let $\sigma:\mathbb{R}\rightarrow\mathbb{R}$ be a universal activation function in $C^{r}(\mathbb{R})$ such that $\sigma$ and $\sigma^{\prime}$ are bounded. Suppose that the class of branch net $\mathcal{B}$ and pre-net $\mathcal{P}$ has a bounded Sobolev norm(i.e., $\|\beta\|_{r}^{s}\leq l_{1},\forall\beta\in\mathcal{B}$, and $\|\rho\|_{r}^{s}\leq l_{3},\forall\rho\in\mathcal{P}$), and any neural network in the class of trunk net $\mathcal{T}$ is Lipschitz continuous with constant $l_{2}$. If any non-constant $\psi\in\mathcal{W}_{r,n}$ does not belong to any class of neural network, then the number of parameters in $\mathcal{T}$ is $\Omega(\epsilon^{-d_{y}/r})$ when $d(\mathcal{F}_{\text{flexDeepONet}}(\mathcal{B},\mathcal{T});\mathcal{W}_{r,d_{y}+m})\leq\epsilon$. ###### Proof. The main difference between flexDeepONet and Shift-DeepONet, which is not mentioned earlier, is that the branch net affects the bias of the output layer. However, adding the values of the two neurons can be implemented in a neural network by adding only one weight of value 1 for each neuron, so all of the previous discussion are valid. ∎ In fact, NOMAD can be substituted with the embedding method handled by Galanti & Wolf (2020). Suppose that the branch net of NOMAD is continuously differentiable. Let’s also assume that the Lipschitz constant of branch net and trunk net is bounded. We would like to briefly cite the relevant theorem here. ###### Theorem 6. (Galanti & Wolf (2020)) Suppose that $\sigma$ is a universal activation in $C^{1}(\mathbb{R})$ such that $\sigma^{\prime}$ is a bounded variation on $\mathbb{R}$. Additionally, suppose that there is no class of neural network that can represent any function in $\mathcal{W}_{1,d_{y}+m}$ other than a constant function. If the weight on the first layer of target network in NOMAD is bounded with respect to $L^{1}$-norm, then $d(\mathcal{N};\mathcal{W}_{1,d_{y}+m})\leq\epsilon$ implies the number of parameters in $\mathcal{N}$ is $\Omega(\epsilon^{-min(d_{y}+m,2\cdot m_{y})})$ where $\mathcal{N}$ denotes the class of function contained as a target network of NOMAD. ## Appendix D Chunked embedding method The HyperDeepONet may suffer from the large complexity of the hypernetwork when the size of the target network increases. Although even a small target network can learn various operators with proper performance, a larger target network will be required for more accurate training. To take into account this case, we employ a chunk embedding method which is developed by von Oswald et al. (2020). The original hypernetwork was designed to generate all of the target network’s weights so that the complexity of hypernetwork could be larger than the complexity of the target network. Such a problem can be overcome by using a hypernetwork with smaller outputs. Figure 9: Chunk Embedding Method More specifically, Figure 9 describes how the chunk embedding method reduces the number of learnable parameters. First, they partition the weights and biases of the target network. The hypernetwork then creates the weights of each parameter group by using the sensor values $\left\\{u(x_{i})\right\\}_{i=1}^{m}$ with a latent vector $z_{j}$. All groups share the hypernetwork so that the complexity decreases by a factor of the number of groups. Since the latent vectors $\\{z_{j}\\}_{j=1}^{N_{c}}$ learn the characteristics of each group during the training period, the chunked embedding method preserves the expressivity of the hypernetwork. The chunked architecture is a universal approximator for the set of continuous functions with the existence of proper partitions (Proposition 1 in von Oswald et al. (2020)). We remark that the method can also generate the additional weights and discard the unnecessary ones when the number of the target network’s parameters is not multiple of $N_{C}$, which is the number of group. ## Appendix E Experimental details | DeepONet | Shift | Flex | NOMAD | Hyper(ours) ---|---|---|---|---|--- Identity | 0.0005 | 0.0002 | 0.0005 | 0.0001 | 0.0001 Differentiation | 0.0005 | 0.0002 | 0.0005 | 0.0001 | 0.0001 Advection | 0.0005 | 0.0001 | 0.0001 | 0.0002 | 0.0005 Burgers | 0.0001 | 0.0005 | 0.0002 | 0.0001 | 0.0001 Shallow | 0.0001 | 0.0005 | - | 0.0001 | 0.0005 Table 4: Setting of the decay rate for each operator problem In most experiments, we follow the hyperparameter setting in Lu et al. (2019; 2021; 2022). We use ADAM in Kingma & Ba (2015) as an optimizer with a learning rate of $1e-3$ and zero weight decay. In all experiments, an InverseTimeDecay scheduler was used, and the step size was fixed to 1. In the experiments of identity and differentiation operators, grid search was performed using the sets 0.0001, 0.0002, 0.0005, 0.001, 0.002, 0.005 for decay rates. The selected values of the decay rate for each model can be found in Table 4. ### E.1 Identity As in the text, we developed an experiment to learn the identity operator for the 20th-order Chebyshev polynomials (Figure 10). Note that the absolute value of the coefficients of all orders is less than or equal to $1/4$. We discretize the domain $[-1,1]$ with a spatial resolution of 50. For the experiments described in the text, we construct all of the neural networks with tanh. We use 1,000 training pairs and 200 pairs to validate our experiments. The batch size during the training is determined to be 5,000, which is one-tenth the size of the entire dataset. ### E.2 Differentiation In this experiment, we set functions whose derivatives are 20th-order Chebyshev polynomials as input functions. As mentioned above, all coefficients of the Chebyshev polynomial are between $-1/4$ and $1/4$. We use the 100 uniform grid on the domain [-1, 1]. The number of training and test samples is 1,000 and 200, respectively. We use a batch size of 10,000, which is one-tenth the size($100,000=100\cdot 1,000$) of the entire dataset. Figure 10: One test data example of differentiation operator problem. ### E.3 Advection equation We consider the linear advection equation on the torus $\mathbb{T}:=\mathbb{R}/\mathbb{Z}$ as follows: $\begin{cases}\frac{\partial w}{\partial t}+c\cdot\frac{\partial w}{\partial x}=0.\\\ w(x,0)=w_{0}(x),&x\in\mathbb{T},\end{cases}$ (13) where $c$ is a constant which denotes the propagation speed of $w$. By constructing of the domain as $\mathbb{T}$, we implicitly assume the periodic boundary condition. In this paper, we consider the case when $c=1$. Our goal is to learn the operator which maps $w_{0}(x)(=w(0,x))$ to $w(0.5,x)$. We use the same data as in Lu et al. (2022). We discretize the domain $[0,1]$ with a spatial resolution of 40. The number of training samples and test samples is 1,000 and 200, respectively. We use the full batch for training so that the batch size is $40\cdot 1,000=40,000$. ### E.4 Burgers’ equation We consider the 1D Burgers’ equation which describes the movement of the viscous fluid $\begin{cases}\frac{\partial w}{\partial t}=-u\cdot\frac{\partial w}{\partial x}+\nu\frac{\partial^{2}w}{\partial x^{2}},&(x,t)\in(0,1)\times(0,1],\\\ w(x,0)=w_{0}(x),&x\in(0,1),\end{cases}$ (14) where $w_{0}$ is the initial state and $\nu$ is the viscosity. Our goal is to learn the nonlinear solution operator of the Burgers’ equation, which is a mapping from the initial state $w_{0}(x)(=w(x,0))$ to the solution $w(x,1)$ at $t=1$. We use the same data of Burgers’ equation provided in Li et al. (2021). The initial state $w_{0}(x)$ is generated from the Gaussian random field $\mathcal{N}(0,5^{4}(-\Delta+25I)^{-2})$ with the periodic boundary conditions. The split step and the fine forward Euler methods were employed to generate a solution at $t=1$. We set viscosity $\nu$ and a spatial resolution to 0.1 and $2^{7}=128$, respectively. The size of the training sample and test sample we used are 1,000 and 200, respectively. We take the ReLU activation function and the InverseTimeDecay scheduler to experiment with the same setting as in Lu et al. (2022). For a fair comparison, all experiments on DeepONet retained the hyperparameter values used in Lu et al. (2022). We use the full batch so that the batch size is 1,280,000. ### E.5 Shallow Water equation The shallow water equations are hyperbolic PDEs which describe the free- surface fluid flow problems. They are derived from the compressible Navier- Stokes equations. The physical conservation laws of the mass and the momentum holds in the shock of the solution. The specific form of the equation can be written as $\begin{cases}\frac{\partial h}{\partial t}+\frac{\partial}{\partial x}(hu)+\frac{\partial}{\partial y}(hv)=0,\\\ \frac{\partial(hu)}{\partial t}+\frac{\partial}{\partial x}(u^{2}h+\frac{1}{2}gh^{2})+\frac{\partial}{\partial y}(huv)=0,\\\ \frac{\partial(hv)}{\partial t}+\frac{\partial}{\partial y}(v^{2}h+\frac{1}{2}gh^{2})+\frac{\partial}{\partial x}(huv)=0,\\\ h(0,x_{1},x_{2})=h_{0}(x_{1},x_{2}),\end{cases}$ (15) for $t\in[0,1]$ and $x_{1},x_{2}\in[-2.5,2.5]$ where $h(t,x_{1},x_{2})$ denotes the height of water with horizontal and vertical velocity $(u,v)$. $g$ denotes the gravitational acceleration. In this paper, we aim to learn the operator $h_{0}(x_{1},x_{2})\mapsto\left\\{h(t,x_{1},x_{2})\right\\}_{t\in[1/4,1]}$ without the information of $(u,v)$. For the sampling of initial conditions and the corresponding solutions, we directly followed the setting of Takamoto et al. (2022). The 2D radial dam break scenario is considered so that the initialization of the water height is generated as a circular bump in the center of the domain. The initial condition is generated by $h(t=0,x_{1},x_{2})=\begin{cases}2.0,&\text{for $r<\sqrt{x_{1}^{2}+x_{2}^{2}}$}\\\ 1.0,&\text{for $r\geq\sqrt{x_{1}^{2}+x_{2}^{2}}$}\end{cases}$ (16) with the radius $r$ randomly sampled from $U[0.3,0.7]$. The spatial domain is determined to be a 2-dimensional rectangle $[-2.5,2.5]\times[-2.5,2.5]$. We use $256=16^{2}$ grids for the spatial domain. We train the models with three snapshots at $t=0.25,0.5,0.75$, and predict the solution $h(t,x_{1},x_{2})$ for four snapshots $t=0.25,0.5,0.75,1$ on the same grid. We use 100 training samples and the batch size is determined to be 25,600. ## Appendix F Additional experiments ### F.1 Comparison under various conditions The experimental results under various conditions are included in Table 5 by modifying the network structure, activation function, and number of sensor points. Although the DeepONet shows good performance in certain settings, the proposed HyperDeepONet shows good performance without dependency on the various conditions. Activation: ReLU, $M=30$ | DeepONet | Shift | Flex | NOMAD | Hyper(Ours) ---|---|---|---|---|--- Target network | $d_{y}$-30-30-30-1 | 0.16797 | 1.30852 | 1.04292 | 0.27209 | 0.02059 $d_{y}$-50-50-1 | 0.04822 | 1.08760 | 1.11957 | 0.21391 | 0.05562 Activation: ReLU, $M=100$ | DeepONet | Shift | Flex | NOMAD | Hyper(Ours) ---|---|---|---|---|--- Target network | $d_{y}$-30-30-30-1 | 0.02234 | 1.08310 | 1.03741 | 0.19089 | 0.01743 $d_{y}$-50-50-1 | 0.07255 | 1.47373 | 1.13217 | 0.14020 | 0.04645 Activation: PReLU, $M=30$ | DeepONet | Shift | Flex | NOMAD | Hyper(Ours) ---|---|---|---|---|--- Target network | $d_{y}$-30-30-30-1 | 0.11354 | 1.09395 | 1.03502 | 0.25651 | 0.02844 $d_{y}$-50-50-1 | 0.00873 | 1.14073 | 1.06947 | 0.04054 | 0.04302 Activation: PReLU, $M=100$ | DeepONet | Shift | Flex | NOMAD | Hyper(Ours) ---|---|---|---|---|--- Target network | $d_{y}$-30-30-30-1 | 0.01035 | 1.05080 | 1.07791 | 0.16592 | 0.01083 $d_{y}$-50-50-1 | 0.07255 | 1.47373 | 1.13217 | 0.14020 | 0.04645 Table 5: The relative $L^{2}$ test errors for experiments on training the identity operator under various conditions ### F.2 Varying the number of layers in branch net and hypernetwork Figure 11 compares the relative $L^{2}$ error of the training data and test data for the DeepONet and the HyperDeepONet by varying the number of layers in the branch net and the hypernetwork while maintaining the same small target network. Note that the bottom 150 training and test data with lower errors are selected to observe trends cleary. The training and test error for the DeepONet is not reduced despite the depth of the branch net becoming larger. This is a limitation of DeepONet’s linear approximation. DeepONet approximates the operator with the dot product of the trunk net’s output that approximates the basis of the target function and the branch net’s output that approximates the target function’s coefficient. Even if a more accurate coefficient is predicted by increasing the depth of the branch net, the error does not decrease because there is a limit to approximating the operator with a linear approximation using the already fixed trunk net. The HyperDeepONet approximates the operator with a low test error in all cases with a different number of layers. Figure 11 shows that the training error of the HyperDeepONet remains small as the depth of the hypernetwork increases, while the test error increases. The increasing gap between the training and test errors is because of overfitting. HyperDeepONet overfits the training data because the learnable parameters of the model are more than necessary to approximate the target operator. Figure 11: Varying the number of layers of branch net and hypernetwork in DeepONet and HyperDeepONet for identity operator problem (left) and differentiation operator problem (right). ### F.3 Comparison of HyperDeepONet with Fourier Neural Operator The Fourier Neural Operator (FNO) (Li et al., 2021) is a well-known method for operator learning. Lu et al. (2022) consider 16 different tasks to explain the relative performance of the DeepONet and the FNO. They show that each method has its advantages and limitations. In particular, DeepONet has a great advantage over FNO when the input function domain is complicated, or the position of the sensor points is not uniform. Moreover, the DeepONet and the HyperDeepONet enable the inference of the solution of time-dependent PDE even in a finer time grid than a time grid used for training, e.g.the continuous- in-time solution operator of the shallow water equation in our experiment. Since the FNO is image-to-image based operator learning model, it cannot obtain a continuous solution operator over time $t$ and position $x_{1},x_{2}$. In this paper, while retaining these advantages of DeepONets, we focused on overcoming the difficulties of DeepONets learning complex target functions because of linear approximation. Therefore, we mainly compared the vanilla DeepONet and its variants models to learn the complex target function without the result of the FNO. Model | HyperDeepONet (ours) | Fourier Neural Operator ---|---|--- Mode 2 | Mode 4 | Mode 8 | Mode 16 Identity | 0.0358 | 0.0005 | 0.0004 | 0.0003 | 0.0004 Differentiation | 0.1268 | 0.8256 | 0.6084 | 0.3437 | 0.0118 $\\#$Param | | 15741(or 16741) --- 20993 | 29185 | 45569 | 78337 Table 6: The relative $L^{2}$ test errors and the number of parameters for the identity and differentiation operator problems using HyperDeepONet and FNO with different number of modes. $\\#$Param denote the number of learnable parameters. Table 6 shows the simple comparison of the HyperDeepONet with the FNO for the identity operator and differentiation operator problems. Although the FNO structure has four Fourier layers, we use only one Fourier layer with 2,4,8, and 16 modes for fair comparison using similar number of parameters. The FNO shows a better performance than the HyperDeepONet for the identity operator problem. Because the FNO has a linear transform structure with a Fourier layer, the identity operator is easily approximated even with the 2 modes. In contrast, the differentiation operator is hard to approximate using the FNO with 2, 4, and 8 modes. Although the FNO with mode 16 can approximate the differentiation operator with better performance than the HyperDeepONet, it requires approximately 4.7 times as many parameters as the HyperDeepONet. Model | DeepONet | Shift | Flex | NOMAD | Hyper(Ours) ---|---|---|---|---|--- Advection | $\\#$Param | 274K | 281K | 282K | 270K | 268K Rel error | 0.0046 | 0.0095 | 0.0391 | 0.0083 | 0.0048 Burgers | $\\#$Param | 115K | 122K | 122K | 117K | 114K Rel error | 0.0391 | 0.1570 | 0.1277 | 0.0160 | 0.0196 Shallow | $\\#$Param | 107K | 111K | - | 117K | 101K Rel error | 0.0279 | 0.0299 | - | 0.0167 | 0.0148 Shallow w/ small param | $\\#$Param | 6.5K | 8.5K | - | 6.4K | 5.6K Rel error | 0.0391 | 0.0380 | - | 0.0216 | 0.0209 Table 7: The relative $L^{2}$ test errors and the number of parameters for the solution operators of PDEs experiments. $\\#$Param denote the number of learnable parameters. Note that the all five models use the similar number of parameters for each problem. ### F.4 Performance of other baselines with the same number of learnable parameters For three different PDEs with complicated target functions, we compare all the baseline methods in Table 7 to evaluate the performances. We analyze the model’s computation efficiency based on the number of parameters and fix the model’s complexity for each equation. All five models demonstrated their prediction abilities for the advection equation. DeepONet shows the greatest performance in this case, and other variants can no longer improve the performance. For the Burgers’ equation, NOMAD and HyperDeepONet are the two outstanding algorithms from the perspective of relative test error. NOMAD seems slightly dominant to our architectures, but the two models compete within the margin of error. Furthermore, HyperDeepONet improves its accuracy using the chunk embedding method, which enlarge the target network’s size while maintaining the complexity. Finally, HyperDeepONet and NOMAD outperform the other models for 2-dimensional shallow water equations. The HyperDeepONet still succeeds in accurate prediction even with a few parameters. It can be observed from Table 7 that NOMAD is slightly more sensitive to an extreme case using a low-complexity model. Because of the limitation in computing 3-dimensional rotation, FlexDeepONet cannot be applied to this problem. Figure 13 shows the results on prediction of shallow water equations’ solution operator using the DeepONet and the HyperDeepONet. The overall performance of the DeepONet is inferior to that of the HyperDeepONet, which is consistent with the result in Figure 12. In particular, the DeepONet has difficulty matching the overall circular shape of the solution when the number of parameters is small. This demonstrates the advantages of the HyperDeepONet when the computational resource is limited. Figure 12: The test $L^{2}$ relative errors of four methods during training for the solution operator of shallow water equations. | | Training time (s) --- (per 1 epoch) Inference time (ms) Same target (Differentiation) | DeepONet | 1.018 | 0.883 HyperDeepONet | 1.097 | 1.389 Same $\\#$param (Advection) | DeepONet | 0.466 | 0.921 HyperDeepONet | 0.500 | 1.912 Table 8: The training time and inference time for the differentiation operator problem and the solution operator of advection equation problem using DeepONet and HyperDeepONet. ### F.5 Comparison of training time and inference time Table 8 shows the training time and the inference time for the DeepONet and the HyperDeepONet for two different operator problems. When the same small target network is employed for the DeepONet and the HyperDeepONet, the training time and inference time for the HyperDeepONet are larger than for the DeepONet. However, in this case, the time is meaningless because DeepONet does not learn the operator with the desired accuracy at all (Table 1 and Figure 6). Even when both models use the same number of training parameters, HyperDeepONet takes slightly longer to train for one epoch than the DeepONet. However, the training complex operator using the HyperDeepONet takes fewer epochs to get the desired accuracy than DeepONet, as seen in Figure 7. This phenomenon can also be observed for the shallow water problem in Figure 12. It shows that the HyperDeepONet converges to the desired accuracy faster than any other variants of DeepONet. The HyperDeepONet also requires a larger inference time because it can infer the target network after the hypernetwork is used to generate the target network’s parameters. However, when the input function’s sensor values are already fixed, the inference time to predict the output of the target function for various query points is faster than that of the DeepONet. This is because the size of the target network for HyperDeepONet is smaller than that of the DeepONet, although the total number of parameters is the same. Figure 13: The examples of predictions on the solution operator of shallow water equations using the DeepONet and the HyperDeepONet. The first column represents the exact solution generated in Takamoto et al. (2022), and the other four columns denote the predicted solutions using the corresponding methods. The four rows shows the predictions $h(t,x_{1},x_{2})$ at four snapshots $t=[0.25,0.5,0.75,1]$.
# Influence of Crystalline Nanoprecipitates on Shear-Band Propagation in Cu- Zr-Based Metallic Glasses Tobias Brink<EMAIL_ADDRESS>Fachgebiet Materialmodellierung, Institut für Materialwissenschaft, Technische Universität Darmstadt, Jovanka- Bontschits-Straße 2, D-64287 Darmstadt, Germany Martin Peterlechner Institut für Materialphysik, Westfälische Wilhelms-Universität Münster, Wilhelm-Klemm- Straße 10, D-48149 Münster, Germany Harald Rösner Institut für Materialphysik, Westfälische Wilhelms-Universität Münster, Wilhelm-Klemm- Straße 10, D-48149 Münster, Germany Karsten Albe Fachgebiet Materialmodellierung, Institut für Materialwissenschaft, Technische Universität Darmstadt, Jovanka-Bontschits-Straße 2, D-64287 Darmstadt, Germany Gerhard Wilde Institut für Materialphysik, Westfälische Wilhelms-Universität Münster, Wilhelm-Klemm-Straße 10, D-48149 Münster, Germany (May 6, 2016) ###### Abstract The interaction of shear bands with crystalline nanoprecipitates in Cu-Zr- based metallic glasses is investigated by a combination of high-resolution TEM imaging and molecular-dynamics computer simulations. Our results reveal different interaction mechanisms: Shear bands can dissolve precipitates, can wrap around crystalline obstacles, or can be blocked depending on the size and density of the precipitates. If the crystalline phase has a low yield strength, we also observe slip transfer through the precipitate. Based on the computational results and experimental findings, a qualitative mechanism map is proposed that categorizes the various processes as a function of the critical stress for dislocation nucleation, precipitate size, and distance. Published in: T. Brink et al., Phys. Rev. Applied 5, 054005 (2016) DOI: 10.1103/PhysRevApplied.5.054005 This article is available under the terms of the Creative Commons Attribution 3.0 License. Further distribution of this work must maintain attribution to the authors and the published article’s title, journal citation, and DOI. Materials Science, Mechanics ## I Introduction Metallic glasses (MGs) have advantageous mechanical properties, such as a high yield strength and a large elastic limit, but suffer from brittle failure, especially under tension, at temperatures significantly below the glass transition.Inoue (2000); Ashby and Greer (2006) Under compression, improved ductility is found for composites of MGs and crystalline secondary phases, namely, for Cu-Zr-based,Fan and Inoue (1997); Lee _et al._ (2006); Hajlaoui _et al._ (2007); Fornell _et al._ (2010); Li _et al._ (2011) Cu-Ti- based,Calin _et al._ (2003) and Zr-Ti-based MGs.Hays _et al._ (2000, 2001); Hofmann _et al._ (2008) Under tension, a small ductility with 1% to 2% strain is observed for Cu-Zr composites containing nanocrystals.Pauly _et al._ (2010a); Barekar _et al._ (2010); Pauly _et al._ (2010b) With a higher volume fraction of the crystalline phase, not only compressive but also significant tensile ductility is reported for Zr-Ti-based,Hofmann _et al._ (2008); Qiao _et al._ (2013) Ti-based,Kim _et al._ (2011) and Cu-Zr-based MGs.Liu _et al._ (2012); Wu _et al._ (2014) For dendritic precipitates, there is a correlation between the location of dendrites and the occurrence of shear-band patterns.Hays _et al._ (2000, 2001) The improved ductility is generally ascribed to the increased number of shear bands and their limited length given by the constraints of the crystalline phase.Hofmann _et al._ (2008) Thus, a high volume fraction of ductile crystalline phase improves the ductility in compression and microindentation tests, while a brittle secondary phase does not.Fu _et al._ (2007) This is confirmed by Song et al., who suggest a crystalline volume fraction between $40\%$ and $80\%$ in Cu-Zr-based MGs for obtaining good mechanical properties,Song _et al._ (2013) which is consistent with the fact that tensile ductility is observed only in glasses with high volume fractions of ductile crystalline phases.Hofmann _et al._ (2008); Kim _et al._ (2011); Liu _et al._ (2012); Wu _et al._ (2014) While the enhancement of macroscopic ductility of MGs with high volume fractions of a ductile crystalline phase can be explained by simple composite models, the influence of nanoprecipitates on the mechanical properties of MGs containing a much lower volume fraction of crystalline matter is still not clear. Similar to the case of dendritic precipitates, shear-band patterns were also observed in glasses containing small spherical crystallites with sizes around $2\,\mathrm{nm}$.Hajlaoui _et al._ (2007) These nanoprecipitates can grow during deformation in certain metallic glasses.Lee _et al._ (2006); Cao _et al._ (2007); Pauly _et al._ (2010a); Barekar _et al._ (2010); Pauly _et al._ (2010b); Fornell _et al._ (2010) An increased growth rate of nanocrystallites in shear bands is observed,Chen _et al._ (2006) which has been related to enhanced atomic mobility inside shear bands.Wilde and Rösner (2011); Bokeloh _et al._ (2011) Deformation-grown nanocrystallites are observed to contain twins,Cao _et al._ (2007); Pauly _et al._ (2010a, b) which occur only in larger crystallites, e.g., with a size greater than $20\,\mathrm{nm}$ in a Cu-Zr-Al MG.Pauly _et al._ (2010b) These deformation- grown precipitates are the possible reason for strain hardening during nanoindentation, Fornell _et al._ (2010) as well as increased plastic strain during compression.Lee _et al._ (2006) It is proposed that the participation of the crystallites in the plastic deformation is the reason for the enhanced ductility: Wu et al. demonstrate that reducing the stacking-fault energy of B2 CuZr by alloying leads to increased twinning and higher ductility under tension.Wu _et al._ (2012) Pauly et al. propose that a martensitic transformation from the B2 phase to the B$19^{\prime}$ phase with a subsequent volume change is responsible for toughening in Cu-Zr-based metallic glasses.Pauly _et al._ (2010a) This interpretation, however, is not generally accepted. Corteen et al., in contrast, note that the volume change of the martensitic transformation is very small and cannot contribute significantly to toughening.Corteen _et al._ (2011) They instead suggest that precipitates increase plasticity by favoring the nucleation of new shear bands over the growth of critical shear bands. Indeed, recent simulation and experimental results provide evidence for the fact that crystal–glass interfaces serve as nucleation sites for shear bands and are therefore responsible for the simultaneous nucleation of multiple shear bands.Albe _et al._ (2013); Zaheri _et al._ (2014); Wang _et al._ (2014a) In tensile tests and corresponding molecular dynamics (MD) simulations of nanolaminates of copper nanocrystals separated by thin Cu-Zr glass layers, the crystal–glass interface acts as a source or sink for dislocations. Shear transformation zones (STZs) are activated by interactions with dislocations.Wang _et al._ (2007); Arman _et al._ (2011); Brandl _et al._ (2013) Figure 1: An undeformed Zr53.8Cu31.6Ag7.0Al7.6 sample after annealing in a TEM bright-field image. An overview of the sample is shown in (a), with clearly visible crystalline precipitates. The inset (b) shows a magnified view, indicating the transition from globular to dendritic morphologies during precipitate growth. The different brightness of parts of one precipitate in (b) is most likely due to the different geometrical orientations of different parts of the same precipitate with respect to the incident electron beam. The strong black-white contrast change observed for some nanocrystals in (a) is most likely due to twinning. Computational studies on the interaction of crystalline precipitates with shear bands provide further insights into nanoscale mechanisms. Lund and Schuh conduct quasi-2D molecular-statics simulations of a binary Lennard-Jones system with a nanocrystal inclusion.Lund and Schuh (2007) They identify three mechanisms of deformation, depending on the ratio of shear-band thickness to crystal size. For small crystals, the deformation is accommodated either in the interface (for example, by rotation) or by dissolution of the crystal. For wide shear bands and intermediate crystal sizes, dislocations in the crystal nucleate at the interface. Finally, for crystals larger than the shear band, they observe homogeneous dislocation nucleation due to stress building in the nanocrystal center. However, it is somewhat unclear how the observed homogeneous dislocation nucleation depends on the artificially induced shear band and the resulting stress state in the system. Shi and Falk conduct molecular-dynamics simulations on a monoatomic amorphous model system with a high fraction of bcc nanocrystallites.Shi and Falk (2008) They find that deformation is induced at the interfaces and that shear bands bend around crystallites away from a direction of maximum resolved shear stress. They also observe blocking of shear bands by crystallites. Because of the high fraction of crystalline phase, the system more closely resembles a nanocrystalline structure. The observation of the initiation of plastic deformation at interfaces still matches the simulations by Albe et al. Albe _et al._ (2013) and underlines the importance of the crystal–glass interface in these composite systems. While it has been shown that interfaces promote shear-band nucleation and that precipitates can act as obstacles or can deform together with the matrix, there is no comprehensive study that investigates the influence of the size and number density of the precipitates. Furthermore, some mechanisms governing the interaction between a propagating shear band and a preexisting precipitate have been observed but not investigated and discussed in detail. Therefore the goal of this study is to investigate the interaction of a shear band with preexisting precipitates in Cu-Zr-based MGs. In the experimental part of the study, we anneal Zr-Cu-Ag-Al melt-spun ribbons to induce the formation of nanocrystalline precipitates. We present transmission electron microscopy (TEM) images of the samples before and after deformation by cold rolling and identify the effects of crystalline precipitates on the shear-band propagation. Using molecular-dynamics computer simulations, we model composite systems with a metallic-glass matrix and crystalline precipitates. We control the initiation of a shear band using a stress concentrator and put precipitates in its propagation path. The focus of the simulations is on the size effects of “hard” precipitates that do not partake in the plastic deformation. Additionally, we study the shear-band interaction with “soft,” plastically deformable precipitates. Finally, we derive a deformation map from the combined observations of simulations and experiments that classifies the observed mechanisms. ## II Experiment ### II.1 Experimental setup We prepare metallic-glass samples of nominal composition Zr53.8Cu31.6Ag7.0Al7.6 from pure components (Cu: 99.999%, Zr: 99.998%, Ag: 99.999%, Al: 99.999%; all in at.%) by prealloying using arc melting. After repeated arc melting with intermittent turning of the specimen to enhance homogenization, the entire ingots are inserted into quartz-glass crucibles for melt spinning. The weight loss during alloying is minimal and subsequent composition analyses by energy-dispersive x-ray diffraction (EDX) confirm that the composition of the material is equal to the nominal composition within the accuracy of the measurement. For melt spinning, the ingots are inductively melted under an Ar atmosphere and the melt is ejected onto a rotating Cu wheel (tangential wheel velocity: $30\,\mathrm{m/s}$), resulting in completely amorphous thin ribbon samples of approximately $80$-$\mathrm{\upmu{}m}$ thickness. X-ray diffraction on the as-quenched ribbon samples does not indicate the presence of any crystalline fraction exceeding the sensitivity threshold of this method. Parts of the ribbon samples are cut to perform differential scanning calorimetry (DSC) measurements in a Perkin Elmer Diamond DSC device. Both isochronal and isothermal measurements under a purified Ar gas flow are conducted and in conjunction with microstructure analyses the time and temperature dependence of the evolving crystalline fraction is determined. On the basis of these results, the samples for deformation processing are annealed in the DSC device at $410\,^{\circ}\mathrm{C}$ for $3\,\mathrm{h}$. This thermal treatment results in the formation of nanocrystalline precipitates with an average diameter of about $70\,\mathrm{nm}$ and a number density on the order of $10^{20}\,\mathrm{m^{-3}}$. The resulting microstructure is shown in Fig. 1. Given that the width of shear bands is on the order of $10\,\mathrm{nm}$,Rösner _et al._ (2014) there is no straight path for propagating shear bands to avoid the interaction with nanoprecipitates in these samples. Figure 2: TEM image of a shear band in a Zr-Cu-Ag-Al ribbon after deformation. The green arrows mark the positions of crystalline precipitates. After the DSC heat treatment, the partially crystalline material is deformed by cold rolling at room temperature in one step to true strain values of about $\varepsilon=5\%$ at room temperature, applying a strain rate of the order of $\dot{\varepsilon}=1\,\mathrm{s^{-1}}$. Subsequently, specimens for TEM investigations are prepared by grinding, dimpling, and finally precision ion polishing (PIPS, Gatan), using a low acceleration voltage of $2.5\,\mathrm{kV}$ and low incidence angles ($<4^{\circ}$) to minimize damage by the preparation process. The electron transparent samples are then analyzed in dark-field, bright-field, and high-resolution transmission electron microscopy modes in a Zeiss Libra 200FE TEM operated at an acceleration voltage of $200\,\mathrm{kV}$. Since intrinsic shear bands formed upon cold rolling could not be found during the TEM inspections of the thin foil regions, the shear bands generated as a result of sample preparation and/or subsequent sample manipulation are studied instead. Thus, the final sample state resembles the state in an in situ TEM experiment comparable to the work in Ref. Wilde and Rösner, 2011, where the shear bands are generated at crack tips in the thin TEM foil. ### II.2 Experimental results Figure 3: TEM and HRTEM images of a shear band. In the upper left corner, the electron transparent hole is visible, which stems from the sample preparation for TEM. (a–b) Bright-field images of the same area, using a different tilt angle. The shear band appears as a white stripe, with part of it already cracked (bright white). (c) Dark-field image of the same area: The nanoprecipitates are visible and marked by circles. The crack stops at precipitate I; the shear band continues from there to precipitate II. (d) The power spectrum of precipitate II along the $\langle\overline{1}10\rangle$ zone axis revealing superlattice reflections of the martensitic B$19^{\prime}$ structure. (e) HRTEM image of precipitate II, showing the shear band stopping at the precipitate. The contrast in (c) and (e) is enhanced to improve the visibility of the precipitates and the shear band. Figure 2 shows a shear band in a sample after TEM preparation. It is noticeable that the shear band switches propagation directions in the vicinity of precipitates. Between precipitates 2 and 3, the shear band has an additional bend. The propagation path, in general, suggests that the shear band is “attracted” to the precipitates, possibly due to a stress field resulting from the density change on crystallization, which explains the change of direction between precipitates 2 and 3. Because of the processing of the samples, it can be excluded that the repeated bending of the shear band results from a change of the external stress state: The deformation by cold rolling is performed in a single step, and the observed shear bands were created during TEM preparation, resembling an in situ experiment. As the path change in the presence of precipitates is rather large (in Fig. 2 around $45^{\circ}$) and correlated to the position of the precipitates, it is most likely induced by the presence of the precipitates. The literature supports this, as shear bands in homogeneous metallic glasses (Cu-Zr-based or otherwise) are straight on the length scale presented here.Kumar _et al._ (2007); Wang _et al._ (2008); Rösner _et al._ (2014); Schmidt _et al._ (2015) It is therefore clear that the precipitates play a major role in influencing shear-band propagation and thereby the macroscopic plastic deformation of the material. TEM images of a second shear band, shown in Fig. 3, shed more light on this interaction. The shear band interacts with two precipitates: A crack follows the path of the shear band up to the first precipitate and the shear band continues from the first to the second precipitate. Because of the visible crack opening near the first precipitate, it can be ruled out that the shear band originates from the second precipitate. While the first precipitate is passed, the second precipitate, which is encountered centrally, stops the shear band. This can also be observed at the end of the shear band in Fig. 2. A detailed analysis of the gray-scale intensity distribution of the high-resolution TEM (HRTEM) image in Fig. 3(e) indicates that the shear band changes its path slightly near the crystalline precipitate but does not proceed further or shows slip transfer into the precipitate. Additionally, a smaller, shear-band-like region emerges almost perpendicular to the previous propagation direction (yellow arrow). This indicates either a shear-band deflection, or a nucleation of a new, perpendicular shear band. Based on the highly local nature of the intensity distribution and based on the comparison of the contrast of other precipitates, preparation artifacts can be excluded. Thus, this type of interaction observed here is part of the intrinsic interaction mechanism between precipitates and advancing shear bands. The power spectrum in Fig. 3(d) indicates a precipitate with B$19^{\prime}$ crystal structure,Schryvers _et al._ (1997) which is consistent with prior observations in Cu-Zr-based MGs.Pauly _et al._ (2010a) It is clear that the interaction of a propagating shear band with a distribution of crystalline precipitates depends on the actual stress state near the shear-band tip, the already-accommodated stress by the shear-band propagation and the details of the local distribution of crystalline precipitates, their sizes, and the residual stresses in the glass matrix due to the formation of the precipitates. Depending on these factors, there are various explanations for the observed paths: The path changes are caused either by a deflection of the shear band, by the blocking and subsequent renucleation of a new shear band, or by several nascent shear bands growing together. The unhindered “passing” of a precipitate can be explained either by a temporary path change of the shear band or by the participation of the precipitate in the plastic deformation. Because of the necessarily limited amount of data that can be obtained in the experiment, we undertake MD simulations to test the aforementioned hypotheses and investigate their relation to the sample geometry. ## III Simulation ### III.1 Simulation setup and analysis Figure 4: Schematic simulation setup. The picture shows a cut through the three-dimensional simulation box in the $xz$ plane at $l/2$. A notch is inserted to control the origin of the shear band (yellow). The spherical precipitate is shown in blue. The box has open boundaries in the $x$ direction and is otherwise periodic. A constant strain rate is applied in the $z$ direction. Figure 5: Snapshots of a simulation with a $30$-$\mathrm{nm}$ CuZr precipitate. The shear band wraps around the precipitate and continues unhindered. The glass matrix is colored according to the atomic strain. The precipitate atoms are shown in blue if they appear in the B2 structure; no defects are visible. The left column shows a cut through the middle of the precipitate. On the right, all atoms with $\eta_{i}<0.3$ are deleted. A video version is provided in Video 1. List of myvideos 1 Simulation of a shear band wrapping around a $30$-$\mathrm{nm}$ CuZr precipitate as shown in Fig. 5. In order to gain more insights into the nanoscale mechanisms of shear-band interaction with precipitates, we perform a number of MD computer simulations, which allow for an “in situ” observation of shear-band propagation. We use the software lammps Plimpton (1995) to quench metallic-glass samples, insert precipitates, and perform mechanical testing. The simulated metallic glass is a $\mathrm{Cu_{64}Zr_{36}}$ alloy modeled with a Finnis–Sinclair-type potential by Mendelev et al.Mendelev _et al._ (2009) Metallic-glass samples with dimensions $10\times 10\times 10\,\mathrm{nm}^{3}$ and $20\times 20\times 20\,\mathrm{nm}^{3}$ are prepared by melting the material at $2000\,\mathrm{K}$ and subsequent quenching to $50\,\mathrm{K}$ with a cooling rate of $0.01\,\mathrm{K/ps}$. We use the sample geometry illustrated in Fig. 4 to investigate the influence of preexisting precipitates on an approaching shear band. Size effects are studied by varying the diameter $d$ of the precipitates and by controlling the distance between periodic precipitate images by varying the box width $l$. Sample sizes are $120\,\mathrm{nm}\times l\times 60\,\mathrm{nm}$, with $l=10\,\mathrm{nm}$, $20\,\mathrm{nm}$, $30\,\mathrm{nm}$, and $40\,\mathrm{nm}$. With a single precipitate per simulation box, this corresponds to number densities for the precipitates of $1.4\times 10^{22}/\mathrm{m^{3}}$, $6.9\times 10^{21}/\mathrm{m^{3}}$, $4.6\times 10^{21}/\mathrm{m^{3}}$, and $3.5\times 10^{21}/\mathrm{m^{3}}$, respectively. The Cu-Zr glass samples are replicated to reach the desired box dimensions, and spherical precipitates with diameters from $3\,\mathrm{nm}$ to $40\,\mathrm{nm}$ were inserted. For this, a hole is cut into the glass matrix with a size chosen to accommodate the precipitate without overlapping atoms. For the CuZr precipitate, we use the experimentally observed B2 structure.Pauly _et al._ (2007); Sun _et al._ (2005); Das _et al._ (2007); Jiang _et al._ (2007) The B$19^{\prime}$ structure, which was found in the precipitates in the experimental part of this paper, is a distortion of the B2 structure.Schryvers _et al._ (1997) A notch controls the origin of the shear band and makes sure that it always hits the precipitate. Periodic boundary conditions are applied in the $y$ and $z$ directions and open boundaries in the $x$ direction. The resulting structure is equilibrated at $50\,\mathrm{K}$ for $2\,\mathrm{ns}$ with a barostat at ambient pressure in periodic directions. After equilibration, no long ranging stress field around the precipitate is left; any mismatches are accommodated by the glass during the interface creation. The resulting composite samples are deformed at $50\,\mathrm{K}$ under a constant tensile strain rate of $\dot{\varepsilon}=4\times 10^{7}/\mathrm{s}$ in the $z$ direction up to a total strain of at least $10\%$. The trajectories from equilibrated to fully deformed samples are analyzed to observe the shear- band propagation path. The shear band is identified using the von Mises local shear invariant $\eta_{i}$ Shimizu _et al._ (2007) as implemented in the visualization tool Ovito.Stukowski (2010) Atoms with a local shear greater than around $0.2$ are assigned to the shear band. To observe plastic deformation events in the crystalline phase, we perform atomic structure identification, which can identify crystal structures, stacking faults, and other defects.Stukowski (2012) ### III.2 Wrapping and blocking Figure 6: Comparison of the Orowan mechanism (a) with a shear band wrapping around precipitates (b). The slip plane is shown in gray, the precipitates in blue, the dislocation as a black line, and the shear band as a red plane. The precipitates can stop a dislocation because it must remain in a defined slip plane, while the shear band can temporarily leave its slip plane and continue unhindered afterwards. The first observed interaction mechanism between a propagating shear band and a preexisting precipitate is shown in Fig. 5 and Video 1. Here, the shear band wraps around the precipitate like a carpet moving over a small obstacle. The particle does not deform and simply moves along with one half of the glass matrix. We call this mechanism the _wrapping mechanism_. This kind of athermal mechanism is virtually unknown in crystalline materials. The closest analog in a crystal—dislocation climb—is purely thermally activated. It is therefore instructive to compare the two material classes, as shown in Fig. 6. In a crystalline material, a dislocation moves on defined slip planes. A change of slip plane is possible only for the screw components of the dislocation,Cottrell (1953) is connected with a high energy barrier, and is usually observed only in stage III work hardening.Hirth and Lothe (1982) Therefore, the Orowan mechanism applies: The dislocation is bent around the obstacle and finally forms dislocation rings [Fig. 6(a)].Gottstein (2004) These rings may pile up, thereby hardening the material. In a metallic glass, as in any isotropic material, all slip directions are equivalent. Only an applied external stress differentiates the directions. Under tensile stress, the planes oriented in $45^{\circ}$ angles towards the tensile axis experience the highest resolved stress. An obstacle can be avoided simply by temporarily and locally changing the slip path, thereby wrapping around the obstacle [Fig. 6(b)]. Depending on the precipitate distance, this wrapping mechanism is observed for precipitates with diameters smaller than $25\,\mathrm{nm}$ to $35\,\mathrm{nm}$ in our simulations. Figure 7: Snapshots of a simulation with a $37.5$-$\mathrm{nm}$ CuZr- precipitate. The shear band is blocked by the precipitate, while a second shear band is immediately nucleated in another plane of high resolved shear stress. The glass matrix is colored according to the atomic strain. The precipitate atoms are shown in blue if they appear in the B2 structure; no defects are visible. The left column shows a cut through the middle of the precipitate. On the right, all atoms with $\eta_{i}<0.3$ are deleted. A video version is provided in Video 2. List of myvideos 2 Simulation of a shear band being blocked by a $37.5\,\mathrm{nm}$ CuZr precipitate as shown in Fig. 7. An alternative mechanism appears for increasing diameters and decreasing distances between precipitates: The shear band is _blocked_ by the precipitate. This causes the simultaneous nucleation of a second shear band perpendicular to the first one on the opposite side of the precipitate as shown in Fig. 7 and in Video 2. Again, there is no slip transfer to the crystalline phase. The paths of the shear bands in this simulation and in the HRTEM image of the experimental sample [Fig. 3(e)] are comparable. Both show what looks like a shear band that starts to wrap around the precipitate but does not propagate further. In the experiment, though, no fully formed shear band appears perpendicular to the original one. This may be a result of the more complex stress state or the fact that a new shear band can nucleate at another precipitate that is not visible in the images. Still, a region resembling a nascent shear band appears perpendicular to the original propagation direction, strengthening the agreement with the simulation results. A systematic investigation of the parameters of precipitate distance and size reveal a clear correlation with the mechanism. For a more quantitative analysis, we define an empirical parameter $\Lambda$, which is given by the ratio of the cross-sectional area $A$ of the precipitate divided by the distance of precipitate centers $l$ as shown in Fig. 8(a). Similar to the derivation of the Orowan stress, the crystalline volume fraction $f=V_{\mathrm{precipitates}}/V$ can be estimated from the average crystallite distance by the relationGottstein (2004) $l=\frac{d/2}{\sqrt{f}}\quad\Leftrightarrow\quad f=\frac{d^{2}}{4l^{2}}.$ (1) Using that, we can express $\Lambda$ only in terms of volume fraction and precipitate geometry: $\Lambda=\frac{A}{l}=\frac{A\sqrt{f}}{d/2}.$ (2) For nonoverlapping precipitates ($d<l$), it is $A=\pi d^{2}/4$ and $\Lambda$ reduces to $\Lambda=\pi\frac{d}{2}\sqrt{f}.$ (3) When the precipitates overlap ($d>l$), we reduce the area $A$ to remove the overlapping circle segments. In the limit $l\rightarrow 0$, $\Lambda$ corresponds to $A/l$ of an infinite cylinder parallel to the shear-band front. Figure 8(b) shows a contour plot of the $\Lambda$ parameter as a function of the precipitate diameter and distance. The data points in the plot represent results from our MD simulations, divided into those showing the wrapping and those showing the blocking mechanism. For the simulation geometry used in this work, there exists a given $\Lambda$ that clearly separates the two mechanisms: $\Lambda_{\mathrm{crit}}\approx 12.65\,\mathrm{nm}$ (4) This $\Lambda_{\mathrm{crit}}$ is not universal, as, for example, the distance between precipitate and notch is not varied. A test simulation finds that increasing the distance from the notch and therefore increasing shear-band length favors the wrapping mechanism. Figure 8: Influence of the precipitate size and distance on the mechanism. (a) Explanation of the parameter $l$, distance of the precipitates, and $A$, the area of a cut through the precipitate. (b) shows a contour plot of the empirical geometry factor $\Lambda=A/l$. The dashed line shows the critical value of $\Lambda$ for a transition from the wrapping to the blocking mechanism. Data points are MD simulations. The data points at $l=0\,\mathrm{nm}$ are infinite cylinders along the $y$ axis, representing the case of overlapping spheres with infinitesimally small distances between their centers. Figure 9: Stress-strain curves of samples which exhibit the wrapping or blocking mechanisms. The arrows indicate when the shear band hits the precipitate. Figure 10: Mechanical dissolution of a $3$-$\mathrm{nm}$ copper particle in a shear band. The gray-scale color coding shows atomic strain $\eta_{i}$ from $0.0$ (black) to $1.0$ (white). The precipitate is shown in color: Light blue atoms are in the fcc structure, green atoms are in a stacking fault, and dark blue atoms are disordered. The arrows indicate the shear direction. Subfigure (a) shows a cut through the middle of the nanoprecipitate, while (b) shows only the atoms that initially belonged to the nanoprecipitate. Using these formulas, we can also estimate $\Lambda$ for the experimental results. Given a number density of precipitates $n=10^{20}\,\mathrm{m^{-3}}$ and particle diameters of around $70\,\mathrm{nm}$, we obtain $\displaystyle f$ $\displaystyle=\frac{V_{\mathrm{precipitates}}}{V}=n\frac{4}{3}\pi(35\,\mathrm{nm})^{3}\approx 1.8\%$ (5) $\displaystyle\Lambda_{\mathrm{exp}}$ $\displaystyle=\pi\times 35\,\mathrm{nm}\times\sqrt{1.8\%}\approx 14.7\,\mathrm{nm}>\Lambda_{\mathrm{crit}}.$ (6) This is consistent with the fact that the shear bands that are observed in the glass samples are blocked by the precipitates. Still, the value is close to $\Lambda_{\mathrm{crit}}$, which means that slightly smaller precipitates (or precipitates that are not hit centrally, virtually decreasing the prefactor $d$) may be susceptible to wrapping. Figure 11: Stress-strain curves for the composites with fcc copper precipitates and for a $10$-$\mathrm{nm}$ copper nanowire. (a) Curves from simulations with the Mendelev potential. The copper precipitates do not deform plastically and show similar behavior to the CuZr precipitates discussed earlier. The curve of the copper nanowire explains this: The yield stress, and therefore the critical shear stress, is higher than the highest resolved shear stress in the steady state in the composite. To be able to compare the tensile stress in the composite with the shear stress in the nanowire, the tensile stress axis is scaled with a factor $0.5$ corresponding to the Schmid factor for the plane of highest resolved shear stress. (b) Curves from simulations with the Ward potential. Now the critical stress for heterogeneous dislocation nucleation is lower than the steady-state stress in the composite and the precipitate deforms plastically. (c) The simulation setup for the nanowire. The wire is sheared in the $[110]$ direction of the fcc crystal structure on the $(111)$ plane. The red atoms are fixed and the atoms on the top are shifted with a constant velocity to shear the nanowire. Figure 9 shows examples of the stress-strain curves of samples that exhibit either the wrapping or the blocking mechanism. Neither a pronounced ductility nor significant strain hardening can be observed. This lack of strain hardening is in accordance with experimental data for tensile tests on Cu-Zr- based metallic glasses with crystalline precipitates.Pauly _et al._ (2010a); Barekar _et al._ (2010); Pauly _et al._ (2010b) It also fits with a recent study of compression tests of Cu-Zr-based metallic glass, that finds an effect of particle size on mechanical parameters but no particle hardening.Wang _et al._ (2014b) ### III.3 Plastic deformation of the crystalline phase Figure 12: Snapshots of a simulation with a $30$-$\mathrm{nm}$ copper precipitate. This simulation uses the Ward potential, in which the critical stress for dislocation nucleation is realistic. As the precipitate is sufficiently soft, the shear band can cut through it. The glass matrix is colored according to the atomic strain (the same scale as Figs. 5 and 7). Defects in the fcc crystal structure are colored according to the legend. The left column shows a cut through the middle of the precipitate. On the right, all atoms with $\eta_{i}<0.3$ and all fcc-coordinated atoms are deleted. A video version is provided in Video 3. List of myvideos 3 Simulation of a shear band cutting a $30$-$\mathrm{nm}$ copper precipitate as shown in Fig. 12. The precipitates discussed until now were all “hard,” i.e., not susceptible to plastic deformation under the simulation conditions: Because of the high antiphase-boundary energy in the B2 structure, superdislocations or twinning with respect to a martensitic transformation would be needed for a plastic deformation of the precipitates. The stress available at the shear-band tip is not sufficient to nucleate these defects, which explains why no plastic deformation of the crystalline phase can be observed in our simulations. In experiments, deformation-grown precipitates show twinning defects which may cause softening effects and make the precipitate susceptible to plastic deformation.Cao _et al._ (2007) As a model for a softer precipitate, we thus exchange the B2 crystal phase for fcc copper. For precipitates that are small relative to the shear-band width, the nanocrystals undergo mechanical dissolution.Lund and Schuh (2007); Albe _et al._ (2013) This is also observed in our setup with $3$-$\mathrm{nm}$ particles as shown in Fig. 10. In samples with larger diameters, the precipitates do not deform plastically but instead show the same wrapping and blocking interactions as described before for the “hard” precipitates. Even if the $(111)$ glide plane is oriented parallel to the shear-band direction to maximize resolved shear stress on the preferred fcc slip plane, we do not observe slip transfer into the nanoprecipitate. The corresponding stress- strain curves are shown in Fig. 11(a). To explain this, we estimate the critical stress for heterogeneous dislocation nucleation in fcc copper by shearing a nanowire on the $(111)$ plane in the $[110]$ direction [see Fig. 11(c) for the simulation setup]. For this, we hold the lower layer of atoms fixed and move the top layer of atoms with a constant velocity in the $[110]$ direction to achieve volume-conserving shear. The diameter of the nanowire is $10\,\mathrm{nm}$, i.e., on the order of the smaller precipitates to maximize surface effects. The resulting shear stress over shear curve is also plotted in Fig. 11(a). The yield stress of the nanowire, $\tau_{\mathrm{crit}}$, is an estimate for the stress needed for heterogeneous dislocation nucleation at the glass-precipitate interface. For comparing the shear stress in the nanowire with the tensile stress in the composite, we assume a Schmid factor of $0.5$ and scale the tensile axis by a factor $0.5$ compared to the shear stress axis. This graphically estimates an upper bound of the shear stress $\tau_{m}$ that arrives at the precipitate, corresponding values are given in Table 1. The upper bound of the estimated maximum resolved shear stress in the composite is comparable to the lower bound for heterogeneous dislocation nucleation. While this suggests that dislocation nucleation may be possible, it is important to keep in mind that the nanowire shear test provides only a lower bound in an idealized case and that the actual shear stress available to nucleate a dislocation may be lower than $0.5\sigma_{z}$, which is why no plastic deformation of the precipitate is observed. Evidently, the value of $\tau_{\mathrm{crit}}$ in the Mendelev potential is much too high, and therefore the Cu precipitates are “harder” than expected, which is a deficiency of the potential model for pure Cu. Table 1: Comparison of maximum stress $\sigma_{m}$ and the corresponding resolved shear stress $\tau_{m}$ with the critical shear stress $\tau_{\text{crit}}$ for heterogeneous dislocation nucleation in copper in the Mendelev and Ward potentials. The calculation of the resolved shear stress assumes a Schmid factor of $0.5$. | $\sigma_{m}$ (GPa) | $\tau_{m}$ (GPa) | $\tau_{\text{crit}}$ (GPa) ---|---|---|--- Mendelev | $3.2$ | $1.6$ | $1.55$ Ward | $2.9$ | $1.45$ | $0.75$ Because of this, we switch to a different potential, which provides a better description of crystalline Cu. The Finnis–Sinclair-type potential by Ward et al. Ward _et al._ (2012) is created by using preexisting potentials for the elemental phases Zhou _et al._ (2004) and fitting the cross terms to the intermetallic phases. As shown in the Appendix, this potential has a more realistic unstable stacking-fault energy and critical stress for homogeneous dislocation nucleation than the Mendelev potential. As shown in Fig. 11(b) and Table 1, the critical stress for heterogeneous nucleation is much lower than even the steady-state stress in the composite, easily allowing plastic deformation of the particle. The results for a $30$-$\mathrm{nm}$ copper precipitate are shown in Fig. 12 and in Video 3. The precipitate is _cut_ by the shear band, and slip transfer through the particle can be observed. This mechanism replaces the previously discussed blocking of the shear band if the nanoprecipitates are “soft”: For the crystal to partake in the plastic deformation, dislocation nucleation must be possible at shear stresses below the highest resolved shear stress in the metallic glass at yield. Despite the participation of the crystalline phase in the plastic deformation, the stress-strain curve in Fig. 11(b) shows the distinctive stress drop connected with a single critical shear band and no strain hardening. The reason is that in this setup the crystalline phase accounts only for roughly $5\,\mathrm{vol\%}$ of the sample. This means that the macroscopic mechanical properties are still dominated by the metallic glass. As the shear band can simply cut through the crystal, the precipitate poses no obstacle to the percolation of the critical shear band. For a larger crystalline volume, a ductile crystalline phase could possibly also constrain the shear bands.Hofmann _et al._ (2008) ## IV Discussion Using TEM imaging, we observe shear-band bending around or close to precipitates, an attraction of shear bands to the precipitates, and shear bands being blocked by precipitates. In our MD simulations, we find four mechanisms of interaction between shear bands and precipitates: 1. (i) precipitates that are small relative to the shear-band width dissolve mechanically, 2. (ii) shear bands can wrap around precipitates, 3. (iii) shear bands are blocked by precipitates, and 4. (iv) shear bands cut through precipitates, and slip transfer into the crystalline phase takes place. Which of these mechanisms is active for a given precipitate depends on the competition between the propagation of the existing shear band, the heterogeneous nucleation of a new shear band, and the heterogeneous dislocation nucleation in the precipitate. The wrapping-to-blocking transition can be quantified by the parameter $\Lambda=A/l\propto A\sqrt{f}/d$. Below $\Lambda=\Lambda_{\mathrm{crit}}$, the wrapping mechanism is favored. This value can be be explained by the following simple argument. When the shear band reaches a precipitate which does not deform, the shearing of the sample momentarily stops. The stress $\tau_{\mathrm{SB}}$ in the shear band resulting from the externally applied tensile stress $\sigma_{\mathrm{ext}}$ amounts to $\tau_{\mathrm{SB}}=\frac{1}{2}\sigma_{\mathrm{ext}}.$ (7) This stress acts mainly on the shear-band front, allowing us to write $F_{\mathrm{ext}}\approx\tau_{\mathrm{SB}}\times l\times h_{\mathrm{SB}}=\frac{\sigma_{\mathrm{ext}}}{2}\,l\,h_{\mathrm{SB}},$ (8) where $h_{\mathrm{SB}}$ is the width of the shear band. At the moment that the shear band hits the precipitate, the force $F_{\mathrm{ext}}$ must be equal to a reaction force $F_{\mathrm{back}}$ from the precipitate (actio est reactio). Using the projected precipitate area $A$ (cf. Fig. 13), we can convert that force into a normal stress: $\sigma_{A}^{n}=\frac{F_{\mathrm{back}}}{A/2}.$ (9) Figure 13: Projection of the forces and stresses acting around the precipitate (blue circle) onto the $xz$ plane. Shear bands are shown as hatched areas, where yellow signifies the arriving shear band, green the path for wrapping, and red the site for the nucleation of a new shear band. Figure 14: Schematic view of different mechanisms for the interaction of a shear band with a crystalline precipitate. With increasing precipitate sizes, the dissolution of the precipitate is first replaced by the wrapping mechanism. Depending on the critical stress for dislocation nucleation, wrapping is replaced by blocking or cutting. Wrapping can be favored by increasing the precipitate distance. We assume that $F_{\mathrm{ext}}$ predominantly acts on one half of the obstacle (area $A/2$), which is supported by the deformation pattern of the plastically deformed particle (Fig. 12 and Video 12). This back stress results in a shear stress $\tau_{\mathrm{wrap}}\approx 0.5\sigma_{A}^{n}$ in the plane of the wrapping shear band (green shear band in Fig. 13). With $F_{\mathrm{back}}=F_{\mathrm{ext}}$, it is $\sigma_{A}^{n}=\frac{F_{\mathrm{ext}}}{A/2}=\frac{\sigma_{\mathrm{ext}}\,l\,h_{\mathrm{SB}}}{A}=2\tau_{\mathrm{wrap}}.$ (10) $\tau_{\mathrm{wrap}}$ must surpass a critical value $\tilde{\tau}_{\mathrm{wrap}}$ to allow the initiation of the wrapping mechanism; otherwise the shear band simply stops propagating. This is not observed in our simulations, suggesting that $\sigma_{\mathrm{ext}}$ at yield is greater than $\tilde{\sigma}_{\mathrm{ext}}=\frac{2A\tilde{\tau}_{\mathrm{wrap}}}{l\,h_{\mathrm{SB}}}.$ (11) The competing mechanism, blocking the shear band and nucleating a new one (red shear band in Fig. 13), can simply be expressed by a critical shear stress $\tilde{\tau}_{\mathrm{nucl}}$. Because of the low temperature in the simulation and a stress close to the yield stress, we consider only the athermal case and do not invoke a nucleation term which takes into account the relative volume of the interface. With $\tau_{\mathrm{nucl}}=0.5\sigma_{\mathrm{ext}}$, the transition from wrapping to blocking takes place where $\displaystyle 2\tilde{\tau}_{\mathrm{nucl}}=\tilde{\sigma}_{\mathrm{ext}}=\frac{2A\tilde{\tau}_{\mathrm{wrap}}}{l\,h_{\mathrm{SB}}},\quad\text{giving}$ (12) $\displaystyle\frac{A}{l}=\Lambda_{\mathrm{crit}}=\frac{\tilde{\tau}_{\mathrm{nucl}}}{\tilde{\tau}_{\mathrm{wrap}}}h_{\mathrm{SB}}.$ (13) This derivation also works in the case of externally applied shear stress, by replacing $\sigma_{\mathrm{ext}}$ with $2\tau_{\mathrm{ext}}$. Assuming that nucleating a new shear band at the interface and propagating the wrapping shear band along the interface have similar critical stresses, we can simplify Eq. 13 to $\Lambda_{\mathrm{crit}}\approx h_{\mathrm{SB}}.$ (14) Shear bands in Cu-Zr-based glasses have widths of around $10\,\mathrm{nm}$,Ritter and Albe (2011) which fits to the $\Lambda_{\mathrm{crit}}=12.65\,\mathrm{nm}$ observed in our model systems. As stated earlier, the wrapping mechanism becomes more favorable again if the shear band is longer before it hits the precipitate. The reason is that this gives the shear band time to deviate slightly from its path, so that it does not hit the precipitate centrally, thus effectively reducing $A$. In practice, this is not a big problem, as the free shear-band length is constrained to approximately $l$ anyway due to the distribution of precipitates in the sample. While this derivation is only approximate, it seems to be sufficient to explain the observed phenomena and guide future efforts in tuning the mechanical behavior of crystal–glass composites. Furthermore, it is easily possible to explain the fourth mechanism, a slip transfer into the crystalline phase. The critical stress for heterogeneous nucleation of a dislocation in the precipitate $\tilde{\tau}_{\mathrm{disl}}$ must be provided by the shear band via $\tau_{\mathrm{SB}}=0.5\sigma_{\mathrm{ext}}$. If $\tilde{\tau}_{\mathrm{disl}}<\tilde{\tau}_{\mathrm{nucl}}$, we can simply replace $\tilde{\tau}_{\mathrm{nucl}}$ by $\tilde{\tau}_{\mathrm{disl}}$ in Eqs. 12 and 13, thereby replacing the blocking mechanism with the plastic deformation of the precipitate. A lowered $\tilde{\tau}_{\mathrm{disl}}$ also lowers $\Lambda_{\mathrm{crit}}$. A simple deflection of the shear band is not observed and seems unlikely, as any deviation of the shear band from its path leads to a reduced resolved shear stress and thereby to a driving force to put it back “on track.” Contrary to the experiment, a change of shear-band path towards the precipitate was also not observed. Because the precipitates in our simulations are inserted artificially and do not have a large stress field around them, this seems reasonable. In thermally grown precipitates, a stress field due to density mismatch between glass and crystal seems likely. Still, due to the geometry of the MD simulations, the shear band has two equivalent propagation pathways from the notch but always chooses the one leading towards the precipitate. This seems to be a weaker form of the attraction observed experimentally. With these results, we can attempt an explanation of the experimentally observed phenomena. First of all, the blocking of the shear band is a one-to- one correspondence between simulation and experiment. Comparing Fig. 3(e) with Fig. 7, we can see that the path of the shear band looks identical. The shear band wraps partly around the precipitate but is then blocked and does not propagate. Contrary to the simulation, no fully formed shear band but only a small shear-band-like region appears at the opposite crystal–glass interface. This may be a result either of the more complex stress state in the experiment or the fact that other precipitates are available at which the new shear band may nucleate. For the winding shear-band path, we can now exclude a simple deflection as discussed above. A possible explanation would be the concurrent nucleation of nascent shear bands at the crystal–glass interfaces which grow together into a single mature shear band. While the interfaces are known to be nucleation sources for shear bands,Albe _et al._ (2013); Zaheri _et al._ (2014); Wang _et al._ (2014a) our simulations show that the nucleation of a shear band at a stress concentrator like a notch or a crack always takes precedence to nucleation at interfaces or surfaces. The shear bands shown in the TEM images all originate from crack tips, making it unlikely that the shear band shown in Fig. 2 consists of several concurrently nucleated shear bands. This leaves the explanation that this winding path is a series of subsequent blocking and renucleation events. The observed $\Lambda_{\mathrm{crit}}$ corresponds to precipitate diameters somewhere between $20\,\mathrm{nm}$ and $40\,\mathrm{nm}$, depending on the interparticle distance. This critical diameter is on the order of magnitude reported in several experimental studies for twinning in B2 crystals in Cu-Zr- Al-based metallic glasses of $20\pm 5\,\mathrm{nm}$.Pauly _et al._ (2010b); Kuo _et al._ (2014) It also fits an experimental work on Al-based glasses, where crystallites growing during deformation are sheared apart when they reach a critical size of about $10\,\mathrm{nm}$.Hebert _et al._ (2006) Cu50Zr45Ti5 metallic glasses exhibit a critical size of about $9\,\mathrm{nm}$ for twinning of B2 precipitates.Wang _et al._ (2014b) The interparticle distances in these experiments are on the same order of magnitude as for our simulations. This supports an explanation of a transition from wrapping to slip transfer in these systems. Figure 14 summarizes the competition between the different mechanisms. Mechanical dissolution of the crystalline particles occurs only if their size is comparable to the shear-band size.Lund and Schuh (2007) With further increasing precipitate sizes, the shear band can still wrap around the obstacle until the size reaches a threshold value. This critical size also depends on the precipitate distance, as discussed before, expressed in the parameter $\Lambda_{\mathrm{crit}}(d,l)$. If wrapping is no longer possible, a precipitate which reacts only elastically to the applied stress will block the shear band. If the precipitate is susceptible to plastic deformation, slip transfer into the precipitate will take place. Concerning the mechanical performance of such in situ composites with crystalline precipitates that originate from nucleation and growth within the glass, the current results suggest that the discussed geometrical effects serve to improve the macroscopic mechanical performance. None of the presented mechanisms seem to inhibit the percolation of a critical shear band, yet, catastrophic slip along a shear band leading to complete failure is delayed in the case of winding shear bands due to the increase of the shear-band path length as well as the raised activation barriers for slip along shear bands that have a more complex topology. The wrapping mechanism does not pose an obstacle to shear-band propagation, but can be avoided by appropriate adjustment of the crystalline volume fraction and precipitate diameter. “Soft” precipitates additionally open possibilities to adjust the plastic deformation by participating in it. Consistently, by increasing the volume fraction of the ductile crystalline phase, the constraints on shear-band propagation can be increased, immediate failure can be prevented (cf. Ref. Hofmann _et al._ , 2008), and the composite displays macroscopic mechanical behavior according to a mixing rule, further allowing one to tailor the properties. ## V Conclusions Using TEM imaging, we observe shear-band bending around or close to precipitates, an attraction of shear bands to the precipitates, and shear bands being blocked by precipitates. MD simulations reveal that the shear-band bending is most likely the result of the subsequent blocking and renucleation of shear bands. Moreover, we identify shear bands wrapping around precipitates and slip transfer into the crystalline phase. By describing the competition between the critical stress for wrapping, the nucleation of a new shear band, and the nucleation of dislocations in the crystal, we could derive a mechanism map for metallic glasses with nanocrystalline precipitates. This detailed description of shear-band propagation not only helps to understand the mechanical failure of these composites but also aids in tuning them. Figure 15: Generalized stacking-fault energies (top) and the resulting shear stresses (bottom) for fcc copper. The DFT values for the stress curve are from Ogata et al.,Ogata _et al._ (2002) and the corresponding stacking-fault energies are approximated by using a numerical integration of the stress data. ###### Acknowledgements. The authors thank J. Bünz and M. Gerlitz for help with sample preparation and DSC characterization. Financial support by the Deutsche Forschungsgemeinschaft (DFG) through project grants nos. AL 578/13-1, AL 578/6-2, and WI 1899/12-1 is gratefully acknowledged. Computing time was granted by the John von Neumann Institute for Computing (NIC) and provided on the supercomputer JUROPA at Jülich Supercomputing Centre (JSC), as well as by the Gauss Centre for Supercomputing (GCS) through the NIC on the GCS share of the supercomputer JUQUEEN at JSC. GCS is the alliance of the three national supercomputing centers HLRS (Universität Stuttgart), JSC (Forschungszentrum Jülich), and LRZ (Bayerische Akademie der Wissenschaften), funded by the German Federal Ministry of Education and Research and the German State Ministries for Research of Baden-Württemberg, Bayern, and Nordrhein-Westfalen. Additional computing time was made available by the Technische Universität Darmstadt on the Lichtenberg cluster. * ## Appendix A Generalized stacking-fault energy in fcc copper The generalized stacking-fault energy in fcc copper is calculated by using both Mendelev Mendelev _et al._ (2009) and Ward Zhou _et al._ (2004); Ward _et al._ (2012) potentials. Density-functional theory (DFT) calculations from Ogata et al. Ogata _et al._ (2002) using the same method are used for comparison. This is plotted in Fig. 15. The Ward potential much more accurately describes the stacking-fault energy than the Mendelev potential, which has a critical stress for homogeneous dislocation nucleation which is more than two times too high. ## References * Inoue (2000) A. Inoue, Acta Mater. 48, 279 (2000). * Ashby and Greer (2006) M. F. Ashby and A. L. Greer, Scr. Mater. 54, 321 (2006). * Fan and Inoue (1997) C. Fan and A. Inoue, Mater. Trans. JIM 38, 1040 (1997). * Lee _et al._ (2006) S.-W. Lee, M.-Y. Huh, E. Fleury, and J.-C. Lee, Acta Mater. 54, 349 (2006). * Hajlaoui _et al._ (2007) K. Hajlaoui, A. R. Yavari, A. LeMoulec, W. J. Botta, F. G. Vaughan, J. Das, A. L. Greer, and A. Kvick, J. Non-Cryst. Solids 353, 327 (2007). * Fornell _et al._ (2010) J. Fornell, E. Rossinyol, S. Suriñach, M. D. Baró, W. H. Li, and J. Sort, Scr. Mater. 62, 13 (2010). * Li _et al._ (2011) J. B. Li, J. S. C. Jang, S. R. Jian, K. W. Chen, J. F. Lin, and J. C. Huang, Mater. Sci. Eng. A 528, 8244 (2011). * Calin _et al._ (2003) M. Calin, J. Eckert, and L. Schultz, Scr. Mater. 48, 653 (2003). * Hays _et al._ (2000) C. C. Hays, C. P. Kim, and W. L. Johnson, Phys. Rev. Lett. 84, 2901 (2000). * Hays _et al._ (2001) C. C. Hays, C. P. Kim, and W. L. Johnson, Mater. Sci. Eng. A 304–306, 650 (2001). * Hofmann _et al._ (2008) D. C. Hofmann, J.-Y. Suh, A. Wiest, G. Duan, M.-L. Lind, M. D. Demetriou, and W. L. Johnson, Nature 451, 1085 (2008). * Pauly _et al._ (2010a) S. Pauly, S. Gorantla, G. Wang, U. Kühn, and J. Eckert, Nat. Mater. 9, 473 (2010a). * Barekar _et al._ (2010) N. S. Barekar, S. Pauly, R. B. Kumar, U. Kühn, B. K. Dhindaw, and J. Eckert, Mater. Sci. Eng. A 527, 5867 (2010). * Pauly _et al._ (2010b) S. Pauly, G. Liu, S. Gorantla, G. Wang, U. Kühn, D. H. Kim, and J. Eckert, Acta Mater. 58, 4883 (2010b). * Qiao _et al._ (2013) J. W. Qiao, T. Zhang, F. Q. Yang, P. K. Liaw, S. Pauly, and B. S. Xu, Sci. Rep. 3, 2816 (2013). * Kim _et al._ (2011) C. P. Kim, Y. S. Oh, S. Lee, and N. J. Kim, Scr. Mater. 65, 304 (2011). * Liu _et al._ (2012) Z. Liu, R. Li, G. Liu, K. Song, S. Pauly, T. Zhang, and J. Eckert, AIP Adv. 2, 032176 (2012). * Wu _et al._ (2014) F.-F. Wu, K. C. Chan, S.-T. Li, and G. Wang, J. Mater. Sci. 49, 2164 (2014). * Fu _et al._ (2007) X. L. Fu, Y. Li, and C. A. Schuh, Scr. Mater. 56, 617 (2007). * Song _et al._ (2013) K. K. Song, S. Pauly, B. A. Sun, J. Tan, M. Stoica, U. Kühn, and J. Eckert, AIP Adv. 3, 012116 (2013). * Cao _et al._ (2007) Q. P. Cao, J. F. Li, Y. Hu, A. Horsewell, J. Z. Jiang, and Y. H. Zhou, Mater. Sci. Eng. A 457, 94 (2007). * Chen _et al._ (2006) M. Chen, A. Inoue, W. Zhang, and T. Sakurai, Phys. Rev. Lett. 96, 245502 (2006). * Wilde and Rösner (2011) G. Wilde and H. Rösner, Appl. Phys. Lett. 98, 251904 (2011). * Bokeloh _et al._ (2011) J. Bokeloh, S. V. Divinski, G. Reglitz, and G. Wilde, Phys. Rev. Lett. 107, 235503 (2011). * Wu _et al._ (2012) Y. Wu, D. Q. Zhou, W. L. Song, H. Wang, Z. Y. Zhang, D. Ma, X. L. Wang, and Z. P. Lu, Phys. Rev. Lett. 109, 245506 (2012). * Corteen _et al._ (2011) J. Corteen, M. Rainforth, and I. Todd, Scr. Mater. 65, 524 (2011). * Albe _et al._ (2013) K. Albe, Y. Ritter, and D. Şopu, Mech. Mater. 67, 94 (2013). * Zaheri _et al._ (2014) A. Zaheri, F. Abdeljawad, and M. Haataja, Model. Simul. Mater. Sc. 22, 085008 (2014). * Wang _et al._ (2014a) Y. S. Wang, G. J. Hao, Y. Zhang, J. P. Lin, L. Song, and J. W. Qiao, Surf. Interface Anal. 46, 293 (2014a). * Wang _et al._ (2007) Y. Wang, J. Li, A. V. Hamza, and T. W. Barbee, Proc. Natl. Acad. Sci. 104, 11155 (2007). * Arman _et al._ (2011) B. Arman, C. Brandl, S. N. Luo, T. C. Germann, A. Misra, and T. Çağin, J. Appl. Phys. 110, 043539 (2011). * Brandl _et al._ (2013) C. Brandl, T. C. Germann, and A. Misra, Acta Mater. 61, 3600 (2013). * Lund and Schuh (2007) A. C. Lund and C. A. Schuh, Philos. Mag. Lett. 87, 603 (2007). * Shi and Falk (2008) Y. Shi and M. L. Falk, Acta Mater. 56, 995 (2008). * Rösner _et al._ (2014) H. Rösner, M. Peterlechner, C. Kübel, V. Schmidt, and G. Wilde, Ultramicroscopy 142, 1 (2014). * Kumar _et al._ (2007) G. Kumar, T. Ohkubo, T. Mukai, and K. Hono, Scr. Mater. 57, 173 (2007). * Wang _et al._ (2008) K. Wang, T. Fujita, Y. Zeng, N. Nishiyama, A. Inoue, and M. Chen, Acta Mater. 56, 2834 (2008). * Schmidt _et al._ (2015) V. Schmidt, H. Rösner, M. Peterlechner, G. Wilde, and P. M. Voyles, Phys. Rev. Lett. 115, 035501 (2015). * Schryvers _et al._ (1997) D. Schryvers, G. S. Firstov, J. W. Seo, J. V. Humbeeck, and Y. N. Koval, Scr. Mater. 36, 1119 (1997). * Plimpton (1995) S. Plimpton, J. Comp. Phys. 117, 1 (1995), http://lammps.sandia.gov/. * Mendelev _et al._ (2009) M. I. Mendelev, M. J. Kramer, R. T. Ott, D. J. Sordelet, D. Yagodin, and P. Popel, Philos. Mag. 89, 967 (2009). * Pauly _et al._ (2007) S. Pauly, J. Das, C. Duhamel, and J. Eckert, Adv. Eng. Mater. 9, 487 (2007). * Sun _et al._ (2005) Y. F. Sun, B. C. Wei, Y. R. Wang, W. H. Li, T. L. Cheung, and C. H. Shek, Appl. Phys. Lett. 87, 051905 (2005). * Das _et al._ (2007) J. Das, S. Pauly, C. Duhamel, B. C. Wei, and J. Eckert, J. Mater. Res. 22, 326 (2007). * Jiang _et al._ (2007) F. Jiang, D. H. Zhang, L. C. Zhang, Z. B. Zhang, L. He, J. Sun, and Z. F. Zhang, Mater. Sci. Eng. A 467, 139 (2007). * Shimizu _et al._ (2007) F. Shimizu, S. Ogata, and J. Li, Mater. Trans. 48, 2923 (2007). * Stukowski (2010) A. Stukowski, Model. Simul. Mater. Sc. 18, 015012 (2010), http://ovito.org/. * Stukowski (2012) A. Stukowski, Model. Simul. Mater. Sc. 20, 045021 (2012). * Cottrell (1953) A. H. Cottrell, _Dislocations and Plastic Flow in Crystals_ (Clarendon Press, 1953). * Hirth and Lothe (1982) J. P. Hirth and J. Lothe, _Theory of Dislocations_ , 2nd ed. (Krieger Publishing Company, 1982). * Gottstein (2004) G. Gottstein, _Physical Foundations of Materials Science_ (Springer, 2004). * Wang _et al._ (2014b) G. Wang, S. Pauly, S. Gorantla, N. Mattern, and J. Eckert, J. Mater. Sci. Technol. 30, 609 (2014b). * Ward _et al._ (2012) L. Ward, A. Agrawal, K. M. Flores, and W. Windl, preprint , (2012), arXiv:1209.0619 [cond-mat.mtrl-sci]. * Zhou _et al._ (2004) X. W. Zhou, R. A. Johnson, and H. N. G. Wadley, Phys. Rev. B 69, 144113 (2004). * Ritter and Albe (2011) Y. Ritter and K. Albe, Acta Mater. 59, 7082 (2011). * Kuo _et al._ (2014) C. N. Kuo, J. C. Huang, J. B. Li, J. S. C. Jang, C. H. Lin, and T. G. Nieh, J. Alloys Compd. 590, 453 (2014). * Hebert _et al._ (2006) R. J. Hebert, J. H. Perepezko, H. Rösner, and G. Wilde, Scr. Mater. 54, 25 (2006). * Ogata _et al._ (2002) S. Ogata, J. Li, and S. Yip, Science 298, 807 (2002).
Staggered Fermions Maarten Golterman Department of Physics and Astronomy, San Francisco State University, San Francisco, CA 94132, USA > These notes are based on a series of lectures on staggered fermions given at > the Centre de Physique Théorique, Luminy, in Marseille, France, January > 17-25, 2024. ## I Introduction These lecture notes present the lattice formulation of vectorlike gauge theories such as QED and QCD employing staggered fermions, which are extensively used in numerical simulations of QCD. While many aspects of the use of staggered fermions are discussed, these notes are not intended as a comprehensive review. Several topics and alternative formulations are only briefly mentioned, or not covered at all (see Sec. IX for pointers to additional reading). Likewise, while important references on staggered fermions and related topics have been included in the bibliography, the list of references is not complete. Staggered fermions were first introduced in a hamiltonian formalism [2, 3], but here we will only be concerned with the euclidean theory. These notes complement the review of Ref. [4]. ## II Staggered fermion action In this section we begin with the naive fermion lattice action, from which we derive the staggered fermion action through spin diagonalization. We then derive the Feynman rules, show that the classical continuum limit reproduces the expected gauge theory, and we list and discuss the symmetries of staggered fermions. We end with the formulation of the staggered theory on the so-called “taste basis.” ### II.1 Naive fermions The naive fermion action is given by $S={1\over 2}\sum_{x\mu}\left(\overline{\psi}(x)\gamma_{\mu}U_{\mu}(x)\psi(x+\mu)-\overline{\psi}(x+\mu)\gamma_{\mu}U^{\dagger}_{\mu}(x)\psi(x)\right)+m\sum_{x}\overline{\psi}(x)\psi(x)\ .$ (1) Points on the hypercubic lattice are labeled by $x$, and $x+\mu$ is the point one lattice spacing away from $x$ in the $+\mu$ direction. Most of the time we will set the lattice spacing $a=1$; if not, this point will be denoted as $x+a\mu$. $\psi(x)$ is a Dirac spinor transforming in the fundamental irreducible representation (irrep) of $SU(3)$, gauged by the link variables $U_{\mu}(x)$. We are in four-dimensional euclidean space, and the Dirac matrices $\gamma_{\mu}$ are hermitian. The parameter $m$ is the bare mass; later on, we will generalize this to a mass matrix $M$ when we consider $N_{f}>1$ flavors of staggered fermions. For now, we take $N_{f}=1$. Let us obtain the free fermion propagator in momentum space. We define the Fourier transform of $\psi$ and $\overline{\psi}$ by $\psi(x)=\int_{p}e^{ipx}\tilde{\psi}(p)\ ,\qquad\overline{\psi}(x)=\int_{p}e^{-ipx}\tilde{\overline{\psi}}(p)\ ,$ (2) where $\int_{p}\equiv\int_{-\pi}^{\pi}\frac{d^{4}p}{(2\pi)^{4}}\ .$ (3) In momentum space, the free action, obtained by setting $U_{\mu}(x)=1$, is $S_{\rm free}=\int_{p}\sum_{\mu}\tilde{\overline{\psi}}(p)\,i\gamma_{\mu}\sin(p_{\mu})\tilde{\psi}(p)+m\int_{p}\tilde{\overline{\psi}}(p)\tilde{\psi}(p)\ .$ (4) We see that the inverse propagator (at $m=0$) not only has a zero at $p=0$, but at all momenta $p=\pi_{A}\ ,$ (5) where $\pi_{A}$ has components $0$ or $\pi$. There are 16 such vectors, and the index $A$ thus runs from 1 to 16. For $p_{\mu}=\tilde{p}_{\mu}+\pi_{A}$, $\sin(p_{\mu})=\sin(\tilde{p}_{\mu}+\pi_{A\mu})=S^{A}_{\mu}\sin(\tilde{p}_{\mu})\ ,$ (6) which defines the signs $S^{A}_{\mu}=e^{i\pi^{A}_{\mu}}$. Therefore, the inverse propagator near $p=\pi_{A}$ for small $\tilde{p}$ equals $S^{-1}(p)=\sum_{\mu}i\gamma_{\mu}S^{A}_{\mu}\tilde{p}_{\mu}+m\ ,$ (7) showing that the continuum limit of the naive fermion action contains 16 relativistic fermions. These are the so-called fermion “species doublers.” In particular, the matrices $\gamma^{A}_{\mu}=\gamma_{\mu}S^{A}_{\mu}$ (8) satisfy the Dirac algebra $\\{\gamma_{\mu}^{A},\gamma_{\nu}^{A}\\}=2\delta_{\mu\nu}$, and are unitarily equivalent to the original set. For the relativistic fermion near $p=\pi_{A}$ we can define $\gamma_{5}^{A}=\gamma^{A}_{1}\gamma^{A}_{2}\gamma^{A}_{3}\gamma^{A}_{4}=S_{1}^{A}S_{2}^{A}S_{3}^{A}S_{4}^{A}\gamma_{1}\gamma_{2}\gamma_{3}\gamma_{4}=\pm\gamma_{5}\ ,$ (9) with the plus (minus) sign if the number of components of $\pi_{A}$ equal to $\pi$ is even (odd). This implies that half of the doublers have axial charge $+1$, and half of them have axial charge $-1$. The multiplet formed by the 16 species is thus anomaly free. This explains how the naive fermion action can regulate a Dirac fermion while maintaining exact chiral symmetry (for $m=0$) [5]. We note that the action (1) is invariant under hypercubic rotations in the $\mu\nu$ plane with the matrix $R_{1\over 2}=e^{{1\over 2}\omega\gamma_{\mu}\gamma_{\nu}}\ ,\qquad\mu<\nu$ (10) with $\omega=\pi/2$ acting on the Dirac spinor $\psi$.111Note that the generators ${1\over 2}\gamma_{\mu}\gamma_{\nu}$, $\mu<\nu$, are anti- hermitian. ### II.2 Spin diagonalization We now carry out the unitary basis transformation $\psi(x)=\gamma_{1}^{x_{1}}\gamma_{2}^{x_{2}}\gamma_{3}^{x_{3}}\gamma_{4}^{x_{4}}\chi(x)\ ,\qquad\overline{\psi}(x)=\bar{\chi}(x)\gamma_{4}^{x_{4}}\gamma_{3}^{x_{3}}\gamma_{2}^{x_{2}}\gamma_{1}^{x_{1}}\ .$ (11) This brings the action (1) into the form $S={1\over 2}\sum_{x\mu}\eta_{\mu}(x)\left(\bar{\chi}(x)U_{\mu}(x)\chi(x+\mu)-\bar{\chi}(x+\mu)U^{\dagger}_{\mu}(x)\chi(x)\right)+m\sum_{x}\bar{\chi}(x)\chi(x)\ ,$ (12) with the phase factors $\eta_{\mu}(x)$ defined by $\eta_{\mu}(x)=(-1)^{x_{1}+\dots+x_{\mu-1}}\ .$ (13) We can make different choices in Eq. (11) by reordering the $\gamma$ matrices, and this will lead to different results for the phase factors $\eta_{\mu}(x)$. However, all essential properties of these phase factors that we will use below are independent of this choice. In Eq. (12) the spin matrices $\gamma_{\mu}$ have been diagonalized. We now change the theory by dropping the spin index on the fields $\chi$ and $\bar{\chi}$. This defines the staggered fermion action, with the “one- component” fields222Of course, these fields still carry a color index and possibly a flavor index, so “one-component” here only refers to spin. $\chi$ and $\bar{\chi}$ (for a discussion of the euclidean action and earlier references, see Ref. [6]). The reduction from the naive to the staggered theory reduces the number of degrees of freedom by a factor 4, so one might expect the staggered theory to only have 4 species doublers in the continuum limit. However, as this reduction was obtained by spin diagonalization, it is not obvious that this is the correct interpretation of the staggered theory. In order to demonstrate this, one first should show that the action (12) has the correct classical continuum limit, and then that quantum corrections do not change this conclusion. If the theory is asymptotically free, so that the continuum limit is taken at the gaussian fixed point, the latter question can be studied in weak-coupling perturbation theory (WCPT), and we will do so in the next chapter. In order to see that the staggered action (12) may have a sensible continuum limit, we will define several operators on the field $\chi$. First, we define $T_{\pm\mu}:\ \chi(x)\to\eta_{\mu}(x)\chi(x\pm\mu)\ ,$ (14) from which it follows that $T_{\mu}T_{\nu}:\ \chi(x)\to\eta_{\mu}(x)\eta_{\nu}(x+\mu)\chi(x+\mu+\nu)=\varepsilon_{\mu\nu}(-1)^{x_{1}+\dots+x_{\mu-1}+x_{1}+\dots+x_{\nu-1}}\chi(x+\mu+\nu)\ ,$ (15) with $\varepsilon_{\mu\nu}=\Biggl{\\{}\begin{array}[]{cc}+1\ ,&\mu\geq\nu\\\ -1\ ,&\mu<\nu\end{array}\ .$ (16) From this, it follows that $\\{T_{\mu},T_{\nu}\\}:\ \chi(x)\to 2\delta_{\mu\nu}\chi(x+2\mu)\ .$ (17) We see that the square of $T_{\mu}$ is a normal translation over two lattice spacings in the $\mu$ direction. Moreover the $T_{\mu}$ themselves satisfy the Dirac algebra modulo these translations by $2\mu$. The irreps of the group generated by the $T_{\mu}$ live in momentum space, as $T_{\mu}$ involves a translation. If we define $\hat{T}_{\mu}$ as the translation (14) modding out the translations over $2\mu$, the $\hat{T}_{\mu}$ generate a 32-element finite group isomorphic with that generated by the Dirac matrices $\gamma_{\mu}$. We will denote this group as $\Gamma_{4}$.333Not to be confused with the matrix $\Gamma_{4}$ that will be introduced later on. The context should help avoid confusion. This group has 16 one-dimensional irreps (obtained by mapping the matrices $\gamma_{\mu}$ onto $\pm 1$ in all possible ways), and one four-dimensional irrep generated by the $\gamma_{\mu}$. Since the 16 fields $\chi(x)$ within a hypercube form a 16-dimensional representation of $\Gamma_{4}$, this representation is reducible (but it will be irreducible under the full lattice symmetry group, see Sec. II.4). We thus expect that a basis transformation exists such that the $\hat{T}_{\mu}$ can be written as $\hat{T}_{\mu}=\gamma_{\mu}\times 1_{4}\ ,$ (18) with $1_{4}$ the $4\times 4$ unit matrix. We expect similar transformations to exist of the form $1_{4}\times\xi_{\nu}$, with $\xi_{\nu}$ also satisfying the Dirac algebra. Indeed, we can define shifts [7, 8] $S_{\pm\mu}:\ \chi\to\zeta_{\mu}(x)\chi(x\pm\mu)\ ,$ (19) and require the new phase factors $\zeta_{\mu}(x)$ to be such that $[S_{\mu},T_{\nu}]=0\ .$ (20) This implies that $\zeta_{\mu}(x)\eta_{\nu}(x+\mu)=\eta_{\nu}(x)\zeta_{\mu}(x+\nu)\ ,$ (21) which can be rewritten as $\zeta_{\mu}(x+\nu)=\eta_{\nu}(x)\eta_{\nu}(x+\mu)\zeta_{\mu}(x)=\varepsilon_{\mu\nu}\zeta_{\mu}(x)\ ,$ (22) and thus $\zeta_{\mu}(x)=(-1)^{x_{\mu+1}+\dots+x_{4}}\ .$ (23) With $S_{\mu}$ acting on the gauge fields as $S_{\mu}:\ U_{\nu}(x)\to U_{\nu}(x+\mu)\ ,$ (24) $S_{\mu}$ is a symmetry of the action (12), and $\\{S_{\mu},S_{\nu}\\}:\ \chi(x)\to 2\delta_{\mu\nu}\chi(x+2\mu)\ .$ (25) We note the similarity to Eq. (17). We will refer to this symmetry as shift symmetry. It is a crucial symmetry for staggered fermions, as we will see below. In contrast, the $T_{\mu}$ are not a symmetry of the theory, but they are helpful for the interpretation of the theory: as we will see in more detail below, they “generate” the spin matrices $\gamma_{\mu}$ needed to interpret the continuum limit as a theory of four Dirac fermions. It is instructive to recast the operators $T_{\mu}$ and $S_{\mu}$ in momentum space. We first define a subset of the vectors $\pi_{A}$: $\displaystyle\pi_{\eta_{1}}$ $\displaystyle=$ $\displaystyle(0,0,0,0)\ ,$ (26) $\displaystyle\pi_{\eta_{2}}$ $\displaystyle=$ $\displaystyle(\pi,0,0,0)\ ,$ $\displaystyle\pi_{\eta_{3}}$ $\displaystyle=$ $\displaystyle(\pi,\pi,0,0)\ ,$ $\displaystyle\pi_{\eta_{4}}$ $\displaystyle=$ $\displaystyle(\pi,\pi,\pi,0)\ ,$ in terms of which $\eta_{\mu}(x)=e^{i\pi_{\eta_{\mu}}x}\ .$ (27) We then have $\eta_{\mu}(x)\chi(x+\mu)=\int_{p}e^{ipx+i\pi_{\eta_{\mu}}x+ip_{\mu}}\tilde{\chi}(p)=\int_{p}e^{ipx}e^{ip_{\mu}}\tilde{\chi}(p+\pi_{\eta_{\mu}})\ ,$ (28) where we used that $(\pi_{\eta_{\mu}})_{\mu}=0$ and that the exponential has a periodicity $2\pi$ in each direction because $x_{\mu}$ is integer on the lattice. We now define the full Brillouin zone (BZ) as $-\frac{\pi}{2}<p_{\mu}\leq\frac{3\pi}{2}\ ,$ (29) and the reduced BZ as $-\frac{\pi}{2}<\tilde{p}_{\mu}\leq\frac{\pi}{2}\ .$ (30) We write $p=\tilde{p}+\pi_{A}$ and define $\phi_{A}(\tilde{p})=\chi(\tilde{p}+\pi_{A})\ .$ (31) It follows that $T_{\mu}:\ \phi_{A}(\tilde{p})=\chi(\tilde{p}+\pi_{A})\to S^{A}_{\mu}e^{i\tilde{p}_{\mu}}\chi(\tilde{p}+\pi_{A}+\pi_{\eta_{\mu}})=S_{\mu}^{A}(\hat{\pi}_{\eta_{\mu}})_{AB}e^{i\tilde{p}_{\mu}}\phi_{B}(\tilde{p})\ ,$ (32) with $(\hat{\pi}_{\eta_{\mu}})_{AB}=\overline{\delta}(\pi_{\eta_{\mu}}+\pi_{A}+\pi_{B})\ ,$ (33) and $\overline{\delta}(\pi_{A})=1$ for $\pi_{A}=0$ (mod $2\pi$) and zero otherwise. Defining also $(S_{\mu})_{AB}=S_{\mu}^{A}\delta_{AB}$ (34) and $\Gamma_{\mu}=S_{\mu}\hat{\pi}_{\eta_{\mu}}\ ,$ (35) Eq. (32) can be rewritten as $T_{\mu}:\ \phi(\tilde{p})\to e^{i\tilde{p}_{\mu}}\Gamma_{\mu}\phi(\tilde{p})\ .$ (36) From Eq. (17) it then follows that $\\{\Gamma_{\mu},\Gamma_{\nu}\\}=2\delta_{\mu\nu}\ ,$ (37) i.e., the $16\times 16$ matrices $\Gamma_{\mu}$ obey the Dirac algebra. Analogously, we can define $\displaystyle\pi_{\zeta_{1}}$ $\displaystyle=$ $\displaystyle(0,\pi,\pi,\pi)\ ,$ (38) $\displaystyle\pi_{\zeta_{2}}$ $\displaystyle=$ $\displaystyle(0,0,\pi,\pi)\ ,$ $\displaystyle\pi_{\zeta_{3}}$ $\displaystyle=$ $\displaystyle(0,0,0,\pi)\ ,$ $\displaystyle\pi_{\zeta_{4}}$ $\displaystyle=$ $\displaystyle(0,0,0,0)\ ,$ and $\displaystyle(\hat{\pi}_{\zeta_{\mu}})_{AB}$ $\displaystyle=$ $\displaystyle\overline{\delta}(\pi_{\zeta_{\mu}}+\pi_{A}+\pi_{B})\ ,$ (39) $\displaystyle\Xi_{\mu}$ $\displaystyle=$ $\displaystyle S_{\mu}\hat{\pi}_{\zeta_{\mu}}\ ,$ to rewrite Eq. (19) in momentum space as $S_{\mu}:\ \phi(\tilde{p})\to e^{i\tilde{p}_{\mu}}\Xi_{\mu}\phi(\tilde{p})\ ,$ (40) with $\Xi_{\mu}$ another set of $16\times 16$ matrices. From Eqs. (20) and (25), it follows that $\\{\Xi_{\mu},\Xi_{\nu}\\}=2\delta_{\mu\nu}\ ,\qquad[\Gamma_{\mu},\Xi_{\nu}]=0\ .$ (41) As in Eq. (18), we can find a basis transformation such that $\Gamma_{\mu}\Xi_{\nu}\to\gamma_{\mu}\times\xi_{\nu}\ ,$ (42) with $\gamma_{\mu}$ and $\xi_{\nu}$ both sets of $4\times 4$ matrices satisfying the Dirac algebra. We note that $\displaystyle[S_{\mu},\hat{\pi}_{\eta_{\mu}}]=[S_{\mu},\hat{\pi}_{\zeta_{\mu}}]=0\ ,$ $\displaystyle S_{\mu}=S_{\mu}^{T}\ ,\quad\hat{\pi}_{\eta_{\mu}}=\hat{\pi}_{\eta_{\mu}}^{T}\ ,\quad\hat{\pi}_{\zeta_{\mu}}=\hat{\pi}_{\zeta_{\mu}}^{T}\quad\Rightarrow\quad\Gamma_{\mu}=\Gamma_{\mu}^{T}\ ,\quad\Xi_{\mu}=\Xi_{\mu}^{T}\ .$ For an explicit representation, see Ref. [8]. ### II.3 Feynman rules and classical continuum limit Using Eq. (27), the free staggered action can be written in momentum space as $S_{\rm free}={1\over 2}\int_{p}\int_{q}\tilde{\overline{\chi}}(q)\left(\sum_{\mu}\delta(p-q+\pi_{\eta_{\mu}})(e^{ip_{\mu}}-e^{-iq_{\mu}})+m\delta(p-q)\right)\tilde{\chi}(p)\ ,$ (44) from which we obtain the inverse propagator (using $(\pi_{\eta_{\mu}})_{\mu}=0$ again) $G^{-1}(p,-q)=\sum_{\mu}\delta(p-q+\pi_{\eta_{\mu}})i\sin(p_{\mu})+m\delta(p-q)\ ,$ (45) where $\delta$ is the periodic delta function with period $2\pi$ in each direction. We now set $p=\tilde{p}+\pi_{A}\ ,\qquad q=\tilde{q}+\pi_{B}\ ,$ (46) to obtain $\displaystyle G^{-1}(\tilde{p}+\pi_{A},-\tilde{q}-\pi_{B})$ $\displaystyle=$ $\displaystyle\sum_{\mu}\delta(\tilde{p}-\tilde{q}+\pi_{A}+\pi_{B}+\pi_{\eta_{\mu}})\,i\sin(p_{\mu})$ $\displaystyle\hskip 28.45274pt+m\delta(\tilde{p}-\tilde{q}+\pi_{A}+\pi_{B})$ $\displaystyle=$ $\displaystyle\delta(\tilde{p}-\tilde{q})\left(iS^{A}_{\mu}(\hat{\pi}_{\eta_{\mu}})_{AB}\sin(\tilde{p}_{\mu})+m\delta_{AB}\right)$ $\displaystyle\equiv$ $\displaystyle\delta(\tilde{p}-\tilde{q})S^{-1}_{AB}(\tilde{p})\ .$ The second line follows, because, from Eq. (30), $-\frac{\pi}{2}<\tilde{p}_{\mu}-\tilde{q}_{\mu}<\frac{\pi}{2}$, with strict inequalities. We thus find for the free propagator $S(\tilde{p})=\left(\sum_{\mu}i\Gamma_{\mu}\sin(\tilde{p}_{\mu})+m\right)^{-1}\ ,$ (48) and the free action becomes $S_{\rm free}=\int_{\tilde{p}}\sum_{AB}\overline{\phi}_{A}(\tilde{p})\left(\sum_{\mu}i(\Gamma_{\mu})_{AB}\sin(\tilde{p}_{\mu})+m\delta_{AB}\right)\phi_{B}(\tilde{p})\ .$ (49) This action is invariant under a group $U(4)$ generated by the $\Xi_{\mu}$ (all products of these matrices precisely give the generators of the group $U(4)$), because this group commutes with $\Gamma_{\mu}$. Free staggered fermions contain 4 degenerate species of Dirac spinors in the continuum limit. We will refer to these 4 species as “tastes,” and we find that the free theory has an exact $U(4)$ taste symmetry on the lattice. In the interacting theory, the product of this $U(4)$ and lattice translations is broken to the discrete symmetries $S_{\mu}$, cf. Eqs. (19), (24) and (40). To obtain the one-gluon vertex, we expand $U_{\mu}(x)=e^{igA_{\mu}(x)}=1+igA_{\mu}(x)-{1\over 2}\,g^{2}A_{\mu}^{2}(x)+\dots\ ,$ (50) and Fourier transform $A_{\mu}(x)=\int_{k}e^{ikx}\tilde{A}_{\mu}(k)\ .$ (51) The one-gluon part of the action is thus $S_{\rm 1-gluon}=ig\int_{p}\int_{q}\int_{k}\sum_{\mu}\tilde{\overline{\chi}}(q)\delta(p+k-q+\pi_{\eta_{\mu}}){1\over 2}(e^{ip_{\mu}}+e^{-iq_{\mu}})\tilde{A}_{\mu}(k)\tilde{\chi}(p)\ ,$ (52) from which we read off the one-gluon vertex $V^{m}_{\mu}(p,-q,k)=-ig\,\delta(p+k-q+\pi_{\eta_{\mu}})\,{1\over 2}(e^{ip_{\mu}}+e^{-iq_{\mu}})\,t_{m}\ ,$ (53) where $\tilde{A}_{\mu}(k)=\sum_{m}\tilde{A}_{\mu}^{m}(k)t_{m}$, with $t_{m}$ the gauge group generators. We can take the classical continuum limit by setting $p=\tilde{p}+\pi_{A}$, $q=\tilde{q}+\pi_{B}$ with $\tilde{p}$, $\tilde{q}$ and $k$ small. Using ${1\over 2}(e^{ip_{\mu}}+e^{-iq_{\mu}})={1\over 2}(S_{\mu}^{A}e^{i\tilde{p}_{\mu}}+S_{\mu}^{B}e^{-i\tilde{q}_{\mu}})\approx{1\over 2}(S_{\mu}^{A}+S_{\mu}^{B})$ (54) and $\delta(\tilde{p}-\tilde{q}+k+\pi_{A}+\pi_{B}+\pi_{\eta_{\mu}})=\delta(\tilde{p}-\tilde{q}+k)(\hat{\pi}_{\eta_{\mu}})_{AB}$ (55) for $\tilde{p}$, $\tilde{q}$ and $k$ small, the one-gluon vertex becomes $V^{m}_{\mu}(\tilde{p}+\pi_{A},-\tilde{q}-\pi_{B},k)\approx- ig\,\delta(\tilde{p}-\tilde{q}+k)(\Gamma_{\mu})_{AB}t_{m}\ ,$ (56) which is the correct classical continuum limit. In particular, the classical continuum limit is invariant under $U(4)_{\rm taste}$. The two-gluon vertex can be read off using the $O(g^{2})$ part of the expansion (50), with $\ell$ the momentum of the second gluon $V^{m_{1}m_{2}}_{\mu_{1}\mu_{2}}(p,-q,k,\ell)={1\over 2}\,g^{2}\sum_{\mu}\delta(p+k+\ell-q+\pi_{\eta_{\mu}})\,{1\over 2}(e^{ip_{\mu}}-e^{-iq_{\mu}})\,\delta_{\mu\mu_{1}}\delta_{\mu\mu_{2}}\,{1\over 2}\\{t_{m_{1}},t_{m_{2}}\\}\ .$ (57) Because of the minus sign between the momentum-dependent phase factors, it is straightforward to show that this vertex vanishes in the classical continuum limit. We end this subsection with the observation that if we take $k_{\mu}=\pi$ in Eq. (53), and use ${1\over 2}(e^{ip_{\mu}}+e^{-iq_{\mu}})={1\over 2}(e^{ip_{\mu}}-e^{-ip_{\mu}})=i\sin(p_{\mu})=iS_{\mu}^{A}\sin(\tilde{p}_{\mu})$ (58) for this choice of $k$, we find that the one-gluon vertex for a gluon with momentum $k_{\mu}=\pi$ is suppressed for small $\tilde{p}_{\mu}$. We will return to this point in Sec. VIII. For the complete Feynman rules, we refer to Ref. [8]. ### II.4 More symmetries We have already identified one symmetry of the staggered action, the shifts defined by Eqs. (19) and (24). There are more symmetries of course, which we will discuss in this subsection. For more detail, see Ref. [8]. #### II.4.1 Hypercubic rotations The action is invariant under hypercubic rotations, which are rotations over an angle $\pi/2$ in each plane. Clearly, the action (49) is invariant under rotations $R_{\rho\sigma}$ in the $\rho\sigma$ plane over an angle $\pi/2$ which takes $\tilde{p}_{\rho}\to\tilde{p}_{\sigma}$ and $\tilde{p}_{\sigma}\to-\tilde{p}_{\rho}$, and $\phi(\tilde{p})\to e^{\frac{\pi}{4}\Gamma_{\rho}\Gamma_{\sigma}}\phi(R_{\rho\sigma}^{-1}\tilde{p})\qquad(\mbox{tentative})\ .$ (59) However, this is in the free theory, which has more symmetry than the interacting theory. Noting that the operators $T_{\mu}$ and $S_{\mu}$ are expected to transform similarly under hypercubic rotations, we replace Eq. (59) by $\phi(\tilde{p})\to e^{\frac{\pi}{4}\Gamma_{\rho}\Gamma_{\sigma}}e^{\frac{\pi}{4}\Xi_{\rho}\Xi_{\sigma}}\phi(R_{\rho\sigma}^{-1}\tilde{p})\ .$ (60) Of course, the gauge fields have to be rotated accordingly, and Eq. (60) can be translated back to position space. The form of rotations on the original staggered fields $\chi(x)$ and $\bar{\chi}(x)$ as well as the links $U_{\mu}(x)$ is given in Ref. [8]. Equation (60), along with the transformation on the gauge fields, constitutes the hypercubic rotational symmetry of the staggered action. These rotations form the discrete hypercubic subgroup of a group $SO(4)$, which is the diagonal subgroup of $SO(4)_{\rm spin}\times SO(4)_{\rm taste}$, with $SO(4)_{\rm taste}$ the unique $SO(4)$ subgroup of $SU(4)_{\rm taste}$. The field $\phi$ transforms in a bosonic irrep of Eq. (60). However, we can define a different group of lattice symmetries under which it transforms as a fermion field. Since we have shift and rotational symmetry, we can define a composite symmetry $\tilde{R}_{\rho\sigma}=R_{\rho\sigma}S_{\rho}\ ,$ (61) which can be visualized as rotations around the center of the hypercube based at the origin. Using that $R_{\rho\sigma}^{-1}S_{\rho}R_{\rho\sigma}=S_{\sigma}\ ,\qquad R_{\rho\sigma}^{-1}S_{\sigma}R_{\rho\sigma}=-S_{\rho}\ ,$ (62) one can prove that, modulo lattice translations over an even number of lattice spacings in any direction, $(\tilde{R}_{\rho\sigma})^{4}=-1\ ,$ (63) i.e., that a rotation over $2\pi$ exists which is equal to $-1$, as one would expect for a fermionic theory. In contrast, the fourth power of $R_{\rho\sigma}$ equals $+1$. In the construction of the symmetry (61) shift symmetry plays a key role. In momentum space, we find that $\tilde{R}_{\rho\sigma}=e^{i\tilde{p}_{\rho}}e^{\frac{\pi}{4}\Gamma_{\rho}\Gamma_{\sigma}}\frac{1}{\sqrt{2}}(\Xi_{\rho}-\Xi_{\sigma})=-ie^{i\tilde{p}_{\rho}}e^{\frac{\pi}{4}\Gamma_{\rho}\Gamma_{\sigma}}e^{i\frac{\pi}{2}\frac{1}{\sqrt{2}}(\Xi_{\rho}-\Xi_{\sigma})}\ ,$ (64) using $\left(\frac{1}{\sqrt{2}}(\Xi_{\rho}-\Xi_{\sigma})\right)^{2}=1\ .$ (65) #### II.4.2 Axis reversal Axis reversal $I_{\rho}$ in the $\rho$ direction takes $x_{\rho}\to-x_{\rho}$, or $\tilde{p}_{\rho}\to-\tilde{p}_{\rho}$ in momentum space. Again, we expect symmetry under the interchange $\Gamma_{\mu}\leftrightarrow\Xi_{\mu}$, and in momentum space, $\phi(\tilde{p})\to\Gamma_{\rho}\Gamma_{5}\Xi_{\rho}\Xi_{5}\phi(I_{\rho}\tilde{p})\ ,$ (66) where $\Gamma_{5}=\Gamma_{1}\Gamma_{2}\Gamma_{3}\Gamma_{4}\ ,\qquad\Xi_{5}=\Xi_{1}\Xi_{2}\Xi_{3}\Xi_{4}\ .$ (67) This form follows because we need the matrix acting on $\phi$ to commute with $\Gamma_{\mu}$ for $\mu\neq\rho$ and anticommute with $\Gamma_{\rho}$, because $\sin(I_{\rho}\tilde{p}_{\rho})=-\sin(\tilde{p}_{\rho})$ (cf. Eq. (49)). In position space, we have that [8] $I_{\rho}:\quad\chi(x)\to(-1)^{x_{\rho}}\chi(I_{\rho}x)\ .$ (68) We can define a parity transformation, $I_{s}=I_{1}I_{2}I_{3}\ ,$ (69) but we note that this involves a taste transformation: $I_{s}:\ \phi(\tilde{p})\to\Gamma_{4}\Xi_{4}\phi(I_{s}\tilde{p})\ .$ (70) Thus, “tasteless” parity $P$ can be defined by combining this with $S_{4}$, $P:\ \phi(\tilde{p})\to e^{i\tilde{p}_{4}}\Gamma_{4}\phi(I_{s}\tilde{p})\ ,$ (71) or, in position space, $P:\ \chi(x)\to(-1)^{x_{1}+x_{2}+x_{3}}\chi(I_{s}x+4)=\eta_{4}(x)\chi(I_{s}x+4)\ .$ (72) Of course, the gauge field also transforms non-trivially under axis reversal [8]. #### II.4.3 Fermion number Because the action is bilinear in $\chi$ and $\bar{\chi}$, it is invariant under $\chi\to e^{i\alpha}\chi\ ,\qquad\bar{\chi}\to\bar{\chi}e^{-i\alpha}\ .$ (73) #### II.4.4 $U(1)_{\epsilon}$ symmetry For $m=0$, the action is invariant under [5] $\chi(x)\to e^{i\beta\epsilon(x)}\chi(x)\ ,\qquad\bar{\chi}(x)\to\bar{\chi}(x)e^{i\beta\epsilon(x)}\ ,$ (74) in which $\epsilon(x)=(-1)^{x_{1}+x_{2}+x_{3}+x_{4}}=e^{i\pi_{\epsilon}x}\ ,\qquad\pi_{\epsilon}=(\pi,\pi,\pi,\pi)\ .$ (75) Since this is only a symmetry of the action for $m=0$, we can interpret this symmetry as a chiral symmetry. (As always, this interpretation depends on the form of the mass term. For instance, for a one-link mass term, $U(1)_{\epsilon}$ is a vectorlike symmetry [8]!) Using that $T_{4}T_{3}T_{2}T_{1}S_{-1}S_{-2}S_{-3}S_{-4}:\ \chi(x)\to\epsilon(x)\chi(x)\ ,$ (76) and using the fact that $T_{\pm\mu}$ ($S_{\pm\mu}$) generates a $\Gamma_{\mu}$ ($\Xi_{\mu}$), we find that $\hat{\pi}_{\epsilon}=\Gamma_{5}\Xi_{5}\ ,$ (77) and thus that in momentum space $U(1)_{\epsilon}$ symmetry takes the form $\phi(\tilde{p})\to e^{i\beta\Gamma_{5}\Xi_{5}}\phi(\tilde{p})\ ,\qquad\overline{\phi}(\tilde{p})\to\overline{\phi}(\tilde{p})e^{i\beta\Gamma_{5}\Xi_{5}}\ .$ (78) Indeed, this looks like an non-singlet axial transformation. #### II.4.5 Charge conjugation The staggered action is invariant under charge conjugation $C_{0}$: $\displaystyle\chi(x)\to i\epsilon(x)\bar{\chi}^{T}(x)\ ,\qquad\bar{\chi}(x)\to i\epsilon(x)\chi^{T}(x)\ ,$ (79) $\displaystyle U_{\mu}(x)\to U^{*}_{\mu}(x)\ ,$ where the transpose acts on the color (and possible flavor) indices. The factors $i\epsilon(x)$ are needed to make both the mass term and the kinetic term invariant under this symmetry. This implies that charge conjugation involves a taste transformation; in momentum space it looks like $\phi(\tilde{p})\to i\Gamma_{5}\Xi_{5}\overline{\phi}^{T}(-\tilde{p})\ ,\qquad\overline{\phi}(\tilde{p})\to i\Gamma_{5}\Xi_{5}\phi^{T}(-\tilde{p})\ .$ (80) The hope is, of course, that the lattice symmetries are sufficient to restore the full $SO(4)_{\rm rot}\,\times\,SU(4)_{\rm taste}$ in the continuum limit without any fine tuning of counter terms. We have noted already that the hypercubic rotations (60) are a discrete subgroup of the diagonal $SO(4)_{\rm diag}$ of $SO(4)_{\rm rot}\,\times\,SO(4)_{\rm taste}$, with $SO(4)_{\rm taste}\,\subset\,SU(4)_{\rm taste}$. Likewise, the group $\Gamma_{4}$ generated by shifts (modulo even translations) is a discrete subgroup of $SU(4)_{\rm taste}$. Finally, $U(1)_{\epsilon}$ provides one exact (non- singlet, in taste space) chiral symmetry. At the classical level, we have seen that indeed the continuum limit coincides with QCD. We will revisit this question at the one-loop level in the next chapter, and in terms of the Symanzik effective theory in Sec. IV. ### II.5 Taste basis Another basis that can sometimes be useful has been introduced a long time ago, the so-called taste basis [9, 10]. We will review the construction here for the free theory. We define a unitary transformation $\psi_{\alpha a}(y)=\frac{1}{2^{3/2}}\sum_{A}(\gamma_{A})_{\alpha a}\chi(2y+A)\ ,\qquad\overline{\psi}_{a\alpha}(y)=\frac{1}{2^{3/2}}\sum_{A}\bar{\chi}(2y+A)(\gamma_{A})^{\dagger}_{a\alpha}\ .$ (81) Here $A$ runs over the 16 vectors with components $0$ or $1$. The $\psi$ fields live on a “coarse” lattice, labeled by lattice sites $y$, and $\gamma_{A}=\gamma_{1}^{A_{1}}\gamma_{2}^{A_{2}}\gamma_{3}^{A_{3}}\gamma_{4}^{A_{4}}\ .$ (82) The index $\alpha$ will be interpreted as a spin index, and the index $a$ as a taste index below. In order to write the (free) one-component staggered fermion action in terms of the fields $\psi$ and $\overline{\psi}$, we need to invert Eq. (82). Using ${\rm tr}\,(\gamma_{B}^{\dagger}\gamma_{A})=4\delta_{AB}\ ,$ (83) we find that $\chi(2y+A)=\frac{1}{2^{5/2}}{\rm tr}\,(\gamma_{A}^{\dagger}\psi(y))\ ,\qquad\bar{\chi}(2y+A)=\frac{1}{2^{5/2}}{\rm tr}\,(\gamma_{A}\overline{\psi}(y))\ .$ (84) We will also need the completeness relation $\sum_{A}(\gamma_{A})_{\alpha a}(\gamma_{A})^{\dagger}_{b\beta}=4\delta_{\alpha\beta}\delta_{ab}\ .$ (85) Using this, it is straightforward to show that $m\sum_{x}\bar{\chi}(x)\chi(x)=\frac{1}{8}m\sum_{y}\overline{\psi}(y)\psi(y)\ .$ (86) The kinetic term is a little more involved: $\displaystyle{1\over 2}\sum_{x\mu}\eta_{\mu}(x)(\bar{\chi}(x)\chi(x+\mu)-\bar{\chi}(x+\mu)\chi(x))$ $\displaystyle=$ $\displaystyle{1\over 2}\sum_{yA\mu}\eta_{\mu}(A)(\bar{\chi}(2y+A)\chi(2y+A+\mu)-\bar{\chi}(2y+A+\mu)\chi(2y+A))$ $\displaystyle=$ $\displaystyle{1\over 2}\sum_{yA\mu,A_{\mu}=0}\eta_{\mu}(A)(\bar{\chi}(2y+A)\chi(2y+A+\mu)-\bar{\chi}(2y+A+\mu)\chi(2y+A))$ $\displaystyle+{1\over 2}\sum_{yA\mu,A_{\mu}=1}\eta_{\mu}(A)(\bar{\chi}(2y+A)\chi(2y+A+\mu)-\bar{\chi}(2y+A+\mu)\chi(2y+A))$ $\displaystyle=$ $\displaystyle\frac{1}{2^{6}}\sum_{yA\mu,A_{\mu}=0}\eta_{\mu}(A)({\rm tr}\,(\gamma_{A}\overline{\psi}(y)){\rm tr}\,(\gamma_{A+\mu}^{\dagger}\psi(y))-{\rm tr}\,(\gamma_{A+\mu}\overline{\psi}(y)){\rm tr}\,(\gamma_{A}^{\dagger}\psi(y)))$ $\displaystyle+\frac{1}{2^{6}}\sum_{yA\mu,A_{\mu}=0}\eta_{\mu}(A)({\rm tr}\,(\gamma_{A+\mu}\overline{\psi}(y)){\rm tr}\,(\gamma_{A}^{\dagger}\psi(y+\mu))-{\rm tr}\,(\gamma_{A}\overline{\psi}(y+\mu)){\rm tr}\,(\gamma_{A+\mu}^{\dagger}\psi(y)))\ .$ We now use that $\displaystyle\sum_{A,A_{\mu}=0}\eta_{\mu}(A)(\gamma_{A+\mu})_{\alpha a}(\gamma_{A}^{\dagger})_{b\beta}$ $\displaystyle=$ $\displaystyle\sum_{A}\eta_{\mu}(A)(\gamma_{A+\mu})_{\alpha a}(\gamma_{A}^{\dagger})_{b\beta}{1\over 2}((1+(-1)^{A_{\mu}})$ $\displaystyle={1\over 2}\sum_{A}(\gamma_{\mu})_{\alpha\gamma}(\gamma_{A})_{\gamma a}(\gamma^{\dagger}_{A})_{b\beta}+{1\over 2}\sum_{A}(-1)^{A_{\mu}}\eta_{\mu}(A)\zeta_{\mu}(A)(\gamma_{A})_{\alpha c}(\gamma_{\mu})_{ca}(\gamma^{\dagger}_{A})_{b\beta}$ $\displaystyle={1\over 2}\sum_{A}(\gamma_{\mu})_{\alpha\gamma}(\gamma_{A})_{\gamma a}(\gamma^{\dagger}_{A})_{b\beta}+{1\over 2}\sum_{A}\epsilon(A)(\gamma_{A})_{\alpha c}(\gamma_{\mu})_{ca}(\gamma^{\dagger}_{A})_{b\beta}$ $\displaystyle={1\over 2}\sum_{A}(\gamma_{\mu})_{\alpha\gamma}(\gamma_{A})_{\gamma a}(\gamma^{\dagger}_{A})_{b\beta}+{1\over 2}\sum_{A}(\gamma_{5})_{\alpha\gamma}(\gamma_{A})_{\gamma d}(\gamma_{5})_{dc}(\gamma_{\mu})_{ca}(\gamma^{\dagger}_{A})_{b\beta}$ $\displaystyle=2(\gamma_{\mu})_{\alpha\beta}\delta_{ab}+2(\gamma_{5})_{\alpha\beta}(\gamma_{5}\gamma_{\mu})_{ba}\ ,$ and, likewise, $\sum_{A,A_{\mu}=0}\eta_{\mu}(A)(\gamma_{A})_{\alpha a}(\gamma_{A+\mu}^{\dagger})_{b\beta}=2(\gamma_{\mu})_{\alpha\beta}\delta_{ab}-2(\gamma_{5})_{\alpha\beta}(\gamma_{5}\gamma_{\mu})_{ba}\ .$ (89) This allows us to write Eq. (II.5) as $\displaystyle\frac{1}{2^{4}}{1\over 2}\sum_{y\mu}\Biggl{(}{\rm tr}\,(\overline{\psi}(y)\gamma_{\mu}\psi(y+\mu)-\overline{\psi}(y+\mu)\gamma_{\mu}\psi(y))$ (90) $\displaystyle\hskip 36.98866pt+{\rm tr}\,(\overline{\psi}(y)\gamma_{5}\psi(y+\mu)\gamma_{5}\gamma_{\mu}+\overline{\psi}(y+\mu)\gamma_{5}\psi(x)\gamma_{5}\gamma_{\mu}-2\overline{\psi}(y)\gamma_{5}\psi(y)\gamma_{5}\gamma_{\mu})\Biggr{)}\ .$ We see that in this form, the free action is that of a naive fermion, with the addition of a spin ($\gamma_{5}$) and taste ($\gamma_{5}\gamma_{\mu}$) non- diagonal Wilson-like mass term, which removes the doublers. The fermion field $\psi$ is thus undoubled, and shows the explicit four-fold degeneracy of staggered fermions in the continuum limit. One can also translate the shift symmetry to the new basis. One obtains, in the free theory, that $S_{\mu}:\ \psi(y)\to{1\over 2}(\psi(x)+\psi(x+\mu))\gamma_{\mu}+{1\over 2}\gamma_{5}\gamma_{\mu}(\psi(x)-\psi(x+\mu))\gamma_{5}\approx\psi(y)\gamma_{\mu}\ ,$ (91) where the approximate equality on the right applies in the continuum limit. We can identify a taste matrix $\xi_{\mu}=\gamma_{\mu}^{T}$ acting on the taste index of the field $\psi$. It is tempting to gauge Eq. (90) on the coarse lattice, instead of starting from the one-component staggered action. However, it is clear that doing this breaks shift symmetry,444Gauging also Eq. (91) does not make it into a symmetry of the gauged version of Eq. (90). and we would thus end up with a different interacting theory, with fewer lattice symmetries. We will see in the next chapter that this would undo a key property of staggered fermions, which is that the mass renormalizes multiplicatively. ## III One loop calculations In this section, we calculate the one-loop vacuum polarization, and demonstrate that it satisfies the Ward–Takahashi identity (WTI), which follows from the gauge invariance of the lattice theory. We also consider the one-loop fermion self energy, and show how staggered symmetries limit possible counter terms. These examples illustrate the conjecture that lattice QCD (or QED) with staggered fermions has a well-defined relativistic continuum limit, with no need for additional counter terms in the lattice action. ### III.1 Vacuum polarization In Fig. 1 we show the two diagrams contributing at one loop to the vacuum polarization. For simplicity, we consider the case of QED with gauge coupling $g$. Using the Feynman rules of Sec. II.3, the tadpole diagram is equal to $\Pi^{\rm tadpole}_{\mu\nu}(k,\ell)=-g^{2}\,\delta_{\mu\nu}\int_{pq}\delta(q-p+k-\ell+\pi_{\eta_{\mu}})\,{1\over 2}\left(e^{i(q+k-\ell)_{\mu}}-e^{-iq_{\mu}}\right)G(p,-q)\ .$ (92) Setting $p=\tilde{p}+\pi_{A}$ and $q=\tilde{q}+\pi_{B}$ this becomes $-g^{2}\delta_{\mu\nu}e^{\frac{i}{2}(k-\ell)_{\mu}}\\!\\!\int_{\tilde{p}\tilde{q}}\sum_{AB}\delta(\tilde{q}-\tilde{p}+k-\ell+\pi_{A}+\pi_{B}+\pi_{\eta_{\mu}})i\sin(\tilde{q}_{\mu}+\mbox{$\small{{1\over 2}}$}(k_{\mu}-\ell_{\mu}))S_{\mu}^{B}G(\tilde{p}+\pi_{A},-\tilde{q}-\pi_{B})\ .$ (93) We now use that (cf. Eq. (II.3)) $G(\tilde{p}+\pi_{A},-\tilde{q}-\pi_{B})=\delta(\tilde{p}-\tilde{q})S_{AB}(\tilde{p})\ ,$ (94) which sets the combination $\tilde{q}-\tilde{p}$ in $\delta(\tilde{q}-\tilde{p}+k-\ell+\pi_{A}+\pi_{B}+\pi_{\eta_{\mu}})$ equal to zero. Then, we use that $k$ and $\ell$, being external momenta, are taken small, so that we can factorize $\delta(k-\ell+\pi_{A}+\pi_{B}+\pi_{\eta_{\mu}})=\delta(k-\ell)(\hat{\pi}_{\eta_{\mu}})_{AB}\ ,$ (95) and use Eq. (35) to arrive at $\displaystyle\Pi^{\rm tadpole}_{\mu\nu}(k)$ $\displaystyle=$ $\displaystyle-g^{2}\delta_{\mu\nu}\int_{\tilde{q}}{\rm tr}\,\left(\Gamma_{\mu}\frac{-i\sum_{\kappa}\Gamma_{\kappa}\sin(\tilde{q}_{\kappa})+m}{\sum_{\kappa}\sin^{2}(\tilde{q}_{\kappa})+m^{2}}\right)i\sin(\tilde{q}_{\mu})$ $\displaystyle=$ $\displaystyle-16g^{2}\delta_{\mu\nu}\int_{\tilde{q}}\frac{\sin^{2}(\tilde{q}_{\mu})}{\sum_{\kappa}\sin^{2}(\tilde{q}_{\kappa})+m^{2}}\ ,$ where we omitted $\delta(k-\ell)$. We see that $\Pi^{\rm tadpole}_{\mu\nu}(k)$ does not actually depend on $k$, as one would expect for a tadpole diagram. $\mu$$\nu$$\to k$$\to\ell$$p$$q$$\mu$$\nu$$q^{\prime}$$p^{\prime}$$p$$q$$\to k$$\to\ell$ Figure 1: The two diagrams contributing to the one-loop vacuum polarization. The left panel shows the tadpole diagram, the right panel the sunset diagram. The other diagram in Fig. 1 is $\displaystyle\Pi^{\rm sunset}_{\mu\nu}(k,\ell)$ $\displaystyle=$ $\displaystyle g^{2}\int_{pqp^{\prime}q^{\prime}}\delta(q^{\prime}-p+k+\pi_{\eta_{\mu}})\,{1\over 2}\left(e^{i(q^{\prime}+k)_{\mu}}+e^{-iq^{\prime}_{\mu}}\right)G(p,-q)$ $\displaystyle\hskip 34.14322pt\times\delta(q-p^{\prime}-\ell+\pi_{\eta_{\nu}})\,{1\over 2}\left(e^{i(q-\ell)_{\nu}}+e^{-iq_{\nu}}\right)G(p^{\prime},-q^{\prime})\ .$ Integrating over $p^{\prime}$ and $q^{\prime}$ and writing $p=\tilde{p}+\pi_{A}$, $q=\tilde{q}+\pi_{B}$ makes this equal to $\displaystyle\Pi^{\rm sunset}_{\mu\nu}(k,\ell)$ $\displaystyle=$ $\displaystyle g^{2}\int_{\tilde{p}\tilde{q}}\sum_{AB}{1\over 2}S_{\mu}^{A}\left(e^{i\tilde{p}_{\mu}}+e^{-i(\tilde{p}-k)_{\mu}}\right)\delta(\tilde{p}-\tilde{q})S_{AB}(\tilde{p})$ $\displaystyle\times{1\over 2}S_{\nu}^{B}\left(e^{i(\tilde{q}-\ell)_{\nu}}+e^{-i\tilde{q}_{\nu}}\right)G(\tilde{q}-\ell+\pi_{\eta_{\nu}}+\pi_{B},-\tilde{p}+k+\pi_{\eta_{\mu}}+\pi_{A})\ ,$ where we used the inverse of Eq. (II.3). Using that we can set $\tilde{q}=\tilde{p}$ because of the $\delta(\tilde{p}-\tilde{q})$ inside the integrand, we use Eqs. (45) to write $\displaystyle G^{-1}(\tilde{p}-\ell+\pi_{\eta_{\nu}}+\pi_{B},-\tilde{p}+k+\pi_{\eta_{\mu}}+\pi_{A})=$ (99) $\displaystyle\hskip 28.45274pt\sum_{\kappa}\delta(-\ell+k+\pi_{\eta_{\kappa}}+\pi_{\eta_{\nu}}+\pi_{B}+\pi_{\eta_{\mu}}+\pi_{A})iS_{\kappa}^{\eta_{\nu}}S_{\kappa}^{B}\sin(\tilde{p}_{\kappa}-\ell_{\kappa})$ $\displaystyle\hskip 28.45274pt+m\delta(-\ell+k+\pi_{\eta_{\nu}}+\pi_{B}+\pi_{\eta_{\mu}}+\pi_{A})\ .$ Now, using that $k$ and $\ell$ are small, $\displaystyle\delta(-\ell+k+\pi_{\eta_{\nu}}+\pi_{B}+\pi_{\eta_{\mu}}+\pi_{A})$ $\displaystyle=$ $\displaystyle\delta(k-\ell)\overline{\delta}(\pi_{\eta_{\nu}}+\pi_{B}+\pi_{\eta_{\mu}}+\pi_{A})$ $\displaystyle=$ $\displaystyle\delta(k-\ell)\sum_{C}\overline{\delta}(\pi_{\eta_{\nu}}+\pi_{B}+\pi_{C})\overline{\delta}(\pi_{\eta_{\mu}}+\pi_{A}+\pi_{C})$ $\displaystyle=$ $\displaystyle\delta(k-\ell)(\hat{\pi}_{\eta_{\nu}}\hat{\pi}_{\eta_{\mu}})_{BA}\ ,$ and, likewise, $\displaystyle\delta(-\ell+k+\pi_{\eta_{\kappa}}+\pi_{\eta_{\nu}}+\pi_{B}+\pi_{\eta_{\mu}}+\pi_{A})S_{\kappa}^{\eta_{\nu}}S_{\kappa}^{B}=$ (101) $\displaystyle\hskip 28.45274pt=\delta(k-\ell)\overline{\delta}(\pi_{\eta_{\nu}}+\pi_{B}+\pi_{\eta_{\kappa}}+\pi_{\eta_{\mu}}+\pi_{A})S_{\kappa}^{\eta_{\nu}}S_{\kappa}^{B}$ $\displaystyle\hskip 28.45274pt=\delta(k-\ell)\sum_{CD}\overline{\delta}(\pi_{\eta_{\nu}}+\pi_{B}+\pi_{D})\overline{\delta}(\pi_{\eta_{\kappa}}+\pi_{D}+\pi_{C})\overline{\delta}(\pi_{\eta_{\mu}}+\pi_{C}+\pi_{A})S_{\kappa}^{D}$ $\displaystyle\hskip 28.45274pt=\delta(k-\ell)(\hat{\pi}_{\eta_{\nu}}\Gamma_{\kappa}\hat{\pi}_{\eta_{\mu}})_{BA}\ .$ We thus obtain $G(\tilde{p}-\ell+\pi_{\eta_{\nu}}+\pi_{B},-\tilde{p}+k+\pi_{\eta_{\mu}}+\pi_{A})=\delta(k-\ell)(\hat{\pi}_{\eta_{\mu}}S(\tilde{p}-\ell)\hat{\pi}_{\eta_{\nu}})_{BA}\ .$ (102) Combining with the factors $S_{\nu}^{B}$ and $S_{\mu}^{A}$ in Eq. (III.1), we arrive at $\Pi^{\rm sunset}_{\mu\nu}(k,\ell)=g^{2}e^{\frac{i}{2}(k_{\mu}-\ell_{\nu})}\delta(k-\ell)\int_{\tilde{p}}{\rm tr}\,\left(S(\tilde{p})\Gamma_{\mu}S(\tilde{p}-k)\Gamma_{\nu}\right)\cos(\tilde{p}_{\mu}-\mbox{$\small{{1\over 2}}$}k_{\mu})\cos(\tilde{p}_{\nu}-\mbox{$\small{{1\over 2}}$}k_{\nu})\ .$ (103) Using ${\rm tr}\,(\Gamma_{\kappa}\Gamma_{\nu}\Gamma_{\lambda}\Gamma_{\mu})=16(\delta_{\kappa\nu}\delta_{\lambda\mu}-\delta_{\kappa\lambda}\delta_{\mu\nu}+\delta_{\kappa\mu}\delta_{\lambda\nu})\ ,$ (104) this equals $\displaystyle\Pi^{\rm sunset}_{\mu\nu}(k)=16g^{2}e^{\frac{i}{2}(k_{\mu}-k_{\nu})}$ (105) $\displaystyle\int_{\tilde{p}}\frac{\delta_{\mu\nu}(\sum_{\kappa}\sin(\tilde{p}_{\kappa})\sin(\tilde{p}_{\kappa}-k_{\kappa})+m^{2})-\sin(\tilde{p}_{\mu})\sin(\tilde{p}_{\nu}-k_{\nu})-\sin(\tilde{p}_{\nu})\sin(\tilde{p}_{\mu}-k_{\mu})}{(\sum_{\kappa}\sin^{2}(\tilde{p}_{\kappa})+m^{2})(\sum_{\lambda}\sin^{2}(\tilde{p}_{\lambda}-k_{\lambda})+m^{2})}$ $\displaystyle\phantom{\int_{\tilde{p}}}\times\cos(\tilde{p}_{\mu}-\mbox{$\small{{1\over 2}}$}k_{\mu})\cos(\tilde{p}_{\nu}-\mbox{$\small{{1\over 2}}$}k_{\nu})\ ,$ where again we omitted $\delta(k-\ell)$. Formally, in the continuum we would have obtained one quarter of this result (with $\sin(\tilde{p}_{\mu})\to\tilde{p}_{\mu}$, $\cos(\tilde{p}_{\mu}-{1\over 2}k_{\mu})\to 1$, etc.), confirming that there are four species doublers (tastes) in the staggered theory. Defining $\Pi_{\mu\nu}(k)=\Pi^{\rm sunset}_{\mu\nu}(k)+\Pi^{\rm tadpole}_{\mu\nu}(k)\ ,$ (106) we first show that this vanishes at $k=0$, thus demonstrating that the tadpole diagram removes the quadratic divergence present in the sunset diagram. At $k=0$, $\Pi^{\rm sunset}_{\mu\nu}(0)=16g^{2}\int_{\tilde{p}}\left(-2\frac{\sin(\tilde{p}_{\mu})\sin(\tilde{p}_{\nu})\cos(\tilde{p}_{\mu})\cos(\tilde{p}_{\nu})}{(\sum_{\kappa}\sin^{2}(\tilde{p}_{\kappa})+m^{2})^{2}}+\delta_{\mu\nu}\frac{\cos^{2}(\tilde{p}_{\mu})}{\sum_{\kappa}\sin^{2}(\tilde{p}_{\kappa})+m^{2}}\right)\ .$ (107) Using $-2\,\frac{\sin(\tilde{p}_{\mu})\cos(\tilde{p}_{\mu})}{(\sum_{\kappa}\sin^{2}(\tilde{p}_{\kappa})+m^{2})^{2}}=\frac{\partial}{\partial\tilde{p}_{\mu}}\frac{1}{\sum_{\kappa}\sin^{2}(\tilde{p}_{\kappa})+m^{2}}$ (108) and partially integrating the first term, one can show that $\Pi^{\rm sunset}_{\mu\nu}(0)$ is equal to minus $\Pi^{\rm tadpole}_{\mu\nu}(0)$. There are no boundary terms, because the integrand is periodic in every direction with period $\pi$. Next, we demonstrate that the WTI is satisfied, which on the lattice takes the form $\sum_{\mu}(1-e^{-ik_{\mu}})\Pi_{\mu\nu}(k)=0\ .$ (109) Using that $(1-e^{-ik_{\mu}})e^{\frac{i}{2}k_{\mu}}\Gamma_{\mu}\cos(\tilde{p}_{\mu}-{1\over 2}k_{\mu})=S^{-1}(\tilde{p})-S^{-1}(\tilde{p}-k)\ ,$ (110) one can simplify $\displaystyle\sum_{\mu}(1-e^{-ik_{\mu}})\Pi^{\rm sunset}_{\mu\nu}(k)$ $\displaystyle\\!\\!=\\!\\!$ $\displaystyle 16g^{2}e^{-\frac{i}{2}k_{\nu}}\int_{\tilde{p}}{\rm tr}\,(\Gamma_{\nu}(S(\tilde{p}-k)-S(\tilde{p}))\cos(\tilde{p}_{\nu}-\mbox{$\small{{1\over 2}}$}k_{\nu})$ $\displaystyle=-16ig^{2}e^{-\frac{i}{2}}\int_{\tilde{p}}\left(\frac{\sin(\tilde{p}_{\nu}-k_{\nu})}{\sum_{\kappa}\sin^{2}(\tilde{p}_{\kappa}-k_{\kappa})+m^{2}}-\frac{\sin(\tilde{p}_{\nu})}{\sum_{\kappa}\sin^{2}(\tilde{p}_{\kappa})+m^{2}}\right)\cos(\tilde{p}_{\nu}-\mbox{$\small{{1\over 2}}$}k_{\nu})$ $\displaystyle=-16ig^{2}e^{-\frac{i}{2}k_{\nu}}\int_{\tilde{p}}\frac{\sin(\tilde{p}_{\nu})}{\sum_{\kappa}\sin^{2}(\tilde{p}_{\kappa})+m^{2}}(\cos(\tilde{p}_{\nu}+\mbox{$\small{{1\over 2}}$}k_{\nu})-\cos(\tilde{p}_{\nu}-\mbox{$\small{{1\over 2}}$}k_{\nu})$ $\displaystyle=16g^{2}(1-e^{-ik_{\nu}})\int_{\tilde{p}}\frac{\sin^{2}(\tilde{p}_{\nu})}{\sum_{\kappa}\sin^{2}(\tilde{p}_{\kappa})+m^{2}}\ ,$ where we again used that the integrand is periodic in $\tilde{p}_{\mu}$ with period $\pi$. This equals $-\sum_{\mu}(1-e^{-ik_{\mu}})\Pi^{\rm tadpole}_{\mu\nu}(k)$, thus proving Eq. (109). Finally, in this subsection, we will indicate how one may actually go about completing the calculation of the vacuuum polarization at one-loop. We start from the expression $\displaystyle\Pi_{\mu\nu}(k)=16g^{2}e^{\frac{i}{2}(k_{\mu}-k_{\nu})}$ (112) $\displaystyle\int_{\tilde{p}}\Biggl{(}\frac{\delta_{\mu\nu}(\sum_{\kappa}\sin(\tilde{p}_{\kappa})\sin(\tilde{p}_{\kappa}-k_{\kappa})+m^{2})-\sin(\tilde{p}_{\mu})\sin(\tilde{p}_{\nu}-k_{\nu})-\sin(\tilde{p}_{\nu})\sin(\tilde{p}_{\mu}-k_{\mu})}{(\sum_{\kappa}\sin^{2}(\tilde{p}_{\kappa})+m^{2})(\sum_{\lambda}\sin^{2}(\tilde{p}_{\lambda}-k_{\lambda})+m^{2})}$ $\displaystyle\phantom{\int_{\tilde{p}}}\times\cos(\tilde{p}_{\mu}-\mbox{$\small{{1\over 2}}$}k_{\mu})\cos(\tilde{p}_{\nu}-\mbox{$\small{{1\over 2}}$}k_{\nu})$ $\displaystyle\phantom{\int_{\tilde{p}}}-\frac{\delta_{\mu\nu}(\sum_{\kappa}\sin^{2}(\tilde{p}_{\kappa})+m^{2})-2\sin(\tilde{p}_{\mu})\sin(\tilde{p}_{\nu})}{(\sum_{\kappa}\sin^{2}(\tilde{p}_{\kappa})+m^{2})^{2}}\cos(\tilde{p}_{\mu})\cos(\tilde{p}_{\nu})\Biggr{)}\ ,$ where the last line, obtained by setting $k=0$ in the middle two lines, is equal to the tadpole contribution because $\Pi_{\mu\nu}(0)=0$. This form is useful because it shows explicitly that $\Pi_{\mu\nu}(k)$ is logarithmically divergent; the last line explicitly subtracts the quadratic divergence present in $\Pi^{\rm sunset}_{\mu\nu}(k)$. Moreover, by shifting $\tilde{p}-{1\over 2}k\to\tilde{p}$, which is allowed in an integral that is only logarithmically divergent, one can show that the integral is quadratic in $k$ in the continuum limit (modulo, of course, a logarithmic dependence on $k$). Returning to Eq. (112), we split the integral into two regions: $\int_{\tilde{p}}=\int_{|\tilde{p}|<\delta}+\int_{|\tilde{p}|>\delta}\ ,$ (113) where we take $|ak|\sim|am|\ll\delta\ll 1\ ,$ (114) and we take the limit $a\to 0$ first, followed by $\delta\to 0$. (In Eq. (114) we wrote $k$ and $m$ in physical units, temporarily restoring the lattice spacing $a$.) We will refer to these two regions as the “inner” and “outer” regions. In the inner region, we may replace the integrand by its covariant form: $\displaystyle\Pi^{\rm inner}_{\mu\nu}(k)$ $\displaystyle=$ $\displaystyle 16g^{2}\int_{|\tilde{p}|<\delta}\Biggl{(}\frac{\delta_{\mu\nu}(\tilde{p}(\tilde{p}-k)+m^{2})-\tilde{p}_{\mu}(\tilde{p}_{\nu}-k_{\nu})-\tilde{p}_{\nu}(\tilde{p}_{\mu}-k_{\mu})}{(\tilde{p}^{2}+m^{2})((\tilde{p}-k)^{2}+m^{2})}$ $\displaystyle\phantom{16g^{2}\int_{|\tilde{p}|<\delta}\Biggl{(}}-\frac{\delta_{\mu\nu}(\tilde{p}^{2}+m^{2})-2\tilde{p}_{\mu}\tilde{p}_{\nu}}{(\tilde{p}^{2}+m^{2})^{2}}\Biggr{)}$ $\displaystyle=$ $\displaystyle\frac{1}{3\pi^{2}}g^{2}(k_{\mu}k_{\nu}-\delta_{\mu\nu}k^{2})\log(\delta^{2})+\mbox{finite\ terms}\ .$ Since $\Pi_{\mu\nu}(k)$ does not depend on $\delta$, the same logarithmic divergence (up to a sign) also has to be present in the outer region integral. We handle this by adding and subtracting $\frac{1}{3}g^{2}(k_{\mu}k_{\nu}-\delta_{\mu\nu}k^{2})\int_{|p|>\delta}\frac{1}{(\sum_{\mu}\sin^{2}({1\over 2}p_{\mu}))^{2}}\ ,$ (116) because, as we will show, $I\equiv\lim_{\delta\to 0}\left(\int_{|p|>\delta}\frac{1}{(\sum_{\mu}\sin^{2}({1\over 2}p_{\mu}))^{2}}+\frac{1}{\pi^{2}}\log(\delta^{2})\right)$ (117) is finite and can be explicitly calculated. This means that the sum of Eq. (116) and the inner region is finite, and thus that the difference of the outer region integral and Eq. (116) is finite as well. Since, in the outer region, $|ak|\sim|am|\ll\delta<|\tilde{p}|$, we can expand the subtracted outer-region integral in powers of $ak_{\mu}$ and $am$, keep those powers that survive in the continuum limit, and then take $\delta$ to zero, leading to purely numerical integrals that can be computed numerically. All non-analytic dependence on $k$ and $m$ resides in the inner-region integral. What remains to be done is to calculate Eq. (117). We begin rewriting this as $\lim_{\delta\to 0}\left(\lim_{m\to 0}\left(\int_{p}\frac{1}{(\sum_{\mu}\sin^{2}({1\over 2}p_{\mu})+\frac{1}{4}m^{2})^{2}}-\int_{|p|<\delta}\frac{16}{(p^{2}+m^{2})^{2}}\right)+\frac{1}{\pi^{2}}\log(\delta^{2})\right)\ .$ (118) The integral over the components $p_{\mu}$ of $p$ in the first integral runs from $-\pi\to\pi$. We use $\displaystyle\int_{p}\frac{1}{\sum_{\mu}\sin^{2}({1\over 2}p_{\mu})+\frac{1}{4}m^{2}}$ $\displaystyle=$ $\displaystyle 2\int_{0}^{\infty}dx\int_{p}e^{-{1\over 2}xm^{2}-2x\sum_{\mu}\sin^{2}({1\over 2}p_{\mu})}$ $\displaystyle=$ $\displaystyle 2\int_{0}^{\infty}dx\,e^{-{1\over 2}xm^{2}}e^{-4x}I_{0}^{4}(x)\ ,$ where $I_{0}(x)$ is the modified Bessel function, and hence $\displaystyle\int_{p}\frac{1}{(\sum_{\mu}\sin^{2}({1\over 2}p_{\mu})+\frac{1}{4}m^{2})^{2}}$ $\displaystyle=$ $\displaystyle-4\,\frac{\partial}{\partial m^{2}}\int_{p}\frac{1}{\sum_{\mu}\sin^{2}({1\over 2}p_{\mu})+\frac{1}{4}m^{2}}$ $\displaystyle=$ $\displaystyle 4\int_{0}^{\infty}dx\,x\,e^{-{1\over 2}xm^{2}}e^{-4x}I_{0}^{4}(x)\ .$ Furthermore, $-16\int_{|p|<\delta}\frac{1}{(p^{2}+m^{2})^{2}}=-16\int_{|p|<1}\frac{1}{(p^{2}+\frac{m^{2}}{\delta^{2}})^{2}}=-\frac{1}{\pi^{2}}\log\frac{\delta^{2}}{m^{2}}+O\left(\frac{m^{2}}{\delta^{2}}\right)\ ,$ (121) showing that the double limit in Eq. (118) is finite. We then calculate the integral on the left of Eq. (121) in a way similar to the calculation in Eq. (III.1), obtaining $-16\int_{|p|<\delta}\frac{1}{(p^{2}+m^{2})^{2}}=-\frac{1}{\pi^{2}}\int_{0}^{\infty}\frac{dx}{x}\,e^{-{1\over 2}x\frac{m^{2}}{\delta^{2}}}\left(1-(1+{1\over 2}x)e^{-{1\over 2}x}\right)\ .$ (122) We can now combine all terms in Eq. (118), take first $m$ and then $\delta$ to zero, obtaining $I=4\int_{0}^{\infty}dx\,x\left(e^{-4x}I_{0}^{4}(x)-\frac{1}{4\pi^{2}x^{2}}\left(1-(1+{1\over 2}x)e^{-{1\over 2}x}\right)\right)=0.4855321\ .$ (123) ### III.2 Fermion self energy The one-loop fermion self energy can be calculated similarly. The steps are similar to those followed in Sec. III.1 (for details, see Ref. [8]), and here we just give the result (in Feynman gauge): $\displaystyle\Sigma(\tilde{p})$ $\displaystyle=$ $\displaystyle g^{2}\sum_{\mu}\int_{k}{1\over 2}(1+\cos(k_{\mu}+2\tilde{p}_{\mu}))\Gamma_{\mu}S(k+\tilde{p})\Gamma_{\mu}D(k)$ $\displaystyle-g^{2}\sum_{\mu}\Gamma_{\mu}\,i\sin(\tilde{p}_{\mu})\int_{k}D(k)\ ,$ $\displaystyle D(k)$ $\displaystyle=$ $\displaystyle\frac{1}{4\sum_{\mu}\sin^{2}({1\over 2}k_{\mu})}\ .$ The term on the second line comes from a gauge-field tadpole diagram. This result can be further evaluated by splitting into inner and outer regions, etc., as we did for the vacuum polarization. $\Sigma(\tilde{p})$ is a $16\times 16$ matrix, and thus a linear combination of products of the matrices $\Gamma_{\mu}$ and $\Xi_{\nu}$. In the continuum limit, non-analytic dependence on $\tilde{p}$ and $m$ is the same as we would obtain in any other regularization, and comes from the inner-region integral. By dimensional analysis, the most general contact term one can find in the continuum limit has the form $C\equiv\tilde{p}_{\mu}X_{\mu}+\frac{1}{a}\,Y_{0}+mY\ ,$ (125) where again we restored the lattice spacing, and where $X_{\mu}$, $Y_{0}$ and $Y$ are $16\times 16$ matrices. We will now see how lattice symmetries restrict the form of $C$. First, shift symmetry implies that $C$ should commute with all $\Xi_{\nu}$, and therefore $X_{\mu}$, $Y_{0}$ and $Y$ cannot contain any $\Xi_{A}$. Then (hypercubic) rotational symmetry implies that $X_{\mu}$ has to be a vector, which implies that $X_{\mu}\propto\Gamma_{\mu}$ or $X_{\mu}\propto\Gamma_{\mu}\Gamma_{5}$, while both $Y_{0}$ and $Y$ have to be proportional to the $16\times 16$ unit matrix or $\Gamma_{5}$. Reflection symmetry $I_{s}$ of Eq. (70) excludes $\Gamma_{5}$ from appearing in any of these, and thus we find that $C=\alpha\,\tilde{p}_{\mu}\Gamma_{\mu}+\beta\,\frac{1}{a}+\gamma\,m\ ,$ (126) with $\alpha$, $\beta$ and $\gamma$ constants. Finally, we use that for $m=0$ we have $U(1)_{\epsilon}$ symmetry, which implies that, for $m=0$, $C$ should anti-commute with $\Gamma_{5}\Xi_{5}$. This sets $\beta=0$, and we thus see that $m$ renormalizes multiplicatively, while the contact terms take on the form expected in the continuum theory [8] (with coefficients specific to the lattice regularization, of course). We note that, in this argument, shift symmetry played a crucial role. Suppose we were to gauge the staggered theory on the taste basis, with action (90) plus (86) on the coarse lattice, losing shift symmetry in the process because the gauge fields $U_{\mu}(y)$ would now live on the coarse lattice. We would expect a counter term of the form $\sum_{\mu}{\rm tr}\,(\overline{\psi}(y)\gamma_{5}\psi(y)\gamma_{5}\gamma_{\mu})$ (127) because of the form of the “Wilson” term in Eq. (90). On our momentum basis, this translates into a counter term $Y_{0}\propto\sum_{\mu}\Gamma_{5}\Xi_{5}\Xi_{\mu}$. Indeed, this anti-commutes with $\Gamma_{5}\Xi_{5}$, so it is not excluded by $U(1)_{\epsilon}$ symmetry! And, indeed, such a counter term appears at one loop in the theory on the taste basis if it is gauged on the coarse lattice [11]. One can verify that the action on the taste basis is invariant under hypercubic rotations of the form (64). In particular, $\tilde{R}^{-1}_{\rho\sigma}\left(\sum_{\mu}\Gamma_{5}\Xi_{5}\Xi_{\mu}\right)\tilde{R}_{\rho\sigma}=\sum_{\mu}\Gamma_{5}\Xi_{5}\Xi_{\mu}\ ,$ (128) showing that a counter term of the form (127) is consistent with rotational symmetry. ## IV Symanzik effective theory for staggered fermions The Symanzik effective theory (SET) is a continuum effective theory that reproduces correlation functions of the underlying lattice theory expanded in powers of the lattice spacing, for momenta $\Lambda_{\rm QCD}\ll|p|\ll 1/a$, where $p$ is the typical momentum of these correlation functions.555In these notes we ignore anomalous dimensions. However, these could become important with high precision computations near the continuum limit [12]. In this chapter, we discuss the SET for QCD with $N_{f}$ flavors of staggered fermions, for $N_{f}>1$. The SET is formulated in terms of continuum quark fields $q_{i}$ ($i$ is the flavor index; we leave spin and color indices implicit) and gluons $A_{\mu}$, with field strength $G_{\mu\nu}$ [13, 14], because at these momenta quarks and gluons are still the relevant degrees of freedom. At leading order, the SET consists of all 4-dimensional operators consistent with the lattice symmetries. As we have seen in the previous chapter, the fermion mass renormalizes multiplicatively, and there are no dimension-3 operators.666In the previous chapter, we demonstrated this only to one loop, but the symmetry arguments we used generalize to higher loops. As we have seen, a crucial role in arriving at this conclusion is played by shift and $U(1)_{\epsilon}$ symmetries. Dimensional analysis, gauge invariance, and staggered symmetries imply that the dimension-4 operators in the SET are just those defining continuum QCD. Before we go on to discuss operators with dimension $d$ larger than 4 (which are thus multiplied by $a^{d-4}$, with $a$ the lattice spacing), we make a general observation on how shift symmetry works in the SET, using the fact that this is a continuum theory, which, by assumption, is invariant under continuous translation symmetry [15]. In momentum space, a shift symmetry takes the form $\phi(p)\to S_{\mu}\phi(p)=e^{iap_{\mu}}\Xi_{\mu}\phi(p)\ ,$ (129) where $-\pi/(2a)<p\leq\pi/(2a)$ is the physical momentum of the field $\phi$ on which $S_{\mu}$ acts, and we made the lattice spacing dependence explicit. The SET is invariant under shift symmetry because the underlying lattice theory is. At the same time, it is also invariant under continuous translations of the form $\phi(p)\to e^{ipr}\phi(p)\ ,$ (130) where $r$ is the vector over which we translate the field. Combining these two symmetries, choosing $r_{\mu}=-a$, $r_{\nu\neq\mu}=0$, we find that the SET is invariant under $\phi(p)\to\Xi_{\mu}\phi(p)\ .$ (131) This is convenient, as this symmetry does not mix operators of different dimensionality in the SET. Below we will make extensive use of Eq. (131), and we will still refer to this symmetry as “shift” symmetry. For additional arguments supporting this claim, diagramatically or using the taste basis, see Ref. [15]. In the discussion below, we will assume that a basis transformation has been carried out such that $\Gamma_{\mu}\to\gamma_{\mu}1_{4}\ ,\qquad\Xi_{\mu}\to 1_{4}\xi_{\mu}\ ,$ (132) where we omit the direct-product sign (i.e., $\gamma_{\mu}1_{4}=\gamma_{\mu}\times 1_{4}$, etc.), and we will work with the spin matrices $\gamma_{\mu}$ and the taste matrices $\xi_{\mu}$, both in the 4-dimensional irrep of the Dirac algebra in four dimensions. We will also drop the index $4$ on the $4\times 4$ unit matrix; in fact, we will mostly not write the unit matrix at all, when no confusion is possible. ### IV.1 Dimension-5 operators We begin with operators of dimension 5, which would be of order $a$ in the Symanzik expansion [16]. These operators can at most contain one bilinear in $q_{i}$ and $\overline{q}_{i}$. They cannot contain any non-trivial taste matrix, because of the symmetry (131). Bilinears of the form $\overline{q}_{i}\sigma_{\mu\nu}G_{\mu\nu}q_{i}\ ,\qquad\overline{q}_{i}D_{\mu}D_{\mu}q_{i}$ (133) (where $D_{\mu}$ is the color-covariant derivative) are ruled out immediately by $U(1)_{\epsilon}$ symmetry. In order to consider other dimension-5 operators, we first define $\displaystyle q_{Li}\\!\\!\\!$ $\displaystyle=$ $\displaystyle\\!\\!\\!{1\over 2}(1-\gamma_{5}\xi_{5})q_{i}\ ,\qquad q_{Ri}={1\over 2}(1+\gamma_{5}\xi_{5})q_{i}\ ,$ (134) $\displaystyle\overline{q}_{Li}\\!\\!\\!$ $\displaystyle=$ $\displaystyle\\!\\!\\!\overline{q}_{i}{1\over 2}(1+\gamma_{5}\xi_{5})\ ,\qquad\overline{q}_{Ri}=\overline{q}_{i}{1\over 2}(1-\gamma_{5}\xi_{5})\ .$ The lattice theory has a $U(N_{f})_{L}\times U(N_{f})_{R}$ symmetry,777Note that this symmetry is smaller than the $U(4N_{f})_{L}\times U(4N_{f})_{R}$ symmetry of the continuum limit. under which these fields transform as $\displaystyle q_{L}\to V_{L}q_{L}\ ,\qquad$ $\displaystyle q_{R}\to V_{R}q_{R}\ ,$ $\displaystyle\overline{q}_{L}\to\overline{q}_{L}V_{L}^{\dagger}\ ,\qquad$ $\displaystyle\overline{q}_{R}\to\overline{q}_{R}V_{R}^{\dagger}\ ,$ with $V_{L}=e^{\frac{i}{2}\alpha^{a}T^{a}(1-\gamma_{5}\xi_{5})}\ ,\qquad V_{R}=e^{\frac{i}{2}\beta^{a}T^{a}(1+\gamma_{5}\xi_{5})}\ ,$ (136) in which $T^{a}$ are the $U(N_{f})$ generators. This symmetry is a generalization of $U(1)\times U(1)_{\epsilon}$ symmetry for one flavor to the case of $N_{f}$ flavors. Note that other symmetries, such as shift and hypercubic symmetry, do not enlarge with the presence of more flavors, because the gauge field transforms non-trivially under these symmetries. Mass terms are not invariant under this symmetry, but can be made invariant by making the $N_{f}\times N_{f}$ mass matrix ${\cal M}$ into a spurion field (i.e., a non-dynamical field) that transforms as ${\cal M}\to V_{R}{\cal M}V_{L}^{\dagger}\ ,$ (137) so that the 4-dimensional staggered mass term $\bar{\chi}_{R}(x){\cal M}\chi_{L}(x)+\bar{\chi}_{L}(x){\cal M}^{\dagger}\chi_{R}(x)$ (138) is invariant (with the symmetry $U(N_{f})_{L}\times U(N_{f})_{R}$ acting on the staggered fields $\chi$ and $\bar{\chi}$ as in Eq. (136) with $\gamma_{5}\xi_{5}$ replaced by $\epsilon(x)$). The idea is to use the spurion form of ${\cal M}$ to see what terms can show up in the SET, and then, once we have found all such terms, set ${\cal M}$ equal to the physical mass matrix $M$. The dimension-4 mass term in the SET, $\overline{q}_{R}{\cal M}q_{L}+\overline{q}_{L}{\cal M}^{\dagger}q_{R}$ (139) is also invariant. We are now ready to return to dimension-5 quark bilinears containing one or more masses. Using the transformations (IV.1) and (137), the symmetry $U(N_{f})_{L}\times U(N_{f})_{R}$ rules out terms of the form $\overline{q}M\hbox to0.0pt{\hbox{$\mskip 1.0mu/$}\hss}\\!Dq\ ,\qquad\overline{q}M^{2}q\ ,\qquad{\rm tr}\,(M)G_{\mu\nu}G_{\mu\nu}\ .$ (140) This exhausts all possible dimension-5 contributions to the SET, and we thus conclude that dependence on the lattice spacing starts at order $a^{2}$, i.e., at dimension 6. We note that no use of equations of motion or field redefinitions was made to rule out all dimension-5 operators, and our conclusion thus holds for off-shell correlation functions as well. We end this subsection with a comment on the taste basis of Sec. II.5. The action (90) has a Wilson term, which is, in fact, a dimension-5 operator. However, this action is invariant under shift symmetry in the form (91), but not under the symmetry $\psi\to\psi\gamma_{\mu}$, which is the continuum form equivalent to Eq. (131). As it is the latter symmetry which is relevant for restricting the form of the SET, there is thus no contradiction. ### IV.2 Dimension-6 operators In this section, we will discuss the order-$a^{2}$ part of the SET, i.e., all dimension-6 operators that contribute, with the focus on operators that break $U(4)_{\rm taste}$ (which is, of course, only a symmetry of the continuum limit). These operators were constructed in Refs. [16, 17] (see also Ref. [18]). There are no taste-breaking bilinears. This follows from the fact that any bilinear has to be invariant under the discrete group $\Gamma_{4}$, generated by the $\xi_{\mu}$. (There exist taste-invariant bilinear operators, such as $\overline{q}\,\hbox to0.0pt{\hbox{$\mskip 1.0mu/$}\hss}\\!D^{3}q$, see the end of this subsection.) Many 4-quark operators can be constructed, and if we require them to be invariant under $SU(N_{f})$, they take the form888They can be color-mixed or -unmixed. $(\overline{q}_{i}Xq_{i})(\overline{q}_{j}Yq_{j})\ ,$ (141) in which sums over $i$ and $j$ are implied, and where $X=\gamma_{A}\xi_{B}$ and $Y=\gamma_{C}\xi_{D}$, with $\displaystyle\gamma_{A}\ ,\gamma_{C}$ $\displaystyle\in$ $\displaystyle\\{1,\ \gamma_{\mu},\ i\gamma_{\mu}\gamma_{\nu},\ i\gamma_{5}\gamma_{\mu},\ \gamma_{5}\\}\ ,$ (142) $\displaystyle\xi_{B}\ ,\xi_{D}$ $\displaystyle\in$ $\displaystyle\\{1,\ \xi_{\mu},\ i\xi_{\mu}\xi_{\nu},\ i\xi_{5}\xi_{\mu},\ \xi_{5}\\}\ ,$ in which $\mu<\nu$. Since no mass matrix can appear in these operators, $U(N_{f})_{L}\times U(N_{f})_{R}$ symmetry implies that both $\gamma_{A}\xi_{B}$ and $\gamma_{C}\xi_{D}$ have to anti-commute with $\gamma_{5}\xi_{5}$, which, in turn, implies that $\displaystyle\gamma_{A}\in\\{1\ \mbox{(S)},\ i\gamma_{\mu}\gamma_{\nu}\ \mbox{(T)},\ \gamma_{5}\ \mbox{(P)}\\}\quad$ $\displaystyle\Rightarrow\quad\xi_{B}\in\\{\xi_{\mu}\ \mbox{(V)},\ i\xi_{5}\xi_{\mu}\ \mbox{(A)}\\}\ ,$ $\displaystyle\gamma_{A}\in\\{\gamma_{\mu}\ \mbox{(V)},\ i\gamma_{5}\gamma_{\mu}\ \mbox{(A)}\\}\quad$ $\displaystyle\Rightarrow\quad\xi_{B}\in\\{1\ \mbox{(S)},\ i\xi_{\mu}\xi_{\nu}\ \mbox{(T)},\ \xi_{5}\ \mbox{(P)}\\}\ .$ Here we denoted the various tensor structures with S (scalar), T (tensor), P (pseudo-scalar), V (vector) and A (axial-vector), for later convenience. Both bilinears thus have to be “odd,” which means that the spin-taste matrix has to have an odd number of Lorentz indices. Translating this back to lattice operators implies that the corresponding lattice bilinears have to be odd-link operators, since each $\gamma_{\mu}$ or $\xi_{\mu}$ is generated by a shift over one lattice spacing in the $\mu$ direction. Next, $\Gamma_{4}$ symmetry implies that $\xi_{B}=\xi_{D}$, because, if $\xi_{B}$ were not equal to $\xi_{D}$, we could find a $\xi_{X}$ such that $\xi_{B}\xi_{X}=-\xi_{X}\xi_{B}$ but $\xi_{D}\xi_{X}=+\xi_{X}\xi_{D}$, and the operator would not be invariant under $q\to\xi_{X}q$, $\overline{q}\to\overline{q}\xi_{X}$. Our operators thus have the form $(\overline{q}_{i}\gamma_{A}\xi_{B}q_{i})(\overline{q}_{j}\gamma_{C}\xi_{B}q_{j})\ ,$ (144) with the constraint (IV.2). A similar argument sets $\gamma_{A}=\gamma_{C}$. Rotations over 180 degrees take the form (cf. Eq. (60)) $q\to\gamma_{\rho}\gamma_{\sigma}\xi_{\rho}\xi_{\sigma}q\ ,\quad\rho\neq\sigma\ ,$ (145) and if we combine these with shifts in the $\rho$ and $\sigma$ directions, it follows that the SET has to be invariant under $q(x)\to\gamma_{\rho}\gamma_{\sigma}q(R^{-1}_{\rho\sigma}(\pi)x),\quad\rho\neq\sigma\ .$ (146) Likewise, we can combine $I_{s}$ of Eq. (70) with a shift in the 4th direction, resulting in a SET symmetry $q(x)\to\gamma_{4}q(I_{s}x)\ .$ (147) Together, these imply that $\gamma_{A}$ has to be equal to $\gamma_{C}$ (with Eq. (147) excluding $\gamma_{C}=\gamma_{A}\gamma_{5}$). Our 4-fermion operators thus take the form $(\overline{q}_{i}\gamma_{A}\xi_{B}q_{i})(\overline{q}_{j}\gamma_{A}\xi_{B}q_{j})\ ,$ (148) with the proviso that $\gamma_{A}\xi_{B}$ has to be odd. It remains to consider rotations over 90 degrees. In order to make sure our 4-fermion operators are invariant under these, we need to sum over Lorentz indices, to make all operators into scalars under 90-degree rotations. There will be two types of operators [16]: those operators which are invariant not only under hypercubic rotations, but also under $SO(4)$, which we will refer to as “type A,” and those which are not invariant under $SO(4)$, but only under hypercubic rotations, “type B.” To show how this works, it is easiest to give some examples. Examples of type-A operators are $\displaystyle\sum_{\mu}(\bar{\gamma}_{i}\xi_{\mu}q_{i})(\overline{q}_{j}\xi_{\mu}q_{j})\ ,$ (149) $\displaystyle\sum_{\mu,\nu\neq\rho}(\overline{q}_{i}\gamma_{\mu}\,i\xi_{\nu}\xi_{\rho}q_{i})(\overline{q}_{j}\gamma_{\mu}\,i\xi_{\nu}\xi_{\rho}q_{j})\ .$ These operators are both invariant under $SO(4)$. The first of these two operators we refer to as $S\times V$, since $\gamma_{A}$ is scalar and $\xi_{B}$ is vector; the second we refer to as $V\times T$, because $\gamma_{A}$ is vector and $\xi_{B}$ is tensor. An example of a type-B operator is $\sum_{\mu\neq\nu}(\overline{q}_{i}\gamma_{\mu}\,i\xi_{\mu}\xi_{\nu}q_{i})(\overline{q}_{j}\gamma_{\mu}\,i\xi_{\mu}\xi_{\nu}q_{j})\ .$ (150) Because the index $\mu$ is repeated four times, this operator is invariant under hypercubic rotations, but not under $SO(4)$. We may refer to this operator as $V_{\mu}\times T_{\mu}$, where the repeated $\mu$ reminds us of the four times repeated index $\mu$ in Eq. (150). This type can only occur if either $\gamma_{A}$ or $\xi_{B}$ is tensor (and thus the other is vector or axial-vector). In fact, the type-B operator of our example can be taken to be $\sum_{\mu\neq\nu}\left((\overline{q}_{i}\gamma_{\mu}\,i\xi_{\mu}\xi_{\nu}q_{i})(\overline{q}_{j}\gamma_{\mu}\,i\xi_{\mu}\xi_{\nu}q_{j})-(\overline{q}_{i}\gamma_{\mu}\,i\xi_{\mu}\xi_{\nu}\xi_{5}q_{i})(\overline{q}_{j}\gamma_{\mu}\,i\xi_{\mu}\xi_{\nu}\xi_{5}q_{j})\right)\ ,$ (151) because the orthogonal combination $\displaystyle\sum_{\mu\neq\nu}\left((\overline{q}_{i}\gamma_{\mu}\,i\xi_{\mu}\xi_{\nu}q_{i})(\overline{q}_{j}\gamma_{\mu}\,i\xi_{\mu}\xi_{\nu}q_{j})+(\overline{q}_{i}\gamma_{\mu}\,i\xi_{\mu}\xi_{\nu}\xi_{5}q_{i})(\overline{q}_{j}\gamma_{\mu}\,i\xi_{\mu}\xi_{\nu}\xi_{5}q_{j})\right)$ $\displaystyle=$ $\displaystyle\sum_{\mu\neq\nu}\left((\overline{q}_{i}\gamma_{\mu}\,i\xi_{\mu}\xi_{\nu}q_{i})(\overline{q}_{j}\gamma_{\mu}\,i\xi_{\mu}\xi_{\nu}q_{j})+(\overline{q}_{i}\gamma_{\mu}\,i\xi_{\rho}\xi_{\sigma}q_{i})(\overline{q}_{j}\gamma_{\mu}\,i\xi_{\rho}\xi_{\sigma}q_{j})_{\rho\sigma\neq\mu\nu}\right)$ $\displaystyle=$ $\displaystyle\sum_{\mu,\rho\neq\sigma}(\overline{q}_{i}\gamma_{\mu}\,i\xi_{\rho}\xi_{\sigma}q_{i})(\overline{q}_{j}\gamma_{\mu}\,i\xi_{\rho}\xi_{\sigma}q_{j})\ ,$ which is type A. We see that the requirement that the 4-quark operators are invariant under the staggered symmetry group leads to a huge reduction in the number of 4-quark operators. With $\gamma_{A}$, $\xi_{B}$, $\gamma_{C}$ and $\xi_{D}$ unrestricted, there are $16^{4}=65536$ choices! For $N_{f}=1$ there is a subtlety, based on the fact that the flavor symmetry group is just $U(1)\times U(1)_{\epsilon}$. It is straightforward to verify that, e.g., $(\overline{q}q)(\overline{q}q)-(\overline{q}\gamma_{5}\xi_{5}q)(\overline{q}\gamma_{5}\xi_{5}q)$ (153) is invariant under this group. Using our shorthand notation, we can label this operator as $S\times S-P\times P$. The fermion bilinears in this combination are even, and this operator thus does not fit into the general pattern we established above. However, it can be made to fit, by a Fierz transformation [16], as we will now show. A Fierz transformation is based on identities of the form $(\gamma_{A})_{\alpha\beta}(\gamma_{A})_{\gamma\delta}=\sum_{B}c(A,B)(\gamma_{B})_{\alpha\delta}(\gamma_{B})_{\gamma\beta}\ ,$ (154) with numerical coefficients $c(A,B)$ that depend on $A$ and $B$. For instance, $1_{\alpha\beta}1_{\gamma\delta}=\frac{1}{4}\sum_{B}(\gamma_{B})_{\alpha\delta}(\gamma_{B})_{\gamma\beta}\ .$ (155) Similar identities hold, of course, for the $\xi_{A}$. Using the abbreviations $S$, $V$, $T$, $A$ and $P$ introduced above, we can use Eq. (155) for both the $\gamma$ and $\xi$ matrices to write $\displaystyle S\times S$ $\displaystyle=$ $\displaystyle\frac{1}{16}(S+V+T+A+P)\times(S+V+T+A+P)\ ,$ (156) $\displaystyle P\times P$ $\displaystyle=$ $\displaystyle\frac{1}{16}(S-V+T-A+P)\times(S-V+T-A+P)\ .$ Subtracting these two, we find that $\displaystyle S\times S-P\times P=\frac{1}{8}\Bigl{(}S\times V+V\times S+T\times V+V\times T+P\times V+V\times P$ $\displaystyle\hskip 56.9055pt+S\times A+A\times S+T\times A+A\times T+P\times A+A\times P\Bigr{)}\ ,$ (157) which allows us to write the operator (153) in the general form we derived above for arbitrary $N_{f}$. Of course, Fierzing interchanges color-mixed and color-unmixed operators, but both types occur in the general analysis as well. There are also dimension-6 operators with just one quark bilinear. Staggered and flavor symmetries restrict these to the list999The operator $\sum_{\mu}\overline{q}_{i}(D_{\mu}^{2}\,\hbox to0.0pt{\hbox{$\mskip 1.0mu/$}\hss}\\!D-\,\hbox to0.0pt{\hbox{$\mskip 1.0mu/$}\hss}\\!DD_{\mu}^{2})q_{i}$ has negative charge parity [16]. $\displaystyle\overline{q}_{i}\,\hbox to0.0pt{\hbox{$\mskip 1.0mu/$}\hss}\\!D^{3}q_{i}\ ,\quad\sum_{\mu}\overline{q}_{i}(D_{\mu}^{2}\,\hbox to0.0pt{\hbox{$\mskip 1.0mu/$}\hss}\\!D+\,\hbox to0.0pt{\hbox{$\mskip 1.0mu/$}\hss}\\!DD_{\mu}^{2})q_{i}\ ,\quad\sum_{\mu}\overline{q}_{i}D_{\mu}\,\hbox to0.0pt{\hbox{$\mskip 1.0mu/$}\hss}\\!DD_{\mu}q_{i}\ ,\quad\sum_{\mu}\overline{q}_{i}\gamma_{\mu}D_{\mu}^{3}q_{i}\ ,$ $\displaystyle\overline{q}_{i}M\,\hbox to0.0pt{\hbox{$\mskip 1.0mu/$}\hss}\\!D^{2}q_{i}\ ,\quad\overline{q}_{i}MD_{\mu}^{2}q_{i}\ ,\quad\overline{q}_{i}M^{2}\,\hbox to0.0pt{\hbox{$\mskip 1.0mu/$}\hss}\\!Dq_{i}\ ,\quad\overline{q}_{i}M^{3}q_{i}\ .$ (158) In particular, the fermion bilinear cannot contain any taste matrix, because of shift symmetry. The operators containing the mass matrix $M$ can be constructed using Eqs. (IV.1) and (137). For example, the last operator is obtained from $\overline{q}_{Ri}{\cal M}{\cal M}^{\dagger}{\cal M}q_{Li}+\overline{q}_{Li}{\cal M}^{\dagger}{\cal M}{\cal M}^{\dagger}q_{Ri}\ ,$ (159) and setting ${\cal M}={\cal M}^{\dagger}=M$. Finally, there are purely gluonic dimension-6 operators, shared with other discretizations of QCD, which we do not list here. ### IV.3 Operators with dimension larger than 6 We briefly discuss what happens beyond dimension 6, i.e., at order $a^{3}$ and higher. #### IV.3.1 Dimension 7 Let us start with dimension-7 operators. We again restrict ourselves to operators containing quark fields (no purely gluonic operators with odd dimension exist). Dimension-7 operators, if they exist, can contain one or two quark bilinears. First, consider operators with only one quark bilinear. Starting from the list in Eq. (IV.2), we need to add one mass dimension, which can be done by adding an extra covariant derivative $D_{\mu}$, or an extra mass matrix $M$. However, if we insert a $D_{\mu}$, Lorentz indices still have to balance, and the only way to do that is to remove or add a $\gamma_{\mu}$ inside the bilinear. But this would change the properties of this bilinear under $U(N_{f})_{L}\times U(N_{f})_{R}$, so this does not work. An analogous argument excludes adding a mass matrix $M$ (using again a spurion-based argument). Essentially the same argument applies to operators with two quark bilinears. Inserting a $D_{\mu}$ leads to an unmatched Lorentz index, or a change in the symmetry properties of one fermion bilinear, making it inconsistent with $U(N_{f})_{L}\times U(N_{f})_{R}$. A similar argument applies with inserting a mass matrix $M$. If we replace one of the two bilinears $\overline{q}_{i}Xq_{i}$ with $\overline{q}_{i}MYq_{i}$, we need to choose $Y$ such that $\overline{q}_{i}MYq_{i}$ is invariant under $U(N_{f})_{L}\times U(N_{f})_{R}$ (after promoting $M$ to the spurions ${\cal M}$ and ${\cal M}^{\dagger}$). But this implies that $Y\neq X$, and thus shift and/or lattice rotation/reflection symmetries would be broken. We conclude that no dimension-7 operators exist that are part of the SET. #### IV.3.2 Dimension 8 In contrast, many operators of dimension 8 exist, and they are easily constructed from dimension-6 operators, e.g., by inserting $D_{\mu}^{2}$ or $\hbox to0.0pt{\hbox{$\mskip 1.0mu/$}\hss}\\!D^{2}$, or by multiplying with ${\rm tr}\,(M^{2})$. However, all continuum symmetries are already broken down to the staggered symmetry group at the level of dimension-6 operators (because of type-B operators!), so no new symmetry breaking occurs at order $a^{4}$ [19].101010In particular, no new taste spurions occur, because the number of quark bilinears is still restriced to two. See next section. #### IV.3.3 Odd dimension $\geq 9$ At dimension 9, we should consider operators with three quark bilinears. Because of dimensional reasons, these cannot contain any derivatives or masses. Each of the bilinears needs to be invariant under $U(N_{f})_{L}\times U(N_{f})_{R}$, which implies that each bilinear has to be odd—it has an odd number of Lorentz indices. But this is impossible, because all Lorentz indices need to be contracted because of hypercubic invariance. In general, any operator, at any dimension, has to have an even number of Lorentz indices. What about dimension-9 operators with only one quark bilinear? Either the bilinear is odd, which means it has 1 or 3 Lorentz indices, or, if it contains a mass matrix $M$, it carries an even number of Lorentz indices, because of $U(N_{f})_{L}\times U(N_{f})_{R}$ symmetry. (If there are two mass matrix insertions, it has to be odd again; if there are three mass matrices, it has to be even again, etc.) The rest of the operator has to be made out of $k$ appearances of $D_{\mu}$ and/or $\ell$ appearances of $G_{\kappa\lambda}$. This part of the operator thus has dimension $d=k+2\ell$, and it has $k+2\ell$ mod 2 Lorentz indices. If the quark bilinear has dimension 3 (no $M$), the complete operator thus has dimension $d=3+k+2\ell$, and it has $1+k+2\ell$ mod 2 Lorentz indices. For odd $d$, we thus need $k+2\ell$ even, but that implies that the number of Lorentz indices is odd. Likewise, if the bilinear contains one factor of $M$, we need $d=4+k+2\ell$ to be odd, and thus that $k+2\ell$ is odd. But now the quark bilinear needs to be even, so again the number of Lorentz indices is odd. This argument generalizes to operators with more insertions of $M$. A very similar argument takes care of dimension-9 operators with two quark bilinears. The general pattern is that the dimension of the gluonic part to be added is always such that the number of Lorentz indices is odd. The conclusion appears to be that no operators with odd dimension, and thus an odd power of the lattice spacing, can appear in the SET. ### IV.4 What causes taste breaking? Finally, in this chapter, let us have a closer look at what causes taste- breaking operators to appear in the SET. As they first appear at dimension 6 through 4-quark operators, we expect them to come from quark-quark scattering, as depicted in Fig. 2. Indeed, when the gluon momentum exchanged between the two quarks has momentum $ak=\pi_{E}\neq 0$, with $\pi_{E}$ one of the momentum vectors with components equal to 0 or $\pi$, this short-distance gluon exchange generates taste-breaking 4-quark operators, as we will show below. Note that in this figure, we restored the lattice spacing $a$. $ak=\pi_{E}$$a\tilde{p}^{\prime}+\pi_{C}$$a\tilde{q}^{\prime}+\pi_{D}$$a\tilde{p}+\pi_{A}$$a\tilde{q}+\pi_{B}$ Figure 2: Exchange of a gluon with momentum equal to $\pi_{E}$ between two staggered quarks, in momentum space. In Feynman gauge, the gluon propagator $D(k)$ in Eq. (III.2), for such momenta, is equal to $D_{\mu\nu}(k=\pi_{E}/a)=\frac{\delta_{\mu\nu}}{\sum_{\mu}\frac{4}{a^{2}}\sin^{2}({1\over 2}ak_{\mu})}\Bigg{|}_{k=\pi_{E}/a}=\frac{a^{2}}{4n}\,\delta_{\mu\nu}\ ,$ (160) where $n$ is the number of components of $\pi_{E}$ equal to $\pi$. Using Eq. (53), and working in Feynman gauge, the diagram is thus equal to (retaining the explicit lattice spacing $a$ only in the prefactor and omitting color generators) $\displaystyle-g^{2}\frac{a^{2}}{4n}\sum_{\mu}\delta(\tilde{p}+\pi_{A}-\tilde{p}^{\prime}+\pi_{C}+\pi_{E}+\pi_{\eta_{\mu}}){1\over 2}(S_{\mu}^{A}e^{i\tilde{p}_{\mu}}+S_{\mu}^{C}e^{-i\tilde{p}^{\prime}_{\mu}})$ $\displaystyle\phantom{-g^{2}\frac{a^{2}}{4n}\sum_{\mu}}\times\delta(\tilde{q}+\pi_{B}-\tilde{q}^{\prime}+\pi_{D}+\pi_{E}+\pi_{\eta_{\mu}}){1\over 2}(S_{\mu}^{B}e^{i\tilde{q}_{\mu}}+S_{\mu}^{D}e^{-i\tilde{q}^{\prime}_{\mu}})$ $\displaystyle=$ $\displaystyle-g^{2}\frac{a^{2}}{4n}\sum_{\mu}\delta(\tilde{p}-\tilde{p}^{\prime})\delta(\tilde{q}-\tilde{q}^{\prime}){1\over 2}(S_{\mu}^{A}e^{i\tilde{p}_{\mu}}+S_{\mu}^{C}e^{-i\tilde{p}_{\mu}}){1\over 2}(S_{\mu}^{B}e^{i\tilde{q}_{\mu}}+S_{\mu}^{D}e^{-i\tilde{q}_{\mu}})$ $\displaystyle\phantom{-g^{2}\frac{a^{2}}{4n}\sum_{\mu}}\times\overline{\delta}(\pi_{A}+\pi_{C}+\pi_{E}+\pi_{\eta_{\mu}})\overline{\delta}(\pi_{B}+\pi_{D}+\pi_{E}+\pi_{\eta_{\mu}})$ $\displaystyle\approx$ $\displaystyle-g^{2}\frac{a^{2}}{4n}\frac{1}{4}\sum_{\mu}\delta(\tilde{p}-\tilde{p}^{\prime})\delta(\tilde{q}-\tilde{q}^{\prime})\Biggl{(}(\hat{\pi}_{E}\Gamma_{\mu})_{CA}(\hat{\pi}_{E}\Gamma_{\mu})_{DB}+(\Gamma_{\mu}\hat{\pi}_{E})_{CA}(\Gamma_{\mu}\hat{\pi}_{E})_{DB}$ $\displaystyle\phantom{-g^{2}\frac{a^{2}}{4n}\frac{1}{4}\sum_{\mu}}+(\hat{\pi}_{E}\Gamma_{\mu})_{CA}(\Gamma_{\mu}\hat{\pi}_{E})_{DB}+(\Gamma_{\mu}\hat{\pi}_{E})_{CA}(\hat{\pi}_{E}\Gamma_{\mu})_{DB}\Biggr{)}\ ,$ where in the last expression we approximated $e^{i\tilde{p}_{\mu}}\approx 1$, etc., and we used $\overline{\delta}(\pi_{A}+\pi_{C}+\pi_{E}+\pi_{\eta_{\mu}})=\sum_{F}\overline{\delta}(\pi_{E}+\pi_{C}+\pi_{F})\overline{\delta}(\pi_{\eta_{\mu}}+\pi_{F}+\pi_{A})=(\hat{\pi}_{E}\hat{\pi}_{\eta_{\mu}})_{CA}=(\hat{\pi}_{\eta_{\mu}}\hat{\pi}_{E})_{CA}$ (162) and likewise for $\overline{\delta}(\pi_{B}+\pi_{D}+\pi_{E}+\pi_{\eta_{\mu}})$. (We recall that $(\pi_{\eta_{\mu}})_{\mu}=0$.) We now use that $\hat{\pi}_{E}\in\\{\Gamma_{\mu}\Xi_{\mu},\ \Gamma_{\mu}\Gamma_{\nu}\Xi_{\mu}\Xi_{\nu},\ \Gamma_{\mu}\Gamma_{5}\Xi_{\mu}\Xi_{5},\ \Gamma_{5}\Xi_{5}\\}\ .$ (163) This is easy to see, because in these combinations, the factors $S_{\mu}$ in Eqs. (35) and (39) cancel, so $\Gamma_{\mu}\Xi_{\mu}=\hat{\pi}_{\eta_{\mu}}\hat{\pi}_{\zeta_{\mu}}$, etc., and hence the matrices in Eq. (163) generate “pure” $\hat{\pi}$ matrices, a total of 15 of them. Equation (IV.4) thus generates the 4-quark operators with spin-taste structure $\displaystyle\hat{\pi}_{E}=$ $\displaystyle\Gamma_{\mu}\Xi_{\mu}:\qquad$ $\displaystyle S\times V\ ,$ $\displaystyle\Gamma_{\nu}\Xi_{\nu}|_{\nu\neq\mu}:\qquad$ $\displaystyle T_{\nu}\times V_{\nu}\ ,$ $\displaystyle\Gamma_{\mu}\Gamma_{\nu}\Xi_{\mu}\Xi_{\nu}:\qquad$ $\displaystyle V_{\nu}\times T_{\nu}\ ,$ $\displaystyle\Gamma_{\rho}\Gamma_{\sigma}\Xi_{\rho}\Xi_{\sigma}|_{\rho,\sigma\neq\mu}:\qquad$ $\displaystyle A_{\nu}\times T_{\rho\sigma}\ (\nu\neq\rho,\sigma)\ ,$ $\displaystyle\Gamma_{\mu}\Gamma_{5}\Xi_{\mu}\Xi_{5}:\qquad$ $\displaystyle P\times A\ ,$ $\displaystyle\Gamma_{\nu}\Gamma_{5}\Xi_{\nu}\Xi_{5}|_{\nu\neq\mu}:\qquad$ $\displaystyle T_{\rho\sigma}\times A_{\nu}\ (\nu\neq\rho,\sigma)\ ,$ $\displaystyle\Gamma_{5}\Xi_{5}:\qquad$ $\displaystyle A\times P\ .$ As expected, these are all odd operators, because the $U(N_{f})_{L}\times U(N_{f})_{R}$ symmetry is respected by the kinetic term in the staggered action from which the Feynman rule used above derives. We note that we do not obtain all 4-quark operators from this diagram, but other processes, for instance with more than one gluon exchanged, can occur. The scale of the coupling at the quark-gluon vertex in Fig. 2 is that of the exchanged gluon, and we thus expect taste breaking to be suppressed by a factor of order $\alpha_{s}(\pi/a)$ (with $\alpha_{s}$ the strong coupling). This already follows from the fact that the free theory has exact $U(4)$ taste symmetry, so internal gluons are needed to produce taste-breaking operators. ## V Staggered chiral perturbation theory The continuum limit of QCD with $N_{f}$ staggered quarks has $4N_{f}$ quarks in the continuum limit, with flavor symmetry group $U(1)\times SU(4N_{f})_{L}\times SU(4N_{f})_{R}$. It follows that this theory is expected to have $(4N_{f})^{2}-1$ Nambu–Goldstone bosons (NGBs). Of course, the number of NGBs is too large, and this is why, in practical numerical computations, the 4th root of the staggered determinant is taken, in order to reduce the number of quark degrees of freedom by a factor of 4.111111The staggered tastes can also be interpreted as physical flavors, and an action can be constructed that completely breaks the taste degeneracy in the continuum limit [8]. However, this construction leads to a practical problems, and has not been used in practice. While we will not review the arguments supporting the “4th- root trick,” we can accommodate this trick in both WCPT and chiral perturbation theory (ChPT) by employing the replica trick. (For a review of the rooting trick and further references, see Ref. [20]. See also Ref. [15].) We introduce $n_{r}$ replicas of each staggered fermion, so that now we have $N=4n_{r}N_{f}$ quarks in the continuum limit, and we thus expect $(4n_{r}N_{f})^{2}-1$ NGBs. Setting, at the end of any calculation (in WCPT or ChPT) $n_{r}=\frac{1}{4}$ thus yields effectively $N_{f}^{2}-1$ NGBs, i.e., the desired number. We will thus develop staggered ChPT (SChPT) for $N$ continuum quarks, and see in an example how the replica trick works. Of course, if we wish to consider unrooted staggered quarks, we would set $n_{r}=1$. ### V.1 Construction of the SChPT lagrangian The chiral lagrangian is, like the SET, a continuum effective theory, with lattice spacing dependence taken into account in terms of an expansion in powers of $a^{2}$. It is formulated in terms of a non-linear field $\Sigma=e^{\frac{i}{f}\pi}\ ,$ (165) where $\pi$ is a traceless, hermitian $4n_{r}N_{f}\times 4n_{r}N_{f}$ matrix describing the $(4n_{r}N_{f})^{2}-1$ NGBs. The field $\Sigma$ transforms as $\Sigma_{ia,jb}\sim q_{Lia}\overline{q}_{Rjb}\ ,$ (166) with $i,\ j=1,\dots,n_{r}N_{f}$ replica-flavor indices, and $a,\ b=1,\dots,4$ taste indices. In this chapter, the left-handed and right-handed projections are defined as in continuum QCD: $\displaystyle q_{L}\\!\\!\\!$ $\displaystyle=$ $\displaystyle\\!\\!\\!{1\over 2}(1-\gamma_{5})q\ ,\qquad q_{R}={1\over 2}(1+\gamma_{5})q\ ,$ (167) $\displaystyle\overline{q}_{L}\\!\\!\\!$ $\displaystyle=$ $\displaystyle\\!\\!\\!\overline{q}{1\over 2}(1+\gamma_{5})\ ,\qquad\overline{q}_{R}=\overline{q}{1\over 2}(1-\gamma_{5})\ .$ We embed the taste matrices $\xi_{A}$ into $4n_{r}N_{f}\times 4n_{r}N_{f}$ matrices by simply repeating the $4\times 4$ matrices $\xi_{A}$ $n_{r}N_{f}$ times along the main diagonal, thus constructing a block-diagonal matrix $\xi_{A}^{(n_{r}N_{F})}$. However, we will drop the superscript $(n_{r}N_{F})$ below, when no confusion is possible. For instance, since $\Sigma$ is a $4n_{r}N_{f}\times 4n_{r}N_{f}$ matrix, we have that $\xi_{A}\Sigma\equiv\xi_{A}^{(n_{r}N_{F})}\Sigma\ .$ (168) For reviews of continuum ChPT and applications to various discretizations of QCD, see Refs. [21, 22]. At leading order (LO) the chiral lagrangian is given by ${\cal L}_{LO}=\frac{1}{8}f^{2}\,{\rm tr}\,(\partial_{\mu}\Sigma^{\dagger}\partial_{\mu}\Sigma)-\frac{1}{4}Bf^{2}\,{\rm tr}\,({\cal M}^{\dagger}\Sigma+{\cal M}\Sigma^{\dagger})\ .$ (169) The trace is over replica/flavor and taste indices (cf. Eq. (166)). This lagrangian is invariant under $\Sigma\to V_{L}\Sigma V^{\dagger}_{R}\ ,\qquad{\cal M}\to V_{L}{\cal M}V_{R}^{\dagger}\ ,$ (170) with now $V_{L}\in SU(N)_{L}$ and $V_{R}\in SU(N)_{R}$, with $N=4n_{r}N_{f}$. As before, ${\cal M}$ is a mass spurion, to be set equal to the quark mass matrix $M$ after the construction of the chiral lagrangian. $f$ and $B$ are low-energy constants (LECs), to be determined by matching with the underlying theory, QCD.121212$f$, of course, is the pion decay constant, normalized such that at LO, $f_{\pi}=f=130.4$ MeV. We will choose the quark mass matrix $M$ to be diagonal and positive. The first four diagonal elements will be equal to the up quark mass $m_{u}$, the next four will be equal to the down quark mass $m_{d}$, etc. (As before, the four tastes produced by each staggered fermion are degenerate in the continuum limit.) To incorporate terms of order $a^{2}$ into ChPT, we use a spurion trick just like for the mass term in Eq. (169). It is easiest to explain this through some examples. Our first example starts from the following operator appearing in the SET lagrangian: $a^{2}(\overline{q}\xi_{5\nu}q)(\overline{q}\xi_{5\nu}q)=a^{2}(\overline{q}_{R}\xi_{5\nu}q_{L}+\overline{q}_{L}\xi_{5\nu}q_{R})^{2}\to(\overline{q}_{R}X_{R}q_{L}+\overline{q}_{L}X_{L}q_{R})^{2}\ ,$ (171) where $\xi_{5\nu}=i\xi_{5}\xi_{\nu}$ and $X_{L,R}$ are new spurions transforming as $X_{L}\to V_{L}X_{L}V^{\dagger}_{R}\ ,\qquad X_{R}\to V_{R}X_{R}V^{\dagger}_{L}\ .$ (172) We can take $X_{R}=X_{L}^{\dagger}$, but this is not necessary. To recover the 4-quark operator we started from, we have to set $X_{L}=X_{R}=a\xi_{5\nu}\ .$ (173) The next step is simply to write down all operators involving the non-linear field $\Sigma$ and two spurions $X_{L}$ and/or $X_{R}$, such that these operators are invariant under Eqs. (170) and (172). The possibilities are $\displaystyle{\rm tr}\,(X_{L}\Sigma^{\dagger})\,{\rm tr}\,(X_{R}\Sigma)\ ,$ (174) $\displaystyle\left({\rm tr}\,(X_{L}\Sigma^{\dagger})\right)^{2}+\left({\rm tr}\,(X_{R}\Sigma)\right)^{2}\ ,$ $\displaystyle{\rm tr}\,(X_{L}\Sigma^{\dagger}X_{L}\Sigma^{\dagger})+{\rm tr}\,(X_{R}\Sigma X_{R}\Sigma)\ .$ The plus signs follow from parity symmetry (Eq. (173) requires parity to exchange $X_{L}$ and $X_{R}$).131313We note that parity $q(x)\to\gamma_{4}\xi_{4}q(I_{s}x)$ is a symmetry of the SET. Setting the spurions equal to their physical values, Eq. (173), yields the operators $\displaystyle a^{2}{\rm tr}\,(\xi_{5\nu}\Sigma^{\dagger})\,{\rm tr}\,(\xi_{5\nu}\Sigma)\ ,$ (175) $\displaystyle a^{2}\left({\rm tr}\,(\xi_{5\nu}\Sigma^{\dagger})\right)^{2}+a^{2}\left({\rm tr}\,(\xi_{5\nu}\Sigma)\right)^{2}\ ,$ $\displaystyle a^{2}{\rm tr}\,(\xi_{5\nu}\Sigma^{\dagger}\xi_{5\nu}\Sigma^{\dagger})+a^{2}{\rm tr}\,(\xi_{5\nu}\Sigma\xi_{5\nu}\Sigma)\ .$ Each of these operators should be added to Eq. (169) with an arbitary new coefficient, thus introducing new LECs into the chiral lagrangian. Before we go on to our next example, we note that we could proceed slightly differently. Instead of the spurions $X_{L,R}$ we could first write $a^{2}(\overline{q}\xi_{5\nu}q)(\overline{q}\xi_{5\nu}q)=a^{2}\left((\overline{q}_{R}\xi_{5\nu}q_{L})^{2}+(\overline{q}_{R}\xi_{5\nu}q_{L})(\overline{q}_{L}\xi_{5\nu}q_{R})+(\overline{q}_{L}\xi_{5\nu}q_{R})^{2}\right)\ ,$ (176) and introduce ${\cal O}(a^{2})$ spurions $\displaystyle X_{LL}$ $\displaystyle\to$ $\displaystyle a^{2}\xi_{5\nu}\times\xi_{5\nu}\ ,$ (177) $\displaystyle X_{LR}$ $\displaystyle\to$ $\displaystyle a^{2}\xi_{5\nu}\times\xi_{5\nu}\ ,$ $\displaystyle X_{RR}$ $\displaystyle\to$ $\displaystyle a^{2}\xi_{5\nu}\times\xi_{5\nu}\ ,$ for each of the three terms on the right-hand side of Eq. (176), thus working only with spurions of order $a^{2}$. The resulting terms in the chiral lagrangian will be the same, so instead we will use the simpler ${\cal O}(a)$ spurions above, with the extra rule that all terms need to contain two of the spurions $X_{L}$ and $X_{R}$. Our second example starts from $a^{2}(\overline{q}\gamma_{\mu}\xi_{5}q)(\overline{q}\gamma_{\mu}\xi_{5}q)=a^{2}(\overline{q}_{L}\gamma_{\mu}\xi_{5}q_{L}+\overline{q}_{R}\gamma_{\mu}\xi_{5}q_{R})^{2}\to((\overline{q}_{L}\gamma_{\mu}Y_{L}q_{L}+\overline{q}_{R}\gamma_{\mu}Y_{R}q_{R})^{2}\ ,$ (178) where we introduced two new spurions $Y_{L}$ and $Y_{R}$ transforming as $Y_{L}\to V_{L}Y_{L}V_{L}^{\dagger}\ ,\qquad Y_{R}\to V_{R}Y_{R}V_{R}^{\dagger}\ .$ (179) Their physical values are $Y_{L}=Y_{R}=a\xi_{5}\ .$ (180) Again, we need the rule that any term in the chiral lagrangian should contain two $Y$ spurions, conforming with Eq. (178). The only operator that can be constructed from $\Sigma$ and two $Y$-spurions is ${\rm tr}\,(Y_{L}\Sigma Y_{R}\Sigma^{\dagger})\to a^{2}{\rm tr}\,(\xi_{5}\Sigma\xi_{5}\Sigma^{\dagger})\ ,$ (181) which thus introduces one new LEC into the chiral lagrangian. Performing this procedure for all dimension-6 operators we constructed in Sec. IV.2 leads to an order-$a^{2}$ “potential” $a^{2}{\cal V}=a^{2}({\cal U}+{\cal U}^{\prime})$ (182) to be added to Eq. (169), with [16, 18] $\displaystyle-{\cal U}$ $\displaystyle=$ $\displaystyle C_{1}\,{\rm tr}\,(\xi_{5}\Sigma\xi_{5}\Sigma^{\dagger})+{1\over 2}C_{3}\sum_{\nu}\left({\rm tr}\,(\xi_{\nu}\Sigma\xi_{\nu}\Sigma)+\mbox{h.c.}\right)$ $\displaystyle+{1\over 2}C_{4}\sum_{\nu}\left({\rm tr}\,(\xi_{5\nu}\Sigma\xi_{5\nu}\Sigma)+\mbox{h.c.}\right)+C_{6}\sum_{\mu<\nu}{\rm tr}\,(\xi_{\mu\nu}\Sigma\xi_{\mu\nu}\Sigma^{\dagger})\ ,$ $\displaystyle-{\cal U}^{\prime}$ $\displaystyle=$ $\displaystyle\frac{1}{4}C_{2V}\sum_{\nu}\left(({\rm tr}\,(\xi_{\nu}\Sigma))^{2}+\mbox{h.c.}\right)+\frac{1}{4}C_{2A}\sum_{\nu}\left(({\rm tr}\,(\xi_{5\nu}\Sigma))^{2}++\mbox{h.c.}\right)$ $\displaystyle+{1\over 2}C_{5V}\sum_{\nu}{\rm tr}\,(\xi_{\nu}\Sigma)\,{\rm tr}\,(\xi_{\nu}\Sigma^{\dagger})+{1\over 2}C_{5A}\sum_{\nu}{\rm tr}\,(\xi_{5\nu}\Sigma)\,{\rm tr}\,(\xi_{\nu}\Sigma^{\dagger})\ ,$ in which $\xi_{\mu\nu}=i\xi_{\mu}\xi_{\nu}$. We find that the operators $S\times V$, $S\times A$, $P\times V$, $P\times A$, $T\times V$, and $T\times A$ are like example 1 above, and contribute to the LECs $C_{3}$, $C_{4}$, $C_{2V,A}$ and $C_{5V,A}$. The operators $V\times P$, $A\times P$, $V\times T$, and $A\times T$ are like example 2 above, and contribute to $C_{1}$ and $C_{6}$. The operators $V\times S$ and $A\times S$ do not break taste symmetry, and thus do not contribute any new operators to the chiral lagrangian. However, they do lead to order-$a^{2}$ corrections to LECs already present in the continuum lagrangian. The lagrangian obtained by adding ${\cal V}$ to Eq. (169) defines staggered ChPT (SChPT) at lowest-order. Operators like $V_{\mu}\times T_{\mu}$, i.e., operators of type B, do contribute at order $a^{2}$ to the SET, but not at this order to SChPT. The reason is that operators that break $SO(4)$ (while preserving hypercubic rotations) can occur only at higher orders in SChPT. For example one can make a hypercubic invariant $\sim\sum_{\mu}\partial_{\mu}^{4}$ which is not an $SO(4)$ invariant. Such an operator is $O(a^{2}p^{4})$ in ChPT power counting (the $a^{2}$ has to be present because the continuum theory respects $SO(4)$ rotational invariance), and thus always of higher order than the LO lagrangian. This has an important consequence: To order $a^{2}$, the physics of NGBs is invariant under the $SO(4)_{\rm taste}$ subgroup of $SU(4)_{\rm taste}$, and not just under the smaller staggered symmetry group [16]! Other examples are operators of the schematic form $\sim\sum_{\mu}\partial_{\mu}^{2}\xi_{\mu}\xi_{\mu}$ or $\sum_{\mu}\xi_{\mu}\xi_{\mu}\xi_{\mu}\xi_{\mu}$, which also respect only the staggered hypercubic invariance. We will discuss such examples toward the end of this subsection; such operators are also higher order than the LO lagrangian. This point brings us to the issue of power counting. SChPT has three small parameters, the typical momentum of physical low-energy pions, the quark masses in $M_{ia,jb}=m_{i}\delta_{ij}\delta_{ab}$, and the parameter $a^{2}$. Since for physical momenta, $p^{2}=-m_{\pi}^{2}$, and, as we will see shortly, $m_{\pi}^{2}\sim m_{i}$, the first two expansion parameters are related.141414However, for an alternative scenario, which appears not to be realized in nature, see Ref. [23]. But $a^{2}$ and the $m_{i}$ are in principle independent, and it depends on the actual numerical computations carried out in lattice QCD which power counting is most appropriate. Most staggered computations have NGB (squared-mass) taste splittings comparable in size to the NGB (squared) masses. We will thus assume a power counting $a^{2}\Lambda_{\rm QCD}^{2}\sim\frac{m_{i}}{\Lambda_{\rm QCD}}\sim\frac{p^{2}}{\Lambda_{\rm QCD}^{2}}\ .$ (184) Of course, if the lattice spacing is small enough, it may be possible to ignore lattice spacing artifacts, in which case one can use continuum ChPT. On the other hand, if the lattice spacing is too large, SChPT may break down, even if the masses $m_{i}$ are small enough. We can now look at the NGB masses predicted by LO SChPT. We write the pion matrix in Eq. (165) as $\pi_{ij}=\sum_{A}\xi_{A}\pi^{A}_{ij}\ ,\qquad\xi_{A}\in\\{\xi_{5},\ \xi_{5\mu},\ \xi_{\mu\nu},\ \xi_{\mu},\ 1\\}$ (185) and expand the LO lagrangian to second order in the pion fields. This allows us to read off the masses at lowest order. For $i\neq j$ we find $m_{Aij}^{2}=B(m_{i}+m_{j})+a^{2}\Delta(\xi_{A})\ ,$ (186) with $\displaystyle\Delta(\xi_{5})$ $\displaystyle=$ $\displaystyle 0\ ,$ (187) $\displaystyle\Delta(\xi_{5\mu})$ $\displaystyle=$ $\displaystyle\frac{16}{f^{2}}(C_{1}+3C_{3}+C_{4}+3C_{6})\ ,$ $\displaystyle\Delta(\xi_{\mu\nu})$ $\displaystyle=$ $\displaystyle\frac{16}{f^{2}}(2C_{3}+2C_{4}+4C_{6})\ ,$ $\displaystyle\Delta(\xi_{\mu})$ $\displaystyle=$ $\displaystyle\frac{16}{f^{2}}(C_{1}+C_{3}+3C_{4}+3C_{6})\ ,$ $\displaystyle\Delta(1)$ $\displaystyle=$ $\displaystyle\frac{16}{f^{2}}(4C_{3}+4C_{4})\ .$ We see that there is a degeneracy predicted by the accidental $SO(4)$ symmetry that is present in the chiral lagrangian at LO, i.e., to order $a^{2}$. In the next section, we will see that the flavor non-diagonal NGBs form eight different multiplets (instead of the five seen in Eq. (187)) if taste symmetry is fully broken. We also note that the $\xi_{5}$ NGBs are massless when all quark masses vanish. This is the NGB associated with the exactly conserved axial currents for the $U(N_{f})_{L}\times U(N_{f})_{R}$ symmetry group of Eqs. (IV.1) and (136). The LECs in the double-trace part of the potential, ${\cal U}^{\prime}$, contribute only to flavor-diagonal NGB masses. To quadratic order, ${\cal U}^{\prime}_{\rm quad}=\frac{32}{f^{2}}(C_{2V}-C_{5V})\sum_{ij}\pi^{V}_{ii}\pi^{V}_{jj}+\frac{32}{f^{2}}(C_{2A}-C_{5A})\sum_{ij}\pi^{A}_{ii}\pi^{A}_{jj}\ ,$ (188) leading to corresponding contributions to (the squares of) their masses. It is important to note that, even if one restricts oneself to external pions that are flavor off-diagonal, flavor diagonal pions still contribute to loops. For examples, see Ref. [18]. If any of the $\Delta(\xi_{A})$ would be negative, this could induce flavor/taste symmetry breaking when the quark masses $m_{i}$ are small enough. However, QCD inequalities (see e.g. Ref. [24]) prevent this from happening. There is, however, no argument that the combinations $C_{2V,A}-C_{5V,A}$ cannot be negative. For further discussion, see Ref. [25]. In practice no spontaneous breaking of flavor/taste symmetries has been observed. At the next order in SChPT, $SO(4)_{\rm taste}$ breaks down to $\Gamma_{4}$, and rotational symmetry breaks down to the hypercubic group. We give a few examples [19]. A single insertion of a 4-quark operator can lead to a ChPT operator of the form $a^{2}\sum_{\mu}{\rm tr}\,(\Sigma\partial_{\mu}\Sigma^{\dagger}\xi_{\mu}\Sigma^{\dagger}\partial_{\mu}\Sigma\xi_{\mu})=\frac{4a^{2}}{f^{2}}\sum_{\mu}{\rm tr}\,(\partial_{\mu}\pi\xi_{\mu}\partial_{\mu}\pi\xi_{\mu})+\dots\ ,$ (189) in which the first $\xi_{\mu}$ comes from a spurion like $X_{L}$ and the second from a spurion like $X_{R}$. This operator leads to direction dependence of the dispersion relation. This contribution appears at order $a^{2}p^{2}$, i.e., at next-to-leading order (NLO). Another example comes from a double insertion of two 4-quark operators, and takes the form $a^{4}\sum_{\mu}\sum_{\nu\neq\mu}{\rm tr}\,(\xi_{\mu\nu}\Sigma\xi_{\mu}\Sigma\xi_{\mu\nu}\Sigma^{\dagger}\xi_{\mu}\Sigma^{\dagger})\ ,$ (190) with the taste matrices coming from spurions $Y_{L}$, $X_{R}$, etc. Because of the four spurions, this contribution appears at order $a^{4}$, which in our power counting is NLO. At tree level, it only contributes to processes with four or more pions. This is easy to see: if we set the last $\Sigma^{\dagger}$ equal to one, we can use $\xi_{\mu}\xi_{\mu\nu}=i\xi_{\nu}$ to write the operator as $a^{4}\sum_{\mu}\sum_{\nu\neq\mu}i\,{\rm tr}\,(\xi_{\nu}\Sigma\xi_{\mu}\Sigma\xi_{\mu\nu}\Sigma^{\dagger})\ ,$ (191) which has $SO(4)$ taste symmetry. Similar simplifications happen when one of the other non-linear fields is set equal to one. At next-to-next-to-leading order (NNLO) this operator does contribute to two-point functions through loops. For a complete construction of the NLO SChPT lagrangian, see Ref. [19]. Finally, in this section, we discuss whether the accidental $SO(4)$ symmetry at order $a^{2}$ also occurs for other hadrons than pions. As an example, we consider taste breaking for the $\rho$ meson [16], introducing a “heavy” $\rho$ field $\rho_{\mu}$ [26]. In the infinite-mass limit, its 4-velocity $v_{\mu}$ is fixed. We first choose $v_{\mu}$ to be a spurion, transforming like a vector under hypercubic rotations, and once the effective lagrangian for the $\rho$ mass has been constructed, we set $v_{4}=1$ and $\vec{v}=0$, choosing the $\rho$’s restframe. Using this and staggered symmetries, we can write down the most general quadratic term in the field $\rho_{\mu ij}=\sum_{A}\xi_{A}\rho^{A}_{\mu ij}$: $\displaystyle{\cal L}_{\rho\rm-mass}$ $\displaystyle=$ $\displaystyle R_{1}\sum_{\mu}{\rm tr}\,(\rho_{\mu}^{\dagger}\rho_{\mu})+R_{2}\sum_{\mu}{\rm tr}\,(\rho_{\mu}^{\dagger}\xi_{5}\rho_{\mu}\xi_{5})+R_{3}\sum_{\mu\nu}{\rm tr}\,(\rho_{\mu}^{\dagger}\xi_{\nu}\rho_{\mu}\xi_{\nu})$ $\displaystyle+$ $\displaystyle R_{4}\sum_{\mu\nu}{\rm tr}\,(\rho_{\mu}^{\dagger}\xi_{5\nu}\rho_{\mu}\xi_{5\nu})+R_{5}\sum_{\mu\ \nu<\kappa}{\rm tr}\,(\rho_{\mu}^{\dagger}\xi_{\nu\kappa}\rho_{\mu}\xi_{\nu\kappa})+R_{6}\sum_{\mu}{\rm tr}\,(\rho_{\mu}^{\dagger}\xi_{\mu}\rho_{\mu}\xi_{\mu})$ $\displaystyle+$ $\displaystyle R_{7}\sum_{\mu}{\rm tr}\,(\rho_{\mu}^{\dagger}\xi_{5\mu}\rho_{\mu}\xi_{5\mu})+R_{8}\sum_{\mu\neq\nu}{\rm tr}\,(\rho_{\mu}^{\dagger}\xi_{\mu\nu}\rho_{\mu}\xi_{\mu\nu})+R_{9}\sum_{\mu}{\rm tr}\,(\rho_{\mu}^{\dagger}\xi_{4}\rho_{\mu}\xi_{4})$ $\displaystyle+$ $\displaystyle R_{10}\sum_{\mu}{\rm tr}\,(\rho_{\mu}^{\dagger}\xi_{45}\rho_{\mu}\xi_{45})+R_{11}\sum_{\mu\nu}{\rm tr}\,(\rho_{\mu}^{\dagger}\xi_{4\nu}\rho_{\mu}\xi_{4\nu})+R_{12}\sum_{\mu}{\rm tr}\,(\rho_{\mu}^{\dagger}\xi_{4\mu}\rho_{\mu}\xi_{4\mu})\ .$ For example, the term with $R_{9}$ is obtained from $\sum_{\mu\nu}{\rm tr}\,(\rho_{\mu}^{\dagger}\xi_{\nu}\rho_{\mu}\xi_{\nu})v_{\nu}v_{\nu}\big{|}_{v=(0,0,0,1)}=\sum_{\mu}{\rm tr}\,(\rho_{\mu}^{\dagger}\xi_{4}\rho_{\mu}\xi_{4})\ .$ (193) There are 11 different quark bilinear operators that create a $\rho$ with the form $\displaystyle\overline{q}\gamma_{k}\xi_{A}q\ ,$ (194) $\displaystyle\xi_{A}=(\xi_{5},\xi_{m},\xi_{4k},\xi_{m5},\xi_{45},\xi_{k5},\xi_{m4},\xi_{k4},\xi_{km},\xi_{\ell m})\ ,$ where $k\neq\ell\neq m\neq k$ [27]. Each of these transforms in an irrep of the relevant staggered symmetry group (see next section). Thus, the 11 operators with LECs $R_{2}$ to $R_{12}$ in Eq. (V.1) completely break the degeneracy (the operator with LEC $R_{1}$ gives the same mass to all tastes).151515It appears that the term with LEC $R_{12}$ was missed in Ref. [16]. ### V.2 Example: Electromagnetic current correlator at NLO As an example of SChPT at work, we will calculate the electromagnetic (EM) current two-point function to one loop, which is the leading order for this correlation function in ChPT. We usually refer to this as NLO, since it involves a loop, but there is no tree-level contribution, as we will see. We will use SChPT with two flavors, and $n_{r}$ replicas. This theory has $4n_{r}$ up quarks, and $4n_{r}$ down quarks in the continuum limit; we will consider the isospin-symmetric case with $m_{u}=m_{d}$. At the end of the calculation, we will set $n_{r}=1/4$, thus implementing the rooting trick. In the continuum limit, this will yield the expected result, which, at this order, is that one would obtain in scalar QED (where the complex scalar is the charged pion). This validates, in this example, the rooting trick. In two-flavor QCD, the continuum EM current for $n_{r}$ replicas of two light flavors is $j_{\mu}=\sum_{i}\left(\frac{2}{3}\,\overline{u}_{i}\gamma_{\mu}u_{i}-\frac{1}{3}\,\overline{d}_{i}\gamma_{\mu}d_{i}\right)\ ,$ (195) where the index $i$ is now the replica index. To couple SChPT to electromagnetism, we introduce the covariant derivative $D_{\mu}\Sigma=\partial_{\mu}\Sigma-i\ell_{\mu}\Sigma+i\Sigma r_{\mu}\ ,$ (196) where $\ell_{\mu}$ and $r_{\mu}$ are external vector fields gauging the groups $U(N)_{L}$ and $U(N)_{R}$, respectively, with $N=8n_{r}$ (since have now specialized to $N_{f}=2$). To obtain the coupling of the pions to the EM field, we take $\ell_{\mu}=r_{\mu}=Qv_{\mu}$, with $Q=\left(\begin{array}[]{cc}\frac{2}{3}I&0\\\ 0&-\frac{1}{3}I\end{array}\right)\ ,$ (197) with $I$ the $4n_{r}\times 4n_{r}$ unit matrix. We thus have $\displaystyle D_{\mu}\Sigma$ $\displaystyle=$ $\displaystyle\partial_{\mu}\Sigma-iv_{\mu}[Q,\Sigma]\ ,$ (198) $\displaystyle(D_{\mu}\Sigma)^{\dagger}$ $\displaystyle=$ $\displaystyle\partial_{\mu}\Sigma^{\dagger}-iv_{\mu}[Q,\Sigma^{\dagger}]\ .$ The kinetic term in the chiral lagrangian becomes ${\cal L}_{\rm kin}=\frac{f^{2}}{8}{\rm tr}\,\left((D_{\mu}\Sigma)^{\dagger}D_{\mu}\Sigma\right),$ (199) and the ChPT form of the EM current (to lowest order) is obtained from the terms linear in $v_{\mu}$: $\displaystyle j_{\mu}$ $\displaystyle=$ $\displaystyle\frac{f^{2}}{8}\,{\rm tr}\,\left(-i[Q,\Sigma^{\dagger}]\partial_{\mu}\Sigma-i\partial_{\mu}\Sigma^{\dagger}[Q,\Sigma]\right)$ $\displaystyle=$ $\displaystyle-\frac{i}{4}\,{\rm tr}\,\left(Q(\pi\partial_{\mu}\pi-\partial_{\mu}\pi\pi)\right)+\dots\ ,$ where in the second line we expanded $\Sigma=1+\frac{i}{f}\pi+\dots$. Writing the pion fields as $\pi=\left(\begin{array}[]{cc}\frac{1}{\sqrt{2}}\pi^{0}&\pi^{+}\\\ \pi^{-}&-\frac{1}{\sqrt{2}}\pi^{0}\end{array}\right)\ ,$ (201) and using $\pi^{\pm}_{ij}=\sum_{A}\xi_{A}\pi^{\pm A}_{ij}$, this can be written as $j_{\mu}=i\left(\pi^{-A}_{ij}\partial_{\mu}\pi^{+A}_{ji}-\pi^{+A}_{ij}\partial_{\mu}\pi^{-A}_{ji}\right)\ .$ (202) It is straightforward to check that this current is conserved by using the equation of motion $\raisebox{-1.72218pt}{\large$\Box$}\pi^{\pm A}_{ij}=m_{A}^{2}\pi^{\pm A}_{ij}$. This current corresponds to the conserved staggered current $\displaystyle J_{\mu}^{\rm cons}$ $\displaystyle=$ $\displaystyle{1\over 2}\frac{2}{3}\left(\bar{\chi}_{u}(x)\eta_{\mu}(x)U_{\mu}(x)\chi_{u}(x+\mu)+\bar{\chi}_{u}(x+\mu)\eta_{\mu}(x)U^{\dagger}_{\mu}(x)\chi_{u}(x)\right)$ $\displaystyle-{1\over 2}\frac{1}{3}\left(\bar{\chi}_{d}(x)\eta_{\mu}(x)U_{\mu}(x)\chi_{d}(x+\mu)+\bar{\chi}_{d}(x+\mu)\eta_{\mu}(x)U^{\dagger}_{\mu}(x)\chi_{d}(x)\right)\ .$ The conserved staggered current is not the only discretization of the EM current we may consider. Another example, which is not conserved on the lattice, is the local current $J_{\mu}^{\rm local}=\frac{2}{3}\,\bar{\chi}_{u}(x)\eta_{\mu}(x)\zeta_{\mu}(x)\chi_{u}(x)-\frac{1}{3}\,\bar{\chi}_{d}(x)\eta_{\mu}(x)\zeta_{\mu}(x)\chi_{d}(x)\ .$ (204) In Eqs. (V.2) and (204) we left the replica index implicit. The continuum limit of $J_{\mu}^{\rm local}$ is161616We use $J$ for lattice currents, and $j$ for continuum currents. $j_{\mu}^{\rm local}=\frac{2}{3}\,\overline{u}_{i}\gamma_{\mu}\xi_{\mu}u_{i}-\frac{1}{3}\,\overline{d}_{i}\gamma_{\mu}\xi_{\mu}d_{i}=\overline{q}\gamma_{\mu}Q\xi_{\mu}^{(2n_{r})}q\ ,$ (205) where in the second step we grouped the $u_{i}$ and $d_{i}$ quarks together into the field $q$. To translate this current into ChPT, we set $\ell_{\mu}=r_{\mu}=Q\xi_{\mu}^{(2n_{r})}v_{\mu}$. In general, we can define “tasteful” currents (as opposed to the conserved current, which is tasteless) by generalizing this to $j_{\mu}^{A}=\frac{2}{3}\,\overline{u}_{i}\gamma_{\mu}\xi_{A}u_{i}-\frac{1}{3}\,\overline{d}_{i}\gamma_{\mu}\xi_{A}d_{i}=\overline{q}\gamma_{\mu}Q\xi_{A}^{(2n_{r})}q\ ,$ (206) with the local current corresponding to $\xi_{A}=\xi_{\mu}$, and the conserved current corresponding to $\xi_{A}=1$. In SChPT, the corresponding currents are $\displaystyle j_{\mu}^{A}$ $\displaystyle=$ $\displaystyle-i\,\frac{f^{2}}{4}\,{\rm tr}\,\left(Q\xi_{A}^{(2n_{r})}[\Sigma^{\dagger},\partial_{\mu}\Sigma]\right)$ $\displaystyle=$ $\displaystyle-\frac{i}{4}\,{\rm tr}\,\left(Q\xi_{A}^{(2n_{r})}[\pi,\partial_{\mu}\pi]\right)+\dots\ ,$ where in the second line we again expanded in the pion fields. Instead of Eq. (201), we will now write $\pi=\left(\begin{array}[]{cc}U&\pi^{+}\\\ \pi^{-}&D\end{array}\right)\ ,$ (208) which, as we will see, is convenient. In terms of the fields introduced in Eq. (208), $j_{\mu}^{A}=-\frac{i}{6}\,{\rm tr}\,\left(\xi_{A}\left(U{\overleftrightarrow{\partial}_{\\!\\!\\!\mu}}U+\pi^{+}{\overleftrightarrow{\partial}_{\\!\\!\\!\mu}}\pi^{-}\right)\right)+\frac{i}{12}\,{\rm tr}\,\left(\xi_{A}\left(\pi^{-}{\overleftrightarrow{\partial}_{\\!\\!\\!\mu}}\pi^{+}+D{\overleftrightarrow{\partial}_{\\!\\!\\!\mu}}D\right)\right)\ .$ (209) We used the notation $f{\overleftrightarrow{\partial}_{\\!\\!\\!\mu}}g=f\partial_{\mu}g-(\partial_{\mu}f)g\ .$ (210) Note that in these expressions, we have replaced $\xi_{A}^{(n_{r})}$ by $\xi_{A}$, for simplicity. If $\xi_{A}=1$, we have that ${\rm tr}\,(U{\overleftrightarrow{\partial}_{\\!\\!\\!\mu}}U)=0\ ,\qquad{\rm tr}\,(D{\overleftrightarrow{\partial}_{\\!\\!\\!\mu}}D)=0\ ,\qquad{\rm tr}\,(\pi^{-}{\overleftrightarrow{\partial}_{\\!\\!\\!\mu}}\pi^{+})=-{\rm tr}\,(\pi^{+}{\overleftrightarrow{\partial}_{\\!\\!\\!\mu}}\pi^{-})\ ,$ (211) and we recover Eq. (202). If $\xi_{A}\neq 1$, ${\rm tr}\,(\xi_{A}U{\overleftrightarrow{\partial}_{\\!\\!\\!\mu}}U)=\sum_{BC}{\rm tr}\,(\xi_{A}[U^{B}\xi_{B},\partial_{\mu}U^{C}\xi_{C}]=\sum_{BC}{\rm tr}\,(\xi_{A}[\xi_{B},\xi_{C}])U^{B}_{ij}\partial_{\mu}U^{C}_{ji}\neq 0$ (212) in general, so the terms with $U$ and $D$ in Eq. (209) do not vanish. However, if $\xi_{B}=1$, the commutator $[\xi_{B},\xi_{C}]=[1,\xi_{C}]=0$, and thus the contribution with $U^{B}=U^{S}$ (with $U^{S}$ the coefficient of $1$ in $U=\sum_{B}\xi_{B}U^{B}$) does in fact vanish, and hence the neutral singlet fields $U^{S}$ and $D^{S}$ do not contribute to the current (209). Since there is no mixing between any of the $U^{A}$ and any of the $D^{B}$ for $A,\ B\neq S$, we can work on the basis of Eq. (208). We have that $\displaystyle\partial_{\mu}j_{\mu}^{A}$ $\displaystyle=$ $\displaystyle-\frac{i}{6}\sum_{BC}{\rm tr}\,(\xi_{A}[\xi_{B},\xi_{C}])U^{B}_{ij}m_{\pi C}^{2}U^{C}_{ji}+\dots$ $\displaystyle=$ $\displaystyle-\frac{i}{12}\sum_{BC}{\rm tr}\,(\xi_{A}[\xi_{B},\xi_{C}])U^{B}_{ij}(m_{\pi C}^{2}-m_{\pi B}^{2})U^{C}_{ji}+\dots\neq 0\ ,$ we see that indeed the current $j_{\mu}^{A}$ is also not conserved in SChPT, unless $\xi_{A}=1$. Of course, in the continuum limit, where all pions become degenerate become equal, all currents are conserved. We note that $Q={1\over 2}\tau_{3}+\frac{1}{6}Y={1\over 2}\left(\begin{array}[]{cc}1&0\\\ 0&-1\end{array}\right)+\frac{1}{6}\left(\begin{array}[]{cc}1&0\\\ 0&1\end{array}\right)\ .$ (214) Since only the commutator $[Q,\Sigma]$ appears in the conserved current, the hypercharge part $Y$ drops out, and thus in SChPT, the conserved EM current is the isospin-1 current. For the other currents, with a non-trivial $\xi_{A}$, the situation is different, as we will discuss shortly. First, let us return to the $\xi_{A}=1$ current $j_{\mu}^{S}=j_{\mu}$, in the continuum, at the quark level. With $\langle\dots\rangle_{U_{\mu}}$ denoting the average over the gauge fields $\langle(\overline{q}(x)\gamma_{\mu}{1\over 2}\tau_{3}q(x))(\overline{q}(y)\gamma_{\nu}{1\over 2}\tau_{3}q(y))\rangle=-{1\over 2}\langle{\rm tr}\,(\gamma_{\mu}S(x,y;U_{\mu})\gamma_{\nu}S(y,x;U_{\mu})\rangle_{U_{\mu}}\ ,$ (215) when $m_{u}=m_{d}$, because any quark-disconnected parts cancel. Here $S(x,y;U_{\mu})$ is the quark propagator (up or down) in a gauge-field background. The trace is over spin, taste and replica indices. We also have $\langle(\overline{q}(x)\gamma_{\mu}{1\over 2}\tau_{3}q(x))(\overline{q}(y)\gamma_{\nu}\frac{1}{6}Yq(y))\rangle=0\ ,$ (216) and, finally, $\displaystyle\langle(\overline{q}(x)\gamma_{\mu}\frac{1}{6}Yq(x))(\overline{q}(y)\gamma_{\nu}\frac{1}{6}Yq(y))\rangle$ $\displaystyle=$ $\displaystyle-\frac{1}{18}\langle{\rm tr}\,(\gamma_{\mu}S(x,y;U_{\mu})\gamma_{\nu}S(y,x;U_{\mu})\rangle_{U_{\mu}}$ $\displaystyle+\frac{1}{9}\langle{\rm tr}\,(\gamma_{\mu}S(x,x;U_{\mu}))\,{\rm tr}\,(\gamma_{\nu}S(y,y;U_{\mu}))\rangle_{U_{\mu}}\ .$ The hypercharge current two-point function is the sum of a quark-connected (first line) and quark-disconnected (second line) part. It follows that the light-quark connected part of the EM current two-point function is equal to $-\frac{5}{9}\langle{\rm tr}\,(\gamma_{\mu}S(x,y;U)\gamma_{\nu}S(y,x;U)\rangle_{U}=\frac{10}{9}\times\langle(\overline{q}(x)\gamma_{\mu}{1\over 2}\tau_{3}q(x))(\overline{q}(y)\gamma_{\nu}{1\over 2}\tau_{3}q(y))\rangle\ ,$ (218) i.e., the light-quark connected part of the EM current two-point function is equal to $10/9$ times the $I=1$ current two-point function. Now consider a similar exercise for the current $j_{\mu}^{A}$ with $\xi_{A}\neq 1$. In this case, the relations (215) and (V.2) hold, with a $\xi_{A}$ inserted next to every $\gamma_{\mu}$ and $\gamma_{\nu}$. For this current, there is no quark-disconnected part, because ${\rm tr}\,(\xi_{A})=0$! Therefore the $j_{\mu}^{A}$ two-point function is equal to its light-quark connected part, and, in the continuum limit, the light-quark connected parts of all currents, including the one with $\xi_{A}$, are equal. Therefore, in the continuum limit, $\langle j_{\mu}^{A}(x)j_{\nu}^{A}(y)\rangle\big{|}_{A\neq S}=\frac{10}{9}\langle j_{\mu}^{S}(x)j_{\nu}^{S}(y)\rangle\ .$ (219) We thus expect to reproduce this relation in the continuum limit of SChPT as well. Next, we carry out the actual calculation, beginning with the $U$ part of the current (209). As we will be interested in $C^{A}(t)=\frac{1}{3}\sum_{i=1}^{3}\sum_{\vec{x}}\langle j_{i}^{A}(\vec{x},t)j_{i}^{A}(0)\rangle\ ,$ (220) we calculate, using $U_{k\ell}(x)=\sum_{X}\xi_{X}U^{X}_{k\ell}(x)$ (where $k$ and $\ell$ are replica indices), $\displaystyle\langle j_{\mu}^{A}(x)j_{\nu}^{A}(y)\rangle_{U\mbox{\tiny- part}}$ $\displaystyle=$ $\displaystyle-\frac{1}{36}\sum_{XX^{\prime}YY^{\prime}}\langle(U^{X}_{mn}(x){\overleftrightarrow{\partial}_{\\!\\!\\!\mu}}U^{X^{\prime}}_{nm}(x))(U^{Y}_{k\ell}(y){\overleftrightarrow{\partial}_{\\!\\!\\!\mu}}U^{Y^{\prime}}_{\ell k}(y))\rangle$ $\displaystyle\phantom{-\frac{1}{36}\sum_{XX^{\prime}YY^{\prime}}}\times\,{\rm tr}\,(\xi_{A}\xi_{X}\xi_{X^{\prime}})\,{\rm tr}\,(\xi_{A}\xi_{Y}\xi_{Y^{\prime}})$ $\displaystyle=$ $\displaystyle-\frac{n_{r}^{2}}{36}\int\frac{dp_{4}}{2\pi}\int\frac{dq_{4}}{2\pi}\frac{1}{V}^{2}\sum_{\vec{p},\vec{q}}\sum_{XX^{\prime}}\frac{e^{ip(x-y)}e^{iq(y-x)}}{(p^{2}+m_{X}^{2})(q^{2}+m_{X^{\prime}}^{2})}(p_{\mu}+q_{\mu})^{2}$ $\displaystyle\phantom{-}\times\Bigl{(}{\rm tr}\,(\xi_{A}\xi_{X}\xi_{X^{\prime}})\,{\rm tr}\,(\xi_{A}\xi_{X}\xi_{X^{\prime}})-{\rm tr}\,(\xi_{A}\xi_{X}\xi_{X^{\prime}})\,{\rm tr}\,(\xi_{A}\xi_{X^{\prime}}\xi_{X})\Bigr{)}\ ,$ where $V$ is the spatial volume (we take the time direction infinite), and $\langle U^{X}_{mn}(x)U_{k\ell}^{Y}(y)\rangle=\delta_{XY}\delta_{m\ell}\delta_{nk}\int\frac{dp_{4}}{2\pi}\frac{1}{V}\sum_{\vec{p}}\frac{e^{ip(x-y)}}{p^{2}+m_{X}^{2}}\ .$ (222) Setting $\mu=i$, summing over $\vec{x}$, $\vec{y}$ and $i$, and using $\displaystyle\int\frac{dp_{4}}{2\pi}\frac{e^{ip_{4}t}}{p^{2}+m_{X}^{2}}$ $\displaystyle=$ $\displaystyle\frac{1}{2E_{X}(\vec{p})}\,e^{-E_{X}(\vec{p})|t|}\ ,$ (223) $\displaystyle E_{X}(\vec{p})$ $\displaystyle=$ $\displaystyle\sqrt{{\vec{p}}^{2}+m_{X}^{2}}\ ,$ we obtain $\displaystyle 3C^{A}(t)_{U\mbox{\tiny-part}}$ $\displaystyle=$ $\displaystyle-\frac{n_{r}^{2}}{36V}\sum_{\vec{p}}\sum_{XX^{\prime}}\frac{e^{-(E_{X}(\vec{p})+E_{X^{\prime}}(\vec{p}))|t|}}{E_{X}(\vec{p})E_{X^{\prime}}(\vec{p})}\,{\vec{p}}^{2}$ $\displaystyle\phantom{-}\times\Bigl{(}{\rm tr}\,(\xi_{A}\xi_{X}\xi_{X^{\prime}})\,{\rm tr}\,(\xi_{A}\xi_{X}\xi_{X^{\prime}})-{\rm tr}\,(\xi_{A}\xi_{X}\xi_{X^{\prime}})\,{\rm tr}\,(\xi_{A}\xi_{X^{\prime}}\xi_{X})\Bigr{)}\ .$ We will refer to the first double trace over taste matrices as the “unmixed” part, and to the second double trace as the “mixed” part. We note that for $\xi_{A}=1$ this vanishes, consistent with the fact that there is no $U$ contribution to the conserved current. For the $D$ contribution, we get $1/4$ of this, and thus the sum of the $U$ and $D$ contributions is equal to $36\times\left(1/36+1/144\right)=5/4$ times Eq. (V.2). It remains to add the contribution from the charged-pion part of the current. One can verify that choosing the term $\pi^{+}{\overleftrightarrow{\partial}_{\\!\\!\\!\mu}}\pi^{-}$ in both currents yields 1 times the mixed term in Eq. (V.2), choosing the term $\pi^{-}{\overleftrightarrow{\partial}_{\\!\\!\\!\mu}}\pi^{+}$ in both currents yields $1/4$ times the mixed term in Eq. (V.2), while the cross terms yields $-1$ times the unmixed term in Eq. (V.2). Adding all contributions leads to $\displaystyle 3C^{A}(t)$ $\displaystyle=$ $\displaystyle\frac{n_{r}^{2}}{144V}\sum_{\vec{p}}\sum_{XX^{\prime}}\frac{e^{-(E_{X}(\vec{p})+E_{X^{\prime}}(\vec{p}))|t|}}{E_{X}(\vec{p})E_{X^{\prime}}(\vec{p})}\,{\vec{p}}^{2}$ $\displaystyle\times\Bigl{(}10{\rm tr}\,(\xi_{A}\xi_{X}\xi_{X^{\prime}})\,{\rm tr}\,(\xi_{A}\xi_{X^{\prime}}\xi_{X})-{\rm tr}\,(\xi_{A}\xi_{X}\xi_{X^{\prime}})\,{\rm tr}\,(\xi_{A}\xi_{X}\xi_{X^{\prime}})\Bigr{)}\ .$ For $\xi_{A}=1$ the traces set $X^{\prime}=X$, and setting all masses $m_{X}$ equal (i.e., taking the continuum limit), the factor on the second line of this result is equal to $9\times 4^{2}\times 16=144\times 16$ (a 4 for each trace; the 16 comes from the $\sum_{X}$). With $n_{r}=1/4$, the total prefactor becomes equal to one, which is what one would have obtained directly in the continuum. For $a>0$, the taste pion $\pi_{X}$ runs around the loop, and these loops are averaged over all tastes. When $\xi_{A}\neq 1$, a little more work is needed. First, we have that ${\rm tr}\,(\xi_{A}\xi_{X}\xi_{X^{\prime}})={\rm tr}\,((\xi_{A}\xi_{X}\xi_{X^{\prime}})^{T})={\rm tr}\,(\xi_{X^{\prime}}^{T}\xi_{X}^{T}\xi_{A}^{T})={\rm tr}\,(\xi_{A}\xi_{X^{\prime}}\xi_{X}^{\prime})^{*}\ ,$ (226) and thus ${\rm tr}\,(\xi_{A}\xi_{X}\xi_{X^{\prime}})\,{\rm tr}\,(\xi_{A}\xi_{X^{\prime}}\xi_{X})=16\delta_{X^{\prime},X+A}\ .$ (227) This takes care of the mixed trace in Eq. (V.2). We still get the factor 16, and the current inserts taste $A$ into the pion loop. For the unmixed trace, we use that $\xi_{A}\xi_{X}=(-1)^{|A||X|+A\cdot X}\xi_{X}\xi_{A}\ ,$ (228) with $|X|=\sum_{\mu}X_{\mu}$ and $A\cdot X=\sum_{\mu}A_{\mu}X_{\mu}$. Thus $({\rm tr}\,(\xi_{A}\xi_{X}\xi_{X^{\prime}}))^{2}=16(-1)^{|A||X|+A\cdot X}\delta_{X^{\prime},X+A}\ .$ (229) For $\xi_{A}=1$ the right-hand side is equal to $\delta_{X,X^{\prime}}$, as before. For any $\xi_{A}\neq 1$, $(-1)^{|A||X|+A\cdot X}=+1$ for half of the tastes $X$, and $-1$ for the other half. Thus, when all masses become degenerate in the continuum limit, the trace term on the second line of Eq. (V.2) becomes equal to 11 when $(-1)^{|A||X|+A\cdot X}=-1$, and 9 when $(-1)^{|A||X|+A\cdot X}=+1$. The total prefactor thus becomes $\frac{n_{r}^{2}}{144}\times 4^{2}\times 8\times(11+9)=\frac{10}{9}\times 16n_{r}^{2}\ ,$ (230) which is $10/9$ times the continuum limit for $\xi_{A}=1$, as we expect from Eq. (219). Here the factors 4 again come from the traces, and the factor $8\times(11+9)$ from the $\sum_{X}$. Again, the current $j_{\mu}^{A}$ inserts a taste $A$ into the pion loop. This verifies Eq. (219) at this order SChPT, in the continuum limit. For $a>0$, the sums in Eq. (V.2) have to be carried out to obtain the complete result in explicit form. The Kronecker delta in Eq. (227) reduces the sum over $X$ and $X^{\prime}$ to a single sum. ## VI Staggered states and operators In this chapter, we first consider the spectrum of QCD with staggered fermions and general properties of staggered correlation functions, after which we discuss the construction of operators that create and annihilate states in the staggered Hilbert space. Most of the chapter will deal with mesons, but in the final section we will briefly discuss aspects of baryonic states in QCD with staggered fermions. The material of the first three sections can be found in Refs. [27, 28, 29], while the material of the final section can be found in Refs. [28, 30]. Another useful reference is Ref. [31]. ### VI.1 Staggered spectrum: states and operators We will consider states in momentum space, because the irreducible representations of a theory with staggered fermions live in momentum space. Operators acting on the staggered Hilbert space will be denoted with a hat. Thus, $\hat{S}_{\mu}$ is the shift operator acting on the Hilbert space. Defining $\hat{T}_{\mu}=\hat{S}_{\mu}^{2}$, $\hat{T}_{\mu}$ is a normal translation over two lattice spacings in the $\mu$ direction. In momentum space, the eigenvalues are $\hat{T}_{k}:\quad e^{2ip_{k}}\ ,\qquad\hat{T}_{4}:\quad e^{-2E}\ ,$ (231) where $p_{k}$ with $-\pi/2<p_{k}\leq\pi/2$ are the components of the physical spatial momentum of a state, and $E$ is the energy. The operator $\hat{T}_{4}$ can be taken as the transfer matrix of the theory. With Eq. (231) we can define the operators $\hat{T}_{\mu}^{-{1\over 2}}$ as having eigenvalues $e^{-ip_{k}}$ and $e^{E}$, and thus we can define $\hat{\Xi}_{\mu}=\hat{S}_{\mu}\hat{T}_{\mu}^{-{1\over 2}}\ ,$ (232) with $\hat{\Xi}_{\mu}^{2}=1\ .$ (233) We will consider states with $\vec{p}=0$ (except in Sec. VI.3). Such states transform in representations of the geometrical restframe group ($GRF$) $GRF=G(\Xi_{\mu},R^{(k\ell)},I_{s})\ ,$ (234) where the spatial rotations $R^{(k\ell)}=R^{(k\ell)}(\pi/2)$ generate the little group of $\vec{p}=0$ states, and we recall that $I_{s}=I_{1}I_{2}I_{3}=\Gamma_{4}\Xi_{4}$ in the defining representation. The notation $G(X)$ denotes the finite group generated by the elements $X$. Since $\Xi_{4}\in GRF$, also parity $P=I_{s}\Xi_{4}=\Gamma_{4}\in GRF$, and, with $P$ commuting with all other elements of the group, $GRF=G(\Xi_{\mu},R^{(k\ell)})\times\\{1,\ P\\}\ .$ (235) A maximally commuting set of elements of $GRF$ is the set $\\{\Xi_{4},\ \Xi_{1}\Xi_{2},\ R^{(12)},\ P\\}$. We will label states in the Hilbert space by the eigenvalues $E$, the eigenvalue $\sigma$ of $P$, and the eigenvalue $\sigma_{t}$ of $\Xi_{4}$: $\displaystyle\hat{S}_{4}|Er\sigma;\sigma_{t}\rangle$ $\displaystyle=$ $\displaystyle\sigma_{t}e^{-E}|Er\sigma;\sigma_{t}\rangle\ ,$ (236) $\displaystyle\Xi_{4}|Er\sigma;\sigma_{t}\rangle$ $\displaystyle=$ $\displaystyle\sigma_{t}|Er\sigma;\sigma_{t}\rangle\ ,$ $\displaystyle P|Er\sigma;\sigma_{t}\rangle$ $\displaystyle=$ $\displaystyle\sigma|Er\sigma;\sigma_{t}\rangle\ ,$ $\displaystyle I_{s}|Er\sigma;\sigma_{t}\rangle$ $\displaystyle=$ $\displaystyle\sigma\sigma_{t}|Er\sigma;\sigma_{t}\rangle\equiv\sigma_{s}|Er\sigma;\sigma_{t}\rangle\ .$ The label $r$ denotes an irrep of $G(\Xi_{\mu},R^{(k\ell)})$; within such an irrep, we can label the state with the eigenvalue $\sigma_{t}$ of $\Xi_{4}$. The final equation defines the parity $\sigma_{s}$ of $I_{s}$. Next, we introduce local operators $\hat{\phi}$ and $\hat{\overline{\phi}}$ at fixed time $t$, which thus transform in representations of a smaller group, the geometric time-slice group $GTS=G(\Xi_{m},R^{(k\ell)},I_{s})\ ,$ (237) which does not contain $\Xi_{4}$. We define time-dependent operators (i.e., we use the Heisenberg picture) through $\hat{\phi}(t)=\hat{S}_{4}^{-t}\hat{\phi}\hat{S}_{4}^{t}\ ,\qquad\hat{\overline{\phi}}(t)=\hat{S}_{4}^{-t}\hat{\overline{\phi}}\hat{S}_{4}^{t}\ .$ (238) There is a $U(1)$ charge operator $\hat{Q}$ (quark number) with $[\hat{Q},\hat{\phi}]=q\hat{\phi}\ ,\qquad[\hat{Q},\hat{\overline{\phi}}]=-q\hat{\overline{\phi}}\ ,$ (239) with $q$ the charge of $\hat{\phi}$ (we can take $q=1$ for the operator $\hat{\chi}$), and a charge conjugation operator $\hat{C}_{0}$ with $\hat{C}_{0}^{-1}\hat{\phi}\,\hat{C}_{0}=\tau\hat{\overline{\phi}}\ ,\qquad\hat{C}_{0}^{-1}\hat{\overline{\phi}}\,\hat{C}_{0}=\overline{\tau}\hat{\phi}\ .$ (240) The phases $\tau$ and $\overline{\tau}$ depend on convention, but the phase $\sigma_{c}=\tau\overline{\tau}$ defined by $\hat{C}_{0}^{-2}\hat{\phi}\,\hat{C}_{0}^{2}=\sigma_{c}\hat{\phi}$ (241) is not. Using the definition of $C_{0}$ in Sec. II.4 we find that $\sigma_{c}=-1$ for the staggered quark field $\chi$, and thus that it is equal to $+1$ for mesonic operators, and $-1$ for baryonic operators. The fields $\hat{\phi}$ and $\hat{\overline{\phi}}$ transform in a representation $r$ of $GTS$, which is isomorphic to $G(\Xi_{\mu},R^{(k\ell)})$ and can thus be taken to be the same irrep as in Eq. (236). We note that $GRF=GTS\times\\{1,\ P\\}$, but the fields $\hat{\phi}$ and $\hat{\overline{\phi}}$ do not have $P$ quantum numbers, since the definition of $P$ involves a translation in time. We can consider operators $\phi$ and $\overline{\phi}$ with a well-defined value $\sigma_{s}$ of $I_{s}$: $\hat{I}_{s}^{-1}\hat{\phi}\hat{I}_{s}=\sigma_{s}\hat{\phi}\ ,\qquad\hat{I}_{s}^{-1}\hat{\overline{\phi}}\hat{I}_{s}=\sigma_{s}\hat{\overline{\phi}}\ .$ (242) ($\hat{\phi}$ and $\hat{\overline{\phi}}$ have the same $I_{s}$ quantum number because $\hat{I}_{s}$ commutes with $\hat{C}_{0}$.) If $\hat{\phi}$, $\hat{\overline{\phi}}$ have the $I_{s}$ quantum number $\sigma_{s}$, we have that $\langle Er\sigma;\sigma_{t}|\hat{\overline{\phi}}|0\rangle=0\qquad\mbox{unless}\ \sigma_{s}=\sigma\sigma_{t}\ ,$ (243) because $\hat{I}_{s}=\hat{P}\hat{\Xi}_{4}$. An example of $\hat{\phi}$ is $\sum_{\vec{m}}\hat{\chi}(\vec{x}+2\vec{m})\ ,$ (244) which transforms in an 8-dimensional irrep of $GTS$. This irrep decomposes as ${\bf 8}\to A_{1}^{+}+A_{1}^{-}+F_{1}^{+}+F_{1}^{-}\ ,$ (245) where $A_{1}$ is the trivial irrep of the cubic group and $F_{1}$ the vector irrep. The superscript denotes the value of $I_{s}$. These irreps of the cubic group can be projected out as [28] $\displaystyle A_{1}^{+}:\qquad$ $\displaystyle\sum_{\vec{x}\ {\rm even}}\hat{\chi}(\vec{x})\ ,$ $\displaystyle A_{1}^{-}:\qquad$ $\displaystyle\sum_{\vec{x}\ {\rm even}}\hat{\chi}(\vec{x}+1+2+3)\ ,$ $\displaystyle F_{1}^{-}:\qquad$ $\displaystyle\sum_{\vec{x}\ {\rm even}}\left(\begin{array}[]{c}\hat{\chi}(\vec{x}+1)\\\ \hat{\chi}(\vec{x}+2)\\\ \hat{\chi}(\vec{x}+3)\end{array}\right)\ ,$ (250) $\displaystyle F_{1}^{+}:\qquad$ $\displaystyle\sum_{\vec{x}\ {\rm even}}\left(\begin{array}[]{c}\hat{\chi}(\vec{x}+2+3)\\\ -\hat{\chi}(\vec{x}+1+3)\\\ \hat{\chi}(\vec{x}+1+2)\end{array}\right)\ .$ (254) The appearance of the irreps $A_{1}$ and $F_{1}$ for a fermionic field can be understood as follows. In the rest frame, the relevant continuum group is $SU(2)_{\rm spin}\times SU(2)_{\rm taste}$. The group $G(R^{(k\ell)})$ is a discrete subgroup of the diagonal $SU(2)$ subgroup of the continuum symmetry group, and the relevant decomposition is, for the groups $SU(2)_{\rm spin}\times SU(2)_{\rm taste}\supset SU(2)_{\rm diag}\supset G(R^{(k\ell)})$, $\left({\scriptstyle{{1\over 2}}}_{\rm spin},\ {\scriptstyle{{1\over 2}}}_{\rm taste}\right)\to{\bf 0}+{\bf 1}\to A_{1}+F_{1}\ .$ (255) Now, let us consider a correlation function of the form $C(t)=\langle 0|\hat{\phi}(t_{1})\hat{\overline{\phi}}(t_{2})|0\rangle\ ,$ (256) with $t=t_{1}-t_{2}>0$. Using Eqs. (236), (238) and (243), this can be written as (note that $\hat{S}_{4}|0\rangle=|0\rangle$) $\displaystyle C(t)$ $\displaystyle=$ $\displaystyle\sum_{E\sigma}\langle 0|\hat{\phi}(t_{1})|Er\sigma;\sigma_{t}\rangle\langle Er\sigma;\sigma_{t}|\hat{\overline{\phi}}(t_{2})|0\rangle$ $\displaystyle=$ $\displaystyle\sum_{E\sigma}\langle 0|\hat{\phi}\hat{S}_{4}^{t_{1}}|Er\sigma;\sigma_{t}\rangle\langle Er\sigma;\sigma_{t}|\hat{S}_{4}^{-t_{2}}\hat{\overline{\phi}}|0\rangle$ $\displaystyle=$ $\displaystyle\sum_{E\sigma}\sigma_{t}^{t}e^{-Et}\langle 0|\hat{\phi}|Er\sigma;\sigma_{t}\rangle\langle Er\sigma;\sigma_{t}|\hat{\overline{\phi}}|0\rangle$ $\displaystyle=$ $\displaystyle\sum_{E_{+}}|\langle 0|\hat{\phi}|E_{+}r;\sigma_{t}=\sigma_{s}\rangle|^{2}\sigma_{s}^{t}e^{-E_{+}t}$ $\displaystyle+\sum_{E_{-}}|\langle 0|\hat{\phi}|E_{-}r;\sigma_{t}=-\sigma_{s}\rangle|^{2}(-\sigma_{s})^{t}e^{-E_{-}t}\ ,$ where in the first term the sum is over states with positive parity $\sigma=\sigma_{t}\sigma_{s}$ and energies $E_{+}$, and in the second term the sum is over states with negative parity and energies $E_{-}$. The subscript on $E_{\pm}$ indicates the parity $\sigma$. If, for example, the operator $\hat{\phi}$ has $I_{s}$ parity $\sigma_{s}=+1$, the negative $P$ parity states yield a rapidly oscillating contribution to $C(t)$. The correlator gets contributions from both parities because $P$ is not well-defined for operators localized in time. The result (VI.1) is valid when the time extent of the lattice is infinite. If, instead, it is $2T$ (it is convenient to take the lattice even in all directions for staggered fermions), then, instead of considering $\langle 0|\hat{\phi}(t_{1})\hat{\overline{\phi}}(t_{2})|0\rangle$, we should consider ${\rm Tr}\,(\hat{S}_{4}^{2T-t}\hat{\phi}\hat{S}_{4}^{t}\hat{\overline{\phi}})$, with ${\rm Tr}\,$ the trace in Hilbert space. Let us instead derive the correction to Eq. (VI.1) more intuitively. If a particle can travel from $t_{2}$ to $t_{1}$, an anti-particle can travel from $t_{1}$ to $t_{2}^{\prime}=t_{2}+2T$. This leads to the replacement $e^{-Et}\to e^{-E(t_{2}^{\prime}-t_{1})}=e^{-E(2T-t)}$. With anti- periodic/periodic boundary conditions for the staggered quarks, we pick up a phase factor $\kappa=+1$, respectively, $\kappa=-1$ for fermionic operators, while always $\kappa=+1$ for bosonic operators. Furthermore, $\langle 0|\hat{\phi}|Er\sigma;\sigma_{t}\rangle\langle Er\sigma;\sigma_{t}|\hat{\overline{\phi}}|0\rangle=\sigma_{c}\langle 0|\hat{C}_{0}^{-1}\hat{\overline{\phi}}\hat{C}_{0}|Er\sigma;\sigma_{t}\rangle\langle Er\sigma;\sigma_{t}|\hat{C}_{0}^{-1}\hat{\phi}\hat{C}_{0}|0\rangle\ .$ (258) Using invariance under charge-conjugation, Eq. (VI.1) gets replaced by $\displaystyle C(t)$ $\displaystyle=$ $\displaystyle\sum_{E_{+}}|\langle 0|\hat{\phi}|E_{+}r;\sigma_{t}=\sigma_{s}\rangle|^{2}\sigma_{s}^{t}\left(e^{-E_{+}t}+\sigma_{c}\kappa\,e^{-(2T-t)E_{+}}\right)$ $\displaystyle+\sum_{E_{-}}|\langle 0|\hat{\phi}|E_{-}r;\sigma_{t}=-\sigma_{s}\rangle|^{2}(-\sigma_{s})^{t}\left(e^{-E_{-}t}+\sigma_{c}\kappa\,e^{-(2T-t)E_{-}}\right)+\dots\ .$ Contributions proportional to $|\langle n|\hat{\phi}|m\rangle|^{2}$ with both $|n\rangle$ and $|m\rangle$ not equal to $|0\rangle$ are not shown. Note that if $\langle 0|\hat{\phi}|0\rangle\neq 0$, we need to subtract $\langle 0|\hat{\phi}|0\rangle\langle 0|\hat{\overline{\phi}}|0\rangle$ to get the connected part of $C(t)$. ### VI.2 Mesons In this subsection, we will specialize to meson operators, in particular, operators bilinear in $\chi$ and $\bar{\chi}$ [27]. If $X_{\mu}=D(\Xi_{\mu})$ is the representation of $\Xi_{\mu}$ on a mesonic operator, they commute, because $D(\Xi_{\mu})D(\Xi_{\nu})=e^{iq\pi}D(\Xi_{\nu})D(\Xi_{\mu})\ ,$ (259) and $q=0$ for mesonic operators. (This can be easily checked on the explicit operators we will construct below.) Defining $\tilde{X}_{m}=X_{1}X_{2}X_{3}X_{m}\ ,$ (260) the restframe symmetry group for mesonic states is $(RF)_{\rm mesons}=G(\tilde{X}_{m},R^{(k\ell)})\times\\{1,\ X_{4}\\}\times\\{1,\ X_{1}X_{2}X_{3}\\}\times\\{1,\ P\\}\times\\{1,\ C_{0}\\}\ .$ (261) Mesonic states can thus be labeled by $|E,\overline{r}^{\sigma_{t}\sigma_{123}},\sigma,\tau_{0}\rangle\ ,$ (262) with $\overline{r}$ an irrep of $G(\tilde{X}_{m},R^{(k\ell)})$, and $\sigma_{t}$, $\sigma_{123}$, $\sigma$ and $\tau_{0}$ the $X_{4}$, $X_{123}$, $P$ and $C_{0}$ parities, respectively. Defining171717After constructing meson operators, they can be made gauge covariant by inserting averages over the shortest Wilson lines between $\chi$ and $\bar{\chi}$. $D_{k}\chi(x)={1\over 2}\left(\chi(x+k)+\chi(x-k)\right)\ ,$ (263) we define classes of bilinear operators (dropping hats on $\chi$ and $\bar{\chi}$) on a fixed time slice $t$: $\displaystyle M_{0}$ $\displaystyle=$ $\displaystyle\sum_{\vec{x}\ {\rm even}}\bar{\chi}(x)\chi(x)\ ,$ (264) $\displaystyle M_{1}$ $\displaystyle=$ $\displaystyle\sum_{\vec{x}\ {\rm even}}\bar{\chi}(x)D_{1}\chi(x)\ ,$ $\displaystyle M_{12}$ $\displaystyle=$ $\displaystyle\sum_{\vec{x}\ {\rm even}}\bar{\chi}(x)D_{1}D_{2}\chi(x)\ ,$ $\displaystyle M_{123}$ $\displaystyle=$ $\displaystyle\sum_{\vec{x}\ {\rm even}}\bar{\chi}(x)D_{1}D_{2}D_{3}\chi(x)\ ,$ where all members of each class are obtained by taking the operator explicitly
# Performance Bounds for Policy-Based Average Reward Reinforcement Learning Algorithms Yashaswini Murthy Electrical and Computer Engineering University of Illinois Urbana-Champaign Urbana, IL 61801 <EMAIL_ADDRESS> &Mehrdad Moharrami Electrical and Computer Engineering University of Illinois Urbana-Champaign Urbana, IL 61801 <EMAIL_ADDRESS>&R. Srikant Electrical and Computer Engineering University of Illinois Urbana-Champaign Urbana, IL 61801 <EMAIL_ADDRESS> ###### Abstract Many policy-based reinforcement learning (RL) algorithms can be viewed as instantiations of approximate policy iteration (PI), i.e., where policy improvement and policy evaluation are both performed approximately. In applications where the average reward objective is the meaningful performance metric, discounted reward formulations are often used with the discount factor being close to $1,$ which is equivalent to making the expected horizon very large. However, the corresponding theoretical bounds for error performance scale with the square of the horizon. Thus, even after dividing the total reward by the length of the horizon, the corresponding performance bounds for average reward problems go to infinity. Therefore, an open problem has been to obtain meaningful performance bounds for approximate PI and RL algorithms for the average-reward setting. In this paper, we solve this open problem by obtaining the first finite-time error bounds for average-reward MDPs, and show that the asymptotic error goes to zero in the limit as policy evaluation and policy improvement errors go to zero. ## 1 Introduction Reinforcement Learning algorithms can be broadly classified into value-based methods and policy-based methods. In the case of discounted-reward Markov Decision Processes (MDPs), value-based methods such as Q-learning [WD92, Tsi94, JJS93, SB18, BT96], fitted value iteration [MS08] and target network Q-learning [CCM22] can be viewed as approximately solving the fixed point of the Bellman optimality equation to find the value function and an approximately optimal control policy. In other words value iteration is approximately implemented for discounted reward MDPs [Ber11, Ber12, Put14]. In policy-gradient methods [AKLM21], a gradient step is used to improve the policy, somewhat similar to the policy improvement step in policy iteration for discounted-reward MDPs. In a recent paper [CM22], it has been shown that many policy-based methods for discounted-reward MDPs can be viewed as special cases of approximate policy iteration. The classical results on approximate policy iteration [BT96] assume that the error in the policy evaluation and improvement steps are constant, independent of the iteration. The key idea in [CM22] is to show that a simple modification of the proof in [BT96] can be used to allow for iteration-dependent policy evaluation and improvement errors, which is then used to make the connection to policy-based methods. Our goal in this paper is to derive similar results for average-reward problems. Average-reward problems [ABB01, WNS21, ZWSW21, Mah96, YB09, Gos04, Sin94] are, in some sense, harder to study than their discounted-reward counterparts. We now discuss why this is so by recalling the error bounds for discounted-reward problems and examining them in an appropriate limit to study their applicability to average-reward problems. The fundamental result on approximate policy iteration for discounted reward MDPs which allows the connection to policy-based methods is the following [BT96]: $\limsup_{k\to\infty}\|J_{\mu_{k}}-J_{*}^{\alpha}\|_{\infty}\leq H_{\alpha}^{2}(\epsilon+2\alpha\delta),$ (1) where $\alpha$ is the discount factor, $H_{\alpha}=1/(1-\alpha),$ $J_{*}^{\alpha}$ is the optimal value function, $J_{\mu_{k}}$ is the value function associated with the policy obtained after $k$ iterations of approximate policy iteration, $\epsilon$ is the policy improvement error, and $\delta$ is the policy evaluation error. Thus, as $\epsilon,\delta\rightarrow 0,$ we recover the result that standard policy iteration converges. On the other hand, to understand whether the above bound is useful to study average- reward problems, we write Equation 1 as $\limsup_{k\to\infty}\frac{1}{H_{\alpha}}\|J_{\mu_{k}}-J_{*}^{\alpha}\|_{\infty}\leq H_{\alpha}(\epsilon+2\alpha\delta).$ Under mild conditions [Ber11, Ros14], it is well known that each element of the value function vector approaches the average-reward for each fixed policy and for the optimal policy, in the limit $\alpha\rightarrow 1$. Hence, the left-hand side of the above equation is an approximation to the error in approximate policy iteration for average-reward MDPs; the right-hand side which gives an upper bound on this error blows up to infinity, i.e., when $\alpha\rightarrow 1.$ Thus, unlike in the discounted-reward case, the above bound fails to even recover the well-known convergence of standard policy iteration in average-reward case in the limit $\epsilon,\delta\rightarrow 0$ (since we have to let $\alpha\rightarrow 1$ before letting $\epsilon,\delta\rightarrow 0$ ). However, as mentioned in [BT96], it is also well known that approximate policy iteration performs much better than the bound suggests. The main goal of our paper is to resolve this discrepancy between theory and practice. ### 1.1 Contributions It is well known that the error bound for discounted-reward, approximate policy iteration is tight for unichain MDPs [BT96]. Thus, it is impossible to improve upon the bound in general. However, as will be explained in a later section, it is easy to convert most reasonable unichain MDPs to MDPs where every stationary policy results in an irreducible Markov chain with an arbitrarily small loss of reward. So the natural question to ask is whether the bound can be dramatically improved for MDPs where every stationary policy results in an irreducible Markov chain. To the best of our knowledge, no such bound was available in the prior literature. Our main contributions are as follows: * • Under the assumption that every stationary policy induces an irreducible Markov chain, we first perform a Schweitzer transformation of the MDP [Sch71] and obtain finite-time error bounds for average-reward approximate policy iteration. * • Using the above finite-time error bounds, we prove an error bound of the form $\limsup_{k\to\infty}J^{*}-J_{\mu_{k}}\leq f(\epsilon,\delta),$ for average-reward approximate policy iteration for a function $f$ such that $f(\epsilon,\delta)\rightarrow 0$ in the limit as $\epsilon,\delta\rightarrow 0.$ Note that this is in sharp contrast to the result that one can obtain from the discounted-reward analysis where the error bound blows up to infinity when applied to discount factors approaching $1,$ which is the appropriate limit to convert discounted-reward problems to average-reward problems. * • The main difficulty in obtaining the above bound when compared to the discounted-reward case is the lack of infinity-norm contraction property for the Bellman operator in the average-reward case. In most analysis of average- reward problems, this difficulty is circumvented by the use of a span- contraction property [Put14]. However, the span contraction property is insufficient for our purpose and therefore, we use a technique due to [VdW80] for the study of a different algorithm called modified policy iteration in [Put14]. * • Next, we extend the above analysis to obtain finite-iteration error bounds for the case where the policy evaluation and improvement errors can potentially vary from iteration to iteration. Further, we allow these errors to be random (as would be the case, for example, when one uses TD learning for policy evaluation and soft policy improvement) and obtain expected error bounds on approximate policy iteration after a finite number of iterations. * • Our error bounds are in a form such that one can then apply them to many policy-based RL algorithms, i.e., we can simply plug in error bounds for specific policy evaluation or policy improvement algorithms. We illustrate the applicability of the results to several different policy-based RL methods (softmax update, greedy update and mirror descent update) studied in the literature [MXSS20, BT96, AYBB+19, AKLM21]. * • While the main focus of our paper is on offline learning, we provide connections to regret bounds for online learning in average-reward MDPS as in [AYBB+19, LYAYS21]. ### 1.2 Related Work Our work in this paper is most closely related to the work in [CM22]. The main contribution there was to recognize that an extension of the approximate policy iteration results in [BT96] can be used to derive finite-time error bounds for many RL-based algorithms. The idea of considering iteration- dependent policy evaluation and improvement bounds was also used to study policy iteration with function approximation in discounted-reward MDPs in [WLLS21]. Our contribution here is to derive the first error bound for approximate policy iteration for average-reward MDPs and then write it in a form which can be applied to policy-based RL methods for average-reward TD learning. To the best of our knowledge, no known error bounds existed in prior literature for average-reward approximate policy iteration due to the fact that the corresponding bounds for discounted MDPs increase proportional to the square of length of the horizon. Thus, in the limit as the horizon increases to infinity, we have essentially reduced the dependence of the error from the square of the length of the horizon to linear in the length of the horizon. RL algorithms have not been studied as extensively for the average-reward case as they are for discounted-reward MDPs. There is recent work [ZZM21] on average-reward TD-learning which we leverage to illustrate the applicability of our results. There are also asymptotic convergence results for policy gradient and actor-critic methods for average-reward MDPs [TVR99, KT99, BSGL09, MMT00]. However, these results only prove convergence to a stationary point and there are no global convergence results. The original paper on natural policy gradient (NPG) is written for average-reward MDPs, but the extensive performance analysis of NPG in subsequent work such as [AKLM21] seems to only consider the discounted-reward case. Recent work [AYBB+19, LYAYS21] considers mirror descent in average reward RL but do not provide performance results for other RL algorithms. In a later section, we compare our results to [AYBB+19, LYAYS21] in the special case of mirror descent-type policy updates ## 2 Model and Preliminaries ### 2.1 Average Reward Formulation We consider the class of infinite horizon MDPs with finite state space $\mathcal{S}$, finite action space $\mathcal{A}$, and transition kernel $\mathbb{P}$, where $|\mathcal{S}|=n$ and $|\mathcal{A}|=m$. Let $\Delta\mathcal{A}$ denote the probability simplex over actions. We consider the class of randomized policies $\Pi=\\{\mu:\mathcal{S}\to\Delta\mathcal{A}\\}$, so that a policy $\mu$ assigns a probability vector over actions to each state. Given a policy $\mu$, the transition kernel for the underlying Markov process is denoted by $\mathbb{P}_{\mu}:\mathcal{S}\to\mathcal{S}$, where $\mathbb{P}_{\mu}(s^{\prime}|s):=\sum_{a\in\mathcal{A}}\mu(a|s)\mathbb{P}(s^{\prime}|s,a)$ is the probability of moving to state $s^{\prime}\in\mathcal{S}$ from $s\in\mathcal{S}$ upon taking action $\mu(s)\in\mathcal{A}$. Associated with each state-action pair $(s,a)$, is a one-step reward which is denoted by $r_{\mu}(s):=\mathbb{E}_{a\sim\mu(s)}r(s,a)$. Let $J_{\mu}\in\mathbb{R}$ be the average reward associated with the policy $\mu\in\Pi$, i.e. $J_{\mu}$ is defined as: $J_{\mu}=\lim_{T\to\infty}\frac{\mathbb{E}_{\mu}\left(\sum_{i=0}^{T-1}r_{\mu}(s_{i})\right)}{T}.$ Here the expectation is taken with respect to the measure $\mathbb{P}_{\mu}$ associated with the policy $\mu$. Let $h_{\mu}\in\mathbb{R}^{n}$ be the relative value function associated with the policy $\mu$. Defining $\boldsymbol{1}\in\mathbb{R}^{n}$ to be the vector of all ones, the pair $(J_{\mu},h_{\mu})$ satisfies the following average reward Bellman equation: $J_{\mu}\boldsymbol{1}+h_{\mu}=r_{\mu}+\mathbb{P}_{\mu}h_{\mu}.$ (2) Let $\pi_{\mu}\in\mathbb{R}^{n}$ denote the stationary distribution associated with the kernel $\mathbb{P}_{\mu}$. Let $\mathbb{P}^{*}_{\mu}=\boldsymbol{1}\pi_{\mu}^{\top}\in\mathbb{R}^{n\times n}$ denote the matrix whose rows are $\pi_{\mu}^{\top}$. Using $\mathbb{P}^{*}_{\mu}=\mathbb{P}_{\mu}^{*}\mathbb{P}_{\mu}$, we get the following characterization of the average reward by multiplying both sides of Equation 2 with $\mathbb{P}^{*}_{\mu}$: $J_{\mu}\boldsymbol{1}=\mathbb{P}^{*}_{\mu}r_{\mu}$. Let $J^{*}:=\max_{\mu\in\Pi}J_{\mu}$ be the optimal average reward. From standard MDP theory, there exists $h^{*}\in\mathbb{R}^{n}$, for which the pair ($J^{*}$,$h^{*}$) satisfies the following Bellman optimality equation: $J^{*}\boldsymbol{1}+h^{*}=\max_{\mu\in\Pi}r_{\mu}+\mathbb{P}_{\mu}h^{*}.$ (3) Let $\mu^{*}$ be the optimizing policy in Equation 3. Then, similar reasoning shows that $J^{*}\boldsymbol{1}=\mathbb{P}^{*}_{\mu^{*}}r_{\mu^{*}}.$ The goal of dynamic programming and reinforcement learning is to determine this optimal average reward $J^{*}$ and its corresponding optimal policy $\mu^{*}$. Policy iteration is one of the most widely used dynamic programming algorithms for this purpose. It consists of two steps: (i) evaluation of the value function associated with a policy, and (ii) determining a greedy policy with respect to this evaluation. There are primarily two challenges that arise when applying such an algorithm: (i) high memory and time complexity when the state space is too large, and (ii) the unknown transition probability kernel that governs the dynamics of the MDP. When the state space is too large, it may be computationally infeasible to perform policy iteration exactly. Existing literature suggests that approximate policy iteration in the context of discounted reward MDPs yields good performance. However, such an algorithm has not been studied in the context of average reward MDPs. In fact, the results obtained for the discounted reward MDPs yield vacuous performance bounds for average reward MDPs given the standard relationship between discounted reward and average reward MDPs. We explore this further in Section 2.2. ### 2.2 Relationship to the Discounted Reward MDP To provide some background, we first discuss the approximate policy iteration algorithm in the context of discounted reward MDPs. Let $\alpha\in[0,1)$ denote the discount factor. The discounted reward associated with a deterministic policy $\mu\in\Pi$ starting from some state $s\in\mathcal{S}$ is denoted $J_{\mu}^{\alpha}(s)$ and is defined as: $J_{\mu}^{\alpha}(s)=\mathbb{E}_{\mu}\left[\sum_{i=0}^{\infty}\alpha^{i}r(s_{i},\mu(s_{i}))\Big{|}s_{0}=s\right].$ The value function $J_{\mu}^{\alpha}\in\mathbb{R}^{n}$ is the unique fixed point of the Bellman operator $\mathsf{T}^{\alpha}_{\mu}:\mathbb{R}^{n}\to\mathbb{R}^{n}$ associated with the policy $\mu$, which is defined as $\mathsf{T}^{\alpha}_{\mu}J=r_{\mu}+\alpha\mathbb{P}_{\mu}J$. For each state $s\in\mathcal{S}$, let $J_{*}^{\alpha}(s)$ denote the optimal discounted reward, i.e., $J_{*}^{\alpha}(s)=\max_{\mu\in\Pi}J_{\mu}^{\alpha}(s).$ Similarly, $J_{*}^{\alpha}\in\mathbb{R}^{n}$ is the unique fixed point of the optimality Bellman operator $\mathsf{T}^{\alpha}:\mathbb{R}^{n}\to\mathbb{R}^{n}$, which is defined as $\mathsf{T}^{\alpha}J=\max_{\mu\in\Pi}\left(r_{\mu}+\alpha\mathbb{P}_{\mu}J\right)$. Algorithm 1 is Approximate Policy Iteration for the discounted reward MDP: Algorithm 1 Approximate Policy Iteration: Discounted Reward Require $J_{0}\in\mathbb{R}^{n}$ for $k=0,1,2,\ldots$ do 1\. Compute $\mu_{k+\\!1}\\!\\!\in\\!\Pi$ such that $\|\mathsf{T}^{\alpha}J_{k}\\!-\\!\mathsf{T}^{\alpha}_{\mu_{k+1}}\\!J_{k}\|_{\infty}\\!\leq\\!\epsilon$ $\triangleright$ Approximate Policy Improvement 2\. Choose $J_{k+1}$ such that $\|J_{k+1}-J_{\mu_{k+1}}\|_{\infty}\leq\delta$ $\triangleright$ Approximate Policy Evaluation where $J_{\mu_{k+1}}=\mathsf{T}^{\alpha}_{\mu_{k+1}}J_{\mu_{k+1}}$ end for ###### Theorem 2.1. Let $\mu_{k}$ be the sequence of policies generated from the approximate policy iteration algorithm (Algorithm 1). Then the performance error is bounded as: $\limsup_{k\to\infty}\|J_{\mu_{k}}-J_{*}^{\alpha}\|_{\infty}\leq\frac{\epsilon+2\alpha\delta}{(1-\alpha)^{2}}.$ (4) ###### Proof. The proof of this theorem can be found in [Ber11] ∎ From literature [Ros14, Ber11], we know that the average reward $J_{\mu}$ associated with any policy $\mu$ is related to its discounted reward $J^{\alpha}_{\mu}(s)$ counterpart as follows: $J_{\mu}=\lim_{\alpha\to 1}(1-\alpha)J^{\alpha}_{\mu}(s).$ Note that the above relation is independent of the state $s\in\mathcal{S}$, as the average reward does not depend on the initial state. Multiplying Equation 4 with $(1-\alpha)$, and letting $\alpha\to 1$, still yields an approximation performance bound with $(1-\alpha)$ in the denominator. This term blows up to infinity as $\alpha\to 1$. Note that this bound is known to be tight under unichain Markov structure of the probability transition kernel [Ber11]. However, in practice it is observed that approximate policy iteration works well in the average reward MDP case although these theoretical bounds are not representative of this performance. To the best of our knowledge, we are unaware of an average reward approximate policy iteration performance bound, when the policies induce an irreducible Markov chain. We bridge this gap between theory and practice by providing a theoretical analysis of approximate policy iteration in the average reward MDP setting, with non trivial performance bounds. ## 3 Approximate Policy Iteration for Average Reward A crucial component of the proof of convergence of approximate policy iteration in the context of discounted reward MDPs is the contraction due to the discount factor $\alpha$ which is absent in the average reward setting (since $\alpha=1$). To get some source of contraction, we have to make some assumptions on the MDP. ###### Assumption 3.1. We make the following assumptions: 1. (a) Every deterministic policy $\mu\in\Pi$ induces an irreducible Markov Chain $\mathbb{P}_{\mu}$. 2. (b) For all policies, the diagonal elements of the probability transition matrix are positive, i.e., the Markov chain stays in the same state with non-zero probability. Assumption (a) need not be satisfied by all MDPs. However, in order to satisfy this assumption, we consider a modified MDP where at every time step with probability $\varepsilon,$ an action is chosen from the set of all possible actions with equal probability. Simultaneously, with probability $1-\varepsilon$, we choose an action dictated by some policy. The problem then is to choose this policy optimally. For most MDPs of interest, this small modification will satisfy our assumption with a small $O(\varepsilon)$ loss in performance. It is straightforward to show the $O(\varepsilon)$ loss, but we include the proof in the appendix for completeness. Assumption (b) is without loss of generality in the following sense: there exists a simple transformation which essentially leaves the MDP unchanged but ensures that this assumption is satisfied. This transformation known as the aperiodicity transformation or Schweitzer transformation was introduced in [Sch71]. For more details, the reader is referred to the appendix. In the remainder of this paper, we assume that all MDPs are transformed accordingly to ensure the assumptions are satisfied. One consequence of Assumption (b) is the following lemma which will be useful to us later. ###### Lemma 3.2. There exists a $\gamma>0$ such that, for any policy $\mu\in\Pi$, $\min_{i\in\mathcal{S}}\pi_{\mu}(i)\geq\gamma,$ where $\pi_{\mu}$ is stationary distribution over states associated with the policy $\mu$. ###### Proof. This lemma has been proved in a more general sense in the context of deterministic policies in [VdW80]. Since the aperiodicity transformation ensures a non-zero probability of staying in the same state, the expected return times are bounded and the lemma holds true for randomized policies as well. ∎ Prior to presenting the algorithm, consider the following definitions. The average-reward Bellman operator $\mathsf{T}_{\mu}:\mathbb{R}^{n}\to\mathbb{R}^{n}$ corresponding to a policy $\mu$ is defined as $\mathsf{T}_{\mu}h={r}_{\mu}+{\mathbb{P}}_{\mu}h.$ The average-reward optimal Bellman operator $\mathsf{T}:\mathbb{R}^{n}\to\mathbb{R}^{n}$ is defined as $\mathsf{T}h=\max_{\mu\in\Pi}{r}_{\mu}+{\mathbb{P}}_{\mu}h.$ The value function $h_{\mu}$ associated with a policy $\mu$ satisfies the Bellman Equation 2 with single step reward ${r}_{\mu}$ and the transition kernel ${\mathbb{P}}_{\mu}$. Note that $h_{\mu}$ is unique up to an additive constant. Solving for $\left({J}_{\mu},{h}_{\mu}\right)$ using Equation 2 involves setting the value function for some fixed state $x^{*}$ as $0$ (since this reduces the system from being underdetermined to one with a unique solution). Further, the value function $h_{\mu}$ can be alternatively expressed as the $\lim_{m\to\infty}\mathsf{T}^{m}_{\mu}h_{0}$ for any $h_{0}\in\mathbb{R}^{n}$ as is the case with discounted reward MDPs. However, unlike the discounted reward Bellman Equation, there is no discount factor $\alpha<1$ preventing the value function from exploding to infinity. Hence, we consider value function computed using the relative Bellman operator $\widetilde{\mathsf{T}}_{\mu}$ defined as $\widetilde{\mathsf{T}}_{\mu}h={r}_{\mu}+{\mathbb{P}}_{\mu}h-{r}_{\mu}(x^{*})\boldsymbol{1}-\left({\mathbb{P}}_{\mu}h\right)(x^{*})\boldsymbol{1}.$ We now state the algorithm for Approximate Policy Iteration for Average Reward MDPs. ### 3.1 Approximate Policy Iteration Algorithm for Average Reward MDPs Algorithm 2 Approximate Policy Iteration: Average Reward 1:Require $h_{0}\in\mathbb{R}^{n}$ 2:for $k=0,1,2,\ldots$ do 3: 1\. Compute $\mu_{k+1}\\!\in\\!\Pi$ such that $\|\mathsf{T}h_{k}\\!-\\!\mathsf{T}_{\mu_{k+1}}h_{k}\|_{\infty}\leq\epsilon$ $\triangleright$ Approx. Policy Improvement 4: 2\. Compute $h_{k+1}$ such that $\|h_{k+1}-h_{\mu_{k+1}}\|_{\infty}\leq\delta$ $\triangleright$ Approx. Policy Evaluation 5: where $h_{\mu_{k+1}}=\lim_{m\to\infty}\widetilde{\mathsf{T}}^{m}_{\mu_{k+1}}h_{k}$ 6:end for ### 3.2 Performance bounds for Average Reward Approximate Policy Iteration We are now ready to present the main result of this work. ###### Theorem 3.3. Let ${J}^{*}$ be the optimal average reward of an MDP that satisfies 3.1. The sequence of policies $\mu_{k}$ generated by Algorithm 2 and their associated average rewards ${J}_{\mu_{k}}$ satisfy the following bound: $\displaystyle\Big{(}{J}^{*}-$ $\displaystyle{J}_{\mu_{k+1}}\Big{)}\leq\underbrace{\frac{\left(1-\left(1-\gamma\right)^{k}\right)}{\gamma}\left(\epsilon\left(\gamma+1\right)+2\delta\right)}_{\text{approximation error}}+\underbrace{\left(1-\gamma\right)^{k}\left({J}^{*}-{\min_{i}\left(\mathsf{T}h_{0}-h_{0}\right)(i)+\epsilon}\right)}_{\text{initial condition error}}.$ Interpretation of the Bound: Since $0<\gamma<1$, as $k\to\infty$, the error due to initial condition $h_{0}$ drops to zero. The approximation error consists of two components, $\epsilon$ and $\delta$. Here, $\epsilon$ represents the approximation error due to a suboptimal policy and $\delta$ represents the approximation error associated with evaluating the value function corresponding to a policy. The limiting behavior of the average reward associated with the policies obtained as a consequence of the algorithm is captured in the following corollary. ###### Corollary 3.4. The asymptotic performance of the policies obtained from Algorithm 2 satisfies: $\limsup_{k\to\infty}{\left({J}^{*}-{J}_{\mu_{k}}\right)}\leq\frac{(1+\gamma)\epsilon+2\delta}{\gamma}.$ (5) Note that the asymptotic bound in Equation 5 is in a very similar form as the asymptotic performance bound for the discounted reward MDP in Equation 4 where $\gamma$ plays the role of $(1-\alpha)^{2}$. However, when $\epsilon$ and $\delta$ got to zero, this bound also goes to zero which is not the case if we try to approximate average-reward problems by using the results for discounted-reward problems in the limit where the horizon goes to infinity. ## 4 Application to Reinforcement Learning Approximate policy iteration is closely related to RL. In general, approximate policy improvement and approximate policy evaluation in Algorithm 2 are executed through approximations to policy improvement (such as greedy update, mirror descent, softmax) and TD learning for value function approximation. First, we present a generic framework to analyze policy-based RL algorithms by building upon the theory presented in the last section. For this purpose, we define the state-action relative value function $Q$ (instead of the relative state value function $h$ as before) to evaluate a policy. Let $\mathsf{T}^{\mathsf{Q}}_{\mu}Q$ denote the Bellman operator with respect to $Q$ and policy $\mu$ where $(\mathsf{T}^{\mathsf{Q}}_{\mu}Q)(s,a)=r(s,a)+\left(\mathbb{Q}_{\mu}Q\right)(s,a),$ where $\mathbb{Q}_{\mu}(s^{\prime},a^{\prime}|s,a)=\mu(a^{\prime}|s^{\prime})\mathbb{P}(s^{\prime}|s,a).$ The relative state action value function corresponding to policy $\mu$ is represented by $Q_{\mu}$ and is the solution to the following Bellman equation: $J_{\mu}+Q_{\mu}(s,a)=r(s,a)+\sum_{s^{\prime}\in\mathcal{S},a^{\prime}\in\mathcal{A}}\mathbb{Q}_{\mu}(s^{\prime},a^{\prime}|s,a)Q_{\mu}(s^{\prime},a^{\prime})$, for all state action pairs $(s,a)\in(\mathcal{S},\mathcal{A})$. We present a generic policy-based algorithm for average-reward problem below. Algorithm 3 Generic Policy Based Algorithm: $Q$-function Average Reward Require $Q_{0}\in\mathbb{R}^{|\mathcal{S}||\mathcal{A}|}$ for $k=0,1,2,\ldots,T$ do 1\. Determine $\mu_{k+1}\\!\in\\!\Pi$ as a function of $Q_{k}$ using a possibly random policy improvement step 2\. Compute $Q_{k+1}$ as an approximation to $Q_{\mu_{k+1}}$ using (state, action, reward) samples from a trajectory generated by policy $\mu_{k+1}$ end for Note that the error in Steps 1 (policy improvement) and 2 (policy evaluation) of the above algorithm could be random. Thus, the analysis for Theorem 3.3 has to be adapted to Algorithm 3. The resulting expected deviation from optimality is characterized in the lemma below. ###### Lemma 4.1. Let $\mu_{k}$ and $Q_{k}$ be the sequence of policies and relative state- action value function iterates generated by Algorithm 3. For all $k\in 0,\ldots,T$, we have: $\displaystyle\\!\\!\Big{(}{J}^{*}-\left(\min_{(s,a)}\left(\mathsf{T}^{\mathsf{Q}}Q_{k}-Q_{k}\right)(s,a)\right)\\!\Big{)}$ $\displaystyle\leq\left(1-\gamma\right)\left({J}^{*}-\left(\min_{(s,a)}\left(\mathsf{T}^{\mathsf{Q}}Q_{k-1}-Q_{k-1}\right)(s,a)\right)\right)$ $\displaystyle+\left\|\mathsf{T}^{\mathsf{Q}}Q_{k-\\!1}\\!-\\!\mathsf{T}^{\mathsf{Q}}_{\mu_{k}}Q_{k-\\!1}\\!\right\|_{\infty}\\!\\!\\!+\\!2\\!\left\|Q_{k}\\!-\\!Q_{\mu_{k}}\\!\right\|_{\infty}\\!,$ where $\gamma=\min_{\begin{subarray}{c}\mu\in{\mu_{1},\mu_{2},\ldots,\mu_{T}}\\\ (s,a)\in\mathcal{S}\times\mathcal{A}\end{subarray}}\mathbb{Q}_{\mu}^{*}(s,a)>0,$ and $\mathbb{Q}_{\mu}^{*}$ is a matrix with identical rows corresponding to the invariant distribution of $\mathbb{Q}_{\mu}$. Iteratively applying Lemma 4.1 yields the following proposition for a generic policy-based algorithm. ###### Proposition 4.2. For any $T>0$, the iterates of Algorithm 3 satisfy $\displaystyle\mathbb{E}\big{[}{J}^{*}\\!-\\!{J}_{\mu_{T+1}}\big{]}\\!$ $\displaystyle\leq\\!\underbrace{\left(1\\!-\\!\gamma\right)^{T}\mathbb{E}\\!\left[{J}^{*}\\!-\\!\min_{i}\left(\mathsf{T}^{\mathsf{Q}}Q_{0}\\!-\\!Q_{0}\right)(i)\right]}_{\text{Initial condition error}}\\!+\\!\underbrace{2\sum_{\ell=0}^{T-1}\left(1\\!-\\!\gamma\right)^{\ell}\mathbb{E}\\!\left[\left\|Q_{T-\ell}\\!-\\!Q_{\mu_{T-\ell}}\right\|_{\infty}\right]}_{\text{Policy evaluation error}}$ $\displaystyle\\!+\\!\underbrace{\\!\sum_{\ell=1}^{T-1}\\!\left(1-\gamma\right)^{\ell-1}\mathbb{E}\\!\left[\left\|\mathsf{T}^{\mathsf{Q}}_{\mu_{T+1-\ell}}\\!Q_{T-\ell}\\!-\\!\mathsf{T}^{\mathsf{Q}}Q_{T-\ell}\right\|_{\infty}\right]+\mathbb{E}\left[{\left\|\mathsf{T}^{\mathsf{Q}}_{\mu_{T+1}}Q_{T}-\mathsf{T}^{\mathsf{Q}}Q_{T}\right\|_{\infty}}\right]}_{\text{Policy improvement error}}.$ We now illustrate the applicability of the above result to specific RL algorithms. More specifically, we characterize finite time performance bounds when state-action relative value functions are learnt through TD-Learning [ZZM21], while several different update rules are employed for policy improvement. ### 4.1 Performance bounds for RL algorithms For every iteration of Algorithm 3, it is necessary to observe trajectories of length $\tau$ in order to learn the state action value function associated with any policy. Using linear function approximations for value functions, the error in policy evaluation at iteration $k$ in Algorithm 3 can be expressed in the following form: $\|Q_{k}-Q_{\mu_{k}}\|_{\infty}\leq\|\Phi(\theta_{k}-\theta^{*}_{\mu_{k}})\|_{\infty}+\|Q_{\mu_{k}}-\Phi\theta^{*}_{\mu_{k}}\|_{\infty},$ (6) where $\Phi\in\mathbb{R}^{|\mathcal{S}||\mathcal{A}|\times d}$ corresponds to the feature vector matrix and $\theta\in\mathbb{R}^{d}$ is the parameter vector. Further, $\theta_{k}$ is the parameter vector obtained as a result of the TD learning algorithm. The projection of $Q_{\mu_{k}}$ onto the span of $\Phi$ is given by $\Phi\theta^{*}_{\mu_{k}}$. The last term in Equation 6 corresponds to the function approximation error which depends on the span of the feature matrix $\Phi$. We denote this error term by $\delta_{0,k}.$ The error due to TD learning (first term in Equation 6) is denoted by $\delta_{\text{TD},k}$. Assuming $\max_{(s,a)\in\mathcal{S}\times\mathcal{A}}\|\phi(s,a)\|_{2}\leq 1$, and $\|\theta_{\mu_{k}}^{*}\|_{2}$ is bounded uniformly in $k$, [ZZM21] shows that there exists $C>0$ such that the length of the sample trajectory $\tau$ and $\delta_{\text{TD},k}$ are related as follows: $\mathbb{E}\left[\delta_{\text{TD},k}\right]=\frac{C}{\sqrt{\tau}}.$ (7) Let $c_{0}=\left(J^{*}-\min_{i\in\mathcal{S}}\left(\mathsf{T}^{\mathsf{Q}}Q_{0}-Q_{0}\right)(i)\right)$ be the error due to initial condition and $\overline{\delta}_{0}=\max_{t}\delta_{0,t}$ be the maximum function approximation error across all iterations. Then we obtain the following performance bounds when TD learning is used for policy evaluation. ###### Corollary 4.3. (Greedy policy update) Let $\tilde{a}(s)=\operatorname*{argmax}_{a\in\mathcal{A}}Q_{k}(s,a^{\prime}).$ Let $\beta>0$ be such that the sequence of policies in Algorithm 3 corresponding to all state action pairs $(s,a)$ are obtained using the following algorithm, which we call the greedy policy update. $\displaystyle\mu_{k+1}(a|s)=\begin{cases}\frac{1}{\beta|\mathcal{A}|},&\text{if }a\neq\tilde{a}(s)\\\ \frac{1}{\beta|\mathcal{A}|}+1-\frac{1}{\beta},&\text{if }a=\tilde{a}(s),\end{cases}$ (8) Let $\eta=\max_{\begin{subarray}{c}s\in\mathcal{S},a\in\mathcal{A}\\\ k\in 1\ldots T\end{subarray}}|Q_{k}(s,a)|.$ Using TD Learning with linear value function approximation for policy evaluation, we get the following finite time performance bound: $\displaystyle\mathbb{E}\left[J^{*}-J_{\mu_{T+1}}\right]$ $\displaystyle\leq(1-\gamma)^{T}c_{0}+\left(\frac{1+\gamma}{\gamma}\right)\frac{2\eta}{\beta}+\frac{2}{\gamma}\left(\overline{\delta}_{0}+\frac{C}{\sqrt{\tau}}\right).$ (9) ###### Corollary 4.4. (Softmax policy update) Let $\beta>0$ be such that the sequences of policies in Algorithm 3 corresponding to all state action pairs $(s,a)$ are obtained using the following algorithm, which we call the softmax policy update. $\mu_{k+1}(a|s)=\frac{\exp\left(\beta Q_{k}(s,a)\right)}{\sum_{a^{\prime}\in\mathcal{A}}\exp\left(\beta Q_{k}(s,a^{\prime})\right)}.$ (10) Using TD Learning with linear value function approximation for policy evaluation, we get the following finite time performance bound: $\displaystyle\mathbb{E}\left[J^{*}-J_{\mu_{T+1}}\right]$ $\displaystyle\leq(1-\gamma)^{T}c_{0}+\left(\frac{1+\gamma}{\gamma}\right)\frac{\log\left(|\mathcal{A}|\right)}{\beta}+\frac{2}{\gamma}\left(\overline{\delta}_{0}+\frac{C}{\sqrt{\tau}}\right).$ (11) ###### Corollary 4.5. (Mirror descent update) Let $\beta>0$ be such that the sequences of policies in Algorithm 3 corresponding to all state action pairs $(s,a)$ are obtained using the following algorithm, which we call the mirror descent policy update. $\mu_{k+1}(a|s)\propto\mu_{k}(a|s)\exp\\{\beta Q_{k}(s,a)\\}.$ (12) Let $\omega=\min_{\begin{subarray}{c}k\in 1\ldots T\\\ s\in\mathcal{S}\end{subarray}}\mu_{k}(a^{*}(s)|s)$ where $a^{*}(s)$ is the optimal action at state $s.$ Using TD Learning with linear value function approximation for policy evaluation, we get the following finite time performance bound: $\displaystyle\mathbb{E}\left[J^{*}-J_{\mu_{T+1}}\right]$ $\displaystyle\leq(1-\gamma)^{T}c_{0}+\left(\frac{1+\gamma}{\gamma\beta}\right)\log\left(\frac{1}{\omega}\right)+\frac{2}{\gamma}\left(\overline{\delta}_{0}+\frac{C}{\sqrt{\tau}}\right).$ (13) ###### Remark 4.6. The following remarks are in order: * • In each of the above performance bounds, there is an inevitable constant error $(\overline{\delta}_{0})$ due to function approximation. If the value function lies in the class of functions chosen, then this error will be zero. * • The error due to policy update (characterized by $\beta$) can be made arbitrarily small by choosing $\beta$ large. However, choosing a large $\beta$ may increase $C$ due to the fact that $C$ is a function of the mixing time of each policy, which is affected by the choice of $\beta.$ However, $C$ cannot be arbitrarily large due to Assumption (a). * • The error due to the initial condition decays exponentially fast but the rate of decay can be slow because $\gamma$ can be very small. In this regard, it is interesting to compare our work with the results in [AYBB+19, LYAYS21]. We do so in the next subsection. ### 4.2 Comparison with [AYBB+19, LYAYS21] Our model for approximate policy iteration is intended to be a general framework to study policy-based RL algorithms. A specific class of such algorithms has been studied in [AYBB+19, LYAYS21], where regret bounds are presented for a mirror descent like algorithm referred to as Politex. Although our model is intended for offline algorithms, it is possible to adapt the techniques in [AYBB+19, LYAYS21] to obtain regret bounds for our model as well. Let $r(s_{t},a_{t})$ be the reward obtained at time $t.$ Let $K$ be the total number of time units for which Algorithm 3 is executed. Hence $K=T\tau$, where $\tau$ is the length of the sample trajectory for every iteration of policy evaluation. The regret, defined as $R(K)=\sum_{t=1}^{K}\left(J^{*}-r(s_{t},a_{t})\right)$ can be further decomposed as: $R(K)=\sum_{t=1}^{K}\left(J^{*}-J_{\mu_{t}}\right)+\left(J_{\mu_{t}}-r(s_{t},a_{t})\right).$ (14) The first term i.e., $\sum_{t=1}^{K}\left(J^{*}-J_{\mu_{t}}\right)$ is referred to as the pseudo regret $R_{\text{PS}}(K)$. The second term is essentially bounding the average value from its estimate and can be bounded using a martingale analysis identical to [AYBB+19, LYAYS21]. Here, we focus on obtaining an upper bound on the pseudo regret with mirror descent update and TD learning. The pseudo regret can be further decomposed as: $R_{\text{PS}}(K)=\tau\sum_{t=1}^{T}\left(J^{*}-J_{\mu_{t}}\right).$ For the mirror descent policy update in Corollary 4.5, we get the following expected regret bound $R_{\text{PS}}(K)\leq\tau\sum_{t=1}^{T}(1-\gamma)^{T}c_{0}+\frac{1+\gamma}{\gamma\beta}\log\left(\frac{1}{\omega}\right)+\frac{2}{\gamma}\left(\overline{\delta}_{0}+\frac{c_{3}}{\sqrt{\tau}}\right).$ (15) Let $\beta=\sqrt{\tau}$, then the above regret can be expressed as: $R_{\text{PS}}(K)\leq\frac{1}{\gamma}\left({\tau c_{0}}+{2K\overline{\delta}_{0}}+\frac{Kc_{5}}{\sqrt{\tau}}\right).$ Optimizing for the regret yields $\tau=O(K^{\frac{2}{3}})$ and the following pseudo regret upper bound, $R_{\text{PS}}(K)\leq\frac{1}{\gamma}\left(2K\overline{\delta}_{0}+c_{6}K^{\frac{2}{3}}\right).$ (16) We now compare our regret bounds to the ones in [AYBB+19, LYAYS21]. Since all of the bounds will have an inevitable error $(\overline{\delta}_{0})$ due to function approximation, the comparison is with respect to the other terms in the regret bound. * • We get a better order-wise bound on the regret compare to [AYBB+19] which obtains a bound of $O(K^{3/4}).$ However, they have better constants which are in terms of mixing-time coefficients rather than in terms of $\gamma.$ * • [LYAYS21] gets a better regret bound $(O(\sqrt{K}))$ than us. The results in [AYBB+19, LYAYS21] use Monte Carlo policy evaluation, which allows for a closed-form expression for the value function estimate. The structure of the closed-form expressions allows them to exploit the fact that the policy varies slowly from one iteration to the next (which is specific to the mirror descent policy update) and hence $\gamma$ does not appear in their analysis. However, the Monte Carlo policy evaluation algorithms use matrix inversion which is often too expensive to implement in practice unlike the TD learning algorithm in our analysis which is widely used. Nevertheless, the ideas in [AYBB+19, LYAYS21] provide a motivation to investigate whether one can extend their analysis to more general policy evaluation and policy improvement algorithms. This is an interesting topic for future research. ## 5 Conclusions We present a general framework for analyzing policy-based reinforcement learning algorithms for average-reward MDPs. Our main contribution is to obtain performance bounds without using the natural source of contraction available in discounted reward formulations. We apply our general framework to several well-known RL algorithms to obtain finite-time error bounds on the average reward. We conclude with a comparative regret analysis and an outline of possible future work. ## References * [ABB01] Jinane Abounadi, Dimitri Bertsekas, and Vivek S Borkar. Learning algorithms for markov decision processes with average cost. SIAM Journal on Control and Optimization, 40(3):681–698, 2001. * [AKLM21] Alekh Agarwal, Sham M Kakade, Jason D Lee, and Gaurav Mahajan. On the theory of policy gradient methods: Optimality, approximation, and distribution shift. The Journal of Machine Learning Research, 22(1):4431–4506, 2021\. * [AYBB+19] Yasin Abbasi-Yadkori, Peter Bartlett, Kush Bhatia, Nevena Lazic, Csaba Szepesvari, and Gellért Weisz. Politex: Regret bounds for policy iteration using expert prediction. In International Conference on Machine Learning, pages 3692–3702. PMLR, 2019. * [Ber11] Dimitri P Bertsekas. Dynamic programming and optimal control 3rd edition, volume ii. Belmont, MA: Athena Scientific, 2011. * [Ber12] Dimitri Bertsekas. Dynamic programming and optimal control: Volume I, volume 1. Athena scientific, 2012. * [BSGL09] Shalabh Bhatnagar, Richard S Sutton, Mohammad Ghavamzadeh, and Mark Lee. Natural actor–critic algorithms. Automatica, 45(11):2471–2482, 2009. * [BT96] Dimitri Bertsekas and John N Tsitsiklis. Neuro-dynamic programming. Athena Scientific, 1996. * [CCM22] Zaiwei Chen, John Paul Clarke, and Siva Theja Maguluri. Target network and truncation overcome the deadly triad in $q$-learning. arXiv preprint arXiv:2203.02628, 2022. * [CM22] Zaiwei Chen and Siva Theja Maguluri. Sample complexity of policy-based methods under off-policy sampling and linear function approximation. In International Conference on Artificial Intelligence and Statistics, pages 11195–11214. PMLR, 2022. * [Gos04] Abhijit Gosavi. Reinforcement learning for long-run average cost. European journal of operational research, 155(3):654–674, 2004\. * [JJS93] Tommi Jaakkola, Michael Jordan, and Satinder Singh. Convergence of stochastic iterative dynamic programming algorithms. Advances in neural information processing systems, 6, 1993. * [KT99] Vijay Konda and John Tsitsiklis. Actor-critic algorithms. Advances in neural information processing systems, 12, 1999. * [LYAYS21] Nevena Lazic, Dong Yin, Yasin Abbasi-Yadkori, and Csaba Szepesvari. Improved regret bound and experience replay in regularized policy iteration. In International Conference on Machine Learning, pages 6032–6042. PMLR, 2021. * [Mah96] Sridhar Mahadevan. Average reward reinforcement learning: Foundations, algorithms, and empirical results. Recent advances in reinforcement Learning, pages 159–195, 1996\. * [MMT00] Peter Marbach, Oliver Mihatsch, and John N Tsitsiklis. Call admission control and routing in integrated services networks using neuro-dynamic programming. IEEE Journal on selected areas in communications, 18(2):197–208, 2000. * [MS08] Rémi Munos and Csaba Szepesvári. Finite-time bounds for fitted value iteration. Journal of Machine Learning Research, 9(5), 2008. * [MXSS20] Jincheng Mei, Chenjun Xiao, Csaba Szepesvari, and Dale Schuurmans. On the global convergence rates of softmax policy gradient methods. In International Conference on Machine Learning, pages 6820–6829. PMLR, 2020. * [Put14] Martin L Puterman. Markov decision processes: discrete stochastic dynamic programming. John Wiley & Sons, 2014. * [Ros14] Sheldon M Ross. Introduction to stochastic dynamic programming. Academic press, 2014. * [SB18] Richard S Sutton and Andrew G Barto. Reinforcement learning: An introduction. MIT press, 2018. * [Sch71] Paul J Schweitzer. Iterative solution of the functional equations of undiscounted markov renewal programming. Journal of Mathematical Analysis and Applications, 34(3):495–501, 1971. * [Sin94] Satinder P Singh. Reinforcement learning algorithms for average-payoff markovian decision processes. In AAAI, volume 94, pages 700–705, 1994. * [Tsi94] John N Tsitsiklis. Asynchronous stochastic approximation and q-learning. Machine learning, 16:185–202, 1994. * [TVR99] John N Tsitsiklis and Benjamin Van Roy. Average cost temporal-difference learning. Automatica, 35(11):1799–1808, 1999. * [VdW80] J Van der Wal. Successive approximations for average reward markov games. International Journal of Game Theory, 9(1):13–24, 1980. * [WD92] Christopher JCH Watkins and Peter Dayan. Q-learning. Machine learning, 8:279–292, 1992. * [WLLS21] Anna Winnicki, Joseph Lubars, Michael Livesay, and R Srikant. The role of lookahead and approximate policy evaluation in reinforcement learning with linear value function approximation. arXiv preprint arXiv:2109.13419, 2021. * [WNS21] Yi Wan, Abhishek Naik, and Richard S Sutton. Learning and planning in average-reward markov decision processes. In Marina Meila and Tong Zhang, editors, Proceedings of the 38th International Conference on Machine Learning, volume 139 of Proceedings of Machine Learning Research, pages 10653–10662. PMLR, 18–24 Jul 2021. * [YB09] Huizhen Yu and Dimitri P Bertsekas. Convergence results for some temporal difference methods based on least squares. IEEE Transactions on Automatic Control, 54(7):1515–1531, 2009. * [ZWSW21] Shangtong Zhang, Yi Wan, Richard S Sutton, and Shimon Whiteson. Average-reward off-policy policy evaluation with function approximation. In Marina Meila and Tong Zhang, editors, Proceedings of the 38th International Conference on Machine Learning, volume 139 of Proceedings of Machine Learning Research, pages 12578–12588. PMLR, 18–24 Jul 2021. * [ZZM21] Sheng Zhang, Zhe Zhang, and Siva Theja Maguluri. Finite sample analysis of average-reward td learning and $q$-learning. Advances in Neural Information Processing Systems, 34:1230–1242, 2021. ## 6 Appendix ### 6.1 Discussion on Assumption 3.1 #### 6.1.1 Assumption 3.1(a) In order to ensure every policy induces an irreducible Markov chain, we modify the MDP where at every time step with probability $\epsilon$, an action is chosen from the set of all possible actions with equal probability. Simultaneously, with probability $1-\epsilon$, we choose an action dictated by some policy. Let the transition kernel of the MDP associated with policy $\mu$ be denoted as $\mathbb{P}_{\mu}$. We denote the modified kernel by $\widehat{\mathbb{P}}_{\mu}$, where the transition probability from state $i$ to $j$ is given by: $\widehat{\mathbb{P}}_{\mu}(j|i)=(1-\epsilon)\mathbb{P}_{\mu}(j|i)+\epsilon\left(\frac{1}{|\mathcal{A}|}\sum_{a\in\mathcal{A}}\mathbb{P}(j|i,a)\right)$ Let $\mathbb{P}_{\rho}(j|i)=\frac{1}{|\mathcal{A}|}\sum_{a\in\mathcal{A}}\mathbb{P}(j|i,a),$ where $\rho$ represents randomized policy, that is $\rho(a|i)=\frac{1}{|\mathcal{A}|.}$ Let $(J_{\mu},V_{\mu})$ be the average reward and state value function vector associated with the policy $\mu.$ Then they the satisfy the following Bellman Equation: $J_{\mu}+V_{\mu}=r_{\mu}+\mathbb{P}_{\mu}V_{\mu},$ where $r_{\mu}(i)=\sum_{a\in\mathcal{A}}\mu(a|i)r(i,a).$ Similarly $(\widehat{J}_{\mu},\widehat{V}_{\mu})$ be the average reward and state value function vector that satisfy the following Bellman Equation, $\widehat{J}_{\mu}+\widehat{V}_{\mu}=\widehat{r}_{\mu}+\widehat{\mathbb{P}}_{\mu}\widehat{V}_{\mu}$ where $\widehat{r}_{\mu}=(1-\epsilon)r_{\mu}+\epsilon r_{\rho}$. Since $\widehat{\mathbb{P}}_{\mu}=(1-\epsilon)\mathbb{P}_{\mu}+\epsilon\mathbb{P}_{\rho},$ the above equation can be rewritten as: $\widehat{J}_{\mu}=\left(1-\epsilon\right)\left(r_{\mu}+\mathbb{P}_{\mu}\widehat{V}_{\mu}-\widehat{V}_{\mu}\right)+\epsilon\left(r_{\rho}+\mathbb{P}_{\rho}\widehat{V}_{\mu}-\widehat{V}_{\mu}\right)$ Multiplying the above equation by the vector $\mathbb{P}^{*}_{\mu}$, which is the invariant distribution over the state space due to policy $\mu$, we obtain, $\widehat{J}_{\mu}=\left(1-\epsilon\right)\left(\mathbb{P}^{*}_{\mu}r_{\mu}\right)+\epsilon\mathbb{P}^{*}_{\mu}\left(r_{\rho}+\mathbb{P}_{\rho}\widehat{V}_{\mu}-\widehat{V}_{\mu}\right)$ since $\mathbb{P}^{*}_{\mu}\mathbb{P}_{\mu}=\mathbb{P}^{*}_{\mu}$. We also know that $J_{\mu}=\mathbb{P}^{*}_{\mu}r_{\mu}.$. Hence we obtain, $\widehat{J}_{\mu}=\left(1-\epsilon\right)J_{\mu}+\epsilon\mathbb{P}^{*}_{\mu}\left(r_{\rho}+\mathbb{P}_{\rho}\widehat{V}_{\mu}-\widehat{V}_{\mu}\right)$ Therefor we obtain the following result, $J_{\mu}-\widehat{J}_{\mu}=\epsilon\left(J_{\mu}-\mathbb{P}^{*}_{\mu}\left(r_{\rho}+\mathbb{P}_{\rho}\widehat{V}_{\mu}-\widehat{V}_{\mu}\right)\right)$ Since the rewards are bounded, so are $J_{\mu},r_{\rho}$ and $\mathbb{P}_{\rho}\widehat{V}_{\mu}-\widehat{V}_{\mu}$ for all $\mu$. Hence we obtain, $J_{\mu}-\widehat{J}_{\mu}=O(\epsilon)$ The difference in the average reward associated with the original MDP and the modified MDP vary by $O(\epsilon).$ Let the optimal policy corresponding to original MDP be $\mu^{*}$ and the optimal policy corresponding to modified MDP be $\widehat{\mu}^{*}$. Then, $J_{\mu^{*}}-\widehat{J}_{\widehat{\mu}^{*}}=J_{\mu^{*}}-\widehat{J}_{\mu^{*}}+\widehat{J}_{\mu^{*}}-\widehat{J}_{\widehat{\mu}^{*}}$ Since $\widehat{\mu}^{*}$ is the optimizing policy for the modified MDP, we have $\widehat{J}_{\mu^{*}}-\widehat{J}_{\widehat{\mu}^{*}}\leq 0$. Hence, we obtain, $J_{\mu^{*}}-\widehat{J}_{\widehat{\mu}^{*}}\leq J_{\mu^{*}}-\widehat{J}_{\mu^{*}}=O(\epsilon).$ Similarly, $J_{\mu^{*}}-\widehat{J}_{\widehat{\mu}^{*}}=J_{\mu^{*}}-J_{\widehat{\mu}^{*}}+J_{\widehat{\mu}^{*}}-\widehat{J}_{\widehat{\mu}^{*}}$ Since $\mu^{*}$ is the optimizing policy for the original MDP, we have $J_{\mu^{*}}-J_{\widehat{\mu}^{*}}\geq 0$. Hence we obtain, $J_{\mu^{*}}-\widehat{J}_{\widehat{\mu}^{*}}\geq J_{\widehat{\mu}^{*}}-\widehat{J}_{\widehat{\mu}^{*}}=O(\epsilon).$ Thus the optimal average reward of the original MDP and modified MDP differ by $O(\epsilon).$ That is, $|J_{\mu^{*}}-\widehat{J}_{\widehat{\mu}^{*}}|\leq O(\epsilon)$ #### 6.1.2 Assumption 3.1 (b) To ensure Assumption 3.1 (b) is satisfied, an aperiodicity transformation can be implemented. Under this transformation, the transition probabilities of the original MDP are modified such that the probability of staying in any state under any policy is non-zero. In order to compensate for the change in transition kernel, the single step rewards are analogously modified such that the Bellman equation corresponding to the original and transformed dynamics are scaled versions of one another. This thus ensures that the optimality of a policy remains the same irrespective of this transformation, along with yielding quantifiable convergence properties. Mathematically, this transformation is described below. ###### Definition 6.1 (Aperiodicity Transformation). Let $\kappa\in\left(0,1\right)$. For every policy $\mu\in\Pi$ and for all states $i,j\in\mathcal{S}$, consider the following transformation: $\displaystyle\widehat{\mathbb{P}}_{\mu}\left(i|i\right)$ $\displaystyle=\kappa+(1-\kappa)\mathbb{P}_{\mu}\left(i|i\right)$ (17) $\displaystyle\widehat{\mathbb{P}}_{\mu}\left(j|i\right)$ $\displaystyle=(1-\kappa)\mathbb{P}_{\mu}\left(j|i\right),\qquad j\neq i$ (18) $\displaystyle\widehat{r}(i,\mu(i))$ $\displaystyle=(1-\kappa)r(i,\mu(i))$ (19) ###### Theorem 6.2. Given the transformation in Equation 17-Equation 19, for every $\mu\in\Pi$, let $(J_{\mu},h_{\mu})$ and $(\widehat{J}_{\mu},\widehat{h}_{\mu})$ be the solution to the Bellman Equation (Equation 2) corresponding to the original MDP $(\mathbb{P}_{\mu},r_{\mu})$ and the transformed MDP $(\widehat{\mathbb{P}}_{\mu},\widehat{r}_{\mu})$. Then, $\left(\widehat{J}_{\mu},\widehat{h}_{\mu}\right)=\left((1-\kappa)J_{\mu},h_{\mu}\right)$ ###### Proof. The proof of this theorem can be found in [Sch71]. ∎ ###### Remark 6.3. Due to this bijective relationship between the original and the transformed problem, it suffices to solve the transformed problem. Given that the trajectory used for algorithms such as TD Learning corresponds to the original system, such a transformation necessitates a mild change in how samples from a given policy are utilized in TD learning. More specifically, for each (state, action) sample, with probability $\kappa,$ we have to add the same sample to the data set for the next time instant. With probability $1-\kappa$ choose the next state in the trajectory. Hence, this transformation is easy to incorporate in RL algorithms. ### 6.2 Proof of Theorem 3.3 Prior to presenting the proof, we define $\displaystyle u_{k}=\max_{i}(\mathsf{T}h_{k}-h_{k})(i),$ (20) $\displaystyle l_{k}=\min_{i}(\mathsf{T}h_{k}-h_{k})(i).$ A key lemma in the proof of convergence of approximate policy iteration in the context of average reward is: ###### Lemma 6.4. Let $J^{*}$ be the optimal average reward associated with the transformed MDP. For all $k\in\mathbb{N}$: $l_{k}-\epsilon\leq J_{\mu_{k+1}}\leq{J}^{*}\leq u_{k}$ ###### Proof. From definition, $\displaystyle l_{k}\boldsymbol{1}$ $\displaystyle\leq\mathsf{T}h_{k}-h_{k}$ Since $\|\mathsf{T}h_{k}-\mathsf{T}_{\mu_{k+1}}h_{k}\|_{\infty}\leq\epsilon$, $\displaystyle l_{k}\boldsymbol{1}$ $\displaystyle\leq\mathsf{T}_{\mu_{k+1}}h_{k}-h_{k}+\epsilon\boldsymbol{1}$ $\displaystyle=r_{\mu_{k+1}}+\mathbb{P}_{\mu_{k+1}}h_{k}-h_{k}+\epsilon\boldsymbol{1}$ $\displaystyle\mathbb{P}_{\mu_{k+1}}^{*}l_{k}\boldsymbol{1}$ $\displaystyle\leq\mathbb{P}_{\mu_{k+1}}^{*}r_{\mu_{k+1}}+\mathbb{P}_{\mu_{k+1}}^{*}\mathbb{P}_{\mu_{k+1}}h_{k}-\mathbb{P}_{\mu_{k+1}}^{*}h_{k}+\mathbb{P}_{\mu_{k+1}}^{*}\epsilon\boldsymbol{1}$ where $\mathbb{P}_{\mu_{k+1}}^{*}$ is a matrix whose rows are identical and are the invariant distribution associated with the probability transition kernel $\mathbb{P}_{\mu_{k+1}}$. Since $\mathbb{P}_{\mu_{k+1}}^{*}\mathbb{P}_{\mu_{k+1}}=\mathbb{P}_{\mu_{k+1}}^{*}$, $\displaystyle l_{k}\boldsymbol{1}$ $\displaystyle\leq\mathbb{P}^{*}_{\mu_{k+1}}r_{\mu_{k+1}}+\epsilon\boldsymbol{1}$ $\displaystyle l_{k}\boldsymbol{1}-\epsilon\boldsymbol{1}$ $\displaystyle\leq{J}_{\mu_{k+1}}\boldsymbol{1}$ $\displaystyle l_{k}-\epsilon$ $\displaystyle\leq{J}_{\mu_{k+1}}$ From definition, $\displaystyle{J}^{*}=\max_{\mu\in\Pi}{J}_{\mu}\geq{J}_{\mu_{k+1}}$ Let $\mu^{*}$ be the policy corresponding to the optimal average reward ${J}^{*}$ and value function $h^{*}$, ie., $\displaystyle{J}^{*}+h^{*}$ $\displaystyle=\max_{\mu}r_{\mu}+\mathbb{P}_{\mu}h^{*}$ $\displaystyle=r_{\mu^{*}}+\mathbb{P}_{\mu^{*}}h^{*}$ Then, $\displaystyle u_{k}\boldsymbol{1}$ $\displaystyle\geq\mathsf{T}h_{k}-h_{k}$ $\displaystyle\stackrel{{\scriptstyle(a)}}{{\geq}}\mathsf{T}_{\mu^{*}}h_{k}-h_{k}$ $\displaystyle=r_{\mu^{*}}+\mathbb{P}_{\mu^{*}}h_{k}-h_{k}$ $\displaystyle\mathbb{P}_{\mu^{*}}^{*}u_{k}\boldsymbol{1}$ $\displaystyle\geq\mathbb{P}_{\mu^{*}}^{*}r_{\mu^{*}}+\mathbb{P}_{\mu^{*}}^{*}\mathbb{P}_{\mu^{*}}h_{k}-\mathbb{P}_{\mu^{*}}^{*}h_{k}$ $\displaystyle u_{k}\boldsymbol{1}$ $\displaystyle\geq\mathbb{P}^{*}_{\mu^{*}}r_{\mu^{*}}$ $\displaystyle u_{k}$ $\displaystyle\geq{J}^{*}$ where $(a)$ is due to the fact that $\mathsf{T}$ is the maximising Bellman operator. Hence $\forall k\in\mathbb{N}$, $\displaystyle l_{k}-\epsilon\leq{J}_{\mu_{k+1}}\leq{J}^{*}\leq u_{k}$ ∎ Without any approximation, the above lemma indicates that there exists a state whose value function changes by an amount less than the average reward associated with the optimizing policy at that iteration, and also that there exists a state whose value function increases by an amount larger than the optimal average reward. The proof proceeds by showing a geometric convergence rate for $l_{k}$. More precisely, the following lemma is proved. ###### Lemma 6.5. For all $k\in\mathbb{N}$, it is true that: $\left({J}^{*}-l_{k}\right)\leq\left(1-\gamma\right)\left({J}^{*}-l_{k-1}\right)+\epsilon+2\delta$ ###### Proof. Define: $g_{k}(i)=\left(\mathsf{T}h_{k}-h_{k}\right)(i)$ $\displaystyle g_{k}(i)$ $\displaystyle\geq\left(\mathsf{T}_{\mu_{k}}h_{k}-h_{k}\right)(i)$ Since $\mathsf{T}$ is the maximizing Bellman operator. We know that: 1. 1. $\left\|h_{k}-h_{\mu_{k}}\right\|_{\infty}\leq\delta$ 2. 2. For all constant $p\in\mathbb{R}$: $\displaystyle\mathsf{T}_{\mu_{k}}\left(h_{k}+p\boldsymbol{1}\right)$ $\displaystyle=r_{\mu_{k}}+\mathbb{P}_{\mu_{k}}\left(h_{k}+p\boldsymbol{1}\right)$ $\displaystyle=r_{\mu_{k}}+\mathbb{P}_{\mu_{k}}h_{k}+p\boldsymbol{1}$ $\displaystyle=\mathsf{T}_{\mu_{k}}h_{k}+p\boldsymbol{1}$ We thus obtain, $\displaystyle g_{k}(i)$ $\displaystyle\geq\mathsf{T}_{\mu_{k}}(h_{\mu_{k}}-\delta)-(h_{\mu_{k}}+\delta)$ (21) $\displaystyle\geq\left(\mathsf{T}_{\mu_{k}}h_{\mu_{k}}-h_{\mu_{k}}\right)(i)-2\delta$ (22) Recall that $h_{\mu_{k}}=\lim_{m\to\infty}\widetilde{T}^{m}_{\mu_{k+1}}h_{k}$ $\displaystyle\widetilde{\mathsf{T}}_{\mu_{k}}^{m}h_{k-1}$ $\displaystyle=\widetilde{\mathsf{T}}_{\mu_{k}}^{m-1}\widetilde{\mathsf{T}}_{\mu_{k}}h_{k-1}$ $\displaystyle=\widetilde{\mathsf{T}}_{\mu_{k}}^{m-1}\left(r_{\mu_{k}}+\mathbb{P}_{\mu_{k}}h_{k-1}-r_{\mu_{k}}(x^{*})\boldsymbol{1}-\left(\mathbb{P}_{\mu_{k}}h_{k-1}\right)(x^{*})\boldsymbol{1}\right)$ $\displaystyle=\widetilde{\mathsf{T}}_{\mu_{k}}^{m-1}\left(\mathsf{T}_{\mu_{k}}h_{k-1}-\left(\mathsf{T}_{\mu_{k}}h_{k-1}\right)(x^{*})\boldsymbol{1}\right)$ $\displaystyle=\widetilde{\mathsf{T}}_{\mu_{k}}^{m-2}\widetilde{\mathsf{T}}_{\mu_{k}}\left(\mathsf{T}_{\mu_{k}}h_{k-1}-\left(\mathsf{T}_{\mu_{k}}h_{k-1}\right)(x^{*})\boldsymbol{1}\right)$ $\displaystyle=\widetilde{\mathsf{T}}_{\mu_{k}}^{m-2}\left(r_{\mu_{k}}+\mathbb{P}_{\mu_{k}}\left(\mathsf{T}_{\mu_{k}}h_{k-1}-\mathsf{T}_{\mu_{k}}h_{k-1}(x^{*})\right)-r_{\mu_{k}}(x^{*})\boldsymbol{1}-\left(\mathbb{P}_{\mu_{k}}\left(\mathsf{T}_{\mu_{k}}h_{k-1}-\mathsf{T}_{\mu_{k}}h_{k-1}(x^{*})\right)\right)(x^{*})\boldsymbol{1}\right)$ $\displaystyle=\widetilde{\mathsf{T}}_{\mu_{k}}^{m-2}\left(r_{\mu_{k}}+\mathbb{P}_{\mu_{k}}\left(\mathsf{T}_{\mu_{k}}h_{k-1}\right)-\mathsf{T}_{\mu_{k}}h_{k-1}(x^{*})-r_{\mu_{k}}(x^{*})\boldsymbol{1}-\mathbb{P}_{\mu_{k}}\left(\mathsf{T}_{\mu_{k}}h_{k-1}\right)(x^{*})\boldsymbol{1}+\mathsf{T}_{\mu_{k}}h_{k-1}(x^{*})\right)$ $\displaystyle=\widetilde{\mathsf{T}}_{\mu_{k}}^{m-2}\left(\mathsf{T}^{2}_{\mu_{k}}h_{k-1}-\mathsf{T}^{2}_{\mu_{k}}h_{k-1}(x^{*})\boldsymbol{1}\right)$ Iterating, we thus obtain, $h_{\mu_{k}}=\lim_{m\to\infty}\mathsf{T}^{m}_{\mu_{k}}h_{k-1}-\mathsf{T}^{m}_{\mu_{k}}h_{k-1}(x^{*})\boldsymbol{1}$ Substituting back in Equation 21, $\displaystyle g_{k}(i)$ $\displaystyle\geq\left(\mathsf{T}_{\mu_{k}}\left(\lim_{m\to\infty}\mathsf{T}^{m}_{\mu_{k}}h_{k-1}-\mathsf{T}^{m}_{\mu_{k}}h_{k-1}(x^{*})\boldsymbol{1}\right)-\lim_{m\to\infty}\mathsf{T}^{m}_{\mu_{k}}h_{k-1}-\mathsf{T}^{m}_{\mu_{k}}h_{k-1}(x^{*})\boldsymbol{1}\right)(i)-2\delta$ $\displaystyle=\left(\mathsf{T}_{\mu_{k}}\left(\lim_{m\to\infty}\mathsf{T}^{m}_{\mu_{k}}h_{k-1}\right)-\lim_{m\to\infty}\mathsf{T}^{m}_{\mu_{k}}h_{k-1}(x^{*})\boldsymbol{1}-\lim_{m\to\infty}\mathsf{T}^{m}_{\mu_{k}}h_{k-1}-\lim_{m\to\infty}\mathsf{T}^{m}_{\mu_{k}}h_{k-1}(x^{*})\boldsymbol{1}\right)(i)-2\delta$ $\displaystyle=\left(\mathsf{T}_{\mu_{k}}\left(\lim_{m\to\infty}\mathsf{T}^{m}_{\mu_{k}}h_{k-1}\right)-\lim_{m\to\infty}\mathsf{T}^{m}_{\mu_{k}}h_{k-1}\right)(i)-2\delta$ Since $\mathsf{T}_{\mu_{k}}$ is a continuous operator, it is possible to take the limit outside. We thus obtain, $\displaystyle g_{k}(i)$ $\displaystyle\geq\lim_{m\to\infty}\left(\mathsf{T}^{m+1}_{\mu_{k}}h_{k-1}-\mathsf{T}^{m}_{\mu_{k}}h_{k-1}\right)(i)-2\delta$ $\displaystyle=\lim_{m\to\infty}\left(\mathsf{T}^{m}_{\mu_{k}}\mathsf{T}_{\mu_{k}}h_{k-1}-\mathsf{T}^{m}_{\mu_{k}}h_{k-1}\right)(i)-2\delta$ $\displaystyle=\lim_{m\to\infty}\left(\mathbb{P}^{m}_{\mu_{k}}\left(\mathsf{T}_{\mu_{k}}h_{k-1}-h_{k-1}\right)\right)(i)-2\delta$ $\displaystyle=\mathbb{P}^{*}_{\mu_{k}}\left(\mathsf{T}_{\mu_{k}}h_{k-1}-h_{k-1}\right)(i)-2\delta$ Since $\left\|T_{\mu_{k}}h_{k-1}-Th_{k-1}\right\|_{\infty}\leq\epsilon$, $\displaystyle g_{k}(i)$ $\displaystyle\geq\mathbb{P}^{*}_{\mu_{k}}\left(\mathsf{T}h_{k-1}-h_{k-1}\right)(i)-2\delta-\epsilon$ $\displaystyle=\left(\mathbb{P}^{*}_{\mu_{k}}g_{k-1}\right)(i)-2\delta-\epsilon$ Note that from Lemma 3.2 in the mainpaper, we know that $\min_{i,j\in\mathcal{S}}{\mathbb{P}}^{*}_{\mu_{k}}(j|i)>\gamma>0$ Also, since every policy is assumed to induce an irreducible Markov Process, we have a limiting distribution with all positive entries. We thus obtain the following crucial relationship, $\displaystyle g_{k}(i)\geq(1-\gamma)l_{k-1}+\gamma u_{k-1}-2\delta-\epsilon$ (23) Since the above inequation is true for all states $i$, it has to also be true for $\operatorname*{argmin}_{i}\left(\mathsf{T}h_{k}-h_{k}\right)(i)$. Hence, we obtain, $\displaystyle l_{k}\geq(1-\gamma)l_{k-1}+\gamma u_{k-1}-2\delta-\epsilon$ Note that from Lemma 6.4, we know that $u_{k}\geq{J}^{*}$ for all $k\in\mathbb{N}$. Hence we obtain, $\displaystyle l_{k}\geq(1-\gamma)l_{k-1}+\gamma{J}^{*}-2\delta-\epsilon$ Upon rearranging we obtain the result, that is, $\displaystyle\left({J}^{*}-l_{k}\right)\leq(1-\gamma)\left({J}^{*}-l_{k-1}\right)+2\delta+\epsilon$ ∎ We now present the proof of Theorem 3.3. From Lemma 6.5, we thus have, $\left({J}^{*}-l_{k}\right)\leq(1-\gamma)\left({J}^{*}-l_{k-1}\right)+2\delta+\epsilon$ (24) However, we know that, $l_{k}-\epsilon\leq{J}^{*}$ In order to iterate Equation 24, need to ensure the terms are non-negative. Rearranging the terms thus yields, $\left({J}^{*}-l_{k}+\epsilon\right)\leq(1-\gamma)\left({J}^{*}-l_{k-1}+\epsilon\right)+2\delta+(1+\gamma)\epsilon$ Let $\omega=2\delta+(1+\gamma)\epsilon$ and $a_{k}={J}^{*}-l_{k}+\epsilon.$ Then upon iterating, we obtain, $\displaystyle a_{k}$ $\displaystyle\leq(1-\gamma)a_{k-1}+\omega$ $\displaystyle\leq(1-\gamma)((1-\gamma)a_{k-2}+\omega)+\omega$ $\displaystyle\leq\omega\frac{1-(1-\gamma)^{k}}{\gamma}+(1-\gamma)^{k}a_{0}$ Substituting back, we obtain, $\displaystyle\left({J}^{*}-l_{k}+\epsilon\right)\leq\left(\frac{1-(1-\gamma)^{k}}{\gamma}\right)\left(2\delta+(1+\gamma)\epsilon\right)+(1-\gamma)^{k}({J}^{*}-l_{0}+\epsilon)$ (25) $\displaystyle\left({J}^{*}-l_{k}\right)\leq\left(\frac{1-(1-\gamma)^{k}}{\gamma}\right)\left(2\delta+(1+\gamma)\epsilon\right)-\epsilon+(1-\gamma)^{k}({J}^{*}-l_{0}+\epsilon)$ Note that from Lemma 6.4 we know that $l_{k}\leq{J}_{\mu_{k+1}}+\epsilon\leq{J}^{*}+\epsilon$ Hence we get, $\left({J}^{*}-{J}_{\mu_{k+1}}\right)\leq\left({J}^{*}-l_{k}+\epsilon\right)$ Thus we obtain, $\left({J}^{*}-{J}_{\mu_{k+1}}\right)\leq\left(\frac{1-(1-\gamma)^{k}}{\gamma}\right)\left(2\delta+(1+\gamma)\epsilon\right)+(1-\gamma)^{k}({J}^{*}-l_{0}+\epsilon)$ Theorem 3.3 presents an upper bound on the error in terms of the average reward. Recall that, for the optimal relative value function $h$, the following relationship holds: $Th-h=J^{*}\boldsymbol{1}.$ Thus, it is also interesting to understand how $Th_{k}-h_{k}$ behaves under approximate policy iteration. The following proposition characterizes the behavior of this term. ###### Proposition 6.6. The iterates generated from approximate policy iteration algorithm 2 satisfy the following bound: $\displaystyle\left(u_{k-1}-l_{k-1}\right)$ $\displaystyle\leq\underbrace{\left(\frac{1-(1-\gamma)^{k}}{\gamma^{2}}\right)\left(2\delta+(1+\gamma)\epsilon\right)+\frac{2\delta+\epsilon}{\gamma}}_{\text{approximation error}}+\underbrace{\frac{(1-\gamma)^{k}({J}^{*}-l_{0}+\epsilon)}{\gamma}}_{\text{initial condition error}},$ where $u_{k},l_{k}$ are defined in Equation 20. ###### Proof. It is known from Equation 23 that, $g_{k}(i)\geq(1-\gamma)l_{k-1}+\gamma u_{k-1}-2\delta-\epsilon$ This further yields, $l_{k}\geq(1-\gamma)l_{k-1}+\gamma u_{k-1}-2\delta-\epsilon$ From Lemma 6.4, we know that, $l_{k}\leq{J}^{*}+\epsilon$ ${J}^{*}+\epsilon\geq(1-\gamma)l_{k-1}+\gamma u_{k-1}-2\delta-\epsilon$ ${J}^{*}-l_{k}+\epsilon\geq\gamma(u_{k-1}-l_{k-1})-2\delta-\epsilon$ $\gamma(u_{k-1}-l_{k-1})\leq\left({J}^{*}-l_{k}+\epsilon\right)+2\delta+\epsilon$ From Equation 25, we thus obtain, $\gamma(u_{k-1}-l_{k-1})\leq\left(\frac{1-(1-\gamma)^{k}}{\gamma}\right)\left(2\delta+(1+\gamma)\epsilon\right)+(1-\gamma)^{k}({J}^{*}-l_{0}+\epsilon)+2\delta+\epsilon$ We thus obtain the result in Proposition 3.11, $\left(u_{k-1}-l_{k-1}\right)\leq\left(\frac{1-(1-\gamma)^{k}}{\gamma^{2}}\right)\left(2\delta+(1+\gamma)\epsilon\right)+\frac{2\delta+\epsilon}{\gamma}+\frac{(1-\gamma)^{k}({J}^{*}-l_{0}+\epsilon)}{\gamma}$ ∎ ###### Corollary 6.7. The asymptotic behavior of the relative value function iterates is given by $\limsup_{k\to\infty}\left(u_{k}-l_{k}\right)\leq\frac{\epsilon\left(1+2\gamma\right)+2\delta\left(1+\gamma\right)}{\gamma^{2}}$ ##### Comment of the Novelty of our Proof Technique: As mentioned in the main body of the paper, our proof is inspired by the proof of convergence of modified policy iteration in [VdW80]. However, since our algorithm is quite different from modified policy iteration due to the presence of errors in each step of the algorithm, special care is needed to obtain useful performance bounds. Specifically, the impact of the errors at each step have to be carefully bounded to ensure that the performance bounds do not blow up to infinity as it does when we obtain a similar result for the discounted-reward case and let the discount factor go to one. ### 6.3 Proofs from Section 4 Algorithm 2 and its analysis, adapted to the context of $Q$ function with time dependent approximation errors is presented in this section. Most RL algorithms use the state-action relative value function $Q$ instead of the relative state value function $h$ to evaluate a policy. The corresponding Bellman Equation in terms of $Q$ associated with any state-action pair $(s,a)$ is given by ${J}_{\mu}+Q_{\mu}(s,a)=r(s,a)+\left(\mathbb{Q}_{\mu}Q_{\mu}\right)(s,a)$ where $\mathbb{Q}_{\mu}(s^{\prime},a^{\prime}|s,a)=\mu(a^{\prime}|s^{\prime})\mathbb{P}(s^{\prime}|s,a).$ Since we are interested in randomized policies for exploration reasons, the irreducibility assumptions imposed on $\mathbb{P}_{\mu}$ also hold true for the transition kernel $\mathbb{Q}_{\mu}$. The state relative value function $h_{\mu}$ and state-action relative value function $Q_{\mu}$ are related as $h_{\mu}(s)=\sum_{a}\mu(a|s)Q_{\mu}(s,a)$. Consider the following similar definitions of the Bellman operators corresponding to the state-action value function: $(\mathsf{T}^{\mathsf{Q}}_{\mu}Q)(s,a)=r(s,a)+\left(\mathbb{Q}_{\mu}Q\right)(s,a)$ and $(\mathsf{T}^{\mathsf{Q}}Q)(s,a)=r(s,a)+\max_{\mu\in\Pi}\left({\mathbb{Q}}_{\mu}Q\right)(s,a).$ Let $(s^{*},a^{*})$ represent some fixed state. Then the relative Bellman operator, relative to $(s^{*},a^{*})$ is defined as: $\left(\widetilde{\mathsf{T}}^{\mathsf{Q}}_{\mu}Q\right)(s,a)=r(s,a)+\left(\mathbb{Q}_{\mu}Q\right)(s,a)-r(s^{*},a^{*})-\left(\mathbb{Q}_{\mu}Q\right)(s^{*},a^{*}).$ The algorithm, thus adapted to state action relative value function is given below: Algorithm 4 Approximate Policy Iteration: Average Reward 1:Require $Q_{0}\in\mathbb{R}^{n}$ 2:for $k=0,1,2,\ldots$ do 3: 1\. Compute $\mu_{k+1}\in\Pi$ such that $\|\mathsf{T}Q_{k}-\mathsf{T}_{\mu_{k+1}}Q_{k}\|_{\infty}\leq\epsilon_{k}$ 4: 2\. Compute $Q_{k+1}$ such that $\|Q_{k+1}-Q_{\mu_{k+1}}\|_{\infty}\leq\delta_{k+1}$ 5: where $Q_{\mu_{k+1}}=\lim_{m\to\infty}\left({\widetilde{\mathsf{T}}^{\mathsf{Q}}_{\mu_{k+1}}}\right)^{m}Q_{k}$ 6:end for Define $\displaystyle g_{k}(s,a)=\left(\mathsf{T}^{\mathsf{Q}}Q_{k}-Q_{k}\right)(s,a),\qquad\forall(s,a)\in\mathcal{S}\times\mathcal{A},$ and set $\displaystyle l_{k}=\min_{(s,a)\in\mathcal{S}\times\mathcal{A}}g_{k}(s,a),\qquad u_{k}=\max_{(s,a)\in\mathcal{S}\times\mathcal{A}}g_{k}(s,a).$ We prove the convergence in three parts: 1. 1. We prove that $l_{k}-\epsilon_{k}\leq{J}_{\mu_{k+1}}\leq{J}^{*}\leq u_{k}$. 2. 2. We use the result from 1 to prove Lemma 4.1. 3. 3. We iteratively employ the result in Lemma 4.1 to prove Proposition 4.2. #### 6.3.1 Proof of part 1 By definition, we have $\displaystyle l_{k}\boldsymbol{1}$ $\displaystyle\leq\mathsf{T}^{\mathsf{Q}}Q_{k}-Q_{k}.$ Since $\|\mathsf{T}^{\mathsf{Q}}Q_{k}-\mathsf{T}^{\mathsf{Q}}_{\mu_{k+1}}Q_{k}\|_{\infty}\leq\epsilon_{k}$, it follows that $\displaystyle l_{k}\boldsymbol{1}$ $\displaystyle\leq\mathsf{T}^{\mathsf{Q}}_{\mu_{k+1}}Q_{k}-Q_{k}+\epsilon_{k}\boldsymbol{1}$ $\displaystyle=r+\mathbb{Q}_{\mu_{k+1}}Q_{k}-Q_{k}+\epsilon_{k}\boldsymbol{1}.$ Multiplying by $\mathbb{Q}^{*}_{\mu_{k+1}}$, we get $\displaystyle\mathbb{Q}_{\mu_{k+1}}^{*}l_{k}\boldsymbol{1}$ $\displaystyle\leq\mathbb{Q}_{\mu_{k+1}}^{*}r+\mathbb{Q}_{\mu_{k+1}}^{*}\mathbb{Q}_{\mu_{k+1}}Q_{k}-\mathbb{Q}_{\mu_{k+1}}^{*}Q_{k}+\mathbb{Q}_{\mu_{k+1}}^{*}\epsilon_{k}\boldsymbol{1}$ $\displaystyle=\mathbb{Q}_{\mu_{k+1}}^{*}r+\epsilon_{k}\boldsymbol{1}$ (26) where $\mathbb{Q}_{\mu_{k+1}}^{*}$ is a matrix whose rows are identical and are the invariant distribution associated with the probability transition kernel $\mathbb{Q}_{\mu_{k+1}}$. Note that $\mathbb{Q}_{\mu_{k+1}}^{*}\boldsymbol{1}=\boldsymbol{1}$, and that $\mathbb{Q}_{\mu_{k+1}}^{*}\mathbb{Q}_{\mu_{k+1}}=\mathbb{Q}_{\mu_{k+1}}^{*}$. Consider the Bellman Equation corresponding to the state-action relative value function associated with the policy $\mu_{k+1}$: $\displaystyle{J}_{\mu_{k+1}}+Q_{\mu_{k+1}}=r+\mathbb{Q}_{\mu_{k+1}}Q_{\mu_{k+1}}$ It follows that $\displaystyle\mathbb{Q}^{*}_{\mu_{k+1}}{J}_{\mu_{k+1}}+\mathbb{Q}^{*}_{\mu_{k+1}}Q_{\mu_{k+1}}=\mathbb{Q}^{*}_{\mu_{k+1}}r+\mathbb{Q}^{*}_{\mu_{k+1}}\mathbb{Q}_{\mu_{k+1}}Q_{\mu_{k+1}}.$ Since $\mathbb{Q}_{\mu_{k+1}}^{*}\mathbb{Q}_{\mu_{k+1}}=\mathbb{Q}^{*}_{\mu_{k+1}}$, we have that $\displaystyle{J}_{\mu_{k+1}}\boldsymbol{1}=\mathbb{Q}_{\mu_{k+1}}^{*}r.$ Hence, Equation 26 yields $\displaystyle l_{k}\leq{J}_{\mu_{k+1}}+\epsilon_{k}.$ Equivalently, we have $\displaystyle l_{k}-\epsilon_{k}\leq{J}_{\mu_{k+1}}.$ From definition, $\displaystyle{J}^{*}=\max_{\mu\in\Pi}{J}_{\mu}\geq{J}_{\mu_{k+1}}.$ Let $\mu^{*}$ be the policy corresponding to the optimal average reward ${J}^{*}$. Let $Q^{*}$ denote the state-action relative value function associated with the policy $\mu^{*}$. Note that $({J}^{*},Q^{*})$ is the solution of the Bellman optimality equation, i.e., $\displaystyle{J}^{*}+Q^{*}$ $\displaystyle=r+\max_{\mu\in\Pi}\mathbb{Q}_{\mu}Q^{*}$ $\displaystyle=r+\mathbb{Q}_{\mu^{*}}Q^{*}.$ Then, $\displaystyle u_{k}\boldsymbol{1}$ $\displaystyle\geq\mathsf{T}^{\mathsf{Q}}Q_{k}-Q_{k}$ $\displaystyle\stackrel{{\scriptstyle(a)}}{{\geq}}\mathsf{T}^{\mathsf{Q}}_{\mu^{*}}Q_{k}-Q_{k}$ $\displaystyle=r+\mathbb{Q}_{\mu^{*}}Q_{k}-Q_{k},$ where (a) is due to the fact that $\mathsf{T}^{\mathsf{Q}}$ is the Bellman optimality operator. Therefore, we have $\displaystyle\mathbb{Q}_{\mu^{*}}^{*}u_{k}\boldsymbol{1}$ $\displaystyle\geq\mathbb{Q}_{\mu^{*}}^{*}r+\mathbb{Q}_{\mu^{*}}^{*}\mathbb{Q}_{\mu^{*}}Q_{k}-\mathbb{Q}_{\mu^{*}}^{*}Q_{k}$ Equivalently, $\displaystyle u_{k}\boldsymbol{1}$ $\displaystyle\geq\mathbb{Q}^{*}_{\mu^{*}}r$ Therefore, we conclude that $\displaystyle u_{k}$ $\displaystyle\geq{J}^{*}$ Hence, for all $k\in\mathbb{N}$, $\displaystyle l_{k}-\epsilon_{k}\leq{J}_{\mu_{k+1}}\leq{J}^{*}\leq u_{k}$ (27) #### 6.3.2 Proof of Lemma 4.1 Recall that $\displaystyle g_{k}(s,a)=\left(\mathsf{T}^{\mathsf{Q}}Q_{k}-Q_{k}\right)(s,a)\geq\left(\mathsf{T}^{\mathsf{Q}}_{\mu_{k}}Q_{k}-Q_{k}\right)(s,a),$ where the inequality follows by the fact that $\mathsf{T}^{\mathsf{Q}}$ is the Bellman optimality operator. Note that the Bellman operator $\mathsf{T}^{\mathsf{Q}}_{\mu_{k}}$ is shift-invariant, i.e., for all $p\in\mathbb{R}$: $\displaystyle\mathsf{T}^{\mathsf{Q}}_{\mu_{k}}\left(Q_{k}+p\boldsymbol{1}\right)$ $\displaystyle=r+\mathbb{Q}_{\mu_{k}}\left(Q_{k}+p\boldsymbol{1}\right)$ $\displaystyle=r+\mathbb{Q}_{\mu_{k}}Q_{k}+p\boldsymbol{1}$ $\displaystyle=\mathsf{T}^{\mathsf{Q}}_{\mu_{k}}Q_{k}+p\boldsymbol{1}$ Recall that $\left\|Q_{k}-Q_{\mu_{k}}\right\|_{\infty}\leq\delta_{k}$. Hence, we have $\displaystyle g_{k}(s,a)$ $\displaystyle\geq\mathsf{T}^{\mathsf{Q}}_{\mu_{k}}(Q_{\mu_{k}}-\delta_{k}\boldsymbol{1})(s,a)-(Q_{\mu_{k}}(s,a)+\delta_{k})$ $\displaystyle\geq\left(\mathsf{T}^{\mathsf{Q}}_{\mu_{k}}Q_{\mu_{k}}-Q_{\mu_{k}}\right)(s,a)-2\delta_{k}$ (28) Recall that $Q_{\mu_{k}}=\lim_{m\to\infty}\left(\widetilde{\mathsf{T}}^{\mathsf{Q}}_{\mu_{k+1}}\right)^{m}Q_{k}$. Therefore, $\displaystyle\left(\widetilde{\mathsf{T}}^{\mathsf{Q}}_{\mu_{k}}\right)^{m}Q_{k-1}$ $\displaystyle=\left(\widetilde{\mathsf{T}}^{\mathsf{Q}}_{\mu_{k}}\right)^{m-1}\left(\widetilde{\mathsf{T}}^{\mathsf{Q}}_{\mu_{k}}\right)Q_{k-1}$ $\displaystyle=\left(\widetilde{\mathsf{T}}^{\mathsf{Q}}_{\mu_{k}}\right)^{m-1}\left(r+\mathbb{Q}_{\mu_{k}}Q_{k-1}-r(s^{*},a^{*})\boldsymbol{1}-\left(\mathbb{Q}_{\mu_{k}}Q_{k-1}\right)(s^{*},a^{*})\boldsymbol{1}\right)$ $\displaystyle=\left(\widetilde{\mathsf{T}}^{\mathsf{Q}}_{\mu_{k}}\right)^{m-1}\left(\mathsf{T}^{\mathsf{Q}}_{\mu_{k}}Q_{k-1}-\left(\mathsf{T}^{\mathsf{Q}}_{\mu_{k}}Q_{k-1}\right)(s^{*},a^{*})\boldsymbol{1}\right)$ $\displaystyle=\left(\widetilde{\mathsf{T}}^{\mathsf{Q}}_{\mu_{k}}\right)^{m-2}\left(\widetilde{\mathsf{T}}^{\mathsf{Q}}_{\mu_{k}}\right)\left(\mathsf{T}^{\mathsf{Q}}_{\mu_{k}}Q_{k-1}-\left(\mathsf{T}^{\mathsf{Q}}_{\mu_{k}}Q_{k-1}\right)(s^{*},a^{*})\boldsymbol{1}\right)$ $\displaystyle=\left(\widetilde{\mathsf{T}}^{\mathsf{Q}}_{\mu_{k}}\right)^{m-2}\bigg{(}r+\mathbb{Q}_{\mu_{k}}\left(\mathsf{T}^{\mathsf{Q}}_{\mu_{k}}Q_{k-1}-\left(\mathsf{T}^{\mathsf{Q}}_{\mu_{k}}Q_{k-1}\right)(s^{*},a^{*})\boldsymbol{1}\right)$ $\displaystyle\qquad\qquad\qquad-r(s^{*},a^{*})\boldsymbol{1}-\left(\mathbb{Q}_{\mu_{k}}\left(\mathsf{T}^{\mathsf{Q}}_{\mu_{k}}Q_{k-1}-\left(\mathsf{T}^{\mathsf{Q}}_{\mu_{k}}Q_{k-1}\right)(s^{*},a^{*})\boldsymbol{1}\right)\right)(s^{*},a^{*})\boldsymbol{1}\bigg{)}$ $\displaystyle=\left(\widetilde{\mathsf{T}}^{\mathsf{Q}}_{\mu_{k}}\right)^{m-2}\bigg{(}r+\mathbb{Q}_{\mu_{k}}\left(\mathsf{T}^{\mathsf{Q}}_{\mu_{k}}Q_{k-1}\right)-\left(\mathsf{T}^{\mathsf{Q}}_{\mu_{k}}Q_{k-1}\right)(s^{*},a^{*})\boldsymbol{1}$ $\displaystyle\qquad\qquad\qquad-r(s^{*},a^{*})\boldsymbol{1}-\mathbb{Q}_{\mu_{k}}\left(\mathsf{T}^{\mathsf{Q}}_{\mu_{k}}Q_{k-1}\right)(s^{*},a^{*})\boldsymbol{1}+\left(\mathsf{T}^{\mathsf{Q}}_{\mu_{k}}Q_{k-1}\right)(s^{*},a^{*})\boldsymbol{1}\bigg{)}$ $\displaystyle=\left(\widetilde{\mathsf{T}}^{\mathsf{Q}}_{\mu_{k}}\right)^{m-2}\left(\left(\mathsf{T}^{\mathsf{Q}}_{\mu_{k}}\right)^{2}Q_{k-1}-\left(\mathsf{T}^{\mathsf{Q}}_{\mu_{k}}\right)^{2}Q_{k-1}(s^{*},a^{*})\boldsymbol{1}\right)$ Iterating, we thus obtain, $Q_{\mu_{k}}=\lim_{m\to\infty}(\mathsf{T}^{\mathsf{Q}}_{\mu_{k}})^{m}Q_{k-1}-(\mathsf{T}^{\mathsf{Q}}_{\mu_{k}})^{m}Q_{k-1}(s^{*},a^{*})\boldsymbol{1}$ Substituting back in Equation 28, $\displaystyle g_{k}(s,a)$ $\displaystyle\geq\bigg{(}\mathsf{T}^{\mathsf{Q}}_{\mu_{k}}\left(\lim_{m\to\infty}(\mathsf{T}^{\mathsf{Q}}_{\mu_{k}})^{m}Q_{k-1}-(\mathsf{T}^{\mathsf{Q}}_{\mu_{k}})^{m}Q_{k-1}(s^{*},a^{*})\boldsymbol{1}\right)$ $\displaystyle\qquad\qquad\qquad\qquad-\lim_{m\to\infty}\left(\mathsf{T}^{\mathsf{Q}}_{\mu_{k}}\right)^{m}Q_{k-1}-\left(\mathsf{T}^{\mathsf{Q}}_{\mu_{k}}\right)^{m}Q_{k-1}(s^{*},a^{*})\boldsymbol{1}\bigg{)}(s,a)-2\delta_{k}$ $\displaystyle=\bigg{(}\mathsf{T}^{\mathsf{Q}}_{\mu_{k}}\left(\lim_{m\to\infty}(\mathsf{T}^{\mathsf{Q}}_{\mu_{k}})^{m}Q_{k-1}\right)-\lim_{m\to\infty}(\mathsf{T}^{\mathsf{Q}}_{\mu_{k}})^{m}Q_{k-1}(s^{*},a^{*})\boldsymbol{1}$ $\displaystyle\qquad\qquad\qquad\qquad-\lim_{m\to\infty}\left(\mathsf{T}^{\mathsf{Q}}_{\mu_{k}}\right)^{m}Q_{k-1}-\lim_{m\to\infty}\left(\mathsf{T}^{\mathsf{Q}}_{\mu_{k}}\right)^{m}Q_{k-1}(s^{*},a^{*})\boldsymbol{1}\bigg{)}(s,a)-2\delta_{k}$ $\displaystyle=\left(\mathsf{T}^{\mathsf{Q}}_{\mu_{k}}\left(\lim_{m\to\infty}\left(\mathsf{T}^{\mathsf{Q}}_{\mu_{k}}\right)^{m}Q_{k-1}\right)-\lim_{m\to\infty}\left(\mathsf{T}^{\mathsf{Q}}_{\mu_{k}}\right)^{m}Q_{k-1}\right)(s,a)-2\delta_{k}$ Since $\mathsf{T}^{\mathsf{Q}}_{\mu_{k}}$ is a continuous operator, it is possible to take the limit outside. We thus obtain, $\displaystyle g_{k}(s,a)$ $\displaystyle\geq\lim_{m\to\infty}\left(\left(\mathsf{T}^{\mathsf{Q}}_{\mu_{k}}\right)^{m+1}Q_{k-1}-\left(\mathsf{T}^{\mathsf{Q}}_{\mu_{k}}\right)^{m}Q_{k-1}\right)(s,a)-2\delta_{k}$ $\displaystyle=\lim_{m\to\infty}\left(\left(\mathsf{T}^{\mathsf{Q}}_{\mu_{k}}\right)^{m}\mathsf{T}^{\mathsf{Q}}_{\mu_{k}}Q_{k-1}-\left(\mathsf{T}^{\mathsf{Q}}_{\mu_{k}}\right)^{m}Q_{k-1}\right)(s,a)-2\delta_{k}$ $\displaystyle=\lim_{m\to\infty}\left(\mathbb{Q}^{m}_{\mu_{k}}\left(\mathsf{T}^{\mathsf{Q}}_{\mu_{k}}Q_{k-1}-Q_{k-1}\right)\right)(s,a)-2\delta_{k}$ $\displaystyle=\mathbb{Q}^{*}_{\mu_{k}}\left(\mathsf{T}^{\mathsf{Q}}_{\mu_{k}}Q_{k-1}-Q_{k-1}\right)(s,a)-2\delta_{k}.$ Since $\left\|\mathsf{T}^{\mathsf{Q}}_{\mu_{k}}Q_{k-1}-\mathsf{T}^{\mathsf{Q}}Q_{k-1}\right\|_{\infty}\leq\epsilon_{k-1}$, $\displaystyle g_{k}(s,a)$ $\displaystyle\geq\mathbb{Q}^{*}_{\mu_{k}}\left(\mathsf{T}^{\mathsf{Q}}Q_{k-1}-Q_{k-1}\right)(s,a)-2\delta_{k}-\epsilon_{k-1}$ $\displaystyle=\left(\mathbb{Q}^{*}_{\mu_{k}}g_{k-1}\right)(s,a)-2\delta_{k}-\epsilon_{k-1}.$ Recall the definition of $\gamma>0$ in Lemma 4.1. Note that all entries of $\mathbb{Q}^{*}_{\mu_{k}}$ are bounded from below by $\gamma$. Hence, we have $\displaystyle g_{k}(s,a)\geq(1-\gamma)l_{k-1}+\gamma u_{k-1}-2\delta_{k}-\epsilon_{k-1}.$ Since the above inequality is true for all states $(s,a)\in\mathcal{S}\times\mathcal{A}$, we obtain, $\displaystyle l_{k}\geq(1-\gamma)l_{k-1}+\gamma u_{k-1}-2\delta_{k}-\epsilon_{k-1}.$ The above inequality combined with Equation 27, yields $\displaystyle l_{k}\geq(1-\gamma)l_{k-1}+\gamma{J}^{*}-2\delta_{k}-\epsilon_{k-1}.$ Upon rearranging the above inequality, we obtain the result, that is, $\displaystyle\left({J}^{*}-l_{k}\right)\leq(1-\gamma)\left({J}^{*}-l_{k-1}\right)+2\delta_{k}+\epsilon_{k-1}.$ Subsequently, $\displaystyle\left({J}^{*}-\min_{(s,a)}\left(\mathsf{T}^{\mathsf{Q}}Q_{k}-Q_{k}\right)(s,a)\right)$ $\displaystyle\leq(1-\gamma)\left({J}^{*}-\min_{(s,a)}\left(\mathsf{T}^{\mathsf{Q}}Q_{k-1}-Q_{k-1}\right)(s,a)\right)$ $\displaystyle\qquad\qquad\qquad+2\left\|Q_{k}-Q_{\mu_{k}}\right\|_{\infty}+\left\|\mathsf{T}^{\mathsf{Q}}Q_{k-1}-\mathsf{T}^{\mathsf{Q}}_{\mu_{k}}Q_{k-1}\right\|_{\infty}$ #### 6.3.3 Proof of Proposition 4.2 We now prove Proposition 4.2. From Lemma 4.1, we have $\left({J}^{*}-l_{k}\right)\leq(1-\gamma)\left({J}^{*}-l_{k-1}\right)+2\delta_{k}+\epsilon_{k-1}.$ (29) To iterate over Equation 29, we need to ensure the terms are non-negative. Note that by the first part, $l_{k}-\epsilon_{k}\leq{J}^{*}.$ Hence, rearranging the terms thus yields, $\left({J}^{*}-l_{k}+\epsilon_{k}\right)\leq(1-\gamma)\left({J}^{*}-l_{k-1}+\epsilon_{k-1}\right)+\epsilon_{k}+2\delta_{k}+\epsilon_{k-1}-(1-\gamma)\epsilon_{k-1}.$ This is equivalent to $\displaystyle\left({J}^{*}-l_{k}+\epsilon_{k}\right)\leq\left(1-\gamma\right)\left({J}^{*}-l_{k-1}+\epsilon_{k-1}\right)+\left(\epsilon_{k}+\gamma\epsilon_{k-1}\right)+2\delta_{k}.$ Let $a_{k}={J}^{*}-l_{k}+\epsilon_{k}.$ Then, we obtain, $\displaystyle a_{k}$ $\displaystyle\leq(1-\gamma)a_{k-1}+\epsilon_{k}+\gamma\epsilon_{k-1}+2\delta_{k}$ $\displaystyle\leq(1-\gamma)\left[(1-\gamma)a_{k-2}+\epsilon_{k-1}+\gamma\epsilon_{k-2}+2\delta_{k-1}\right]+\epsilon_{k}+\gamma\epsilon_{k-1}+2\delta_{k}$ $\displaystyle=(1-\gamma)^{k}a_{0}+\sum_{\ell=0}^{k-1}(1-\gamma)^{\ell}\epsilon_{k-\ell}+\gamma\sum_{\ell=0}^{k-1}(1-\gamma)^{\ell}\epsilon_{k-1-\ell}+2\sum_{\ell=0}^{k-1}(1-\gamma)^{\ell}\delta_{k-\ell}$ Since $\epsilon_{0}=0$, $\displaystyle a_{k}\leq\left(1-\gamma\right)^{k}a_{0}+\epsilon_{k}+\sum_{\ell=1}^{k-1}\left(1-\gamma\right)^{\ell-1}\epsilon_{k-\ell}+2\sum_{\ell=0}^{k-1}\left(1-\gamma\right)^{\ell}\delta_{k-\ell}$ Substituting for $a_{k}$, we get $\displaystyle\left({J}^{*}-l_{k}\right)\leq\left(1-\gamma\right)^{k}\left({J}^{*}-l_{0}\right)+\sum_{\ell=1}^{k-1}\left(1-\gamma\right)^{\ell-1}\epsilon_{k-\ell}+2\sum_{\ell=0}^{k-1}\left(1-\gamma\right)^{\ell}\delta_{k-\ell}.$ Since from part 1, we know that ${J}^{*}-{J}_{\mu_{k+1}}\leq{J}^{*}-l_{k}+\epsilon_{k}$, we have $\displaystyle{J}^{*}-{J}_{\mu_{k+1}}$ $\displaystyle\leq\left(1-\gamma\right)^{k}\left[{J}^{*}-\min_{i}\left(\mathsf{T}^{\mathsf{Q}}Q_{0}-Q_{0}\right)(i)\right]$ $\displaystyle+\sum_{\ell=1}^{k-1}\left(1-\gamma\right)^{\ell-1}\left[\left\|\mathsf{T}^{\mathsf{Q}}_{\mu_{k+1-\ell}}Q_{k-\ell}-\mathsf{T}^{\mathsf{Q}}Q_{k-\ell}\right\|_{\infty}\right]+{\left\|\mathsf{T}^{\mathsf{Q}}_{\mu_{k+1}}Q_{k}-\mathsf{T}^{\mathsf{Q}}Q_{k}\right\|_{\infty}}$ $\displaystyle+2\sum_{\ell=0}^{k-1}\left(1-\gamma\right)^{\ell}\left[\left\|Q_{k-\ell}-Q_{\mu_{k-\ell}}\right\|_{\infty}\right].$ Taking expectation, we have $\displaystyle\mathbb{E}\left[{J}^{*}-{J}_{\mu_{k+1}}\right]$ $\displaystyle\leq\left(1-\gamma\right)^{k}\mathbb{E}\left[{J}^{*}-\min_{i}\left(\mathsf{T}^{\mathsf{Q}}Q_{0}-Q_{0}\right)(i)\right]$ $\displaystyle+\sum_{\ell=1}^{k-1}\left(1-\gamma\right)^{\ell-1}\mathbb{E}\left[\left\|\mathsf{T}^{\mathsf{Q}}_{\mu_{k+1-\ell}}Q_{k-\ell}-\mathsf{T}^{\mathsf{Q}}Q_{k-\ell}\right\|_{\infty}\right]+\mathbb{E}\left[\left\|\mathsf{T}^{\mathsf{Q}}_{\mu_{k+1}}Q_{k}-\mathsf{T}^{\mathsf{Q}}Q_{k}\right\|_{\infty}\right]$ $\displaystyle+2\sum_{\ell=0}^{k-1}\left(1-\gamma\right)^{\ell}\mathbb{E}\left[\left\|Q_{k-\ell}-Q_{\mu_{k-\ell}}\right\|_{\infty}\right].$ #### 6.3.4 TD Learning for $Q$ function One RL algorithm to estimate the state-action relative value function with linear function approximation, associated with a policy $\mu$ is TD($\lambda$). Finite-time sample complexity of TD($\lambda$) has been recently studied in [ZZM21]. Invoking their results, restating it in terms of the state-action relative value function rather than the state relative value function, we characterize the sample complexity required to achieve certain accuracy in the approximation. We estimate $Q_{\mu}(s,a)$ linearly by $\phi(s,a)^{\top}\theta^{*}_{\mu}$, for some $\theta^{*}_{\mu}\in\mathbb{R}^{d}$, where $\phi(s,a)=$ $[\phi_{1}(s,a),\cdots,\phi_{d}(s,a)]^{T}\in\mathbb{R}^{d}$ is the feature vector associated with $(s,a)\in\mathcal{S}\times\mathcal{A}$. Here, $\theta^{*}_{\mu}$ is a fixed point of $\Phi\theta=\Pi_{D,W_{\phi}}\mathsf{T}^{\mathsf{Q},\lambda}_{\mu}\Phi\theta$, where $\Phi$ is an $|\mathcal{S}|\times d$ matrix whose $k$-th column is $\phi_{k}$, $\mathsf{T}^{\mathsf{Q},\lambda}_{\mu}=(1-\lambda)\sum_{m=0}^{\infty}\lambda^{m}\left(\widehat{\mathsf{T}}^{\mathsf{Q}}_{\mu}\right)^{m+1}$ where $\widehat{\mathsf{T}}^{\mathsf{Q}}_{\mu}=\mathsf{T}^{\mathsf{Q}}_{\mu}-{J}_{\mu}\boldsymbol{1}$, $D$ is the diagonal matrix with diagonal elements given by the stationary distribution of the policy $\mu$ on $\mathcal{S}\times\mathcal{A}$, and $\Pi_{D,W_{\phi}}=\Phi(\Phi^{\top}D\Phi)^{-1}\Phi^{\top}D$ is the projection matrix onto $W_{\Phi}=\\{\Phi\theta|\theta\in\mathbb{R}^{d}\\}$ with respect to the norm $\|\cdot\|_{D}$. Note that this fixed point equation may have multiple solutions. In particular, if $\Phi\theta_{e}=\boldsymbol{1}$, then $\theta^{*}_{\mu}+p\theta_{e}$ is also a solution for any $p\in\mathbb{R}$. Hence, we focus on the set of equivalent classes $E$ where we say $\theta_{1}\sim\theta_{2}$ if $\Phi(\theta_{1}-\theta_{2})=\boldsymbol{1}$. The value of $\mathbb{E}[\|\Pi_{2,E}(\theta-\theta^{*}_{\mu})\|_{2}^{2}]$ characterizes the accuracy of our approximation, where $\Pi_{2,E}$ is the projection onto $E$ with respect to $\|\cdot\|_{2}$. Suppose that $\\{\phi_{1},\phi_{2},\cdots,\phi_{d}\\}$ are linearly independent and that $\max_{(s,a)\in\mathcal{S}\times\mathcal{A}}\|\phi(s,a)\|_{2}\leq 1$. Upon making a new observation $(s_{t+1},a_{t+1})$ for $t\geq 0$, the average reward $TD(\lambda)$ uses the following update equations: $\displaystyle\text{Eligibility trace: }z_{t}=\lambda z_{t-1}+\phi(s_{t},a_{t})$ (30) $\displaystyle\text{TD error: }d_{t}=r(s_{t},a_{t})-{J}_{t}+\phi(s_{t+1},a_{t+1})^{\top}\theta_{t}$ $\displaystyle\qquad\qquad\qquad\qquad-\phi(s_{t},a_{t})^{\top}\theta_{t}$ $\displaystyle\text{average-reward update: }{J}_{t+1}={J}_{t}+c_{\alpha}\beta_{t}(r(s_{t},a_{t})-{J}_{t}),$ $\displaystyle\text{parameter update: }\theta_{t+1}=\theta_{t}+\beta_{t}\delta_{t}(\theta)z_{t},$ where $\beta_{t}$ is the scalar step size, $c_{\alpha}>0$ is a constant, and $z_{-1}$ is initialized to be zero. Following the same argument as in [ZZM21], we get the following theorem which characterizes the number of samples required to get a certain accuracy in the approximation. ###### Theorem 6.8. Let $\beta_{t}=\frac{c_{1}}{t+c_{2}}$, and suppose that the positive constants $c_{1}$, $c_{2}$, and $c_{\alpha}$ are properly chosen. Then, the number of samples required to ensure an approximation accuracy of $\mathbb{E}[\|\Pi_{2,E}(\theta-\theta^{*}_{\mu})\|_{2}^{2}]\leq\delta$, is given by $\displaystyle\tau=O\left(\frac{K\log(\frac{1}{\Delta})\|\theta^{*}_{\mu}\|_{2}^{2}}{\Delta^{4}(1-\lambda)^{4}\delta^{2}}\right),$ where $K$ is the minimum mixing time constant across all probability transition matrices induced by the policies, $\displaystyle\Delta=\min_{\|\theta\|_{2}=1,\theta\in E}\theta^{T}\Phi^{T}D(I-\mathbb{Q}^{\lambda})\Phi\theta>0,$ and $\mathbb{Q}^{\lambda}=(1-\lambda)\sum_{m=0}^{\infty}\lambda^{m}\mathbb{Q}^{m+1}$. ###### Proof. The proof is a simple adaptation of the proof in [ZZM21] for the state relative value function. ∎ Let $Q_{k}=\Phi\theta_{k}$, where $\theta_{k}$ is the output of TD($\lambda$) algorithm with $T$ samples. Then, we have $\displaystyle\|Q_{k}-Q_{\mu_{k}}\|_{\infty}\leq\|\Phi(\theta_{k}-\theta_{\mu_{k}}^{*})\|_{\infty}+\|\Phi\theta_{\mu_{k}}^{*}-Q_{\mu_{k}}\|_{\infty},$ where the second term above is the error associated with the choice of $\Phi$. Since $\max_{(s,a)\in\mathcal{S}\times\mathcal{A}}\|\phi(s,a)\|_{2}\leq 1$, we get $\displaystyle\delta_{k}=d\|\theta_{k}-\theta_{\mu_{k}}^{*}\|_{\infty}+\|\Phi\theta_{\mu_{k}}^{*}-Q_{\mu_{k}}\|_{\infty}.$ Using Theorem 6.8, we get a bound on the value of $\|\theta_{k}-\theta_{\mu_{k}}^{*}\|_{\infty}$. Note that if $\Phi\theta_{e}=\boldsymbol{1}$ for some $\theta_{e}\in\mathbb{R}^{d}$, then we can pick $\theta_{k+1}$ and $\theta_{\mu_{k+1}}^{*}$ so that $\phi(s^{*},a^{*})^{T}\theta_{k+1}=\phi(s^{*},a^{*})^{T}\theta_{\mu_{k+1}}^{*}=0$. Otherwise, there is a unique choice for $\theta_{\mu_{k+1}}^{*}$, and we have $E=\mathbb{R}^{d}$. The expected value of learning error can thus be expressed as: $\displaystyle\mathbb{E}\left[\delta_{k}\right]=\mathbb{E}\left[\delta_{\mathsf{TD},k}\right]+\mathbb{E}\left[\delta_{0,k}\right]$ where ${\delta_{\mathsf{TD},k}}=d\|\theta_{k+1}-\theta_{\mu_{k+1}}^{*}\|_{\infty}$ represents the TD learning error of the parameter vector $\theta_{\mu_{k+1}}^{*}$ and ${\delta_{0,k}}=\|\Phi\theta_{\mu_{k+1}}^{*}-Q_{\mu_{k+1}}\|_{\infty}$ represents the function approximation error associated with the span of the feature vector matrix $\Phi$. ${\delta_{\mathsf{TD},k}}$ has a direct dependence on the number of samples utilized for TD learning and the mixing time of the corresponding Markov Chain induced by the policy $\mu_{k}.$ More precisely, from Theorem 3.1, $\displaystyle\mathbb{E}\left[\delta_{\mathsf{TD},k}\right]=O\left(\sqrt{\frac{K\log(\frac{1}{\Delta})\|\theta^{*}_{\mu_{k}}\|_{2}^{2}}{\Delta^{4}(1-\lambda)^{4}\tau}}\right),$ Hence as long as the mixing constant $\Delta$ is uniformly greater than zero, and the feature vectors $\theta^{*}_{\mu_{k}}$ are such that they are uniformly upper bounded in $k$, it is true that $\displaystyle\mathbb{E}\left[\delta_{\mathsf{TD},k}\right]=\frac{C}{\sqrt{\tau}},$ for some constant $C>0$. Next, we prove the corollaries for specific policy improvement algorithms. Since the policy improvement part does not depend on whether the problem is a discounted-reward problem or an average reward problem, we can borrow the results from [CM22] to identify $\epsilon_{k}$ in each of the corollaries. #### 6.3.5 Proof of Corollary 4.3 Recall the greedy update rule. Let $a^{*}=\operatorname*{argmax}_{a^{\prime}}Q_{k}(s,a^{\prime})$. Given a parameter $\beta>0$, at any time instant $k$, the greedy policy update $\mu_{k+1}$ is given by: $\displaystyle\mu_{k+1}(a|s)=\begin{cases}\frac{1}{\beta|\mathcal{A}|},&\text{if }a\neq a^{*}\\\ \frac{1}{\beta|\mathcal{A}|}+1-\frac{1}{\beta},&\text{if }a=a^{*}\end{cases}$ (31) Let $\eta_{k}=\max_{\begin{subarray}{c}s^{\prime}\in\mathcal{S}\\\ a^{\prime}\in\mathcal{A}\end{subarray}}|\mathbb{Q}_{k}(s^{\prime},a^{\prime})|$. The policy improvement approximation can be shown to be the following: $\displaystyle\epsilon_{k}$ $\displaystyle=\left(\mathsf{T}^{\mathsf{Q}}Q_{k}-\mathsf{T}^{\mathsf{Q}}_{\mu_{k+1}}Q_{k}\right)(s,a)$ $\displaystyle\leq\sum_{s^{\prime}\in\mathcal{S}}\mathbb{P}(s^{\prime}|s,a)\frac{2}{\beta}\max_{a^{\prime}\in\mathcal{A}}|Q_{k}(s^{\prime},a^{\prime})|$ $\displaystyle\leq\frac{2\eta_{k}}{\beta}.$ Recall the error due to TD Learning $\displaystyle\mathbb{E}\left[\delta_{k}\right]=\mathbb{E}\left[\delta_{\mathsf{TD},k}\right]+\mathbb{E}\left[\delta_{0,k}\right].$ Let $\overline{\delta}_{0}=\max_{t}\delta_{0,t}$. Then we obtain the following: $\displaystyle\mathbb{E}\left[\delta_{k}\right]=\overline{\delta}_{0}+\frac{C}{\sqrt{\tau}}.$ Substituting for expressions in Proposition 4.2, we obtain, $\displaystyle\mathbb{E}\left[J^{*}-J_{\mu_{T+1}}\right]$ $\displaystyle\leq\left(1-\gamma\right)^{T}\mathbb{E}\left[{J}^{*}-\min_{i}\left(\mathsf{T}^{\mathsf{Q}}Q_{0}-Q_{0}\right)(i)\right]$ $\displaystyle+2\sum_{\ell=0}^{T-1}\left(1-\gamma\right)^{\ell}\left(\overline{\delta}_{0}+\frac{C}{\sqrt{\tau}}\right)+\sum_{\ell=1}^{T-1}\left(1-\gamma\right)^{\ell-1}\frac{2\eta_{T-\ell}}{\beta}+\frac{2\eta_{T}}{\beta}.$ Let $c_{0}=\mathbb{E}\left[{J}^{*}-\min_{i}\left(\mathsf{T}^{\mathsf{Q}}Q_{0}-Q_{0}\right)(i)\right]$ be the error associated with the initial condition. Let $\overline{\eta}=\max_{t}\eta_{t}$ ($\overline{\eta}$ can be the uniform upper bound of the estimates $Q_{k}$ of relative state action value function $Q_{\mu_{k}}$ over $k.$) Then we obtain the result in the corollary, $\displaystyle\mathbb{E}\left[J^{*}-J_{\mu_{T+1}}\right]$ $\displaystyle\leq(1-\gamma)^{T}c_{0}+\frac{2}{\gamma}\left(\overline{\delta}_{0}+\frac{C}{\sqrt{\tau}}\right)+\left(\frac{1+\gamma}{\gamma}\right)\frac{2\overline{\eta}}{\beta}.$ (32) #### 6.3.6 Proof of Corollary 4.4 Recall the softmax policy update. Given a parameter $\beta>0$, at any time instant $k$, the Softmax update policy $\mu_{k+1}$ is: $\mu_{k+1}(a|s)=\frac{\exp\left(\beta Q_{k}(s,a)\right)}{\sum_{a^{\prime}\in\mathcal{A}}\exp\left(\beta Q_{k}(s,a^{\prime})\right)}.$ (33) The policy improvement approximation for this update rule turns out to be time independent and is given by: $\displaystyle\epsilon_{k}=\left(\mathsf{T}^{\mathsf{Q}}Q_{k}-\mathsf{T}^{\mathsf{Q}}_{\mu_{k+1}}Q_{k}\right)(s,a)\leq\frac{\log|\mathcal{A}|}{\beta}.$ Given $\overline{\delta}_{0}=\max_{t}\delta_{0,t}$ the following is true, $\displaystyle\mathbb{E}\left[\delta_{k}\right]=\overline{\delta}_{0}+\frac{C}{\sqrt{\tau}}.$ Substituting for expressions in Proposition 4.2, we obtain, $\displaystyle\mathbb{E}\left[J^{*}-J_{\mu_{T+1}}\right]$ $\displaystyle\leq\left(1-\gamma\right)^{T}\mathbb{E}\left[{J}^{*}-\min_{i}\left(\mathsf{T}^{\mathsf{Q}}Q_{0}-Q_{0}\right)(i)\right]$ $\displaystyle+2\sum_{\ell=0}^{T-1}\left(1-\gamma\right)^{\ell}\left(\overline{\delta}_{0}+\frac{C}{\sqrt{\tau}}\right)+\sum_{\ell=1}^{T-1}\left(1-\gamma\right)^{\ell-1}\frac{\log|\mathcal{A}|}{\beta}+\frac{\log|\mathcal{A}|}{\beta}.$ Let $c_{0}=\mathbb{E}\left[{J}^{*}-\min_{i}\left(\mathsf{T}^{\mathsf{Q}}Q_{0}-Q_{0}\right)(i)\right]$ be the error associated with the initial condition. Then we obtain the result in the corollary, $\displaystyle\mathbb{E}\left[J^{*}-J_{\mu_{T+1}}\right]$ $\displaystyle\leq(1-\gamma)^{T}c_{0}+\frac{2}{\gamma}\left(\overline{\delta}_{0}+\frac{C}{\sqrt{\tau}}\right)+\left(\frac{1+\gamma}{\gamma}\right)\frac{\log\left(|\mathcal{A}|\right)}{\beta}.$ (34) #### 6.3.7 Proof of Corollary 4.5 Recall the mirror descent update. Given $\beta>0$ the mirror descent update is given by: $\displaystyle\mu_{k+1}(a|s)=\frac{\mu_{k}(a|s)e^{\beta Q_{k}(s,a)}}{\sum_{a^{\prime}\in\mathcal{A}}\mu_{k}(a^{\prime}|s)e^{\beta Q_{k}(s,a^{\prime})}}$ (35) The policy improvement error for this update rule and is given by: $\displaystyle\epsilon_{k}$ $\displaystyle=\left(\mathsf{T}^{\mathsf{Q}}Q_{k}-\mathsf{T}^{\mathsf{Q}}_{\mu_{k+1}}Q_{k}\right)(s,a)$ $\displaystyle\leq\frac{1}{\beta}\log\frac{1}{\min_{s\in\mathcal{S}}\mu_{k+1}\left(a^{*}(s)|s\right)}.$ where $a^{*}(s)$ is the optimal action at state $s.$ Let $\omega_{k+1}=\min_{s\in\mathcal{S}}\mu_{k+1}\left(a^{*}(s)|s\right)$ Given that the TD learning error is of the form $\displaystyle\mathbb{E}\left[\delta_{k}\right]=\overline{\delta}_{0}+\frac{C}{\sqrt{\tau}},$ Substituting for expressions in Proposition 4.2, we obtain, $\displaystyle\mathbb{E}\left[J^{*}-J_{\mu_{T+1}}\right]$ $\displaystyle\leq\left(1-\gamma\right)^{T}\mathbb{E}\left[{J}^{*}-\min_{i}\left(\mathsf{T}^{\mathsf{Q}}Q_{0}-Q_{0}\right)(i)\right]$ $\displaystyle+2\sum_{\ell=0}^{T-1}\left(1-\gamma\right)^{\ell}\left(\overline{\delta}_{0}+\frac{C}{\sqrt{\tau}}\right)+\sum_{\ell=1}^{T-1}\left(1-\gamma\right)^{\ell-1}\frac{2}{\beta}\log{\frac{1}{\omega_{T-\ell}}}+\frac{2}{\beta}\log{\frac{1}{\omega_{T}}}.$ Let $c_{0}=\mathbb{E}\left[{J}^{*}-\min_{i}\left(\mathsf{T}^{\mathsf{Q}}Q_{0}-Q_{0}\right)(i)\right]$ be the error associated with the initial condition. Let $\underline{\omega}=\min_{t}\omega_{t}$ Then we obtain the result in the corollary, $\displaystyle\mathbb{E}\left[J^{*}-J_{\mu_{T+1}}\right]$ $\displaystyle\leq(1-\gamma)^{T}c_{0}+\frac{2}{\gamma}\left(\overline{\delta}_{0}+\frac{C}{\sqrt{\tau}}\right)+\left(\frac{1+\gamma}{\gamma\beta}\right)\log\left(\frac{1}{\underline{\omega}}\right).$ (36) #### 6.3.8 Regret Analysis Recall the pseudo regret defined as follows: $\displaystyle R_{\text{PS}}(K)=\sum_{t=1}^{K}\left(J^{*}-J_{\mu_{t}}\right),$ (37) where $\mu_{t}$ us the policy utilized at time $t$ and obtained through mirror descent, and $K$ is the time horizon. Since $K=\tau T$, we have $\displaystyle R_{\text{PS}}(K)=\tau\sum_{t=1}^{T}\left(J^{*}-J_{\mu_{t}}\right).$ (38) Substituting for $J^{*}-J_{\mu_{t}}$, it is true that $\displaystyle R_{\text{PS}}(K)=\tau\left(\sum_{t=1}^{T}(1-\gamma)^{T}c_{0}+\frac{1+\gamma}{\gamma\beta}\log\left(\frac{1}{\omega}\right)+\frac{2}{\gamma}\left(\overline{\delta}_{0}+\frac{c_{3}}{\sqrt{\tau}}\right)\right).$ (39) Let $\beta=\sqrt{\tau}$. The above regret can be simplified as: $\displaystyle R_{\text{PS}}(K)=\frac{\tau c_{0}}{\gamma}+\frac{K}{\gamma\sqrt{\tau}}\left((1+\gamma)\log\left(\frac{1}{\omega}\right)+2c_{3}\right)+\frac{2K\overline{\delta}_{0}}{\gamma}$ (40) Let $c_{5}=(1+\gamma)\log(1/\omega)+2c_{3}$. We then have $\displaystyle R_{\text{PS}}(K)=\frac{\tau c_{0}}{\gamma}+\frac{Kc_{5}}{\gamma\sqrt{\tau}}+\frac{2K\overline{\delta}_{0}}{\gamma}.$ (41) Optimizing for regret involves equating $\frac{\tau c_{0}}{\gamma}$ and $\frac{Kc_{5}}{\gamma\sqrt{\tau}}$. This yields $\tau=\left(\frac{Kc_{5}}{c_{0}}\right)^{2/3}$. Further substituting for $\tau$ yields $\displaystyle R_{\text{PS}}(K)$ $\displaystyle=\left(\frac{Kc_{3}}{c_{0}}\right)^{2/3}\cdot\frac{c_{0}}{\gamma}+\frac{Kc_{5}c_{0}^{1/3}}{\gamma(Kc_{5})^{1/3}}+\frac{2K\overline{\delta}_{0}}{\gamma}$ (42) $\displaystyle=\frac{(Kc_{5})^{2/3}\cdot c_{0}^{1/3}}{\gamma}+\frac{(Kc_{5})^{2/3}\cdot c_{0}^{1/3}}{\gamma}+\frac{2K\overline{\delta}_{0}}{\gamma}$ (43) $\displaystyle=\frac{2c_{5}^{2/3}\cdot c_{0}^{1/3}\cdot K^{2/3}}{\gamma}+\frac{2K\overline{\delta}_{0}}{\gamma}.$ (44) Let $c_{6}=2c_{5}^{2/3}c_{0}^{1/3}$. We have $\displaystyle R_{\text{PS}}(K)=\frac{K^{2/3}c_{6}}{\gamma}+\frac{2K\overline{\delta_{0}}}{\gamma}.$ (45)
# Improving Mitosis Detection Via UNet-based Adversarial Domain Homogenizer ††thanks: * Indicates equal contribution. 1st Tirupati Saketh Chandra Electrical Engineering Indian Institute of Technology Bombay Mumbai, India 0000-0002-0325-5821 2nd Sahar Almahfouz Nasser Electrical Engineering Indian Institute of Technology Bombay Mumbai, India 0000-0002-5063-9211 3rd Nikhil Cherian Kurian Electrical Engineering Indian Institute of Technology Bombay Mumbai, India 0000-0003-1713-0736 4th Amit Sethi Electrical Engineering Indian Institute of Technology Bombay Mumbai, India 0000-0002-8634-1804 ###### Abstract The effective localization of mitosis is a critical precursory task for deciding tumor prognosis and grade. Automated mitosis detection through deep learning-oriented image analysis often fails on unseen patient data due to inherent domain biases. This paper proposes a domain homogenizer for mitosis detection that attempts to alleviate domain differences in histology images via adversarial reconstruction of input images. The proposed homogenizer is based on a U-Net architecture and can effectively reduce domain differences commonly seen with histology imaging data. We demonstrate our domain homogenizer’s effectiveness by observing the reduction in domain differences between the preprocessed images. Using this homogenizer, along with a subsequent retina-net object detector, we were able to outperform the baselines of the 2021 MIDOG challenge in terms of average precision of the detected mitotic figures. ###### Index Terms: MIDOG, domain generalization, mitosis detection, domain homogenizer, auto- encoder. ## I Introduction and Related Work The demand for robust models which perform well on unseen testing data has increased rationally in the last two years, especially in the medical domain. In many practical applications of machine learning models domain shift occurs after training, wherein the characteristics of the test data are different from the training data. Particularly in the application of deep neural networks (DNNs) to pathology images, the test data may have different colors, stain concentration, and magnification compared to what the DNN was trained on due to changes in scanner, staining reagents, and sample preparation protocols. MIDOG2021 [20] was the first MICCAI challenge that addressed the problem of domain shift as it is one of the reasons behind the failure of machine learning models after training, including those for mitosis detection techniques when tested on data from a different domain (scanner) than the one used during the training. MIDOG2021 was the first MICCAI challenge that addressed the problem of domain shift due to change in scanner [20]. Domain generalization is the set of techniques that improve the prediction accuracy of machine learning models on data from new domains without assuming access to those data during training. Proposing and testing various domain generalization techniques was the main goal of the MIDOG2021 challenge. In this section, we introduce the most significant notable solutions for the MIDOG20201 challenge before introducing our proposed method. In [1] the authors modified RetinaNet network for mitosis detection [2] for mitosis detection by adding a domain classification head and a gradient reversal layer to encourage domain agnosticism. In this work, they used a pre- trained Resnet18 for the encoder. For their discriminator, it was a simple sequence of three convolutional blocks and a fully connected layer. The domain classifier was placed at the bottleneck of the encoder. Breen et al [3] proposed a U-Net type architecture that outputs the probability map of the mitotic figures. These probabilities get converted into bounding boxes around the mitotic figures. They used a neural style transfer (NST) as a domain adaptation technique. This technique casts the style of one image on the content of another. The method proposed by [4] consists of two parts, a patch selection and a style transfer module. To learn the styles of images from different scanners, they used a StarGAN. A two steps domain-invariant mitotic detection method was proposed by [5]. This method is based on Fast RCNN [6]. For domain generalization purposes they used StainTools software [17] to augment the images. StainTools package decomposes the image into two matrices, a concentration matrix C and a stain matrix S. By combining the C and S matrices from different images they produced the augmented images. A cascaded pipeline of a Mask RCNN [9] followed by a classification ensemble was proposed by [7] to detect mitotic candidates. A Cycle GAN [8] was used to transfer every scanner domain to every other scanner domain. In [10] the authors used a stain normalization method proposed by [11] as a preprocessing step for the images. Others like [12] merged Hard negative mining with immense data augmentation for domain generalization was proposed by [12]. Stain normalization techniques such as [14] and [15] were used in [13] to account for the domain difference between images. Almahfouz Nasser et al., [16] proposed an autoencoder trained adversarially on the sources of domain variations. This autoencoder makes the appearance of images uniform across different domains. In this paper, we present our work which is an extension of our proposed method for MIDOG2021 [16]. Our contribution is three-fold and can be summarised in as follows. Firstly, we modified [16] by shifting the domain classifier from the latent space to the end of the autoencoder, which improved the results drastically. Secondly, we showed the importance of perceptual loss in preserving the semantic information which affects the final accuracy of the object detection part. Finally, unlike our previous work, training the auto- encoder along with the object detection network end-to-end improved the quality of the homogenized outputs substantially. ## II Methodology ### II-A Notations In domain generalization there are source (seen) domains which are shown to the model during training, and there are some target (unseen) domains which are used only during testing. Labelled samples from the source domains are represented by $D_{ls}$=$\\{(x^{ls}_{i},y^{ls}_{i})\\}^{N_{ls}}_{i=1}$, unlabelled source domains are represented by $D_{us}$=$\\{(x^{us}_{i})\\}^{N_{us}}_{i=1}$ and labelled target domains are represented by $D_{lus}$=$\\{(x^{lus}_{i},y^{lus}_{i})\\}^{N_{lus}}_{i=1}$. Let the unlabelled images from all subsets be represented by $D_{all}$=$D_{ls}$ $\cup$ $D_{us}$ $\cup$ $D_{lus}$. ### II-B Adversarial end-to-end trainable architecture Inspired by the work of [18], we have used an encoder-decorder network to translate the patches from different domains (scanners) to a common space. The translated images are then passed through RetinaNnet [2] for object detection. The architecture also consists of an adversarial head with domain classification as an auxiliary task. This head encourages the encoder-decoder network to erase all the domain- specific information. The architecture of our method is as shown in figure 1. Figure 1: The pipeline of our proposed method for mitosis detection. ### II-C Training Objectives The object detection loss consists of bounding box loss ($\mathcal{L}_{bb}$) and instance classification loss ($\mathcal{L}_{inst}$). The bounding box loss ($\mathcal{L}_{bb}$) is computed as smooth L1 loss and the focal loss function is used for the instance classification ($\mathcal{L}_{inst}$). In order to ensure that the images translated by encoder-decoder network contains the semantic information we have used a perceptual loss ($\mathcal{L}_{percp}$). We have used the perceptual loss based on pretrained VGG-16, which is proposed in [19]. At the end of the adversarial head we have used standard cross entropy loss ($\mathcal{L}_{CE}$) for domain classification. The overall loss for the end-to-end training is given by, $\mathcal{L}=\mathcal{L}_{bb}+\mathcal{L}_{inst}+\lambda_{1}\mathcal{L}_{percep}+\lambda_{2}\mathcal{L}_{CE}$ (1) ## III Data and Experiments ### III-A Dataset The experiments were conducted on MIDOG 2021 dataset [20] which consists of 50 whole slide images of breast cancer from four scanners namely Hamamatsu XR NanoZoomer 2.0, Hamamatsu S360, Aperio ScanScope CS2, and Leica GT450 forming four domains. Two classes of objects are to be detected namely mitotic figures and hard negatives. The whole slide images from scanners other than the Leica GT450 areis labelled. Small patches of size 512 x 512 are mined for supervised end-to-end training such that the cells belonging to at least one of the mitotic figures or hard negatives are present in the patch. The seen and unseen domains i.e., the scanners are $D_{ls}$={Hamamatsu XR NanoZoomer 2.0, the Hamamatsu S360}, $D_{us}$={Leica GT450}, $D_{lus}$={Aperio ScanScope CS2} (refer II-A for notations.) ### III-B Implementation Details The model is implemented using Pytorch [21] library [21]. For supervised end- to-end training a batch size of 12 is used with equal number of patches being included from each scanner. Here the model is trained using FastAI [22] library default settings with an initial learning rate of $1e^{-4}$. In the equation 1 we have set the values of hyperparameters as $\lambda_{1}$=10 and $\lambda_{2}$=25. ### III-C Results Two classes of objects – hard negatives, and mitotic figures – are detected. The models are evaluated on $D_{ls}\cup D_{lus}$. Average precision at IoU threshold of 0.5 is used as metric for evaluation. End-to-end training with (AEC_RetinaNet + Pecp) and without using perceptual loss (AEC_RetinaNet) were tried. The results are compared with the reference algorithm DA_RetinaNet [23], RetinaNet [24] with and without data augmentation. The results obtained are as shown in table I. TABLE I: Results obtained using end -to -end training of models Model | AP-Hard Neg | AP-Mitotic figures | mAP ---|---|---|--- RetinaNet | 0.196 | 0.352 | 0.274 RetinaNet + Aug | 0.238 | 0.619 | 0.429 DA_RetinaNet | 0.347 | 0.655 | 0.501 AEC_RetinaNet | 0.289 | 0.448 | 0.369 AEC_RetinaNet + Pecp | 0.248 | 0.72 | 0.484 The count of mitotic figures is an important clinical goal. So, the performance on the class of mitotic figures was our focused. The results in the above table show that the newly designed end-to-end training architectures performs better than the reference algorithm and the basic Retina-Net based algorithms. The best performance is in terms of recall metric, which is evident from the plot shown in figure 2. The perceptual loss added at the output of the decoder helps in retaining the semantic information. This information which helps in better object detection. This is also validated by higher AP score obtained when perceptual loss component is added. As shown in figure 3 the modified domain homogenizer produced much more plausible images than the original domain homogenizer [16]. Besides, figure 4 shows the detection accuracy of our proposed method. Figure 2: The precision vs recall plot shows that the newly designed end-to- end model performs better than the baselines especially for the recall values. ## IV Conclusions In this paper, we proposed a modified version of our previous domain homogenizer proposed by us and tested it on the data from for the MIDOG 2021 challenge 2021. We showed that the great impact of the position of the domain classifier has a significant impact on the performance of the homogenizer. [EXPLAIN HOW AND WHY] Additionally, our experiments revealed that training the homogenizer along with the object detection network end-to-end improves the detection accuracy by a significant margin. Finally, we showed that our method substantially improves upon the baseline of the MIDOG challenge in terms of mitotic figures detection. Figure 3: A visual comparison of the performances of the domain homogenizer and the modified domain homogenizer(proposed method). Figure 4: Two examples show the detection accuracy of our proposed method. ## References * [1] Wilm, Frauke, et al. ”Domain Adversarial RetinaNet as a Reference Algorithm for the MItosis DOmain Generalization Challenge.” International Conference on Medical Image Computing and Computer-Assisted Intervention. Springer, Cham, 2021. * [2] Marzahl, Christian, et al. ”Deep learning-based quantification of pulmonary hemosiderophages in cytology slides.” Scientific reports 10.1 (2020): 1-10. * [3] Breen, Jack, et al. ”Assessing domain adaptation techniques for mitosis detection in multi-scanner breast cancer histopathology images.” International Conference on Medical Image Computing and Computer-Assisted Intervention. Springer, Cham, 2021. * [4] Chung, Youjin, Jihoon Cho, and Jinah Park. ”Domain-Robust Mitotic Figure Detection with Style Transfer.” International Conference on Medical Image Computing and Computer-Assisted Intervention. Springer, Cham, 2021. * [5] Nateghi, Ramin, and Fattaneh Pourakpour. ”Two-Step Domain Adaptation for Mitotic Cell Detection in Histopathology Images.” International Conference on Medical Image Computing and Computer-Assisted Intervention. Springer, Cham, 2021. * [6] Girshick, Ross. ”Fast r-cnn.” Proceedings of the IEEE international conference on computer vision. 2015. * [7] Fick, Rutger HJ, et al. ”Domain-Specific Cycle-GAN Augmentation Improves Domain Generalizability for Mitosis Detection.” International Conference on Medical Image Computing and Computer-Assisted Intervention. Springer, Cham, 2021. * [8] Zhu, Jun-Yan, et al. ”Unpaired image-to-image translation using cycle-consistent adversarial networks.” Proceedings of the IEEE international conference on computer vision. 2017. * [9] He, Kaiming, et al. ”Mask r-cnn.” Proceedings of the IEEE international conference on computer vision. 2017. * [10] Jahanifar, Mostafa, et al. ”Stain-robust mitotic figure detection for the Mitosis Domain Generalization Challenge.” International Conference on Medical Image Computing and Computer-Assisted Intervention. Springer, Cham, 2021. * [11] Vahadane, Abhishek, et al. ”Structure-preserving color normalization and sparse stain separation for histological images.” IEEE transactions on medical imaging 35.8 (2016): 1962-1971. * [12] Dexl, Jakob, et al. ”MitoDet: Simple and robust mitosis detection.” International Conference on Medical Image Computing and Computer-Assisted Intervention. Springer, Cham, 2021. * [13] Long, Xi, et al. ”Domain Adaptive Cascade R-CNN for MItosis DOmain Generalization (MIDOG) Challenge.” International Conference on Medical Image Computing and Computer-Assisted Intervention. Springer, Cham, 2021. * [14] Reinhard, Erik, et al. ”Color transfer between images.” IEEE Computer graphics and applications 21.5 (2001): 34-41. * [15] Vahadane, Abhishek, et al. ”Structure-preserved color normalization for histological images.” 2015 IEEE 12th International Symposium on Biomedical Imaging (ISBI). IEEE, 2015. * [16] Almahfouz Nasser, Sahar, Nikhil Cherian Kurian, and Amit Sethi. ”Domain Generalisation for Mitosis Detection Exploting Preprocessing Homogenizers.” International Conference on Medical Image Computing and Computer-Assisted Intervention. Springer, Cham, 2021. * [17] StainTools Homepage. https://github.com/Peter554/StainTools. Accessed 03 Aug 2022 * [18] Yaroslav Ganin and Victor Lempitsky. “Unsupervised Domain Adaptation by Backpropagation”. In: Proceedings of the 32nd International Conference on Machine Learning. Ed. by Francis Bach and David Blei. Vol. 37. Proceedings of Machine Learning Research. Lille, France: PMLR, July 2015, pp. 1180–1189. url: https://proceedings.mlr.press/v37/ganin15.html * [19] Justin Johnson, Alexandre Alahi, and Li Fei-Fei. “Perceptual losses for real-time style transfer and super-resolution”. In: European Conference on Computer Vision. 2016 * [20] Marc Aubreville et al. MItosis DOmain Generalization Challenge. Mar. 2021. doi: 10.5281/zenodo.4573978. url: https://doi.org/10.5281/zenodo.4573978 * [21] Adam Paszke et al. “PyTorch: An Imperative Style, High-Performance Deep Learning Library”. In: Advances in Neural Information Processing Systems. Ed. by H. Wallach et al. Vol. 32. Curran Associates, Inc., 2019. url: https://proceedings.neurips.cc/paper/2019/file/bdbca288fee7f92f2bfa9f7012727740-Paper.pdf * [22] Jeremy Howard et al. fastai. https://github.com/fastai/fastai. 2018 * [23] Frauke Wilm et al. “Domain Adversarial RetinaNet as a Reference Algorithm for the MItosis DOmain Generalization Challenge”. In: Biomedical Image Registration, Domain Generalisation and Out-of-Distribution Analysis. Ed. by Marc Aubreville, David Zimmerer, and Mattias Heinrich. Cham: Springer International Publishing, 2022, pp. 5–13. isbn: 978-3-030-97281-3. * [24] Tsung-Yi Lin et al. “Focal Loss for Dense Object Detection”. In: 2017 IEEE International Conference on Computer Vision (ICCV). 2017, pp. 2999–3007. doi: 10.1109/ICCV.2017.324.
# Progress on Local Properties Problems of Difference Sets††thanks: This research project was done as part of the 2020 NYC Discrete Math REU, supported by NSF awards DMS-1802059, DMS-1851420, DMS-1953141, and DMS-2028892. Anqi Li Massachusetts Institute of Technology, 77 Massachusetts Ave, Cambridge, MA 02139<EMAIL_ADDRESS>Supported by MIT Department of Science. ###### Abstract We derive several new bounds for the problem of difference sets with local properties, such as establishing the super-linear threshold of the problem. For our proofs, we develop several new tools, including a variant of higher moment energies and a Ramsey-theoretic approach for the problem. ## 1 Introduction Erdős and Shelah [3] initiated the study of graphs with local properties. For parameters $n,k,$ and $\ell$, they considered edge colorings of the complete graph $K_{n}$, such that every $K_{k}$ subgraph contains at least $\ell$ colors. They defined $f(n,k,\ell)$ as the minimum number of colors that such an edge coloring can have. For example, $f(n,3,2)$ is the minimum number of colors that is required to color $K_{n}$, so that it does not contain a monochromatic triangle. The value of $f(n,k,2)$ can be thought of as an inverse of Ramsey’s theorem. In Ramsey’s theorem, we look for the largest $n$ such that there exists a coloring of the edges of $K_{n}$ with $c$ colors and no monochromatic $K_{k}$. When studying $f(n,k,2)$, we are given the number of vertices $n$, and instead look for the number of colors $c$. By definition, $1\leq f(n,k,\ell)\leq\binom{n}{2}$. We usually consider $n$ to be asymptotically large, while $k$ and $\ell$ are fixed constants. For asymptotically large $n$ and a fixed $k$, the _linear threshold_ is the smallest $\ell$ for which $f(n,k,\ell)=\Omega(n)$. This value of $\ell$ is expressed as a function of $k$. Similarly, the _quadratic threshold_ is the smallest $\ell$ that satisfies $f(n,k,\ell)=\Theta(n^{2})$. The _polynomial threshold_ is the smallest $\ell$ that satisfies $f(n,k,\ell)=\Omega(n^{\varepsilon})$ for some ${\varepsilon}>0$. Erdős and Gyárfás [4] identified the linear and quadratic thresholds to be $\ell=\binom{k}{2}-k+3$ and $\ell=\binom{k}{2}-\lfloor k/2\rfloor+2$, respectively. In the same paper, they also observed that $f(n,k,k)$ is polynomial in $n$. Conlon, Fox, Lee and Sudakov [2] proved that $f(n,k,k-1)$ is sub-polynomial in $n$, which established the polynomial threshold $\ell=k$. Beyond these thresholds, almost tight bounds for $f(n,k,\ell)$ are known in other regimes. Particularly, Pohoata and Sheffer [9] showed that, for any integers $k>m\geq 2$, $f\left(n,k,\binom{k}{2}-m\cdot\left\lfloor\frac{k}{m+1}\right\rfloor+m+1\right)=\Omega\left(n^{1+\frac{1}{m}}\right).$ (1) Probabilistic constructions of Erdős and Gyárfás [4] show that (1) is asymptotically tight up to sub-polynomial factors. Zeev Dvir recently suggested an additive combinatorics variant of the local properties problem. As stated in [4, Section 9], Erdős and Sós also studied this problem, but did not publish their results. We define the _difference set_ of a set $A\subset\mathbb{R}$ as $A-A=\\{a-a^{\prime}\ :\ a,a^{\prime}\in A\ \text{ and }\ a-a^{\prime}>0\\}.$ The standard definition of a difference set does not include the condition $a-a^{\prime}>0$. However, this less common definition makes the following local properties problem nicer to study. For parameters $n,k,$ and $\ell$, we consider sets $A$ of $n$ real numbers, such that every subset $A^{\prime}\subset A$ of size $k$ satisfies that $|A^{\prime}-A^{\prime}|\geq\ell$. Let $g(n,k,\ell)$ be the minimum size of $\lvert A-A\rvert$, over all such sets $A$. In the current work, we study $g(n,k,\ell)$. The trivial bounds for this problem are $n-1\leq g(n,k,\ell)\leq\binom{n}{2}$. As before, for a fixed $k$, the quadratic threshold is the smallest $\ell$ for which $g(n,k,\ell)=\Theta(n^{2})$. We define the _super-linear threshold_ as the largest $\ell$ for which $g(n,k,\ell)=O(n)$. Previous bounds. Currently, not much is known about the behavior of $g(n,k,\ell)$. The unpublished work of Erdős and Sós proved that $g(n,4,5)\geq\binom{n}{2}-n+2$. We also have the bound $g(n,k,\ell)\geq f(n,k,\ell)$, for every $n,k,$ and $\ell$. For example, the bounds of (1) also hold after replacing $f(\cdot)$ with $g(\cdot)$. Indeed, we can construct a complete graph with a vertex for every element of $A$. The color of an edge corresponds to the difference of the numbers that correspond to the two vertices. If every subset $A^{\prime}\subset A$ of size $k$ satisfies that $|A^{\prime}-A^{\prime}|\geq\ell$, then every subgraph $K_{k}$ spans at least $\ell$ colors. Similarly, the total number of colors is $|A-A|$. Fish, Pohoata, and Sheffer [6] observed that Roth’s theorem implies the following bound. ###### Theorem 1.1. For every $\epsilon>0$, there exists $c$ that satisfies the following. For every sufficiently large integer $k$, we have that $g\left(n,k,c\cdot k\cdot\log^{1/4-\epsilon}k\right)=n2^{O(\sqrt{\log n})}.$ Fish, Lund, and Sheffer [5] obtained the following result by projecting a hypercube onto a line. ###### Theorem 1.2. For every $n>k\geq 0$ where $n$ is a power of two, we have that $g\left(n,k,\frac{k^{\log_{2}(3)}-1}{2}\right)=O\left(n^{\log_{2}(3)}\right).$ Our results. We derive several new bounds for $g(n,k,\ell)$. First, we identify the super-linear threshold of $g(n,k,\ell)$. Recall that $\omega(\cdot)$ means “asymptotically strictly larger than”. ###### Theorem 1.3. For every $k$ and sufficiently large $n$, we have that $\displaystyle g(n,k,k-1)=n-1\qquad\text{ and }\qquad g(n,k,k)=\omega(n).$ After establishing the super-linear threshold, it is natural to ask how quickly the bound changes as $\ell$ grows. We obtain the following bound via a probabilistic argument. ###### Theorem 1.4. For every $c\geq 2$ and integer $k>(c^{2}+1)^{2}$, we have that $g(n,k,ck+1)=O\Big{(}n^{1+\frac{c^{2}+1}{k}}\Big{)}.$ When fixing $c\geq 2$, the bound of Theorem 1.4 becomes arbitrarily close to $O(n)$ as $k$ grows. We note that the assumption $k>(c^{2}+1)^{2}$ can be removed from the statement of this theorem. Indeed, when $k\leq(c^{2}+1)^{2}$, the bound of the theorem is trivial. Additive combinatorics provides many tools for studying sets that have a small difference set. We rely on such tools to study the case of small $\ell$, in the proofs of Theorems 1.3 and 1.4. To study sets with larger values of $\ell$, we have to develop new tools. In particular, we obtain the following result by studying a new variant of higher moment energies. For the definition of higher moment energies and of our new variant, see Section 2. ###### Theorem 1.5. Every $k\geq 8$ that is a multiple of 8 satisfies $g\left(n,k,\frac{9k^{2}}{32}-\frac{9k}{8}+5\right)=\Omega(n^{4/3}).$ By (1), for the graphs local properties problem, we have that $f\left(n,k,\binom{k}{2}-3\cdot\left\lfloor\frac{k}{4}\right\rfloor+4\right)=\Omega\left(n^{4/3}\right).$ This bound is tight up to subpolynomial factors. Theorem 1.5 shows that, in the additive combinatorics variant, a much weaker local restriction leads to $\Omega(n^{4/3})$, up to subpolynomial factors. Unlike the graph variant, in this case the coefficient of $k^{2}$ is smaller than $1/2$. Theorem 1.2 provides an upper bound for the local properties problem by introducing a construction. Surprisingly, we show that the same construction can be used to obtain a proof for a lower bound. ###### Theorem 1.6. For every $k\geq 8$ that is a power of two and sufficiently large $n$, we have that $g\left(n,k,\frac{k^{\log_{2}(3)}+1}{2}\right)=\Omega\left(n^{1+\frac{2}{k-2}}\right).$ The proof of Theorem 1.6 is in some sense Ramsey theoretic: We show that a set with many repeated differences must contain a copy of the construction from Theorem 1.2. Such a subset contradicts the local property, and thus cannot exist. Finally, we mention the best bound that we managed to obtain for the quadratic threshold. ###### Claim 1.7. When $k$ is a multiple of two, we have that $g\left(n,k,\frac{3k^{2}}{8}-\frac{3k}{4}+2\right)=\Omega(n^{2}).$ It is not difficult to prove Claim 1.7, and it is possible that the quadratic threshold is much smaller. We include Claim 1.7 for completeness. That is, to document all the current best bounds for the arithmetic variant of the local properties problem. Lastly, we note that while most results in this paper apply to the setting of a general Abelian group, we decided to not pursue this direction and leave this to future work. Outline. In Section 2, we define our new energy variant and use it to prove Theorem 1.5. In Section 3, we prove our other results for large values of $\ell$: Theorem 1.6 and Claim 1.7. In Section 4, we prove our bounds for small values of $\ell$: Theorem 1.3 and Theorem 1.4. In Section 5, we pose some questions for future work. Acknowledgements. We would like to thank MIT Department of Science for the generous support. We would also like to express our gratitude towards Adam Sheffer for his mentorship and support throughout the NYC Discrete Math REU, as well as his invaluable help in the preparation of this paper. In addition, we would like to thank Sara Fish for many helpful discussions as well as Ilya Shkredov for his enlightening suggestion that led to an improvement in the bound of Theorem 1.5. Lastly, we thank the referees for their many useful comments and suggestions. ## 2 Dumbbell energy and Theorem 1.5 In this section, we prove Theorem 1.5. Our proof relies on a new energy variant. Before we introduce this energy, we briefly recall the notion of higher moment energies. The _additive energy_ of a finite set $A\subset\mathbb{R}$ is $E^{+}(A)=\lvert\\{(a_{1},a_{2},a_{3},a_{4})\in A^{4}\ :\ a_{1}-a_{2}=a_{3}-a_{4}\\}\rvert.$ Additive energy is a common tool in additive combinatorics. For example, it provides the following lower bound for the size of the difference set: $|A-A|\geq\frac{|A|^{4}}{E^{+}(A)}.$ In the past decade, several variants of additive energy led to interesting new results. One of those variants is _higher moment energies_. For an integer $\ell>1$, the $\ell$th moment energy of $A$ is $E_{\ell}^{+}(A)=\lvert\\{(a_{1},\ldots,a_{2\ell})\in A^{2\ell}\ :\ a_{1}-a_{2}=a_{3}-a_{4}=\cdots=a_{2\ell-1}-a_{2\ell}\\}\rvert.$ We note that $E_{2}^{+}(A)$ is the standard additive energy $E^{+}(A)$. For applications of higher moment energies, see for example [11, 12]. Figure 1: (a) A dumbbell. (b) An arithmetic dumbbell. We now define a new variant of additive energy, which we call _dumbbell energy_. We say that a quadruple $(a_{1},a_{2},a_{3},a_{4})\in\mathbb{R}$ forms a _dumbbell_ if $a_{1}<a_{2}<a_{3}<a_{4}$ and $a_{2}-a_{1}=a_{4}-a_{3}$. See Figure 1. For distances $d,d^{\prime}>0$, we define $D(d,d^{\prime})$ as the congruence class of all dumbbells $(a_{1},a_{2},a_{3},a_{4})$ that satisfy $a_{2}-a_{1}=a_{4}-a_{3}=d$ and $a_{3}-a_{2}=d^{\prime}$. That is, two dumbbells are congruent if there exists a translation of $\mathbb{R}$ that takes one dumbbell to the other. We can think of the $\ell$th additive energy as the number of $\ell$-tuples of congruent intervals. Similarly, we consider an energy variant that counts $\ell$-tuples of congruent dumbbells. We define the $\ell$th _dumbbell energy_ of $A\subset\mathbb{R}$ as $\displaystyle E^{D}_{\ell}(A)=\Big{|}\Big{\\{}(a_{1},\ldots,a_{4\ell})\in A^{4\ell}\ :\ $ $\displaystyle\text{ For }1\leq i\leq\ell,\text{ the quadruples }(a_{4i-3},a_{4i-2},a_{4i-1},a_{4i})$ $\displaystyle\text{ form }\ell\text{ distinct dumbbells from the same congruence class }\Big{\\}}\Big{|}.$ For $d,d^{\prime}\in A-A$, we set $r^{D}_{A}(d,d^{\prime})=|\\{(a_{1},a_{2},a_{3},a_{4})\in A^{4}\ :\ a_{2}-a_{1}=a_{4}-b_{3}=d\text{ and }a_{3}-a_{2}=d^{\prime}\\}|.$ In other words, $r^{D}_{A}(d,d^{\prime})$ is the number of dumbbells in the congruence class $D(d,d^{\prime})$ that are spanned by $A$. We note that a dumbbell $(a_{1},a_{2},a_{3},a_{4})$ of $D(d,d^{\prime})$ is uniquely determined by $a_{1}$. Thus, $r^{D}_{A}(d,d^{\prime})\leq|A|$ for all $d,d^{\prime}\in A-A$. We also note that $E^{D}_{\ell}(A)=\sum_{d,d^{\prime}\in A-A}\binom{r^{D}_{A}(d,d^{\prime})}{\ell}.$ For a difference $d\in A-A$, we set $r^{-}_{A}(d)=|\\{(a,a^{\prime})\in A^{2}\ :\ a-a^{\prime}=d\\}|.$ In other words, $r^{-}_{A}(d)$ is the number of representations that the difference $d$ has in $A$. The following lemma states a connection between the $\ell$th dumbbell energy $E^{D}_{\ell}(A)$ and the size of the difference set $|A-A|$. ###### Lemma 2.1. The following holds for every even constant $\ell\geq 2$. Every set $A$ of $n$ reals satisfies that $|A-A|=\Omega\left(n^{4/3}\right)$ or that $E^{D}_{\ell}(A)=\Omega\left(\frac{n^{4\ell}}{|A-A|^{3\ell-2}}\right).$ ###### Proof. As a first step, we will prove that $\sum_{d,d^{\prime}\in A-A}r^{D}_{A}(d,d^{\prime})^{2}=\Omega\left(\frac{n^{8}}{|A-A|^{4}}\right).$ Let $\mathbf{1}_{B}(x)=\begin{cases}1&\text{if }x\in B,\\\ 0&\text{otherwise},\end{cases}$ be the indicator function for a set $B\subset\mathbb{R}$. We may write $S:=\sum_{d,d^{\prime}\in A-A}r_{A}^{D}(d,d^{\prime})=\sum_{a\in A}\sum_{d,d^{\prime}\in A-A}\mathbf{1}_{A\cap(A-d)}(a)\mathbf{1}_{A\cap(A-d)}(a+d^{\prime})=\sum_{a\in A}\sum_{d\in A-A}\mathbf{1}_{A\cap(A-d)}(a)r_{A}^{-}(d).$ Note that $\sum_{d\in A-A}r_{A}^{-}(d)=\sum_{a\in A}\sum_{d\in A-A}\mathbf{1}_{A\cap(A-d)}(a)=\binom{|A|}{2}=\binom{n}{2}$. Consider $\mathcal{D}=\left\\{d:|r_{A}^{-}(d)|\geq S/4\cdot\binom{n}{2}^{-1}\right\\}$. It is not difficult to see that $\mathcal{D}$ captures a large fraction of the above sum: since $\sum_{a\in A}\sum_{d\not\in\mathcal{D}}\mathbf{1}_{A\cap(A-d)}(a)r_{A}^{-}(d)<\max_{d\not\in\mathcal{D}}r_{A}^{-}(d)\cdot\binom{n}{2}=\frac{S}{4},$ it follows that $\sum_{a\in A}\sum_{d\in\mathcal{D}}\mathbf{1}_{A\cap(A-d)}(a)r_{A}^{-}(d)\geq\frac{3S}{4}.$ We also have the following bound $|\mathcal{D}|\cdot S/4\cdot\binom{n}{2}^{-1}\leq\sum_{d\in A-A}r_{A}^{-}(d)=\binom{n}{2}.$ That is, $|\mathcal{D}|\leq 4/S\cdot\binom{n}{2}^{2}$. Let $\mathcal{B}\subset(A-A)\times(A-A)\subset\mathbb{R}^{2}$ be the set which projects onto a copy of $\mathcal{D}$ on each coordinate axes. By the symmetry in $d,d^{\prime}$, the above inequality implies that $\sum_{d,d^{\prime}\in\mathcal{B}}r_{A}^{D}(d,d^{\prime})\geq\frac{3S}{4}.$ Furthermore, the Loomis-Whitney inequality gives the upper bound $|\mathcal{B}|\leq|\mathcal{D}|^{2}\leq 16/S^{2}\cdot\binom{n}{2}^{4}$. Consequently, the Cauchy-Schwarz inequality implies that $\frac{9S^{2}}{16}\leq\left(\sum_{d,d^{\prime}\in\mathcal{B}}r_{A}^{D}(d,d^{\prime})\right)^{2}\leq|\mathcal{B}|\sum_{d,d^{\prime}\in\mathcal{B}}r_{A}^{D}(d,d^{\prime})^{2}\leq 16/S^{2}\cdot\binom{n}{2}^{4}\sum_{d,d^{\prime}\in A}r_{A}^{D}(d,d^{\prime})^{2}.$ Equivalently, $\sum_{d,d^{\prime}\in A-A}r^{D}_{A}(d,d^{\prime})^{2}\geq\frac{9S^{4}}{256\cdot\binom{n}{2}^{4}}.$ (2) By definition, we have that $\displaystyle S=\sum_{d,d^{\prime}\in A-A}r^{D}_{A}(d,d^{\prime})=\sum_{d\in A-A}\binom{r^{D}_{A}(d)}{2}$ $\displaystyle=\sum_{d\in A-A}\frac{r^{D}_{A}(d)^{2}}{2}-\sum_{d\in A-A}\frac{r^{D}_{A}(d)}{2}$ $\displaystyle=\sum_{d\in A-A}\frac{r^{D}_{A}(d)^{2}}{2}-\frac{1}{2}\binom{n}{2}.$ Combining the above with the Cauchy-Schwarz inequality leads to $\sum_{d,d^{\prime}\in A-A}r^{D}_{A}(d,d^{\prime})\geq\frac{(\sum_{d\in A-A}r^{D}_{A}(d))^{2}}{2|A-A|}-\frac{1}{2}\binom{n}{2}=\binom{n}{2}^{2}/2|A-A|-\frac{1}{2}\binom{n}{2}.$ If $|A-A|=\Theta(n^{2})$ then we are done. We may thus assume that $S=\Omega\left(\frac{n^{4}}{|A-A|}\right).$ (3) Combining this Equation 2, we get the desired lower bound $\sum_{d,d^{\prime}\in A-A}r^{D}_{A}(d,d^{\prime})^{2}=\Omega\left(\frac{n^{8}}{|A-A|^{4}}\right).$ Lastly, by Hölder’s inequality and since $\ell$ is constant, we obtain that $\displaystyle E_{\ell}^{D}(A)$ $\displaystyle=\sum_{d,d^{\prime}\in A-A}\binom{r_{A}^{D}(d,d^{\prime})}{\ell}\geq\sum_{d,d^{\prime}\in A-A}\left(\frac{r_{A}^{D}(d,d^{\prime})}{\ell}\right)^{\ell}\geq\frac{(\sum_{d,d^{\prime}\in A-A}r_{A}^{D}(d,d^{\prime})^{2})^{\ell/2}}{|A-A|^{2(\ell/2-1)}\ell^{\ell}}$ $\displaystyle=\Omega\left(\frac{(n^{8}/|A-A|^{4})^{\ell/2}}{|A-A|^{\ell-2}\ell^{\ell}}\right)=\Omega\left(\frac{n^{4\ell}}{|A-A|^{3\ell-2}}\right).$ ∎ We next introduce a variant of the energy graph that is defined in [6]. For a finite $A\subset\mathbb{R}$ and $\ell$ even, we define the $\ell$th _dumbell energy graph_ $G$ of $A$, as follows. The graph $G$ has a vertex for every $2\ell$-tuple from $A^{2\ell}$. There is an edge between the vertices $(a_{1},\ldots,a_{2\ell})$ and $(b_{1},\ldots,b_{2\ell})$ if the tuple $(a_{1},\ldots,a_{2\ell},b_{1},\ldots,b_{2\ell})$ contributes to $E^{D}_{\ell}$. Equivalently, if each vertex consists of $\ell/2$ dumbbells and all $\ell$ dumbbells are congruent. By definition, the number of edges in $G$ is $E_{\ell}^{D}(A)$. We are now ready to prove Theorem 1.5. We first recall the statement of this theorem. Theorem 1.5. _Every $k\geq 8$ that is a multiple of 8 satisfies_ $g\left(n,k,\frac{9k^{2}}{32}-\frac{9k}{8}+5\right)=\Omega(n^{4/3}).$ ###### Proof. Let $A$ be a set of $n$ reals, such that every subset $A^{\prime}\subset A$ of size $k$ satisfies that $|A^{\prime}-A^{\prime}|\geq\frac{5k^{2}}{16}-k+5$. Let $G$ be the $k/4$ dumbbell energy graph of $A$. Consider three vertices $v_{1},v_{2},v_{3}$ of $G$, such that there is an edge between $v_{1}$ and $v_{2}$ and another edge between $v_{2}$ and $v_{3}$. Then each of the three vertices consists of $k/8$ dumbbells, and all $3k/8$ dumbbells are congruent. Thus, there is also an edge between $v_{1}$ and $v_{3}$. We conclude that every connected component of $G$ is a clique. We may discard connected components with a single vertex without changing the number of edges. Every remaining connected component corresponds to a distinct dumbbell congruency class. Since there are $O(|A-A|^{2})$ distinct (non- congruent) dumbbells, the number of remaining connected components in $G$ is $O(|A-A|^{2})$. Assume for contradiction that $A$ spans $k/4$ disjoint congruent dumbbells. Let $A^{\prime}\subset A$ be the set of $k$ numbers that form these $k/4$ dumbbells. At most four distinct differences are spanned by pairs of numbers from the same dumbbell. In Figure 1(b), these distances are $d,d^{\prime},d+d^{\prime}$, and $2d+d^{\prime}$. Figure 2: The nine distances between the two dumbbells are $d_{3},d_{1}+d_{3},2d_{1}+d_{3},d_{1}+d_{2}+d_{3},2d_{1}+d_{2}+d_{3},3d_{1}+d_{2}+d_{3},2d_{1}+2d_{2}+d_{3},3d_{1}+2d_{2}+d_{3},4d_{1}+2d_{2}+d_{3}$. There are $\binom{k/4}{2}$ pairs of distinct dumbbells as described above. Each such pair spans at most 9 distinct differences with one number from each dumbbell. See Figure 2. We conclude that $|A^{\prime}-A^{\prime}|\leq 4+\binom{k/4}{2}\cdot 9=\frac{9k^{2}}{32}-\frac{9k}{8}+4.$ This contradicts the assumption, so $A$ does not span $k/4$ disjoint congruent dumbbells. Let $S\subset A$ be a dumbbell. We wish to derive an upper bound for the number of dumbbells $S^{\prime}$ that are congruent to $S$ but are not disjoint to it. There are at most four choices for the element of $S$ that also appears in $S^{\prime}$. Then, there are at most three possible positions for this element in $S^{\prime}$. Since $S^{\prime}$ is congruent to $S$, the other three coordinates of $S^{\prime}$ are uniquely determined. Thus, at most 12 dumbbells are congruent to $S$ and not disjoint to it. By combining the two preceding paragraphs, we conclude that $A$ cannot span $3k$ congruent dumbbells. For a fixed dumbbell congruence class, every vertex in the corresponding connected component of $G$ consists of $k/8$ copies of that dumbbell, possibly with repetitions. Since there are fewer than $3k$ such dumbbells to choose from, this connected component consists of fewer than $(3k)^{k/8}$ vertices. The number of edges in the connected component is smaller than $(3k)^{k/4}$. This holds for every connected component of $G$. Recalling that the number of remaining connected components in $G$ is at most $|A-A|^{2}$, we conclude that the number of edges in $G$ is $O_{k}(|A-A|^{2})$. Since the number of edge in $G$ is $E_{k/4}^{D}(A)$, we have that $E_{k/4}^{D}(A)=O_{k}(|A-A|^{2})$. By Lemma 2.1, we have that $E^{D}_{k/4}(A)=\Omega\left(\frac{n^{k}}{|A-A|^{3k/4-2}}\right).$ Combining the two above bounds for $E^{D}_{k/4}(A)$ implies that $|A-A|=\Omega(n^{4/3})$. ∎ ###### Remark 2.2. When performing the above analysis with $E_{2}^{D}(A)$ instead of $E^{D}_{k/4}(A)$, we obtain the weaker bound $|A-A|=\Omega(n^{8/7})$ with the same local condition. ###### Remark 2.3. We say that an 8-tuple $(a_{1},a_{2},a_{3},a_{4},a_{5},a_{6},a_{7},a_{8})\in\mathbb{R}^{8}$ forms a _double dumbbell_ if $a_{1}<a_{2}<a_{3}<a_{4}<a_{5}<a_{6}<a_{7}<a_{8}$ and $a_{2}-a_{1}=a_{4}-a_{3}=a_{6}-a_{5}=a_{8}-a_{7}$ and $a_{3}-a_{2}=a_{7}-a_{6}$. A natural generalization of the dumbbell energy studied above is the $\ell$th _double dumbbell energy_ , defined for $A\subset\mathbb{R}$ as $\displaystyle\mathcal{E}^{D}_{\ell}(A)=\Big{|}\Big{\\{}(a_{1},\ldots,a_{8\ell})\in A^{8\ell}\ :\ $ $\displaystyle\text{ For }1\leq i\leq\ell,\text{ the quadruples }(a_{8i-7},a_{8i-6},\ldots,,a_{8i})$ $\displaystyle\text{ form }\ell\text{ distinct double dumbbells from the same congruence class }\Big{\\}}\Big{|}.$ As before, two double dumbbells are congruent if there exists a translation of $\mathbb{R}$ that takes one double dumbbell to the other. An identical argument as that for Theorem 1.5 gives the bound $g\left(n,k,\frac{31k^{2}}{128}-\frac{31k}{16}+9\right)=\Omega(n^{8/7})$ for every $k$ a multiple of 16. Here we have slightly weaker local condition than that in Theorem 1.5 at the expense of a worse exponent in $\Omega\left(n^{8/7}\right)$. We chose to present Theorem 1.5 because it is less calculation intensive and better illustrates the main ideas of the proof. ## 3 Additional proofs for large values of $\ell$ In this section, we prove our bounds for large values of $\ell$ that are not related to dumbbell energy. That is, we prove Theorem Theorem 1.6 and Claim 1.7. We begin with the proof of Claim 1.7, since this simple proof is a nice warm- up. The proof relies on the observation that $a-b=c-d$ implies $a-c=b-d$. Recall that for a set $A$ and a difference $d\in A-A$ we set $r^{-}_{A}(d)=|\\{(a,a^{\prime})\in A^{2}\ :\ a-a^{\prime}=d\\}|.$ Claim 1.7. _When $k$ is a multiple of two, we have that_ $g\left(n,k,\frac{3k^{2}}{8}-\frac{3k}{4}+2\right)=\Omega(n^{2}).$ ###### Proof. Let $A$ be a set of $n$ reals, such that every subset $A^{\prime}\subset A$ of size $k$ satisfies that $|A^{\prime}-A^{\prime}|\geq\frac{3k^{2}}{8}-\frac{3k}{4}+2$. Assume for contradiction that there exists $d\in A-A$ for which $r_{A}^{-}(d)\geq k$. Since every number participates in at most two representations of $d$, there exist distinct $a_{1},\ldots,a_{k/2},b_{1},\ldots,b_{k/2}$ such that $a_{1}-b_{1}=\cdots=a_{k/2}-b_{k/2}=d$. We set $S=\\{a_{1},\ldots,a_{k/2},b_{1},\ldots,b_{k/2}\\}$. We may assume, without loss of generality, that $a_{1}>a_{2}>\cdots>a_{k/2}$. This implies that $b_{1}>b_{2}>\cdots>b_{k/2}$. For every $1\leq i<j\leq k/2$, we have that $a_{i}-b_{i}=a_{j}-b_{j}$, which in turn implies that $a_{i}-a_{j}=b_{i}-b_{j}$. Since there are $\binom{k/2}{2}$ such choices of $i$ and $j$, there are $\binom{k/2}{2}$ equations of the form $a_{i}-a_{j}=b_{i}-b_{j}$. We conclude that $|S-S|\leq\binom{k}{2}-\left(\frac{k}{2}-1\right)-\binom{k/2}{2}=\frac{3k^{2}}{8}-\frac{3k}{4}+1.$ Since the above contradicts the assumption about $A$, we get that every $d\in A-A$ satisfies $r_{A}^{-}(d)<k$. Since $k$ is constant, we have that $|A-A|>\frac{1}{k}\binom{n}{2}=\Theta(n^{2}).$ ∎ We now move to prove Theorem 1.6. We rely on the following configuration from [5]. Fish, Lund and Sheffer utilize the configuration to obtain an upper bound on $g(n,k,k^{2/3})$. We will instead use this construction to prove a lower bound on $g(n,k,\ell)$ where $\ell=\Theta(k^{2/3})$. For an integer $i\geq 1$ and positive reals $a,\delta_{1},\ldots,\delta_{i}$, we define a _projected- $i$-cube_ as $P(a,\delta_{1},\ldots,\delta_{i})=\\{a+x_{1}\delta_{1}+x_{2}\delta_{2}+\cdots+x_{i}\delta_{i}\ :\ x_{1},\ldots,x_{i}\in\\{0,1\\}\\}.$ We note that a projected-$i$-cube is a set of at most $2^{i}$ real numbers. A projected-$i$-cube can be thought of as a projection of an $i$-dimensional hypercube onto $\mathbb{R}$, which explains the name of this object. For $i>1$, every projected-$i$-cube is the union of two projected-$(i-1)$-cubes that are identical up to a translation of distance $p_{i}$. The following lemma states another useful property of projected-$i$-cubes. ###### Lemma 3.1. If $A$ is a projected-$i$-cube then $\lvert A-A\rvert\leq\frac{3^{i}-1}{2}.$ For comparison, a set $A$ of $2^{i}$ random reals is expected to satisfy $\lvert A-A\rvert=\Theta(4^{i})$. ###### Proof. The proof is by induction on $i$. For the induction basis, we consider the case of $i=1$. In this case $A$ is a set of two numbers, so $|A-A|=1=(3^{1}-1)/2$. For the induction step, we assume that the claim holds for projected-$i$-cubes and consider a projected-$(i+1)$-cube $A=P(a,\delta_{1},\ldots,\delta_{i+1})$. Then $A$ is the union of the two projected-$i$-cubes $A_{1}=P(a,\delta_{1},\ldots,\delta_{i})$ and $A_{2}=P(a+\delta_{i+1},\delta_{1},\ldots,\delta_{i})$. These two projected-$i$-cubes have the same difference set $D=A_{1}-A_{1}=A_{2}-A_{2}$. The induction hypothesis implies that $|D|\leq(3^{i}-1)/2$. Next, we note that every difference in $(A-A)\setminus\\{\delta_{i+1}\\}$ is either formed by two numbers from $A_{1}$ or by one number from $A_{1}$ and another from $A_{2}$. By the preceding paragraph, there are at most $(3^{i}-1)/2$ differences of the former type. Every difference of the latter type is in $D\pm\delta_{i+1}$. The number of such differences is at most $2\cdot|D|=3^{i}-1$. In total, $|A-A|\leq(3^{i}-1)/2+(3^{i}-1)+1=(3^{i+1}-1)/2.$ This completes the proof of the induction step and thus the proof of the theorem. ∎ We define the _left endpoint_ of a projected-$i$-cube as the smallest number of that cube. The _right endpoint_ is the largest number of the cube. Theorem 1.6. _For every $k\geq 8$ that is a power of two and sufficiently large $n$, we have that_ $g\left(n,k,\frac{k^{\log_{2}(3)}+1}{2}\right)=\Omega\left(n^{1+\frac{2}{k-2}}\right).$ ###### Proof. We consider a set $A\subset\mathbb{R}$ such that $|A|=n$ and $|A-A|\leq n^{1+\frac{2}{k-2}}/9$. We show that such a set must contain a projected-$(\log_{2}(k)-2)$-cube. Then, Lemma 3.1 implies that $A$ does not satisfy the local property in the statement of the theorem, which completes the proof. For an integer $i$, we set $\eta_{i}=1-(2^{i+1}-2)/(k-2)$. We prove by induction on $1\leq i\leq\log_{2}(k)-2$ that $A$ contains at least $n^{\eta_{i}}$ disjoint translated copies of a projected-$i$-cube with exactly $2^{i}$ elements. Combining this with the preceding paragraph completes the proof of the theorem. There are $\binom{n}{2}>n^{2}/3$ pairs of elements from $A$ and $|A-A|\leq n^{1+\frac{2}{k-2}}/9$. By the pigeonhole principle there exists a distance that occurs at least $3n^{(k-4)/(k-2)}$ times. We arbitrarily fix one such distance and denote it as $d_{1}$. We set $A_{1}$ to be the set of elements of $A$ that are at distance $d_{1}$ from another element of $A$. Let $G_{1}$ be the graph whose vertices are the elements of $A_{1}$. Two vertices are connected by an edge if they span the difference $d_{1}$. We note that $G_{1}$ contains at least $3n^{(k-4)/(k-2)}$ edges. Since every vertex is of degree at most two, we can obtain a matching by repeatedly choosing an arbitrary edge $e$ and discarding at most two other edges that share a vertex with $e$. This implies that $G_{1}$ contains a matching of size at least $n^{(k-4)/(k-2)}$. Since a projected-1-cube consists of two points and $\eta_{1}=(k-4)/(k-2)$, we obtain the case of $i=1$. This completes the induction basis. For the induction step, we fix $1\leq i<\log_{2}(k)-2$ and assume that the claim holds for $i$. That is, $A$ contains at least $n^{\eta_{i}}$ disjoint translated copies of a projected-$i$-cube. We note that, in this range, $\eta_{i}>0$. We fix exactly $n^{\eta_{i}}$ translated copies and set $A_{i+1}$ to be the set of endpoints of these. By definition, we have that $|A_{i+1}|=2n^{\eta_{i}}$. The number of pairs of endpoints that do not belong to the same translated copy is $\frac{1}{2}\cdot 2n^{\eta_{i}}\cdot\left(2n^{\eta_{i}}-2\right)=2n^{2\eta_{i}}-2n^{\eta_{i}}.$ Each difference is spanned by $A_{i+1}$ fewer than $2n^{\eta_{i}}$ times. Thus, the number of above pairs that span a difference that is also spanned by the projected-$i$-cube is $O(n^{\eta_{i}})$. When $n$ is sufficiently large, we may assume that the number of pairs that span a distance that is not in the projected-$i$-cube is at least $n^{2\eta_{i}}$. Since $A_{i+1}\subset A$ and $|A-A|\leq n^{1+\frac{2}{k-2}}/9$, there exists a difference that is spanned by at least $9n^{2\eta_{i}-1-2/(k-2)}=9n^{\eta_{i+1}}$ of these pairs. We fix such a difference and denote it as $d_{i+1}$. In each of the corresponding pairs, the difference $d_{i+1}$ occurs between the two left endpoints, between the two right endpoints, or between one right endpoint and one left endpoint. By the pigeonhole principle, at least one third of the pairs are of the same type. Thus, $A_{i+1}$ contains at least $3n^{\eta_{i+1}}$ translatd copies of the same projected-$(i+1)$-cube. The above translated copies of a projected-$(i+1)$-cube might not be disjoint. In particular, a projected projected-$i$-cube from $A_{i}$ might participate in two projected-$(i+1)$-cubes. By repeating the matching argument from the induction basis, we obtain that at least one third of the projected-$(i+1)$-cubes are disjoint. That is $A_{i+1}$ contains at least $n^{\eta_{i+1}}$ disjoint translated copies of the same projected-$(i+1)$-cube. This concludes the proof of the induction step and thus the proof of the theorem. ∎ ## 4 Small values of $\ell$ In this section, we prove our bounds for small values of $\ell$: Theorem 1.3 and Theorem 1.4. We first briefly go over the tools that we require from additive combinatorics. Tools from additive combinatorics. Szemerédi’s theorem states that every sufficiently large subset of $\\{1,2,\ldots,n\\}$ contains a long arithmetic progression. We rely on the following variant of this result, due to Gowers [7]. ###### Theorem 4.1. For every $k$, there exists $c>0$ that satisfies the following. Every set of $n/(\log\log n)^{c}$ numbers from $\\{1,2,\ldots,n\\}$ contains a $k$-term arithmetic progression. Moreover, $c$ can be taken to be $2^{-2^{k+9}}$. A _generalized arithmetic progression of dimension_ $d$ is defined as $\Big{\\{}a+\sum_{j=1}^{d}k_{j}b_{j}\ :\ a,b_{1},\ldots,b_{d}\in\mathbb{R}\text{ and with integer }0\leq k_{j}\leq n_{j}-1\text{ for every }1\leq j\leq d\Big{\\}}.$ (4) The size of a generalized arithmetic progression is the number of elements in it. We note that an arithmetic progression is a generalized arithmetic progression of dimension one. One of the most common tools for studying sets with a small different set is _Freiman’s theorem._ The following is a variant of this theorem over the reals. For example, see Sanders [10, Theorem 1.3]. ###### Theorem 4.2. For every sufficiently large finite set $A\subset\mathbb{R}$ and sufficiently large $d>0$, if $|A-A|\leq d|A|$ then $A$ is contained in a generalized arithmetic progression of dimension at most $d\cdot\log^{4}d$ and size at most $|A|\cdot e^{d\cdot\log^{4}d}$. Similarly to a difference set $A-A$, we define the _sum set_ of a set $A\subset\mathbb{R}$ as $A+A=\\{a+a^{\prime}\ :\ a,a\in A\\}.$ The following _Plünnecke-Ruzsa estimate_ (for example, see [13, Corollary 6.29]) shows one connection between $A+A$ and $A-A$. ###### Lemma 4.3. If a finite $A\subset\mathbb{R}$ satisfies $|A-A|\leq c|A|$ then $|A+A|\leq c^{2}|A|.$ The following is a special case of a result of Green and Morris [8] about sum sets. ###### Theorem 4.4. For every $c>0$, every sufficiently large $m$ satisfies the following. For every integer $n$, the number of subsets $A\subset\\{1,2,\ldots,n\\}$ that satisfy $|A|=m$ and $|A+A|\leq c|A|$ is at most $2^{m}\cdot n^{c+1}\cdot\binom{cm/2}{m}.$ Our proofs. We now derive the super-linear threshold for our local properties problem. We first recall the statement of this result. Theorem 1.3. _For every $k$ and sufficiently large $n$, we have that_ $\displaystyle g(n,k,k-1)=n-1\qquad\text{ and }\qquad g(n,k,k)=\omega(n).$ ###### Proof. To see that $g(n,k,k-1)=n-1$, we set $A=\\{1,2,\ldots,n\\}$ and note that $|A-A|=n-1$. The local property trivially holds, since every $k$ numbers span at least $k-1$ differences. It remains to prove that $g(n,k,k)=\omega(n)$. For a constant $c>0$, we consider a set $A$ such that $|A|=n$ and $|A-A|\leq cn$. By Theorem 4.2, the set $A$ is contained in a generalized arithmetic progression $G$ of size at most $Cn$ and dimension at most $D$. The constants $C$ and $D$ depend on $c$ but not on $n$. We define $G$ as in (4). Without loss of generality, we may assume that $n_{1}\geq n^{1/D}$. We may think of $G$ as the union of disjoint arithmetic progressions with step $b_{1}$, each of size at least $n_{1}\geq n^{1/D}$. Since $|A|/|G|\geq 1/C$, there exists such an arithmetic progression $S$ such that $|A\cap S|/|S|\geq 1/C$. We translate and scale $\mathbb{R}$ such that $S$ becomes $\\{1,2,3,...,|S|\\}$. Then, $A$ contains at least $|S|/C$ elements of $\\{1,2,3,...,|S|\\}$. Theorem 4.1 implies that $A$ contains a $k$-term arithmetic progression. We proved that the following holds for every constant $c>0$, when $n$ is sufficiently large: Every set $A$ of $n$ reals that satisfies $|A-A|\leq cn$ contains $k$ numbers that span $k-1$ differences. Thus, a set that satisfies the local restrictions of $g(n,k,\ell)$ also satisfies that $|A-A|>cn$ for every $c$. ∎ ###### Remark 4.5. Quantitatively, the proof above gives the bound $g(n,k,k)\leq n\cdot\sqrt{\log(\log\log n)^{c}}$, where $c=2^{-2^{k+9}}$. For our probabilistic arguments, we recall the following Chernoff bounds (for example, see [1, Corollary A.1.14]). ###### Theorem 4.6. For every ${\varepsilon}>0$, there exists $c_{\varepsilon}$ that satisfies the following. Let $Y$ be a sum of independent indicator random variables and let $\mu$ be the expected value of $Y$. Then $\Pr[|Y-\mu|>{\varepsilon}\mu]<2e^{-e_{\varepsilon}\mu}.$ We may set $c_{\varepsilon}=\min\\{\ln(e^{-{\varepsilon}}(1+{\varepsilon})^{1+{\varepsilon}}),{\varepsilon}^{2}/2\\}.$ We now prove our upper bound for small values of $\ell$. We first recall the statement of this result. Theorem 1.4. _For every $c\geq 2$ and integer $k>(c^{2}+1)^{2}$, we have that_ $g(n,k,ck+1)=O\Big{(}n^{1+\frac{c^{2}+1}{k}}\Big{)}.$ ###### Proof. Let $N$ be a sufficiently large constant multiple of $n$, which is determined below. We define the probability $p=3N^{-(c^{2}+1)/k}$ and set $M=\big{\\{}1,2,\ldots,N^{1+(c^{2}+1)/k}\big{\\}}$. Let $A^{\prime}$ be a set that is obtained by selecting each element of $M$ independently with probability $p$. The expected size of $A^{\prime}$ is $p\cdot N^{1+(c^{2}+1)/k}=3N.$ By applying Theorem 4.6 with $\mu=3N$ and ${\varepsilon}=1/2$, we obtain that $\Pr\left[\frac{3N}{2}\leq\lvert A\rvert\leq\frac{9N}{2}\right]>1-2e^{-N/4}.$ (5) By Theorem 4.4, the number of subsets $B\subset M$ that satisfy $|B|=k$ and $|B+B|\leq c^{2}k$ is at most $2^{k}\cdot(N^{1+(c^{2}+1)/k})^{c^{2}+1}\cdot\binom{c^{2}k/2}{k}=O_{c,k}(N^{c^{2}+1+(c^{2}+1)^{2}/k}).$ By the contrapositive of Lemma 4.3, the number of subsets $B\subset M$ that satisfy $|B|=k$ and $|B-B|\leq ck$ is $O_{c,k}(N^{c^{2}+1+(c^{2}+1)^{2}/k})$. We refer to such a set as a _set with small doubling_. The probability that a fixed set with small doubling is in $A$ is $p^{k}$. Thus, the expected number of sets with small doubling in $A$ is $p^{k}\cdot O_{c,k}(N^{c^{2}+1+(c^{2}+1)^{2}/k})=O_{c,k}(N^{(c^{2}+1)^{2}/k}).$ By the assumption $k>(c^{2}+1)^{2}$, the expected number of sets with small doubling in $A$ is $o_{c,k}(N)$. Combining this with (5) implies that, with positive probability, we have that $\frac{3N}{2}\leq\lvert A\rvert\leq\frac{9N}{2}$ and that the number of sets with small doubling is $o_{c,k}(N)$. Thus, there exists $A$ that satisfies both of these properties. We fix such a set $A$ and arbitrarily remove one number from each of the remaining sets with small doubling. This leads to $|A|=\Theta(N)$ and no sets with small doubling in $A$. In other words, every $k$ elements from $A$ span at least $ck+1$ distinct differences. Since $A\subset M$, we have that $|A-A|<N^{1+(c^{2}+1)/k}$. By taking $N$ to be a sufficiently large constant mulitple of $n$, we obtain that $|A|\geq n$. We arbitrarily remove elements from $A$ until $|A|=n$. This leads to the assertion of the theorem. ∎ ## 5 Future work We are still far from understanding the behavior of $g(n,k,\ell)$ an have more questions than answers. One main open problem is to identify the quadratic threshold of $g(n,k,\ell)$. Since this turned out to be challenging, we suggest the following problem as a step towards the quadratic threshold. ###### Question 5.1. For what values of $\ell$ can we prove a bound of the form $g(n,k,\ell)=\Omega(n^{c})$ for some $4/3<c<2$? New ideas would be needed for this problem: as mentioned in Remark 2.3, though the idea of dumbbell energies can be generalized to “higher” dumbbell energies, the bounds obtained that way would be of the form $g(n,k,\ell)=\Omega(n^{c})$ for some $c<4/3$. In terms of quadratic lower bounds, we pose the following question. ###### Question 5.2. Is it true that when $k$ is a multiple of four, we have that $g\left(n,k,\frac{k^{2}}{4}\right)=\Omega(n^{2})?$ There is also a natural generalization of the construction of the projected-$(i+1)$-cube, where instead of projecting an $i$-dimensional Boolean hypercube onto $\mathbb{R}$, we could instead project an $i$-dimensional $b$-ary hypercube onto $\mathbb{R}$. Specifically, let $A_{1}=\\{1,2,\ldots,b\\}$ and $D=\\{0,1,\ldots,b-1\\}$. For every $i>1$, we let $s_{i}=4b\cdot\max\\{A_{i-1}-A_{i-1}\\}$ and $T_{i}=s_{i}\cdot D$. We then set $A_{i}=T_{i}+A_{i-1}.$ We note that $1\in A_{i}$ for every $i\geq 1$. This implies that $s_{i}\geq 4b(\max\\{A_{i}\\}-1)$. Since the nonzero elements of $T_{i}$ are significantly larger than the elements of $A_{i}$, we get that $|A_{i}|=b\cdot|A_{i-1}|$. Since $|A_{1}|=b$, we conclude that $|A_{i}|=b^{i}$ for every $i\geq 1$. By definition, every element of $A_{i}-A_{i}$ can be represented as $x+y$ where $x\in T_{i}-T_{i}$ and $y\in A_{i-1}-A_{i-1}$. By our choice of $s_{i}$, we have that $x,x^{\prime}\in T_{i}-T_{i}$ and $y,y^{\prime}\in A_{i-1}-A_{i-1}$ satisfy that $x+y=x^{\prime}+y^{\prime}$ if and only if $(x,y)=(x^{\prime},y^{\prime})$. Indeed, if $x\neq x^{\prime}$ then $\lvert x-x^{\prime}\rvert\geq 4b\max\\{|y|,|y^{\prime}|\\}\geq 2b\lvert y-y^{\prime}\rvert$. In particular, it can be easily checked that as a consequence we have $\lvert A_{i}-A_{i}\rvert=\frac{(2b-1)^{i}-1}{2}$. We conjecture that the natural generalization of [5] should hold in this case. ###### Conjecture 5.3. When $k$ is a power of $b$, $g\left(n,k,\frac{k^{\log_{b}(2b-1)}+1}{2}\right)=O\left(k^{\log_{b}(2b-1)}\right)$. In the same flavor as Theorem 1.6, we may ask the following question. ###### Question 5.4. Can we prove a lower bound for $g\left(n,k,\frac{k^{\log_{b}(2b-1)}+1}{2}\right)$ when $k$ is a power of $b$? ## References * [1] Noga Alon and Joel H Spencer. The probabilistic method. John Wiley & Sons, 2004. * [2] David Conlon, Jacob Fox, Choongbum Lee, and Benny Sudakov. The Erdős–Gyárfás problem on generalized Ramsey numbers. Proceedings of the London Mathematical Society, 110(1):1–18, 2015\. * [3] Paul Erdos. Problems and results on finite and infinite graphs. In Recent advances in graph theory (Proc. Second Czechoslovak Sympos., Prague, 1974), pages 183–192, 1975. * [4] Paul Erdős and András Gyárfás. A variant of the classical ramsey problem. Combinatorica, 17(4):459–467, 1997. * [5] Sara Fish, Ben Lund, and Adam Sheffer. A construction for difference sets with local properties. European Journal of Combinatorics, 79:237–243, 2019. * [6] Sara Fish, Cosmin Pohoata, and Adam Sheffer. Local properties via color energy graphs and forbidden configurations. SIAM Journal on Discrete Mathematics, 34(1):177–187, 2020. * [7] William Timothy Gowers. A new proof of Szemerédi’s theorem. Geometric & Functional Analysis GAFA, 11(3):465–588, 2001. * [8] Ben Green and Robert Morris. Counting sets with small sumset and applications. Combinatorica, 36(2):129–159, 2016. * [9] Cosmin Pohoata and Adam Sheffer. Local properties in colored graphs, distinct distances, and difference sets. Combinatorica, 39(3):705–714, 2019. * [10] Tom Sanders. The structure theory of set addition revisited. Bull. Amer. Math. Soc. (N.S.), 50(1):93–127, 2013. * [11] Tomasz Schoen. New bounds in Balog-Szemerédi-Gowers theorem. Combinatorica, 35(6):695–701, 2015. * [12] Tomasz Schoen and Ilya D Shkredov. Higher moments of convolutions. Journal of Number Theory, 133(5):1693–1737, 2013. * [13] Terence Tao and Van H Vu. Additive combinatorics, volume 105. Cambridge University Press, 2006.
# Termination Analysis of Programs with Multiphase Control-Flow††thanks: This paper is a summary of the corresponding invited talk of the second author given at HCVS 2021, and is based on the PhD thesis of the first author [15] and on collaborations with Amir M. Ben-Amram [9, 5] and John P. Gallagher [16]. This work was funded partially by the Spanish MCIU, AEI and FEDER (EU) project RTI2018-094403-B-C31, by the CM project S2018/TCS-4314 co-funded by EIE Funds of the European Union, and by the UCM CT42/18-CT43/18 grant. Jesús J. Domenech1 Samir Genaim1,2 1 Universidad Complutense de Madrid, Spain2 Instituto de Tecnología del Conocimiento, Madrid, Spain<EMAIL_ADDRESS><EMAIL_ADDRESS> ###### Abstract Programs with multiphase control-flow are programs where the execution passes through several (possibly implicit) phases. Proving termination of such programs (or inferring corresponding runtime bounds) is often challenging since it requires reasoning on these phases separately. In this paper we discuss techniques for proving termination of such programs, in particular: (1) using multiphase ranking functions, where we will discuss theoretical aspects of such ranking functions for several kinds of program representations; and (2) using control-flow refinement, in particular partial evaluation of Constrained Horn Clauses, to simplify the control-flow allowing, among other things, to prove termination with simpler ranking functions. ## 1 Introduction Proving that a program will eventually terminate, i.e., that it does not go into an infinite loop, is one of the most fundamental tasks of program verification, and has been the subject of voluminous research. Perhaps the best known, and often used, technique for proving termination is that of _ranking functions_ , which has already been used by Alan Turing in his early work on program verification [34]. This consists of finding a function $\rho$ that maps program states into the elements of a well-founded ordered set, such that $\rho(s)\succ\rho(s^{\prime})$ holds for any consecutive states $s$ and $s^{\prime}$. This implies termination since infinite descent is impossible in a well-founded order. Besides proving termination, Turing [34] mentions that ranking functions can be used to bound the length computations as well. This is useful in applications such as cost analysis and loop optimisation [17, 4, 2, 12]. Unlike termination of programs in general, which is undecidable, the algorithmic problems of detection (deciding the existence) or generation (synthesis) of a ranking function can well be solvable, given certain choices of the program representation, and the class of ranking functions. There is a considerable amount of research in this direction, in which different kinds of ranking functions for different kinds of program representations were considered. In some cases, the algorithmic problems have been completely settled, and efficient algorithms provided, while other cases remain open. A common program representation in this context is _Single-path Linear- Constraint_ (SLC) loops, where a state is described by the values of numerical variables, and the effect of a transition (one iteration) is described by a conjunction of _linear constraints_. Here is an example of this loop representation; primed variables $x^{\prime}$ and $y^{\prime}$ refer to the state following the transition: $while\leavevmode\nobreak\ (x\leq y)\leavevmode\nobreak\ do\leavevmode\nobreak\ x^{\prime}=x+1,\,y^{\prime}\leq y$ (1) Note that $x^{\prime}=x+1$ is an equation, not an assignment. The description of a loop may involve linear inequalities rather than equations, such as $y^{\prime}\leq y$ above, and consequently, be non-deterministic. Note that for a SLC loop with $n$ variables, a transition can be seen as a point $\bigl{(}\begin{smallmatrix}{\mathbf{x}}\hfill\\\ {\mathbf{x}^{\prime}}\hfill\end{smallmatrix}\bigr{)}\in\mathbb{Q}^{2n}$, where its first $n$ components correspond to $\mathbf{x}$ and its last $n$ components to $\mathbf{x}^{\prime}$. We denote the set of all transitions by ${\mathcal{Q}}$, which is a polyhedron. A more general program representation is _Transition Systems_ (TSs), which are defined by _Control-Flow Graphs_ (CFGs) with numerical variables, consisting of nodes representing program locations and edges annotated with linear constraints (polyhedra) describing how values of variables change when moving from one location to another. Figure 1 includes a program and its corresponding TS ${\cal T}$ to its right. Primed variables in the linear constraints refer to the state following the transition, exactly as in the case of SLC loops. In both program representations mentioned above, the domain of variables is also important as it typically affects the complexity of the underlying decision and synthesis problems. Although these program representations allow only numerical variables and linear constraints, data structures can be handled using _size abstractions_ , e.g., length of lists, depth of trees, etc. [27, 24, 13, 33, 28, 20]. In such case, variables represent sizes of corresponding data structures. ⬇ 1void phases1(int x, int y, int z) { 2 while( x >= 1 ){ 3 if ( y <= z - 1 ) { 4 y = y + 1; 5 } else { 6 x = x - 1; 7 } 8 } 9} $\mathtt{n_{\\_}0}$ $\mathtt{n_{\\_}1}$ $\mathtt{n_{\\_}2}$ $\mathtt{n_{\\_}3}$ ${\cal T}$ $\scriptscriptstyle{\mathcal{Q}}_{\\_}0$ $\scriptscriptstyle{\mathcal{Q}}_{\\_}1$ $\scriptscriptstyle{\mathcal{Q}}_{\\_}2$ $\scriptscriptstyle{\mathcal{Q}}_{\\_}3$ $\scriptscriptstyle{\mathcal{Q}}_{\\_}4$ $\mathtt{n_{\\_}0}$ $\mathtt{n_{\\_}1^{3}}$ $\mathtt{n_{\\_}3^{2}}$ $\mathtt{n_{\\_}2^{2}}$ $\mathtt{n_{\\_}1^{2}}$ ${\cal T}_{\\_}{pe}$ $\mathtt{n_{\\_}2^{1}}$ $\mathtt{n_{\\_}3^{1}}$ $\mathtt{n_{\\_}1^{1}}$ $\scriptscriptstyle{\mathcal{Q}}_{\\_}0$ $\scriptscriptstyle{\mathcal{Q}}_{\\_}1$ $\scriptscriptstyle{\mathcal{Q}}_{\\_}2$ $\scriptscriptstyle{\mathcal{Q}}_{\\_}6$ $\scriptscriptstyle{\mathcal{Q}}_{\\_}5$ $\scriptscriptstyle{\mathcal{Q}}_{\\_}7$ $\scriptscriptstyle{\mathcal{Q}}_{\\_}8$ $\scriptscriptstyle{\mathcal{Q}}_{\\_}{6}$ $\scriptscriptstyle{\mathcal{Q}}_{\\_}{1}$ $\begin{array}[]{r@{}llllll}{\mathcal{Q}}_{\\_}0\equiv&\\{&&x^{\prime}=x,&y^{\prime}=y,&z^{\prime}=z&\\}\\\ {\mathcal{Q}}_{\\_}1\equiv&\\{x\geq 1,&&x^{\prime}=x,&y^{\prime}=y,&z^{\prime}=z&\\}\\\ {\mathcal{Q}}_{\\_}2\equiv&\\{x\leq 0,&&x^{\prime}=x,&y^{\prime}=y,&z^{\prime}=z&\\}\\\ {\mathcal{Q}}_{\\_}3\equiv&\\{y\leq z-1,&&x^{\prime}=x,&y^{\prime}=y+1,&z^{\prime}=z&\\}\\\ {\mathcal{Q}}_{\\_}4\equiv&\\{y\geq z,&&x^{\prime}=x-1,&y^{\prime}=y,&z^{\prime}=z&\\}\\\ {\mathcal{Q}}_{\\_}5\equiv&\\{x\geq 1,&y\leq z-1,&x^{\prime}=x,&y^{\prime}=y+1,&z^{\prime}=z&\\}\\\ {\mathcal{Q}}_{\\_}6\equiv&\\{x\geq 1,&y\geq z,&x^{\prime}=x-1,&y^{\prime}=y,&z^{\prime}=z&\\}\\\ {\mathcal{Q}}_{\\_}7\equiv&\\{x\geq 1,&y\geq z,&x^{\prime}=x,&y^{\prime}=y,&z^{\prime}=z&\\}\\\ {\mathcal{Q}}_{\\_}8\equiv&\\{x\leq 0,&y\geq z,&x^{\prime}=x,&y^{\prime}=y,&z^{\prime}=z&\\}\\\ \end{array}$ Figure 1: A loop with $2$ phases, it corresponding TS ${\cal T}$, and the TS ${\cal T}_{\\_}{pe}$ after applying CFR. Due to practical considerations, termination analysis tools typically focus on classes of ranking functions that can be synthesised efficiently. This, however, does not mean that theoretical aspects of such classes are put aside, as understanding theoretical limits and properties of the underlying problems is necessary for developing practical algorithms. The most popular class of ranking functions in this context is probably that of _Linear Ranking Functions_ (LRFs). A LRF is a function $\rho(x_{\\_}1,\dots,x_{\\_}n)=a_{\\_}1x_{\\_}1+\dots+a_{\\_}nx_{\\_}n+a_{\\_}0$ such that any transition from $\mathbf{x}$ to $\mathbf{x}^{\prime}$ satisfies (i) $\rho(\mathbf{x})\geq 0$; and (ii) $\rho(\mathbf{x})-\rho(\mathbf{x}^{\prime})\geq 1$. For example, $\rho(x,y)=y-x$ is a LRF for Loop (1). Several polynomial-time algorithms to find a LRF using linear programming exist [4, 14, 17, 29, 30, 32]. These algorithms are complete111Complete means that if there is a LRF, they will find one. for TSs with rational-valued variables, but not with integer-valued variables. Ben-Amram and Genaim [6] showed how completeness for the integer case can be achieved, and also classified the corresponding decision problem as co-NP complete. Despite their popularity, LRFs do not suffice for all programs, and a natural question is what to do when a LRF does not exist; and a natural answer is to try a richer class of ranking functions. Of particular importance is the class of _Lexicographic-Linear Ranking Functions_ (LLRFs). These are tuples of linear functions that decrease lexicographically over a corresponding well- founded ordered set. For example, the program depicted in Figure 1 does not have a LRF, but it has the LLRF $\langle z-y,x\rangle$ since: for the _then_ branch $z-y$ decreases and for the _else_ branch $x$ decreases while $z-y$ does not change. LLRFs might be necessary even for SLC loops. For example, the following SLC loop $while\leavevmode\nobreak\ (x\geq 0)\leavevmode\nobreak\ do\leavevmode\nobreak\ x^{\prime}=x+y,\,y^{\prime}=y+z,z^{\prime}=z-1$ (2) does not have a LRF, but can be proved terminating using the LLRF $\langle z,y,x\rangle$. There are several definitions for LLRFs in the literature [4, 7, 11, 23] and they have different power, i.e., some can prove termination of a program while others fail. They have corresponding polynomial-time synthesis algorithms for the case of rational variables, and the underlying decision problems are co-NP complete for the case of integer variables [7, 8]. The definition of Larraz et al. [23] is the most general in this spectrum of LLRFs, its complexity classification is not known yet and corresponding complete synthesis algorithms do not exist. The LLRF $\langle z,y,x\rangle$ used above for Loop (2) belongs to a class of LLRFs that is know as _Multiphase-Linear Ranking Functions_ (M$\Phi$RFs). Ranking functions in this class are characterised by the following behaviour: the first component _always_ decreases, and when it becomes negative the second component starts to decrease (and will keep decreasing during the rest of the execution), and when it becomes negative the third component starts to decrease, and so on. This behaviour defines phases through which executions pass. This is different from LLRFs in general since once a component becomes negative it cannot be used anymore. Also the LLRF $\langle z-y,x\rangle$ for the program of Figure 1 induces a similar multiphase behaviour, since once $z-y$ becomes negative it cannot be used anymore. However, it is not a M$\Phi$RF since $z-y$ stops decreasing once we move to the second phase. In the rest of this paper we overview works on termination analysis of multiphase programs, in particular: in Section 2 we discuss techniques for proving termination using M$\Phi$RFs; and in Section 3 we discuss techniques for proving termination using _Control-Flow Refinement_ (CFR). Section 4 ends the paper with some concluding remarks and discusses further research directions. ## 2 Termination analysis using multiphase ranking functions A M$\Phi$RF is a tuple $\langle\rho_{\\_}1,\ldots,\rho_{\\_}d\rangle$ of linear functions that define phases of the program that are linearly ranked, where $d$ is the _depth_ of the M$\Phi$RF, intuitively the number of phases. The decision problem _Existence of a M $\Phi$RF_ asks to determine whether a program has a M$\Phi$RF. The _bounded_ decision problem restricts the search to M$\Phi$RFs of depth $d$, where $d$ is part of the input. For the case of SLC loops, the complexity and algorithmic aspects of the bounded version of the M$\Phi$RF problem were settled by Ben-Amram and Genaim [9]. The decision problem is $\mathtt{PTIME}$ for SLC loops with rational- valued variables, and $\mathtt{coNP}$-complete for SLC loops with integer- valued variables; synthesising M$\Phi$RFs, when they exist, can be performed in polynomial and exponential time, respectively. We note that the proof of the rational case is done by showing that M$\Phi$RFs and _nested_ ranking functions [25] (a strict subclass of M$\Phi$RFs for which a polynomial-time algorithm exists) have the same power for SLC loops. Besides, they show that for SLC loops M$\Phi$RFs have the same power as general lexicographic-linear ranking functions, and that M$\Phi$RFs induce linear iteration bounds. The problem of deciding if a given SLC loop admits a M$\Phi$RF, without a given bound on the depth, is still open. In practice, termination analysis tools search for M$\Phi$RFs incrementally, starting by depth $1$ and increase the depth until they find one, or reach a predefined limit, after which the returned answer is _don’t know_. Finding a theoretical upper-bound on the depth of a M$\Phi$RF, given the loop, would also settle this problem, however, as shown by Ben-Amram and Genaim [9] such bound must depend not only on the number of constraints or variables, as for other classes of LLRFs [4, 7, 11], but also on the coefficients used in the corresponding constraints. Yuan et al. [35] proposed an incomplete method to bound the depth of M$\Phi$RFs for SLC loops. Ben-Amram et al. [5] have done a significant progress towards solving the problem of _existence of a M $\Phi$RF_ for SLC loop, i.e., seeking a M$\Phi$RF without a given bound on the depth. In particular, they present an algorithm for seeking M$\Phi$RFs that reveals novel insights on the structure of these ranking functions. In a nutshell, the algorithm starts from the set of transitions of the given SLC loop, which is a polyhedron, and iteratively removes transitions $\bigl{(}\begin{smallmatrix}{\mathbf{x}}\hfill\\\ {\mathbf{x}^{\prime}}\hfill\end{smallmatrix}\bigr{)}$ such that $\rho(\mathbf{x})-\rho(\mathbf{x}^{\prime})>0$ for some function $\rho(\mathbf{x})=\vec{a}\cdot\mathbf{x}+b$ that is _non-negative on all enabled states_. The process continues iteratively, since after removing some transitions, more functions $\rho$ may satisfy the non-negativity condition, and they may eliminate additional transitions in the next iteration. When all transitions are eliminated in a finite number of iterations, one can construct a M$\Phi$RF using the $\rho$ functions; and when reaching a situation in which no transition can be eliminated, the remaining set of transitions, which is a polyhedron, is actually a recurrent set that witnesses non-termination. The algorithm always finds a M$\Phi$RF if one exists, and in many cases, it finds a recurrent set when the loop is non-terminating, however, it is not a decision procedure as it diverges in some cases. Nonetheless, the algorithm provides important insights into the structure of M$\Phi$RFs. Apart from revealing a relation between seeking M$\Phi$RFs and seeking recurrent sets, these insights are useful for finding classes of SLC loops for which, when terminating, there is always a M$\Phi$RF and thus have linear run-time bound. This result was proven for two kinds of SLC loops, both considered in previous work, namely _octagonal relations_ and _affine relations with the finite- monoid property_ – for both classes, termination has been proven decidable [22]. Ben-Amram et al. [5] have also suggested a new representation for SLC loops, called the _displacement_ representation, that provides new tools for studying termination of SLC loops in general, and the existence of a M$\Phi$RF in particular. In this representation, a transition $\bigl{(}\begin{smallmatrix}{\mathbf{x}}\hfill\\\ {\mathbf{x}^{\prime}}\hfill\end{smallmatrix}\bigr{)}$ is represented as $\bigl{(}\begin{smallmatrix}{\mathbf{x}}\hfill\\\ {\mathbf{y}}\hfill\end{smallmatrix}\bigr{)}$ where $\mathbf{y}=\mathbf{x}^{\prime}-\mathbf{x}$. Using this representation the algorithm described above can be formalised in a simple way that avoids computing the $\rho$ functions mentioned above, and reduces the existence of a M$\Phi$RF of depth $d$ to unsatisfiability of a certain linear constraint system. Moreover, any satisfying assignment for this linear constraint system is a witness that explains why the loop has no M$\Phi$RF of depth $d$. As an evidence on the usefulness of this representation in general, they showed that some non-trivial observations on termination of bounded SLC loops (i.e., the set of transitions is a bounded polyhedron) are made straightforward in this representation, while they are not easy to see in the normal representation, in particular: a bounded SLC loop terminates if and only if it has a LRF, and it does not terminate if and only if it has a fixpoint transition $\bigl{(}\begin{smallmatrix}{\mathbf{x}}\hfill\\\ {\mathbf{x}}\hfill\end{smallmatrix}\bigr{)}$. The works discussed above are limited to the case of SLC loops. The case of general TSs has been considered by Leike and Heizmann [25] and Li et al. [26], where both translate the existence of a M$\Phi$RF of a given depth $d$ to solving a corresponding non-linear constraint problem, which is complete for the case of rational variables. However, while complete, these approaches do not provide any insights on the complexity of the underlying decision problems. The technique of Borralleras et al. [10] can be used to infer, among other things, M$\Phi$RFs for TSs without a given bound on the depth. It is based on solving corresponding safety problems using Max-SMT, and it is not complete. ## 3 Termination analysis using control-flow refinement As we have seen in Section 1, there are TSs with multiphase behaviour that do not admit M$\Phi$RFs, but rather a different notion of LLRF that does not require the components to keep decreasing after turning negative. To prove termination of such TSs one can use other classes of LLRFs [4, 7, 11, 23], however, this might not be enough due to complex control-flow where abstract properties of several implicit execution paths are merged. To overcome this imprecision, not only for termination but for program analysis in general, one approach is to simplify the control-flow in order to make implicit execution paths explicit, and thus makes it possible to infer the desired properties with a weaker version of the analysis. This is known as _Control-Flow Refinement_ (CFR). Let us see how CFR can simplify the kind of ranking functions needed to proved termination, and thus improve the precision of the underlying termination analyser. Consider the program depicted Figure 1 again, and recall that we failed to prove its termination when using only LRFs. Examining this program carefully, we can see that any execution passes in two phases: in the first one, $\mathtt{y}$ is incremented until it reaches the value of $\mathtt{z}$, and in the second phase $\mathtt{x}$ is decremented until it reaches $\mathtt{0}$. Lets us transform the program into a semantically equivalent one such that the two phases are separate and explicit: ⬇ 1 while (x >= 0 and y <= z-1) y = y + 1; 2 while (x >= 0 and y >= z) x = x - 1; Now we can prove termination of this program using LRFs only: for the first loop $z-y$ is a LRF, and for the second loop $x$ is a LRF. Moreover, cost analysis tools that are based on bounding loop iterations using LRFs [2] would infer a linear bound for this program while they would fail on the original one. Apart from simplifying the termination proof, there are also cases where it is not possible to prove termination without such transformations even when using LLRFs. In addition, CFR can help in inferring more precise invariants, without the need for expensive disjunctive abstract domains, which can benefit any analysis that uses such invariants, e.g., termination and cost analysis. CFR has been considered by Gulwani et al. [21] and Flores-Montoya and Hähnle [18] to improve the precision of cost analysis, and by Sharma et al. [31] to improve the precision of invariants in order to prove program assertions. While all these techniques can automatically obtain the transformed program above, they are developed from scratch and tailored to some analysis of interest. Recently, CFR has also been considered by Albert et al. [3] to improve cost analysis as well, but from a different perspective that uses termination witnesses to guide CFR. Since CFR is, in principle, a program transformation that specialises programs to distinguish different execution scenarios, Domenech et al. [16] explored the use of general-purpose specialisation techniques for CFR, in particular the techniques of Gallagher [19] for partial evaluation of Constrained Horn Clauses (CHCs). Basing CFR on partial evaluation has the clear advantage that soundness comes for free because partial evaluation guarantees semantic equivalence between the original program and its transformed version. Moreover, this way we obtain a CFR procedure that is not tailored for a particular purpose, but rather can be tuned depending on the application domain. Domenech et al. [16] developed such a CFR procedure for TSs, by transforming TSs into CHCs and using the partial evaluator of Gallagher [19], and integrated it in a termination analysis algorithm in a way that allows applying CFR at different levels of granularity, and thus controlling the trade-off between precision and performance. This is done by suggesting different schemes for applying CFR, not only as a preprocessing step but also on specific parts of the TS which we could not prove terminating. Moreover, they developed heuristics for automatically configuring partial evaluation (i.e., inferring properties to guides specialisation) in order to achieve the desired CFR. Experimental evaluation provides a clear evidence to that their CFR procedure significantly improves the precision of termination analysis, cost analysis, and invariants generation. Let use demonstrate this CFR procedure on the TS ${\cal T}$ that is depicted in Figure 1. In a first step, the TS ${\cal T}$ is translated into the following (semantically) equivalent CHC program: $\begin{array}[]{rlllll}q_{\\_}{\mathtt{n_{\\_}0}}(x,y,z)\leftarrow&q_{\\_}{\mathtt{n_{\\_}1}}(x,y,z).\\\ q_{\\_}{\mathtt{n_{\\_}1}}(x,y,z)\leftarrow&\\{x\geq 1\\},&q_{\\_}{\mathtt{n_{\\_}2}}(x,y,z).\\\ q_{\\_}{\mathtt{n_{\\_}1}}(x,y,z)\leftarrow&\\{x\leq 0\\},&q_{\\_}{\mathtt{n_{\\_}3}}(x,y,z).\\\ q_{\\_}{\mathtt{n_{\\_}2}}(x,y,z)\leftarrow&\\{y\leq z-1,&y^{\prime}=y+1\\},&q_{\\_}{\mathtt{n_{\\_}1}}(x,y^{\prime},z).\\\ q_{\\_}{\mathtt{n_{\\_}2}}(x,y,z)\leftarrow&\\{y\geq z,&x^{\prime}=x-1\\},&q_{\\_}{\mathtt{n_{\\_}1}}(x^{\prime},y,z).\\\ \end{array}$ Then the partial evaluator of Gallagher [19] is applied, using the (automatically inferred) properties $\\{x\geq 1,y\geq z\\}$ for the loop head predicate $q_{\\_}{\mathtt{n_{\\_}1}}(x,y,z)$, which results in the following CHC program: $\begin{array}[]{rllllll}q_{\\_}{\mathtt{n_{\\_}0}}(x,y,z)\leftarrow&q_{\\_}{\mathtt{n_{\\_}1^{3}}}(x,y,z).\\\ q_{\\_}{\mathtt{n_{\\_}1^{3}}}(x,y,z)\leftarrow&\\{x\leq 0\\},&q_{\\_}{\mathtt{n_{\\_}3^{2}}}(x,y,z).\\\ q_{\\_}{\mathtt{n_{\\_}1^{3}}}(x,y,z)\leftarrow&\\{x\geq 1\\},&q_{\\_}{\mathtt{n_{\\_}2^{2}}}(x,y,z).\\\ q_{\\_}{\mathtt{n_{\\_}2^{2}}}(x,y,z)\leftarrow&\\{x\geq 1,&y\leq z-1,&y^{\prime}=y+1\\},&q_{\\_}{\mathtt{n_{\\_}1^{2}}}(x,y^{\prime},z).\\\ q_{\\_}{\mathtt{n_{\\_}2^{2}}}(x,y,z)\leftarrow&\\{x\geq 1,&y\geq z,&x^{\prime}=x-1\\},&q_{\\_}{\mathtt{n_{\\_}1^{1}}}(x^{\prime},y,z).\\\ q_{\\_}{\mathtt{n_{\\_}1^{2}}}(x,y,z)\leftarrow&\\{x\geq 1\\},&q_{\\_}{\mathtt{n_{\\_}2^{2}}}(x,y,z).\\\ q_{\\_}{\mathtt{n_{\\_}1^{1}}}(x,y,z)\leftarrow&\\{x\leq 0,&y\geq z\\},&q_{\\_}{\mathtt{n_{\\_}3^{1}}}(x,y,z).\\\ q_{\\_}{\mathtt{n_{\\_}1^{1}}}(x,y,z)\leftarrow&\\{x\geq 1,&y\geq z\\},&q_{\\_}{\mathtt{n_{\\_}2^{1}}}(x,y,z).\\\ q_{\\_}{\mathtt{n_{\\_}2^{1}}}(x,y,z)\leftarrow&\\{x\geq 1,&y\geq z,&x^{\prime}=x-1\\},&q_{\\_}{\mathtt{n_{\\_}1^{1}}}(x^{\prime},y,z).\\\ \end{array}$ Now translating this CHC program into a TS results in the TS ${\cal T}_{\\_}{pe}$ that is depicted in Figure 1. Note that the two phases are now separated into two different strongly connected components, which can be proven terminating using only LRFs. ## 4 Concluding Remarks In this paper we have discussed techniques for proving termination of programs with multiphase control-flow, where the execution passes through several (possibly implicit) phases. In particular, we have discussed techniques that correspond to our recent work: (1) the use of multiphase ranking functions [9, 5]; and (2) the use of control-flow refinement [16]. As a byproduct of our research, we have developed an open-source termination analyser called iRankFinder 222http://irankfinder.loopkiller.com that implements all these techniques as well as other state-of-the-art techniques. Some of the components of iRankFinder can be used independently, in particular the CFR component that can be used to incorporate CFR in static analysers with little effort. ### Future Work. For M$\Phi$RFs, an obvious future direction is to study the problem of deciding whether a TS has a M$\Phi$RF, both from algorithmic and theoretical complexity perspectives. In initial unpublished work, we have proven that the corresponding decision problem is NP-hard for the rational setting, but we could not obtain a further classification. Further exploration of the M$\Phi$RF problem for SLC loops is also required since it is not solved for the general case yet. For CFR, we have developed some heuristics for the automatic inference of properties, which is crucial for obtaining the desired transformations, but further research in this direction is required. We concentrated on CFR of TSs, and in a future direction one could apply our CFR techniques for program representations that allow recursion as well. Technically, this would not require much work since the partial evaluation techniques of Gallagher [19] specialise CHCs that include recursion already. We also concentrated on numerical programs, and a possible future direction can concentrate on using CFR for program analysis tools where the data is not necessarily numerical. Here one should also adapt the partial evaluation techniques to support such specialisations, which seems doable for the partial evaluation techniques of Gallagher [19] since it is based on using abstract properties like those used in abstract domains of program analysis. ## References * [1] * [2] Elvira Albert, Puri Arenas, Samir Genaim & Germán Puebla (2011): _Closed-Form Upper Bounds in Static Cost Analysis_. Journal of Automated Reasoning 46(2), pp. 161–203, 10.1007/s10817-010-9174-1. * [3] Elvira Albert, Miquel Bofill, Cristina Borralleras, Enrique Martin-Martin & Albert Rubio (2019): _Resource Analysis driven by (Conditional) Termination Proofs_. Theory and Practice of Logic Programming 19(5-6), pp. 722–739, 10.1017/S1471068419000152. * [4] Christophe Alias, Alain Darte, Paul Feautrier & Laure Gonnord (2010): _Multi-dimensional Rankings, Program Termination, and Complexity Bounds of Flowchart Programs_. In Radhia Cousot & Matthieu Martel, editors: Static Analysis Symposium, SAS’10, LNCS 6337, Springer, pp. 117–133, 10.1007/978-3-642-15769-18. * [5] Amir M. Ben-Amram, Jesús J. Doménech & Samir Genaim (2019): _Multiphase-Linear Ranking Functions and Their Relation to Recurrent Sets_. In Bor-Yuh Evan Chang, editor: Proceedings of the 26th International Symposium on Static Analysis, SAS’19, Lecture Notes in Computer Science 11822, Springer, pp. 459–480, 10.1007/978-3-030-32304-222. * [6] Amir M. Ben-Amram & Samir Genaim (2013): _On the Linear Ranking Problem for Integer Linear-Constraint Loops_. In: Principles of programming languages, POPL’13, ACM, pp. 51–62, 10.1145/2480359.2429078. * [7] Amir M. Ben-Amram & Samir Genaim (2014): _Ranking Functions for Linear-Constraint Loops_. Journal of the ACM 61(4), pp. 26:1–26:55, 10.1145/2629488. * [8] Amir M. Ben-Amram & Samir Genaim (2015): _Complexity of Bradley-Manna-Sipma Lexicographic Ranking Functions_. In Daniel Kroening & Corina S. Păsăreanu, editors: Computer Aided Verification, CAV’14, LNCS 9207, Springer, pp. 304–321, 10.1007/978-3-319-21668-318. * [9] Amir M. Ben-Amram & Samir Genaim (2017): _On Multiphase-Linear Ranking Functions_. In Rupak Majumdar & Viktor Kuncak, editors: Computer Aided Verification, CAV’17, LNCS 10427, Springer, pp. 601–620, 10.1007/978-3-319-63390-932. * [10] Cristina Borralleras, Marc Brockschmidt, Daniel Larraz, Albert Oliveras, Enric Rodríguez-Carbonell & Albert Rubio (2017): _Proving Termination Through Conditional Termination_. In Axel Legay & Tiziana Margaria, editors: Tools and Algorithms for the Construction and Analysis of Systems, TACAS’17, LNCS 10205, pp. 99–117, 10.1007/978-3-662-54577-56. * [11] Aaron R. Bradley, Zohar Manna & Henny B. Sipma (2005): _Linear Ranking with Reachability_. In Kousha Etessami & Sriram K. Rajamani, editors: Computer Aided Verification, CAV’05, LNCS 3576, Springer, pp. 491–504, 10.1007/1151398848. * [12] Marc Brockschmidt, Fabian Emmes, Stephan Falke, Carsten Fuhs & Jürgen Giesl (2016): _Analyzing Runtime and Size Complexity of Integer Programs_. ACM Transactions on Programming Languages and Systems, TOPLAS’16 38(4), pp. 13:1–13:50. Available at https://doi.org/10.1145/2866575. * [13] Maurice Bruynooghe, Michael Codish, John P. Gallagher, Samir Genaim & Wim Vanhoof (2007): _Termination analysis of logic programs through combination of type-based norms_. ACM Transactions on Programming Languages and Systems, TOPLAS’07 29(2), pp. 10–54. Available at https://doi.org/10.1145/1216374.1216378. * [14] Michael Colón & Henny Sipma (2001): _Synthesis of Linear Ranking Functions_. In Tiziana Margaria & Wang Yi, editors: Tools and Algorithms for the Construction and Analysis of Systems, TACAS’01, LNCS 2031, Springer, pp. 67–81, 10.1007/3-540-45319-96. * [15] Jesús J. Domenech (2021): _Termination Analysis of Programs with Complex Control-Flow_. Ph.D. thesis, Facultad de Informática, Universidad Complutense de Madrid. * [16] Jesús J. Doménech, John P. Gallagher & Samir Genaim (2019): _Control-Flow Refinement by Partial Evaluation, and its Application to Termination and Cost Analysis_. Theory Pract. Log. Program. 19(5-6), pp. 990–1005, 10.1017/S1471068419000310. * [17] Paul Feautrier (1992): _Some efficient solutions to the affine scheduling problem. I. One-dimensional time_. International Journal of Parallel Programming 21(5), pp. 313–347. Available at https://doi.org/10.1007/BF01407835. * [18] Antonio Flores-Montoya & Reiner Hähnle (2014): _Resource Analysis of Complex Programs with Cost Equations_. In Jacques Garrigue, editor: Asian Symposium on Programming Languages and Systems, APLAS’14, LNCS 8858, Springer, pp. 275–295. Available at https://doi.org/10.1007/978-3-319-12736-1_15. * [19] John P. Gallagher (2019): _Polyvariant program specialisation with property-based abstraction_. In Alexei Lisitsa & Andrei P. Nemytykh, editors: Verification and Program Transformation, VPT’19, 299, Open Publishing Association, pp. 34–48. Available at https://doi.org/10.4204/eptcs.299.6. * [20] Jürgen Giesl, Cornelius Aschermann, Marc Brockschmidt, Fabian Emmes, Florian Frohn, Carsten Fuhs, Jera Hensel, Carsten Otto, Martin Plücker, Peter Schneider-Kamp, Thomas Ströder, Stephanie Swiderski & René Thiemann (2017): _Analyzing Program Termination and Complexity Automatically with AProVE_. Journal of Automated Reasoning 58(1), pp. 3–31. Available at https://doi.org/10.1007/s10817-016-9388-y. * [21] Sumit Gulwani, Sagar Jain & Eric Koskinen (2009): _Control-flow refinement and progress invariants for bound analysis_. In Michael Hind & Amer Diwan, editors: Programming Language Design and Implementation, PLDI’09, 44, ACM, pp. 375–385. Available at https://doi.org/10.1145/1542476.1542518. * [22] Radu Iosif, Filip Konečný & Marius Bozga (2014): _Deciding Conditional Termination_. Logical Methods in Computer Science 10(3), 10.2168/lmcs-10(3:8)2014. Available at http://doi.org/10.2168/LMCS-10(3:8)2014. * [23] Daniel Larraz, Albert Oliveras, Enric Rodríguez-Carbonell & Albert Rubio (2013): _Proving termination of imperative programs using Max-SMT_. In: Formal Methods in Computer-Aided Design, FMCAD’13, IEEE, pp. 218–225. Available at https://doi.org/10.1109/FMCAD.2013.6679413. * [24] Chin Soon Lee, Neil D. Jones & Amir M. Ben-Amram (2001): _The size-change principle for program termination_. In Chris Hankin & Dave Schmidt, editors: Principles of Programming Languages, POPL’01, ACM, pp. 81–92. Available at https://doi.org/10.1145/360204.360210. * [25] Jan Leike & Matthias Heizmann (2015): _Ranking Templates for Linear Loops_. Logical Methods in Computer Science 11(1), pp. 1–27, 10.2168/LMCS-11(1:16)2015. * [26] Yi Li, Guang Zhu & Yong Feng (2016): _The L-Depth Eventual Linear Ranking Functions for Single-Path Linear Constraint Loops_. In: Theoretical Aspects of Software Engineering, TASE’16, IEEE, pp. 30–37. Available at https://doi.org/10.1109/TASE.2016.8. * [27] Naomi Lindenstrauss & Yehoshua Sagiv (1997): _Automatic Termination Analysis of Logic Programs_. In Lee Naish, editor: International Conference on Logic Programming, ICLP’97, MIT Press, pp. 64–77. Available at https://ieeexplore.ieee.org/document/6279160. * [28] Stephen Magill, Ming-Hsien Tsai, Peter Lee & Yih-Kuen Tsay (2010): _Automatic numeric abstractions for heap-manipulating programs_. In Manuel V. Hermenegildo & Jens Palsberg, editors: Principles of Programming Languages, POPL’10, ACM, pp. 211–222. Available at https://doi.org/10.1145/1706299.1706326. * [29] Frédéric Mesnard & Alexander Serebrenik (2008): _Recurrence with affine level mappings is P-time decidable for CLP(R)_. Theory and Practice of Logic Programming, TPLP’08 8(1), pp. 111–119. Available at https://doi.org/10.1017/S1471068407003122. * [30] Andreas Podelski & Andrey Rybalchenko (2004): _A Complete Method for the Synthesis of Linear Ranking Functions_. In Bernhard Steffen & Giorgio Levi, editors: Verification, Model Checking, and Abstract Interpretation, VMCAI’04, LNCS 2937, Springer, pp. 239–251, 10.1007/978-3-540-24622-020. * [31] Rahul Sharma, Isil Dillig, Thomas Dillig & Alex Aiken (2011): _Simplifying Loop Invariant Generation Using Splitter Predicates_. In Ganesh Gopalakrishnan & Shaz Qadeer, editors: Computer Aided Verification, CAV’11, LNCS 6806, Springer, pp. 703–719, 10.1007/s10990-006-8609-1. * [32] Kirack Sohn & Allen Van Gelder (1991): _Termination Detection in Logic Programs using Argument Sizes_. In Daniel J. Rosenkrantz, editor: Principles of Database Systems, PoDS’91, ACM Press, pp. 216–226. Available at https://doi.org/10.1145/113413.113433. * [33] Fausto Spoto, Fred Mesnard & Étienne Payet (2010): _A termination analyzer for Java bytecode based on path-length_. ACM Transactions on Programming Languages and Systems, TOPLAS’10 32(3). Available at https://doi.org/10.1145/1709093.1709095. * [34] Alan M. Turing (1949): _Checking a Large Routine_. In: Report on a Conference on High Speed Automatic Computation, June 1949, University Mathematical Laboratory, Cambridge University, pp. 67–69. Available at http://www.turingarchive.org/browse.php/B/8. * [35] Yue Yuan, Yi Li & Wenchang Shi (2021): _Detecting multiphase linear ranking functions for single-path linear-constraint loops_. Int. J. Softw. Tools Technol. Transf. 23(1), pp. 55–67, 10.1007/s10009-019-00527-1.
# Neutrino meets ultralight dark matter: $\boldsymbol{0\nu\beta\beta}$ decay and cosmology Guo-yuan Huang<EMAIL_ADDRESS>Max-Planck-Institut für Kernphysik, Saupfercheckweg 1, 69117 Heidelberg, Germany Newton Nath <EMAIL_ADDRESS>Istituto Nazionale di Fisica Nucleare, Via Orabona 4, 70126 Bari, Italy ###### Abstract We explore the neutrinoless double beta ($0\nu\beta\beta$) decay induced by an ultralight dark matter field coupled to neutrinos. The effect on $0\nu\beta\beta$ decay is significant if the coupling violates the lepton number, for which the $\Delta L=2$ transition is directly driven by the dark matter field without further suppression of small neutrino masses. As the ultralight dark matter can be well described by a classical field, the effect features a periodic modulation pattern in decay events. However, we find that in the early Universe such coupling will be very likely to alter the standard cosmological results. In particular, the requirement of neutrino free- streaming before the matter-radiation equality severely constrains the parameter space, such that the future $0\nu\beta\beta$ decay experiments can hardly see any signal even with a meV sensitivity to the effective neutrino mass. ## I Introduction In spite of its remarkable success, the Standard Model (SM) of particle physics fails to address several fundamental issues, e.g., the nature of dark matter (DM). The mass of viable DM candidates can cover a huge landscape spanning from the primordial black holes Bird:2016dcv ; Carr:2016drx ; Chen:2016pud ; Georg:2017mqk to the ultralight regime Hu:2000ke . At present, the DM direct detection experiments such as Xenon1T XENON:2020gfr , PandaX-II PandaX-II:2017hlx come out with null signal of weakly interacting massive particle (WIMP), one of the most promising DM candidates, hence setting stringent limits on the available WIMP parameter space. Another particular DM class, the ultralight DM Ferreira:2020fam ; Urena-Lopez:2019kud with a mass ranging from $\mathcal{O}(10^{-22})~{}{\rm eV}$ 111This is often called fuzzy dark matter Hu:2000ke ; Hui:2016ltb . The term “fuzzy” corresponds to a huge Compton wavelength, $\lambda=2\pi/m_{\phi}\simeq 0.4\,{\rm pc}\times(10^{-22}\,\text{eV}/m_{\phi})$ with $m_{\phi}$ being the DM mass. to $\mathcal{O}(1)~{}{\rm eV}$, for example the well-motivated quantum chromodynamics (QCD) axion Kim:2008hd ; Graham:2015ouw ; Irastorza:2018dyq ; Marsh:2015xka , is receiving renewing attention. Different from WIMP, the ultralight DM is produced non-thermally in the early Universe and can be well described by a classical-number field. Beside DM, the nature of massive neutrinos, i.e., whether they are Majorana or Dirac fermions, is yet to be answered. At present, the only feasible process that can uncover the nature of neutrinos is the neutrinoless double beta ($0\nu\beta\beta$) decay, which violates the lepton number by two units DellOro:2016tmg ; Vergados:2016hso . There have been many attempts to connect the mentioned two mysterious sectors, i.e., neutrinos and ultralight DM Berlin:2016woy ; Brdar:2017kbt ; Krnjaic:2017zlz ; Liao:2018byh ; Capozzi:2018bps ; Reynoso:2016hjr ; Huang:2018cwo ; Pandey:2018wvh ; Farzan:2018pnk ; Choi:2019ixb ; Baek:2019wdn ; Choi:2019zxy ; Choi:2020ydp ; Dev:2020kgz ; Baek:2020ovw ; Losada:2021bxx ; Smirnov:2021zgn ; Alonso-Alvarez:2021pgy ; Huang:2021zzz . In this work, we investigate $0\nu\beta\beta$ decays in the presence of a coupling between neutrinos and the ultralight DM with a general mass spanning from $10^{-22}~{}{\rm eV}$ to sub-eV. Among general interaction forms, the scalar and tensor couplings are able to violate the lepton number (or equivalently, the DM particle carries two units of lepton number), and hence important for $0\nu\beta\beta$-decay experiments. The Lagrangian reads $\displaystyle-\mathcal{L}^{\rm M}_{\rm int}=\begin{dcases}\frac{g_{\phi}^{\alpha\beta}}{2}\,\phi\,\overline{\nu_{\rm L\alpha}}\nu^{\rm c}_{\rm L\beta}+{\rm h.c.}\;,\\\\[2.15277pt] \frac{g_{T}^{\alpha\beta}}{2}T^{\mu\nu}\overline{\nu_{\rm L\alpha}}\sigma_{\mu\nu}\nu^{\rm c}_{\rm L\beta}+{\rm h.c.}\;,\end{dcases}$ (1) where $\alpha,\beta=e,\mu,\tau$ are the flavor indices, and $\phi$ and $T^{\mu\nu}$ represent real scalar and tensor fields with neutrino coupling constants $g_{\phi}^{\alpha\beta}$ and $g_{T}^{\alpha\beta}$, respectively. We have assumed that neutrinos are Majorana particles, such that only the field $\nu_{\rm L}$ is involved at low energies. In analogy with the electromagnetic force, the tensor interaction can be generated by a dark photon field $A^{\prime}$ coupled to neutrinos with a dark magnetic dipole moment Capozzi:2018bps . However, for Majorana neutrinos, only transition magnetic moment exists and the diagonal part of the tensor coupling constant $g_{T}^{\alpha\beta}$ is vanishing, thus not contributing to the $0\nu\beta\beta$ process which exchanges two electron neutrinos. Hence, we will restrict our following discussions to only the ultralight scalar DM. It is worth mentioning that the transition magnetic moment can be relevant for the lepton-number-violating meson decays by exchanging different flavors of neutrinos. For Majorana neutrinos, $\nu\equiv\nu_{\rm L}+{\nu^{\rm c}_{\rm L}}$, in the background of scalar DM, the Hamiltonian density of the system is $\displaystyle\mathcal{H}=\frac{1}{2}(\partial\phi)^{2}+\frac{1}{2}m^{2}_{\phi}\phi^{2}+\frac{1}{2}\overline{\nu}(\mathrm{i}\not{\partial}+\widehat{M}_{\nu}+g\phi)\nu\;,$ (2) where $m_{\phi}$ is the DM mass, $\widehat{M}_{\nu}$ is the neutrino mass matrix in vacuum, and the coupling constant $g$ is a real symmetric matrix in general. Note that the pseudoscalar coupling has been ignored in the expression, which will induce a similar effect. The corresponding equation of motion is given by $\displaystyle(\partial^{2}+m^{2}_{\phi})\phi=-\frac{1}{2}g\overline{\nu}\nu\;,$ (3) $\displaystyle i\not{\partial}\nu=(\widehat{M}_{\nu}+g\phi)\nu\;.$ (4) As long as the neutrino number density is not too large in Eq. (3), the ultralight DM evolves as free classical field in the Universe as $\phi=\phi_{0}\cos{m_{\phi}t}$ after the scalar mass “freeze-in”, i.e., $H<m_{\phi}$, with $H$ being the Hubble expansion rate. The field strength is given by $\phi_{0}=\sqrt{2\rho}/m_{\phi}$, where $\rho\approx m_{\phi}n_{\phi}\approx 0.3~{}{\rm GeV\cdot cm^{-3}}$ is the DM energy density in our local galaxy. The large occupation number of local ultralight DM, namely $N=n_{\phi}\lambda^{3}\approx 6\times 10^{36}\,(10^{-10}~{}{\rm eV}/m_{\phi})^{4}$ within the cube of Compton wavelength, guarantees the classical-field description as an excellent approximation to the more fundamental quantum field. The formation of the classical field in our local galaxy will generate a time-varying effective Majorana mass $m_{\nu,{\rm eff}}=g\phi$ to neutrinos, which is very similar to the Higgs mechanism. At the particle level, the $0\nu\beta\beta$ decays induced by ultralight DM are shown as diagrams in Fig. 1. The illustrated leading-order contribution is accurate enough since the coupling of our concern is set to be extremely small. The emission process has been discussed before in the context of Majoron model Blum:2018ljv ; Brune:2018sab ; Cepedello:2018zvr . However, different from the “spontaneous” Majoron emission process, in our case both the absorption and emission of scalars benefit from the large occupation number of ultralight DM, which gives rise to a dramatic stimulation enhancement. Figure 1: Feynman diagrams of $0\nu\beta\beta$ decays induced by the absorption and emission of ultralight scalar DM. Because these transitions take place in a background of identical scalar bosons, there will be an enhancement of stimulation. In the rest of the work, we first describe how the coupled system of ultralight DM and neutrinos evolves in the early Universe in Sec. II. Based on the robust cosmic microwave background (CMB) observations, we confine the allowed parameter space of the mass and coupling of ultralight DM. In Sec. III, we explore the possible impacts on $0\nu\beta\beta$ decays. Conclusions are presented in Sec. IV. ## II Cosmological evolution A popular way to generate the coherent scalar field in the early Universe is via the misalignment mechanism, which is well known for the QCD axion Kim:2008hd ; Graham:2015ouw ; Irastorza:2018dyq ; Marsh:2015xka . The production is typically associated with a global symmetry, which is broken spontaneously and results in a Goldstone mode. The later dynamics will break the shift-symmetry of the Goldstone by selecting a preferred phase. A mass term will then be generated to the field, which populates the DM energy density after $H\lesssim m_{\phi}$. The initial condition of a misaligned scalar in our scenario reads Marsh:2015xka $\displaystyle\phi(t_{\rm i},x)=f_{\phi}\theta_{\phi}\;,~{}\dot{\phi}(t_{\rm i},x)=0\;,$ (5) where $t_{\rm i}$ is the time set by the spontaneous symmetry breaking of the global symmetry, $f_{\phi}$ represents the energy scale of the symmetry breaking, $\theta_{\phi}$ is the initial phase, and $\dot{\phi}\equiv\mathrm{d}\phi/\mathrm{d}t$. At the quantum field level, the ultralight DM produced by the misalignment mechanism (from the dynamical symmetry breaking) should be descried by a coherent state of the form Marsh:2015xka ; Ferreira:2020fam ; Miransky1994 ; Ferreira:lecture ; Davidson:2014hfa ; Veltmaat2020GalaxyFW ; Guth:2014hsa ; Zhang:1999is ; Barnett:1997 $\displaystyle\left|\phi\right\rangle=\mathrm{e}^{-\frac{N}{2}}\mathrm{e}^{\int\frac{\mathrm{d}^{3}k}{(2\pi)^{3}}\tilde{\phi}(k)\hat{a}^{\dagger}_{k}}\left|0\right\rangle\;,$ (6) where $\mathrm{e}^{-\frac{N}{2}}$ is a normalization factor, $\tilde{\phi}(k)$ is a general complex field with the wave-vector $k$, and $\hat{a}^{\dagger}_{k}$ is the corresponding particle creation operator. The coherent state $|\phi\rangle$ has been thought of as the most field-like solution of quantum states. The property of this state can be largely captured by the expectation value of the field operator $\hat{\phi}$, namely $\displaystyle\phi(t,x)\equiv\left\langle\phi\right|\hat{\phi}\left|\phi\right\rangle$ $\displaystyle=$ $\displaystyle\int\frac{\mathrm{d}^{3}k}{(2\pi)^{3}}\frac{1}{\sqrt{2E_{k}}}\left[\tilde{\phi}(k)\mathrm{e}^{-i(\omega t-k\cdot x)}+\right.$ (7) $\displaystyle\left.+\tilde{\phi}^{*}(k)\mathrm{e}^{i(\omega t-k\cdot x)}\right]\;,$ which is a classical-number field. This non-vanishing expectation value can be interpreted in a way similar to the Higgs mechanism. The evolution of the classical field $\phi(t,x)$ right follows the equation of motion in Eq. (3). For the isotropic DM field without kinematic energy, i.e., $k=0$, we have $\left|\phi\right\rangle=\mathrm{e}^{-\frac{N}{2}}\mathrm{e}^{\sqrt{N}\mathrm{e}^{\mathrm{i}\theta}\hat{a}^{\dagger}_{0}}\left|0\right\rangle$. It should be emphasized that the coherent state $|\phi\rangle$ is an eigenstate of $\hat{a}_{0}$, i.e., $\hat{a}_{0}|\phi\rangle=\sqrt{N}\mathrm{e}^{\mathrm{i}\theta}|\phi\rangle$, but not that of the particle number $\hat{N}=\hat{a}^{\dagger}_{0}\hat{a}_{0}$ in the Fock space. However, for a large enough ensemble of DM, there is a well-defined mean particle number $\langle\hat{a}^{\dagger}_{0}\hat{a}_{0}\rangle=N$ with a quantum fluctuation proportional to $1/\sqrt{N}$. In fact, for the state with a definitive particle number, e.g., $\hat{a}^{\dagger}_{0}\hat{a}_{0}|N\rangle=N|N\rangle$, the field expectation value $\left\langle N\right|\hat{\phi}\left|N\right\rangle$ is vanishing. The difference between $|\phi\rangle$ and $|{N}\rangle$ is critical and should be highlighted, as they will induce very distinct signals experimentally. For instance, for $|{N}\rangle$ the absolute quantum phase $\mathrm{exp}(-\mathrm{i}\,m_{\phi}t)$ is meaningless in practice, but the classical phase of $|\phi\rangle$ is physical and will lead to time-varying signals DirectMeasurementofLightWaves ; Barnett1996 . In the early Universe, the problem is ascribed to solving the evolution of classical field. Because the ultralight DM should be homogeneously produced, the evolution equation of Eq. (3) can be recast to Esteban:2021ozz $\displaystyle\ddot{\phi}+3H\dot{\phi}+m^{2}_{\phi}\phi=-g\int\frac{\mathrm{d}^{3}\bm{p}}{(2\pi)^{3}}\frac{g\phi}{\sqrt{\bm{p}^{2}+(g\phi)^{2}}}f_{\nu}(a|\bm{p}|),\hskip 14.22636pt$ (8) where $a$ is the scale factor, connected to the redshift $z$ via $a\equiv a_{0}/(z+1)$ with $a_{0}$ being the scale factor at present. Here, the neutrino distribution function simply follows $f_{\nu}(a|\bm{p}|)=D_{\nu}/(\mathrm{e}^{|\bm{p}|/T_{\nu}(a)}+1)$ with the temperature $T_{\nu}(a)\propto 1/a$ and the factor $D_{\nu}=6$ accommodating three generations of neutrinos and antineutrinos. The source term on the right-hand side of Eq. (8) modifies the evolution of a free scalar. Note that we have assumed the vacuum neutrino mass to be vanishing in the above equation, and a larger effect of neutrino source will be expected if the vacuum neutrino mass is taken into account. ### II.1 Observable impacts If $T_{\nu}\gg g\phi$ is satisfied (i.e., neutrinos in the plasma are relativistic), the source term is equivalent to a screening mass of the scalar, with Davoudiasl:2018hjw ; Esteban:2021ozz $\displaystyle m^{2}_{\phi,\rm s}(a)=\int\frac{\mathrm{d}^{3}\bm{p}}{(2\pi)^{3}}\frac{g^{2}}{|\bm{p}|}f_{\nu}(a|\bm{p}|)\sim g^{2}T^{2}_{\nu}.\hskip 14.22636pt$ (9) The screening mass describes how the neutrino plasma should response to the change of the scalar potential, and its main effect will be altering the oscillating frequency of scalar field. In this case, we have $\displaystyle\ddot{\phi}+3H\dot{\phi}+\left[m^{2}_{\phi}+m^{2}_{\phi,\rm s}(a)\right]\phi=0\;,\hskip 14.22636pt$ (10) with an effective mass varying as the Universe expands, $m_{\phi,{\rm eff}}=[m^{2}_{\phi}+m^{2}_{\phi,\rm s}(a)]^{1/2}$. For the numerical convenience, it is useful to replace the physical time $t$ with the scale factor $a=m\,x$, where $m$ is an arbitrary scale and $x$ is dimensionless. The equation of motion can be recast into $\displaystyle x^{2}\phi^{\prime\prime}+\left(4x+x^{2}\frac{H^{\prime}}{H}\right)\phi^{\prime}+\frac{m^{2}_{\phi,{\rm eff}}}{H^{2}}\phi=0\;,$ (11) where ${\phi}^{\prime}\equiv\mathrm{d}\phi/\mathrm{d}x$. The scalar field $\phi$ starts to oscillate when $H\lesssim m_{\phi,{\rm eff}}$. If $\phi$ is not coupled to neutrinos, its amplitude will simply scale as $\phi_{0}\propto a^{-3/2}$ Marsh:2015xka , and the energy density $\rho_{\phi}\approx m^{2}_{\phi}\phi^{2}_{0}/2$ will be proportional to $a^{-3}$, which justifies its capability of being the cold DM. However, if the scalar mass is significantly screened, this benefit will be lost due to the behavior $m_{\phi,{\rm s}}\propto a^{-1}$. In the completely screened case, i.e. $m_{\phi,{\rm s}}\gg m_{\phi}$, we find 222In the fast-oscillating regime, to obtain the evolution of the amplitude we decompose the scalar field as $\phi(t)=\phi_{0}(t)\cos{(m_{\phi,{\rm eff}}\cdot t)}$ Marsh:2015xka . Keeping the leading order in the assumption of $H\sim\dot{\phi}_{0}\ll m_{\phi,{\rm eff}}$, we obtain $2\dot{\phi_{0}}=-3H\phi_{0}-\phi_{0}\dot{m}_{\phi,{\rm eff}}/m_{\phi,{\rm eff}}$. For $m_{\phi,{\rm eff}}\propto a^{-1}$, we have $\dot{\phi_{0}}=-H\phi_{0}$, and hence $\phi_{0}\propto a^{-1}$. $\displaystyle\phi_{0}\propto a^{-1}\;,~{}\rho_{\phi}\sim\frac{m^{2}_{\phi,{\rm s}}\phi^{2}_{0}}{2}\propto a^{-4}\;,$ (12) which indicates that $\phi$ evolves more like a radiation instead of cold DM with the screened mass. The screening effect is thus to be constrained by the cosmological observations if we require the scalar to be DM. An example of $\phi^{2}$ ($\geq 0$) evolution with respect to the scale factor $a$ is given in Fig. 2 by numerically solving Eq. (11). The lighter fast- oscillating curves stand for the magnitude of $\phi^{2}$ with (red) or without (blue) the screening effect, and the darker ones are produced by averaging over fast-oscillating modes to see the overall amplitude. As the screening mass $m_{\phi,{\rm s}}$ is dominant for $a<500$, the scalar field scales as $a^{-1}$ instead of $a^{-3/2}$ and contributes to the radiation component. After $a=500$ the vacuum mass starts to dominate over the screening one, and the scalar field plays the role of DM as normal. Figure 2: A demonstration of evolution of ultralight DM as a function of the scale factor. For demonstration we give $\phi^{2}$ in the logarithmic scale. The case with the screening effect of neutrinos is shown as the red curve, while that without the screening effect is given as the blue one. The lighter curves are the actual solutions, and the darker ones are averaged over certain scale factor to see the magnitude of the field. Another effect to avoid is the impact on neutrino velocity $v_{\nu}$, which is related to the neutrino free-streaming length before the decoupling of CMB. As has been mentioned, in the background of ultralight DM, neutrinos acquire an effective mass $m_{\nu,{\rm eff}}=g\phi$. The neutrino velocity in the early Universe will be given by $\displaystyle\frac{v_{\nu}}{c}=\frac{|\bm{p}|}{\sqrt{\bm{p}^{2}+m^{2}_{\nu,{\rm eff}}}}\;,$ (13) where the average momentum of neutrino plasma reads $\langle|\bm{p}|\rangle\sim 3T_{\nu}$. Notice that the effective neutrino mass $m_{\nu,{\rm eff}}\propto a^{-3/2}$ increases faster than $T_{\nu}\propto a^{-1}$ as we go to higher redshift $z$. This means that neutrinos will become more and more non-relativistic and hence less and less free-streaming at higher $z$. In comparison, the CMB data are consistent with neutrino free- streaming with a speed of light up to a redshift as high as $z=10^{5}$ Hannestad:2004qu ; Hannestad:2005ex ; Bell:2005dr ; Basboll:2008fx ; Archidiacono:2013dua ; Forastieri:2015paa ; Cyr-Racine:2013jua ; Oldengott:2014qra ; Forastieri:2017oma ; Oldengott:2017fhy ; Choudhury:2020tka ; Brinckmann:2020bcn ; Das:2020xke ; Mazumdar:2020ibx . For simplicity, in the following we will consider neutrinos as free-streaming with the speed of light if $T_{\nu}>g\phi$, and vise versa. It is worth noting that the CMB constraint on the sum of neutrino masses is not as competitive as the above two effects. The combined analysis of $Planck~{}{\rm TT,TE,EE+lowE+lensing+BAO}$ data leads to $\Sigma m_{\nu}<0.12~{}{\rm eV}$ Planck:2018vyg , which actually takes account of data at very low redshift long after the recombination $z_{\rm rc}=1100$. In fact, around the recombination era neutrinos with a mass $\mathcal{O}(0.1)~{}{\rm eV}$ are still relativistic. Since $m_{\nu,{\rm eff}}\propto a^{-3/2}$, the impact of effective neutrino masses turns out to be less and less important at lower $z$. Hence, we do not attempt to set the bound according to $\Sigma m_{\nu,\rm eff}(z)<0.12~{}{\rm eV}$ at recombination $z_{\rm rc}=1100$, which has been considered previously. Figure 3: The evolution of average neutrino velocity in terms of the redshift in the early Universe. The colorful curves stand for the cases with a coupling between neutrinos and the ultralight DM, with the vacuum neutrino mass vanishing and DM-induced neutrino mass $g\phi$ at present indicated close to the curve. The black curve represents the neutrino velocity of the standard case with a mass $m_{\nu}=0.04~{}{\rm eV}$. ### II.2 Updated CMB constraints Depending on the parameter choice of $g$ and $m_{\phi}$, there can be four different scenarios in the early Universe for the evolution equation of Eq. (8): * • Case I: $g\phi<T_{\nu}$ and $g^{2}T^{2}_{\nu}<m^{2}_{\phi}$. The effective neutrino mass is negligible compared to the neutrino temperature and the screening effect is also not important. In this case, the scalar field evolves like matter with $\phi_{0}\propto a^{-3/2}$ while $T_{\nu}\propto a^{-1}$, and these two conditions will maintain themselves as the Universe expands. * • Case II: $g\phi<T_{\nu}$ and $g^{2}T^{2}_{\nu}>m^{2}_{\phi}$. Neutrinos are relativistic but the screening scalar mass $m_{\phi,\rm s}\sim gT_{\nu}$ dominates the oscillation of $\phi$. In such scenario, the scalar field evolves like radiation with $\phi_{0}\propto a^{-1}$. This state is not stable and will collapse into Case I eventually. * • Case III: $g\phi>T_{\nu}$ and $gT^{3}_{\nu}<m^{2}_{\phi}\phi$. The screening effect is not important compared to the vacuum mass of $\phi$, but neutrinos are non-relativistic (thus not free-streaming with $c$) due to the large effective mass. This case will also end up into Case I as the Universe expands. * • Case IV: $g\phi>T_{\nu}$ and $gT^{3}_{\nu}>m^{2}_{\phi}\phi$. Both the screening effect and the effective neutrino mass are dominant factors, which is to be avoided over the observable history of CMB. For Cases I and III, in terms of redshift, the neutrino temperature and the scalar field strength read as $T_{\nu}\approx 1.68\times 10^{-4}~{}{\rm eV}(z+1)$ and $\phi\approx 2.15\times 10^{-3}~{}{\rm eV}^{2}/m_{\phi}\left[(z+1)/45\right]^{3/2}$, respectively, assuming the ultralight scalar accommodates all the DM abundance. The evolution of the Universe can be in a sequence of Case III$\to$I or II$\to$I. In Fig. 3, we illustrate the neutrino velocity evolution $\langle v_{\nu}\rangle/c$, i.e., Eq. (13), in terms of the redshift $z$ from Case III to I, where the screening effect is not important but the neutrino effective mass is severely affected. This can be achieved by properly choosing the DM parameter. It can be noticed that the neutrino velocity drops rapidly as we go back to higher $z$, which should be avoided to be in accordance with CMB free- streaming requirement. In comparison, we also give the neutrino velocity in the standard case with a constant mass $m_{\nu}=0.04~{}{\rm eV}$ corresponding to the Planck limit $\Sigma m_{\nu}<0.12~{}{\rm eV}$. In this case, the neutrino velocity evolves in an opposite manner compared to that with a coupling to the ultralight DM. As has been mentioned, around the recombination era $z_{\rm rc}\approx 1100$, neutrinos of the Planck limit are still ultra- relativistic, and it is the data at lower $z$ that contribute to this limit. In order not to significantly alter the CMB observations, two criteria should be satisfied: (1) the ultralight scalar must evolve as matter before the matter-radiation equality around $z_{\rm eq}=3000$; (2) neutrinos should be free-streaming up to the redshift $z_{\rm eq}=3000$ (conservative) or $z_{\rm fs}=10^{5}$ (aggressive). The aggressive redshift for neutrino free-streaming is inspired by the investigation of neutrino self-interactions Hannestad:2004qu ; Hannestad:2005ex ; Bell:2005dr ; Basboll:2008fx ; Archidiacono:2013dua ; Forastieri:2015paa ; Cyr-Racine:2013jua ; Oldengott:2014qra ; Forastieri:2017oma ; Oldengott:2017fhy ; Choudhury:2020tka ; Brinckmann:2020bcn ; Das:2020xke ; Mazumdar:2020ibx , which similarly shrinks the neutrino free-streaming length. One may argue that the scalar becomes a dynamical field (acquiring the shift-symmetry breaking terms) only after a certain redshift, e.g., $z=10^{4}$, such that the aggressive free- streaming limit does not apply in general. But there is no way to avoid the robust limit set at $z_{\rm eq}=3000$, because the CMB data require the dominance of matter component after that redshift. If these two criteria are satisfied, during the observable history of CMB the ultralight scalar will play the role of cold DM and neutrinos will stream as free particles. We can evolve the current Universe back to higher redshift for the CMB constraints. To be more specific, we require $g\phi<T_{\nu}$ and $g^{2}T^{2}_{\nu}<m^{2}_{\phi}$ before $z=3000$ for the conservative limit, and $g\phi<T_{\nu}$ before $z=10^{5}$ for the aggressive one. We obtain our CMB limits as $\displaystyle g$ $\displaystyle<$ $\displaystyle 0.4\cdot\frac{m_{\phi}}{\rm eV}\hskip 14.22636pt({\rm conservative}),$ (14) $\displaystyle g$ $\displaystyle<$ $\displaystyle 0.09\cdot\frac{m_{\phi}}{\rm eV}\hskip 9.67383pt({\rm aggressive}),$ (15) which correspond to the limits on the effective neutrino mass $g\phi$ in our local galaxy $\displaystyle m_{\nu,{\rm eff}}$ $\displaystyle<$ $\displaystyle 9\times 10^{-4}~{}{\rm eV}\hskip 11.38092pt({\rm conservative}),$ (16) $\displaystyle m_{\nu,{\rm eff}}$ $\displaystyle<$ $\displaystyle 2\times 10^{-4}~{}{\rm eV}\hskip 11.38092pt({\rm aggressive}).$ (17) These limits are dominated by the neutrino free-streaming requirement. We find that, for the aggressive limit, as we go to higher redshift the state will transit from Case I to II around $z=6\times 10^{4}$, after which the free- streaming condition $g\phi<T_{\nu}$ becomes stabilized as $\phi_{0}\propto a^{-1}$ for Case II. For $g$ larger than the aggressive limit Case I will transit to III instead before $z_{\rm fs}=10^{5}$, and the free-streaming requirement is spoiled. We need to point out that a thorough analysis of the CMB spectrum by simulating the evolution of perturbations in this context will make the results more robust, and we leave it as a possible future work. ## III Implication for $\boldsymbol{0\nu\beta\beta}$ decays There are many dedicated experiments that are looking for the signatures of $0\nu\beta\beta$ decays, e.g., GERDA Phase-II Agostini:2018tnm , CUORE Alduino:2017ehq , SuperNEMO Barabash:2011aa , KamLAND-Zen KamLAND-Zen:2016pfg and EXO Agostini:2017jim . The discovery of lepton-number-violating processes in the future can reveal the Majorana nature of massive neutrinos Dolinski:2019nrj . As has been shown in Fig. 1, even without the vacuum Majorana neutrino mass the $0\nu\beta\beta$ transition can still take place in the background of ultralight DM. The induced transition is very similar to the Majoron emission process in vacuum. However, because the vacuum in our case is occupied by dense coherent scalars, the dominant process will be the “stimulated” emission and absorption of the scalar particle, in analogy with the atomic transition in a laser field. Because the occupation number of scalars $N=n_{\phi}\lambda^{3}$ is very large, such a transition will leave the DM field almost unaffected. This is equivalent to simply taking the expectation value for all relevant operators in the Lagrangian or scattering matrix. In the presence of ultralight DM, the effective neutrino mass ($m_{ee}$) will hence receive an additional time- varying contribution $\displaystyle{m}^{\prime}_{ee}(t)=\langle\phi|m_{ee}+g\hat{\phi}(t,x)|\phi\rangle=m_{ee}+g{\phi}(t,x)\;,$ (18) The $0\nu\beta\beta$-decay process typically takes place with a very small uncertainty of time. It is therefore a good approximation to take the factor $\cos{m_{\phi}t}$ as a constant during the transition. But the long-term events will feature a time-varying pattern if enough data have been collected. The above result should be able to be reproduced from the more fundamental quantum field theory. However, it is not apparent to obtain the time-varying effect from a view of the Feynman diagrams as illustrated in Fig. 1, where the emission and absorption processes are themselves time-independent. To understand this, we note that the mass of the ultralight DM ($<{\rm eV}$), namely the change in the energy of final electrons, is much smaller than the energy resolution of $0\nu\beta\beta$-decay experiments ($\sim{\rm keV}$). Hence, it is not possible at the quantum level to distinguish between the emission and absorption diagrams, which will lead to an interference. One can find that it is this interference that gives rise to the time variation. To be more precise, after the $0\nu\beta\beta$ transition via the vertex $(m_{ee}+g\hat{\phi})\overline{\nu^{\rm c}_{\rm L}}\nu_{\rm L}/2$ the scalar state will be taken to $\displaystyle\left[m_{ee}+g\hat{\phi}(t,x)\right]\left|\phi\right\rangle$ $\displaystyle=\left[m_{ee}+g\sqrt{\frac{2}{m_{\phi}V}}\left(\hat{a}\mathrm{e}^{-\mathrm{i}m_{\phi}t}+\hat{a}^{\dagger}\mathrm{e}^{\mathrm{i}m_{\phi}t}\right)\right]|\phi\rangle\;.$ (19) Note that $|\phi\rangle$, $\hat{a}|\phi\rangle$ and $\hat{a}^{\dagger}|\phi\rangle$ almost completely overlap with each other, e.g., $\langle\phi|\hat{a}\,\hat{a}|\phi\rangle=N\mathrm{e}^{2\mathrm{i}\theta}$. Together with other terms, this feature results in a probability proportional to $\Sigma_{N}|\left\langle N\right|m_{ee}+g\hat{\phi}(t,x)|\phi\rangle|^{2}$, namely $\left|m_{ee}+g{\phi}(t,x)\right|^{2}$. The above observations are also applicable to the impact on neutrino oscillation experiments. As for our $0\nu\beta\beta$ process, the half-life of a given isotope can be expressed as Rodejohann:2011mu $\dfrac{1}{T^{0\nu}_{1/2}}=G_{0\nu}\cdot\left|{M}_{0\nu}\right|^{2}\cdot\frac{\left|m^{\prime}_{ee}(t)\right|^{2}}{M^{2}_{e}}\;,$ (20) where $G_{0\nu}$ represents the phase-space factor, $M_{0\nu}$ is the nuclear matrix element (NME), and $M_{e}\approx 0.51~{}{\rm MeV}$ is the electron mass. The present lower limit on the half-life of $0\nu\beta\beta$ decays from KamLAND-Zen collaboration KamLAND-Zen:2016pfg is $T^{0\nu}_{1/2}>1.07\times 10^{26}$ yr at 90% C. L. Using the phase-space factor $G_{0\nu}=3.793\times 10^{-14}~{}{\rm yr^{-1}}$ Kotila:2012zza with $g=1.27$ and NME $M_{0\nu}\in(1.5-4.2)$ Guzowski:2015saa for ${}^{136}{\rm Xe}$, the upper bound on the effective neutrino mass in standard three-flavor neutrino scenario is obtained as $|m_{ee}|<(61-169)$ meV. Here, $m_{ee}$ is given by $\displaystyle{m}_{ee}$ $\displaystyle\equiv$ $\displaystyle m_{1}\cos^{2}\theta_{13}\cos^{2}\theta_{12}e^{{\rm i}\rho}+m_{2}\cos^{2}\theta_{13}\sin^{2}\theta_{12}$ (21) $\displaystyle+m_{3}\sin^{2}\theta_{13}e^{{\rm i}\sigma}\;,$ where $m_{i}$ (for $i=1,2,3$) stand for the absolute masses of three neutrinos, $\\{\theta_{12},\theta_{13}\\}$ are the leptonic mixing angles, and $\\{\rho,\sigma\\}$ are the Majorana CP phases. Figure 4: Exclusion and sensitivities of various experiments on the parameter space of the ultralight DM mass $m_{\phi}$ and the coupling to neutrinos $g$. The exclusion using the result of KamLAND-Zen are given by cyan lines, while the CMB bounds in present work are shown as the dashed brown and solid black lines. See the text for details of other limits. To find out how the ultralight scalar exactly modifies the rate, we expand the vacuum and DM contributions $\displaystyle|{m}^{\prime}_{ee}|^{2}$ $\displaystyle=$ $\displaystyle\left|m_{ee}\right|^{2}+g^{2}\phi^{2}_{0}\cos^{2}{m_{\phi}t}$ (22) $\displaystyle+2|m_{ee}|g\phi_{0}\cos\eta\cos{m_{\phi}t}\;,$ where $\eta$ is the relative phase between $m_{ee}$ and $g\phi$, the second term is induced solely by the DM field, and the third term is the interference with the vacuum neutrino mass. Such time-varying effect is in principle observable, if enough $0\nu\beta\beta$-decay events with proper time resolution have been obtained. When the ultralight DM oscillates fast enough, the long-term decay rate depends on the time average of $|{m}^{\prime}_{ee}|^{2}$, i.e., $\displaystyle\overline{|{m}^{\prime}_{ee}|^{2}}=\left|m_{ee}\right|^{2}+\frac{1}{2}g^{2}\phi^{2}_{0}\;.$ (23) Note that the interference term has been averaged out in such case. The coupling constant $g$ is in general not related to $m_{ee}$. Hence, when the cancellation of effective neutrino mass occurs, i.e., $m_{ee}=0$ for normal ordering, the DM-induced term can dominate the transition. We present our noteworthy results as Fig. 4 in the $m_{\phi}$-$g$ parameter plane. The exclusion regions of $0\nu\beta\beta$-decay experiments and cosmology are shown along with bounds and sensitivities arising from different experimental searches. The vertical blue band shows the constraint based on the results of Lyman-$\alpha$ forest. It has been pointed out (see for detailed analysis in Ref. Hui:2016ltb ) that fuzzy DM lighter than $\sim 10^{-21}$ eV is in tension with observations of the Lyman-$\alpha$ forest. The gray shaded region gives the limit from Super-K and Sudbury Neutrino Observatory (SNO) neutrino oscillation experiments Berlin:2016woy ; Super- Kamiokande:2003snd ; SNO:2005ftm . The sensitivity from forthcoming neutrino oscillation experiments Krnjaic:2017zlz ; Dev:2020kgz like DUNE DUNE:2015lol and JUNO JUNO:2015zny have also been presented using the magenta shaded region and orange dot-dashed line, respectively. The constraint from $0\nu\beta\beta$-decay searches (KamLAND-Zen KamLAND- Zen:2016pfg ) is shown as the solid cyan line. The projection with an ultimate sensitivity goal of $0\nu\beta\beta$-decay experiments $|m^{\prime}_{ee}|=1~{}{\rm meV}$ Ge:2016tfx ; Penedo:2018kpc ; Cao:2019hli ; Huang:2020mkz is shown as the dotted cyan line. In order to translate results into the $m_{\phi}$-$g$ plane, we have used Eq. (23). We present our CMB bounds as the dashed brown and solid black lines, respectively. Two different limits corresponding to conservative and aggressive considerations are shown, as given by Eqs. (14) and (15), respectively. It can be noticed that JUNO can cover the limit of $0\nu\beta\beta$-decay experiment for most of the parameter spcae. However, for the region where the ultralight DM mass is greater than $10^{-11}~{}{\rm eV}$, $0\nu\beta\beta$-decay experiments set tighter bound than neutrino oscillation experiments. On the other hand, Super-K and SNO put the best constraint for ultralight DM below $10^{-17}~{}{\rm eV}$, whereas the CMB analysis provides the most stringent limit for ultralight DM above $10^{-17}$ eV. Figure 5: The effective Majorana neutrino mass as a function of the lightest active neutrino mass $m_{1}$. The orange region stands for the standard case with normal mass ordering for neutrinos, while the blue region is for the case with $g\phi_{0}=0.01~{}{\rm eV}$. Finally, as a visual guide for the ultralight DM searches at $0\nu\beta\beta$-decay experiments, we present the effective Majorana neutrino mass, in presence of ultralight DM, in Fig. 5 Graf:2020cbf . The horizontal green band indicates the current experimental limit from KamLAND-Zen, $(61-165)~{}{\rm meV}$ KamLAND-Zen:2016pfg . The dashed lines correspond to the future sensitivities projected by SNO+ with $(19-46)~{}{\rm meV}$ Andringa:2015tza , LEGEND with $(10.7-22.8)~{}{\rm meV}$ Abgrall:2017syy , and nEXO with $(5.7-17.7)~{}{\rm meV}$ Albert:2017hjq , respectively. The vertical gray band represents the current neutrino mass limit of cosmological data from the Planck collaboration Aghanim:2018eyx . The orange region depicts the standard three-flavor neutrino scenario. The impact of ultralight DM has been demonstrated using the blue region for a benchmark choice of $g\phi_{0}=0.01~{}{\rm eV}$. It turns out that even the averaged effect of ultralight DM can be tested by the next generation of $0\nu\beta\beta$-decay experiments. ## IV Conclusion In this work, we investigate the impact of ultralight DM on $0\nu\beta\beta$-decay experiments. We find that the scalar DM can directly induce the $0\nu\beta\beta$-decay transition, even without the vacuum neutrino mass term. The effect can be significant for very small couplings between the ultralight DM and neutrinos, due to the stimulated absorption and emission of scalars in the phase space filled with identical particles. As the consequence, the vacuum Majorana neutrino mass will be modified with a time- varying term. We also explore the consequence of such a scalar field on cosmology. There are mainly two effects. First, the neutrino plasma will induce a screening mass to the ultralight DM, such that the DM field evolves as radiation instead of matter component. Second, the dense ultralight DM in the early Universe will shrink the neutrino free-streaming length significantly and is in tension with the CMB observations. The constraints and sensitivities of different experiments are given in Fig. 4. We conclude that $0\nu\beta\beta$-decay experiments can give stringent constraint compared to the forthcoming neutrino oscillation experiments like DUNE and JUNO for the DM mass above $\thicksim 10^{-11}$ eV. In addition, the CMB observations have the best constraining capability for the mass range above $\thicksim 10^{-17}$ eV. Furthermore, we observe from Fig. 5 that our framework of ultralight DM can be tested even without analyzing the time- varying effect by the next generation of $0\nu\beta\beta$-decay experiments. ###### Acknowledgements. Authors are thankful to Werner Rodejohann for his valuable comments and careful reading of the manuscript. Authors would also like to thank Eligio Lisi, Pablo Martínez-Miravé, Manibrata Sen, and Shun Zhou for fruitful discussions. GYH is supported by the Alexander von Humboldt Foundation. NN is supported by the Istituto Nazionale di Fisica Nucleare (INFN) through the “Theoretical Astroparticle Physics” (TAsP) project. ## References * (1) S. Bird, I. Cholis, J. B. Muñoz, Y. Ali-Haïmoud, M. Kamionkowski, E. D. Kovetz, A. Raccanelli, and A. G. Riess, “Did LIGO detect dark matter?,” Phys. Rev. Lett. 116 (2016) no. 20, 201301, arXiv:1603.00464. * (2) B. Carr, F. Kuhnel, and M. Sandstad, “Primordial Black Holes as Dark Matter,” Phys. Rev. D 94 (2016) no. 8, 083504, arXiv:1607.06077. * (3) L. Chen, Q.-G. Huang, and K. Wang, “Constraint on the abundance of primordial black holes in dark matter from Planck data,” JCAP 12 (2016) 044, arXiv:1608.02174. * (4) J. Georg and S. Watson, “A Preferred Mass Range for Primordial Black Hole Formation and Black Holes as Dark Matter Revisited,” JHEP 09 (2017) 138, arXiv:1703.04825. * (5) W. Hu, R. Barkana, and A. Gruzinov, “Cold and fuzzy dark matter,” Phys. Rev. Lett. 85 (2000) 1158–1161, arXiv:astro-ph/0003365. * (6) XENON, E. Aprile et al., “Search for Coherent Elastic Scattering of Solar 8B Neutrinos in the XENON1T Dark Matter Experiment,” Phys. Rev. Lett. 126 (2021) 091301, arXiv:2012.02846. * (7) PandaX-II, X. Cui et al., “Dark Matter Results From 54-Ton-Day Exposure of PandaX-II Experiment,” Phys. Rev. Lett. 119 (2017) no. 18, 181302, arXiv:1708.06917. * (8) E. G. M. Ferreira, “Ultra-light dark matter,” Astron. Astrophys. Rev. 29 (2021) no. 1, 7, arXiv:2005.03254. * (9) L. A. Ureña López, “Brief Review on Scalar Field Dark Matter Models,” Front. Astron. Space Sci. 6 (2019) 47. * (10) L. Hui, J. P. Ostriker, S. Tremaine, and E. Witten, “Ultralight scalars as cosmological dark matter,” Phys. Rev. D 95 (2017) no. 4, 043541, arXiv:1610.08297. * (11) J. E. Kim and G. Carosi, “Axions and the Strong CP Problem,” Rev. Mod. Phys. 82 (2010) 557–602, arXiv:0807.3125. [Erratum: Rev.Mod.Phys. 91, 049902 (2019)]. * (12) P. W. Graham, I. G. Irastorza, S. K. Lamoreaux, A. Lindner, and K. A. van Bibber, “Experimental Searches for the Axion and Axion-Like Particles,” Ann. Rev. Nucl. Part. Sci. 65 (2015) 485–514, arXiv:1602.00039. * (13) I. G. Irastorza and J. Redondo, “New experimental approaches in the search for axion-like particles,” Prog. Part. Nucl. Phys. 102 (2018) 89–159, arXiv:1801.08127. * (14) D. J. E. Marsh, “Axion Cosmology,” Phys. Rept. 643 (2016) 1–79, arXiv:1510.07633. * (15) S. Dell’Oro, S. Marcocci, M. Viel, and F. Vissani, “Neutrinoless double beta decay: 2015 review,” Adv. High Energy Phys. 2016 (2016) 2162659, arXiv:1601.07512. * (16) J. D. Vergados, H. Ejiri, and F. Šimkovic, “Neutrinoless double beta decay and neutrino mass,” Int. J. Mod. Phys. E 25 (2016) no. 11, 1630007, arXiv:1612.02924. * (17) A. Berlin, “Neutrino Oscillations as a Probe of Light Scalar Dark Matter,” Phys. Rev. Lett. 117 (2016) no. 23, 231801, arXiv:1608.01307. * (18) V. Brdar, J. Kopp, J. Liu, P. Prass, and X.-P. Wang, “Fuzzy dark matter and nonstandard neutrino interactions,” Phys. Rev. D 97 (2018) no. 4, 043001, arXiv:1705.09455. * (19) G. Krnjaic, P. A. N. Machado, and L. Necib, “Distorted neutrino oscillations from time varying cosmic fields,” Phys. Rev. D 97 (2018) no. 7, 075017, arXiv:1705.06740. * (20) J. Liao, D. Marfatia, and K. Whisnant, “Light scalar dark matter at neutrino oscillation experiments,” JHEP 04 (2018) 136, arXiv:1803.01773. * (21) F. Capozzi, I. M. Shoemaker, and L. Vecchi, “Neutrino Oscillations in Dark Backgrounds,” JCAP 07 (2018) 004, arXiv:1804.05117. * (22) M. M. Reynoso and O. A. Sampayo, “Propagation of high-energy neutrinos in a background of ultralight scalar dark matter,” Astropart. Phys. 82 (2016) 10–20, arXiv:1605.09671. * (23) G.-Y. Huang and N. Nath, “Neutrinophilic Axion-Like Dark Matter,” Eur. Phys. J. C 78 (2018) no. 11, 922, arXiv:1809.01111. * (24) S. Pandey, S. Karmakar, and S. Rakshit, “Interactions of Astrophysical Neutrinos with Dark Matter: A model building perspective,” JHEP 01 (2019) 095, arXiv:1810.04203. * (25) Y. Farzan and S. Palomares-Ruiz, “Flavor of cosmic neutrinos preserved by ultralight dark matter,” Phys. Rev. D 99 (2019) no. 5, 051702, arXiv:1810.00892. * (26) K.-Y. Choi, J. Kim, and C. Rott, “Constraining dark matter-neutrino interactions with IceCube-170922A,” Phys. Rev. D 99 (2019) no. 8, 083018, arXiv:1903.03302. * (27) S. Baek, “Dirac neutrino from the breaking of Peccei-Quinn symmetry,” Phys. Lett. B 805 (2020) 135415, arXiv:1911.04210. * (28) K.-Y. Choi, E. J. Chun, and J. Kim, “Neutrino Oscillations in Dark Matter,” Phys. Dark Univ. 30 (2020) 100606, arXiv:1909.10478. * (29) K.-Y. Choi, E. J. Chun, and J. Kim, “Dispersion of neutrinos in a medium,” arXiv:2012.09474. * (30) A. Dev, P. A. N. Machado, and P. Martínez-Miravé, “Signatures of ultralight dark matter in neutrino oscillation experiments,” JHEP 01 (2021) 094, arXiv:2007.03590. * (31) S. Baek, “A connection between flavour anomaly, neutrino mass, and axion,” JHEP 10 (2020) 111, arXiv:2006.02050. * (32) M. Losada, Y. Nir, G. Perez, and Y. Shpilman, “Probing scalar dark matter oscillations with neutrino oscillations,” arXiv:2107.10865. * (33) A. Y. Smirnov and V. B. Valera, “Resonance refraction and neutrino oscillations,” arXiv:2106.13829. * (34) G. Alonso-Álvarez and J. M. Cline, “Sterile neutrino dark matter catalyzed by a very light dark photon,” arXiv:2107.07524. * (35) G.-y. Huang and W. Rodejohann, “Tritium beta decay with modified neutrino dispersion relations: KATRIN in the dark sea,” arXiv:2110.03718. * (36) K. Blum, Y. Nir, and M. Shavit, “Neutrinoless double-beta decay with massive scalar emission,” Phys. Lett. B 785 (2018) 354–361, arXiv:1802.08019. * (37) T. Brune and H. Päs, “Massive Majorons and constraints on the Majoron-neutrino coupling,” Phys. Rev. D 99 (2019) no. 9, 096005, arXiv:1808.08158. * (38) R. Cepedello, F. F. Deppisch, L. González, C. Hati, and M. Hirsch, “Neutrinoless Double-$\beta$ Decay with Nonstandard Majoron Emission,” Phys. Rev. Lett. 122 (2019) no. 18, 181801, arXiv:1811.00031. * (39) V. A. Miransky, Dynamical Symmetry Breaking In Quantum Field Theories. World Scientific, 1994. * (40) “Lecture on ultralight dark matter,”. https://wwwmpa.mpa-garching.mpg.de/~komatsu/lecturenotes/Elisa_Ferreira_on_SuperfluidDarkMatter.pdf. * (41) S. Davidson, “Axions: Bose Einstein Condensate or Classical Field?,” Astropart. Phys. 65 (2015) 101–107, arXiv:1405.1139. * (42) J. Veltmaat, “Galaxy Formation With Ultralight Bosonic Dark Matter,” 2020\. https://d-nb.info/121070269X/34. * (43) A. H. Guth, M. P. Hertzberg, and C. Prescod-Weinstein, “Do Dark Matter Axions Form a Condensate with Long-Range Correlation?,” Phys. Rev. D 92 (2015) no. 10, 103513, arXiv:1412.5930. * (44) W.-M. Zhang, “Coherent states in field theory,” arXiv:hep-th/9908117. * (45) S. Barnett, K. Burnett, and J. Vaccaro, “Why a Condensate Can Be Thought of as Having a Definite Phase,” Journal of Research of the National Institute of Standards and Technology 101 . * (46) E. Goulielmakis, M. Uiberacker, R. Kienberger, A. Baltuska, V. Yakovlev, A. Scrinzi, T. Westerwalbesloh, U. Kleineberg, U. Heinzmann, M. Drescher, and F. Krausz, “Direct Measurement of Light Waves,” Science 305 (2004) no. 5688, 1267–1269. * (47) S. Barnett, K. Burnett, and A. Vacarro, “Why a Condensate can be Thought of Having a Definite Phase,,” J. Res. Natl. Inst. Stand. Technol. 101 (1996) 593. * (48) I. Esteban and J. Salvado, “Long Range Interactions in Cosmology: Implications for Neutrinos,” JCAP 05 (2021) 036, arXiv:2101.05804. * (49) H. Davoudiasl, G. Mohlabeng, and M. Sullivan, “Galactic Dark Matter Population as the Source of Neutrino Masses,” Phys. Rev. D 98 (2018) no. 2, 021301, arXiv:1803.00012. * (50) S. Hannestad, “Structure formation with strongly interacting neutrinos - Implications for the cosmological neutrino mass bound,” JCAP 02 (2005) 011, arXiv:astro-ph/0411475. * (51) S. Hannestad and G. Raffelt, “Constraining invisible neutrino decays with the cosmic microwave background,” Phys. Rev. D 72 (2005) 103514, arXiv:hep-ph/0509278. * (52) N. F. Bell, E. Pierpaoli, and K. Sigurdson, “Cosmological signatures of interacting neutrinos,” Phys. Rev. D 73 (2006) 063523, arXiv:astro-ph/0511410. * (53) A. Basboll, O. E. Bjaelde, S. Hannestad, and G. G. Raffelt, “Are cosmological neutrinos free-streaming?,” Phys. Rev. D 79 (2009) 043512, arXiv:0806.1735. * (54) M. Archidiacono and S. Hannestad, “Updated constraints on non-standard neutrino interactions from Planck,” JCAP 07 (2014) 046, arXiv:1311.3873. * (55) F. Forastieri, M. Lattanzi, and P. Natoli, “Constraints on secret neutrino interactions after Planck,” JCAP 07 (2015) 014, arXiv:1504.04999. * (56) F.-Y. Cyr-Racine and K. Sigurdson, “Limits on Neutrino-Neutrino Scattering in the Early Universe,” Phys. Rev. D 90 (2014) no. 12, 123533, arXiv:1306.1536. * (57) I. M. Oldengott, C. Rampf, and Y. Y. Y. Wong, “Boltzmann hierarchy for interacting neutrinos I: formalism,” JCAP 04 (2015) 016, arXiv:1409.1577. * (58) F. Forastieri, M. Lattanzi, G. Mangano, A. Mirizzi, P. Natoli, and N. Saviano, “Cosmic microwave background constraints on secret interactions among sterile neutrinos,” JCAP 07 (2017) 038, arXiv:1704.00626. * (59) I. M. Oldengott, T. Tram, C. Rampf, and Y. Y. Y. Wong, “Interacting neutrinos in cosmology: exact description and constraints,” JCAP 11 (2017) 027, arXiv:1706.02123. * (60) S. Roy Choudhury, S. Hannestad, and T. Tram, “Updated constraints on massive neutrino self-interactions from cosmology in light of the $H_{0}$ tension,” arXiv:2012.07519. * (61) T. Brinckmann, J. H. Chang, and M. LoVerde, “Self-interacting neutrinos, the Hubble parameter tension, and the Cosmic Microwave Background,” arXiv:2012.11830. * (62) A. Das and S. Ghosh, “Flavor-specific Interaction Favours Strong Neutrino Self-coupling,” arXiv:2011.12315. * (63) A. Mazumdar, S. Mohanty, and P. Parashari, “Flavour specific neutrino self-interaction: $H_{0}$ tension and IceCube,” arXiv:2011.13685. * (64) Planck, N. Aghanim et al., “Planck 2018 results. VI. Cosmological parameters,” Astron. Astrophys. 641 (2020) A6, arXiv:1807.06209. [Erratum: Astron.Astrophys. 652, C4 (2021)]. * (65) GERDA, M. Agostini et al., “Improved Limit on Neutrinoless Double-$\beta$ Decay of 76Ge from GERDA Phase II,” Phys. Rev. Lett. 120 (2018) no. 13, 132503, arXiv:1803.11100. * (66) CUORE, C. Alduino et al., “First Results from CUORE: A Search for Lepton Number Violation via $0\nu\beta\beta$ Decay of 130Te,” Phys. Rev. Lett. 120 (2018) no. 13, 132501, arXiv:1710.07988. * (67) A. S. Barabash, “SeperNEMO double beta decay experiment,” J. Phys. Conf. Ser. 375 (2012) 042012, arXiv:1112.1784. * (68) KamLAND-Zen, A. Gando et al., “Search for Majorana Neutrinos near the Inverted Mass Hierarchy Region with KamLAND-Zen,” Phys. Rev. Lett. 117 (2016) no. 8, 082503, arXiv:1605.02889. [Addendum: Phys.Rev.Lett. 117, 109903 (2016)]. * (69) M. Agostini, G. Benato, and J. Detwiler, “Discovery probability of next-generation neutrinoless double- $\beta$ decay experiments,” Phys. Rev. D 96 (2017) no. 5, 053001, arXiv:1705.02996. * (70) M. J. Dolinski, A. W. P. Poon, and W. Rodejohann, “Neutrinoless Double-Beta Decay: Status and Prospects,” Ann. Rev. Nucl. Part. Sci. 69 (2019) 219–251, arXiv:1902.04097. * (71) W. Rodejohann, “Neutrino-less Double Beta Decay and Particle Physics,” Int. J. Mod. Phys. E 20 (2011) 1833–1930, arXiv:1106.1334. * (72) J. Kotila and F. Iachello, “Phase space factors for double-$\beta$ decay,” Phys. Rev. C 85 (2012) 034316, arXiv:1209.5722. * (73) P. Guzowski, L. Barnes, J. Evans, G. Karagiorgi, N. McCabe, and S. Soldner-Rembold, “Combined limit on the neutrino mass from neutrinoless double-$\beta$ decay and constraints on sterile Majorana neutrinos,” Phys. Rev. D 92 (2015) no. 1, 012002, arXiv:1504.03600. * (74) Super-Kamiokande, J. Yoo et al., “A Search for periodic modulations of the solar neutrino flux in Super-Kamiokande I,” Phys. Rev. D 68 (2003) 092002, arXiv:hep-ex/0307070. * (75) SNO, B. Aharmim et al., “A Search for periodicities in the B-8 solar neutrino flux measured by the Sudbury neutrino observatory,” Phys. Rev. D 72 (2005) 052010, arXiv:hep-ex/0507079. * (76) DUNE, R. Acciarri et al., “Long-Baseline Neutrino Facility (LBNF) and Deep Underground Neutrino Experiment (DUNE): Conceptual Design Report, Volume 2: The Physics Program for DUNE at LBNF,” arXiv:1512.06148. * (77) JUNO, F. An et al., “Neutrino Physics with JUNO,” J. Phys. G 43 (2016) no. 3, 030401, arXiv:1507.05613. * (78) S.-F. Ge and M. Lindner, “Extracting Majorana properties from strong bounds on neutrinoless double beta decay,” Phys. Rev. D 95 (2017) no. 3, 033003, arXiv:1608.01618. * (79) J. T. Penedo and S. T. Petcov, “The $10^{-3}$ eV frontier in neutrinoless double beta decay,” Phys. Lett. B 786 (2018) 410–417, arXiv:1806.03203. * (80) J. Cao, G.-Y. Huang, Y.-F. Li, Y. Wang, L.-J. Wen, Z.-Z. Xing, Z.-H. Zhao, and S. Zhou, “Towards the meV limit of the effective neutrino mass in neutrinoless double-beta decays,” Chin. Phys. C 44 (2020) no. 3, 031001, arXiv:1908.08355. * (81) G.-y. Huang and S. Zhou, “Tentative sensitivity of future $0\nu\beta\beta$-decay experiments to neutrino masses and Majorana CP phases,” JHEP 03 (2021) 084, arXiv:2010.16281. * (82) L. Graf, S. Jana, M. Lindner, W. Rodejohann, and X.-J. Xu, “Flavored neutrinoless double beta decay,” Phys. Rev. D 103 (2021) no. 5, 055007, arXiv:2010.15109. * (83) SNO+, S. Andringa et al., “Current Status and Future Prospects of the SNO+ Experiment,” Adv.High Energy Phys. 2016 (2016) 6194250, arXiv:1508.05759. * (84) LEGEND, N. Abgrall et al., “The Large Enriched Germanium Experiment for Neutrinoless Double Beta Decay (LEGEND),” vol. 1894, p. 020027. 2017\. arXiv:1709.01980. * (85) nEXO, J. Albert et al., “Sensitivity and Discovery Potential of nEXO to Neutrinoless Double Beta Decay,” Phys.Rev. C97 (2018) 065503, arXiv:1710.05075. * (86) Planck, N. Aghanim et al., “Planck 2018 results. VI. Cosmological parameters,” arXiv:1807.06209.
# CEHR-GPT: Generating Electronic Health Records with Chronological Patient Timelines Chao Pang111indicates equal contribution Department of Biomedical Informatics, Columbia University Irving Medical Center Observational Health Data Sciences and Informatics Xinzhuo Jiang∗ Department of Biomedical Informatics, Columbia University Irving Medical Center Observational Health Data Sciences and Informatics Nishanth Parameshwar Pavinkurve Department of Biomedical Informatics, Columbia University Irving Medical Center Observational Health Data Sciences and Informatics Krishna S. Kalluri Department of Biomedical Informatics, Columbia University Irving Medical Center Observational Health Data Sciences and Informatics Elise L. Minto Department of Biomedical Informatics, Columbia University Irving Medical Center Observational Health Data Sciences and Informatics Jason Patterson Department of Biomedical Informatics, Columbia University Irving Medical Center Observational Health Data Sciences and Informatics Linying Zhang Institute for Informatics, Data Science, and Biostatistics. Washington University in St Louis Observational Health Data Sciences and Informatics George Hripcsak Department of Biomedical Informatics, Columbia University Irving Medical Center Medical Informatics Services, New York-Presbyterian Hospital Observational Health Data Sciences and Informatics Noémie Elhadad Department of Biomedical Informatics, Columbia University Irving Medical Center Medical Informatics Services, New York-Presbyterian Hospital Observational Health Data Sciences and Informatics Karthik Natarajan Department of Biomedical Informatics, Columbia University Irving Medical Center Medical Informatics Services, New York-Presbyterian Hospital Observational Health Data Sciences and Informatics ## Abstract Synthetic Electronic Health Records (EHR) have emerged as a pivotal tool in advancing healthcare applications and machine learning models, particularly for researchers without direct access to healthcare data. Although existing methods, like rule-based approaches and generative adversarial networks (GANs), generate synthetic data that resembles real-world EHR data, these methods often use a tabular format, disregarding temporal dependencies in patient histories and limiting data replication. Recently, there has been a growing interest in leveraging Generative Pre-trained Transformers (GPT) for EHR data. This enables applications like disease progression analysis, population estimation, counterfactual reasoning, and synthetic data generation. In this work, we focus on synthetic data generation and demonstrate the capability of training a GPT model using a particular patient representation derived from CEHR-BERT, enabling us to generate patient sequences that can be seamlessly converted to the Observational Medical Outcomes Partnership (OMOP) data format. ## Keywords Generative Pre-trained Transformer, Synthetic Electronic Health Records, Patient Representation, Observational Medical Outcomes Partnership - Common Data Model, Observational Health Data Sciences and Informatics ## 1 Introduction Access to electronic health records (EHRs) is fundamental to healthcare research, drug surveillance, clinical machine learning, and system development. However, the use of real-world EHR data comes with considerable challenges such as privacy and security issues, institutional consent, and restrictions on data sharing. Synthetic data emerges as a promising solution, offering a more expedient and secure pathway to healthcare information, which could accelerate progress across various sectors, including academic research, clinical settings, and the pharmaceutical industry [1]. Synthetic data is not real data, that is, it doesn’t relate to any specific individual. However, it mimics the statistical characteristics and journeys of specific patient populations. Synthetic data enables a broader range of researchers to answer their questions of interest without going through the cumbersome process of accessing real data and worrying about patient privacy [2]. In recent years, many machine learning, specifically deep learning and generative artificial intelligence (AI) models, have been developed to derive synthetic data from real EHR data. [3] However, most existing methods for synthetic EHR data generation fail to adequately capture the temporal dependencies that are often critical in medical scenarios. These temporal aspects, such as medication schedules, symptom progression, and lab result timelines, are vital for a comprehensive understanding of patient health trajectories and for developing effective treatment strategies. An ideal synthetic dataset derived from institutional data should maintain the inherent correlations among time-series features, thus enabling researchers to externally validate machine learning models in different populations. Crucially, the synthetic dataset must preserve accurate patient timelines, as predictive tasks are highly susceptible to temporal variations. A synthetic dataset is considered to exhibit comparable machine learning utility to the original data if it meets two key criteria: 1) it demonstrates similar outcome prevalence to the source data; 2) machine learning models trained on the synthetic data achieve performance metrics akin to those trained with the original data. The majority of the existing research focuses on developing new deep learning models in generative EHR research but without adequate emphasis on retaining accurate temporal information [4, 5, 6]. Unfortunately, synthetic EHR datasets developed as such will not support use cases that require the accurate construction of a patient timeline e.g., 30-day readmission, one-year risk of heart failure, and disease progression. This limits existing work to only perform simple code prediction in their evaluations instead of comprehensive phenotype predictions. Another challenge in using synthetic EHR data in practice is its difficulty with dissemination due to a lack of standards. Synthetic patient sequences cannot be widely adopted for analyses without use of a common data model, however, none of the existing works have included such a component in their frameworks to present the synthetic data in an easy-to- consume fashion. In our view, time-series synthetic data should not only capture the underlying characteristics of heterogeneous EHRs but also satisfy the following temporal requirements, 1) a matching distribution of the starting age; 2) a matching distribution of the starting year; 3) a matching distribution of the inpatient duration; 4) a matching distribution of time intervals between neighboring visits. Furthermore, synthetic EHR data should be stored in common data models such as the Observational Medical Outcomes Partnership (OMOP) Common Data Model, which is used in many large data networks, [7] for easy dissemination and consumption. Although creating such a time-series EHR dataset seems to be a challenging task, we think that this problem can be solved through a patient representation approach. The key is to focus on designing a good patient representation rather than creating a sophisticated architecture to model time and medical events simultaneously. In this paper, we present the CEHR-GPT framework for building an end-to-end workflow to generate time-series synthetic EHR data. Our contributions are summarized below, * • We design a novel patient representation that not only keeps track of the visit types and discharge facility for inpatient visits but also preserves all temporal information including patients’ starting year, starting age, time intervals between visits, and inpatient visit spans. To the best of our knowledge, this is the first time that temporal information is fully retained. * • We treat patient sequence generation as a language modeling problem, which allowed us to use the state-of-the-art language model Generative Pre-trained Transformers (GPT) to learn the distribution of patient sequences to generate new synthetic sequences [8, 9]. * • We converted synthetic sequences to the common data format OMOP with almost no loss of temporal information. Synthetic OMOP can be easily evaluated using the OHDSI tools and disseminated to others. * • We evaluated the synthetic EHR data on three levels, dimension-wise distribution (marginal distribution), co-occurrence relationship, and machine learning model performance metrics. ## 2 Related work With the adoption of Generative Adversarial Networks (GANs) [10], researchers have found creative ways to generate synthetic EHR data. Since 2017, several groups have applied GANs to tabular EHRs and developed several evaluation and privacy metrics to quantify the performance of GANs [3]. Despite the success, one limitation is that the tabular format fails to capture the temporal nature of EHR data because they are constructed from patient histories using a bag- of-words approach. It was not until 2020 that researchers started developing new GAN architectures to tackle the time series data. Dual adversarial autoencoder (DAAE) [6] used a combination of a variational autoencoder (VAE) and two GAN components, where the inner GAN was trained to replicate the encoded representation generated by the encoder, and the outer GAN was trained to generate realistic-looking patient sequences in addition to the reconstruction error. Another model called EHR-M-GAN [5] employed a similar autoencoder architecture with two main differences 1) they used a dual-VAE framework to handle the continuous and discrete valued features, where the continuous and discrete representations were generated by encoders first and then collectively used for decoding; 2) in the GAN generator, they used two parallel recurrent neural networks (one for sampling continuous noise and one for discrete noise) with a so-called Bilateral LSTM cell to allow the continuous and discrete noise vectors to interact with each other to generate better sampling vectors. Although these GANs took the temporal order of events into consideration, they did not generate timestamps for visits or medical events, therefore limiting the use of the synthetic data. Improving upon the previous works, a two-stage learning algorithm (dependency learning and conditional simulation) named SynTEG was proposed to generate timestamped synthetic data [11]. In dependency learning, transformer encoders learned visit representations, which were used to feed into an recurrent neural network (RNN) model to learn the dependencies between visits. Two self- supervising learning tasks were used for training – prediction of the timestamp and diagnoses at the next visit. In the conditional simulation, the learned visit representations generated from the previous step were used to train a conditional GAN to generate the diagnosis codes at each visit. This approach achieved superior performance in the time-sensitive evaluations over the previous methods. However, there are still ongoing challenges that have not been addressed in their work, 1) other EHR data was not utilized because only diagnosis codes were included in training; 2) the visits were assumed to start and end on the same day, which would fail to model in-patient visits that normally span days; therefore, the constructed patient timeline would be inaccurate; 3) the synthetic data didn’t include visit types and discharge facilities. Until now, almost all the existing approaches used some variation of GAN for learning the data distribution, but unfortunately, GANs are notoriously difficult to train and easily subject to mode collapse. Despite the recent advancements in optimization techniques such as Wasserstein-GAN [12], it would require a significant amount of time to tune hyperparameters to properly train GANs. As the previous works have demonstrated [13, 14, 15], patient sequence generation could be conceptually represented as a language modeling problem. Foresight [16] adapted GPT to forecast patient trajectories. Their methodused a name entity recognition (NER) tool to extract medical concepts from discharge summaries, based on which a patient sequence was constructed chronologically, and then they trained a standard GPT model using all patient sequences constructed from the previous step. To forecast future events, they fed a patient history to prompt the model and employed a Monte Carlo sampling strategy to calculate the probability of developing certain conditions. One limitation of this work was that the model could not predict when a certain condition would happen due to a lack of temporal information in their patient sequence. Nevertheless, their work demonstrated the potential of using GPT for modeling patient sequences. We aim to address this limitations in our work. ## 3 Methods In Figure 1, we present a framework for generating synthetic EHR data from an OMOP source. To retain the temporal dependencies, we opted to work directly with time-series patient sequences instead of using a bag-of-words (BOW) representation by building on our previous work [13]. We first encoded the OMOP data into patient sequences using a specific patient representation, described later in this section. Secondly, we trained a generative model on the converted patient sequences and utilized it to generate new synthetic patient sequences. Figure 1: Overall architecture. The OMOP data is first converted to patient sequences by an OMOP encoder based on the patient representation that preserves demographics, visit types, and temporal intervals between visits. Then a generative model is trained to learn the sequence distribution in order to generate new sequences. Next, the generated sequences are converted back to the OMOP format using an OMOP decoder. Finally, we fed the synthetic patient sequences into an OMOP decoder to create a synthetic OMOP dataset. Furthermore, an evaluation procedure was developed to assess the similarity between the synthetic OMOP and the source OMOP data. ### 3.1 Patient Representation We designed a patient representation in CEHR-BERT [13] that captures medically relevant events and their timelines while exhibiting certain characteristics of a sentence. In order to fully leverage Large Language Models (LLM) on patient sequences, we further extended this patient representation to include demographic information, patient history, and temporal dependencies as shown in Figure 2. The start of the sequence defines the demographic prompt containing EHR start year, age, gender, and race. It is followed by visit blocks separated by artificial time tokens (ATT) representing time intervals in days, e.g., $D_{1}$ represents an interval of 1 day. For time intervals surpassing 1080 days, we grouped these into a single Long Term (LT) token, a decision guided by the low occurrence rate in this time frame. Each visit block starts with a visit type token (VTT) to signify the type of visit and is then followed by domain records arranged in chronological order. In the case of inpatient visits, a distinct inpatient ATT (IATT) was inserted between neighboring inpatient spans, defined as the groups of records that occurred on the same day. These are distinct from ATT and are used for capturing the time between multiple events characterized by concepts within the same visit. In addition, a discharge facility code (e.g. discharge home and long-term care) was inserted at the end of the inpatient visit. The use of IATT, which is distinct from ATT, was necessary since they are attributed to two different contexts, and resulted in better performance than using the same ATT across both contexts. This patient representation allows us to convert from any common data model (e.g. OMOP) to patient sequences and vice versa without any loss of temporal information. To formulate this property, let’s denote $D_{i}$ to be data associated with the $i$th patient in the source format, $P_{i}$ to be the patient sequence converted from $D_{i}$, and $F$ to be the function that converts source data to patient sequences, represented as $P_{i}=F(D_{i})$. Let’s then denote $D_{i}^{\prime}$ to be the reconstructed source data and $F^{\prime}$ to be the inverse function that converts patient sequences back to original source format. This indicates $D_{i}^{\prime}=F^{\prime}(P_{i})$. Finally, let’s denote $T$ to be a function that extracts all the dates for a set of patient records. The patient representation is said to preserve temporal information perfectly if and only if the following statement is true for every single patient, $T(D_{i})=T(D_{i}^{\prime})+C_{i}$, where $C_{i}$ is a constant that represents a consistent time shift e.g., $C_{i}=4$ days. Figure 2: The patient representation preserves demographics, visit types, and temporal intervals between visits and inpatient duration. It’s designed to have the demographic prompt at the beginning including year at the first visit, age at the first visit, gender and race tokens, then followed by a series of visit blocks to represent the complete patient timeline. An artificial time token (ATT) is inserted between the neighboring visit blocks to keep track of the time intervals in days. In each visit block, all the essential information is retained including the visit type and domain records. In the case of inpatient visits, the inpatient ATT tokens (representing time intervals in days) are inserted between groups of concepts that occur on the same day, in addition, a discharge token is provided at the end of the visit block. ### 3.2 OMOP Encoder To create a patient sequence, we began by generating a demographic prompt using data from the OMOP person and visit tables. This prompt included essential demographic information such as the patient’s age at their initial visit, the year of their first visit, their gender, and their race. Subsequently, we constructed a series of visit blocks to represent the patient’s entire medical history. We inserted an ATT token between these visit blocks to signify the time intervals between them. Within each visit block, we gathered all relevant records from OMOP domain tables (e.g., condition) and arranged them chronologically based on their respective timestamps. In cases where there are timestamp ties, we sorted the concepts. Additionally, three artificial tokens (VS, VE, and VTT) were added at the beginning and end of each visit block to denote the start, end, and type of the visit, as illustrated in Figure 2. For inpatient visit blocks, extra processing steps were necessary. Initially, we grouped records by their timestamps to identify all inpatient spans, arranging them in chronological order. Next, we inserted IATT between these spans, aligning them with the respective time intervals. Finally, the discharge facility code was extracted from the OMOP visit table and appended to the end of the block. ### 3.3 Generative Model We used a GPT model with standard transformer decoders, where the input layer utilized concept embedding and trainable positional embedding. The model was trained using the Next Word Prediction learning objective. When generating a patient sequence, we randomly sampled a demographic prompt from our source sequences, which served as the input to the GPT model. Using these prompts, the entire patient history was generated autoregressively by sampling tokens from the predictive distribution at the final layer. ### 3.4 OMOP Decoder The patient sequence was converted back to the OMOP format using the OMOP decoder. The start-year prompt determined the EHR history’s beginning, using January 1st as the default. Demographic data was stored in the person table, while concepts were transformed into condition, drug, and procedure tables. A date cursor was used to represent the “current time” as we were processing each patient sequence, it was initially set to the star year and was updated whenever an ATT token was encountered. We first parsed out the number of days represented by the ATT token, and then moved the data cursor by the same number of days to the future. During the sequence processing, the VS token marked the start of a new visit block. We, therefore, extracted the token corresponding to the visit type (the token immediately followed by VS) and created a new visit record with the corresponding type. All the tokens subsumed by this visit block were converted to condition, drug, and procedure records and linked to the current visit. For outpatient visits, we assumed that all data points were generated on the same day; therefore, we set the start and end dates of the visit to the current value of the date cursor. Similarly, all the domain records were set to the date cursor as well. When processing inpatient visits, the date cursor was updated inside the visit block based on the time intervals represented by IATT tokens between inpatient spans. Domain records were generated with the current value of the date cursor. Towards the end of the inpatient visit block, we extracted the last token (the token right before VE) corresponding to a discharge facility and updated the visit end date using the current value of the date cursor (as the date cursor was frequently updated inside an inpatient visit block). This allowed us to preserve the complete information about inpatient visits. If generated sequences do not follow patterns presented in the patient representation, they will be discarded to ensure the quality of the synthetic data. ## 4 Experiments and Results ### 4.1 Data and Preprocessing The source patient sequences were generated from the OMOP database derived from Columbia University Irving Medical Center-New York Presbyterian Hospital EHR data, which includes 3.7 million unique patients’ medical histories including condition, medication, and procedure. Unknown concepts (i.e., $concept\\_id=0$) were removed from all domains except for the visit type when constructing the patient sequences using the proposed patient representation. Patients with less than 20 tokens were removed from the training dataset, and approximately 2.3 million patients were included for training whereas 75,000 patients were held out for privacy evaluations. For the GPT model, we used a context window of 512, 16 transformer decoders, 8 attention heads with a dropout rate of 0.1, and 128 dimensions for both the embedding and hidden units. All patients with longer than 512 tokens were post-truncated to fit the context window. The statistics of training data was summarized in Table 1. We trained the model for 2 epochs on 2 Nvidia 2080 TI GPUs with a batch size of 32 and a learning rate of 0.0002. The model checkpoint was created every 10,000 steps. | No. of visits per patient | Sequence length per patient ---|---|--- mean | 16 | 148 std | 19 | 154 min | 2 | 20 25% | 4 | 38 50% | 8 | 78 75% | 21 | 198 max | 102 | 512 Table 1: Summary statistics of the CUIMC-NYP OMOP training data During the first epoch, we used a standard data generator strategy, where every training example was fed to the model; however, we switched to a random sampling strategy to draw training examples during the second epoch. For synthetic data generation, we used the 10th model snapshot because early experiments showed its superior performance compared to other snapshots. Furthermore, we used several sampling hyper-parameters including top k=300, top k=200, top k=100, top p=90%, top=95%, and top=100% to generate different synthetic OMOP datasets. Using specific top p/k values is a common technique in language models for data generation. For instance, the top k approach limits the selection of the k most probable tokens in the prediction distribution during sampling. On the other hand, the top p method selects a set of most likely tokens whose combined probability reaches p% (e.g. 90%) in the predictive distribution. For each sampling strategy, 1M synthetic sequences were generated and converted to OMOP. On average, 98% of the generated sequences passed the validation and were converted to OMOP. ### 4.2 Evaluations We followed the evaluation procedures proposed by [17] to compute the data utility metrics including dimension-wise distribution, co-occurrence relationship, and machine learning model performance. Some of the metrics were originally designed for tabular EHR data; therefore, we adapted them to the time-series setting. When using Kullback-Leibler (KL) divergence to evaluate source and synthetic datasets, we use concept probabilities defined as, $P_{prob}(c)=\frac{\sum_{i}^{n}\mathbbm{1}\Big{[}c\in h_{i}\Big{]}}{\sum_{i}^{n}\sum_{j}^{m}\mathbbm{1}\Big{[}c_{j}\in h_{i}\Big{]}}$ where $c$ denotes the target concept, $h_{i}$ denotes the $ith$ patient history, $n$ and $m$ denote the total number of patients and concepts respectively. Due to the small probability values, we opted to use the prevalence instead for data visualization using a slightly modified formula below, $P_{prev}(c)=\frac{\sum_{i}^{n}\mathbbm{1}\Big{[}c\in h_{i}\Big{]}}{n}$ The only difference between the prevalence and the probability is the normalization constant used as the denominator. For baseline comparison, we added three variants of GPT models trained on slightly different patient representations. The first baseline model was trained on the adjusted CEHR-BERT representation, which differs from the proposed patient representation by 1) CEHR-BERT ATT tokens contain a mix of day/week/month/year tokens to represent time intervals while CEHR-GPT only used the day tokens and 2) the IATT tokens and discharge facility tokens were not used in the CEHR-BERT representation. The second baseline model used the proposed patient representation with IATT tokens removed. We will refer to this baseline as GPT-OUTPAT. The last baseline GPT was trained on the patient representation widely used in time-series EHR research, where the concepts were simply ordered chronologically and put in a sequence without any additional artificial tokens. To use such sequences for comparison, we assumed that all medical events in the patient sequence belonged to a single visit. This baseline model will be referred to as GPT-Vanilla. We only used the top_p=95% sampling strategy to generate synthetic OMOPs for these baseline models. The comparison between these patient representations can be seen in Supplementary Figure 9. #### 4.2.1 Dimension-wise Distribution KL divergence was assessed to compare the concept probability distributions between synthetic and source datasets among the entire population. In Figure 3, synthetic datasets generated by different patient representations and sampling strategies were evaluated against real patient data. The results showed that baseline models CEHR-BERT, GPT-Vanilla and GPT-OUTPAT with sampling strategy using the threshold of top_p$=95\%$ diverged the least from real concept probability distributions followed by GPT models trained with top_p$=100\%$ and top_p$=95\%$. GPT models with threshold of top_k$=300$ and top_k$=200$ with relatively similar divergence. However, GPT models with threshold of top_k$=100$ sampling strategy had the largest KL divergence. Figure 3: KL divergence for comparing concept probability distribution between synthetic data and real data. The probabilities of concepts were calculated on the scale of the entire population. Furthermore, we conducted a qualitative analysis using the synthetic data top_p=95% to gain a more comprehensive understanding. The dimension-wise distributions between the synthetic and source datasets were compared at three distinct levels: the entire population, specific sub-groups (e.g., female population), and particular cohorts (e.g., hospitalization cases). Figure 4 illustrates the concept prevalence comparison between the original OMOP dataset and the generated OMOP dataset, using a threshold of top_p$=95\%$. In the high-frequency regions, most data points cluster closely around the diagonal line, indicating a strong agreement between the source and synthetic data. Conversely, data points appear more dispersed for low-frequency concepts. Notably, in the subplot representing female conditions (located in the first column, second row), there is an unusual cluster of concepts positioned above the diagonal line. Further examination revealed that these concepts were male-specific and should not appear in the female population. Although there are a few instances of such cases in the source data, GPT amplified such cases in synthetic data. Figure 4: Concept prevalence comparison between the source OMOP and generated OMOP using top $p=95\%$ in the log scale stratified by domain in columns and by population in rows, where x-axis and y-axis represent the source and the synthetic data respectively, and each dot represents a concept In addition, we conducted a detailed comparison of the visit tables between the two OMOP datasets. Our specific focus was on performing demographic breakdowns and analyses for gender, race, and age group. The supplementary Figures 10, 12, 11 highlight the top 10 most prevalent visits for each demographic breakdown, showcasing that the trends in both the source and generated visit tables exhibited notable similarities. #### 4.2.2 Co-occurrence Relationship To measure how closely the generated datasets resemble the source, we computed the KL divergence between their co-occurrence matrices. The matrix was constructed temporally with the following logic: 1) for each concept in the patient sequence, only the future concepts were taken into consideration for creating concept pairs; 2) each patient could only contribute to the same concept pair once; 3) the matrix was normalized into a proper probability distribution by dividing the occurrences of each concept pair by the overall number of pairs. Additionally, we set benchmarks for our analysis: a lower- bound and an upper-bound. The lower bound was determined by applying the KL divergence method to two random samples from the source data. The upper bound was established by creating a theoretical co-occurrence matrix under the assumption that all concepts in the source data were independent. The KL divergence was then applied to this hypothetical matrix to calculate the upper bound. Figure 5 illustrates that the datasets with CEHR-BERT and top_k=300 most closely approach the lower bound, with the top_p=95% and GPT-OUTPAT baselines coming next. Datasets with top_k=100 and top_k=200 had marginally higher KL divergence values, while the top_p=100%, the GPT-Vanilla datasets exhibited the largest KL divergence. To thoroughly examine the similarities between the original and the synthetic data, we carried out a qualitative analysis using one of the synthetic datasets. This involved comparing the most frequent pairs of co-occurring concepts within each category, such as condition-condition (interpreted as one condition concept followed by another condition). Figure 5: KL divergence associated with different synthetic data. The closer to the lower bound in the bottom left corner, the better the synthetic data. Figure 6: Top 100 co-occurring concept pairs for each co-occurrence category e.g. condition-condition interpreted as a condition concept followed by another condition concept. The x and y axes represent the synthetic and source data respectively The results, illustrated in Figure 6, display the top 100 pairs for each type of co-occurrence relationship. The analysis revealed that the synthetic data accurately mirrored the co-occurrence patterns in most categories, with the exception of the categories ending with a drug concept (as shown in the second column of Figure 6), where the data points were more scattered. Notably, a majority of the points in condition-condition and procedure-procedure pairs aligned along the diagonal line. Figure 7: Comparison of temporal co-occurrence networks around Type 2 Diabetes Mellitus (T2DM) for real (left) and synthetic data (right). Networks were constructed with T2DM’s (large circle) top 5 co-occurring condition concepts and their own top 5 co-occurring condition concepts. The metrics Prevalence (top) and Pointwise Mutual Information-3 (PMI3) (bottom) were used to quantify co-occurrence. Gold edges indicate edges that are shared by both real and synthetic data, edge thickness indicates the strength of co-occurrence, and arrow direction indicates the direction of the temporal association. For an in-depth analysis, we focused on examining the co-occurrence relationships associated with Type 2 Diabetes Mellitus (T2DM). Figure 7 qualitatively compares the real and top_p=95% synthetic data networks of condition concepts that co-occur around T2DM. The degree of co-occurrence is calculated with prevalence, which evaluates the frequency of concept co- occurrence, and Pointwise Mutual Information-3 (PMI3), which evaluates the probabilistic association of concepts. There is much overlap between the real and synthetic networks when prevalence is used to construct them, but connections are more disparate when PMI3 is used. #### 4.2.3 Predictive performance For this analysis, we constructed five prediction tasks using the method described in [13] and the Book of OHDSI [18]. Table 2 shows the cohort and the corresponding definition. Cohort | Definition ---|--- HF readmission | HF patients who have a 30-day all-cause readmission. Observation window: 360 days, Prediction windows 30 days Hospitalization | 2-year risk of hospitalization starting from the 3rd year since the initial entry into the EHR system. Observation window: 540 days, hold-off window: 180 days, prediction windows: 720 days COPD readmission | COPD patients who have a 30-day all-cause readmission. Observation window: 360 days, prediction windows: 30 days Afib ischemic stroke | Afib patients with 1-year risk since the initial diagnosis of afib ischemic stroke. Observation window: 720 days, prediction windows: 360 days CAD CABG | Patients initially diagnosed with Coronary Arterial Disease (CAD) without any prior stent graft will receive the Coronary artery bypass surgery (CABG) treatment. Observation window: 720 days, prediction windows: 360 days Table 2: Cohort definitions To extract features, we first rolled up the medical concepts using ontological hierarchies to reduce dimensionality [see supplementary materials] and used the bag-of-word (BOW) approach, where we counted the frequency of each concept in a given observation window. For each task, we split the cohort data (both synthetic and real) into training and testing sets with a split ratio of 85:15. We ran logistic regression using Sklearn’s implementation with the default configuration. Finally, the area under the receiver operating characteristics curve (AUC) was calculated using the test set. In addition, we reported PR-AUC (precision-recall) due to the class imbalance often present in EHR data. Table 3 shows the prevalence of the positive cases, ROC-AUC, and PR- AUC for each synthetic data. The metrics associated with the baseline CEHR- BERT are reported in the Supplementary Materials Table 7. For easy comparison of synthetic datasets, we defined a consolidated distance metric as the weighted average of the relative differences of the three aforementioned metrics, $dist=\frac{|\delta_{Pre}|\times 0.5}{Pre_{true}}+\frac{|\delta_{AUC}|\times 0.25}{AUC_{true}}+\frac{|\delta_{PR}|\times 0.25}{PR_{true}}$ where $\delta_{Pre}$, $\delta_{AUC}$, and $\delta_{PR}$ represent the differences in prevalence, ROC-AUC, and PR-AUC between the source and the synthetic data; $Pre_{true}$, $AUC_{true}$, and $PR_{true}$ denote the ground truth metrics generated from the source data. Figure 8 presents the distance metrics for various synthetic datasets across different cohorts. The dataset created with top_k=300 displayed the best performance in HF and COPD Readmission but was less effective in other cohorts. The top_p=95% dataset maintained consistent performance levels, ranking as the best in Hospitalization and Afib ischemic stroke, and as second-best in HF and COPD readmission. In comparison, the GPT-OUTPAT baseline dataset exhibited a slightly higher divergence across the Hospitalization, Afib Ischemic Stroke, and CAD CABG groups. However, the CEHR-BERT baseline showed similar patterns in most cohorts, with a notably lower divergence in CAD CABG. It is noted that both GPT-OUTPAT and CEHR-BERT were not included in the HF Readmission and COPD Readmission analyses due to the absence of synthetic patients meeting the selection criteria for these cohorts. Other synthetic datasets, specifically those generated with top_p=100% and top_k=200 lagged behind in performance. Interestingly, top_k=100 showed a unique pattern, being the closest in the distance for the CAD CABG cohort but underperforming in the others. Figure 8: The consolidated distance metrics for different synthetic datasets stratified by cohort. GPT-Vanilla and CEHR-BERT were omitted as the cohorts couldn’t be constructed due to the loss of temporal information. GPT-OUTPAT was omitted from HF readmission and COPD readmission as the cohorts for the same reason. Cohort | Real | p=95% | p=100% | k=100 | k=200 | k=300 | GPT-OUTPAT ---|---|---|---|---|---|---|--- HF readmission | Pre = 25.7 AUC = 65.7 PR = 39.3 | Pre = 27.6 AUC = 69.2 PR = 45.7 | Pre =27.7 AUC = 52.4 PR = 29.0 | Pre = 30.7 AUC = 68.1 PR = 47.8 | Pre = 29.3 AUC = 54.0 PR = 32.9 | Pre = 26.5 AUC = 61.1 PR = 33.8 | Pre =100.0 AUC = NA PR = NA Hospitalization | Pre = 5.6 AUC = 75.3 PR = 19.5 | Pre = 5.2 AUC = 77.1 PR = 21.4 | Pre = 7.4 AUC = 71.3 PR = 20.2 | Pre = 2.8 AUC = 87.0 PR = 22.1 | Pre = 5.2 AUC = 84.2 PR = 20.8 | Pre = 6.3 AUC = 78.7 PR = 24.6 | Pre = 5.2 AUC = 70.2 PR = 14.3 COPD readmission | Pre = 34.5 AUC = 74.2 PR = 83.8 | Pre = 37.8 AUC = 76.4 PR = 84.4 | Pre = 47.2 AUC = 74.1 PR = 67.2 | Pre = 26.4 AUC = 75.9 PR = 90.3 | Pre = 28.3 AUC = 70.1 PR = 82.8 | Pre = 34.5 AUC = 68.8 PR = 80.2 | Pre = NA AUC = NA PR = NA Afib ischemic stroke | Pre = 8.7 AUC = 84.0 PR = 48.5 | Pre = 10.2 AUC = 78.9 PR = 41.2 | Pre = 10.4 AUC = 70.7 PR = 39.1 | Pre = 16.6 AUC = 77.1 PR = 50.5 | Pre = 15.8 AUC =68.9 PR = 36.6 | Pre = 10.8 AUC = 76.8 PR = 38.5 | Pre = 9.7 AUC = 67.2 PR = 27.2 CAD CABG | Pre = 7.1 AUC = 88.4 PR = 55.9 | Pre = 4.1 AUC = 81.5 PR = 25.2 | Pre = 4.4 AUC = 52.9 PR = 4.3 | Pre = 7.2 AUC = 84.7 PR = 31.3 | Pre = 4.9 AUC = 73.5 PR = 24.3 | Pre = 4.0 AUC = 79.0 PR = 24.1 | Pre = 3.5 AUC = 81.5 PR = 44.4 Table 3: Logistic regression model performance across different datasets. In each cell, three numbers were reported including the prevalence of the positive cases, ROC-AUC, and PR-AUC ### 4.3 Privacy Evaluations We adopted the privacy evaluation framework outlined in [17], focusing on quantifying the risk of privacy breaches through attribute and member inference attacks. The attribute inference risk assesses the likelihood of inferring sensitive attributes from a synthetic dataset by utilizing targeted patients’ demographic information and common diagnoses found in real datasets available to the adversary. In a membership inference attack, the adversary tries to determine if the target record was in the training set given all or partial attributes. #### 4.3.1 Membership Inference Attack We simulated two different attack scenarios, namely the dataset attack and the model attack. In a dataset attack, attackers only have access to the synthetic dataset, whereas they can query the model itself in a model attack. Their goal is to find a way to infer whether or not a real patient record was used for training [19]. For example, the model might output a high probability associated with the training data points and a low probability for the non- training ones, similarly, the synthetic data might be more similar to the training data than the non-training ones. The attack dataset was constructed based on the following procedure 1) we created the negative examples by taking the 75000 holdout set and labeling them as negative; 2) we created the positive examples by randomly selecting 75000 patients from the training set and assigning them as positive. In the dataset attack, for each attack data point, we found the best match from the synthetic data using a hamming distance metric, which sums up the absolute differences in age, number of visits, and concepts [19], finally, the median distance was used as a threshold, below which patients were predicted as positive and negative otherwise. In the model attack, we computed the average loss of each patient sequence by feeding the attack dataset to the model directly, similarly, the median loss was used to infer the positive/negative cases, where the positive cases were assigned to those below the threshold and positive otherwise. Finally, we calculated recall/precision/f1/accuracy using the ground truth labels against those inferred predictions. Table 4 shows that the accuracy of both attacks is slightly less than 50%, indicating that the performance of such attacks is worse than a random guess. | Accuracy | Recall | Precision | F1 ---|---|---|---|--- Model Attack | 0.4955 | 0.4955 | 0.4942 | 0.4949 Data Attack | 0.4941 | 0.4996 | 0.4959 | 0.4978 Table 4: Membership Inference Attack metrics #### 4.3.2 Attribute Inference Attack In attribute inference attack, we assumed the attacker has access to a group of target patients along with their demographic data, including age, gender, race, the year of their initial clinical visit, and prevalent clinical conditions like hypertension, abdominal pain, chest pain, etc. With these features, the attacker identifies the patient in the synthetic dataset with the closest attribute resemblance to the target individual. Then the attacker aims to use the sensitive attributes of matched patient from synthetic dataset to infer the corresponding sensitive attributes of the target patient. To quantitatively evaluate the attribute inference risk, we randomly sampled 150,000 patients from the real patient dataset as a target group. The search group compromised 1 million generated synthetic patients. To find most similar patients from the search group, we created a set of common attributes including the demographic data along with the top 1% of the most prevalent condition concepts, represented as one-hot encoded features. Then a k-nearest neighbors (KNN with k = 1) algorithm was applied to each target patient and a synthetic patient with the smallest euclidean distance was found. Finally we extracted the sensitive attributes (condition concepts not in the top 1% tier) from matched target and synthetic patients. F1 scores were computed for the sensitive attributes of each matched patient pair and aggregated across all matched patients. A baseline analysis was performed substituting real patients for synthetic patients. A result lower than the baseline suggests that the likelihood of finding a synthetic data similar to a real patient is lower than finding a real patient similar to a real patient who share sensitive attributes. This implies a lower attribute inference risk which could be acceptable. Our training set was randomly divided into two halves, with 1 million real patients assigned as the target group, and the remaining 1 million as the search group. The matching process was consistently applied and an aggregated F1 score was computed. Table 5 shows that the F1 score of synthetic vs real is less than real vs real scenario. But it’s still higher than other models who didn’t capture the real patterns as effectively. [19] | Recall | Precision | F1 ---|---|---|--- Synthetic vs Real | 0.0350 | 0.0343 | 0.0271 Real vs Real | 0.0612 | 0.0468 | 0.0421 Table 5: Attribute Inference Attack metrics ## 5 Discussion To the best of our knowledge, this is the first attempt to utilize GPT for generating time-series heterogeneous EHR data, while preserving patient privacy. Our main contribution lies in designing a novel patient representation that preserves a complete timeline of the patient’s history, along with crucial visit details, thereby enabling GPT to create realistic patient sequences. Importantly, this representation facilitates seamless conversion back to the OMOP format, simplifying dissemination and analysis. This patient representation could serve as an effective messenger for transferring information across various standard data models. At present, our system is tailored to the OMOP format. However, it is designed with adaptability in mind, enabling us to seamlessly integrate new encoder/decoder pairs. This flexibility would facilitate the conversion of patient sequences to other widely-used data models, such as i2b2[20]. The study undertakes a three-tiered evaluation approach, systematically comparing synthetic and real datasets based on their marginal (column-wise distribution), conditional (co-occurrence relationship), and joint distributions (predictive performance). Concurrently, as the evaluation progresses through these three levels, there is a corresponding escalation in the complexity and challenge of the tasks involved. The outcomes of the KL divergence analysis revealed a nuanced relationship between the top k/p sampling strategies and the performance across the evaluation levels. Specifically, an increase in top k/p values enhanced performance in level 1 concept prevalence. However, an excessively high or low top k/p value adversely affects both level 2 co-occurrence metrics and level 3 machine learning predictions. This pattern suggests that including more tokens in the predictive distribution introduces greater uncertainty and a wider array of potential patient trajectory variations in the data generation process, thereby escalating the difficulty of achieving comparable performance outcomes. The sampling strategies of top_p=95% and top_k=300 seem to be most effective for generating the synthetic data. For instance, the synthetic data created with top_p=95% demonstrates the second smallest divergence in both dimension- wise distribution and co-occurrence relationship. Simultaneously, the corresponding synthetic cohorts successfully replicated the performance metrics in all predictive tasks, with the exception of CAD CABG. Finally, the significance of this patient representation transcends synthetic data generation; we believe it has the potential to establish the groundwork for integrating time into patient representations across diverse EHR-based deep- learning models. ### 5.1 Loss of Temporal Information The reason that CEHR-GPT replicated the performance metrics of the prediction tasks can be attributed to the use of time tokens in its underlying patient representation. The majority of the prediction problems are phrased as “For a group of target patients who share similar characteristics, who would experience a particular medical event in one year from the index event?” in EHR research, therefore maintaining a complete patient timeline is crucial for time-sensitive cohort constructions [18]. We claimed that the proposed patient representation had almost zero loss of temporal information, although this makes intuitive sense, there does not exist a formal metric to quantify this. To bridge this gap, we conceived a new metric named loss of temporal information (LOTI) to estimate the shrinkage of the patient timeline due to the use of the patient representation in an EHR dataset. Let’s denote $T$ to be the time interval measured in days, ATT to be an artificial time token that represents a time interval ($W_{0}$), $F$ to be a function that maps $T$ to an ATT token (four days to $W_{0}$). In addition, let $G$ be the inverse function of $F$ that converts an ATT to $T$, moreover, we impose the constraint on G such that it takes the lower bound of ATT e.g., $W_{0}\implies 0$ days. Formally, we define LOTI as the expected difference between the original time interval $T$ and the reconstructed time interval $G(F(T))$ as the following, $LOTI=E_{p(T)}\Big{[}T-G\big{(}F(T)\big{)}\Big{]}$ where $P(T)$ is the probability of $T$ observed in the training data defined as, $P(T)=\frac{\textit{Freq of T}}{\Sigma\textsuperscript{T}\textit{Freq of T}}$ We computed LOTI for the patient representations utilized in CEHR-GPT and baseline models, shown in Figure 9. As Table 6 shows, CEHR-GPT has the least LOTI compared to the other patient representations while GPT-OUTPAT has a slightly higher time shrinkage because the inpatient duration was not retained. CEHR-BERT has a relatively large LOTI compared to the previous two representations due to the use of coarse ATT tokens. Finally, GPT-Vanilla has the most LOTI, which is equal to the expected length of the timeline in the source population due to the complete collapse of the timeline. Representation | Between visit ATT token | Between inpatient span ATT token | LOTI ---|---|---|--- CEHR-GPT | Day token for $T\leq 1080$ LT token for $T>1080$ | Day token | 7.739 GPT-OUTPAT | Day token for $T\leq 1080$ LT token for $T>1080$ | N/A | 7.962 CEHR-BERT | Day token for $T<7$ Week token for $7\leq T<30$ Month token for $30\leq T<360$ LT token $T\geq 360$ | N/A | 31.482 GPT-Vanilla | N/A | N/A | 111.164 Table 6: Loss of Temporal Information for Different Patient Representations ### 5.2 Time Invariance and Sensitivity Across all synthetic datasets, the dimension-wise distribution (marginal distribution) and co-occurrence relationship were well preserved, regardless of the patient timeline’s integrity within the models used. Even with a high LOTI, baseline models such as CEHR-BERT accurately mirrored both marginal distribution and co-occurrence relationship, yielding results similar to those from CEHR-GPT. This implies that these two measures may be largely time- invariant, unaffected by any shrinkage in the patient timeline. The rationale behind this is rooted in their construction methods, which either disregard or marginalize the temporal factor. Marginal distribution was constructed by counting the unique number of patients associated with the target concept, which was then normalized by a constant. The construction disregarded temporality, as the placement of a concept on the timeline was not a factor of consideration. The co-occurrence matrix was created using time initially but was marginalized after all co-occurring pairs were collected from the patient population. On the contrary, the predictive performance is extremely sensitive to any change made to the patient timeline as the cohort construction requires the integrity of the patient timeline. For example, HF readmission requires a 30-day prediction window from the index event (defined as the hospitalization episode with a heart failure diagnosis). Any shrinkage to the patient timeline will disrupt the construction of this cohort. The synthetic HF readmission cohort produced by CEHR-BERT showed a readmission rate of 100% due to the shrinkage of the timeline in this patient representation; whereas, the actual expected rate of readmission should be approximately 25%. Compared to CEHR- GPT, GPT-OUTPAT encoded time intervals between visits but did not preserve the duration of inpatient visits, therefore having a slightly higher LOTI. As a consequence, it showed reasonable performance in the cohorts (Hospitalization and Afib ischemic stroke), which had a 360-day prediction window and were thus less impacted by timeline shrinkage. However, in the case of HF Readmission (where a 100% readmission rate was observed) and COPD Readmission (which identified no patients), GPT-OUTPAT was less successful. These cohorts used a short 30-day prediction window, making them highly sensitive to any distortions in the timeline, which likely led to synthetic patients not meeting the cohort selection criteria. Interestingly, the CAD CABG cohort presents a notable deviation from the general trend, where the CEHR-GPT dataset with top_k=100 outperformed both the top_p=95% and top_k=300 configurations. Additionally, the CEHR-BERT synthetic data accurately replicated the machine learning performance metrics as well. This indicates that the CAD CABG cohort was less affected by time shrinkage. Therefore, selecting the appropriate patient representation is pivotal in maintaining specific properties of the source data when generating synthetic data. The choice hinges on the intended application of synthetic data, ensuring that critical features and patterns inherent to the patient information are accurately reflected and retained. ### 5.3 Time Sensitive Forecasting Because the patient representation encodes all the temporal information in the sequence, the trained GPT model could be used potentially for time-sensitive forecasting. We could prompt the trained GPT model with a patient history and estimate the time of the next visit via a Monte Carlo Sampling approach shown in the following equation, $P(\delta_{t}|h)\approx\frac{\sum^{n}_{i=1}\mathbbm{1}\Big{[}M_{gpt}(h)=\delta_{t}\Big{]}}{n}$ where $M_{gpt}$ denotes the GPT model, $h$ denotes a patient history, $\delta_{t}$ denotes any time interval, and $n$ represents the number of samples. Then we can use the expectation $E\big{(}\delta_{t}\big{)}$ as the predicted time interval. In addition, we can quantify the confidence by calculating the standard deviation e.g. $sd(\delta_{t})$. Similarly, we can predict the visit type ($v$) using the same Monte Carlo approach. $P(v|E\big{(}\delta_{t}\big{)},h)\approx\frac{\sum^{n}_{i=1}\mathbbm{1}\Big{[}M_{gpt}\Big{(}E\big{(}\delta_{t}\big{)},h\Big{)}=v\Big{]}}{n}$ Finally, we can predict the most likely medical event ($c$) given the predicted visit type $v$, $P(c|v,E\big{(}\delta_{t}\big{)},h)\approx\frac{\sum^{n}_{i=1}\mathbbm{1}\Big{[}M_{gpt}\Big{(}v,E\big{(}\delta_{t}\big{)},h\Big{)}=c\Big{]}}{n}$ This approach goes beyond conventional prediction methods by not only forecasting future medical events but also determining the timing of the next visit and the specific medical events associated with that visit type. This predictive model could provide a more detailed and actionable timeline for patient care. ### 5.4 Limitations While synthetic datasets demonstrated a high degree of similarity to source data, they are subject to certain known constraints. Firstly, a selection bias was present in the training data due to the constraints on sequence length, ranging from 20 to 512. This limitation resulted in the partial inclusion of patients with chronic conditions, which typically require a longer context window for accurate representation. While extending the context window of the model could potentially address this issue, it may introduce unforeseen effects. Finding the optimal configuration to accommodate a broader context window would require comprehensive experiments. Secondly, identifying an optimal sampling strategy for generating synthetic data remains a challenge due to the presence of numerous hyperparameters such as temperature, top_p, and top_k. These parameters, when used in conjunction, could yield a wide array of configurations. While the top_p=95% strategy showed the least divergence from the source data, it was unable to accurately replicate performance metrics for CAD CABG. As an interim solution, we may publish multiple versions of synthetic data, along with their corresponding performance metrics. This approach would allow researchers to select the most suitable dataset for their specific use case. Thirdly, the GPT model showed a propensity to over-represent prevalent concepts, skewing towards those with higher frequencies in the dataset. An illustration of this is seen in the synthetic data, where 78% of patients had at least one outpatient visit, in contrast to the actual data where this figure was 73%. This discrepancy indicates a bias in the model towards more common occurrences. In addition, the co-occurrence analysis also demonstrated that the synthetic reconstruction faithfully represents the frequent concept pairs in the original data but may be less effective at recovering the finer associations between rare concepts as shown in Figure 7. To address the over- representation of prevalent concepts by the GPT model, future work will look into implementing regularization techniques. One promising approach is adaptive regularization, which can be outlined in several steps: 1) Model Training: Begin by training the GPT model for a predetermined number of steps; 2) Sequence Generation and Analysis: Generate a sample of patient sequences from the trained model and calculate the marginal distribution of the concepts within these sequences; 3) Distribution Comparison and Adjustment Score Calculation: Compare the model-generated distribution to the empirical distribution derived from the actual data. From this comparison, calculate an adjustment score for each concept; 4) Logit Adjustment: Modify the logits for each concept in the model according to the calculated adjustment scores. By implementing this procedure, the influence of each concept on the model’s learning process would be adaptively modified during back-propagation, allowing for an update in the model parameters that takes into account the disparity between the generated and actual data distributions. This should help in reducing the bias towards over-represented concepts. Furthermore, improving patient representation is an area for further development. Currently, the model’s representation is limited to daily intervals and does not capture more precise measurements like hours or minutes. This limitation is particularly relevant for intensive care unit (ICU) data where time-sensitive decisions are critical. Furthermore, the current model framework assigns the first visit of every synthetic patient to the start of the year, as it only includes a year token to denote the commencement of patient history. To refine the accuracy of patient history commencement, integrating a month token alongside the year token is being considered to properly represent seasonality. This would provide a more accurate reconstruction of the starting point for a patient’s first visit in the generated data. Lastly, there is a necessity to incorporate the death event within patient sequences. Including this event would allow the synthetic data to more accurately represent mortality, enhancing its utility for predictions related to patient outcomes and lifespan. These enhancements aim to create a more precise and clinically relevant synthetic dataset that better mirrors the complexities of real-world patient trajectories. ## 6 Conclusion To our knowledge, this is the first attempt to utilize GPT for generating time-series EHR data. Our main contribution lies in the design of a patient representation that captures temporal dependencies among token types, enabling GPT to generate realistic patient sequences. Moreover, this representation allows for easy conversion back to the OMOP format. Comprehensive evaluations showed that the synthetic data effectively captures the intricate patterns present in EHR data. ## References * [1] Murray, R. E., Ryan, P. B. & Reisinger, S. J. Design and validation of a data simulation model for longitudinal healthcare data. In _AMIA Annual Symposium Proceedings_ , vol. 2011, 1176 (American Medical Informatics Association, 2011). * [2] Ghosheh, G., Li, J. & Zhu, T. A review of generative adversarial networks for electronic health records: applications, evaluation measures and data sources (2022). URL https://arxiv.org/abs/2203.07018v2. * [3] Choi, E. _et al._ Generating multi-label discrete patient records using generative adversarial networks. In Doshi-Velez, F. _et al._ (eds.) _Proceedings of the 2nd Machine Learning for Healthcare Conference_ , vol. 68 of _Proceedings of Machine Learning Research_ , 286–305 (PMLR, 2017). URL https://proceedings.mlr.press/v68/choi17a.html. * [4] Li, Y., Bengio, S. & Du, N. Time-dependent representation for neural event sequence prediction. _6th International Conference on Learning Representations, ICLR 2018 - Workshop Track Proceedings_ (2017). URL https://arxiv.org/abs/1708.00065v4. * [5] Li, J., Cairns, B. J., Li, J. & Zhu, T. Generating synthetic mixed-type longitudinal electronic health records for artificial intelligent applications. _npj Digital Medicine 2023 6:1_ 6, 1–18 (2023). URL https://www.nature.com/articles/s41746-023-00834-7. * [6] Lee, D. _et al._ Generating sequential electronic health records using dual adversarial autoencoder. _Journal of the American Medical Informatics Association : JAMIA_ 27, 1411–1419 (2020). URL https://pubmed.ncbi.nlm.nih.gov/32989459/. * [7] Hripcsak, G. & Albers, D. J. Next-generation phenotyping of electronic health records. _Journal of the American Medical Informatics Association : JAMIA_ 20, 117–121 (2013). URL https://pubmed.ncbi.nlm.nih.gov/22955496/. * [8] Openai, A. R., Openai, K. N., Openai, T. S. & Openai, I. S. Improving language understanding by generative pre-training URL https://gluebenchmark.com/leaderboard. * [9] Vaswani, A. _et al._ Attention is all you need . * [10] Goodfellow, I. _et al._ Generative adversarial networks. _Communications of the ACM_ 63, 139–144 (2014). URL https://arxiv.org/abs/1406.2661v1. * [11] Zhang, Z., Yan, C., Lasko, T. A., Sun, J. & Malin, B. A. Synteg: a framework for temporal structured electronic health data simulation. _Journal of the American Medical Informatics Association_ 28, 596–604 (2021). URL https://dx.doi.org/10.1093/jamia/ocaa262. * [12] Gulrajani, I., Ahmed, F., Arjovsky, M., Dumoulin, V. & Courville, A. Improved training of wasserstein gans. _Advances in Neural Information Processing Systems_ 2017-Decem, 5768–5778 (2017). URL https://arxiv.org/abs/1704.00028v3. * [13] Pang, C. _et al._ Cehr-bert: Incorporating temporal information from structured ehr data to improve prediction tasks. In Roy, S. _et al._ (eds.) _Proceedings of Machine Learning for Health_ , vol. 158 of _Proceedings of Machine Learning Research_ , 239–260 (PMLR, 2021). URL https://proceedings.mlr.press/v158/pang21a.html. * [14] Li, Y. _et al._ Behrt: Transformer for electronic health records. _Scientific Reports 2020 10:1_ 10, 1–12 (2020). URL https://www.nature.com/articles/s41598-020-62922-y. * [15] Rasmy, L., Xiang, Y., Xie, Z., Tao, C. & Zhi, D. Med-bert: pretrained contextualized embeddings on large-scale structured electronic health records for disease prediction. _npj Digital Medicine 2021 4:1_ 4, 1–13 (2021). URL https://www.nature.com/articles/s41746-021-00455-y. * [16] Kraljevic, Z. _et al._ Foresight – generative pretrained transformer (gpt) for modelling of patient timelines using ehrs (2022). URL https://arxiv.org/abs/2212.08072v2. * [17] Yan, C. _et al._ A multifaceted benchmarking of synthetic electronic health record generation models. _Nature Communications_ 13 (2022). URL https://doi.org/10.1038%2Fs41467-022-35295-1. * [18] OHDSI. _The Book of OHDSI: Observational Health Data Sciences and Informatics_ (OHDSI, 2019). URL https://books.google.com/books?id=JxpnzQEACAAJ. * [19] Theodorou, B., Xiao, C. & Sun, J. Synthesize high-dimensional longitudinal electronic health records via hierarchical autoregressive language model. _Nature Communications 2023 14:1_ 14, 1–13 (2023). URL https://www.nature.com/articles/s41467-023-41093-0. * [20] Murphy, S. N. _et al._ Serving the enterprise and beyond with informatics for integrating biology and the bedside (i2b2). _Journal of the American Medical Informatics Association_ 17, 124–130 (2010). ## 7 Supplementary Materials Figure 9: The comparison of the patient representations. Cohort | Real | GPT-OUTPAT | CEHR-BERT ---|---|---|--- HF readmission | Pre = 25.7 AUC = 65.7 PR = 39.3 | Pre =100.0 AUC = NA PR = NA | Pre =100.0 AUC = NA PR = NA Hospitalization | Pre = 5.6 AUC = 75.3 PR = 19.5 | Pre = 5.3 AUC = 70.2 PR = 14.3 | Pre = 0.3 AUC = 76.0 PR = 1.0 COPD readmission | Pre = 34.5 AUC = 74.2 PR = 83.8 | Pre = NA AUC = NA PR = NA | Pre = NA AUC = NA PR = NA Afib ischemic stroke | Pre = 8.7 AUC = 84.0 PR = 48.5 | Pre = 9.7 AUC = 67.2 PR = 27.2 | Pre = 19.5 AUC = 72.4 PR = 60.0 CAD CABG | Pre = 7.1 AUC = 88.4 PR = 55.9 | Pre = 3.5 AUC = 81.5 PR = 44.4 | Pre = 5.4 AUC = 77.2 PR = 44.1 Table 7: Logistic regression (LR) model performance across baseline synthetic datasets. In each cell, three numbers were reported including the prevalence of the positive cases, ROC-AUC, and PR-AUC. If the prevalence=NA, this indicates that $0$ patients were identified in the cohort. AUC=NA and PR=NA indicate the LR model could not be run successfully due to either 100% or 0% prevalence. Figure 10: The visit prevalence stratified by age group Figure 11: The visit prevalence stratified by race Figure 12: The visit prevalence stratified by gender
Ronald A. Remmerswaal Arthur E.P. Veldman Bernoulli Institute, University of Groningen PO Box 407, 9700 AK Groningen, The Netherlands two-phase flow volume of fluid method parabolic reconstruction This work is part of the research programme SLING, which is (partly) financed by the Netherlands Organisation for Scientific Research (NWO).
# Formation of moons and equatorial ridge around top-shaped asteroids after surface landslide Ryuki Hyodo ISAS/JAXA, Sagamihara, Kanagawa, Japan<EMAIL_ADDRESS>Keisuke Sugiura Earth-Life Science Institute, Tokyo Institute of Technology, Meguro-ku, Tokyo 152-8550, Japan ###### Abstract Top-shaped asteroids have been observed among near-Earth asteroids. About half of them are reported to have moons (on the order of $\sim 1$wt.% of the top- shaped primary) and many of them have an equatorial ridge. A recent study has shown that the enigmatic top-shaped figure of asteroids (e.g., Ryugu, Bennu, and Didymos) could result from an axisymmetric landslide of the primary during a fast spin-up near the breakup rotation period. Such a landslide would inevitably form a particulate disk around an asteroid with a short timescale ($\sim 3$ hours). However, the long-term full dynamical evolution is not investigated. Here, we perform a continuous simulation ($\sim 700$ hours) that investigates the sequence of events from the surface landslide that forms a top-shaped asteroid and a particulate disk to disk evolution. We show that the disk quickly spreads and produces moons (within $\sim 300$ hours). The mass of the formed moon is consistent with what is observed around the top-shaped asteroids. We also demonstrate that an equatorial ridge is naturally formed because a fraction of the disk particles re-accretes selectively onto the equatorial region of the primary. We envision that Ryugu and Bennu could once have an ancient moon that was later lost due to a successive moon’s orbital evolution. Alternatively, at the top-shaped asteroid that has a moon, such as Didymos, no significant orbital evolution of the moon has occurred that would result in its loss. Our study would also be qualitatively applicable to any rubble-pile asteroids near the breakup rotation period. Asteroids (72); Near-Earth objects (1092) ## 1 Introduction Figure 1: Schematic summary of our paper. Panel (a): a rubble-pile asteroid spins up due to, for example, the YORP effect, small impacts, a close encounter with a planet, or re-accumulation after a catastrophic impact. Panel (b): a surface landslide occurs when a critical spin-state is realized and a top-shaped figure is formed (Sugiura et al., 2021). Panel (c): a particulate disk spreads due to inelastic collisions and gravitational interactions among particles. Panel (d): a moon is gravitationally accreted outside the Roche limit of the central top-shaped body and an axisymmetric equatorial ridge is formed due to the re-accretion of disk particles. Panels (e) and (f): the formed moon is lost or remains, depending on the long-term orbital evolution between the moon and the primary. More and more growing interests are on near-Earth asteroids (NEAs) together with recent activities of asteroid exploration missions. Top-shaped asteroids may exist ubiquitously among NEAs (Walsh & Jacobson, 2015; Margot et al., 2015). These include asteroid 162173 Ryugu (Watanabe et al., 2019), 101955 Bennu (Lauretta et al., 2019), 65803 Didymos, 1999 KW4 (Ostro et al., 2006), and 2001 SN263 (Becker et al., 2015). Their diameters are less than several kilometers, and recent in-situ observations by spacecrafts showed that they might be rubble-pile bodies (e.g., Watanabe et al., 2019). Because of their small size, Yarkovsky-O’Keefe-Radzievskii-Paddack (YORP) effect (Rubincam, 2000), small impacts (Takeda & Ohtsuki, 2009), and/or a close encounter with a planet (Hyodo et al., 2016), for example, could efficiently change their spin states. The acceleration rate of spin-up (or spin-down) depends on each physical process (e.g., roughly instantaneous by an impact and $\gtrsim 100$ Kyr by YORP effect on a kilometer-sized body). Interestingly, about half of the reported top-shaped asteroids have moons around them (e.g., 2001 NE263; Becker et al. 2015, 1999 KW4; Ostro et al. 2006, 1994 CC; Brozović et al. 2011, Didymos; Naidu et al. 2020). The masses of the moons are roughly on the order of $\sim 1$wt.% of the host top-shaped asteroids. Furthermore, many of the top-shaped asteroids are reported to have an equatorial ridge (Benner et al., 2015). There are several proposed mechanisms to form a top-shaped asteroid. The top- shaped figure may be formed during re-accretion after catastrophic disruption of a parent body (Michel et al., 2020). Alternatively, a mass movement with reshaping or a landslide of surface materials of a rubble-pile asteroid due to fast spin-up may form the top-shaped figure (Walsh et al., 2008, 2012; Harris et al., 2009; Hirabayashi et al., 2020; Sugiura et al., 2021). Recently, Sugiura et al. (2021) used the Smoothed Particle Hydrodynamics (SPH) method to study rotational deformation during spin-up, which has also been used to model the shape deformation due to impacts (Jutzi & Asphaug, 2015). In their study, they included effective bulk friction of rubble pile bodies as a parameter (i.e., effective friction angle $\phi_{\rm fri}$), which could effectively include, for example, the effect of cohesion (see Sec. 2.2 in Sugiura et al. 2021). They demonstrated that, for $\phi_{\rm fri}\geq 70$ deg with a spin-up timescale of $\lesssim$ a few days (defined as the elapsed time of spin-up from the rotation period of 3.5 h to 3.0 h), an axisymmetric set of surface landslides would occur and top-shaped rubble-pile bodies would be correspondingly formed (panels (a)-(b) in Fig. 1). The story of the landslide hypothesis, however, should not have ended here, although Sugiura et al. (2021) did not study the fate of the surface materials of the landslide. In this study, we first show that the surface materials are distributed around the newly formed top-shaped central body, forming a transient particle disk (panel (c) in Fig. 1). Here, the disk mass is typically found to be about $\sim 10$wt.% of the top-shaped primary body. In this study, following the numerical approach used in Sugiura et al. (2021), we investigate a full long-term dynamical evolution – from the landslide to the end of the disk evolution – to understand the fate of the surface materials that are distributed around the newly formed top-shaped body (Fig. 1). We show that moons and an axisymmetric equatorial ridge are naturally and inevitably formed around the top-shaped primary (panel (d) in Fig. 1). In Sect. 2, we describe our numerical models. In Sect. 3, we show our numerical results. Section 4 discusses the long-term orbital evolution between a moon and the central primary due to the binary YORP (BYORP) effect, which may lead to diverse systems of the top-shaped asteroids with and without a companion moon(s) (panel (e) and panel (f) in Fig. 1). Finally, Section 5 summarizes this paper. ## 2 Numerical method Our SPH simulations are the same as those in Sugiura et al. (2021), and we refer the readers to their Sec. 2 for more details. Our simulations solve hydrodynamic equations (the equations of continuity and motion) with self- gravity and yield processes with the aid of a friction model that determines the shear strength of granular material (Jutzi, 2015). In a similar context, a Discrete Element Methods (DEMs) with cohesionless particles were used in Walsh et al. (2008). Shape deformation using this approach could be numerically affected due to the inconsistency between the realistic particle size and that used in their simulations. In the SPH approach, in contrast, a rubble-pile body was constructed by a continuum of granular material and we explicitly set a specific value of the friction angle of the material. We expect that the SPH method has the advantage of precise control of bulk friction of rubble piles (see also Jutzi & Asphaug, 2015). We also note that there is a difference in the process of the moon formation between their study and ours; in Walsh et al. (2008), the moon formation occurred via a step-by-step accumulation of discretely ejected particles, while, in our study, the moon formation occurs because of particle disk evolution. The rubble-pile body initially had a uniform and spherical body with a radius of $R_{\rm ini}=500$ m, which was then numerically spun up with the angular acceleration of $8.954\times 10^{-10}$ rad s-2. We stopped the acceleration when 1wt.% of the primary is ejected; this prevents further artificial deformation after the landslide, and the arbitrary choice of 1wt.% does not affect our results as long as the landslide occurs with a shorter timescale (here $\sim 3$ h) than the spin-up timescale (here $\sim 30$ h). The density of particles at uncompressed states is $\rho_{\rm ini}=1.19$ g cm-3. Particle mass is given as $m=M_{\rm tot}/N_{\rm tot}$, where $M_{\rm tot}\equiv(4/3)\pi\rho_{\rm ini}R_{\rm ini}^{3}$ and $N_{\rm tot}=25470$ is the total number of the SPH particles. In this study, following Sugiura et al. (2021), we specifically focused on a set of parameters (i.e., $\phi_{\rm fri}=80$ deg; see panels (k)-(o) in their Fig. 1), which successfully resulted in the formation of an axisymmetric top- shaped figure as seen at asteroids Ryugu, Bennu, and Didymos. Lowering $\phi_{\rm fri}$, for example, resulted in an elongated lemon-shaped primary via internal deformation (see panels (a)-(e) and (f)-(j) in Fig. 1 of Sugiura et al. 2021), which eventually led to distribute particles as disks with a variety of initial disk mass. The dependences on the disk parameters (e.g., initial disk mass) and tidal parameters are studied in detail using an independent numerical approach (1D fluid simulations) in our companion paper (Madeira et al. submitted). Here, we used the direct and continuous approach – from the landslide to the end of the disk evolution – and focused on studying the details regarding the relevant consequences of the top-shape formation. Below, we show that the surface materials of the primary are distributed in a disk-like structure and such a particulate disk eventually forms moon(s) around the top-shaped primary. We numerically distinguished member particles of the primary as follows. We start from a randomly chosen particle and then iteratively detect any particles within a critical distance in a bottom-up fashion (Hyodo & Genda, 2018). This procedure continues until no more new particles are detected. As a critical distance, we used 1.5 times the smoothing length of an SPH particle to be consistent with Sugiura et al. (2021), although this specific choice does not affect our results. In this study, we continued our simulation up to $t\sim 25\times 10^{5}$ s ($\sim 700$ hours). This is much after the landslide had occurred; Sugiura et al. (2021) stopped their simulations at $t\sim 1\times 10^{5}$ s ($\sim 30$ hours) when the top-shaped figure was formed. We may need to be careful about the angular momentum (AM) conservation in SPH simulations, particularly for a rotating system (such as disk evolution). In our calculations, after an artificial spin-up of the primary (i.e., artificial AM increase in the system) was stopped, the cumulative AM error was $\sim 2.6$% during and after the disk formation until the end of the simulation (i.e., from $t\sim 1.5\times 10^{5}$ s to $t\sim 25\times 10^{5}$ s). Although our SPH simulations contain some error in AM conservation, independent numerical simulations of particle disk evolutions of the Earth’s Moon formation (e.g., $N$-body simulations; e.g., Kokubo et al. 2000 and 1D semi-analytical fluid simulations; e.g., Salmon & Canup 2012) used similar initial disk conditions (e.g., in terms of disk to primary mass ratio) and their results (e.g., resultant moon to primary mass ratio) were consistent with those obtained here. This is because the gravitational disk evolution can be well characterized by the mass ratio between the disk and the primary (Kokubo et al., 2000; Salmon & Canup, 2012). Furthermore, using $N$-body simulations, Hyodo et al. (2015) showed a scaling law for the mass of the largest moon formed via disk spreading as a function of the initial disk mass (see their Fig. 12) and it is generally consistent with that obtained in this study. We note that, although these studies generally assumed particulate disks, a more realistic Moon-forming disk would be vapor-rich and thus the Earth’s Moon formation process would be more complex (e.g., Thompson & Stevenson, 1988; Nakajima & Stevenson, 2014; Lock et al., 2018) Figure 2: Overall disk evolution after the landslide. Gray particles represent member particles of the top-shaped primary. Dark-gray particles represent disk particles and member particles of gravitational clumps (including the formed moon at later epochs). The black circle indicates the Roche limit of the primary ($r_{\rm R}\sim 2.5R_{\rm c}$ where $R_{\rm c}=500$m is used). Figure 3: Same as Fig. 2, but edge on views. ## 3 Numerical results: moon formation Figure 4: Evolutions of the disk mass (left panel; not including the mass of the largest moons) and the masses of the largest three moons (right panel). At around $t\sim 1\times 10^{5}$s with a time-interval of $\Delta t\sim 10^{4}$s, the landslide occurs and a particle disk is formed (left panel). The disk spreads and starts to form gravitational clumps with masses larger than $\sim 0.001M_{\rm c}$ at $t\sim 5.5\times 10^{5}$s, where $M_{\rm c}$ is the mass of the central primary. Two large moons are formed by the end of our simulations ($t=25\times 10^{5}$s; $\sim 700$h). Figures 2 and 3 show an overall evolution of a particle disk including the landslide. The landslide (i.e., disk formation) occurs with a very short timescale ($\sim 10^{4}$ s $\simeq 3$ hours) compared to the disk evolution timescale ($\sim 10^{6}$ $\simeq 300$ hours; see Fig. 4). Because of the landslide, a top-shaped figure of the primary is formed (see more details in Sugiura et al. 2021 and their Fig.1). The surface materials of the primary are distributed around the primary, forming a particle disk. The initial mass of the disk is $M_{\rm disk}\sim 20$wt.% of the top-shaped primary (Fig. 4). Most of the disk mass is initially within the Roche limit of the primary, defined as $r_{\rm R}\equiv 2.456(\rho_{\rm m}/\rho_{\rm c})^{-1/3}R_{\rm c}$ where $\rho_{\rm m}$ and $\rho_{\rm c}$ are densities of the moon and the central body. $R_{\rm c}$ is the radius of the central body. The disk mass is large enough to become gravitationally unstable and the spiral arm structures are formed as a result of gravitational instability. This leads to an efficient angular momentum transfer and the outer disk spreads further radially outward (Takeda & Ida, 2001), distributing the disk materials beyond the Roche limit. Because of the angular momentum conservation, the inner part of the disk spreads inward, resulting in re- accretion onto the primary. The materials scattered beyond the Roche limit start to coagulate via their own gravity, forming gravitational clumps (Figs. 2 and 3). With time, accretion among clumps proceeds, forming larger gravitational bodies (see also the right panel of Fig. 4 for the mass evolution of the largest objects). After $\sim 300$ hours since the landslide, large moons around the top-shaped primary asteroid are formed. The total mass of the formed moons is $\sim 4$wt.% of the primary. Here, two large moons are formed (Figs. 2, 3, and 4). The number of the large moon is the result of the stochastic nature of the accretion processes that can be seen in the 3D simulations, especially when the disk is massive to the central body; the details of the stochastic nature were reported in independent $N$-body simulations (Hyodo et al., 2015). This means that, if one runs another simulation with the same parameters but with slightly different initial positions of particles, the final outcome could be the formation of a single moon. Thus, the number of large moons is not a decisive parameter that characterizes the general outcome. More importantly, instead, regardless of the number of moons, the total mass of the moons rarely changes and it characterizes the outcome of specific choices of the disk parameters (e.g., initial disk mass; see Ida et al. 1997; Kokubo et al. 2000). ## 4 Discussion ### 4.1 Formation of equatorial ridge Figure 5: Comparison of asteroid shapes between the initial figure ($t=0$h; left panel) and after the disk evolution ($t\sim 700$h; right panel). A black circle indicates a circle with a radius of 500 meters. Our numerical simulations show another interesting feature as a result of disk evolution around a top-shaped asteroid. Because the disk spreads both inward and outward to conserve the angular momentum, the disk particles initially at the inner region re-accrete onto the top-shaped primary. This has led to the formation of an equatorial ridge. Other potential physical mechanisms to form an equatorial ridge are discussed in other literature. These are, for example, rotational reshaping and mass movement (e.g., Walsh et al., 2012; Hirabayashi & Scheeres, 2015; Hirabayashi et al., 2020), re-accumulation of debris after a catastrophic disruption of a parent body (Michel et al., 2020), or selective accumulation of ejecta of small impacts at rapidly rotating asteroid (Ikeya & Hirata, 2021). Figure 5 shows an edge on views of the central rubble-pile body at $t=0$h (left panel) and at the end of our simulation without disk particles and formed moons ($t\sim 700$h; right panel). A significant shape change can be seen from the initial spherical shape to a top-shaped figure with a prominent axisymmetric equatorial ridge. This is a direct consequence of the re- accretion of equatorial disk particles onto the top-shaped primary. Many of the top-shaped asteroids and relatively spheroidal bodies in the near- Earth region are reported to have an equatorial ridge (Benner et al., 2015). JAXA’s Hayabusa2 mission revealed that asteroid Ryugu is a top-shaped asteroid with an equatorial ridge (Watanabe et al., 2019). NASA’s OSIRIS-REx mission also reported that asteroid Bennu has an equatorial ridge (Walsh et al., 2019). Another top-shaped asteroid, Didymos, has also an equatorial ridge as well as a moon, Dimorphos (Naidu et al., 2020). We importantly note that the axis ratio of $c/a$ and the equatorial ridge seen in Fig. 5 are smaller and much more prominent than the observations of, for example, Ryugu, Bennu, and Didymos (Watanabe et al., 2019; Lauretta et al., 2019; Naidu et al., 2020). Our SPH simulations would be limited in their ability to precisely demonstrate the detailed accumulation processes of particles in the equatorial ridge; because our SPH simulations assumed and employed, for example, the same size and simple physical properties among all particles. We also emphasize the importance of studying the long-term geological evolution (e.g., including degradation) of the ridge. Continuous micrometeoroid impacts and/or thermal fatigue would also change the ridge shape. Therefore, further studies of the detailed ridge formation process as well as the post-formation geological processes need to be done to fully validate our ridge formation scenario via the landslides followed by the disk evolution. We leave these points to later work. ### 4.2 On the diversity of top-shaped asteroids Some top-shaped asteroids, e.g., Didymos, are found to have a companion moon(s), while others do not (e.g., Ryugu and Bennu). In the discussion below, we focus on asteroids that have (1) top-shaped figures and (2) equatorial ridges, as background considerations. We, then, additionally consider the existence and non-existence of a companion moon around the top-shaped primary. Our results together with those of Sugiura et al. (2021) indicate that small rubble-pile asteroids may inevitably experience a landslide due to spin-up by the YORP effect or by other physical processes (e.g., re-accumulation, small impact, and/or a close encounter with a planet). Depending on the acceleration rate of spin-up as well as the effective friction angle, $\phi_{\rm fri}$, of constituent particles of an asteroid, the resultant shape of asteroids via a landslide would change (e.g., a lemon-shape or a top-shape; see Fig.1 of Sugiura et al. 2021). When the top-shaped primary was formed with the fast spin-up rate and $\phi_{\rm fri}\gtrsim 70$ deg, Sugiura et al. (2021) numerically demonstrated that about $\sim 10$wt.% of the primary is generally ejected. When the initial disk mass is $\sim 10$wt.% or larger, the mass of the moon would be linearly scaled with the disk mass (Ida et al., 1997; Kokubo et al., 2000), and thus a change in the resultant disk mass within the same order of magnitude does not significantly change the mass of the moon (would be on the order of $\sim 1$wt.% of the primary as observed for the top-shaped binary asteroids). In contrast, when the effective friction angle is smaller than the above value, the shape deformation became more gradual, and a lemon-shaped primary was formed. In this case, the ejected mass tends to be smaller (e.g., $\sim 3$wt.% and $\sim 5$wt.% for $\phi_{\rm fri}=40$ deg and $\phi_{\rm fri}=60$ deg, respectively), although more detailed studies on throughout parameters may be needed (see Fig. 1 of Sugiura et al., 2021). In these cases of smaller initial disk masses, the mass of the resultant moon depends more strongly on the disk mass and much smaller moons can be produced from the disk (Hyodo et al., 2015). Importantly, as a natural consequence of a landslide, a particle disk is formed around the primary. Once a particle disk exists, the disk inevitably spreads and forms moon(s) as long as the disk is massive enough. The final mass and orbital configurations of the moon systems formed through the disk spreading depend on the initial disk mass and tidal parameters. Such dependencies are studied using 1D fluid simulations in our companion paper in the context of the Didymos-Dimorphos system formation (Madeira et al. submitted). In short, the direct consequence of the landslide would be the system of a moon(s) around a central asteroid – regardless of the top-shaped or the lemon-shaped central primary – with an equatorial ridge on the primary. Then, the question now is can we remove the formed moon if the top-shaped figures of Ryugu and/or Benuu are formed via landslides? This is because these top-shaped asteroids today do not have a moon around them. Ejection via tidal evolution would not be promising because its efficiency significantly decreases as the distance between the primary and the moon becomes larger. A potential dominant dynamical mechanism on the binary system may be the binary YORP (BYORP) effect (e.g., Ćuk & Burns, 2005; Ćuk & Nesvorný, 2010; McMahon & Scheeres, 2010; Jacobson & Scheeres, 2011). Although the timescale of orbital separation to eject a companion moon via the BYORP effect can be as small as $\sim 10^{5}$ years (see Sec.4.3 in Sugiura et al. 2021 and Ćuk & Burns, 2005), its timescale and direction of the orbital separation (shrinking or expanding) strongly depend on the details of the surface properties of the asteroids and moons (Ćuk & Burns, 2005; McMahon & Scheeres, 2010). We envision that such a complex dependence on the shapes and surface properties can potentially lead to a diversity of the top-shaped asteroids with and without a companion moon. The moons can be ejected by the BYORP effect in some cases, while the dynamical evolution of the orbital separation may not be efficient in other cases, remaining the system as binary asteroids (Figure 1). ## 5 Summary In this study, using the SPH simulations, we studied a continuous dynamical sequence of events from the surface landslide that forms a top-shaped asteroid and a particulate disk to disk evolution (Figure 1). We numerically demonstrated that the particle disk formed by a surface landslide would quickly spread and produce moons just outside the Roche limit of the top- shaped primary (within $\sim 300$ hours). The mass of the moon is consistent with what is observed around the top-shaped asteroids (on the order of $\sim 1$wt.% of the primary). We also demonstrated that an equatorial ridge would be naturally formed because a fraction of the disk particles re-accrete selectively on the equatorial region of the primary. Tidal interaction as well as the binary YORP (BYORP) effect between moons and the primary would change the orbital separation between them. The timescale and direction of the orbital separation (shrinking or expanding) strongly depend on the details of the surface properties of the moons and the primary. This indicates that a long-term orbital evolution could produce diverse moon systems around the top-shaped asteroids. We envision that top-shaped Ryugu and Bennu could once have an ancient moon that was later lost as a result of a successive moon’s orbital evolution. Alternatively, other top-shaped asteroids today, such as Didymos, have a companion moon. In these cases, no significant orbital evolution of the moon may have occurred that would result in its loss. Our study focused on the top-shaped asteroids. However, our results of the consequences of a surface landslide – the moon formation and the equatorial ridge formation – can be qualitatively (but not always quantitatively) applied to any small rubble-pile asteroids near the breakup rotation period. Indeed, Sugiura et al. (2021) showed that changing the effective friction angle and/or the spin-up rate results in different patterns of the deformation mode with a variety of disk formation (see their Fig. 1). Once a particle disk is formed, the moon and the equatorial ridge would be naturally and inevitably formed. The long-term evolution, then, would lead to a diverse asteroid system (see also Ćuk, 2007; Jacobson & Scheeres, 2011; Ćuk et al., 2021). Further development of the theoretical and modeling research, especially on the BYORP effect, together with a better understanding of the surface and particle properties of individual asteroids would be needed to further constrain and validate the results presented in this study. Further studies on the accumulation and post-formation geological processes of the equatorial ridge are also demanded. Analysis of the return samples by JAXA’s Hayabusa2 and NASA’s OSIRIS-REx as well as data that would be obtained by NASA’s DART and ESA’s HERA missions would help us to better understand the nature and evolution of the top-shaped asteroids. R.H. acknowledges the financial support of MEXT/JSPS KAKENHI (Grant Number JP22K14091). R.H. also acknowledges JAXA’s International Top Young program. K.S. acknowledges the financial support of JSPS, Japan KAKENHI Grant (JP20K14536, JP20J01165). Numerical simulations in this work were carried out on the Cray XC50 supercomputer at the Center for Computational Astrophysics, National Astronomical Observatory of Japan. ## References * Becker et al. (2015) Becker, T. M., Howell, E. S., Nolan, M. C., et al. 2015, Icarus, 248, 499, doi: 10.1016/j.icarus.2014.10.048 * Benner et al. (2015) Benner, L. A. M., Busch, M. W., Giorgini, J. D., Taylor, P. A., & Margot, J. L. 2015, in Asteroids IV, 165–182, doi: 10.2458/azu_uapress_9780816532131-ch009 * Brozović et al. (2011) Brozović, M., Benner, L. A. M., Taylor, P. A., et al. 2011, Icarus, 216, 241, doi: 10.1016/j.icarus.2011.09.002 * Ćuk (2007) Ćuk, M. 2007, ApJ, 659, L57, doi: 10.1086/516572 * Ćuk & Burns (2005) Ćuk, M., & Burns, J. A. 2005, Icarus, 176, 418, doi: 10.1016/j.icarus.2005.02.001 * Ćuk et al. (2021) Ćuk, M., Jacobson, S. A., & Walsh, K. J. 2021, , 2, 231, doi: 10.3847/PSJ/ac3093 * Ćuk & Nesvorný (2010) Ćuk, M., & Nesvorný, D. 2010, Icarus, 207, 732, doi: 10.1016/j.icarus.2009.12.005 * Harris et al. (2009) Harris, A. W., Fahnestock, E. G., & Pravec, P. 2009, Icarus, 199, 310, doi: 10.1016/j.icarus.2008.09.012 * Hirabayashi & Scheeres (2015) Hirabayashi, M., & Scheeres, D. J. 2015, ApJ, 798, L8, doi: 10.1088/2041-8205/798/1/L8 * Hirabayashi et al. (2020) Hirabayashi, M., Nakano, R., Tatsumi, E., et al. 2020, Icarus, 352, 113946, doi: 10.1016/j.icarus.2020.113946 * Hyodo et al. (2016) Hyodo, R., Charnoz, S., Genda, H., & Ohtsuki, K. 2016, ApJ, 828, L8, doi: 10.3847/2041-8205/828/1/L8 * Hyodo & Genda (2018) Hyodo, R., & Genda, H. 2018, ApJ, 856, L36, doi: 10.3847/2041-8213/aab7f0 * Hyodo et al. (2015) Hyodo, R., Ohtsuki, K., & Takeda, T. 2015, ApJ, 799, 40, doi: 10.1088/0004-637X/799/1/40 * Ida et al. (1997) Ida, S., Canup, R. M., & Stewart, G. R. 1997, Nature, 389, 353, doi: 10.1038/38669 * Ikeya & Hirata (2021) Ikeya, R., & Hirata, N. 2021, Icarus, 367, 114590, doi: 10.1016/j.icarus.2021.114590 * Jacobson & Scheeres (2011) Jacobson, S. A., & Scheeres, D. J. 2011, ApJ, 736, L19, doi: 10.1088/2041-8205/736/1/L19 * Jutzi (2015) Jutzi, M. 2015, Planet. Space Sci., 107, 3, doi: 10.1016/j.pss.2014.09.012 * Jutzi & Asphaug (2015) Jutzi, M., & Asphaug, E. 2015, Science, 348, 1355, doi: 10.1126/science.aaa4747 * Kokubo et al. (2000) Kokubo, E., Ida, S., & Makino, J. 2000, Icarus, 148, 419, doi: 10.1006/icar.2000.6496 * Lauretta et al. (2019) Lauretta, D. S., Dellagiustina, D. N., Bennett, C. A., et al. 2019, Nature, 568, 55, doi: 10.1038/s41586-019-1033-6 * Lock et al. (2018) Lock, S. J., Stewart, S. T., Petaev, M. I., et al. 2018, Journal of Geophysical Research (Planets), 123, 910, doi: 10.1002/2017JE005333 * Margot et al. (2015) Margot, J. L., Pravec, P., Taylor, P., Carry, B., & Jacobson, S. 2015, in Asteroids IV, 355–374, doi: 10.2458/azu_uapress_9780816532131-ch019 * McMahon & Scheeres (2010) McMahon, J., & Scheeres, D. 2010, Celestial Mechanics and Dynamical Astronomy, 106, 261, doi: 10.1007/s10569-009-9247-9 * Michel et al. (2020) Michel, P., Ballouz, R. L., Barnouin, O. S., et al. 2020, Nature Communications, 11, 2655, doi: 10.1038/s41467-020-16433-z * Naidu et al. (2020) Naidu, S. P., Benner, L. A. M., Brozovic, M., et al. 2020, Icarus, 348, 113777, doi: 10.1016/j.icarus.2020.113777 * Nakajima & Stevenson (2014) Nakajima, M., & Stevenson, D. J. 2014, Icarus, 233, 259, doi: 10.1016/j.icarus.2014.01.008 * Ostro et al. (2006) Ostro, S. J., Margot, J.-L., Benner, L. A. M., et al. 2006, Science, 314, 1276, doi: 10.1126/science.1133622 * Rubincam (2000) Rubincam, D. P. 2000, Icarus, 148, 2, doi: 10.1006/icar.2000.6485 * Salmon & Canup (2012) Salmon, J., & Canup, R. M. 2012, ApJ, 760, 83, doi: 10.1088/0004-637X/760/1/83 * Sugiura et al. (2021) Sugiura, K., Kobayashi, H., Watanabe, S.-i., et al. 2021, Icarus, 365, 114505, doi: 10.1016/j.icarus.2021.114505 * Takeda & Ida (2001) Takeda, T., & Ida, S. 2001, ApJ, 560, 514, doi: 10.1086/322406 * Takeda & Ohtsuki (2009) Takeda, T., & Ohtsuki, K. 2009, Icarus, 202, 514, doi: 10.1016/j.icarus.2009.03.001 * Thompson & Stevenson (1988) Thompson, C., & Stevenson, D. J. 1988, ApJ, 333, 452, doi: 10.1086/166760 * Walsh & Jacobson (2015) Walsh, K. J., & Jacobson, S. A. 2015, in Asteroids IV, 375–393, doi: 10.2458/azu_uapress_9780816532131-ch020 * Walsh et al. (2008) Walsh, K. J., Richardson, D. C., & Michel, P. 2008, Nature, 454, 188, doi: 10.1038/nature07078 * Walsh et al. (2012) —. 2012, Icarus, 220, 514, doi: 10.1016/j.icarus.2012.04.029 * Walsh et al. (2019) Walsh, K. J., Jawin, E. R., Ballouz, R. L., et al. 2019, Nature Geoscience, 12, 242, doi: 10.1038/s41561-019-0326-6 * Watanabe et al. (2019) Watanabe, S., Hirabayashi, M., Hirata, N., et al. 2019, Science, 364, 268, doi: 10.1126/science.aav8032
# Existence and emergent dynamics of quadratically separable states to the Lohe tensor model Seung-Yeal Ha Department of Mathematical Sciences and Research Institute of Mathematics, Seoul National University, Seoul 08826 and Korea Institute for Advanced Study, Hoegiro 85, Seoul, 02455, Republic of Korea<EMAIL_ADDRESS>, Dohyun Kim School of Mathematics, Statistics and Data Science, Sungshin Women’s University, Seoul 02844, Republic of Korea <EMAIL_ADDRESS>and Hansol Park Department of Mathematical Sciences, Seoul National University, Seoul 08826, Republic of Korea <EMAIL_ADDRESS> ###### Abstract. A tensor is a multi-dimensional array of complex numbers, and the Lohe tensor model is an aggregation model on the space of tensors with the same rank and size. It incorporates previously well-studied aggregation models on the space of low-rank tensors such as the Kuramoto model, Lohe sphere and matrix models as special cases. Due to its structural complexities in cubic interactions for the Lohe tensor model, explicit construction of solutions with specific structures looks daunting. Recently, we obtained completely separable states by associating rank-1 tensors. In this paper, we further investigate another type of solutions, namely “quadratically separable states” consisting of tensor products of matrices and their component rank-2 tensors are solutions to the double matrix model whose emergent dynamics can be studied using the same methodology of the Lohe matrix model. ###### Key words and phrases: Aggregation, double sphere model, gradient flow, Kuramoto model, Lohe matrix model, Lohe tensor model, synchronization ###### 2010 Mathematics Subject Classification: 82C10, 82C22, 35B37 Acknowledgment. The work of S.-Y. Ha is supported by National Research Foundation of Korea (NRF-2020R1A2C3A01003881). ## 1\. Introduction Collective behaviors often appear in large population systems for weakly coupled oscillators or interacting units [3, 6, 7] in diverse scientific disciplines including biology, social sciences, engineering with various space and time scales, for instance, colonies of bacteria [25], school of fish [4, 26], flock of starlings [29], pedestrian dynamics [18], opinion dynamics [17], power grid networks [22], etc. For a brief introduction to collective dynamics, we refer the reader to survey articles and book [1, 5, 10, 24]. Mathematical approach toward the understanding of collective motions has been established in literature by Winfree [28] and Kuramoto [19] in the 1970s, and by Vicsek [27] in the 1990s. After their remarkable works, the aforementioned models have been extended in several directions, particularly in high- dimensional extension to Riemannian manifolds including the hypersurfaces [2, 12] and the matrix Lie group [11] which have attracted lots of interest thanks to its powerful application, for instance, nonconvex optimization. In this work, among high-dimensional models, we are concerned with the Lohe tensor model in [13]. Next, we briefly discuss tensors and an aggregation model on the space of tensors, namely “the Lohe tensor model”. A rank-$m$ complex valued tensor can be represented as a multi-dimensional array of complex numbers with multi- indices. The rank of a tensor is the number of indices, say a rank-$m$ tensor with size $d_{1}\times\cdots\times d_{m}$ is an element of ${\mathbb{C}}^{d_{1}\times\cdots\times d_{m}}$. For example, scalars, vectors and matrices correspond to rank-0, 1 and 2 tensors, respectively. Let $T$ be a rank-$m$ tensor with a size $d_{1}\times\cdots\times d_{m}$. Then, we denote $(\alpha_{1},\cdots,\alpha_{m})$-th component of the tensor $T$ by $[T]_{\alpha_{1}\cdots\alpha_{m}}$, and we set $\overline{T}$ by the rank-$m$ tensor whose components are the complex conjugate of the elements in $T$: $[\overline{T}]_{\alpha_{1}\cdots\alpha_{m}}:=\overline{[T]_{\alpha_{1}\cdots\alpha_{m}}}.$ Let ${\mathcal{T}}_{m}(\mathbb{C};d_{1}\times\cdots\times d_{m})$ be the collection of all rank-$m$ tensors with size $d_{1}\times\cdots\times d_{m}$. Then, it is a complex vector space. Several well-known first-order aggregation models, for instance, the Kuramoto model [19], the swarm sphere model [23] and matrix models [8, 20] can be regarded as aggregation models on ${\mathcal{T}}_{0}(\mathbb{C};1),{\mathcal{T}}_{1}(\mathbb{R};d)$ and ${\mathcal{T}}_{2}(\mathbb{C};d\times d)$, respectively. Let $A_{j}$ be the skew-hermitian rank-$2m$ tensor with size $(d_{1}\times\cdots\times d_{m})\times(d_{1}\times\cdots\times d_{m})$. For simplicity, we introduce handy notation as follows: for $T\in{\mathcal{T}}_{m}(\mathbb{C};d_{1}\times\cdots\times d_{m})$ and $A\in{\mathcal{T}}_{2m}(\mathbb{C};d_{1}\times\cdots\times d_{m}\times d_{1}\times\cdots\times d_{m})$, we set $\displaystyle\begin{aligned} &[T]_{\alpha_{*}}:=[T]_{\alpha_{1}\alpha_{2}\cdots\alpha_{m}},\quad[T]_{\alpha_{*0}}:=[T]_{\alpha_{10}\alpha_{20}\cdots\alpha_{m0}},\quad[T]_{\alpha_{*1}}:=[T]_{\alpha_{11}\alpha_{21}\cdots\alpha_{m1}},\\\ &[T]_{\alpha_{*i_{*}}}:=[T]_{\alpha_{1i_{1}}\alpha_{2i_{2}}\cdots\alpha_{mi_{m}}},\quad[T]_{\alpha_{*(1-i_{*})}}:=[T]_{\alpha_{1(1-i_{1})}\alpha_{2(1-i_{2})}\cdots\alpha_{m(1-i_{m})}},\\\ &[A]_{\alpha_{*}\beta_{*}}:=[A]_{\alpha_{1}\alpha_{2}\cdots\alpha_{m}\beta_{1}\beta_{2}\cdots\beta_{m}}.\end{aligned}$ Then, the Lohe tensor model in a component form reads as follows: $\begin{cases}\displaystyle\dot{[T_{j}]}_{\alpha_{*0}}=[A_{j}]_{\alpha_{*0}\alpha_{*1}}[T_{j}]_{\alpha_{*1}}\\\ \displaystyle\hskip 42.67912pt+\sum_{i_{*}\in\\{0,1\\}^{m}}\kappa_{i_{*}}\Big{(}[T_{c}]_{\alpha_{*i_{*}}}\bar{[T_{j}]}_{\alpha_{*1}}[T_{j}]_{\alpha_{*(1-i_{*})}}-[T_{j}]_{\alpha_{*i_{*}}}\bar{[T_{c}]}_{\alpha_{*1}}[T_{j}]_{\alpha_{*(1-i_{*})}}\Big{)},\\\ \displaystyle\bar{[A_{j}]}_{\alpha_{*0}\alpha_{*1}}=-[A_{j}]_{\alpha_{*1}\alpha_{*0}},\end{cases}$ (1.1) where $\kappa_{i_{*}}$’s are nonnegative coupling strengths. Before we discuss our main issues, we introduce a concept of a “quadratically separable state” for the Lohe tensor model (1.1). ###### Definition 1.1. Let $\\{T_{i}\\}$ be a quadratically separable state to (1.1), if it is decomposed as a tensor product of rank-2 tensors (or matrices): $T_{i}=U_{i}^{1}\otimes U_{i}^{2}\otimes\cdots\otimes U_{i}^{m},\quad U_{i}^{k}\in\mathbb{C}^{d_{1}^{k}\times d_{2}^{k}},\quad\|U_{i}^{k}\|_{\textup{F}}=1,\quad 1\leq i\leq N,\quad 1\leq k\leq m,$ where $\|\cdot\|_{\textup{F}}$ is the Frobenius norm induced by Frobenius inner product: for matrices $A$ and $B$, $\langle A,B\rangle_{\textup{F}}:=\textup{tr}(A^{\dagger}B),\quad\|A\|_{\textup{F}}:=\sqrt{\textup{tr}(A^{\dagger}A)}.$ In this paper, we are interested in the following simple questions: * • (Q1): Are there quadratically separable states for the Lohe tensor model? * • (Q2): If so, do they exhibit collective behaviors under which circumstances? Our main results deal with the raised two questions (Q1) and (Q2). More precisely, our main results of this paper can be summarized as follows. First, we introduce the double matrix model induced from the Lohe tensor model whose elements have rank-4 with a specific condition on natural frequencies $B_{j}$ and $C_{j}$ (see (3.9)): $\displaystyle\begin{cases}\dot{U}_{j}=B_{j}U_{j}+\displaystyle\frac{\kappa_{1}}{N}\sum_{k=1}^{N}\left(\langle V_{j},V_{k}\rangle_{\textup{F}}~{}U_{k}U_{j}^{\dagger}U_{j}-\langle V_{k},V_{j}\rangle_{\textup{F}}~{}U_{j}U_{k}^{\dagger}U_{j}\right)\\\ \hskip 56.9055pt+\displaystyle\frac{\kappa_{2}}{N}\sum_{k=1}^{N}\left(\langle V_{j},V_{k}\rangle_{\textup{F}}~{}U_{j}U_{j}^{\dagger}U_{k}-\langle V_{k},V_{j}\rangle_{\textup{F}}~{}U_{j}U_{k}^{\dagger}U_{j}\right),\\\ \dot{V}_{j}=C_{j}V_{j}+\displaystyle\frac{\kappa_{1}}{N}\sum_{k=1}^{N}\left(\langle U_{j},U_{k}\rangle_{\textup{F}}~{}V_{k}V_{j}^{\dagger}V_{j}-\langle U_{k},U_{j}\rangle_{\textup{F}}~{}V_{j}V_{k}^{\dagger}V_{j}\right)\\\ \hskip 56.9055pt+\displaystyle\frac{\kappa_{2}}{N}\sum_{k=1}^{N}\left(\langle U_{j},U_{k}\rangle_{\textup{F}}~{}V_{j}V_{j}^{\dagger}V_{k}-\langle U_{k},U_{j}\rangle_{\textup{F}}~{}V_{j}V_{k}^{\dagger}V_{j}\right),\\\ \end{cases}$ (1.2) where $B_{j}\in\mathbb{C}^{d_{1}\times d_{2}\times d_{1}\times d_{2}}$ and $C_{j}\in\mathbb{C}^{d_{3}\times d_{4}\times d_{3}\times d_{4}}$ are skew- symmetric rank-4 tensors. For a solution $\\{(U_{i},V_{i})\\}$ to the double matrix model (1.2), a special solution $T_{i}$ to (1.1) can be represented as follows: $T_{i}(t)\equiv U_{i}(t)\otimes V_{i}(t),\quad t>0.$ Precisely, if $T_{i}$ is initially decomposed into the tensor product of two matrices $U_{i}$ and $V_{i}$, then its separability is propagated along the flow for all time. For details, we refer the reader to Section 3. Second, we study emergent dynamics of the double matrix model by investigating several aggregation quantities: $\displaystyle\mathcal{D}(\mathcal{U}(t)):=\max_{1\leq i,j\leq N}\|U_{i}(t)-U_{j}(t)\|_{\textup{F}},\quad\mathcal{S}(\mathcal{U}(t)):=\max_{1\leq i,j\leq N}|n-\langle U_{i},U_{j}\rangle_{\textup{F}}(t)|,$ $\displaystyle\mathcal{D}(\mathcal{V}(t)):=\max_{1\leq i,j\leq N}\|V_{i}(t)-V_{j}(t)\|_{\textup{F}},\quad\mathcal{S}(\mathcal{V}(t)):=\max_{1\leq i,j\leq N}|m-\langle V_{i},V_{j}\rangle_{\textup{F}}(t)|.$ For a homogeneous ensemble (i.e. $B_{i}=B_{j}$ and $C_{i}=C_{j}$ for all $i$ and $j$.) we show that system (1.2) exhibits complete aggregation in which all relative distances for $\\{U_{i}\\}$ and $\\{V_{i}\\}$ tend to zero respectively (see Theorem 4.1). On the other hand, for a heterogeneous ensemble ($B_{i}\neq B_{j}$ and $C_{i}\neq C_{j}$ in general) complete aggregation (one-point collapse) would not be expected. Instead, our concern is dedicated to emergence of locked states in which relative distances converge to positive definite values (see Theorem 4.2). For our analytical results, we need to assume that the size of unitary matrices satisfy $\min(n,m)>4\sqrt{\max(n,m)}$ that requires restriction on $n,m$. In fact, this technical assumption on the sizes is mainly due to the fact that elements are complex-valued. Thus, when the unitary groups $\mathbf{U}(n)$ and $\mathbf{U}(m)$ are replaced by the special orthogonal groups $\textbf{SO}(n)$ and $\textbf{SO}(m)$, such restriction on $n,m$ would be removed (see Theorem C.1 and Theorem C.2). The rest of the paper is organized as follows. In Section 2, we begin with previous results on the relation between the Lohe tensor model and the swarm double sphere model presented in [9]. As a natural extension, we construct the double matrix model in Section 3 and study existence and uniqueness of quadratically separable states. In Section 4, we study emergent dynamics of the double matrix model for both homogeneous and heterogeneous ensembles. Next, the double matrix model is further generalized to the multi matrix model in Section 5. Finally, Section 6 is devoted to a brief summary of the paper and future work. In Appendix A and Appendix B, we provide proofs of Lemma 4.1 and Lemma 4.2, respectively. In Appendix C, emergent dynamics of the double matrix model on $\textbf{SO}(n)\times\textbf{SO}(m)$ is provided. For simplicity of presentation, we use the following abbreviated jargons: * • LT model: Lohe tensor model, LM model: Lohe matrix model, * • SDS model: swarm double sphere model, SMS model: swarm multi-sphere model, * • DM model: double matrix model, DUM model: double unitary matrix model, * • DSOM model: double special orthogonal matrix model, QSS: quadratically separable state, * • MM model: multiple matrix model, MUM model: multiple unitary matrix model. ## 2\. Preliminaries In this section, we review how the SDS model [9, 21] can be related to the LT model via completely separable states, and discuss extension of the SDS model to the SMS model leading to the DM model. ### 2.1. From the LT model to the SDS model In this subsection, we briefly recall the relation between the LT model and the SDS model which was first observed in [9]. In [21], Lohe introduced a first-order aggregation model on the product of two unit spheres $(u_{i},v_{i})\in\mathbb{S}^{d_{1}-1}\times\mathbb{S}^{d_{2}-1}$: $\displaystyle\begin{cases}\dot{u}_{i}=\Omega_{i}u_{i}+\displaystyle\frac{\kappa}{N}\sum_{j=1}^{N}\langle v_{i},v_{j}\rangle(u_{j}-\langle u_{i},u_{j}\rangle u_{i}),\quad t>0,\\\ \dot{v}_{i}=\Lambda_{i}v_{i}+\displaystyle\frac{\kappa}{N}\sum_{j=1}^{N}\langle u_{i},u_{j}\rangle(v_{j}-\langle v_{i},v_{j}\rangle v_{i}),\\\ (u_{i},v_{i})(0)=(u_{i}^{0},v_{i}^{0})\in\mathbb{S}^{d_{1}-1}\times\mathbb{S}^{d_{2}-1},\quad 1\leq i\leq N,\end{cases}$ (2.1) where $\Omega_{i}$ and $\Lambda_{i}$ are skew-symmetric matrices of sizes $d_{1}\times d_{1}$ and $d_{2}\times d_{2}$, respectively: $\Omega_{i}^{\top}=-\Omega_{i},\quad\Lambda_{i}^{\top}=-\Lambda_{i},\quad 1\leq i\leq N,$ and ${\kappa}$ denotes the (uniform) coupling strength. On the other hand, if we choose the following parameters: $m=2,\quad{\kappa}_{00}={\kappa}_{11}=0,\quad{\kappa}_{01}={\kappa}_{10}={\kappa},$ system (1.1) reduces to the generalized Lohe matrix model in [15]: $\begin{cases}\displaystyle{\dot{T}}_{i}=A_{i}T_{i}+\kappa(T_{c}T_{i}^{\dagger}T_{i}-T_{i}T_{c}^{\dagger}T_{i})+\kappa(T_{i}T_{i}^{\dagger}T_{c}-T_{i}T_{c}^{\dagger}T_{i}),\quad t>0,\\\ \displaystyle T_{i}(0)=T_{i}^{0},\quad\|T_{i}^{0}\|_{\textup{F}}=1,\quad T_{c}:=\frac{1}{N}\sum_{k=1}^{N}T_{k},\quad i=1,\cdots,N.\end{cases}$ (2.2) Next, we present how models (2.1) and (2.2) can be viewed as equivalent systems under well-prepared natural frequency tensors and initial data in the following proposition. ###### Proposition 2.1. [9] Systems (2.1) and (2.2) are equivalent in the following sense. 1. (1) Suppose $\\{(u_{i},v_{i})\\}$ is a solution to (2.1). Then, rank-2 real tensors $T_{i}$ defined by $T_{i}:=u_{i}\otimes v_{i}$ is a solution to (2.2) with initial data $T_{i}^{0}=u_{i}^{0}\otimes v_{i}^{0}$ and well-prepared free flow tensors $A_{i}$: $A_{i}T_{i}:=\Omega_{i}T_{i}+T_{i}\Lambda_{i}^{\top}.$ (2.3) 2. (2) Suppose $T_{i}$ is a solution to (2.2)-(2.3) with completely factorized initial data: $T_{i}^{0}=:u_{i}^{0}\otimes v_{i}^{0},\quad 1\leq i,j\leq N,$ for rank-1 real tensors $u_{i}^{0}\in\mathbb{S}^{d_{1}1}$ and $v_{i}^{0}\in\mathbb{S}^{d_{2}-1}$. Then, there exists a pair of unit vectors $(u_{i}(t),v_{i}(t))$ such that $T_{i}(t)=u_{i}(t)\otimes v_{i}(t),\quad t>0,$ where $(u_{i},v_{i})$ is a solution to (2.1) with initial data $(u_{i},v_{i})(0)=(u_{i}^{0},v_{i}^{0})$. By applying the completely separability stated in Proposition 2.1, emergent behaviors for (2.1) and those for (2.2) are exactly the same. Thus, it suffices to investigate the SDS model (2.1). ###### Proposition 2.2. _[9]_ Suppose the initial data $\\{(u_{i}^{0},v_{i}^{0})\\}$ satisfy the following conditions: $\min_{1\leq i,j\leq N}\langle u_{i}^{0},u_{j}^{0}\rangle>0,\quad\min_{1\leq i,j\leq N}\langle v_{i}^{0},v_{j}^{0}\rangle>0,$ and let $\\{(U,V)\\}$ be a solution to system (2.1). Then, we have $\lim_{t\to\infty}\max_{1\leq i,j\leq N}|u_{i}(t)-u_{j}(t)|=0\quad\textup{and}\quad\lim_{t\to\infty}\max_{1\leq i,j\leq N}|v_{i}(t)-v_{j}(t)|=0.$ Now, it is worthwhile mentioning that system (2.1) can be represented as a coupled gradient flow: $\begin{cases}\displaystyle{\dot{u}}_{i}=-\frac{N\kappa}{2}{\mathbb{P}}_{T_{u_{i}}\mathbb{S}^{d_{1}-1}}\Big{(}\nabla_{u_{i}}\mathcal{E}(U,V)\Big{)},\vspace{0.2cm}\\\ \displaystyle{\dot{v}}_{i}=-\frac{N\kappa}{2}{\mathbb{P}}_{T_{v_{i}}\mathbb{S}^{d_{2}-1}}\Big{(}\nabla_{v_{i}}\mathcal{E}(U,V)\Big{)},\end{cases}$ (2.4) where the projection operators ${\mathbb{P}}_{T_{u_{i}}\mathbb{S}^{d_{1}-1}}$ and ${\mathbb{P}}_{T_{v_{i}}\mathbb{S}^{d_{2}-1}}$ onto the tangent spaces of $\mathbb{S}^{d_{1}-1}$ and $\mathbb{S}^{d_{2}-1}$ at $u_{i}$ and $v_{i}$ are defined by the formulae, respectively: for $w_{1}\in\mathbb{R}^{d_{1}}$ and $w_{2}\in\mathbb{R}^{d_{2}}$, $\begin{cases}\displaystyle{\mathbb{P}}_{T_{u_{i}}\mathbb{S}^{d_{1}-1}}(w_{1}):=w_{1}-\langle w_{1},u_{i}\rangle u_{i},\\\ \displaystyle{\mathbb{P}}_{T_{v_{i}}\mathbb{S}^{d_{2}-1}}(w_{2}):=w_{2}-\langle w_{2},v_{i}\rangle v_{i},\end{cases}$ and the potential function $\mathcal{E}(U,V)$ is defined as $\mathcal{E}(U,V):=1-\frac{1}{N^{2}}\sum_{i,j=1}^{N}\langle u_{i},u_{j}\rangle\langle v_{i},v_{j}\rangle.$ (2.5) Thanks to the gradient flow formulation (2.4), any solution to system (2.1) converges to an equilibrium as $t\to\infty$. ### 2.2. From the SDS model to the SMS model In this subsection, we extend the SDS model (2.1) on the product of two unit spheres to an aggregation model on the product of multiple unit spheres, namely, the SMS model. Note that the SDS model can be represented as a gradient flow with a potential function as can be seen in (2.4)–(2.5). Thus, we first generalize the potential function (2.5) as follows: for $u_{i}^{k}\in\mathbb{S}^{d_{k}-1},\quad i=1,\cdots,N,\quad k=1,\cdots,m,$ we set $\mathcal{E}(U^{1},U^{2},\cdots,U^{m}):=1-\frac{1}{N^{2}}\sum_{i,j=1}^{N}\prod_{k=1}^{m}\langle u_{i}^{k},u_{j}^{k}\rangle,\quad U^{k}:=\\{u_{1}^{k},u_{2}^{k},\cdots,u_{N}^{k}\\}.$ (2.6) Using the same spirit for a gradient flow with the potential function (2.6), we propose the SMS model as follows. $\displaystyle\begin{cases}\displaystyle{\dot{u}}_{i}^{k}=\frac{\kappa}{N}\sum_{j=1}^{N}\left(\prod_{l\neq k}\langle u_{i}^{l},u_{j}^{l}\rangle\right)\Big{(}u_{j}^{k}-\langle u_{i}^{k},u_{j}^{k}\rangle u_{i}^{k}\Big{)},\quad t>0,\\\ u_{i}^{k}(0)=u_{i}^{k,0}\in\mathbb{S}^{d_{k}-1}\qquad i\in\\{1,2,\cdots,N\\},\quad k\in\\{1,2,\cdots,m\\}.\end{cases}$ (2.7) As in Section 2.1, we set rank-$m$ real tensor $T_{i}$: $T_{i}:=u_{i}^{1}\otimes\cdots\otimes u_{i}^{m},\quad i=1,\cdots,N.$ Then, it is easy to check that $T_{i}$ satisfies $[\dot{T}_{p}]_{\alpha_{*0}}=\frac{\kappa}{N}\sum_{k=1}^{m}\sum_{\ell=1}^{N}\left([T_{\ell}]_{\alpha_{*i_{*}^{k}}}[\bar{T}_{p}]_{\alpha_{*1}}[T_{p}]_{\alpha_{*(1-i_{*}^{k})}}-[T_{p}]_{\alpha_{*i_{*}^{k}}}[\bar{T}_{\ell}]_{\alpha_{*1}}[T_{p}]_{\alpha_{*(1-i_{*}^{k})}}\right).$ (2.8) It should be noted that (2.8) can be derivable from the Lohe tensor model (1.1) with the following conditions: $\displaystyle\kappa_{i_{*}}=\begin{cases}\kappa\quad&\text{when }i_{*}=i_{*}^{k},\quad 1\leq k\leq m,\\\ 0&\text{otherwise},\end{cases}\quad\text{where}\quad i_{*}^{k}:=\underbrace{(0,\cdots,0,1,0,\cdots,0)}_{\text{only $k^{th}$ index is $1$}}.$ Hence, systems (2.7) and (2.8) can be related in view of a completely separable state, and since the emergent dynamics of the LT model has been discussed in literature [13, 14, 15], we conclude that system (2.7) exhibits complete aggregation under suitable circumstances. ###### Proposition 2.3. _[9]_ Suppose that initial data $T^{0}=\\{T_{i}^{0}\\}$ are completely factorized as a tensor product of rank-1 real tensors: $\displaystyle\begin{aligned} &T_{i}^{0}=u_{i}^{1,0}\otimes u_{i}^{2,0}\otimes\cdots\otimes u_{i}^{m,0},\quad i=1,\cdots,N,\\\ &{\mathcal{A}}(U^{k,0}):=\min_{1\leq i,j\leq N}\langle u^{k,0}_{i},u^{k,0}_{j}\rangle>0,\quad k=1,\cdots,m,\end{aligned}$ and let $T=\\{T_{i}\\}$ be a solution to system (2.8) and $\\{U^{1},\cdots,U^{d}\\}$ be a solution to system (2.7). Then, the following assertions hold. 1. (1) $T_{i}=T_{i}(t)$ is completely separable in the sense that $T_{i}(t)=u_{i}^{1}(t)\otimes u_{i}^{2}(t)\otimes\cdots\otimes u_{i}^{m}(t),\quad t>0,\quad i=1,\cdots,N.$ 2. (2) The solution exhibits the complete aggregation: $\lim_{t\to\infty}\max_{1\leq i,j\leq N}\|T_{i}(t)-T_{j}(t)\|_{\textup{F}}=0.$ ###### Proof. For a proof, we refer the reader to Theorem 6.2 and Proposition 7.1 in [9]. ∎ ## 3\. Existence of rank-4 quadratically separable states In this section, we present existence of the QSS for the LT model with rank-4 tensors and introduce the DM model than can be induced from the LT model. ### 3.1. The DM model In this subsection, we propose the DM model consisting of two generalized Lohe matrix model on the rectangular matrices with possibly different sizes: $U_{j}\in\mathbb{C}^{d_{1}\times d_{2}}\quad\mbox{and}\quad V_{j}\in\mathbb{C}^{d_{3}\times d_{4}},\quad j=1,\cdots,N.$ Below, we sketch our strategy how to derive the DM model from the LT model: * • Step A (A homogeneous ensemble): we present a DM model for a homogeneous ensemble (Section 3.1.1). * • Step B (A heterogeneous ensemble): by adding natural frequency tensors with suitable structure conditions, we derive the DM model from the LT model (Section 3.1.2). #### 3.1.1. A homogeneous ensemble Let $T_{j}\in\mathbb{C}^{d_{1}\times d_{2}\times d_{3}\times d_{4}}$ be a rank-4 tensor that is a solution to (1.1) with zero natural frequency tensors $A_{j}\equiv O$: $[\dot{T}_{j}]_{\alpha_{*0}}=\displaystyle\sum_{i_{*}\in\\{0,1\\}^{4}}\left[\frac{\kappa_{i_{*}}}{N}\sum_{k=1}^{N}\left([T_{k}]_{\alpha_{*i_{*}}}[\bar{T}_{j}]_{\alpha_{*1}}[T_{j}]_{\alpha_{*(1-i_{*})}}-[T_{j}]_{\alpha_{*i_{*}}}[\bar{T}_{k}]_{\alpha_{*1}}[T_{j}]_{\alpha_{*(1-i_{*})}}\right)\right].$ (3.1) For a given solution $T_{j}$ to (3.1), we assume that there exist two matrices $U_{j}\in\mathbb{C}^{d_{1}\times d_{2}}$ and $V_{j}\in\mathbb{C}^{d_{3}\times d_{4}}$ such that $T_{j}=U_{j}\otimes V_{j},\quad[T_{j}(t)]_{\alpha\beta\gamma\delta}=[U_{j}(t)]_{\alpha\beta}[V_{j}(t)]_{\gamma\delta}\quad\mbox{in a component form}.$ Next, we rewrite cubic interaction terms in (3.1) in terms of $U_{j}$ and $V_{j}$. For this, we decompose the index vectors $i_{*}$ and $\alpha_{*i_{*}}$ as $i_{*}:=(i_{1},i_{2},i_{3},i_{4}),\quad\alpha_{*i_{*}}:=(\beta_{*j_{*}},\gamma_{*k_{*}}),\quad j_{*}:=(i_{1},i_{2}),\quad k_{*}:=(i_{3},i_{4}),$ where $j_{*}$ and $k_{*}$ correspond to the index vectors for $U_{j}$ and $V_{j}$, respectively. We now observe $\displaystyle\begin{aligned} &[T_{i}]_{\alpha_{*i_{*}}}[\bar{T}_{j}]_{\alpha_{*1}}[T_{k}]_{\alpha_{*(1-i_{*})}}\\\ &\hskip 14.22636pt=[T_{i}]_{(\beta_{*j_{*}},\gamma_{*k_{*}}}[\bar{T}_{j}]_{(\beta_{*1},\gamma_{*1})}[T_{k}]_{(\beta_{*(1-j_{*})},\gamma_{*(1-k_{*})})}\\\ &\hskip 14.22636pt=[U_{i}]_{\beta_{*j_{*}}}[V_{i}]_{\gamma_{*k_{*}}}[\bar{U}_{j}]_{\beta_{*1}}[\bar{V}_{j}]_{\gamma_{*1}}[U_{k}]_{\beta_{*(1-j_{*})}}[V_{k}]_{\gamma_{*(1-k_{*})}}\\\ &\hskip 14.22636pt=\left([U_{i}]_{\beta_{*j_{*}}}[\bar{U}_{j}]_{\beta_{*1}}[U_{k}]_{\beta_{*(1-j_{*})}}\right)\left([V_{i}]_{\gamma_{*k_{*}}}[\bar{V}_{j}]_{\gamma_{*1}}[V_{k}]_{\gamma_{*(1-k_{*})}}\right).\end{aligned}$ By interchanging the roles of $j\leftrightarrow k$, the term inside of the summation in the right-hand side of (3.1) becomes $\displaystyle[T_{k}]_{\alpha_{*i_{*}}}[\bar{T}_{j}]_{\alpha_{*1}}[T_{j}]_{\alpha_{*(1-i_{*})}}-[T_{j}]_{\alpha_{*i_{*}}}[\bar{T}_{k}]_{\alpha_{*1}}[T_{j}]_{\alpha_{*(1-i_{*})}}$ $\displaystyle\hskip 28.45274pt=\left([U_{k}]_{\beta_{*j_{*}}}[\bar{U}_{j}]_{\beta_{*1}}[U_{j}]_{\beta_{*(1-j_{*})}}\right)\left([V_{k}]_{\gamma_{*k_{*}}}[\bar{V}_{j}]_{\gamma_{*1}}[V_{j}]_{\gamma_{*(1-k_{*})}}\right)$ $\displaystyle\hskip 42.67912pt-\left([U_{j}]_{\beta_{*j_{*}}}[\bar{U}_{k}]_{\beta_{*1}}[U_{j}]_{\beta_{*(1-j_{*})}}\right)\left([V_{j}]_{\gamma_{*k_{*}}}[\bar{V}_{k}]_{\gamma_{*1}}[V_{j}]_{\gamma_{*(1-k_{*})}}\right).$ Since the left-hand side of (3.1) has the form of $\dot{T}_{j}=\dot{U}_{j}\otimes V_{j}+U_{j}\otimes\dot{V}_{j},$ (3.2) one should impose either ${\kappa}_{*}=(1,1)$ or $j_{*}=(1,1)$ to derive the restriction on ${\kappa}_{i_{*}}$: $\kappa_{i_{*}}=0\quad\textup{for all $i_{*}\in\\{0,1\\}^{4}$ with $(i_{1},i_{2})\neq(1,1)$ and $(i_{3},i_{4})\neq(1,1)$.}$ Then, the right-hand side of (3.1) further reduces to $\displaystyle\begin{aligned} &\displaystyle\sum_{i_{*}\in\\{0,1\\}^{4}}\left[\frac{\kappa_{i_{*}}}{N}\sum_{k=1}^{N}\left([T_{k}]_{\alpha_{*i_{*}}}[\bar{T}_{j}]_{\alpha_{*1}}[T_{j}]_{\alpha_{*(1-i_{*})}}-[T_{j}]_{\alpha_{*i_{*}}}[\bar{T}_{k}]_{\alpha_{*1}}[T_{j}]_{\alpha_{*(1-i_{*})}}\right)\right]\\\ &\hskip 28.45274pt=\displaystyle\sum_{j_{*}\in\\{0,1\\}^{2}}\biggl{[}\frac{\kappa_{(j_{*},1,1)}}{N}\sum_{k=1}^{N}\Big{(}[T_{k}]_{\beta_{*j_{*}}\gamma_{*1}}[\bar{T}_{j}]_{\beta_{*1}\gamma_{*1}}[T_{j}]_{\beta_{*(1-j_{*})}\gamma_{*0}}\\\ &\hskip 142.26378pt-[T_{j}]_{\beta_{*j_{*}}\gamma_{*1}}[\bar{T}_{k}]_{\beta_{*1}\gamma_{*1}}[T_{j}]_{\beta_{*(1-j_{*})}\gamma_{*0}}\Big{)}\biggr{]}\\\ &\hskip 42.67912pt+\displaystyle\sum_{k_{*}\in\\{0,1\\}^{2}}\biggl{[}\frac{\kappa_{(1,1,k_{*})}}{N}\sum_{k=1}^{N}\Big{(}[T_{k}]_{\beta_{*1}\gamma_{*k_{*}}}[\bar{T}_{j}]_{\beta_{*1}\gamma_{*1}}[T_{j}]_{\beta_{*0}\gamma_{*(1-k_{*})}}\\\ &\hskip 142.26378pt-[T_{j}]_{\beta_{*1}\gamma_{*k_{*}}}[\bar{T}_{k}]_{\beta_{*1}\gamma_{*1}}[T_{j}]_{\beta_{*0}\gamma_{*(1-k_{*})}}\Big{)}\biggr{]}\\\ &\hskip 28.45274pt=[V_{j}]_{\gamma_{*0}}\displaystyle\sum_{j_{*}\in\\{0,1\\}^{2}}\biggl{[}\frac{\kappa_{(j_{*},1,1)}}{N}\sum_{k=1}^{N}\Big{(}\langle V_{j},V_{k}\rangle_{\textup{F}}[U_{k}]_{\beta_{*j_{*}}}[\bar{U}_{j}]_{\beta_{*1}}[U_{j}]_{\beta_{*(1-j_{*})}}\\\ &\hskip 142.26378pt-\langle V_{k},V_{j}\rangle_{\textup{F}}[U_{j}]_{\beta_{*j_{*}}}[\bar{U}_{k}]_{\beta_{*1}}[U_{j}]_{\beta_{*(1-j_{*})}}\Big{)}\biggr{]}\\\ &\hskip 42.67912pt+[U_{j}]_{\beta_{*0}}\displaystyle\sum_{k_{*}\in\\{0,1\\}^{2}}\biggl{[}\frac{\kappa_{(1,1,k_{*})}}{N}\sum_{k=1}^{N}\Big{(}\langle U_{j},U_{k}\rangle_{\textup{F}}[V_{k}]_{\gamma_{*k_{*}}}[\bar{V}_{j}]_{\gamma_{*1}}[V_{j}]_{\gamma_{*(1-k_{*})}}\\\ &\hskip 142.26378pt-\langle U_{k},U_{j}\rangle_{\textup{F}}[V_{j}]_{\gamma_{*k_{*}}}[\bar{V}_{k}]_{\gamma_{*1}}[V_{j}]_{\gamma_{*(1-k_{*})}}\Big{)}\biggr{]}.\end{aligned}$ By comparing $\cdot\otimes V_{j}$ and $U_{j}\otimes\cdot$ in (3.2), one has $\begin{cases}[\dot{U}_{j}]_{\beta_{*0}}=\displaystyle\sum_{j_{*}\in\\{0,1\\}^{2}}\biggl{(}\frac{\kappa_{(j_{*},1,1)}}{N}\sum_{k=1}^{N}\Big{(}\langle V_{j},V_{k}\rangle_{\textup{F}}[U_{k}]_{\beta_{*j_{*}}}[\bar{U}_{j}]_{\beta_{*1}}[U_{j}]_{\beta_{*(1-j_{*})}}\\\ \displaystyle\hskip 170.71652pt-\langle V_{k},V_{j}\rangle_{\textup{F}}[U_{j}]_{\beta_{*j_{*}}}[\bar{U}_{k}]_{\beta_{*1}}[U_{j}]_{\beta_{*(1-j_{*})}}\Big{)}\biggr{)},\\\ [\dot{V}_{j}]_{\gamma_{*0}}=\displaystyle\sum_{k_{*}\in\\{0,1\\}^{2}}\biggl{(}\frac{\kappa_{(1,1,k_{*})}}{N}\sum_{k=1}^{N}\Big{(}\langle U_{j},U_{k}\rangle_{\textup{F}}[V_{k}]_{\gamma_{*k_{*}}}[\bar{V}_{j}]_{\gamma_{*1}}[V_{j}]_{\gamma_{*(1-k_{*})}}\\\ \displaystyle\hskip 170.71652pt-\langle U_{k},U_{j}\rangle_{\textup{F}}[V_{j}]_{\gamma_{*k_{*}}}[\bar{V}_{k}]_{\gamma_{*1}}[V_{j}]_{\gamma_{*(1-k_{*})}}\Big{)}\biggr{)}.\\\ \end{cases}$ If we choose the coupling strengths as for remaining $i_{*}$: $\displaystyle\kappa_{(0,1,1,1)}=\kappa_{(1,1,0,1)}=\kappa_{1},\quad\kappa_{(1,0,1,1)}=\kappa_{(1,1,1,0)}=\kappa_{2},\quad\kappa_{i_{*}}=0,$ (3.3) we obtain the desired system for $(U_{j},V_{j})$: $\displaystyle\begin{cases}\dot{U}_{j}=\displaystyle\frac{\kappa_{1}}{N}\sum_{k=1}^{N}\left(\langle V_{j},V_{k}\rangle_{\textup{F}}~{}U_{k}U_{j}^{\dagger}U_{j}-\langle V_{k},V_{j}\rangle_{\textup{F}}~{}U_{j}U_{k}^{\dagger}U_{j}\right)\\\ \hskip 28.45274pt+\displaystyle\frac{\kappa_{2}}{N}\sum_{k=1}^{N}\left(\langle V_{j},V_{k}\rangle_{\textup{F}}~{}U_{j}U_{j}^{\dagger}U_{k}-\langle V_{k},V_{j}\rangle_{\textup{F}}~{}U_{j}U_{k}^{\dagger}U_{j}\right),\\\ \dot{V}_{j}=\displaystyle\frac{\kappa_{1}}{N}\sum_{k=1}^{N}\left(\langle U_{j},U_{k}\rangle_{\textup{F}}~{}V_{k}V_{j}^{\dagger}V_{j}-\langle U_{k},U_{j}\rangle_{\textup{F}}~{}V_{j}V_{k}^{\dagger}V_{j}\right)\\\ \hskip 28.45274pt+\displaystyle\frac{\kappa_{2}}{N}\sum_{k=1}^{N}\left(\langle U_{j},U_{k}\rangle_{\textup{F}}~{}V_{j}V_{j}^{\dagger}V_{k}-\langle U_{k},U_{j}\rangle_{\textup{F}}~{}V_{j}V_{k}^{\dagger}V_{j}\right).\\\ \end{cases}$ (3.4) #### 3.1.2. A heterogeneous ensemble Similar to several aggregation models such as the Lohe matrix model [20] and the Lohe tensor model [13], a natural candidate for heterogeneous (or non- identical) extension of (3.4) would be the model (3.4) together with natural frequency tensors $B_{j}$ and $C_{j}$ whose ranks are 4. Thus, the DM model for a heterogeneous ensemble reads as $\displaystyle\begin{cases}\dot{U}_{j}=B_{j}U_{j}+\displaystyle\frac{\kappa_{1}}{N}\sum_{k=1}^{N}\left(\langle V_{j},V_{k}\rangle_{\textup{F}}~{}U_{k}U_{j}^{\dagger}U_{j}-\langle V_{k},V_{j}\rangle_{\textup{F}}~{}U_{j}U_{k}^{\dagger}U_{j}\right)\\\ \hskip 28.45274pt+\displaystyle\frac{\kappa_{2}}{N}\sum_{k=1}^{N}\left(\langle V_{j},V_{k}\rangle_{\textup{F}}~{}U_{j}U_{j}^{\dagger}U_{k}-\langle V_{k},V_{j}\rangle_{\textup{F}}~{}U_{j}U_{k}^{\dagger}U_{j}\right),\\\ \dot{V}_{j}=C_{j}V_{j}+\displaystyle\frac{\kappa_{1}}{N}\sum_{k=1}^{N}\left(\langle U_{j},U_{k}\rangle_{\textup{F}}~{}V_{k}V_{j}^{\dagger}V_{j}-\langle U_{k},U_{j}\rangle_{\textup{F}}~{}V_{j}V_{k}^{\dagger}V_{j}\right)\\\ \hskip 28.45274pt+\displaystyle\frac{\kappa_{2}}{N}\sum_{k=1}^{N}\left(\langle U_{j},U_{k}\rangle_{\textup{F}}~{}V_{j}V_{j}^{\dagger}V_{k}-\langle U_{k},U_{j}\rangle_{\textup{F}}~{}V_{j}V_{k}^{\dagger}V_{j}\right),\\\ \end{cases}$ (3.5) where $B_{j}\in\mathbb{C}^{d_{1}\times d_{2}\times d_{1}\times d_{2}}$ and $C_{j}\in\mathbb{C}^{d_{3}\times d_{4}\times d_{3}\times d_{4}}$ are rank-4 tensors satisfying skew-symmetric properties: for suitable indices, $[B_{j}]_{\alpha_{1}\beta_{1}\alpha_{2}\beta_{2}}=-[\bar{B}_{j}]_{\alpha_{2}\beta_{2}\alpha_{1}\beta_{1}},\quad[C_{j}]_{\gamma_{1}\delta_{1}\gamma_{2}\delta_{2}}=-[\bar{C}_{j}]_{\gamma_{2}\delta_{2}\gamma_{1}\delta_{1}}.$ (3.6) Moreover, $U_{j}B_{j}$ and $C_{j}V_{j}$ can be understood as tensor contractions between rank-4 and rank-2 tensors: $[B_{j}U_{j}]_{\alpha\beta}=[B_{j}]_{\alpha\beta\gamma\delta}[U_{j}]_{\gamma\delta},\quad[C_{j}V_{j}]_{\alpha\beta}=[C_{j}]_{\alpha\beta\gamma\delta}[V_{j}]_{\gamma\delta}.$ Next, we find a condition for $A_{j}$ in (1.1) in terms of $B_{j}$ and $C_{j}$ in (3.5) so that models (1.1) and (3.5) are equivalent. For this, it suffices to focus on the free flows by setting ${\kappa}_{1}={\kappa}_{2}=0$, i.e., $\dot{U}_{j}=B_{j}U_{j},\quad\dot{V}_{j}=C_{j}V_{j},\quad j=1,\cdots,N.$ If we use the relations above and $(U_{j}\otimes V_{j})^{\prime}=(B_{j}U_{j})\otimes V_{j}+U_{j}\otimes(C_{j}V_{j}),$ then we can find $A_{j}(U_{j}\otimes V_{j})=(B_{j}U_{j})\otimes V_{j}+U_{j}\otimes(C_{j}V_{j}).$ (3.7) In addition, if we rewrite (3.7) as a component form, then it becomes $\displaystyle\begin{aligned} 0&=[A_{j}]_{\alpha_{1}\beta_{1}\gamma_{1}\delta_{1}\alpha_{2}\beta_{2}\gamma_{2}\delta_{2}}[U_{j}]_{\alpha_{2}\beta_{2}}[V_{j}]_{\gamma_{2}\delta_{2}}-[B_{j}]_{\alpha_{1}\beta_{1}\alpha_{2}\beta_{2}}[U_{j}]_{\alpha_{2}\beta_{2}}[V_{j}]_{\gamma_{1}\delta_{1}}\\\ &\hskip 14.22636pt-[U_{j}]_{\alpha_{1}\beta_{1}}[C_{j}]_{\gamma_{1}\delta_{1}\gamma_{2}\delta_{2}}[V_{j}]_{\gamma_{2}\delta_{2}}\\\ &=\Big{(}[A_{j}]_{\alpha_{1}\beta_{1}\gamma_{1}\delta_{1}\alpha_{2}\beta_{2}\gamma_{2}\delta_{2}}-[B_{j}]_{\alpha_{1}\beta_{1}\alpha_{2}\beta_{2}}\delta_{\gamma_{1}\gamma_{2}}\delta_{\delta_{1}\delta_{2}}-[C_{j}]_{\gamma_{1}\delta_{1}\gamma_{2}\delta_{2}}\delta_{\alpha_{1}\alpha_{2}}\delta_{\beta_{1}\beta_{2}}\Big{)}\\\ &\hskip 14.22636pt\times[U_{j}]_{\alpha_{2}\beta_{2}}[V_{j}]_{\gamma_{2}\delta_{2}}.\end{aligned}$ (3.8) Since (3.8) holds for arbitrary $U_{j}$ and $V_{j}$, we can find an explicit relation between $A_{j}$ and $B_{j}$, $C_{j}$: $\displaystyle[A_{j}]_{\alpha_{1}\beta_{1}\gamma_{1}\delta_{1}\alpha_{2}\beta_{2}\gamma_{2}\delta_{2}}=[B_{j}]_{\alpha_{1}\beta_{1}\alpha_{2}\beta_{2}}\delta_{\gamma_{1}\gamma_{2}}\delta_{\delta_{1}\delta_{2}}+[C_{j}]_{\gamma_{1}\delta_{1}\gamma_{2}\delta_{2}}\delta_{\alpha_{1}\alpha_{2}}\delta_{\beta_{1}\beta_{2}}.$ (3.9) So far, we have derived the DM model (3.5) with (3.9) from the LT model (1.1). However, if we use the same argument reversely and recall that system (3.5) admits a unique solution, then one can identify the LT model from the DM model. Thus, we can say that system (3.5)–(3.9) and system (1.1) are equivalent in some sense. Below, we summarize the argument above in the following proposition. ###### Proposition 3.1. 1. (1) Suppose $\\{(U_{i},V_{i})\\}$ is a solution to (3.5). Then, a rank-4 tensor $T_{i}$ defined by $T_{i}:=U_{i}\otimes V_{i}$ is a QSS to (1.1) with a well- prepared free flow tensor $A_{i}$ satisfying (3.9). 2. (2) Suppose a rank-4 tensor $T_{i}$ is a solution to (1.1) with (3.9) and quadratically separable initial data: $T_{i}^{0}=:U_{i}^{0}\otimes V_{i}^{0},\quad 1\leq i\leq N,$ for rank-2 tensors $U_{i}^{0}\in\mathbb{C}^{d_{1}\times d_{2}}$ and $V_{i}^{0}\in\mathbb{C}^{d_{3}\times d_{4}}$ with unit norms. Then, there exist two matrices with unit norms $U_{i}=U_{i}(t)$ and $V=V_{i}(t)$ such that $T_{i}(t)=U_{i}(t)\otimes V_{i}(t),\quad t>0,$ where $(U_{i},V_{i})$ is a solution to (3.5) with $(U_{i},V_{i})(0)=(U_{i}^{0},V_{i}^{0})$. ### 3.2. Gradient flow formulation of the DM model In this subsection, we show that system (3.5) can be formulated as a gradient flow with a suitable analytical potential induced from the Lohe tensor model. Recall from [14] that the following functional can be associated with the LT model: $\mathcal{V}(T)=1-\frac{1}{N^{2}}\sum_{i,j=1}^{N}\langle T_{i},T_{j}\rangle_{\textup{F}}.$ (3.10) If we use the decomposition $T_{i}=U_{i}\otimes V_{i}$, then we find the corresponding functional for (3.5): $\mathcal{E}(U,V):=1-\frac{1}{N^{2}}\sum_{i,j=1}^{N}\langle U_{i}\otimes V_{i},U_{j}\otimes V_{j}\rangle_{\textup{F}}=1-\frac{1}{N^{2}}\sum_{i,j=1}^{N}\langle U_{i},U_{j}\rangle_{\textup{F}}\langle V_{i},V_{j}\rangle_{\textup{F}}.$ (3.11) In the following lemma, we show that $\mathcal{E}(U,V)$ is non-increasing along the flow (3.5). ###### Lemma 3.1. Let $\\{(U_{i},V_{i})\\}$ be a solution to (3.5) with $(B_{j},C_{j})=(O,O)$. Then, the functional $\mathcal{E}(U,V)$ is non-increasing in time: $\displaystyle\frac{{\textup{d}}}{{\textup{d}t}}\mathcal{E}(U,V)=-\frac{\kappa_{1}}{N}\sum_{j=1}^{N}\left\|\frac{1}{N}\sum_{i=1}^{N}\Big{(}\langle V_{j},V_{i}\rangle_{\textup{F}}U_{i}U_{j}^{\dagger}-\langle V_{i},V_{j}\rangle_{\textup{F}}U_{j}U_{i}^{\dagger}\Big{)}\right\|_{\textup{F}}^{2}$ $\displaystyle\hskip 19.91684pt-\frac{\kappa_{1}}{N}\sum_{j=1}^{N}\left\|\frac{1}{N}\sum_{i=1}^{N}\left(\langle U_{j},U_{i}\rangle_{\textup{F}}V_{i}V_{j}^{\dagger}-\langle U_{i},U_{j}\rangle_{\textup{F}}V_{j}V_{i}^{\dagger}\right)\right\|_{\textup{F}}^{2}$ $\displaystyle\hskip 19.91684pt-\frac{\kappa_{2}}{N}\sum_{j=1}^{N}\left\|\frac{1}{N}\sum_{i=1}^{N}\Big{(}\langle V_{j},V_{i}\rangle_{\textup{F}}U_{j}^{\dagger}U_{i}-\langle V_{i},V_{j}\rangle_{\textup{F}}U_{i}^{\dagger}U_{j}\Big{)}\right\|_{\textup{F}}^{2}$ $\displaystyle\hskip 19.91684pt-\frac{\kappa_{2}}{N}\sum_{j=1}^{N}\left\|\frac{1}{N}\sum_{i=1}^{N}\left(\langle U_{j},U_{i}\rangle_{\textup{F}}V_{j}^{\dagger}V_{i}-\langle U_{i},U_{j}\rangle_{\textup{F}}V_{i}^{\dagger}V_{j}\right)\right\|_{\textup{F}}^{2}.$ ###### Proof. It follows from Proposition 4.1 in [14] that $\mathcal{V}(T)$ in (3.10) for the LT model (1.1) satisfies $\frac{{\textup{d}}}{{\textup{d}t}}\mathcal{V}(T)=-\frac{1}{N}\sum_{j=1}^{N}\sum_{i_{*}\in\\{0,1\\}^{4}}\kappa_{i_{*}}\|M^{i_{*}}(T_{c})M^{i_{*}}(T_{j})^{\dagger}-M^{i_{*}}(T_{j})M^{i_{*}}(T_{c})^{\dagger}\|_{\textup{F}}^{2}.$ (3.12) Here, we recall from [14, Definition 3.1] that for a rank-4 tensor $T$, $\mathcal{M}^{i_{*}}(T)$ is defined as a rank-2 tensor reshaped from $T$ or it can be understood as a bijective linear map which conserves the Frobenius norm. Since we only consider four types among $i_{*}\in\\{0,1\\}^{4}$ as in (3.3), we set $\kappa_{(0,1,1,1)}=\kappa_{(1,1,0,1)}=\kappa_{1},\quad\kappa_{(1,0,1,1)}=\kappa_{(1,1,1,0)}=\kappa_{2},\quad\kappa_{i_{*}}=0\quad\text{for other }i_{*}.$ (3.13) On the other hand, in (3.12), we are concerned with the term $M^{i_{*}}(T_{i})M^{i_{*}}(T_{j})^{\dagger}$ for $T_{i}=U_{i}\otimes V_{i}$ and $i_{*}$ in (3.13), for instance, if $i_{*}=(0,1,1,1)$, then we have $\displaystyle[M^{i_{*}}(T_{i})M^{i_{*}}(T_{j})^{\dagger}]_{\alpha_{20}\alpha_{21}}$ $\displaystyle=[T_{i}]_{\alpha_{11}\alpha_{20}\alpha_{31}\alpha_{41}}[\bar{T}_{j}]_{\alpha_{11}\alpha_{21}\alpha_{31}\alpha_{41}}$ $\displaystyle=[U_{i}]_{\alpha_{11}\alpha_{20}}[V_{i}]_{\alpha_{31}\alpha_{41}}[\bar{U}_{j}]_{\alpha_{11}\alpha_{21}}[\bar{V}_{j}]_{\alpha_{31}\alpha_{41}}=\langle V_{j},V_{i}\rangle_{\textup{F}}[U_{j}^{\dagger}U_{i}]_{\alpha_{21}\alpha_{20}}.$ If we repeat similar calculations as above, we obtain the desired dissipative estimate. ∎ ### 3.3. The DM model In this subsection, we further reduce the DM model (3.5) to the model defined on the product of two unitary matrices $\mathbf{U}(n)\times\mathbf{U}(m)$ which we call as the DUM model. By the construction of model (3.5), we know that the Frobenius norms of $U_{j}$ and $V_{j}$ are conserved. In addition to the conservation of the Frobenius norms, we show that unitarity is also preserved, when rectangular matrices are replaced by square matrices. ###### Lemma 3.2. Suppose that the system parameters and initial data satisfy $d_{1}=d_{2}=n,\quad d_{3}=d_{4}=m,\quad U_{j}^{0}\in\mathbf{U}(n),\quad V_{j}^{0}\in\mathbf{U}(m),$ and let $\\{(U_{j},V_{j})\\}$ be a solution to (3.5). Then, one has $\displaystyle U_{j}(t)\in\mathbf{U}(n),\quad V_{j}(t)\in\mathbf{U}(m),\quad t>0,\quad j=1,\cdots,N.$ ###### Proof. Since $U_{j}$ and $V_{j}$ have the same structure, we focus only on $U_{j}$. We rewrite the dynamics of $U_{j}$ as a simpler form: $\dot{U}_{j}=B_{j}U_{j}+(P_{j}-P_{j}^{\dagger})U_{j}+U_{j}(Q_{j}-Q_{j}^{\dagger}),$ where $P_{j}$ and $Q_{j}$ are defined as $\displaystyle P_{j}:=\frac{{\kappa}_{1}}{N}\sum_{k=1}^{N}\langle V_{j},V_{k}\rangle_{\textup{F}}U_{k}U_{j}^{\dagger},\quad Q_{j}:=\frac{{\kappa}_{2}}{N}\sum_{k=1}^{N}\langle V_{j},V_{k}\rangle_{\textup{F}}U_{j}^{\dagger}U_{k}.$ Then, one has $\displaystyle\frac{{\textup{d}}}{{\textup{d}t}}(U_{j}U_{j}^{\dagger})$ $\displaystyle=B_{j}U_{j}U_{j}^{\dagger}-U_{j}U_{j}^{\dagger}B_{j}+(P_{j}-P_{j}^{\dagger})U_{j}U_{j}^{\dagger}+U_{j}(Q_{j}-Q_{j}^{\dagger})U_{j}^{\dagger}$ $\displaystyle\hskip 14.22636pt- U_{j}U_{j}^{\dagger}(P_{j}-P_{j}^{\dagger})-U_{j}(Q_{j}-Q_{j}^{\dagger})U_{j}^{\dagger}$ $\displaystyle=(B_{j}+P_{j}-P_{j}^{\dagger})U_{j}U_{j}^{\dagger}-U_{j}U_{j}^{\dagger}(B_{j}+P_{j}-P_{j}^{\dagger}).$ By straightforward calculation, we find $\frac{{\textup{d}}}{{\textup{d}t}}\|I_{n}-U_{j}U_{j}^{\dagger}\|_{\textup{F}}=0.$ (3.14) Since we assume $U_{j}^{0}\in\mathbf{U}(n)$, i.e., $I_{n}-U_{j}U_{j}^{\dagger}=O$, the relation (3.14) yields the desired result. ∎ Due to the unitarity, system (3.5) can further be reduced to the model on $\mathbf{U}(n)\times\mathbf{U}(m)$ by using the relations: $U_{j}U_{j}^{\dagger}=I_{n}=U_{j}^{\dagger}U_{j},\quad V_{j}V_{j}^{\dagger}=I_{m}=V_{j}^{\dagger}V_{j},\quad j=1,\cdots,N.$ Thus, the DM model reads as $\displaystyle\begin{cases}\dot{U}_{j}=B_{j}U_{j}+\displaystyle\frac{\kappa}{N}\sum_{k=1}^{N}\left(\langle V_{j},V_{k}\rangle_{\textup{F}}~{}U_{k}-\langle V_{k},V_{j}\rangle_{\textup{F}}~{}U_{j}U_{k}^{\dagger}U_{j}\right),\\\ \dot{V}_{j}=C_{j}V_{j}+\displaystyle\frac{\kappa}{N}\sum_{k=1}^{N}\left(\langle U_{j},U_{k}\rangle_{\textup{F}}~{}V_{k}-\langle U_{k},U_{j}\rangle_{\textup{F}}~{}V_{j}V_{k}^{\dagger}V_{j}\right),\end{cases}$ where $\kappa=\kappa_{1}+\kappa_{2}$. Note that natural frequency tensors $B_{j}$ and $C_{j}$ have rank-4 satisfying skew-symmetric properties as in (3.6). In order to give a meaning of Hamiltonian, we associate two Hermitian matrices, namely, $H_{j}\in\mathbb{C}^{n\times n}$ and $G_{j}\in\mathbb{C}^{m\times m}$: $[B_{j}]_{\alpha_{1}\beta_{1}\alpha_{2}\beta_{2}}=:[-\mathrm{i}H_{j}]_{\alpha_{1}\alpha_{2}}\delta_{\beta_{1}\beta_{2}},\quad[C_{j}]_{\gamma_{1}\delta_{1}\gamma_{2}\delta_{2}}=:[-\mathrm{i}G_{j}]_{\gamma_{1}\gamma_{2}}\delta_{\delta_{1}\delta_{2}}.$ Then, system (3.10) reduces to the model on $\mathbf{U}(n)\times\mathbf{U}(m)$: $\displaystyle\begin{cases}\dot{U}_{j}=-\mathrm{i}H_{j}U_{j}+\displaystyle\frac{\kappa}{N}\sum_{k=1}^{N}\left(\langle V_{j},V_{k}\rangle_{\textup{F}}~{}U_{k}-\langle V_{k},V_{j}\rangle_{\textup{F}}~{}U_{j}U_{k}^{\dagger}U_{j}\right),\\\ \dot{V}_{j}=-\mathrm{i}G_{j}V_{j}+\displaystyle\frac{\kappa_{1}}{N}\sum_{k=1}^{N}\left(\langle U_{j},U_{k}\rangle_{\textup{F}}~{}V_{k}-\langle U_{k},U_{j}\rangle_{\textup{F}}~{}V_{j}V_{k}^{\dagger}V_{j}\right),\end{cases}$ (3.15) where $H_{j}U_{j}$ and $G_{j}V_{j}$ are now usual matrix products. Then as in Lemma 3.1, system (3.15) also satisfies the dissipative energy estimate. Since the proof can be directly obtained from Lemma 3.1, we omit it. ###### Corollary 3.1. Let $\\{(U_{i},V_{i})\\}$ be a solution to (3.15) with $H_{j}=G_{j}\equiv O$. Then the Lyapunov functional (5.1) satisfies $\frac{{\textup{d}}}{{\textup{d}t}}\mathcal{E}(U,V)=-\frac{\kappa}{N}\sum_{j=1}^{N}\left(\|\dot{U}_{j}\|_{\textup{F}}^{2}+\|\dot{V}_{j}\|_{\textup{F}}^{2}\right),\quad t>0.$ ## 4\. Emergent dynamics of rank-4 quadratically separable states In this section, we study the emergent behavior of the QSS for the Lohe tensor model by investigating the dynamics of the DUM model which reads as follows: $\begin{cases}\dot{U}_{j}=\displaystyle-\mathrm{i}H_{j}U_{j}+\frac{\kappa}{N}\sum_{k=1}^{N}\left(\langle V_{j},V_{k}\rangle_{\textup{F}}~{}U_{k}-\langle V_{k},V_{j}\rangle_{\textup{F}}~{}U_{j}U_{k}^{\dagger}U_{j}\right),\quad t>0,\\\ \dot{V}_{j}=\displaystyle-\mathrm{i}G_{j}V_{j}+\frac{\kappa}{N}\sum_{k=1}^{N}\left(\langle U_{j},U_{k}\rangle_{\textup{F}}~{}V_{k}-\langle U_{k},U_{j}\rangle_{\textup{F}}~{}V_{j}V_{k}^{\dagger}V_{j}\right),\\\ (U_{j},V_{j})(0)=(U_{j}^{0},V_{j}^{0})\in\mathbf{U}(n)\times\mathbf{U}(m),\quad 1\leq j\leq N,\end{cases}$ (4.1) where $H_{j}\in\mathbb{C}^{n\times n}$ and $G_{j}\in\mathbb{C}^{m\times m}$ are Hermitian matrices which play roles of natural frequencies for each oscillator. First, we recall several definitions of emergent behaviors for (4.1). ###### Definition 4.1. _[16]_ Let $(\mathcal{U},\mathcal{V}):=\\{U_{j},V_{j}\\}_{j=1}^{N}$ be a solution to (4.1). 1. (1) System (4.1) exhibits complete aggregation if the following estimate holds: $\lim_{t\to\infty}\max_{1\leq i,j\leq N}\Big{(}\|U_{i}(t)-U_{j}(t)\|_{\textup{F}}+\|V_{i}(t)-V_{j}(t)\|_{\textup{F}}\Big{)}=0.$ (4.2) 2. (2) The state $(\mathcal{U},\mathcal{V})$ tends to a phase-locked state if the following relation holds: $\exists~{}\lim_{t\to\infty}U_{i}(t)U_{j}(t)^{\dagger}\quad\textup{and}\quad\exists~{}\lim_{t\to\infty}V_{i}(t)V_{j}(t)^{\dagger}.$ In order to investigate emergent behaviors for (4.1), we denote the following quantities for notational simplicity: $\displaystyle\begin{aligned} &X_{ij}:=U_{i}U_{j}^{\dagger},\quad S_{ij}:=I_{n}-U_{i}U_{j}^{\dagger},\quad d_{ij}:=\langle U_{i},U_{j}\rangle_{\textup{F}},\\\ &Y_{ij}:=V_{i}V_{j}^{\dagger},\quad T_{ij}:=I_{m}-V_{i}V_{j}^{\dagger}\quad c_{ij}:=\langle V_{i},V_{j}\rangle_{\textup{F}}.\end{aligned}$ (4.3) It follows from simple observations that $\displaystyle\|U_{j}\|_{\textup{F}}=n,\quad\|V_{j}\|_{\textup{F}}=m,\quad|d_{ij}|\leq n,\quad|c_{ij}|\leq m,$ $\displaystyle\|S_{ij}\|^{2}_{\textup{F}}=\|U_{i}-U_{j}\|_{\textup{F}}^{2}=2\textup{Re}(n-d_{ij}),\quad\|T_{ij}\|_{\textup{F}}^{2}=\|V_{i}-V_{j}\|_{\textup{F}}^{2}=2\textup{Re}(m-c_{ij}).$ Then, it is easy to see that $\|U_{i}-U_{j}\|_{\textup{F}}\to 0\Longleftrightarrow|n-d_{ij}|\to 0,\quad\|V_{i}-V_{j}\|_{\textup{F}}\to 0\Longleftrightarrow|m-c_{ij}|\to 0.$ Thus, the complete aggregation in (4.2) can be represented in terms of the quantities in (4.3): $\displaystyle\lim_{t\to\infty}\max_{1\leq i,j\leq N}\Big{(}\|S_{ij}(t)\|_{\textup{F}}+\|T_{ij}(t)\|_{\textup{F}}\Big{)}=0\quad\textup{or}\quad\lim_{t\to\infty}\max_{1\leq i,j\leq N}\Big{(}|n-d_{ij}(t)|+|m-c_{ij}(t)|\Big{)}=0.$ In this regard, we define aggregation quantities: for $t>0$, $\displaystyle\mathcal{D}(\mathcal{U}(t)):=\max_{1\leq i,j\leq N}\|U_{i}(t)-U_{j}(t)\|_{\textup{F}},\quad\mathcal{S}(\mathcal{U}(t)):=\max_{1\leq i,j\leq N}|n-d_{ij}(t)|,$ $\displaystyle\mathcal{D}(\mathcal{V}(t)):=\max_{1\leq i,j\leq N}\|V_{i}(t)-V_{j}(t)\|_{\textup{F}},\quad\mathcal{S}(\mathcal{V}(t)):=\max_{1\leq i,j\leq N}|m-c_{ij}(t)|.$ Note that $\mathcal{S}(\mathcal{U})$ has a second-order with respect to the state $\mathcal{U}$, whereas $\mathcal{D}(\mathcal{U})$ has a first-order. In the following two subsections, we present emergent dynamics for homogeneous and heterogeneous ensembles respectively. ### 4.1. Homogeneous ensemble In this subsection, we deal with the emergent dynamics of (4.1) with a homogeneous ensemble: $H_{j}\equiv H,\quad G_{j}\equiv G,\quad j=1,\cdots,N,$ and by the solution splitting property, we may assume $H=O,\quad G=O.$ In this setting, system (3.5) becomes $\displaystyle\begin{cases}\dot{U}_{j}=\displaystyle\frac{\kappa}{N}\sum_{k=1}^{N}\left(\langle V_{j},V_{k}\rangle_{\textup{F}}~{}U_{k}-\langle V_{k},V_{j}\rangle_{\textup{F}}~{}U_{j}U_{k}^{\dagger}U_{j}\right),\quad t>0,\\\ \dot{V}_{j}=\displaystyle\frac{\kappa}{N}\sum_{k=1}^{N}\left(\langle U_{j},U_{k}\rangle_{\textup{F}}~{}V_{k}-\langle U_{k},U_{j}\rangle_{\textup{F}}~{}V_{j}V_{k}^{\dagger}V_{j}\right),\\\ (U_{j},V_{j})(0)=(U_{j}^{0},V_{j}^{0})\in\mathbf{U}(n)\times\mathbf{U}(m).\end{cases}$ (4.4) Without loss of generality, we may assume $n\geq m.$ (4.5) Our goal of this subsection is to find a sufficient condition under which $\lim_{t\to\infty}\mathcal{L}(t)=0,\quad\mathcal{L}(t):=\mathcal{D}(\mathcal{U}(t))+\mathcal{D}(\mathcal{V}(t))+\mathcal{S}(\mathcal{U}(t))+\mathcal{S}(\mathcal{V}(t)),$ where $\mathcal{L}=\mathcal{L}(t)$ is called as a total aggregation functional. In [16], it suffices to study the temporal evolutions of $\mathcal{D}(\mathcal{U})$. However in our case, time evolutions of $\mathcal{S}(\mathcal{U})$ as well as $\mathcal{S}(\mathcal{V})$ are needed to achieve complete aggregation estimates. Below, we derive a differential inequality for $\mathcal{L}$. ###### Lemma 4.1. Let $\\{(U_{j},V_{j})\\}$ be a solution to (4.4) with (4.5). Then, the total aggregation functional $\mathcal{L}$ satisfies $\displaystyle\begin{aligned} \dot{\mathcal{L}}\leq-2{\kappa}(m-4\sqrt{n})\mathcal{L}+{\kappa}(4n+9)\mathcal{L}^{2}+{\kappa}\left(2n+\frac{8}{3}\right)\mathcal{L}^{3},\quad t>0.\end{aligned}$ (4.6) ###### Proof. Since a proof is lengthy, we provide it in Appendix A. ∎ We are now ready to provide a sufficient condition leading to the complete aggregation for (4.4). ###### Theorem 4.1. Suppose that the system parameters and initial data satisfy $\displaystyle\begin{aligned} &\textup{(i)}~{}~{}n\geq m>4\sqrt{n}.\\\ &\textup{(ii)}~{}~{}\mathcal{L}^{0}<\alpha_{n,m}:=\frac{-(12n+27)+\sqrt{(12n+27)^{2}+48(m-4\sqrt{n})(3n+4)}}{4(3n+4)},\end{aligned}$ (4.7) and let $\\{(U_{j},V_{j})\\}$ be a solution to (4.4). Then, we have $\lim_{t\to\infty}\mathcal{L}(t)=0.$ Moreover, the convergence rate is at least exponential. In other words, system (4.4) exhibits complete aggregation with an exponential convergence. ###### Proof. Consider an auxiliary quadratic polynomial: $f(s):=\left(2n+\frac{8}{3}\right)s^{2}+(4n+9)s-2(m-4\sqrt{n}).$ Since $m>4\sqrt{n}$, the algebraic relation $f=0$ admits a unique positive root $\alpha_{n,m}$ defined in (4.7)(ii). Then, the relation (4.6) is rewritten in terms of $f$: $\frac{{\textup{d}}}{{\textup{d}t}}\mathcal{L}\leq{\kappa}\mathcal{L}f(\mathcal{L}),\quad t>0.$ Since initial data satisfy (4.7)(ii), the desired result follows from dynamical systems theory. ∎ ###### Remark 4.1. (i) In (4.7)(i), we have assumed that $n\geq m>4\sqrt{n}$ which imposes a restriction on the size of $U_{i}$ such as $n>16.$ It also should be mentioned that such restriction arises, for instance, when we estimate the terms $\mathcal{I}_{15}$ in (A.7) and $\mathcal{I}_{25}$ in (A.10). We indeed show in Appendix C that this technical assumption would be removed. (ii) Since $m\leq n$, we have $\limsup_{n\to\infty}\alpha_{n,m}\leq\lim_{n\to\infty}\frac{-(12n+27)+\sqrt{(12n+27)^{2}+48(n-4\sqrt{n})(3n+4)}}{4(3n+4)}=\frac{1}{2}.$ ### 4.2. Heterogeneous ensemble In this subsection, we study the case of heterogeneous Hamiltonians in which $H_{j}$ and $G_{j}$ in (4.1) are given to be different in general. In order to establish the emergence of the phase-locked state, we will follow a strategy developed in [16]. For any two solutions $\\{U_{j},V_{j}\\}$ and $\\{\tilde{U}_{j},\tilde{V}_{j}\\}$ to (4.1), we define the diameters measuring the dissimilarity of two configurations: $\displaystyle\begin{aligned} &d(U,\tilde{U})(t):=\max_{1\leq i,j\leq N}\|U_{i}(t)U_{j}^{\dagger}(t)-\tilde{U}_{i}(t)\tilde{U}_{j}^{\dagger}(t)\|_{\textup{F}},\\\ &d(V,\tilde{V})(t):=\max_{1\leq i,j\leq N}\|V_{i}(t)V_{j}^{\dagger}(t)-\tilde{V}_{i}(t)\tilde{V}_{j}^{\dagger}(t)\|_{\textup{F}}.\end{aligned}$ (4.8) Then, we will show that the diameters above converge to zero: $\lim_{t\to\infty}\Big{(}d(U,\tilde{U})(t)+d(V,\tilde{V})(t)\Big{)}=0.$ (4.9) As a next step, since our system is autonomous, for any $T>0$, we choose $\tilde{U}_{j}$ and $\tilde{V}_{j}$ as $\tilde{U}_{j}(t)=U_{j}(t+T),\quad\tilde{V}_{j}(t)=V_{j}(t+T).$ By discretizing the time $t\in\mathbb{R}_{+}$ as $n\in\mathbb{Z}_{+}$ and setting $T=m\in\mathbb{Z}_{+}$, we conclude that $\\{U_{i}(n)U_{j}^{\dagger}(n)\\}_{n\in\mathbb{Z}_{+}}$ and $\\{V_{i}(n)V_{j}^{\dagger}(n)\\}_{n\in\mathbb{Z}_{+}}$ are indeed Cauchy sequences in the complete spaces $\mathbf{U}(n)$ and $\mathbf{U}(m)$, respectively. Hence, there exist two constant unitary matrices $U_{ij}^{\infty}\in\mathbf{U}(n)$ and $V_{ij}^{\infty}\in\mathbf{U}(m)$ such that $\lim_{t\to\infty}\|U_{i}(t)U_{j}^{\dagger}(t)-U_{ij}^{\infty}\|_{\textup{F}}=0,\quad\lim_{t\to\infty}\|V_{i}(t)V_{j}^{\dagger}(t)-V_{ij}^{\infty}\|_{\textup{F}}=0.$ Hence, we aim to find a sufficient condition under which (4.9) holds. To show (4.9), we associate another diameters measuring the difference between two solution configurations $\\{U_{j},V_{j}\\}$ and $\\{\tilde{U}_{j},\tilde{V}_{j}\\}$: $\displaystyle\begin{aligned} &\mathcal{S}(U,\tilde{U})(t):=\max_{1\leq i,j\leq N}|\langle U_{i},U_{j}\rangle(t)-\langle\tilde{U}_{i},\tilde{U}_{j}\rangle(t)|,\\\ &\mathcal{S}(V,\tilde{V})(t):=\max_{1\leq i,j\leq N}|\langle V_{i},V_{j}\rangle(t)-\langle\tilde{V}_{i},\tilde{V}_{j}\rangle(t)|.\end{aligned}$ (4.10) Note that $\mathcal{S}(U,\tilde{U})\leq\sqrt{n}d(U,\tilde{U}),\quad\mathcal{S}(V,\tilde{V})\leq\sqrt{m}d(V,\tilde{V}).$ To this end, our goal of this subsection is to find a sufficient framework leading to $\lim_{t\to\infty}\mathcal{F}(t)=0,\quad\mathcal{F}(t):=d(U,\tilde{U})(t)+d(V,\tilde{V})(t)+\mathcal{S}(U,\tilde{U})(t)+\mathcal{S}(V,\tilde{V})(t).$ (4.11) Below, we derive a differential inequality for $\mathcal{F}$ in (4.11). ###### Lemma 4.2. Let $\\{(U_{i},V_{i})\\}$ and $\\{(\tilde{U}_{i},\tilde{V}_{i})\\}$ be any two solutions to (4.1) with (4.5), respectively. Then, the aggregation functional $\mathcal{F}$ satisfies $\frac{{\textup{d}}}{{\textup{d}t}}{\mathcal{F}}\leq-{\kappa}\left(2m-8\sqrt{n}-\frac{\max\\{\mathcal{D}(\mathcal{H}),\mathcal{D}(\mathcal{G})\\}}{{\kappa}}\right)\mathcal{F}+{\kappa}(4n+22)\mathcal{L}\mathcal{F}+20{\kappa}\mathcal{L}^{2}\mathcal{F},$ (4.12) where $\mathcal{D}(\mathcal{H})$ and $\mathcal{D}(\mathcal{G})$ are defined as $\mathcal{D}(\mathcal{H}):=\max_{1\leq i,j\leq N}\|H_{i}-H_{j}\|_{\infty},\quad\mathcal{D}(\mathcal{G}):=\max_{1\leq i,j\leq N}\|G_{i}-G_{j}\|_{\infty}.$ ###### Proof. We postpone its proof in Appendix B. ∎ In what follows, using the differential inequality (4.12), we provide a sufficient condition leading to the phase-locked state. First, in order to make a leading coefficient of (4.12) negative, we assume that a coupling strength ${\kappa}$ is sufficiently large: ${\kappa}>\frac{\max\\{\mathcal{D}(\mathcal{H}),\mathcal{D}(\mathcal{G})\\}}{2(m-4\sqrt{n})}.$ For a handy notation, we denote $\Lambda:=2(m-4\sqrt{n})-\frac{\max\\{\mathcal{D}(\mathcal{H}),\mathcal{D}(\mathcal{G})\\}}{{\kappa}}.$ Next, we show that we can make the total aggregation functional $\mathcal{L}$ small as we wish by controlling the coupling strength ${\kappa}$. ###### Proposition 4.1. Suppose system parameters and initial data satisfy $\displaystyle\begin{aligned} &\textup{(i)}~{}~{}n\geq m,\quad\mathcal{D}(\mathcal{H})\geq\mathcal{D}(\mathcal{G}),\qquad\textup{(ii)}~{}~{}{\kappa}\geq{\kappa}_{\textup{c}},\quad\mathcal{L}^{0}\ll\nu_{2},\end{aligned}$ (4.13) where ${\kappa}_{\textup{c}}$ and $\nu_{2}$ are specified later in (4.18) and (4.19), respectively, and let $\\{(U_{j},V_{j})\\}$ be a solution to (4.1). Then, we have $\lim_{{\kappa}\to\infty}\limsup_{t\to\infty}\mathcal{L}(t)=0.$ ###### Proof. Since we assumed (4.13)(i), it follows from (4.6) that $\dot{\mathcal{L}}\leq-2{\kappa}(m-4\sqrt{n})\mathcal{L}+{\kappa}(4n+9)\mathcal{L}^{2}+{\kappa}\left(2n+\frac{8}{3}\right)\mathcal{L}^{3}+2(1+3\sqrt{n})\mathcal{D}(\mathcal{H}).$ (4.14) Now, we introduce an auxiliary cubic polynomial: $g(s):=2(m-4\sqrt{n})s-(4n+9)s^{2}-\left(2n+\frac{8}{3}\right)s^{3}.$ (4.15) Then, (4.14) can be rewritten as $\dot{\mathcal{L}}\leq{\kappa}\left(\frac{2(1+3\sqrt{n})\mathcal{D}(\mathcal{H})}{{\kappa}}-g(\mathcal{L})\right).$ By investigating roots of the polynomial $g$ in (4.15), we deduce that for a sufficient large ${\kappa}$, the polynomial $g$ admits one negative root, say, $\nu_{0}<0$, and two positive roots, say, $0<\nu_{1}<\nu_{2}$ with continuous dependence of ${\kappa}$, i.e., $\lim_{{\kappa}\to\infty}\nu_{1}({\kappa})=0,\quad\lim_{{\kappa}\to\infty}\nu_{2}({\kappa})=\alpha_{*},\quad\textup{$\alpha_{*}$: a unique positive root of $g$}.$ (4.16) Since we assume (4.13)(ii), there exists a finite entrance time $T_{*}>0$ such that $\mathcal{L}(t)<\nu_{1},\quad t>T_{*}.$ (4.17) Finally, we combine (4.16) and (4.17) to obtain the desired estimate: $\lim_{{\kappa}\to\infty}\limsup_{t\to\infty}\mathcal{L}(t)=0.$ ∎ ###### Remark 4.2. For an explicit value for ${\kappa}_{\textup{c}}$ in (4.13)(ii), by simple calculus, we know that the polynomial $g=g(s)$ in (4.15) admits a global maximum in $\mathbb{R}_{+}$ at $s=s_{*}$: $s_{*}:=\frac{-(4n+9)+\sqrt{(4n+9)^{2}+2(6n+8)(m-4\sqrt{n})}}{6n+8}.$ Thus, in order to guarantee the existence of a positive root for $g$, one should impose $g(s_{*})>\frac{2(1+3\sqrt{n})\mathcal{D}(\mathcal{H})}{{\kappa}},\quad\textup{i.e.,}\quad{\kappa}>\frac{2(1+3\sqrt{n})\mathcal{D}(\mathcal{H})}{g(s_{*})}=:{\kappa}_{\textup{c}}.$ (4.18) For this ${\kappa}_{\textup{c}}$, $\nu_{2}$ in (4.13)(ii) can be explicitly determined as the largest positive root of $g(s)=\frac{2(1+3\sqrt{n})\mathcal{D}(\mathcal{H})}{{\kappa}}\quad\textup{when ${\kappa}>{\kappa}_{\textup{c}}$}.$ (4.19) It follows from Proposition 4.1 that under the assumption on the smallness of the initial data $\mathcal{L}^{0}$ and largeness of the coupling strength ${\kappa}$, we have $\lim_{{\kappa}\to\infty}\limsup_{t\to\infty}\mathcal{L}(t)=0.$ Hence, there exists a (large) coupling strength ${\kappa}_{\textup{p}}$ larger than ${\kappa}_{\textup{c}}$ such that for $t\gg 1$, $(4n+22)\mathcal{L}(t)+20\mathcal{L}(t)^{2}<\frac{\Lambda}{2},\quad\Lambda=2(m-4\sqrt{n})-\frac{\max\\{\mathcal{D}(\mathcal{H}),\mathcal{D}(\mathcal{G})\\}}{{\kappa}}.$ Then, (4.12) becomes $\displaystyle\frac{{\textup{d}}}{{\textup{d}t}}\mathcal{F}$ $\displaystyle\leq-\Lambda{\kappa}\mathcal{F}+{\kappa}(4n+22)\mathcal{L}\mathcal{F}+20{\kappa}\mathcal{L}^{2}\mathcal{F}\leq-\Lambda{\kappa}\mathcal{F}+\frac{\Lambda{\kappa}}{2}\mathcal{F}=-\frac{\Lambda{\kappa}}{2}\mathcal{F}.$ This yields a desired exponential decay for $\mathcal{F}$. The argument above can be stated and shown as follows. ###### Theorem 4.2. Suppose system parameters and initial data satisfy $\displaystyle\begin{aligned} &\textup{(i)}~{}~{}n\geq m,\quad\mathcal{D}(\mathcal{H})\geq\mathcal{D}(\mathcal{G}),\qquad&\textup{(ii)}~{}~{}{\kappa}>{\kappa}_{\textup{p}}>{\kappa}_{\textup{c}},\quad\max\\{\mathcal{L}^{0},\tilde{\mathcal{L}}^{0}\\}\leq\nu_{2},\end{aligned}$ and let $\\{(U_{i},V_{i})\\}$ and $\\{(\tilde{U}_{i},\tilde{V}_{i})\\}$ be any two solutions to (4.1), respectively. Then, the following assertions hold. 1. (1) The functional $\mathcal{F}$ converges to zero with an exponential rate. 2. (2) The normalized velocities $\mathrm{i}\dot{U}_{j}U_{j}^{\dagger}$ and $\mathrm{i}\dot{V}_{j}V_{j}^{\dagger}$ synchronize: $\lim_{t\to\infty}\|\mathrm{i}\dot{U}_{j}U_{j}^{\dagger}-\mathrm{i}\dot{\tilde{U}}_{j}\tilde{U}_{j}^{\dagger}\|_{\textup{F}}=0,\quad\lim_{t\to\infty}\|\mathrm{i}\dot{V}_{j}V_{j}^{\dagger}-\mathrm{i}\dot{\tilde{V}}_{j}\tilde{V}_{j}^{\dagger}\|_{\textup{F}}=0.$ 3. (3) There exist unitary matrices $X_{\infty}\in\mathbf{U}(n)$ and $Y_{\infty}\in\mathbf{U}(m)$ such that $\displaystyle\lim_{t\to\infty}U_{i}^{\dagger}(t)\tilde{U}_{i}(t)=X_{\infty},\quad\lim_{t\to\infty}\|\tilde{U}_{i}(t)-U_{i}(t)X_{\infty}\|_{\textup{F}}=0,$ $\displaystyle\lim_{t\to\infty}V_{i}^{\dagger}(t)\tilde{V}_{i}(t)=Y_{\infty},\quad\lim_{t\to\infty}\|\tilde{V}_{i}(t)-V_{i}(t)Y_{\infty}\|_{\textup{F}}=0.$ 4. (4) Asymptotic phase-locking emerge: for any indices $i$ and $j$, $\exists~{}\lim_{t\to\infty}U_{i}(t)U_{j}^{\dagger}(t)\quad\textup{and}\quad\exists~{}\lim_{t\to\infty}U_{i}(t)U_{j}^{\dagger}(t).$ Moreover, there exist phase-locked state $\mathcal{X}^{\infty}:=\\{X_{i}^{\infty}\\}_{i=1}^{N}$ and $\mathcal{Y}^{\infty}:=\\{Y_{i}^{\infty}\\}_{i=1}^{N}$, and unitary matrices $P\in\mathbf{U}(n)$ and $Q\in\mathbf{U}(m)$ such that $\displaystyle\lim_{t\to\infty}\|U_{i}(t)-X_{i}^{\infty}P\|_{\textup{F}}=0=\lim_{t\to\infty}\|V_{i}(t)-Y_{i}^{\infty}Q\|_{\textup{F}}=0,\quad\max\\{\mathcal{D}(\mathcal{X}^{\infty}),\mathcal{D}(\mathcal{Y}^{\infty})\\}<\nu_{2}.$ ###### Proof. (i) It follows from (4.12) in Proposition 4.1 that $\mathcal{F}$ satisfies $\frac{{\textup{d}}}{{\textup{d}t}}{\mathcal{F}}\leq-{\kappa}\left(2m-8\sqrt{n}-\frac{\max\\{\mathcal{D}(\mathcal{H}),\mathcal{D}(\mathcal{G})\\}}{{\kappa}}\right)\mathcal{F}+{\kappa}(4n+22)\mathcal{L}\mathcal{F}+20{\kappa}\mathcal{L}^{2}\mathcal{F}.$ (4.20) Since $\mathcal{L}$ can be sufficiently small under (4.13), i.e., $\mathcal{L}(t)<\nu_{1},\quad t>T_{*},$ we choose ${\kappa}_{\textup{p}}>{\kappa}_{\textup{c}}$ sufficiently large such that $(4n+22)\mathcal{L}+20\mathcal{L}^{2}<\frac{\Lambda}{2},\quad\Lambda=2(m-4\sqrt{n})-\frac{\max\\{\mathcal{D}(\mathcal{H}),\mathcal{D}(\mathcal{G})\\}}{{\kappa}}.$ Thus, (4.20) becomes $\frac{{\textup{d}}}{{\textup{d}t}}\mathcal{F}\leq-\frac{\Lambda{\kappa}}{2}\mathcal{F},\quad t>T_{*},$ (4.21) and the relation (4.21) yields the desired exponential convergence of $\mathcal{F}$ toward zero. (ii) For the second assertion, we claim: $\|\mathrm{i}\dot{U}_{j}U_{j}^{\dagger}-\mathrm{i}\dot{\tilde{U}}_{j}\tilde{U}_{j}^{\dagger}\|_{\textup{F}}\leq 2{\kappa}(m+n)\mathcal{F},\qquad\|\mathrm{i}\dot{V}_{j}V_{j}^{\dagger}-\mathrm{i}\dot{\tilde{V}}_{j}\tilde{V}_{j}^{\dagger}\|_{\textup{F}}\leq 2{\kappa}(m+n)\mathcal{F}.$ (4.22) Once the relations (4.22) hold, it follows from the first assertion to derive the second assertion. Note that $\displaystyle\begin{aligned} &\|\mathrm{i}\dot{U}_{j}U_{j}^{\dagger}-\mathrm{i}\dot{\tilde{U}}_{j}\tilde{U}_{j}^{\dagger}\|_{\textup{F}}=\left\|\frac{{\kappa}}{N}\sum_{k=1}^{N}\Big{(}\langle V_{j},V_{k}\rangle_{\textup{F}}U_{k}U_{j}^{\dagger}-\langle V_{k},V_{j}\rangle_{\textup{F}}U_{j}U_{k}^{\dagger}\Big{)}\right.\\\ &\hskip 142.26378pt\left.-\frac{{\kappa}}{N}\sum_{k=1}^{N}\Big{(}\langle\tilde{V}_{j},\tilde{V}_{k}\rangle_{\textup{F}}\tilde{U}_{k}\tilde{U}_{j}^{\dagger}-\langle\tilde{V}_{k},\tilde{V}_{j}\rangle_{\textup{F}}\tilde{U}_{j}\tilde{U}_{k}^{\dagger}\Big{)}\right\|_{\textup{F}}.\end{aligned}$ (4.23) By algebraic manipulations, one has $\displaystyle\|\langle V_{j},V_{k}\rangle_{\textup{F}}U_{k}U_{j}^{\dagger}-\langle\tilde{V}_{j},\tilde{V}_{k}\rangle_{\textup{F}}\tilde{U}_{k}\tilde{U}_{j}^{\dagger}\|_{\textup{F}}$ $\displaystyle\hskip 14.22636pt\leq\|\langle V_{j},V_{k}\rangle_{\textup{F}}(U_{k}U_{j}^{\dagger}-\tilde{U}_{k}\tilde{U}_{j}^{\dagger})\|_{\textup{F}}+\|(\langle V_{j},V_{k}\rangle_{\textup{F}}-\langle\tilde{V}_{j},\tilde{V}_{k}\rangle_{\textup{F}})\tilde{U}_{k}\tilde{U}_{j}^{\dagger}\|_{\textup{F}}\leq(m+n)\mathcal{F}.$ Thus, the relation (4.23) yields $\|\mathrm{i}\dot{U}_{j}U_{j}^{\dagger}-\mathrm{i}\dot{\tilde{U}}_{j}\tilde{U}_{j}^{\dagger}\|_{\textup{F}}\leq 2{\kappa}(m+n)\mathcal{F}.$ This shows the desired synchronization of the normalized velocities and the exactly same argument is applied to $\mathrm{i}\dot{V}_{j}V_{j}^{\dagger}$. (iii) Since system (4.1) is autonomous, we directly use Theorem 2(3) of [16]. Hence, we briefly sketch a proof. We observe $\displaystyle\left\|\frac{{\textup{d}}}{{\textup{d}}s}(U_{i}^{\dagger}\tilde{U}_{i})\right\|_{\textup{F}}=\|\dot{U}_{i}U_{i}^{\dagger}-\dot{\tilde{U}}_{i}\tilde{U}_{i}^{\dagger}\|_{\textup{F}}\leq 2{\kappa}(m+n)\mathcal{F}.$ Since $\mathcal{F}$ tends to zero exponentially, $\lim_{t\to\infty}(U_{i}^{\dagger}(t)\tilde{U}_{i}(t))=U_{i}^{0,{\dagger}}\tilde{U}_{i}^{0}+\int_{0}^{\infty}\frac{{\textup{d}}}{{\textup{d}}s}(U_{i}^{\dagger}(s)\tilde{U}_{i}(s)){\textup{d}}s.$ (4.24) In addition, since $d(U,\tilde{U})$ converges to zero, one deduces that the right-hand side of (4.24) does not depend on the index $i$. This establishes the desired assertion. (iv) We first show the existence of the asymptotic limit of $U_{i}U_{j}^{\dagger}$ and $V_{i}V_{j}^{\dagger}$. For any $T>0$, since system (4.1) is autonomous, we choose $\tilde{U}_{i}$ and $\tilde{V}_{i}$ as $\tilde{U}_{i}(t)=U_{i}(t+T),\quad V_{i}(t)=V_{i}(t+T).$ If we discretize the time $t\in\mathbb{R}_{+}$ and $n\in\mathbb{Z}_{+}$ and choose $T=m\in\mathbb{Z}_{+}$, we use the convergence of $d(U,\tilde{U})$ and $d(V,\tilde{V})$ to conclude that $\\{U_{i}(n)U_{j}^{\dagger}(n)\\}_{n\in\mathbb{Z}_{+}}$ and $\\{V_{i}(n)V_{j}^{\dagger}(n)\\}_{n\in\mathbb{Z}_{+}}$ become Cauchy sequences in the complete spaces $\mathbf{U}(n)$ and $\mathbf{U}(m)$, respectively. Thus, the limits of $U_{i}U_{j}^{\dagger}$ and $V_{i}V_{j}^{\dagger}$ exist. In particular, if we denote $X_{i}^{\infty}:=\lim_{t\to\infty}U_{i}(t)U_{1}^{\dagger}(t),$ then one has $\lim_{t\to\infty}U_{i}(t)U_{j}^{\dagger}(t)=\lim_{t\to\infty}U_{i}U_{1}^{\dagger}(t)(U_{j}(t)U_{1}^{\dagger}(t))^{\dagger}=X_{i}^{\infty}X_{j}^{\infty,{\dagger}},\quad\lim_{t\to\infty}\langle U_{i},U_{j}\rangle_{\textup{F}}=\langle X_{i}^{\infty},X_{j}^{\infty}\rangle_{\textup{F}}.$ On the other hand, we recall (A.2) $\displaystyle\begin{aligned} &\frac{{\textup{d}}}{{\textup{d}t}}(U_{i}U_{j}^{\dagger})=\mathrm{i}(U_{i}U_{j}^{\dagger}H_{j}-H_{i}U_{i}U_{j}^{\dagger})\\\ &+\frac{{\kappa}}{N}\sum_{k=1}^{N}\Big{(}\langle V_{i},V_{k}\rangle_{\textup{F}}U_{k}U_{j}^{\dagger}-\langle V_{k},V_{i}\rangle_{\textup{F}}U_{i}U_{k}^{\dagger}U_{i}U_{j}^{\dagger}+\langle V_{k},V_{j}\rangle_{\textup{F}}U_{i}U_{k}^{\dagger}-\langle V_{j},V_{k}\rangle_{\textup{F}}U_{i}U_{j}^{\dagger}U_{k}U_{j}^{\dagger}\Big{)}.\end{aligned}$ (4.25) In (4.25), if we let $t\to\infty$ and apply Barbalat’s lemma, the left-hand side of (4.25) vanishes and consequently, the relation (4.25) becomes $\displaystyle\begin{aligned} O&=\mathrm{i}(X_{i}^{\infty}X_{j}^{\infty,{\dagger}}H_{j}-H_{i}X_{i}^{\infty}X_{j}^{\infty,{\dagger}})\\\ &\hskip 14.22636pt+\frac{{\kappa}}{N}\sum_{k=1}^{N}\Big{(}\langle Y_{i}^{\infty},Y_{k}^{\infty}\rangle_{\textup{F}}X_{k}^{\infty}X_{j}^{\infty,{\dagger}}-\langle Y_{k}^{\infty},Y_{i}^{\infty,{\dagger}}\rangle_{\textup{F}}X_{i}^{\infty}X_{k}^{\infty,{\dagger}}X_{i}^{\infty}X_{j}^{\infty,{\dagger}}\\\ &\hskip 14.22636pt+\langle Y_{k}^{\infty},Y_{j}^{\infty}\rangle_{\textup{F}}X_{i}^{\infty}X_{k}^{\infty,{\dagger}}-\langle Y_{j}^{\infty},Y_{k}^{\infty,{\dagger}}\rangle_{\textup{F}}X_{i}^{\infty}X_{j}^{\infty,{\dagger}}X_{k}^{\infty}X_{j}^{\infty,{\dagger}}\Big{)}.\end{aligned}$ (4.26) By performing left-multiplication $X_{i}^{\infty,{\dagger}}$ and right- multiplication $X_{j}^{\infty}$ with (4.26), we obtain $\displaystyle\begin{aligned} &-\mathrm{i}X_{j}^{\infty,{\dagger}}H_{j}X_{j}^{\infty}+\frac{{\kappa}}{N}\sum_{k=1}^{N}\Big{(}\langle Y_{j}^{\infty},Y_{k}^{\infty}\rangle_{\textup{F}}X_{j}^{\infty,{\dagger}}X_{k}^{\infty}-\langle Y_{k}^{\infty},Y_{j}^{\infty}\rangle_{\textup{F}}X_{k}^{\infty,{\dagger}}X_{j}^{\infty}\Big{)}\\\ &\hskip 14.22636pt=-\mathrm{i}X_{i}^{\infty,{\dagger}}H_{i}X_{i}^{\infty}+\frac{{\kappa}}{N}\sum_{k=1}^{N}\Big{(}\langle Y_{i}^{\infty},Y_{k}^{\infty}\rangle_{\textup{F}}X_{i}^{\infty,{\dagger}}X_{k}^{\infty}-\langle Y_{k}^{\infty},Y_{i}^{\infty}\rangle_{\textup{F}}X_{k}^{\infty,{\dagger}}X_{i}^{\infty}\Big{)}.\end{aligned}$ (4.27) Since the relation(4.27) does not depend on the index, we can set $-\mathrm{i}\Theta:=-\mathrm{i}X_{j}^{\infty,{\dagger}}H_{j}X_{j}^{\infty}+\frac{{\kappa}}{N}\sum_{k=1}^{N}\Big{(}\langle Y_{j}^{\infty},Y_{k}^{\infty}\rangle_{\textup{F}}X_{j}^{\infty,{\dagger}}X_{k}^{\infty}-\langle Y_{k}^{\infty},Y_{j}^{\infty}\rangle_{\textup{F}}X_{k}^{\infty,{\dagger}}X_{j}^{\infty}\Big{)}.$ This yields $X_{j}^{\infty}\Theta X_{j}^{\infty,{\dagger}}=H_{j}+\frac{{\kappa}}{N}\sum_{k=1}^{N}\Big{(}\langle Y_{k}^{\infty},Y_{j}^{\infty}\rangle_{\textup{F}}X_{k}^{\infty,{\dagger}}X_{j}^{\infty}-\langle Y_{j}^{\infty},Y_{k}^{\infty}\rangle_{\textup{F}}X_{j}^{\infty,{\dagger}}X_{k}^{\infty}\Big{)}.$ Similarly for $\\{V_{i}\\}$, we can define a matrix $\Gamma$ independent of the index such that $Y_{j}^{\infty}\Gamma Y_{j}^{\infty,{\dagger}}=G_{j}++\frac{{\kappa}}{N}\sum_{k=1}^{N}\Big{(}\langle X_{k}^{\infty},X_{j}^{\infty}\rangle_{\textup{F}}Y_{k}^{\infty,{\dagger}}Y_{j}^{\infty}-\langle X_{j}^{\infty},X_{k}^{\infty}\rangle_{\textup{F}}Y_{j}^{\infty,{\dagger}}Y_{k}^{\infty}\Big{)}.$ Then, $\\{\\{X_{i}^{\infty}\\},\Theta\\}$ and $\\{\\{Y_{i}^{\infty}\\},\Gamma\\}$ indeed consist of the phase-locked state. Moreover, one has $\mathcal{D}(\mathcal{X}^{\infty})=\max_{1\leq i,j\leq N}\|X_{i}^{\infty}X_{j}^{\infty,{\dagger}}-I_{n}\|=\lim_{t\to\infty}\max_{1\leq i,j\leq N}\|U_{i}(t)U_{j}^{\dagger}(t)-I_{n}\|_{\textup{F}}=\lim_{t\to\infty}\mathcal{D}(\mathcal{U})<\nu_{2}.$ Exactly the same estimate holds for $\mathcal{D}(\mathcal{Y}^{\infty})$ as well. ∎ ###### Remark 4.3. In [20], the phase-locked state of (4.1) are defined to be of the following form: $U_{i}(t)=U_{i}^{\infty}e^{-\mathrm{i}\Gamma_{U}t},\quad V_{i}(t)=V_{i}^{\infty}e^{-\mathrm{i}\Gamma_{V}t},$ where $U_{i}^{\infty}\in\mathbf{U}(n)$ and $V_{i}^{\infty}\in\mathbf{U}(m)$ are unitary matrices, and $\Lambda_{U}$ and $\Lambda_{V}$ satisfy $\displaystyle\begin{aligned} &U_{i}^{\infty}\Gamma_{U}U_{i}^{\infty,{\dagger}}=H_{i}+\frac{\mathrm{i}{\kappa}}{N}\sum_{k=1}^{N}\Big{(}\langle V_{k}^{\infty},V_{j}^{\infty}\rangle_{\textup{F}}U_{k}^{\infty}U_{i}^{\infty,{\dagger}}-\langle V_{j}^{\infty},V_{k}^{\infty}\rangle_{\textup{F}}U_{i}^{\infty}U_{k}^{\infty,{\dagger}}\Big{)},\\\ &V_{i}^{\infty}\Gamma_{V}V_{i}^{\infty,{\dagger}}=G_{i}+\frac{\mathrm{i}{\kappa}}{N}\sum_{k=1}^{N}\Big{(}\langle U_{k}^{\infty},U_{j}^{\infty}\rangle_{\textup{F}}V_{k}^{\infty}V_{i}^{\infty,{\dagger}}-\langle U_{j}^{\infty},U_{k}^{\infty}\rangle_{\textup{F}}V_{i}^{\infty}V_{k}^{\infty,{\dagger}}\Big{)}.\end{aligned}$ ## 5\. Rank-$2m$ quadratically separable state In this section, we study a quadratically separable state of the LT model for rank-$2m$ tensors by introducing the MM model whose solution configuration is given as follows. $\\{\mathcal{U}^{1},\mathcal{U}^{2},\cdots,\mathcal{U}^{m}\\},\quad\mathcal{U}^{p}:=(U_{1}^{p},\cdots,U_{N}^{p}),\quad p=1,\cdots,m.$ In the following two subsections, we introduce extended models for the DM model (3.5) and the DUM model (3.15). Since the procedures are similar as those in Section 3, we omit details. ### 5.1. The MM model In this subsection, we introduce the MM model: $\displaystyle\begin{cases}\displaystyle\dot{U}_{j}^{p}=B_{j}^{p}U_{j}^{p}+\displaystyle\frac{\kappa_{1}}{N}\sum_{k=1}^{N}\left(\prod_{\begin{subarray}{c}\ell=1\\\ \ell\neq p\end{subarray}}^{m}\langle U_{j}^{\ell},U_{k}^{\ell}\rangle_{\textup{F}}U_{k}^{p}\big{(}U_{j}^{p}\big{)}^{\dagger}U_{j}^{p}-\prod_{\begin{subarray}{c}\ell=1\\\ \ell\neq p\end{subarray}}^{m}\langle U_{k}^{l},U_{j}^{l}\rangle_{\textup{F}}U_{j}^{p}\big{(}U_{k}^{p}\big{)}^{\dagger}U_{j}^{p}\right)\\\ \hskip 28.45274pt\displaystyle+\frac{\kappa_{2}}{N}\sum_{k=1}^{N}\left(\prod_{\begin{subarray}{c}\ell=1\\\ \ell\neq p\end{subarray}}^{m}\langle U_{j}^{\ell},U_{k}^{\ell}\rangle_{\textup{F}}U_{j}^{p}\big{(}U_{j}^{p}\big{)}^{\dagger}U_{k}^{p}-\prod_{\begin{subarray}{c}\ell=1\\\ \ell\neq p\end{subarray}}^{m}\langle U_{k}^{\ell},U_{j}^{\ell}\rangle_{\textup{F}}U_{j}^{p}\big{(}U_{k}^{p}\big{)}^{\dagger}U_{j}^{p}\right),\end{cases}$ (5.1) where $B_{j}^{p}$ is a rank-$4m$ tensor satisfying the skew-symmetric property. For a solution $\\{\mathcal{U}^{p}\\}_{p=1}^{m}$ to (5.1), we set a rank-$2m$ tensor denoted by $T_{i}$: $T_{i}(t)=U_{i}^{1}(t)\otimes U_{i}^{2}(t)\otimes\cdots\otimes U_{i}^{m}(t),\quad U_{i}^{p}\in\mathbb{C}^{d_{1}^{p}\times d_{2}^{p}},\quad i=1,\cdots,N,\quad p=1,\cdots,m,$ (5.2) which can be written in a component form: $[T_{i}]_{\alpha_{1}\beta_{1}\alpha_{2}\beta_{2}\cdots\alpha_{m}\beta_{m}}=[U_{i}^{1}]_{\alpha_{1}\beta_{1}}[U_{i}^{2}]_{\alpha_{2}\beta_{2}}\cdots[U_{i}^{m}]_{\alpha_{m}\beta_{m}}.$ In order to relate the model (5.1) with the LT model (1.1), we consider the index vector $i_{*}$ in ${\kappa}_{i_{*}}$. Then, we introduce two subsets of $\\{0,1\\}^{2d}$: for $q=1,\cdots,m$, $\displaystyle\Lambda_{1}$ $\displaystyle=\\{i_{*}\in\\{0,1\\}^{2d}:\text{0 appears once at $2q-1$-th coordinate}\\},$ $\displaystyle\Lambda_{2}$ $\displaystyle=\\{i_{*}\in\\{0,1\\}^{2d}:\text{0 appears once at $2q$-th coordinate}\\}.$ Note that $|\Lambda_{1}|=|\Lambda_{2}|=m$ and for instance with $m=2$, $(0,1,1,1),(1,1,0,1)\in\Lambda_{1},\quad(1,0,1,1),(1,1,1,0)\in\Lambda_{2}.$ In this regard, we choose the index vector as $\displaystyle\kappa_{i_{*}}=\begin{cases}\kappa_{1}\quad&\text{if }i_{*}\in\Lambda_{1},\\\ \kappa_{2}\quad&\text{if }i_{*}\in\Lambda_{2},\\\ 0&\text{otherwise},\end{cases}$ (5.3) which generalizes (3.3) for the DM model. Next, for the natural frequency tensors, we use $\\{B_{j}^{p}\\}_{p=1}^{m}$ to associate a rank-$4m$ tensor $A_{j}$ as $\displaystyle[A_{j}]_{\alpha_{1}\alpha_{2}\cdots\alpha_{2m}\beta_{1}\beta_{2}\cdots\beta_{2m}}=\sum_{k=1}^{m}\left([B_{j}^{k}]_{\alpha_{2k-1}\alpha_{2k}\beta_{2k-1}\beta_{2k}}\prod^{d}_{\begin{subarray}{c}\ell=1\\\ \ell\neq k\end{subarray}}\left(\delta_{\alpha_{2\ell-1}\beta_{2\ell-1}}\delta_{\alpha_{2\ell}\beta_{2\ell}}\right)\right),$ (5.4) which corresponds to (3.9) for the DM model. Then, it follows from straightforward calculation that the Lohe tensor model with (5.3) and (5.4) reduces to (5.1) whose solutions can be related by (5.2). The argument above is summarized in the following proposition analogous to Proposition 3.1. ###### Proposition 5.1. The following assertions hold. 1. (1) Suppose $\\{\mathcal{U}^{p}\\}_{p=1}^{m}$ is a solution to (5.1). Then, a rank-$2m$ tensor defined by $T_{i}:=U_{i}^{1}\otimes U_{i}^{2}\otimes\cdots\otimes U_{i}^{m}$ is the QSS to (1.1) with well- prepared initial data and free flow tensors $A_{i}$ satisfying (5.4). 2. (2) Suppose a rank-$2m$ tensor $T_{i}$ is a solution to (1.1) with (5.4) and quadratically separable initial data: $T_{i}^{0}:=U_{i}^{1,0}\otimes U_{i}^{2,0}\otimes\cdots\otimes U_{i}^{p,0},\quad 1\leq i\leq N,$ for rank-2 tensors $U_{i}^{p,0}\times\mathbb{C}^{d_{1}^{p}\times d_{2}^{p}}$ with unit norms. Then, there exist matrices $\\{U_{i}^{p}\\}$ with unit norms such that $T_{i}(t)=U_{i}^{1}(t)\otimes U_{i}^{2}(t)\otimes\cdots\otimes U_{i}^{p}(t),\quad t>0,$ where $\\{U_{i}^{p}\\}$ is a solution to (5.1) with $U_{i}^{p,0}=U_{i}^{p}(0)$. ### 5.2. The MUM model In this subsection, we further reduce to the MM model to the model on the product of the unitary groups. Since the unitary group is concerned, we set $d_{1}^{p}=d_{2}^{p}=:d_{p},\quad 1\leq p\leq m.$ (5.5) For the modeling of natural frequencies, we also define Hermitian matrices with a size $d_{p}\times d_{p}$: $[-\mathrm{i}H_{j}^{p}]_{\alpha_{1}\alpha_{2}}\delta_{\beta_{1}\beta_{2}}:=[B_{j}^{p}]_{\alpha_{1}\beta_{1}\alpha_{2}\beta_{2}}.$ In addition, $\\{H_{j}^{p}\\}$ satisfy $\displaystyle[A_{j}]_{\alpha_{1}\alpha_{2}\cdots\alpha_{2m}\beta_{1}\beta_{2}\cdots\beta_{2m}}$ $\displaystyle=\sum_{k=1}^{m}\left([-\mathrm{i}H_{j}^{k}]_{\alpha_{2k-1}\beta_{2k-1}}\delta_{\alpha_{2k}\beta_{2k}}\prod^{m}_{\begin{subarray}{c}\ell=1\\\ \ell\neq k\end{subarray}}\left(\delta_{\alpha_{2\ell-1}\beta_{2\ell-1}}\delta_{\alpha_{2\ell}\beta_{2\ell}}\right)\right).$ (5.6) By Lemma 3.2, one can verify that system (5.1) with (5.5) conserves the unitarity of $U_{j}^{p}$. Thus, system (5.1) reduces to the following model on the unitary group: $\displaystyle\begin{cases}\displaystyle\dot{U}_{j}^{p}=-\mathrm{i}H_{j}^{p}U_{j}^{p}+\displaystyle\frac{\kappa}{N}\sum_{k=1}^{N}\left(\prod_{\begin{subarray}{c}\ell=1\\\ \ell\neq p\end{subarray}}^{m}\langle U_{j}^{l},U_{k}^{l}\rangle_{\textup{F}}U_{k}^{p}-\prod_{\begin{subarray}{c}\ell=1\\\ \ell\neq p\end{subarray}}^{m}\langle U_{k}^{\ell},U_{j}^{\ell}\rangle_{\textup{F}}U_{j}^{p}\big{(}U_{k}^{p}\big{)}^{\dagger}U_{j}^{p}\right),\\\ U_{j}^{p}(0)=U_{j}^{p,0}\in\mathbf{U}(d_{p}),\end{cases}$ (5.7) where $\kappa=\kappa_{1}+\kappa_{2}$ and $H_{j}^{p}U_{j}^{p}$ is a usual matrix product. As in Proposition 5.1, existence and uniqueness of the QSS for (5.7) can be stated as follows. ###### Proposition 5.2. The following assertions hold. 1. (1) Suppose $\\{\mathcal{U}^{p}\\}_{p=1}^{m}$ is a solution to (5.7). Then, a rank-$2m$ tensor defined by $T_{i}:=U_{i}^{1}\otimes U_{i}^{2}\otimes\cdots\otimes U_{i}^{m}$ is a quadratically separable state to (1.1) with a well-prepared free flow tensor $A_{i}$ satisfying (5.6). 2. (2) Suppose a rank-$2m$ tensor $T_{i}$ is a solution to (1.1) with (5.6) and quadratically separable initial data: $T_{i}^{0}:=U_{i}^{1,0}\otimes U_{i}^{2,0}\otimes\cdots\otimes U_{i}^{p,0},\quad 1\leq i\leq N,$ for rank-2 tensors $U_{i}^{p,0}\times\mathbb{C}^{n_{1}^{p}\times n_{2}^{p}}$ with unit norms. Then, there exist matrices $\\{U_{i}^{p}\\}$ with unit norms such that $T_{i}(t)=U_{i}^{1}(t)\otimes U_{i}^{2}(t)\otimes\cdots\otimes U_{i}^{p}(t),\quad t>0,$ where $\\{U_{i}^{p}\\}$ is a solution to (5.7) with $U_{i}^{p,0}=U_{i}^{p}(0)$. ###### Remark 5.1. In Section 4, we provided the emergent dynamics of the double unitary matrix model (4.1). However for its generalized model (5.7), emergent dynamics will not be studied, since it can be straightforwardly obtained from the results in Section 4. ## 6\. Conclusion In this paper, we have studied the existence and emergent dynamics of quadratically separable states for the Lohe tensor model which incorporates several well-known low-rank aggregation models such as the Kuramoto model, the Lohe sphere model and the Lohe matrix model, etc. In our previous work [9], we obtained completely separable states as special solutions to the Lohe tensor model defined as tensor products of rank-1 real tensors (or vectors). In analogy with the previous work, we consider the states in which a solution can be decomposed as a tensor product of rank-2 tensors (or matrices), namely, a quadratically separable state. Precisely, if initial data are quadratically separable, then such separability is preserved along the Lohe tensor flow. Moreover, by introducing and analyzing double matrix and unitary matrix models, we are able to study asymptotic behavior of the quadratically separable states to the Lohe tensor model. Of course, there are several issues that are not discussed in this work. For example, one can naturally consider the state consisting of tensors with possibly different ranks and sizes. We explore this issue in a future work. ## Appendix A Proof of Lemma 4.1 In this appendix, we provide a proof of Lemma 4.1 in which a differential inequality for the aggregation functional $\mathcal{L}=\mathcal{D}(\mathcal{U})+\mathcal{D}(\mathcal{V})+\mathcal{S}(\mathcal{U})+\mathcal{S}(\mathcal{V})$ is derived. We divide a proof into two steps: * • Step A: we derive differential inequalities for $\mathcal{D}(\mathcal{U})$ and $\mathcal{D}(\mathcal{V})$ (see Lemma A.1). * • Step B: we derive differential inequalities for $\mathcal{S}(\mathcal{U})$ and $\mathcal{S}(\mathcal{V})$ (see Lemma A.2). ###### Lemma A.1. Let $\\{(U_{i},V_{i})\\}$ be a solution to (4.4). Then, $\mathcal{D}(\mathcal{U})$ and $\mathcal{D}(\mathcal{V})$ satisfy $\displaystyle\begin{aligned} &\frac{{\textup{d}}}{{\textup{d}t}}\mathcal{D}(\mathcal{U})\leq-2m{\kappa}\mathcal{D}(\mathcal{U})+m{\kappa}\mathcal{D}(\mathcal{U})^{3}+6{\kappa}\mathcal{S}(\mathcal{V})\mathcal{D}(\mathcal{U})+2{\kappa}\mathcal{S}(\mathcal{V})\mathcal{D}(\mathcal{U})^{2}+4{\kappa}\sqrt{n}S(\mathcal{V}),\\\ &\frac{{\textup{d}}}{{\textup{d}t}}\mathcal{D}(\mathcal{V})\leq-2n{\kappa}\mathcal{D}(\mathcal{V})+n{\kappa}\mathcal{D}(\mathcal{V})^{3}+6{\kappa}\mathcal{S}(\mathcal{U})\mathcal{D}(\mathcal{V})+2{\kappa}\mathcal{S}(\mathcal{U})\mathcal{D}(\mathcal{V})^{2}+4{\kappa}\sqrt{m}S(\mathcal{U}).\end{aligned}$ (A.1) ###### Proof. By straightforward calculations, one finds a differential equation for $G_{ij}=U_{i}U_{j}^{\dagger}$: $\frac{{\textup{d}}}{{\textup{d}t}}G_{ij}=\frac{{\kappa}}{N}\sum_{k=1}^{N}(c_{ik}G_{kj}-c_{ki}G_{ik}G_{ij}+c_{kj}G_{ik}-c_{jk}G_{ij}G_{kj}).$ (A.2) By using the relation $G_{ij}=I_{n}-S_{ij}$, we see that $S_{ij}$ satisfies $\displaystyle\begin{aligned} \frac{{\textup{d}}}{{\textup{d}t}}S_{ij}&=\frac{{\kappa}}{N}\sum_{k=1}^{N}\Big{[}c_{ik}S_{kj}-c_{ki}S_{ik}-c_{ki}S_{ij}+c_{i}S_{k}S_{ij}+c_{kj}S_{ik}-c_{jk}S_{kj}-c_{jk}S_{ij}+c_{jk}S_{ij}S_{kj}\\\ &\hskip 71.13188pt+(c_{ki}-c_{ik}+c_{jk}-c_{kj})I_{n}\Big{]}.\end{aligned}$ (A.3) After algebraic manipulation, we rewrite (A.3) in terms of $S_{ij}$ and $c_{ij}-m$ that are expected to converge to zero: $\displaystyle\begin{aligned} \frac{{\textup{d}}S_{ij}}{dt}&=-2m{\kappa}S_{ij}+\frac{m{\kappa}}{N}\sum_{k=1}^{N}\Big{(}S_{ik}S_{ij}+S_{ij}S_{kj}\Big{)}\\\ &\hskip 5.69046pt+\frac{{\kappa}}{N}\sum_{k=1}^{N}\Big{[}(c_{ik}-m)S_{kj}-(c_{ki}-m)S_{ik}-(c_{ik}-m)S_{ij}+(c_{ki}-m)S_{ik}S_{ij}\Big{]}\\\ &\hskip 5.69046pt+\frac{{\kappa}}{N}\sum_{k=1}^{N}\Big{[}(c_{kj}-m)S_{ik}-(c_{jk}-m)S_{kj}-(c_{jk}-m)S_{ij}+(c_{jk}-m)S_{ij}S_{kj}\Big{]}\\\ &\hskip 5.69046pt+\frac{{\kappa}}{N}\sum_{k=1}^{N}\Big{[}(c_{ki}-c_{ik}+c_{jk}-c_{kj})I_{n}\Big{]}.\end{aligned}$ (A.4) On the other hand for an $n\times n$ matrix $A$, one has $\frac{1}{2}\frac{{\textup{d}}}{{\textup{d}t}}\|A\|_{\textup{F}}^{2}=\frac{1}{2}\frac{{\textup{d}}}{{\textup{d}t}}\textup{tr}(AA^{\dagger})=\frac{1}{2}\textup{tr}(\dot{A}A^{\dagger}+A\dot{A}^{\dagger})=\textup{Re}\textup{tr}(\dot{A}A^{\dagger}).$ We multiply (A.4) with $S_{ij}^{\dagger}$ to find $\displaystyle\begin{aligned} &\frac{1}{2}\frac{{\textup{d}}}{{\textup{d}t}}\|S_{ij}\|_{\textup{F}}^{2}=-2m{\kappa}\|S_{ij}\|_{\textup{F}}^{2}+\frac{m{\kappa}}{N}\sum_{k=1}^{N}\textup{Re}\textup{tr}(S_{ik}S_{ij}S_{ij}^{\dagger}+S_{ij}S_{kj}S_{ij}^{\dagger})\\\ &\hskip 14.22636pt+\frac{{\kappa}}{N}\sum_{k=1}^{N}\textup{Re}[(c_{ik}-m)\textup{tr}(S_{kj}S_{ij}^{\dagger})]-\textup{Re}[(c_{ki}-m)\textup{tr}(S_{ik}S_{ij}^{\dagger})]-\textup{Re}(c_{ik}-m)\|S_{ij}\|_{\textup{F}}^{2}\\\ &\hskip 14.22636pt+\frac{{\kappa}}{N}\sum_{k=1}^{N}\textup{Re}[(c_{kj}-m)\textup{tr}(S_{ik}S_{ij}^{\dagger})]-\textup{Re}[(c_{jk}-m)\textup{tr}(S_{kj}S_{ij}^{\dagger})]-\textup{Re}(c_{jk}-m)\|S_{ij}\|_{\textup{F}}^{2}\\\ &\hskip 14.22636pt+\frac{{\kappa}}{N}\sum_{k=1}^{N}\textup{Re}[(c_{ki}-m)\textup{tr}(S_{ik}S_{ij}S_{ij}^{\dagger})]+\textup{Re}[(c_{jk}-m)\textup{tr}(S_{ij}S_{kj}S_{ij}^{\dagger})]\\\ &\hskip 14.22636pt+\frac{{\kappa}}{N}\sum_{k=1}^{N}\textup{Re}[(c_{ki}-c_{ik}+c_{jk}-c_{kj})\textup{tr}(S_{ij}^{\dagger})]\\\ &=:-2m{\kappa}\|S_{ij}\|_{\textup{F}}^{2}+\mathcal{I}_{11}+\mathcal{I}_{12}+\mathcal{I}_{13}+\mathcal{I}_{14}+\mathcal{I}_{15}.\end{aligned}$ (A.5) Below, we present estimates for $\mathcal{I}_{1k},~{}k=1,\cdots,5$, respectively. $\bullet$ (Estimate of $\mathcal{I}_{11}$): We use (A.6) and $S_{ij}+S_{ji}=S_{ij}S_{ji},\quad S_{ij}^{\dagger}=S_{ji}$ (A.6) to derive $\displaystyle\textup{Re}\textup{tr}(S_{ik}S_{ij}S_{ji})$ $\displaystyle=\frac{1}{2}\textup{tr}(S_{ik}S_{ij}S_{ji}+S_{ij}S_{ji}S_{ki})=\frac{1}{2}\textup{tr}(S_{ij}S_{ji}(S_{ki}+S_{ik}))$ $\displaystyle=\frac{1}{2}\textup{tr}(S_{ij}S_{ji}S_{ki}S_{ik})=\frac{1}{2}\|S_{ki}S_{ij}\|_{\textup{F}}^{2}.$ Similarly, one has $\textup{Re}\textup{tr}(S_{ij}S_{kj}S_{ij}^{\dagger})=\frac{1}{2}\|S_{ij}S_{jk}\|_{\textup{F}}^{2}.$ Hence, $\mathcal{I}_{11}$ satisfies $\mathcal{I}_{11}=\frac{m{\kappa}}{N}\sum_{k=1}^{N}\textup{Re}\textup{tr}(S_{ik}S_{ij}S_{ij}^{\dagger}+S_{ij}S_{kj}S_{ij}^{\dagger})=\frac{m{\kappa}}{2N}\sum_{k=1}^{N}\big{(}\|S_{ki}S_{ij}\|_{\textup{F}}^{2}+\|S_{ij}S_{jk}\|_{\textup{F}}^{2}\big{)}\leq m{\kappa}\mathcal{D}(\mathcal{U})^{4}.$ $\bullet$ (Estimates of $\mathcal{I}_{12}$ and $\mathcal{I}_{13}$): By the maximality of $\mathcal{D}(\mathcal{U})$ and $\mathcal{S}(\mathcal{V})$, we have $\displaystyle\mathcal{I}_{12}$ $\displaystyle=\frac{{\kappa}}{N}\sum_{k=1}^{N}\textup{Re}[(c_{ik}-m)\textup{tr}(S_{kj}S_{ij}^{\dagger})]-\textup{Re}[(c_{ki}-m)\textup{tr}(S_{ik}S_{ij}^{\dagger})]-\textup{Re}(c_{ik}-m)\|S_{ij}\|_{\textup{F}}^{2}$ $\displaystyle\leq 3{\kappa}\mathcal{S}(\mathcal{V})\mathcal{D}(\mathcal{U})^{2},$ $\displaystyle\mathcal{I}_{13}$ $\displaystyle=\frac{{\kappa}}{N}\sum_{k=1}^{N}\textup{Re}[(c_{kj}-m)\textup{tr}(S_{ik}S_{ij}^{\dagger})]-\textup{Re}[(c_{jk}-m)\textup{tr}(S_{kj}S_{ij}^{\dagger})]-\textup{Re}(c_{jk}-m)\|S_{ij}\|_{\textup{F}}^{2}$ $\displaystyle\leq 3{\kappa}\mathcal{S}(\mathcal{V})\mathcal{D}(\mathcal{U})^{2}.$ $\bullet$ (Estimate of $\mathcal{I}_{14}$): By straightforward calculation, one has $\displaystyle\mathcal{I}_{14}$ $\displaystyle=\frac{{\kappa}}{N}\sum_{k=1}^{N}\textup{Re}[(c_{ki}-m)\textup{tr}(S_{ik}S_{ij}S_{ij}^{\dagger})]+\textup{Re}[(c_{jk}-m)\textup{tr}(S_{ij}S_{kj}S_{ij}^{\dagger})]\leq 2{\kappa}\mathcal{S}(\mathcal{V})\mathcal{D}(\mathcal{U})^{3}.$ $\bullet$ (Estimate of $\mathcal{I}_{15}$): Note that for an $n\times n$ matrix $A$, $\textup{tr}(A)=\textup{tr}(I_{n}A)\leq\|I_{n}\|_{\textup{F}}\|A\|_{\textup{F}}=\sqrt{n}\|A\|_{\textup{F}}.$ This yields $\mathcal{I}_{15}=\frac{{\kappa}}{N}\sum_{k=1}^{N}\textup{Re}[(c_{ki}-c_{ik}+c_{jk}-c_{kj})\textup{tr}(S_{ij}^{\dagger})]\leq 4{\kappa}\sqrt{n}\mathcal{S}(\mathcal{V})\mathcal{D}(\mathcal{U}).$ (A.7) In (A.5), we collect all the estimates for $\mathcal{I}_{1k},~{}k=1,\cdots,5$ to obtain $\frac{1}{2}\frac{{\textup{d}}}{{\textup{d}t}}\|S_{ij}\|_{\textup{F}}^{2}\leq-2m{\kappa}\|S_{ij}\|_{\textup{F}}^{2}+m{\kappa}\mathcal{D}(\mathcal{U})^{4}+6{\kappa}\mathcal{S}(\mathcal{V})\mathcal{D}(\mathcal{U})^{2}+2{\kappa}\mathcal{S}(\mathcal{V})\mathcal{D}(\mathcal{U})^{3}+4{\kappa}\sqrt{n}\mathcal{S}(\mathcal{V})\mathcal{D}(\mathcal{U}).$ Hence, $\mathcal{D}(\mathcal{U})$ satisfies $\frac{{\textup{d}}}{{\textup{d}t}}\mathcal{D}(\mathcal{U})\leq-2m{\kappa}\mathcal{D}(\mathcal{U})+m{\kappa}\mathcal{D}(\mathcal{U})^{3}+6{\kappa}\mathcal{S}(\mathcal{V})\mathcal{D}(\mathcal{U})+2{\kappa}\mathcal{S}(\mathcal{V})\mathcal{D}(\mathcal{U})^{2}+4{\kappa}\sqrt{n}\mathcal{S}(\mathcal{V}).$ Similarly, one can find a differential inequality for $\mathcal{D}(\mathcal{V})$ by exchanging the roles of $\mathcal{U}$ and $\mathcal{V}$: $\frac{{\textup{d}}}{{\textup{d}t}}\mathcal{D}(\mathcal{V})\leq-2n{\kappa}\mathcal{D}(\mathcal{V})+n{\kappa}\mathcal{D}(\mathcal{V})^{3}+6{\kappa}\mathcal{S}(\mathcal{U})\mathcal{D}(\mathcal{V})+2{\kappa}\mathcal{S}(\mathcal{U})\mathcal{D}(\mathcal{V})^{2}+4{\kappa}\sqrt{m}\mathcal{S}(\mathcal{U}).$ ∎ In (A.1), note that $\mathcal{S}(\mathcal{U})$ and $\mathcal{S}(\mathcal{V})$ appear in the differential inequalities for $\mathcal{D}(\mathcal{U})$ and $\mathcal{D}(\mathcal{V})$. Hence, we derive the differential inequalities for $\mathcal{S}(\mathcal{U})$ and $\mathcal{S}(\mathcal{V})$ below. ###### Lemma A.2. Let $\\{(U_{i},V_{i})\\}$ be a solution to (4.4). Then, $\mathcal{S}(\mathcal{U})$ and $\mathcal{S}(\mathcal{V})$ satisfy $\displaystyle\begin{aligned} &\frac{{\textup{d}}}{{\textup{d}t}}\mathcal{S}(\mathcal{U})\leq-2m{\kappa}\mathcal{S}(\mathcal{U})+2m{\kappa}\mathcal{D}(\mathcal{U})^{2}+6{\kappa}\mathcal{S}(\mathcal{U})\mathcal{S}(\mathcal{V})+2{\kappa}\mathcal{S}(\mathcal{V})\mathcal{D}(\mathcal{U})^{2}+4{\kappa}\sqrt{n}\mathcal{S}(\mathcal{V}),\\\ &\frac{{\textup{d}}}{{\textup{d}t}}\mathcal{S}(\mathcal{V})\leq-2n{\kappa}\mathcal{S}(\mathcal{V})+2n{\kappa}\mathcal{D}(\mathcal{V})^{2}+6{\kappa}\mathcal{S}(\mathcal{U})\mathcal{S}(\mathcal{V})+2{\kappa}\mathcal{S}(\mathcal{U})\mathcal{D}(\mathcal{V})^{2}+4{\kappa}\sqrt{m}\mathcal{S}(\mathcal{U}).\end{aligned}$ (A.8) ###### Proof. In (A.4), we take trace to obtain $\displaystyle\begin{aligned} &\frac{{\textup{d}}}{{\textup{d}t}}(n-d_{ij})=-2m{\kappa}(n-d_{ij})+\frac{m{\kappa}}{N}\sum_{k=1}^{N}\textup{tr}(S_{ik}S_{ij}+S_{ij}S_{kj})\\\ &\hskip 28.45274pt+\frac{{\kappa}}{N}\sum_{k=1}^{N}(c_{ik}-m)(n-d_{kj})-(c_{ki}-m)(n-d_{ik})-(c_{ik}-m)(n-d_{ij})\\\ &\hskip 28.45274pt+\frac{{\kappa}}{N}\sum_{k=1}^{N}(c_{kj}-m)(n-d_{ik})-(c_{jk}-m)(n-d_{kj})-(c_{jk}-m)(n-d_{ij})\\\ &\hskip 28.45274pt+\frac{{\kappa}}{N}\sum_{k=1}^{N}(c_{ki}-m)\textup{tr}(S_{ik}S_{ij})+(c_{jk}-m)\textup{tr}(S_{ij}S_{kj})\\\ &\hskip 28.45274pt+\frac{{\kappa}}{N}\sum_{k=1}^{N}[(c_{ki}-c_{ik})+(c_{jk}-c_{kj})]\sqrt{n}\\\ &\hskip 28.45274pt=:-2m{\kappa}(n-d_{ij})+\mathcal{I}_{21}+\mathcal{I}_{22}+\mathcal{I}_{23}+\mathcal{I}_{24}+\mathcal{I}_{25}.\end{aligned}$ (A.9) Below, we present estimates of $\mathcal{I}_{2k},~{}k=1,\cdots,5$, separately. $\bullet$ (Estimate of $\mathcal{I}_{21}$): We use the definition of $\mathcal{D}(\mathcal{U})$ to find $\displaystyle\mathcal{I}_{21}=\frac{m{\kappa}}{N}\sum_{k=1}^{N}\textup{tr}(S_{ik}S_{ij}+S_{ij}S_{kj})\leq 2m{\kappa}\mathcal{D}(\mathcal{U})^{2}.$ $\bullet$ (Estimates of $\mathcal{I}_{22}$ and $\mathcal{I}_{23}$): It is easy to see that $\displaystyle\mathcal{I}_{22}=\frac{{\kappa}}{N}\sum_{k=1}^{N}(c_{ik}-m)(n-d_{kj})-(c_{ki}-m)(n-d_{ik})-(c_{ik}-m)(n-d_{ij})\leq 3{\kappa}\mathcal{S}(\mathcal{U})\mathcal{S}(\mathcal{V}),$ $\displaystyle\mathcal{I}_{23}=\frac{{\kappa}}{N}\sum_{k=1}^{N}(c_{kj}-m)(n-d_{ik})-(c_{jk}-m)(n-d_{kj})-(c_{jk}-m)(n-d_{ij})\leq 3{\kappa}\mathcal{S}(\mathcal{U})\mathcal{S}(\mathcal{V}).$ $\bullet$ (Estimate of $\mathcal{I}_{24}$): Similar to $\mathcal{I}_{21}$, one has $\displaystyle\mathcal{I}_{24}=\frac{{\kappa}}{N}\sum_{k=1}^{N}(c_{ki}-m)\textup{tr}(S_{ik}S_{ij})+(c_{jk}-m)\textup{tr}(S_{ij}S_{kj})\leq 2{\kappa}\mathcal{S}(\mathcal{V})\mathcal{D}(\mathcal{U})^{2}.$ $\bullet$ (Estimate of $\mathcal{I}_{25}$): We find $\mathcal{I}_{25}=\frac{{\kappa}}{N}\sum_{k=1}^{N}[(c_{ki}-c_{ik})+(c_{jk}-c_{kj})]\sqrt{n}\leq 4{\kappa}\sqrt{n}\mathcal{S}(\mathcal{V}).$ (A.10) In (A.9), we combine all estimates to obtain $\frac{{\textup{d}}}{{\textup{d}t}}\mathcal{S}(\mathcal{U})\leq-2m{\kappa}\mathcal{S}(\mathcal{U})+2m{\kappa}\mathcal{D}(\mathcal{U})^{2}+6{\kappa}\mathcal{S}(\mathcal{U})\mathcal{S}(\mathcal{V})+2{\kappa}\mathcal{S}(\mathcal{V})\mathcal{D}(\mathcal{U})^{2}+4{\kappa}\sqrt{n}\mathcal{S}(\mathcal{V}).$ By exchanging the roles of $\mathcal{U}$ and $\mathcal{V}$, we derive a differential inequality for $\mathcal{S}(\mathcal{V})$: $\frac{{\textup{d}}}{{\textup{d}t}}\mathcal{S}(\mathcal{V})\leq-2n{\kappa}\mathcal{S}(\mathcal{V})+2n{\kappa}\mathcal{D}(\mathcal{V})^{2}+6{\kappa}\mathcal{S}(\mathcal{U})\mathcal{S}(\mathcal{V})+2{\kappa}\mathcal{S}(\mathcal{U})\mathcal{D}(\mathcal{V})^{2}+4{\kappa}\sqrt{m}\mathcal{S}(\mathcal{U}).$ ∎ ###### Remark A.1. For homogeneous Hamiltonians, we add $\eqref{AA-12}_{1}$ and $\eqref{AA-12}_{2}$ to find $\displaystyle\frac{{\textup{d}}}{{\textup{d}t}}(\mathcal{S}(\mathcal{U})+\mathcal{S}(\mathcal{V}))$ $\displaystyle\leq-2{\kappa}(m-2\sqrt{m})\mathcal{S}(\mathcal{U})-2{\kappa}(n-2\sqrt{n})\mathcal{S}(\mathcal{V})+\mathcal{O}((\mathcal{S}(\mathcal{U})+\mathcal{S}(\mathcal{V}))^{2}).$ Hence, in order to obtain the desired convergence, we assume $n>2\sqrt{n},\quad m>2\sqrt{m},\quad\textup{i.e.,}\quad n,m>4,$ which requires the restriction on the size of $U_{i}$ and $V_{j}$. This technical assumption arises from the estimate of $\mathcal{I}_{25}$ in (A.10). Now, we are ready to present a proof of Lemma 4.1 using Lemma A.1 and Lemma A.2. Proof of Lemma 4.1: Recall the inequalities in (A.1) and (A.8): $\displaystyle\begin{aligned} &\frac{{\textup{d}}}{{\textup{d}t}}\mathcal{D}(\mathcal{U})\leq-2m{\kappa}\mathcal{D}(\mathcal{U})+m{\kappa}\mathcal{D}(\mathcal{U})^{3}+6{\kappa}\mathcal{S}(\mathcal{V})\mathcal{D}(\mathcal{U})+2{\kappa}\mathcal{S}(\mathcal{V})\mathcal{D}(\mathcal{U})^{2}+4{\kappa}\sqrt{n}\mathcal{S}(\mathcal{V}),\\\ &\frac{{\textup{d}}}{{\textup{d}t}}\mathcal{D}(\mathcal{V})\leq-2n{\kappa}\mathcal{D}(\mathcal{V})+n{\kappa}\mathcal{D}(\mathcal{V})^{3}+6{\kappa}\mathcal{S}(\mathcal{U})\mathcal{D}(\mathcal{V})+2{\kappa}\mathcal{S}(\mathcal{U})\mathcal{D}(\mathcal{V})^{2}+4{\kappa}\sqrt{m}\mathcal{S}(\mathcal{U}),\\\ &\frac{{\textup{d}}}{{\textup{d}t}}\mathcal{S}(\mathcal{U})\leq-2m{\kappa}\mathcal{S}(\mathcal{U})+2m{\kappa}\mathcal{D}(\mathcal{U})^{2}+6{\kappa}\mathcal{S}(\mathcal{U})\mathcal{S}(\mathcal{V})+2{\kappa}\mathcal{S}(\mathcal{V})\mathcal{D}(\mathcal{U})^{2}+4{\kappa}\sqrt{n}\mathcal{S}(\mathcal{V}),\\\ &\frac{{\textup{d}}}{{\textup{d}t}}\mathcal{S}(\mathcal{V})\leq-2n{\kappa}\mathcal{S}(\mathcal{V})+2n{\kappa}\mathcal{D}(\mathcal{V})^{2}+6{\kappa}\mathcal{S}(\mathcal{U})\mathcal{S}(\mathcal{V})+2{\kappa}\mathcal{S}(\mathcal{U})\mathcal{D}(\mathcal{V})^{2}+4{\kappa}\sqrt{m}\mathcal{S}(\mathcal{U}).\end{aligned}$ (A.11) Without loss of generality, we may assume $n\geq m$, and add all the inequalities in (A.11) to find $\displaystyle\begin{aligned} \frac{{\textup{d}}}{{\textup{d}t}}\mathcal{L}&\leq-2{\kappa}(m-4\sqrt{n})\mathcal{L}+n{\kappa}(\mathcal{D}(\mathcal{U})^{3}+2\mathcal{D}(\mathcal{U})^{2}+\mathcal{D}(\mathcal{V})^{3}+2\mathcal{D}(\mathcal{V})^{2})\\\ &\hskip 14.22636pt+6{\kappa}(\mathcal{D}(\mathcal{U})\mathcal{S}(\mathcal{V})+\mathcal{D}(\mathcal{V})\mathcal{S}(\mathcal{U})+2\mathcal{S}(\mathcal{U})\mathcal{S}(\mathcal{V}))+4{\kappa}(\mathcal{D}(\mathcal{U})^{2}\mathcal{S}(\mathcal{V})+\mathcal{D}(\mathcal{V})^{2}\mathcal{S}(\mathcal{U}))\\\ &=:-2{\kappa}(m-4\sqrt{n})\mathcal{L}+n{\kappa}\mathcal{I}_{31}+6{\kappa}\mathcal{I}_{32}+4{\kappa}\mathcal{I}_{33}.\end{aligned}$ (A.12) In the sequel, we provide estimates for $\mathcal{I}_{3k},~{}k=1,2,3$, respectively. $\bullet$ (Estimate of $\mathcal{I}_{31}$): We use a rough estimate to find $\displaystyle\mathcal{I}_{31}=\mathcal{D}(\mathcal{U})^{3}+2\mathcal{D}(\mathcal{U})^{2}+\mathcal{D}(\mathcal{V})^{3}+2\mathcal{D}(\mathcal{V})^{2}\leq\mathcal{L}^{3}+2\mathcal{L}^{2}.$ $\bullet$ (Estimate of $\mathcal{I}_{32}$): By straightforward calculation, one has $\displaystyle\mathcal{I}_{32}$ $\displaystyle=\mathcal{D}(\mathcal{U})\mathcal{S}(\mathcal{V})+\mathcal{D}(\mathcal{V})\mathcal{S}(\mathcal{U})+2\mathcal{S}(\mathcal{U})\mathcal{S}(\mathcal{V})$ $\displaystyle\leq\frac{1}{2}\mathcal{D}(\mathcal{U})^{2}+\frac{1}{2}\mathcal{D}(\mathcal{V})^{2}+\frac{3}{2}\mathcal{S}(\mathcal{U})^{2}+\frac{3}{2}\mathcal{S}(\mathcal{V})^{2}\leq\frac{3}{2}\mathcal{L}^{2}.$ $\bullet$ (Estimate of $\mathcal{I}_{33}$): We use Young’s inequality that for $a,b>0$, $a^{2}b\leq\frac{2}{3}a^{3}+\frac{1}{3}b^{3}$ to find $\displaystyle\mathcal{I}_{33}=\mathcal{D}(\mathcal{U})^{2}\mathcal{S}(\mathcal{V})+\mathcal{D}(\mathcal{V})^{2}\mathcal{S}(\mathcal{U})\leq\frac{2}{3}\mathcal{D}(\mathcal{U})^{3}+\frac{1}{3}\mathcal{S}(\mathcal{V})^{3}+\frac{2}{3}\mathcal{D}(\mathcal{V})^{3}+\frac{1}{3}\mathcal{S}(\mathcal{U})^{3}\leq\frac{2}{3}\mathcal{L}^{3}.$ In (A.12), we combine all the estimates to obtain the desired inequality for $\mathcal{L}$: $\displaystyle\frac{{\textup{d}}}{{\textup{d}t}}\mathcal{L}$ $\displaystyle\leq-2{\kappa}(m-4\sqrt{n})\mathcal{L}+2{\kappa}n(\mathcal{L}^{3}+2\mathcal{L}^{2})+9{\kappa}\mathcal{L}^{2}+\frac{8{\kappa}}{3}\mathcal{L}^{3}$ $\displaystyle=-2{\kappa}(m-4\sqrt{n})\mathcal{L}++{\kappa}(4n+9)\mathcal{L}^{2}+{\kappa}\left(2n+\frac{8}{3}\right)\mathcal{L}^{3}.$ ## Appendix B Proof of Lemma 4.2 In this appendix, we present a proof of Lemma 4.2 in which the differential inequality for the aggregation functional $\mathcal{F}$ in (4.11) will be derived. ###### Lemma B.1. Let $\\{(U_{i},V_{i})\\}$ be a solution to (4.1). Then, $d(U,\tilde{U})$ and $d(V,\tilde{V})$ satisfy $\displaystyle\begin{aligned} \frac{{\textup{d}}}{{\textup{d}t}}d(U,\tilde{U})&\leq-2m{\kappa}d(U,\tilde{U})+4m{\kappa}\mathcal{L}d(U,\tilde{U})+6\mathcal{L}(d(U,\tilde{U})+\mathcal{S}(V,\tilde{V}))\\\ &\hskip 14.22636pt+2\mathcal{L}^{2}(4d(U,\tilde{U})+\mathcal{S}(V,\tilde{V}))+4\sqrt{n}\mathcal{S}(V,\tilde{V}).\\\ \frac{{\textup{d}}}{{\textup{d}t}}d(V,\tilde{V})&\leq-2n{\kappa}d(V,\tilde{V})+4n{\kappa}\mathcal{L}d(V,\tilde{V})+6\mathcal{L}(d(V,\tilde{V})+\mathcal{S}(U,\tilde{U}))\\\ &\hskip 14.22636pt+2\mathcal{L}^{2}(4d(V,\tilde{V})+\mathcal{S}(U,\tilde{U}))+4\sqrt{m}\mathcal{S}(U,\tilde{U}).\end{aligned}$ (B.1) ###### Proof. First, we recall (A.4): $\displaystyle\frac{{\textup{d}}}{{\textup{d}t}}S_{ij}$ $\displaystyle=-2m{\kappa}S_{ij}+\frac{m{\kappa}}{N}\sum_{k=1}^{N}S_{ik}S_{ij}+S_{ij}S_{kj}+\mathrm{i}(H_{i}-H_{j})+\mathrm{i}(S_{ij}H_{j}-H_{i}S_{ij})$ $\displaystyle\hskip 14.22636pt+\frac{{\kappa}}{N}\sum_{k=1}^{N}(c_{ik}-m)S_{kj}-(c_{ki}-m)S_{ik}-(c_{ik}-m)S_{ij}+(c_{ki}-m)S_{ik}S_{ij}$ $\displaystyle\hskip 14.22636pt+\frac{{\kappa}}{N}\sum_{k=1}^{N}(c_{kj}-m)S_{ik}-(c_{jk}-m)S_{kj}-(c_{jk}-m)S_{ij}+(c_{jk}-m)S_{ij}S_{kj}$ $\displaystyle\hskip 14.22636pt+\frac{{\kappa}}{N}\sum_{k=1}^{N}(c_{ki}-c_{ik}+c_{jk}-c_{kj})I_{n}.$ Thus, $S_{ij}-\tilde{S}_{ij}$ satisfies $\displaystyle\begin{aligned} &\frac{{\textup{d}}}{{\textup{d}t}}(S_{ij}-\tilde{S}_{ij})=-2m{\kappa}(S_{ij}-\tilde{S}_{ij})+\frac{m{\kappa}}{N}\sum_{k=1}^{N}\underbrace{(S_{ik}S_{ij}+S_{ij}S_{kj}-\tilde{S}_{ik}\tilde{S}_{ij}-\tilde{S}_{ij}\tilde{S}_{kj})}_{=:\mathcal{I}_{41}}\\\ &\hskip 28.45274pt+\underbrace{\mathrm{i}((S_{ij}-\tilde{S}_{ij})H_{j}-H_{i}(S_{ij}-\tilde{S}_{ij}))}_{=:\mathcal{I}_{42}}+\frac{{\kappa}}{N}\sum_{k=1}^{N}(\mathcal{I}_{43}-\tilde{\mathcal{I}}_{43})+\frac{{\kappa}}{N}\sum_{k=1}^{N}(\mathcal{I}_{44}-\tilde{\mathcal{I}}_{44})\\\ &\hskip 28.45274pt+\frac{{\kappa}}{N}\sum_{k=1}^{N}\underbrace{((c_{ki}-c_{ik}+c_{jk}-c_{kj})-(\tilde{c}_{ki}-\tilde{c}_{ik}+\tilde{c}_{jk}-\tilde{c}_{kj}))I_{n}}_{=:\mathcal{I}_{45}},\end{aligned}$ (B.2) where $\mathcal{I}_{43}$ and $\mathcal{I}_{44}$ are defined as $\displaystyle\mathcal{I}_{43}:=(c_{ik}-m)S_{kj}-(c_{ki}-m)S_{ik}-(c_{ik}-m)S_{ij}+(c_{ki}-m)S_{ik}S_{ij},$ $\displaystyle\mathcal{I}_{44}:=(c_{kj}-m)S_{ik}-(c_{jk}-m)S_{kj}-(c_{jk}-m)S_{ij}+(c_{jk}-m)S_{ij}S_{kj}.$ Below, we provide the estimates for $\mathcal{I}_{4k},~{}k=1,\cdots,5$, respectively. $\bullet$ (Estimate of $\mathcal{I}_{41}$): Note that $\displaystyle\mathcal{I}_{41}$ $\displaystyle=S_{ik}S_{ij}+S_{ij}S_{kj}-\tilde{S}_{ik}\tilde{S}_{ij}-\tilde{S}_{ij}\tilde{S}_{kj}$ $\displaystyle=S_{ik}(S_{ij}-\tilde{S}_{ij})+(S_{ik}-\tilde{S}_{ik})\tilde{S}_{ij}+S_{ij}(S_{kj}-\tilde{S}_{kj})+(S_{ij}-\tilde{S}_{ij})\tilde{S}_{kj}.$ This yields $\|\mathcal{I}_{41}\|_{\textup{F}}\leq 4\mathcal{L}d(U,\tilde{U}).$ (B.3) $\bullet$ (Estimate of $\mathcal{I}_{42}$): Note that for a skew-Hermitian matrix $\Omega$ and a matrix $A$, one has $\displaystyle\textup{tr}(\Omega AA^{\dagger})=0.$ This implies $\displaystyle\textup{Re}\textup{tr}(\mathcal{I}_{42}(S_{ij}-\tilde{S}_{ij})^{\dagger})=\textup{Re}\textup{tr}(\mathrm{i}H_{j}(S_{ij}-\tilde{S}_{ij})(S_{ij}-\tilde{S}_{ij})^{\dagger})=0.$ $\bullet$ (Estimate of $\mathcal{I}_{43}$): We observe $\displaystyle\begin{aligned} &\|(c_{ik}-m)S_{kj}-(\tilde{c}_{ik}-m)\tilde{S}_{kj}\|_{\textup{F}}\\\ &\hskip 28.45274pt=\|(c_{ik}-m)(S_{kj}-\tilde{S}_{kj})+(c_{ik}-\tilde{c}_{ik})\tilde{S}_{kj}\|_{\textup{F}}\\\ &\hskip 28.45274pt\leq\mathcal{L}d(U,\tilde{U})+\mathcal{L}\mathcal{S}(V,\tilde{V})=\mathcal{L}(d(U,\tilde{U})+\mathcal{S}(V,\tilde{V})).\end{aligned}$ Moreover, we use (B.3) $\displaystyle\|(c_{ki}-m)S_{ik}S_{ij}-(\tilde{c}_{ki}-m)\tilde{S}_{ik}\tilde{S}_{ij}\|_{\textup{F}}$ $\displaystyle\hskip 28.45274pt=\|(c_{ki}-m)(S_{ik}S_{ij}-\tilde{S}_{ik}\tilde{S}_{ij})+(c_{ik}-\tilde{c}_{ik})\tilde{S}_{ik}\tilde{S}_{ij}\|_{\textup{F}}$ $\displaystyle\hskip 28.45274pt\leq\mathcal{L}\cdot 4\mathcal{L}d(U,\tilde{U})+\mathcal{S}(V,\tilde{V})\mathcal{L}^{2}=\mathcal{L}^{2}(4d(U,\tilde{U})+\mathcal{S}(V,\tilde{V}))$ to obtain $\displaystyle\|\mathcal{I}_{43}-\tilde{\mathcal{I}}_{43}\|_{\textup{F}}\leq 3\mathcal{L}(d(U,\tilde{U})+\mathcal{S}(V,\tilde{V}))+\mathcal{L}^{2}(4d(U,\tilde{U})+\mathcal{S}(V,\tilde{V})).$ $\bullet$ (Estimate of $\mathcal{I}_{44}$): Similar to $\mathcal{I}_{43}$, one finds $\displaystyle\|\mathcal{I}_{44}-\tilde{\mathcal{I}}_{44}\|_{\textup{F}}\leq 3\mathcal{L}(d(U,\tilde{U})+\mathcal{S}(V,\tilde{V}))+\mathcal{L}^{2}(4d(U,\tilde{U})+\mathcal{S}(V,\tilde{V})).$ $\bullet$ (Estimate of $\mathcal{I}_{45}$): We directly find $\displaystyle\|\mathcal{I}_{45}\|_{\textup{F}}\leq 4\sqrt{n}\mathcal{S}(V,\tilde{V}).$ In (B.2), we combine all the estimates to obtain $\displaystyle\frac{{\textup{d}}}{{\textup{d}t}}d(U,\tilde{U})$ $\displaystyle\leq-2m{\kappa}d(U,\tilde{U})+4m{\kappa}\mathcal{L}d(U,\tilde{U})+6\mathcal{L}(d(U,\tilde{U})+\mathcal{S}(V,\tilde{V}))$ $\displaystyle\hskip 14.22636pt+2\mathcal{L}^{2}(4d(U,\tilde{U})+\mathcal{S}(V,\tilde{V}))+4\sqrt{n}\mathcal{S}(V,\tilde{V}).$ By exchanging the roles of $\mathcal{U}$ and $\mathcal{V}$, we find the differential inequality for $d(V,\tilde{V})$: $\displaystyle\frac{{\textup{d}}}{{\textup{d}t}}d(V,\tilde{V})$ $\displaystyle\leq-2n{\kappa}d(V,\tilde{V})+4n{\kappa}\mathcal{L}d(V,\tilde{V})+6\mathcal{L}(d(V,\tilde{V})+\mathcal{S}(U,\tilde{U}))$ $\displaystyle\hskip 14.22636pt+2\mathcal{L}^{2}(4d(V,\tilde{V})+\mathcal{S}(U,\tilde{U}))+4\sqrt{m}\mathcal{S}(U,\tilde{U}).$ ∎ Below, we derive differential inequalities for $\mathcal{S}(U,\tilde{U})$ and $\mathcal{S}(V,\tilde{V})$. ###### Lemma B.2. Let $\\{(U_{i},V_{i})\\}$ be a solution to (4.1). Then, $\mathcal{S}(U,\tilde{U})$ and $\mathcal{S}(V,\tilde{V})$ satisfy $\displaystyle\begin{aligned} \frac{{\textup{d}}}{{\textup{d}t}}\mathcal{S}(U,\tilde{U})&\leq-2m{\kappa}\mathcal{S}(U,\tilde{U})+4\mathcal{L}d(U,\tilde{U})+\mathcal{D}(\mathcal{H})d(U,\tilde{U})+6\mathcal{L}(\mathcal{S}(U,\tilde{U})+\mathcal{S}(V,\tilde{V}))\\\ &\hskip 14.22636pt+8\mathcal{L}^{2}d(U,\tilde{U})+2\mathcal{L}^{2}\mathcal{S}(V,\tilde{V})+4\sqrt{n}\mathcal{S}(V,\tilde{V}).\\\ \frac{{\textup{d}}}{{\textup{d}t}}\mathcal{S}(V,\tilde{V})&\leq-2n{\kappa}\mathcal{S}(V,\tilde{V})+4\mathcal{L}d(V,\tilde{V})+\mathcal{D}(\mathcal{H})d(V,\tilde{V})+6\mathcal{L}(\mathcal{S}(V,\tilde{V})+\mathcal{S}(U,\tilde{U}))\\\ &\hskip 14.22636pt+8\mathcal{L}^{2}d(V,\tilde{V})+2\mathcal{L}^{2}\mathcal{S}(U,\tilde{U})+4\sqrt{m}\mathcal{S}(U,\tilde{U}).\end{aligned}$ (B.4) ###### Proof. We recall (A.9): $\displaystyle\begin{aligned} \frac{{\textup{d}}}{{\textup{d}t}}(n-d_{ij})&=-2m{\kappa}(n-d_{ij})+\frac{m{\kappa}}{N}\sum_{k=1}^{N}\textup{tr}(S_{ik}S_{ij}+S_{ij}S_{kj})\\\ &\hskip 14.22636pt+\mathrm{i}\textup{tr}(H_{i}-H_{j})+\mathrm{i}\textup{tr}(S_{ij}H_{j}-H_{i}S_{ij})\\\ &\hskip 14.22636pt+\frac{{\kappa}}{N}\sum_{k=1}^{N}(c_{ik}-m)(n-d_{kj})-(c_{ki}-m)(n-d_{ik})-(c_{ik}-m)(n-d_{ij})\\\ &\hskip 14.22636pt+\frac{{\kappa}}{N}\sum_{k=1}^{N}(c_{kj}-m)(n-d_{ik})-(c_{jk}-m)(n-d_{kj})-(c_{jk}-m)(n-d_{ij})\\\ &\hskip 14.22636pt+\frac{{\kappa}}{N}\sum_{k=1}^{N}(c_{ki}-m)\textup{tr}(S_{ik}S_{ij})+(c_{jk}-m)\textup{tr}(S_{ij}S_{kj})\\\ &\hskip 14.22636pt+\frac{{\kappa}}{N}\sum_{k=1}^{N}[(c_{ki}-c_{ik})+(c_{jk}-c_{kj})]\sqrt{n}.\end{aligned}$ We denote $p_{ij}:=n-d_{ij},\quad q_{ij}:=m-c_{ij}.$ Then, $p_{ij}-\tilde{p}_{ij}$ satisfies $\displaystyle\begin{aligned} &\frac{{\textup{d}}}{{\textup{d}t}}(p_{ij}-\tilde{p}_{ij})=-2m{\kappa}(p_{ij}-\tilde{p}_{ij})+\frac{m{\kappa}}{N}\sum_{k=1}^{N}\underbrace{\textup{tr}(S_{ik}S_{ij}+S_{ij}S_{kj}-\tilde{S}_{ik}\tilde{S}_{ij}-\tilde{S}_{ij}\tilde{S}_{kj})}_{=:\mathcal{I}_{51}}\\\ &\hskip 5.69046pt+\underbrace{\mathrm{i}\textup{tr}((S_{ij}-\tilde{S}_{ij})(H_{j}-H_{i}))}_{=:\mathcal{I}_{52}}+\frac{{\kappa}}{N}\sum_{k=1}^{N}(\mathcal{I}_{53}-\tilde{\mathcal{I}}_{53})+\frac{{\kappa}}{N}\sum_{k=1}^{N}(\mathcal{I}_{54}-\tilde{\mathcal{I}}_{54})\\\ &\hskip 5.69046pt+\frac{{\kappa}}{N}\sum_{k=1}^{N}(\mathcal{I}_{55}-\tilde{\mathcal{I}}_{55})+\frac{{\kappa}}{N}\sum_{k=1}^{N}\underbrace{(p_{ik}-\tilde{p}_{ik}+p_{kj}-\tilde{p}_{kj})+(\tilde{p}_{ki}-\tilde{p}_{ki})+(\tilde{p}_{jk}-p_{jk})}_{=:\mathcal{I}_{56}}\sqrt{n}.\end{aligned}$ (B.5) Below, we present the estimates for $\mathcal{I}_{5k},~{}k=1,\cdots,6$, separately. $\bullet$ (Estimate of $\mathcal{I}_{51}$): We recall (B.3) to find $|\mathcal{I}_{51}|\leq 4\mathcal{L}d(U,\tilde{U}).$ $\bullet$ (Estimate of $\mathcal{I}_{52}$): we set $\mathcal{D}(\mathcal{H}):=\max_{1\leq i,j\leq N}\|H_{i}-H_{j}\|_{\infty},$ Then, one has $|\mathcal{I}_{52}|\leq\mathcal{D}(\mathcal{H})d(U,\tilde{U}).$ $\bullet$ (Estimate of $\mathcal{I}_{53}$): Note that $\displaystyle|(c_{ik}-m)(n-d_{kj})-(\tilde{c}_{ik}-m)(n-\tilde{d}_{kj})|$ $\displaystyle\hskip 28.45274pt\leq|p_{ik}-\tilde{p}_{ik}||q_{ik}|+|\tilde{p}_{ik}||q_{kj}-\tilde{q}_{kj}|\leq\mathcal{L}(\mathcal{S}(U,\tilde{U})+\mathcal{S}(V,\tilde{V})).$ Thus, we find $\displaystyle\|\mathcal{I}_{53}-\tilde{\mathcal{I}}_{53}\|_{\textup{F}}\leq 3\mathcal{L}(\mathcal{S}(U,\tilde{U})+\mathcal{S}(V,\tilde{V})).$ $\bullet$ (Estimate of $\mathcal{I}_{54}$): We observe $\displaystyle\|\mathcal{I}_{54}-\tilde{\mathcal{I}}_{54}\|_{\textup{F}}\leq 3\mathcal{L}(\mathcal{S}(U,\tilde{U})+\mathcal{S}(V,\tilde{V})).$ $\bullet$ (Estimate of $\mathcal{I}_{55}$): Note that $\displaystyle q_{ki}\textup{tr}(S_{ik}S_{ij}-\tilde{q}_{ki}\textup{tr}(\tilde{S}_{ik}\tilde{S}_{ij})$ $\displaystyle\hskip 14.22636pt=q_{ij}\textup{tr}(S_{ik}S_{ij}-\tilde{S}_{ik}\tilde{S}_{ij})+(q_{ij}-\tilde{q}_{ij})\textup{tr}(\tilde{S}_{ik}\tilde{S}_{ij})\leq 4\mathcal{L}^{2}d(U,\tilde{U})+\mathcal{L}^{2}\mathcal{S}(V,\tilde{V}).$ Thus, we have $\displaystyle\|\mathcal{I}_{55}-\tilde{\mathcal{I}}_{55}\|_{\textup{F}}\leq 8\mathcal{L}^{2}d(U,\tilde{U})+2\mathcal{L}^{2}\mathcal{S}(V,\tilde{V}).$ $\bullet$ (Estimate of $\mathcal{I}_{56}$): We directly find $\displaystyle|\mathcal{I}_{56}|\leq 4\sqrt{n}\mathcal{S}(V,\tilde{V}).$ In (B.5), we collect all the estimates to find $\displaystyle\frac{{\textup{d}}}{{\textup{d}t}}\mathcal{S}(U,\tilde{U})$ $\displaystyle\leq-2m{\kappa}\mathcal{S}(U,\tilde{U})+4\mathcal{L}d(U,\tilde{U})+\mathcal{D}(\mathcal{H})d(U,\tilde{U})+6\mathcal{L}(\mathcal{S}(U,\tilde{U})+\mathcal{S}(V,\tilde{V}))$ $\displaystyle\hskip 14.22636pt+8\mathcal{L}^{2}d(U,\tilde{U})+2\mathcal{L}^{2}\mathcal{S}(V,\tilde{V})+4\sqrt{n}\mathcal{S}(V,\tilde{V}).$ By exchanging the roles of $\mathcal{U}$ and $\mathcal{V}$, we can derive a differential inequality for $\mathcal{S}(V,\tilde{V})$: $\displaystyle\frac{{\textup{d}}}{{\textup{d}t}}\mathcal{S}(V,\tilde{V})$ $\displaystyle\leq-2n{\kappa}\mathcal{S}(V,\tilde{V})+4\mathcal{L}d(V,\tilde{V})+\mathcal{D}(\mathcal{H})d(V,\tilde{V})+6\mathcal{L}(\mathcal{S}(V,\tilde{V})+\mathcal{S}(U,\tilde{U}))$ $\displaystyle\hskip 14.22636pt+8\mathcal{L}^{2}d(V,\tilde{V})+2\mathcal{L}^{2}\mathcal{S}(U,\tilde{U})+4\sqrt{m}\mathcal{S}(U,\tilde{U}).$ ∎ Now, we are ready to provide a proof of Lemma 4.2 with the help of Lemma B.1 and Lemma B.2. Proof of Lemma 4.2: recall the inequalities in (B.1) and (B.4): $\displaystyle\begin{aligned} \frac{{\textup{d}}}{{\textup{d}t}}d(U,\tilde{U})&\leq-2m{\kappa}d(U,\tilde{U})+4m{\kappa}\mathcal{L}d(U,\tilde{U})+6\mathcal{L}(d(U,\tilde{U})+\mathcal{S}(V,\tilde{V}))\\\ &\hskip 14.22636pt+2\mathcal{L}^{2}(4d(U,\tilde{U})+\mathcal{S}(V,\tilde{V}))+4\sqrt{n}\mathcal{S}(V,\tilde{V}),\\\ \frac{{\textup{d}}}{{\textup{d}t}}d(V,\tilde{V})&\leq-2n{\kappa}d(V,\tilde{V})+4n{\kappa}\mathcal{L}d(V,\tilde{V})+6\mathcal{L}(d(V,\tilde{V})+\mathcal{S}(U,\tilde{U}))\\\ &\hskip 14.22636pt+2\mathcal{L}^{2}(4d(V,\tilde{V})+\mathcal{S}(U,\tilde{U}))+4\sqrt{m}\mathcal{S}(U,\tilde{U}),\\\ \frac{{\textup{d}}}{{\textup{d}t}}\mathcal{S}(U,\tilde{U})&\leq-2m{\kappa}\mathcal{S}(U,\tilde{U})+4\mathcal{L}d(U,\tilde{U})+\mathcal{D}(\mathcal{H})d(U,\tilde{U})+6\mathcal{L}(\mathcal{S}(U,\tilde{U})+\mathcal{S}(V,\tilde{V}))\\\ &\hskip 14.22636pt+8\mathcal{L}^{2}d(U,\tilde{U})+2\mathcal{L}^{2}\mathcal{S}(V,\tilde{V})+4\sqrt{n}\mathcal{S}(V,\tilde{V}),\\\ \frac{{\textup{d}}}{{\textup{d}t}}\mathcal{S}(V,\tilde{V})&\leq-2n{\kappa}\mathcal{S}(V,\tilde{V})+4\mathcal{L}d(V,\tilde{V})+\mathcal{D}(\mathcal{G})d(V,\tilde{V})+6\mathcal{L}(\mathcal{S}(V,\tilde{V})+\mathcal{S}(U,\tilde{U}))\\\ &\hskip 14.22636pt+8\mathcal{L}^{2}d(V,\tilde{V})+2\mathcal{L}^{2}\mathcal{S}(U,\tilde{U})+4\sqrt{m}\mathcal{S}(U,\tilde{U}).\end{aligned}$ (B.6) We add all the inequalities in (B.6) to obtain the desired inequality for $\mathcal{F}$: $\displaystyle\frac{{\textup{d}}\mathcal{F}}{dt}$ $\displaystyle\leq-{\kappa}\left(2m-8\sqrt{n}-\frac{\mathcal{D}(\mathcal{H})}{{\kappa}}\right)\mathcal{F}+4(n+1){\kappa}\mathcal{L}(d(U,\tilde{U})+d(V,\tilde{V}))+6{\kappa}\mathcal{L}\mathcal{F}$ $\displaystyle\hskip 14.22636pt+12{\kappa}\mathcal{L}(\mathcal{S}(U,\tilde{U})+\mathcal{S}(V,\tilde{V}))+16{\kappa}\mathcal{L}^{2}(d(U,\tilde{U})+d(V,\tilde{V}))+4{\kappa}\mathcal{L}^{2}(\mathcal{S}(U,\tilde{U})+\mathcal{S}(V,\tilde{V}))$ $\displaystyle\leq-{\kappa}\left(2m-8\sqrt{n}-\frac{\max\\{\mathcal{D}(\mathcal{H}),\mathcal{D}(\mathcal{G})\\}}{{\kappa}}\right)\mathcal{F}+{\kappa}(4n+22)\mathcal{L}\mathcal{F}+20{\kappa}\mathcal{L}^{2}\mathcal{F}.$ ## Appendix C Emergent dynamics of the DSOM model In this section, we replace the product of two unitary groups $\mathbf{U}(n)\times\mathbf{U}(m)$ by the product of two special orthogonal groups $\mathbf{SO}(n)\times\mathbf{SO}(m)$. Since all elements of special orthogonal matrices are real-valued, one has for any (real-valued) square matrices $A$ and $B$, $\langle A,B\rangle_{\textup{F}}=\textup{tr}(AB^{\top})=\textup{tr}(BA^{\top})=\langle B,A\rangle_{\textup{F}}.$ (C.1) Thus, model (3.5) reduces to $\begin{cases}\displaystyle\dot{U}_{j}=B_{j}U_{j}+\frac{{\kappa}}{N}\sum_{k=1}^{N}\langle V_{j},V_{k}\rangle_{\textup{F}}(U_{k}-U_{j}U_{k}^{\top}U_{j}),\quad t>0,\\\ \displaystyle\dot{V}_{j}=C_{j}V_{j}+\frac{{\kappa}}{N}\sum_{k=1}^{N}\langle U_{j},U_{k}\rangle_{\textup{F}}(V_{k}-V_{j}V_{k}^{\top}V_{j}),\\\ \displaystyle(U_{j},V_{j})(0)=(U_{j}^{0},V_{j}^{0})\in\mathbf{SO}(n)\times\mathbf{SO}(m),\quad 1\leq j\leq N,\end{cases}$ (C.2) where $B_{j}\in\mathbb{R}^{n\times n\times n\times n}$ and $C_{j}\in\mathbb{R}^{m\times m\times m\times m}$ are given rank-4 tensors satisfying skew-symmetric properties: $[B_{j}]_{\alpha_{1}\beta_{1}\alpha_{2}\beta_{2}}=-[B_{j}]_{\alpha_{2}\beta_{2}\alpha_{1}\beta_{1}},\quad[C_{j}]_{\gamma_{1}\delta_{1}\gamma_{2}\delta_{2}}=-[{C}_{j}]_{\gamma_{2}\delta_{2}\gamma_{1}\delta_{1}}.$ ###### Remark C.1. Due to the symmetric property of the Frobenius inner product on real-valued matrices, one has $d_{ij}=\langle U_{i},U_{j}\rangle_{\textup{F}}=\langle U_{j},U_{i}\rangle_{\textup{F}}=d_{ji},\quad c_{ij}=\langle V_{i},V_{j}\rangle_{\textup{F}}=\langle V_{j},V_{i}\rangle_{\textup{F}}=c_{ji}.$ (C.3) Due to the relation (C.3), for example, the last term $c_{ki}-c_{ik}+c_{jk}-c_{kj}$ in (A.3) vanishes. Thus, it would be expected that the relation (C.3) relaxes the condition such as the dimension condition $n\geq m>4\sqrt{n}$ in (4.7)(i) (see also Remark A.1). The contents of this appendix are exactly the same as those of Section 4, and we only provide a sketch of the proof. In Section C.1, we are concerned with the identical (or homogeneous) Hamiltonian, and in Section C.2, the non- identical (or heterogeneous) Hamiltonian is considered to exhibit the phase- locked state. ### C.1. Homogeneous ensemble In this subsection, thanks to the solution splitting property, we assume that $B_{j}\equiv O,\quad C_{j}\equiv O,\quad j=1,\cdots,N.$ (C.4) Below, we state the main theorem for an identical ensemble under which the complete aggregation occurs. For this, we recall diameters and aggregation functional: $\displaystyle\begin{aligned} &\mathcal{D}(\mathcal{U}(t))=\max_{1\leq i,j\leq N}\|U_{i}(t)-U_{j}(t)\|_{\textup{F}},\quad\mathcal{S}(\mathcal{U}(t))=\max_{1\leq i,j\leq N}|n-d_{ij}(t)|,\\\ &\mathcal{D}(\mathcal{V}(t))=\max_{1\leq i,j\leq N}\|V_{i}(t)-V_{j}(t)\|_{\textup{F}},\quad\mathcal{S}(\mathcal{V}(t))=\max_{1\leq i,j\leq N}|m-c_{ij}(t)|,\\\ &\mathcal{L}=\mathcal{D}(\mathcal{U})+\mathcal{D}(\mathcal{V})+\mathcal{S}(\mathcal{U})+\mathcal{S}(\mathcal{V}).\end{aligned}$ ###### Theorem C.1. Suppose initial data satisfy $\mathcal{L}^{0}<\alpha_{m,n}=\frac{-(12n+27)+\sqrt{(12n+27)^{2}+24m(3n+4)}}{4(4n+3)},$ (C.5) and let $\\{(U_{i},V_{i})\\}$ be a solution to (C.2) with (C.4). Then, we have $\lim_{t\to\infty}\mathcal{L}(t)=0.$ ###### Proof. By recalling the inequalities in (A.1) with the relation (C.1), we see $\displaystyle\begin{aligned} &\frac{{\textup{d}}}{{\textup{d}t}}\mathcal{D}(\mathcal{U})\leq-2m{\kappa}\mathcal{D}(\mathcal{U})+m{\kappa}\mathcal{D}(\mathcal{U})^{3}+6{\kappa}\mathcal{S}(\mathcal{V})\mathcal{D}(\mathcal{U})+2{\kappa}\mathcal{S}(\mathcal{V})\mathcal{D}(\mathcal{U})^{2},\\\ &\frac{{\textup{d}}}{{\textup{d}t}}\mathcal{D}(\mathcal{V})\leq-2n{\kappa}\mathcal{D}(\mathcal{V})+n{\kappa}\mathcal{D}(\mathcal{V})^{3}+6{\kappa}\mathcal{S}(\mathcal{U})\mathcal{D}(\mathcal{V})+2{\kappa}\mathcal{S}(\mathcal{U})\mathcal{D}(\mathcal{V})^{2}.\end{aligned}$ (C.6) Similarly, we use (A.8) to find $\displaystyle\begin{aligned} &\frac{{\textup{d}}}{{\textup{d}t}}\mathcal{S}(\mathcal{U})\leq-2m{\kappa}\mathcal{S}(\mathcal{U})+2m{\kappa}\mathcal{D}(\mathcal{U})^{2}+6{\kappa}\mathcal{S}(\mathcal{U})\mathcal{S}(\mathcal{V})+2{\kappa}\mathcal{S}(\mathcal{V})\mathcal{D}(\mathcal{U})^{2},\\\ &\frac{{\textup{d}}}{{\textup{d}t}}\mathcal{S}(\mathcal{V})\leq-2n{\kappa}\mathcal{S}(\mathcal{V})+2n{\kappa}\mathcal{D}(\mathcal{V})^{2}+6{\kappa}\mathcal{S}(\mathcal{U})\mathcal{S}(\mathcal{V})+2{\kappa}\mathcal{S}(\mathcal{U})\mathcal{D}(\mathcal{V})^{2}.\end{aligned}$ (C.7) Now, we add (C.6) and (C.7) to obtain a differential inequality for $\mathcal{L}$: $\frac{{\textup{d}}}{{\textup{d}t}}\mathcal{L}\leq-2{\kappa}m\mathcal{L}+{\kappa}(4n+9)\mathcal{L}^{2}+{\kappa}\left(2n+\frac{8}{3}\right)\mathcal{L}^{3}.$ (C.8) If we introduce an auxiliary polynomial: $f(s):=\left(2n+\frac{8}{3}\right)s^{2}+(4n+9)s-2m,$ (C.9) then (C.8) can be written as $\frac{{\textup{d}}}{{\textup{d}t}}\mathcal{L}\leq{\kappa}\mathcal{L}f(\mathcal{L}).$ Furthermore, we notice that $\alpha_{m,n}$ in (C.5) becomes a unique positive root of the quadratic polynomial $f$ in (C.9). Hence, the desired zero convergence of $\mathcal{L}$ directly follows from the dynamical systems theory. ∎ ### C.2. Heterogeneous ensemble In this subsection, we are concerned with the heterogeneous Hamiltonians where the natural frequency tensors are different in general. Similar to Section 4.2, there exist two skew-symmetric matrices $\Omega_{j}\in\mathbb{R}^{n\times n}$ and $\Psi_{j}\in\mathbb{R}^{m\times m}$ such that $B_{j}U_{j}=\Omega_{j}U_{j},\quad C_{j}V_{j}=\Psi_{j}V_{j},\quad 1\leq j\leq N.$ Thus, our model reads as $\begin{cases}\displaystyle\dot{U}_{j}=\Omega_{j}U_{j}+\frac{{\kappa}}{N}\sum_{k=1}^{N}\langle V_{j},V_{k}\rangle_{\textup{F}}(U_{k}-U_{j}U_{k}^{\top}U_{j}),\\\ \displaystyle\dot{V}_{j}=\Psi_{j}V_{j}+\frac{{\kappa}}{N}\sum_{k=1}^{N}\langle U_{j},U_{k}\rangle_{\textup{F}}(V_{k}-V_{j}V_{k}^{\top}V_{j}),\\\ \displaystyle(U_{j},V_{j})(0)=(U_{j}^{0},V_{j}^{0})\in\mathbf{SO}(n)\times\mathbf{SO}(m),\quad 1\leq j\leq N,\end{cases}$ (C.10) A crucial estimate used for the emergence of the phase-locked state is (4.12) stated in Lemma 4.2 where the aggregation functional $\mathcal{F}$ tends to zero. Here, $\mathcal{F}$ measures the inter-distance between any two solution configurations $\\{(U_{i},V_{i})\\}$ and $\\{(\tilde{U}_{i},\tilde{V}_{i})\\}$: $\mathcal{F}(t)=d(U,\tilde{U})(t)+d(V,\tilde{V})(t)+\mathcal{S}(U,\tilde{U})(t)+\mathcal{S}(V,\tilde{V})(t),$ where the diameters are defined in (4.8) and (4.10). We also denote the diameters for natural frequencies: $\mathcal{D}(\Omega):=\max_{1\leq i,j\leq N}\|\Omega_{i}-\Omega_{j}\|_{\infty},\quad\mathcal{D}(\Psi):=\max_{1\leq i,j\leq N}\|\Psi_{i}-\Psi_{j}\|_{\infty}.$ Without loss of generality, we may assume $\mathcal{D}(\Omega)\geq\mathcal{D}(\Psi).$ In the following lemma, we provide a temporal evolution of $\mathcal{F}$. ###### Lemma C.1. Let $\\{(U_{i},V_{i})\\}$ and $\\{(\tilde{U}_{i},\tilde{V}_{i})\\}$ be any two solutions to (C.10). Then, the aggregation functionals $\mathcal{L}$ and $\mathcal{F}$ satisfy $\displaystyle\begin{aligned} &\frac{{\textup{d}}\mathcal{L}}{dt}\leq 2(1+3\sqrt{n})\mathcal{D}(\Omega)-2{\kappa}m\mathcal{L}+{\kappa}(4n+9)\mathcal{L}^{2}+{\kappa}\left(2n+\frac{8}{3}\right)\mathcal{L}^{3},\quad t>0,\\\ &\frac{{\textup{d}}\mathcal{F}}{dt}\leq-{\kappa}\left(2m-\frac{\max\\{\mathcal{D}(\Omega),\mathcal{D}(\Psi)\\}}{{\kappa}}\right)\mathcal{F}+{\kappa}(4n+22)\mathcal{L}\mathcal{F}+20{\kappa}\mathcal{L}^{2}\mathcal{F}.\end{aligned}$ (C.11) ###### Proof. For the inequality for $\mathcal{L}$, it directly follows from (C.8) in Theorem C.1. Similarly, if we closely follow a proof of Lemma 4.2 presented in Appendix B, then we find the desired inequality for $\mathcal{F}$. ∎ By applying $\eqref{Y-15}_{2}$, we establish the desired practical aggregation estimate for (C.10). ###### Proposition C.1. Suppose that the system parameters and initial data satisfy $\mathcal{D}(\Omega)\geq\mathcal{D}(\Psi),\quad{\kappa}>{\kappa}_{\textup{c}},\quad\mathcal{L}^{0}<\nu_{2},$ (C.12) where ${\kappa}_{\textup{c}}$ and $\nu_{2}$ are specified in (C.14) and (C.15), respectively, and let $\\{(U_{i},V_{i})\\}$ be a solution to (C.10). Then, one has $\lim_{{\kappa}\to\infty}\limsup_{t\to\infty}\mathcal{L}(t)=0.$ ###### Proof. We introduce an auxiliary polynomial: $g(s):=2ms-(4n+9)s^{2}-\left(2n+\frac{8}{3}\right)s^{3}.$ (C.13) Then, we notice that $g$ has three roots, say, $\alpha_{0}<0<\alpha_{1}$. Since $\eqref{Y-15}_{2}$ is rewritten as $\dot{\mathcal{L}}\leq{\kappa}\left(\frac{2(1+3\sqrt{n})\mathcal{D}(\mathcal{H})}{{\kappa}}-g(\mathcal{L})\right)=:{\kappa}p(s),$ for a sufficiently large ${\kappa}>0$, $p$ admits one negative root, say, $\nu_{0}<0$ and two positive roots, say, $0<\nu_{1}<\nu_{2}$ with continuous dependence on ${\kappa}$: for $\nu_{1}=\nu_{1}({\kappa})$ and $\nu_{2}=\nu_{2}({\kappa})$, $\lim_{{\kappa}\to\infty}\nu_{1}({\kappa})=0,\quad\lim_{{\kappa}\to\infty}\nu_{2}({\kappa})=\alpha_{1}.$ Since we assume $\eqref{Y-18}_{3}$, it follows from dynamical systems theory that there exists a finite entrance time $T_{*}>0$ such that $\mathcal{L}(t)<\nu_{1},\quad t>T_{*}.$ Hence, this yields the desired result. ∎ ###### Remark C.2. For the explicit value of ${\kappa}_{\textup{c}}$, we see that the cubic polynomial $g$ in (C.13) admits the local maximum at $s=s_{*}$: $s_{*}=\frac{-(4n+9)+\sqrt{(4n+9)^{2}+4m(3n+4)}}{6n+8}.$ Hence, ${\kappa}_{\textup{c}}$ is chosen to be ${\kappa}_{\textup{c}}:=\frac{2(1+3\sqrt{n})\mathcal{D}(\Omega)}{g(s_{*})}\quad\textup{so that}\quad g(s_{*})>\frac{2(1+3\sqrt{n})}{{\kappa}_{\textup{c}}}.$ (C.14) For this ${\kappa}_{\textup{c}}$, $\nu_{2}=\nu_{2}({\kappa})$ is completely determined as the largest positive root of $p(s)=0\quad\textup{for ${\kappa}>{\kappa}_{\textup{c}}$}.$ (C.15) It follows from Proposition 5.1 that we can make $\mathcal{L}$ sufficiently small by increasing the value of ${\kappa}$ for $t>T_{*}$. Thus, there exists ${\kappa}_{\textup{p}}>{\kappa}_{\textup{c}}$ such that for ${\kappa}>{\kappa}_{\textup{p}}$, $(4n+22)\mathcal{L}+20\mathcal{L}^{2}<\frac{1}{2}\left(2m-\frac{\max\\{\mathcal{D}(\Omega),\mathcal{D}(\Psi)\\}}{{\kappa}}\right)=:\frac{1}{2}\Lambda.$ Hence, $\eqref{Y-15}_{2}$ yields $\frac{d\mathcal{F}}{dt}\leq-\frac{1}{2}\Lambda\mathcal{F},\quad t>T_{*}.$ Since we have established the zero convergence of $\mathcal{F}$, we can obtain the same result in Theorem 4.2. Thanks to exactly the same proof, we only state the results. ###### Theorem C.2. Suppose that the system parameters and initial data satisfy $n\geq m,\quad\mathcal{D}(\Omega)\geq\mathcal{D}(\Psi),\quad{\kappa}>{\kappa}_{\textup{p}},\quad\max\\{\mathcal{L}^{0},\tilde{\mathcal{L}}^{0}\\}\leq\nu_{2},$ and let $\\{(U_{i},V_{i})\\}$ and $\\{(\tilde{U}_{i},\tilde{V}_{i})\\}$ be any two solutions to (C.10). Then, the following assertions hold. 1. (1) The aggregation functional $\mathcal{F}$ converges to zero exponentially. 2. (2) The normalized velocities $\dot{U}_{i}U_{i}^{\top}$ and $\tilde{\dot{U}}_{i}\tilde{U}_{i}^{\top}$ synchronize: $\|\dot{U}_{i}U_{i}^{\top}-\tilde{\dot{U}}_{i}\tilde{U}_{i}^{\top}\|_{\textup{F}}\leq 2{\kappa}(m+n)\mathcal{F}.$ 3. (3) There exist special orthogonal matrices $X_{\infty}\in\mathbf{SO}(n)$ and $Y_{\infty}\in\mathbf{SO}(m)$ such that $\displaystyle\lim_{t\to\infty}U_{i}^{\top}(t)\tilde{U}_{i}(t)=X_{\infty},\quad\lim_{t\to\infty}\|\tilde{U}_{i}(t)-U_{i}(t)X_{\infty}\|_{\textup{F}}=0,$ $\displaystyle\lim_{t\to\infty}V_{i}^{\top}(t)\tilde{V}_{i}(t)=Y_{\infty},\quad\lim_{t\to\infty}\|\tilde{V}_{i}(t)-V_{i}(t)Y_{\infty}\|_{\textup{F}}=0.$ 4. (4) System (C.10) exhibits asymptotic phase-locking: for any indices $i$ and $j$, $\exists~{}~{}\lim_{t\to\infty}U_{i}(t)U_{j}^{\top}(t)\quad\textup{and}\quad\exists~{}~{}\lim_{t\to\infty}V_{i}(t)V_{j}^{\top}(t).$ ## References * [1] Albi, G., Bellomo, N., Fermo, L., Ha, S.-Y., Kim, J., Pareschi, L., Poyato, D. and Soler, J.: Vehicular traffic, crowds, and swarms: from kinetic theory and multiscale methods to applications and research perspectives. Math. Models Methods Appl. Sci. 29 (2019), 1901–2005. * [2] Aydogu, Aylin, McQuade, S. T. and Duteil, N. P.: Opinion dynamics on a general compact Riemannian manifold. Netw. Heterog. Media 12 (2017), 489-523. * [3] Buck, J. and Buck, E.: Biology of synchronous flashing of fireflies. Nature 211 (1966), 562–564. * [4] Calovi, D. S., Lopez, U., Ngo, S., SIre, C., Chate, H. and Theraulaz, G.: Swarming, schooling, milling: phase diagram of a data-driven fish school model. New J. Phys. 16 (2014), 015206. * [5] Choi, Y.-P., Ha, S.-Y. and Li, Z.: Emergent dynamics of the Cucker-Smale flocking model and its variants. Active particles. Vol. 1. Advances in theory, models, and applications, 299–331, Model. Simul. Sci. Eng. Technol., Birkhäuser/Springer, Cham, 2017. * [6] Cucker, F. and Smale, S.: Emergent behavior in flocks. IEEE Trans. Automat. Control 52 (2007), 852–862. * [7] Danino, T., Mondragon-Palomino, O., Tsimring, L. and Hasty, J.: A synchronized quorum of genetic clocks. Nature 463 (2010), 326–330. * [8] DeVille, L.: Synchronization and stability for quantum Kuramoto. J. Stat. Phys. 174 (2019), 160-187. * [9] Ha, S.-Y., Kim, D. and Park, H.: On the completely separable state for the Lohe tensor model. To appear in J. Stat. Phys. * [10] Ha, S.-Y., Ko, D., Park, J. and Zhang, X.: Collective synchronization of classical and quantum oscillators. EMS Surv. Math. Sci. 3 (2016), 209–267. * [11] Ha, S.-Y., Ko, D. and Ryoo, S.: Emergent dynamics of a generalized Lohe model on some class of Lie groups. J. Stat. Phys. 168 (2017), 171–207. * [12] Ha, S.-Y., Ko, D. and Ryoo, S.: On the relaxation dynamics of Lohe oscillators on the Riemannian manifold. J. Stat. Phys. 172 (2018), 1427–1478. * [13] Ha, S.-Y. and Park, H.: Emergent behaviors of Lohe tensor flocks. J. Stat. Phys. 178 (2020), 1268–1292. * [14] Ha, S.-Y. and Park, H.: Complete aggregation of the Lohe Tensor model with the same free flow. J. Math. Phys 61 (2020), 102702. * [15] Ha, S.-Y. and Park, H.: Emergent behaviors of the generalized Lohe matrix model. To appear in Discrete Contin. Dyn. Syst. B. * [16] Ha, S.-Y. and Ryoo, S. W.: On the emergence and orbital stability of phase-locked states for the Lohe model. J. Stat. Phys. 163 (2016), 411–439. * [17] Hegselmann, R. and Krause, U.: Opinion dynamics and bounded confidence models, analysis, and simulation. J. Artif. Soc. Soc. Simul. 5 (2002). * [18] Helbing, D., Buzna, L., Johansson, A. and Werner, T.: Self-organized pedestrian crowd dynamics: experiments, simulations, and design solutions. Transport Sci. 39 (2005), 1–24. * [19] Kuramoto, Y.: Self-entrainment of a population of coupled non-linear oscillators. In International symposium on mathematical problems in mathematical physics. Lecture notes in theoretical physics 30 (1975), 420–422. * [20] Lohe, M. A.: Non-Abelian Kuramoto model and synchronization. J. Phys. A 42 (2009), 395101. * [21] Lohe, M. A.: On the double sphere model of synchronization. Phys. D. 412 (2020), 132642. * [22] Motter, A.E., Myers, S.A., Anghel, M. and Nishikawa, T.: Spontaneous synchrony in power-grid networks. Nature Phys. 9 (2013), 191–197. * [23] Olfati-Saber, R.: Swarms on sphere: A programmable swarm with synchronous behaviors like oscillator networks. Proc. of the 45th IEEE Conference on Decision and Control (2006), 5060–5066. * [24] Pikovsky, A., Rosenblum, M. and Kurths, J.: Synchronization: A universal concept in nonlinear sciences. Cambridge University Press, Cambridge, 2001. * [25] Polezhaev, A. A., Pashkov, R. A., Lobanov, A. I. and Petrov, I. B.: Spatial patterns formed by chemotactic bacteria Escherichia coli. Int. J. Dev. Biol. 50 (2006), 309–314. * [26] Tunstrom, K., Katz, Y., Ioannou, C. C., Huepe, C., Lutz, M. J. and Couzin, I. D.: Collective states, multistability and transitional behavior in schooling fish. PLoS Comput. Biol. 9 (2013) e1002915. * [27] Vicsek, T., Czirók, A., Ben-Jacob, E., Cohen, I. and Shochet, O.: Novel type of phase transition in a system of self-driven particles. Phys. Rev. Lett. 75 (1995), 1226–1229. * [28] Winfree, A. T.: Biological rhythms and the behavior of populations of coupled oscillators. J. Theor. Biol. 16 (1967), 15–42. * [29] Young, G. F., Scardovi, L., Cavagna, A., Giardina, I. and Leonard, N. E.: Starling flock networks manage uncertainty in consensus at low cost. PLoS Comput. Biol. 9 (2013) e1002894.
Test of the Orbital-Based LI3 Index as a Predictor of the Height of the 3MLCT $\rightarrow$ 3MC Transition-State Barrier for Gas-Phase [Ru(N∧N)3]2+ Polypyridine Complexes Denis Magero School of Science, Technology and Engineering, Department of Chemistry and Biochemistry, Alupe University, P.O. Box 845-50400, Busia, Kenya e-mail<EMAIL_ADDRESS> Ala Aldin M. H. M. Darghouth College of Sciences, University of Mosul, Al Majmoaa Street, Mosul, 41002 Iraq. e-mail<EMAIL_ADDRESS> Mark E. Casida Laboratoire de Spectrométrie, Interactions et Chimie théorique (SITh), Département de Chimie Moléculaire (DCM, UMR CNRS/UGA 5250), Institut de Chimie Moléculaire de Grenoble (ICMG, FR-2607), Université Grenoble Alpes (UGA) 301 rue de la Chimie, CS40700, F-38058 Grenoble Cedex 9, France e-mail<EMAIL_ADDRESS> Abstract Ruthenium(II) polypyridine compounds often luminesce but the luminescence lifetime depends upon the precise nature of the ligands. This luminescence lifetime is thought to be controlled by the barrier to conversion from an initial phosphorescent triplet metal-ligand charge transfer (3MLCT) state to a nonluminscent triplet metal-centered (3MC) state which decays nonradiatively. Earlier work [J. Photochem. Photobiol. A 348, 305 (2017)] took room temperature and liquid nitrogen (77 K) lifetimes from a large previously- published database [Coord. Chem. Rev. 84, 85 (1988)] and extracted empirical average 3MLCT $\rightarrow$ 3MC transition state (TS) barrier heights ($E_{ave}$s) which were not believed to be quantitative but which were believed to capture the trends in the true barrier heights correctly. These were then used together with information from partial density of states calculations [J. Photochem. Photobiol. A 276, 8 (2014)] to derive several orbital-based luminescence indices of which the third (LI3) was based upon frontier-molecular-orbital-like ideas and correlated linearly with values of $E_{ave}$. As it is known that $E_{ave}$ is a large underestimate of the true 3MLCT $\rightarrow$ 3MC TS barrier height in the case of the trisbipyridine ruthenium(II) cation $\\{$ [Ru(bpy)3]2+ $\\}$, but accurate TS barrier heights are difficult to obtain experimentally, it was judged useful to verify the ideas used to derive the LI3 index by calculating the energetics of the gas- phase 3MLCT $\rightarrow$ 3MC reaction for a series of ruthenium(II) tris bipyridine complexes using the same density functional and basis sets used in calculating LI3. Specifically, four closely-related bipyradine complexes $\\{$ [Ru(N∧N)3]2+ with N∧N = bpy (6), 4,4’-dm-bpy (70), 4,4’-dph-bpy (73), and 4,4’-DTB-bpy (74) $\\}$ were used for these calculations. We examine the trans dissociation mechanism in great detail at the B3LYP/6-31G+LANLDZ(Ru) level of detail and uncover a two part mechanism. In the first part, the electron is transferred to a single ligand rather than symmetrically to all three ligands. It is the two Ru-N bonds to this ligand which are equally elongated in the transition state. The intrinsic reaction coordinate then continues down a ridge in hyperspace and bifurcates into one of two symmetry-equivalent 3MC structures with elongated trans bonds. Interestingly, no significant difference is found for the TS barriers for the four complexes treated here. Instead, LI3 is linearly correlated with the energy difference $\Delta E=E(\mbox{${}^{3}$MLCT})-E(\mbox{${}^{3}$MC})$. While this work shows that LI3 predicts the total energy difference and this does correlate well with the luminescence lifetime, we do not have a detailed understanding yet of how this happens and are also wary of oversimplicity as there are likely other excited- state reactions occurring on similar time scales which may also impact the luminescence lifetimes. Graphical Abstract ## 1 Introduction Luminescent ruthenium(II) complexes elicit immense interest [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12] owing to their wide range of applications, including photochemical molecular devices, biological sensors, organic light emitting diodes, and biomedical applications, ††margin: LFT among others. The commonly assumed ligand-field theory (LFT) mechanism explaining this luminescence begins with the $d^{6}$ ruthenium complex initially in ††margin: GS its closed-shell ground-state (1GS) configuration, $\mbox{metal: }\begin{array}[]{c}\underbrace{[\,\,\,,\,\,\,][\,\,\,,\,\,\,]}_{e_{g}^{*}}\\\ \underbrace{[\uparrow,\downarrow][\uparrow,\downarrow][\uparrow,\downarrow]}_{t_{2g}}\end{array}+\mbox{ligands: }\underbrace{[\,\,\,,\,\,\,][\,\,\,,\,\,\,]\cdots}_{\pi^{*}}\,.$ (1) (The 1X and 3X pre-exponents denote respectively that the state X ††margin: 1X , 3X is a singlet or triplet. Note that the $t_{2g}$ orbital is nonbonding but that the $e_{g}^{*}$ orbitals are antibonding. We have also indicated some unoccupied $\pi^{*}$ orbitals on the ligands.) Figure 1 emphasizes that these $\pi^{*}$ ligand orbitals are within the LFT gap of the metal. Figure 1: Pseudo-octahedral ligand field theory diagram for ruthenium(II) complexes. Upon absorption of a photon, an electron is excited from the metal to the ligand to create a high-lying singlet ††margin: MLCT metal-ligand-charge- transfer (1MLCT) configuration, $\mbox{metal: }\begin{array}[]{c}\underbrace{[\,\,\,,\,\,\,][\,\,\,,\,\,\,]}_{e_{g}^{*}}\\\ \underbrace{[\uparrow,\downarrow][\uparrow,\downarrow][\downarrow,\,\,\,]}_{t_{2g}}\end{array}+\mbox{ligands: }\underbrace{[\uparrow,\,\,\,][\,\,\,,\,\,\,]\cdots}_{\pi^{*}}\,,$ (2) which relaxes by non-radiative relaxation to the lowest 3MLCT state, $\mbox{metal: }\begin{array}[]{c}\underbrace{[\,\,\,,\,\,\,][\,\,\,,\,\,\,]}_{e_{g}^{*}}\\\ \underbrace{[\uparrow,\downarrow][\uparrow,\downarrow][\uparrow,\,\,\,]}_{t_{2g}}\end{array}+\mbox{ligands: }\underbrace{[\uparrow,\,\,\,][\,\,\,,\,\,\,]\cdots}_{\pi^{*}}\,.$ (3) its closed-shell ground-state (1GS) configuration, $\mbox{metal: }\begin{array}[]{c}\underbrace{[\,\,\,,\,\,\,][\,\,\,,\,\,\,]}_{e_{g}^{*}}\\\ \underbrace{[\uparrow,\downarrow][\uparrow,\downarrow][\uparrow,\downarrow]}_{t_{2g}}\end{array}+\mbox{ligands: }\underbrace{[\,\,\,,\,\,\,][\,\,\,,\,\,\,]\cdots}_{\pi^{*}}\,.$ (4) Energetically close to this state is a triplet metal-centered (3MC) state which mixes with the 3MLCT state to form an avoided crossing, allowing repopulation of the $e_{g}^{*}$ orbitals via back transfer of an electron, $\mbox{metal: }\begin{array}[]{c}\underbrace{[\uparrow,\,\,\,][\,\,\,,\,\,\,]}_{e_{g}^{*}}\\\ \underbrace{[\uparrow,\downarrow][\uparrow,\downarrow][\uparrow,\,\,\,]}_{t_{2g}}\end{array}+\mbox{ligands: }\underbrace{[\,\,\,,\,\,\,][\,\,\,,\,\,\,]\cdots}_{\pi^{*}}\,.$ (5) As this 3MC state does not luminesce, its formation quenches the phosphoresence of the ruthenium(II) complex leading to shortened luminescence lifetimes. Here it is important to realize that, while we talk of states, the picture in most chemical physicists’/physical chemists’ minds is that of LFT orbitals. However most transition metal complex calculations are carried out using ††margin: DFT density-functional theory (DFT) where extracting LFT orbitals is far from obvious. Previous work showed how a LFT-like understanding of ruthenium(II) polypyridine complexes could be ††margin: PDOS obtained using a partial density of state (PDOS) method [13], in particular allowing assignment of the energy of the antibonding $e_{g}^{*}$ state which had posed a problem in previous theoretical work. This, in turn, allowed the construction of an orbital-based luminscence ††margin: LI3 index of which the third try (LI3) correlated well with experimentally-derived 3MLCT $\rightarrow$ 3MC activation energies $E_{\mbox{ave}}$ for about one hundred compounds [14]. However we have no reason to believe that $E_{\mbox{ave}}$ is a quantitative measure of the 3MLCT $\rightarrow$ 3MC barrier height, even though it is derived from experiment and so reflects general luminescence trends in ruthenium(II) polypyridine complexes. This is further confirmed by comparison of $E_{\mbox{ave}}$ with the best available experimental and theoretical values for the same barrier height [14]. It thus makes sense to carry out a more stringent test of LI3 using barriers calculated at the same level as the original PDOS calculations used to calculate LI3, namely gas- phase B3LYP/6-31G+LANLDZ(Ru) (see Sec. 2 for computational details). Such calculations are both computer-resource and human-time intensive and so we restricted our calculations to only a few ruthenium(II) polypyridine complexes. Several aspects of the results turned out to be rather surprising! (Vide infra!) The 3MLCT $\rightarrow$ 3MC process may be thought of as a chemical reaction involving the metal $d$ orbitals and the ligand $\pi^{*}$ orbitals. Just as frontier molecular orbital ††margin: FMOT theory (FMOT) has been successful in explaining many chemical reactions, similar arguments were used to motivate the LI3 orbital-based luminescence index [14] $\mbox{LI3}=\frac{\left[\left(\epsilon_{{e}_{g}^{*}}+\epsilon_{\pi^{*}}\right)/2\right]^{2}}{\epsilon_{{e}_{g}^{*}}-\epsilon_{\pi^{*}}}\,.$ (6) Notice how the form of the LI3 descriptor is typical of FMOT with an orbital energy difference in the denominator. The numerator represents the square of the off-diagonal term in the orbital hamiltonian in the Helmholtz approximation just as in familiar semi-empirical applications of FMOT, but neglecting the orbital overlap term. As shown in Fig. 2, this descriptor works remarkably well for predicting $E_{\mbox{ave}}$ for the compounds considered in the present article. It still works very well when considering a much larger number of compounds though there is necessarily more scatter in the corresponding graph and a small number of compounds appear far enough away from the line that we might start to question whether they follow the same luminescence mechanism as the other compounds [14]. This is not entirely surprising because not all bonds are Figure 2: Performance of LI3 for complexes of formula [RuX3]${}^{2}+$. Numbers in parentheses designate the compounds shown in Fig. 3 which, in turn, are named after the ligands shown in Fig. 4. symmetry equivalent in many of these compounds. Given the difficulty of a thorough investigation of reaction barriers, we decided to focus first on only a small series of similar complexes for which results are reported in the present article. These complexes are shown in Fig. 3 and the Lewis dot structures of the corresponding ligands are shown in Fig. 4. | ---|--- (a) N∧N = bpy (6) | (b) N∧N = 4,4’-dm-bpy (70) | (c) N∧N = 4,4’-dph-bpy (73) | (d) N∧N = 4,4’-DTB-bpy (74) Figure 3: Structures of [Ru(N∧N)3]2+. Hydrogen atoms have been suppressed for clarity. We have chosen to use the $\Delta$ stereoisomers, though similar results are expected for the corresponding $\Lambda$ stereoisomers. --- (a) bpy: 2,2’-bipyridine (6) (b) 4,4’-dm-bpy: 4,4’-dimethyl-2,2’-bipyridine (70) (c) 4,4’-dph-bpy: 4,4’-diphenyl-2,2’-bipyridine (73) (d) 4,4’-DTB-bpy: 4,4’-di-tert-butyl-2,2’-bipyridine (74) Figure 4: Ligand list. This series of complexes was chosen as small and “simple” as possible to ††margin: TS minimize the problem of multiple transition states (TSs) because locating TSs and determining their barrier heights can be difficult. Let us look in more detail at the conventional mechanism for ruthenium(II) photoluminescence. On the basis of experimental results and LFT group theoretical arguments, Adamson has given a set of rules for photodissociation rules for $O_{h}$ complexes [15]: > Rule 1. Trans dissociation will occur for those opposing ligands with the > weakest average ligand field strength. > Rule 2. In the case of nonidentical ligands, the ligand of greater field > strength aquates first. As Rule 1 appears to be stronger than Rule 2, ruthenium(II) polypyridine complexes have been almost universally assumed to undergo trans dissociation. This coordinate provides a 1D cut allowing us to sketch ††margin: PEC an approximate potential energy curve (PEC). The luminsence mechanism described in the first paragraph is then easy to understand in terms of the diagram in Fig. 5. Note that the 1GS Figure 5: The diagram shows the principle potential energy curves in our model. The abscissa corresponds to a reaction pathway involving partial removal of a ligand while the ordinate represents the state energy. The dashed lines indicate diabatic states whose avoided crossing leads to the energetic barrier on the adiabatic surface between the 3MLCT and 3MC minima. and 3MLCT geometries are represented as very similar, radiationless 1MLCT $\rightarrow$ 3MLCT intersystem crossing is very easy because the two curves have very similar energies and because of spin-orbit coupling due to the heavy ruthenium atom, the 1MC and 3MC geometries are also expected to be similar but with longer ruthenium-ligand bond lengths because of the occupation of the antibonding $e^{*}_{g}$ orbital, there is an avoided crossing between the 3MC and 3MLCT states, and there is easy radiationless de-activation of the 3MC state back to the 1GS. Such a picture is subject to experimental verification but obtaining barrier heights experimentally requires careful rate measurements at a series of temperatures and extraction using an appropriate model, usually assuming an Arrhenius rate law. An accurate experimental value for the 3MLCT $\rightarrow$ 3MC barrier height of 3800 cm-1 is known for the prototypical trisbipyridine ruthenium(II) cation complex [Ru(bpy)3]2+ in acetonitrile [16] (3960 cm-1 in propionitrile/buylronitile (4:5 v/v) is quoted in Ref. [14]), but for very few other complexes. The problem of obtaining experimental barrier heights has been discussed in more detail in Ref. [14] where a compromise was made in calculating ††margin: $E_{\mbox{ave}}$ $E_{\mbox{ave}}$ so as to be able to use data covering a large number of compounds. The value $E_{\mbox{ave}}$ = 132 cm-1 was found for [Ru(bpy)3]2+ which clearly underestimates the previously quoted experimental barriers, but the hope was that trends in barrier heights would be accurately captured by the experimentally-derived $E_{\mbox{ave}}$. This hypothesis is exactly what we wish to test in the present work. It might seem that determining reaction paths and barrier heights might be easier than determining them experimentally. However such complexes are expected to have a plethera of TSs. Adamson’s trans dissociation of $O_{h}$ complexes with six identical monodentate ligands is completely consistent with the expected Jahn-Teller distortion of an open-shell $O_{h}$ complex. However there are three symmetry-equivalent trans dissociation pathways, each of which should be associated with three ††margin: PES different TSs on a Mexican hat potential energy (hyper)surface (PES) [17]. This has also been discussed in the context of cobalt(II) imine complexes having a single type of bidentate ligand (e.g., Fig. 7 of Ref. [18]). In addition to the TSs of each of the three symmetry-equivalent trans dissociation products, we may also expect TSs between the symmetry-equivalent 3MLCT geometries and between the symmetry- equivalent 3MC geometries. In the case of three identical bidentatate ligands, our $O_{h}$ complex has $\Lambda$ and $\Delta$ enantiomers. These also interconvert via TSs. Interconversion between the isomers has been proposed to happen either via Bailar or by Ray-Dutt twists [19], each of which has associated TSs. And there is the question of whether the excited electron in the 3MLCT is transferred equally to all three bidentate ligands or preferentially to a single bidentate ligand? Certainly photodissociation experiments which observe the replacement of a bpy ligand with water in [Ru(bpy)3]2+ to form [Ru(bpy)2(H2O)2]2+ [20] seem consistent with the idea that symmetry breaking is occurring in such a way that one of the bpy is treated differently than the others. And this reaction also has its own TS. All of which means that we must expect many probably close-lying TSs to complicate our investigation. Of course we are not the first to carryout a theoretical investigation of 3MCLT $\rightarrow$ 3MC PECs, but the number seems to still be rather small [21, 22, 23, 24] and relatively recent. In their 2015 paper [21], Yoshikawa et al. report the energies and geometries of 3MCLT $\rightarrow$ 3MC transition states for nine [Ru(bpy)2(phen derivative)]2+, where “phen derivative” refers to either (1) phen, (2) 4-phenyl-phen, (3) 4,7-diphenyl-phen, (4) 2,9-dimethyl-4,7-diphenyl-phen, (5) 2,9-dimethyl-phen,(6) 3,4,7,8-tetramethyl- phen, (7) 5-amino-phen, (8) 5-methyl-phen, and (9) 3-phenylethynyl-phen. However no PECs were reported. In their 2016 paper [22, 23], Zhou et al. report 3MCLT $\rightarrow$ 3MC PECs for four derivatives of fac- tris(1-methyl-5-phenyl-3-n-propyl-[1,2,4]-triazolyl)iridium(III). In their 2017 paper [25], Dixon et al. report 3MCLT $\rightarrow$ 3MC PECs for [Ru(bpy)(btz)2]2+. In their 2017 paper [24], Sun et al. report 3MCLT $\rightarrow$ 3MC PECs for [Ru(bpy)3]2+ and for [Ru(mphen)3]2+. Finally, in their 2018 papers [26, 25], Soupart et al. report the 3MCLT $\rightarrow$ 3MC potential energy curves for trans [25] and for cis [26] dissociation. Fumanal et al. report transition states for [Mn(im)(CO)3(phen)]+ but without detailed potential energy curves [27]. Some of the most extensive theoretical work has been done studying [Ru(bpy)3]2+ [28, 29, 30, 25, 26] where it is clearly seen that the PEC of Fig. 5 is an oversimplification which sould be replaced with a possibly complex set of competing processes. To begin with, the 3MLCT state of [Ru(bpy)3]2+ has more than one minimum, including a minimum with $D_{3}$ symmetry where the ruthenium electron has been transferred equally to the three ligands and a minimum with $C_{3v}$ symmetry where the ruthenium electron has been transferred to a single bpy ligand [28]. The same reference reports a 3MC in order to remove the spatial degeneracy of the 3MLCT state. Spin-orbit effects can explain radiationless relaxation to the ground state via S${}_{0}/^{3}$MC mixing as in shown in Fig. 5 [29]. However LFT considerations [15] suggest that cis dissociation might also occur. Indeed theoretical calculations show that cis triplet dissociation has a comparable but higher barrier than does trans triplet dissociation [30, 25]. The trans 3MCLT $\rightarrow$ 3MC pathway has been mapped out in gas phase and solvent for [Ru(bpy)3]2+ [26]. While the overall conclusion is that complex processes are going on, the dominant one still seems to be trans dissociation. We emphasize that we have consulted with the authors of Ref. [26] and use the same techniques, but have pushed the methodology a little further in the direction of answering our driving question about LI3 and the energetics of trans dissociation. Although the present work is restricted to a small family of closely related molecules, we have nevertheless realized that great care was needed in our calculations because something strange seemed to be happening. Complex 6 ([Ru(bpy)3]2+) has $N=61$ atoms and therefore $F=3N-6=177$ internal degrees of freedom. Even though some of these degrees of freedom are fairly rigid, this still leaves a lot of possibilities for processes to occur which are different from those initially imagined! So we have made a very careful effort to be systematic: 1. 1. Our first step was to make 2D scans of the PES by fixing two trans bond lengths at different fixed values and minimizing all the other degrees of freedom. This gives us our first crude view of the reaction mechanism, but scans are well-known to be misleading when it comes to finding TSs and reaction paths (Ref. [31] gives a particularly clear explanation of why this is so.) It does, however, provide us with first guesses that are subsequently used to optimize the 3MC and 3MCLT geometries for trans dissociation. ††margin: NEB 2. 2. We then used the nudged elastic band (NEB) method [32, 33] to start from an interpolated path for trans dissociation and find a first approximation to the reaction path. This is important as only ††margin: MEP the NEB, and not the 2D scan, was able to give us a maximum energy point (MEP) close enough to a TS to be useful. 3. 3. The MEP geometry still has to be optimized using a TS optimizer. Special methods must then be used to locate the TS and to determine the reaction pathways [34, 35]. These are typically computationally resource intensive both because they require multiple single point quantum chemical calculations but also because of the need to calculate first and second derivatives of the PESs [34, 35, 36, 37, 38, 39, 40]. 4. 4. The optimized TS still has to be confirmed as linking the 3MC and 3MLCT minima and this is done by following the intrinsic reaction ††margin: IRC coordinate (IRC) in both directions along the direction of the imaginary vibrational coordinate to verify that the TS does indeed correspond to the desired reaction path. It is also possible to follow the change in character of the molecule along the IRC by using Mulliken population analysis to follow the spin- and charge- density along the reaction path. Given the surprising nature and complexity of what we found, we prefer to give a schematic of our main conclusions and then later justify them. This schematic PES is shown in Fig. 6. The details of the notation will be explained later. Briefly the initial 3MLCT state is quite symmetric so that it is difficult to distinguish three symmetry equivalent Jahn-Teller distorted geometries. The IRC is consistent with an electron being transferred to a particular bidentate ligand whose two bonds to the central ruthenium atom elongate simultaneously so as to maintain $C_{2}$ symmetry at the TS and beyond. The reaction path is then on a ridge and can bifurcate in either of two directions to find one of two symmetry-equivalent minima. This accounts for small indications here and there in the literature (and our own experience) that, and explains why, the IRC is unusually difficult to calculate for this reaction. (a) | ---|--- (b) | | Figure 6: Schematic PES representing our best understanding of the 3MLCT $\rightarrow$ 3MC reaction path: A represents three nearly identical 3MLCT minima, the Bi,j represent three TS which lead to a ridge descending to three TSs (Ci,j) representing the barrier to interconversion between symmetry- equivalent 3MC minima (Di and Dj). Somewhere along the ridge between Bi,j and Ci,j is a valley-ridge inflection (VLI) point defined mathematically as the point at which the eigenvalue of the Hessian matrix is zero and the gradient vector is perpendicular to the corresponding eigenvector) [41, 42]. The VLI may be thought of as (close to) where the 3MLCT $\rightarrow$ 3MC reaction path bifurcates [part (a) of this figure]. The rest of this paper is organized as follows: Our computational methods are described in the next section. Section 3 justifies the general picture seen in Fig. 6 and then goes on to compare calculated barrier and reaction energies with LI3. Section 4 summarizes our conclusions. Additional information is available as Supplementary Information ††margin: SI (SI) (Sec. Supplementary Information). ## 2 Computational Details The four-step procedure that was used in the present work has already been described in the introduction. This section goes into greater technical detail. We used two programs to carryout this procedure, namely version 09 revision D.01 of the Gaussian code [43] and version 5.0.2 of the ORCA code [44]. All calculations were gas phase calculations with the same basis sets and the same functionals. ††margin: 6-31G The basis consisted of the valence double-$\zeta$ quality 6-31G basis set for H [45], N [46], and C [46]. The Ru atom requires an effective core potential (ECP) in order to include relativistic effects. We used ††margin: LANL2DZ ECP the LANL2DZ ECP and the corresponding double-$\zeta$ basis set [47]. effective core potential (ECP) [47] was used to characterize Ru. H, N and C atoms were described by 6-31G basis set. ††margin: B3LYP The density functional approximation used is the B3LYP (Becke exchange, 3 parameter, Lee-Yang-Parr correlation) functional described in Ref. [48] and initially programmed in Gaussian. Unfortunately this functional has a history of confusion which can ††margin: VWN be traced back to the implementation of the Vosko-Wilke-Nusair (VWN) ††margin: LDA parameterization of the local density approximation (LDA) in Gaussian. Their parameterization of Ceperley and Alder’s quantum Monte Carlo results (found in the caption of Fig. 5 of Ref. [49]) should have been used as this is what had been referred to as VWN in works by all previous authors. However the Vosko-Wilke-Nusair parameterization ††margin: RPA of a random phase approximation (RPA) result for the electron gas result was called VWN in Gaussian. Instead of correcting the error, Gaussian refers to the original parameterization of the Ceperley- Alder ††margin: VWN5 results as VWN5. The implementation of B3LYP in Gaussian uses VWN for the LDA. However B3LYP in ORCA uses the orginal VWN (i.e., ††margin: B3LYP/G VWN5) in its implementation of B3LYP and requires the specification of B3LYP/G in order to get the original B3LYP. We used the original B3LYP in our calculations througout this paper. While there may be some value in using other functionals and better basis sets in implicit solvent, such changes would defeat our goal of calculating barriers to compare with the LI3 indices already calculated at the B3LYP/6-31G+LANLDZ(Ru) level used in Ref. [14]. So we decided to stick with the original level of calculation. In Gaussian, all calculations were carried out without any symmetry contraints, using an ultrafine grid, and the spin-unrestricted formalism. ††margin: SCF DIIS Self-consistent field (SCF) convergience was achieved using the direct inversion of iterative subspace (DIIS) extrapolation algorithm. Frequency calculations were performed on the optimized geometries so as to check whether they were the true minima (with no imaginary frequencies) or a transition state (with one imaginary frequency). For all the calculations, (pop=full) was set to full so as to extract all the information. In ORCA, keywords used included NORI (no approximation is used), TightSCF, TightOpt, SlowConv, NumFreq and an ultra-fine grid. We explicitly confirmed that Gaussian and Orca gave the same results for the same basis sets and functionals in single-point calculations. Figure 7: The $m$t$n$ labeling system for the N atoms around the central Ru. We also need some system of N atom numbering in order to be able to discuss the photochemical reaction pathways. One possibility is to just use the numbers shown in Fig. 3. However we found that we could be a little less arbitrary by using a system where each of the N∧N ligands with the numbers 1,2, and 3. Each of the N in a given N∧N is ††margin: $n$t$m$ trans to a different N∧N, so we may label as $n$t$m$ the $N$ in the $n$th N∧N that is trans to the $m$th N∧N. This is a great improvement over using the atom numbers in Fig. 3, but still is complicated by the fact that there are still six different ways to permute the three numbers (1,2,3). The possible alternative numberings are shown in Table 1. With a little practice and an occasional sketch (Fig. 7), it is relatively easy to master. Usually it is the 2t1-1t2 or the 3t1-1t3 trans stretch which interests us in our calculations. We will now take a step-by-step look at how we carried out our calculations. Many of these steps could have been carried out using either the Gaussian or the ORCA codes, but we chose to allow different people use the codes with which they were most familiar once we had assured ourselves that the two codes did indeed give the same results. Complex | Ru-N Bond | (1,2,3) | (1,3,2) | (2,1,3) | (2,3,1) | (3,1,2) | (3,2,1) ---|---|---|---|---|---|---|--- [Ru(bpy)3]2+ | 1-2 | 3t2 | 2t3 | 3t1 | 1t3 | 2t1 | 1t2 (6) | 1-3 | 3t1 | 2t1 | 3t2 | 1t2 | 2t3 | 1t3 | 1-22 | 2t1 | 3t1 | 1t2 | 3t2 | 1t3 | 2t3 | 1-23 | 2t3 | 3t2 | 1t3 | 3t1 | 1t2 | 2t1 | 1-42 | 1t3 | 1t2 | 2t3 | 2t1 | 3t2 | 3t1 | 1-43 | 1t2 | 1t3 | 2t1 | 2t3 | 3t1 | 3t2 [Ru(4,4’-dm-bpy)3]2+ | 1-2 | 3t2 | 2t3 | 3t1 | 1t3 | 2t1 | 1t2 (70) | 1-3 | 2t1 | 3t1 | 1t2 | 3t2 | 1t3 | 2t3 | 1-4 | 1t3 | 1t2 | 2t3 | 2t1 | 3t1 | 3t1 | 1-29 | 3t1 | 2t1 | 3t2 | 3t2 | 2t3 | 1t3 | 1-30 | 1t2 | 1t3 | 2t1 | 2t3 | 2t1 | 3t2 | 1-31 | 2t3 | 1t3 | 1t3 | 3t1 | 1t2 | 2t1 [Ru(4,4’-dph-bpy)3]2+ | 1-2 | 1t2 | 1t3 | 2t1 | 2t3 | 3t1 | 3t2 (73) | 1-3 | 1t3 | 1t2 | 2t3 | 2t1 | 3t2 | 3t1 | 1-4 | 3t2 | 2t3 | 3t1 | 1t3 | 2t1 | 1t2 | 1-29 | 2t1 | 3t1 | 1t2 | 3t2 | 1t3 | 2t3 | 1-39 | 2t3 | 3t2 | 1t3 | 3t1 | 2t1 | 2t1 | 1-41 | 3t1 | 2t1 | 3t2 | 1t2 | 2t3 | 1t3 [Ru(4,4’-DTB-bpy)3]2+ | 1-2 | 3t1 | 2t1 | 3t2 | 1t2 | 2t3 | 1t3 (74) | 1-3 | 1t2 | 1t3 | 2t1 | 2t3 | 3t1 | 3t2 | 1-4 | 2t3 | 3t2 | 1t3 | 3t1 | 2t1 | 2t1 | 1-68 | 3t2 | 2t3 | 3t1 | 1t3 | 2t1 | 1t2 | 1-69 | 2t1 | 3t1 | 1t2 | 3t2 | 1t3 | 2t3 | 1-70 | 1t3 | 1t2 | 2t3 | 2t1 | 3t2 | 3t1 Table 1: The six different ways to relabel N atoms as in the $n$th N∧N unit trans to the $m$th N∧N unit. With a little practice, this notation becomes fairly easy to use. #### Step 1 consists of making a 2D scan of the PES. This was done using Gaussian with the objective of finding a suitable 3MLCT minimum energy geometry. This was achieved by first doing a vertical excitation ††margin: FC from the 1GS minimum to obtain the Franck-Condon (FC) 3MLCT. This FC 3MLCT geometry was then optimized to get the 3MLCT local minimum. (It is not a global minimum!) This minimum was confirmed by checking for the absence of imaginary frequencies. We then made a simultaneous 2D-scan by independently stretching two trans Ru-N bonds and made a contour plot. This strategy was adopted from [28] which suggested that the ground state molecular orbitals are of the $d_{z^{2}}$-type and the population of this $d_{z^{2}}$-like orbital in the triplet excited state would thus result in the bond elongations (i.e., the $e^{*}_{g}$ LFT orbital is antibonding). Of course all three trans distortions should give the same reaction path. Two axial bonds that were trans to each other were independently elongated from the initial bond length in the optimized geometry of the 3MLCT state to 2.500 Å in steps of 0.002 Å, letting all other geometric parameters relax to give the lowest energy with only the two trans Ru-N distances constrained. The resulting energies had to be resorted before plotting because Gaussian uses a boustrophedon (as an ox plows a field) order when it does a 2D-scan in order to minimize changes in geometries during the scan. The lowest energy point with long trans bonds is a first guess at the 3MC geometry. This was then optimized once again to obtain the minimum energy geometry and the minima was confirmed by checking for the absence of any imaginary frequencies. Mulliken spin density analysis was used to confirm of the nature of the excited-state {0.9 excess spin-up electron on Ru [Eq. (2)] for the 3MLCT state and 1.8 excess spin-up electron on Ru for a 3MC state [Eq. (3)]}. #### Step 2 Having found our 3MLCT and 3MC end points, we then carried out an NEB calculation [32, 33] with ORCA to find a best first guess at the IRC. #### Step 3 The MEP from the NEB calculation in ORCA was further optimized using ORCA’s NEB-TS algorithm. The resultant TS was reoptimized with a different algorithm with Gaussian. #### Step 4 Finally Gaussian was used to follow the IRC from the calculated TS and insure that it is indeed connected to the originally input 3MLCT and 3MC input geometries. ## 3 Results and Discussion We are now prepared to investigate the question initially posed, namely how well the orbital-based luminescence index LI3 predicts the 3MLCT $\rightarrow$ 3MC barrier height. A first subsection will focus on the mechanism and energetics of the 3MLCT $\rightarrow$ 3MC reaction. The method of calculation of the barrier height was explained earlier as was the reason for the choice of the level of our calculations to be consistent with our earlier calculations of LI3. While we will focus on results for complex 70 (the results for the other three complexes are in the SI), we also want to reveal the correctness of the schematic PES of Fig. 6 as we believe this to be a new and novel contribution to the literature of this well-studied and important family of molecules. A second, and final, subsection will then examine the question of what we can say about the predictive value of LI3. ### 3.1 3MLCT $\rightarrow$ 3MC Mechanism and Energetics In the first instance, we focus on complex 70 and show the results for the various steps of our reaction mechanism investigation. Figure 8: Contour plot of 2D-Scan of Ru1-N3 (2t1) and Ru1-N30 (1t2) for complex 70. #### Step 1 The calculated contour plot is shown in Fig. 8. Inspection of a model shows that this plot should be symmetric about the 45∘ line representing interchange between the 2t1 and 1t2 N atoms. It is approximately but not exactly symmetric because of the existance of multiple internal degrees of freedom in this system which may be trapped in different local minima due to the boustrophedon nature of Gaussian’s 2D scan. This phenomenon is well understood [31]. Nevertheless the contour plot is nearly symmetric. It allows us to see that there is an energetically high-lying 3MLCT minimum and an energetically lower- lying 3MC minimum, separated by a barrier. Importantly, this allows us to find a first guess of the 3MC structure which is subsequently optimized. Key optimized Ru-N bond lengths are collected in Table 2. Ru-N | Ground State | 3MLCT | TS | 3MC ---|---|---|---|--- Bond | Gaussian | Orca | Gaussian | Orca | Gaussian | Orca | Gaussian | Orca [Ru(bpy)3]2+ (6) 3t2 | 2.110 (s) | 2.111 (s) | 2.096 (s) | 2.098 (s) | 2.101 (s) | 2.100 (s) | 2.115 (s) | 2.115 (s) 3t1 | 2.110 (s) | 2.111 (s) | 2.097 (s) | 2.098 (s) | 2.102 (s) | 2.102 (s) | 2.112 (s) | 2.115 (s) 2t1 | 2.110 (s) | 2.110 (s) | 2.118 (s) | 2.119 (s) | 2.211 (m) | 2.214 (m) | 2.463 (l) | 2.464 (l) 2t3 | 2.110 (s) | 2.110 (s) | 2.118 (s) | 2.119 (s) | 2.215 (m) | 2.212 (m) | 2.176 (m) | 2.170 (m) 1t3 | 2.110 (s) | 2.110 (s) | 2.097 (s) | 2.099 (s) | 2.103 (s) | 2.102 (s) | 2.164 (m) | 2.170 (m) 1t2 | 2.110 (s) | 2.110 (s) | 2.095 (s) | 2.097 (s) | 2.098 (s) | 2.101 (s) | 2.454 (l) | 2.464 (l) [Ru(4,4’-dm-bpy)3]2+ (70) 3t2 | 2.109 (s) | 2.109 (s) | 2.097 (s) | 2.097 (s) | 2.098 (s) | 2.096 (s) | 2.129 (s) | 2.112 (s) 2t1 | 2.109 (s) | 2.109 (s) | 2.097 (s) | 2.097 (s) | 2.227 (m) | 2.225 (m) | 2.437 (l) | 2.465 (l) 1t3 | 2.109 (s) | 2.109 (s) | 2.097 (s) | 2.097 (s) | 2.103 (s) | 2.103 (s) | 2.139 (m) | 2.186 (m) 3t1 | 2.109 (s) | 2.109 (s) | 2.097 (s) | 2.097 (s) | 2.103 (s) | 2.104 (s) | 2.104 (s) | 2.112 (s) 1t2 | 2.109 (s) | 2.109 (s) | 2.097 (s) | 2.097 (s) | 2.098 (s) | 2.097 (s) | 2.387 (l) | 2.465 (l) 2t3 | 2.109 (s) | 2.109 (s) | 2.097 (s) | 2.097 (s) | 2.226 (m) | 2.225 (m) | 2.222 (m) | 2.168 (m) [Ru(4,4’-dph-bpy)3]2+ (73) 1t2 | 2.104 (s) | 2.104 | 2.114 (s) | 2.094 (s) | 2.251 (m) | 2.251 (m) | 2.165 (m) | 2.459 (l) 1t3 | 2.104 (s) | 2.104 | 2.115 (s) | 2.091 (s) | 2.258 (m) | 2.258 (m) | 2.458 (l) | 2.165 (m) 3t2 | 2.104 (s) | 2.104 | 2.091 (s) | 2.098 (s) | 2.103 (s) | 2.103 (s) | 2.165 (m) | 2.108 (s) 2t1 | 2.104 (s) | 2.104 | 2.087 (s) | 2.094 (s) | 2.099 (s) | 2.099 (s) | 2.108 (s) | 2.457 (l) 2t3 | 2.104 (s) | 2.104 | 2.094 (s) | 2.091 (s) | 2.106 (s) | 2.106 (s) | 2.108 (s) | 2.165 (m) 3t1 | 2.104 (s) | 2.104 | 2.086 (s) | 2.098 (s) | 2.101 (s) | 2.101 (s) | 2.456 (l) | 2.108 (s) [Ru(4,4’-DTB-bpy)3]2+ (74) 3t1 | 2.107 (s) | 2.107 | 2.130 (s) | 2.130 (s) | 2.229 (m) | 2.230 (m) | 2.357 (l) | 2.465 (l) 1t2 | 2.107 (s) | 2.107 | 2.092 (s) | 2.092 (s) | 2.102 (s) | 2.102 (s) | 2.106 (s) | 2.166 (s) 2t3 | 2.107 (s) | 2.107 | 2.075 (s) | 2.073 (s) | 2.090 (s) | 2.090 (s) | 2.210 (m) | 2.110 (m) 3t2 | 2.107 (s) | 2.107 | 2.130 (s) | 2.130 (s) | 2.231 (m) | 2.231 (m) | 2.357 (l) | 2.167 (m) 2t1 | 2.107 (s) | 2.107 | 2.092 (s) | 2.092 (s) | 2.102 (s) | 2.102 (s) | 2.105 (s) | 2.109 (s) 1t3 | 2.107 (s) | 2.107 | 2.075 (s) | 2.073 (s) | 2.090 (s) | 2.090 (s) | 2.210 (m) | 2.459 (l) Table 2: Key Ru-N bond lengths (Å) for different compounds as computed with Gaussian and with ORCA. In parentheses: “s” stands for “short” ($\sim$ 2.0 Å), “m” for “medium length” ($\sim$ 2.1 Å), and “l” for “long” ($\sim$ 2.4 Å). Figure 9: NEB-TS optimization for complex 70. The initial guess IRC is in black, the final converged IRC is in red, and the intermediate IRCs obtained during the NEB procedure are shown in grey. #### Step 2 Figure 9 shows the iterative convergence of the NEB IRC from its initial guess to its final result. This is an expensive calculation involving the simultaneous displacement of 24 points (the end points are fixed) in a parallelized calculation. It is clear that its convergence is approximate. Figure 10: Plot of the NEB IRC for complex 70. The red circles are on top of the black squares. The blue point is the energy of the optimized TS. #### Step 3 Figure 10 shows the IRC obtained from the NEB calculation. In this particular case, the MEP optimizes to a point which gives a considerably lower barrier. The geometries of the optimized TSs obtained with Gaussian and Orca may compared in Table 2. Agreement is to within 0.002 Å. Figure 11: Plot of the Gaussian total energy in Hartree against the IRC for complex 70. #### Step 4 Figure 11 shows the more accurate IRC obtained using Gaussian. Note that this figure evolves from left to right in the 3MLCT $\rightarrow$ 3MC direction while Fig 10 evolves from left to right in the 3MC $\rightarrow$ 3MLCT direction. These calculations confirmed that the TS is connected by the IRC to the originally input 3MC and 3MLCT structures. Figure 12: Plot of bond lengths along the IRC for complex 70 as calculated with Orca. Figure 13: Plot of bond lengths along the IRC for complex 70 as calculated with Gaussian. #### Reaction Mechanism We have also looked at spin densities, but we have found the evolution of Ru-N bond lengths along the IRC to be the most valuable indicator of how the reaction mechanism proceeds. Figures 12 and 13 show how these bond lengths vary along the IRC during the photoreaction. The NEB calculation (Fig. 12) from left to right follows the 3MC $\rightarrow$ 3MLCT reaction. It is less accurate than the IRC calculation carried out with Gaussian (Fig. 13) which from left to right follows the 3MLCT $\rightarrow$ 3MC direction. Only the Gaussian IRC is accurate enough to understand what is going on. From right to left: The initial 3MLCT Ru-N bond lengths are nearly equal until the IRC is reduced to about 3, at which point the 2t1 and 2t3 bond lengths remain equal but increase while the other bond lengths continue to stay about the same. This means that ligand 2 is moving away from the metal in a way that might be expected if the $e_{g}^{*}$ electron on Ru had been transferred to a $\pi^{*}$ orbital on ligand 2. This continues beyond the TS (at IRC coordinate 0 in Fig. 13) until we get to an IRC coordinate of about -5. So, up to this point, one of the ligands is coming off by elongating both of its Ru-N bonds equally. When the IRC is less than -5, there is a sudden change in the bond lengths, so that the two trans Ru-N pairs (2t1 and 1t2) become longest while two other pairs (3t2 and 2t2) reduce to a medium bond length. Two Ru-N pairs (3t2 and 3t1) remain short. Such behavior is consistent with the PES scheme shown in Fig. 6. Such a PES should show a bifurcation such that different trans bond lengths could lengthen in different calculations. This is seen in Table 2 for complexes 73 and 74, confirming our overall interpretation of the trans dissociation reaction mechanism. ### 3.2 What Does LI3 Predict? We established in our earlier work that the orbital-based luminescence index LI3 correlates well with $E_{\mbox{ave}}$ values of the 3MLCT $\rightarrow$ 3MC barrier derived from experimental luminescence data obtained at room temperature and at the boiling point of liquid nitrogen [14]. This has been illustrated in Fig. 2. The question is not whether LI3 works (it does!) but how it is able to do this. We will address this question more fully in this subsection. Complex | $\tau$/$\mu$s (77K) | $\tau$/$\mu$s (RT) | $\Delta E_{\mbox{ave}}$/cm-1 ---|---|---|--- (6) | 5.23 | 0.845 | 132. (70) | 4.6 | 0.525 | 157. (73) | 4.79 | 1.31 | 94. (74) | 9.93 | 0.673 | 194. Table 3: Luminescence lifetimes at both room temperature (RT) and liquid nitrogen temperature (77K) along with the empirically-derived $\Delta E_{\mbox{ave}}$ as reported in Tables 10 and 11 of Ref. [14]. It is perhaps appropriate to recall that the fundamental 3MLCT $\rightarrow$ 3MC reaction involves an electron transfer. One famous theory of electron transfer is Marcus theory [50] and Marcus theory has indeed been used in connection with studying electrogenerated chemiluminescence in tris(polypyridine) ruthenium complexes [51]. Marcus theory is valid when the Massey parameter, $\Gamma=\frac{2\pi|H_{1,2}|^{2}}{|d(E_{1}-E_{2})/dt|}\,,$ (7) is small, where $E_{1}$ and $E_{2}$ are the energies of the two diabatic curves (3MLCT and 3MC here) and $H_{1,2}$ is the adiabatic coupling. Diagonalizing the corresponding $2\times 2$ matrix gives the adiabatic curves, corresponding to the PESs treated in the present work. Marcus theory assumes rapid crossing of the avoided crossing region in the weak coupling case (i.e., when $E_{1}\approx E_{2}$). A detailed derivation [52] shows that Marcus theory assumes rapid oscillation between the reactants and products. The final formula for the reaction rate constant, expressed in terms of Gibb’s free energies $G$ is, $k=|G_{R,P}|^{2}\sqrt{\frac{\pi}{\lambda k_{B}T}}e^{-\frac{\left(\lambda+\Delta G^{0}\right)^{2}}{4\lambda k_{B}T}}\,,$ (8) where the relaxation parameter $\lambda$ is the difference between the product diabatic curve at the reactant minimum and the same curve at the product minimum. In contrast, the strong coupling limit combined with a no-recrossing rule, leads to Eyring’s transition state theory [53] and the equally well- known formula, $k=\frac{\kappa k_{B}T}{2\pi}e^{-{\frac{\Delta G^{\ddagger}}{RT}}}\,,$ (9) which explains the famous Arrhenius equation, $k=Ae^{-\frac{E_{a}}{RT}}\,,$ (10) familiar from first-year University chemistry courses. Marcus theory and Eyring transition state theory are, in fact, just two limiting cases of the more general theory of charge transfer reactions [54]. We have not defined every parameter in these well-known equations because we do not need them in the present work and because we are sure that the interested reader can easily find suitable references to fill in any missing details. But it is evident that temperature $T$ is in these equations which may at first seem a little strange for a photochemical reaction. A justification is that the luminescence lifetimes of our complexes are on the order of microseconds (Table 3) whereas a typical vibrational time is on the order of picoseconds, which suggests that some degree of thermodynamic equilibration may occur, even if it is not necessarily complete. In turn, this explains why the usual analysis of barriers for these complexes is carried out in terms of the Arrhenius equation with some additional terms involving equilibrium between near-lying electronic states and a melting term (see Ref. [14] and references therein). In our case, we are only interested in the energetics of the 3MLCT $\rightarrow$ 3MC reaction and the extent to which it correlates with LI3. We will assume the Arrhenius picture with the same freqency factor $A$ for all our complexes so that only the activation energy $E_{a}$ is important. Of course, we are also limited by the fact that we are using the same level of calculation of the PESs that we used to calculate our LI3s. We have to accept in advance that this level of calculation may or may not be adequate for directly explaining the experimental observations, even if we expect it to shed light upon our problem. Energy\Complex | 6 | 70 | 73 | 74 ---|---|---|---|--- LI3 | 16.78 eV | 13.78 eV | 9.68 eV | 11.97 eV 3MLCT | -1579.29913 Ha | -1815.18022 Ha | -2965.37532 Ha | -2522.64515 Ha 3MC | -1579.30474 Ha | -1815.18504 Ha | -2965.37816 Ha | -2522.64934 Ha 3MEP | -1579.29661 Ha | -1815.17613 Ha | -2965.37120 Ha | -2522.64201 Ha 3TS | -1579.29769 Ha | -1815.17880 Ha | -2965.37392 Ha | -2522.64328 Ha 3MEP-3MLCT | 0.00252 Ha | 0.00409 Ha | 0.00412 Ha | 0.00314 Ha | 553.08 cm-1 | 897.65 cm-1 | 904.24 cm-1 | 689.15 cm-1 3TS-3MLCT | 0.00144 Ha | 0.00142 Ha | 0.0014 Ha | 0.00187 Ha | 316.04 cm-1 | 311.65 cm-1 | 307.26 cm-1 | 410.42 cm-1 3MEP-3MC | 0.00813 Ha | 0.00891 Ha | 0.00696 Ha | 0.00733 Ha | 1784.33 cm-1 | 1955.52 cm-1 | 1527.54 cm-1 | 1608.75 cm-1 3TS-3MC | 0.00705 Ha | 0.00624 Ha | 0.00424 Ha | 0.00606 Ha | 1547.52 cm-1 | 1369.52 cm-1 | 930.57 cm-1 | 1330.02 cm-1 3MC-3MLCT | 0.005610 Ha | 0.00482 Ha | 0.00284 Ha | 0.004190 Ha | 1231.25 cm-1 | 1057.87 cm-1 | 623.31 cm-1 | 919.60 cm-1 Table 4: Energies for 3MLCT, 3MC and TS of 4 compounds obtained from NEB calculations as implemented in the ORCA code. Table 4 gives energies for the 3MLCT states, 3MC states, and TSs for the four compounds obtained from NEB calculations as implemented in the Orca code. Values in parenthesis are in cm-1. Except for the 3MEP values which come from NEB calculations, the same numbers are obtained with Gaussian. It is immediately clear that the 3MLCT $\rightarrow$ 3MC barrier (3MEP-3MLCT or, for more accuracy, 3TS-3MC) is really very small and on the order of about 300-400 cm-1 (0.86-1.14 kcal/mol). Chemical accuracy is usually cited as 1 kcal/mol (349.757 cm-1) and is very hard to achieve even with the best quantum chemical methods. Figure 14: Graph of the 3MLCT-3MC energy difference versus LI3. Figure 14 shows that the 3MLCT-3MC energy difference correlates very well with L3—in fact even better than the correlation of $\Delta E_{\mbox{ave}}$ with LI3 (Fig. 2). This is consistent with the origin of the FMOT origin of the orbital-based luminescence index LI3. Figure 15: Explanation of the Bell-Evans-Polanyi model [55, 56] in the present context: RE, reactant PEC, $E_{R}(x)=(k_{R}/2)(x-x_{R})^{2}+E_{R}^{0}$; LRE, linearized RE, $E_{R}(x)=m_{R}(x-x_{R})+E_{R}^{0}$; PE, product PEC, $E_{P}(x)=(k_{P}/2)(x-x_{P})^{2}+E_{P}^{0}$; LPE, linearized PE, $E_{P}(x)=m_{P}(x-x_{P})+E_{P}^{0}$. Here $x$ is the reaction coordinate. versus There is an argument that the 3TS-3MLCT energy barrier should vary linearly with the 3MLCT-3MC energy difference, $E[\mbox{${}^{3}$TS}]-E[\mbox{${}^{3}$MLCT}]=\alpha\left(E[\mbox{${}^{3}$MC}]-E[\mbox{${}^{3}$MLCT}]\right)+E_{0}\,,$ (11) and hence that the 3TS-3MLCT energy barrier should also correlate quite nicely with L3. This can be understood in terms of two parabollic intersecting diabatic curves as shown in Fig. 15. The intersection point $(x_{a},E_{a}+E_{R}^{0})$ defines the activation energy ($E_{a}$). The corresponding reaction coordinate may be found by solving the equation, $x_{a}=\frac{x_{P}+x_{R}}{2}+\frac{\Delta E^{0}+\left(\frac{\Delta k}{2}\right)(x_{a}-x_{R})^{2}}{k_{P}(x_{P}-x_{R})}\,,$ (12) where $\Delta E^{0}=E_{P}^{0}-E_{R}^{0}$ and $\Delta k=k_{P}-k_{R}$. Equation (12) may be solved for $x_{a}$ in many different ways, but taking $\Delta k=0$ as an initial guess and then interating leads to rapid convergence in the case we tried. Back in 1936, Bell, Evans, and Polanyi [55, 56] presented an argument which is easily used in the present context. It requires linearlization of the parabolas (dotted lines in Fig. 15) and then solving, which gives, $E_{a}=\underbrace{\frac{m_{R}}{m_{R}-m_{P}}}_{\alpha}\Delta E^{0}+\underbrace{E_{R}^{0}+\frac{m_{R}m_{P}}{m_{P}-m_{R}}(x_{P}-x_{R})}_{E_{0}}\,.$ (13) Identification of the reactant as the 3MLCT state and the product as the 3MC state, then gives Eq. (11). Solving with the full parabolas (which is a step in the derivation of Marcus theory) would lead to an additional quadratic term, which we will ignore here. But this derivation was important here because it underlines that we need a PEC that resembles two intersecting parabolas, which is evidently not the case in Fig. 11 because of the plateau in the IRC curve caused by the presence of a bifurcation. For this reason, we do not expect the 3TS-3MLCT energy to follow the Bell-Evans-Polanyi postulate, even approximately. However the first guess IRC of the NEB does not go through the bifurcation region and so in in approximate agreement with the Bell-Evans-Polanyi postulate as shown in Fig. 16. As might be expected, after a little reflection, the slope of the first guess NEB MEPs opposite would be opposite to that seen in Fig. 2 for $E_{\mbox{ave}}$. Also, as previously mentioned our best results (namely 3TS-3MLCT) are actually essentially constant. Figure 16: 3MLCT $\rightarrow$ 3MC barrier heights: black square, accurate TS; black diamond, converged NEB MEP; black downward facing triangle, initial NEB guess MEP. We are now left with a final conundrum. All our calculations for the four complexes studied in this article suggest that trans dissociation kinetics is actually unable to explain the observed correlation between the experimentally-derived $E_{\mbox{ave}}$ and LI3, because our theoretically- calculated 3TS-3MLCT is complex-independent to within the expected accuracy of our methodology. Nevertheless LI3 is positively correlated with the 3MLCT-3MC energy difference as expected from FMOT and this leads to a strong correlation between $E_{\mbox{ave}}$ and the 3MLCT-3MC energy difference $\Delta E^{0}$. Why? Could it be that our experimentally-derived $E_{\mbox{ave}}$ reflects something other than kinetics? Afterall, the measured luminescence lifetimes are orders of magnitude longer than typical vibrational lifetimes. This might allow some sort of quasiequilibrium to take place. Indeed, our observations up to this point are more consistent with the hypothesis that 3MLCT states are being removed via a fast 3MLCT = 3MC equilibrium process whose equilibrium constant is roughly, $\frac{[\mbox{${}^{3}$MC}]}{[\mbox{${}^{3}$MLCT}]}=K_{eq}=e^{-\Delta G^{0}/RT}\propto e^{-\Delta E^{0}/RT}\,.$ (14) Hence, all other things being equal, a larger 3MLCT-3MC means a more negative value of $\Delta E^{0}$ and hence a larger ratio of 3MC product in comparison to the 3MLCT reactant, effectively removing the luminescent 3MLCT state in favor of the nonluminescent 3MC state and leading to shorter observed luminescence lifetimes which were previously interpretted in terms of larger values of $E_{\mbox{ave}}$. Of course, we cannot rule out the presence of other competing pathways leading to different products and transition states and hence the possibility of completing kinetics and quasiequilibria. ## 4 Conclusion This may be regarded as the third in a series of articles seeking an orbital- based explanation of luminescence life-times in ruthenium(II) polypyridine complexes. Article I [13] introduced the use of the partial density of states (PDOS) for the extraction of a ligand-field theory (LFT) like picture from density-functional theory (DFT) calculations. Direct visualization of molecular orbitals (MOs) is insufficient for this purpose because the $e_{g}^{*}$ metal MOs mix heavily with ligand MOs. However the PDOS analysis allows the assignment of an energy to these orbitals in a quantitative, albeit fuzzy and both basis set and functional dependent, fashion. This allowed the investigation of on the order of 100 molecules in Article II [14] to develop an orbital-based luminescence index, called LI3, which correlated strongly with an average “activation energy” $E_{\mbox{ave}}$ calculated from experimental luminescence lifetimes measured in solution at room temperature and liquid nitrogen temperature. The work in Articles I and II is based upon the widely assumed 3MLCT $\rightarrow$ 3MC trans distortion mechanism and made use of ideas from frontier molecular orbital theory (FMOT). However it was made clear in Article II that $E_{\mbox{ave}}$ is quite different from an accurately determined experimental 3MLCT $\rightarrow$ 3MC activation energy for [Ru(bpy)3]2+, complex 6 in the present article, for which $E_{\mbox{ave}}$ = 132 cm-1 [14], but for which experimental studies in solution gave 3800 cm-1 [16] and 3960 cm-1 depending upon the solvent and details of the analysis. The present article corrects this somewhat unsatisfactory state of affairs by carrying out calculations of the transition states (TSs) and intrinsic reaction coordinates (IRCs) for four closely related ruthenium(II) bipyridine complexes under the same conditions as used to calculate LI3 in Article II, namely gas phase calculations with the B3LYP functional and the 6-31G+LANLDZ(Ru) basis set. Our calculated 3MLCT $\rightarrow$ 3MC TS barrier is 316.04 cm-1 which may be compared with the previously calculated theoretical gas phase value of 700 cm-1 [11]. It is important to note that we used the same nudged elastic band (NEB) methodology as in Ref. [11], though the basis set and functional differ, but made a more thorough investigation of our four molecules using more accurate TS and IRC searches. The result carries some surprises. For example, our NEB maximum energy point (MEP) gives a TS barrier of 553 cm-1 (closer to the 700 cm-1 value quoted in Ref. [11], particularly when considering that chemical accuracy is often quoted as 1 kcal/mol = 350 cm-1). But we realized that the NEB MEP is not sufficiently accurate for our purposes and so used specific TS optimizers and proved the correctness of our optimized TS by calculation of the IRC. Examination of Ru-N bond lengths along the IRC showed some surprising results which can only be explained by a TS corresponding to one of the ligands symmetrically lengthening its Ru-N bonds, hence going through a sort of cis TS consistent with electron transfer to a single ligand, rather than symmetrically to all three ligands. The trans dissociation then continues along a ridge and then bifurcates into one of two symmetry-equivalent minima. Once this is taken into account, not only is a lower TS barrier obtained, but a much richer and more complex description is obtained of the trans dissociation mechanism. Returning to LI3 and the energetics of trans dissociation, we find that the 3MLCT-3MC energy correlates linearly with LI3 but that the 3TS-3MLCT is, to within the accuracy of our calculation, molecule independent. Attempts to make alternative kinetic arguments, such as by using the Bell-Evans-Polanyi postulate, only lead to ideas in contradiction with the observed relation between $E_{\mbox{ave}}$ and LI3. This suggests that it is the total energy difference of the reaction, rather than the barrier height, which determines the luminescence lifetime. That is, we are looking at a quasielquilibrium property rather than a kinetic property. Given that the measured luminescence lifetimes are on the order of $\mu$s which is much several orders of magnitude longer than a typical vibrational time, then such an assumption is not unreasonable. Our picture of what is going on in our study is limited by it being a gas- phase study with a particular basis set and a particular functional. It is also limited by the fact that only one luminescence quenching mechanism has been studied in a family of molecules which is rich in degrees of freedom and hence also rich in dissociation mechanisms as shown by the brief review of related work given in the introduction of this article. Nevertheless we think that our work brings a new and more detailed picture of part of what could be going on in the luminescence mechanism for this important family of molecules. ## Acknowledgement DM and MEC gratefully acknowledge helpful funding from the African School on Electronic Structure Methods and Applications (ASESMA), ASESMANet, and the US- Africa Initiative. We thank the following people for insightful discussions: Ragnar Björnsson, Isabelle M. Dixon, Jean Louis Heully, Valid Mwatati, Max Daku Latévi Lawson, Sandeep Sharma. In particular, Isabelle M. Dixon is gratefully acknowledged for sharing unpublished computational results with us early on in our project. We would like to thank Pierre Girard and Sébastien Morin for technical support in the context of the Grenoble Centre d’Experimentation du Calcul Intensif en Chimie (CECIC) computers used for the calculations reported here. ## Supplementary Information Only plots for complex 70 have been used in the main article. The corresponding plots for complexes 6, 73, and 74 are available on-line: 1. 1. Contour plots 2. 2. NEB optimizations 3. 3. Converged NEB Minimum Energy Paths 4. 4. IRC Energy Profiles 5. 5. Variation of Metal-Ligand Bond Lengths Along the NEB Reaction Path 6. 6. Variation of Metal-Ligand Bond Lengths Along the IRC 7. 7. Author contributions ## References * [1] D. Magero, T. Mestiri, K. Alimi, and M. E. Casida, Computational studies of ruthenium and iridium complexes for energy sciences and progress on greener alternatives, in Green Chemistry and Computational Chemistry: Shared Lessons in Sustainability, edited by L. Mammino, pages 115–145, Elsevier, 2020, preprint: https://arxiv.org/abs/2004.03345. * [2] J. P. Sauvage, J. Collin, J. C. Chambron, S. Guillerez, and C. Coudret, Ruthenium(II) and osmium(II) bis(terpyridine) complexes in covalently-linked multicomponent systems: Synthesis, electrochemical behavior, absorption spectra, and photochemical and photophysical properties, Chem. Rev. 94, 993 (1994). * [3] K. Nakamaru, Synthesis, luminescence quantum yields, and lifetimes of trischelated ruthenium (II) mixed-ligand complexes including 3,3’-dimethyl-2,2’-bipyridyl, Bull. Chem. Soc. Jpn. 55, 2697 (1982). * [4] G. Liebsch, I. Klimant, and O. S. Wolfbeis, Luminescence lifetime temperature sensing based on sol-gels and poly(acrylonitrile)s dyed with ruthenium metal-ligand complexes, Adv. Mat. 11, 1296 (1999). * [5] V. Balzani and A. Juris, Photochemistry and photophysics of Ru (II) polypyridine complexes in the bologna group. From early studies to recent developments, Coordin. Chem. Rev. 211, 97 (2001). * [6] M. Duati, S. Tasca, F. C. Lynch, H. Bohlen, J. G. Vos, S. Stagni, and M. D. Ward, Enhancement of luminescence lifetimes of mononuclear ruthenium(II)-terpyridine complexes by manipulation of the $\sigma$-donor strength of ligands, J. Inorg. Chem. 42, 8377 (2003). * [7] A. Harriman, A. Khatyr, and R. Ziessel, Extending the luminescence lifetime of ruthenium(II) poly(pyridine) complexes in solution at ambient temperature, Dalton Trans. 10, 2061 (2003). * [8] E. A. Medlycott and G. S. Hanan, Designing tridentate ligands for ruthenium(II) complexes with prolonged room temperature luminescence lifetimes, Chem. Soc. Rev. 34, 133 (2005). * [9] L. J. Nurkkala, R. O. Steen, H. K. J. Friberg, J. A. Häggström, P. V. Bernhardt, M. J. Riley, and S. J. Dunne, The effects of pendant vs. fused thiophene attachment upon the luminescence lifetimes and electrochemistry of tris(2,2’-bipyridine)ruthenium(II) complexes, Eur. J. Inorg. Chem. 26, 4101 (2008). * [10] S. Ji, W. Wu, W. Wu, P. Song, K. Han, Z. Wang, S. Liu, H. Guo, and J. Zhao, Tuning the luminescence lifetimes of ruthenium(II) polypyridine complexes and its application in luminescent oxygen sensing, J. Mater. Chem. 20, 1953 (2010). * [11] A. Soupart, F. Alary, J. Heully, P. I. P. Elliot, and I. M. Dixon, Theoretical study of the full photosolvolysis mechanism of [Ru(bpy)3]2+: Providing a general mechanistic roadmap for the photochemistry of [Ru(N$\wedge$N)3]2+-type complexes toward both cis and trans photoproducts, Inorg. Chem. 59, 14679 (2020). * [12] V. Balzani, P. Ceroni, A. Credi, and M. Venturi, Ruthenium tris(bipyridine) complexes: Interchange between photons and electrons in molecular-scale devices and machines, Coord. Chem. Rev. 433, 213758 (2021). * [13] C. M. Wawire, D. Jouvenot, F. Loiseau, P. Baudin, S. Liatard, L. Njenga, G. Kamau, and M. E. Casida, Density-functional study of luminescence in polypyrine ruthenium complexes, J. Photochem. and Photobiol. A 276, 8 (2014). * [14] D. Magero, M. E. Casida, G. Amolo, N. Makau, and L. Kituyi, Partial density of states ligand field theory (PDOS-LFT): Recovering a LFT-like picture and application to photoproperties of ruthenium(II) polypyridine complexes, J. Photochem. Photobiol. A 348, 305 (2017). * [15] L. G. Vanquickenborne and A. Ceulemans, Ligand-field models and the photochemistry of coordination compounds, Coord. Chem. Rev. 48, 157 (1983). * [16] J. V. Caspar and T. J. Meyer, Photochemistry of tris(2,2’-bipyridine) ruthenium(ii) ion ([Ru(bpy)3]2+). Solvent effects., J. Am. Chem. Soc. 105, 5583 (1983). * [17] I. B. Bersuker, Modern aspects of the jahn-teller effect theory and applications to molecular problems, Chem. Rev. 101, 1067 (2001). * [18] I. Krivokapic, M. Zerara, M. Lawson Daku, A. Vargas, C. Enachescu, C. Ambrus, P. Tregenna-Piggott, N. Amstutz, E. Krausz, and A. Hauser, Spin-crossover in cobalt(II) imine complexes, Coord. Chem. Rev. 251, 364 (2007). * [19] D. C. Ashley and E. Jakubikova, Ray-Dutt and Bailar twists in Fe(II)-tris(2,2’-bipyridine): Spin states, sterics, and Fe-N bond strengths, Inorg. Chem. 57, 5585 (2018). * [20] A. Vaidyalingam and P. K. Dutta, Analysis of the photodecomposition products of Ru(bpy)${}_{3}^{2+}$ in various buffers and upon zeolite encapsulation, Anal. Chem. 72, 5219 (2000). * [21] N. Yoshikawa, S. Yamabe, S. Sakaki, N. Kanehisa, T. Inoue, and H. Takashima, Transition states of the 3MLCT to 3MC conversion in Ru(bpy)2(phen derivative)2+ complexes, J. Mol. Struct. 1094, 98 (2015). * [22] X. W. Zhou, P. L. Burn, and B. J. Powell, Bond fission and non-radiative decay in iridium(III) complexes, Inorg. Chem. 55, 5266 (2016). * [23] X. W. Zhou, P. L. Burn, and B. J. Powell, Correction to “bond fission and non-radiative decay in iridium(III) complexes”, Inorg. Chem. 56, 7574 (2017). * [24] Q. Sun, B. Dereka, E. Vauthey, L. M. L. Daku, and A. Hauser, Ultrafast transient IR spectroscopy and DFT calculations of ruthenium(II) polypyridyl complexes, Chem. Sci. 8, 223 (2017). * [25] A. Soupart, I. M. Dixon, F. Alary, and J. Heully, DFT rationalization of the room-temperature luminescence properties of Ru(bpy)${}_{3}^{2+}$ and Ru(tpy)${}_{3}^{2+}$: 3MLCT-3MC minimum energy path from neb calculations and emission spectra from vres calculations, Theor. Chem. Acc. 137, 37 (2018). * [26] A. Soupart, F. Alary, J. Heully, P. I. P. Elliott, and I. M. Dixon, Exploration of the uncharted 3PES territory for [Ru(bpy)3]2+] A new 3MC minimum prone to ligand loss photochemistry, Inorg. Chem. 57, 3192 (2018). * [27] M. Fumanal, Y. Harabuchi, E. Gindensperger, S. Maeda, and C. Daniel, Excited-state reactivity of [Mn(im)(CO)3(phen)]+: A structural exploration, J. Comput. Chem. 40, 72 (2019). * [28] F. Alary, J. L. Heully, L. Bijeire, and P. Vicendo, Is the 3MLCT the only photoreactive state of polypyridyl complexes?, Inorg. chem. 46, 3154 (2007). * [29] J. Heully, F. Alary, and M. Boggio-Pasqua, Spin-orbit effects on the photophysical properties of [Ru(bpy)3]2+, J. Chem. Phys. 131, 184308 (2009). * [30] I. M. Dixon, J. Heully, F. Alary, and P. I. P. Elliott, Theoretical illumination of highly original photoreactive 3MC states and the mechanism of the photochemistry of Ru(II) tris(bidentate) complexes, Phys. Chem. Chem. Phys. 19, 27765 (2017). * [31] K. Müller, Reaction paths on multidimensional energy hypersurfaces, Angew. Chem. Int. Ed. Engl. 19, 1 (1980). * [32] H. Jónsson, G. Mills, and K. W. Jacobsen, Nudged elastic band method for finding minimum energy paths of transitions, in Classical and quantum dynamics in condensed phase simulations, pages 385–404, World Scientific, 1998. * [33] V. Ásgeirsson, B. O. Birgisson, R. Bjornsson, U. Becker, F. Neese, C. Riplinger, and H. Jónsson, Nudged elastic band method for molecular reactions using energy-weighted springs combined with eigenvector following, Journal of chemical theory and computation 17, 4929 (2021). * [34] H. B. Schlegel, Exploring potential energy surfaces for chemical reactions: An overview of some practical methods, J. Comput. Chem. 24, 1514 (2003). * [35] H. B. Schlegel, Geometry optimization, WIREs Comput. Mol. Sci. 1, 790 (2011). * [36] C. Peng and H. B. Schlegel, Combining synchronous transit and quasi-Newton methods to find transition states, Israel J. Chem. 33, 449 (1993). * [37] C. Peng, P. Y. Ayala, H. B. Schlegel, and M. J. Frisch, Using redundant internal coordinates to optimize equilibrium geometries and transition states, J. Comput. Chem. 17, 49 (1996). * [38] A. Heyden, A. T. Bell, and F. J. Keil, Efficient methods for finding transition states in chemical reactions: Comparison of improved dimer method and partitionned rational function optimization method, J. Chem. Phys. 123, 224101 (2005). * [39] S. M. Sharada, P. M. Zimmerman, A. T. Bell, and M. Head-Gordon, Automated transition state searches without evaluating the Hessian, J. Chem. Theory Comput. 8, 5166 (2012). * [40] A. C. Vaucher and M. Reiher, Minimum energy paths and transition states by curve optimization, J. Chem. Theory Comput. 14, 3091 (2018). * [41] S. Maeda, Y. Harabuchi, Y. Ono, T. Taketsugu, and K. Morokuma, Intrinsic reaction coordinate: Calculation, bifurcation, and automated search, Int. J. Quant. Chem. 115, 258 (2015). * [42] D. H. Ess, S. E. Wheeler, R. G. Iafe, L. Xu, N. Celebi-Oelcuem, and K. N. Houk, Bifurcations on potential energy surfaces of organic reactions, Angewandte Chemie International Edition 47, 7592 (2008). * [43] M. J. Frisch et al., Gaussian 09 Revision D.01, Gaussian Inc. Wallingford CT 2009. * [44] F. Neese, F. Wennmohs, U. Becker, and C. Riplinger, The orca quantum chemistry program package, The Journal of chemical physics 152, 224108 (2020). * [45] R. Ditchfield, W. Hehre, and J. A. Pople, Self-consistent molecular-orbital methods. ix. an extended gaussian-type basis for molecular-orbital studies of organic molecules, J. Chem. Phys. 54, 724 (1971). * [46] W. J. Hehre, R. Ditchfield, and J. A. Pople, Self—consistent molecular orbital methods. xii. further extensions of gaussian—type basis sets for use in molecular orbital studies of organic molecules, J. Chem. Phys. 56, 2257 (1972). * [47] P. J. Hay and W. R. Wadt, Ab initio effective core potentials for molecular calculations. Potentials for K to Au including the outermost core orbitals, J. Chem. Phys. 82, 299 (1985), Ab initio effective core potentials for molecular calculations. Potentials for K to Au including the outermost core orbitals. * [48] P. J. Stephens, F. J. Devlin, C. F. Chabalowski, and M. J. Frisch, Ab initio calculation of vibrational absorption and circular dichroism spectra using density functional force fields, J. Phys. Chem. 98, 11623 (1994). * [49] S. H. Vosko, L. Wilk, and M. Nusair, Accurate spin-dependent electron liquid correlation energies for local spin density calculations: a critical analysis, Can. J. Phys. 58, 1200 (1980). * [50] R. A. Marcus, On the theory of oxidation-reduction reactions involving electron transfer. I, J. Chem. Phys. 24, 966 (1956). * [51] F. Kanoufi and A. J. Bard, Electrogenerated chemiluminescence. 65. An investigation of the oxidation of oxalate by tris(polypyridine) ruthenium complexes and the effect of the electrochemical steps on the emission intensity, J. Chem. Phys. B 103, 10469 (1999). * [52] G. F. Bertsch, Derivations of Marcus’s formula, http://www.int.washington.edu/users/bertsch/marcus.1.pdf, last accessed 6 May 2020. * [53] K. J. Laldler and M. C. King, The development of transition-state theory, J. Phys. Chem. 87, 2657 (1983). * [54] P. F. Barbara, T. J. Meyer, and M. A. Ratner, Contemporary issues in electron transfer research, J. Phys. Chem. 100, 13148 (1996). * [55] R. P. Bell, The theory of reactions involving proton transfers, Proc. R. Soc. London, Ser. A 154, 414 (1936). * [56] M. G. Evans and M. Polanyi, Further considerations on the thermodynamics of chemical equilibria and reaction rates, J. Chem. Soc., Faraday Trans. 32, 1333 (1936).
 # Efficiency of Nonthermal Particle Acceleration in Magnetic Reconnection Masahiro Hoshino Department of Earth and Planetary Science, The University of Tokyo, Tokyo 113-0033, Japan<EMAIL_ADDRESS> ###### Abstract The nonthermal particle acceleration during magnetic reconnection remains a fundamental topic in several astrophysical phenomena, such as solar flares, pulsar wind, magnetars, etc, for more than half a century, and one of the unresolved questions is its efficiency. Recently, nonthermal particle acceleration mechanisms during reconnection have been extensively studied by particle-in-cell simulations, yet it is an intriguing enigma as to how the magnetic field energy is divided into thermally heated plasmas and nonthermal particles. Here we study both non-relativistic and relativistic magnetic reconnections using large-scale particle-in-cell simulation for a pair plasma, and indicate that the production of the nonthermal particle becomes efficient with increasing the plasma temperature. In the relativistic hot plasma case, we determine that the heated plasmas by reconnection can be approximated by a kappa distribution function with the kappa index of approximately $3$ or less (equivalent to $2$ or less for the power-law index), and the nonthermal energy density of reconnection is approximately over $95\%$ of the total internal energy in the downstream exhaust. magnetic reconnection — plasmas — particle acceleration ## I Introduction Recently, the rapid magnetic energy release by magnetic reconnection, as well as the supersonic bulk flow energy release by collisionless shock waves, is regarded as an important particle acceleration process, and reconnection has garnered significant attention in various plasma environments such as the solar flares, Earth’s magnetosphere, pulsar magnetospheres, and magnetars [e.g., 1, 2, 3, 4, 5]. In these hot and rarefied magneto-active plasmas, the magnetic energy stored in a plasma sheet with an anti-parallel magnetic field component can be efficiently converted into not only the Alfvénic bulk flow energy, but also the energies of plasma heating and nonthermal particle acceleration by magnetic reconnection. While the reconnection paradigm has been established as a powerful magnetic energy dissipation process in the plasma universe [e.g., 6, 7, 8, 9, 10], it is recently being recognized that the reconnection plays a much more important role than before in the astrophysical phenomena, because it can efficiently generate the nonthermal particles, and the nonthermal energy spectra observed in various astrophysical objects often indicate a hard power-law spectrum, which cannot be easily explained by the conventional acceleration processes such as the diffusive Fermi shock acceleration [e.g. 11, 12]. From the satellite observations of the solar flares and substorms in the Earth’s magnetotail, it is recognized that not only the plasma heating, but also the nonthermal particle acceleration is an ubiquitous process of reconnection, and it is suggested that the reconnection process with a relativistic hot gas temperature such as gamma-ray flares in the Crab nebula can generate nonthermal particles as well. In fact, by employing particle-in- cell (PIC) simulation studies, it is discussed that a hard non-thermal energy spectrum $N(\varepsilon)\propto\varepsilon^{-s}$ with its power-law index $s\leq 2$ can be formed for a relativistic reconnection using 2-dimensional PIC simulations [e.g. 13, 14]. It is revealed in three-dimensional PIC simulations [e.g. 15, 16, 18, 17] that such a hard power-spectrum can be formed for a large magnetization parameter $\sigma=B^{2}/(4\pi\rho c^{2})$, whereas the power-law slope becomes softer as the magnetization parameter $\sigma$ decreases, where $B$ and $\rho$ are the magnetic field and mass density, respectively. While the magnetic reconnection is understood as an efficient particle accelerator in plasma universe, one of the remaining challenges is to quantitatively understand the energy conversion rate into nonthermal particles. The efficiency of nonthermal particle production against the total internal plasma energy during the magnetic energy dissipation is not necessarily understood in a consistent way. Ball et al. [18] discussed the nonthermal particle efficiency in trans-relativistic reconnection using PIC simulations, and determined that the efficiency increases with increasing magnetization parameter $\sigma$. However, it is interesting to know the efficiency of the nonthermal production for a wider range of $\sigma$ values, from non-relativistic cold plasmas to relativistic hot plasma reconnections. It is also important to carefully discuss the spectral feature from the low- energy to high-energy nonthermal range, and to understand the energy partition between the thermal and nonthermal high energies by analyzing the reconnection energy spectrum. The energy partition between the thermal plasma and nonthermal particles is still an important open question. To understand the energy partition between the thermal and nonthermal energies, we investigate magnetic reconnection using particle-in-cell simulations by focusing on the time and spatial evolutions of the energy spectrum. The energy spectrum obtained by the reconnection simulation, in general, comprises the thermal and non-thermal populations, and a model fitting of the energy spectrum by the combination of Maxwellian and kappa distribution functions is made as well. [e.g. 19]. We quantify the thermal plasma and nonthermal populations as the function of the initial plasma temperature, both in non-relativistic and relativistic regimes, and prove that the production of the nonthermal particle becomes efficient with an increase in plasma temperature. In the relativistic hot plasma case, we determine that the nonthermal power-law index $s$ can be close to $2$, and the nonthermal energy density of reconnection reaches over approximately $95\%$ of the total heated plasma in the exhaust. ## II Overview of Simulation Study We study the plasma heating and particle acceleration of magnetic reconnection using two-dimensional PIC simulation [20, 21, 22] with a periodic boundary in the $x$-direction for $x=-215\lambda$ and $x=215\lambda$, and the conducting walls for the upper and lower boundaries at $y=\pm 215\lambda$, where $\lambda$ is the thickness of the initial plasma sheet. The total system size for our fiducial runs is $L_{x}\times L_{y}=430\lambda\times 430\lambda$ and the computational grid size is $5376\times 5376$. The total number of particles is $8.4\times 10^{10}$, to calculate the high-$\beta$ plasma sheet as accurately as possible. For simplicity, we adopt the Harris solution [23] for the pair plasma with the same mass $m$ and temperature $T$ uniform in space, and focus on the energy partition of an idealized magnetic reconnection in a collisionless plasma system. The magnetic field ${\bf B}=B_{x}(y){\bf e_{x}}$ and plasma density $N(y)$ are given by $B_{x}(y)=B_{0}\tanh(y/\lambda),$ (1) and $N(y)=N_{0}\cosh^{-2}(y/\lambda)+N_{\rm b},$ (2) respectively, where $N_{\rm b}$ is the background uniform plasma density adopted to demonstrate continuous plasma injection from the outside plasma sheet to the reconnection exhaust. The background plasma density $N_{\rm b}$ is set to be $5\%$ of the maximum Harris plasma density $N_{0}$. Note that the relationships of the pressure balance $B_{0}^{2}/8\pi=2N_{0}T$ and force balance $2T/\lambda=|e|u_{d}B_{0}/c$ are satisfied, where $u_{d}$ is the drift velocity , and the initial electric field ${\bf E}$ is zero. We study eight different cases of plasma temperature from a cold non- relativistic to hot relativistic temperature of $T/mc^{2}=10^{n/2-2}\quad{\rm with}\quad n=0,1,2,..,7.$ (3) The above temperatures correspond to the magnetization parameter $\sigma=B_{0}^{2}/(8\pi N_{0}mc^{2})=2\times 10^{n/2-2}$, if we use the number density for the value of the plasma sheet $N_{0}$, owing to the pressure balance between the gas pressure inside the plasma sheet and magnetic pressure outside the plasma sheet. Note that the Alfvén speed $v_{A}$ is given by $v_{A}=c\sqrt{\sigma/(1+\sigma)}$ in a cold plasma limit, where the density is used for the value of the central plasma sheet and the magnetic field is the value of the outside plasma sheets. We keep the ratio of the gyro-radius $r_{g}$ and thickness of the plasma sheet $\lambda$ for all simulation runs, and set $r_{g}/\lambda=0.45$, where $r_{g}=mc^{2}\sqrt{\gamma_{th}^{2}-1}/(eB_{0})$, because a thin plasma sheet whose thickness is almost equal to the gyro-radius has been formed before the explosive growth of magnetic reconnection in the pair plasma [22] and has been observed before the onset of substorms in the Earth’s magnetotail [e.g. 24, 25]. For simplicity, we have assumed $(\gamma_{th}-1)mc^{2}/T=1$, and then the drift speed $u_{d}$ can be calculated by $u_{d}/c=2(r_{g}/\lambda)/\sqrt{1+2mc^{2}/T}=0.9/\sqrt{1+2\times 10^{2-n/2}}$ . We add a small initial perturbation for the vector potential $\delta A_{z}$ to the Harris equilibrium, which is given by $\delta A_{z}(x,y)\propto\exp(-(\frac{x}{\lambda_{x}})^{2}-(\frac{y}{\lambda_{y}})^{2})\cosh(\frac{y}{\lambda_{y}}),$ where $\lambda_{x}=2\lambda$, $\lambda_{y}=\lambda/2$, and the initial amplitude of the reconnected magnetic field is set to be $\max(B_{y}/B_{0})=10^{-2}$ in the neutral sheet. Figure 1: Time evolution of the collisionless magnetic reconnection for a Harris current sheet obtained by PIC simulation for $T/mc^{2}=10^{-2}$. The top panel ($t/\tau_{A}=119$) shows the early stage of reconnection, and the X-type neutral point is formed at the center of the simulation box. The middle panel ($t/\tau_{A}=176$) is the nonlinear stage where several plasmoids are coalescing and forming the larger plasmoids, and the bottom panel ($t/\tau_{A}=308$) is the almost final stage when two large plasmoids are just coalescing into one large plasmoid in the system. The thick white lines indicate the magnetic field lines that pass through the X-type neutral line, i.e., the separatrix between the downstream exhaust and the upstream plasma sheet. The color bars in the right-hand side indicate that the logarithmic plasma density normalized the initial plasma sheet density. Figure 2: Energy spectra for the downstream exhaust region for $T/mc^{2}=10^{-2}$ (left), $T/mc^{2}=1/\sqrt{10}$ (middle) and $T/mc^{2}=10$ (right), respectively. The color lines indicate the time evolution of spectra, whose time stages are indicated in the right-hand side bar. The dark blue line is the initial state, and the red one is the final stage when a large plasmoid is formed in the simulation box. The thick solid lines are the model fitting for the composed function of the Maxwellian and kappa distributions $N_{{\rm M}+\kappa}(\gamma)$. The black dashed lines in the lower energy regime represent the Maxwellian part of the fitting result $N_{\rm M}(\gamma)$, while the other dashed lines in the higher energy regime are the kappa distribution parts $N_{\kappa}(\gamma)$. The bottom panels indicate the errors of the model fitting, which show the difference between the fitting model function and simulation result. The dashed lines in the bottom plot are depicted as reference of an acceptable level of error. Illustrated in Figure 1 are three-time stages of the reconnection structure obtained by the particle-in-cell simulation for a non-relativistic reconnection with the temperature of $T/mc^{2}=0.01$. The plasma density and magnetic field lines are indicated by the color contour and white lines, respectively. The thick white lines indicate the separatrix of the magnetic reconnection for the X-type neutral point formed around the center $x/\lambda\sim 0$, i.e., the most recently reconnected magnetic field lines. The downstream of the separatrix is occupied by the reconnection heated plasmas. These lines of the separatrix can be obtained by the contour lines of the vector potential $A_{z}$ that has the minimum value in the neutral line of $y=0$. A part of the simulation system is illustrated as well. The top panel is the early time stage at $t/\tau_{A}=119$ when the reconnection just started at the center, where $\tau_{A}=\lambda/v_{A}$ is the Alfvén transit time, and several small scale plasmoids have been formed along $y=0$. The middle panel indicates the time stage when several plasmoids grow rapidly and coalesce with each other. As we imposed the initial small perturbation $\delta A_{z}$ at $x/\lambda=0$, the most prominent reconnection occurs at the center, but in addition to that, we can observe several other small size plasmoids embedded in the downstream domain of the main reconnection. The bottom panel is the approximate final stage of the active reconnection in our periodic system after the two large plasmoids have merged into one large magnetic island. ## III Model Fitting of Energy Spectrum As our objective is to understand the energy partition between the thermal and nonthermal populations during reconnection, we study the energy spectrum in the downstream of the separatrix lines in details. Figure 2 illustrates the time evolutions of three energy spectra for $T/mc^{2}=10^{-2}$ (left), $1/\sqrt{10}$ (middle), and $10$ (right), integrated over the reconnection region sandwiched by the separatrix, i.e., the thick white lines defined in Figure 1. Note that if the width of the separatrix is smaller than the thickness of the plasma sheet $\lambda$, the integration is conducted for the plasma sheet with the size $\lambda$. The horizontal axis is the particle energy $(\gamma-1)mc^{2}$, and the vertical axis is the number density $N(\gamma)d\gamma$, where $\gamma=1/\sqrt{1-(v/c)^{2}}$. The color lines indicate the time evolution of the energy spectra, whose time stages are indicated in the right-hand color bars. The blueish and reddish colors correspond to the earlier and later time stages, respectively. As we assumed a uniform plasma temperature for both the $\cosh$-type Harris density population and background populations in Equation (2), the dark blue color spectra in Figure 2 illustrate approximately the initial plasma state with the initial Maxwellian distribution functions. For the relativistic case in Figure 2c, however, one-dimensional energy spectrum is modified by the effect of a high- speed drift velocity with $u_{d}/c\sim 0.82$. As time goes on, we can clearly observe those Maxwellian plasmas are gradually heated owing to the reconnection heating process, and the high-energy components are further accelerated and form the nonthermal population. During the time evolution, by looking at the energy interval of the time evolution of the energy spectra, it is determined that the energy intervals are wider in the early acceleration stage compared with those in the later stage, suggesting that the rapid energy gain happens in the early stage. As the spectra with the red curve is approximately stable and at the final stage, we discuss the spectra behavior for the red curve by making a model function fitting in detail. It is evident that the energy spectra cannot be simply approximated by a Maxwell distribution function given by, $N_{M}(\gamma)\propto\gamma\sqrt{\gamma^{2}-1}\exp(-\frac{\gamma-1}{T_{M}/mc^{2}}),$ (4) where, for simplicity, a three-dimensional Maxwell distribution function has been projected into a one-dimensional distribution in the simulation frame as the function of the particle energy $\gamma$. To make a better model fitting of the observed distribution, we first tested a kappa distribution function given by, $N_{\kappa}(\gamma)\propto\gamma\sqrt{\gamma^{2}-1}\left(1+\frac{\gamma-1}{\kappa T_{\kappa}/mc^{2}}\right)^{-(1+\kappa)}f_{cut}(\gamma).$ (5) The kappa distribution comprises a thermal Maxwellian at low energies and a nonthermal population approximated by a power-law function at high energies [e.g. 19]. It is known that the observed solar wind distribution function can be well modeled by a kappa ($\kappa$) distribution function [e.g 19, 26], and the kappa modeling is widely applied for the space plasma investigation [e.g 27, 28]. In our analysis, we added a high energy cutoff function $f_{cut}$ to represent a possible high energy cutoff in the simulation data [29, 30], which is given by, $f_{cut}(\gamma)=\left\\{\begin{array}[]{@{\,}ll}1&\mbox{for $\gamma\leq\gamma_{cut}$ }\\\ \exp\left(-(\gamma-\gamma_{cut})/\gamma_{cut}\right)&\mbox{for $\gamma>\gamma_{cut}$ }\end{array}\right.$ The high energy cutoff may come from the finite time evolution of reconnection under the limited system size [30]. Note that the high energies of $N_{\kappa}$ can be approximated by a power-law function of $N_{\kappa}\propto\gamma^{-\kappa+1}$ with the power-law index $s=\kappa-1$ for $\gamma\gg T_{\kappa}/mc^{2}$. Although the model function was not necessarily a good approximation, we determined that this kappa model function is better than the Maxwellian fitting. As the spectra for the red lines in Figure 2 can be observed, the contributions of the cold plasma freshly transported from the outside plasma sheet is not necessarily negligible. Note that the energetic particles generated in the X-type region/magnetic diffusion region can escape along the magnetic field line, and simultaneously, the pre-heated plasma is transported from the outside plasma sheet by $E\times B$ drift motion in association with the magnetic field lines. Therefore, the separatrix region contains both the cold and hot plasmas [e.g. 31]. Having the above plasma transport, we finally determined that the best-fitted model function $N_{M+\kappa}$ is a combination of the Maxwellian and kappa distribution functions, i.e., $N_{M+\kappa}(\gamma)=N_{M}(\gamma)+N_{\kappa}(\gamma).$ (6) In the top three panels in Figure 2, the sold lines are the model fitting curves for $N_{M+\kappa}$ at the final time stage, while the two dashed lines are the fitting curve of $N_{M}$ at lower energies and $N_{\kappa}$ for higher energies, respectively. It is determined that three simulation spectra in the downstream exhaust for $T/mc^{2}=10^{-2}$ (left), $1/\sqrt{10}$ (middle), and $10$ (right) can be approximated by the composed model function of Maxwellian and kappa distributions $N_{{\rm M}+\kappa}(\gamma)$. We have also confirmed that five other simulation cases of $T/mc^{2}=10^{n/2-2}$ with $n=1,2,4,5$ and $7$ express the similar good model fitting. The bottom three panels indicate the error of the model fitting given by, $\displaystyle{\rm Error}(\gamma)=|N_{data}(\gamma)-N_{M+\kappa}(\gamma)|/$ $\displaystyle{\rm min}(N_{data}(\gamma),N_{M+\kappa}(\gamma)),$ (7) where ${\rm min()}$ means a function to return the minimum element from $N_{data}$ and $N_{M+\kappa}$. If $\frac{1}{2}N_{data}<N_{M+\kappa}<2N_{data}$, then Error becomes less than $1$, as depicted by the dashed line in the bottom panel of Figure 2. We can observe that the errors in the wide energy ranges are less than $10^{-1}$, and the model fitting of $N_{M+\kappa}$ can well describe the energy spectra obtained in the simulation runs. Note that a small bump seen around $\gamma\sim 8$ for $T/mc^{2}=10^{-2}$ is the contribution from the small-scale plasmoids behind the main large-scale plasmoid, but the total energy contribution is not necessarily large. Note that the bumpy structure in the energy spectrum is not a permanent structure and disappears after the small scale plasmoid is absorbed into the large plasmoid. For the model fitting for $T/mc^{2}=10$ for the right-hand panel in Figure 2, we can observe a large discrepancy between the simulation data and model fitting at high energies. By choosing another cutoff function given by $f_{cut}(\gamma)=\exp(-((\gamma-\gamma_{cut})/\gamma_{cut})^{2})$, we obtain the better fitting result (not indicated here). Figure 3: The model fitting results for eight different initial plasma temperatures of $T/mc^{2}=10^{-2}\sim 10^{3/2}$. (Left) The temperatures of the Maxwellian part $T_{M}$ (red dots) and kappa distribution part $T_{\kappa}$ (blue dots) as the function of the initial plasma temperature. The kappa index $\kappa$ (middle) and cutoff energy $\gamma_{\rm cut}$ (right) as the function of the initial plasma temperature. We quantitatively discuss the energy partition between the thermal and nonthermal plasmas. Figure 3 illustrates the model fitting results of (a) the temperatures $T_{M}$ (red dots) and $T_{\kappa}$ (blue dots), (b) the $\kappa$ index with a power-law tail, and (c) the cutoff energy $\gamma_{cut}$ as the function of the initial plasma temperature $T/mc^{2}$, respectively. We determine that the temperature of $T_{M}$ remains approximately the same as the initial background temperature $T$, suggesting that the exhaust region sandwiched by the separatrix, as indicated by the thick white lines in Figure 1, still contains a part of non-heated plasmas, and that those temperatures may represent the flux tubes just transported into the reconnection exhaust, and are not heated strongly [31]. Note that during the transport from the outside plasma sheet into the plasma sheet, the plasmas are slightly cooled down owing to the adiabatic expansion of the magnetic flux tube, because the slow mode expansion waves are emitted from the X-type reconnection region. However, the temperatures of the kappa distribution function $T_{\kappa}$ are higher than that of the Maxwellian function $T_{M}$. The ratio of $T_{\kappa}/T_{M}$ is large for the non-relativistic reconnection cases for $T/mc^{2}<1$, while those for the relativistic reconnection regime of $T/mc^{2}>1$ become less than 2. $\kappa$ index, which is equivalent to the power-law index $s=\kappa-1$ is illustrated in Figure 3b. For the non- relativistic case of $T/mc^{2}=10^{-2}$, $\kappa$ indicates a large value, suggesting that the spectrum is close to a Maxwell distribution and the efficiency of the nonthermal particle production is very low, whereas for the relativistic regime of $T/mc^{2}>1$, the $\kappa$ values are approximately $3$ and the nonthermal tails are well developed. Note that $\kappa=3$ can be rephrased as the power-law index of $2$. Figure 3c is the cutoff energy $\gamma_{cut}$. Except for the case of $T/mc^{2}=10^{-2}$, we determine that the ratio of $\gamma_{cut}/(T_{\kappa}/mc^{2})$ is approximately $50\sim 10^{2}$, suggesting that the nonthermal tails are well extended above the thermal population of the kappa distribution function. It would be interesting to mention the relationship between the gyro-radius of the accelerated particle and the size of the magnetic island. The initial gyro-radius was set to be 0.45 $\lambda$ for all runs, and we can find in Figure 2 that the highest energies after reconnection attained to several $10^{2}$ \- $10^{3}$ times larger than the initial thermal energies. Therefore, we can estimate that the gyro-radius of the highest energy particles is about $100-200\lambda$, which is almost same as the size of the magnetic island. It would be better to make a short comment on the model fitting error to estimate the fitting parameters. As can be observed in Figure 2, the spectral profile in the final stages of the simulation have approximately the same spectral structure in the logarithmic scale, but the fitting parameters may have approximately 10% variations depending on the time stages. ## IV Efficiency of Nonthermal Density and Energy Production Figure 4: The model fitting results for the same data set as Figure 3. (Left) The percentage in the number density of the kappa distribution part $\xi_{\rm den}$ (red) and percentage in the energy density $\xi_{\rm ene}$ (blue). (Right) The percentage in the number density of the nonthermal population of the kappa distribution against the total distribution $\varepsilon_{\rm den}$ (red) and percentage in the energy density of the nonthermal population $\varepsilon_{\rm ene}$ (blue). Based on the above model fitting, we can study the efficiency of the nonthermal particle acceleration in reconnection. In the previous section, we discovered that the energy spectrum in the exhaust region can be approximated by the combination of Maxwellian and kappa distribution functions. To quantify the efficiency of the production of the nonthermal particles, we calculate the percentage of the nonthermal population from the fitted model function. As illustrated in Figure 4, the left-hand panel a is the percentage of the number density and the energy density for the population of the $\kappa$ distribution against the entire population of the energy spectrum as the function of the initial plasma temperature $T/mc^{2}$. The solid red and dashed blue lines are the number and energy densities defined by, $\displaystyle\xi_{\rm den}=\int_{1}^{\infty}N_{\kappa}(\gamma)d\gamma/\int_{1}^{\infty}N_{{\rm M}+\kappa}(\gamma)d\gamma,$ and $\displaystyle\xi_{\rm ene}=\int_{1}^{\infty}(\gamma-1)N_{\kappa}(\gamma)d\gamma/\int_{1}^{\infty}(\gamma-1)N_{{\rm M}+\kappa}(\gamma)d\gamma,$ respectively. We discussed that the energy spectrum can be approximated by the combination of Maxwellian and kappa distribution functions in Figure 2, and the above denominator indicates the total contribution to the energy spectrum coming from both the Maxwellian and $\kappa$ portions. The percentages of the number and energy densities of the population of the kappa distribution function weakly depend on the plasma temperature $T/mc^{2}$, and we determine that the kappa distributions occupy more than $95\%$ and $70\%$ the total energy and total number densities, respectively. The exhaust region contains the freshly injected plasma from the outside plasma sheet, but the contribution of the cold Maxwellian population is less than $5\%$ in the total energy density. Figure 4b illustrates the ratios of the nonthermal to total number densities (solid red line), and the nonthermal energy to total energy densities (dashed blue line) as the function of the initial plasma temperature $T/mc^{2}$, respectively. We estimate the percentage of the nonthermal number density by the following definition of $\displaystyle\varepsilon_{\rm den}=\int_{1}^{\infty}(N_{\kappa}(\gamma)-N_{\kappa}^{\rm M}(\gamma))d\gamma/\int_{1}^{\infty}N_{{\rm M}+\kappa}(\gamma)d\gamma,$ where $N_{\kappa}^{\rm M}(\gamma)$ represents the portion of Maxwellian distribution function out of the $\kappa$ distribution function, i.e., the $\kappa$ value is replaced by $\kappa=\infty$ by keeping other fitted parameters. Therefore, the numerator of $N_{\kappa}(\gamma)-N_{\kappa}^{\rm M}$ corresponds to only the nonthermal power-law component. Note that a kappa distribution function approaches a Maxwellian distribution as $\kappa\to\infty$, as follows, $\displaystyle\lim_{\kappa\to\infty}\left(1+\frac{\gamma-1}{\kappa T_{\kappa}}\right)^{-(\kappa+1)}\simeq\exp\left(-\frac{\gamma-1}{T_{\kappa}}\right).$ (8) Similarly, the nonthermal energy density can be calculated by, $\displaystyle\varepsilon_{\rm ene}=\int_{1}^{\infty}(\gamma-1)(N_{\kappa}(\gamma)-N_{\kappa}^{\rm M}(\gamma))d\gamma/$ $\displaystyle\int_{1}^{\infty}(\gamma-1)N_{{\rm M}+\kappa}(\gamma)d\gamma.$ From Figure 4b, it is determined that the nonthermal number and energy densities increase with an increase in the plasma temperature in the non- relativistic regime of $T/mc^{2}<1$, while the number and energy densities are approximately constant in the relativistic regime of $T/mc^{2}>1$. We determine that the relativistic reconnection can efficiently generate a lot of nonthermal particles compared with the non-relativistic reconnection. The percentage of the nonthermal number density saturates at $60\%$, while that of the energy density is approximately $95\%$ of the total internal energy in the downstream exhaust. This result suggests that the energy density of the nonthermal population dominates for the relativistic reconnection with a relativistic temperature, and the thermal energy density population is negligible. The similar nonthermal efficiency has been discussed for trans- relativistic reconnection by Ball et al. [18]. ## V Evolution of Nonthermal Plasma in Association with Magnetic Flux Tube Transport So far, we discussed the nonthermal energy spectra integrated all particles in the exhaust plasma sheet, which is sandwiched by the last reconnected magnetic field line; however, it would be interesting to discuss where and how those nonthermal particles are accelerated in the plasma sheet. It is known that the exhaust plasma sheet has several distinct structures responsible for particle acceleration: (i) the magnetic diffusion region with weak magnetic fields, where the fresh particles transported from the outside plasma sheet are picked up and accelerated by the reconnection electric field, (ii) the plasma sheet boundary layer between the outside and exhaust plasma sheets where the plasma gas pressure suddenly increases, (iii) the plasmoids where the high- temperature and high-density plasmas are confined by the closed magnetic flux rope, and (iv) the magnetic field pile-up region where the Alfvénic jet from the diffusion region collides with the pre-existing plasma sheet/plasmoid etc. In association with those distinct regions, their roles on plasma heating and acceleration processes have been discussed by paying attention to the individual particle motion [e.g 34, 35, 36, 31, 37, 38, 39]. Figure 5: The colour contour of the plasma density structure of magnetic reconnection for $T/mc^{2}=10^{-2}$. The logarithmic density contour is indicated in the right-hand side bar. The white lines indicate the magnetic field lines, i.e., the contour lines of the vector potential $A_{z}(x,y)$. The interval of the contour line is chosen in such a way that the downstream exhaust region is divided by ten magnetic field lines. The thick white lines are depicted with the flux tube number for every five lines, which corresponds to the spatial evolution of the energy spectrum indicated by the right-hand color box in Figure 6. Figure 6: Energy spectra for $T/mc^{2}=10^{-2}$ (left), $T/mc^{2}=1/\sqrt{10}$ (middle) and $T/mc^{2}=10$ (right), respectively. The same data set as Figure 2. The color lines correspond to the energy spectra for the 19 divided regions illustrated in Figure 5. The dark blue color spectrum is taken for the outermost region, while the red one is the innermost region. The right-hand side color bar indicates the region number. The solid black lines are the energy spectra for the downstream exhaust regions, which are the same spectra as illustrated in Figure 2. The black dashed lines are the downstream exhaust spectra in which the contributions in the inner most flux tube are subtracted. The relative importance of the efficiency of nonthermal particle production in those different regions, however, has not necessarily been discussed in a systematic manner. To understand the relative importance, we focus on the spatial evolution of the nonthermal energy spectrum in association with the magnetic flux transportation. We divide the exhaust plasma sheet with 10 magnetic flux tubes by the equal interval of the vector potential $\Delta A_{z}=(A_{z}^{\rm max}-A_{z}^{\rm min})/10$, where $A_{z}^{\rm max}$ and $A_{z}^{\rm min}$ correspond to the vector potential for the last reconnected magnetic field line at the X-type neutral point, and that for the minimum bottom of the O-type point in the plasmoid, respectively. The reconnection structure divided by the 19 magnetic flux tubes for $T/mc^{2}=10^{-2}$ is illustrated in Figure 5, and we mark labels on the magnetic field lines from 1 – 19 from inside to outside. The last reconnected magnetic field line corresponds to the label number of 10. The underlying assumption of this analysis is that the thermal and energetic particles are transported together with the magnetic flux, because we have confirmed that the frozen-in condition is approximately satisfied [32]. Figure 6 illustrates the energy spectra for those 19 divided regions labeled by the rainbow colors indicated by the right-hand side color bar. To focus on the nonthermal tail, the energy range in the horizontal axis is chosen to be three orders of magnitudes. The red line is the energy spectrum for the flux tube number 1, which corresponds to the innermost region almost occupied by the yellow color in Figure 5. The green line of the label number 10 corresponds to the flux tube just attached to the X-type neutral point, and the dark blue is that of outside the plasma sheet. Evidently, the blueish lines are basically the thermal Maxwellian spectrum with the initial plasma temperature almost maintained, thus suggesting that no plasma heating and acceleration occurs during the plasma transport from the outside plasma sheet toward the separatrix region. (Note that this statement is not necessarily correct if the evolution of the velocity distribution function is discussed in detail, because the adiabatic plasma process can be observed in association with the change of flux tube, which is initiated by the slow mode expansion waves generated around the magnetic diffusion region [32]). However, when the magnetic flux tube is attached to the separatrix, the rapid increase in the nonthermal particle population can be observed, regardless of the initial plasma temperature. The flux numbers of 10 and 11 with the greenish colors corresponds to this region, and the flux number 11 has been highlighted by the thick green color. This behavior suggests that the initial significant particle acceleration can happen during the reconnection of two different magnetic flux tube. The flux tube number 11 is just situated outside the separatrix, but the magnetic diffusion region at the X-type neutral point has a finite width in $y$-direction, and the flux tube is contaminated by those accelerated particles around the diffusion region. The main energy gain in the flux tube numbers 10 and 11 comes from the meandering particles in the diffusion region [35, 37] where the meandering particles are accelerated by the inductive electric field $E_{z}$. Note that the resonant acceleration in the diffusion region can be enhanced in a relativistic plasma due to the relativistic inertia effect [13]. In addition to the acceleration in the diffusion region, the beam plasma heating can occur in the plasma sheet boundary, because a part of the accelerated particles around the diffusion region may have a small pitch-angle against the local magnetic field, and those particles can escape outward along the magnetic field line. The interaction of the field-aligned beam component and cold plasma transported by $E\times B$ motion from the outside plasma sheet may lead to the plasma heating together with the generation of Alfvénic waves [33]. After rapid energization in the magnetic diffusion region and along the separatrix magnetic field lines, those flux tubes are convected downward, and further gradual energization can be observed, as indicated by the yellowish and reddish lines. However, we determine that those energy gains are not necessarily strong, compared with those in the last reconnecting magnetic flux tube. Several bumpy structures in the spectra may arise from the dynamical evolution of plasmoids. The thick black solid curves are the energy spectra for the downstream exhaust region integrated from the flux tube number 1 – 10, which is the same as the final stage illustrated in Figure 2. By looking at the spatial evolution of the energy spectra from No.10 (flux tube of the separatrix) to No.2 (inner flux tube), we more or less observe the gradual heating and particle acceleration. However, we determine that the spectra for the innermost flux tube No. 1 has a slightly different behavior, because of the pre-existed and pre-heated plasma before the onset of reconnection. To distinguish the plasma heating between the pre-existed plasma and freshly injected particles, we indicate that the energy spectra integrated the region from the flux tube of No.2 – No.10 as the black dashed line, i.e., we excluded the innermost flux tube. As can be observed, the difference appears only around the middle energy range between the thermal Maxwellian part and nonthermal power-law part, and the kappa index becomes slightly harder as well. Regarding the case of the relativistic reconnection, we obtain the kappa index of $\kappa<2$, and the result is roughly consistent with the previous studies [13, 14, 15, 16]. The efficiency of nonthermal density discussed in the previous section, however, does not change so much. Figure 7: Time evolution of energy spectra for fixed magnetic flux tubes for $T/mc^{2}=10^{-2}$ (left), $T/mc^{2}=1/\sqrt{10}$ (middle) and $T/mc^{2}=10$ (right), respectively. The same data set as Figure 2. The color lines indicate the time evolution of spectra, whose time stages are indicated in the right- hand side bar. The dark blue line is the initial state when the magnetic flux tube is situated in the upstream. On the other hand, the red one is the final stage when the flux tube is transported into the downstream exhaust. The thick solid lines are the model fitting for the composed function of the Maxwellian and kappa distributions $N_{{\rm M}+\kappa}(\gamma)$. The black dashed lines in the lower energy regime represent the Maxwellian part of the fitting result $N_{\rm M}(\gamma)$, while the other dashed lines in the higher energy regime are the kappa distribution parts $N_{\kappa}(\gamma)$. The bottom panels indicate the errors of the model fitting in the same format as Figure 2. Although we discussed the spatial evolution of the energy spectra in Figure 6, it is also interesting to mention the time evolution of the energy spectrum for a fixed magnetic flux tube, which is transported from the outside plasma sheet into the downstream exhaust region during reconnection. Figure 7 illustrates the time evolution of the energy spectra for the fix magnetic flux tubes sandwiched between $\left|y\right|/\lambda=9$ and $12$ at the initial state $t=0$. As could be easily expected, the behavior of the energization is basically same as Figure 6. The dark blue color spectra are taken when the flux tubes are situated in the upstream, and those temperatures almost remain the initial Maxwellian. As time goes on, we observed the rapid energization and formation of the nonthermal tail when two flux tubes just reconnect at the X-type neutral point at $t/\tau_{A}\sim 120$. The timing of the attachment of two flux tubes normalized by the Alfvén transit time $\tau_{A}$ was almost same for three cases. After the attachment of the two magnetic flux tubes, those reconnected magnetic flux tubes are further transported into the downstream, and the high energy particles are gradually produced. The red lines are the almost final stages of the magnetic reconnection. The thick solid lines are the model fitting for the composed function of the Maxwellian and kappa distributions $N_{{\rm M}+\kappa}(\gamma)$. The black dashed lines in the lower energy regime represent the Maxwellian part of the fitting result $N_{\rm M}(\gamma)$, while the other dashed lines in the higher energy regime are the kappa distribution parts $N_{\kappa}(\gamma)$. The fitted parameters of ($T_{M}/mc^{2}$, $T_{\kappa}/mc^{2}$, $\kappa$, $\gamma_{cut}$) are ($0.013$, $0.078$, $6.1$, $6.2\times 10^{2}$), ($0.79$, $0.44$, $3.0$, $2.2\times 10^{1}$), and ($15.$, $9.0$, $2.6$, $4.7\times 10^{2}$) for $T/mc^{2}=10^{-2}$, $1/\sqrt{10}$, and $10$, respectively. These fitted parameters are basically same as those obtained in Figures 2 and 3. The contributions of the Maxwellian part, however, are smaller than the kappa distribution part compared with Figure 2, because the magnetic flux tubes are convected into the deep downstream and the upstream cold plasma are heated. The bottom panels indicate the errors of the model fitting in the same format as Figure 2. We find that the model fittings are very good for a wide energy ranges. ## VI Discussions and Summary During the last two decades, it has been discussed that magnetic reconnection can produce not only the hot thermal plasma, but also the high-energy supra- thermal particles. Specifically, relativistic reconnection can efficiently produce nonthermal particles, whose energy spectrum is approximated by a hard power-law function. However, how the efficiency changes between the non- relativistic and relativistic reconnection has remained an open question. Meanwhile, the detailed acceleration process focusing on the individual particle motion has been well discussed by several people [e.g. 31, 35, 36, 37, 38, 39]. The interaction between the reconnection electric field $E_{z}$ and meandering motion under the anti-parallel magnetic field $B_{x}$ in the neutral sheet is known to play an important role as the primary acceleration process, and the accelerated particles in and around the X-type neutral line can be ejected outward by the Lorentz force because of the reconnecting magnetic field [34, 37]. Zenitani and Hoshino [13] investigated this process to the relativistic reconnection, and argued that the interaction efficiency between the reconnection electric field and meandering particle can be enhanced in the relativistic reconnection, because of the relativistic inertia effect during the Speiser motion [34]. In addition to the particle acceleration in and around the magnetic diffusion region, several other possible mechanisms have been proposed. The collision between the Alfvénic jet ejected from the magnetic diffusion region and magnetic field pile-up region ahead of the plasmoid can provide the betatron- like acceleration owing to the gradient-B drift motion [e.g 35]. Drake et al. [36] proposed that the particles trapped by the plasmoid can be energized by a Fermi process, which is the curvature drift motion in association of the time evolution of the plasmoid. Drake et al. [36], Dahlin et al. [38] investigated the above mechanisms responsible for energy gain using the guiding center formalism. Oka et al. [39] argued that the particles can be accelerated during coalescence of two plasmoids with the similar process that occurs in the magnetic diffusion region, but with a stronger inductive electric field in association with the dynamic coalescence. We did not analyze such detail acceleration mechanisms in this study, rather focused on the evolution of energy spectrum during the plasma transport in association with the magnetic flux tube. We not only observed the primary particle acceleration around the magnetic diffusion region, but also the additional acceleration during the plasma transport in the exhaust. These heating and acceleration are basically consistent with the previous studies. In addition, we determined that the energy spectra obtained by PIC simulations can be well approximated by a kappa distribution function for the high energy population. By examining the model fitting results, we determined that the efficiency of the nonthermal particle acceleration increases with increasing plasma temperature, and the nonthermal energy density reaches over $95\%$ of the total internal energy in the relativistic temperature of the exhaust, whereas the nonthermal number density remains approximately $40\%\sim 60\%$ for the total number density in the exhaust. Our simulation study in a two-dimensional system with the anti-parallel magnetic field topology is only a first step toward understanding the energy partition between thermal and nonthermal plasmas. We investigated the efficiency of the nonthermal particle production for a pair plasma, but it is important to understand the acceleration efficiency for an ion-electron plasma. Recently, several people discussed the energy partition between ion and electron, and it is revealed that ions are preferentially heated, compared to electrons during magnetic reconnection by PIC simulations [40, 41, 42]. It has been discussed that the ratio of the ion heating to the electron $T_{i}/T_{e}$ can be approximated by $(m_{i}/m_{e})^{1/4}$ for the magnetic reconnection without guide field, where $m_{i}$ and $m_{e}$ are the ion and electron mass, respectively [32], whereas the electrons receive larger energies than the ions for guide field reconnection [43]. Although ions seem to receive much thermal energies compared with electrons, suprathermal electrons seem to be efficiently generated, while the proton nonthermal spectrum seems not to be necessarily well developed in particle simulations. Significant attention has been paid to electron nonthermal particle acceleration, but the study of ion nonthermal production is limited. We are probably required to study a longer time evolution of reconnection for a large system size. It is important to check the dependence of the energy spectrum on the system size of our PIC simulation. So far, we fixed the system size with (a) $430\lambda\times 430\lambda$, but we studied other cases with (b) $215\lambda\times 215\lambda$, (c) $860\lambda\times 215\lambda$, and (d) $1720\lambda\times 430\lambda$ for $T/mc^{2}=0.1$. The fitting results for $T_{M}$, $T_{\kappa}$, $\kappa$, $\varepsilon_{\rm den}$ and $\varepsilon_{\rm ene}$ does not strongly depend on the system size, and we found that those differences of $T_{M}$, $T_{\kappa}$, $\kappa$ on the system size were less than $15\%$. However, the cutoff energy seems to weakly depend on the system size, except for case (b). Regarding case (b), $\gamma_{\rm cut}$ was approximately 6.7, but other cases of (a),(c) and (d) were $\gamma_{\rm cut}\sim 9-12$. It would be important to mention the effect of the plasma temperature in the outside plasma sheet, which is denoted by $N_{b}$ in Equation (2). So far, we assumed that the background temperature is the same as that for the Harris plasma component $N_{0}$, but we can choose any temperature to keep an equilibrium state. In astrophysical settings, it is likely that the plasma temperature outside the plasma sheet is colder than the Harris plasma temperature. In fact, for the case of the Earth’s magnetotail, the temperature in the lobe is known to be colder than that in the plasma sheet [e.g. 24, 25]. We have studied the cold background plasma temperature $T_{b}/mc^{2}=10^{-4}$ by keeping other plasma parameters the same. We determine that the efficiency of the nonthermal energy density is almost the same as the case of the uniform background plasma temperature discussed in this study. We will discuss this behavior in a separate paper. Our final comment is on the effect of the background density $N_{b}$. It is well known that the reconnection rate increases with an increase in the Alfvén speed in the inflow region. Furthermore, the reconnection electric field $E_{z}$ becomes larger with a decrease in the background density, and the magnetic energy dissipation rate being converted into the thermal and nonthermal plasmas would be enhanced as well. Although the nonthermal particle production would be expected to increase with a decrease in the background plasma density, the energy partition between the thermal and nonthermal plasma production should be carefully studied. In our preliminary study for the low background density with $N_{b}/N_{0}=1/80$, we obtained the nonthermal $\kappa$ values of $2.2$, $2.6$, $2.7$, and $4.3$ for $T/mc^{2}=10$, $1$, $10^{-1}$ and $10^{-2}$, respectively. We find that these $\kappa$ values are about $30\%$ smaller than those of $N_{b}/N_{0}=1/20$ shown in Figure 3b. It would be also interesting to study the effects of background temperature and density for the generalized magnetization parameter $\sigma=B^{2}/(4\pi w)$ and Alfvén speed $v_{A}=c/\sqrt{1+4\pi(e+p)/B^{2}}$, where $w$, $e$ and $p$ are the enthalpy density, the total energy density and the gas pressure, respectively. We will discuss these effects in a separate paper together with the effect of the guide field. ###### Acknowledgements. This study was supported by JSPS KAKENHI, Grant Number 20K20908. ## DATA AVAILABILITY The data support the findings of this study are available from the corresponding author upon reasonable request. ## Appendix A Nonthermal contribution in $\kappa$ distribution function We shortly discuss the basic behavior of $\kappa$ distribution function by focusing on the thermal and non-thermal populations. As the $\kappa$ distribution function are given by two parameters of the temperature $T$ and $\kappa$ index, the nonthermal contribution can be given by two parameters. By assuming that the thermal population of the $\kappa$ distribution function is given by keeping the same temperature but by setting $\kappa=\infty$, we can easily estimate the percentages of the nonthermal number and nonthermal energy densities. The top and bottom panels in Figure 8 illustrate the nonthermal number density $\varepsilon_{\rm den}$ and the nonthermal energy density $\varepsilon_{\rm ene}$, respectively. The horizontal axes are the $\kappa$ index and plasma temperature normalized by the rest mass energy $T/mc^{2}$. We can observe that the nonthermal contributions become small for a large $\kappa$ value. For the case of a small $\kappa$ and relativistic hot plasma, we determine that the nonthermal contributions become significant. Regarding $\kappa=3$, the nonthermal number density contribution $\varepsilon_{\rm den}$ becomes $\varepsilon_{\rm den}\rightarrow 7/9$ for $T/mc^{2}\rightarrow\infty$, whereas the nonthermal energy density contribution is $\varepsilon_{\rm ene}\rightarrow 1$ for $T/mc^{2}\rightarrow\infty$. For $\kappa=4$ and $T/mc^{2}\rightarrow\infty$, $\varepsilon_{\rm den}\rightarrow 5/8$ and $\varepsilon_{\rm ene}\rightarrow 29/32$. Figure 8: The percentage of the nonthermal number density $\varepsilon_{\rm den}$ (top) and nonthermal energy density $\varepsilon_{\rm ene}$ (bottom) as the functions of the temperature $T/mc^{2}$ and kappa index $\kappa$. It is also interesting to note the effect of the cutoff energy function introduced in the simulation analysis. In our simulation results, the cutoff energies were much larger than the heated plasma temperature, i.e., $\gamma_{cut}-1\sim 10^{2}\times T/mc^{2}$. By assuming that $\gamma_{cut}-1=10^{2}\times T/mc^{2}$, we determined that the difference between the nonthermal contribution with and without the cut-off energy is less than several $\%$ for $\kappa>3$. ## References * Birn & Priest [2007] Birn, J. & Priest, E. R. 2007, Reconnection of magnetic fields : magnetohydrodynamics and collisionless theory and observations / edited by J. Birn and E. R. Priest. Cambridge : Cambridge University Press, 2007. ISBN: 9780521854207 (hbk.) * Zweibel & Yamada [2009] Zweibel, E. G. & Yamada, M. 2009, Ann. Rev. Astron. Astrophys., 47, 291. doi:10.1146/annurev-astro-082708-101726 * Hoshino & Lyubarsky [2012] Hoshino, M., & Lyubarsky, Y. 2012, Space Science Reviews, 173, 521 * Uzdensky [2016] Uzdensky, D. A. 2016, Magnetic Reconnection: Concepts and Applications, 473 * Blandford et al. [2017] Blandford, R., Yuan, Y., Hoshino, M., & Sironi, L., 2017, Space Science Reviews, 207, 291 * Coppi et al. [1966] Coppi, B., Laval, G., & Pellat, R. 1966, Phys. Rev. Lett. , 16, 1207 * Hoh [1966] Hoh, F. C. 1966, Physics of Fluids, 9, 277 * Schindler [1966] Schindler, K. 1966, Phenomena in Ionized Gases, Volume II, VII International Conference, 736 * Biskamp & Schindler [1971] Biskamp, D. & Schindler, K. 1971, Plasma Physics, 13, 1013. doi:10.1088/0032-1028/13/11/003 * Galeev & Zelenyǐ [1976] Galeev, A. A. & Zelenyǐ, L. M. 1976, Soviet Journal of Experimental and Theoretical Physics, 43, 1113 * Bell [1978] Bell, A. R. 1978, Monthly Notices of the Royal Astronomical Society, 182, 147 * Blandford and Ostriker [1978] Blandford, R. D., & Ostriker, J. P. 1978, The Astrophysical Journal, 221, L29 * Zenitani and Hoshino [2001] Zenitani, S., & Hoshino, M. 2001, The Astrophysical Journal, 562, L63 * Jaroschek et al. [2004] Jaroschek, C. H., Lesch, H., & Treumann, R. A. 2004, The Astrophysical Journal, 605, L9 * Guo et al. [2014] Guo, F., Li, H., Daughton, W., & Liu, Y.-H. 2014, Physical Review Letters, 113, 155005 * Sironi et al. [2014] Sironi, L., & Spitkovsky, A. 2014, The Astrophysical Journal, 783, L21 * Guo et al. [2020] Guo, F., Liu, Y.-H., Li, X., Li, H., Daughton, W., and Kilian, P. 2020, Physics of Plasmas, 27, 080501. doi:10.1063/5.0012094 * Ball et al. [2018] Ball D., Sironi L., Özel F., 2018, ApJ, 862, 80. doi:10.3847/1538-4357/aac820 * Vasyliunas [1968] Vasyliunas, V. M. 1968, J. Geophys. Res., 73, 2839. doi:10.1029/JA073i009p02839 * Hoshino [1987] Hoshino, M. 1987, J. Geophys. Res., 92, 7368. doi:10.1029/JA092iA07p07368 * Hoshino [2020] Hoshino, M. 2020, Astrophys. J. , 900, 66. doi:10.3847/1538-4357/aba59d * Hoshino [2021] Hoshino, M. 2021, Physics of Plasmas, 28, 062106. doi:10.1063/5.0050389 * Harris [1962] Harris, E. G. 1962, Il Nuovo Cimento, 23, 115 * Asano et al. [2003] Asano, Y., Mukai, T., Hoshino, M., Saito, Y., Hayakawa, H., & Nagai, T. 2003, Journal of Geophysical Research (Space Physics), 108, 1189. doi:10.1029/2002JA009785 * Sharma et al. [2008] Sharma, A. S., Nakamura, R., Runov, A., Grigorenko, E. E., Hasegawa, H., Hoshino, M., Louarn, P., Owen, C. J., Petrukovich, A., Sauvaud, J. -A., et al. 2008, Annales Geophysicae, 26, 955. doi:10.5194/angeo-26-955-2008 * Feldman et al. [1975] Feldman, W. C., Asbridge, J. R., Bame, S. J., Montgomery, M. D., & Gary, S. P. 1975, J. Geophys. Res., 80, 4181. doi:10.1029/JA080i031p04181 * Livadiotis & McComas [2013] Livadiotis, G. & McComas, D. J. 2013, Space Sci. Rev., 175, 183. doi:10.1007/s11214-013-9982-9 * Lazar et al. [2017] Lazar, M., Pierrard, V., Shaaban, S. M., Fichtner, H., & Poedts, S. 2017, A & A, 602, A44. doi:10.1051/0004-6361/201630194 * Werner et al. [2016] Werner, G.R., Uzdensky, D. A., Cerutti, B., Nalewajko, K., & Begelman, M. C. 2016, Astrophys. J. Lett., 816, L8. doi:10.3847/2041-8205/816/1/L8 * Petropoulou & Sironi [2018] Petropoulou, M. & Sironi, L. 2018, Mon. R. Astron. Soc., 481, 5687. doi:10.1093/mnras/sty2702 * Hoshino et al. [1998] Hoshino, M., Mukai, T., Yamamoto, T., & Kokubun, S. 1998, J. Geophys. Res., 103, 4509. doi:10.1029/97JA01785 * Hoshino [2018] Hoshino, M. 2018, Astrophys. J. Lett., 868, L18. doi:10.3847/2041-8213/aaef3a * Hesse et al. [2018] Hesse, M., Norgren, C., Tenfjord, P., Burch, J. L., Liu, Y.-H., Chen, L.-J., Bessho, N., Wang, S., Nakamura, R., Eastwood, J. P., et al. 2018, Physics of Plasmas, 25, 122902. doi:10.1063/1.5054100 * Speiser [1965] Speiser, T. W. 1965, J. Geophys. Res., 70, 4219. doi:10.1029/JZ070i017p04219 * Hoshino et al. [2001] Hoshino, M., Mukai, T., Terasawa, T., & Shinohara, I. 2001, J. Geophys. Res., 106, 25979. doi:10.1029/2001JA900052 * Drake et al. [2006] Drake, J. F., Swisdak, M., Schoeffler, K. M., Rogers, B. N., & Kobayashi, S. 2006, Geophys. Res. Lett., 33, L13105. doi:10.1029/2006GL025957 * Pritchett [2001] Pritchett, P. L. 2001, J. Geophys. Res., 106, 3783. doi:10.1029/1999JA001006 * Dahlin et al. [2014] Dahlin, J. T., Drake, J. F., & Swisdak, M. 2014, Physics of Plasmas, 21, 092304. doi:10.1063/1.4894484 * Oka et al. [2010] Oka, M., Phan, T.-D., Krucker, S., Fujimoto, M., & Shinohara, I. 2010, Astrophys. J. , 714, 915. doi:10.1088/0004-637X/714/1/915 * Shay et al. [2014] Shay, M. A., Haggerty, C. C., Phan, T. D., Drake, J. F., Cassak, P. A., Wu, P., Oieroset, M., Swisdak, M., & Malakit, K. 2014, Physics of Plasmas, 21, 122902. doi:10.1063/1.4904203 * Haggerty et al. [2015] Haggerty, C. C., Shay, M. A., Drake, J. F., Phan, T. D., & McHugh, C. T. 2015, Geophys. Res. Lett., 42, 9657. doi:10.1002/2015GL065961 * Rowan et al. [2017] Rowan, M. E., Sironi, L., & Narayan, R. 2017, Astrophys. J. , 850, 29. doi:10.3847/1538-4357/aa9380 * Rowan et al. [2019] Rowan, M. E., Sironi, L., & Narayan, R. 2019, Astrophys. J. , 873, 2. doi:10.3847/1538-4357/ab03d7
11institutetext: Istanbul Technical University, Istanbul, Turkey 22institutetext: Mehmet Akif Ersoy Thoracic and Cardiovascular Surgery Training and Research Hospital, Istanbul, Turkey 33institutetext: Acibadem University, Istanbul, Turkey 44institutetext: King’s College London ∗Corresponding author: 44email<EMAIL_ADDRESS> # Shifted Windows Transformers for Medical Image Quality Assessment Caner Özer ** 1 1 Arda Güler 22 Aysel Türkvatan Cansever 22 Deniz Alis 33 Ercan Karaarslan 33 İlkay Öksüz 1144 ###### Abstract To maintain a standard in a medical imaging study, images should have necessary image quality for potential diagnostic use. Although CNN-based approaches are used to assess the image quality, their performance can still be improved in terms of accuracy. In this work, we approach this problem by using Swin Transformer, which improves the poor-quality image classification performance that causes the degradation in medical image quality. We test our approach on Foreign Object Classification problem on Chest X-Rays (Object-CXR) and Left Ventricular Outflow Tract Classification problem on Cardiac MRI with a four-chamber view (LVOT). While we obtain a classification accuracy of 87.1% and 95.48% on the Object-CXR and LVOT datasets, our experimental results suggest that the use of Swin Transformer improves the Object-CXR classification performance while obtaining a comparable performance for the LVOT dataset. To the best of our knowledge, our study is the first vision transformer application for medical image quality assessment. ###### Keywords: Visual Transformers Medical Image Quality Assessment Chest X-Rays Cardiac MRI. ## 1 Introduction It is the responsibility of technicians to ensure that the acquired medical scan should be of good quality so that the physician, who asked for the medical scan, will be able to diagnose accurately. However, this condition may not be satisfied due to the medical image quality problems such as patient breathing or arrhythmia in Cardiac Magnetic Resonance Imaging (MRI), ringing artifact presence in Computed Tomography (CT), and low dose or foreign object appearance on the Chest X-Rays (CXR). These image quality issues would also constitute a critical problem for automated processing of medical scans within a clinical decision support system consisting of big data. Maintaining a high standard in image quality is imperative also for a machine learning-guided medical image analysis downstream tasks. More specifically, existence of poor quality images may lead a machine learning methodology to learn spurious set of features which are not clinically useful. Hence, identifying, eliminating or correcting the samples with poor image quality is necessary for downstream tasks such as segmentation and disease diagnosis. The reasons for an image that should be eliminated from the pipeline include but are not limited to artefacts in MR Imaging [11], foreign object existence [15], or undesirable region appearance [12] within the scans. For this purpose, several datasets on a variety of modalities have been developed to overcome such problems. Ma et. al. [11] has developed a brain MRI dataset in which the patient data has been labelled according to the level of motion artefact. The authors have also developed a 4-layer convolutional neural network model to classify the samples according to their motion degradation level. JFHealthcare [8] proposed a MIDL 2020 challenge which comes up with a dataset that aims to localize the foreign objects within chest X-Rays. Swin-Transformers are becoming more common in medical image analysis, especially in the field of diagnosis. For instance, Zhang et al. [18] uses a cascaded mechanism for chest CT COVID-19 diagnosis which uses a U-Net to extract lungs as a foreground and background segmentation mechanism, and Swin- Transformer uses these segmented lungs to find out whether the patient has COVID-19 or not. Chen et al. [2] combines a statistical significance-based slice sampling scheme and Swin-Transformers to propose a framework that can detect COVID-19 from chest CT scans by using a single slice. In addition to diagnosis, Swin-Transformers are becoming more common to be used for other problems, such as brain tumor segmentation [1] [5], and cardiac CT reconstruction [13]. However, despite being promising for these problems in medical imaging, Swin-Transformers’ have yet to be studied for medical image quality assessment tasks. In this work, we propose a SwinTransformer-based [10] approach for medical image quality analysis where our contributions are as follows: * • We train a variety of SwinTransformer models in order to classify X-Ray images with foreign objects and Cardiac MRI images with a 4-chamber view with Left Ventricular Outflow Tract (LVOT). * • We compare our approach with deep neural network (DNN) based approaches, such as ResNet [6] and EfficientNet [14]. * • To our best knowledge, this is the first paper to perform medical image quality assessment through SwinTransformers. ## 2 Methodology Figure 1: Swin Transformer pipeline for medical image quality analysis. ### 2.1 Pre-processing In order to train our model, we apply a number of preprocessing procedures prior to feeding the input images to the network. In this regard, instead of manually defining the data augmentations and their corresponding hyperparameters, we use RandAugment [3] which uses a previously found automated data augmentation strategy to train a DNN. After that, we resize these images to a pre-determined resolution of $H\times W$. Finally, we normalize the images by using ImageNet statistics that they become ready-to- use in a pre-trained model. ### 2.2 Swin Transformer Following the pre-processing procedure, we used the Swin Transformer model that is being demonstrated in Fig. 1. Given an input image, Swin Transformer first divides the image into patches using a Patch Partition layer that generates $48$-dimensional representations for each of the $4\times 4$ patches on the image. After that, as shown in Stage 1, the Linear Embedding layer transforms $48$-d representations into $C$ dimensions. The second component of Stage 1 is the Swin Transformer Blocks which has been inherited from the general Visual Transformer (ViT) [4] architecture. Despite having the same high-level block structure as the one in ViT, multiheaded self-attention (MSA) blocks are replaced with an alternate ordering of window- based MSA (W-MSA) and shifted window-based MSA (SW-MSA), sequentially. In W-MSA, the feature map is divided into windows and MSA computes the output on each of them. However, this brings a cost of missing global connectivity, since this was not an issue for MSA in ViT that operates globally. For this reason, cyclic shifting operation is used to overcome this problem with SW-MSA that, the feature map is circularly shifted on the height and width axes by the half-length of the patch size. Then, W-MSA is executed on the resulting feature map, and a reverse circular shift operation is used to obtain a reverted feature map without any shifts. In all of Stages 2, 3, and 4, we have a Patch Merging layer and a varying number of Swin Transformer blocks. In the Patch Merging layer, each of the pixels located on the $2\times 2$ patches on the feature map are concatenated across the channel axis, i.e., $\frac{H}{2}\times\frac{W}{2}\times C$ to $\frac{H}{4}\times\frac{W}{4}\times 4C$. Then, after applying the layer normalization layer, a linear layer is used to reduce the dimensionality to half, i.e., to $\frac{H}{4}\times\frac{W}{4}\times 2C$. The output is then transferred to the Swin Transformer blocks for further processing. Lastly, after using layer normalization followed by an average pooling layer to calculate the features of the Swin Transformer model, the output is used for classification in a single Linear layer. This model classifies if there is anything present in the input image that can affect the image quality adversely. The hyperparameter settings can be found in Table 1. Stochastic depth refers to the probability of disregarding the transformation that is applied on a sample within a batch during training, while being processed within a Swin Transformer block, embedding dims $C$ is the dimensionality of the patch-wise embeddings, depths and MSA heads correspond to the number of repeated Swin Transformer blocks and head count of the MSA layer for each stage, respectively. Table 1: Hyperparameter settings for different Swin Transformer models. Model Name | | Stochastic Depth --- | Embedding Dims $(C)$ --- Depths | MSA Heads Swin-Tiny | 0.2 | 96 | 2, 2, 6, 2 | 3, 6, 12, 24 Swin-Small | 0.3 | 96 | 2, 2, 18, 2 | 3, 6, 12, 24 Swin-Base | 0.2 | 128 | 2, 2, 18, 2 | 4, 8, 16, 32 ### 2.3 Training the Model We adopt the official code of Swin Transformers111https://github.com/microsoft/Swin-Transformer in which we performed some modifications to run it on PyTorch 1.10.2. We also used TorchVision 0.11.3 to inherit the pre-trained models for EfficientNet and ResNet pre-trained architectures. We apply the learning rate warm-up scheme and learning rate scheduling and leverage color jittering, stochastic depth (only for Swin Transformer) [7], Random Erasing [19] MixUp [17], CutMix [16] and RandAugment [3] data augmentation methods for promoting model generalization. For the foreign object classification model, we set the batch size to 16, gradient accumulation steps to 8 and window size to 8. Instead for the LVOT classification model, we set the batch size to 64, gradient accumulation steps to 2 and window size to 7, due to lower input image resolution comparing chest X-Rays. We employ an AdamW [9] optimizer with weight-decay of $10^{-8}$ to fine-tune both of these models by using an ImageNet pretrained model, as we set the initial learning rate to $0.00006$ and use a linear learning-rate scheduler. The fine-tuning procedure takes $60$ epochs with $5$ epochs of a warm-up period. We also trained the models from scratch for $300$ epochs that, the learning rate is set to $0.0005$ and the objective function is Cross Entropy. You can access to our repository by using the following link: https://github.com/canerozer/qct. ## 3 Experimental Results ### 3.1 Datasets We demonstrate our results on the Object-CXR [8], a challenge dataset where the aim is to classify the Chest X-Ray scans with foreign objects and to localize the foreign objects, and the LVOT classification dataset, which is constructed by selecting a subset of the Cardiac MR (CMR) scans with 4-chamber view that were acquired in the Mehmet Akif Ersoy Thoracic and Cardiovascular Surgery Training and Research Hospital between the years 2016 and 2019. Object-CXR dataset contains 8,000 samples for training and 1,000 samples for each of the validation and testing splits. There is an equal number of positive and negative samples for each split which does and does not contain foreign objects. We only use the class label, but not the localization ground truth for this dataset during training. We set the fixed input size to $H=W=1024$ since the images can reach a spatial resolution of $4048\times 4932$. On the other hand, LVOT classification dataset contains 272 positive and 278 negative samples that are used for training, 34 positive and 35 negative samples for validation, and 35 positive and 35 negative samples for testing. These images have relatively way less spatial resolution than the ones in Object-CXR, hence, we set $H=W=224$. In addition, since our pipeline is capable of processing 2D images, we slice the 3D CMR patient scans into 2D images, which is also helpful in terms of increasing the number of samples for each split by a factor of 25. ### 3.2 Left Ventricular Outflow Tract Classification Results First, we start by comparing the performances of different Swin Transformer models that are named as Swin-Tiny, Swin-Small, and Swin-Base in Table 2. We additionally measured the differences between using a pretrained model and training from scratch. We notice that, the accuracy is the highest for the Swin-Tiny model with an accuracy of $95.59\%$. Nevertheless, Swin-Small’s AUC (Area Under Curve) score is performing the best among these three models with a score of $0.980$. For this reason, we decided to include both of the Swin- Small and Swin-Tiny models to compare them with the other baselines. Meanwhile, switching from pre-trained model to training from scratch significantly decreases the accuracy of at least around $12.35\%$ and AUC score of $0.059$. We relate this to the data-hungry regime of Swin Transformers, as they need abundant data samples to capture the inductive bias during training. Hence, it becomes necessary to use the pretrained models or to increase the number of samples to the order of tens of thousands. Table 2: Comparison among Swin Transformer models with different capacity for the LVOT classification task. Model | Pretrained? | Acc | AUC | | # of params --- GFLOPs | | Throughput (img/sec) --- Swin-Tiny | Yes | 95.59 | 0.969 | 27,520,892 | 4.5 | 515 No | 72.93 | 0.757 Swin-Small | Yes | 95.48 | 0.980 | 48,838,796 | 8.7 | 313 No | 83.13 | 0.921 Swin-Base | Yes | 93.86 | 0.976 | 86,745,274 | 15.4 | 222 No | 67.01 | 0.738 Table 3: Our results on the LVOT dataset indicate that the performance is comparable to the other DNN-based methods. Model | Acc | AUC | | # of params --- GFLOPs | Throughput EfficientNet-B0 | 93.74 | 0.979 | 4,010,110 | 0.4 | 2274 EfficientNet-B1 | 93.68 | 0.971 | 6,515,746 | 0.6 | 1565 EfficientNet-B2 | 94.61 | 0.971 | 7,703,812 | 0.7 | 1480 ResNet-34 | 95.19 | 0.974 | 21,285,698 | 3.7 | 1963 ResNet-50 | 95.77 | 0.979 | 23,512,130 | 4.1 | 1005 ResNet-101 | 95.59 | 0.969 | 42,504,258 | 7.8 | 587 ResNet-152 | 96.00 | 0.988 | 58,147,906 | 11.6 | 409 Swin-Tiny | 95.59 | 0.969 | 27,520,892 | 4.5 | 515 Swin-Small | 95.48 | 0.980 | 48,838,796 | 8.7 | 313 Then, we compare our approach with other deep learning baselines as ResNet [6] and EfficientNet [14] as shown in Table 3. We see that, despite the differences in the number of parameters and GFLOPs, the performance varies insignificantly ($p>0.05$) with outstanding accuracy and AUC metrics. We relate this to the triviality of the LVOT classification task, where the goal is to check if fifth chamber exists on the Cardiac MRI scan - generally located at the centre. As a result, it becomes possible to obtain testing accuracy of over $95\%$ and AUC scores exceeding $0.969$ for all ResNet architectures and Swin-Small without overfitting. Still, these models are ready to be deployed for a real-world scenario for alerting LVOT regions on 4-chamber Cardiac MRIs almost instantaneously. ### 3.3 Foreign Object Classification Results Table 4: Results for the Object-CXR dataset. Model | Acc | AUC | | # of params --- GFLOPs | Throughput EfficientNet-B0 | 82.1 | 0.873 | 4,010,110 | 8.3 | 119 EfficientNet-B1 | 83.0 | 0.882 | 6,515,746 | 12.3 | 83 EfficientNet-B2 | 83.2 | 0.890 | 7,703,812 | 14.2 | 78 ResNet-34 | 82.9 | 0.895 | 21,285,698 | 91.8 | 94 ResNet-50 | 82.8 | 0.897 | 23,512,130 | 102.7 | 42 ResNet-101 | 84.1 | 0.906 | 42,504,258 | 195.8 | 26 ResNet-152 | 85.1 | 0.909 | 58,147,906 | 241.5 | 22 Swin-Base | 87.1 | 0.922 | 86,766,330 | 324.6 | 11 The results show Swin Transformer’s strength on a higher dimensional input, namely, on the Chest X-Ray data for foreign object classification. The model complexity is highly correlated with the model performance that, the best performance is obtained by the Swin-Base model with $87.1\%$ testing accuracy and $0.922$ testing AUC score. The closest performance to the best model is the ResNet-152 model, which has $1.5\times$ less number of parameters and $25\%$ less floating point operations, at a cost of $2\%$ less testing accuracy and $0.013$ less AUC score. Despite Swin-Base is working slightly slower than real-time due to its huge input size, its significant performance gains ($p<0.005$) makes it a promising approach. ### 3.4 Qualitative Analysis Prior to utilizing our model in a clinical study, we further examine our model under extreme conditions. Firstly, we start with the Object-CXR classification task as some cases are shown in Fig. 2. Regardless of the orientation and contrast factors, our model can correctly classify the X-Ray images which does and does not contain foreign objects as demonstrated in Fig. 2a and 2b, respectively. What is interesting especially for the rightmost image of Fig. 2b is that, even though the foreign objects may not be visible to human eye without altering the contrast, our model can still successfully classify the image as it contains foreign objects. Foreign objects are marked with a blue bounding box in Fig. 2b. Figure 2: Some true challenging samples due to their medical image quality factors, from Object-CXR dataset. We also perform a qualitative analysis on the classification results as some of the edge cases are shown in Fig. 3 for the LVOT classification task. In Fig. 3a, where all 3 images are labelled as good quality images without LVOT appearance, the model can correctly classify even when the images are subject to motion artefacts and local noise upto some level. Also, as shown in Fig. 3b, our model can correctly classify the images containing LVOT regions under low contrast. We show the regions with LVOT with a blue bounding box in Fig. 3b. Figure 3: Some true challenging samples due to their medical image quality factors, from LVOT dataset. ## 4 Discussion & Conclusion In this work, we propose to use Swin Transformers for medical image quality analysis, where we validated our approach on two datasets working on two different modalities. We outperform a variety of ResNet and EfficientNet baselines on the Object-CXR dataset by obtaining a testing accuracy of $87.1\%$ while obtaining a comparable performance for the LVOT classification objective. Comparing Swin Transformer to the other baselines and performing a qualitative analysis demonstrate its potential for utilizing it in a clinical setting. Although the problem setting has been restricted to foreign object classification in Chest X-Rays and LVOT classification for Cardiac MRI’s, it is possible to generalize the approach for other medical image quality problems such as motion artefact grade classification. In addition, the interpretability aspect of the Swin Transformers is a critical avenue of research for the safe transition of an automated medical image analysis pipeline. ## Acknowledgments This paper has been produced benefiting from the 2232 International Fellowship for Outstanding Researchers Program of TUBITAK (Project No: 118C353). However, the entire responsibility of the publication/paper belongs to the owner of the paper. The financial support received from TUBITAK does not mean that the content of the publication is approved in a scientific sense by TUBITAK. ## References * [1] Cao, H., Wang, Y., Chen, J., Jiang, D., Zhang, X., Tian, Q., Wang, M.: Swin-unet: Unet-like pure transformer for medical image segmentation (2021) * [2] Chen, G.L., Hsu, C.C., Wu, M.H.: Adaptive distribution learning with statistical hypothesis testing for covid-19 ct scan classification. In: 2021 IEEE/CVF International Conference on Computer Vision Workshops (ICCVW). pp. 471–479 (2021). https://doi.org/10.1109/ICCVW54120.2021.00057 * [3] Cubuk, E.D., Zoph, B., Shlens, J., Le, Q.: Randaugment: Practical automated data augmentation with a reduced search space. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems. vol. 33, pp. 18613–18624. Curran Associates, Inc. (2020), https://proceedings.neurips.cc/paper/2020/file/d85b63ef0ccb114d0a3bb7b7d808028f-Paper.pdf * [4] Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S., Uszkoreit, J., Houlsby, N.: An image is worth 16x16 words: Transformers for image recognition at scale. In: 9th International Conference on Learning Representations, ICLR 2021, Virtual Event, Austria, May 3-7, 2021. OpenReview.net (2021), https://openreview.net/forum?id=YicbFdNTTy * [5] Hatamizadeh, A., Nath, V., Tang, Y., Yang, D., Roth, H., Xu, D.: Swin UNETR: swin transformers for semantic segmentation of brain tumors in MRI images. CoRR abs/2201.01266 (2022), https://arxiv.org/abs/2201.01266 * [6] He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016, Las Vegas, NV, USA, June 27-30, 2016. pp. 770–778. IEEE Computer Society (2016). https://doi.org/10.1109/CVPR.2016.90, https://doi.org/10.1109/CVPR.2016.90 * [7] Huang, G., Sun, Y., Liu, Z., Sedra, D., Weinberger, K.Q.: Deep networks with stochastic depth. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) Computer Vision - ECCV 2016 - 14th European Conference, Amsterdam, The Netherlands, October 11-14, 2016, Proceedings, Part IV. Lecture Notes in Computer Science, vol. 9908, pp. 646–661. Springer (2016). https://doi.org/10.1007/978-3-319-46493-0_39, https://doi.org/10.1007/978-3-319-46493-0_39 * [8] JFHealthcare: Automatic detection of foreign objects on chest x-rays. https://academictorrents.com/details/fdc91f11d7010f7259a05403fc9d00079a09f5d5 (2020), accessed on 11/27/2020 * [9] Kingma, D.P., Ba, J.: Adam: A method for stochastic optimization. In: Bengio, Y., LeCun, Y. (eds.) 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Conference Track Proceedings (2015), http://arxiv.org/abs/1412.6980 * [10] Liu, Z., Lin, Y., Cao, Y., Hu, H., Wei, Y., Zhang, Z., Lin, S., Guo, B.: Swin transformer: Hierarchical vision transformer using shifted windows. In: Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV). pp. 10012–10022 (October 2021) * [11] Ma, J.J., Nakarmi, U., Kin, C.Y.S., Sandino, C.M., Cheng, J.Y., Syed, A.B., Wei, P., Pauly, J.M., Vasanawala, S.S.: Diagnostic image quality assessment and classification in medical imaging: Opportunities and challenges. In: 17th IEEE International Symposium on Biomedical Imaging, ISBI 2020, Iowa City, IA, USA, April 3-7, 2020. pp. 337–340. IEEE (2020). https://doi.org/10.1109/ISBI45749.2020.9098735, https://doi.org/10.1109/ISBI45749.2020.9098735 * [12] Oksuz, I., Ruijsink, B., Puyol-Antón, E., Sinclair, M., Rueckert, D., Schnabel, J.A., King, A.P.: Automatic left ventricular outflow tract classification for accurate cardiac mr planning. In: 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018). pp. 462–465 (2018). https://doi.org/10.1109/ISBI.2018.8363616 * [13] Pan, J., Wu, W., Gao, Z., Zhang, H.: Mist-net: Multi-domain integrative swin transformer network for sparse-view CT reconstruction. CoRR abs/2111.14831 (2021), https://arxiv.org/abs/2111.14831 * [14] Tan, M., Le, Q.V.: Efficientnet: Rethinking model scaling for convolutional neural networks. In: Chaudhuri, K., Salakhutdinov, R. (eds.) Proceedings of the 36th International Conference on Machine Learning, ICML 2019, 9-15 June 2019, Long Beach, California, USA. Proceedings of Machine Learning Research, vol. 97, pp. 6105–6114. PMLR (2019), http://proceedings.mlr.press/v97/tan19a.html * [15] Xue, Z., Candemir, S., Antani, S., Long, L.R., Jaeger, S., Demner-Fushman, D., Thoma, G.R.: Foreign object detection in chest x-rays. In: 2015 IEEE International Conference on Bioinformatics and Biomedicine (BIBM). pp. 956–961 (2015). https://doi.org/10.1109/BIBM.2015.7359812 * [16] Yun, S., Han, D., Chun, S., Oh, S.J., Yoo, Y., Choe, J.: Cutmix: Regularization strategy to train strong classifiers with localizable features. In: 2019 IEEE/CVF International Conference on Computer Vision, ICCV 2019, Seoul, Korea (South), October 27 - November 2, 2019. pp. 6022–6031. IEEE (2019). https://doi.org/10.1109/ICCV.2019.00612, https://doi.org/10.1109/ICCV.2019.00612 * [17] Zhang, H., Cissé, M., Dauphin, Y.N., Lopez-Paz, D.: mixup: Beyond empirical risk minimization. In: 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018, Conference Track Proceedings. OpenReview.net (2018), https://openreview.net/forum?id=r1Ddp1-Rb * [18] Zhang, L., Wen, Y.: A transformer-based framework for automatic covid19 diagnosis in chest cts. In: 2021 IEEE/CVF International Conference on Computer Vision Workshops (ICCVW). pp. 513–518 (2021). https://doi.org/10.1109/ICCVW54120.2021.00063 * [19] Zhong, Z., Zheng, L., Kang, G., Li, S., Yang, Y.: Random erasing data augmentation. In: The Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, The Thirty-Second Innovative Applications of Artificial Intelligence Conference, IAAI 2020, The Tenth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2020, New York, NY, USA, February 7-12, 2020. pp. 13001–13008. AAAI Press (2020), https://aaai.org/ojs/index.php/AAAI/article/view/7000
# Field-free anomalous junction and superconducting diode effect in spin split superconductor/topological insulator junctions T.H. Kokkeler<EMAIL_ADDRESS>Donostia International Physics Center (DIPC), 20018 Donostia–San Sebastián, Spain Interfaces and Correlated Electron Systems, Faculty of Science and Technology, University of Twente, Enschede, The Netherlands A.A. Golubov Interfaces and Correlated Electron Systems, Faculty of Science and Technology, University of Twente, Enschede, The Netherlands F. S. Bergeret Centro de Física de Materiales (CFM-MPC) Centro Mixto CSIC- UPV/EHU, E-20018 Donostia-San Sebastián, Spain Donostia International Physics Center (DIPC), 20018 Donostia–San Sebastián, Spain ###### Abstract We study the transport properties of a diffusive Josephson junction between two spin-split superconductors made of superconductor-ferromagnetic insulator bilayers (FIS) on top of a 3D topological insulator (TI). We derive the corresponding Usadel equation describing the quasiclassical Green’s functions in these systems and first solve the equation analytically in the weak- proximity case. We demonstrate the appearance of an anomalous phase in the absence of an external magnetic field. We also explore non-reciprocal electronic transport. Specifically, we calculate the junction’s diode efficiency $\eta$ by solving the Usadel equation. We obtain a sizable diode effect even at zero applied magnetic field. We discuss how the diode efficiency $\eta$ depends on the different parameters and find a non-monotonic behavior of $\eta$ with temperature. ††preprint: APS/123-QED ## I Introduction Recent advances attracting attention in superconductivity research are effects related to non-reciprocal charge transport [1, 2, 3, 4, 5, 6, 7, 8, 9, 10], particularly the superconducting diode effect [11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27]. A mesoscopic superconducting junction is called a superconducting diode if the critical current is different for opposite current directions. That is, for a superconducting diode the minimum, $I_{-}=\min_{\phi}{I_{s}}(\phi)$, and maximum, $I_{+}=\max_{\phi}{I_{s}}(\phi)$, of the current phase relation (CPR) are unequal in magnitude. If a current $I$ flows in such a junction, with $\min(|I_{-}|,I_{+})<I<\max(|I_{-}|,I_{+})$, in one direction it is a supercurrent, whereas in the other direction the current dissipates. In a conventional Josephson junction the critical current is the same in both directions, the CPR has the following symmetry: $I(-\phi)=-I(\phi)$ [28]. This symmetry holds if either time reversal symmetry or inversion symmetry is present in the system [29, 30]. The superconducting diode effect can thus only be obtained if both time reversal symmetry and inversion symmetry are broken [11, 12]. Time reversal symmetry breaking can be achieved by a magnetic field. On the other hand inversion symmetry can be broken intrinsically, such as in topological insulators [31, 32, 33] or superconductors with Rashba spin orbit coupling [34, 35, 36, 37, 38, 39, 40, 41]. Inversion symmetry can also be broken by using an asymmetric junction geometry [42] or asymmetry of the device originated in the fabrication [16]. If both time-reversal symmetry and inversion-symmetry are broken, in general $I(-\phi)\neq-I(\phi)$ and thus possibly $I(\phi=0)\neq 0$. Such junctions, for which the current vanishes at a nonzero phase difference, are called $\phi_{0}$-junctions [35]. In weak coupling Josephson junctions, the CPR is proportional to $\sin(\phi+\phi_{0})$; hence, in this regime, one cannot observe the diode effect. However, if higher harmonics contribute to the current, in general $I_{+}\neq|I_{-}|$, and the diode effect can be observed [15, 13, 17, 18]. One way of breaking time-reversal symmetry without an external magnetic field is to attach a ferromagnetic insulator (FI) to the superconductor. FI-S systems have been discussed extensively in the literature [43, 44, 45, 46]. The exchange interaction between the localized magnetic moments of the FI and the itinerant electrons in the SC leads to a spin split in the density of states of the latter. Spin split superconductors form an active field of research with varying directions [47, 48, 49, 50, 51, 52, 53]. In this work, we investigate the diode effect in a Josephson junction made of spin-split superconducting electrodes on the 2D surface of a disordered 3D topological insulator (TI). Specifically, our setup consists of two spin-split superconductors (FIS) placed on top of a topological insulator, see Fig. 1a. The spin-splitting in the superconductor breaks time-reversal symmetry, whereas inversion symmetry is broken because we consider only the top surface of the topological insulator. Thus, the conditions to have a $\phi_{0}$-effect are fulfilled even without an external magnetic field. Using the linearized Usadel equation, we first show analytically that such CPR exhibits the $\phi_{0}$-effect. Going beyond the linear regime we compute the diode efficiency. Even in the case of low transmission FIS/TI interfaces we obtain an efficiency of 1%. By increasing the interface transmission the efficiency can reach values larger than 7%. We also find that for short junctions the efficiency is maximized at a finite temperature independent of the strength of the exchange field. The article is structured as follows. In section II we introduce the setup and the basic equations. We derive the Usadel equation for a diffusive topological insulator in proximity with a spin-split superconductor. In section III, we focus on analytical results that can be obtained by linearizing the Usadel equation, which is valid under the assumption that the proximity effect is small. Even though in this limit the CPR contains only the first harmonic, and thus has no diode effect, we can determine the condition for observation of the anomalous Josephson currents. The latter is the precursor of a diode effect in the non-linearized equation. We also show that the $\phi_{0}$\- effect is suppressed by impurity scattering. In section IV we go beyond the linear regime and present our numerical results for the nonlinear equation and the diode efficiency as a function of the temperature for different values of the exchange field, length of the junction, and transparency of the interfaces. Finally in section V we conclude and propose real material combinations to test our predictions. Throughout the paper it is assumed that $\hbar=k_{B}=1$. ## II The system and basic equations We consider the FIS-TI-FIS system shown in Fig. 1a. A FIS can be realized by placing a ferromagnetic insulator on top of a superconducting film (SC) with a conventional s-wave pair potential. The thickness of the latter is assumed to be small compared to its coherence length, so that we can assume a homogeneous splitting field in the SC induced by the magnetic proximity effect [54]. We neglect the inverse proximity effect of the topological insulator on the superconductor. In this setup a current can flow through the top surface of the TI from one FIS to the other. Because the system is finite in $x$ direction, no current can flow at $x=\pm L/2$, where $L$ is the length of the TI. We denote by $L_{1}$ the length of the TI between the two FIS electrodes. (a) (b) Figure 1: (a) Sketch of the junction under consideration. (b) The CPR of the FIS-TI-FIS junction for different temperatures. The parameters chosen are $\frac{\gamma_{0}}{E_{\text{Th}}}=\frac{\Delta_{0}}{E_{\text{Th}}}=\frac{5h}{2E_{\text{Th}}}=25$, $\frac{L_{1}}{L}=\frac{1}{10}$ and $\frac{l}{L}=0.08\ll 1$. Inset: zoom in of the CPR around $\phi=0$. We assume that the transport at the TI surface is diffusive and can be described by the Usadel equation: the derivation of this equation for our system closely resembles the derivation of the Usadel equation for a TI in an exchange field as presented in Refs. [55, 56]. However, whereas superconductivity in the systems discussed in these papers is introduced as an effective pair potential, for the spin split superconductor a different approach is taken. We incorporate the effect of the spin-split superconductor as a self-energy term $\bar{\Sigma}_{s}$. For the self-energy we follow the approach similar to [57, 58], in which the self-energy, up to second order in the tunneling parameter $T_{1}$ between the TI and the superconductor, is introduced as $\displaystyle\bar{\Sigma}_{s}=T_{1}^{2}\rho_{S}\check{\tau}_{3}\sigma_{0}\bar{G^{\prime}}_{S}\sigma_{0}\check{\tau}_{3}=T_{1}^{2}\rho_{S}\bar{G}_{S},$ (1) where $\rho_{S}$ is the density of states in the superconductor, $\sigma_{0}$ is the identity matrix in spin space and $\check{\tau}_{3}$ is the third Pauli matrix in electron-hole space. $\bar{G^{\prime}}_{S}$ is the momentum integrated Green’s function in the spin-split superconductor and $\bar{G}_{S}=\check{\tau}_{3}\sigma_{0}\bar{G^{\prime}}_{S}\sigma_{0}\check{\tau}_{3}$ is introduced to shorten notation. Note that the only effect of this transformation by $\check{\tau}_{3}\sigma_{0}$ is to negate the pair amplitudes. The self-energy term appears as an added term in the commutator on the left hand side of the Eilenberger equation. The Eilenberger equation thus reads, using the same presentation in spin-Nambu space as in [55]: $\displaystyle\frac{v_{F}}{2}\\{\\{\eta_{j},\nabla_{j}\check{g}(1+\vec{n}_{F}\cdot\vec{\eta})\\}\\}$ $\displaystyle=[\check{g}(1+\vec{n}_{F}\cdot\eta),\omega_{n}\check{\tau}_{3}+\Gamma(x)\bar{G}_{S}+\frac{\langle\check{g}(1+\vec{n}_{F}\cdot\vec{\eta})\rangle}{2\tau}],$ (2) where $\\{\cdot,\cdot\\}$ denotes the anticommutator and $[\cdot,\cdot]$ the commutator. In our notation $\check{g}$ is the quasiclassical Green’s function, $\omega_{n}$ is the nth Matsubara frequency, $\vec{n}_{F}$ is the direction of the momentum at the Fermi surface, $v_{F}$ is the magnitude of the Fermi velocity, $\tau$ is the collision time, $\check{\tau}_{3}$ is the third Pauli matrix in Nambu space, $\mu$ is the Fermi energy and $\vec{\eta}=(-\sigma_{2},\sigma_{1},0)$. The tunneling parameter $T_{1}$ is nonzero only in FIS regions. To reflect this we introduce the boundary parameter $\Gamma(x)$: $\Gamma(x)=\gamma_{0}\Theta(|x|-\frac{L_{1}}{2})\Theta(\frac{L}{2}-|x|),$ (3) where $\Theta$ denotes the Heaviside function and $\gamma_{0}=T_{1}^{2}\rho_{S}$. The Green’s function is written as $\check{g}\frac{1}{2}(1+\vec{n}_{F}\cdot\vec{\eta})$ to reflect the strong coupling between spin and direction of momentum in a topological insulator. In this work a 2D surface of a 3D TI is studied. Therefore scattering is not prohibited, unlike in a 1D edge of a 2D TI. We assume the junction is in the dirty limit, that is, the inverse scattering time $\frac{1}{\tau}$ is much larger than any energy scale other than the chemical potential $\mu$. In that case the Green’s function $g$ is almost isotropic. Thus, it is a good approximation to keep only the zeroth and first term in the expansion in angular momentum, that is, $\check{g}\approx\check{g}_{s}+\vec{n}_{F}\cdot\vec{\check{g}}_{a},$ (4) where the zeroth order $\check{g}_{s}$ and the first order angular momentum $\vec{\check{g}}_{a}$ satisfy $\check{g}_{s}^{2}=\mathbf{1}$ and $\check{g}_{s}\vec{\check{g}}_{a}+\vec{\check{g}}_{a}\check{g}_{s}=\vec{0}$ in order to satisfy the normalisation condition $\bar{g}^{2}=\mathbf{1}$ up to second order in $\tau$. The Green’s functions $\check{g}_{s}$ and $\vec{\check{g}}_{a}$ do not have any degrees of freedom in momentum space nor in spin space and are thus functions which map position into the space of 2 by 2 matrices. Using the expansion in angular momentum, the Usadel equation can be derived. The strategy followed to derive the Usadel equation for our structure is very similar to the strategy used in [55], [56]. To this end, we first write $\bar{G}_{S}=\check{G}_{S0}\sigma_{0}+\vec{\check{G}}_{S1}\cdot\vec{\sigma},$ (5) where $\check{G}_{S0}$ and $\vec{\check{G}}_{S1}$ are matrix functions without spin degrees of freedom, and $\vec{\sigma}$ is the vector of Pauli matrices in spin space. It is assumed that the exchange field in the FI always points in the same direction, so that there are no domain walls which may affect the density of states significantly [53]. The position independent Green’s function in the spin-split superconductor is $\displaystyle\bar{G}_{S}$ $\displaystyle=\frac{1}{2}(1+\vec{b}\cdot\vec{\sigma})\check{G}_{\uparrow}+\frac{1}{2}(1-\vec{b}\cdot\vec{\sigma})\check{G}_{\downarrow}$ (6) $\displaystyle\check{G}_{\uparrow,\downarrow}$ $\displaystyle=g_{\uparrow,\downarrow}\tau_{3}+f_{\uparrow,\downarrow}(\cos{\phi}\tau_{1}+\sin\phi\tau_{2}),$ (7) where $g_{\uparrow,\downarrow}=(\omega_{n}\pm ih)/{\sqrt{(\omega_{n}\pm ih)^{2}+\Delta^{2}}}$ are the normal parts and $f_{\uparrow,\downarrow}=\Delta/{\sqrt{(\omega_{n}\pm ih)^{2}+\Delta^{2}}}$ are the anomalous parts [59]. The + sign is used for the spin-up component and the - sign for the spin down component, $\tau_{1,2,3}$ are the Pauli matrices in Nambu space, $h$ is the magnitude of the exchange field $\vec{h}=h\vec{b}$, $\Delta$ is the superconducting potential calculated self-consistently and $\phi$ is the phase of the superconductor. Combining Eq. (5) with Eqs. (6) and (7), we may write $\displaystyle\check{G}_{S0,1}$ $\displaystyle=g_{S0,1}\tau_{3}+f_{S0,1}(\cos{\phi}\tau_{1}+\sin{\phi}\tau_{2}),$ (8) $\displaystyle\vec{\check{G}}_{S1}$ $\displaystyle=\check{G}_{S1}\vec{b}\cdot\vec{\sigma},$ (9) where $g_{s0,1}=(g_{\uparrow}\pm g_{\downarrow})/2$ and $f_{S0,1}=(f_{\uparrow}\pm f_{\downarrow})/2$. The component $f_{S0}$ of the condensate is the usual singlet component, whereas $f_{S1}$ is the odd- frequency triplet component [59]. The Usadel equation for the angular averaged Green’s function $g_{s}$ without spin degrees of freedom in the TI is obtained analogous to the approach laid out in [55]. We combine the spin-trace of the equation obtained by angular averaging and the equation obtained by angular averaging after multiplication by $\vec{n}_{F}$. The resulting Usadel equation is $D\hat{\nabla}(\check{g}_{s}\hat{\nabla}\check{g}_{s})=[\omega_{n}\check{\tau}_{3}+\frac{\Gamma(x)}{2}\check{G}_{S0},\check{g}_{s}],$ (10) with $D=v_{F}^{2}\tau$. Eq. (10) is similar to the Usadel equation in normal metals. However, the derivative is replaced by a generalized derivative: $\displaystyle\hat{\nabla}=\nabla+\frac{\Gamma(x)}{2v_{F}}[\cdot,\check{G}_{S1y}]e_{x}-\frac{\Gamma(x)}{2v_{F}}[\cdot,\check{G}_{S1x}]e_{y},$ (11) For a spin split superconductor this becomes $\displaystyle\hat{\nabla}=\nabla+\frac{\Gamma(x)}{2v_{F}}(b_{y}e_{x}-b_{x}e_{y})[\cdot,\check{G}_{S1}].$ (12) This derivative is similar to the derivative presented in [56] for a topological insulator with an exchange field, in fact, Eq. (10) reduces to this expression if $\check{G}_{S1}=h\check{\tau_{3}}$. Throughout this work we will assume that the magnetic field is oriented perpendicular to the current direction, so that $b_{y}=1$ and $b_{x}=0$. The equation is accompanied by the boundary conditions $\displaystyle\hat{\nabla}G_{S1}(x=\pm\frac{L}{2})=0.$ (13) In this paper it is assumed that the ferromagnetic insulator is either very thin or very thick, so that there is no y nor z-dependence in the problem. As a consequence the equation becomes effectively one-dimensional. From the solutions of Eqs. (10),(12) and (13) one can determine the current: $\displaystyle I=\frac{\sigma_{N}}{2}T\sum_{n}\text{Tr}\Big{(}\tau_{3}\bar{G}(x^{*},\omega_{n})\hat{\nabla}\bar{G}(x^{*},\omega_{n})\Big{)}\;,$ (14) where $\sigma_{N}$ is the normal state conductance, $T$ is the temperature entering the Matsubara frequencies $\omega_{n}=(2n+1)\pi T$, and $x^{*}$ is any position for which $\Gamma(x^{*})=0$. The quantities of interest in this article are the maximum supercurrents $I_{c}^{+}$ and $|I_{c}^{-}|$ in both directions, and the diode efficiency, defined by $\displaystyle\eta=\frac{I_{c}^{+}-|I_{c}^{-}|}{I_{c}^{+}+|I_{c}^{-}|}.$ (15) Before showing the numerical solution of the above boundary problem, in the next section, we study the linearized equation in the limit of a small proximity effect. As discussed in the introduction, in this limiting case the diode effect vanishes, but the anomalous phase $\phi_{0}$ , can be studied analytically. The anomalous current is a strong indication for the diode effect to appear. ## III Linearized case: The $\phi_{0}$-junction To get an understanding of the physics behind the new Usadel equation, Eq. (10), we focus first on the case of a weak proximity effect. In this case the anomalous parts of the GF are much smaller than the normal one, that is, $\text{Tr}(\tau_{1,2}G)\ll\text{Tr}(\tau_{3}G)$, and $G$ can be approximated by $G(i\omega_{n})\approx\begin{bmatrix}\text{sgn}(\omega_{n})&F\\\ \tilde{F}&-\text{sgn}(\omega_{n})\end{bmatrix}$, where $|F|,|\tilde{F}|\ll 1$. Using this approximation the Usadel equation reduces to a linear equation. We assume here that the system in Fig. 1a is infinite in $x$ direction. The superconductor is only absent in the region $(-\frac{L_{1}}{2},\frac{L_{1}}{2})$ and present everywhere outside this region. In this case Eq. (10) can be written in the three separate spatial regions: $\displaystyle\begin{cases}D\partial_{xx}F=2|\omega_{n}|F-\gamma_{0}f_{S0}e^{-i\frac{\phi}{2}}&x<-\frac{L_{1}}{2}\\\ D\partial_{xx}F=2|\omega_{n}|F&|x|<\frac{L_{1}}{2}\\\ \end{cases}$ (16) For $x\xrightarrow{}\pm\infty$ the Green’s function should commute with $\omega\tau_{3}+\check{G}_{S0}$. This implies that the pair potential is given by $\displaystyle\lim_{x\xrightarrow{}\pm\infty}F=\frac{\gamma_{0}f_{S0}}{2|\omega_{n}|}e^{\pm i\frac{\phi}{2}}.$ (17) These are exactly the same equations as for the conventional SNS junction. However, the equations which joint the solutions at $x=\pm\frac{L_{1}}{2}$, are different compared to the conventional SNS junction. Requiring continuity of both the Green’s function and the current through the junction yields $\displaystyle F\left(\pm\frac{L_{1}}{2}+0^{+}\right)=F\left(\pm\frac{L_{1}}{2}+0^{-}\right)$ (18) $\displaystyle\derivative{F}{x}\left(\pm\frac{L_{1}}{2}+0^{\pm}\right)+\frac{\gamma_{0}}{v_{F}}f_{S1}\text{sgn}(\omega_{n})e^{\pm i\frac{\phi}{2}}=\derivative{F}{x}\left(\pm\frac{L_{1}}{2}+0^{\mp}\right).$ (19) This expression differs from the expression for the SNS junction in the appearance of the $f_{S1}$-term on the right hand side of Eq. (19). The CPR following from these equations is $\displaystyle I(\phi)$ $\displaystyle=\sum_{n}\frac{1}{4}e^{-2\sqrt{\frac{2|\omega_{n}|}{D}}}\text{Im}((A_{n}-iB_{n})e^{i\frac{\phi}{2}})^{2}$ (20) $\displaystyle=\frac{1}{4}\sum_{n}e^{-2\sqrt{\frac{2|\omega_{n}|}{D}}}\left(A_{n}^{2}-B_{n}^{2}\sin{\phi}+2A_{n}B_{n}\cos{\phi}\right)$ (21) where $A_{n}$ and $B_{n}$ are real coefficients given by $A_{n}=\frac{\gamma_{0}f_{S0}}{2|\omega_{n}|}$ and $B_{n}=-i\gamma_{0}f_{S1}\sqrt{\frac{D\omega_{n}}{2}}\frac{1}{|\omega_{n}|}$. This implies that $\displaystyle\phi_{0}=\arctan 2\frac{\sum_{n=-\infty}^{\infty}A_{n}B_{n}e^{-2\sqrt{\frac{2|\omega_{n}|}{D}}}}{\sum_{n=-\infty}^{\infty}(A_{n}^{2}-B_{n}^{2})e^{-2\sqrt{\frac{2|\omega_{n}|}{D}}}}$ (22) The $\phi_{0}$-effect is largest if $A_{n}=B_{n}$ and small if $|\frac{B_{n}}{A_{n}}|$ is not close to 1. This ratio is given by $\displaystyle|\frac{B_{n}}{A_{n}}|$ $\displaystyle=\frac{\gamma_{0}|f_{S1}|\sqrt{\frac{D}{2|\omega_{n}|}}2|\omega_{n}|}{\gamma_{0}f_{S0}v_{F}}=\frac{|f_{S1}|}{f_{S0}}\frac{1}{v_{F}}\sqrt{2|\omega_{n}|D}$ (23) Recall that the diffusion constant is given by $D=v_{F}^{2}\tau$. This means that $\frac{1}{v_{F}}\sqrt{2|\omega_{n}|D}=O(\sqrt{|\omega_{n}|\tau})$, which is small in the diffusive regime. In fact, in the derivation of the Usadel equation, it is assumed that $\frac{1}{\tau}\gg|\Delta|$ and thus $\frac{1}{\tau}\gg|\omega_{n}|$ for every $n$ that contributes significantly to the current. This means the effect can only be large if $f_{S1}\gg f_{S0}$. This constraint can only be satisfied if, $h\gg\sqrt{\omega_{n}^{2}+\Delta^{2}}$ for all Matsubara frequencies that have a significant contribution to the critical current. However, in the setup discussed in this paper the condition $h\gg\Delta$ can not be realized, since the magnetisation is induced via the superconductor and a high magnetisation destroys the superconductivity. The $\phi_{0}$-effect is suppressed by a factor $\sqrt{|\omega_{n}|\tau}$ in the linearized case. Next the temperature dependence is discussed. Because the $\phi_{0}$-effect is small we can simplify the equation for $\phi_{0}$ to $\displaystyle\phi_{0}\approx 2\frac{\sum_{n=-\infty}^{\infty}A_{n}B_{n}e^{-2\sqrt{\frac{2|\omega_{n}|}{D}}}}{\sum_{n=-\infty}^{\infty}(A_{n}^{2})e^{-2\sqrt{\frac{2|\omega_{n}|}{D}}}}.$ (24) Now, consider the behaviour at low temperatures. Since the triplet component $f_{S1}$ is odd in frequency, whereas the singlet component $f_{S0}$ is even in frequency, $|\frac{B_{n}}{A_{n}}|$ is small for small Matsubara frequencies. As the temperature is decreased these terms become more and more dominant in the sum. Thus, at low temperatures, the $\phi_{0}$-effect is increases with increasing temperature. On the other hand, for large Matsubara frequencies the ratio between triplet and singlet components $|\frac{f_{s1}}{f_{s0}}|=O(\frac{h}{|\omega_{n}|})$. This means that at high temperatures the $\phi_{0}$-effect decreases. Therefore, the $\phi_{0}$-effect must be non-monotonic as a function of temperature, it attains a maximum. Moreover, since $\sqrt{\tau}$ is an $\omega_{n}$-independent prefactor it can not determine the maximum. The temperature at which the maximum is attained is determined by only two dimensionless quantities, $\frac{\Delta}{E_{\text{Th}}}$ and $\frac{h}{\Delta}$. An interesting limit is the limit in which $\sqrt{\frac{2\pi T}{D}}L_{1}\ll 1$ and $h\ll\Delta$ so that the exponential suppression can be ignored to first order in $\sqrt{\frac{2\pi T}{D}}L_{1}$. The following expression is obtained 111Strictly speaking, the linearized case leads to a divergency as T goes to 0. We therefore replace $\Delta/|\omega_{n}|$ by $\Delta/\sqrt{\omega^{2}+\gamma_{0}^{2}}$, based on the non-linearized case: $\displaystyle\phi_{0}\approx$ $\displaystyle\frac{\sum_{n=0}^{\infty}A_{n}B_{n}}{\sum_{n=0}^{\infty}A_{n}^{2}}=h\sqrt{\tau}\frac{\sum_{n=0}^{\infty}\frac{1}{(\omega_{n}^{2}+\Delta^{2})^{2}\sqrt{\omega_{n}}}}{\sum_{n=0}^{\infty}\frac{1}{\omega_{n}^{4}+\Delta^{2}\omega_{n}^{2}}}$ (25) The multiplication of the sum with $\sqrt{\tau}$ signals the dirty limit suppression of the $\phi_{0}$-effect. The resulting expression is evaluated numerically as a function of temperature. Numerical evaluation confirmed the non-monotonicity, see Fig. 2. Figure 2: The $\phi_{0}$-effect as a function of temperature as calculated using Eq. (25). The $\phi_{0}$-effect is suppressed at low temperatures. The $\phi_{0}$-effect is given in units of the small quantity $h\sqrt{\tau/\Delta}$. In the following sections we discuss the full non-linear equation, and we show that the diode-effect is non-monotonic with temperature for short junctions. ## IV Non-linearized case: The superconducting diode effect To investigate the diode effect one needs to go beyond the linear approach and solve numerically the Usadel equation. In this section we present our numerical results for the supercurrent in the FIS-TI-FIS junction. As a first step it is convenient to write the Usadel equation, Eq. (10), in dimensionless form, normalising $x$ by the total length of the junction. The obtained equation is $\displaystyle\hat{\nabla}(\check{G}\hat{\nabla}\check{G})=[\frac{\omega_{n}}{E_{\text{Th}}}\check{\tau}_{3}+\frac{\Gamma(x)}{2E_{\text{Th}}}\check{G}_{S0},\check{G}],$ (26) $\displaystyle\hat{\nabla}=\derivative{x}+\frac{\Gamma(x)L}{2v_{F}}\tilde{b}_{y}[\check{G}_{S1},\cdot]$ (27) Thus, all energies are given in units of the Thouless energy $E_{\text{Th}}=\frac{D}{L^{2}}$, whereas lengths are given in units of $L$. The strength of the proximity effect is described by the dimensionless parameter $\frac{\gamma_{0}L}{v_{F}}=\frac{\gamma_{0}}{E_{\text{Th}}}\frac{l}{L}$, where $l=v_{F}\tau$ is the mean free path, which in the diffusive limit must be the shortest length involved in the problem. This puts a constraint on the magnitude of the new dimensionless quantity, it must be much smaller than $\frac{\gamma_{0}}{E_{\text{Th}}}$. However, without this term the equations reduce to the equations for the SNS junction, which is known not to have a diode effect [28], it has no time-reversal symmetry breaking. Thus, if the new quantity is very small, the diode effect is very small. Therefore, $\frac{\gamma_{0}}{E_{\text{Th}}}$ must be chosen large, that is, the contact between the superconductor and the TI must be good to have a large $\gamma_{0}$. To solve the non-linearized Usadel equation, Eq. (10), the so-called Riccati parametrisation is used, $\check{G}=\frac{1}{1+\bar{\gamma}\tilde{\gamma}}\begin{bmatrix}1-\bar{\gamma}\tilde{\gamma}&2\gamma\\\ 2\tilde{\gamma}&-1+\bar{\gamma}\tilde{\gamma}\end{bmatrix},$ (28) where $\bar{\gamma}$ and $\tilde{\gamma}$ are the Riccati parameters. In principle, the pair potential $\Delta$ has to be determined self- consistently since it is suppressed by the exchange field [61]. In the numerical calculations we choose values of the exchange field smaller than $\frac{h}{\Delta_{0}}=\frac{2}{5}$. Other parameters are set as follows: $\frac{\gamma_{0}}{E_{\text{Th}}}=25$, whereas $\frac{L_{1}}{L}$ is chosen to be $\frac{1}{10}$ and $\frac{l}{L}=0.08\ll 1$. Fig. 1b shows our numerical results for the CPR obtained from the non- linearized Eq. (10). One can see a finite current value at $\phi=0$ associated with the appearance of the anomalous phase $\phi_{0}$. Moreover, even though small, there is a difference in the absolute value of the maximum and minimum of the current. This asymmetry reflects the diode effect. By increasing the temperature both the current at zero phase and the critical current decrease. We now study the temperature dependence of the diode effect for different exchange fields and sizes of the junction. The numerical results for the diode efficiency are shown in Fig. 3. Interestingly, we find a non-monotonic behaviour with a maximum efficiency at a finite temperature, $T_{d}$. It is important to notice, that by computing $\eta$ in Fig. 3, the self-consistency of the pair potential is ignored, because of the reduction of computational costs. However, we verify that for all values of $h$ considered here, the self-consistency leaves the magnitude of the gap almost unchanged for temperatures of the order of $T=T_{d}$. The critical temperatures for all cases shown in Fig. 3a lie right of the dashed vertical line. If the exchange field is increased, Fig. 3a, the diode efficiency becomes larger. $\eta$ increases approximately linearly with $h$. The temperature at which the diode efficiency is maximal is almost independent of the exchange field, with $T_{d}\approx 0.18$. We also investigate the influence of the distance between the leads, $L_{1}$ on $\eta$, see Fig. 3b. Specifically, $E_{Th1}=D/L_{1}^{2}$ is varied, whereas the quantities $\frac{l}{L}$ and $\frac{L_{1}}{L}$ are held constant. As the Thouless energy is decreased, the diode efficiency decreases. Moreover, the temperature $T_{d}$ at which the diode effect is maximal decreases with the length of the junction. For enough long junctions the dependence of $\eta$ on temperature becomes monotonic. (a) (b) Figure 3: The temperature dependence of the diode efficiency $\eta$ for different values of the exchange field strength $h$ (a), and Thouless energy $E_{\text{Th}1}=\frac{D}{L_{1}^{2}}$ of the TI part (b). The critical temperature is for all magnitudes of the exchange field considered here larger than $0.9T_{c}$, highlighted by the black dotted line. So far, we have considered disordered systems with low transparent S/FI interfaces. One can, though, increase the $\phi_{0}$ and diode effects by relaxing these conditions. On the one hand, our analytical results in section III indicate that the $\phi_{0}$-effect can be increased by increasing $\tau$, see for example Eq. (25). We also verified numerically that the diode effect increases if the degree of disorder decreases. On the other hand, we also investigated the effect of increasing the ratio $\frac{L_{1}}{L}$. Our numerical calculations demonstrate that whereas $I(\phi=0)/I_{c}$ increases, the diode effect decreases with increasing $\frac{L_{1}}{L}$. This can be explained as follows. By increasing the distance $L_{1}$ between the electrodes the CPR becomes more sinus-like. To be precise, when increasing $\frac{L_{1}}{L}$ from $\frac{1}{10}$ to $\frac{1}{2}$, the ratio between the magnitudes of the second and first harmonics decreases from $\approx\frac{1}{6}$ to $\approx\frac{1}{10}$. As we discussed before, besides the breaking of time-reversal and inversion symmetries, the diode effect relies crucially on the contribution of higher harmonics to the CPR. A way to increase the higher harmonics contribution, is to increase the coupling between the superconducting correlations from the left and right electrodes. This can be achieved by increasing the transparency of the FIS/TI interfaces, as shown in Fig. 4. Figure 4: The diode efficiency as a function of temperature for different values of $\gamma_{0}$. Finally, another way to increase the diode effect is by increasing the exchange field, as shown in Fig. 3. In our junction, however, the value of $h$ is limited by the critical field of the superconductor. To increase the strength of the exchange field without suppressing superconductivity in the S electrodes one could add an additional ferromagnetic insulating layers directly on top of the TI between the two superconductors, similar to the situation investigated in Refs. [55, 62, 63]. In that case the exchange field can be larger than the superconducting gap and the diode effect may increase. ## V Conclusions We present a study of the $\phi_{0}$ and diode effects in a FIS-TI-FIS Josephson junction. Though disorder tend to suppress them [40], we found, even in the diffusive limit, sizable effects without applying any external field. We found that by increasing the FIS/TI interface transparency and the magnetic field one can increase the diode effect. For short junctions the diode effect is non-monotonic as a function of temperature. By increasing the distance between the electrodes the $\phi_{0}$-effect is enhanced, however the diode effect is suppressed due to the loss of higher harmonics. From the point of view of materials the proposed structure can be fabricated with well studied material combinations. On the one hand the use of topological insulators in Josephson junctions is well understood [64, 65, 66, 67, 68, 69, 70, 71, 72]. On the other hand, spin-split superconductivity has been measured in several experiments on ferromagnetic insulator/superconductor bilayers, as for example EuS/Al structures [73, 74, 51, 54]. Moreover, good interfaces between TI and FI has been reported in Ref. [75]. ## Acknowledgements We thank Stefan Ilic for useful discussions. We acknowledge financial support from Spanish AEI through project PID2020-114252GB-I00 (SPIRIT). FSB acknowledges financial support from the European Union’s Horizon 2020 Research and Innovation Framework Programme under Grant No. 800923 (SUPERTED), the A. v. Humboldt Foundation, and the Basque Government through grant IT-1591-22. ## References * Zazunov _et al._ [2009] A. Zazunov, R. Egger, T. Jonckheere, and T. Martin, Anomalous josephson current through a spin-orbit coupled quantum dot, Physical review letters 103, 147004 (2009). * Tanaka _et al._ [2009] Y. Tanaka, T. Yokoyama, and N. Nagaosa, Manipulation of the majorana fermion, andreev reflection, and josephson current on topological insulators, Physical review letters 103, 107002 (2009). * Liu and Chan [2010] J.-F. Liu and K. S. Chan, Anomalous josephson current through a ferromagnetic trilayer junction, Physical Review B 82, 184533 (2010). * Alidoust and Linder [2013] M. Alidoust and J. Linder, $\varphi$-state and inverted fraunhofer pattern in nonaligned josephson junctions, Physical Review B 87, 060503 (2013). * Alidoust [2020] M. Alidoust, Critical supercurrent and $\varphi$ 0 state for probing a persistent spin helix, Physical Review B 101, 155123 (2020). * Alidoust _et al._ [2021] M. Alidoust, C. Shen, and I. Žutić, Cubic spin-orbit coupling and anomalous josephson effect in planar junctions, Physical Review B 103, L060503 (2021). * Szombati _et al._ [2016] D. Szombati, S. Nadj-Perge, D. Car, S. Plissard, E. Bakkers, and L. Kouwenhoven, Josephson $\phi$0-junction in nanowire quantum dots, Nature Physics 12, 568 (2016). * Brunetti _et al._ [2013] A. Brunetti, A. Zazunov, A. Kundu, and R. Egger, Anomalous josephson current, incipient time-reversal symmetry breaking, and majorana bound states in interacting multilevel dots, Physical Review B 88, 144515 (2013). * Campagnano _et al._ [2015] G. Campagnano, P. Lucignano, D. Giuliano, and A. Tagliacozzo, Spin–orbit coupling and anomalous josephson effect in nanowires, Journal of Physics: Condensed Matter 27, 205301 (2015). * Lu _et al._ [2015] B. Lu, K. Yada, A. A. Golubov, and Y. Tanaka, Anomalous josephson effect in d-wave superconductor junctions on a topological insulator surface, Physical Review B 92, 100503 (2015). * Wakatsuki _et al._ [2017] R. Wakatsuki, Y. Saito, S. Hoshino, Y. M. Itahashi, T. Ideue, M. Ezawa, Y. Iwasa, and N. Nagaosa, Nonreciprocal charge transport in noncentrosymmetric superconductors, Science advances 3, e1602390 (2017). * Wakatsuki and Nagaosa [2018] R. Wakatsuki and N. Nagaosa, Nonreciprocal current in noncentrosymmetric Rashba superconductors, Physical Review Letters 121, 026601 (2018). * Ando _et al._ [2020] F. Ando, Y. Miyasaka, T. Li, J. Ishizuka, T. Arakawa, Y. Shiota, T. Moriyama, Y. Yanase, and T. Ono, Observation of superconducting diode effect, Nature 584, 373 (2020). * Pal _et al._ [2022] B. Pal, A. Chakraborty, P. K. Sivakumar, M. Davydova, A. K. Gopi, A. K. Pandeya, J. A. Krieger, Y. Zhang, S. Ju, N. Yuan, _et al._ , Josephson diode effect from cooper pair momentum in a topological semimetal, Nature Physics , 1 (2022). * Baumgartner _et al._ [2022] C. Baumgartner, L. Fuchs, A. Costa, S. Reinhardt, S. Gronin, G. C. Gardner, T. Lindemann, M. J. Manfra, P. E. Faria Junior, D. Kochan, _et al._ , Supercurrent rectification and magnetochiral effects in symmetric Josephson junctions, Nature Nanotechnology 17, 39 (2022). * Hou _et al._ [2022] Y. Hou, F. Nichele, H. Chi, A. Lodesani, Y. Wu, M. F. Ritter, D. Z. Haxell, M. Davydova, S. Ilić, F. S. Bergeret, _et al._ , Ubiquitous superconducting diode effect in superconductor thin films, arXiv preprint arXiv:2205.09276 (2022). * Lin _et al._ [2022] J.-X. Lin, P. Siriviboon, H. D. Scammell, S. Liu, D. Rhodes, K. Watanabe, T. Taniguchi, J. Hone, M. S. Scheurer, and J. Li, Zero-field superconducting diode effect in small-twist-angle trilayer graphene, Nature Physics , 1 (2022). * Lyu _et al._ [2021] Y.-Y. Lyu, J. Jiang, Y.-L. Wang, Z.-L. Xiao, S. Dong, Q.-H. Chen, M. V. Milošević, H. Wang, R. Divan, J. E. Pearson, _et al._ , Superconducting diode effect via conformal-mapped nanoholes, Nature Communications 12, 1 (2021). * Wu _et al._ [2021] H. Wu, Y. Wang, P. K. Sivakumar, C. Pasco, S. S. Parkin, Y.-J. Zeng, T. McQueen, and M. N. Ali, Realization of the field-free Josephson diode, arXiv preprint arXiv:2103.15809 (2021). * Silaev _et al._ [2014] M. Silaev, A. Y. Aladyshkin, M. Silaeva, and A. Aladyshkina, The diode effect induced by domain-wall superconductivity, Journal of Physics: Condensed Matter 26, 095702 (2014). * Yuan and Fu [2022] N. F. Yuan and L. Fu, Supercurrent diode effect and finite-momentum superconductors, Proceedings of the National Academy of Sciences 119, e2119548119 (2022). * Tanaka _et al._ [2022] Y. Tanaka, B. Lu, and N. Nagaosa, Theory of diode effect in d-wave superconductor junctions on the surface of topological insulator, arXiv preprint arXiv:2205.13177 (2022). * Pal and Benjamin [2019] S. Pal and C. Benjamin, Quantized josephson phase battery, EPL (Europhysics Letters) 126, 57002 (2019). * Kopasov _et al._ [2021] A. Kopasov, A. Kutlin, and A. Mel’nikov, Geometry controlled superconducting diode and anomalous josephson effect triggered by the topological phase transition in curved proximitized nanowires, Physical Review B 103, 144520 (2021). * Margaris _et al._ [2010] I. Margaris, V. Paltoglou, and N. Flytzanis, Zero phase difference supercurrent in ferromagnetic josephson junctions, Journal of Physics: Condensed Matter 22, 445701 (2010). * Chen _et al._ [2018] C.-Z. Chen, J. J. He, M. N. Ali, G.-H. Lee, K. C. Fong, and K. T. Law, Asymmetric josephson effect in inversion symmetry breaking topological materials, Physical Review B 98, 075430 (2018). * He _et al._ [2022] J. J. He, Y. Tanaka, and N. Nagaosa, A phenomenological theory of superconductor diodes, New Journal of Physics 24, 053014 (2022). * Golubov _et al._ [2004] A. A. Golubov, M. Y. Kupriyanov, and E. Il’Ichev, The current-phase relation in Josephson junctions, Reviews of Modern Physics 76, 411 (2004). * Silaev _et al._ [2017] M. Silaev, I. Tokatly, and F. Bergeret, Anomalous current in diffusive ferromagnetic Josephson junctions, Physical Review B 95, 184508 (2017). * Zhang _et al._ [2021] Y. Zhang, Y. Gu, J. Hu, and K. Jiang, General theory of Josephson diodes, arXiv preprint arXiv:2112.08901 (2021). * Hasan and Kane [2010] M. Z. Hasan and C. L. Kane, Colloquium: topological insulators, Reviews of Modern Physics 82, 3045 (2010). * Dolcini _et al._ [2015] F. Dolcini, M. Houzet, and J. S. Meyer, Topological Josephson $\phi$ 0 junctions, Physical Review B 92, 035428 (2015). * Karabassov _et al._ [2022] T. Karabassov, I. Bobkova, A. Golubov, and A. Vasenko, Hybrid helical state and superconducting diode effect in s/f/ti heterostructures, arXiv preprint arXiv:2203.15608 (2022). * Reynoso _et al._ [2008] A. Reynoso, G. Usaj, C. Balseiro, D. Feinberg, and M. Avignon, Anomalous Josephson current in junctions with spin polarizing quantum point contacts, Physical Review Letters 101, 107001 (2008). * Buzdin [2008] A. Buzdin, Direct coupling between magnetism and superconducting current in the Josephson $\varphi$ 0 junction, Physical Review Letters 101, 107005 (2008). * Yokoyama _et al._ [2014] T. Yokoyama, M. Eto, and Y. V. Nazarov, Anomalous Josephson effect induced by spin-orbit interaction and zeeman effect in semiconductor nanowires, Physical Review B 89, 195407 (2014). * Konschelle _et al._ [2015] F. Konschelle, I. V. Tokatly, and F. S. Bergeret, Theory of the spin-galvanic effect and the anomalous phase shift $\varphi$ 0 in superconductors and Josephson junctions with intrinsic spin-orbit coupling, Physical Review B 92, 125443 (2015). * Bergeret and Tokatly [2015] F. Bergeret and I. Tokatly, Theory of diffusive $\varphi$0 Josephson junctions in the presence of spin-orbit coupling, EPL (Europhysics Letters) 110, 57005 (2015). * Daido _et al._ [2022] A. Daido, Y. Ikeda, and Y. Yanase, Intrinsic superconducting diode effect, Physical Review Letters 128, 037001 (2022). * Ilić and Bergeret [2021] S. Ilić and F. S. Bergeret, Effect of disorder on superconducting diodes, arXiv preprint arXiv:2108.00209 (2021). * Jeon _et al._ [2022] K.-R. Jeon, J.-K. Kim, J. Yeon, J.-C. Jeon, H. Han, A. Cottet, T. Kontos, and S. S. Parkin, Zero-field polarity-reversible Josephson supercurrent diodes enabled by a proximity-magnetized Pt barrier, Nature Materials , 1 (2022). * Souto _et al._ [2022] R. S. Souto, M. Leijnse, and C. Schrade, The Josephson diode effect in supercurrent interferometers, arXiv preprint arXiv:2205.04469 (2022). * Hao _et al._ [1991] X. Hao, J. S. Moodera, and R. Meservey, Thin-film superconductor in an exchange field, Physical Review Letters 67, 1342 (1991). * Meservey and Tedrow [1994] R. Meservey and P. Tedrow, Spin-polarized electron tunneling, Physics Reports 238, 173 (1994). * Tokuyasu _et al._ [1988] T. Tokuyasu, J. A. Sauls, and D. Rainer, Proximity effect of a ferromagnetic insulator in contact with a superconductor, Physical Review B 38, 8823 (1988). * Tanaka and Kashiwaya [1997] Y. Tanaka and S. Kashiwaya, Theory of Josephson effect in superconductor-ferromagnetic-insulator-superconductor junction, Physica C: Superconductivity 274, 357 (1997). * Heikkilä _et al._ [2019] T. T. Heikkilä, M. Silaev, P. Virtanen, and F. S. Bergeret, Thermal, electric and spin transport in superconductor/ferromagnetic-insulator structures, Progress in Surface Science 94, 100540 (2019). * Bergeret _et al._ [2018] F. S. Bergeret, M. Silaev, P. Virtanen, and T. T. Heikkilä, Colloquium: Nonequilibrium effects in superconductors with a spin-splitting field, Reviews of Modern Physics 90, 041001 (2018). * Rouco _et al._ [2019] M. Rouco, S. Chakraborty, F. Aikebaier, V. N. Golovach, E. Strambini, J. S. Moodera, F. Giazotto, T. T. Heikkilä, and F. S. Bergeret, Charge transport through spin-polarized tunnel junction between two spin-split superconductors, Physical Review B 100, 184501 (2019). * Liu _et al._ [2019] Y. Liu, S. Vaitiekenas, S. Martí-Sánchez, C. Koch, S. Hart, Z. Cui, T. Kanne, S. A. Khan, R. Tanta, S. Upadhyay, _et al._ , Semiconductor–ferromagnetic insulator–superconductor nanowires: Stray field and exchange field, Nano Letters 20, 456 (2019). * Strambini _et al._ [2017] E. Strambini, V. Golovach, G. De Simoni, J. Moodera, F. Bergeret, and F. Giazotto, Revealing the magnetic proximity effect in EuS/Al bilayers through superconducting tunneling spectroscopy, Physical Review Materials 1, 054402 (2017). * Ojajärvi _et al._ [2021] R. Ojajärvi, T. T. Heikkilä, P. Virtanen, and M. Silaev, Giant enhancement to spin battery effect in superconductor/ferromagnetic insulator systems, Physical Review B 103, 224524 (2021). * Hijano _et al._ [2022] A. Hijano, V. N. Golovach, and F. S. Bergeret, Quasiparticle density of states and triplet correlations in superconductor/ferromagnetic-insulator structures across a sharp domain wall, Physical Review B 105, 174507 (2022). * Hijano _et al._ [2021] A. Hijano, S. Ilić, M. Rouco, C. González-Orellana, M. Ilyn, C. Rogero, P. Virtanen, T. Heikkilä, S. Khorshidian, M. Spies, _et al._ , Coexistence of superconductivity and spin-splitting fields in superconductor/ferromagnetic insulator bilayers of arbitrary thickness, Physical Review Research 3, 023131 (2021). * Zyuzin _et al._ [2016] A. Zyuzin, M. Alidoust, and D. Loss, Josephson junction through a disordered topological insulator with helical magnetization, Physical Review B 93, 214502 (2016). * Bobkova _et al._ [2016] I. Bobkova, A. Bobkov, A. A. Zyuzin, and M. Alidoust, Magnetoelectrics in disordered topological insulator Josephson junctions, Physical Review B 94, 134506 (2016). * Tkachov [2013] G. Tkachov, Suppression of surface p-wave superconductivity in disordered topological insulators, Physical Review B 87, 245422 (2013). * Bobkova and Bobkov [2017] I. Bobkova and A. Bobkov, Electrically controllable spin filtering based on superconducting helical states, Physical Review B 96, 224505 (2017). * Bergeret _et al._ [2005] F. Bergeret, A. F. Volkov, and K. B. Efetov, Odd triplet superconductivity and related phenomena in superconductor-ferromagnet structures, Reviews of Modern Physics 77, 1321 (2005). * Note [1] Strictly speaking, the linearized case leads to a divergency as T goes to 0. We therefore replace $\Delta/|\omega_{n}|$ by $\Delta/\sqrt{\omega^{2}+\gamma_{0}^{2}}$, based on the non-linearized case. * Buzdin [2005] A. I. Buzdin, Proximity effects in superconductor-ferromagnet heterostructures, Reviews of Modern Physics 77, 935 (2005). * Rabinovich _et al._ [2020] D. Rabinovich, I. Bobkova, and A. Bobkov, Electrical response of superconductor/ferromagnet/topological-insulator/superconductor junctions to magnetic texture dynamics, Physical Review B 101, 054517 (2020). * Bobkova _et al._ [2020] I. Bobkova, A. Bobkov, I. Rahmonov, A. Mazanik, K. Sengupta, and Y. M. Shukrinov, Magnetization reversal in superconductor/insulating ferromagnet/superconductor josephson junctions on a three-dimensional topological insulator, Physical Review B 102, 134505 (2020). * Veldhorst _et al._ [2012] M. Veldhorst, M. Snelder, M. Hoek, T. Gang, V. Guduru, X. Wang, U. Zeitler, W. G. van der Wiel, A. Golubov, H. Hilgenkamp, _et al._ , Josephson supercurrent through a topological insulator surface state, Nature Materials 11, 417 (2012). * Veldhorst _et al._ [2013] M. Veldhorst, M. Snelder, M. Hoek, C. Molenaar, D. P. Leusink, A. A. Golubov, H. Hilgenkamp, and A. Brinkman, Magnetotransport and induced superconductivity in Bi based three-dimensional topological insulators, Physica Status Solidi (RRL)–Rapid Research Letters 7, 26 (2013). * Snelder _et al._ [2014] M. Snelder, C. Molenaar, Y. Pan, D. Wu, Y. Huang, A. de Visser, A. Golubov, W. van der Wiel, H. Hilgenkamp, M. Golden, _et al._ , Josephson supercurrent in a topological insulator without a bulk shunt, Superconductor Science and Technology 27, 104001 (2014). * Maier _et al._ [2012] L. Maier, J. B. Oostinga, D. Knott, C. Brüne, P. Virtanen, G. Tkachov, E. M. Hankiewicz, C. Gould, H. Buhmann, and L. W. Molenkamp, Induced superconductivity in the three-dimensional topological insulator HgTe, Physical Review Letters 109, 186806 (2012). * Sochnikov _et al._ [2015] I. Sochnikov, L. Maier, C. A. Watson, J. R. Kirtley, C. Gould, G. Tkachov, E. M. Hankiewicz, C. Brüne, H. Buhmann, L. W. Molenkamp, _et al._ , Nonsinusoidal current-phase relationship in Josephson junctions from the 3d topological insulator HgTe, Physical Review Letters 114, 066801 (2015). * Wiedenmann _et al._ [2016] J. Wiedenmann, E. Bocquillon, R. S. Deacon, S. Hartinger, O. Herrmann, T. M. Klapwijk, L. Maier, C. Ames, C. Brüne, C. Gould, _et al._ , 4$\pi$-periodic Josephson supercurrent in HgTe-based topological Josephson junctions, Nature Communications 7, 1 (2016). * Oostinga _et al._ [2013] J. B. Oostinga, L. Maier, P. Schüffelgen, D. Knott, C. Ames, C. Brüne, G. Tkachov, H. Buhmann, and L. W. Molenkamp, Josephson supercurrent through the topological surface states of strained bulk HgTe, Physical Review X 3, 021007 (2013). * Wang _et al._ [2018] A.-Q. Wang, C.-Z. Li, C. Li, Z.-M. Liao, A. Brinkman, and D.-P. Yu, 4$\pi$-periodic supercurrent from surface states in cd 3 as 2 nanowire-based Josephson junctions, Physical Review Letters 121, 237701 (2018). * Mandal _et al._ [2022] P. Mandal, N. Taufertshöfer, L. Lunczer, M. P. Stehno, C. Gould, and L. Molenkamp, Finite field transport response of a dilute magnetic topological insulator-based Josephson junction, Nano Letters (2022). * Moodera _et al._ [1988] J. Moodera, X. Hao, G. Gibson, and R. Meservey, Electron-spin polarization in tunnel junctions in zero applied field with ferromagnetic EuS barriers, Physical Review Letters 61, 637 (1988). * Xiong _et al._ [2011] Y. Xiong, S. Stadler, P. Adams, and G. Catelani, Spin-resolved tunneling studies of the exchange field in EuS/Al bilayers, Physical Review Letters 106, 247001 (2011). * Wei _et al._ [2013] P. Wei, F. Katmis, B. A. Assaf, H. Steinberg, P. Jarillo-Herrero, D. Heiman, and J. S. Moodera, Exchange-coupling-induced symmetry breaking in topological insulators, Physical Review Letters 110, 186807 (2013).
# One-class Recommendation Systems with the Hinge Pairwise Distance Loss and Orthogonal Representations Ramin Raziperchikolaei Young-joo Chung ###### Abstract In one-class recommendation systems, the goal is to learn a model from a small set of interacted users and items and then identify the positively-related user-item pairs among a large number of pairs with unknown interactions. Most previous loss functions rely on dissimilar pairs of users and items, which are selected from the ones with unknown interactions, to obtain better prediction performance. This strategy introduces several challenges such as increasing training time and hurting the performance by picking ”similar pairs with the unknown interactions” as dissimilar pairs. In this paper, the goal is to only use the similar set to train the models. We point out three trivial solutions that the models converge to when they are trained only on similar pairs: collapsed, partially collapsed, and shrinking solutions. We propose two terms that can be added to the objective functions in the literature to avoid these solutions. The first one is a hinge pairwise distance loss that avoids the shrinking and collapsed solutions by keeping the average pairwise distance of all the representations greater than a margin. The second one is an orthogonality term that minimizes the correlation between the dimensions of the representations and avoids the partially collapsed solution. We conduct experiments on a variety of tasks on public and real-world datasets. The results show that our approach using only similar pairs outperforms state-of- the-art methods using similar pairs and a large number of dissimilar pairs. ## 1 Introduction Feedback in recommendation systems (RSs) can be explicit or implicit. In the case of explicit feedback [20, 24, 35, 12], users provide rating values after interacting with the items, which shows how satisfied they are. In the case of implicit feedback [20, 11, 15], which is a common scenario in the real world, we only know if the user interacted with an item or not (such as clicked, purchased, etc.). The goal of one-class recommendation systems is to solve the implicit feedback prediction problem. It’s called a one-class problem since a ”no interaction” between a user and an item in the training set does not necessarily mean the user does not like that item. It just means we do not have enough information about their interaction. That’s because the set of the items in RSs is huge and the users can’t see all the items, i.e., users can only see a small subset of the items and then interact with a few of them. There are three main steps in learning a RS model to predict the implicit feedback. The first step is to learn user and item representations. This can be done by learning user and item embedding matrices from their IDs [24, 35, 12, 8, 9], learning user and item multi-layer perceptrons (MLPs) from the interaction vectors [21, 30, 5] or side information [23, 34, 4], or learning graph neural networks from the bipartite user-item graph [25, 31, 32, 10]. The second step is to model the interaction score from the user and item representations. The common functions are dot product [11, 15, 24, 12, 4, 32, 10, 18], cosine similarity [30, 31], and neural networks over the concatenated user and item representations [8, 9, 5, 17, 16]. The third step is to optimize a loss function, which gets smaller values as the model gives a larger score to the similar pairs of the user and items compared to the dissimilar ones. Different types of loss functions have been used in the literature. Mean squared error (MSE) loss [11, 15, 24, 12, 4, 32, 17, 18] and Binary cross-entropy (BCE) [30, 8, 5] directly minimize the difference between the predicted and the actual scores. The Bayesian personalized rank (BPR) loss [19, 9, 10] tries to make the interaction score of the similar pairs greater than the dissimilar ones, instead of directly mapping them to the actual scores. The contrastive learning loss tries to put representations of the similar pairs close to each other and put the dissimilar ones far away [18, 14, 29]. All the above loss functions need both similar and dissimilar pairs of the users and items to learn a model. If we train these loss functions only using the similar pairs, we get a collapsed solution: all representations will be mapped to the same point in the latent space and the model predicts the same interaction score for all the pairs. The performance of the collapsed solution is as bad as assigning random representations to the users and items. So the dissimilar sets are essential in training RS models to avoid the collapsed solution. In one-class recommendation systems, we only have access to the implicit (known) interactions, and the rest of the interactions are unknown. To create a dissimilar set of users and items, the common approach is to randomly select a set of user and item pairs with the unknown interactions and consider them dissimilar [30, 8, 5, 4, 32, 18]. Another strategy is to find out the hard- negatives: the pairs with the unknown interactions that the model has difficulty with classifying as dissimilar [33, 2, 3]. Creating dissimilar pairs from unknown ones is problematic for two main reasons. First, as we show in experiments, we might need a large number of dissimilar pairs to achieve reasonable results, which makes the training slow. Second, using a large number of dissimilar pairs increases the chance of converting a ”similar pair with an unknown interaction” to a dissimilar pair, which hurts the performance. This issue is more severe in the case of the hard negative approach, since ”similar pairs with unknown interactions” are by definition difficult to be classified as dissimilar, and will be mistakenly taken as hard negatives. The main goal of this paper is to propose a new objective function that avoids the collapsed solution without using the dissimilar pairs. As we define it mathematically in Section 3, at the collapsed solution, the average pairwise distance between all the user-user, item-item, and user-item representations is $0$. To avoid the collapsed solution, we propose a hinge pairwise distance loss which keeps the pairwise distance above a margin. We show that this new loss function alone is not enough to generate high-quality representations, as the solution might converge to a ”partially collapsed solution”, where all users and items converge to exactly two unique representations. To avoid this solution, which performs as poorly as the collapsed solution, we propose to minimize the correlation between the dimensions of representations by making them orthogonal. We conduct extensive experiments on several real-world and public datasets in both warm-start and cold-start settings. The experimental results in Section 4 show that our approach, which only uses similar pairs, has a better prediction performance than state-of-the-art methods, which use both similar and dissimilar pairs. #### Notations. We denote the sparse interaction matrix by $\mathbf{R}\in\mathbb{R}^{m\times n}$, where $m$ and $n$ are the number of users and items, respectively, $R_{jk}>0$ is the interaction value of the user $j$ on the item $k$, and $R_{jk}=0$ means the interaction is unknown. The goal is to predict the unknown interactions in $\mathbf{R}$. The $i$th row of a matrix $\mathbf{H}$ is shown by $\mathbf{H}_{i,:}$ and the $j$th column is shown by $\mathbf{H}_{:,j}$. The $d$-dimensional representations of the all users and all items are denoted by $\mathbf{Z}^{u}\in\mathbb{R}^{m\times d}$ and $\mathbf{Z}^{i}\in\mathbb{R}^{n\times d}$, respectively. The representation of the $j$th user and $k$th item are denoted by $\mathbf{z}_{j}^{u}=\mathbf{Z}_{j,:}^{u}$ and $\mathbf{z}_{k}^{i}=\mathbf{Z}_{k,:}^{i}$, respectively. ## 2 Related works Since our main contribution is proposing a new objective function, we review the different loss functions used in the RS literature. Most of these loss functions are defined based on the predicted user-item interactions from the representations. The most common mappings from representations to predicted interactions are dot product [11, 15, 24, 12, 4, 32, 10, 18], cosine similarity [30, 31], and neural networks [8, 9, 5, 17, 16]: $\text{dot product:}\hat{R}_{jk}=(\mathbf{z}_{j}^{u})^{T}\mathbf{z}_{k}^{i},\quad\text{cosine:}\hat{R}_{jk}=\frac{(\mathbf{z}_{j}^{u})^{T}\mathbf{z}_{k}^{i}}{||\mathbf{z}_{j}^{u}||||\mathbf{z}_{k}^{i}||},\quad\text{NNs:}\hat{R}_{jk}=h([\mathbf{z}_{j}^{u},\mathbf{z}_{k}^{i}])$ (1) where $[\cdot,\cdot]$ merges the two vectors and $h()$ is a neural network. In [8, 5],$[\cdot,\cdot]$ is concatenation and $h()$ is an MLP. In [9], $[\cdot,\cdot]$ is outer product and $h()$ is a convolutional network. To define the loss functions in RSs, previous works use sets of similar and dissimilar pairs of the users and items. The similar set $S^{+}$ contains all the users and items that interacted with each other, i.e., $(j,k)\in S^{+}$ if $R_{jk}=1$. The dissimilar set $S^{-}$ contains a subset of the users and items with unknown interactions, i.e., $R_{jk}=0$. Two popular loss functions in the literature are mean squared error (MSE) [11, 15, 24, 12, 4, 32, 17, 18] and binary cross-entropy (BCE) [30, 8, 5], which directly minimize the difference between the predicted and actual interactions: $\displaystyle l_{\text{BCE}}(\mathbf{z}_{j}^{u},\mathbf{z}_{k}^{i})=-(R_{jk}\log\hat{R}_{jk}+(1-R_{jk})\log(1-\hat{R}_{jk})),$ $\displaystyle l_{\text{MSE}}(\mathbf{z}_{j}^{u},\mathbf{z}_{k}^{i})=(\hat{R}_{jk}-R_{jk})^{2},$ (2) where $j,k\in S^{+}\cup S^{-}$. On the other hand, the BPR loss [19, 9, 10] is defined based on the difference between the predicted interactions of the similar and dissimilar pairs: $l_{\text{BPR}}(\mathbf{z}_{j}^{u},\mathbf{z}_{k}^{i})=-\ln{\sigma(\hat{R}_{jk}-\hat{R}_{jl})},$ (3) where $j,k\in S^{+}$ and $j,l\in S^{-}$, i.e., user $j$ is similar to item $k$ and dissimilar to item $l$. Note that the above loss functions are optimized over the user and item representations, $\mathbf{z}_{j}^{u}$ and $\mathbf{z}_{k}^{i}$, which are used to generate the predicted interaction. Contrastive loss functions [6] are also used in recommendation systems [18, 14, 29]. The idea is to map representations of similar pairs of the users and items close to each other and the dissimilar ones far away: $l_{\text{contrastive}}(\mathbf{z}_{j}^{u},\mathbf{z}_{k}^{i})=R_{jk}||\mathbf{z}^{u}_{j}-\mathbf{z}^{i}_{k}||^{2}+(1-R_{jk})\lambda\max{(0,m_{d}-||\mathbf{z}^{u}_{j}-\mathbf{z}^{i}_{k})||)}^{2},$ (4) where $m_{d}$ is a margin and $j,k\in S^{+}\cup S^{-}$. As we explain in the next section, without the dissimilar sets, all the above loss functions will converge to a collapsed solution, which performs as bad as a random assignment of the representations. The goal of this paper is to introduce new terms to let the above loss functions be trained only on similar sets and still avoid the collapse solution. ## 3 Proposed method As mentioned in the introduction, there are several issues in using the dissimilar pairs in one-class RSs. First, we are never sure whether the selected pairs with the unknown interactions are dissimilar or not. A ”no interaction” does not necessarily mean that the user did not like the item. Selecting similar pairs and using them as dissimilar pairs can hurt the performance. Second, as we show in experiments, previous works use a large number of dissimilar pairs to get reasonable performance, which will increase the training time. Our goal is to learn a recommendation system model by discarding the dissimilar set of users and items, training only on the similar pairs, avoiding the collapsed solution, and achieving state-of-the-art results. In this section, we first explain how the previous objective functions converge to a collapsed solution when they only use similar pairs. Then, we explain our proposed method by considering the $E_{\text{cont}}(\mathbf{Z}^{u},\mathbf{Z}^{i})$ in Eq. (LABEL:e:con_reg) as the main term of the objective function. We show how we avoid the collapsed, partially collapsed, and shrinking solutions by proposing new terms. Then, we will explain how our objective function works with other loss functions. Finally, we go over the computational complexity of the proposed objective function. ### 3.1 The collapsed solution without the dissimilar set The loss functions introduced in Section 2 use both similar and dissimilar pairs to predict the actual representations. Let us explain what happens if we only use the similar pairs and discard the dissimilar ones in two cases: 1) using the loss function $l_{\text{MSE}}$ in Eq. (2) and dot product mapping function in Eq. (1), and 2) using the constrastive loss defined in Eq. (4): $\begin{split}&E_{\text{MSE}}(\mathbf{Z}^{u},\mathbf{Z}^{i})=\sum_{j,k\in S^{+}}((\mathbf{z}^{u}_{j})^{T}\mathbf{z}^{i}_{k}-1)^{2},\\\ &E_{\text{cont}}(\mathbf{Z}^{u},\mathbf{Z}^{i})=\sum_{j,k\in S^{+}}||\mathbf{z}^{u}_{j}-\mathbf{z}^{i}_{k}||^{2}.\end{split}$ (5) Both objective functions above are defined on similar pairs $S^{+}$ and optimizing them will lead to a collapsed solution, where all the users and item representations are mapped to a certain point in the latent space. The optimal solution of the $E_{\text{MSE}}$ is achieved by mapping all the user and item representations to any $d$-dimensional vector with a unit L2 norm. That’s because the dot product of all pairs becomes $1$, i.e., $\hat{R}_{jk}=(\mathbf{z}^{u}_{j})^{T}\mathbf{z}^{i}=1$, which makes the loss value $0$ for all the terms. In the case of $E_{\text{cont}}$, the optimal (and collapsed) solution is achieved by mapping all the user and item representations to any $d$-dimensional vector. At the collapsed solution, which is the result of removing dissimilar pairs from the loss functions, the model always returns the same prediction, no matter what the input pairs are. This works as poorly as a random model. Note that different combinations of mapping functions and loss functions can give us different objective functions. In all cases, without dissimilar pairs, the result is a collapsed solution. ### 3.2 Avoiding the collapsed solution: hinge pairwise distance loss Let us assume the $d$-dimensional representations of the users and items is denoted by $\mathbf{Z}^{u}\in\mathbb{R}^{m\times d}$ and $\mathbf{Z}^{i}\in\mathbb{R}^{n\times d}$, respectively. The joint user-item representation is achieved by vertically concatenating the user and item representations, $\mathbf{Z}=[\mathbf{Z}^{u},\mathbf{Z}^{i}]\in\mathbb{R}^{(m+n)\times d}$. The pairwise distance between all the representations in $\mathbf{Z}$ is computed as: $d_{p}=E_{\text{cont}}(\mathbf{Z},\mathbf{Z})=\frac{1}{(m+n)^{2}}\sum_{l,s=1}^{m+n}||\mathbf{z}_{l}-\mathbf{z}_{s}||^{2}.$ (6) Note that $d_{p}$ computes distance between all the user-user, item-item, and user-item representations, which is different from $E_{\text{cont}}(\mathbf{Z}^{u},\mathbf{Z}^{i})$ that computes the distance between similar pairs of the users and items. Mathematically, at the collapsed solution, we have $d_{p}=0$. To avoid the collapsed solution, we propose a hinge pairwise distance loss that keeps the average pairwise distance $d_{p}$ greater than a margin $m_{p}$. The new objective function can be written as: $E=E_{\text{cont}}(\mathbf{Z}^{u},\mathbf{Z}^{i})+E_{d_{p}}(\mathbf{Z})=\sum_{j,k\in S^{+}}||\mathbf{z}^{u}_{j}-\mathbf{z}^{i}_{k}||^{2}+\max{(0,m_{p}-d_{p})}^{2}.$ (7) Note that $d_{p}$ involves computing the distances between all the pairs, which could be very time-consuming. As we show below, $d_{p}$ is equivalent to the summation of the variance of each dimension, which can be computed way faster. Let us denote the $q$th dimension of the $l$th representation as $z_{l,q}$, and the pairwise distance of the $q$th dimension as $d_{p}^{q}$. Then, $d_{p}$ in Eq. (6) can be separated over the $d$ dimensions: $d_{p}=\sum_{q=1}^{d}d_{p}^{q}=\frac{1}{(m+n)^{2}}\sum_{q=1}^{d}\sum_{l,s=1}^{m+n}(z_{lq}-z_{sq})^{2}.$ (8) We can rewrite $d_{p}^{q}$ as: $d_{p}^{q}=\frac{1}{(m+n)^{2}}\sum_{l,s=1}^{m+n}(z_{lq}-z_{sq})^{2}=\frac{1}{(m+n)^{2}}\sum_{l,s=1}^{m+n}z_{lq}{{}^{2}}+z_{sq}{{}^{2}}-2z_{lq}z_{sq}=\\\ \frac{2}{(m+n)^{2}}\sum_{l,s=1}^{m+n}z_{lq}{{}^{2}}-\frac{2}{(m+n)^{2}}\sum_{l=1}^{m+n}z_{lq}\sum_{s=1}^{m+n}z_{sq}=\frac{2}{(m+n)}\sum_{l=1}^{m+n}z_{lq}{{}^{2}}-2\bar{\mathbf{Z}}_{:,q}^{2}=2\text{var}(\mathbf{Z}_{:,q}).$ (9) So twice the variance of a dimension is equivalent to the pairwise distances of the representations in that dimension. This means that at the collapse solution, the variance of each dimension is $0$, and to avoid this solution, we just need the summation of the variance of the dimensions to be greater than a margin. At the end of this section, we explain how the new formulation helps the computational complexity. ### 3.3 Avoiding the partially collapsed solution: orthogonality term While the objective function of Eq. (7) avoids the collapsed solution, it might give us low-quality representations by converging to a ”partially collapsed solution”. This solution returns only two representations for the whole set of users and items: a portion of the users and items are mapped to one representation and the rest of them are mapped to a second representation. Below, we explain when the objective function converges to this solution and how we avoid it by proposing an orthogonality term. To explain the partially collapsed solution, we consider the user-item bipartite graph. This graph contains two disjoint sets of nodes: set $I$ contains the items and set $U$ contains the users. The edges are determined by $S^{+}$, i.e., there is an edge between nodes of user $j$ and item $k$ if $(j,k)\in S^{+}$. Now, let us assume the graph can be partitioned into $V>1$ disjoint components $\\{C_{v}\\}_{v=1}^{V}$, where the edges in each component are denoted by $S_{v}^{+}$. Since it’s a disjoint partition, each node from $U$ and $I$ appears in exactly one component, $\cup\\{S_{v}^{+}\\}=S^{+}$, and there is no edge across the nodes of the two different components. The interaction matrix is more than $99\%$ sparse in real world scenarios, which can easily lead to having multiple disjoint partitions. The optimal solution of our objective function in the setting above can be achieved by assigning representation $\mathbf{p}_{1}$ to all the nodes in some of the components and $\mathbf{p}_{2}$ to all the nodes in the rest of the components. First, let us show how the above solution makes makes $E_{\text{cont}}=0$. Here, for simplicity, let us assign $\mathbf{p}_{1}$ to the components $1,\dots,\frac{v}{2}$ and assign $\mathbf{p}_{2}$ to the components $\frac{v}{2}+1,\dots,V$. Then we have: $E_{\text{con}}=\sum_{j,k\in S^{+}}||\mathbf{z}^{u}_{j}-\mathbf{z}^{i}_{k}||^{2}=\sum_{v=1}^{V}\sum_{j,k\in S_{v}^{+}}||\mathbf{z}^{u}_{j}-\mathbf{z}^{i}_{k}||^{2}=\\\ \sum_{v=1}^{\frac{V}{2}}\sum_{j,k\in S_{v}^{+}}||\mathbf{z}^{u}_{j}-\mathbf{z}^{i}_{k}||^{2}+\sum_{v={\frac{V}{2}+1}}^{V}\sum_{j,k\in S_{v}^{+}}||\mathbf{z}^{u}_{j}-\mathbf{z}^{i}_{k}||^{2}=\\\ \sum_{v=1}^{\frac{V}{2}}\sum_{j,k\in S_{v}^{+}}||\mathbf{p}_{1}-\mathbf{p}_{1}||^{2}+\sum_{v={\frac{V}{2}+1}}^{V}\sum_{j,k\in S_{v}^{+}}||\mathbf{p}_{2}-\mathbf{p}_{2}||^{2}=0.$ (10) We can easily prove that $E_{\text{con}}$ becomes $0$ for the other ways of assigning the two representations to the components. To make the hinge pairwise distance loss $E_{d_{p}}=0$, we just need to put the two points $\mathbf{p}_{1}$ and $\mathbf{p}_{2}$ far away from each other, which makes the average variance of the dimensions greater than the margin $m_{d}$. In the extreme case, consider each dimension of $\mathbf{p}_{1}$ goes to $+\infty$ and each dimension of $\mathbf{p}_{2}$ goes to$-\infty$. The partially collapsed solution is (almost) as bad as the collapsed solution. It assigns the same representation to all the nodes in the same component which makes them inseparable, while the nodes in each component should be separated based on their interactions with the other nodes. To avoid the partially collapsed solution, we look at the correlation coefficients between the dimensions of the representations. If we have only two unique representations in the user-item representations matrix $\mathbf{Z}$, then there is a linear relationship between the dimension of the $\mathbf{Z}$, and by definition the correlation between the dimensions becomes maximum ($-1$ or $+1$). To avoid this, we add an orthogonality term to the objective function to make the representations orthogonal: $E_{ours}=\lambda_{1}E_{\text{cont}}(\mathbf{Z}^{u},\mathbf{Z}^{i})+\lambda_{2}E_{d_{p}}(\mathbf{Z})+\lambda_{3}E_{\text{orth}}(\mathbf{Z}),$ (11) where $E_{\text{cont}}(\mathbf{Z}^{u},\mathbf{Z}^{i})$ and $E_{d_{p}}(\mathbf{Z})$ are defined in Eq. (7) and we defined $E_{\text{orth}}(\mathbf{Z})$ as follows: $E_{\text{orth}}(\mathbf{Z})=\sum_{q=1}^{d}\sum_{s=q+1}^{d}\hat{\mathbf{Z}}_{:,q}^{T}\hat{\mathbf{Z}}_{:,s},$ (12) where $\hat{\mathbf{Z}}$ is achieved by subtracting the mean of each dimension from $\mathbf{Z}$, and $\hat{\mathbf{Z}}_{:,q}$ is the $q$th column of $\hat{\mathbf{Z}}$. The orthogonality term let the objective avoid the partially collapsed solution by making the off-diagonal values of the covariance matrix small, which consequently makes the correlation of the dimensions of the representations small. ### 3.4 Shrinking solution: why we need both terms As explained above, if we only optimize the hinge pairwise distance loss $E_{d_{p}}(\mathbf{Z})$ and $E_{\text{cont}}(\mathbf{Z}^{u},\mathbf{Z}^{i})$ together, without using the orthogonality term $E_{\text{orth}}(\mathbf{Z})$, it will converge to a partially collapsed solution. On the other hand, optimizing the orthogonality term $E_{\text{orth}}(\mathbf{Z})$ and $E_{\text{cont}}(\mathbf{Z}^{u},\mathbf{Z}^{i})$, without $E_{d_{p}}(\mathbf{Z})$, is enough to avoid both the collapsed and partially collapsed solutions. Then, a question arises here: do we need both hinge pairwise distance loss and the orthogonality term in the same objective function? Or is it enough to only enforce representation orthogonality to make the correlation small? In other words, where does the solution converge if we set $\lambda_{2}=0$ in Eq. (11)? To answer this question, let us consider a trivial (but not collapsed) solution, where all users’ and items’ d-dimensional representations are generated from a multivariate uniform distribution. More specifically, each of the $d$ dimensions is drawn from the interval $[-r,r]$, independently from other dimensions. In this scenario, the dimensions are uncorrelated, so the orthogonality term $E_{\text{orth}}(\mathbf{Z})$ is almost $0$ in practice. The maximum distance between any two representations is $4dr^{2}$, which is the upper-bound for $E_{cont}(\mathbf{Z}^{u},\mathbf{Z}^{i})$. So by setting $r$ to an extremely small number (but larger than $0$), we can make $E_{cont}(\mathbf{Z}^{u},\mathbf{Z}^{i})$ extremely small. So this solution, which generates low-quality random representations from a multivariate uniform distribution, can make the objective function of Eq. (11) very small if we don’t include the hinge pairwise distance loss. That’s because as we make $r$ smaller, the space of the representation shrinks, and the variance of each dimension becomes very small. The hinge pairwise distance loss makes sure that the average variance is larger than a margin and avoids this shrinking solution. ### 3.5 Other objective functions in RSs In this section, we explore the impact of the proposed terms on the different objective functions. We explain what happens if we change the objective from $E_{\text{cont}}$ to another one, achieved by combining loss functions in Eq. (2) and mappings in Eq. (1). Is it necessary to use both $E_{d_{p}}$ and $E_{\text{orth}}$ with the objective functions other than $E_{\text{cont}}$? Let us explain what happens if we replace $E_{\text{cont}}$ by $E_{\text{MSE}}$, defined in Eq. (LABEL:e:con_reg). By optimizing $E_{\text{MSE}}$ on the similar pairs, the collapsed solution is achieved by assigning all users and items to any point $\mathbf{p}\in\mathbb{R}^{d}$ such that $||\mathbf{p}||_{2}=1$. Let us assume the user-item graph can be partitioned into $V$ components. The partially collapsed solution is achieved by assigning all representations to any two points $\mathbf{p}_{1}\in\mathbb{R}^{d}$ and $\mathbf{p}_{2}\in\mathbb{R}^{d}$, where $||\mathbf{p}_{1}||_{2}=1$, $||\mathbf{p}_{2}||_{2}=1$, such that all the users and items in the same component get the same representation. Unlike the previous case where we used $E_{\text{cont}}$, we do not get the shrinking solution if we draw each dimension of each representation from a multivariate uniform distribution in $[-r,r]$. That’s because as we make $r$ smaller, the absolute value of the dimensions of the representations decreases, the dot product of the representations becomes smaller and further away from $1$, and the loss function $E_{\text{MSE}}$ increases. So the loss function $E_{\text{MSE}}$ itself avoids the shrinking solution. Since we can prevent the collapsed and partially collapsed solutions using the orthogonality term, without converging to a shrinking solution, we cannot gain much by adding $E_{d_{p}}$. The results on CiteULike dataset in our experimental results section confirm this. For other objective functions in the literature, one needs to first verify the existence of the collapsed, partially collapsed, and shrinking solutions and then decide which terms to use to prevent them. ### 3.6 Computational complexity and batch-wise training Here, we analyze the computational complexity of each term in our objective function in Eq. (11). The time complexity of $E_{\text{cont}}$ and $E_{\text{orth}}$ are $O(|S^{+}|d)$ and $O((m+n)d)$, respectively. For $E_{d_{p}}$, if we use Eq. (6), the complexity is $O((m+n)^{2}d)$. If we use Eq. (9) to compute the variance, then the time complexity of computing $E_{d_{p}}$ will decrease to $O((m+n)d)$, which is linear in the total number of users and items. In our training, we use neural networks of the previous works to extract user and item representations. Since we use batches to train the model and update the representations, the three terms are computed on a batch and will have a much smaller time complexity, which depends on the number of users and items in the batch. ## 4 Experiments ### 4.1 Experimental setup We conduct experiments in both warm-start and item cold-start settings. In the warm-start setting, the interaction feedback of the test users and items is available in training data and no new user or item will be added to the system at test time. In the item cold-start setting, there is no feedback history for the test items and we only have access to their side information. In the cold-start setting, we use the following two datasets: * • CiteULike [27]. In this dataset, each user has a set of saved articles and the goal is to recommend new (cold-start) articles to the users. We use the subset provided by [26], which contains $5\,551$ users, $16\,980$ articles, and $204\,986$ user-article pairs. The interaction matrix $\mathbf{R}$ is $99.8$% sparse and $R_{jk}=1$ means user $j$ saved article $k$. A set of $3\,396$ articles are removed from the training data and used as the test cold-start articles. * • Ichiba10m. This dataset contains $10$ million purchases from a specific category of the Rakuten Ichiba 111www.rakuten.co.jp. There are around $844\,000$ items, $1.4$M users, and $2\,000$ cold-start items. The dataset is $99.9991$% sparse. The model uses the user interaction vector and item side information, which includes item’s category, price, and title. The train/test split is done based on the time period. The cold start items are the items that didn’t appear during the training period. In Ichiba10m dataset, the goal is to find interested users for a newly released (cold-start) item. In CiteULike dataset, the goal is to assign a set of cold-start articles to each user. In both datasets, we report recall to evaluate the performance of the methods, which shows what portion of the retrieved items/articles is relevant to the users. Since ranking all users is time-consuming in Ichiba, we randomly sample 1 000 unobserved interactions for each item and report recall by returning top 50 users with highest score. We use the following two datasets in the warm-start setting: * • Lastfm222https://www.last.fm/. The dataset333http://ocelma.net/MusicRecommendationDataset/lastfm-1K.html contains $69\,149$ interactions from $1\,741$ users on $2\,665$ songs. The dataset is $98.5$% sparse. * • AMusic [7]. The dataset contains $1\,700$ users, $13\,000$ musics, and $46\,000$ interactions. The dataset is $99.8$% sparse. We use the pre-processed datasets provided by Dong et al. [5], which is publicly available for non-commercial use444https://github.com/familyld/DeepCF. We report Hit Ratio (HR) to evaluate the implicit feedback prediction performance. We follow the same protocols as [5, 8] and truncate the rank list at $10$ for both metrics. HR measures whether the actual test item exists in the top-ranked list. #### Baselines. We compare our method with the state-of-the-art methods in both cold-start and warm-start settings: * • Shared [18]. It uses a shared item network to learn both user and item representations. The item network utilizes side information to create item representations. The user representations are generated from the representations of the items purchased by the user. * • Shared attention [18]. Same the the Shared model above, except that it uses attention mechanism to combine the item representations. * • DropoutNet-WMF and DropoutNet-CDL [26]. The key idea is to apply input dropout to the neural network model such that it can handle the missing input data in the cold-start setting. * • ACCM [22]. It applies the attention mechanism to hybrid RSs to adjust the importance of the input sources. * • BUIR-ID [13]. This is the first work to utilize the stop-gradient approach [1] in RSs. Similar to SimPDO, BUIR-ID only uses positive pairs. The difference from SimPDO is that BUIR-ID has two encoders and uses the momentum mechanism to update the parameters of one encoder and to avoid the collapse solution. * • CFNet [5]. It uses interaction vector as the input and learns two user and two item representations by two neural network models, which are then combined and mapped to the final interaction by an MLP. * • NeuMF [8]. It uses IDs as the input and learns two user and two item representations, which are then combined by an MLP to predict the final interaction. * • DMF [30]. It learns user and item representations by training a neural networks which uses the explicit feedback as the input and the dot-product as the mapping function. Ichiba10m --- | CiteULike | Figure 1: Impact of each term of our objective function on the results in Ichiba10m (top row) and CiteULike (bottom row) datasets. We report the average variance of each dimension, the average correlation between the dimensions, and recall, as we train the models. The model that uses all three terms achieves the best results. #### RSs models to learn the user and item representations. The specific models to map users and items to the low-dimensional representations are taken from previous works since the novelty of our work is in designing a new objective function, which works no matter what the model is. In the cold-start setting, we use the Shared model proposed by Raziperchikolaei et al. [18] and in the warm-start setting we use the neural collaborative framework proposed by Wang et al. [28]. #### Implementation details. We implemented our method using Keras with TensorFlow 2 backend. We ran all the experiments on one Nvidia Tesla V100-SXM2 32GB GPU in the internal cluster. We use grid search to set the hyper-parameters using a subset of training set and a small validation set. We set the maximum number of epochs to $50$. We use SGD with the learning rate of $0.5$ for our method in all experiments. We set the batch size to $32$ in CiteULike and $128$ in other datasets. We set $\lambda_{1}=0.01$, $\lambda_{1}=1$, and $\lambda_{3}=1$ in all datasets. The margin $m_{p}$ is $0.1$ in the CiteULike and $0.01$ in all other datasets. The dimension of the user and item embeddings are $100$ in the CiteULike and Ichiba10m dataset and $1\,000$ in AMusic and Lastfm. Figure 2: SimPDO vs Shared [18] as we change the number of training pairs from $1$M to $10$M and report recall@50 in Ichiba10m dataset. SimPDO achieves significantly better results when we use a smaller number of training pairs. ### 4.2 Experimental results Our proposed objective function uses Similar pairs, Pairwise Distance loss, and Orthogonality loss, and is denoted by SimPDO. #### Impact of each term of SimPDO. Our proposed objective function has three terms. The first term, $E_{\text{cont}}$ or $E_{\text{MSE}}$, tries to generate a high interaction score for the similar pairs, either by putting their representations close to each other or by maximizing the dot product of the representations. The second term, $E_{d_{p}}$ defined in Eq. (6), tries to maximize the pairwise distance of all the pairs (or equivalently maximize the variance) and avoid the collapsed and shrinking solutions. The third term, $E_{\text{orth}}$ in Eq. (11), avoid the partially collapsed solution by making the representations orthogonal. In Fig. 1, we investigate the impact of each term by reporting: 1) the average variance of the dimensions of all the representations, 2) the average correlation between the dimensions of the representations, and 3) the recall, as we train the models. In the top row, we report the results on Ichiba10m dataset. We can see that the method with all three terms achieves the maximum recall because: 1) without $E_{\text{cont}}$, the similarity between the similar pairs will not be preserved, 2) without $E_{d_{p}}$, as we can see in the second column, the variance drops significantly towards $0$, which is a sign of the shrinking or collapsed solution, and 3) without $E_{\text{orth}}$, as we can see in the third column, the average correlation increase, which is a sign of the partially collapsed solution. In the bottom row of the Fig. 1, we show the results on CiteULike, where we use $E_{\text{MSE}}$ instead of the $E_{\text{cont}}$. The results on the CiteULike confirm that the model with all three terms performs best. Note that in CiteULike the curve without $E_{d_{p}}$ performs almost as well as using all the terms. In addition, note that even without $E_{d_{p}}$, the variance does not drop significantly and it is a bit smaller compared to the other curves. That’s because $E_{\text{orth}}$ avoids the collapsed and partially collapsed solution and the shrinking solution does not happen when we use $E_{\text{MSE}}$. #### Training only on similar pairs: better results with fewer training points. In Ichiba10m dataset, we fix the test set to $2\,000$ cold-start items, change the number of training pairs from $1$M to $10$M, and report the recall of our method (SimPDO) and Shared model [18] in Fig. 2. SimPDO only uses similar pairs in training and optimizes the objective function of Eq. (11). Shared model [18] uses both similar and dissimilar pairs and optimizes the objective function of Eq. (4). Table 1: Comparison with the state-of-the-art cold-start methods on CiteULike dataset. The * in front of the methods’ name means we report the results from the Shared model paper [18]. Our method outperforms the competitors. We run SimPDO three times and report the mean and standard deviation of the results. method | recall@100 ---|--- SimPDO | $\mathbf{66.8}\pm 0.11$% Shared* [18] | $65.7$% Shared attention* [18] | $66.4$% DropoutNet-WMF* [26] | $65.2$% DropoutNet-CDL* [26] | $62.9$% ACCM [22] | $63.1$% In SimPDO, all training pairs are similar pairs so we have no dissimilar pairs. In Shared, the training pairs are divided between the similar and dissimilar pairs in different portions. In Fig. 2, we considered two cases, where the number of dissimilar pairs is equal to (first case) and three times more than (second case) the number of similar pairs. There are three remarkable points about the results of Fig. 2. First, SimPDO performs better than the Shared model no matter how many training pairs are used, which shows the advantage of training only on similar pairs. Second, the gap between the Shared and SimPDO becomes smaller as we increase the portion of the dissimilar pairs compared to the similar ones, which shows the importance of using a large number of dissimilar pairs. Third, SimPDO is significantly better than the Shared model using a smaller number of training pairs. This is a big advantage of SimPDO when the datasets have billions of pairs and training on all of them is time-consuming: SimPDO can be trained on a smaller training set and still achieve reasonable results. Table 2: Comparison with the state-of-the-art warm-start methods on AMusic and Lastfm datasets. The * in front of the methods’ name means we report the results from the DeepCF paper [5]. Our method achieves competitive results using smaller number of training pairs. We put $-$ for HR@5 of some of the methods since only HR@10 is reported in [5]. We run SimPDO, BUIR-ID, and CFNet three times and report the mean and standard deviation of the results. | AMusic | Lastfm ---|---|--- method | HR@5 | HR@10 | HR@5 | HR@10 SimPDO | $\mathbf{32.66\pm 0.4}$ | $\mathbf{43.47\pm 0.09}$ | $75.4\pm 0.1$ | $\mathbf{89.6\pm 0.18}$ BUIR-ID [13] | $25.76\pm 0.2$ | $36.5\pm 0.7$ | $70.16\pm 0.6$ | $84.15\pm 0.9$ CFNet [5] | $29.7\pm 0.08$ | $38.4\pm 0.003$ | $\mathbf{77.1\pm 0.17}$ | $88.6\pm 0.02$ CFNet* [5] | - | $41.2$ | - | $89.9$ CFNet-ml* [5] | - | $40.7$ | - | $88.3$ CFNet-rl* [5] | - | $39.4$ | - | $88.4$ NeuMF* [8] | - | $38.9$ | - | $88.6$ DMF* [30] | - | $37.4$ | - | $88.4$ Table 3: We report number of similar and dissimilar pairs to train different methods and training time per epoch of different methods. SimPDO and BUIR-ID use a smaller number of pairs and can be trained much faster than the other methods. | AMusic | Lastfm | CiteULike ---|---|---|--- method | numS | numD | training time | numS | numD | training time | numS | numD | training time SimPDO | $46$K | $0$ | $13$s | $69$K | $0$ | $15$s | $200$K | $0$M | $123$s BUIR-ID | $46$K | $0$ | $8$s | $69$K | $0$ | $11$s | $200$K | $0$M | $123$s CFNet | $46$K | $184$K | $36$s | $69$K | $276$K | $36$s | - | - | - Shared | - | - | - | - | - | - | $200$K | $2$M | $1871$s Shared_attention | - | - | - | - | - | - | $200$K | $2$M | $1963$s #### Comparison with the state-of-the-art methods. In Table 1, we compare the methods on CiteULike dataset in the cold-start setting. The dataset and the Shared code are available online555https://github.com/rakutentech/shared_rep. We use the code and processed data provided in this repository to train our model. Note that our method uses the same user and item models as the shared model, but uses the new objective function. This makes it clear that the better performance of the SimPDO comes from the proposed objective function that is only using similar pairs. In Table 2, we compare the methods on AMusic and Lastfm datasets in the warm- start setting. The dataset and the DeepCF code are available online666https://github.com/familyld/DeepCF. We use the code and processed data provided in this repository to train our model. We also trained CFNet and BUIR-ID from the publicly available code and put the results in the second and third row of Table 2. Results of Table 1 and Table 2 show that SimPDO outperforms most state-of-the- art methods and achieves very competitive results. Note that SimPDO and BIUR use a significantly smaller number of training pairs compared to the other methods, such as Shared and CFNet. We have put the number of similar and dissimilar pairs used in training the methods with the best results in Table 3. SimPDO and BUIR-ID use a smaller number of pairs and can be trained much faster than the other works. ## 5 Conclusion In this paper, we proposed SimPDO, a new objective function that enables training of one-class recommendation system models without dissimilar pairs. We showed that by only using similar pairs, the optimal solution of existing objective functions becomes a collapsed solution, where every representation is mapped to the same point in the latent space. We avoided the collapsed solution by providing a hinge loss for pairwise distance which is equivalent to the average variance of each dimension of representations. We also showed that we need an orthogonality term to avoid collapsed and partially collapsed solutions, where the optimal solution under the pairwise loss returns only two representations. The orthogonality term minimizes the correlation between each dimension of representations and forces the objective function to return various representations. Finally, we showed that both terms are necessary to learn meaningful representations. The results demonstrated that SimPDO outperformed the existing RS objective functions without using dissimilar pairs. We also showed that SimPDO can be trained more efficiently with a smaller number of training pairs. ## References * Chen and He [2021] Xinlei Chen and Kaiming He. Exploring simple siamese representation learning. In _Proceedings of the 2021 IEEE Computer Society Conference Computer Vision and Pattern Recognition (CVPR)_ , 2021. * Ding et al. [2019] Jingtao Ding, Yuhan Quan, Xiangnan He, Yong Li, and Depeng Jin. Reinforced negative sampling for recommendation with exposure data. In _Proceedings of the 28th International Joint Conference on Artificial Intelligence (IJCAI)_ , 2019. * Ding et al. [2020] Jingtao Ding, Yuhan Quan, Quanming Yao, Yong Li, and Depeng Jin. Simplify and robustify negative sampling for implicit collaborative filtering. In _Proceedings of the 34th International Conference on Neural Information Processing Systems (NeurIPS)_ , 2020. * Dong et al. [2017] Xin Dong, Lei Yu, Zhonghuo Wu, Yuxia Sun, Lingfeng Yuan, and Fangxi Zhang. A hybrid collaborative filtering model with deep structure for recommender systems. In _Proceedings of the 31st AAAI Conference on Artificial Intelligence_ , 2017. * Dong et al. [2019] Zhi-Hong Dong, Ling Huang, Chang-Dong Wang, Jian-Huang Lai, and Philip S Yu. DeepCF: a unified framework of representation learning and matching function learning in recommender system. In _Proceedings of the 33rd AAAI Conference on Artificial Intelligence_ , 2019. * Hadsell et al. [2006] Raia Hadsell, Sumit Chopra, and Yann LeCun. Dimensionality reduction by learning an invariant mapping. In _Proceedings of the 200614 IEEE Computer Society Conference Computer Vision and Pattern Recognition (CVPR)_ , 2006. doi: 10.1109/cvpr.2006.100. * He and McAuley [2016] Ruining He and Julian McAuley. Ups and downs: Modeling the visual evolution of fashion trends with one-class collaborative filtering. In _Proceedings of the 25th International Conference on World Wide Web (WWW)_ , 2016. doi: 10.1145/2872427.2883037. * He et al. [2017] Xiangnan He, Lizi Liao, Hanwang Zhang, Liqiang Nie, Xia Hu, and Tat-Seng Chua. Neural collaborative filtering. In _Proceedings of the 26th International Conference on World Wide Web (WWW)_ , 2017. * He et al. [2018] Xiangnan He, Xiaoyu Du, Xiang Wang, Feng Tian, Jinhui Tang, and Tat-Seng Chua. Outer product-based neural collaborative filtering. In _Proceedings of the 27th International Joint Conference on Artificial Intelligence (IJCAI)_ , 2018. * He et al. [2020] Xiangnan He, Kuan Deng, Xiang Wang, Yan Li, Yongdong Zhang, and Meng Wang. LightGCN: Simplifying and powering graph convolution network for recommendation. In _Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval_ , 2020. * Hu et al. [2008] Yifan Hu, Yehuda Koren, and Chris Volinsky. Collaborative filtering for implicit feedback datasets. In _Proceedings of the 8th IEEE International Conference Data Mining (ICDM)_ , 2008. * Koren et al. [2009] Yehuda Koren, Robert Bell, and Chris Volinsky. Matrix factorization techniques for recommender systems. _Computer_ , 42(8):30–37, 2009. doi: 10.1109/mc.2009.263. * Lee et al. [2021] Dongha Lee, SeongKu Kang, Hyunjun Ju, Chanyoung Park, and Hwanjo Yu. Bootstrapping user and item representations for one-class collaborative filtering. In _Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval_. ACM, jul 2021\. doi: 10.1145/3404835.3462935. * Lin et al. [2022] Zihan Lin, Changxin Tian, Yupeng Hou2, and Wayne Xin Zhao. Improving graph collaborative filtering withneighborhood-enriched contrastive learning. In _Proceedings of the 31st International Conference on World Wide Web (WWW)_ , 2022. * Pan et al. [2008] Rong Pan, Yunhong Zhou, Bin Cao, Nathan N. Liu, and Rajan Lukose. One-class collaborative filtering. In _Proceedings of the 8th IEEE International Conference Data Mining (ICDM)_ , 2008. * Raziperchikolaei and joo Chung [2022] Ramin Raziperchikolaei and Young joo Chung. Simultaneous learning of the inputs and parameters inneural collaborative filtering. _arXiv preprint arXiv:2203.07463_ , 2022. * Raziperchikolaei et al. [2021a] Ramin Raziperchikolaei, Tianyu Li, and Young joo Chung. Neural representations in hybrid recommender systems: Prediction versus regularization. In _Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval_ , 2021a. * Raziperchikolaei et al. [2021b] Ramin Raziperchikolaei, Guannan Liang, and Young joo Chung. Shared neural item representations for completely cold start problem. In _Proceedings of the 15th ACM Conference on Recommender Systems (RecSys)_ , 2021b. * Rendle et al. [2009] Steffen Rendle, Christoph Freudenthaler, Zeno Gantner, and Lars Schmidt-Thieme. BPR: bayesian personalized ranking from implicit feedback. In _Proceedings of the 25th Conference on Uncertainty in Artificial Intelligence (UAI)_ , 2009. * Ricci et al. [2011] Francesco Ricci, Lior Rokach, Bracha Shapira, and Paul B. Kantor, editors. _Recommender Systems Handbook_. Springer US, 2011. doi: 10.1007/978-0-387-85820-3. * Sedhain et al. [2015] Suvash Sedhain, Aditya Krishna Menon, Scott Sanner, and Lexing Xie. AutoRec: Autoencoders meet collaborative filtering. In _Proceedings of the 24th International Conference on World Wide Web (WWW)_ , 2015. doi: 10.1145/2740908.2742726. * Shi et al. [2018] Shaoyun Shi, Min Zhang, Yiqun Liu, and Shaoping Ma. Attention-based adaptive model to unify warm and cold starts recommendation. In _Proceedings of the 27th ACM International on Conference on Information and Knowledge Management (CIKM)_ , 2018. doi: 10.1145/3269206.3271710. * Strub et al. [2016] Florian Strub, Romaric Gaudel, and Jérémie Mary. Hybrid recommender system based on autoencoders. In _Proceedings of the 1st Workshop on Deep Learning for Recommender Systems (DLSRS)_ , 2016. doi: 10.1145/2988450.2988456. * Takács et al. [2008] Gábor Takács, István Pilászy, Bottyán Németh, and Domonkos Tikk. Matrix factorization and neighbor based algorithms for the netflix prize problem. In _Proceedings of the 2nd ACM Conference on Recommender Systems (RecSys)_ , 2008. doi: 10.1145/1454008.1454049. * van den Berg et al. [2018] Rianne van den Berg, Thomas N. Kipf, and Max Welling. Graph convolutional matrix completion. In _Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD)_ , 2018. * Volkovs et al. [2017] Maksims Volkovs, Guangwei Yu, and Tomi Poutanen. DropoutNet: Addressing cold start in recommender systems. In _Proceedings of the 31st International Conference on Neural Information Processing Systems (NIPS)_ , 2017. * Wang and Blei [2011] Chong Wang and David M. Blei. Collaborative topic modeling for recommending scientific articles. In _Proceedings of the 17th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD)_ , 2011. * Wang et al. [2019] Xiang Wang, Xiangnan He, Meng Wang, Fuli Feng, and Tat-Seng Chua. Neural graph collaborative filtering. In _Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR)_ , 2019\. doi: 10.1145/3331184.3331267. * Wu et al. [2019] Ga Wu, Maksims Volkovs, and Chee Loong Soon. Noise contrastive estimation forone-class collaborative filtering. In _Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval_ , 2019. * Xue et al. [2017] Hong-Jian Xue, Xinyu Dai, Jianbing Zhang, Shujian Huang, and Jiajun Chen. Deep matrix factorization models for recommender systems. In _Proceedings of the 26th International Joint Conference on Artificial Intelligence (IJCAI)_ , 2017. doi: 10.24963/ijcai.2017/447. * Ying et al. [2018] Rex Ying, Ruining He, Kaifeng Chen, Pong Eksombatchai, William L. Hamilton, and Jure Leskovec. Graph convolutional neural networks for web-scale recommender systems. In _Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD)_ , 2018. doi: 10.1145/3219819.3219890. * Zhang et al. [2019] Jiani Zhang, Xingjian Shi, and Shenglin Zhaoand Irwin King. STAR-GCN: Stacked and reconstructed graph convolutional networks for recommender systems. In _Proceedings of the 33rd AAAI Conference on Artificial Intelligence_ , 2019. * Zhang et al. [2013] Weinan Zhang, Tianqi Chen, Jun Wang, and Yong Yu. Optimizing top-n collaborative filtering via dynamic negative item sampling. In _Proceedings of the 36th International ACM SIGIR Conference on Research and Development in Information Retrieval_ , 2013. * Zhang et al. [2017] Yongfeng Zhang, Qingyao Ai, Xu Chen, and W. Bruce Croft. Joint representation learning for Top-N recommendation with heterogeneous information sources. In _Proceedings of the 26th ACM International on Conference on Information and Knowledge Management (CIKM)_ , 2017. doi: 10.1145/3132847.3132892. * Zhou et al. [2008] Yunhong Zhou, Dennis Wilkinson, Robert Schreiber, and Rong Pan. Large-scale parallel collaborative filtering for the netflix prize. In _Algorithmic Aspects in Information and Management_ , pages 337–348. Springer Berlin Heidelberg, 2008. doi: 10.1007/978-3-540-68880-8˙32.
††thanks: M. A. N., H. K. and W. B. C. contributed equally to this work.††thanks: M. A. N., H. K. and W. B. C. contributed equally to this work.††thanks: M. A. N., H. K. and W. B. C. contributed equally to this work. # Iterative assembly of 171Yb atom arrays in cavity-enhanced optical lattices M. A. Norcia H. Kim W. B. Cairncross M. Stone A. Ryou M. Jaffe M. O. Brown K. Barnes P. Battaglino T. C. Bohdanowicz A. Brown K. Cassella C.-A. Chen R. Coxe D. Crow J. Epstein C. Griger E. Halperin F. Hummel A. M. W. Jones J. M. Kindem J. King K. Kotru J. Lauigan M. Li M. Lu E. Megidish J. Marjanovic M. McDonald T. Mittiga J. A. Muniz S. Narayanaswami C. Nishiguchi T. Paule K. A. Pawlak L. S. Peng K. L. Pudenz D. Rodríguez Pérez A. Smull D. Stack M. Urbanek R. J. M. van de Veerdonk Z. Vendeiro L. Wadleigh T. Wilkason T.-Y. Wu X. Xie E. Zalys-Geller X. Zhang B. J. Bloom Atom Computing, Inc ###### Abstract Assembling and maintaining large arrays of individually addressable atoms is a key requirement for continued scaling of neutral-atom-based quantum computers and simulators. In this work, we demonstrate a new paradigm for assembly of atomic arrays, based on a synergistic combination of optical tweezers and cavity-enhanced optical lattices, and the incremental filling of a target array from a repetitively filled reservoir. In this protocol, the tweezers provide microscopic rearrangement of atoms, while the cavity-enhanced lattices enable the creation of large numbers of deep optical potentials that allow for rapid low-loss imaging of atoms. We apply this protocol to demonstrate deterministic filling (99% per-site occupancy) of 1225-site arrays. Because the reservoir is repeatedly filled with fresh atoms, the array can be maintained in a filled state indefinitely. We anticipate that this protocol will be compatible with mid-circuit reloading, which will be a key capability for running large-scale error-corrected quantum computations whose durations exceed the lifetime of a single atom in the system. ## I Introduction Individually controlled neutral atoms provide a promising platform for quantum information processing and simulation, and expanding the size and control of these systems represent key challenges in the ongoing push to access regimes beyond the capabilities of classical simulation. Tweezer arrays and optical lattices have emerged as core technologies for optically trapping and manipulating cold atoms, each with complementary capabilities. Optical lattices provide a well-defined potential landscape with features on the sub- wavelength scale, which can enable tight confinement for imaging and a clean optical potential for simulations involving itinerant atoms [1]. Optical tweezers provide a means of selectively manipulating individual atoms [2]. Notably, tweezer arrays have gained prominence for deterministic assembly of atomic arrays [3, 4, 5] with up to hundreds of atoms [6, 7] and inter-atomic spacing suitable for enabling interactions between atoms through excitation to Rydberg states [8, 9, 10, 6, 7, 11, 12]. Recently, optical tweezers and lattices have been combined to create programmable initial conditions in a Hubbard-regime system [13, 14], and to enable enhanced scaling of atom number for metrology [15] and quantum simulation[16]. The merits of optical lattices can be further enhanced through the use of optical buildup cavities. Optical cavities enhance the lattice depth by enabling the constructive interference of many overlapped reflections of laser light, in turn enabling the creation of arrays of many deep traps [17, 18]. Figure 1: Conceptual diagram of repeated loading sequence. A “reservoir” optical tweezer array (smaller left grid) is repeatedly filled with 171Yb atoms transported from a spatially separated magneto-optical trap (MOT), and ultimately transferred into a “target” tweezer array (larger right grid) using a “rearrangement” tweezer. The rearrangement moves required for the transfer are determined from low-loss images obtained by transferring atoms from the tweezers into a cavity-enhanced optical lattice and performing site-resolved fluorescence detection. The cavity-enhanced lattice allows for the scalable generation of large numbers of deep traps. The call-out shows example images prior to the 20th cycle of rearrangement, and after the 70th cycle, with the final reservoir reloading step omitted. This protocol allows us to load over 1200 atoms into 1225 target sites. In this work, we combine the capabilities of tweezers and cavity-enhanced optical lattices to demonstrate an iterative approach to creating large arrays of individually-controlled atoms. Typically, tweezer rearrangement is performed by stochastically loading up to a single atom into each trap within an array [19], imaging the atoms to determine trap occupancy, and then rearranging atoms within the array to create a deterministically occupied sub- array [4, 5]. Crucially, the number of atoms contained in the final array with this approach is no greater than the number initially loaded. Further, because the initial loading is stochastic, the number of sites in the array must generally be substantially larger than desired final sub-array (though under certain conditions, near-deterministic loading can be achieved [20, 21], and several initial arrays can be used to increase the number of available atoms [22]). Recently, repeated loading of a “buffer” array from an optical dipole trap “reservoir” has demonstrated that one can decouple the filling of a six- site target array from a single loading of a cold reservoir [23]. In this work, we extend this concept to repeated loading of a reservoir array, from which we create a deterministically filled target array (typically 99% occupancy) of over 1200 171Yb atoms in 1225 sites. This is made possible by combining optical tweezer arrays with a cavity-enhanced optical lattice to provide both microscopic control and the large number of deep traps required for rapid, high-fidelity, low-loss imaging of large numbers of atoms. A single cycle of loading proceeds as follows (See figure 1): Atoms are collected and cooled in a magneto-optical trap 30 cm below a “science region” where we create a cavity-enhanced optical lattice, as well as tweezer arrays projected through a high-numerical-aperture imaging. The atoms are then transported into the science region, where they are loaded into a set of tweezers that form the reservoir. The reservoir sits adjacent to the target tweezer array, which on all loading cycles but the first is already partially occupied with atoms loaded on previous cycles. The atoms from both reservoir and target tweezer arrays are then transferred into the trapping potential of the cavity-enhanced optical lattice, where we perform site-resolved nondestructive fluorescence imaging to determine which sites of each tweezer array were occupied. We then transfer the atoms back from the cavity lattice to the tweezers, and use a single separate “rearrangement” tweezer to move atoms from the reservoir into empty target tweezers. This defines one loading cycle. Over the course of multiple cycles, the number of atoms in the target array increases until the target array is filled. At this point, further operations may be performed on the atoms, and subsequent loading cycles are applied to maintain a filled array. Because we reload the reservoir on each loading cycle from a fresh magneto- optical trap, we can continue loading the target array indefinitely. In the near-term, this allows for relatively high data-rates for quantum simulation and computation with large system sizes, and could also be of benefit for optical clocks for high statistical precision with low dead-time [24]. Ultimately, the ability to reload new atoms while maintaining both the presence and coherence of existing atoms will be a key capability for performing error-corrected quantum computations, where execution of an algorithm may take much longer than the lifetime of any given atom in the system. When combined with site-selective hiding [25, 26] and mid-circuit rearrangement techniques already demonstrated in a similar system [26], we anticipate that this protocol will be compatible with mid-circuit reloading of atoms into the array. As a means of maintaining a fully filled array, our approach represents an alternative to the interleaved use of two atomic species [27, 28]. Unlike the two-species approach, the method presented here does not require simultaneous replacement of the entire array, and so may be expected to ease requirements for scaling to large arrays, and would not require inter-species gates [29] in order to replenish the array during a quantum computation. ## II Trapping and imaging atoms in a 3D cavity enhanced lattice Figure 2: Cavity-enhanced optical lattices for rapid low-loss imaging. (a) The intersection of two cavity modes provide three-dimensional confinement for atoms within the field of view of our high-numerical-aperture imaging system. Confinement in the $x$ and $y$ directions is provided by a self-intersecting standing-wave cavity, while confinement in the $z$ direction is provided by a separate self-intersecting running-wave cavity mode. Imaging light (not shown for clarity) is incident along the mode of the running-wave lattice. (b) Observed counts during a 7 ms long image of single atoms within the optical lattice, showing well-resolved peaks for occupied and unoccupied sites. For these conditions, per-image survival is 99.8(1)%. (c) Lattice homogeneity across the target array as characterized by light shifts on the 1S0 $m_{f}=1/2$ to 3P1, $F=3/2$ $m_{f}=-1/2$ transition. Atoms within 13.4 $\mu$m squares are averaged together for statistics and display clarity. (d, e) Light shift contributions from each cavity averaged over rows (d) and columns (e). Red and orange markers and lines represent the measured and predicted profiles of the XY (red) and Z (orange) cavities, with the prediction centered on the array. Rapid low-loss imaging of the locations of atoms is essential to our repeated loading protocol. Because atoms are heated by the photon scattering required for imaging, deep optical traps and imaging protocols with low equilibrium temperatures are required to prevent atom loss. Optical tweezers represent a common method for achieving deep optical traps – the tight foci can provide high intensities for moderate powers. However, the total power requirement for a large array of tweezers presents scaling limitations due to limited available laser power. Cavity-enhanced optical lattices offer an alternative path to achieving many deep traps. While the laser intensity in a lattice is spread over a larger cross-sectional area than in optical tweezers both because it spans the regions between utilized sites, and because the Gaussian mode profile must be large compared to the size of the array to achieve uniform trapping, interference of many reflected paths in a cavity-supported lattice can enable more power-efficient generation of traps. We employ two intersecting optical cavities to provide tight confinement in three dimensions, as shown in figure 2a. The first cavity – which we call the XY cavity – is a retro-reflected four-mirror cavity whose mode intersects itself at the location of the atoms. The polarization is perpendicular to the plane of propagation, providing interference between the two crossed portions of the cavity mode [30]. Compared to using two crossed non-interfering lattices [31, 18], and assuming equal cavity finesse, this configuration provides an eight-fold enhancement in trap depth (defined as the potential barrier between adjacent lattice sites) per unit of total laser power. The XY cavity provides tight confinement in the directions perpendicular to the high numerical aperture optical axis used for the tweezers and imaging. The second cavity, which we call the Z cavity, is a four-mirror running wave cavity whose mode intersects itself at the location of the atoms [32] at 15 degree angles from the XY plane. This provides confinement along the tweezer axis. The polarization of the Z cavity is also oriented perpendicular to its intersection plane, enabling complete interference of the crossing modes. Each cavity has two highly reflecting mirrors, and two matched partially transmitting mirrors. The finesse of the XY and Z cavities are 2900(100) and 3000(100) respectively, and the design mode waists at the location of the atoms are 268 $\mu$m and 183 $\mu$m respectively. For imaging, we typically operate with XY (Z) trap frequencies near 160 kHz (50 kHz), corresponding to 330 $\mu$K (260 $\mu$K) deep traps. We image atoms using the the narrow-linewidth (180 kHz) 1S0 to 3P1 transition, which simultaneously provides cooling and scattering of photons for detection [33, 34, 26]. We use a single beam, detuned several hundred kHz from the $F=3/2$, $m_{f}=3/2$ state, with projection onto both the X and Z directions. For applications like array assembly where we wish to determine the presence of an atom but not the state of its nuclear spin (the qubit we use for quantum information applications), we simultaneously apply optical pumping on the 1S0 $m_{f}=-1/2$ to 3P1 $F=1/2$, $m_{f}=1/2$ transition with light incident along the magnetic field ($x$ direction). At our benchmark imaging parameters, we collect 50 photons in 7 ms, and distinguish occupied from unoccupied sites with a fidelity of approximately 99.95%, as determined from the overlap error in a double-Gaussian fit to the bimodal distribution of counts. Atom loss during such an image is $2(1)\times 10^{-3}$. Spin-selective imaging requires us to omit the optical pumping, and rely on the combination of frequency and polarization selectivity within the 3P1 manifold to create a large imbalance of scattering between the two nuclear spin states, as demonstrated in optical tweezers in refs. [34, 26]. The omission of the optical pumping does not modify the photon scattering rate appreciably, but does introduce a finite spin flip probability of of $4(1)\times 10^{-3}$ for our typical imaging parameters, measured as an additional apparent loss in repeated images of 1S0 $m_{f}=1/2$. These spin flips are likely due to the fact that the lattice polarization is perpendicular to the magnetic field, and so can induce spin- changing transitions within the 3P1 manifold [34]. In order to ensure uniform detuning of the imaging transition across the array and between different motional states, we operate with a trapping wavelength that has equal polarizability for the ground and excited states of the imaging transition – a so-called magic wavelength. For the 1S0 $m_{f}=1/2$ to 3P1 $m_{f}=3/2$ transition, we find a magic condition near 783.8 nm. We characterize the homogeneity of the optical lattice potential by probing the 1S0 $m_{f}=1/2$ to 3P1 $F=3/2$, $m_{f}=-1/2$ transition, for which the excited state has approximately 40% higher polarizability than the ground state. We infer the trap depth from the transition light shifts, measured using optical pumping from the 1S0 $m_{f}=1/2$ state, followed by spin- selective imaging (fig. 2c). By varying the power in each cavity, we can extract their independent contributions, displayed averaged over rows and columns in figure 2d,e. Over our 115 $\mu$m square array, peak deviations are within 20% for the Z lattice and 10% for the XY lattice. Because we operate with a magic wavelength condition for critical operations, these deviations do not present a major limitation. ## III Repeated loading to build up a large array Figure 3: (a) Timing diagram for our continuous loading sequence. See main text for a detailed description of the procedure. The beam paths referenced in the call-out are illustrated in figure 4. (b) Number of atoms and fill fraction as a function of loading iteration number during initial loading (left panel) and while maintaining a filled array (right panel). Black points represent atom number inferred from images preceding the rearrangement step. Red points represent atom number inferred from additional images inserted after rearrangement, representing atoms available for computation. In this section, we describe our continuous loading protocol in detail, and characterize its performance for loading large arrays. Our loading cycle (illustrated in figure 3a) begins with collecting atoms from a pre-cooled atomic beam into a magneto-optical trap (MOT) using a “core- shell” configuration [35]. In this MOT, each beam consists of a ring of 399 nm light near resonance with the broad linewidth 1S0 to 1P1 transition surrounding a central region with only 556 nm light near resonance with the narrow linewidth 1S0 to 3P1 transition. The shell is primarily responsible for capturing atoms from the incident beam, while the low Doppler temperature of the narrow transition used in the core enables cooling to several 10 $\mu$K. Because atoms do not scatter light from the broad transition except during initial capture, the core-shell design poses less risk of generating scattered photons that may be absorbed by atoms in the target array – the total scattering rate on the narrow transition is much lower, and the large magnetic field present in the science region dramatically suppresses reabsorption of photons via the narrow-line transition. In our operating condition, the 399 nm photon scattering rate is about 2 orders of magnitude lower than the broad- line stage of a sequential MOT. We find that the core-shell MOT configuration increases the MOT loading rate by roughly a factor of two, compared to a more standard sequential MOT configuration in a static magnetic field. From the core-shell MOT, atoms are loaded into a vertically oriented standing- wave lattice formed by counterpropagating 532 nm laser beams focused at the location of the atoms. Loading is achieved by overlapping the lattice with the center of the MOT, followed by reducing the detuning and intensity of the 556 nm light. We then extinguish the MOT light and transport the atoms vertically by synchronously applying a frequency offset to the upward-going and downward- going lattice beams, and translating their foci [26]. This brings the atoms 30 cm vertically into the vicinity of the target array, at which point we transfer the atoms into a reservoir array formed by 483 nm tweezers (another magic wavelength for the 1S0 to 3P1 transition [26]). To avoid disturbing atoms already loaded into the target array, this transfer takes place at a location displaced 170 $\mu$m horizontally from the target array. No dissipation is required to transfer atoms from the transport lattice to the tweezers – loading is achieved by simply increasing the intensity of the tweezers while the lattice is overlapped and then decreasing the intensity of the lattice. Once loaded, the reservoir array is translated to be directly adjacent to the target array by changing the angle of a galvo mirror that is imaged onto the entrance pupil of the objective. The target array consists of a rectangular array of tweezer spots formed with light at 459.5960(5) nm wavelength, a magic wavelength for the 1S0 to 3P0 clock transition [26, 36], and a set of tweezer spots formed by light at 423.31 nm, which is near resonance with a transition between 3P1 to a higher- lying 6S8S 3S1 state. Both wavelengths provide large light-shifts to atoms in 3P1, which prevents unwanted scattering from atoms loaded in these tweezer arrays, and can be used interchangeably for this work. For this work, we operate with a 35x35-site target array with 3.3 $\mu$m spacing, comprised of two separate rectangular arrays each using one of the two previously mentioned wavelengths. We have also operated with overlapped full-size arrays to achieve sufficient trap depth, and obtained similar results. However, the overlapped condition is more sensitive to alignments, and so we use spatially separated arrays for all results presented here. While translating the reservoir array adjacent to the target array, both arrays are illuminated with light resonant with the 1S0 to 3P1 $F=3/2$, $m_{f}=3/2$ transition. This induces light-assisted collisions between atoms in the reservoir tweezers, resulting in either zero or one atom in each reservoir tweezer with approximately equal probability [19]. Importantly, the tweezer light illuminating the atoms in the target array prevents them from scattering the light used to induce collisions, which might otherwise cause loss of existing atoms in the target array. Next, atoms from both tweezer arrays are transferred into the optical lattice and imaged, as described above. The imaging leaves the atoms at a temperature of 20 $\mu$K, which is too warm to efficiently transfer into the relatively shallow tweezers of the target array. Following each image, we apply Doppler and Raman sideband cooling to reach an average of $\sim$ 0.1 motional quanta in each direction (Appendix A). Once cooled, the atoms are transferred back into the tweezer arrays, and a single optical tweezer derived from the same laser as the reservoir light is used to transfer atoms from filled sites of the reservoir array into empty sites of the target array. This completes a single loading cycle. In order to shorten the time required to build up an atomic array, and to limit the losses associated with background gas collisions, we load the MOT at the same time as performing operations in the science chamber. This is made possible by the spatial separation between MOT and science regions, and by the fact that we operate with static magnetic fields: we can simultaneously operate with the magnetic field gradients required by the MOT and the large 500 Gauss homogeneous bias field that is desirable for imaging and other operations required in the science regions. The low scattering rate of 399 nm photons, combined with the large magnetic field and spatial separation between the MOT and science regions prevents light scattered in the MOT from being reabsorbed by atoms in the science region, which could cause heating or decoherence. Figure 3b shows examples of the atom number increasing as a function of load cycles, with 105 sites in the reservoir and 1225 sites in the target array. At first, the atom number increases approximately linearly as roughly 45 atoms are transferred into the target array per cycle. The loading rate decreases slightly as the array fills, as some atoms from the reservoir are used to counteract loss. Once the number of vacancies in the target array drops below the typical number of atoms in the reservoir, an equilibrium is reached with the vacancy fraction set by the per cycle probability of atom loss. In figure 3b, the atom number is inferred from the image taken before rearrangement, so the measured fill fraction is sensitive to all losses that occur between such images. This fill fraction is typically 98%, with the largest single loss contribution coming from vacuum loss during the typical 300 ms time between images due to our approximately 30 s vacuum lifetime. We can add additional diagnostic images directly after rearrangement to determine the atom number available for computation, and observe typical filling fractions of 99%, again dominated by vacuum loss during the 150 ms typically allocated for rearrangement. Our experimental control allocates a variable duration for rearrangement that depends on the number and length of moves to be performed, as well as the time required to compute the moves. The durations above are typical for the 1225-site arrays studied in this work. Other sources of loss or potential loss are described in more detail in Appendix B, and include the handoff between the tweezer arrays and the lattice, and imaging loss and infidelity. ## IV Conclusions and outlooks We have demonstrated an iterative method for assembling large arrays of individually controlled neutral atoms, suitable for quantum computation, simulation, and metrology. Through this approach, we decouple the final size of the array from the number of atoms that can be loaded at once, and the need to create many tweezer traps with sufficient depth for loading and imaging atoms. In particular, our reservoir and lattice traps are much deeper than the target tweezer traps, allowing for the use of large numbers of sites with moderate laser power. Our current arrays have percent-level defects, which could be straightforwardly reduced with improved vacuum. For quantum computing applications, our iterative loading may be combined with recently demonstrated mid-circuit measurement techniques [25, 26] in order to achieve continuous mid-circuit refilling of array defects. This capability will be critical for mitigating the effect of atom loss during execution of complex error-corrected circuits that may extend beyond the lifetime of an individual atom in the array. ## V Appendix A: Raman Sideband Cooling Figure 4: Raman sideband cooling (RSC). (a) Diagram of relevant directions and polarizations. RSC beams RSC1 and RSC2 are combined to cool the $z$ direction, while RSC1 and RSC3 are combined to cool the $x$ and $y$ directions. Optical pumping (beam OP) is applied along the $x$ direction, which is along the applied magnetic field. (b) Level diagram for Raman sideband cooling. Raman transitions detuned by 35 MHz from 3P1 $F=3/2$ $m_{f}=1/2$ are driven from 1S0 $m_{f}=1/2$ to 1S0 $m_{f}=-1/2$, while reducing the number of motional quanta along the addressed direction(s) from $n$ to $n-1$. Optical pumping through 3P1 $F=1/2$ $m_{f}=1/2$ returns atoms to 1S0 $m_{f}=1/2$ to enable further cycles. (c) Sideband spectra in the $x$, $y$ plane. The sideband imbalance indicates an average number of motional quanta along the differential momentum vector of the RSC1 and RSC3 beams of $\bar{n}_{xy}=0.08(1)$. (d) Sideband spectra in the $z$ direction. The sideband imbalance indicates an average number of motional quanta along the $z$ direction of $\bar{n}_{z}=0.12(1)$. (e) Heating in the lattice, as measured from the sideband imbalance in the $x$, $y$ plane, with a fit to an exponential growth profile, returning a time- constant of 190(40) ms. Transferring atoms from the lattice into shallow tweezer traps with low loss requires that the atoms be much colder than the depth of the tweezer traps. By operating with colder atoms, we can use shallower tweezers, and in turn scale to larger array sizes. To accomplish this, we use Raman sideband cooling (RSC) between the two ground nuclear spin states [21, 25]. For the motional Raman transitions, we use two pairs of beams oriented along the $x$ and $y$ directions, and along the $x$ and $z$ directions to provide cooling along all three directions (fig. 4a). The Raman transitions transfer atoms from 1S0 $m_{f}=1/2$ to 1S0 $m_{f}=-1/2$, state, and are detuned from the motional carrier transition by the appropriate trap oscillation frequency in order to reduce the motional state by one quanta (fig. 4b). Optical pumping is provided by a beam oriented along $x$ and the magnetic field that addresses the 1S0 $m_{f}=-1/2$ to 1P1 $F=1/2$, $m_{f}=1/2$ transition. We perform 20 iterations of cooling, with each iteration consisting of a $\pi$ pulse on the red motional sideband for each pair, followed by optical pumping. The pulse durations are 200 $\mu$s for the $xy$ pair and 100 $\mu$s for the $xz$ pair (tuned to the $z$ direction sidband), and 50 $\mu$s for optical pumping. The total cooling sequence lasts 8 ms. In typical sequences, we apply a 2 ms Doppler cooling pulse prior to the RSC sequence using the same beam as for the imaging pulse, but with lower intensity and farther red-detuning. The Doppler cooling results in 10 $\mu$K temperature in the cavity as measured by a release and recapture protocol [37]. This roughly corresponds to 1 motional quanta occupancy in the cavity. Subsequent RSC reduces the average motional quanta to in each direction to $\sim$ 0.1, as evident from the sideband imbalance following cooling (fig. 4c, d). We measure heating in the $x$, $y$ plane while holding atoms in the lattice by performing Raman sideband spectroscopy to extract the average motional quantum number $\bar{n}_{xy}$ as a function of time. We fit this quantity to a function representing exponential growth, as expected for parametric heating due to intensity fluctuations of the lattice [38], and extract an exponential time-constant of 190(40) ms. This is likely due to conversion of laser frequency noise to amplitude by the cavity resonance, and could be improved by optimizing the lock of lasers to the cavity. ## VI Appendix B: Sources of Per-Cycle Loss Figure 5: Characterization of per-cycle loss mechanisms. (a), (b). Loss per pair of handoffs between the 459 nm (a) and 423 nm (b) tweezer arrays and lattice, measured by performing 25 handoffs with cooling in between. The black line represents the estimated contribution from vacuum loss, with the red band representing day-to-day drifts in this value. Tweezer depths are normalized to their default values, which correspond to roughly 50 $\mu$K. (c) Final vacancy fraction versus rearrangment tweezer power, normalized to the default tweezer depth of approximately 150 $\mu$K. For low powers, atoms are not efficiently transferred from the reservoir to target array. For too high of powers, we observe increased loss of loaded atoms due to the rearrangement tweezer passing nearby. Vacuum loss due to the finite duration of loading cycles is the single largest source of loss in our system, contributing percent-level per-cycle loss. Because our loading cycle is of variable length, and because the vacuum level in our system can fluctuate from day to day, it is difficult to estimate its exact contribution. However, typical values for the loading cycle duration and vacuum lifetime are 300 ms and 30 s respectively, so 1% represents a typical value for this loss. In this section, we describe other loss sources that contribute to a lesser degree, especially if parameters are not carefully optimized. In general, the loss mechanisms described below both limit the size of array we can ultimately load (as the reservoir must be large enough to replace lost atoms), and limit the final fill fraction of the array. As described in the main text, typical imaging loss is $2(1)\times 10^{-3}$ where roughly 30 % is the vacuum loss and the rest is the Raman scattering out of 3P1 due to the trap light , and the discrimination infidelity is typically at or below the $10^{-3}$ level. If an atom within the target array is lost during the image after scattering enough photons to be identified as present, the defect will not be filled during the subsequent rearrangement step, leading to a defect in the array. If a site within the array is incorrectly identified as empty, an atom will likely be added to that site from the reservoir, and subsequently undergo lossy collisions with the original occupant. Empty sites within the target or reservoir arrays that are mistakenly identified as full will lead to a previous defect not being repaired. However, because defects are rare, this mechanism is less problematic. We currently operate with a discrimination threshold that balances the correct identification of empty and full sites. In a system where imaging losses and infidelity become dominant loss sources (a system with better vacuum), it may be advantageous to bias the threshold to minimize the more problematic forms of imaging error. Each image requires handing off atoms from the tweezer arrays into the lattice and back. We find that this process is sensitive to the alignment of the two arrays on the scale of a single lattice site, to the depth of the tweezer arrays, and to the temperature of the atoms. Atom temperature and alignment are described in Appendix A and C, respectively. We isolate the effect of trap depth on handoff for cold atoms and well-aligned arrays by performing 25 subsequent handoff pairs between tweezers and lattice in figure 5a, b. We perform our usual Raman sideband cooling (RSC) each time the atoms are in the lattice between handoffs. For 459 nm tweezer depths below 75% of our typical operating conditions, we observe significant loss. Above this power level, we observe a constant loss rate of 0.0006(1), which is consistent with data taken with the handoff omitted, and with our typical vacuum losses. The 423 nm tweezers show a similar behavior, though with a slightly higher loss rate of 0.0015(2) at high powers, likely due to worse alignment or optical aberrations. If the tweezers are aligned to the trough of the XY cavity intensity, the loss can be as high as 0.02. In principle, moving new atoms in the target array may cause loss of existing atoms in the array, as the rearrangement tweezer moves near the occupied target sites. We assess this possibility in figure 5c by measuring the final fill fraction versus the depth of the rearrangement tweezer. For a rearrangement tweezer that is too shallow, we observe a lower loading rate and so a lower final fill fraction. For too deep a tweezer, we also see a reduction in the final fill fraction, as the rearrangement leads to loss of existing atoms. Near our typical operating conditions, we observe a region with weak dependency of the fill fraction on the tweezer power. ## VII Appendix C: Alignment of arrays Figure 6: Alignment of the tweezer arrays to the cavity lattice is performed by monitoring atomic survival after repeated handoffs, while scanning the position offset of the tweezer array. For the representative alignment scan of the 459 nm tweezer array shown here, we perform 25 handoff pairs between the tweezers and lattice, with no cooling in between. Data is averaged over the full utilized field of view – the presence of a visible fringe in the average image indicates that the sites across the array can be aligned simultaneously. Precise alignment of the different optical potentials is critical for the performance of our repeated loading protocol. The alignment of the target tweezer arrays to the optical lattice is particularly sensitive, as misalignment here can lead to increased handoff loss that can limit both the largest array that can be loaded, and the final fill fraction of the array. Alignment of the arrays involves first matching the spacing and tilt of the lattice – quantities which do not drift appreciably over time – and then at a higher frequency aligning the $x$ and $y$ offsets. We obtain the correct spacings and tilts by populating the lattice with atoms transferred from an expanded but sparse version of the reservoir array. We image the atoms in the lattice to determine the lattice grid, and then in the reservoir tweezers to determine required corrections to the tweezer array. We then use a camera whose imaging system is corrected for chromatic shifts between the different tweezer wavelengths to register the target array tweezers to the reservoir tweezers. We then optimize the translation alignment by performing repeated handoffs between the tweezers and the lattice and scanning the position offset of the tweezer arrays. We observe a periodic modulation in the atomic survival (fig. 6), which we fit to determine the optimal alignment. This process is repeated as necessary to ensure alignment of the arrays. ## References * Gross and Bloch [2017] C. Gross and I. Bloch, Quantum simulations with ultracold atoms in optical lattices, Science 357, 995 (2017). * Kaufman and Ni [2021] A. M. Kaufman and K.-K. Ni, Quantum science with optical tweezer arrays of ultracold atoms and molecules, Nature Physics 17, 1324 (2021). * Kim _et al._ [2016] H. Kim, W. Lee, H.-g. Lee, H. Jo, Y. Song, and J. Ahn, In situ single-atom array synthesis using dynamic holographic optical tweezers, Nature Communications 7, 13317 (2016). * Barredo _et al._ [2016] D. Barredo, S. De Léséleuc, V. Lienhard, T. Lahaye, and A. Browaeys, An atom-by-atom assembler of defect-free arbitrary two-dimensional atomic arrays, Science 354, 1021 (2016). * Endres _et al._ [2016] M. Endres, H. Bernien, A. Keesling, H. Levine, E. R. Anschuetz, A. Krajenbrink, C. Senko, V. Vuletic, M. Greiner, and M. D. Lukin, Atom-by-atom assembly of defect-free one-dimensional cold atom arrays, Science 354, 1024 (2016). * Scholl _et al._ [2021] P. Scholl, M. Schuler, H. J. Williams, A. A. Eberharter, D. Barredo, K.-N. Schymik, V. Lienhard, L.-P. Henry, T. C. Lang, T. Lahaye, _et al._ , Quantum simulation of 2d antiferromagnets with hundreds of Rydberg atoms, Nature 595, 233 (2021). * Ebadi _et al._ [2021] S. Ebadi, T. T. Wang, H. Levine, A. Keesling, G. Semeghini, A. Omran, D. Bluvstein, R. Samajdar, H. Pichler, W. W. Ho, _et al._ , Quantum phases of matter on a 256-atom programmable quantum simulator, Nature 595, 227 (2021). * Wilk _et al._ [2010] T. Wilk, A. Gaëtan, C. Evellin, J. Wolters, Y. Miroshnychenko, P. Grangier, and A. Browaeys, Entanglement of two individual neutral atoms using Rydberg blockade, Phys. Rev. Lett. 104, 010502 (2010). * Isenhower _et al._ [2010] L. Isenhower, E. Urban, X. L. Zhang, A. T. Gill, T. Henage, T. A. Johnson, T. G. Walker, and M. Saffman, Demonstration of a neutral atom controlled-NOT quantum gate, Phys. Rev. Lett. 104, 010503 (2010). * Madjarov _et al._ [2020] I. S. Madjarov, J. P. Covey, A. L. Shaw, J. Choi, A. Kale, A. Cooper, H. Pichler, V. Schkolnik, J. R. Williams, and M. Endres, High-fidelity entanglement and detection of alkaline-earth Rydberg atoms, Nature Physics 16, 857 (2020). * Schine _et al._ [2022] N. Schine, A. W. Young, W. J. Eckner, M. J. Martin, and A. M. Kaufman, Long-lived bell states in an array of optical clock qubits, Nature Physics 18, 1067 (2022). * Bluvstein _et al._ [2023] D. Bluvstein, S. J. Evered, A. A. Geim, S. H. Li, H. Zhou, T. Manovitz, S. Ebadi, M. Cain, M. Kalinowski, D. Hangleiter, _et al._ , Logical quantum processor based on reconfigurable atom arrays, Nature , 1 (2023). * Young _et al._ [2022] A. W. Young, W. J. Eckner, N. Schine, A. M. Childs, and A. M. Kaufman, Tweezer-programmable 2d quantum walks in a Hubbard-regime lattice, Science 377, 885 (2022). * Young _et al._ [2023] A. W. Young, S. Geller, W. J. Eckner, N. Schine, S. Glancy, E. Knill, and A. M. Kaufman, An atomic boson sampler, arXiv preprint arXiv:2307.06936 (2023). * Young _et al._ [2020] A. W. Young, W. J. Eckner, W. R. Milner, D. Kedar, M. A. Norcia, E. Oelker, N. Schine, J. Ye, and A. M. Kaufman, Half-minute-scale atomic coherence and high relative stability in a tweezer clock, Nature 588, 408 (2020). * Tao _et al._ [2023] R. Tao, M. Ammenwerth, F. Gyger, I. Bloch, and J. Zeiher, High-fidelity detection of large-scale atom arrays in an optical lattice, arXiv preprint arXiv:2309.04717 (2023). * Bloom _et al._ [2014] B. Bloom, T. Nicholson, J. Williams, S. Campbell, M. Bishof, X. Zhang, W. Zhang, S. Bromley, and J. Ye, An optical lattice clock with accuracy and stability at the 10-18 level, Nature 506, 71 (2014). * Park _et al._ [2022] A. J. Park, J. Trautmann, N. Šantić, V. Klüsener, A. Heinz, I. Bloch, and S. Blatt, Cavity-enhanced optical lattices for scaling neutral atom quantum technologies to higher qubit numbers, PRX Quantum 3, 030314 (2022). * Schlosser _et al._ [2001] N. Schlosser, G. Reymond, I. Protsenko, and P. Grangier, Sub-Poissonian loading of single atoms in a microscopic dipole trap, Nature 411, 1024 (2001). * Brown _et al._ [2019] M. O. Brown, T. Thiele, C. Kiehl, T.-W. Hsu, and C. A. Regal, Gray-molasses optical-tweezer loading: Controlling collisions for scaling atom-array assembly, Phys. Rev. X 9, 011057 (2019). * Jenkins _et al._ [2022] A. Jenkins, J. W. Lis, A. Senoo, W. F. McGrew, and A. M. Kaufman, Ytterbium nuclear-spin qubits in an optical tweezer array, Phys. Rev. X 12, 021027 (2022). * Pause _et al._ [2024] L. Pause, L. Sturm, M. Mittenbühler, S. Amann, T. Preuschoff, D. Schäffner, M. Schlosser, and G. Birkl, Supercharged two-dimensional tweezer array with more than 1000 atomic qubits, Optica 11, 222 (2024). * Pause _et al._ [2023] L. Pause, T. Preuschoff, D. Schäffner, M. Schlosser, and G. Birkl, Reservoir-based deterministic loading of single-atom tweezer arrays, Physical Review Research 5, L032009 (2023). * Dick [1989] G. J. Dick, Local oscillator induced instabilities in trapped ion frequency standards, in _Proceedings of the 19th Annual Precise Time and Time Interval Systems and Applications Meeting_ (1989) pp. 133–147. * Lis _et al._ [2023] J. W. Lis, A. Senoo, W. F. McGrew, F. Rönchen, A. Jenkins, and A. M. Kaufman, Midcircuit operations using the omg architecture in neutral atom arrays, Phys. Rev. X 13, 041035 (2023). * Norcia _et al._ [2023] M. A. Norcia, W. B. Cairncross, K. Barnes, P. Battaglino, A. Brown, M. O. Brown, K. Cassella, C.-A. Chen, R. Coxe, D. Crow, J. Epstein, C. Griger, A. M. W. Jones, H. Kim, J. M. Kindem, J. King, S. S. Kondov, K. Kotru, J. Lauigan, M. Li, M. Lu, E. Megidish, J. Marjanovic, M. McDonald, T. Mittiga, J. A. Muniz, S. Narayanaswami, C. Nishiguchi, R. Notermans, T. Paule, K. A. Pawlak, L. S. Peng, A. Ryou, A. Smull, D. Stack, M. Stone, A. Sucich, M. Urbanek, R. J. M. van de Veerdonk, Z. Vendeiro, T. Wilkason, T.-Y. Wu, X. Xie, X. Zhang, and B. J. Bloom, Midcircuit qubit measurement and rearrangement in a ${}^{171}\mathrm{Yb}$ atomic array, Phys. Rev. X 13, 041034 (2023). * Singh _et al._ [2022] K. Singh, S. Anand, A. Pocklington, J. T. Kemp, and H. Bernien, Dual-element, two-dimensional atom array with continuous-mode operation, Physical Review X 12, 011040 (2022). * Singh _et al._ [2023] K. Singh, C. Bradley, S. Anand, V. Ramesh, R. White, and H. Bernien, Mid-circuit correction of correlated phase errors using an array of spectator qubits, Science 380, 1265 (2023). * Anand _et al._ [2024] S. Anand, C. E. Bradley, R. White, V. Ramesh, K. Singh, and H. Bernien, A dual-species rydberg array, arXiv preprint arXiv:2401.10325 (2024). * Sebby-Strabley _et al._ [2006] J. Sebby-Strabley, M. Anderlini, P. S. Jessen, and J. V. Porto, Lattice of double wells for manipulating pairs of cold atoms, Phys. Rev. A 73, 033605 (2006). * Heinz _et al._ [2021] A. Heinz, J. Trautmann, N. Šantić, A. J. Park, I. Bloch, and S. Blatt, Crossed optical cavities with large mode diameters, Optics Letters 46, 250 (2021). * Cai _et al._ [2020] Y. Cai, D. G. Allman, J. Evans, P. Sabharwal, and K. C. Wright, Monolithic bowtie cavity traps for ultracold gases, JOSA B 37, 3596 (2020). * Saskin _et al._ [2019] S. Saskin, J. T. Wilson, B. Grinkemeyer, and J. D. Thompson, Narrow-line cooling and imaging of ytterbium atoms in an optical tweezer array, Phys. Rev. Lett. 122, 143002 (2019). * Huie _et al._ [2023] W. Huie, L. Li, N. Chen, X. Hu, Z. Jia, W. K. C. Sun, and J. P. Covey, Repetitive readout and real-time control of nuclear spin qubits in ${}^{171}\mathrm{Yb}$ atoms, PRX Quantum 4, 030337 (2023). * Lee _et al._ [2015] J. Lee, J. H. Lee, J. Noh, and J. Mun, Core-shell magneto-optical trap for alkaline-earth-metal-like atoms, Physical Review A 91, 053405 (2015). * Höhn _et al._ [2023] T. O. Höhn, E. Staub, G. Brochier, N. D. Oppong, and M. Aidelsburger, State-dependent potentials for the 1S0 and 3P0 clock states of neutral ytterbium atoms, arXiv preprint arXiv:2305.20084 (2023). * Tuchendler _et al._ [2008] C. Tuchendler, A. M. Lance, A. Browaeys, Y. R. P. Sortais, and P. Grangier, Energy distribution and cooling of a single atom in an optical tweezer, Phys. Rev. A 78, 033425 (2008). * Savard _et al._ [1997] T. Savard, K. O’Hara, and J. Thomas, Laser-noise-induced heating in far-off resonance optical traps, Physical Review A 56, R1095 (1997).
# Inference for a Large Directed Acyclic Graph with Unspecified Interventions Chunlin Li<EMAIL_ADDRESS> School of Statistics, University of Minnesota, Minneapolis, MN 55455, USA Xiaotong Shen<EMAIL_ADDRESS> School of Statistics, University of Minnesota, Minneapolis, MN 55455, USA Wei Pan<EMAIL_ADDRESS> Division of Biostatistics, University of Minnesota, Minneapolis, MN 55455, USA ###### Abstract Inference of directed relations given some unspecified interventions, that is, the target of each intervention is unknown, is challenging. In this article, we test hypothesized directed relations with unspecified interventions. First, we derive conditions to yield an identifiable model. Unlike classical inference, testing directed relations requires identifying ancestors and relevant interventions of hypothesis-specific primary variables. Towards this end, we propose a peeling algorithm based on nodewise regressions to establish a topological order of primary variables. Moreover, we prove that the peeling algorithm yields a consistent estimator in low order polynomial time. Second, we propose a likelihood ratio test integrated with a data perturbation scheme to account for the uncertainty of identifying ancestors and interventions. Also, we show that the distribution of a data perturbation test statistic converges to the target distribution. Numerical examples demonstrate the utility and effectiveness of the proposed methods, including an application to infer gene regulatory networks. The implementation of the proposed methods is available at https://github.com/chunlinli/intdag. Keywords: High-dimensional inference, data perturbation, structure learning, identifiability, peeling algorithm ## 1 Introduction Directed relations are essential to explaining pairwise dependencies among multiple interacting units. In gene network analysis, regulatory gene-to-gene relations are a focus of biological investigation (Sachs et al., 2005), while in a human brain network, scientists investigate causal influences among regions of interest to understand how the brain functions through the regional effects of neurons (Liu et al., 2017). In such a situation, inferring directed effects without other information is generally impossible because of the lack of identifiability in a Gaussian directed acyclic graph (DAG) model, and hence external interventions are introduced to treat a non-identifiable situation (Heinze-Deml et al., 2018). For instance, the genetic variants such as single- nucleotide polymorphisms (SNPs) can be, and indeed are increasingly, treated as external interventions to infer inter-trait causal relations in a quantitative trait network (Brown and Knowles, 2020) and gene interactions in a gene regulatory network (Teumer, 2018; Molstad et al., 2021). In neuroimaging analysis, scientists use randomized experimental stimuli as interventions to identify causal relations in a functional brain network (Grosse-Wentrup et al., 2016; Bergmann and Hartwigsen, 2021). However, the interventions in these studies often have unknown targets and off-target effects (Jackson et al., 2003; Eaton and Murphy, 2007). Consequently, inferring directed relations while identifying useful interventions for inference is critical. This paper focuses on the simultaneous inference of directed relations subject to unspecified interventions, that is, without known targets. In a DAG model, the research has been centered on the reconstruction of directed relations in observational and interventional studies (van de Geer and Bühlmann, 2013; Oates et al., 2016; Heinze-Deml et al., 2018; Zheng et al., 2018; Yuan et al., 2019). For inference, Bayesian methods (Friedman and Koller, 2003; Luo and Zhao, 2011; Viinikka et al., 2020) have been popular. Yet, statistical inference remains under-studied, especially for interventional models in high dimensions (Peters et al., 2016; Rothenhäusler et al., 2019). Recently, for observational data, Janková and van de Geer (2018) propose a debiased test of the strength of a single directed relation and Li et al. (2019) derives a constrained likelihood ratio test for multiple directed relations. Despite progress, challenges remain. First, inferring directed relations requires identifying a certain DAG topological order (van de Geer and Bühlmann, 2013), while the identifiability in a Gaussian DAG with unspecified interventions remains under-explored. Second, the inferential results should agree with the acyclicity requirement for a DAG. As a result, degenerate and intractable situations can occur, making inference greatly different from the classical ones. Third, likelihood-based methods for learning the DAG topological order often use permutation search (van de Geer and Bühlmann, 2013) or continuous optimization subject to the acyclicity constraint (Zheng et al., 2018; Yuan et al., 2019), where a theoretical guarantee of the actual estimate (instead of the global optimum) has not been established for these approaches. Recently, an important line of work (Ghoshal and Honorio, 2018; Rajendran et al., 2021; Rolland et al., 2022) has focused on order-based algorithms with computational and statistical guarantees. However, in Gaussian DAGs, existing methods often rely on some error scale assumptions (Peters and Bühlmann, 2014), which is sensitive to variable scaling like the common practice of standardizing variables. This drawback could limit their applications, especially in causal inference, as causal relations are typically invariant to scaling. To address the above issues, we develop structure learning and inference methods for a Gaussian DAG with unspecified additive interventions. Unlike the existing approach where structure discovery and subsequent inference are treated separately, our proposal integrates DAG structure learning and testing directed relations (also known as the natural direct effect in causal inference), accounting for the uncertainty of structure learning for inference. With suitable interventions called instrumental variables (IVs), the proposed approach removes the restrictive error scale assumptions and delivers creditable outcomes with theoretical guarantees in low order polynomial time. This result indicates IVs, a well-known tool in causal inference, can play important roles in structure learning even if some interventions do not meet the IV criterion. More specifically, our contributions are summarized as the following aspects. * • For modeling, we establish the identifiability conditions for a Gaussian DAG with unspecified interventions. In particular, the conditions allow interventions on more than one target, which is suitable for multivariate causal analysis (Murray, 2006). Moreover, we introduce the concepts of nondegeneracy and regularity for hypothesis testing under a DAG model to characterize the behavior of a test. * • For methodology, we develop likelihood ratio tests for directed edges and pathways in a super-graph of the true DAG, where the super-graph is formed by ancestral relations and candidate interventional relations, offering the topological order for inference. Furthermore, we reconstruct the super-graph by the peeling algorithm, which automatically meets the acyclicity requirement. By integrating structure learning with inference, we account for the uncertainty of the super-graph estimation for the proposed tests via a novel data perturbation (DP) scheme, which effectively controls the type-I error while enjoying high statistical power. * • For theory, we prove that the proposed peeling algorithm based on nodewise regressions yields consistent results in $O(p\times\log\kappa_{\max}^{\circ}\times(q^{3}+nq^{2}))$ operations almost surely, where $p,q$ are the numbers of primary and intervention variables, $n$ is the sample size, and $\kappa_{\max}^{\circ}$ is the sparsity. On this basis, we justify the proposed DP inference method by establishing the convergence of the DP likelihood ratio to the null distribution and desired power properties. * • The numerical studies and real data analysis demonstrate the utility and effectiveness of the proposed methods. The implementation of the proposed tests and structure learning method is available at https://github.com/chunlinli/intdag. The rest of the article is structured as follows. Section 2 establishes model identifiability and introduces the concepts of nondegeneracy and regularity. Section 3 develops the proposed methods for structure learning and statistical inference. Section 4 presents statistical theory to justify the proposed methods. Section 5 performs simulation studies, followed by an application to infer gene pathways with gene expression and SNP data in Section 6. Section 7 concludes the article. The Appendix contains illustrative examples and technical proofs. ## 2 Gaussian directed acyclic graph with additive interventions To infer directed relations among primary variables (variables of primary interest) $\bm{Y}=(Y_{1},\ldots,Y_{p})^{\top}$, consider a structural equation model with additive interventions: $\displaystyle\bm{Y}=\bm{U}^{\top}\bm{Y}+\bm{W}^{\top}\bm{X}+\bm{\varepsilon},\quad\bm{\varepsilon}\sim N(\bm{0},\bm{\Sigma}),\quad\bm{\Sigma}=\text{Diag}(\sigma^{2}_{1},\ldots,\sigma^{2}_{p}),$ (1) where $\bm{X}=(X_{1},\ldots,X_{q})^{\top}$ is a vector of additive intervention variables, $\bm{U}\in\mathbb{R}^{p\times p}$ and $\bm{W}\in\mathbb{R}^{q\times p}$ are unknown coefficient matrices, and $\bm{\varepsilon}=(\varepsilon_{1},\ldots,\varepsilon_{p})^{\top}$ is a vector of random errors with $\sigma_{j}^{2}>0$; $j=1,\ldots,p$. In (1), $\bm{\varepsilon}$ is independent of $\bm{X}$ but components of $\bm{X}$ can be dependent. The matrix $\bm{U}$ specifies the directed relations among $\bm{Y}$, where $U_{kj}\neq 0$ if $Y_{k}$ is a direct cause of $Y_{j}$, denoted by $Y_{k}\rightarrow Y_{j}$, and $Y_{k}$ is called a parent of $Y_{j}$ or $Y_{j}$ a child of $Y_{k}$. Thus, $\bm{U}$ represents a directed graph, which is required to be acyclic to ensure the validity of the local Markov property (Spirtes et al., 2000). The matrix $\bm{W}$ specifies the targets and strengths of interventions, where $W_{lj}\neq 0$ indicates $X_{l}$ intervenes on $Y_{j}$, denoted by $X_{l}\to Y_{j}$. In (1), no directed edge from a primary variable $Y_{j}$ to an intervention variable $X_{l}$ is permissible. In what follows, we will focus on the DAG $G$ with nodes $\bm{Y}=(Y_{1},\ldots,Y_{p})$, primary variable edges $\mathcal{E}=\\{(k,j):U_{kj}\neq 0\\}$, and intervention edges $\mathcal{I}=\\{(l,j):W_{lj}\neq 0\\}$. ### 2.1 Identifiability Model (1) is generally non-identifiable without interventions ($\bm{W}=\bm{0}$) when errors do not meet some requirements such as the equal- variance assumption (Peters and Bühlmann, 2014) and its variants (Ghoshal and Honorio, 2018; Rajendran et al., 2021). Moreover, a model can be identified when $\bm{\varepsilon}$ in (1) is replaced by non-Gaussian errors (Shimizu et al., 2006) or linear relations are replaced by nonlinear ones (Hoyer et al., 2009). Regardless, suitable interventions can make (1) identifiable. When intervention targets are known, the identifiability issue has been studied (Oates et al., 2016; Chen et al., 2018). However, it is less so when the exact targets and strengths of interventions are unknown, as in many biological applications (Jackson et al., 2003; Kulkarni et al., 2006), referring to the case of _unspecified_ or _uncertain_ interventions (Heinze-Deml et al., 2018; Eaton and Murphy, 2007; Squires et al., 2020). We now categorize interventions as instruments and invalid instruments. ###### Definition 1 (Instrument) An intervention is an instrument in a DAG if it satisfies: 1. (A) (Relevance) it intervenes on at least one primary variable; 2. (B) (Exclusion) it does not intervene on more than one primary variable. Otherwise, it is an invalid instrument. Here, (A) requires an intervention to be active, while (B) prevents simultaneous interventions of a single intervention variable on multiple primary variables. This is critical to identifiability because in this situation an instrument on a cause variable $Y_{1}$ helps reveal its directed effect on an outcome variable $Y_{2}$, which would otherwise be impossible. Thus, the instruments are introduced to break the symmetry that results in non-identifiability of directed relations $Y_{1}\rightarrow Y_{2}$ and $Y_{2}\rightarrow Y_{1}$. Next, we make some assumptions on intervention variables to yield an identifiable model, where dependencies among intervention variables are permissible. ###### Assumption 1 Assume that model (1) satisfies the following conditions. 1. (1A) (Positive definiteness) $\operatorname{\mathbb{E}}\bm{X}\bm{X}^{\top}$ is positive definite, where $\operatorname{\mathbb{E}}$ is the expectation. 2. (1B) (Local faithfulness) $\operatorname{Cov}(Y_{j},X_{l}\mid\bm{X}_{\\{1,\ldots,q\\}\setminus\\{l\\}})\neq 0$ when $X_{l}$ intervenes on $Y_{j}$, where $\operatorname{Cov}$ denotes the covariance. 3. (1C) (Instrumental sufficiency) Each primary variable is intervened by at least one instrument. Assumption 1A imposes mild distributional restrictions on $\bm{X}$, permitting discrete variables such as SNPs. Assumption 1B requires interventional effects not to cancel out as multiple targets from an invalid instrument are permitted. Importantly, if either Assumption 1B or Assumption 1C fails, model (1) is generally not identifiable, as shown in Example 2 of Appendix A.1. In Section 5, we empirically examine the situation when Assumption 1C is not met. ###### Proposition 1 (Identifiability) Under Assumption 1, the parameters $(\bm{U},\bm{W},\bm{\Sigma})$ in model (1) are identifiable. Proposition 1 (proved in Appendix B.1) is derived for a DAG model with unspecified interventions. This is in contrast to Proposition 1 of Chen et al. (2018), which proves the identifiability of the parameters in a directed graph with target-known instruments on each primary variable. Moreover, it is worth noting that the estimated graph in Chen et al. (2018) may be cyclic and lacks the local Markov property for causal interpretation (Spirtes et al., 2000). ### 2.2 Likelihood-based inference for a DAG Our primary goal is to perform statistical inference of directed edges and pathways. Let $\mathcal{H}\subseteq\\{(k,j):k\neq j,1\leq k,j\leq p\\}$ be an edge set among primary variables $\\{Y_{1},\ldots,Y_{p}\\}$, where $(k,j)\in\mathcal{H}$ specifies a (hypothesized) directed edge $Y_{k}\to Y_{j}$ in (1). We shall focus on two types of testing with null and alternative hypotheses $H_{0}$ and $H_{a}$. For simultaneous testing of directed edges, $\displaystyle H_{0}:U_{kj}=0;\text{ for all }(k,j)\in\mathcal{H}\quad\mbox{ versus }\quad H_{a}:U_{kj}\neq 0\text{ for some }(k,j)\in\mathcal{H};$ (2) for simultaneous testing of pathways, $\displaystyle H_{0}:U_{kj}=0;\text{ for some }(k,j)\in\mathcal{H}\quad\mbox{ versus }\quad H_{a}:U_{kj}\neq 0\text{ for all }(k,j)\in\mathcal{H},$ (3) where $(\bm{U}_{\mathcal{H}^{c}},\bm{W},\bm{\Sigma})$ are unspecified nuisance parameters and c is the set complement. Note that $H_{0}$ in (3) is a composite hypothesis that can be decomposed into sub-hypotheses $H_{0,\nu}:U_{k_{\nu},j_{\nu}}=0,\quad\mbox{ versus }\quad{H}_{a,\nu}:U_{k_{\nu},j_{\nu}}\neq 0;\quad\nu=1,\ldots,|\mathcal{H}|,$ where $\mathcal{H}=\\{(k_{1},j_{1}),\ldots,(k_{|\mathcal{H}|},j_{|\mathcal{H}|})\\}$ and each sub-hypothesis is a directed edge test. Now consider likelihood inference for (2). Given an independent and identically distributed sample $(\bm{Y}_{i},\bm{X}_{i})_{i=1}^{n}$ from (1), the log-likelihood is $L(\bm{\theta},\bm{\Sigma})=-\frac{1}{2}\sum_{i=1}^{n}\|\bm{\Sigma}^{-1/2}((\bm{I}-\bm{U}^{\top})\bm{Y}_{i}-\bm{W}^{\top}\bm{X}_{i})\|^{2}_{2}-n\log\sqrt{\det(\bm{\Sigma})},$ (4) where $\bm{\theta}=(\bm{U},\bm{W})$, $\bm{\Sigma}=\text{Diag}(\sigma_{1}^{2},\ldots,\sigma_{p}^{2})$, and $\bm{U}$ is subject to the acyclicity constraint (Zheng et al., 2018; Yuan et al., 2019) that no directed cycle is permissible in a DAG. The likelihood ratio is defined as $\text{Lr}=L(\widehat{\bm{\theta}}^{(1)},\widehat{\bm{\Sigma}})-L(\widehat{\bm{\theta}}^{(0)},\widehat{\bm{\Sigma}})$, where $\widehat{\bm{\theta}}^{(1)}$ is a maximum likelihood estimate (MLE) in (4) subject to the acyclicity constraint, $\widehat{\bm{\theta}}^{(0)}$ is an MLE subject to an additional requirement $\bm{U}_{\mathcal{H}}=\bm{0}$, and $\widehat{\bm{\Sigma}}$ is an estimator of $\bm{\Sigma}$. In many statistical models, a likelihood ratio has a nondegenerate and tractable distribution, for instance, a chi-squared distribution with degrees of freedom $|\mathcal{H}|$. However, in (1), because of the acyclicity constraint, some hypothesized edges in $\mathcal{H}$ can be absent given the edge subset $\mathcal{E}\cap\mathcal{H}^{c}$ defined by the nonzero elements of nuisance parameter $\bm{U}_{\mathcal{H}^{c}}$, where $\mathcal{E}$ denotes the edge set of the whole DAG. As a result, Lr may converge to a distribution with degrees of freedom less than $|\mathcal{H}|$ and the distribution may be even intractable, making inference for a DAG greatly different from the classical ones, as illustrated by Example 1. Figure 1: A true DAG structure of five primary variables $Y_{1},\ldots,Y_{5}$ and five intervention variables $X_{1},\ldots,X_{5}$, where directed edges are represented by solid arrows while dependencies among $\bm{X}$ are not displayed. ###### Example 1 Consider the likelihood ratio test under null $H_{0}$ and alternative $H_{a}$ for the DAG displayed in Figure 1. Let $\mathcal{E}$ be the edge set of the DAG. * • $H_{0}:U_{21}=0$ versus $H_{a}:U_{21}\neq 0$, where $\mathcal{H}=\\{(2,1)\\}$. Here, $(2,1)$ forms a cycle together with the edges in $\mathcal{E}\cap\mathcal{H}^{c}$ (namely the edges not considered by the hypothesis), violating the acyclicity constraint. Note that the maximum likelihood value $L(\widehat{\bm{\theta}},\widehat{\bm{\Sigma}})$ with $U_{21}\neq 0$ tends to be smaller than that with $U_{21}=0$ when a random sample is obtained under $H_{0}$, especially so when the asymptotics kicks in as the sample size increases. Consequently, the likelihood ratio Lr becomes zero, constituting a degenerate situation. * • $H_{0}:U_{45}=U_{53}=0$ versus $H_{a}:U_{45}\neq 0$ or $U_{53}\neq 0$, where $\mathcal{H}=\\{(4,5),(5,3)\\}$. In this case, $\\{(4,5),(5,3)\\}$ forms a cycle with the edges in $\mathcal{E}\cap\mathcal{H}^{c}$, violating the acyclicity constraint. Since the likelihood value $L({\bm{\theta}},\bm{\Sigma})$ tends to be maximized under the true graph when data is sampled under $H_{0}$, we have $L(\widehat{\bm{\theta}}^{(1)},\widehat{\bm{\Sigma}})=\max(L(\widehat{U}_{45},\widehat{U}_{53}=0,\widehat{\bm{U}}_{\mathcal{H}^{c}},\widehat{\bm{W}},\widehat{\bm{\Sigma}}),L(\widehat{U}_{45}=0,\widehat{U}_{53},\widehat{\bm{U}}_{\mathcal{H}^{c}},\widehat{\bm{W}},\widehat{\bm{\Sigma}}))$. As a result, the likelihood ratio distribution becomes complicated in this situation due to the dependence between the two components in $L(\widehat{\bm{\theta}}^{(1)},\widehat{\bm{\Sigma}})$. Motivated by Example 1, we introduce the concepts of nondegeneracy and regularity. ###### Definition 2 (Nondegeneracy and regularity with respect to DAG) 1. (A) An edge $(k,j)\in\mathcal{H}$ is nondegenerate with respect to DAG $G$ if $\\{(k,j)\\}\cup\mathcal{E}$ contains no directed cycle, where $\mathcal{E}$ denotes the edge set of $G$. Otherwise, $(k,j)$ is degenerate. Let $\mathcal{D}\subseteq\mathcal{H}$ be the set of all nondegenerate edges with respect to $G$. A null hypothesis $H_{0}$ is nondegenerate with respect to DAG $G$ if $\mathcal{D}\neq\emptyset$. Otherwise, $H_{0}$ is degenerate. 2. (B) A null hypothesis $H_{0}$ is said to be regular with respect to DAG $G$ if $\mathcal{D}\cup\mathcal{E}$ contains no directed cycle, where $\mathcal{E}$ denotes the edge set of $G$. Otherwise, $H_{0}$ is called irregular. ###### Remark 1 In practice, $\mathcal{D}$ is unknown and needs to be estimated from data. Nondegeneracy ensures nonnegativity of the likelihood ratio. In testing (2), regularity excludes intractable situations for the null distribution. In testing (3), if $H_{0}$ is irregular, then $\mathcal{D}\cup\mathcal{E}$ has a directed cycle, which means the hypothesized directed pathway cannot exist due to the acyclicity constraint. Thus, regularity excludes the degenerate situations in testing (3). In what follows, we mainly focus on nondegenerate and regular hypotheses. For the degenerate case, the p-value is defined to be one. For the irregular case of edge test (2), we decompose the hypothesis into regular sub-hypotheses and conduct multiple testing. For the irregular case of pathway test (3), the p-value is defined to be one. More discussions on the implementation in irregular cases are provided in Appendix A.5. Finally, we introduce some notations for a DAG $G$ with primary variables $\\{Y_{1},\ldots,Y_{p}\\}$, intervention variables $\\{X_{1},\ldots,X_{q}\\}$, a directed edge set $\mathcal{E}\subseteq\\{(k,j):k\neq j,1\leq k,j\leq p\\}$, and an intervention edge set $\mathcal{I}\subseteq\\{(l,j):1\leq l\leq q,1\leq j\leq p\\}$. If there is a directed path $Y_{k}\to\cdots\to Y_{j}$ in $G$, $Y_{k}$ is called an ancestor of $Y_{j}$ or $Y_{j}$ a descendant of $Y_{k}$. Let $\mathcal{A}=\\{(k,j):Y_{k}\to\cdots\to Y_{j}\\}$ be the set of all ancestral relations. In a DAG $G$, let $\textnormal{{pa}}_{G}(j)=\\{k:(k,j)\in\mathcal{E}\\}$, $\textnormal{{an}}_{G}(j)=\\{k:(k,j)\in\mathcal{A}\\}$, $\textnormal{{d}}_{G}(j)=\\{k:(k,j)\in\mathcal{D}\\}$, and $\textnormal{{in}}_{G}(j)=\\{l:(l,j)\in\mathcal{I}\\}$ be the parent, ancestor, nondegenerate hypothesized parent, and intervention sets of $Y_{j}$, respectively. ## 3 Methodology This section develops the main methodology, including the peeling algorithm for structure learning and the data perturbation inference for simultaneous testing of directed edges (2) and pathways (3). One major challenge to the likelihood approach is optimization subject to the acyclicity requirement. This constraint imposes difficulty on not only computation but also asymptotic theory. As a result, there is a gap between the asymptotic distribution based on a global maximum and that on the actual estimate which could be a local maximum (Janková and van de Geer, 2018; Li et al., 2019). Moreover, the actual estimate may give an imprecise topological order, tending to impact adversely on inference. To circumvent the acyclicity requirement, we propose using a super-graph $S$ of the true DAG $G$, where $S$ is constructed by a novel peeling method without explicitly imposing the acyclicity constraint. As a result, our method is scalable with a statistical guarantee of the actual estimate. We will focus on a specific super-graph $S$ consisting of primary variable edges $\mathcal{E}_{S}:=\mathcal{A}\supseteq\mathcal{E}$ and candidate intervention edges $\mathcal{I}_{S}:=\mathcal{C}\supseteq\mathcal{I}$, where $\mathcal{E}$ and $\mathcal{I}$ are sets of primary variable edges and intervention edges in the original graph $G$, and set $\mathcal{C}$ is to be defined shortly after (6). Note that $S$, whose primary variable edges consist of all ancestral relations of $G$, is also a DAG. Given the super-graph $S$, the true DAG $G$ can be reconstructed by nodewise constrained regressions. Let $\bm{\theta}_{S}=\big{(}\bm{U}_{\mathcal{A}},\bm{W}_{\mathcal{C}}\big{)}$ be $\bm{\theta}$ restricted to $S$. Note that $\bm{\theta}_{S^{c}}=(\bm{U}_{\mathcal{A}^{c}},\bm{W}_{\mathcal{C}^{c}})=\bm{0}$ in the original graph $G$. Given $S$, the log-likelihood $L(\bm{\theta},\bm{\Sigma})$ in (4) can be rewritten as $\begin{split}L_{S}(\bm{\theta}_{S},\bm{\Sigma})&=-\sum_{j=1}^{p}\Big{(}\text{RSS}_{j}(\bm{\theta}_{S})/2\sigma_{j}^{2}+\log\sqrt{2\pi\sigma_{j}^{2}}\Big{)},\\\ \text{RSS}_{j}(\bm{\theta}_{S})&=\sum_{i=1}^{n}\Big{(}Y_{ij}-\sum_{(k,j)\in\mathcal{A}}U_{kj}Y_{ik}-\sum_{(l,j)\in\mathcal{C}}W_{lj}X_{il}\Big{)}^{2},\\\ \end{split}$ (5) which involves $|\mathcal{A}|+|\mathcal{C}|$ parameters for $\bm{\theta}_{S}$. Our plan is as follows. In Section 3.1, we hierarchically construct ${S}$ via nodewise regressions of $Y_{j}$ over $\bm{X}$; $j=1,\ldots,p$ without the acyclicity constraint for $\bm{U}$. Then in Section 3.2, on the ground of constructed super-graph $S$ and (5), we develop likelihood ratio tests for hypotheses (2) and (3). ### 3.1 Structure learning via peeling This section develops a novel structure learning method based on a peeling algorithm and nodewise constrained regressions to construct $S=(\mathcal{A},\mathcal{C})$ in a hierarchical manner. First, we observe an important connection between primary variables and intervention variables. Rewrite (1) as $\bm{Y}=\bm{V}^{\top}\bm{X}+\bm{\varepsilon}_{V},\quad\bm{\varepsilon}_{V}=(\bm{I}-\bm{U}^{\top})^{-1}\bm{\varepsilon}\sim N(\bm{0},\bm{\Omega}^{-1}),$ (6) where $\bm{\Omega}=(\bm{I}-\bm{U})\bm{\Sigma}^{-1}(\bm{I}-\bm{U}^{\top})$ is a precision matrix and $\bm{V}=\bm{W}(\bm{I}-\bm{U})^{-1}$. Now, define the set of candidate interventional relations $\mathcal{C}:=\\{(l,j):V_{lj}\neq 0\\}$. ###### Proposition 2 Suppose Assumption 1 is satisfied. Then $\mathcal{C}\supseteq\mathcal{I}$. Moreover, 1. (A) if $V_{lj}\neq 0$, then $X_{l}$ intervenes on $Y_{j}$ or an ancestor of $Y_{j}$; 2. (B) $Y_{j}$ is a leaf variable (having no child) if and only if there is an instrument $X_{l}$ such that $V_{lj}\neq 0$ and $V_{lj^{\prime}}=0$ for $j^{\prime}\neq j$. The proof of Proposition 2 is deferred to Appendix B.2. Intuitively, $V_{lj}\neq 0$ implies the dependence of $Y_{j}$ on $X_{l}$ through a directed path $X_{l}\to\cdots\to Y_{j}$, and hence that $X_{l}$ intervenes on $Y_{j}$ or an ancestor of $Y_{j}$. Thus, the instruments on a leaf variable are independent of the other primary variables conditional on the rest of interventions. This observation suggests a method to reconstruct the DAG topological order by sequentially identifying the leaves in the DAG and removing the identified leaf variables. Next, we discuss the estimation of $\bm{V}$ and construction of $S$. #### 3.1.1 Nodewise constrained regressions We propose nodewise $\ell_{0}$-constrained regressions to estimate nonzero elements of $\bm{V}$: $\widehat{\bm{V}}_{\cdot j}=\mathop{\rm arg\,min\,}_{\bm{V}_{\cdot j}}\ \sum_{i=1}^{n}(Y_{ij}-\bm{V}_{\cdot j}^{\top}\bm{X}_{i})^{2}\quad\text{s.t.}\quad\sum_{l=1}^{q}\operatorname{I}(V_{lj}\neq 0)\leq\kappa_{j};\quad j=1,\ldots,p,$ (7) where $1\leq\kappa_{j}\leq q$ is an integer-valued tuning parameter controlling the sparsity of $\bm{V}_{\cdot j}$ and can be chosen by BIC or cross-validation. To solve (7), we use $J_{\tau_{j}}(z)=\min(|z|/\tau_{j},1)$ as a surrogate of the $\ell_{0}$-function $\operatorname{I}(z\neq 0)$ (Shen et al., 2012) and develop a difference-of-convex (DC) algorithm with the $\ell_{0}$-projection to improve globality of the solution of (7). At $(t+1)$th iteration, we solve a weighted Lasso problem, $\displaystyle\widetilde{\bm{V}}_{\cdot j}^{[t+1]}=\mathop{\rm arg\,min\,}_{\bm{V}_{\cdot j}}\ \sum_{i=1}^{n}(Y_{ij}-\bm{V}_{\cdot j}^{\top}\bm{X}_{i})^{2}+2n\gamma_{j}\tau_{j}\sum_{l=1}^{q}\operatorname{I}\big{(}|\widetilde{V}^{[t]}_{lj}|\leq\tau_{j}\big{)}|V_{lj}|;\quad j=1,\ldots,p,$ (8) where $\gamma_{j}>0$ is a tuning parameter and $\widetilde{\bm{V}}^{[t]}_{\cdot j}$ is the solution of (8) at the $t$-th iteration. The DC algorithm terminates at $\widetilde{\bm{V}}_{\cdot j}=\widetilde{\bm{V}}_{\cdot j}^{[t]}$ such that $\|\widetilde{\bm{V}}_{\cdot j}^{[t+1]}-\widetilde{\bm{V}}_{\cdot j}^{[t]}\|_{\infty}\leq\sqrt{\text{tol}}$, where tol is the machine precision. Then, we obtain the solution $\widehat{\bm{V}}_{\cdot j}$ of (7) by projecting $\widetilde{\bm{V}}_{\cdot j}$ onto the $\ell_{0}$-constrained set $\\{\|\bm{V}_{\cdot j}\|_{0}\leq\kappa_{j}\\}$. 1 2Specify $\gamma_{j}>0$, $\tau_{j}>0$, and $1\leq\kappa_{j}\leq q$. Initialize $\widetilde{\bm{V}}_{\cdot j}^{[0]}$ with $|\\{l:\widetilde{V}^{[0]}_{lj}>\tau_{j}\\}|\leq\kappa_{j}$. 3Compute $\widetilde{\bm{V}}_{\cdot j}=\widetilde{\bm{V}}^{[t]}_{\cdot j}$ such that $\|\widetilde{\bm{V}}_{\cdot j}^{[t+1]}-\widetilde{\bm{V}}_{\cdot j}^{[t]}\|_{\infty}\leq\sqrt{\text{tol}}$. 4($\ell_{0}$-projection) Let $B_{j}=\\{l:\sum_{l^{\prime}=1}^{q}\operatorname{I}(|\widetilde{V}_{l^{\prime}j}|\geq|\widetilde{V}_{lj}|)\leq\kappa_{j}\\}$. Set $\widehat{\bm{V}}_{\cdot j}=\mathop{\rm arg\,min\,}_{\bm{V}_{\cdot j}}\ \|\bm{\mathsf{Y}}_{j}-\bm{V}_{\cdot j}^{\top}\bm{\mathsf{X}}\|_{2}^{2}\quad\text{s.t.}\quad V_{lj}=0\text{ for }l\notin B_{j}.$ Algorithm 1 Constrained minimization via DC programming + $\ell_{0}$ projection 1 2Let $\widehat{\bm{V}}^{[0]}=\widehat{\bm{V}}$. Begin iteration with $h=0,\ldots$: at iteration $h$, 3In the current sub-graph identify all instrument-leaf pairs $X_{l}\to Y_{j}$ satisfying $l=\mathop{\rm arg\,min\,}_{l^{\prime}:\|\widehat{\bm{V}}^{[h]}_{l^{\prime}\cdot}\|_{0}\neq 0}\|\widehat{\bm{V}}^{[h]}_{l^{\prime}\cdot}\|_{0}$ and $j=\mathop{\rm arg\,max\,}_{j^{\prime}}|\widehat{V}^{[h]}_{lj^{\prime}}|$. Let $h_{\widehat{G}}(j)=h$. 4For a leaf $Y_{k}$, identify all directed relations $Y_{k}\to Y_{j}$ with $h_{\widehat{G}}(j)=h-1$ when $\widehat{V}_{lj}\neq 0$ for each instrument $X_{l}$ of $Y_{k}$ in the current sub-graph. 5Peeling-off all the leaves $Y_{j}$ by deleting $j$th columns from $\widehat{\bm{V}}^{[h]}$ to yield $\widehat{\bm{V}}^{[h+1]}$. 6Repeat Steps 2-4 until all primary variables are removed. 7Compute $\widehat{S}=(\widehat{\mathcal{A}},\widehat{\mathcal{C}})$ and $\widehat{\mathcal{D}}$: $\begin{split}\widehat{\mathcal{A}}&=\Big{\\{}(k,j):Y_{k}\to\cdots\to Y_{j}\Big{\\}},\quad\widehat{\mathcal{D}}=\Big{\\{}(k,j)\in\mathcal{H}:(j,k)\notin\widehat{\mathcal{A}}\Big{\\}},\\\ \widehat{\mathcal{C}}&=\Big{\\{}(l,j):|\widehat{V}_{lj}|>\tau_{j}\Big{\\}},\text{ where $\tau_{j}$ is specified in Algorithm \ref{algorithm:nodewise- regression}.}\end{split}$ Algorithm 2 Reconstruction of topological layers by peeling #### 3.1.2 Peeling Now, we describe a _peeling_ algorithm to estimate $S=(\mathcal{A},\mathcal{C})$ based on $\bm{V}$. Before proceeding, we introduce the notion of _topological height_ to characterize the DAG topological order. Given a DAG $G$, the height $h_{G}(j)$ of $Y_{j}$ is defined as the maximal length of a directed path from $Y_{j}$ to a leaf variable, where $0\leq h_{G}(j)\leq p-1$ with a leaf having zero height. Then the primary variables of the same height form a topological layer and the layers constitute a partition of primary variables in $G$. For instance, in Figure 1, $\\{Y_{4},Y_{5}\\},\\{Y_{3}\\},\\{Y_{2}\\},\\{Y_{1}\\}$ are topological layers with height $0,1,2,3$, respectively. Observe that the topological layer of height $h$ are leaves in the sub-graph with primary variables of height no less than $h$. Based on Proposition 2, the peeling algorithm sequentially identifies and peels off the primary variables of height $h$; $h=0,1,2,\ldots$. Note that instruments in the sub-graph may not be necessarily the instruments in the graph $G$ after some primary variables are removed. However, the next proposition indicates that the instruments in the corresponding sub-graph remain useful to reconstruct ancestral relations. ###### Proposition 3 Let $Y_{k}$ have height $h_{G}(k)$. Let $G^{\prime}$ be the sub-graph with primary variables of height $\geq h_{G}(k)$. Under Assumption 1, for a pair $(k,j)$ with $h_{G}(j)=h_{G}(k)-1$, we have $Y_{k}\to Y_{j}$ if and only if $V_{lj}\neq 0$ for each instrument $X_{l}$ of $Y_{k}$ in $G^{\prime}$. Proposition 3 (proved in Appendix B.3) permits the construction of directed relations between the topological layers of heights $h$ and $h-1$; $h\geq 1$ and thus ancestral relations. The peeling algorithm is summarized in Algorithm 2 and a detailed illustration is presented in Example 3 of Appendix A.2. ###### Remark 2 (Recovery of true graph $G$) $\widehat{\mathcal{A}}$ estimates a superset of directed relations of $\bm{U}$ and $\widehat{\mathcal{C}}$ a superset of interventional relations of $\bm{W}$. Therefore, under some technical conditions, with suitable tuning parameters, the true graph $G$ can be reconstructed by estimating $(\bm{U},\bm{W})$ through $\ell_{0}$-constrained regressions given the estimated super-graph $\widehat{S}=(\widehat{\mathcal{A}},\widehat{\mathcal{C}})$, $\begin{split}\min_{(\bm{U}_{\textnormal{{an}}_{\widehat{S}}(j),j},\bm{W}_{\textnormal{{in}}_{\widehat{S}}(j),j})}\sum_{i=1}^{n}\left(Y_{ij}-\bm{U}_{\textnormal{{an}}_{\widehat{S}}(j),j}^{\top}\bm{Y}_{i,\textnormal{{an}}_{\widehat{S}}(j)}-\bm{W}_{\textnormal{{in}}_{\widehat{S}}(j),j}^{\top}\bm{X}_{i,\textnormal{{in}}_{\widehat{S}}(j)}\right)^{2}\\\ \text{s.t.}\quad\sum_{k\in\textnormal{{an}}_{\widehat{S}}(j)}\operatorname{I}(U_{kj}\neq 0)+\sum_{l\in\textnormal{{in}}_{\widehat{S}}(j)}\operatorname{I}(W_{lj}\neq 0)\leq\kappa_{j}^{\prime}.\end{split}$ (9) The final estimates $\widehat{\bm{U}}_{\cdot j}=\Big{(}\widehat{\bm{U}}_{\textnormal{{an}}_{\widehat{S}}(j),j},\bm{0}_{\textnormal{{an}}_{\widehat{S}}(j)^{c}}\Big{)}$ and $\widehat{\bm{W}}_{\cdot j}=\Big{(}\widehat{\bm{W}}_{{\textnormal{{in}}}_{\widehat{S}}(j),j},\bm{0}_{{\textnormal{{in}}}_{\widehat{S}}(j)^{c}}\Big{)}$; $j=1,\ldots,p$. ### 3.2 Likelihood ratio test via data perturbation This subsection proposes an inference method for testing (2) and (3). We first construct a likelihood ratio $\text{Lr}(S,\bm{\Sigma})$ in given $(S,\bm{\Sigma})$ and then plug-in an estimate $(\widehat{S},\widehat{\bm{\Sigma}})$. Next we perform a test via data perturbation, accounting for the uncertainty of constructing $S$. #### 3.2.1 Likelihood ratio given $S$ Consider the test in (2). From (5), the likelihood ratio given $(S,\bm{\Sigma})$ is $\text{Lr}(S,\bm{\Sigma})=L_{S}(\widehat{\bm{\theta}}_{S}^{(1)},\bm{\Sigma})-L_{S}(\widehat{\bm{\theta}}_{S}^{(0)},\bm{\Sigma})=\sum_{j=1}^{p}\frac{\text{RSS}_{j}(\widehat{\bm{\theta}}^{(0)}_{S})-\text{RSS}_{j}(\widehat{\bm{\theta}}^{(1)}_{S})}{2\sigma_{j}^{2}}$ where $\widehat{\bm{\theta}}_{S}^{(1)}$ is the MLE given $(S,\bm{\Sigma})$, and $\widehat{\bm{\theta}}_{S}^{(0)}$ is the MLE given $(S,\bm{\Sigma})$ subject to $\bm{U}_{\mathcal{H}}=0$. Now, recall that $\textnormal{{d}}_{G}(j)=\\{k:(k,j)\in\mathcal{D}\\}$, where $\mathcal{D}$ is the set of nondegenerate edges of $H_{0}$ with respect to $G$ (Definition 2). Of note, $\textnormal{{d}}_{S}(j)=\textnormal{{d}}_{G}(j)$ since an edge is nondegenerate with respect to $S$ if and only if it is nondegenerate with respect to $G$. To eliminate the nuisance parameters, observe that if $\textnormal{{d}}_{S}(j)=\emptyset$, then ${\text{RSS}}_{j}(\widehat{\bm{\theta}}^{(0)}_{S})-{\text{RSS}}_{j}(\widehat{\bm{\theta}}^{(1)}_{S})=0$, because no $H_{0}$ constraint is imposed for $Y_{j}$. Hence, $\text{Lr}(S,\bm{\Sigma})$ only summarizes the contributions from the primary variables with nondegenerate hypothesized edges, $\text{Lr}(S,\bm{\Sigma})=\sum_{j:\textnormal{{d}}_{S}(j)\neq\emptyset}\frac{{\text{RSS}}_{j}(\widehat{\bm{\theta}}^{(0)}_{S})-{\text{RSS}}_{j}(\widehat{\bm{\theta}}^{(1)}_{S})}{2\sigma_{j}^{2}},$ (10) where the nuisance parameters for $Y_{j}$ with $\textnormal{{d}}_{S}(j)=\emptyset$ are eliminated. The likelihood ratio test statistic becomes $\text{Lr}=\text{Lr}(\widehat{S},\widehat{\bm{\Sigma}})$. Here, $\widehat{S}$ is constructed by Algorithm 2 in Section 3.1 and $\widehat{\bm{\Sigma}}=\text{Diag}(\widehat{\sigma}^{2}_{1},\ldots,\widehat{\sigma}^{2}_{p})$, $\displaystyle\widehat{\sigma}^{2}_{j}=\frac{\text{RSS}_{j}(\widehat{\bm{\theta}}^{(1)}_{\widehat{S}})}{n-|{\textnormal{{an}}}_{\widehat{S}}(j)|-|{\textnormal{{in}}}_{\widehat{S}}(j)|},\quad j=1,\ldots,p,$ (11) where ${\textnormal{{an}}}_{\widehat{S}}(j)$ and ${\textnormal{{in}}}_{\widehat{S}}(j)$ are the estimated ancestor and candidate intervention sets of $Y_{j}$. #### 3.2.2 Testing directed edges (2) via data perturbation The likelihood ratio requires an estimation of $S$, where we must account for the uncertainty of $\widehat{S}$ for accurate finite-sample inference. To proceed, we consider the test statistic Lr based on a “correct” super-graph $S\supseteq S^{\circ}$, where $S=(\mathcal{A},\mathcal{C})\supseteq S^{\circ}=(\mathcal{A}^{\circ},\mathcal{C}^{\circ})$ means that $\mathcal{A}\supseteq\mathcal{A}^{\circ}$ and $\mathcal{C}\supseteq\mathcal{C}^{\circ}$ and ∘ denotes the truth. Intuitively, a “correct” super-graph distinguishes descendants and nondescendants and thus can help to infer the true directed relations defined by the local Markov property (Spirtes et al., 2000) without introducing model errors, yet may lead to a less powerful test when $S$ is much larger than $S^{\circ}$. By comparison, a “wrong” graph $S\not\supseteq S^{\circ}$ provides an incorrect topological order, and a test based on a “wrong” graph may be biased, accompanied by an inflated type-I error. Now, let $\bm{\mathsf{Z}}_{n\times(p+q)}=(\bm{\mathsf{X}},\bm{\mathsf{Y}})$ be the data matrix, where $\bm{\mathsf{Y}}_{n\times p}=(\bm{Y}_{1},\ldots,\bm{Y}_{n})^{\top}$ and $\bm{\mathsf{X}}_{n\times q}=(\bm{X}_{1},\ldots,\bm{X}_{n})^{\top}$. Moreover, let $\bm{\mathsf{e}}_{j}\in\mathbb{R}^{n}$ be the $j$th column of error matrix $\bm{\mathsf{e}}_{n\times p}=(\bm{\varepsilon}_{1},\ldots,\bm{\varepsilon}_{n})^{\top}$. From (10), assuming $\widehat{S}\supseteq S^{\circ}$ is a “correct” super-graph, the likelihood ratio becomes $\begin{split}\text{Lr}&=\sum_{j:{\textnormal{{d}}}_{\widehat{S}}(j)\neq\emptyset}\frac{\bm{\mathsf{Y}}_{j}^{\top}(\bm{\mathsf{P}}_{\widehat{A}_{j}}-\bm{\mathsf{P}}_{\widehat{B}_{j}})\bm{\mathsf{Y}}_{j}}{2\widehat{\sigma}_{j}}\\\ &=\sum_{j:{\textnormal{{d}}}_{\widehat{S}}(j)\neq\emptyset}\frac{(\bm{\mathsf{Y}}_{{\textnormal{{d}}}_{\widehat{S}}(j)}\bm{U}_{{\textnormal{{d}}}_{\widehat{S}}(j),j}+\bm{\mathsf{e}}_{j})^{\top}(\bm{\mathsf{P}}_{\widehat{A}_{j}}-\bm{\mathsf{P}}_{\widehat{B}_{j}})(\bm{\mathsf{Y}}_{{\textnormal{{d}}}_{\widehat{S}}(j)}\bm{U}_{{\textnormal{{d}}}_{\widehat{S}}(j),j}+\bm{\mathsf{e}}_{j})}{2\widehat{\sigma}_{j}},\\\ \widehat{\sigma}_{j}&=\frac{\bm{\mathsf{Y}}_{j}^{\top}(\bm{I}-\bm{\mathsf{P}}_{\widehat{A}_{j}})\bm{\mathsf{Y}}_{j}}{n-|\widehat{A}_{j}|}=\frac{\bm{\mathsf{e}}_{j}^{\top}(\bm{I}-\bm{\mathsf{P}}_{\widehat{A}_{j}})\bm{\mathsf{e}}_{j}}{n-|\widehat{A}_{j}|};\quad 1\leq j\leq p,\end{split}$ (12) where $\bm{\mathsf{P}}_{A}=\bm{\mathsf{Z}}_{A}(\bm{\mathsf{Z}}_{A}^{\top}\bm{\mathsf{Z}}_{A})^{-1}\bm{\mathsf{Z}}_{A}^{\top}$ is the projection matrix onto the span of columns $A$ in $\bm{\mathsf{Z}}$, $\widehat{A}_{j}={\textnormal{{an}}}_{\widehat{S}}(j)\cup{\textnormal{{in}}}_{\widehat{S}}(j)\cup{\textnormal{{d}}}_{\widehat{S}}(j)$, and $\widehat{B}_{j}=({\textnormal{{an}}}_{\widehat{S}}(j)\cup{\textnormal{{in}}}_{\widehat{S}}(j))\setminus{\textnormal{{d}}}_{\widehat{S}}(j)$. Of note, $\bm{\mathsf{Y}}_{{\textnormal{{d}}}_{S}(j)}\bm{U}_{{\textnormal{{d}}}_{S}(j),j}=\bm{0}$ for all $j$ under $H_{0}$, while $\bm{\mathsf{Y}}_{{\textnormal{{d}}}_{S}(j)}\bm{U}_{{\textnormal{{d}}}_{S}(j),j}\neq\bm{0}$ for some $j$ under $H_{a}$. We propose a data perturbation (DP) method (Shen and Ye, 2002; Breiman, 1992) to approximate the null distribution of Lr in (12). The idea behind DP is to assess the sensitivity of the estimates through perturbed data $\bm{Y}^{*}_{i}=\bm{Y}_{i}+\bm{\varepsilon}^{*}_{i}$, where $\bm{\varepsilon}_{i}^{*}\sim N(0,\widehat{\bm{\Sigma}})$; $i=1,\ldots,n$. Let $(\bm{\mathsf{Z}}^{*},\bm{\mathsf{e}}^{*})=(\bm{\mathsf{X}},\bm{\mathsf{Y}}^{*},\bm{\mathsf{e}}^{*})$ be perturbed data, where $\bm{\mathsf{e}}^{*}_{n\times p}=(\bm{\mathsf{e}}^{*}_{1},\ldots,\bm{\mathsf{e}}^{*}_{p})=(\bm{\varepsilon}^{*}_{1},\ldots,\bm{\varepsilon}^{*}_{n})^{\top}$ denotes perturbation errors. Given $(\bm{\mathsf{Z}}^{*},\bm{\mathsf{e}}^{*})$, we obtain a perturbation estimate $\widehat{S}^{*}$. In (12), under $H_{0}$, $\text{Lr}=f(\bm{\mathsf{X}},\bm{\mu},\bm{\mathsf{e}})$ is a function of intervention data $\bm{\mathsf{X}}$, conditional mean $\bm{\mu}=\operatorname{\mathbb{E}}(\bm{\mathsf{Y}}\mid\bm{\mathsf{X}})\in\mathbb{R}^{n\times p}$ and unobserved error $\bm{\mathsf{e}}$. However, the perturbation error $\bm{\mathsf{e}}^{*}$ is _known_ given the DP sample $(\bm{\mathsf{Z}}^{*},\bm{\mathsf{e}}^{*})$, suggesting a form $\text{Lr}^{*}$ based on Lr: $\begin{split}\text{Lr}^{*}&\equiv f(\bm{\mathsf{X}},\bm{\mathsf{Y}},\bm{\mathsf{e}}^{*})=\sum_{j:{\textnormal{{d}}}_{\widehat{S}}(j)\neq\emptyset}\frac{(\bm{\mathsf{e}}_{j}^{*})^{\top}(\bm{\mathsf{P}}_{\widehat{A}^{*}_{j}}-\bm{\mathsf{P}}_{\widehat{B}^{*}_{j}})\bm{\mathsf{e}}^{*}_{j}}{2\widetilde{\sigma}_{j}^{*}},\quad\widetilde{\sigma}^{*}_{j}=\frac{(\bm{\mathsf{e}}_{j}^{*})^{\top}(\bm{I}-\bm{\mathsf{P}}_{\widehat{A}^{*}_{j}})\bm{\mathsf{e}}^{*}_{j}}{n-|\widehat{A}^{*}_{j}|},\quad 1\leq j\leq p,\end{split}$ (13) which mimics (12) when $\bm{\mathsf{Y}}_{{\textnormal{{d}}}_{S}(j)}\bm{U}_{{\textnormal{{d}}}_{S}(j)}=\bm{0}$ under $H_{0}$, with $\bm{\mathsf{e}}_{j}$ replaced by $\bm{\mathsf{e}}^{*}_{j}$. As a result, when $\widehat{S}^{*}\supseteq S^{\circ}$, the conditional distribution of $\text{Lr}^{*}$ given $\bm{\mathsf{Z}}$ well approximates the null distribution of Lr, where the model selection effect is accounted for by assessing the variability of $\widehat{A}^{*}_{j}$ and $\widehat{B}^{*}_{j}$ over different realizations of $(\bm{\mathsf{Z}}^{*},\bm{\mathsf{e}}^{*})$. In practice, we use Monte-Carlo to approximate the distribution of $\text{Lr}^{*}$ given $\bm{\mathsf{Z}}$ in (13). In particular, we generate $M$ perturbed samples $(\bm{\mathsf{Z}}^{*}_{m},\bm{\mathsf{e}}_{m}^{*})_{m=1}^{M}$ independently and compute $\text{Lr}^{*}_{m}$; $m=1,\ldots,M$. Then, we examine the condition $\widehat{S}^{*}_{m}\supseteq S^{\circ}$ by checking its empirical counterpart $\widehat{S}^{*}_{m}\supseteq\widehat{S}$. The DP p-value of edge test in (2) is defined as $\text{Pval}=\left(\sum_{m=1}^{M}\operatorname{I}(\text{Lr}_{m}^{*}\geq\text{Lr},\widehat{S}^{*}_{m}\supseteq\widehat{S})\right)/\left(\sum_{m=1}^{M}\operatorname{I}(\widehat{S}^{*}_{m}\supseteq\widehat{S})\right),$ (14) where $\operatorname{I}(\cdot)$ is the indicator function. ###### Remark 3 Assume $\widehat{S}^{*}\supseteq S^{\circ}$ is a “correct” super-graph. A naive likelihood ratio based on perturbed data $(\bm{\mathsf{Z}}^{*},\bm{\mathsf{e}}^{*})$ is $\text{Lr}(\widehat{S}^{*},\widehat{\bm{\Sigma}}^{*})=\sum_{j:{\textnormal{{d}}}_{\widehat{S}}(j)\neq\emptyset}\frac{(\bm{\mathsf{Y}}_{{\textnormal{{d}}}_{\widehat{S}}(j)}\bm{U}_{{\textnormal{{d}}}_{\widehat{S}}(j),j}+\bm{\mathsf{e}}_{j}+\bm{\mathsf{e}}^{*}_{j})^{\top}(\bm{\mathsf{P}}_{\widehat{A}^{*}_{j}}-\bm{\mathsf{P}}_{\widehat{B}^{*}_{j}})(\bm{\mathsf{Y}}_{{\textnormal{{d}}}_{\widehat{S}}(j)}\bm{U}_{{\textnormal{{d}}}_{\widehat{S}}(j),j}+\bm{\mathsf{e}}_{j}+\bm{\mathsf{e}}^{*}_{j})}{2\widehat{\sigma}_{j}^{*}}.$ Note that the error $\bm{\mathsf{e}}_{j}$ given $\bm{\mathsf{Z}}$ is deterministic and does not vanish under either $H_{0}$ or $H_{a}$; $j=1,\ldots,p$. Thus, the conditional distribution of $\text{Lr}(\widehat{S}^{*},\widehat{\bm{\Sigma}}^{*})$ given $\bm{\mathsf{Z}}$ does not approximate the null distribution of Lr in (12), in contrast to the DP likelihood ratio $\text{Lr}^{*}$ in (13). #### 3.2.3 Extension to hypothesis testing for a pathway Next, we extend the DP inference for (3). Denote $\mathcal{H}=\\{(k_{1},j_{1}),\ldots,(k_{|\mathcal{H}|},j_{|\mathcal{H}|})\\}$. Then the test of pathways in (3) can be reduced to testing sub-hypotheses $H_{0,\nu}:U_{k_{\nu},j_{\nu}}=0,\quad\mbox{ versus }\quad{H}_{a,\nu}:U_{k_{\nu},j_{\nu}}\neq 0;\quad\nu=1,\ldots,|\mathcal{H}|,$ where each sub-hypothesis is a directed edge test. Given $(S,\bm{\Sigma})$, the likelihood ratio for $\text{H}_{0,\nu}$ is $\text{Lr}_{\nu}(S,\bm{\Sigma})=L_{S}(\widehat{\bm{\theta}}_{S}^{(1)},\bm{\Sigma})-L_{S}(\widehat{\bm{\theta}}_{S}^{(0,\nu)},\bm{\Sigma})$, where $\widehat{\bm{\theta}}_{S}^{(0,\nu)}$ is the MLE under the constraint that $U_{k_{\nu},j_{\nu}}=0$. Then for $\widehat{S}\supseteq S^{\circ}$, we have $\begin{split}\text{Lr}_{\nu}&=\frac{1}{2}\frac{(\bm{\mathsf{Y}}_{k_{\nu}}U_{k_{\nu},j_{\nu}}+\bm{\mathsf{e}}_{j_{\nu}})^{\top}(\bm{\mathsf{P}}_{\widehat{A}_{j_{\nu}}}-\bm{\mathsf{P}}_{\widehat{B}_{j_{\nu}}})(\bm{\mathsf{Y}}_{k_{\nu}}U_{k_{\nu},j_{\nu}}+\bm{\mathsf{e}}_{j_{\nu}})}{(\bm{\mathsf{e}}_{j_{\nu}})^{\top}(\bm{I}-\bm{\mathsf{P}}_{\widehat{A}_{j_{\nu}}})\bm{\mathsf{e}}_{j_{\nu}}/(n-|\widehat{A}_{j_{\nu}}|)},\\\ \text{Lr}_{\nu}^{*}&=\frac{1}{2}\frac{(\bm{\mathsf{e}}^{*}_{j_{\nu}})^{\top}(\bm{\mathsf{P}}_{\widehat{A}^{*}_{j_{\nu}}}-\bm{\mathsf{P}}_{\widehat{B}^{*}_{j_{\nu}}})\bm{\mathsf{e}}^{*}_{j_{\nu}}}{(\bm{\mathsf{e}}^{*}_{j_{\nu}})^{\top}(\bm{I}-\bm{\mathsf{P}}_{\widehat{A}^{*}_{j_{\nu}}})\bm{\mathsf{e}}^{*}_{j_{\nu}}/(n-|\widehat{A}^{*}_{j_{\nu}}|)},\end{split}$ (15) where the distributions of $\text{Lr}_{\nu}^{*}$ given $\bm{\mathsf{Z}}$ approximates the null distributions of $\text{Lr}_{\nu}$. Finally, define the p-value of pathway test in (3) as $\text{Pval}=\max_{1\leq\nu\leq|\mathcal{H}|}\left(\sum_{m=1}^{M}\operatorname{I}(\text{Lr}_{\nu,m}^{*}\geq\text{Lr}_{\nu},\widehat{S}^{*}_{m}\supseteq\widehat{S})\right)/\left(\sum_{i=1}^{M}\operatorname{I}(\widehat{S}^{*}_{m}\supseteq\widehat{S})\right).$ (16) Note that if any $\text{H}_{0,\nu}$ is degenerate, then $\text{Pval}=1$. Algorithm 3 summarizes the DP method for hypothesis testing. 1Specify the Monte Carlo size $M$. 2Compute Lr and $\widehat{\bm{\Sigma}}=\text{Diag}(\widehat{\sigma}^{2}_{1},\ldots,\widehat{\sigma}^{2}_{p})$ via Algorithm 2 with original data $\bm{\mathsf{Z}}$. 3(Parallel DP) Generate perturbed data $(\bm{\mathsf{Z}}^{*}_{m},\bm{\mathsf{e}}_{m}^{*})=(\bm{\mathsf{X}},\bm{\mathsf{Y}}_{m}^{*},\bm{\mathsf{e}}_{m}^{*})$ in parallel, where $\bm{\varepsilon}_{m,i}^{*}\sim N(\bm{0},\widehat{\bm{\Sigma}})$. For (2), compute $\text{Lr}^{*}_{m}$ based on $(\bm{\mathsf{Z}}^{*}_{m},\bm{\mathsf{e}}_{m}^{*})$, while, for (3), compute $\text{Lr}^{*}_{\nu,m}$; $\nu=1,\ldots,|\mathcal{H}|$ based on $(\bm{\mathsf{Z}}^{*}_{m},\bm{\mathsf{e}}_{m}^{*})$; $m=1,\ldots,M$. 4Compute the p-value of a test as (14) or (16) accordingly. Algorithm 3 DP likelihood ratio test ###### Remark 4 (Computation) To accelerate the computation, we consider a parallelizable procedure in Algorithm 3. In addition, we use the original estimate $\widehat{\bm{\theta}}$ as a warm-start initialization for the DP estimates, which effectively reduces the computing time. ###### Remark 5 (Connection with bootstrap) One may consider parametric or nonparametric bootstrap for Lr. The parametric bootstrap requires a good initial estimate of $(\bm{U},\bm{W})$. Yet, it is rather challenging to correct the bias of this estimate because of the acyclicity constraint. By comparison, DP does not rely on such an estimate. On the other hand, nonparametric bootstrap resamples the original data with replacement. In a bootstrap sample, only about 63% distinct observations in the original data are used in model selection and fitting, leading to deteriorating performance (Kleiner et al., 2012), especially in a small sample. As a result, nonparametric bootstrap may not well-approximate the distribution of Lr, while DP provides a better approximation of Lr, taking advantage of a full sample. ## 4 Theory This section provides the theoretical justification for the proposed methods. ### 4.1 Convergence and consistency of structure learning First, we introduce some technical assumptions to derive statistical and computational properties of Algorithms 1 and 2. Let $(\bm{\mathsf{Y}},\bm{\mathsf{X}})\in\mathbb{R}^{n\times(p+q)}$ be the data matrix. Denote by $\bm{\zeta}$ a generic vector and $\bm{\zeta}_{A}\in\mathbb{R}^{|A|}$ be the projection of $\bm{\zeta}$ onto coordinates in $A$. Let $\kappa^{\circ}_{j}=\|\bm{V}^{\circ}_{\cdot j}\|_{0}$ and $\kappa^{\circ}_{\max}=\max_{1\leq j\leq p}\kappa_{j}^{\circ}$. ###### Assumption 2 (Restricted eigenvalues) For a constant $c_{1}>0$, $\min_{A:|A|\leq 2\kappa^{\circ}_{\max}}\min_{\\{\bm{\zeta}:\|\bm{\zeta}_{A^{c}}\|_{1}\leq 3\|\bm{\zeta}_{A}\|_{1}\\}}\frac{\|\bm{\mathsf{X}}\bm{\zeta}\|_{2}^{2}}{n\|\bm{\zeta}\|_{2}^{2}}\geq c_{1}.$ ###### Assumption 3 (Interventions) For a constant $c_{2}>0$, $\max\Big{(}\max_{1\leq l\leq q}n^{-1}(\bm{\mathsf{X}}^{\top}\bm{\mathsf{X}})_{ll},\max_{A:|A|\leq 2\kappa^{\circ}_{\max}}\max_{1\leq l\leq|A|}n((\bm{\mathsf{X}}_{A}^{\top}\bm{\mathsf{X}}_{A})^{-1})_{ll}\Big{)}\leq c_{2}^{2}.$ ###### Assumption 4 (Nuisance signals) $\min_{V^{0}_{lj}\neq 0}|V^{0}_{lj}|\geq 100\frac{c_{2}}{c_{1}}\max_{1\leq j\leq p}\sqrt{\Omega_{jj}^{-1}\Big{(}\frac{\log(q)}{n}+\frac{\log(n)}{n}\Big{)}}.$ (17) Assumptions 2-3, as a replacement of Assumption 1A, are satisfied with a probability tending to one for isotropic subgaussian or bounded $\bm{X}$ (Rudelson and Zhou, 2012). Assumption 2 is a common condition for proving the convergence rate of Lasso regression (Bickel et al., 2009; Zhang et al., 2014). Assumption 4, as an alternative to Assumption 1B, specifies the minimal signal strength over candidate interventions. A signal strength requirement like Assumption 4 is a condition for establishing selection consistency in high dimensional variable selection (Fan et al., 2014; Loh et al., 2017; Zhao et al., 2018). Moreover, Assumption 4 can be further relaxed to a less intuitive condition, Assumption 6; see Appendix A.3 for details. ###### Theorem 4 (Finite termination and consistency) Under Assumptions 1-4 with constants $c_{1}<6c_{2}$, if the machine precision $\text{tol}\ll 1/n$ is negligible, and tuning parameters of Algorithm 1 satisfy: 1. (1) $\gamma_{j}\leq c_{1}/6$, 2. (2) $\sqrt{32c_{2}^{2}\Omega_{jj}^{-1}(\log(q)+\log(n))/n}/\gamma_{j}\leq\tau_{j}\leq 0.4\min_{V^{0}_{lj}\neq 0}|V^{0}_{lj}|$, 3. (3) $\kappa_{j}=\kappa_{j}^{\circ}$, for $j=1,\ldots,p$, then Algorithm 1 yields the global minimizer $\widehat{\bm{V}}_{\cdot j}$ of (7) in at most $1+\lceil\log(\kappa_{\max}^{\circ})/\log 4\rceil$ DC iterations almost surely, where $\lceil\cdot\rceil$ is the ceiling function. Moreover, Algorithm 2 recovers $\mathcal{A}^{\circ}$, $\mathcal{C}^{\circ}\equiv\\{l:V^{\circ}_{lj}\neq 0\\}\supseteq\mathcal{I}^{\circ}$, and $\mathcal{D}^{\circ}$ almost surely as $n\to\infty$. By Theorem 4 (proved in Appendix B.4), the computational complexity of Algorithm 1 is the number of iterations multiplied by that of solving a weighted Lasso regression, $O(\log\kappa^{\circ}_{\max})\times O(q^{3}+nq^{2})$ (Efron et al., 2004). Also note that in Steps 2-6 of Algorithm 2, no heavy computation is involved. Thus, the overall time complexity of Algorithm 2 is bounded by $O(p\times\log\kappa^{\circ}_{\max}\times(q^{3}+nq^{2}))$ for computing $\widehat{\bm{V}}$ matrix. Finally, the peeling algorithm does not apply to observational data ($\bm{W}=\bm{0}$). In a sense, interventions play an essential role. Although Theorem 4 establishes the consistent reconstruction of the peeling algorithm for the super-graph $S$, it does not provide any uncertainty measure of the presence of some directed edges of the true DAG. In what follows, we will develop a theory for hypothesis tests to make valid inference concerning directed edges of interest. ### 4.2 Inferential theory ###### Assumption 5 (Hypothesis-specific dimension restriction) For a constant $\rho$, $\max_{j:\textnormal{{d}}_{G^{\circ}}(j)\neq\emptyset}\frac{|\textnormal{{an}}_{G^{\circ}}(j)|+|\textnormal{{d}}_{G^{\circ}}(j)|+{\kappa_{j}^{\circ}}}{n}\leq\rho<1,\quad\text{ as }n\rightarrow\infty.$ Assumption 5 is a hypothesis-specific condition restricting the underlying dimension of the testing problem. Usually, $|\textnormal{{an}}_{G^{\circ}}(j)|\asymp\kappa_{j}^{\circ}\ll p$; $j=1,\ldots,p$, which dramatically relaxes the condition $n\gg{|\mathcal{D}|^{1/2}p\log(p)}$ for the constrained likelihood ratio test (Li et al., 2019). ###### Theorem 5 (Empirical p-values) Suppose Assumptions 1-5 are met and $H_{0}$ is regular. If the tuning parameters in Algorithm 1 satisfy (1)-(3) in Theorem 4, then, 1. (A) (Test of directed edges (2)) $\begin{split}\lim_{\begin{subarray}{c}n\to\infty\\\ \bm{\theta}^{\circ}\textnormal{ satisfies H}_{0}\textnormal{ in }\eqref{equation:link- test}\end{subarray}}\lim_{M\rightarrow\infty}\operatorname{\mathbb{P}}_{\bm{\theta}^{\circ}}(\textnormal{Pval}<\alpha)=\alpha,&\quad\text{ if }H_{0}\text{ is nondegenerate};\\\ \lim_{\begin{subarray}{c}n\to\infty\\\ \bm{\theta}^{\circ}\textnormal{ satisfies H}_{0}\textnormal{ in }\eqref{equation:link- test}\end{subarray}}\lim_{M\rightarrow\infty}\operatorname{\mathbb{P}}_{\bm{\theta}^{\circ}}(\textnormal{Pval}=1)=1,&\quad\text{ if }H_{0}\text{ is degenerate}.\end{split}$ 2. (B) (Test of directed pathways (3)) $\begin{split}\limsup_{\begin{subarray}{c}n\to\infty\\\ \bm{\theta}^{\circ}\textnormal{ satisfies H}_{0}\textnormal{ in }\eqref{equation:pathway- test}\end{subarray}}\lim_{M\rightarrow\infty}\operatorname{\mathbb{P}}_{\bm{\theta}^{\circ}}(\textnormal{Pval}<\alpha)=\alpha,&\quad\text{ if }H_{0}\text{ is nondegenerate with }|\mathcal{D}|=|\mathcal{H}|;\\\ \limsup_{\begin{subarray}{c}n\to\infty\\\ \bm{\theta}^{\circ}\textnormal{ satisfies H}_{0}\textnormal{ in }\eqref{equation:pathway- test}\end{subarray}}\lim_{M\rightarrow\infty}\operatorname{\mathbb{P}}_{\bm{\theta}^{\circ}}(\textnormal{Pval}=1)=1,&\quad\text{ if }|\mathcal{D}|<|\mathcal{H}|.\end{split}$ By Theorem 5, the DP likelihood ratio test yields a valid p-value for (2) and (3). Note that $|\mathcal{D}|$ is permitted to depend on $n$. Moreover, Proposition 6 summarizes the asymptotics for directed edge test (2). ###### Proposition 6 (Asymptotic distribution of Lr of (2)) Assume that the assumptions of Theorem 5 are met. Under a regular $H_{0}$, as $n\to\infty$, $\begin{split}2\textnormal{Lr}\overset{d}{\longrightarrow}\chi^{2}_{|\mathcal{D}|},&\text{ if }H_{0}\text{ is nondegenerate with }|\mathcal{D}|>0\text{ being fixed};\\\ \frac{2\textnormal{Lr}-|\mathcal{D}|}{\sqrt{2|\mathcal{D}|}}\overset{d}{\longrightarrow}N(0,1),&\text{ if }H_{0}\text{ is nondegenerate with }|\mathcal{D}|\rightarrow\infty,\ \frac{|\mathcal{D}|\log|\mathcal{D}|}{n}\rightarrow 0.\end{split}$ The proofs of Theorem 5 and Proposition 6 are given in Appendix B.5 and B.6. ###### Remark 6 (Local structure for hypothesis testing) Theorem 5 or Proposition 6 only requires correct identification of the local structures of the DAG such as, $\textnormal{{an}}_{S}(j)=\textnormal{{an}}_{G}(j)$, $\textnormal{{d}}_{S}(j)=\textnormal{{d}}_{G}(j)$, and $\textnormal{{in}}_{S}(j)$, for variables $Y_{j}$ with $\textnormal{{d}}_{S}(j)\neq\emptyset$, as opposed to those of entire $S$. Next, we analyze the local limit power of the proposed tests for (2) and (3). Assume $\bm{\theta}^{\circ}=(\bm{U}^{\circ},\bm{W}^{\circ})$ satisfies $H_{0}$. Let $\bm{\Delta}\in\mathbb{R}^{p\times p}$ satisfy $\bm{\Delta}_{\mathcal{D}^{c}}=\bm{0}$ so that $\bm{U}^{\circ}+\bm{\Delta}$ represents a DAG. For nondegenerate and regular $H_{0}$, consider an alternative $H_{a}$: $\bm{U}_{\mathcal{H}}=\bm{U}_{\mathcal{H}}^{\circ}+\bm{\Delta}_{\mathcal{H}}$, and define the power function as $\beta(\bm{\theta}^{\circ},\bm{\Delta})=\operatorname{\mathbb{P}}_{H_{a}}(\text{Pval}<\alpha).$ (18) ###### Proposition 7 (Local power of edge test) Assume that $H_{0}$ is nondegenerate and regular. Let $\|\bm{\Delta}\|_{F}=\|\bm{\Delta}_{\mathcal{H}}\|_{F}=n^{-1/2}\delta$ when $|\mathcal{D}|>0$ is fixed and $\|\bm{\Delta}\|_{F}=|\mathcal{D}|^{1/4}n^{-1/2}h$ when $|\mathcal{D}|\to\infty$, where $\delta>0$ and $\|\cdot\|_{F}$ is the matrix Frobenius norm. If the assumptions of Theorem 5 are met, then under $H_{a}$, as $n,M\to\infty$, $\beta(\bm{\theta}^{\circ},\bm{\Delta})\geq\begin{cases}\operatorname{\mathbb{P}}\Big{(}\|\bm{Z}+c_{l}\sqrt{n}\bm{\Delta}\|_{2}^{2}>\chi^{2}_{|\mathcal{D}|,1-\alpha}\Big{)}&\text{if }|\mathcal{D}|>0\text{ is fixed};\\\ \operatorname{\mathbb{P}}\Big{(}Z>z_{1-\alpha}-{c_{l}\|\bm{\Delta}\|^{2}_{2}}/{\sqrt{2|\mathcal{D}|}}\Big{)}&\text{if }|\mathcal{D}|\to\infty,\frac{|\mathcal{D}|\log|\mathcal{D}|}{n}\to 0,\end{cases}$ where $\bm{Z}\sim N(\bm{0},\bm{I}_{|\mathcal{D}|})$, $Z\sim N(0,1)$, and $\chi^{2}_{|\mathcal{D}|,1-\alpha}$ and $z_{1-\alpha}$ are the $(1-\alpha)$th quantile of distributions $\chi^{2}_{|\mathcal{D}|}$ and $N(0,1)$, respectively. In particular, $\lim_{\delta\to\infty}\lim_{n\to\infty}\beta(\bm{\theta}^{\circ},\bm{\Delta})=1$. ###### Proposition 8 (Local power of pathway test) Assume that $H_{0}$ is nondegenerate and regular with $|\mathcal{D}|=|\mathcal{H}|$. Let $\min_{(k,j)\in\mathcal{H}}|U^{\circ}_{kj}+\Delta_{kj}|=n^{-1/2}\delta$ when $|\mathcal{H}|>0$ is fixed and $\min_{(k,j)\in\mathcal{H}}|U^{\circ}_{kj}+\Delta_{kj}|=n^{-1/2}\delta\sqrt{\log|\mathcal{H}|}$ when $|\mathcal{H}|\to\infty$. If the assumptions of Theorem 5 are met, then under $H_{a}$, as $n,M\to\infty$, $\beta(\bm{\theta}^{\circ},\bm{\Delta})\geq 1-\frac{|\mathcal{H}|}{\sqrt{2\pi}}\exp\Big{(}-\Big{(}\delta\sqrt{\log|\mathcal{H}|}/\max_{1\leq j\leq p}\Omega_{jj}-\sqrt{\chi^{2}_{1,1-\alpha}}\Big{)}^{2}/2\Big{)},$ where $Z\sim N\Big{(}\delta^{2}/\max_{1\leq j\leq p}\Omega_{jj},1\Big{)}$ and $\chi^{2}_{1,1-\alpha}$ is the $(1-\alpha)$th quantile of distribution $\chi^{2}_{1}$. Then, $\lim_{\delta\to\infty}\lim_{n\to\infty}\beta(\bm{\theta}^{\circ},\bm{\Delta})=1$. The proofs of Propositions 7 and 8 are deferred to Appendix B.7 and B.8. ## 5 Simulations This section investigates the operating characteristics of the proposed tests and the peeling algorithm via simulations. In simulations, we consider two setups for generating $\bm{U}\in\mathbb{R}^{p\times p}$, representing random and hub DAGs, respectively. * • Random graph. The upper off-diagonal entries $U_{kj}$; $k<j$ are sampled independently from $\\{0,1\\}$ according to Bernoulli$(1/p)$, while other entries are zero. This generates a random graph with a sparse neighborhood. * • Hub graph. Set $U_{1,2j+1}=1$ and $U_{2,2j+2}=1$ for $j=1,\ldots,\lfloor p/2\rfloor-2$, while other entries are zero. This generates a hub graph, where nodes 1 and 2 are hub nodes with a dense neighborhood. Moreover, we consider three setups for intervention matrix $\bm{W}\in\mathbb{R}^{q\times p}$, representing different scenarios. Setups A and B are designed for inference, whereas Setup C in Section 5.2 is designed to compare with the method of Chen et al. (2018) for structure learning. Let $\bm{W}=(\bm{A}^{\top},\bm{B}^{\top},\bm{0}^{\top})^{\top}$, where $\bm{A},\bm{B}\in\mathbb{R}^{p\times p}$ and $\bm{0}\in\mathbb{R}^{(q-2p)\times p}$. * • Setup A. Set $A_{jj}=B_{jj}=B_{j,j+1}=1$; $j=1,\ldots,p-1$, $A_{pp}=1$, while other entries of $\bm{A}$, $\bm{B}$ are zero. Then, $X_{1},\ldots,X_{p}$ are instruments for $Y_{1},\ldots,Y_{p}$, respectively, $X_{p+1},\ldots,X_{2p-1}$ are invalid instruments with two targets, and $X_{2p},\ldots,X_{q}$ represent inactive interventions. * • Setup B. Set $A_{jj}=A_{j,j+1}=B_{jj}=B_{j,j+1}=1$; $j=1,\ldots,p-1$, $A_{pp}=1$, while other entries of $\bm{A}$, $\bm{B}$ are zero. Here, the only valid instrument is $X_{p}$ on $Y_{p}$, and the other intervention variables either have two targets or are inactive. Importantly, the instrumental sufficiency (Assumption 1C) is not met. To generate $(\bm{X},\bm{Y})$ for each setup, we sample $\bm{X}\sim N(\bm{0},\bm{\Sigma}_{X})$ with $(\Sigma_{X})_{ll^{\prime}}=0.5^{|l-l^{\prime}|}$; $1\leq l,l^{\prime}\leq q$ and sample $\bm{Y}$ according to (1) with $(\bm{U},\bm{W},\sigma_{1}^{2},\ldots,\sigma_{p}^{2})$, where $\sigma_{1}^{2},\ldots,\sigma_{p}^{2}$ are set to be equally spaced from $0.5$ to $1$. ### 5.1 Inference We compare three tests in empirical type-I errors and powers in simulated examples, namely, the DP likelihood ratio test (DP-LR) in Algorithm 3, the asymptotic likelihood ratio test (LR), and the oracle likelihood ratio test (OLR). Here LR uses $\text{Lr}(\widehat{S},\widehat{\bm{\Sigma}})$, while OLR uses $\text{Lr}(S^{\circ},\widehat{\bm{\Sigma}})$ assuming that the super- graph $S^{\circ}$ were known in advance. The p-values of LR and OLR are computed via Proposition 6. The implementation details of these tests are in Appendix C. For the empirical type-I error of a test, we compute the percentage of times rejecting $H_{0}$ out of 500 simulations when $H_{0}$ is true. For the empirical power of a test, we report the percentage of times rejecting $H_{0}$ out of 100 simulations when $H_{a}$ is true under alternative hypotheses $H_{a}$. * • Test of directed edges. For (2), we examine two different hypotheses: (i) $H_{0}:U_{1,20}=0$ versus $H_{a}:U_{1,20}\neq 0$. In this case, $|\mathcal{D}|=1$. (ii) $H_{0}:\bm{U}_{\mathcal{H}}=\bm{0}$ versus $H_{a}:\bm{U}_{\mathcal{H}}\neq\bm{0}$, where $\mathcal{H}=\\{(k,20):k=1,\ldots,15\\}$. In this case, $|\mathcal{D}|=15$. Moreover, five alternatives $H_{a}:U_{1,20}=0.1l$ and $\bm{U}_{\mathcal{H}\setminus\\{(1,20)\\}}=\bm{0}$; $l=1,2,3,4,5$, are used for the power analysis in (i) and (ii). The data are generated by modifying $\bm{U}$ accordingly. * • Test of directed pathways. We examine the test of a directed path $Y_{1}\to Y_{5}\to Y_{10}\to Y_{15}\to Y_{20}$, namely $\mathcal{H}=\\{(1,5),(5,10),(10,15),(15,20)\\}$ in (3). Since (3) is a test of composite null hypothesis, the data are generated under a graph with parameters $(\bm{U},\bm{W},\bm{\Sigma})$ satisfying $H_{0}$, where $\bm{U}_{\mathcal{H}}=\bm{0}$. Five alternatives $H_{a}:\bm{U}_{\mathcal{H}}=\bm{0.1}l$; $l=1,2,3,4,5$ are used for the power analysis. Figure 2: Empirical type-I errors and powers of tests of directed edges. The black dotted line marks the nominal level of significance $\alpha=0.05$. Figure 3: Empirical type-I errors and powers of tests of a directed pathway. The black dotted line marks the nominal level of significance $\alpha=0.05$. For testing directed edges, as displayed in Figure 2, DP-LR and LR perform well compared to the ideal test OLR in Setup A with Assumption 1 satisfied. In setup B with Assumption 1C not fulfilled, DP-LR appears to have control of type-I error, whereas LR has an inflated empirical type-I error compared to the nominal level $\alpha=0.05$. This discrepancy is attributed to the data perturbation scheme accounting for the uncertainty of identifying $S$. However, both suffer a loss of power compared to the oracle test OLR in this setup. This observation suggests that without Assumption 1C, the peeling algorithm tends to yield an estimate $\widehat{S}\supseteq S^{\circ}$, which overestimates $S^{\circ}$, resulting in a power loss. For testing directed pathways, as indicated in Figure 3, we observe similar phenomena as in the previous directed edge tests. Of note, both LR and DP-LR are capable in controlling type-I error of directed path tests. In summary, DP-LR has a suitable control of type-I error when there are invalid instruments and Assumption 1C is violated. Concerning the power, DP-LR and LR are comparable in all scenarios and their powers tend to one as the sample size $n$ or the signal strength of tested edges increases. Moreover, DP-LR and LR perform nearly as well as the oracle test OLR when Assumption 1 is satisfied. These empirical findings agree with our theoretical results. ### 5.2 Structure learning This subsection compares the peeling algorithm with the Two-Stage Penalized Least Squares (2SPLS, Chen et al. (2018)) in terms of the structure learning accuracy. For peeling, we consider Algorithm 2 with an additional step (9) for structure learning of $\bm{U}$. For 2SPLS, we use the R package BigSEM. Figure 4: Structural Hamming distances (SHDs) for the peeling algorithm and 2SPLS, where a smaller value of SHD indicates a better result. 2SPLS requires that all the intervention variables to be target-known instruments in addition to Assumption 1C. Thus, we consider an additional Setup C. * • Setup C. Let $\bm{W}=(\bm{I}_{p\times p},\bm{0})^{\top}\in\mathbb{R}^{q\times p}$. Then $X_{1},\ldots,X_{p}$ are valid instruments for $Y_{1},\ldots,Y_{p}$, respectively, and other intervention variables are inactive. For 2SPLS, we assign each active intervention variable to its most correlated primary variable in Setups A-C. In Setup C, this assignment yields a correct identification of valid instruments, meeting all the requirements of 2SPLS. For each scenario, we compute the average structural Hamming distance $\text{SHD}(\widehat{\bm{U}},\bm{U}^{\circ})=\sum_{k,j}|\operatorname{I}(\widehat{U}_{kj}\neq 0)-\operatorname{I}(U^{\circ}_{kj}\neq 0)|$ over 100 runs. As shown in Figure 4, the peeling algorithm outperforms 2SPLS, especially when there are invalid instruments and Assumption 1C is violated. ### 5.3 Comparison of inference and structure learning Figure 5: Rejection rates for $H_{0}:U_{1,20}=0$ versus $H_{a}:U_{1,20}=1/\sqrt{n}$ by DP inference and structure learning. For structure learning, $H_{0}$ is rejected if $\widehat{U}_{1,20}\neq 0$. The red lines indicate the results of DP inference using the significance level $\alpha=0.05$. The other colored lines display the results of structure learning using different sparsity parameter values $\kappa=1,2,3,4,5$. The simulation is repeated for 500 times and $\kappa=2$ is chosen by BIC in over 90% cases. This subsection compares the proposed DP testing method against the proposed structure learning method in (9) in terms of inferring the true graph structure. To this end, we consider Setup A in Section 5.1 with $p=30,q=100$, and the hypotheses $H_{0}:U_{1,20}=0$ versus $H_{a}:U_{1,20}=1/\sqrt{n}$. For DP inference, we use $\alpha=0.05$ and choose the tuning parameters by BIC as in previous experiments; see Appendix C for details. For structure learning, we reject the null hypothesis when $\widehat{U}_{1,20}\neq 0$. As displayed in Figure 5, when the null hypothesis $H_{0}$ is true, the DP testing method controls type-I error very close to the nominal level of 0.05, whereas the type-I error of the structure learning varies greatly depending on the tuning parameter selection. Under the alternative hypothesis $H_{a}$, the DP inference enjoys high statistical power than structure learning methods when $n\geq 200$. Interestingly, the power of structure learning diminishes as $n$ increases. This observation is in agreement with our theoretical results in Theorem 4, suggesting that consistent reconstruction requires the smallest size of nonzero coefficients to be of order $\sqrt{\log(n)/n}$ with the tuning parameter $\tau$ of the same order (fixing $p,q$). In this case, the edge $U_{1,20}$ is of order $1/\sqrt{n}$, which is less likely to be reconstructed as $n$ increases. In contrast, Proposition 7 indicates that a DP test has a non-vanishing power when the hypothesized edges are of order $1/\sqrt{n}$. Figure 5 demonstrates some important distinctions between inference and structure learning. When different tuning parameters are used, the structure learning results correspond to different points on an ROC curve. Although it is asymptotically consistent when optimal tuning parameters are used, structure learning lacks an uncertainty measure of graph structure identification. As a result, it is nontrivial for structure learning methods to trade-off the false discovery rate and detection power in practice. This makes the interpretation of such results hard, especially when they heavily rely on hyperparameters as in Figure 5. By comparsion, DP inference aims to maximize statistical power while controlling type-I error at a given level, offering a clear interpretation of its result. This observation agrees with the discussions in the literature on variable selection and inference (Wasserman and Roeder, 2009; Meinshausen and Bühlmann, 2010; Lockhart et al., 2014; Candes et al., 2018) and it justifies the demand for inferential tools for directed graphical models. ## 6 ADNI data analysis This section applies the proposed tests to analyze an Alzheimer’s Disease Neuroimaging Initiative (ADNI) dataset. In particular, we infer gene pathways related to Alzheimer’s Disease (AD) to highlight some gene-gene interactions differentiating patients with AD/cognitive impairments and healthy individuals. The raw data are available in the ADNI database (https://adni.loni.usc.edu), including gene expression, whole-genome sequencing, and phenotypic data. After cleaning and merging, we have a sample size of 712 subjects. From the KEGG database (Kanehisa and Goto, 2000), we extract the AD reference pathway (hsa05010, https://www.genome.jp/pathway/hsa05010), including 146 genes in the ADNI data. For data analysis, we first regress the gene expression levels on five covariates – gender, handedness, education level, age, and intracranial volume, and then use the residuals as gene expressions in the following analysis. Next, we extract the genes with at least one SNP at a marginal significance level below $10^{-3}$, yielding $p=63$ genes as primary variables. For these genes, we further extract their marginally most correlated two SNPs, resulting in $q=63\times 2=126$ SNPs as unspecified intervention variables, in subsequent data analysis. All gene expression levels are normalized. The data contains individuals in four groups, namely, Alzheimer’s Disease (AD), Early Mild Cognitive Impairment (EMCI), Late Mild Cognitive Impairment (LMCI), and Cognitive Normal (CN). For our purpose, we treat 247 CN individuals as controls while the remaining 465 individuals as cases (AD-MCI). Then, we use the gene expressions and the SNPs to reconstruct the ancestral relations and infer gene pathways for 465 AD-MCI and 247 CN control cases, respectively. (a) AD-MCI (b) CN (c) AD-MCI (d) CN Figure 6: Display of the subnetworks associated with genes APP and CASP3. (a) and (b): Solid/dashed arrows indicate significant/insignificant edges at $\alpha=0.05$ after adjustment for multiplicity by the Bonferroni-Holm correction. (c) and (d): Solid arrows indicate the reconstructed edges using 2SPLS (Chen et al., 2018). (a) AD-MCI (b) CN Figure 7: The p-values of pathway tests (3) by the proposed tests for the AD- MCI and CN groups, where p-values are adjusted for multiplicity by the Bonferroni-Holm correction and solid/dashed arrows indicate significant/insignificant pathways at $\alpha=0.05$. In the literature, genes APP, CASP3, and PSEN1 are well-known to be associated with AD, reported to play different roles in AD patients and healthy subjects (Julia and Goate, 2017; Su et al., 2001; Kelleher and Shen, 2017). For this dataset, we conduct hypothesis testing on edges and pathways related to genes APP, CASP3, and PSEN1 in the KEGG AD reference (hsa05010) to evaluate the proposed DP inference by checking if DP inference can discover the differences that are reported in the biomedical literature. First, we consider testing $H_{0}:U_{kj}=0$ versus $H_{a}:U_{kj}\neq 0$, for each edge $(k,j)$ as shown in Figure 6 (a) and (b). Moreover, we consider two hypothesis tests of pathways $H_{0}:U_{kj}=0$ for some $(k,j)\in\mathcal{P}_{\ell}$ versus $H_{a}:U_{kj}\neq 0$ for all $(k,j)\in\mathcal{P}_{\ell}$; $\ell=1,2$, where the two pathways are specified by $\mathcal{P}_{1}=\\{\text{PSEN1}\to\text{CAPN1}\to\text{CDK5R1}\\}$, and $\mathcal{P}_{2}=\\{\text{PSEN1}\to\text{CAPN2}\to\text{CDK5R1}\\}$. See Figure 7. Of note, for clear visualization, Figure 6 (a), (b) and Figure 7 only display the edges related to hypothesis testing, whereas the ancestral relations are reconstructed using $p=63$ genes and $q=126$ SNPs for AD-MCI and CN groups separately. In Figures 6-7, the significant results under the level $\alpha=0.05$ after the Holm-Bonferroni adjustment for $2\times(9+2)=22$ tests are displayed. In Figures 6, the edge test in (2) exhibits a strong evidence for the presence of directed connectivity $\\{$APP $\to$ APBB1, APP $\to$ GSK3B, FADD $\to$ CASP3$\\}$ in the AD-MCI group, but no evidence in the CN group. Meanwhile, this test suggests the presence of connections $\\{$TNFRSF1A $\to$ CASP3, FADD $\to$ CASP8$\\}$ in the CN group but not so in the AD-MCI group. In both groups, we identify directed connections $\\{$TNFRSF1A $\to$ FADD, TNFRSF1A $\to$ CASP8$\\}$. In Figure 7, the pathway test (3) supports the presence of a pathway PSEN1 $\to$ CAPN1 $\to$ CDK5R1 in the AD-MCI group with a p-value of $0.044$ but not in the CN group with a p-value of $0.33$. The pathway PSEN1 $\to$ CAPN2 $\to$ CDK5R1 appears insignificant at $\alpha=0.05$ for both groups. Also noted is that some of our discoveries agree with the literature according to the AlzGene database (alzgene.org) and the AlzNet database (https://mips.helmholtz-muenchen.de/AlzNet-DB). Specifically, GSK3B differentiates AD patients from normal subjects; as shown in Figures 6, our result indicates the presence of connection APP $\to$ GSK3B for the AD-MCI group, but not for the CN group, the former of which is confirmed by Figure 1 of Vanleuven (2011). The connection APP $\to$ APBB1 also differs in AD-MCI and CN groups, which appears consistent with Figure 3 of Bu (2009). Moreover, the connection CAPN1 $\to$ CDK5R1, in the pathway PSEN1 $\to$ CAPN1 $\to$ CDK5R1 discovered in AD-MCI group, is found in the AlzNet database (interaction-ID 24614, https://mips.helmholtz-muenchen.de/AlzNet-DB/entry/show/1870). Finally, as suggested by Figure 8, the normality assumption in (1) is adequate for both groups. By comparison, as shown in Figure 6 (c) and (d), gene APP in the reconstructed networks by 2SPLS (Chen et al., 2018) is not connected with other genes, indicating no regulatory relation of APP with other genes in the AD-MCI and CN groups. However, as a well-known gene associated with AD, APP is reported to play different roles in controlling the expressions of other genes for AD patients and healthy people (Matsui et al., 2007; Julia and Goate, 2017). Our results in Figure 6 (a) and (b) are congruous with the studies: the connections of APP with other genes are different in our estimated networks for AD-MCI and CN groups. Figure 8: Normal quantile-quantile plots of studentized residuals of the AD- MCI and CN groups. In summary, our findings seem to agree with those in the literature (Julia and Goate, 2017; Su et al., 2001; Kelleher and Shen, 2017), where the subnetworks of genes APP, CASP3 in Figure 6 and PSEN1 in Figure 7 differentiate the AD-MCI from the CN groups. Furthermore, the pathway PSEN1 $\to$ CAPN1 $\to$ CDK5R1 in Figure 7 seems to differentiate these groups, which, however, requires validation in biological experiments. ## 7 Summary This article proposes structure learning and inference methods for a Gaussian DAG with interventions, where the targets and strengths of interventions are unknown. A likelihood ratio test is derived based on an estimated super-graph formed by ancestral relations and candidate interventional relations. This test accounts for the statistical uncertainty of the construction of the super-graph based on a novel data perturbation scheme. Moreover, we develop a peeling algorithm for the super-graph construction. The peeling algorithm allows scalable computing and yields a consistent estimator. The numerical studies justify our theory and demonstrate the utility of our methods. The proposed methods can be extended to many practical situations beyond biological applications with independent and identically distributed data. An instance is to infer directed relations between multiple autoregressive time series (Pamfil et al., 2020), where the lagged variables and covariates can serve as interventions for each time series. Acknowledgments The authors would like to thank the action editor and anonymous referees for helpful comments and suggestions. The research is supported by NSF grants DMS-1712564, DMS-1952539 and NIH grants R01GM126002, R01HL116720, R01HL105397, R01AG069895, R01AG065636, R01AG074858, and U01AG073079. ## A Illustrative examples and discussions ### A.1 Identifiability of model (1) and Assumption 1 The parameter space for model (1) is $\\{(\bm{U},\bm{W},\bm{\Sigma}):\bm{U}\in\mathbb{R}^{p\times p}\text{ represents a DAG},\ \bm{W}\in\mathbb{R}^{p\times q},\ \bm{\Sigma}=\text{diag}(\sigma_{1}^{2},\ldots,\sigma^{2}_{p})\\}.$ As suggested by Proposition 1, Assumption 1 (1A-1C) suffices for identification of every parameter value in the parameter space. Next, we show by examples that if Assumption 1B or 1C is violated then model (1) is no longer identifiable. Before proceeding, we rewrite (1) as $\bm{Y}=\bm{V}^{\top}\bm{X}+\bm{\varepsilon}_{V},\quad\bm{\varepsilon}_{V}=(\bm{I}-\bm{U}^{\top})^{-1}\bm{\varepsilon}\sim N(\bm{0},\bm{\Omega}^{-1}),$ where $\bm{\Omega}=(\bm{I}-\bm{U})\bm{\Sigma}^{-1}(\bm{I}-\bm{U}^{\top})$ is a precision matrix and $\bm{V}=\bm{W}(\bm{I}-\bm{U})^{-1}$. ###### Example 2 (Identifiability) In model (1), consider two non-identifiable bivariate situations: (1) $p=2$ and $q=4$ and (2) $p=2$ and $q=2$. 1. (1) Model (1) is non-identifiable when Assumption 1B breaks down. Consider two different models with different parameter values: $\displaystyle\bm{\theta}:$ $\displaystyle Y_{1}=X_{1}+X_{2}+X_{3}+\varepsilon_{1},\quad$ $\displaystyle Y_{2}=Y_{1}-X_{2}+X_{3}+X_{4}+\varepsilon_{2},$ (19) $\displaystyle\widetilde{\bm{\theta}}:$ $\displaystyle Y_{1}=0.5Y_{2}+0.5X_{1}+X_{2}-0.5X_{4}+\widetilde{\varepsilon}_{1},\quad$ $\displaystyle Y_{2}=X_{1}+2X_{3}+X_{4}+\widetilde{\varepsilon}_{2},$ (20) where $\varepsilon_{1},\varepsilon_{2}\sim N(0,1)$ are independent, and $\widetilde{\varepsilon}_{1}\sim N(0,0.5)$, $\widetilde{\varepsilon}_{2}\sim N(0,2)$ are independent. As depicted in Figure 9, (19) satisfies Assumption 1C. However, Assumption 1B is violated given that $\operatorname{Cov}(Y_{2},X_{2}\mid\bm{X}_{\\{1,3,4\\}})=0$ and $X_{2}$ is an intervention variable of $Y_{2}$. This is because the direct interventional effect of $X_{2}$ on $Y_{2}$ are canceled out by its indirect interventional effect through $Y_{1}$. Similarly, (20) satisfies Assumption 1C but $\operatorname{Cov}(Y_{1},X_{4}\mid\bm{X}_{\\{1,2,3\\}})=0$ violating Assumption 1B. In this case, it can be verified that $\bm{\theta}$ and $\widetilde{\bm{\theta}}$ correspond to the same distribution $\operatorname{\mathbb{P}}(\bm{Y}\mid\bm{X})$, because they share the same $(\bm{V},\bm{\Omega})$ even with different values of $(\bm{U},\bm{W},\bm{\Sigma})$. Hence, it is impossible to infer the directed relation between $Y_{1}$ and $Y_{2}$. 2. (2) Model (1) is non-identifiable when Assumption 1C breaks down. Consider two different models with different parameter values: $\displaystyle\bm{\theta}:$ $\displaystyle Y_{1}=X_{1}+X_{2}+\varepsilon_{1},\quad$ $\displaystyle Y_{2}=Y_{1}+X_{2}+\varepsilon_{2},$ (21) $\displaystyle\widetilde{\bm{\theta}}:$ $\displaystyle Y_{1}=0.5Y_{2}+0.5X_{1}+\widetilde{\varepsilon}_{1},\quad$ $\displaystyle Y_{2}=X_{1}+2X_{2}+\widetilde{\varepsilon}_{2},$ (22) where $\varepsilon_{1},\varepsilon_{2}\sim N(0,1)$ are independent, and $\widetilde{\varepsilon}_{1}\sim N(0,0.5)$, $\widetilde{\varepsilon}_{2}\sim N(0,2)$ are independent. Note that (21) and (22) satisfy Assumption 1B. In (21), $Y_{2}$ does not have any instrumental intervention although it has an invalid instrument $X_{2}$. Similarly, in (22), neither does $Y_{1}$ have any instrumental intervention while having an invalid instrument $X_{1}$. As in the previous case, $\bm{\theta}$ and $\widetilde{\bm{\theta}}$ yield the same distribution $\operatorname{\mathbb{P}}(\bm{Y}\mid\bm{X})$ because they share the same $(\bm{V},\bm{\Omega})$ even with different values of $(\bm{U},\bm{W},\bm{\Sigma})$. In this case, it is impossible to infer the directed relation between $Y_{1}$ and $Y_{2}$. Figure 9: (a) Display of DAG defined by (19). (b) Display of DAG defined by (20). Figure 10: (a) Display of DAG defined by (21). (b) Display of DAG defined by (22). ### A.2 Illustration of Algorithm 2 We now illustrate Algorithm 2 by Example 3. ###### Example 3 Consider model (1) with $p=q=5$, $\begin{aligned} Y_{1}&=X_{1}+\varepsilon_{1},\\\ Y_{3}&=0.5Y_{2}+X_{5}+\varepsilon_{3},\\\ \end{aligned}\qquad\begin{aligned} Y_{2}&=0.5Y_{1}+X_{3}+\varepsilon_{2},\\\ Y_{4}&=0.5Y_{3}-0.1Y_{1}+X_{2}+\varepsilon_{4},\\\ \end{aligned}\qquad\begin{aligned} Y_{5}&=X_{4}+\varepsilon_{5},\\\ &\end{aligned}$ (23) where $\varepsilon_{1},\ldots,\varepsilon_{5}\sim N(0,1)$ are independent. Then (23) defines a DAG as displayed in Figure 1(a). For illustration, we generate a random sample of size $n=40$ and compute $\widehat{\bm{V}}$ by Algorithm 1. In particular, $\bm{V}^{\circ}=\begin{pmatrix}1&0.5&0.25&0.025&0\\\ 0&0&0&1&0\\\ 0&1&0.5&0.25&0\\\ 0&0&0&0&1\\\ 0&0&1&0.5&0\end{pmatrix},\quad\widehat{\bm{V}}=\begin{pmatrix}0.92&0.48&0.27&0&0\\\ 0&0&0&1.08&0\\\ 0&1.03&0.52&0.21&0\\\ 0&0&0&0&1.06\\\ 0&0&0.98&0.55&0\end{pmatrix}.\quad$ Algorithm 2 proceeds as follows. * • $h=0$: We have the set $\\{X_{2},X_{4}\\}$ of instruments on leaves. * – $X_{2}$ is identified as an instrument of leaf node $Y_{4}$ ($X_{2}\to Y_{4}$) as $\widehat{V}_{24}\neq 0$ is the only nonzero in the row 2 with the smallest (positive) row $\ell_{0}$-norm. * – $X_{4}$ is identified as an instrument of leaf node $Y_{5}$ ($X_{4}\to Y_{5}$) as $\widehat{V}_{45}\neq 0$ is the only nonzero in row 4 with the smallest (positive) row $\ell_{0}$-norm. Then $h_{G}(4)=h_{G}(5)=0$ and $\\{Y_{4},Y_{5}\\}$ are removed. * • $h=1$: We have the set $\\{X_{5}\\}$ of instrument on leaf. * – $X_{5}$ is identified as an instrument of a leaf node $Y_{3}$ ($X_{5}\to Y_{3}$) in the subgraph for $Y_{1},Y_{2},Y_{3}$ given that $\widehat{V}_{53}\neq 0$ is the only nonzero element in the row with the smallest (positive) row $\ell_{0}$-norm of the submatrix for $Y_{1},Y_{2},Y_{3}$. Since $\widehat{V}_{54}\neq 0$ and $h_{G}(4)=0$, we have $Y_{3}\to Y_{4}$ by Proposition 3. Then $h_{G}(3)=1$ and $\\{Y_{3}\\}$ is removed. * • $h=2$: We have the set $\\{X_{3}\\}$ of instrument on leaf. * – $X_{3}$ is identified as an instrument of a leaf node $Y_{2}$ ($X_{3}\to Y_{2}$) similarly in the subgraph for $Y_{1},Y_{2}$ given that $\widehat{V}_{32}\neq 0$ is the largest nonzero element in its row of the submatrix. Since $\widehat{V}_{33}\neq 0$ and $h_{G}(3)=1$, we have $Y_{2}\to Y_{3}$. Then $h_{G}(2)=2$ and $\\{Y_{2}\\}$ is removed. * • $h=3$: We have the set $\\{X_{1}\\}$ of instrument on leaf. * – $X_{1}$ is an instrument of $Y_{1}$ ($X_{1}\to Y_{1}$). Since $\widehat{V}_{12}\neq 0$ and $h_{G}(2)=2$, we have $Y_{1}\to Y_{2}$. Then $h_{G}(1)=3$, $\\{Y_{1}\\}$ is removed, and the peeling process is terminated. Finally, Step 6 of Algorithm 2 identifies the ancestral relations $\widehat{\mathcal{A}}=\\{(1,2),(2,3),(3,4),(1,3),(1,4),(2,4)\\}$ and the candidate interventional relations $\widehat{\mathcal{I}}=\\{(1,1),(1,2),(1,3),(2,4),(3,2),(3,3),(3,4),(4,5),(5,3),(5,4)\\}.$ In Example 3, $\widehat{\bm{V}}\neq\bm{V}^{\circ}$, suggesting that the selection consistency of matrix $\bm{V}^{\circ}$ is unnecessary for Algorithm 2 to correctly reconstruct ancestral relations $\mathcal{A}^{\circ}$; see also Section A.3. ### A.3 Relaxation of Assumption 4 Assumption 4 in Theorem 4 leads to consistent identification for $\bm{V}^{\circ}$. Now, we discuss when Algorithm 2 correctly reconstruct ancestral relations $\mathcal{A}^{\circ}$ without requiring Assumption 4. ###### Assumption 6 There exists $(\kappa_{j}^{*},\tau^{*}_{j})$ such that $\kappa_{j}^{*}=|\\{l:|V^{\circ}_{lj}|\geq\tau_{j}^{*}\\}|$ and $\begin{split}&\min\Big{(}\min_{l:X_{l}\to Y_{j}}|V^{\circ}_{lj}|,\min_{\begin{subarray}{c}l:X_{l}\to Y_{k}\to Y_{j}\\\ \text{ with }h(k)=h(j)+1\end{subarray}}|V^{\circ}_{lj}|\Big{)}\\\ &\geq\tau^{*}_{j}\geq 100\frac{c_{2}}{c_{1}}\sqrt{\Omega_{jj}^{-1}(\kappa^{\circ}_{j}-\kappa^{*}_{j}+1)\Big{(}\frac{\log(q)}{n}+\frac{\log(n)}{n}\Big{)}};\quad 1\leq j\leq p.\end{split}$ (24) Assumption 6 requires a subset of nonzero entries in $\bm{V}^{\circ}$ to exceed a certain signal strength. These signals enable us to reconstruct the topological layers and $\mathcal{A}^{\circ}$. Note that Assumption 6 reduces to Assumption 4 when $\kappa^{*}_{j}=\kappa^{\circ}_{j}$. ###### Theorem 9 Under Assumptions 1-3 and 6 with constants $c_{1}<6c_{2}$, if the machine precision $\text{tol}\ll 1/n$ is negligible and tuning parameters of Algorithm 1 satisfy: 1. (1) $\gamma_{j}\leq c_{1}/6\sqrt{\kappa^{\circ}_{j}-\kappa^{*}_{j}+1}$, 2. (2) $\sqrt{32c_{2}^{2}\Omega_{jj}^{-1}(\log(q)+\log(n))/n}/\gamma_{j}\leq\tau_{j}\leq 0.4\tau^{*}_{j}$, 3. (3) $\kappa_{j}=\kappa_{j}^{\circ}$, for $j=1,\ldots,p$, then Algorithm 1 terminates in at most $1+\lceil\log(\kappa_{\max}^{\circ})/\log 4\rceil$ iterations almost surely. Moreover, $\widehat{\mathcal{A}}=\mathcal{A}^{\circ}$, $\widehat{\mathcal{C}}\supseteq\\{(l,j):|V_{lj}^{\circ}|\geq\tau^{*}_{j}\\}\supseteq\mathcal{I}^{\circ}$, and $\widehat{\mathcal{D}}=\mathcal{D}^{\circ}$ almost surely as $n\to\infty$. The proof of Theorem 9 is given in Appendix B.9. ### A.4 Comparison of strong faithfulness and Assumption 4 (or 6) In the literature, a faithfulness condition is usually assumed for identifiability up to Markov equivalence classes (Spirtes et al., 2000). For discussion, we formally introduce the concepts of faithfulness and strong faithfulness. Consider a DAG $G$ with node variables $(Z_{1},\ldots,Z_{p+q})^{\top}$. Nodes $Z_{i}$ and $Z_{j}$ are adjacent if $Z_{i}\to Z_{j}$ or $Z_{j}\to Z_{i}$. A _path_ (_undirected_) between $Z_{i}$ and $Z_{j}$ in $G$ is a sequence of distinct nodes $(Z_{i},\ldots,Z_{j})$ such that all pairs of successive nodes in the sequence are adjacent. A nonendpoint node $Z_{k}$ on a path $(Z_{i},\ldots,Z_{k-1},Z_{k},Z_{k+1},\ldots,Z_{j})$ is a _collider_ if $Z_{k-1}\to Z_{k}\leftarrow Z_{k+1}$. Otherwise it is a _noncollider_. Let $A\subseteq\\{1,\ldots,p+q\\}$, where $A$ does not contain $i$ and $j$. Then $\bm{Z}_{A}$ _blocks_ a path $(Z_{i},\ldots,Z_{j})$ if at least one of the following holds: (i) the path contains a noncollider that is in $\bm{Z}_{A}$, or (ii) the path contains a collider that is not in $\bm{Z}_{A}$ and has no descendant in $\bm{Z}_{A}$. A node $Z_{i}$ is d-separated from $Z_{j}$ given $\bm{Z}_{A}$ if $\bm{Z}_{A}$ block every path between $Z_{i}$ and $Z_{j}$; $i\neq j$ (Pearl, 2000). According to Uhler et al. (2013), a multivariate Gaussian distribution of $(Z_{1},\ldots,Z_{p+q})^{\top}$ is said to be $\varsigma$-strong faithful to a DAG with node set $\mathcal{V}=\\{1,\ldots,p+q\\}$ if $\min_{A\subseteq\mathcal{V}\backslash\\{i,j\\}}\Big{\\{}|{\operatorname{Corr}}(Z_{i},Z_{j}\mid\bm{Z}_{A})|:Z_{i}\text{ is not d-separated from }Z_{j}\text{ given }\bm{Z}_{A}\Big{\\}}>\varsigma,$ (25) for $1\leq i\neq j\leq p+q$, where $\varsigma\in[0,1)$, ${\operatorname{Corr}}$ denotes the correlation. When $\varsigma=0$, (25) is equivalent to faithfulness. For consistent structure learning (up to Markov equivalence classes), it often requires that $\varsigma\gtrsim\sqrt{s_{0}\log(p+q)/n}$, where $s_{0}$ is a sparsity measure; see Uhler et al. (2013) for a survey. For a pair $(i,j)$, the number of possible sets for $A$ is $2^{(p+q-2)}$. If $Z_{i}\to Z_{j}$, then $\text{Corr}(Z_{i},Z_{j}\mid\bm{Z}_{A})\neq 0$ for any $A$. Therefore, for this $(i,j)$ pair alone, (25) could require exponentially many conditions. Indeed, (25) is very restrictive in high-dimensional situation (Uhler et al., 2013). By comparison, Algorithm 2 yields consistent structure learning based on Assumption 4 or 6 instead of strong faithfulness. In some sense, Assumption 4 or 6 requires the sufficient signal strength that is analogous to the condition for consistent feature selection (Shen et al., 2012). This assumption may be thought of as an alternative to strong faithfulness. As illustrated in Example 4, Assumption 4 or 6 is less stringent than strong faithfulness. ###### Example 4 (Faithfulness) For simplicity, assume $\bm{X}\sim N(\bm{0},\bm{I})$. Consider model (1) with $p=q=3$, $\displaystyle Y_{1}=W_{11}X_{1}+\varepsilon_{1},\quad Y_{2}=U_{12}Y_{1}+W_{22}X_{2}+\varepsilon_{2},\quad Y_{3}=U_{13}Y_{1}+U_{23}Y_{2}+W_{33}X_{3}+\varepsilon_{3},$ where $\varepsilon_{1},\varepsilon_{2},\varepsilon_{3}\sim N(0,1)$ are independent and $U_{12},U_{13},U_{23},W_{11},W_{22},W_{33}\neq 0$. Let $\bm{Z}=(Y_{1},Y_{2},Y_{3},X_{1},X_{2},X_{3})^{\top}$. Since directed relations among $\bm{X}$ are not of interest, (25) becomes $\begin{split}\min_{\begin{subarray}{c}\bm{Z}_{A}=(\bm{Y}_{A_{1}},\bm{X}_{A_{2}}):\\\ A_{1}\subseteq\\{i,j\\}^{c},A_{2}\end{subarray}}&\Big{\\{}|{\operatorname{Corr}}(Y_{i},Y_{j}\mid\bm{Z}_{A})|:Y_{i},Y_{j}\text{ are not d-separated given }\bm{Z}_{A}\Big{\\}}>\varsigma,\\\ \min_{\begin{subarray}{c}\bm{Z}_{A}=(\bm{Y}_{A_{1}},\bm{X}_{A_{2}}):\\\ A_{1}\subseteq\\{j\\}^{c},A_{2}\subseteq\\{l\\}^{c}\end{subarray}}&\Big{\\{}|{\operatorname{Corr}}(Y_{j},X_{l}\mid\bm{Z}_{A})|:Y_{j},X_{l}\text{ are not d-separated given }\bm{Z}_{A}\Big{\\}}>\varsigma,\end{split}$ (26) for each pair $(i,j)$ with $i\neq j$ and each pair $(l,j)$. Then strong- faithfulness in (26) assumes 152 conditions for the correlations. By comparison, (17) requires the absolute values of $V_{11},V_{12},V_{13},V_{22},V_{23},V_{33}\gtrsim\sqrt{\log(q)/n}$, which in turn requires the minimal absolute value of six correlations $\gtrsim\sqrt{\log(q)/n}$, (i) ${\operatorname{Corr}}(Y_{1},X_{1}\mid X_{2},X_{3})$, (ii) ${\operatorname{Corr}}(Y_{2},X_{1}\mid X_{2},X_{3})$, (iii) ${\operatorname{Corr}}(Y_{3},X_{1}\mid X_{2},X_{3})$, (iv) ${\operatorname{Corr}}(Y_{2},X_{2}\mid X_{1},X_{3})$, (v) ${\operatorname{Corr}}(Y_{3},X_{2}\mid X_{1},X_{3})$, (vi) ${\operatorname{Corr}}(Y_{3},X_{3}\mid X_{1},X_{2})$. Importantly, (i)-(vi) are required in (26), suggesting that strong- faithfulness is more stringent than (17). ### A.5 Irregular hypothesis Assume, without loss of generality, that $\widehat{\mathcal{D}}=\mathcal{D}^{\circ}$ and $\widehat{\mathcal{A}}=\mathcal{A}^{\circ}$ are correctly reconstructed in the following discussion. * • For testing of directed edges (2), suppose $H_{0}$ is irregular, namely, $\mathcal{D}^{\circ}\cup\mathcal{E}^{\circ}$ contains a directed cycle. This implies that a directed cycle exists in $\widehat{\mathcal{D}}\cup\widehat{\mathcal{A}}$. In this situation, we decompose $H_{0}$ into sub-hypotheses $H_{0}^{(1)},\ldots,H_{0}^{(\nu)}$, each of which is regular. Then testing $H_{0}$ is equivalent to multiple testing for $H_{0}^{(1)},\ldots,H_{0}^{(\nu)}$. For instance, in Example 1, $H_{0}:U_{45}=U_{53}=0$ is irregular, and $H_{0}$ can be decomposed into $H_{0}^{(1)}:U_{45}=0$ and $H_{0}^{(2)}:U_{53}=0$. * • For testing of directed pathways (3), if $H_{0}$ is irregular, then $\widehat{\mathcal{D}}\cup\widehat{\mathcal{A}}$ has a directed cycle. The p-value is defined to be one in this situation since no evidence supports the presence of the pathway. ## B Technical proofs ### B.1 Proof of Proposition 1 Suppose that $\bm{\theta}=(\bm{U},\bm{W},\bm{\Sigma})$ and $\widetilde{\bm{\theta}}=(\widetilde{\bm{U}},\widetilde{\bm{W}},\widetilde{\bm{\Sigma}})$ render the same distribution of $(\bm{X},\bm{Y})$. We will prove that $\bm{\theta}=\widetilde{\bm{\theta}}$. Denote by $G(\bm{\theta})$ and $G(\widetilde{\bm{\theta}})$ the DAGs corresponding to $\bm{\theta}$ and $\widetilde{\bm{\theta}}$, respectively. First, consider $G(\bm{\theta})$. Without loss of generality, assume $Y_{1}$ is a leaf node. By Assumption 1C, there exists an instrumental intervention with respect to $G(\bm{\theta})$, say $X_{1}$. Then, $\displaystyle\operatorname{Cov}(Y_{j},X_{1}\mid\bm{X}_{\\{2,\ldots,q\\}})=0,$ $\displaystyle\quad j=2,\ldots,p,$ (27) $\displaystyle\operatorname{Cov}(Y_{1},X_{1}\mid\bm{Y}_{A},\bm{X}_{\\{2,\ldots,q\\}})\neq 0,$ $\displaystyle\quad\text{for any }A\subseteq\\{2,\ldots,p\\}.$ (28) By the local Markov property (Spirtes et al., 2000), (28) implies that $X_{1}\to Y_{1}$ in $G(\widetilde{\bm{\theta}})$. Suppose $Y_{1}$ is not a leaf node in $G(\widetilde{\bm{\theta}})$. Assume, without loss of generality, that $Y_{1}\to Y_{2}$ with $h_{G(\widetilde{\bm{\theta}})}(1)=h_{G(\widetilde{\bm{\theta}})}(2)+1$. Then $\operatorname{Cov}(Y_{2},X_{1}\mid\bm{X}_{\\{2,\ldots,q\\}})=0$ but $X_{1}\to Y_{1}$ and $Y_{1}\to Y_{2}$ with $h_{G(\widetilde{\bm{\theta}})}(1)=h_{G(\widetilde{\bm{\theta}})}(2)+1$, which contradicts to Assumption 1B. This implies that if $Y_{1}$ is a leaf node in $G(\bm{\theta})$ then it must be a leaf node in $G(\widetilde{\bm{\theta}})$. In both $G(\bm{\theta})$ and $G(\widetilde{\bm{\theta}})$, the parents and interventions of $Y_{1}$ can be identified by $\begin{split}\operatorname{\mathbb{E}}(Y_{1}\mid\bm{Y}_{\\{2,\ldots,p\\}},\bm{X})&=\operatorname{\mathbb{E}}(Y_{1}\mid\bm{Y}_{\textnormal{{pa}}_{G(\bm{\theta})}(1)},\bm{X})=\operatorname{\mathbb{E}}(Y_{1}\mid\bm{Y}_{\textnormal{{pa}}_{G(\bm{\theta})}(1)},\bm{X}_{\textnormal{{in}}_{G(\bm{\theta})}(1)}),\\\ \operatorname{\mathbb{E}}(Y_{1}\mid\bm{Y}_{\\{2,\ldots,p\\}},\bm{X})&=\operatorname{\mathbb{E}}(Y_{1}\mid\bm{Y}_{\textnormal{{pa}}_{G(\widetilde{\bm{\theta}})}(1)},\bm{X})=\operatorname{\mathbb{E}}(Y_{1}\mid\bm{Y}_{\textnormal{{pa}}_{G(\widetilde{\bm{\theta}})}(1)},\bm{X}_{\textnormal{{in}}_{G(\widetilde{\bm{\theta}})}(1)}).\end{split}$ Consequently, $Y_{1}$ has the same parents and interventions in $G(\bm{\theta})$ and $G(\widetilde{\bm{\theta}})$. The forgoing argument is applied to other nodes sequentially. First, we remove $Y_{1}$ with any directed edges to $Y_{1}$, which does not alter the joint distribution of $(\bm{Y}_{\\{2,\ldots,p\\}},\bm{X})$ and the subgraph of nodes $Y_{2},\ldots,Y_{p}$. By induction, we remove the leafs in $G(\bm{\theta})$ until it is empty, leading to $G(\bm{\theta})=G(\widetilde{\bm{\theta}})$. Finally, $\bm{\theta}=\widetilde{\bm{\theta}}$ because they have the same locations for nonzero elements and these model parameters (or regression coefficients) are uniquely determined under Assumption 1A (Shojaie and Michailidis, 2010). This completes the proof. ### B.2 Proof of Proposition 2 It follows from Assumption 1B that $\mathcal{C}=\\{(l,j):V_{lj}\neq 0\\}\supseteq\mathcal{I}$. #### Proof of (A) Note that the maximal length of a path in a DAG of $p$ nodes is at most $p-1$. Then it can be verified that $\bm{U}$ is nilpotent in that $\bm{U}^{p}=\bm{0}$. An application of the matrix series expansion yields that $(\bm{I}-\bm{U})^{-1}=\bm{I}+\bm{U}+\ldots+\bm{U}^{p-1}$. Using the fact that $\bm{V}=\bm{W}(\bm{I}-\bm{U})^{-1}$ from (6), we have, for any $1\leq l,j\leq p$, $V_{lj}=\sum_{k=1}^{p}W_{lk}\Big{(}I_{kj}+U_{kj}+\ldots+(\bm{U}^{p-1})_{kj}\Big{)},$ where $U_{kj}$ is the $(k,j)$th entry of $\bm{U}$. If $V_{lj}\neq 0$, then there exists $k$ such that $W_{lk}\neq 0$ and $(\bm{U}^{r})_{kj}\neq 0$ for some $0\leq r\leq p-1$. If $r=0$, then we must have $k=j$, and $X_{l}\to Y_{j}$. If $r>0$, then $X_{l}\to Y_{k}$ and $Y_{k}$ is an ancestor of $Y_{j}$. #### Proof of (B) First, for any leaf node variable $Y_{j}$, by Assumption 1, there exists an instrument $X_{l}\to Y_{j}$. If $V_{lj^{\prime}}\neq 0$ for some $j^{\prime}\neq j$, then $Y_{j}$ must be an ancestor of $Y_{j^{\prime}}$, which contradicts the fact that $Y_{j}$ is a leaf node variable. Conversely, suppose that $V_{lj}\neq 0$ and $V_{lj^{\prime}}=0$ for $j^{\prime}\neq j$. If $Y_{j}$ is not a leaf node variable, then there exists a variable $Y_{j^{\prime}}$ such that $Y_{j}\to Y_{j^{\prime}}$ with $h_{G}(j)=h_{G}(j^{\prime})+1$, that is $U_{jj^{\prime}}\neq 0$ and $(\bm{U}^{r})_{jj^{\prime}}=0$ for $r>1$. Then $V_{lj^{\prime}}=W_{lj}U_{jj^{\prime}}\neq 0$, a contradiction. ### B.3 Proof of Proposition 3 Suppose $Y_{k}\to Y_{j}$ and $h_{G}(j)=h_{G}(k)-1$. Let $X_{l}$ be an instrument of $Y_{k}$ in $G^{\prime}$. Then there are two cases: (1) $X_{l}$ intervenes on $Y_{k}$ but does not intervene on $Y_{j}$, namely $X_{l}\to Y_{k}$ but $X_{l}\not\to Y_{j}$; (2) $X_{l}$ intervenes on $Y_{k}$ and $Y_{j}$ simultaneously, namely $X_{l}\to Y_{k}$ and $X_{l}\to Y_{j}$. For (1), $V_{lj}=W_{lk}U_{kj}\neq 0$. For (2), Assumption 1B implies that $V_{lj}\neq 0$. This holds for every instrument of $Y_{k}$ in $G^{\prime}$, and the desired result follows. Conversely, suppose for each instrument $X_{l}$ of $Y_{k}$ in $G^{\prime}$, we have $V_{lj}\neq 0$. Let $X_{l^{\prime}}$ be an instrument of $Y_{k}$ in $G$, which is also an instrument in $G^{\prime}$. Then $0\neq V_{l^{\prime}j}=W_{l^{\prime}k}U_{kj}$, which implies $U_{kj}\neq 0$. This completes the proof. ### B.4 Proof of Theorem 4 The proof proceeds in two steps: (A) Show that $\\{l:\widehat{V}_{lj}\neq 0\\}=\\{l:V_{lj}^{\circ}\neq 0\\}$ almost surely for $j=1,\ldots,p$; and (B) Show that $\widehat{S}=S^{\circ}$ if $\widehat{\bm{V}}$ satisfies the property in (A). As a result, $\widehat{\mathcal{D}}=\mathcal{D}^{\circ}$ follows immediately. #### Proof of (A) Let $A_{j}^{\circ}=\\{l:V^{\circ}_{lj}\neq 0\\}$ and $A^{[t]}_{j}=\\{l:|\widetilde{V}^{[t]}_{lj}|\geq\tau_{j}\\}$ be the estimated support of penalized solution at the $t$-th iteration of Algorithm 1. For the penalized solution, define the false negative set $\text{FN}^{[t]}_{j}=A^{\circ}_{j}\setminus A_{j}^{[t]}$ and the false positive set $\text{FP}^{[t]}_{j}=A_{j}^{[t]}\setminus A_{j}^{\circ}$; $t\geq 0$. Consider a “good” event $\mathscr{E}_{j}=\\{\|\bm{\mathsf{X}}^{\top}\widehat{\bm{\xi}}_{j}/n\|_{\infty}\leq 0.5\gamma_{j}\tau_{j}\\}\cap\\{\|\widehat{\bm{V}}^{\circ}_{\cdot j}-\bm{V}^{\circ}_{\cdot j}\|_{\infty}\leq 0.5\tau_{j}\\}$, where $\widehat{\bm{\xi}}_{j}=\bm{\mathsf{Y}}_{j}-\bm{\mathsf{X}}\widehat{\bm{V}}^{\circ}_{\cdot j}$ is the residual of the least squares estimate $\widehat{\bm{V}}^{\circ}_{\cdot j}$ such that $A_{j}^{\circ}=\\{l:\widehat{V}^{\circ}_{lj}\neq 0\\}$; $j=1,\ldots,p$. We shall show that $\text{FN}^{[t]}_{j}$ and $\text{FP}^{[t]}_{j}$ are eventually empty sets on event $\mathscr{E}_{j}$ which has a probability tending to one. First, we show that if $|A_{j}^{\circ}\cup A^{[t-1]}_{j}|\leq 2\kappa_{\max}^{\circ}$ on $\mathscr{E}_{j}$ for $t\geq 1$, then $|A_{j}^{\circ}\cup A^{[t]}_{j}|\leq 2\kappa^{\circ}_{\max}$, to be used in Assumption 2. Now suppose $|A_{j}^{\circ}\cup A^{[t-1]}_{j}|\leq 2\kappa^{\circ}_{\max}$ on $\mathscr{E}_{j}$ for $t\geq 1$. By the optimality condition of (8), $\begin{split}\Big{(}\widehat{\bm{V}}^{o}_{\cdot j}-\widetilde{\bm{V}}^{[t]}_{\cdot j}\Big{)}^{\top}\Big{(}-\bm{\mathsf{X}}^{\top}(\bm{\mathsf{Y}}_{j}-\bm{\mathsf{X}}\widetilde{\bm{V}}^{[t]}_{\cdot j})/n+\gamma_{j}\tau_{j}\nabla\|\widetilde{\bm{V}}_{(A^{[t-1]}_{j})^{c},j}^{[t]}\|_{1}\Big{)}\geq 0,\end{split}$ where $\widetilde{\bm{V}}^{[t]}$ is defined in (8). Plugging $\widehat{\bm{\xi}}_{j}=\bm{\mathsf{Y}}_{j}-\bm{\mathsf{X}}\widehat{\bm{V}}^{\circ}_{\cdot j}$ into the inequality and rearranging it, we have $\|\bm{\mathsf{X}}(\widetilde{\bm{V}}_{\cdot j}^{[t]}-\widehat{\bm{V}}_{\cdot j}^{\circ})\|_{2}^{2}/n$ is no greater than $\begin{split}&\Big{(}\widetilde{\bm{V}}_{\cdot j}^{[t]}-\widehat{\bm{V}}_{\cdot j}^{\circ}\Big{)}^{\top}\Big{(}\bm{\mathsf{X}}^{\top}\widehat{\bm{\xi}}_{j}/n-\gamma_{j}\tau_{j}\nabla\|\widetilde{\bm{V}}^{[t]}_{(A_{j}^{[t-1]})^{c},j}\|_{1}\Big{)}\\\ &=\Big{(}\widetilde{\bm{V}}_{A_{j}^{\circ}\setminus A_{j}^{[t-1]},j}^{[t]}-\widehat{\bm{V}}_{A_{j}^{\circ}\setminus A_{j}^{[t-1]},j}^{\circ}\Big{)}^{\top}\Big{(}\bm{\mathsf{X}}^{\top}_{A_{j}^{\circ}\setminus A^{[t-1]}_{j}}\widehat{\bm{\xi}}_{j}/n-\gamma_{j}\tau_{j}\nabla\|\widetilde{\bm{V}}^{[t]}_{A_{j}^{\circ}\setminus A^{[t-1]}_{j},j}\|_{1}\Big{)}\\\ &\quad+\Big{(}\widetilde{\bm{V}}_{(A_{j}^{\circ}\cup A^{[t-1]}_{j})^{c},j}^{[t]}-\widehat{\bm{V}}_{(A_{j}^{\circ}\cup A^{[t-1]}_{j})^{c},j}^{\circ}\Big{)}^{\top}\Big{(}\bm{\mathsf{X}}^{\top}\widehat{\bm{\xi}}/n-\gamma_{j}\tau_{j}\nabla\|\widetilde{\bm{V}}^{[t]}_{(A_{j}^{\circ}\cup A^{[t-1]}_{j})^{c},j}\|_{1}\Big{)}\\\ &\quad+\Big{(}\widetilde{\bm{V}}_{A^{[t-1]}_{j}\setminus A_{j}^{\circ},j}^{[t]}-\widehat{\bm{V}}_{A^{[t-1]}_{j}\setminus A_{j}^{\circ},j}^{\circ}\Big{)}^{\top}\bm{\mathsf{X}}^{\top}_{A_{j}^{[t-1]}\setminus A_{j}^{\circ}}\widehat{\bm{\xi}}/n.\end{split}$ (29) Note that $\Big{(}\widetilde{\bm{V}}_{(A_{j}^{\circ}\cup A_{j}^{[t-1]})^{c},j}^{[t]}-\widehat{\bm{V}}_{(A_{j}^{\circ}\cup A_{j}^{[t-1]})^{c},j}^{\circ}\Big{)}^{\top}\nabla\Big{\|}\widetilde{\bm{V}}^{[t]}_{(A_{j}^{\circ}\cup A_{j}^{[t-1]})^{c},j}\Big{\|}_{1}=\Big{\|}\widetilde{\bm{V}}_{(A_{j}^{\circ}\cup A_{j}^{[t-1]})^{c},j}^{[t]}-\widehat{\bm{V}}_{(A_{j}^{\circ}\cup A_{j}^{[t-1]})^{c},j}^{\circ}\Big{\|}_{1}.$ Then (29) is no greater than $\begin{split}&\Big{\|}\widetilde{\bm{V}}_{A_{j}^{\circ}\triangle A^{[t-1]}_{j},j}^{[t]}-\widehat{\bm{V}}_{A_{j}^{\circ}\triangle A^{[t-1]}_{j},j}^{\circ}\Big{\|}_{1}\Big{(}\|\bm{\mathsf{X}}^{\top}\widehat{\bm{\xi}}_{j}/n\|_{\infty}+\gamma_{j}\tau_{j}\Big{)}\\\ &+\Big{\|}\widetilde{\bm{V}}_{(A_{j}^{\circ}\cup A_{j}^{[t-1]})^{c},j}^{[t]}-\widehat{\bm{V}}_{(A_{j}^{\circ}\cup A_{j}^{[t-1]})^{c},j}^{\circ}\Big{\|}_{1}\Big{(}\|\bm{\mathsf{X}}^{\top}\widehat{\bm{\xi}}_{j}/n\|_{\infty}-\gamma_{j}\tau_{j}\Big{)},\end{split}$ (30) where $\triangle$ denotes the symmetric difference. Note that $\|\bm{\mathsf{X}}(\widetilde{\bm{V}}_{\cdot j}^{[t]}-\widehat{\bm{V}}_{\cdot j}^{\circ})\|_{2}^{2}/n\geq 0$. Rearranging the inequality yields that $\begin{split}&\Big{(}\gamma_{j}\tau_{j}-\|\bm{\mathsf{X}}^{\top}\widehat{\bm{\xi}}_{j}/n\|_{\infty}\Big{)}\Big{\|}\widetilde{\bm{V}}_{(A_{j}^{\circ}\cup A_{j}^{[t-1]})^{c},j}^{[t]}-\widehat{\bm{V}}_{(A_{j}^{\circ}\cup A_{j}^{[t-1]})^{c},j}^{\circ}\Big{\|}_{1}\\\ &\leq\Big{(}\|\bm{\mathsf{X}}^{\top}\widehat{\bm{\xi}}_{j}/n\|_{\infty}+\gamma_{j}\tau_{j}\Big{)}\Big{\|}\widetilde{\bm{V}}_{A_{j}^{\circ}\triangle A^{[t-1]}_{j},j}^{[t]}-\widehat{\bm{V}}_{A_{j}^{\circ}\triangle A^{[t-1]}_{j},j}^{\circ}\Big{\|}_{1}.\end{split}$ On event $\mathscr{E}_{j}$, $\|\bm{\mathsf{X}}^{\top}\widehat{\bm{\xi}}_{j}/n\|_{\infty}\leq\gamma_{j}\tau_{j}/2$, implying that $\|\widetilde{\bm{V}}_{(A_{j}^{\circ}\cup A_{j}^{[t-1]})^{c},j}^{[t]}-\widehat{\bm{V}}_{(A_{j}^{\circ}\cup A_{j}^{[t-1]})^{c},j}^{\circ}\|_{1}\leq 3\|\widetilde{\bm{V}}_{A_{j}^{\circ}\triangle A^{[t-1]}_{j},j}^{[t]}-\widehat{\bm{V}}_{A_{j}^{\circ}\triangle A^{[t-1]}_{j},j}^{\circ}\|_{1}\leq 3\|\widetilde{\bm{V}}_{A_{j}^{\circ}\cup A^{[t-1]}_{j},j}^{[t]}-\widehat{\bm{V}}_{A_{j}^{\circ}\cup A^{[t-1]}_{j},j}^{\circ}\|_{1}$. Note that $|A_{j}^{\circ}\cup A^{[t-1]}_{j}|\leq 2\kappa^{\circ}_{\max}$. By Assumption 2 and (30), $\begin{split}c_{1}\|\widetilde{\bm{V}}_{\cdot j}^{[t]}-\widehat{\bm{V}}_{\cdot j}^{\circ}\|_{2}^{2}\leq&\Big{(}\|\bm{\mathsf{X}}^{\top}\widehat{\bm{\xi}}_{j}/n\|_{\infty}+\gamma_{j}\tau_{j}\Big{)}\Big{\|}\widetilde{\bm{V}}_{A_{j}^{\circ}\triangle A_{j}^{[t-1]},j}^{[t]}-\widehat{\bm{V}}_{A_{j}^{\circ}\triangle A_{j}^{[t-1]},j}^{\circ}\Big{\|}_{1}\\\ &+\Big{(}\|\bm{\mathsf{X}}^{\top}\widehat{\bm{\xi}}_{j}/n\|_{\infty}-\gamma_{j}\tau_{j}\Big{)}\Big{\|}\widetilde{\bm{V}}_{(A_{j}^{\circ}\cup A_{j}^{[t-1]})^{c},j}^{[t]}-\widehat{\bm{V}}_{(A_{j}^{\circ}\cup A_{j}^{[t-1]})^{c},j}^{\circ}\Big{\|}_{1}\\\ \leq&\Big{(}\|\bm{\mathsf{X}}^{\top}\widehat{\bm{\xi}}_{j}/n\|_{\infty}+\gamma_{j}\tau_{j}\Big{)}\Big{\|}\widetilde{\bm{V}}_{A_{j}^{\circ}\triangle A_{j}^{[t-1]},j}^{[t]}-\widehat{\bm{V}}_{A_{j}^{\circ}\triangle A_{j}^{[t-1]},j}^{\circ}\Big{\|}_{1}.\end{split}$ (31) By the Cauchy-Schwarz inequality, $\Big{\|}\widetilde{\bm{V}}_{A_{j}^{\circ}\triangle A_{j}^{[t-1]},j}^{[t]}-\widehat{\bm{V}}_{A_{j}^{\circ}\triangle A_{j}^{[t-1]},j}^{\circ}\Big{\|}_{1}\leq\sqrt{A_{j}^{\circ}\triangle A^{[t-1]}_{j}}\Big{\|}\widetilde{\bm{V}}_{\cdot j}^{[t]}-\widehat{\bm{V}}^{\circ}_{\cdot j}\Big{\|}_{2}$. Thus, $c_{1}\|\widetilde{\bm{V}}_{\cdot j}^{[t]}-\widehat{\bm{V}}_{\cdot j}^{\circ}\|_{2}\leq 1.5\gamma_{j}\tau_{j}\sqrt{2\kappa^{\circ}_{\max}}$, since $|A_{j}^{\circ}\triangle A_{j}^{[t-1]}|\leq|A_{j}^{\circ}\cup A_{j}^{[t-1]}|\leq 2\kappa^{\circ}_{\max}$ and $\|\bm{\mathsf{X}}^{\top}\widehat{\bm{\xi}}_{j}/n\|_{\infty}\leq 0.5\gamma_{j}\tau_{j}$. By the condition (1) of Theorem 4, $\gamma_{j}\leq c_{1}/6$ and $\|\widetilde{\bm{V}}_{\cdot j}^{[t]}-\widehat{\bm{V}}_{\cdot j}^{\circ}\|_{2}/\tau_{j}\leq\sqrt{\kappa^{\circ}_{\max}}$. On the other hand, for any $l\in\text{FP}^{[t]}_{j}=A^{[t]}_{j}\setminus A_{j}^{\circ}$, we have $|\widetilde{V}_{lj}^{[t]}-\widehat{V}_{lj}^{\circ}|=|\widetilde{V}_{lj}^{[t]}|>\tau_{j}$. Thus, $\sqrt{|\text{FP}^{[t]}_{j}|}\leq\|\widetilde{\bm{V}}_{\cdot j}^{[t]}-\widehat{\bm{V}}_{\cdot j}^{\circ}\|_{2}/\tau_{j}\leq\sqrt{\kappa^{\circ}_{\max}}$. Consequently, $|A_{j}^{\circ}\cup A^{[t]}_{j}|=|A_{j}^{\circ}|+|\text{FP}^{[t]}_{j}|\leq 2\kappa^{\circ}_{\max}$ on $\mathscr{E}_{j}$ for $t\geq 1$. Next, we estimate the number of iterations required for termination. Note that the machine precision is negligible. The termination criterion is met when $A^{[t]}_{j}=A^{[t-1]}_{j}$ since the weighted Lasso problem (8) remains same at the $t$th and $(t-1)$th iterations. To show that $A^{[t]}_{j}=A^{\circ}_{j}$ eventually, we prove that $|\text{FN}^{[t]}_{j}|+|\text{FP}^{[t]}_{j}|<1$ eventually. Now, suppose $|\text{FN}^{[t]}_{j}|+|\text{FP}_{j}^{[t]}|\geq 1$. For any $l\in\text{FN}^{[t]}_{j}\cup\text{FP}_{j}^{[t]}$, by Assumption 4, $|\widetilde{V}^{[t]}_{lj}-\widehat{V}^{\circ}_{lj}|\geq|\widetilde{V}^{[t]}_{lj}-V^{\circ}_{lj}|-|\widehat{V}^{\circ}_{lj}-V^{\circ}_{lj}|\geq\tau_{j}-0.5\tau_{j}$, so $\sqrt{|\text{FN}^{[t]}_{j}|+|\text{FP}^{[t]}_{j}|}\leq\|\widetilde{\bm{V}}_{\cdot j}^{[t]}-\widehat{\bm{V}}_{\cdot j}^{\circ}\|_{2}/0.5\tau_{j}$. By (31) and the Cauchy-Schwarz inequality, $c_{1}\|\widetilde{\bm{V}}_{\cdot j}^{[t]}-\widehat{\bm{V}}_{\cdot j}^{\circ}\|_{2}\leq 1.5\gamma_{j}\tau_{j}\sqrt{|\text{FN}_{j}^{[t-1]}|+|\text{FP}^{[t-1]}_{j}|}.$ By conditions (1) and (2) for $(\tau_{j},\gamma_{j})$ in Theorem 4, we have $\begin{split}\sqrt{|\text{FN}^{[t]}_{j}|+|\text{FP}^{[t]}_{j}|}\leq\frac{\|\widetilde{\bm{V}}_{\cdot j}^{[t]}-\widehat{\bm{V}}_{\cdot j}^{\circ}\|_{2}}{0.5\tau_{j}}&\leq\frac{3\gamma_{j}}{c_{1}}\sqrt{|\text{FN}_{j}^{[t-1]}|+|\text{FP}_{j}^{[t-1]}|}\\\ &\leq 0.5\sqrt{|\text{FN}_{j}^{[t-1]}|+|\text{FP}_{j}^{[t-1]}|}.\end{split}$ Hence, $\sqrt{|\text{FN}^{[t]}_{j}|+|\text{FP}^{[t]}_{j}|}\leq(1/2)^{t}\sqrt{|A^{\circ}_{j}|+|A^{[0]}_{j}|}$. In particular, for $t\geq 1+\lceil\log\kappa^{\circ}_{j}/\log 4\rceil$, $|\text{FN}_{j}^{[t]}|+|\text{FP}^{[t]}_{j}|<1$ implying that $\text{FN}_{j}^{[t]}=\emptyset$ and $\text{FP}_{j}^{[t]}=\emptyset$. Let $t_{\max}=1+\lceil\log\kappa^{\circ}_{j}/\log 4\rceil$. Then $\text{FN}_{j}^{[t_{\max}]}=\text{FP}^{[t_{\max}]}_{j}=\emptyset$ on event $\mathscr{E}_{j}$. Under condition (3), we have $\\{l:\widetilde{V}^{[t_{\max}]}_{lj}\neq 0\\}=\\{l:V_{lj}^{\circ}\neq 0\\}$. To bound $\operatorname{\mathbb{P}}(\bigcup_{j=1}^{p}\mathscr{E}_{j}^{c})$, let $\bm{\eta}=\bm{\mathsf{X}}^{\top}(\bm{I}-\bm{\mathsf{P}}_{A^{\circ}_{j}}){\bm{\xi}}_{j}$ and $\bm{\eta}^{\prime}=n(\bm{\mathsf{X}}^{\top}_{A^{\circ}_{j}}\bm{\mathsf{X}}_{A^{\circ}_{j}})^{-1}\bm{\mathsf{X}}^{\top}_{A_{j}^{\circ}}{\bm{\xi}}_{j}$, where $\bm{\xi}_{j}=((\varepsilon_{V})_{1j},\ldots,(\varepsilon_{V})_{nj})^{\top}\in\mathbb{R}^{n}$. Then $\bm{\eta}\in\mathbb{R}^{q}$ and $\bm{\eta}^{\prime}\in\mathbb{R}^{|\kappa^{\circ}_{j}|}$ are Gaussian random vectors, where $\mathop{\rm Var}(\eta_{l})\leq\Omega_{jj}^{-1}(\bm{\mathsf{X}}^{\top}\bm{\mathsf{X}})_{ll}\leq n\Omega_{jj}^{-1}c_{2}^{2}$ and $\mathop{\rm Var}(\eta^{\prime}_{l})\leq\Omega_{jj}^{-1}((\bm{\mathsf{X}}^{\top}_{A^{\circ}_{j}}\bm{\mathsf{X}}_{A^{\circ}_{j}})^{-1})_{ll}\leq n\Omega^{-1}_{jj}c_{2}^{2}$. Then $\begin{split}\operatorname{\mathbb{P}}\Big{(}\|\bm{\mathsf{X}}^{\top}\widehat{\bm{\xi}}_{j}/n\|_{\infty}>0.5\gamma_{j}\tau_{j}\Big{)}&=\operatorname{\mathbb{P}}\Big{(}\|\bm{\eta}/n\|_{\infty}>0.5\gamma_{j}\tau_{j}\Big{)}\\\ &\leq\sum_{l=1}^{q}\operatorname{\mathbb{P}}\Big{(}|\eta_{l}|\geq 0.5n\gamma_{j}\tau_{j}\Big{)}\\\ &\leq q\int_{\frac{0.5\sqrt{n\Omega_{jj}}\gamma_{j}\tau_{j}}{c_{2}}}^{\infty}\frac{e^{-t^{2}/2}}{\sqrt{2\pi}}dt\leq q\sqrt{\frac{2}{\pi}}\exp\Big{(}-\frac{n\Omega_{jj}\gamma^{2}_{j}\tau^{2}_{j}}{8c^{2}_{2}}\Big{)},\end{split}$ and similarly, $\begin{split}\operatorname{\mathbb{P}}\Big{(}\|\widehat{\bm{V}}_{\cdot j}^{\circ}-\bm{V}_{\cdot j}^{\circ}\|_{\infty}>0.5\tau_{j}\Big{)}\leq\sum_{l\in A^{\circ}_{j}}\operatorname{\mathbb{P}}\Big{(}|\eta^{\prime}_{l}|>0.5\tau_{j}\Big{)}\leq\kappa^{\circ}_{\max}\sqrt{\frac{2}{\pi}}\exp\Big{(}-\frac{n\Omega_{jj}\tau^{2}_{j}}{8c^{2}_{2}}\Big{)}.\end{split}$ Under conditions (1) and (2) for $(\tau_{j},\gamma_{j})$ in Theorem 4, $\begin{split}&\operatorname{\mathbb{P}}\Big{(}\Big{\\{}l:\widetilde{V}^{[t_{\max}]}_{lj}\neq 0\Big{\\}}=\Big{\\{}l:V_{lj}^{\circ}\neq 0\Big{\\}};\ 1\leq j\leq p\Big{)}\\\ &\geq 1-\operatorname{\mathbb{P}}\Big{(}\bigcup_{j=1}^{p}\mathscr{E}_{j}^{c}\Big{)}\\\ &\geq 1-p\sqrt{\frac{2}{\pi}}\Big{(}e^{-4(\log(q)+\log(n))+\log(q)}+e^{-144c_{1}^{-2}c_{2}^{2}(\log(q)+\log(n))+\log(\kappa^{\circ}_{j})}\Big{)}\geq 1-\sqrt{\frac{2}{\pi}}pq^{-3}n^{-4}.\end{split}$ By Borel-Cantelli lemma, $\\{l:\widetilde{V}^{[t_{\max}]}_{lj}\neq 0\\}=\\{l:V_{lj}^{\circ}\neq 0\\}$; $j=1,\ldots,p$ almost surely as $n\to\infty$. So far, we have $\widetilde{\bm{V}}_{\cdot j}=\widetilde{\bm{V}}^{[t_{\max}]}_{\cdot j}=\widehat{\bm{V}}^{\circ}_{\cdot j}$; $j=1,\ldots,p$, almost surely. It remains to show that $\widehat{\bm{V}}^{\circ}_{\cdot j}$ is a global minimizer of (7); $j=1,\ldots,p$, with a high probability. Note that Assumptions 2 and 4 imply the degree of separation condition (Shen et al., 2013). By Theorem 2 of Shen et al. (2013), $\operatorname{\mathbb{P}}\Big{(}\widehat{\bm{V}}_{\cdot j}^{\circ}\text{ is not a global minimizer of }\eqref{pseudo-likelihood};1\leq j\leq p\Big{)}\leq 3p\exp\Big{(}-2(\log(q)+\log(n))\Big{)}.$ implying that $\widehat{\bm{V}}_{\cdot j}=\widetilde{\bm{V}}^{[t_{\max}]}_{\cdot j}=\widehat{\bm{V}}^{\circ}_{\cdot j}$ is a global minimizer of (7) almost surely, as $n\to\infty$. This completes the proof. #### Proof of (B) Assume $\widehat{\bm{V}}$ satisfies the properties in (A). Then Propositions 2 and 3 holds true for $\widehat{\bm{V}}$. Clearly, we have $\widehat{\mathcal{C}}=\mathcal{C}^{\circ}$. Thus, we need to show $\widehat{\mathcal{A}}=\mathcal{A}^{\circ}$ and $\widehat{\mathcal{D}}=\mathcal{D}^{\circ}$. We shall prove this by induction. For $h\geq 0$, the set of instruments on leaves in the graph $G^{[h]}$ (with primary variables $\bm{Y}_{\\{j:h_{G}(j)\geq h\\}}$ and intervention variables $\bm{X}$) is $\mathscr{I}^{[h]}=\Big{\\{}l:l=\mathop{\rm arg\,min\,}_{l^{\prime}:\|\widehat{\bm{V}}^{[h]}_{l^{\prime}\cdot}\|_{0}\neq 0}\|\widehat{\bm{V}}^{[h]}_{l^{\prime}\cdot}\|_{0}\Big{\\}}=\Big{\\{}l:\|\bm{V}^{[h]}_{l\cdot}\|_{0}=1\Big{\\}}.$ Hence, $X_{l}$ is an instrument of leaf variable $Y_{j}$ in $G^{[h]}$, when $l\in\mathscr{I}^{[h]}$ and $j=\mathop{\rm arg\,max\,}_{j^{\prime}}|V_{lj^{\prime}}|$. Moreover, $h_{G}(j)=h$. By Proposition 3 it also recovers the edges $Y_{k}\to Y_{j}$ such that $h_{G}(k)=h_{G}(j)+1$. In model (1), after removing $\\{Y_{j}:h_{G}(j)\leq h\\}$, the local Markov property among $(\bm{Y}_{\\{j:h_{G}(j)\geq h+1\\}},\bm{X})$ remain intact provided that $\\{j:h_{G}(j)\geq h+1\\}\neq\emptyset$. Then we increment $h$ and repeat the procedure until all primary variables are removed, namely, $\\{j:h_{G}(j)\geq h+1\\}=\emptyset$. As a result, Algorithm 2 correctly identifies the topological height $h_{G}(j)$ for each primary variable $Y_{j}$ and the edges $Y_{k}\to Y_{j}$ such that $h_{G}(k)=h_{G}(j)+1$. Consequently, the ancestral relations can be reconstructed by $\mathcal{A}=\Big{\\{}(k_{0},k_{d}):Y_{k_{0}}\to\cdots\to Y_{k_{d}},\ h_{G}(k_{\nu})=h_{G}(k_{\nu+1})+1,\ \nu=0,\ldots,d-1\Big{\\}}.$ Therefore, Step 6 of Algorithm 2 recovers $\widehat{\mathcal{A}}=\mathcal{A}^{\circ}$ and $\widehat{\mathcal{D}}=\mathcal{D}^{\circ}$ almost surely as $n\to\infty$. This completes the proof of (B). ###### Lemma 10 Let $T_{j}=\frac{\bm{\mathsf{e}}_{j}^{\top}(\bm{\mathsf{P}}_{A^{\circ}_{j}}-\bm{\mathsf{P}}_{B^{\circ}_{j}})\bm{\mathsf{e}}_{j}}{\bm{\mathsf{e}}_{j}^{\top}(\bm{I}-\bm{\mathsf{P}}_{A^{\circ}_{j}})\bm{\mathsf{e}}_{j}/(n-|A^{\circ}_{j}|)},\quad T_{j}^{*}=\frac{(\bm{\mathsf{e}}^{*}_{j})^{\top}(\bm{\mathsf{P}}_{A^{\circ}_{j}}-\bm{\mathsf{P}}_{B^{\circ}_{j}})\bm{\mathsf{e}}^{*}_{j}}{(\bm{\mathsf{e}}^{*}_{j})^{\top}(\bm{I}-\bm{\mathsf{P}}_{A^{\circ}_{j}})\bm{\mathsf{e}}^{*}_{j}/(n-|A^{\circ}_{j}|)},$ (32) where $A^{\circ}_{j}=\textnormal{{an}}_{S^{\circ}}(j)\cup\textnormal{{in}}_{S^{\circ}}(j)\cup\textnormal{{d}}_{S^{\circ}}(j)$, $B^{\circ}_{j}=(\textnormal{{an}}_{S^{\circ}}(j)\cup\textnormal{{in}}_{S^{\circ}}(j))\setminus\textnormal{{d}}_{S^{\circ}}(j)$. Then $(T_{1},\ldots,T_{p})$ are independent and $(T_{1}^{*},\ldots,T_{p}^{*})$ are conditionally independent given $\bm{\mathsf{Z}}$. Moreover, $T^{*}_{j}/(|A^{\circ}_{j}|-|B^{\circ}_{j}|)\mid\bm{\mathsf{Z}}\overset{d}{=}T_{j}/(|A^{\circ}_{j}|-|B^{\circ}_{j}|)\sim F_{|A^{\circ}_{j}|-|B^{\circ}_{j}|,n-|A^{\circ}_{j}|};\quad 1\leq j\leq p,$ where $F_{d_{1},d_{2}}$ denotes the F-distribution with $d_{1}$ and $d_{2}$ degrees of freedom. ### Proof of Lemma 10 Let $\bm{\mathsf{Z}}=(\bm{\mathsf{Y}},\bm{\mathsf{X}})$ as in (12). Given data submatrix $\bm{\mathsf{Z}}_{A^{\circ}_{j}}$, $\frac{\bm{\mathsf{e}}_{j}^{\top}(\bm{\mathsf{P}}_{A^{\circ}_{j}}-\bm{\mathsf{P}}_{B^{\circ}_{j}})\bm{\mathsf{e}}_{j}}{\sigma^{2}_{j}}\mid\bm{\mathsf{Z}}_{A^{\circ}_{j}}\sim\chi^{2}_{|A^{\circ}_{j}|-|B^{\circ}_{j}|},\quad\frac{\bm{\mathsf{e}}_{j}^{\top}(\bm{I}-\bm{\mathsf{P}}_{A^{\circ}_{j}})\bm{\mathsf{e}}_{j}}{\sigma^{2}_{j}}\mid\bm{\mathsf{Z}}_{A^{\circ}_{j}}\sim\chi^{2}_{n-|A^{\circ}_{j}|},$ and they are independent, because $\bm{\mathsf{P}}_{A^{\circ}_{j}}-\bm{\mathsf{P}}_{B^{\circ}_{j}}$ and $\bm{I}-\bm{\mathsf{P}}_{A^{\circ}_{j}}$ are orthonormal projection matrices, $\bm{\mathsf{e}}_{j}\mid\bm{\mathsf{Z}}_{A^{\circ}_{j}}\sim N(\bm{0},\sigma_{j}^{2}\bm{I}_{n})$, and $(\bm{\mathsf{P}}_{A^{\circ}_{j}}-\bm{\mathsf{P}}_{B^{\circ}_{j}})(\bm{I}-\bm{\mathsf{P}}_{A^{\circ}_{j}})=\bm{0}$. Then, for any $t\in\mathbb{R}$, the characteristic function $\operatorname{\mathbb{E}}\exp(\iota tT_{j}/(|A^{0}_{j}|-|B^{0}_{j}|))$ can be written as ($\iota$ is the imaginary unit) $\operatorname{\mathbb{E}}\left(\operatorname{\mathbb{E}}\left(\exp(\iota tT_{j}/(|A^{\circ}_{j}|-|B^{\circ}_{j}|))\mid\bm{\mathsf{Z}}_{A^{\circ}_{j}}\right)\right)=\operatorname{\mathbb{E}}\psi_{|A^{\circ}_{j}|-|B^{\circ}_{j}|,n-|A^{\circ}_{j}|}(t)=\psi_{|A^{\circ}_{j}|-|B^{\circ}_{j}|,n-|A^{\circ}_{j}|}(t),$ where $\psi_{|A^{\circ}_{j}|-|B^{\circ}_{j}|,n-|A^{\circ}_{j}|}$ is the characteristic function of F-distribution with degrees of freedom $|A^{\circ}_{j}|-|B^{\circ}_{j}|,n-|A^{\circ}_{j}|$. Hence, $T_{j}/(|A^{\circ}_{j}|-|B^{\circ}_{j}|)\sim F_{|A^{\circ}_{j}|-|B^{\circ}_{j}|,n-|A^{\circ}_{j}|}$. Similarly, $T_{j}^{*}/(|A^{\circ}_{j}|-|B^{\circ}_{j}|)\sim F_{|A^{\circ}_{j}|-|B^{\circ}_{j}|,n-|A^{\circ}_{j}|}$; $j=1,\ldots,p$. Next, we prove independence for $\bm{T}$ and $\bm{T}^{*}$ via a peeling argument. Let $\bm{t}=(t_{1},\ldots,t_{p})$. Let $j$ be a leaf node of the graph $G^{\circ}$. Then $\bm{T}_{-j}\mid\bm{\mathsf{Y}}_{-j},\bm{\mathsf{X}}$ is deterministic, where $\bm{T}_{-j}$ is the subvector of $\bm{T}$ with the $j$th component removed. The characteristic function of $\bm{t}^{\top}\bm{T}$ is $\begin{split}\operatorname{\mathbb{E}}\exp(\iota\bm{t}^{\top}\bm{T})&=\operatorname{\mathbb{E}}\Big{(}\operatorname{\mathbb{E}}\Big{(}\exp(\iota t_{j}T_{j})\mid\bm{\mathsf{Y}}_{-j},\bm{\mathsf{X}}\Big{)}\exp(\iota\bm{t}^{\top}_{-j}\bm{T}_{-j})\Big{)}\\\ &=\psi_{|A^{\circ}_{j}|-|B^{\circ}_{j}|,n-|A^{\circ}_{j}|}(t_{j})\operatorname{\mathbb{E}}\exp(\iota\bm{t}_{-j}^{\top}\bm{T}_{-j}),\end{split}$ (33) where $\psi_{|A^{\circ}_{j}|-|B^{\circ}_{j}|,n-|A^{\circ}_{j}|}$ is the characteristic function of the F-distribution with degrees of freedom $(|A^{\circ}_{j}|-|B^{\circ}_{j}|,n-|A^{\circ}_{j}|)$. Next, let $j^{\prime}$ be a leaf node of the graph $G^{\prime}$ and apply the law of iterated expectation again, where $G^{\prime}$ is the subgraph of $G^{\circ}$ without node $j$. Repeat this procedure and after $p$ steps $\operatorname{\mathbb{E}}\exp(\iota\bm{t}^{\top}\bm{T})=\prod_{j=1}^{p}\psi_{|A^{\circ}_{j}|-|B^{\circ}_{j}|,n-|A^{\circ}_{j}|}(t_{j})$, which implies $(T_{1},\ldots,T_{p})$ are independent. Similarly, $\bm{T}^{*}$ also has independent components given $\bm{\mathsf{Z}}$ and has the same distribution as $\bm{T}$. This completes the proof. ### B.5 Proof of Theorem 5 #### Proof of (A) Without loss of generality, assume $M$ is sufficiently large. For any $x\in\mathbb{R}$, $\begin{split}&|\operatorname{\mathbb{P}}(\textnormal{Lr}\leq x)-\operatorname{\mathbb{P}}(\textnormal{Lr}(S^{\circ},\widehat{\bm{\Sigma}})\leq x)|\\\ &\leq\operatorname{\mathbb{E}}|\operatorname{I}(\textnormal{Lr}\leq x)-\operatorname{I}(\textnormal{Lr}(S^{\circ},\widehat{\bm{\Sigma}})\leq x)|\\\ &=\operatorname{\mathbb{E}}|\operatorname{I}(\textnormal{Lr}\leq x,\widehat{S}\neq S^{\circ})+\operatorname{I}(\textnormal{Lr}(S^{\circ},\widehat{\bm{\Sigma}})\leq x)(\operatorname{I}(\widehat{S}=S^{\circ})-1)|\\\ &\leq 2\operatorname{\mathbb{P}}(\widehat{S}\neq S^{\circ}).\end{split}$ (34) From (13), $\textnormal{Lr}^{*}(S^{\circ},\widehat{\bm{\Sigma}}^{*})=\sum_{j:\textnormal{{d}}_{S^{\circ}}(j)\neq\emptyset}T_{j}^{*}$. By Lemma 10, $\operatorname{\mathbb{P}}(\textnormal{Lr}(S^{\circ},\widehat{\bm{\Sigma}})\leq x)=\operatorname{\mathbb{P}}(\textnormal{Lr}^{*}(S^{\circ},\widehat{\bm{\Sigma}}^{*})\leq x\mid\bm{\mathsf{Z}})$. Note that $\textnormal{Lr}^{*}=\textnormal{Lr}^{*}(\widehat{S}^{*},\widehat{\bm{\Sigma}}^{*})$. Then for any $x\in\mathbb{R}$, $\begin{split}&|\operatorname{\mathbb{P}}(\textnormal{Lr}^{*}\leq x\mid\bm{\mathsf{Z}})-\operatorname{\mathbb{P}}(\textnormal{Lr}(S^{\circ},\widehat{\bm{\Sigma}})\leq x)|\\\ &=|\operatorname{\mathbb{P}}(\textnormal{Lr}^{*}\leq x,\widehat{S}^{*}\neq S^{\circ}\mid\bm{\mathsf{Z}})+\operatorname{\mathbb{P}}(\textnormal{Lr}^{*}\leq x,\widehat{S}^{*}=S^{\circ}\mid\bm{\mathsf{Z}})-\operatorname{\mathbb{P}}(\textnormal{Lr}^{*}(S^{\circ},\widehat{\bm{\Sigma}}^{*})\leq x\mid\bm{\mathsf{Z}})|\\\ &\leq|\operatorname{\mathbb{P}}(\textnormal{Lr}^{*}\leq x,\widehat{S}^{*}\neq S^{\circ}\mid\bm{\mathsf{Z}})|+|\operatorname{\mathbb{P}}(\textnormal{Lr}^{*}(S^{\circ},\widehat{\bm{\Sigma}}^{*})\leq x,\widehat{S}^{*}\neq S^{\circ}\mid\bm{\mathsf{Z}})|\\\ &\leq 2\operatorname{\mathbb{P}}(\widehat{S}^{*}\neq S^{\circ}\mid\bm{\mathsf{Z}}).\end{split}$ (35) Since (34) and (35) hold uniformly in $x$, we have $\sup_{x\in\mathbb{R}}|\operatorname{\mathbb{P}}(\textnormal{Lr}\leq x)-\operatorname{\mathbb{P}}(\textnormal{Lr}^{*}\leq x\mid\bm{\mathsf{Z}})|\leq 2\operatorname{\mathbb{P}}(\widehat{S}\neq S^{\circ})+2\operatorname{\mathbb{P}}(\widehat{S}^{*}\neq S^{\circ}\mid\bm{\mathsf{Z}}).$ By Theorem 4, we have $\operatorname{\mathbb{P}}(\widehat{S}\neq S^{\circ})\to 0$, which holds uniformly for all $\bm{\theta}$ which satisfy the Assumptions 1C-5. For $\operatorname{\mathbb{P}}(\widehat{S}^{*}\neq S^{\circ}\mid\bm{\mathsf{Z}})$, the error terms of (7) in the perturbed data $\bm{\mathsf{Z}}^{*}$ are rescaled with $\Omega_{jj}^{-1}+\widehat{\sigma}^{2}_{j}\leq 2\Omega_{jj}^{-1}$. By Theorem 4, $\operatorname{\mathbb{P}}(\widehat{S}^{*}\neq S^{\circ})=\operatorname{\mathbb{E}}\operatorname{\mathbb{P}}(\widehat{S}^{*}\neq S^{\circ}\mid\bm{\mathsf{Z}})\to 0$, which implies $\operatorname{\mathbb{P}}(\widehat{S}^{*}\neq S^{\circ}\mid\bm{\mathsf{Z}})\overset{p}{\longrightarrow}0$ as $n\rightarrow\infty$ by the Markov inequality. Consequently, $\sup_{x\in\mathbb{R}}|\operatorname{\mathbb{P}}(\textnormal{Lr}\leq x)-\operatorname{\mathbb{P}}(\textnormal{Lr}^{*}\leq x\mid\bm{\mathsf{Z}})|\to 0$. For $|\mathcal{D}|=0$, $\operatorname{\mathbb{P}}(\textnormal{Lr}=0)\to 1$, $\operatorname{\mathbb{P}}(\textnormal{Lr}^{*}=0\mid\bm{\mathsf{Z}})\to 1$, and $\operatorname{\mathbb{P}}(\textnormal{Pval}=1)\to 1$. For $|\mathcal{D}|>0$, $P(\textnormal{Lr}^{*}\geq\textnormal{Lr}\mid\bm{\mathsf{Z}})\to\textnormal{Unif}(0,1)$ and $\operatorname{\mathbb{P}}(\textnormal{Pval}<\alpha)\to\alpha$. This completes the proof of (A). #### Proof of (B) Let $\textnormal{Pval}_{k}=M^{-1}\sum_{m=1}^{M}\operatorname{I}(\text{Lr}^{*}_{k,m}\geq\text{Lr})$, the p-value of sub-hypothesis $\text{H}_{0,k}$. For $|\mathcal{D}|<|\mathcal{H}|$, there exists an edge $(i_{k},j_{k})\in\mathcal{H}$ but $(i_{k},j_{k})\notin\mathcal{D}$. Then by (A), $\operatorname{\mathbb{P}}(\textnormal{Pval}=\textnormal{Pval}_{k}=1)\to 1$. For $|\mathcal{D}|=|\mathcal{H}|$, note that as $n,M\to\infty$, $\operatorname{\mathbb{P}}\Big{(}\textnormal{Pval}<\alpha\Big{)}=\operatorname{\mathbb{P}}\Big{(}\textnormal{Pval}_{1}<\alpha,\ldots,\textnormal{Pval}_{|\mathcal{H}|}<\alpha\Big{)}\leq\operatorname{\mathbb{P}}\Big{(}\textnormal{Pval}_{1}<\alpha\Big{)}\to\alpha.$ Now, define a sequence $\bm{U}_{\mathcal{H}}^{(r)}$; $r=1,2,\ldots$ such that $\bm{U}_{(i_{1},j_{1})}^{(r)}=0$ and $\min_{k=2}^{|\mathcal{H}|}|\bm{U}_{(i_{k},j_{k})}^{(r)}|\geq c>0$. Thus, $\bm{U}_{\mathcal{H}}^{(r)}$; $r=1,2,\ldots$ satisfies $H_{0}$. By Proposition 7, $\text{Pval}_{k}\overset{p}{\longrightarrow}0$ for $k\geq 2$ as $r\to\infty$ . Hence, $\limsup_{\begin{subarray}{c}n\to\infty\\\ \bm{U}_{\mathcal{H}}^{(r)}\text{ satisfies H}_{0}\end{subarray}}\operatorname{\mathbb{P}}_{\bm{\theta}^{(r)}}\Big{(}\text{Pval}<\alpha\Big{)}\geq\sup_{r}\lim_{n,p,q\to\infty}\operatorname{\mathbb{P}}_{\bm{\theta}^{(r)}}\Big{(}\text{Pval}<\alpha\Big{)}=\alpha.$ This completes the proof. ### B.6 Proof of Proposition 6 Since $\operatorname{\mathbb{P}}(\widehat{S}=S^{\circ})\to 1$, it suffices to consider $\text{Lr}(S^{\circ},\widehat{\bm{\Sigma}})$. For $|\mathcal{D}|=0$, we have $\operatorname{\mathbb{P}}(\text{Lr}=0)\to 1$. Now, assume $|\mathcal{D}|>0$. Then $\begin{split}2\text{Lr}(S^{\circ},\widehat{\bm{\Sigma}})=\underbrace{\sum_{j:\textnormal{{d}}_{S^{\circ}}(j)\neq\emptyset}\frac{\bm{\mathsf{e}}_{j}^{\top}(\bm{\mathsf{P}}_{A^{\circ}_{j}}-\bm{\mathsf{P}}_{B^{\circ}_{j}})\bm{\mathsf{e}}_{j}}{\sigma^{2}_{j}}}_{R_{1}}+\underbrace{\sum_{j:\textnormal{{d}}_{S^{\circ}}(j)\neq\emptyset}\left(\frac{\sigma^{2}_{j}}{\widehat{\sigma}^{2}_{j}}-1\right)\frac{\bm{\mathsf{e}}_{j}^{\top}(\bm{\mathsf{P}}_{A^{\circ}_{j}}-\bm{\mathsf{P}}_{B^{\circ}_{j}})\bm{\mathsf{e}}_{j}}{\sigma_{j}^{2}}}_{R_{2}}.\end{split}$ To derive the asymptotic distribution of $R_{1}$, we apply the peeling strategy with law of iterated expectation as in the proof of Lemma 10. Then we have $\bm{\mathsf{e}}_{j}^{\top}(\bm{\mathsf{P}}_{A^{0}_{j}}-\bm{\mathsf{P}}_{B^{0}_{j}})\bm{\mathsf{e}}_{j}/\sigma^{2}_{j}$; $\textnormal{{d}}_{S^{\circ}}(j)\neq\emptyset$ are independent. Therefore, $R_{1}\sim\chi^{2}_{|\mathcal{D}|}$. To bound $R_{2}$, we apply Lemma 1 of Laurent and Massart (2000). By Assumption 5, $\begin{split}\operatorname{\mathbb{P}}\left(\max_{j:\textnormal{{d}}_{S^{\circ}}(j)\neq\emptyset}\left|\frac{\widehat{\sigma}^{2}_{j}}{\sigma^{2}_{j}}-1\right|\geq 4\sqrt{\frac{\log|\mathcal{D}|}{(1-\rho)n}}+8\frac{\log|\mathcal{D}|}{(1-\rho)n}\right)&\leq 2\exp(-\log|\mathcal{D}|).\end{split}$ Hence, $\max_{j:\textnormal{{d}}_{S^{\circ}}(j)\neq\emptyset}\left|\frac{\sigma^{2}_{j}}{\widehat{\sigma}_{j}^{2}}-1\right|\leq 8\sqrt{\frac{\log|\mathcal{D}|}{(1-\rho)n}}$ with probability tending one as $n,p\rightarrow\infty$. Thus $|R_{2}|\leq|R_{1}|\max_{j:\textnormal{{d}}_{S^{\circ}}(j)\neq\emptyset}\left|\frac{\sigma^{2}_{j}}{\widehat{\sigma}_{j}^{2}}-1\right|\leq O_{\operatorname{\mathbb{P}}}\left(|\mathcal{D}|\sqrt{\frac{\log|\mathcal{D}|}{n}}\right).$ Consequently, as $n,p,q\rightarrow\infty$, if $|\mathcal{D}|>0$ is fixed, then $2\text{Lr}\overset{d}{\longrightarrow}\chi^{2}_{|\mathcal{D}|}$; if $|\mathcal{D}|\to\infty$, then $(2|\mathcal{D}|)^{-1/2}(2\text{Lr}-|\mathcal{D}|)\overset{d}{\longrightarrow}N(0,1)$. This completes the proof. ### B.7 Proof of Proposition 7 Let $\bm{\theta}^{(n)}=(\bm{U}^{\circ}+\bm{\Delta},\bm{W}^{\circ})$. Then $L(\bm{\theta}^{(n)},\bm{\Sigma}^{\circ})-L(\bm{\theta}^{\circ},\bm{\Sigma}^{\circ})=\sum_{j:\textnormal{{d}}^{\circ}(j)\neq\emptyset}\Big{(}\sqrt{n}\bm{\eta}_{j}^{\top}\bm{\Delta}_{\cdot j}-\frac{1}{2}\sqrt{n}\bm{\Delta}_{\cdot j}\Big{(}n^{-1}\bm{\mathsf{Y}}^{\top}_{\textnormal{{d}}^{\circ}(j)}\bm{\mathsf{Y}}_{\textnormal{{d}}^{\circ}(j)}\Big{)}\sqrt{n}\bm{\Delta}_{\cdot j}\Big{)},\\\ $ where $\bm{\eta}_{j}=(n^{-1/2}\bm{\mathsf{Y}}_{\textnormal{{d}}^{\circ}(j)}^{\top}\bm{\mathsf{e}}_{j},\bm{0}_{|\textnormal{{d}}^{\circ}(j)^{c}|})\in\mathbb{R}^{p}$. For the likelihood ratio, it suffices to consider $\textnormal{Lr}(S^{\circ},\widehat{\bm{\Sigma}})$, since $P(\widehat{S}=S^{\circ})\to 1$. Under $H_{0}$, we have $2\text{Lr}=\sum_{j:\textnormal{{d}}^{\circ}(j)\neq\emptyset}T_{j}$. Note that $T_{j}=\sum_{r=1}^{|\textnormal{{d}}^{\circ}(j)|}(\bm{\mathsf{q}}_{j,r}^{\top}\bm{\mathsf{e}}_{j})^{2}+o_{\operatorname{\mathbb{P}}}(1)$ by Proposition 6, where $\bm{\mathsf{P}}_{A^{\circ}_{j}}-\bm{\mathsf{P}}_{B^{\circ}_{j}}=\sum_{r=1}^{|\textnormal{{d}}^{\circ}(j)|}\bm{\mathsf{q}}_{j,r}\bm{\mathsf{q}}_{j,r}^{\top}$. Let $\bm{\mathsf{Q}}_{j}=(\bm{\mathsf{q}}_{j,1},\ldots,\bm{\mathsf{q}}_{j,|\textnormal{{d}}^{\circ}(j)|})\in\mathbb{R}^{n\times|\textnormal{{d}}^{\circ}(j)|}$. Then $\begin{pmatrix}\bm{\mathsf{Q}}_{j}^{\top}\bm{\mathsf{e}}_{j}\\\ \bm{\eta}_{j}\end{pmatrix}\ \vline\ \bm{\mathsf{Y}}_{\textnormal{{an}}(j)},\bm{\mathsf{X}}\sim N\left(\bm{0},\begin{pmatrix}\bm{I}_{|\textnormal{{d}}^{\circ}(j)|}&n^{-1/2}\bm{\mathsf{Q}}_{j}^{\top}\bm{\mathsf{Y}}_{\textnormal{{d}}^{\circ}(j)}\\\ n^{-1/2}\bm{\mathsf{Y}}_{\textnormal{{d}}^{\circ}(j)}^{\top}\bm{\mathsf{Q}}_{j}&n^{-1}\bm{\mathsf{Y}}^{\top}_{\textnormal{{d}}^{\circ}(j)}\bm{\mathsf{Y}}_{\textnormal{{d}}^{\circ}(j)}\end{pmatrix}\right),$ $\bm{\mathsf{Q}}^{\top}_{j}\bm{\mathsf{e}}_{j}\mid\bm{\eta}_{j},\bm{\mathsf{Y}}_{\textnormal{{an}}(j)},\bm{\mathsf{X}}\sim N\Big{(}n^{-1/2}\bm{\mathsf{Y}}_{\textnormal{{d}}^{\circ}(j)}^{\top}\bm{\mathsf{Q}}_{j}\bm{\eta}_{j},\bm{\mathsf{Y}}_{\textnormal{{d}}^{\circ}(j)}^{\top}(\bm{I}_{n}-\bm{\mathsf{Q}}_{j}\bm{\mathsf{Q}}_{j}^{\top})\bm{\mathsf{Y}}_{\textnormal{{d}}^{\circ}(j)}\Big{)}.$ Next, let $k$ be a leaf node of the graph $G^{\circ}$. For fixed $|\mathcal{D}|>0$, after change of measure, $\begin{split}&\beta(\bm{\theta}^{\circ},\bm{\Delta})\\\ &\geq\liminf_{n\to\infty}\operatorname{\mathbb{E}}_{\bm{\theta}^{\circ}}\left(\operatorname{I}\left(\sum_{j=1}^{p}\|\bm{\mathsf{Q}}^{\top}\bm{\mathsf{e}}_{j}\|^{2}_{2}>\chi^{2}_{|\mathcal{D}|,1-\alpha}\right)\exp(L(\bm{\theta}^{(n)},\bm{\Sigma}^{\circ})-L(\bm{\theta}^{\circ},\bm{\Sigma}^{\circ}))\right)\\\
# Deployment of Energy-Efficient Deep Learning Models on Cortex-M based Microcontrollers using Deep Compression Mark Deutel1, Philipp Woller2, Christopher Mutschler2, and Jürgen Teich4 1 Mark Deutel is with Fraunhofer IIS, Fraunhofer Institute for Integrated Circuits IIS, Nuremberg, Germany and Friedrich-Alexander-Universität Erlangen- Nürnberg (FAU), Erlangen, Germany<EMAIL_ADDRESS>Phillip Woller and Christopher Mutschler are with the Fraunhofer IIS, Fraunhofer Institute for Integrated Circuits IIS, Nuremberg, Germany <EMAIL_ADDRESS>Jürgen Teich is with Friedrich- Alexander-Universität Erlangen-Nürnberg (FAU), Erlangen, Germany <EMAIL_ADDRESS> ###### Abstract Large Deep Neural Networks (DNNs) are the backbone of today’s artificial intelligence due to their ability to make accurate predictions when being trained on huge datasets. With advancing technologies, such as the Internet of Things, interpreting large quantities of data generated by sensors is becoming an increasingly important task. However, in many applications not only the predictive performance but also the energy consumption of deep learning models is of major interest. This paper investigates the efficient deployment of deep learning models on resource-constrained microcontroller architectures via network compression. We present a methodology for the systematic exploration of different DNN pruning, quantization, and deployment strategies, targeting different ARM Cortex-M based low-power systems. The exploration allows to analyze trade-offs between key metrics such as accuracy, memory consumption, execution time, and power consumption. We discuss experimental results on three different DNN architectures and show that we can compress them to below 10% of their original parameter count before their predictive quality decreases. This also allows us to deploy and evaluate them on Cortex-M based microcontrollers. ###### Index Terms: Deep Neural Networks, Deep Compression, Network Pruning, Weight Quantization, Microcontrollers ## I Introduction Deep Neural Networks (DNNs) became predominant in many applications that require autonomous decision-making based on environmental information, including audio recognition [1, 2], image classification [3, 4], or human activity monitoring [5]. DNNs are beneficial as they are easy to set up and as they can be trained to detect correlations even when they are confronted with high-dimensional data. However, the execution of DNNs is energy-, resource-, and time-expensive [6, 7]. In situations where the trade-off between resource constraints, execution time, and predictive quality is key, DNNs often struggle to compete with classical machine learning approaches [8]. However, with trends like smart devices and the internet of things (IoT), the demand and interest in deploying DNNs on microcontrollers grows. Figure 1: Overview of our methodology. Deep compression is a relatively young research area that deals with the compression of DNNs. Prominent techniques include DNN graph pruning [9], weight quantization [10, 11], and subspace methods [12]. Their goal is to reduce the resource footprint of a DNN by reducing the number of trainable weights and computational complexity while preserving the original predictive performance. Based on these principles different DNN compression pipelines have been proposed. Most noticeably, Han et al. [13] who proposed a pipeline combining network pruning, integer quantization, and Huffman encoding. Others focus on quantization during network training [10] or on structure-based pruning. This allows for an immediate removal of pruned weights [14, 15]. However, such well-established frameworks only trade compression over predictive accuracy but do not explicitly target energy-efficiency and architecture-specific constraints like memory availability and processing speed that play an important role in many embedded applications. This paper proposes a methodology to systematically train and deploy DNN architectures on Cortex-M-based microcontrollers. We introduce an automated pipeline that covers application-specific DNN training and compression, and that combines it with a target architecture-specific deployment, see Fig. 1. Our proposed pipeline is composed of two major building blocks. First, from an application-specific viewpoint we systematically explore compression techniques, i.e., network pruning and weight quantization, and configurations during the training of DNNs. Second, incorporating an architecture-specific view, we realize the mapping from a high-level graph-based DNN representation to low level code. This step involves an offline code generator and a runtime library. The former takes care of the data layout, plans memory allocation and emits executable code while the latter provides implementations of common DNN layers used by the generated code. Novel methods implemented in our proposed pipeline include ahead-of-time code generation and memory allocation scheduling, which eliminates the need for any form of network interpretation or dynamic memory management at runtime, and the exploitation of sparse matrices generated as part of our pruning techniques using the _Compressed Column Storage (CCS)_ format [16]. In our experiments we evaluate both compression and deployment for three common DNN architectures, i.e., AlexNet [17], ResNet [3] and LeNet [18]. Our objective is a thorough evaluation of the relation between compressed DNNs and their predictive quality. In contrast to previous work our results do not only focus on deployment or compression alone, but provide detailed insight into the relationship between different pruning, quantization, and deployment strategies when applied in combination. Furthermore, we deployed the compressed models on three target systems and discuss their memory consumption, execution time, and power consumption. The rest of this paper is structured as follows. Sec. II discusses related work. Secs. III and IV provide details about our compression and deployment pipeline. Sec. V discusses our experimental results. Sec. VII concludes. ## II Related Work Existing research predominately compresses DNNs via network pruning and weight quantization. These techniques are well understood as research has been conducted exploring the effects of pruning and quantization on a network’s predictive performance [19, 20]. However, when deploying applications for embedded targets, they are defined by the constraints imposed by the platforms they use. As a result, the suitability of DNN models for deployment on microcontrollers is not only determined by their accuracy but also by their memory footprint and inference time. Therefore, this work extends existing findings by analysing the effects of DNN compression not only on accuracy but also on relevant deployment metrics, i.e. memory consumption, latency, and energy consumption. Furthermore, research that focuses on the deployment of DNNs to microcontrollers is often published in an application-oriented way, e.g., to realize real-time drowsiness detection [21] or to perform motor fault diagnosis [22]. Those platforms do not serve as general purpose frameworks as they are tightly optimized to the particular application requirement and as they do not generalize to a broader set of target architectures. Nevertheless, recent scientific work has provided some insight into generalized deployment of DNNs on microcontrollers. The approach most closely related to our proposed methodology is MCUNet [23]. Similar to our pipeline the authors describe a two stage process to seamlessly combine model design (TinyNAS) with an inference engine (TinyEngine). Still MCUNet differs from our approach in the way it generates suitable DNN candidates for deployment. To find networks that meet target platform constraints, MCUNet focuses on neural architecture search (NAS) [24] while our framework starts from well-known existing DNN architectures and then dynamically scales them down during their training using pruning and quantization techniques. A more general approach to a deployment framework for microcontrollers is _tfl-micro_ [25], which supports the execution of quantized tensorflow lite models on ARM Cortex-M-based hardware using ARM’s _CMSIS_ library. However, this also limits the framework as it requires the usage of _tensorflow (TF)_ for model training and also only supports a subset of features implemented in TF. Besides that there are also commercial frameworks focusing on embedded platforms. Noticeable examples are STM’s _X-CUBE-AI_ 111https://www.st.com/en/embedded-software/x-cube-ai.html, that allows for the automatic conversion of pre-trained AI algorithms to deployable code as well as web-service based end-to-end solutions like _Edge Impulse_ 222https://www.edgeimpulse.com/ or _SensiML_ 333https://sensiml.com/. However, such commercial frameworks are often either black boxes (e.g., X-CUBE-AI) or they base themselves on already existing underlying solutions (e.g., _Edge Impulse_ uses _tfl-micro_) and their limitations. ## III Compression and Deployment Pipeline Our pipeline is fully integrated and seamlessly covers the complete DNN training and deployment process. Our methodology uses both network pruning (Sec. III-A) and weight quantization (Sec. III-B) which can both be controlled via a set of additional _hyperparameters_. Furthermore, the trained and compressed DNNs are directly converted from their graph-based representation to executable architecture-specific program code (see Sec. IV). As a result, our pipeline can easily be integrated with existing meta-heuristic optimization frameworks (e.g. Optuna [26]) to conduct automated design space exploration. ### III-A Network Pruning Our pipeline implements configurable elements for network pruning, i.e., (1) pruning techniques, (2) pruning heuristics, and (3) pruning schedule. which we describe in the following. Pruning techniques. Pruning DNNs by removing parameters has been proposed early [27, 9]. While initially being introduced to improve generalization and convergence it recently became a standard size reduction technique with no or low cost of accuracy. Our pipeline implements _element-wise pruning_ and _structural pruning_. Element-wise pruning removes connections from a DNN’s compute graph, i.e., parameters of the network are set to zero. Hence, these parameters do no longer influence the training error and are removed from the scope of the optimizer that trains the network. Structural pruning sets whole structures of parameters to zero. This has shown to be very efficient for pruning filters [14] or channels [15] of 2D-convolutional layers but it can analogously also be applied to rows and columns of linear layers. Its major benefit is the removal of complete structures from the weight tensors at once, which results in a considerable immediate reductions of parameters (which is in contrast to element-wise pruning that only creates sparse weight tensors). Pruning heuristics. A critical aspect of pruning is the selection of elements or structures that, when removed, have the least impact on predictive performance. Oracle pruning [28] finds an optimal selection by removing every single structure and element of a network before evaluating its impact on the prediction quality. In practical applications this approach cannot be applied as it is too resource- and time-consuming. Fortunately, there have been proposed a number of heuristics that approximate optimal element-wise or structural pruning. In our framework we implemented many popular approaches that are based on different criteria such as magnitude [29], L-norm [14], gradient [28] or percentage of zeros found in activations [30] to rank parameters or parameter structures by their approximated importance. Pruning schedules. The pruning schedule determines when, how often, and to what extent a DNN will be pruned during training. We implement two well-known approaches: _One-Shot Pruning_ [9] and _Iterative Pruning_. One-shot pruning first trains a DNN until it achieves a reasonable accuracy on the test dataset, and then prunes the DNN (optionally followed by a few epochs of re- training). Iterative pruning prunes a DNN over the course of training, which allows for an interleaved re-training. Hence, not all weights are removed at the same time but step by step over several pruning iterations (finally enforcing maximal sparsity). We implemented _Automated Gradual Pruning (AGP)_ [31], which gradually increases the number of pruned weights $s_{t}$ starting at $t_{0}$ from an initial sparsity $s_{i}$ to a final sparsity $s_{f}$ over $n$ steps: $s_{t}=s_{f}+(s_{i}-s_{f})\left(1-\frac{t-t_{0}}{n\Delta t}\right)^{3},\ t\in\left\\{t_{0},\dots,t_{0}+n\Delta t\right\\}.$ (1) ### III-B Weight Quantization Quantization reduces the numerical resolution of the parameters and their computation. This not only cuts down the memory footprint but also the computational complexity of a DNN. However, as parameter quantization comes with an additional error on the predictive performance a major challenge lies in the selection of a good trade-off between predictive quality and parameter resolution. Our framework uses an _affine mapping_ scheme that transforms an original floating-point parameter into an 8-bit unsigned integer444See also [32] and https://onnxruntime.ai/docs/how-to/quantization.html.. We apply a function $f(x)$ in combination with additional sets of trainable scale and zero point parameters: $\begin{split}f(x)&=g\left(\left\lfloor\frac{x}{s}\right\rceil+zp\right),\\\ s&=\frac{max_{data}-min_{data}}{255},\ 0\leq zp\leq 255,\end{split}$ (2) where $g(x)$ is the clamp-function to avoid data type overflows: $g(x)=\left\\{\begin{array}[]{rcl}x&\mbox{if}&0\leq x\leq 255\\\ 255&\mbox{if}&x>255\\\ 0&\mbox{if}&x<0.\end{array}\right.$ (3) The scale parameter $s$ defines the step size of the quantization, which is calculated as the ratio between the span in which values are distributed in the original floating-point space and the span covered by the quantized integer space. The zero point parameter $zp$ denotes an exact point in quantization space that represents zero in the original floating-point space. The two parameters can be defined either per tensor or per structure. Quantization can not only be applied to weight tensors but also to activation tensors. We refer to this as _full integer quantization_. During execution most computations can then be performed in integer instead of floating-point space which is beneficial for target systems that are not equipped with floating-point units. We give an example for applying full-integer quantization to matrix-multiplications. The general form is defined as: $\begin{split}c_{ij}=\sum_{k=0}^{n}a_{ik}\cdot b_{ki},\forall i\in\\{0,\dots,m\\},\forall j\in\\{0,\dots,p\\},\end{split}$ (4) where the first line describes how the elements of a matrix $C$ are calculated from the elements of a $m\times n$ matrix $A$ and a $n\times p$ matrix $B$. In a fully-quantized DNN, both matrices $A$ and $B$ contain integer values and we first must de-quantize them by rearranging Equation 2 before we multiply them. As the resulting matrix $C$ is represented in the un-quantized space, we have to quantize it by applying Eq. 2 again. By substituting and rearranging the previous computations we obtain $c_{ij}=g\left(zp_{c}+\left\lfloor{\left(\frac{s_{a}\cdot s_{b}}{s_{c}}\right)\sum_{k=0}^{n}(a_{ik}-zp_{a})(b_{ik}-zp_{b})}\right\rceil\right).$ (5) Note that only the scale parameters $\\{s_{a},s_{b},s_{c}\\}\in\mathbb{R}$ while all other parameters $\in\mathbb{N}_{0}$. Our pipeline implements two popular ways of determining at which point quantization is applied to a DNN. The first method quantizes as a post process (PPQ) [11, 33], i.e., after training has finished, and the second method integrates quantization into the training loop. The latter is denoted by _Quantization-aware Training (QAT)_ [10]. Both techniques come with their advantages and disadvantages: PPQ is extremely easy to integrate as it can be performed completely decoupled from a DNN’s training process and does not requires any invasive changes to a DNN’s architecture (i.e., no re-training to fine-tune quantization parameters). However, this usually comes at the cost of a larger error introduced by quantization as the required scale and zero point parameters are only roughly approximated. In contrast, QAT adapts quantization parameters as part of a DNN’s training process and can hence yield better results. However, QAT only works properly with extensive network augmentation, which leads to a more complex and computationally expensive training process. ## IV Architecture-Specific Deployment Our pipeline provides a deployment framework for targeting microcontrollers, see Fig. 2. We call this framework _dnnruntime_. It uses a platform- independent, offline, and ahead of time conversion tool together with a runtime library. The conversion tool maps pre-trained DNNs stored in the ONNX format to C code (Sec. IV-A), while the runtime library implements platform- specific DNN operators that are subsequently used by the code emitted from the conversion tool (Sec. IV-B). Our implementation is novel in the way that it exploits static properties of trained DNNs (i.e. fixed layer configurations and parameters). and therefore removes the necessity of interpreting the DNN at runtime. This includes dynamic allocation of memory for intermediate tensors which can be simulated offline allowing heap allocation to be conducted at compile time. This not only decreases the computational overhead at runtime but also allows metrics like simulated memory consumption to be directly fed back into the overall optimization process without having to evaluate the model on the target system. Figure 2: Overview of the process implemented by our conversion tool, mapping trained ONNX DNN models to C-Code and finally compiling them to deployable binaries. ### IV-A Conversion Tool The main functionality of the conversion tool is to generate ANSI C code based on a given ONNX model of the DNN to be deployed. This involves two steps: (1) parsing and converting the model to an intermediate representation, and (2) using this representation to determine a suitable data format, simulate memory allocation and generate an implementation describing the model’s structure in code. The ONNX format stores a DNN’s compute graph as a directed, acyclic graph with a set of fixed inputs and outputs. Every node in the graph represents an operation and can have multiple incoming and outgoing edges assigned to it. Edges describe the dataflow during the DNN’s execution. Besides that, based on the type of operation a node represents, additional static parameters tensors can also be assigned to it. #### IV-A1 Mapping ONNX to target-specific intermediate format We first map a given ONNX representation to an architecture-specific intermediate format that can be used to emit program code later on. This involves three consecutive steps. First, we concatenate the static tensors of all nodes into a byte-stream. The single elements of each tensor are stored in the stream by using a little- endian byte order as this is the default memory format on ARM architectures (of course, this can be modified easily). Additionally, we add padding bytes where necessary to avoid triggering the memory protection unit (MPU) when accessing tensor data at runtime. Afterwards, we generate descriptor structures containing the location of each tensor in the byte-stream and additional metadata such as data types and tensor shapes. Sparse tensors are handled as edge cases as they are generated by element pruning during our pipeline’s compression stage. To reduce memory usage, our tool applies a conversion from the original full-sized layout of the tensors to a more compact _Compressed Row Storage (CRS)_ [16] layout, see Fig. 3 for an example. CRS reduces the memory footprint, allows for an optimized implementation of matrix-vector products, and does not pose any requirements on how sparsity is distributed inside a tensor (hence, element pruning during compression can ignore the subsequent space-saving storage of pruned tensors). A disadvantage of CRS is that it can only be applied to 2D-tensors (matrices), which means that we need to map higher dimensional weight tensors of convolutions to 2D space before processing them. $\displaystyle A=\begin{bmatrix}10&0&0&0&1\\\ 0&7&0&2&0\\\ 0&0&8&0&0\\\ 14&0&0&0&6\end{bmatrix}$ values | 10 | 1 | 7 | 2 | 8 | 14 | 6 ---|---|---|---|---|---|---|--- col_ind | 0 | 4 | 1 | 3 | 2 | 0 | 4 row_ptr | 0 | 2 | 4 | 5 | 7 | | Figure 3: Conversion of an asymmetric matrix $A$ (left) into its CRS representation (right): we emit three smaller arrays to the bytes-stream instead of one. The first one contains all non-zero values, the second one contains column indices, and the third one contains row pointers. Second, the conversion tool generates descriptor structures for all dynamic activation tensors. This is more compilicated than it is for static parameter tensors as activation tensors are only represented in the ONNX model’s compute graph as edges. Edges are not required to provide any meta information like data types or shapes. However, this information is mandatory for our conversion tool. Hence, we implement a process called _shape inference_. The idea is to trace the execution of a DNN through its compute graph from input to output nodes, and to use these traces to infer the shapes and types of intermediate tensors assigned to inner edges. Third, we parse and interpret all operator nodes in the ONNX compute graph and bring them into a topological and serialized order. This information is then used during code generation to determine the execution order of operations. #### IV-A2 Code Generation Using the intermediate representation generated by the first step the conversion tool can emit code. We start by estimating the minimal heap size required for storing activation tensors. This information can be queried offline as once the training of a DNN is complete, its structure and dimensionality remains unchanged throughout its lifetime. Using the minimal heap size, we define a fixed-size memory balloon at compile time (eliminating the need for dynamic memory management at runtime). A naive approach that estimates the size of this balloon calculates the product of the shapes of all activation tensors and multiplies them with the byte sizes of their respective data types. However, this is not space-efficient as usually the lifetimes of these tensors are rather short. Hence, parts of the heap memory can be reused for multiple tensors during an inference pass. This may have a big impact on the amount of memory required. We take advantage of this by implementing an offline _memory planning algorithm_ based on graph tracing and using a first-fit allocation strategy therein. We estimate optimal heap re-usage in two steps: First, based on incoming and outgoing edges of nodes (i.e., operators) in the input model, the algorithm creates two lists per operator: The first list contains all tensors that have to be allocated for that operator (i.e., allocation list) and the second list contains all previously allocated tensors that can be discarded (i.e., release list). Second, the algorithm proceeds iterating through the sequence of operators starting with an empty, infinitely sized memory balloon. For each operator, it first iterates the tensors in the corresponding free list and marks their space in the memory balloon as unoccupied. After that, it iterates the allocate list and tries to reserve pieces of memory based on the shapes and data types of the tensors. To find suitable locations in the balloon, the algorithm compares the required sizes with available segments of free memory starting from the beginning (i.e., first fit). Once it found suitable pieces of memory, it marks them as allocated in the balloon and stops searching. During all steps, the algorithm keeps track of the maximum size of the memory balloon. The emitted code implements a two-function API: The first function allows to setup the converted DNN and the second function executes an inference given an input sample. The latter is implemented based on the list of topologically sorted ONNX operator nodes stored in the previously generated intermediate representation. For each operator a function call is emitted. These functions are implemented by the _runtime library_. To give context to these function and pass intermediate results between them, we provide references to constant tensor descriptor structures generated as part of the intermediate representation. All static data associated with weight tensor descriptors is stored in a byte-array in the intermediate representation. Therefore, a constant C-array (i.e. flash memory) containing all the data is emitted, accordingly. The amount of random access memory required for intermediate activation tensors is based on the minimal memory balloon previously estimated by our _memory planning algorithm_. Hence our tool emits another accordingly sized zero initialized non-constant C-array (i.e. heap memory). ### IV-B Runtime Library To perform the operations described by the input ONNX model, the code emitted by our conversion tool relies on additional DNN operator functionality that we implement by a runtime library. Currently, this includes operators such as convolutions, linear transformations, batch-normalization, pooling operations, and activation functions. All our implementations follow the ONNX operator specification555https://github.com/onnx/onnx/blob/master/docs/Operators.md. Where required, we also implement quantized versions of these operators. Based on the target platform there are different possibilities to optimize the execution of DNNs. During profile tests we found that most execution time is spent on computing convolutions or matrix-vector products. Hence, an optimal implementation of these operator types yields significant improvements in both resource consumption and execution time. Less crucial but still significantly, some operations can be removed from a DNN’s compute graph by applying graph optimization, which we apply after DNN training and compression. Notable optimization techniques include _batch normalization folding_ [10] and the fusing of ReLU activation functions into preceding quantized linear or convolutional operations [10]. For our experiments we focus on Cortex-M0+ and Cortex-M4 processors. This is why our implementation heavily makes use of the information of these processor architectures. A major algorithmic optimization that we apply is to unroll 2D convolutions into more CPU- and memory-friendly matrix-vector products (_im2col mapping_). Hence, during mapping we rearrange both the input tensors and the parameter tensors of convolutions. This is a common approach used in digital signal processing.666e.g., Matlab® provides convmtx() and convmtx2() for this task.. In addition, this mapping also enables our conversion tool to apply CRS to convolutional layers. Since DNNs also use matrix-vector products in linear transformations, a nice additional side-effect is that by using the im2col mapping complete inference passes can be described by matrix-vector products and non-linearities alone. Moreover, ARM provides optimized open-source implementations for matrix products in their _CMSIS_ library777https://github.com/ARM-software/CMSIS_5. Using them is especially beneficial on architectures like the Cortex-M4 as it allows to use SIMD instructions provided by ARM’s _Digital Signal Processing (DSP)_ extension, see Sec. V. ## V Evaluation TABLE I: Parameters of the DNN architectures used in our experiments. AlexNet (44.7M) | ResNet (9.4M) | LeNet (1.2M) ---|---|--- Input: [3, 32, 32] | Input: [1, 28, 28] [3, 64, 2, 2] | [64, 64, 3, 1] | [1, 32, 3, 1] [64, 192, 3, 1] | [64, 128, 3, 1] | [32, 64, 3, 1] [192, 384, 3, 1] | [128, 256, 3, 1] | [384, 256, 3, 1] | [256, 512, 3, 1] | MaxPool: [2] [6400, 4096] | [25088,10] | [9216, 128] [4096, 4096] | | [128, 10] [4096, 10] | | Output: [10] SoftMax Tuples describe linear layers in the form of [num. inputs, num. outputs]. Every layer uses ReLU as their non-linearity except the last ones which are followed by SoftMax; AlexNet/LeNet: quadruples describe 2D-convolutions by [channels, filters, kernel size, stride]; ResNet: quadruples describe residual blocks each with two 2D-convolutions in the form of [block in. channels, block out. filters, conv. kernel size, conv. stride]; AlexNet/ResNet: every 2D-convolution is followed by batch normalization. TABLE II: Microcontrollers considered in our evaluation. | Raspberry Pi Pico | Arduino Nano 33 BLE Sense | Raspberry Pi 4B ---|---|---|--- Processor | RP2040, Cortex-M0+ (Armv6-M) | nrf52840, Cortex-M4 (Armv7-M) | BCM2711 SoC, Cortex-A72 (ARMv8-A) Clock | 133 MHz | 64 MHz | 1.5 GHz Flash | 2 MB | 1 MB | 16 GB (SD-Card) RAM | 256 KB (SRAM) | 256 KB (SRAM) | 8 GB (SDRAM) SIMD | No | Yes (ARM DSP) | Yes (ARM Neon) (a) AlexNet on CIFAR-10. (b) ResNet on CIFAR-10. (c) LeNet on MNIST. Figure 4: Element-wise and structural pruning applied to the DNN architectures. The curves describe the relation between theoretical model size and accuracy of compressed models relative to their uncompressed baselines when using different pruning techniques in combination with an iterative AGP pruning schedule. We define theoretical model size to be the number of weights a model features excluding all weights that have been set to zero by pruning. To evaluate our pipeline we selected three popular DNN architectures: (1) a convolutional network similar to the one proposed by Krizhevsky et al. [17] to classify CIFAR-10 images (AlexNet), (2) a residual network [3] (ResNet), and (3) a smaller network architecture initially proposed by LeCun et al. [18] (LeNet) for classifying the MNIST handwritten digit database. See Table I for more details. We trained AlexNet and ResNet on the CIFAR10 [4] dataset for 100 epochs with mini-batches of size 80 and LeNet on the MNIST handwritten digit datasets [34] for 20 epochs with mini-batches of size 48 (as training converges on MNIST considerably faster). On all the models we used stochastic gradient descent (SGD) with a momentum of 0.9 and a learning rate of $1e-3$. We achieved a maximum accuracy of 86.51% for AlexNet, 85.97% for ResNet, and 98.46% for LeNet that serve as baselines for our experiments. To evaluate our deployment pipeline we selected three target systems: (1) a Raspberry Pi Pico, (2) an Arduino Nano 33 BLE Sense, and (3) a Raspberry Pi 4B (that serves as a larger reference system), see Table II. As our runtime library is mainly optimized for Cortex-M architectures, we instead use onnxruntime [32] for the deployment to the Raspberry Pi 4B. We compare the performance of different pruning techniques in Sec. V-A and discuss their combination with quantization in Sec. V-B. We analyze the memory footprint of our compressed DNNs in Sec. V-C. In Sec. V-D we discuss the execution time and power/energy consumption w.r.t. the predictive accuracy from a real-world execution of the compressed DNNs on the target platforms. ### V-A Comparison of Pruning Techniques First, we present results of pruning experiments conducted for each of our three test DNN architectures. We repeated model training from scratch and increased pruning target rates starting with 0% as the un-pruned baseline and ending with 99% (i.e., a relative theoretical model size of 1%) as the most aggressively pruned configuration. For each of the configurations we repeated the experiment five times and report their means and standard deviations. Figure 5: Comparison of pruning schedules for both structural and element-wise pruning with AlexNet on CIFAR-10 (L1: pruning selected parameters based on $\ell_{1}$-norm; Level: selects parameters based on magnitude level). Figs. 4 and 5 show the predictive accuracy of the models (relative to the un- pruned baseline) over percentages of theoretically remaining parameters on the validation dataset. We first investigate the influence of pruning schedules on the achievable pruning rate, see AlexNet on CIFAR-10 in Fig. 5. While we cannot observe major differences between iterative (i.e., Automated Gradual Pruning, AGP) and one-shot schedules for final parameter counts $>10\%$, we see that iterative schedules perform slightly better both for element-wise and structural pruning for parameter counts below $10\%$. We argue that this is because it is more difficult for the network to retrain when a large number of parameters are removed at once than when parameters are removed gradually and retraining is possible in between. Hence, we decided to focus on iterative pruning schedules for all further experiments. With the iterative AGP schedule on the same experimental setup we tested different pruning heuristics for structural and element-wise pruning on all architectures, see Figs. 4(a) to 4(c). We report four different heuristics for structural pruning and one for element-wise pruning alongside a random selection (’Random’) approach as a baseline for the more elaborate heuristics. For structural pruning, we use both the $\ell_{1}$\- and the $\ell_{2}$-norm of parameter structures (’L1’ and ’L2’) as well as their gradient size (’Gradient’) and the average percentage of zeros in their corresponding activation (’Activation’) as heuristics. For element-wise pruning we use a magnitude level to decide which elements to prune (’Level’). While we see a significant improvement of the level-based heuristic over the corresponding random selection for element-wise pruning on all our three DNN architectures, we cannot observe a similar behaviour for structural pruning. Instead, none of the more complex structural pruning heuristics managed to significantly improve over the random selection approach. There is also only little variation in the results of the heuristics. We believe that the main reason for this is the iterative re-training between pruning steps: while introducing pruning during DNN training can cause degradation in the predictive quality of a model, it was very often regained in a short number of epochs when re-training. This is in line with results reported in previous work [29]. In all our experiments we used the same target compression rates for both element-wise and structural pruning. However, we see that the structural pruning experiments resulted in models that exceed their selected targeted compression rate. In some cases this reduces the parameter count to almost 99.9%. For element-wise pruning, we do not see such an effect. The reason for this is related to the removal of structures from DNN models during structural pruning: due to the existence of data dependencies between layers, removing structures from their parameter tensors also influences the shapes of tensors in surrounding layers. For element-wise pruning, parameters are not completely eliminated from the DNN but are instead just set to zero. Therefore, all data dependencies remain in the network, no tensor shape changes occur, and the pruning target rate is more precisely met. ### V-B Combining Pruning and Quantization We present experimental results for weight quantization in combination with pruning for our three test DNN architectures. Fig. 6(a) shows the results for quantization in combination with structural pruning and Fig. 6(b) shows the results for element-wise pruning. The different colors refer to the models we trained and quantization strategies are differentiated with the line and marker style. We aim to give an understanding on how much the additional quantization error alone influences the prediction quality of quantized models. Hence, as before, the $x$-axes show the relative theoretical model size reduction while this time the $y$-axes reports the accuracy decrease of each pruned and quantized model in relation to its pruned but not quantized version.888We want to point out that quantization does not change the number of parameters in a model, only their resolution, so the application of quantization does not affect the relative theoretical model size, although the model effectively becomes smaller. We present our findings on actual size reduction achievable with quantization in Sec. V-C. This allows us to focus on the additional error that is introduced through quantization alone. When looking at the results of structural pruning in combination with both Quantization as a post process after training (PPQ) and Quantization Aware Training (QAT), we see that the techniques work well together with pruning for all our three architectures. In Fig. 6(a) we see that the accuracy decrease consistently is $<5\%$ even when using quantization in combination with aggressive pruning regimes. The only outliers we monitored were part of our experiments on the LeNet architecture. Here, for the two most aggressive pruning configurations, the accuracy decrease between the non-quantized and quantized models went to around 40% for both tested quantization strategies. Moreover, we observed an increase in standard deviation as accuracy decreased, which we believe is related to an increase in variance in the trained weight values that we observed for LeNet at higher compression rates. The higher the variance of the values in a weight tensor, the worse quantization can represent these values in integer space. (a) L1-norm AGP structural pruning (b) Level AGP element-wise pruning Figure 6: Decrease in accuracy on the evaluation dataset observed for our three DNN architectures when combining pruning and quantization techniques. We also tested element-wise pruning in combination with PPQ and QAT, see Fig. 6(b). Other than for structural pruning, where PPQ performed consistently well even in combination with aggressive pruning configurations, for element-wise pruning we observed accuracy decreases of over 70% in comparison to the un- quantized versions. In particular, we observe that PPQ noticeably failed for models that have been compressed by element-wise pruning to 10% or less of their original parameter count. This is despite the fact that the technique performed well for pruning configurations that target compression rates above 10%. In contrast, QAT performed significantly better than PPQ even when used together with aggressive pruning configurations. The technique was able to keep the accuracy decrease very close to 0% during all conducted experiments. Hence, we conclude that PPQ generally seems to perform better when being applied in combination with structural pruning than when used with element pruning. QAT on the other hand performed consistently well, both in combination with structural and element-wise pruning. (a) Flash consumption of the compressed models (left: AlexNet/CIFAR-10; middle: ResNet/CIFAR-10; right: LeNet/MNIST). (b) SRAM consumption of the compressed models (left: AlexNet/CIFAR-10; middle: ResNet/CIFAR-10; right: LeNet/MNIST). Figure 7: Memory required for the deployment of our three DNN architectures (AlexNet, ResNet, LeNet) when applying the proposed runtime environment. (a) shows the required amounts of read-only ”Flash” memory to store the DNNs’ trained weights, while (b) shows the required amounts of random access memory ”SRAM” to store the dynamic activation tensors. The dashed horizontal red and green lines mark the memory limits of our two considered embedded target platforms, the Raspberry Pi Pico and the Arduino Nano 33 BLE Sense. ### V-C From Size Reduction to Memory Consumption As described in Sec. IV the execution of DNNs requires both static read-only and dynamic random access memory. We now discuss the memory footprint of the models when they are being deployed. Fig. 7(a) shows the relation between relative model size (on the $x$-axes) and ROM/flash consumption in Kibibyte (on the $y$-axes) and Fig. 7(b) shows the same relation for SRAM. The curves for the pruned models are drawn with solid lines while the curves for the models where additional quantization was applied are dashed. To also highlight the importance of model compression we added the Flash and SRAM limits of out two microcontroller target platforms (red and green dashed horizontal lines). On both of these platforms flash and SRAM availability are two of the main bottlenecks for deployment. We see a linear correlation between flash consumption and relative theoretical model size for structural pruning in Fig. 7(a) (note that both axes are in logarithmic scale). The reason for this is that pruned structures can be completely removed from the model which immediately decreases its memory consumption in the process. For element-wise pruning, there is no such direct correlation. Instead, when looking at the orange curves in Fig. 7(a) we observe plateaus for higher theoretical model sizes. This is because pruning only generates sparse weight matrices. Our runtime environment utilizes compressed row storage (CRS) as a special decoding technique to store sparse weight matrices space-efficiently. However, a characteristic of the technique is that decoding a tensor will only result in memory savings after a certain percentage of sparsity has been reached. For any amount of sparsity below this threshold, it is better to just use the default memory layout where all elements are saved sequentially. An additional observation we made is that for models that are compressed using quantization and element-wise pruning the threshold at which CRS becomes feasible is much higher than for just pruned models. Again, this is related to the properties of CRS decoding. Instead of all values, only values unequal to zero (or a zero point) are stored. To preserve their position in the original un-decoded matrix, the row and column indices of the values must be stored as well. For larger matrices, like they exist in DNNs, these indices usually require 16- or 32-bit integers to be stored correctly. Therefore, memory savings made by CRS can be considered as a trade-off between storing some elements and their indices versus storing all elements without any indices. When introducing sparsity into matrices with 32-bit floating point values, this quickly becomes a good trade-off. However, since quantized values require only 8 bits, while the index values introduced by CRS are still usually at least 16 bits long, the amount of sparsity that has to be introduced before memory can be saved is higher. Fig. 7(b) shows the relationship between relative theoretical model size and required SRAM. Similar to flash consumption, a correlation between model size and SRAM consumption can be observed for structural pruning. When structures are removed from parameter tensors during pruning, data dependencies between layers are also removed. Therefore, the shapes of the dynamic intermediate activation tensors stored in RAM and shared between succeeding layers reduce as well. The reason why the relation is not perfectly linear is because our runtime library tries to re-use heap memory for several activation tensors. How good this strategy works and how much memory it can save depends on the topology of the DNN that it is applied to. We cannot observe a reduction in SRAM consumption for element-wise pruning. This is because we do not remove any elements during element-wise pruning. While CRS can compress the sparse parameter tensors, it cannot change their original shapes. Therefore, as expected, no data dependencies are removed during decoding and as a result all intermediate activation tensors retain their original sizes. TABLE III: Average ($n=8$) execution time, power and energy consumption observed per inference for compressed version of LeNet targeting two relative accuracy levels. The baseline model achieved an accuracy of $\approx$98% for both float and uint8. rel. accuracy | compression | type | system | execution time (ms) | power (mW) | energy (mWs) ---|---|---|---|---|---|--- $>$99.00% | 99.18% | structure, float | arduino | $44.15$ p m 0.03 | $141.14$ p m 0.07 | $6.23$ p m 0.01 | | | pico | $184.28$ p m 0.17 | $81.42$ p m 0.42 | $15.00$ p m 0.07 | | | pi4b | $1.81$ p m 0.55 | $3754.70$ p m 284.92 | $6.65$ p m 1.35 | | structure, uint8 | arduino | $18.84$ p m 0.04 | $101.06$ p m 1.80 | $1.90$ p m 0.03 | | | pico | $50.06$ p m 0.03 | $79.37$ p m 1.03 | $3.97$ p m 0.05 | | | pi4b | $2.16$ p m 0.60 | $3624.90$ p m 272.88 | $7.68$ p m 1.53 | 95.00% | element, float | arduino | — | — | — | | | pico | $1004.77$ p m 0.07 | $81.44$ p m 0.10 | $81.82$ p m 0.11 | | | pi4b | $5.66$ p m 1.39 | $3772.18$ p m 263.01 | $21.00$ p m 3.34 | | element, uint8 | arduino | $360.95$ p m 0.09 | $142.28$ p m 0.10 | $51.36$ p m 0.04 | | | pico | $392.23$ p m 0.07 | $84.82$ p m 0.37 | $33.27$ p m 0.15 | | | pi4b | $5.10$ p m 1.27 | $3673.70$ p m 332.91 | $18.31$ p m 3.04 $>$97.00% | 99.7% | structure, float | arduino | $22.48$ p m 0.01 | $105.32$ p m 0.27 | $2.37$ p m 0.01 | | | pico | $73.26$ p m 0.33 | $81.72$ p m 0.48 | $5.99$ p m 0.03 | | | pi4b | $1.91$ p m 0.56 | $3554.33$ p m 232.34 | $6.67$ p m 1.48 | | structure, uint8 | arduino | $18.79$ p m 0.03 | $82.54$ p m 0.80 | $1.55$ p m 0.02 | | | pico | $24.97$ p m 0.44 | $79.92$ p m 0.88 | $2.00$ p m 0.04 | | | pi4b | $2.01$ p m 0.53 | $3581.79$ p m 240.22 | $7.11$ p m 1.41 | 97.00% | element, float | arduino | — | — | — | | | pico | $658.70$ p m 0.10 | $82.44$ p m 0.68 | $54.30$ p m 0.45 | | | pi4b | $5.35$ p m 1.23 | $3816.72$ p m 246.08 | $20.12$ p m 2.79 | | element, uint8 | arduino | $303.15$ p m 0.09 | $141.58$ p m 0.05 | $42.92$ p m 0.01 | | | pico | $392.27$ p m 0.10 | $84.67$ p m 0.45 | $33.22$ p m 0.18 | | | pi4b | $4.64$ p m 1.39 | $3787.84$ p m 395.19 | $17.05$ p m 3.16 ### V-D Deployment Results As the last step of evaluating our pipeline, we deployed several of the pruned and quantized models from our previous experiments on our target systems and monitored key runtime metrics. We especially focused on execution time per inference, power, and energy consumption. For measuring these metrics on our test systems, we used an Agilent N6705A DC Power Analyzer to provide them with a regulated power supply. Furthermore, the Power Analyzer allowed us to measure the current and power drawn by the systems. To measure the execution time required for calculating the energy consumption of our DNN models, we used a GPIO signal. We toggled the signal every time an inference started and finished and monitored it using an oscilloscope. We present the results of our measurements for LeNet in Table III and the results for our other two DNN architectures in Table IV. TABLE IV: Average ($n=8$) execution time, power and energy consumption observed per inference for AlexNet and ResNet. To achieve different compression rates and thus different accuracy scores, we used L1-norm structural pruning. In addition, we used post training quantization for all non floating-point measurements. compression | type | rel. accuracy | system | execution time (ms) | power (mW) | energy (mWs) ---|---|---|---|---|---|--- 95.9% | float | 97.79% | arduino | — | — | — | | | pico | — | — | — | | | pi4b | $6.60$ p m 1.85 | $3795.81$ p m 334.67 | $24.43$ p m 4.10 | uint8 | 97.56% | arduino | — | — | — | | | pico | $1766.58$ p m 0.25 | $92.40$ p m 0.03 | $163.23$ p m 0.07 | | | pi4b | $3.96$ p m 1.12 | $3809.41$ p m 250.59 | $14.82$ p m 2.71 98.2% | float | 94.88% | arduino | — | — | — | | | pico | — | — | — | | | pi4b | $2.68$ p m 0.70 | $3687.37$ p m 221.25 | $9.71$ p m 1.76 | uint8 | 94.65% | arduino | $403.11$ p m 0.16 | $139.96$ p m 0.17 | $56.42$ p m 0.07 | | | pico | $669.97$ p m 0.19 | $86.41$ p m 0.02 | $57.89$ p m 0.03 | | | pi4b | $2.93$ p m 0.41 | $4187.75$ p m 89.73 | $12.28$ p m 2.05 99.6% | float | 83.26% | arduino | $252.40$ p m 0.40 | $107.83$ p m 0.42 | $27.22$ p m 0.10 | | | pico | $1435.42$ p m 0.34 | $81.55$ p m 0.07 | $117.06$ p m 0.09 | | | pi4b | $2.45$ p m 0.09 | $4182.59$ p m 66.35 | $10.25$ p m 0.35 | uint8 | 82.33% | arduino | $110.00$ p m 0.01 | $141.17$ p m 0.04 | $15.53$ p m 0.01 | | | pico | $186.57$ p m 0.12 | $86.86$ p m 0.21 | $16.21$ p m 0.04 | | | pi4b | $2.10$ p m 0.14 | $4134.21$ p m 52.03 | $8.68$ p m 0.60 (a) AlexNet, CIFAR10, baseline accuracy $\approx$86% for both float and uint8 compression | type | rel. accuracy | system | execution time (ms) | power (mW) | energy (mWs) ---|---|---|---|---|---|--- 93.2% | float | 96.25% | arduino | — | — | — | | | pico | — | — | — | | | pi4b | $29.41$ p m 2.62 | $4234.81$ p m 121.13 | $124.24$ p m 7.11 | uint8 | 96.43% | arduino | $11\,983.40$ p m 2.05 | $143.74$ p m 0.49 | $1722.48$ p m 5.69 | | | pico | $25\,997.20$ p m 4.00 | $91.94$ p m 0.01 | $2390.24$ p m 0.41 | | | pi4b | $21.32$ p m 3.32 | $4085.92$ p m 213.90 | $86.54$ p m 9.62 98.7% | float | 92.50% | arduino | $6105.20$ p m 1.60 | $105.76$ p m 1.85 | $645.69$ p m 11.16 | | | pico | $46\,266.50$ p m 16.82 | $85.78$ p m 0.02 | $3968.83$ p m 1.49 | | | pi4b | $9.41$ p m 4.27 | $4536.99$ p m 183.46 | $42.06$ p m 16.53 | uint8 | 88.52% | arduino | $2425.50$ p m 0.92 | $102.38$ p m 1.10 | $248.31$ p m 2.70 | | | pico | $4269.78$ p m 2.58 | $87.81$ p m 0.04 | $374.92$ p m 0.36 | | | pi4b | $7.74$ p m 2.78 | $3863.07$ p m 427.70 | $28.82$ p m 6.98 99.6% | float | 83.48% | arduino | $1754.75$ p m 0.31 | $106.57$ p m 2.64 | $187.00$ p m 4.63 | | | pico | $11\,140.25$ p m 1.85 | $83.13$ p m 0.24 | $926.12$ p m 2.61 | | | pi4b | $5.42$ p m 1.66 | $3560.66$ p m 421.12 | $18.70$ p m 4.01 | uint8 | 78.28% | arduino | $806.80$ p m 0.60 | $104.48$ p m 0.09 | $84.29$ p m 0.13 | | | pico | $1214.53$ p m 1.00 | $84.13$ p m 0.06 | $102.18$ p m 0.10 | | | pi4b | $3.45$ p m 0.22 | $3881.06$ p m 72.77 | $13.37$ p m 0.60 (b) ResNet, CIFAR10, baseline accuracy $\approx$85% for float and $\approx$84% for uint8 For the LeNet architecture, we used the models’ accuracy on its evaluation dataset to compare the deployment of different pruning and quantization techniques. The logic behind this was that if models that were compressed using different approaches can achieve a similar accuracy, then they can be seen as direct alternatives and are therefore comparable. In our experimental setup we defined two relative accuracy boundaries for which we selected the smallest compressed models from our previous tests that meet them: $>$99% and $>$97%. In the second column of Table III we can see that LeNet can be compressed to well over 5% of its original parameter count while still passing both accuracy thresholds. We tested all selected models on our three target systems, including not only the pruned models but also their quantized counterparts, see the third and fourth columns. Note that using element pruning we were not able to deploy all selected models on the Arduino. For all deployed LeNet models, we monitored execution time, power and energy consumption over a span of 8 different inferences. Their resulting averages are presented in the remaining columns of Table III. First, we see that the execution time per inference on the Pi 4B is significantly lower than on the Arduino or the Pi Pico. This is expected as the Pi 4B runs between 1.0 to 1.5 GHz while both the Arduino and the Pico run in a lower MHz range. When comparing the Raspberry Pi Pico and the Arduino Nano we observed a higher execution time on the Pico than on the Arduino (consistently). This is even though the Arduino runs with a clock speed only around half as fast as the Pico. On both systems the quantized models always outperformed their floating- point counter parts although much more pronounced on the Pico. This can be explained by the features present on both systems. First, the Arduino’s Cortex-M4 processor implements a real floating point unit while on the Pi Pico’s Cortex-M0+ processor floating point arithmetic has to be simulated. Second, the Arduino’s Cortex-M4 processor supports ARM’s Digital Signal Processing (DSP) extension giving it access to a subset of SIMD instructions to accelerate its integer operations. The Pi Pico does not implement the DSP extension. Considering the power measured during inference for all deployed models, we see that all our tested systems on average draw a constant amount of power while we see a more significant variation in the different samples taken for each model on the Pi 4B (note that the power consumption is much higher than on the other two systems). On the Arduino, the measured power consumption was between 100 to 150 mW on average while on the Pi Pico it was around 80 mW. In contrast to that, the Pi 4B generally consumed around 4W. However, in addition to power, execution time is the second factor in calculating a system’s energy consumption. Looking at the results, we see that for some cases the Pi 4B scored the best energy consumption per inference. It is often followed by the Arduino and then the Pi Pico. This is in reverse to the power consumption and shows that having an excellent runtime can compensate for high power consumption. Besides LeNet, we also deployed our AlexNet and ResNet architectures, see Table IV. For both architectures we evaluated models which were compressed using structural pruning. This is different from our LeNet experiments where we tested both structural and element-wise pruning. The reason we did not do this for AlexNet and ResNet is due to memory limitations on our Arduino and Pi Pico target systems, which made it impossible for us to deploy any element pruned models. For structural pruning the situation is different and we were able to feasibly deploy models. However, we were still forced to select models trained with aggressive compression rates that removed well over 90% of the original parameters. As a consequence we had to accept decreases in accuracy to be able to deploy models, see the third column of Tables 8(a) and 8(b). To measure execution time, power, and energy consumption, we used the same approach as before. We again monitored all three metrics over 8 different inferences and calculated their averages and respective standard deviations. When looking at the results for AlexNet and ResNet, we see the same patterns we discussed for LeNet in Table III. Yet we evaluated the two architectures as a way to explore upper boundaries of feasible DNN deployment. AlexNet features a very high number of trainable parameters while ResNet contains a high number of large and computationally expensive 2D-convolutions. This affects the deployment of the two architectures differently. While for AlexNet, we need to apply high compression rates to shrink the model size far enough to fit the it into the memory of our target microcontrollers (see Sec. V-C) for ResNet execution time is the primary bottleneck. Even after pruning almost 99% of ResNet’s parameters we still measure inference times of around 4 seconds on the Pi Pico, and of around 6 seconds on the Arduino Nano even when applying quantization, see the fifth column of Table 8(b). Only after removing $>$99% of all parameters the execution time ended up in an acceptable range (of around 1 second) on both microcontroller targets. However, these extremely high compression rates were accompanied by a high loss in accuracy, see the first column. This may make the usage of such a complex DNN architecture impractical for microcontroller targets. Therefore, we conclude that not only a model’s parameter count but also its topology decides if it is deployable on a target system from a performance standpoint. ## VI Discussion of Results By using our compression and deployment pipeline, we were able to automatically and feasibly deploy DNNs to microcontrollers. Compressing DNNs allowed us to achieve significant savings in memory consumption, execution time, and energy consumption at runtime, without sacrificing model accuracy for it. For pruning, we achieved the best results and most savings with structural pruning. When comparing different pruning strategies, our experiments indicate that structural pruning offers better opportunities for saving memory and execution time than element-wise pruning. In addition, the execution of DNNs that were compressed using this technique did not require any special support for sparse matrix formats as it is required for element pruning. Furthermore, we observed that for structural pruning, using different state-of-the-art heuristics did not have that much of an impact. Choosing a reasonable pruning schedule and allowing for retraining has proven to be more effective. Additionally, applying weight quantization together with structural pruning resulted in even more savings not only in memory consumption, but also in execution time. This is due to the fact that our target systems were able to process much more efficient in integer- than in floating-point space. Besides that, we noticed that our different compression and deployment strategies had almost no influence on the power drawn by both the Pi Pico and the Arduino during inference. This means that the observed energy savings were mainly the result of execution times. We come to the conclusion that a DNN model is deployed optimally on a microcontroller, if it runs on a system where it fits into the available memory and draws the least amount of power under load while still being able to run inferences in a reasonable time frame. Furthermore, we argue that a DNN’s execution time has to be seen in relation to the frequency at which input samples are generated by connected sensors. ## VII Conclusion In this work we presented a configurable pipeline for compressing DNNs and deploying them on Cortex-M based microcontrollers. To achieve compression, our pipeline utilizes network pruning and weight quantization techniques. The deployment is handled by using a proposed runtime environment, which consists of a code generator for mapping trained networks to executable C-code and a runtime library which provides optimized implementations of common DNN layers. We used the introduced pipeline to compare DNNs compressed with different pruning and quantization techniques. Furthermore, we tested how compression influences runtime performance on multiple target systems. We were able to show that even larger DNN architectures like AlexNet or ResNet can be feasibly deployed on microcontrollers featuring memory footprints of as little as 1-2 MB Flash and 256 Kb SRAM while still achieving good execution time and accuracy results. ## Acknowledgement This work was supported by the Bavarian Ministry for Economic Affairs, Infrastructure, Transport and Technology through the Center for Analytics- Data-Applications (ADA- Center) within the framework of “BAYERN DIGITAL II”. ## References * [1] S. Hershey, S. Chaudhuri, D. P. W. Ellis, J. F. Gemmeke, A. Jansen, R. C. Moore, M. Plakal, D. Platt, R. A. Saurous, B. Seybold, M. Slaney, R. J. Weiss, and K. Wilson, “CNN architectures for large-scale audio classification,” in _2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)_ , Mar. 2017, pp. 131–135. * [2] O. Abdel-Hamid, A.-r. Mohamed, H. Jiang, L. Deng, G. Penn, and D. Yu, “Convolutional Neural Networks for Speech Recognition,” _IEEE/ACM Transactions on Audio, Speech, and Language Processing_ , vol. 22, no. 10, pp. 1533–1545, Oct. 2014. * [3] K. He, X. Zhang, S. Ren, and J. Sun, “Deep Residual Learning for Image Recognition,” in _2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)_ , Jun. 2016, pp. 770–778. * [4] A. Krizhevsky, “Learning Multiple Layers of Features from Tiny Images,” _University of Toronto_ , May 2012. * [5] T. Kautz, B. H. Groh, J. Hannink, U. Jensen, H. Strubberg, and B. M. Eskofier, “Activity recognition in beach volleyball using a Deep Convolutional Neural Network,” _Data Mining and Knowledge Discovery_ , vol. 31, no. 6, pp. 1678–1705, Nov. 2017. * [6] V. Sze, Y.-H. Chen, T.-J. Yang, and J. S. Emer, “Efficient Processing of Deep Neural Networks: A Tutorial and Survey,” _Proceedings of the IEEE_ , vol. 105, no. 12, pp. 2295–2329, Dec. 2017. * [7] A. Canziani, A. Paszke, and E. Culurciello, “An Analysis of Deep Neural Network Models for Practical Applications,” _arXiv:1605.07678 [cs]_ , Apr. 2017. * [8] N. D. Lane, S. Bhattacharya, A. Mathur, P. Georgiev, C. Forlivesi, and F. Kawsar, “Squeezing Deep Learning into Mobile and Embedded Devices,” _IEEE Pervasive Computing_ , vol. 16, no. 3, pp. 82–88, 2017\. * [9] Y. L. Cun, J. S. Denker, and S. A. Solla, “Optimal Brain Damage,” in _Advances in Neural Information Processing Systems_. Morgan Kaufmann, 1990, pp. 598–605. * [10] B. Jacob, S. Kligys, B. Chen, M. Zhu, M. Tang, A. Howard, H. Adam, and D. Kalenichenko, “Quantization and training of neural networks for efficient integer-arithmetic-only inference,” in _Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR)_ , Jun. 2018. * [11] M. Nagel, R. A. Amjad, M. V. Baalen, C. Louizos, and T. Blankevoort, “Up or Down? Adaptive Rounding for Post-Training Quantization,” in _International Conference on Machine Learning_. PMLR, Nov. 2020, pp. 7197–7206. * [12] C. Liu, J. Li, and Y. Gong, “SVD-based universal DNN modeling for multiple scenarios,” in _Sixteenth Annual Conference of the International Speech Communication Association_ , 2015. * [13] S. Han, H. Mao, and W. J. Dally, “Deep compression: Compressing deep neural networks with pruning, trained quantization and huffman coding,” _International Conference on Learning Representations (ICLR)_ , 2016. * [14] H. Li, A. Kadav, I. Durdanovic, H. Samet, and H. P. Graf, “Pruning Filters for Efficient ConvNets,” _arXiv:1608.08710 [cs]_ , Mar. 2017. * [15] S. Anwar, K. Hwang, and W. Sung, “Structured pruning of deep convolutional neural networks,” _J. Emerg. Technol. Comput. Syst._ , vol. 13, no. 3, Feb. 2017. * [16] R. Barrett, M. Berry, T. F. Chan, J. Demmel, J. Donato, J. Dongarra, V. Eijkhout, R. Pozo, C. Romine, and H. V. d. Vorst, _Templates for the Solution of Linear Systems: Building Blocks for Iterative Methods, 2nd Edition_. Philadelphia, PA: SIAM, 1994. * [17] A. Krizhevsky, I. Sutskever, and G. E. Hinton, “ImageNet classification with deep convolutional neural networks,” _Communications of the ACM_ , vol. 60, no. 6, pp. 84–90, May 2017. * [18] Y. LeCun, B. Boser, J. S. Denker, D. Henderson, R. E. Howard, W. Hubbard, and L. D. Jackel, “Backpropagation Applied to Handwritten Zip Code Recognition,” _Neural Computation_ , vol. 1, no. 4, pp. 541–551, Dec. 1989\. * [19] P. Hu, X. Peng, H. Zhu, M. M. S. Aly, and J. Lin, “OPQ: Compressing Deep Neural Networks with One-shot Pruning-Quantization,” _Proceedings of the AAAI Conference on Artificial Intelligence_ , vol. 35, May 2021. * [20] Z. Liu, M. Sun, T. Zhou, G. Huang, and T. Darrell, “Rethinking the value of network pruning,” in _International Conference on Learning Representations_ , 2019. * [21] B. Reddy, Y.-H. Kim, S. Yun, C. Seo, and J. Jang, “Real-Time Driver Drowsiness Detection for Embedded System Using Model Compression of Deep Neural Networks,” in _2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)_ , Jul. 2017, pp. 438–445. * [22] S. Lu, G. Qian, Q. He, F. Liu, Y. Liu, and Q. Wang, “In Situ Motor Fault Diagnosis Using Enhanced Convolutional Neural Network in an Embedded System,” _IEEE Sensors Journal_ , vol. 20, no. 15, pp. 8287–8296, Aug. 2020. * [23] J. Lin, W.-M. Chen, Y. Lin, j. cohn, C. Gan, and S. Han, “MCUNet: Tiny Deep Learning on IoT Devices,” in _Advances in Neural Information Processing Systems_ , vol. 33. Curran Associates, Inc., 2020, pp. 11 711–11 722. * [24] B. Zoph and Q. V. Le, “Neural Architecture Search with Reinforcement Learning,” arXiv, Tech. Rep., Feb. 2017. * [25] R. David, J. Duke, A. Jain, V. J. Reddi, N. Jeffries, J. Li, N. Kreeger, I. Nappier, M. Natraj, S. Regev, R. Rhodes, T. Wang, and P. Warden, “TensorFlow Lite Micro: Embedded Machine Learning on TinyML Systems,” _arXiv:2010.08678 [cs]_ , Mar. 2021. * [26] T. Akiba, S. Sano, T. Yanase, T. Ohta, and M. Koyama, “Optuna: A next-generation hyperparameter optimization framework,” in _Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining_, ser. KDD ’19. New York, NY, USA: Association for Computing Machinery, 2019, p. 2623–2631. * [27] S. J. Hanson and L. Y. Pratt, “Comparing biases for minimal network construction with back-propagation,” in _Proceedings of the 1st International Conference on Neural Information Processing Systems_. Cambridge, MA, USA: MIT Press, Jan. 1988, pp. 177–185. * [28] P. Molchanov, S. Tyree, T. Karras, T. Aila, and J. Kautz, “Pruning Convolutional Neural Networks for Resource Efficient Inference,” _arXiv:1611.06440 [cs]_ , Jun. 2017. * [29] S. Han, J. Pool, J. Tran, and W. J. Dally, “Learning both weights and connections for efficient neural networks,” in _Proceedings of the 28th International Conference on Neural Information Processing Systems - Volume 1_ , ser. NIPS’15. Cambridge, MA, USA: MIT Press, 2015, p. 1135–1143. * [30] H. Hu, R. Peng, Y.-W. Tai, and C.-K. Tang, “Network Trimming: A Data-Driven Neuron Pruning Approach towards Efficient Deep Architectures,” _arXiv:1607.03250 [cs]_ , Jul. 2016. * [31] M. Zhu and S. Gupta, “To prune, or not to prune: exploring the efficacy of pruning for model compression,” _arXiv:1710.01878 [cs, stat]_ , Nov. 2017\. * [32] O. R. developers, “ONNX Runtime,” https://onnxruntime.ai/, 2021, version: 1.8.0. * [33] I. Hubara, Y. Nahshan, Y. Hanani, R. Banner, and D. Soudry, “Improving Post Training Neural Quantization: Layer-wise Calibration and Integer Programming,” _arXiv:2006.10518 [cs, stat]_ , Dec. 2020. * [34] Y. LeCun, C. Cortes, and C. Burges, “MNIST handwritten digit database,” _ATT Labs [Online]. Available: http://yann.lecun.com/exdb/mnist_ , vol. 2, 2010.
# SEA++: Multi-Graph-based High-Order Sensor Alignment for Multivariate Time- Series Unsupervised Domain Adaptation Yucheng Wang, Yuecong Xu, Jianfei Yang, Min Wu, Xiaoli Li, Lihua Xie, Fellow, IEEE, Zhenghua Chen111Corresponding Author Yucheng Wang and Yuecong Xu are with Institute for Infocomm Research, A∗STAR, Singapore and the School of Electrical and Electronic Engineering, Nanyang Technological University, Singapore (Email<EMAIL_ADDRESS>xuyu0014@e.ntu.edu.sg).Zhenghua Chen and Min Wu are with Institute for Infocomm Research, A∗STAR, Singapore (Email<EMAIL_ADDRESS>wumin@i2r.a-star.edu.sg).Xiaoli Li is with Institute for Infocomm Research, A∗STAR, Singapore and the School of Computer Science and Engineering, Nanyang Technological University, Singapore (Email: xlli@i2r.a-star.edu.sg).Jianfei Yang and Lihua Xie are with the School of Electrical and Electronic Engineering, Nanyang Technological University, Singapore (Email<EMAIL_ADDRESS>elhxie@ntu.edu.sg). ###### Abstract Unsupervised Domain Adaptation (UDA) methods have been successful in reducing label dependency by minimizing the domain discrepancy between a labeled source domain and an unlabeled target domain. However, these methods face challenges when dealing with Multivariate Time-Series (MTS) data. MTS data typically consist of multiple sensors, each with its own unique distribution. This characteristic makes it hard to adapt existing UDA methods, which mainly focus on aligning global features while overlooking the distribution discrepancies at the sensor level, to reduce domain discrepancies for MTS data. To address this issue, a practical domain adaptation scenario is formulated as Multivariate Time-Series Unsupervised Domain Adaptation (MTS-UDA). In this paper, we propose SEnsor Alignment (SEA) for MTS-UDA, aiming to reduce domain discrepancy at both the local and global sensor levels. At the local sensor level, we design endo-feature alignment, which aligns sensor features and their correlations across domains. To reduce domain discrepancy at the global sensor level, we design exo-feature alignment that enforces restrictions on global sensor features. We further extend SEA to SEA++ by enhancing the endo- feature alignment. Particularly, we incorporate multi-graph-based high-order alignment for both sensor features and their correlations. High-order statistics are employed to achieve comprehensive alignment by capturing complex data distributions across domains. Meanwhile, a multi-graph alignment technique is introduced to effectively align the evolving distributions of MTS data. Extensive empirical results have demonstrated the state-of-the-art performance of our SEA and SEA++ on public MTS datasets for MTS-UDA. Unsupervised Domain Adaptation (UDA) methods have been successful in reducing label dependency by minimizing the domain discrepancy between a labeled source domain and an unlabeled target domain. However, these methods face challenges when dealing with Multivariate Time-Series (MTS) data. MTS data typically consist of multiple sensors, each with its own unique distribution. This characteristic makes it hard to adapt existing UDA methods, which mainly focus on aligning global features while overlooking the distribution discrepancies at the sensor level, to reduce domain discrepancies for MTS data. To address this issue, a practical domain adaptation scenario is formulated as Multivariate Time-Series Unsupervised Domain Adaptation (MTS-UDA). In this paper, we propose SEnsor Alignment (SEA) for MTS-UDA, aiming to reduce domain discrepancy at both the local and global sensor levels. At the local sensor level, we design endo-feature alignment, which aligns sensor features and their correlations across domains. To reduce domain discrepancy at the global sensor level, we design exo-feature alignment that enforces restrictions on global sensor features. We further extend SEA to SEA++ by enhancing the endo- feature alignment. Particularly, we incorporate multi-graph-based high-order alignment for both sensor features and their correlations. Extensive empirical results have demonstrated the state-of-the-art performance of our SEA and SEA++ on public MTS datasets for MTS-UDA. ###### Index Terms: Multivariate Time-Series Data; Unsupervised Domain Adaptation; Graph Neural Network ## I Introduction Time-Series (TS) data have gained significant attention in various fields due to their wide range of applications. Deep Learning (DL) methods have played a pivotal role in addressing TS-related problems by leveraging their ability to effectively capture latent dependencies within the data [1, 2, 3, 4, 5, 6, 7]. However, the effectiveness of DL models heavily relies on the availability of large-scale labeled TS data, which is often a challenge in real-world applications due to the high cost associated with data labeling. To reduce the cost of labeling, Unsupervised Domain Adaptation (UDA) methods have been proposed to transfer the knowledge from a labeled source domain to an unlabeled target domain, enabling effective learning in the absence of target domain labels [8]. Due to domain shifts, existing UDA methods focused on reducing discrepancy across domains by learning domain-invariant features [9, 10, 11, 12, 13, 14]. These methods have achieved decent performance, showing the effectiveness of UDA in reducing label dependency. Subsequently, researchers applied UDA to TS data, aiming to reduce domain discrepancy by learning domain-invariant temporal features extracted by Recurrent Neural Network (RNN) [15], Long Short-Term Memory (LSTM) [16], and Convolutional Neural Network (CNN) [17]. Figure 1: Comparisons between existing UDA methods and ours, where before and after represent before and after alignment respectively. Left: Only global sensor features are aligned, resulting in the poor alignment of the red sensor. Right: SEA aligns the sensor information between domains, so the feature discrepancy of each sensor can be reduced. While substantial progress has been made, existing TS UDA methods may be inapplicable for real-world applications where multiple sensors are deployed simultaneously. For instance, in Remaining Useful Life (RUL) prediction, various types of sensors are employed to measure different physical parameters. On the other hand, in Human Activity Recognition (HAR), multiple sensors are positioned at various locations on the human body to capture activity data. To address these application scenarios, we formulate a more challenging yet practical problem as Multivariate Time-Series Unsupervised Domain Adaptation (MTS-UDA). Applying current TS UDA methods directly to MTS- UDA poses two key limitations. First, MTS data typically consist of signals from multiple sensors, where each sensor collects data with distinct distributions due to variations in sensor types or deployment locations [18]. When adapting existing TS UDA methods to MTS-UDA, they mainly focus on learning global features to facilitate alignment across domains by treating the signals from various sensors as a whole [16, 19]. However, these methods ignore the diverse distributions across sensors, resulting in misalignment at the sensor level and thus leading to suboptimal solutions for MTS-UDA. As presented in Fig. 1 (Left), existing UDA methods focused on aligning the global features between domains, resulting in the poor alignment of the features from the red sensor and limiting the transferability of the model. Second, MTS data contain essential spatial-temporal dependency information due to its multi-source nature. Specifically, the spatial dependency refers to the correlations between sensors, which capture the critical interactive relationships among them. For instance, the spatial dependency in the context of machine RUL prediction corresponds to the correlation between a temperature sensor and a fan speed sensor, as they exhibit strong correlations. Besides, the temporal dependency refers to the temporal relationships between consecutive timestamps. Existing UDA methods are limited in effectively modelling and transferring both dependencies across domains, thus leading to suboptimal solutions for MTS-UDA. These characteristics within MTS-UDA render it a more complex and challenging task compared to standard TS UDA. To address the above challenges, we propose SEnsor Alignment (SEA) for MTS- UDA. We first aim to reduce domain discrepancy at both the local and global sensor levels. At the local sensor level, we propose endo-feature alignment to align sensor-level information between domains. This alignment process includes aligning both the sensor features and sensor correlations, which capture the characteristics of individual sensors and their interactions, respectively. To achieve this alignment, we introduce contrastive sensor alignment and sensor correlation alignment. Further, at the global sensor level, we design exo-feature alignment to reduce global feature discrepancy. This is achieved by enforcing restrictions on the global features derived from stacked sensor features. Moreover, to capture the spatial-temporal dependencies inherent in MTS data and enable their effective transfer across domains, we introduce a graph-based encoder that is able to capture the dependency information. Given that MTS data may not come with pre-defined graphs that represent the relationships between sensors, we propose a multi- branch self-attention mechanism to model these dependencies. Although SEA presents a feasible solution for sensor alignment in MTS-UDA, it employs a low-order statistics-based scheme for achieving endo-feature alignment. This scheme may be insufficient to achieve comprehensive alignment when dealing with complex data distributions in certain scenarios. Furthermore, SEA can benefit from leveraging the dynamically changing distributions within MTS data to enhance its performance. For instance, as a machine’s health deteriorates, the distribution of a fan speed sensor undergoes dynamic changes due to increased friction. Taking these issues into consideration, we develop SEA++ as an extension of SEA by introducing multi- graph-based high-order alignment for enhancing our endo-feature alignment. It incorporates improved sensor correlation alignment and enhanced contrastive sensor alignment, utilizing high-order statistics to achieve comprehensive sensor-level alignment by capturing complex data distributions across domains. Additionally, SEA++ introduces a multi-graph alignment technique specifically tailored for aligning the evolving distributions within MTS data. This involves the construction of sequential graphs that represent the distributions within local temporal intervals, followed by the alignment of corresponding sequential graphs between domains. To optimize the alignment process, weights for balancing sequential graphs are learned based on distribution discrepancies, allowing us to prioritize alignment for challenging graphs. This module effectively aligns the evolving distributions, enabling adaptation to the changing nature of the data. With the two types of enhancements in the improved version, we represent the name with two plus signs as SEA++. Our contributions can be summarized as following. * • We formulate a challenging scenario of Multivariate Time-Series Unsupervised Domain Adaptation (MTS-UDA) in accordance with the characteristics of MTS data. To our best knowledge, this is the first work to design a UDA method specifically for MTS data. * • We analyze the problems underlying MTS-UDA and propose SEnsor Alignment (SEA) that reduces domain discrepancy at both the local and global sensor levels. Meanwhile, SEA captures the spatial-temporal dependencies within MTS data, enabling effective transfer of the dependency information across domains. * • To cope with scenarios with complex data distributions, we introduce SEA++ by enhancing our endo-feature alignment. High-order statistics alignment methods are employed to achieve comprehensive sensor correlation and sensor feature alignment. * • Considering the dynamically changing distributions within MTS data, SEA++ further incorporates a multi-graph alignment technique. It aligns sequential graphs representing local distributions, effectively aligning evolving distributions between domains. * • We evaluate the effectiveness of SEA and SEA++ with various real-world MTS datasets through extensive experiments, demonstrating that the both methods achieve state-of-the-art performance for MTS-UDA. This journal paper presents an extended version of our previous work [20] by introducing multi-graph-based high-order alignment techniques. First, we introduce high-order statistic alignment to improve endo-feature alignment, aiming to achieve more comprehensive sensor-level alignment. Second, we analyze the properties of MTS data and emphasize the significance of dynamically changing distributions inherent within MTS data. Thus, we introduce a multi-graph alignment technique to align the evolving distributions across domains. Third, we have conducted additional experiments, including more compared methods, comprehensive ablation studies, and extensive sensitivity analysis, to fully assess SEA and SEA++. ## II Related Work ### II-A Unsupervised domain adaptation UDA aims to mitigate the need for labeled data by transferring knowledge from a labeled source domain to an unlabeled target domain. To achieve good performance on the target domain, existing UDA methods have mainly focused on minimizing domain discrepancies by utilizing two categories of approaches. The first category comprises adversarial-based methods, which utilize domain discriminator networks to encourage a feature extractor to learn domain- invariant representations. Prominent models in this category include Domain Adversarial Neural Network (DANN) [10] and Adversarial Discriminative Domain Adaptation (ADDA) [11]. These models have served as foundations for various improved variants that address specific challenges [12, 21, 22, 23, 24, 25]. For instance, CDAN [12] and MADA [21] were both extensions of DANN that incorporate multi-mode structure information for classification tasks. DAAN [22] introduced class-aware adversarial loss to align the conditional distribution using DANN. The second category encompasses metric-based methods, which enforce metric restrictions to enable the learning of invariant features. Commonly employed metrics include Maximum Mean Discrepancy (MMD) [9] and deep CORrelation ALignment (CORAL) [13]. Within this category, several variants have been proposed to tackle specific challenges [14, 26, 27, 28]. For example, JAN [14] was developed based on MMD to align the joint distribution of feature and label distributions. MK-MMD [26] employed MMD to align features in the final layers, enhancing their transferability. Furthermore, high-order moment matching called HoMM [27] was developed as a high-order moment matching technique to align feature distributions in hidden layers, addressing the limitations of low-order alignment methods. In addition, central moment discrepancy [29] and maximum density divergence [30] are another two criterias to align feature distributions in hidden layers. Overall, these methods have contributed to the advancement of UDA by effectively reducing the domain discrepancy and enabling knowledge transfer between different domains. ### II-B Unsupervised domain adaptation for time-series data In recent years, there has been a growing trend of exploring existing UDA methods in various domains, showing their effectiveness in different applications [31, 32]. In the context of TS data, researchers have primarily focused on reducing domain discrepancy by aligning temporal features across different domains [16, 17, 15, 19, 33]. Several notable works have made significant contributions to this area. For instance, VRADA [15] leveraged a variational RNN to learn temporal features and employs adversarial-based methods to reduce the domain discrepancy. Similarly, Zhao et al. [34] introduced a UDA method combining DANN and CNN for EEG classification, which incorporated a center loss to minimize intra-class variation and maximize inter-class distance. CoDATS [35] and CLADA [36], as multi-source UDA methods for TS data, applied DANN and 1D-CNN to handle the scenarios where multiple source domains are available. Moreover, some researchers have specifically designed UDA methods tailored to TS data. ADATIME [19] considered the temporal properties of TS data and evaluated various temporal encoders for TS UDA. Besides, AdvSKM [37] mapped features into a high-dimensional space by proposing adversarial Spectral Kernels and then aligned them using MMD. To consider the invariant associative structure between variables exhibited by MTS data, SASA [16] aligned this structural information for TS UDA. These works are great pioneers for TS UDA, but they may not perform well for the scenarios requiring multiple sensors, i.e., MTS data. In MTS data, signals from different sensors often follow various distributions due to their varied physical locations and measurements. Existing TS UDA methods can be adapted to MTS data by treating all sensors as a whole [38, 19, 15, 34], facilitate the learning of global features for alignment, but they are limited in accounting for the distinct distributions of sensors within MTS data. Meanwhile, they ignored the multi-source nature of MTS data, restricting their ability to model and transfer spatial-temporal dependency information across domains. To address these limitations, we propose our SEA and SEA++ that accounts for aligning sensor-level information and modelling dependency information within MTS data to achieve better MTS-UDA. Figure 2: The overall structure. (1) To transfer spatial-temporal dependencies across domains, source and target domains share the same graph-based encoder, including MSGC, GNN, and LSTM. Each sample is segmented as multiple patches, which are constructed as sequential graphs. With the combination of GNN and LSTM, decent sensor information is learned, including sensor features and correlations. (2) In addition to supervised learning on the source domain, endo-feature alignment and exo-feature alignment are designed to reduce domain discrepancy at the local and global sensor levels. The endo-feature alignment includes aligning sensor features and their correlations between domains by SCA and SFA, both of which are enhanced by incorporating multi-graph alignment and high-order statistic alignment. Exo-feature alignment aligns global features mapped by sensor features. ## III SEA Model ### III-A Problem Definition For MTS-UDA, we are given a source domain with $N_{s}$ labeled samples $\mathcal{D}_{S}=\\{(x_{i}^{s},y_{i}^{s})\\}_{i=1}^{N_{s}}$ and a target domain with $N_{t}$ unlabeled samples $\mathcal{D}_{T}=\\{x_{i}^{t}\\}_{i=1}^{N_{t}}$. Each MTS sample $x_{i}$ (either $x_{i}^{s}$ or $x_{i}^{t}$) originates from $N$ sensors with different distributions, i.e., $x_{i}=\\{x_{im}\\}_{m=1}^{N}\in\mathbb{R}^{N\times{L}}$, where $L$ represents the time length. The goal of SEA is to train an encoder by transferring the knowledge from the source domain to the target domain, enabling the effective learning of features $h_{i}\in\mathbb{R}^{F}$ from the MTS data $x_{i}^{t}$ in the target domain. The features can then be used for downstream tasks such as RUL prediction and HAR. Notably, the SEA framework consists of two stages. In the first stage, we focus on feature extraction from MTS data. To simplify the notation and enhance clarity, we omit the index $i$ in this stage. Thus, each sample is represented as $x^{s}$ and $x^{t}$, and the data from the $m$-th sensor is denoted as $x_{m}^{s}$ and $x_{m}^{t}$. While in the second stage that involves domain alignment, we retain the use of the index to facilitate a clear description of the alignment process. Furthermore, to model the evolving distributions within MTS data, we construct sequential graphs $\\{\mathcal{G}_{T}\\}_{T=1}^{\hat{L}},\mathcal{G}_{T}=(Z_{T},E_{T})$ from the sample $x$, where $Z_{T}=\\{z_{m,T}\\}_{m=1}^{N}$ and $E_{T}=\\{e_{mn,T}\\}_{m,n=1}^{N}$ represent the sensor features and correlations, respectively, in the $T$-th graph. $\\{\mathcal{G}^{s}_{T}\\}_{T=1}^{\hat{L}}$ and $\\{\mathcal{G}^{t}_{T}\\}_{T=1}^{\hat{L}}$ represent the sequential graphs in the source and target domains respectively. ### III-B Preliminary #### III-B1 Graph Neural Network Several previous studies have demonstrated the effectiveness of Graph Neural Network (GNN) in capturing spatial dependencies [39, 40]. Motivated by this capability, we employ GNN to capture the spatial dependencies, i.e., sensor correlations, within our constructed sequential graphs. Specifically, we utilize a variant of GNN known as Message Passing Neural Network (MPNN) [41] to process each graph. The MPNN framework consists of two stages: propagation and updating. During the propagation stage, the features of neighboring nodes are propagated to the central node. Subsequently, the aggregated features are updated using a non-linear function, such as a Multi-Layer Perceptron (MLP) network. With MPNN, we can capture the spatial dependencies between sensors, enabling the effective transfer of the dependency information across domains in the subsequent alignment processes. The details of MPNN are shown in Eq. (1), where sensor $m$ is the central node, and $\mathcal{N}(m)$ is the set of its neighboring nodes. $\begin{split}{h_{m,T}}&=\sum_{j\in\mathcal{N}(m)}{e_{mj,T}}{z_{j,T}},\\\ {z_{m,T}}&=ReLU({h_{m,T}}W_{G}).\end{split}$ (1) #### III-B2 Deep Coral Deep Coral [13] is one of the most typical metric-based methods for UDA. Different from $\mathcal{L}_{2}$ or cosine distances which mainly employ first-order statistics, Deep Coral focuses on minimizing the discrepancy in second-order statistics between domains [42], making it possible to capture complex data distributions for alignment. Given training batches with $n^{s}$ and $n^{t}$ samples in source and target domains respectively, we denote the extracted features from these domains as $\mathcal{H}^{s}\in\mathbb{R}^{n^{s}\times{f}}$ and $\mathcal{H}^{t}\in\mathbb{R}^{n^{t}\times{f}}$, where $f$ represents the feature dimension. With the features, we can compute the discrepancy of second-order statistics with Deep Coral $M_{c}(\mathcal{H}^{s},\mathcal{H}^{t})$ as Eq. (2), where $||\cdot||_{F}^{2}$ is the squared matrix Frobenius norm. By minimizing the Deep Coral discrepancy $\mathcal{L}_{Coral}$, the second-order statistics of the features are aligned between the source and target domains, leading to comprehensive alignment across domains. $\begin{split}&C^{s}=\frac{(\mathcal{H}^{s})^{T}\mathcal{H}^{s}-\frac{1}{n^{s}}(\textbf{1}^{T}\mathcal{H}^{s})(\textbf{1}^{T}\mathcal{H}^{s})}{n^{s}-1},\\\ &C^{t}=\frac{(\mathcal{H}^{t})^{T}\mathcal{H}^{t}-\frac{1}{n^{t}}(\textbf{1}^{T}\mathcal{H}^{t})(\textbf{1}^{T}\mathcal{H}^{t})}{n^{t}-1},\\\ &M_{c}(\mathcal{H}^{s},\mathcal{H}^{t})=\frac{1}{4f^{2}}||C^{s}-C^{t}||_{F}^{2}.\end{split}$ (2) ### III-C Overall Structure The overall structure of SEA is shown in Fig. 2. The superscripts $s$ and $t$ are omitted in Section III-C and III-D, as the forward processes in source and target domains share the same graph-based encoder, including Multi-branch Self-attention based Graph Construction (MSGC), GNN, and LSTM. First, we model the spatial-temporal dependencies within MTS data for effective transfer across domains. Given the sample $x$, we segment it into multiple patches, which are constructed as sequential graphs by MSGC. Then, with the sequential graphs, GNN and LSTM are leveraged to capture the spatial-temporal information for learning sensor-level information, i.e., sensor features and correlations. Second, in addition to supervised learning on the source domain, we incorporate endo-feature alignment and exo-feature alignment techniques to reduce domain discrepancy at both the local and global sensor levels. At the local sensor level, we design endo-feature alignment, which involves aligning sensor features and sensor correlations by proposed Sensor Feature Alignment (SFA) and Sensor Correlation Alignment (SCA) respectively. The two components are further improved by both the high-order statistic alignment and multi- graph alignment to form iSFA and iSCA. The high-order statistic alignment aims to achieve comprehensive alignment by capturing complex data distributions while the multi-graph alignment aims to align evolving distributions at the sensor level across domains. At the global sensor level, we design exo-feature alignment to align global features by enforcing restrictions. We introduce the details of each module in subsequent sections. ### III-D Graph-based Encoder #### III-D1 Segmentation To capture the evolving distributions within MTS data, we employ a segmentation approach. In this approach, a sample $x\in\mathbb{R}^{N\times{L}}$ is segmented into multiple patches, each with a size of $d$, yielding sequential features denoted as $\\{Z_{T}\\}_{T=1}^{\hat{L}}\in\mathbb{R}^{N\times{\hat{L}\times{d}}}$. Here, $\hat{L}=\lfloor{\frac{L}{d}}\rfloor$ represents the number of patches, and $d$ represents the feature dimension. The features within the $T$-th patch $Z_{T}\in\mathbb{R}^{N\times{d}}$ capture the distributions in local temporal intervals. Within each patch, the features of the $m$-th sensor are denoted as $z_{m,T}\in\mathbb{R}^{d}$. Additionally, we incorporate a nonlinear function $f_{L}(\cdot)$ to enhance the nonlinear expressiveness of the model. #### III-D2 MSGC We construct sequential graphs with the sequential features. Traditionally, various methods have been employed to construct graphs for MTS data, such as dot-product[43], Euclidean distance[40], and concatenation [44]. However, the traditional methods are insufficient to model comprehensive relationships of sensors due to the complex topological structures presented within MTS data. To solve the limitation, we propose MSGC, which constructs a graph by combining multiple graphs generated from various weight initializations across multiple branches. Each branch employs differently weighted matrices to model sensor correlations from distinct perspectives, allowing us to capture comprehensive sensor correlations. Specifically, we adopt query and key to learn the edge weights between sensors. Eq. (3) describes the edge between sensors $m$ and $n$ in the $i$-th branch. $\begin{split}q_{m,T}^{i}&=z_{m,T}W_{Q}^{i},k_{n,T}^{i}=z_{n,T}W_{K}^{i},\\\ e_{mn,T}^{i}&=\frac{\frac{q_{m,T}^{i}(k_{n,T}^{i})^{T}}{\sqrt{d}}}{\sum_{j=1}^{N}\frac{q_{m,T}^{i}(k_{j,T}^{i})^{T}}{\sqrt{d}}},\\\ \end{split}$ (3) where we treat the $m$-th sensor $z_{m,T}$ as the query sensor and the $n$-th sensor $z_{n,T}$ as the key sensor, and use this information to learn the edge weight between them for the $i$-th branch, i.e., $e_{mn,T}^{i}$. The edge weights are then normalized using the softmax function to ensure they fall within the range of [0, 1]. To combine the graphs from multiple branches, we calculate the average of the corresponding elements, i.e., $e_{mn,T}=\sum_{i=1}^{n_{b}}\frac{e_{mn,T}^{i}}{n_{b}}$, where $n_{b}$ represents the number of branches. This operation allows us to obtain the sensor correlations $E_{T}=\\{e_{mn,T}\\}_{m,n=1}^{N}\in\mathbb{R}^{N\times{N}}$ for the $T$-th graph. By conducting similar operations in other patches, we can finally obtain the sequential correlations $\\{E_{T}\\}_{T=1}^{\hat{L}}\in\mathbb{R}^{\hat{L}\times{N}\times{N}}$. Figure 3: Left: Sensor correlation alignment is to make the correlations between domains identical. Middle: Sensor feature alignment is to make the features from the corresponding sensors in two domains similar and the data from the different sensors in two domains different. Right: Exo-feature alignment reduces the discrepancy between the global features in different domains. #### III-D3 Capturing Spatial-Temporal Information With the constructed sequential graphs, we employ two models, MPNN and LSTM, to capture the spatial-temporal dependencies within MTS data. First, we utilize MPNN (Eq. (1)) to capture the sensor correlations within each sequential graph. In this process, the graphs are processed with shared weights to save training costs and trainable parameters. Once the features of each sensor in the sequential graphs have been updated using MPNN, our focus shifts to capture temporal dependencies. It is noted that features from corresponding sensors across the sequential graphs show temporal dependencies. To capture these dependencies, we employ LSTM, which captures the dependency information for each sensor by taking its features $z_{m}\in\mathbb{R}^{{\hat{L}}\times{d}}$ across sequential graphs. After GNN and LSTM, we finally obtain updated sequential features $\\{Z_{T}\\}_{T=1}^{\hat{L}}\in\mathbb{R}^{\hat{L}\times{N}\times{d}}$. By combining the capabilities of the two networks, we can effectively capture the spatial-temporal dependencies within MTS data, allowing us to transfer the dependency information across domains through the alignment at the local and global sensor levels in the subsequent stages. ### III-E SEnsor Alignment To reduce the discrepancy between domains for MTS-UDA, we design endo-feature alignment and exo-feature alignment at the local and global sensor levels respectively. #### III-E1 Endo-feature alignment Existing UDA methods reduce domain discrepancy by aligning global features across domains, which is applicable for the data coming from a single source. However, MTS data originate from multiple sensors, with each sensor following various distributions. When adapting existing UDA methods to MTS-UDA, they can consider global distributions by treating all sensors as a whole yet ignore the distinct distributions of each sensor. This results in the misalignment at the sensor level, restricting the performance of the learned model transferring from the source domain to the target domain. To address this limitation, we propose an endo-feature alignment method designed to align sensor-level information and avoid misalignment for each sensor when transferring knowledge across domains. To make better endo-feature alignment, it is necessary to note that the sensor-level information consists of both the sensor features and sensor correlations. Sensor features represent the properties of each sensor, while sensor correlations represent the interactive information between sensors. For illustration, we present a scenario with temperature and fan speed sensors deployed together to monitor the status of a machine. It is desirable for the data from temperature sensors in two domains to follow similar distributions. Additionally, the correlation between the sensors should also demonstrate consistent patterns across domains, such as the fan speed consistently increasing with temperature rises. Based on these considerations, aligning sensor features and sensor correlations across domains simultaneously is important to ensure comprehensive alignment at the sensor level. ##### Sensor Correlation Alignment To explore consistent patterns within the interactions of sensors across domains, we propose to align the sensor correlations. Specifically, our aim is to align the distribution of correlations between any two sensors across domains. To achieve this, we propose sensor correlation alignment (SCA), which involves individually aligning each edge of the graphs as shown in Fig. 3 Left. Given training batches with $n^{s}$ and $n^{t}$ samples in source and target domains, we represent sensor correlations across sequnetial graphs as $\mathcal{E}^{s}\in\mathbb{R}^{n^{s}\times{\hat{L}}\times{N}\times{N}}$ and $\mathcal{E}^{t}\in\mathbb{R}^{n^{t}\times{\hat{L}}\times{N}\times{N}}$ respectively, with $\mathcal{E}_{T}^{s}\in\mathbb{R}^{n^{s}\times{N}\times{N}}$ and $\mathcal{E}_{T}^{t}\in\mathbb{R}^{n^{t}\times{N}\times{N}}$ signifying sensor correlations for the $T$-th graph. Here, $\mathcal{E}_{T}^{s}=\\{\mathcal{E}_{mn,T}^{s}\\}_{m,n=1}^{N}$, wherein $\mathcal{E}_{mn,T}^{s}=\\{e^{s}_{i,mn,T}\\}_{i}^{n^{s}}$. For SCA, alignment of sensor correlations across domains is achieved based on averaged sensor correlations. Specifically, for the edge connecting sensors $m$ and $n$, its mean value across the training batch is computed as $e^{s}_{mn,T}=\sum_{i}^{n^{s}}e^{s}_{i,mn,T}/n^{s}$. Subsequently, the averaged sensor correlations over sequential graphs are calculated as $e^{s}_{mn}=\sum_{T}^{\hat{L}}e^{s}_{mn,T}/{\hat{L}}$. Next, we minimize the averaged sensor correlations across domains as shown in Eq. (4), where $|\cdot|$ is absolute value. This involves minimizing the expected discrepancy of sensor correlations between the source and target domains. By doing so, we can reduce distribution discrepancies for sensor correlations in source and target domains. $\begin{split}e_{mn}^{st}&=e_{mn}^{s}-e_{mn}^{t},\\\ \mathcal{L}_{SCA}&=\mathbb{E}(|e_{mn}^{st}|).\end{split}$ (4) ##### Sensor Feature Alignment Next, we focus on aligning sensor features between domains. Our assumption is that features originating from corresponding sensors in two domains should exhibit similar distributions. Meanwhile, compared with features from corresponding sensors, features from different sensors in two domains should have distinct properties. For example, as shown in Fig. 3 Middle, the distribution of sensor 1 in the source domain should be more similar to the distribution of its corresponding sensor in the target domain than to the distributions of other sensors (e.g., sensors 2, 3, and 4) in the target domain. To achieve this goal, we propose a sensor contrasting mechanism that facilitates the alignment of sensor features between domains. Given training batches with $n^{s}$ and $n^{t}$ samples, we define sensor features as $\mathcal{Z}^{s}\in\mathbb{R}^{n^{s}\times{\hat{L}}\times{N}\times{d}}$ and $\mathcal{Z}^{t}\in\mathbb{R}^{n^{t}\times{\hat{L}}\times{N}\times{d}}$, with $\mathcal{Z}_{T}^{s}\in\mathbb{R}^{n^{s}\times{N}\times{d}}$ and $\mathcal{Z}_{T}^{t}\in\mathbb{R}^{n^{t}\times{N}\times{d}}$ denoting sensor features for the $T$-th graph. Here, $\mathcal{Z}_{T}^{s}=\\{\mathcal{Z}_{n,T}^{s}\\}_{n=1}^{N}$, wherein $\mathcal{Z}_{n,T}^{s}=\\{z^{s}_{i,n,T}\\}_{i}^{n^{s}}$. In SFA, domain contrasting is realized based on averaged sensor features. In Eq. (5), $p^{s}_{m}$ represents the averaged sensor features over multiple graphs within a training batch, calulated as $p^{s}_{m}=\sum_{T}^{\hat{L}}\sum_{i}^{n^{s}}z^{s}_{i,n,T}/n^{s}\hat{L}$. $\begin{split}\mathcal{L}_{SFA}&=-\frac{1}{N}\sum_{n}^{N}log\frac{e^{\varphi(p_{m}^{s},p_{m}^{t})}}{\sum_{p_{j}^{t}\in{P^{t}}}e^{\varphi(p_{m}^{s},p_{j}^{t})}},\\\ \varphi(p_{m}^{s},p_{m}^{t})&=p_{m}^{s}(p_{m}^{t})^{T}.\end{split}$ (5) ##### Exo-feature alignment. To reduce the domain discrepancy at the global level, we enforce restrictions on the global features mapped by sensor features between domains. Specifically, we first need to learn the global features of each sample. This involves stacking all sensor features across the sequential graphs and subsequently employing an MLP network to map the stacked features as global features $\mathcal{P}^{s}$ and $\mathcal{P}^{t}$ in source and target domains, as shown in Fig. 3 Right. With the global features, we can then align them with Deep Coral, expressed as $\mathcal{L}=\mathcal{M}_{c}(\mathcal{P}^{s},\mathcal{P}^{t})$. Finally, the overall loss function (6) is minimized: $\begin{split}\min\mathcal{L}=\mathcal{L}_{C}+\mathcal{L}_{EXO}+\mathcal{L}_{Endo},\\\ \mathcal{L}_{Endo}=\lambda_{SCA}\mathcal{L}_{SCA}+\lambda_{SFA}\mathcal{L}_{SFA}.\end{split}$ (6) where $\mathcal{L}_{C}$ is the cost function (such as mean square error loss or cross-entropy loss, depending on specific tasks) computed by the source data and labels, and $\lambda_{SCA}$ and $\lambda_{SFA}$ are hyperparameters to balance the effect of SCA and SFA respectively. ## IV SEA++ Model While the method described above offers a viable approach for achieving sensor correlation and sensor feature alignment, there is still room for improvement to enhance endo-feature alignment. Two key directions warrant consideration. First, SEA employs first-order statistics for alignment, such as averaged sensor features over a training batch used in the sensor contrastive mechanism. However, the first-order statistical alignment method is limited in dealing with scenarios with complex data distributions. This limitation may restrict the ability to achieve comprehensive alignment across domains, leading to suboptimal performance in certain scenarios. To address that, we propose introducing second-order statistical metrics for alignment to achieve more comprehensive endo-feature alignment. Second, the constructed sequential graphs, comprising sequential features and correlations, capture evolving distributions within local temporal intervals and should be aligned across domains. We consider the example of monitoring machine health status for illustration. As the machine’s health deteriorates, the distribution of a fan speed sensor may change due to increased friction or other factors. Similarly, the interaction between the fan speed sensor and a temperature sensor will also be affected and undergo changes. Specifically, the same fan speeds might lead to higher temperature increases due to a machine in poor health, e.g., friction increasing. Thus, we emphasize the importance of aligning the evolving distributions at the sensor level between domains. However, SEA employs averaged sensor correlations and sensor features over the sequential graphs, limiting its ability to capture the evolving distributions for alignment. To improve this, we introduce a multi-graph alignment technique, aiming to effectively align evolving distributions between domains. Next, we begin by discussing SCA and SFA enhanced by high-order statistical alignment within each sequential graph. Subsequently, we focus on combining the alignment across all sequential graphs by employing our multi-graph alignment. ### IV-A Sensor Correlation Alignment Given training batches with $n^{s}$ and $n^{t}$ samples in source and target domains, we have sensor correlations for the $T$-th graph, denoted as $\mathcal{E}_{T}^{s}\in\mathbb{R}^{n^{s}\times{N}\times{N}}$ and $\mathcal{E}_{T}^{t}\in\mathbb{R}^{n^{t}\times{N}\times{N}}$ respectively. Instead of leveraging averaged sensor correlations within batches, we introduce second-order statistics with Deep Coral, i.e., $M_{c}$ in Eq. (2), to align the sensor correlations between the source and target domains. Building upon this idea, we enhance SCA and introduce improved SCA (iSCA) as shown in Eq. (7). $\begin{split}d_{mn,T}^{st}&=M_{c}(\mathcal{E}_{mn,T}^{s},\mathcal{E}_{mn,T}^{t}),\\\ \mathcal{L}_{iSCA,T}&=\mathbb{E}(d_{mn,T}^{st})=\frac{\sum_{m}^{N}\sum_{n}^{N}d_{mn,T}^{st}}{mn}.\end{split}$ (7) By aligning each edge between domains with the second-order statistical metric and minimizing the distribution discrepancy of sensor correlations, we can achieve more comprehensive alignment for sensor correlations. ### IV-B Sensor Feature Alignment Next, we focus on improving sensor feature alignment. Similar to SCA, the sensor contrasting mechanism in SFA is realized with the averaged sensor features over the training batches in source and target domains. Thus, SFA encounters a challenge similar to SCA. Specifically, the contrasting-based alignment procedure in Eq. (5) relies on first-order statistics, which might not achieve comprehensive alignment when dealing with complex distributions. To overcome this limitation, we enhance SFA and introduce improved SFA (iSFA) by incorporating Deep Coral into the contrasting process. In Eq. (8), iSFA begins by calculating sensor-level distribution discrepancies using Deep Coral, which takes into account second-order statistics. Subsequently, iSFA leverages the sensor-level discrepancies to perform contrasting between domains. This involves minimizing the distribution discrepancies between corresponding sensors across domains while maximizing the distribution discrepancies between different sensors across domains. This approach allows for more comprehensive alignment of sensor features. $\begin{split}\mathcal{L}_{iSFA,T}=-\frac{1}{N}\sum_{n}^{N}log\frac{e^{M_{c}(\mathcal{Z}_{n,T}^{s},\mathcal{Z}_{n,T}^{t}})}{\sum^{N}_{j}e^{M_{c}(\mathcal{Z}_{n,T}^{s},\mathcal{Z}_{j,T}^{t})}}.\end{split}$ (8) ### IV-C Multi-Graph Alignment To combine the alignment for sensor correlations and sensor features across all sequential graph, a straightforward approach would involve computing the average alignment for these graphs. However, the distributions between domains for different graphs might exhibit various discrepancies. To account for this and enhance domain adaptation, we propose a method to learn adaptive weights for adjustment, as shown in Eq. (9). $\begin{split}\mathcal{L}_{iEndo}&=\sum_{T}\mathcal{W}_{T}(\lambda_{SCA}\mathcal{L}_{iSCA,T}+\lambda_{SFA}\mathcal{L}_{iSFA,T}),\\\ \mathcal{W}_{T}&=M_{c}([\mathcal{Z}_{1,T}^{s},...,\mathcal{Z}_{\hat{L},T}^{s}],[\mathcal{Z}_{1,T}^{t},...,\mathcal{Z}_{\hat{L},T}^{t}]),\end{split}$ (9) where $[a,...,b]$ represents concatenating the elements from $a$ to $b$. In Eq. (9), the learnable weights $\mathcal{W}_{T}$ are determined by evaluating the distribution discrepancy of the $T$-th graph between domains. Specifically, $\mathcal{W}_{T}$ will be large when there is a substantial discrepancy in the distributions of the $T$-th graph between the source and target domains. By assigning larger weights to the graphs with larger distribution discrepancies between domains, we prioritize alignment in those graphs. This adaptive weighting scheme allows us to better adapt the model to the evolving distributions between domains. Furthermore, we also introduce $\lambda_{SCA}$ and $\lambda_{SFA}$ to balance the effect of iSCA and iSFA. By combining the exo-feature alignment and the improved endo-feature alignment, the improved overall loss function (10) can be obtained as: $\begin{split}\min\mathcal{L}&=\mathcal{L}_{C}+\mathcal{L}_{EXO}+\mathcal{L}_{iEndo}.\end{split}$ (10) ## V Experiment ### V-A Datasets and Setup To evaluate the performance of SEA and SEA++ under various MTS scenarios, we conducted experiments on two public datasets: C-MAPSS for remaining useful life prediction [45], which collects MTS data from different sensor types, and Opportunity HAR for human activity recognition [46], which collects MTS data from different locations. C-MAPSS describes the degradation process of aircraft engines. The MTS data in C-MAPSS were originated from 14 sensors to measure various physical parameters, such as temperature, fan speed, and pressure. The dataset includes four sub-datasets collected under different working conditions and fault modes, where each sub-dataset represents one domain. We processed the sub- datasets following the data preparation in the previous work [47], and data annotations represent the remaining useful life circle of engines. Opportunity HAR describes human activities. The MTS data in Opportunity HAR were originated from 113 sensors deployed to various locations of the human body, such as hand and leg. After removing the sensors with constant values, 110 sensors were used for evaluation. The dataset includes the data collected from four subjects, each representing one domain. The data annotations include two levels: 1). locomotion representing the low-level tasks including four categories, sitting, standing, walking, and lying down; 2). gestures representing the high-level tasks including 17 various actions. Following the experimental settings in the previous work [48], we adopted the low-level tasks. As some values in the data were missing, we adopted the linear interpolation approach to fill in the missing positions. To construct the training dataset, we adopted a sliding window with a size of 128 and an overlapping of 50% following the previous work [48]. The experiments include three parts, comparisons with state-of-the-art methods, the ablation study, and sensitivity analysis. All experiments were run ten times with average results reported to remove the effect of random initialization. Besides, we set the batch size as 50 and employed the optimizer as Adam with a learning rate of 0.001. Meanwhile, we adopted 20 training epochs for training our model. Furthermore, we built and trained our model based on Pytorch 1.9 and NVIDIA GeForce RTX 3080Ti GPU. We employed different evaluation indicators to assess the performance of our method on the two distinct datasets. For C-MAPSS experiments, which involve predicting the RUL of an engine, we utilized Root Mean Square Error (RMSE) and the Score function [47, 49, 50, 51] as evaluation metrics. Lower values for both indicators indicate better model performance. For Opportunity HAR experiments, which entail a classification task, we employed accuracy as the evaluation metric. High accuracy values signify superior model performance. ### V-B Comparisons with State-of-the-Art TABLE I: The Comparisons with SOTAs in C-MAPSS (R: RMSE; S: Score) Models (R) | 1$\to$2 | 1$\to$3 | 1$\to$4 | 2$\to$1 | 2$\to$3 | 2$\to$4 | 3$\to$1 | 3$\to$2 | 3$\to$4 | 4$\to$1 | 4$\to$2 | 4$\to$3 | Avg. ---|---|---|---|---|---|---|---|---|---|---|---|---|--- Target | 12.93 | 13.33 | 14.66 | 12.36 | 13.33 | 14.66 | 12.36 | 12.93 | 14.66 | 12.36 | 12.93 | 13.33 | 13.32 Source | 28.76 | 35.28 | 27.21 | 21.86 | 39.80 | 33.50 | 37.90 | 30.33 | 24.10 | 33.76 | 25.31 | 21.75 | 29.96 DDC | 43.25 | 39.48 | 42.99 | 40.07 | 39.46 | 43.01 | 40.93 | 43.43 | 43.72 | 41.58 | 43.38 | 39.61 | 41.74 Coral | 16.72 | 26.49 | 25.03 | 14.47 | 37.41 | 31.13 | 32.82 | 23.40 | 19.84 | 28.37 | 32.00 | 22.47 | 25.85 DANN | 17.02 | 23.33 | 24.20 | 14.16 | 30.00 | 25.91 | 19.89 | 20.31 | 19.95 | 31.74 | 24.68 | 17.94 | 22.43 BNM | 48.18 | 41.61 | 46.45 | 56.66 | 48.40 | 62.02 | 49.39 | 48.89 | 42.98 | 40.76 | 43.71 | 41.89 | 47.58 SDAT | 16.50 | 24.11 | 23.07 | 13.57 | 29.76 | 26.38 | 25.34 | 20.29 | 18.63 | 20.13 | 23.23 | 17.65 | 21.55 AdvSKM | 31.09 | 34.16 | 40.74 | 23.93 | 32.41 | 35.73 | 28.47 | 22.17 | 26.45 | 28.56 | 22.75 | 23.51 | 29.16 CoDATs | 15.91 | 26.67 | 25.24 | 15.59 | 24.52 | 25.64 | 21.25 | 21.04 | 19.77 | 19.41 | 18.99 | 17.85 | 20.99 CtsADA | 18.31 | 29.23 | 29.49 | 23.20 | 54.69 | 41.60 | 38.14 | 24.67 | 19.86 | 31.63 | 25.71 | 23.97 | 30.04 CLUDA | 25.31 | 36.80 | 33.24 | 22.51 | 33.82 | 34.63 | 26.48 | 24.09 | 26.29 | 24.33 | 25.14 | 22.18 | 27.90 SEA | 15.63 | 21.09 | 23.21 | 13.29 | 24.62 | 25.32 | 19.69 | 19.44 | 19.37 | 18.92 | 17.70 | 16.45 | 19.56 SEA++ | 15.41 | 19.38 | 22.12 | 13.00 | 24.60 | 24.90 | 19.19 | 19.25 | 17.75 | 18.43 | 18.16 | 16.79 | 19.08 Models (S) | 1$\to$2 | 1$\to$3 | 1$\to$4 | 2$\to$1 | 2$\to$3 | 2$\to$4 | 3$\to$1 | 3$\to$2 | 3$\to$4 | 4$\to$1 | 4$\to$2 | 4$\to$3 | Avg. Target | 678 | 330 | 1027 | 241 | 330 | 1027 | 241 | 678 | 1027 | 241 | 678 | 330 | 569 Source | 9341 | 4751 | 4699 | 2340 | 8335 | 11018 | 24892 | 26878 | 11006 | 16318 | 8771 | 2047 | 10866 DDC | 23323 | 6164 | 34891 | 11581 | 6578 | 35122 | 22236 | 59400 | 80774 | 29592 | 58029 | 15239 | 31911 Coral | 1158 | 1468 | 3613 | 406 | 6241 | 8492 | 15298 | 12661 | 3048 | 6742 | 25763 | 4758 | 7471 DANN | 1073 | 1612 | 3340 | 392 | 3160 | 4667 | 1666 | 3896 | 2381 | 22225 | 4620 | 762 | 4150 BNM | 24265 | 44650 | 19017 | 526203 | 56769 | 29689 | 46262 | 29188 | 39559 | 11618 | 20566 | 65807 | 76133 SDAT | 974 | 1403 | 2729 | 300 | 4360 | 3859 | 3756 | 3134 | 2289 | 966 | 2369 | 678 | 2235 AdvSKM | 13313 | 5012 | 18444 | 2344 | 4406 | 14624 | 6601 | 4516 | 11857 | 6332 | 4474 | 5218 | 8095 CoDATs | 1043 | 2854 | 5074 | 836 | 1533 | 5094 | 1300 | 3993 | 2352 | 1065 | 2205 | 697 | 2337 CtsADA | 1467 | 1965 | 6832 | 4136 | 29153 | 22515 | 47682 | 17418 | 2138 | 13019 | 15436 | 1593 | 13613 CLUDA | 2686 | 4754 | 6337 | 873 | 3501 | 7159 | 2389 | 3232 | 3078 | 1658 | 5566 | 1093 | 3527 SEA | 922 | 1302 | 3786 | 287 | 2090 | 3782 | 1163 | 2990 | 2278 | 939 | 1930 | 635 | 1842 SEA++ | 881 | 803 | 2669 | 295 | 1510 | 3220 | 1117 | 3045 | 1558 | 859 | 2072 | 572 | 1550 TABLE II: The Comparisons with SOTAs in Opportunity HAR Models | 1$\to$2 | 1$\to$3 | 1$\to$4 | 2$\to$1 | 2$\to$3 | 2$\to$4 | 3$\to$1 | 3$\to$2 | 3$\to$4 | 4$\to$1 | 4$\to$2 | 4$\to$3 | Avg. ---|---|---|---|---|---|---|---|---|---|---|---|---|--- Target | 91.90 | 97.95 | 95.60 | 98.40 | 97.95 | 95.60 | 98.40 | 91.90 | 95.60 | 98.40 | 91.90 | 97.95 | 95.96 Source | 47.50 | 64.50 | 60.83 | 73.00 | 71.00 | 65.67 | 64.83 | 56.67 | 72.17 | 66.67 | 65.50 | 74.67 | 65.25 DDC | 40.50 | 58.50 | 55.50 | 43.35 | 58.30 | 55.50 | 45.00 | 40.50 | 55.50 | 45.00 | 40.50 | 58.50 | 49.72 Coral | 78.55 | 82.71 | 78.17 | 81.58 | 79.93 | 79.07 | 78.50 | 75.40 | 84.88 | 78.56 | 80.38 | 87.88 | 80.47 DANN | 69.62 | 85.12 | 79.00 | 85.62 | 81.62 | 74.88 | 76.50 | 75.75 | 86.12 | 84.25 | 74.12 | 86.62 | 79.94 BNM | 44.00 | 61.00 | 52.00 | 45.00 | 61.00 | 52.00 | 45.00 | 44.00 | 52.00 | 45.50 | 44.00 | 61.00 | 50.54 LMMD | 56.25 | 61.17 | 59.08 | 45.33 | 58.50 | 55.50 | 49.42 | 46.83 | 57.33 | 59.00 | 44.75 | 61.75 | 54.58 SDAT | 84.92 | 72.42 | 77.00 | 87.25 | 83.00 | 76.17 | 87.75 | 84.25 | 87.25 | 74.58 | 66.75 | 91.50 | 81.07 AdvSKM | 58.10 | 58.25 | 51.75 | 45.40 | 40.95 | 48.40 | 69.60 | 72.20 | 70.60 | 68.75 | 55.50 | 58.20 | 58.14 CoDATs | 80.17 | 83.33 | 82.33 | 88.67 | 80.50 | 81.00 | 89.25 | 79.12 | 85.75 | 84.67 | 79.17 | 86.17 | 83.34 CtsADA | 80.00 | 73.67 | 76.08 | 69.75 | 78.17 | 79.08 | 67.75 | 77.33 | 64.75 | 74.58 | 52.17 | 75.00 | 72.36 SLARDA | 80.38 | 80.50 | 79.50 | 82.75 | 78.88 | 79.25 | 75.88 | 78.62 | 70.12 | 80.62 | 84.00 | 77.25 | 78.98 CLUDA | 70.45 | 73.75 | 74.40 | 73.20 | 74.25 | 73.95 | 61.65 | 55.10 | 64.30 | 65.75 | 61.05 | 71.55 | 68.28 SEA | 86.35 | 86.15 | 86.00 | 86.35 | 84.35 | 81.75 | 87.25 | 83.05 | 88.70 | 84.90 | 84.25 | 92.25 | 85.42 SEA++ | 86.85 | 90.30 | 90.85 | 90.25 | 83.85 | 85.40 | 89.80 | 83.10 | 88.10 | 91.70 | 86.95 | 92.55 | 88.31 We compared SEA and SEA++ with State-Of-The-Art methods (SOTAs), including general UDA methods and TS UDA methods. General UDA methods were designed to be applicable to various tasks and domains, regardless of the specific data type, including DDC [9], Deep Coral [52], DANN [53], BNM [54], LMMD [55], and SDAT [56]. TS UDA methods were specifically designed for TS data, including AdvSKM [37], CoDATs [35], CtsADA [47], SLARDA [48], and CLUDA [57]. For fair comparisons, all methods were run ten times based on the same feature extractor (i.e., GNN-LSTM). LMMD and SLARDA were designed specifically for multi-class classification UDA, so they only had results on Opportunity HAR. Notably, in the previous SEA paper, different cross-domain scenarios were based on various sample lengths. For example, the cross-domain scenario 1$\to$2 was based on a time length of 72 while 3$\to$4 was based on a time length of 45. However, this training setting may be inapplicable in real-world systems where obtaining samples with a fixed time length is desired. Thus, all results in this paper, particularly in the context of the C-MAPSS dataset, have been re-conducted based on a consistent time length of 60 timestamps. Additionally, we have included the results of Target-only (Target) and Source- only (Source) baselines for comparison. The target-only baseline represents the upper bound of UDA, where training and testing were performed on the target domain only. The source-only baseline represents the lower bound of UDA, where only the source data were used for training without any UDA methods applied. TABLE I shows the RMSE and Score results for 12 cross-domain scenarios in C-MAPSS. From the results, we observe that both SEA and SEA++ achieve better performance than SOTAs in most cross-domain scenarios. Specifically, compared to the methods excluding SEA++, SEA performs the best performance in 9 out of 12 scenarios and the second-best performance in the remaining scenarios with very marginal gaps compared to the best methods. We utilize the RMSE results in C-MAPSS for illustration. In the cross-domain scenario 1$\to$3, SEA outperforms the second-best method by a significant 9.6%. In scenario 1$\to$4, SEA is the second-best method and is only 0.6% weaker than the top-performing method. These results demonstrate the effectiveness of SEA compared to SOTAs. However, SEA exhibits minor limitations that make it slightly less effective in specific cases, motivating the development of SEA++. Regarding the results of SEA++, it not only achieves better performance in scenarios where SEA performs the best but also achieves the best performance in 2 out of 3 scenarios where SEA is the second-best. For example, in the cross-domain scenario 1$\to$4 where SEA is not the top-performing, SEA++ improves by 4.1% compared to the best conventional method. Although SEA++ is slightly weak in 2$\to$3, the gap is further reduced, being only 0.3% weaker than the best method. In terms of average RMSE results, SEA outperforms the best conventional methods by 9.2%. With improved endo-feature alignment and multi-graph alignment, SEA++ further enhances the performance, surpassing SEA by 2.45% and outperforming the best conventional method by 11.4%. Similar improvements can be observed in the Score results of C-MAPSS and Opportunity HAR in TABLE II. Taking the average accuracy in Opportunity HAR for illustration, SEA improves by 2.08% compared to the best conventional method. Additionally, SEA++ outperforms SEA by 2.89% and outperforms the best conventional method by 4.97%. These results indicate the effectiveness of SEA, showing the necessity of aligning sensor-level information. Furthermore, SEA++ addresses the minor limitations of SEA, resulting in improved performance that establishes its state-of-the-art status compared to traditional methods. ### V-C Ablation Study TABLE III: The Ablation Study in C-MAPSS (R: RMSE; S: Score) Variants (R) | 1$\to$2 | 1$\to$3 | 1$\to$4 | 2$\to$1 | 2$\to$3 | 2$\to$4 | 3$\to$1 | 3$\to$2 | 3$\to$4 | 4$\to$1 | 4$\to$2 | 4$\to$3 | Avg. ---|---|---|---|---|---|---|---|---|---|---|---|---|--- w/ Exo (w/o endo) | 16.96 | 27.74 | 28.28 | 14.47 | 33.70 | 28.18 | 32.88 | 25.45 | 19.98 | 30.75 | 23.19 | 20.03 | 25.13 w/ Exo + SCA | 15.96 | 23.83 | 25.40 | 14.12 | 27.32 | 27.13 | 26.89 | 23.97 | 19.87 | 25.94 | 20.62 | 17.97 | 22.42 w/ Exo + SCA + MGA | 15.96 | 21.95 | 22.68 | 13.65 | 28.93 | 26.31 | 27.98 | 21.64 | 18.70 | 26.15 | 20.31 | 17.33 | 21.80 w/ Exo + SCA + 2-order | 15.94 | 21.61 | 23.05 | 13.32 | 28.76 | 26.94 | 23.60 | 20.73 | 18.41 | 22.40 | 19.85 | 17.25 | 20.99 w/ Exo + iSCA | 15.90 | 21.39 | 22.62 | 13.28 | 27.91 | 26.08 | 22.01 | 19.86 | 18.30 | 22.37 | 19.55 | 17.18 | 20.54 w/ Exo + SFA | 16.71 | 22.03 | 24.96 | 14.03 | 29.01 | 26.15 | 26.71 | 21.28 | 19.71 | 25.97 | 19.63 | 17.30 | 21.96 w/ Exo + SFA + MGA | 16.11 | 21.61 | 22.47 | 13.75 | 28.97 | 26.05 | 27.35 | 21.20 | 18.97 | 23.67 | 19.58 | 17.08 | 21.40 w/ Exo + SFA + 2-order | 16.15 | 22.08 | 22.57 | 13.77 | 28.51 | 26.45 | 23.99 | 20.67 | 18.19 | 22.57 | 19.87 | 17.85 | 21.06 w/ Exo + iSFA | 15.90 | 20.37 | 22.47 | 13.59 | 27.80 | 25.34 | 21.44 | 19.95 | 18.10 | 20.54 | 19.30 | 17.28 | 20.17 SEA++ | 15.41 | 19.38 | 22.12 | 13.00 | 24.60 | 24.90 | 19.19 | 19.25 | 17.75 | 18.43 | 18.16 | 16.79 | 19.08 Variants (S) | 1$\to$2 | 1$\to$3 | 1$\to$4 | 2$\to$1 | 2$\to$3 | 2$\to$4 | 3$\to$1 | 3$\to$2 | 3$\to$4 | 4$\to$1 | 4$\to$2 | 4$\to$3 | Avg. w/ Exo (w/o endo) | 1156 | 1863 | 7246 | 406 | 3705 | 5160 | 17374 | 17854 | 2240 | 10247 | 6635 | 1224 | 6259 w/ Exo + SCA | 1101 | 1537 | 6392 | 360 | 2358 | 4657 | 11756 | 7351 | 2458 | 4082 | 3925 | 735 | 3893 w/ Exo + SCA + MGA | 980 | 979 | 2936 | 356 | 2255 | 3893 | 9340 | 7284 | 1974 | 4052 | 3652 | 664 | 3197 w/ Exo + SCA + 2-order | 995 | 1070 | 2838 | 328 | 2517 | 4287 | 3489 | 5954 | 1820 | 1976 | 3207 | 696 | 2431 w/ Exo + iSCA | 978 | 1045 | 2674 | 306 | 2076 | 3704 | 2270 | 3906 | 1571 | 1815 | 3084 | 679 | 2009 w/ Exo + SFA | 1052 | 1445 | 4347 | 368 | 2152 | 3855 | 6638 | 7266 | 2111 | 4406 | 3082 | 654 | 3115 w/ Exo + SFA + MGA | 949 | 1194 | 2634 | 364 | 1873 | 3697 | 6193 | 6023 | 1915 | 2443 | 3048 | 627 | 2580 w/ Exo + SFA + 2-order | 987 | 1099 | 2674 | 367 | 2233 | 4214 | 3654 | 7640 | 1621 | 2212 | 3208 | 707 | 2551 w/ Exo + iSFA | 946 | 1083 | 2634 | 342 | 1873 | 3598 | 1902 | 4515 | 1694 | 1395 | 2814 | 630 | 1952 SEA++ | 881 | 803 | 2669 | 295 | 1510 | 3220 | 1117 | 3045 | 1558 | 859 | 2072 | 572 | 1550 * • **MGA represents Multi-Graph Alignment; 2-order represents high-order statistical alignment. TABLE IV: The Ablation Study in Opportunity HAR Variants | 1$\to$2 | 1$\to$3 | 1$\to$4 | 2$\to$1 | 2$\to$3 | 2$\to$4 | 3$\to$1 | 3$\to$2 | 3$\to$4 | 4$\to$1 | 4$\to$2 | 4$\to$3 | Avg. ---|---|---|---|---|---|---|---|---|---|---|---|---|--- w/ Exo (w/o endo) | 78.55 | 82.71 | 78.17 | 81.58 | 79.93 | 79.07 | 78.50 | 75.40 | 84.88 | 78.56 | 80.38 | 87.88 | 80.47 w/ Exo + SCA | 84.45 | 84.90 | 82.20 | 84.00 | 81.95 | 81.00 | 81.50 | 75.81 | 86.95 | 81.80 | 81.60 | 87.35 | 82.79 w/ Exo + SCA + MGA | 85.00 | 86.75 | 84.05 | 88.25 | 82.60 | 82.20 | 83.25 | 79.75 | 85.80 | 83.90 | 81.85 | 88.35 | 84.31 w/ Exo + SCA + 2-order | 85.20 | 87.40 | 84.90 | 88.35 | 82.30 | 82.00 | 81.10 | 76.75 | 87.15 | 81.35 | 84.40 | 87.30 | 84.02 w/ Exo + iSCA | 85.70 | 87.20 | 88.20 | 89.10 | 81.30 | 84.10 | 86.75 | 79.15 | 88.05 | 89.80 | 85.50 | 87.50 | 86.03 w/ Exo + SFA | 82.94 | 83.62 | 80.38 | 84.25 | 83.45 | 80.55 | 78.80 | 75.38 | 86.20 | 79.05 | 81.90 | 87.50 | 82.00 w/ Exo + SFA + MGA | 85.95 | 87.60 | 85.75 | 87.05 | 82.75 | 82.90 | 79.60 | 79.05 | 87.35 | 83.50 | 79.95 | 88.90 | 84.20 w/ Exo + SFA + 2-order | 84.50 | 87.40 | 84.25 | 87.00 | 84.90 | 82.45 | 82.15 | 80.95 | 87.75 | 85.45 | 83.60 | 88.45 | 84.90 w/ Exo + iSFA | 85.70 | 87.90 | 85.85 | 87.90 | 83.70 | 82.90 | 86.45 | 82.05 | 86.90 | 87.10 | 84.75 | 90.15 | 85.95 SEA++ | 86.85 | 90.30 | 90.85 | 90.25 | 83.85 | 85.40 | 89.80 | 81.50 | 88.10 | 91.70 | 86.95 | 92.55 | 88.18 To comprehensively evaluate the effectiveness of our endo-feature alignment and the associated improvements, we have conducted extensive ablation studies consisting of nine different variants. These variants can be categorized into three aspects. The first aspect compares our method with a baseline that only performs exo-feature alignment without utilizing any sensor-level alignment. In the second and third aspects, we consider variants that combine exo-feature alignment with one of the endo-feature alignment techniques: SCA or SFA. Each aspect further includes four variants based on different improvements. We utilize the variants with SCA for illustration. The “w/ Exo + SCA” variant represents the use of the vanilla SCA for sensor correlation alignment. By introducing Multi-Graph Alignment (MGA), we have the “w/ Exo + SCA + MGA” variant, while still using alignment with first-order statistics. By introducing alignment with second-order statistics without MGA, we have the variant “w/ Exo + SCA + 2-order”. Finally, by incorporating both improvements, i.e., MGA and second-order statistic alignment, we obtain the “w/ Exo + iSCA” variant. The same variations were applied for SFA, resulting in four additional variants. TABLE III and TABLE IV show the ablation study on C-MAPSS and Opportunity HAR respectively. Focusing on the average RMSE results on C-MAPSS for illustration, we consider the variant “w/ Exo” as the baseline. We can first observe that introducing SCA and SFA leads to improvements of 10.8% and 12.6%, respectively, which indicates the effectiveness of endo-feature alignment. As the endo-feature alignment still has minor limitations, we further introduce MGA and second-order statistic alignment to improve the sensor-level alignment. Introducing MGA leads to performance improvements of 13.3% (SCA) and 14.9% (SFA), while introducing second-order statistic alignment improves it by 16.5% (SCA) and 16.2% (SFA). The improvements indicate that MGA and second-order statistic alignment both show effectiveness for better sensor- level alignment. Thus, it is useful to combine them to further improve the performance of our method. Combining both parts, the variants “w/ iSCA” and “w/ iSFA” improve the performance by 18.3% and 19.7% respectively compared to the baseline. Finally, both the sensor features and sensor correlations can be aligned simultaneously. Thus, by combining iSCA and iSFA for endo-feature alignment, we achieve a remarkable improvement of 24.1% compared to the baseline. Similar trends are also observed in the experiments conducted on Opportunity HAR, as shown in TABLE IV. For example, the average accuracy of SEA++ increases by 7.61% compared to the model with exo-feature alignment only. The results in the cross-domain scenarios of Opportunity HAR can further verify the effectiveness of our proposed modules. These experiments highlight the significance of endo-feature alignment in achieving effective alignment at the sensor level for MTS-UDA. Furthermore, the additional modules for endo-feature alignment, namely MGA and high-order statistics alignment, prove to be highly effective in further improving our model’s performance. ### V-D Sensitivity Analysis Our method includes hyperparameters that require sensitivity analysis to evaluate their impact on the method’s performance. We focus on four hyperparameters: $\lambda_{SCA}$, $\lambda_{SFA}$, number of heads, and patch sizes. The first two hyperparameters control the balance between sensor correlation alignment and sensor feature alignment. The number of heads determines the number of graph branches used to model spatial dependencies between sensors. The patch sizes determine the number of sequential graphs constructed. Conducting sensitivity analysis on these hyperparameters helps determine their sensitivity and find optimal values. Notably, the sensitivity analysis presented here pertains specifically to SEA++. Additional details regarding the analysis related to SEA can be found in our appendix. #### V-D1 Analysis for $\lambda_{SCA}$ and $\lambda_{SFA}$ Figure 4: The sensitivity analysis for $\lambda_{SCA}$ on C-MAPSS. Figure 5: The sensitivity analysis for $\lambda_{SFA}$ on C-MAPSS. Figure 6: The sensitivity analysis for $\lambda_{SCA}$ and $\lambda_{SFA}$ on Opportunity HAR. We have conducted experiments with various values of $\lambda_{SCA}$ and $\lambda_{SFA}$ varying from 0.001 to 10 with an interval of $10\times$. Fig. 4 and 5 present the analysis of $\lambda_{SCA}$ and $\lambda_{SFA}$ on C-MAPSS respectively, and Fig. 6 shows the analysis for $\lambda_{SCA}$ and $\lambda_{SFA}$ on Opportunity HAR. From the results, we can find similar trends for the two hyperparameters, which include two points. Firstly, our method exhibits insensitivity to changes in the hyperparameters in most cases. For example, our method is insensitive to $\lambda_{SCA}$ in cross-domain scenarios such as 1$\to$4, 2$\to$1, and 2$\to$4 of C-MAPSS, and 1$\to$3, 2$\to$3, and 4$\to$2 of Opportunity HAR. Similarly, regarding $\lambda_{SFA}$, we can also find the insensitivity in the scenarios 1$\to$2, 2$\to$1, and 3$\to$4 of C-MAPSS, and 1$\to$4, 3$\to$2, and 4$\to$3 of Opportunity HAR. These results indicate that our method can achieve stable performance without requiring extensive hyperparameter tuning, making it applicable to real-world scenarios. Second, in the cases where our method is sensitive to the hyperparameters, our method tends to achieve good performance when their values are within the range of [0.001,1]. Taking the results on C-MAPSS as examples, we observe that our method achieves poor performance when $\lambda_{SCA}$ is set to 10 in the case of 3$\to$2, and when $\lambda_{SCA}$ is set to 0.001 in the case of 4$\to$1\. Similar trends can also be observed for $\lambda_{SFA}$, e.g., $\lambda_{SFA}$=10 in 2$\to$3\. Based on these observations, we recommend setting the values of these hyperparameters within the range of 0.01 to 1, where our method consistently achieves good performance. #### V-D2 Analysis for the number of heads Figure 7: The sensitivity analysis for different heads on C-MAPSS. Figure 8: The sensitivity analysis for different heads on Opportunity HAR. For the construction of each graph with MSGC, we have multiple branches with differently initialized weights to model various aspects of sensor correlations. In this part, we aim to evaluate the effect of the number of branches on the model performance. Fig. 7 and 8 show the results on C-MAPSS and Opportunity HAR respectively. In our analysis, we analyze the effect using a maximum of 4 branches, as we observe that further increasing the number of branches does not yield significant improvements. The results consistently show that SEA++ obtains performance improvements with increasing the number of branches. Specifically, SEA++ with just one branch tends to exhibit poorer performance in most cases, such as in the 2$\to$3 cross-domain scenario of C-MAPSS and the 2$\to$4 scenario of Opportunity HAR. With increasing the number of branches, the performance improves. It is worth noting that the improvements become less pronounced when the number of branches is large enough, particularly when transitioning from 3 to 4 branches. This suggests that using 3 branches is generally sufficient for effectively capturing the spatial dependencies between sensors. Overall, the analysis demonstrates the importance of leveraging multiple branches in MSGC for modeling the complex relationships among sensors. Increasing the number of branches allows our method to capture a more comprehensive understanding of sensor correlations, leading to improved performance for MTS-UDA. #### V-D3 Analysis for patch size Figure 9: The sensitivity analysis for patch sizes on C-MAPSS. Figure 10: The sensitivity analysis for patch sizes on Opportunity HAR. To align the evolving distributions within MTS data, we segment an MTS sample into multiple patches with a size of $d$, which are then utilized to construct sequential graphs and conduct multi-graph alignment. Evaluating the effect of patch size $d$ is important, as the size affects the construction of sequential graphs. Notably, we conducted the analysis with different patch sizes for C-MAPSS and Opportunity HAR as their samples have different lengths. For C-MAPSS, the time length of each sample is fixed as 60, so we use the patch sizes within [4,5,6,7,8,9,10] for patch size analysis. For Opportunity HAR, the time length of each sample is fixed as 128, so we use the patch sizes within [2,4,8,16,32]. Based on the analysis of C-MAPSS in Fig. 9 and Opportunity in Fig. 10, it is evident that relatively large patch sizes tend to result in better performance for our method. This observation holds true for several cross-scenarios, such as 3$\to$1, 3$\to$2, and 3$\to$4 of C-MAPSS, where the best performance is achieved with patch sizes of 8, 9, or 10. Similar observations can be found for Opportunity HAR, such as 1$\to$3, 2$\to$1, and 3$\to$1, where the best performance is achieved with patch sizes of 16 or 32. Conversely, setting too small patch sizes leads to poor performance in most scenarios, such as 1$\to$3 and 1$\to$4 of C-MAPSS, and 1$\to$2 and 1$\to$3 of Opportunity HAR. The reason behind this trend lies in the construction of sequential graphs and the capturing of spatial dependencies for transferring across domains. By segmenting MTS data into patches, we construct sequential graphs whose spatial dependencies are captured for transferring across domains. In this process, too small patch sizes may not contain enough information to fully capture the complex spatial dependencies between sensors for each patch, resulting in suboptimal performance in transferring the dependency information. Therefore, it is recommended to use relatively large patch sizes to obtain better performance using SEA++ for MTS-UDA. ## VI Conclusion In this paper, we formulate Multivariate Time-Series Unsupervised Domain Adaptation (MTS-UDA). We analyze the problems underlying this task and propose SEnsor Alignment (SEA) to address these issues. To reduce the domain discrepancy at both the local and global sensor levels, we design endo-feature alignment and exo-feature alignment. At the local sensor level, endo-feature alignment aligns sensor features and sensor correlations across domains, preventing misalignment at the sensor level. To enhance the endo-feature alignment, we further design SEA++, incorporating high-order statistic alignment and a multi-graph alignment technique. These enhancements are specifically designed to facilitate comprehensive alignment across domains and effectively capture the evolving data distributions within MTS data. At the global sensor level, we enforce restrictions on global sensor features to reduce domain discrepancy. Our extensive experiments demonstrate the effectiveness of SEA and SEA++ for MTS-UDA. ## References * [1] E. Eldele, M. Ragab, Z. Chen, M. Wu, C.-K. Kwoh, X. Li, and C. Guan, “Self-supervised contrastive representation learning for semi-supervised time-series classification,” _IEEE Transactions on Pattern Analysis and Machine Intelligence_ , 2023. * [2] Q. Ma, S. Li, and G. W. Cottrell, “Adversarial joint-learning recurrent neural network for incomplete time series classification,” _IEEE Transactions on Pattern Analysis and Machine Intelligence_ , vol. 44, no. 4, pp. 1765–1776, 2020. * [3] R. Zhao, R. Yan, Z. Chen, K. Mao, P. Wang, and R. X. Gao, “Deep learning and its applications to machine health monitoring,” _Mechanical Systems and Signal Processing_ , vol. 115, pp. 213–237, 2019. * [4] K. Chen, D. Zhang, L. Yao, B. Guo, Z. Yu, and Y. Liu, “Deep learning for sensor-based human activity recognition: Overview, challenges, and opportunities,” _ACM Computing Surveys (CSUR)_ , vol. 54, no. 4, pp. 1–40, 2021. * [5] H. He, Q. Zhang, S. Bai, K. Yi, and Z. Niu, “Catn: Cross attentive tree-aware network for multivariate time series forecasting,” _Proceedings of the AAAI Conference on Artificial Intelligence_ , vol. 36, no. 4, pp. 4030–4038, Jun. 2022. [Online]. Available: https://ojs.aaai.org/index.php/AAAI/article/view/20320 * [6] W. Zheng and J. Hu, “Multivariate time series prediction based on temporal change information learning method,” _IEEE Transactions on Neural Networks and Learning Systems_ , 2022. * [7] A. Garg, W. Zhang, J. Samaran, R. Savitha, and C.-S. Foo, “An evaluation of anomaly detection and diagnosis in multivariate time series,” _IEEE Transactions on Neural Networks and Learning Systems_ , vol. 33, no. 6, pp. 2508–2517, 2021. * [8] S. Zhao, X. Yue, S. Zhang, B. Li, H. Zhao, B. Wu, R. Krishna, J. E. Gonzalez, A. L. Sangiovanni-Vincentelli, S. A. Seshia _et al._ , “A review of single-source deep unsupervised visual domain adaptation,” _IEEE Transactions on Neural Networks and Learning Systems_ , vol. 33, no. 2, pp. 473–493, 2020. * [9] E. Tzeng, J. Hoffman, N. Zhang, K. Saenko, and T. Darrell, “Deep domain confusion: Maximizing for domain invariance,” _arXiv preprint arXiv:1412.3474_ , 2014. [Online]. Available: https://arxiv.org/abs/1412.3474 * [10] Y. Ganin, E. Ustinova, H. Ajakan, P. Germain, H. Larochelle, F. Laviolette, M. Marchand, and V. Lempitsky, “Domain-adversarial training of neural networks,” _J. Mach. Learn. Res._ , vol. 17, no. 1, p. 2096–2030, jan 2016\. * [11] E. Tzeng, J. Hoffman, K. Saenko, and T. Darrell, “Adversarial discriminative domain adaptation,” in _2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)_ , 2017, pp. 2962–2971. * [12] M. Long, Z. Cao, J. Wang, and M. I. Jordan, “Conditional adversarial domain adaptation,” _Advances in neural information processing systems_ , vol. 31, 2018. * [13] B. Sun and K. Saenko, “Deep coral: Correlation alignment for deep domain adaptation,” in _Computer Vision – ECCV 2016 Workshops_ , G. Hua and H. Jégou, Eds. Cham: Springer International Publishing, 2016, pp. 443–450. * [14] M. Long, H. Zhu, J. Wang, and M. I. Jordan, “Deep transfer learning with joint adaptation networks,” in _International conference on machine learning_. PMLR, 2017, pp. 2208–2217. * [15] S. Purushotham, W. Carvalho, T. Nilanon, and Y. Liu, “Variational recurrent adversarial deep domain adaptation,” in _ICLR_ , 2017. * [16] R. Cai, J. Chen, Z. Li, W. Chen, K. Zhang, J. Ye, Z. Li, X. Yang, and Z. Zhang, “Time series domain adaptation via sparse associative structure alignment,” _Proceedings of the AAAI Conference on Artificial Intelligence_ , vol. 35, no. 8, pp. 6859–6867, May 2021. [Online]. Available: https://ojs.aaai.org/index.php/AAAI/article/view/16846 * [17] Q. Liu and H. Xue, “Adversarial spectral kernel matching for unsupervised time series domain adaptation,” in _Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence, IJCAI-21_ , Z.-H. Zhou, Ed. International Joint Conferences on Artificial Intelligence Organization, 8 2021, pp. 2744–2750, main Track. [Online]. Available: https://doi.org/10.24963/ijcai.2021/378 * [18] Y. Shi, X. Ying, and J. Yang, “Deep unsupervised domain adaptation with time series sensor data: A survey,” _Sensors_ , vol. 22, no. 15, p. 5507, 2022\. * [19] M. Ragab, E. Eldele, W. L. Tan, C.-S. Foo, Z. Chen, M. Wu, C.-K. Kwoh, and X. Li, “Adatime: A benchmarking suite for domain adaptation on time series data,” _arXiv preprint arXiv:2203.08321_ , 2022. [Online]. Available: https://arxiv.org/abs/2203.08321 * [20] Y. Wang, Y. Xu, J. Yang, Z. Chen, M. Wu, X. Li, and L. Xie, “Sensor alignment for multivariate time-series unsupervised domain adaptation,” in _Proceedings of the AAAI Conference on Artificial Intelligence_ , vol. 37, no. 8, 2023, pp. 10 253–10 261. * [21] Z. Pei, Z. Cao, M. Long, and J. Wang, “Multi-adversarial domain adaptation,” _Proceedings of the AAAI Conference on Artificial Intelligence_ , vol. 32, no. 1, Apr. 2018. [Online]. Available: https://ojs.aaai.org/index.php/AAAI/article/view/11767 * [22] C. Yu, J. Wang, Y. Chen, and M. Huang, “Transfer learning with dynamic adversarial adaptation network,” in _2019 IEEE International Conference on Data Mining (ICDM)_. IEEE, 2019, pp. 778–786. * [23] Z. Du, J. Li, H. Su, L. Zhu, and K. Lu, “Cross-domain gradient discrepancy minimization for unsupervised domain adaptation,” in _Proceedings of the IEEE/CVF conference on computer vision and pattern recognition_ , 2021, pp. 3937–3946. * [24] Z. Gao, S. Zhang, K. Huang, Q. Wang, and C. Zhong, “Gradient distribution alignment certificates better adversarial domain adaptation,” in _Proceedings of the IEEE/CVF International Conference on Computer Vision_ , 2021, pp. 8937–8946. * [25] L. Chen, H. Chen, Z. Wei, X. Jin, X. Tan, Y. Jin, and E. Chen, “Reusing the task-specific classifier as a discriminator: Discriminator-free adversarial domain adaptation,” in _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition_ , 2022, pp. 7181–7190. * [26] M. Long, Y. Cao, Z. Cao, J. Wang, and M. I. Jordan, “Transferable representation learning with deep adaptation networks,” _IEEE transactions on pattern analysis and machine intelligence_ , vol. 41, no. 12, pp. 3071–3085, 2018. * [27] C. Chen, Z. Fu, Z. Chen, S. Jin, Z. Cheng, X. Jin, and X.-S. Hua, “Homm: Higher-order moment matching for unsupervised domain adaptation,” in _Proceedings of the AAAI conference on artificial intelligence_ , vol. 34, no. 04, 2020, pp. 3422–3429. * [28] S. Li, C. Liu, Q. Lin, B. Xie, Z. Ding, G. Huang, and J. Tang, “Domain conditioned adaptation network,” in _Proceedings of the AAAI Conference on Artificial Intelligence_ , vol. 34, no. 07, 2020, pp. 11 386–11 393. * [29] W. Zellinger, T. Grubinger, E. Lughofer, T. Natschläger, and S. Saminger-Platz, “Central moment discrepancy (cmd) for domain-invariant representation learning,” _arXiv preprint arXiv:1702.08811_ , 2017. * [30] J. Li, E. Chen, Z. Ding, L. Zhu, K. Lu, and H. T. Shen, “Maximum density divergence for domain adaptation,” _IEEE transactions on pattern analysis and machine intelligence_ , vol. 43, no. 11, pp. 3918–3930, 2020. * [31] Y. Xu, J. Yang, H. Cao, K. Wu, M. Wu, Z. Li, and Z. Chen, “Multi-source video domain adaptation with temporal attentive moment alignment network,” _IEEE Transactions on Circuits and Systems for Video Technology_ , 2023. * [32] Y. Xu, H. Cao, K. Mao, Z. Chen, L. Xie, and J. Yang, “Aligning correlation information for domain adaptation in action recognition,” _IEEE Transactions on Neural Networks and Learning Systems_ , 2022. * [33] J. Li, S. Qiu, C. Du, Y. Wang, and H. He, “Domain adaptation for eeg emotion recognition based on latent representation similarity,” _IEEE Transactions on Cognitive and Developmental Systems_ , vol. 12, no. 2, pp. 344–353, 2019. * [34] H. Zhao, Q. Zheng, K. Ma, H. Li, and Y. Zheng, “Deep representation-based domain adaptation for nonstationary eeg classification,” _IEEE Transactions on Neural Networks and Learning Systems_ , vol. 32, no. 2, pp. 535–545, 2020. * [35] G. Wilson, J. R. Doppa, and D. J. Cook, “Multi-source deep domain adaptation with weak supervision for time-series sensor data,” in _Proceedings of the 26th ACM SIGKDD international conference on knowledge discovery & data mining_, 2020, pp. 1768–1778. * [36] ——, “Calda: Improving multi-source time series domain adaptation with contrastive adversarial learning,” _IEEE Transactions on Pattern Analysis and Machine Intelligence_ , 2023. * [37] Q. Liu and H. Xue, “Adversarial spectral kernel matching for unsupervised time series domain adaptation.” in _IJCAI_ , 2021, pp. 2744–2750. * [38] T. Yang, X. Yu, N. Ma, Y. Zhao, and H. Li, “A novel domain adaptive deep recurrent network for multivariate time series prediction,” _Engineering Applications of Artificial Intelligence_ , vol. 106, p. 104498, 2021. [Online]. Available: https://www.sciencedirect.com/science/article/pii/S0952197621003468 * [39] A. Deng and B. Hooi, “Graph neural network-based anomaly detection in multivariate time series,” in _Proceedings of the AAAI Conference on Artificial Intelligence_ , vol. 35, 2021, pp. 4027–4035. * [40] Z. Jia, Y. Lin, J. Wang, R. Zhou, X. Ning, Y. He, and Y. Zhao, “Graphsleepnet: Adaptive spatial-temporal graph convolutional networks for sleep stage classification,” in _Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence, IJCAI-20_ , C. Bessiere, Ed. International Joint Conferences on Artificial Intelligence Organization, 7 2020, pp. 1324–1330, main track. * [41] J. Gilmer, S. S. Schoenholz, P. F. Riley, O. Vinyals, and G. E. Dahl, “Neural message passing for quantum chemistry,” _arXiv preprint arXiv:1704.01212_ , 2017. * [42] H. Guo, R. Pasunuru, and M. Bansal, “Multi-source domain adaptation for text classification via distancenet-bandits,” in _Proceedings of the AAAI conference on artificial intelligence_ , vol. 34, no. 05, 2020, pp. 7830–7838. * [43] Y. Wang, M. Wu, X. Li, L. Xie, and Z. Chen, “Multivariate time series representation learning via hierarchical correlation pooling boosted graph neural network,” _IEEE Transactions on Artificial Intelligence_ , 2023. * [44] R.-G. Cirstea, B. Yang, and C. Guo, “Graph attention recurrent neural networks for correlated time series forecasting,” _MileTS19@ KDD_ , pp. 1–6, 2019\. * [45] A. Saxena, K. Goebel, D. Simon, and N. Eklund, “Damage propagation modeling for aircraft engine run-to-failure simulation,” in _2008 International Conference on Prognostics and Health Management_ , 2008, pp. 1–9. * [46] D. Roggen, A. Calatroni, M. Rossi, T. Holleczek, K. Förster, G. Tröster, P. Lukowicz, D. Bannach, G. Pirkl, A. Ferscha, J. Doppler, C. Holzmann, M. Kurz, G. Holl, R. Chavarriaga, H. Sagha, H. Bayati, M. Creatura, and J. d. R. Millàn, “Collecting complex activity datasets in highly rich networked sensor environments,” in _2010 Seventh International Conference on Networked Sensing Systems (INSS)_ , 2010, pp. 233–240. * [47] M. Ragab, Z. Chen, M. Wu, C. S. Foo, C. K. Kwoh, R. Yan, and X. Li, “Contrastive adversarial domain adaptation for machine remaining useful life prediction,” _IEEE Transactions on Industrial Informatics_ , vol. 17, no. 8, pp. 5239–5249, 2021. * [48] M. Ragab, E. Eldele, Z. Chen, M. Wu, C.-K. Kwoh, and X. Li, “Self-supervised autoregressive domain adaptation for time series data,” _IEEE Transactions on Neural Networks and Learning Systems_ , pp. 1–11, 2022. * [49] Z. Chen, M. Wu, R. Zhao, F. Guretno, R. Yan, and X. Li, “Machine remaining useful life prediction via an attention-based deep learning approach,” _IEEE Transactions on Industrial Electronics_ , vol. 68, no. 3, pp. 2521–2531, 2021. * [50] Q. Xu, Z. Chen, K. Wu, C. Wang, M. Wu, and X. Li, “Kdnet-rul: A knowledge distillation framework to compress deep neural networks for machine remaining useful life prediction,” _IEEE Transactions on Industrial Electronics_ , vol. 69, no. 2, pp. 2022–2032, 2022. * [51] S. Fu, S. Zhong, L. Lin, and M. Zhao, “A novel time-series memory auto-encoder with sequentially updated reconstructions for remaining useful life prediction,” _IEEE Transactions on Neural Networks and Learning Systems_ , vol. 33, no. 12, pp. 7114–7125, 2021. * [52] B. Sun, J. Feng, and K. Saenko, _Correlation Alignment for Unsupervised Domain Adaptation_. Cham: Springer International Publishing, 2017, pp. 153–171. [Online]. Available: https://doi.org/10.1007/978-3-319-58347-1_8 * [53] Y. Ganin, E. Ustinova, H. Ajakan, P. Germain, H. Larochelle, F. Laviolette, M. Marchand, and V. Lempitsky, _Domain-Adversarial Training of Neural Networks_. Cham: Springer International Publishing, 2017, pp. 189–209. [Online]. Available: https://doi.org/10.1007/978-3-319-58347-1_10 * [54] S. Cui, S. Wang, J. Zhuo, L. Li, Q. Huang, and Q. Tian, “Towards discriminability and diversity: Batch nuclear-norm maximization under label insufficient situations,” in _IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)_ , June 2020. * [55] Y. Zhu, F. Zhuang, J. Wang, G. Ke, J. Chen, J. Bian, H. Xiong, and Q. He, “Deep subdomain adaptation network for image classification,” _IEEE Transactions on Neural Networks and Learning Systems_ , vol. 32, no. 4, pp. 1713–1722, 2021. * [56] H. Rangwani, S. K. Aithal, M. Mishra, A. Jain, and R. V. Babu, “A closer look at smoothness in domain adversarial training,” in _Proceedings of the 39th International Conference on Machine Learning_ , 2022. * [57] M. Vayyat, J. Kasi, A. Bhattacharya, S. Ahmed, and R. Tallamraju, “Cluda: Contrastive learning in unsupervised domain adaptation for semantic segmentation,” _arXiv preprint arXiv:2208.14227_ , 2022.
# Unveiling a hidden bar-like structure in NGC 1087: kinematic and photometric evidence using MUSE/VLT, ALMA and JWST Carlos López-Cobá Institute of Astronomy and Astrophysics, Academia Sinica, No. 1, Section 4, Roosevelt Road, Taipei 10617, Taiwan Lihwai Lin Institute of Astronomy and Astrophysics, Academia Sinica, No. 1, Section 4, Roosevelt Road, Taipei 10617, Taiwan Sebastián F. Sánchez Instituto de Astronomía, Universidad Nacional Autonoma de México, Circuito Exterior, Ciudad Universitaria, Ciudad de México 04510, Mexico ###### Abstract We report a faint non-axisymmetric structure in NGC 1087 through the use of JWST Near Infrared Camera (NIRCam), with an associated kinematic counterpart observed as an oval distortion in the stellar velocity map, H$\alpha$ and CO $J=2\rightarrow 1$ velocity fields. This structure is not evident in the MUSE optical continuum images but only revealed in the near-IR with the F200W and F300M band filters at $2\mu$m and $3\mu$m respectively. Due to its elongation, this structure resembles a stellar bar although with remarkable differences with respect to conventional stellar bars. Most of the near-IR emission is concentrated within $6\arcsec~{}\sim 500$ pc with a maximum extension up to 1.2 kpc. The spatial extension of the large-scale non-circular motions is coincident with the bar, which undoubtedly confirms the presence of a non- axisymmetric perturbation in the potential of NGC 1087. The oval distortion is enhanced in CO due to its dynamically cold nature rather than in H$\alpha$. We found that the kinematics in all phases including stellar, ionized and molecular, can be described simultaneously by a model containing a bisymmetric perturbation; however, we find that an inflow model of gas along the bar major axis is also likely. Furthermore the molecular mass inflow rate associated can explain the observed star formation rate in the bar. This reinforces the idea that bars are mechanisms for transporting gas and triggering star formation. This work contributes to our understanding of non-axisymmetry in galaxies using the most sophisticated data so far. ## 1 Introduction Stellar bars or “bars” are one of the most visual signs of non-axisymmetry in galaxies (de Vaucouleurs et al., 1991). Like other morphological structures in galaxies, bars are not well defined, although they are relatively easy to identity in composite images because of their bar-shape structure (e.g., Masters et al., 2011; Cheung et al., 2013), and their prominent bar dust lanes (Athanassoula, 1992). NGC 1087 is a clear example where the combination of high spatial resolution data allowed us to reveal a hidden bar that is only observed in the near-infrared (NIR). The James Webb Space Telescope (JWST) is allowing us to reveal structures not seen before due to the lack of resolution and sensitivity. Their infrared bands are prone to detect the emission from old-stellar structures like bars $\sim 10$ Gyr (Sánchez-Blázquez et al., 2011). In the optical, dust obscuration can prevent their detection, affecting the global statistics of galaxies hosting bars in the nearby Universe (e.g., Sellwood & Wilkinson, 1993). Apart from their optical and NIR characteristics, bars leave imprints of non-axisymmetry in the kinematics of gas and stars (e.g., Wong et al., 2004; Fathi et al., 2005; López-Cobá et al., 2022); thus studying the dynamical effects of bars at local scales is crucial for understanding the role they play in galaxy evolution. A comprehensive study of bars could be addressed if kinematics and wide-range photometric observations were accessible for a considerably large sample of galaxies. Yet, spatial resolution plays an important role in separating the different structural components of galaxies. For example, the large integral field spectroscopic surveys like CALIFA (Sánchez et al., 2012) or MaNGA (Bundy et al., 2015), although providing a large statistical sample of galaxies, their nominal resolution of $2\farcs 5\sim 1$ kpc inhibits a detailed study of individual galaxies, while biasing the sample towards bar-lengths larger than the FWHM resolution. In a similar manner NIR photometric and optical catalogs provide resolution of few arcseconds which again limits the detection of small bars (Sheth et al., 2010). In this work, we make use of the most sophisticated data to unravel the detection of a bar in NGC 1087. This paper is structured as follows: in Section 2 we describe the data and data analysis; in Section 3 we address the detection of a faint stellar bar in the infrared while their ionized and molecular counterpart are addressed in Section 4; Section 5 describes the oval distortion and in Section 6 we present the discussion and conclusions. ## 2 Data and Data analysis NGC 1087 is an intermediate Sc galaxy located at 14 Mpc (e.g., Kourkchi & Tully, 2017); at this distance the physical spatial resolution is 68 pc$\mathrm{~{}arcsec^{-1}}$. This work is based on public data from the Multi-Unit Spectroscopic Explorer (MUSE, Bacon et al., 2010); data from the JWST Near Infrared Camera (NIRCam); and ALMA CO $J=2\rightarrow$1 observations, in particular we used the data products from the PHANGS–ALMA survey (e.g., Leroy et al., 2021a, b), namely moments 0 and 1. The MUSE-VLT data was obtained as part of the recent release of the MUSE-PHANGS datacubes (e.g., Schinnerer, 2021; Emsellem et al., 2022). MUSE is an integral field spectrograph which provides spatially resolved spectra over a $1\arcmin\times 1\arcmin$ field of view (FoV), covering the optical spectrum, with a spectral resolution at full width at half maximum (FWHM) of $\sim 2.6$Å. The estimated spatial resolution of the $2\arcmin\times 3\arcmin$ MUSE mosaic covering NGC 1087 is not worse than $0\farcs 9$/FWHM (e.g., Emsellem et al., 2022). We used fully calibrated data products from JWST from the JWST Science Calibration Pipeline version 1.10.1 (e.g., Bushouse et al., 2022). Specifically, we made use of the PHANGS-JWST first result products (e.g., Rosolowsky, Erik, 2022), namely, the NIRCam F200W filter (FWHM $\sim 0\farcs 05$ and $0\farcs 031/\mathrm{pixel}$ sampling), which has its nominal wavelength at $1.99\mu$m, in addition to the F360M and F335M band filters to trace the stellar continuum. Finally the spatial resolution of the ALMA data is $1\farcs 60$/FWHM following Leroy et al. (2021a). The CO moment 0 map was transformed to surface density ($\mathrm{\Sigma_{mol}}$), assuming a standard Milky Way $\mathrm{\alpha_{\mathrm{CO}}^{1-0}=4.35\,M_{\odot}/pc^{2}}$ conversion factor and $\mathrm{\Sigma_{mol}[M_{\odot}pc^{2}]=6.7\,I_{CO(2-1)}[K\,km/s]}\cos i$ (e.g., Eq. 11 from Leroy et al., 2021a), with $i$ being the disk inclination angle estimated in $44.5^{\circ}$. The MUSE data analysis was made with the pypipe3d tool (e.g., Lacerda et al., 2022). In short, pypipe3d performs a decomposition of the observed stellar spectra into multiple simple stellar populations (SSPs) each with different age and metalicities. For concordance with Emsellem et al. (2022), we use the same stellar libraries based on the extended MILES (E-MILES, Vazdekis et al., 2016), FWHM $=2.5$Å; We convolved the spectral resolution of the SSPs to that of the MUSE line-spread function (e.g., Bacon et al., 2017). To increase the signal-to-noise (SN) of the stellar continuum, we performed a Voronoi binning- segmentation on a 2D-slice centered around the V-band, ensuring bin-sizes of the order of the MUSE-FWHM resolution and SN $\sim 50$. This binned map will serve to compute different properties of the underlying stellar continuum, while the ionized gas properties are analyzed in a spaxel-wise sense. The result of pypipe3d is a set of maps comprising information about the stellar populations (stellar velocity, stellar mass, among others products), and the ionized gas (emission-line fluxes including: H$\alpha$, H$\beta$, [S ii]$\lambda 6717,6731$, [N ii]$\lambda 6584$, [O iii]$\lambda 5007$, [O i]$\lambda 6300$, emission-line velocities, equivalent widths etc); see Lacerda et al. (2022) for a thorough description of the analysis and dataproducts. From this analysis we estimate the stellar mass of this object in $\mathrm{\log M_{\star}/M_{\odot}=9.9}$. Figure 1: Central panel: False color image showing the distribution of the ionized gas and NIR continuum in NGC 1087 with fluxes taken from MUSE-VLT and NIRCam imaging respectively (red: [S ii], yellow: H$\alpha$, blue: [O iii], brown: F360M, orange: F335M, white: F200W). Right panels: A zoom-in of the $30\arcsec\times 20\arcsec$ innermost region is shown to highlight the bar- structure. The upper panel shows a true color image from the MUSE cube (R: i-band, G: r, B: g). The black cross in the middle represents the NIR nucleus. The middle panel shows the JWST-F200W image together with a set of isophotes describing the bar NIR light distribution; the isophotes show a constant orientation in the sky at $\phi^{\prime}_{bar}=304^{\circ}\pm 3^{\circ}$. Two isophotes located at $6\arcsec$ and $14\arcsec$ are highlighted with thicker lines for reference. In each case the FWHM resolution is shown with a yellow circle. Bottom inset: Orientation of the object in the sky. ## 3 FAINT stellar bar The optical continuum image of NGC 1087 exhibits a bright, fuzzy and featureless nucleus with several dust lanes as revealed by the $gri$ MUSE image in Figure 1, because of that it is unclear whether this object shows a stellar bar in the optical. Two bright spots are observed in the central region where the near-IR (NIR) nucleus is found. Unlike the majority of galaxies hosting stellar bars, NGC 1087 does not show a well defined bar, nucleus neither a star forming ring (e.g., de Vaucouleurs et al., 1991). Conversely, the high spatial resolution from the F200W imaging filter allows to resolve a clear elongated and clumpy structure of presumably stellar clusters aligned along a preferential position angle (P.A.), which differs from the disk orientation111From now on primed variables make reference to values measured on the sky plane, otherwise in the disk plane. $\phi_{disk}^{\prime}=359^{\circ}$, as observed in Figure 1. At $2\mu$m this structure resembles a faint stellar bar and it is obscured in the optical due to dust absorption. F200W traces mostly stellar continuum, hence, it is expected that most of the continuum emission along this structure comes from old stars (e.g., Leitherer et al., 1999). Elliptical isophotes with variable position angle and ellipticity ($\varepsilon^{\prime}$) show a preferential alignment of the $2\mu$m emission as observed in the top-right panel from Figure 1. Figure 2: NGC 1087 as observed with the F200W and F300M NIRCam band filters. First panel: zoom around a $30\arcsec\times 20\arcsec$ size window containing the bar-like structure. second panel: Best non-parametric model for the $2\mu$m light distribution using the photometric version of the $\mathtt{XS}$ code (in prep.). The coordinates of the photometric center was estimated in $\alpha_{2000}=2^{h}46^{m}25.170^{s}$, $\delta_{2000}=-00^{\circ}29^{\prime}55.728\arcsec$, while the position angle of the disk describing the bar light distribution was estimated in $\phi_{bar}^{\prime}=305^{\circ}$ with a $0.5$ ellipticity. Third panel: residuals of the modeling (observed-model). Fourth panel: the black line represents the light distribution profile of the model shown in the second panel, while the blue dashed line represents the cumulative light distribution of the model. Distances are measured following these projection angles. Figure 3: Diagnostic diagrams for the ionized gas in the bar structure. Top panel: [O i]6300/H$\alpha$ vs. [O iii]/H$\beta$ diagram. The black dashed line represents the star forming demarcation curve from Kewley et al. (2006). Shock grid models from the updated MAPPINGS-V (Sutherland et al., 2018) computed by Alarie & Morisset (2019) are shown for $\mathrm{n_{e}=10\,cm^{-3}}$, $\mathrm{Z_{ISM}=0.008~{}(equivalent~{}to~{}12+\log O/H=8.55)}$, and a wide range of magnetic fields (blue lines) and pre-shock velocities (red lines). The colors of the points map the F200W light distribution shown in the inset figure. Bottom panel: EW(H$\alpha$) in absolute value vs. [N ii]/H$\alpha$ line ratio. Old and evolved stars are expected to produce EW(H$\alpha$) $<3$Å.. Figure 4: Oxygen abundance in the bar-region computed with the Pilyugin & Grebel (2016) calibrator. As in Figure 1 the white ellipses describe the NIR bar. The innermost ellipse shows an average metallicity of $\mathrm{12+\log O/H=8.5}$. Figure 5: Top: Specific star formation rate computed through the H$\alpha$ based SFR and the stellar mass density maps, $\mathrm{sSFR=SFR/M_{\star}}$. Bottom: Molecular surface density around the bar structure. The H$\alpha$ flux is overlaid with reddish colors for comparison. The yellow circles represent the FWHM spatial resolution in each case. The white ellipses delineate the NIR bar extension. Figure 6: First column: Line-of-sight velocities of CO, H$\alpha$ and the stellar velocity in NGC 1087. Second column: Zoom-in around the bar-structure, with $\pm 25$ km s-1 spaced iso-velocity shown on-top; the black straight line shows the P.A. of the faint bar at $\phi_{bar}^{\prime}=305^{\circ}$. Third column: residual velocities after subtracting the circular rotation in each velocity map. Fourth column: Bisymmetric model after fixing the oval orientation at $\phi_{bar}^{\prime}$. Fifth column: inflow model assuming the gas streams along $\phi_{bar}^{\prime}$; the individual radial profiles and expressions are shown in Figure 7. The root mean square of the models is shown at the bottom right of each panel. All maps share the same color-bar except for the residual maps which cover $\pm 50$ km s-1 following the same colors scheme. ### 3.1 Bar light modeling The method adopted here is similar to the non-parametric method from Reese et al. (2007), which is the same algorithm used by $\mathtt{XS}$ (e.g., Lopez- Coba et al., 2021) for extracting the different velocities in the kinematic models, as we explain in the following sections. Our non-parametric model minimizes the function $\chi^{2}=(I_{obs}-I_{k}(r)w_{k}(r))^{2}$, where $I_{obs}$ is the observed intensity map, $I_{k}$ represents a set of intensities that will be estimated at different annuli, and $w_{k}$ is a set of weighting factors that will serve for performing a linear interpolation between the estimated $I_{k}$ intensities. The implementation on the F200W and F300M band filters is shown in Figure 2. The NIR light from the bar-structure can be successfully modeled by an elliptical light distribution with with a constant orientation in the sky. The P.A. of such ellipse is $\phi_{bar}^{\prime}=305^{\circ}$222Sky angles relative the disk major axis can be translated to angles measured in the galaxy plane and vise versa through, $\tan\phi_{bar}=\tan\Delta\phi/\cos i$, with $\Delta\phi=\phi_{bar}^{\prime}-\phi_{disk}^{\prime}$. and $\varepsilon^{\prime}_{bar}=0.5$, as shown in Figure 2. From this analysis the estimated semi-major axis length of the NIR bar is $a_{bar}^{\prime}=14\arcsec$ or $18\arcsec\sim 1.2$ kpc333Distances measured at different P.A. in the sky plane can be translated to distances in the galaxy plane following: $r_{disk}^{2}=r_{sky}^{2}(\cos^{2}\Delta\phi+\sin^{2}\Delta\phi/\cos^{2}i)$. on the disk at the considered distance. However, this structure is very diffuse, with most of the 2$\mu$m emission $\sim 70\%$, concentrated within the inner $6\arcsec\sim 500$ pc. The latter region encloses the two bright spots observed in the MUSE optical continuum images. ## 4 Ionized and molecular gas The central panel of Figure 1 shows the ionized gas distribution traced by the H$\alpha$, [S ii] and [O iii] emission-line fluxes. In general, galaxies hosting stellar bars do not frequently present cold or warm gas along the bar. This is often explained as a lost of angular momentum after the gas encounters the offset ridges, resulting in an infall of gas towards a central ring (e.g., Athanassoula, 1992). NGC 1087 shows plenty of ionized gas throughout the disk and within the bar region. Furthermore, as we will see later, multiple enhanced CO bloops are spatially coincident with the F200W emission. To investigate the dominant ionizing source along the bar-like structure we made use of line-ratios sensitive to the ionization, in particular the [O i]6300/H$\alpha$ and [O iii]/H$\beta$. The [O i]6300/H$\alpha$ ratio has the advantage of being a good indicator of the presence of shocks. The top panel from Figure 3 shows these line-ratios color-coded with the $2\mu$m emission. The observed ionized gas is compatible with being produced by star-formation (SF) according to the Kewley et al. (2006) demarcation line. However, dust lanes in bars are often associated with shocks as a result of the complex gas dynamics (Athanassoula, 1992). The bar-region in NGC 1087 shows several dust lanes, thus we adopt shock models to investigate whether shock ionization can reproduce the observed line ratios. For this purpose we used the photoionization grids from MAPPINGS V (e.g., Sutherland et al., 2018) computed by Alarie & Morisset (2019); we adopt the Gutkin et al. (2016) metallicities. The latter parameter was set from the oxygen gas-phase abundance, adopting the Pilyugin & Grebel (2016) calibrator based on strong emission-lines (see Figure 4). For the bar-region we find $\mathrm{12+\log O/H\sim 8.5}$ corresponding to $\mathrm{Z_{ISM}=0.008}$ in the Gutkin et al. (2016) models. For this metallicity, the predicted emission-line intensities from shock models can not explain the observed [O i]/H$\alpha$ line ratios as observed in the top panel of Figure 3. Thus, if shocks are happening in the bar-region, then their optical emission is not expected to be the dominant ionizing source. Although there is still a possibility that a combination of SF plus shocks could contribute to the observed line-ratios (e.g., Davies et al., 2016). Additionally, Figure 3 shows the equivalent width of H$\alpha$ (EW($\rm{H}\alpha$)) for the ionized gas in the bar-region. Hot, old, and low- mass evolved stars are characterized for producing EW($\rm{H}\alpha$)$~{}\lesssim 3$Å (e.g., Stasińska et al., 2008; Cid Fernandes et al., 2010; Lacerda et al., 2018). The large values of EW($\rm{H}\alpha$) $\gtrsim 100$ Å found in the bar, can not be explained by the ionizing continuum emitted by this population of stars. Therefore, if most of the $2\mu$m emission observed along the bar is due to old stars, then their ionizing continuum is not sufficient to explain the observed line-ratios. The star formation rate (SFR) was estimated from the H$\alpha$ luminosity (Kennicutt, 1998), after correcting the H$\alpha$ flux from dust extinction adopting the Cardelli et al. (1989) extinction law with $\mathrm{R_{V}=3.1}$, and assuming an intrinsic flux ratio of H$\alpha$/H$\beta$= 2.86 and ionized gas temperature of T $\sim 10^{4}$ K corresponding to case B recombination (e.g., Osterbrock, 1989). The integrated SFR within the NIR bar (i.e., Figure 2) is $\mathrm{SFR(H\alpha)_{bar}}=0.08~{}\mathrm{M_{\odot}/yr}$, while the total $\mathrm{SFR(H\alpha)_{total}}=0.4~{}\mathrm{M_{\odot}/yr}$. The specific star formation rate (sSFR) was obtained from dividing the SFR surface density (=SFR/pixel area) by the stellar mass surface density as shown in Figure 5. We note a clear enhancement in SF along the bar main axis; furthermore, the molecular surface density shows an enhancement in the same region. Overall, our analysis suggests that the ionized gas in the bar structure is mostly associated to SF processes. Following we investigate whether the CO and SF concentration is induced by the bar potential. ## 5 Inner Oval distortion If a perturbation in the gravitational potential, such as that induced by a bar potential, is causing the $2\mu$m light distribution to elongate along a preferred direction; then the particles in such structure are expected to follow quasi-elliptical orbits (e.g., Athanassoula, 1992). Hence kinematics of gas and stars should reflect this perturbation (Pence & Blackman, 1984). Figure 6 shows the CO, ionized and stellar kinematics around the bar region. The gas, being collisional, is more sensitive to non-axisymmetric perturbations. This is clearly reflected in the CO moment 1 map where a strong distortion of the semi-minor axis is observed. The former misalignment is a signature of an oval distortion produced by a bar (e.g., Pence et al., 1988; López-Cobá et al., 2022), here observed at high spatial resolution. The distortion is also observed in the H$\alpha$ velocity map, however, since the ionized gas traced by H$\alpha$ is hotter ($\mathrm{T\sim 10^{4}}$ K), and therefore it presents a larger intrinsic velocity dispersion than the one of the molecular gas traced with CO ($\mathrm{T\sim 10}$ K), it is less pronounced in the ionized gas and is probably affected by local SF. The stellar kinematics although affected by the pixel-coadding during the SSP analysis, it still reveal a clear distortion in the iso-velocities near the bar as noted in Figure 6. So far, it is clear that the stars, the molecular gas traced with CO and the ionized gas traced with H$\alpha$ respond in a similar way to the presence of this bar-like structure detected in NGC 1087. ### 5.1 Kinematic interpretation of the oval distortion The distortion in the velocity field such as the ones observed before could be caused by an elongated potential. However, only a few kinematic models in the literature attempt to describe the flow caused by an oval distortion (e.g., Spekkens & Sellwood, 2007; Maciejewski et al., 2012). A bisymmetric distortion induced by a second order perturbation to the potential, such as that induced by an elongated potential, has been shown to successfully reproduce the velocity field in bars (e.g., Spekkens & Sellwood, 2007). Since the F200W image evidences the presence of a faint bar-like structure oriented at $305^{\circ}$ in the sky, we performed bisymmetric models over the stellar, H$\alpha$ and CO velocity maps, with a fixed orientation of the oval distortion. The bisymmetric model from Spekkens & Sellwood (2007) is described by the following expression: $V_{\mathrm{LoS}}=\sin i\Big{(}V_{t}(r)\cos\theta-V_{2,t}(r)\cos 2\theta_{\mathrm{bar}}\cos\theta\\\ -V_{2,r}(r)\sin 2\theta_{\mathrm{bar}}\sin\theta\Big{)}+V_{\mathrm{sys}}$ (1) where $\theta_{bar}=\theta-\phi_{bar}$; with $\theta$ being the azimuthal angle measured on the disk plane from the line of nodes and $\phi_{bar}$ the position angle of the bar-like structure on the disk plane. $V_{t}$ is the tangential or circular rotation and $V_{2r}$ and $V_{2t}$ represent the radial and tangential velocities that result from a bisymmetric distortion to the gravitational potential. Before this analysis we performed a circular rotation model to obtain the disk projection angles and the rotational curves. In this case and in the subsequent, we used the $\mathtt{XS}$ code for generating the kinematic models (e.g., Lopez-Coba et al., 2021). This code derives an interpolated model over a set of concentric rings evenly spaced $r_{k}$, minimizing the function $\chi^{2}=(V_{obs}-V_{k}(r)w_{k}(r))^{2}/\sigma^{2}$, where $V_{obs}$ is the observed velocity map; $V_{k}$ are the set of velocities inferred at $r_{k}$; $w_{k}$ is a set of weights that depend on the specific kinematic model adopted and will serve to create a 2D model; $\sigma$ is the error velocity map. We refer to Lopez-Coba et al. (2021) for a thorough description of this analysis. It is important to mention that the modeling does not assume any parametric function on the velocity profiles. A pure circular rotation, without non-circular motions, is described by the first term on the right from Equation 1. The disk orientation of NGC 1087 was estimated from modeling the H$\alpha$ and stellar velocity maps with circular rotation only, obtaining $\phi_{disk}^{\prime}=358.9^{\circ}$, $i=44.5^{\circ}$ and $V_{sys}=1526$ km s-1. The residual velocity maps from this model i.e., $V_{obs}-V_{circular}$, are shown in the third column from Figure 6. The circular rotation models leave large-scale residual velocities of the order of $30$ km s-1 with a mirror symmetry about the nucleus; this corresponds to de-projected amplitudes of $\sim 50$ km s-1 for the non- circular velocities on the disk plane. These residual patterns have been observed in larger scales in galaxies with long bars in H$\alpha$ (e.g., Lang et al., 2020; López-Cobá et al., 2022) and molecular gas (e.g., Pence et al., 1988; Mazzalay et al., 2014). The bisymmetric model with a fixed oval orientation at $\phi_{bar}^{\prime}$ is shown in the fourth column in Figure 6. This model reproduces simultaneously the twisted iso-velocities in the three velocity maps. The root mean square (rms) of the models decrease compared with the circular rotation one, therefore, in terms of the residuals, the LoS-velocities observed around the bar region can be reproduced by a kinematic model that considers a bar- like perturbation as the main source of non-circular motions. However, at such high spatial resolution $\sim 70$ pc, there are still residual velocities that the bisymmetric model can not account for, this is reflected in the 15 km s-1 rms in the models, which is similar or larger than the level of turbulence of the ISM (e.g., Moiseev et al., 2015). An alternative interpretation to the observed non-circular motions in the H$\alpha$ and CO velocity maps is the presence of gas inflow induced by the NIR-bar (Mundell & Shone, 1999); in fact, hydro-dynamical models simulating bars predict inflow of gas along the offset ridges, or bar dust lanes, (e.g., Athanassoula, 1992). As observed in Figure 1, the MUSE image of NGC 1087 shows several dust lanes in the central region making difficult the identification of those associated with the bar. Assuming the spiral arms are trailing (see bottom-right panel from Figure 1), NGC 1087 rotates counterclockwise in the sky. Thus, positive (negative) residuals in the near (far) side represent inflow; hence, a radial flow ($V_{rad}$) along the bar major axis is inflowing if $V_{rad}<0$. We implemented a non-axisymmetric model, with a flow streaming along $\phi_{bar}^{\prime}$. The non-axisymmetric inflow model is described by the following expression: $V_{LoS}=(V_{t}(r)\cos\theta+V_{rad}(r)\cos\theta_{bar}\sin\phi_{bar})\sin i\\\ +V_{sys}$ (2) where $V_{rad}$ is the radial velocity of the flow, and $\phi_{bar}$ is the misalignment between the disk and the bar position angles on the disk plane. This model assumes the gas is flowing along $\phi_{bar}$. This expression is similar to Hirota et al. (2009) and Wu et al. (2021) assuming that the gas flows parallel to the bar major axis. In order to consider pixels likely affected by this motion, we consider the elliptical region that better describes the bar-like light distribution and defined in Figure 2. The 2D representation of this model is shown in Figure 6, while the kinematic radial profiles of all models are shown in Figure 7. As observed, the radial velocity $V_{rad}$ results negative in the model, with maximum inflow velocities of the order of 50 km s-1. Figure 7: Radial distribution of the different velocity components for the considered kinematic models. $V_{t}$ is the disk pure circular rotation, or rotational curve; $V_{2r}$ and $V_{2t}$ are the bisymmetric velocities; $V_{rad}$ the radial flow from the non-axisymmetric flow model. In all non- circular models the position angle of the oval distortion was fixed to $\phi_{bar}^{\prime}=305^{\circ}$ (namely, $297^{\circ}$ in the galaxy plane). Lines in red colors represent results of models computed on the H$\alpha$ velocity map, blue lines on the CO moment 1 and black lines on the stellar velocity map. Shaded regions represent $1\sigma$ errors. The non-circular velocities were estimated up to $18\arcsec$, which covers the de-projected length of the bar. A $2\arcsec$ sampling step was adopted, i.e., larger than the FWHM spatial resolution of the data. The molecular mass flow rate ($\dot{M}_{mol}$) associated can be computed following the expression (e..g, Di Teodoro & Peek, 2021): $\dot{M}_{mol}(r)=2\pi r\Sigma_{mol}(r)V_{inflow}(r)$ (3) with $\mathrm{\Sigma_{mol}}$ being the de-projected molecular mass surface density, $V_{inflow}$ the CO inflow velocity and $r$ the galactocentric distance. Figure 8 shows the spatially resolved $\dot{M}_{mol}$ and its average radial profile. We find an average $\dot{M}_{mol}\sim-20\mathrm{M_{\odot}/yr}$. For comparison, spiral arms induce radial flows and radial velocities of the order of $1\mathrm{M_{\odot}/yr}$ and $10$ km s-1 respectively, (e.g., Di Teodoro & Peek, 2021). Figure 8: Molecular mass inflow induced by the bar-like structure. Top panel: 2D distribution of $\dot{M}_{mol}$. Bottom panel: Average radial distribution of $\dot{M}_{mol}$ in black and radial velocity profile of the inflow velocity in red colors. Shadow black regions represent the standard deviation of $\dot{M}_{mol}$ at each radial bin. A $\mathrm{SN>3}$ was applied to the moment 0 map to exclude spurious CO detection. ## 6 Discussion and Conclusions The combination of spatially resolved spectra provided by MUSE-VLT, with the high spatial resolution from JWST and the ALMA CO $J=2\rightarrow 1$ data allowed us to reveal a central kinematic oval distortion, as well as a small scale elongated structure in the NIR continuum in NGC 1087. This elongated structure is not evident in optical images, but only revealed with the NIRCam F200W and F360M filters at $2\mu$m and $3\mu$m respectively, thanks to its exquisite resolution. The optical counterpart of this bar is likely to be affected by dust absorption and by the intense radiation from young stars. Its preferential elongation suggests the presence of a non-axisymmetric perturbation in the gravitational potential, in particular a bar-like type. At first order, the 2$\mu$m light distribution of this structure can be described with an exponential disk profile oriented at $\phi_{bar}^{\prime}=305^{\circ}$, and the ellipticity of the elongated structure is $\varepsilon^{\prime}=0.5$, (see Figure 2), with a half-major axis length of $a_{bar}^{\prime}\sim 14\arcsec$ or $\sim$ 1.2 kpc in the galaxy plane. Overall the shape of this structure resembles to a faint stellar bar, although with remarkable differences with respect to conventional stellar bars, in terms of bar-length, the lack of a SF nuclear ring and not clear associated dust lanes along the bar axis. The same argument however applies for an elongated bulge. The main reason this is not considered is due to the large reservoirs of molecular gas observed, since bulges tend to be quiescent structures with null or low SFRs (Shapiro et al., 2010). Furthermore, the [O i]$/$H$\alpha$ and [O iii]$/$H$\beta$ line-ratios and large EWH$\alpha$ from the bar (see Figure 1) indicate that the ionized gas is product from a recent or ongoing SF event given its spatial coincidence with the CO emission. The ionized and molecular gas kinematics and the stellar velocity, reveal simultaneously the presence of an oval distortion in the velocity maps in concordance with the extension of the faint bar. This evidences undoubtedly the imprints of a non-axisymmetric perturbation in the potential, here captured in different phases and enhanced in dynamically cold gas. Kinematic models for a second order potential perturbation successfully described the LoS velocities in the three velocity maps. We showed that this is achieved just by fixing the orientation of the stream flow to the bar major axis in the models, and the radial profiles of the noncircular velocities are consistent with the residuals from circular rotation, with possible differences in amplitudes due to assymmetric drift and other random motions not considered in the model as we showed in Figure 7. Apart from the elliptical motions considered in the bisymmetric models, radial inflows are expected in bars, and these are observed to occur along the bar-dust lanes (Wu et al., 2021; Sormani et al., 2023). Our implementation of an inflow model is therefore physically motivated. An inflow of cold gas caused by a lose of angular momentum could trigger the SF observed along the bar. Although a fraction of ionized gas could arise by the shock with the dust lanes, the contribution of shocks is expected to be minor. Moreover, the sign of the radial flow suggests the gas is inflowing to the center at a maximum speed of $50$ km s-1. This kinematic model yields the lowest rms despite being the model with the lowest number of free variables. Statistically speaking, the inflow scenario is favored over the bisymmetric model, although given the complex behavior of gas in bars, the contribution of gas oval orbits can not be ruled out. In fact both type of motions are expected to happen simultaneously in bars (Athanassoula, 1992; Regan et al., 1999). To date, there is no a 2D-kinematic model that includes such dynamics of bars. The gas inflow model is in concordance with the observed enhancement in H$\alpha$ flux along the bar. The inferred inflow velocities translate in molecular mass inflow rates of $\dot{M}_{mol}=20~{}\mathrm{M_{\odot}/yr}$. This is larger than the measured SFR(H$\alpha$) along the bar, thus we argue that a major inflow of gas might be feeding the SF in the bar. In fact SF prevents the so- called continuity problem (e.g., Simon et al., 2003; Maciejewski et al., 2012) when often considering radial flow models. This large inflow rate is partially due to the large inflow velocities; although this amplitude of non-circular motions is observed in bars, the global inflow rate expected by hydrodynamical models is lower (Athanassoula, 1992; Mundell & Shone, 1999). To summarize, in this work we have taken the advantage of multiple archival data to reveal a hidden non-axisymmetric structure in the optical, thanks to the high angular resolution of the data. We were able to confirm the presence of an oval distortion and successfully model it with a bar-like flow model. We also show that the SFR in the bar could be explained by an inflow of gas along the bar major-axis. Our results contribute to understand the overall picture of non-axisymmetry in galaxies with the most sophisticated data so far. Additionally, it highlights the importance of using the IR instead of optical bands to detect stellar bars, which could increase the fraction of barred galaxies detected in the local Universe. ## Acknowledgment We thank the anonymous referee for their comments and suggestions which help to improve the quality of this manuscript. CLC acknowledges support from Academia Sinica Institute of Astronomy and Astrophysics. LL acknowledges the Ministry of Science & Technology of Taiwan under the grant NSTC 112-2112-M-001-062 -. Based on data obtained from the ESO Science Archive Facility with DOI(s): https://doi.org/10.18727/archive/47 (catalog https://doi.org/10.18727/archive/47). The specific PHANGS–JWST observations analyzed can be can be found in MAST: https://doi.org/10.17909/9bdf-jn24 (catalog https://doi.org/10.17909/9bdf-jn24). This paper makes use of the following ALMA data: ADS/JAO.ALMA#2018.1.01651.S. ## References * Alarie & Morisset (2019) Alarie, A., & Morisset, C. 2019, Rev. Mexicana Astron. Astrofis., 55, 377, doi: 10.22201/ia.01851101p.2019.55.02.21 * Athanassoula (1992) Athanassoula, E. 1992, MNRAS, 259, 345, doi: 10.1093/mnras/259.2.345 * Bacon et al. (2010) Bacon, R., Accardo, M., Adjali, L., et al. 2010, in Proc. SPIE, Vol. 7735, Ground-based and Airborne Instrumentation for Astronomy III, 773508, doi: 10.1117/12.856027 * Bacon et al. (2017) Bacon, R., Conseil, S., Mary, D., et al. 2017, A&A, 608, A1, doi: 10.1051/0004-6361/201730833 * Bundy et al. (2015) Bundy, K., Bershady, M. A., Law, D. R., et al. 2015, ApJ, 798, 7, doi: 10.1088/0004-637X/798/1/7 * Bushouse et al. (2022) Bushouse, H., Eisenhamer, J., Dencheva, N., et al. 2022, JWST Calibration Pipeline, 1.8.2, Zenodo, doi: 10.5281/zenodo.7229890 * Cardelli et al. (1989) Cardelli, J. A., Clayton, G. C., & Mathis, J. S. 1989, ApJ, 345, 245, doi: 10.1086/167900 * Cheung et al. (2013) Cheung, E., Athanassoula, E., Masters, K. L., et al. 2013, ApJ, 779, 162, doi: 10.1088/0004-637X/779/2/162 * Cid Fernandes et al. (2010) Cid Fernandes, R., Stasińska, G., Schlickmann, M. S., et al. 2010, MNRAS, 403, 1036, doi: 10.1111/j.1365-2966.2009.16185.x * Davies et al. (2016) Davies, R. L., Groves, B., Kewley, L. J., et al. 2016, MNRAS, 462, 1616, doi: 10.1093/mnras/stw1754 * de Vaucouleurs et al. (1991) de Vaucouleurs, G., de Vaucouleurs, A., Corwin, Herold G., J., et al. 1991, Third Reference Catalogue of Bright Galaxies * Di Teodoro & Peek (2021) Di Teodoro, E. M., & Peek, J. E. G. 2021, ApJ, 923, 220, doi: 10.3847/1538-4357/ac2cbd * Emsellem et al. (2022) Emsellem, E., Schinnerer, E., Santoro, F., et al. 2022, A&A, 659, A191, doi: 10.1051/0004-6361/202141727 * Fathi et al. (2005) Fathi, K., van de Ven, G., Peletier, R. F., et al. 2005, MNRAS, 364, 773, doi: 10.1111/j.1365-2966.2005.09648.x * Gutkin et al. (2016) Gutkin, J., Charlot, S., & Bruzual, G. 2016, MNRAS, 462, 1757, doi: 10.1093/mnras/stw1716 * Hirota et al. (2009) Hirota, A., Kuno, N., Sato, N., et al. 2009, PASJ, 61, 441, doi: 10.1093/pasj/61.3.441 * Kennicutt (1998) Kennicutt, Jr., R. C. 1998, ApJ, 498, 541, doi: 10.1086/305588 * Kewley et al. (2006) Kewley, L. J., Groves, B., Kauffmann, G., & Heckman, T. 2006, MNRAS, 372, 961, doi: 10.1111/j.1365-2966.2006.10859.x * Kourkchi & Tully (2017) Kourkchi, E., & Tully, R. B. 2017, ApJ, 843, 16, doi: 10.3847/1538-4357/aa76db * Lacerda et al. (2022) Lacerda, E. A. D., Sánchez, S. F., Mejía-Narváez, A., et al. 2022, New A, 97, 101895, doi: 10.1016/j.newast.2022.101895 * Lacerda et al. (2018) Lacerda, E. A. D., Cid Fernandes, R., Couto, G. S., et al. 2018, MNRAS, 474, 3727, doi: 10.1093/mnras/stx3022 * Lang et al. (2020) Lang, P., Meidt, S. E., Rosolowsky, E., et al. 2020, ApJ, 897, 122, doi: 10.3847/1538-4357/ab9953 * Leitherer et al. (1999) Leitherer, C., Schaerer, D., Goldader, J. D., et al. 1999, ApJS, 123, 3, doi: 10.1086/313233 * Leroy et al. (2021a) Leroy, A. K., Schinnerer, E., Hughes, A., et al. 2021a, ApJS, 257, 43, doi: 10.3847/1538-4365/ac17f3 * Leroy et al. (2021b) Leroy, A. K., Hughes, A., Liu, D., et al. 2021b, ApJS, 255, 19, doi: 10.3847/1538-4365/abec80 * Lopez-Coba et al. (2021) Lopez-Coba, C., Lin, L.-H., & Sanchez Sanchez, S. F. 2021, arXiv e-prints, arXiv:2110.05095. https://arxiv.org/abs/2110.05095 * López-Cobá et al. (2022) López-Cobá, C., Sánchez, S. F., Lin, L., et al. 2022, ApJ, 939, 40, doi: 10.3847/1538-4357/ac937b * Maciejewski et al. (2012) Maciejewski, W., Emsellem, E., & Krajnović, D. 2012, MNRAS, 427, 3427, doi: 10.1111/j.1365-2966.2012.22040.x * Masters et al. (2011) Masters, K. L., Nichol, R. C., Hoyle, B., et al. 2011, MNRAS, 411, 2026, doi: 10.1111/j.1365-2966.2010.17834.x * Mazzalay et al. (2014) Mazzalay, X., Maciejewski, W., Erwin, P., et al. 2014, MNRAS, 438, 2036, doi: 10.1093/mnras/stt2319 * Moiseev et al. (2015) Moiseev, A. V., Tikhonov, A. V., & Klypin, A. 2015, MNRAS, 449, 3568, doi: 10.1093/mnras/stv489 * Mundell & Shone (1999) Mundell, C. G., & Shone, D. L. 1999, MNRAS, 304, 475, doi: 10.1046/j.1365-8711.1999.02330.x * Osterbrock (1989) Osterbrock, D. E. 1989, Astrophysics of gaseous nebulae and active galactic nuclei * Pence & Blackman (1984) Pence, W. D., & Blackman, C. P. 1984, MNRAS, 210, 547, doi: 10.1093/mnras/210.3.547 * Pence et al. (1988) Pence, W. D., Taylor, K., Freeman, K. C., de Vaucouleurs, G., & Atherton, P. 1988, ApJ, 326, 564, doi: 10.1086/166117 * Pilyugin & Grebel (2016) Pilyugin, L. S., & Grebel, E. K. 2016, MNRAS, 457, 3678, doi: 10.1093/mnras/stw238 * Reese et al. (2007) Reese, A. S., Williams, T. B., Sellwood, J. A., Barnes, E. I., & Powell, B. A. 2007, AJ, 133, 2846, doi: 10.1086/516826 * Regan et al. (1999) Regan, M. W., Sheth, K., & Vogel, S. N. 1999, ApJ, 526, 97, doi: 10.1086/307960 * Rosolowsky, Erik (2022) Rosolowsky, Erik. 2022, PHANGS-JWST First Results, STScI/MAST, doi: 10.17909/9BDF-JN24 * Sánchez et al. (2012) Sánchez, S. F., Kennicutt, R. C., Gil de Paz, A., et al. 2012, A&A, 538, A8, doi: 10.1051/0004-6361/201117353 * Sánchez-Blázquez et al. (2011) Sánchez-Blázquez, P., Ocvirk, P., Gibson, B. K., Pérez, I., & Peletier, R. F. 2011, MNRAS, 415, 709, doi: 10.1111/j.1365-2966.2011.18749.x * Schinnerer (2021) Schinnerer, E. 2021, 3D Mapping of the Dynamic Feedback and Chemistry in the Star Formation Process across Galaxy Disks, European Southern Observatory (ESO), doi: 10.18727/ARCHIVE/47 * Sellwood & Wilkinson (1993) Sellwood, J. A., & Wilkinson, A. 1993, Reports on Progress in Physics, 56, 173, doi: 10.1088/0034-4885/56/2/001 * Shapiro et al. (2010) Shapiro, K. L., Falcón-Barroso, J., van de Ven, G., et al. 2010, MNRAS, 402, 2140, doi: 10.1111/j.1365-2966.2009.16111.x * Sheth et al. (2010) Sheth, K., Regan, M., Hinz, J. L., et al. 2010, PASP, 122, 1397, doi: 10.1086/657638 * Simon et al. (2003) Simon, J. D., Bolatto, A. D., Leroy, A., & Blitz, L. 2003, ApJ, 596, 957, doi: 10.1086/378200 * Sormani et al. (2023) Sormani, M. C., Barnes, A. T., Sun, J., et al. 2023, MNRAS, 523, 2918, doi: 10.1093/mnras/stad1554 * Spekkens & Sellwood (2007) Spekkens, K., & Sellwood, J. A. 2007, ApJ, 664, 204, doi: 10.1086/518471 * Stasińska et al. (2008) Stasińska, G., Vale Asari, N., Cid Fernandes, R., et al. 2008, MNRAS, 391, L29, doi: 10.1111/j.1745-3933.2008.00550.x * Sutherland et al. (2018) Sutherland, R., Dopita, M., Binette, L., & Groves, B. 2018, MAPPINGS V: Astrophysical plasma modeling code, Astrophysics Source Code Library, record ascl:1807.005. http://ascl.net/1807.005 * Vazdekis et al. (2016) Vazdekis, A., Koleva, M., Ricciardelli, E., Röck, B., & Falcón-Barroso, J. 2016, MNRAS, 463, 3409, doi: 10.1093/mnras/stw2231 * Wong et al. (2004) Wong, T., Blitz, L., & Bosma, A. 2004, ApJ, 605, 183, doi: 10.1086/382215 * Wu et al. (2021) Wu, Y.-T., Trejo, A., Espada, D., & Miyamoto, Y. 2021, MNRAS, 504, 3111, doi: 10.1093/mnras/stab1087
# Draft Hitting time for Markov decision process Ruihcao Jiang, Javad Tavakoli, Yiqiang Zhao ###### Abstract We define the hitting time for a Markov decision process (MDP). We do not use the hitting time of the Markov process induced by the MDP because the induced chain may not have a stationary distribution. Even it has a stationary distribution, the stationary distribution may not coincide with the (normalized) occupancy measure of the MDP. We observe a relationship between the MDP and the PageRank. Using this observation, we construct an MP whose stationary distribution coincides with the normalized occupancy measure of the MDP and we define the hitting time of the MDP as the hitting time of the associated MP. ## 1 Introduction ††footnotetext: The first version of this paper pointed out some issues in the old version of [3]. The authors of [3] have then addressed these issues according to our suggestions in a new version. We therefore updated our paper, in which we removed contents related to these issues. In [3], Gromov-Wasserstein distance was used to compare different MDPs. In their approach, a distance function is needed on the space of state-action pairs of the underlying MDP. They used Euclidean distance. However, it is not entirely clear how to define the distance between actions. Moreover, it is desirable that the distance reflects some useful information about the MDP. For this reason, we propose to use the first-hitting time quasi-distance in places of the distance function. The rationale is as follows. * • The occupancy measure to the MDP is whats the stationary distribution to an Markov process (MP); * • The stationary distribution of an MP is simply the inverse of the mean of the first-return time; * • The first-return time is simply the diagonal element for the first-hitting time matrix. ## 2 Hitting time of MDP To define the first-hitting time of MDP, we first look at the occupancy measure of MDP. ###### Definition 1 (Occupancy measure). Let $(S,A,R,P,\gamma)$ be an MDP with the initial distribution $\rho_{0}(s)$ on $S$ and the stationary policy $\pi$, the occupancy measure $\rho:S\times A\to\mathbb{R}$ is defined as follows. $\rho_{\pi}(s,a)\coloneqq\sum_{t=0}^{\infty}\gamma^{t}\mathbb{P}\left(s_{t}=s,a_{t}=a|s_{0}\sim\rho_{0},\pi\right).$ (1) We give a new interpretation of the occupancy measure. First, Tulcea’s theorem (Proposition C.10 in [4]) says that a policy $\pi$ induces a Markov Process (MP) on $S\times A$ with the transition probability $P_{\pi}(s^{\prime},a^{\prime}|s,a)\coloneqq P(s^{\prime}|s,a)\pi(a^{\prime}|s^{\prime})$ (2) and the initial distribution $\rho_{0}(s,a)\coloneqq\rho_{0}(s)\pi(a|s)$. This describes a process as follows: Given the state-action pair $(s,a)$, transit to the next state $s^{\prime}$ according to the transition probability of the original MDP; then immediately choose an action $a^{\prime}$ according to the policy $\pi$. Note that $\rho_{\pi}$ is not the stationary distribution (if it exists at all) of $(S\times A,P_{\pi})$. First of all, $\rho_{\pi}$ is not a probability measure as it sums to $\frac{1}{1-\gamma}$. Even after a normalization, $\left(1-\gamma\right)\rho_{\pi}$ may differ from the stationary distribution. Consider a scenario where a car moves straight on a straight road and if it turns left or right, it will go off the road. Then none of the state-action pairs related to steering will be occupied and hence $\rho_{\pi}$ is zero on those pairs. However, for any stationary distribution, each component must be strictly positive. We will see later that it does not matter whether $(S\times A,P_{\pi})$ is ergodic or not. Treating $\rho_{\pi}$ and $\rho_{0}$ as $1\times|S||A|$ vectors and $P_{\pi}$ as an $|S||A|\times|S||A|$ matrix, the following equation is well-known. $\rho_{\pi}=\left(I-\gamma P_{\pi}\right)^{-1}\rho_{0}.$ (3) From Equation (3), we make the following key observation. ###### Observation (PageRank). $(1-\gamma)\rho_{\pi}$ coincides with the personalized PageRank vector on $S\times A$ with transition matrix $P_{\pi}$ and with probability $1-\gamma$ to restart with the initial distribution $\rho_{0}$. Second, we look at the first-hitting time of an MP. ###### Definition 2 (First-hitting time of MP). Let $\\{X_{t}\\}$ be an MP. The first-hitting time $T$ is the random variable defined as follows. $T_{ij}=\begin{cases}\inf_{t\in\mathbb{N}}\\{t\ |\ X_{0}=j,X_{1}\neq i,\cdots,X_{t-1}\neq i,X_{t}=i\\}&\text{if}\ i\neq j\\\ 0,&\text{otherwise.}\end{cases}$ (4) If $i$ is never reached from $j$, $T_{ij}=+\infty$, where we have adopted the convention that $\inf\emptyset=+\infty$. By abuse of the notation, we call the expectation $\mathbb{E}T_{ij}$ of the random variable $T_{ij}$ also the first-hitting time and denote it by $T_{ij}$. $T_{ij}\coloneqq\mathbb{E}T_{ij}=\begin{cases}\sum_{t=1}^{\infty}t\mathbb{P}(T_{ij}=t)&\text{if}\ i\neq j\\\ 0&\text{otherwise}.\end{cases}$ (5) If the MP has transition matrix $P_{\pi}$, a one-step analysis shows that $T_{ij}$ satisfies the following recursive relation. $T_{ij}=\begin{cases}1+\sum_{k}T_{ik}P_{\pi}(k|j)&\text{if}\ i\neq j\\\ 0&\text{otherwise}.\end{cases}$ (6) By our Observation, we reduce the task of defining a first-hitting time for an MDP $(S,A,R,P,\gamma)$ to that of defining a first-hitting time for a PageRank $(S\times A,P_{\pi},1-\gamma)$. A PageRank can be described equivalently by a new process $(S\times A,\tilde{P}_{\pi})$ with the transition matrix $\tilde{P}_{\pi}(x^{\prime}|x)\coloneqq(1-\gamma)\rho_{0}(x^{\prime})+\gamma P_{\pi}(x^{\prime}|x).$ (7) $(S\times A,\tilde{P}_{\pi})$ is obviously Markovian. The initial condition for this new process can be arbitrary, in particular, not necessarily $\rho_{0}$ since it turns out that $(S\times A,\tilde{P}_{\pi})$ is ergodic and its stationary distribution is given by $(1-\gamma)\rho_{\pi}$ (See Proposition 1 and Corollary 1 of [1]). The ergodicity of $(S\times A,\tilde{P}_{\pi})$ implies that the first-hitting time is finite on supp$(\rho_{\pi})$. In this case, Equation (6) reads $T_{ij}=\begin{cases}1+(1-\gamma)\sum_{k}T_{ik}\rho_{0}(k)+\gamma\sum_{k}T_{ik}P_{\pi}(k|j)&\text{if}\ i\neq j\\\ 0&\text{otherwise}.\end{cases}$ (8) There is also a related quantity $L_{ij}$ satisfying the following recursive relation. $L_{ij}=\begin{cases}1+\gamma\sum_{k}L_{ik}P_{\pi}(k|j)&\text{if}\ i\neq j\\\ 0&\text{otherwise}.\end{cases}$ (9) The probabilistic interpretation of $L_{ij}$ is the expectation of the discounted path length from $j$ to $i$ for the first time. $L_{ij}$ and $T_{ij}$ are related by the following formula (Theorem 1(b) of [1]). $T_{ij}=\frac{L_{ij}}{1-(1-\gamma)\sum_{k}L_{ik}\rho_{0}(k)}.$ (10) ###### Remark. $L_{ij}$ is called the (1-$\gamma$)-discounted hitting time in [6] although $L_{ij}$ is not the first-hitting time of $(S\times A,P_{\pi})$, whose first- hitting time is given by Equation (6), nor of $(S\times A,\tilde{P}_{\pi})$, whose first-hitting time is given by Equation (8). However, $L_{ij}$ is easier to estimate empirically. We have shown that * • The occupancy measure of an MDP $(S,A,R,P,\gamma)$ with the initial distribution $\rho_{0}$ coincides with the PageRank vector of the personalized PageRank $(S\times A,P_{\pi},1-\gamma)$ with the initial distribution $\rho_{0}$. * • The PageRank vector of the personalized PageRank $(S\times A,P_{\pi},1-\gamma)$ with the initial distribution $\rho_{0}$ coincides with the stationary distribution of the ergodic MP $(S\times A,\tilde{P}_{\pi})$ with arbitrary initial distribution. Based on the above facts, we propose the following definition. ###### Definition 3 (First-hitting time of MDP). Let $(S,A,R,P,\gamma)$ be an MDP with a stationary policy $\pi$. Let $P_{\pi}$ be the induced transition matrix in Equation (2) and $\tilde{P}_{\pi}$ be the modified transition matrix in Equation (7). The first-hitting time of the MDP is defined as the first-hitting time of $(S\times A,\tilde{P}_{\pi})$ with its expectation given by Equation (8). Since we define the first-hitting time of MDP as the first-hitting time of some MP, we immediately have the following. ###### Proposition 1. The first-hitting time of an MDP gives rise to a quasi-distance $T_{\pi}(i,j)\coloneqq T_{ij}$, i.e. a distance function without satisfying the symmetry condition. ## 3 Hitting time quasidistance induces distance It might be a concern that we use a quasi-distance, not a distance. In this section, we show that applying $\mathcal{GW}$ to a quasi-metric measure spaces still yield a distance (our Theorem 2). We have constructed a triple $(S\times A,\rho_{\pi},T_{\pi})$ from the MDP $(S,A,R,P,\gamma)$. Now we can apply the $\mathcal{GW}$ construction [5] to define a quantity between two MDPs ###### Definition 4. Let $(S_{1},A_{1},R_{1},P_{1},\gamma_{1})$ and $(S_{2},A_{2},R_{2},P_{2},\gamma_{2})$ be MDPs. Denote by $(\mathcal{X},\rho_{\mathcal{X}},T_{\mathcal{X}})$ and $(\mathcal{Y},\rho_{\mathcal{Y}},T_{\mathcal{Y}})$ their induced triples, respectively. Then $\begin{split}&\mathcal{GW}((S_{1},A_{1},R_{1},P_{1},\gamma_{1}),(S_{2},A_{2},R_{2},P_{2},\gamma_{2}))\\\ \coloneqq&\mathcal{GW}((\mathcal{X},(1-\gamma_{1})\rho_{\mathcal{X}},T_{\mathcal{X}}),(\mathcal{Y},(1-\gamma_{2})\rho_{\mathcal{Y}},T_{\mathcal{Y}}))\\\ =&\min_{\mu}\frac{1}{2}\left[\sum_{\mathcal{X}\times\mathcal{Y}}\sum_{\mathcal{X}\times\mathcal{Y}}|T_{\mathcal{X}}(x,x^{\prime})-T_{\mathcal{Y}}(y,y^{\prime})|^{2}\mu(x,y)\mu(x^{\prime}.y^{\prime})\right]^{\frac{1}{2}},\end{split}$ (11) where $\mu\in\mathcal{M}((1-\gamma_{1})\rho_{\mathcal{X}},(1-\gamma_{2})\rho_{\mathcal{Y}})$ is the set of couplings, i.e. $\mathcal{M}((1-\gamma_{1})\rho_{\mathcal{X}},(1-\gamma_{2})\rho_{\mathcal{Y}})=\left\\{\mu\left|\sum_{y\in\mathcal{Y}}\mu=(1-\gamma_{1})\rho_{\mathcal{X}}\ \text{and}\ \sum_{x\in\mathcal{X}}\mu=(1-\gamma_{2})\rho_{\mathcal{Y}}\right\\}\right.$ (12) ###### Remark. It is necessary to normalize the occupancy measures. Otherwise, if $\gamma_{1}\neq\gamma_{2}$, then no coupling exists since $\sum_{x\in\mathcal{X}}\sum_{y\in\mathcal{Y}}\mu=\sum_{x\in\mathcal{X}}\rho_{\mathcal{X}}=\frac{1}{1-\gamma_{1}}\neq\frac{1}{1-\gamma_{2}}=\sum_{y\in\mathcal{Y}}\rho_{\mathcal{Y}}=\sum_{x\in\mathcal{X}}\sum_{y\in\mathcal{Y}}\mu.$ Different MDPs may give rise to the same triple. To eliminate this redundancy, we introduce an equivalent relation on $\mathcal{MDP}$, the set of all MDPs. ###### Definition 5 (Equivalent relation on $\mathcal{MDP}$). $\textup{MDP}_{1}$ and $\textup{MDP}_{2}$ are said to be equivalent, denoted by $\textup{MDP}_{1}\sim\textup{MDP}_{2}$ if they give rise to the same triple, i.e. there exists a bijection $\varphi:S_{1}\times A_{1}\to S_{2}\times A_{2}$ s.t. $T_{1}(x,x^{\prime})=T_{2}(\varphi(x),\varphi(x^{\prime}))$ and $(1-\gamma_{1})\rho_{1}(x)=(1-\gamma_{2})\rho_{2}(\varphi(x))$ for all $x,x^{\prime}\in S_{1}\times A_{1}$. We have not yet called $\mathcal{GW}(\textup{MDP}_{1},\textup{MDP}_{2})$ a distance. We know that $\mathcal{GW}$ is a distance function on the space of metric-measure spaces after quotient out redundancy [5]. However, the first- hitting time is only a quasi-distance, hence $(S\times A,(1-\gamma)\rho_{\pi},T_{\pi})$ is not necessarily a metric-measure space. We prove that $\mathcal{GW}$ is indeed a distance function on $\mathcal{MDP}/\sim$. ###### Theorem 1. $\mathcal{GW}$ is a distance on $\mathcal{MDP}/\sim$. ###### Proof. Theorem 16 in [2] says that if $T_{\mathcal{X}}$ and $T_{\mathcal{Y}}$ are measurable, then $\mathcal{GW}$ is a pseudo-distance. Hence,to show that $\mathcal{GW}$ is a distance, we need to show that $\mathcal{GW}(\textup{MDP}_{1},\textup{MDP}_{2})=0$ iff $\textup{MDP}_{1}\sim\textup{MDP}_{2}$. The “if" part is trivial. We only need to show $\mathcal{GW}(\textup{MDP}_{1},\textup{MDP}_{2})=0\implies\textup{MDP}_{1}\sim\textup{MDP}_{2}$. Let $\mu$ be a coupling s.t. Equation (11) evaluates to $0$. Then $T_{\mathcal{X}}(x,x^{\prime})=T_{\mathcal{Y}}(y,y^{\prime})$ for all pairs of $(x,y)$ and $(x^{\prime},y^{\prime})$ s.t. $\mu(x,y)>0$, $\mu(x^{\prime},y^{\prime})>0$. Let $x=x^{\prime}$. Then $T_{\mathcal{X}}(x,x)=T_{\mathcal{Y}}(y,y^{\prime})=0$. Since $T_{\mathcal{Y}}$ is a quasi-distance, it implies that $y=y^{\prime}$. Hence $\mu$ does not split mass i.e. $\mu$ induces a bijection $\varphi$ s.t. $T_{\mathcal{X}}(x,x^{\prime})=T_{\mathcal{Y}}(\varphi(x),\varphi(x^{\prime}))$. It remains to check measures. By the boundary condition Equation (12), $\begin{split}(1-\gamma_{2})\rho_{\mathcal{Y}}(\varphi(x))&=\sum_{x\in\mathcal{X}}\mu(x,\varphi(x))\\\ &=\sum_{\varphi(x)\in\mathcal{Y}}\mu(x,\varphi(x))\\\ &=(1-\gamma_{1})\rho_{\mathcal{X}}(x)\end{split}$ as desired. ∎ ###### Remark. Intuitively, the definition of $\mathcal{GW}$ involves a minimization over all couplings. If $T_{\pi}(x,x^{\prime})\neq T_{\pi}(x^{\prime},x)$, the minimization will choose the smaller one. ###### Remark. Theorem 16 in [2] only says that their $\mathcal{GW}$ is a pseudo-distance. They explicitly constructs an example in their Remark 15 where their $\mathcal{GW}$ fails to be a distance. This is because they assume nothing about their network weight function $\omega$, except measurability. In particular, $\omega(x,x)$ in their Remark 15 does not evaluate to zero. However, for the first-hitting time, by definition $T(x,x)=0$. ## 4 Conclusion We defined the hitting time for an MDP via a relationship between the MDP and the PageRank. The hitting time is a quasi-distance. By applying $\mathcal{GW}$ construction, the hitting time quasi-distance induces a distance on the set of discrete MDPs. ## References * [1] K. Avrachenkov, A. Piunovskiy and Y. Zhang “Hitting Times in Markov Chains with Restart and their Application to Network Centrality” In _Methodology and Computing in Applied Probability_ 20, 2018, pp. 1173–1188 URL: https://doi.org/10.1007/s11009-017-9600-5 * [2] S. Chowdhury and F. Mémoli “The Gromov–Wasserstein distance between networks and stable network invariants” In _Information and Inference: A Journal of the IMA_ 8, 2019, pp. 757–787 URL: https://doi.org/10.1093/imaiai/iaz026 * [3] Arnaud Fickinger, Samuel Cohen, Stuart Russell and Brandon Amos “Cross-Domain Imitation Learning via Optimal Transport” In _International Conference on Learning Representations_ , 2022 URL: https://openreview.net/forum?id=xP3cPq2hQC * [4] O. Hernández-Lerma and J. Lasserre “Discrete-Time Markov Control Processes”, Applications of Mathematics Springer, 1996 URL: https://doi.org/10.1007/978-1-4612-0729-0 * [5] F. Mémoli “Gromov–Wasserstein distances and the metric approach to object matching” In _Foundations of computational mathematics_ 11, 2011, pp. 417–487 URL: https://doi.org/10.1007/s10208-011-9093-5 * [6] P. Sarkar and G. Gordon “Random Walks with Random Projections” In _Workshop on Analyzing Networks and Learning with Graphs NIPS_ , 2009 URL: http://snap.stanford.edu/nipsgraphs2009/papers/sarkar-paper.pdf
# Robust Environment Perception for Automated Driving: A Unified Learning Pipeline for Visual-Infrared Object Detection Mohsen Vadidar1, Ali Kariminezhad1, Christian Mayr1, Laurent Kloeker2 and Lutz Eckstein2 1The authors are with the Elektronische Fahrwerksysteme GmbH, 85080 Gaimersheim, Germany{mohsen.vadidar, ali.kariminezhad, christian5.mayr}@efs- auto.de2The authors are with the research area Vehicle Intelligence & Automated Driving, Institute for Automotive Engineering, RWTH Aachen University, 52074 Aachen, Germany {laurent.kloeker<EMAIL_ADDRESS>aachen.de ###### Abstract The RGB complementary metal-oxide-semiconductor (CMOS) sensor works within the visible light spectrum. Therefore it is very sensitive to environmental light conditions. On the contrary, a long-wave infrared (LWIR) sensor operating in 8-14 µm spectral band, functions independent of visible light. In this paper, we exploit both visual and thermal perception units for robust object detection purposes. After delicate synchronization and (cross-) labeling of the FLIR [1] dataset, this multi-modal perception data passes through a convolutional neural network (CNN) to detect three critical objects on the road, namely pedestrians, bicycles, and cars. After evaluation of RGB and infrared (thermal and infrared are often used interchangeably) sensors separately, various network structures are compared to fuse the data at the feature level effectively. Our RGB-thermal (RGBT) fusion network, which takes advantage of a novel entropy-block attention module (EBAM), outperforms the state-of-the-art network [2] by 10% with 82.9% mAP. ## I INTRODUCTION A statistical projection of traffic fatalities in the United States for the first half of 2021 shows that an estimated 20,160 people died in motor vehicle traffic crashes. This represents an increase of about 18.4 percent as compared to 17,020 fatalities that were reported in the first half of 2020 [3]. Looking at the fatal accidents of 2019 based on the time, one can see that there are 1,000 more fatal accidents during the night-time compared to the day-time [4]. Given less average traffic during the night-time, the importance of visibility in dark is inevitable. The number of publications on RGB-IR sensor fusion for multi-spectral object detection in the automotive sector has increased within the past two years. However, the lack of data in this research area is still noticeable. There are two main sources of data, namely FLIR thermal dataset [1] and KAIST multi- spectral pedestrian detection benchmark [5], which provide a dataset containing IR and RGB pair images. FLIR mainly provides three classes car, pedestrian, and bicycle, whereas KAIST only contains pedestrians. (a) RGB Frame (b) IR Frame Figure 1: A sample of IR and RGB frames in heavy rain from the same scene: one can notice that the pedestrians are not detected on the RGB frame and the cold vehicle coming out of a parking lot is not recognized on the IR frame. Realizing the complementary data in this multispectral setup motivated us to fuse the information. Nevertheless, the FLIR dataset comes only with IR labels. That introduces the first challenge to researchers. Previously published papers [2], [6] and [7] have made various objections to the dataset. For instance, the usage of different camera resolutions at multiple instances ranging from 0.3 to 3.1 MP, misalignment of resolution/aspect ratio, a vast difference between the field of views and resolutions between the sensors, and no possibility to align the RGB and IR frames without having a correlation matrix. These issues currently made fusing the data on this dataset an unsolved problem. Within this work, we address the FLIR dataset challenges and propose a solution to properly utilize it. Further, we introduce a CNN architecture to fuse RGB-IR data. The proposed network outperforms the state-of-the-art networks for multi-spectral object detection purposes [2]. In section II, the latest advancements and publications on RGB-IR sensor fusion networks will be presented. Later in section III, a cross-labeling algorithm and pre-processing pipeline will be introduced to provide suitable labels for RGB frames. In section IV, we start with an RGB network and compare its detection performance with a monospectral IR network. The best monospectral detector will be considered as our baseline. Afterward, we proceed with the simplest form of fusion and develop our proposed RGBT network step by step. Finally, sections V and VI are dedicated to results, analysis and conclusion of this work. ## II Related Works In July 2020 Ravi Yadav et al. attempted fusing color and thermal images to detect objects for self-driving applications [6]. They have used VGG16 [8] and Faster-RCNN [9] as their encoder and detector, respectively. They achieved a miss rate of 29% on the KAIST benchmark. However, they could not perform well on the FLIR benchmark due to its complications. Later that month, Chaitanya Devaguptapu et al. [10] have used the GAN framework to use the FLIR dataset and produce RGB images from the IR frames. The idea was to borrow knowledge from the data-rich pre-trained RGB domain and use it on IR images. Heng Zhang et. al. introduced cyclic fuse and refine blocks for object detection [7] in September 2020. The FSDD (Feature Fusion Single Shot Multibox Detector) network [11] is used as their object detector and two VGG16 backbones are implemented to work independently from each other for each sensor. No information is given regarding the number of parameters or complexity of the network. However, looking at the designed cyclic module, the number of its occurrences, and considering the used VGG16, which is one of the most expensive feature extractors with 138 million parameters, one can realize that this network has more than 276 million parameters. In this work, a so-called "well-aligned" version [7] of FLIR is released. However, as mentioned in the paper, the misalignment problem made exploiting FLIR dataset for multispectral object detection impractical. Therefore, the pre-processing part, could not tackle all the issues from the dataset. We will have a look at a failed attempt in the next section. Nevertheless, Cyclic Fuse-and-Refine Network (CFR_3) could reduce the miss rate on the KAIST dataset and improve the mAP on the FLIR benchmark to 72.39%. In November 2020 A. Sai Charan et al. attempted to fuse IR and RGB data by giving weight to each sensor using attention mechanism [12]. The attention module is placed in every residual block [13] after the RELU activation function and first batch normalization. They have tried both the late and mid-fusion approaches. Late fusion in this case is done by pooling features of the thermal and RGB camera separately from their respective region proposals and then concatenating them together. Comparing the performance of the two mid and late fusions, mid-fusion has shown a better result with 62.12% mAP. In January 2021 Heng Zhang et. al. proposed guided attentive feature fusion (GAFF) [2]. This paper claims to be the first work that is considering multi-spectral feature fusion as a sub-task in network optimization. While bringing accuracy gains, GAFF has a lower computational cost compared to common addition or concatenation methods. They reduced the miss rate on the KAIST dataset by 2% and achieved an mAP of 72.9% on the FLIR dataset. To the best of our knowledge, this is the highest mAP achieved using the FLIR dataset. Thus, this work will be taken as the state-of-the-art network on this benchmark. ZHANG et. al. in a comprehensive review studied different methods and levels of fusion between visible and IR cameras for the task of tracking [14]. Since the feature extraction process is the same in tracking and object detection, this study can provide a good insight into feature-level fusion using deep learning-based methods. ## III Dataset and Data Pre-processing This section is an important part of our work, without which the rest of our research would not be possible. ### III-A Dataset Table I summarizes open-source datasets including IR frames. Most of the datasets are concerned with pedestrian class only. Those which cover a wider range of objects like the Dense dataset [15] confront other downsides such as non-availability of RGB-IR data pairs, annotation issues, strong misalignment between IR and RGB fields of view (FOV), and lack of extrinsic calibration information (e.g. correlation matrix). TABLE I: Available Datasets including IR Frames (NG stands for not given) Name | Year | Size (Frames) | Downsides ---|---|---|--- FLIR | 2020 | 13813 | IR label only Dense | 2020 | 121500 | Narrow IR FOV ZUT-FIR | 2020 | NG | IR data available only SCUT-CV | 2019 | 211011 | IR data available only | | | Pedestrian class only CAMEL | 2018 | NG | Not on-board frames Multispec. | 2017 | 7512 | Different Ref. Sys. CVC-14 | 2016 | 3695 | Gray RGB | | | Pedestrian class only KAIST | 2015 | 95000 | Crowd label as human | | | Pedestrian class only CVC-09 | 2014 | 5990 | Pedestrian class only ### III-B Data Pre-processing To include the Vulnerable Road Users (VRU) as well as motorized users in our study, we have decided to work with FLIR dataset [1]. The non-availability of a dataset for fusing the IR and RGB frames, led us to put a great deal of effort into developing a reliable cross-labeling algorithm, additional manual labeling of the missing objects, and checking the data one by one, to make sure the entire dataset is reliable and usable. Sensor data fusion between an IR and an RGB camera was a challenge to researchers since there is no open-source dataset providing synchronized and labeled RGB-IR image pairs. Given that the ground truth in FLIR dataset is provided for IR frames only, we introduce a method to cross-label the RGB frames using IR annotations semi-automatically. Figure 2: IR-RGB overlay: Union of the filed of views in green and the transferred labels from IR to RGB frame in red box Algorithm 1 demonstrates our approach for pre-processing the data. We first convert the RGB frames to gray-scale images and extract the edges. IR and RGB frames are matrices of size $640\times 512$ and $1800\times 1600$, respectively. Thus, we introduce a scaling-parameter to resize the IR frame gradually in a loop, until the IR extracted edges fall on RGB edges. The cross-correlation value between the two matrices is a good metric to understand the accuracy of the alignment. The highest value is obtained, when the edges of two frames are completely aligned. Since the two frames are different in size, the cross-correlation between them takes place in a sliding-window manner. As figure 2 illustrates, vertical and horizontal offsets exist between the RGB and IR frames. Figure 3 represents a comparison between our result and the "well-aligned" version [7]. Ultimately, we utilize 6924, 1982, and 3960 pair images for training, validation, and test, respectively. Result: X_Offset, Y_Offset, Scaling_Factor 1 load both frames 2 Smooth both frames by a Gaussian filter 3 Detect the edges using Canny edge detector 4 Use binary threshold to set the edge pixels to one and all the other pixels to zero 5 6Scaling_Factor = 2 # A constant to resize the IR frame 7 while _Width_IR < Width_RGB_ do 8 Width_IR_new := Width_IR $\times$ Scaling_Factor 9 Resize IR binary frame with the Width_IR_new and the corresponding Height_IR_new 10 Cross correlate (CC) the IR and RGB binary frames in a sliding window manner and save the pixel coordinate of the maximum value from the correlation matrix 11 12 if _CC_Value > Best_CC_Value_ then 13 Update Best_CC_Value, Pixel_coordinates, Best_Scaling_Factor 14 15 end if 16 Scaling_Factor += 0.001 17 18 end while 19Best_X_Offset, Best_Y_Offset := Pixel_coordinates 20 return Best_X_Offset, Best_Y_Offset, Best_Scaling_Factor Algorithm 1 Cross Labeling Algorithm (a) "well-aligned" version (b) our version Figure 3: Comparison of two cross-labeling methods: the left image represents the previous attempt [7] to process the FLIR dataset [1] and transfer the labels from IR to RGB coordinate system. The right picture illustrates the result of our proposed cross-labeling algorithm. ## IV Network Structures One of the most accurate networks on MSCOCO benchmark [16], namely Scaled- Yolov4 [17] is selected as a starting point for this work. The performance of each sensor will be evaluated separately and the best monospectral network will be chosen as our baseline. The CSP (Cross-Stage-Partial-Connections) [18] version of the network is used for both sensors. That is possible since the IR sensor is delivering image-like data and the encoder is very good at picking up the features and adapting itself to the IR sensors’ data texture (i.e. domain adaptation). To take the most of each monospectral backbone, we start with our vanilla fusion. The designed network fuses the feature maps across scales as shown in figures 4 and 5. The smallest scale is designed for large objects and the biggest one detects the smaller objects. After concatenation, channels are halved using $1\times 1$ convolution layers to prepare the feature maps for the detector head. The most crucial part of fusing visible and IR sensors is finding a way to weight each backbone, depending on how informative the extracted features are. As each sensor receives a different wavelength, we want to ensure that complementary data from both spectrums increase the chance of detecting an object as well as improving the robustness of the network. The goal is to detect the objects, even in lack of strong emissions within visible-light frequency. To do that, we integrate two Convolutional Block Attention Modules (CBAM) [19], such that more weight is given to more informative features depending on the input data (see figure 5). Figure 4: RGBT Network Structure: Two identical encoders are used for feature extraction from both RGB and IR frames. The feature-level fusion block receives features in three scales from each encoder and fuses them. Then, the data is passed to the detector, which is the same as CSP version of Scaled_Yolov4. At last, three tensor outputs refer to three scales. Elements of the fusion block are demonstrated in figure 5. In what follows, we will discuss our proposed entropy-block attention module (EBAM) and how it is a better substitute for the conventional CBAM [19]. In this regard, we will proceed with channel attention, which can further be generalized to spatial attention. Suppose, average and max pooling for feature map (channel) $k$ are represented by $a_{k}$ and $m_{k}$, as: $\displaystyle a_{k}$ $\displaystyle=\frac{1}{MN}\sum_{i}^{M}\sum_{j}^{N}X^{<k>}_{ij},$ (1) $\displaystyle m_{k}$ $\displaystyle=\max_{i,j}\\{X^{<k>}_{ij}\\},$ (2) respectively, where $\mathbf{X}^{<k>}$ is the matrix of $k$th feature map and $X^{<k>}_{ij}$ is the $k$th feature map grid value for the grid located in $i$th row and $j$th column. Further, we define the vector of average and max poolings as $\displaystyle\mathbf{a}$ $\displaystyle=[a_{1},\cdots,a_{k},\cdots,a_{K}]^{T},$ (3) $\displaystyle\mathbf{m}$ $\displaystyle=[m_{1},\cdots,m_{k},\cdots,m_{K}]^{T}.$ (4) According to the CBAM, the vectors $\mathbf{a}$ and $\mathbf{m}$ pass through a shared multi-layer perceptron, for which the weights are optimized in the backpropagation. Here, we assume a single layer for convenience in further mathematical presentation. This can later be generalized to multi-layer networks. Thus, the channel attention portion of the CBAM with a single-layer perceptron delivers the following weight vector: $\displaystyle\mathbf{c}_{\textrm{CBAM}}=\mathbf{W}(\mathbf{m}+\mathbf{a})+\mathbf{b},$ (5) where the matrix of shared weights and the bias are represented by $\mathbf{W}\in R^{K\times K}$ and $\mathbf{b}\in R^{K}$, respectively. Therefore, the attention weight corresponding to the $k$th feature map can readily be formulated as $\displaystyle c^{<k>}_{\textrm{CBAM}}=\mathbf{w}_{k}(\mathbf{m}+\mathbf{a})+b_{k},$ (6) where $\mathbf{w}_{k}$ is the $k$th row of matrix $\mathbf{W}$. Interestingly, one could observe that the average and max poolings are equally weighted due to the shared layers among them. Consequently, the contribution of the average and max poolings are not captured and only the sum of the values plays a role. For instance, assume a particular feature map consists of multiple very high values, while the majority of the grids contain low values. This way, the average pooling downgrades the importance of that particular feature map. However, due to the shared weights, this phenomenon can not be remedied in the backpropagation by weight optimization. This motivates us to develop a more robust and flexible attention mechanism as a replacement for CBAM. Uncertainty of a data-driven feature is a valuable metric delivering a notion for assessing how much extra information can still be mined from that feature. $\displaystyle H_{k}=-\sum_{i=1}^{M}\sum_{j=1}^{N}P(X^{<k>}_{ij})\log P(X^{<k>}_{ij}),$ (7) where $P(X^{<k>}_{ij})$ are the probabilities delivered by the softmax operation as, $\displaystyle P(X^{<k>}_{ij})=\frac{e^{X^{<k>}_{ij}}}{\sum_{i=1}^{M}\sum_{j=1}^{N}e^{X^{<k>}_{ij}}}.$ (8) Notice that, equal probabilities maximizes the entropy. $\displaystyle P(X^{<k>}_{ij})=\frac{1}{MN},\quad\forall i,j.$ (9) It is important to discern that, the max entropy is reached if the max pooling equals the average pooling of the feature map $k$, i.e., $a_{k}=m_{k}$, which holds if: $\displaystyle X^{<k>}_{ij}=\frac{1}{\sum_{i}^{M}\sum_{j}^{N}X^{<k>}_{ij}},\forall i,j.$ (10) This reflects the fact that in an extreme case the sum of max and average, i.e., $m_{k}+a_{k}$, corresponds with the entropy upper-bound, i.e., $H^{\textrm{max}}_{k}$. This extreme case, i.e., $a_{k}=m_{k}$ can not be differentiated by CBAM, since it considers only the sum. However, entropy captures the interaction between max and average as follows: $\displaystyle\lim_{m_{k}-a_{k}\rightarrow 0}H_{k}$ $\displaystyle=H^{\textrm{max}}_{k},$ (11) $\displaystyle\lim_{m_{k}-a_{k}\rightarrow\infty}H_{k}$ $\displaystyle=0.$ (12) That means, as the max pool gets closer to the average pool, the entropy maximizes and as it gets higher value than the average pool, the entropy approaches zero. It is important to mention that, the entropy of a random process does not reduce except by providing information. Furthermore, notice that more data does not necessarily mean more information. This observation motivated us to design a network, which focuses on informative features in the fusion process. This focus is represented by strongly weighting the highly- informative feature maps. We call this new attention mechanism entropy-block attention module (EBAM), for which the $k$th channel attention weight is given by: $\displaystyle c^{<k>}_{\textrm{EBAM}}=\mathbf{v}_{k}\mathbf{h}+b_{k},$ (13) where $\mathbf{h}=[H_{1},\cdots,H_{k},\cdots,H_{K}]^{T}$. Note that $\mathbf{v}_{k}$ is the channel attention weight vector designated for the $k$th feature map optimized in the backpropagation. In what follows we generalize the EBAM to capture both channel and spatial attention. Let’s define the intermediate feature map $F\in R^{C\times H\times W}$ as a tensor consisting of $\mathbf{X}^{<k>},\ \forall k$. That means, $F=[\mathbf{X}^{<1>},\cdots,\mathbf{X}^{<k>},\cdots,\mathbf{X}^{<K>}]$. This feature map tensor is the input to our attention module, from which attention weights are derived. Hence, the generalized self-attention can be written as: $W(F)=\alpha(F)\>\otimes\>F,$ (14) where $\otimes$ denotes element-wise multiplication. $C,H$ and $W$ refer to channel, height and width, respectively and $\alpha(F):R^{C\times H\times W}\rightarrow R^{C\times H\times W}$ is an attention function of the feature map $F$. Figure 5: Upper block: Vanilla fusion is referred to a simplistic approach to fuse the data by just concatenating the extracted features in each scale and selecting the most valuable information by a convolution layer. Middle block: CBAM is added as an attention module. Lower block: EBAM is designed to select informative features more efficiently. EBAM sequentially infers a 1D channel attention map $A_{c}\in R^{C\times 1\times 1}$ and a 2D spatial attention $A_{s}\in R^{1\times W\times H}$ as illustrated in figure 6. The overall attention process can be summarized as: $\displaystyle F^{\prime}=A_{c}(F)\otimes F,$ (15) $\displaystyle F^{\prime\prime}=A_{s}(F^{\prime})\otimes F^{\prime},$ Figure 6: EBAM: detailed illustration of entropy-block attention module, which contains channel and spatial domains. MLP refers to multilayer perceptron. In the channel attention by giving more weight to more uncertain feature maps (assuming each feature map is represented by a random variable), the network is signaled to extract more information by prioritizing the gradient direction in backpropagation. Therefore, the network is forced to gain more information (i.e. kernel activation) from different kernels. As a matter of fact, the proposed attention module operating only in the channel domain, showed a relatively better performance than a CBAM including both spatial and channel attention. One can calculate the explained channel attention as follows: $\displaystyle A_{c}(F)$ $\displaystyle=\sigma(MLP(H(F))),$ (16) $\displaystyle=\sigma(W_{0}(F^{C}_{H})),$ where $\sigma$ denotes the sigmoid function, $H$ is the entropy matrix, $F$ stands for feature map and $W_{0}$ represents the MLP weights. The more the entropy of a random process, the more uncertain we are about the outcome. In the case of the attention module in the spatial domain, each grid pattern can contribute to detecting an object. As the training process goes on, the grids, which refer to an object will have lower entropy (i.e. lower uncertainty). Hence, if we use the entropy values directly, more weight will be given to grids with high entropy, which would be the background information in this case. Thus, more weights are intended to be allocated for the grids with less entropy values through the feature maps. Therefore, the entropy-based spatial attention can be denoted as follows: $\displaystyle A_{s}(F^{\prime})$ $\displaystyle=\sigma(Conv.(1-\dfrac{H_{ij}}{\max_{\\{i,j\\}}H_{ij}})),$ (17) $\displaystyle=\sigma(W_{1}(F^{S}_{H})),$ The networks are trained to the point, where an overfitting effect appeared. The hyperparameters are kept unchanged as the original Scaled-Yolov4 [17] implementation. ## V Results and Analysis In this chapter, the results of our pre-processing, baseline, and fusion networks are presented. The processed data and the code can be found in the following GitHub link: https://github.com/SamVadidar/RGBT. ### V-A Data Pre-processing Exploiting our proposed pre-processing step (i.e. algorithm 1), not only the accuracy of the bounding boxes is improved, but more data (i.e. 13,848 images) could be recovered from the original dataset. That means 34.6% more data with respect to what Heng Zhang’s version [7] could use in training. Thus, a dataset is produced, where the labels can be referred to both frames, which are $640\times 512$ pixels. After running the pre-processing algorithm on the original dataset, one receives 6924, 1982, and 3960 paired images for train, validation, and test sets, respectively. Since the intention is to compare a CMOS stand-alone sensor with a CMOS-IR sensor pair, the dataset is split into night and day subsets. Therefore, one can test and evaluate a trained network based on these two scenarios. ### V-B Studied Networks To reduce the computational costs, all the network trainings and their comparisons will be done using $320\times 320$ image size and only the final network will be fine-tuned with the original images ($640\times 512$ pixels). Generally speaking, car class being the largest object among the three, shows the best mAP. On the contrary, the person class being the smallest in size is the most challenging object to detect. TABLE II: Summary of studied networks: Fusion+H_C refers to a fusion network with entropy-block attention modules in the channel domain only. RGBT contains the full network structure as discussed in chapter IV. CFR [7] and GAFF [2] illustrate the two best available mAPs (NG stands for not given) on the FLIR dataset [1]. RGBT* is trained and tested using previously published aligned version [7] of the FLIR dataset. Network |<EMAIL_ADDRESS>| Param. ---|---|--- | Person | Bicycle | Car | Overall | RGB Baseline | 39.6% | 50.4% | 79.4% | 56.6% | 52.5 M IR Baseline | 49.6% | 54.9% | 84.4% | 63.0% | 52.5 M Vanilla Fusion | 56.9% | 56.7% | 82.0% | 65.2% | 81.8 M Fusion+CBAM | 57.6% | 60.5% | 83.6% | 67.2% | 82.7 M Fusion+H_C | 62.6% | 65.9% | 86.0% | 71.5% | 82.7 M RGBT | 63.7% | 67.1% | 86.4% | 72.4% | 82.7 M CFR(640) | 74.4% | 57.7% | 84.9% | 72.3% | $>$276 M GAFF(640) | NG | NG | NG | 72.9% | NG RGBT*(640) | 85.3% | 64.0% | 88.6% | 79.3% | 82.7 M RGBT(640) | 80.1% | 76.7% | 91.8% | 82.9% | 82.7 M A very interesting finding is that the infrared camera outperforms the CMOS sensor even as a stand-alone sensor. Its performance in table II shows a 6.4% overall improvement in the mAP metric using the same training pipeline, same network, and an equal number of epochs. One can observe the improvement in all classes, especially for the pedestrian class with a 10% jump in mAP. High contrast in the frames between the objects and surroundings, very high sensor sensitivity of less than 50 mK as well as an operating temperature of -40°C to +85°C, delivers a data-rich infrared frame for feature extraction. Later in this chapter, some qualitative and quantitative analyses will be illustrated to support this statement. The first attempt for fusing the two backbones (i.e. Vanilla Fusion) has shown a 2.2% improvement in the overall mAP compared to the IR network (See table II). The boldest improvement among the classes belongs to the pedestrian class, which has experienced a drastic improvement of 7.3% with respect to the IR network and 17.3% to the RGB network considering the mAP metric. Overall 8.6% mAP rise compared to the RGB network by adding no additional module, motivated us to use the CBAM attention module [19]. By the way the CBAM is implemented, each beneficial sensor data is enhanced and downsides are compensated by focusing on the right and reliable features automatically. Therefore, an additional performance gain of 2.2% compared to the vanilla network is achieved and the bicycle class mAP is improved by 3.8%. After designing the entropy-block, despite spatial attention absence, the Fusion+H_C network could outperform the Fusion+CBAM network. One can see 4.3% mAP improvement overall, and 5.0% for pedestrian class. The proposed RGBT Network contains two complete (i.e. channel and spatial) attention modules. The spatial attention added 0.9% on top of the channel-only attention network and pushed the overall mAP up to 72.4%. The network shows a noticeable overall mAP improvement to RGB and IR baselines with 15.8% and 9.4%, respectively. This shows that a fusion network can have an immense impact on improving the mAP. Since the RGBT network performs better than both IR and RGB object detectors, one can observe the importance of an efficient feature-fusion by selecting the best features from each encoder through the entropy-block attention module. This work could outperform the results of the CFR network [7], by only utilizing half of the pixel resolution. We also have trained and tested our proposed network with the same aligned version of the FLIR dataset [7], which CFR and GAFF used. This helps to differentiate the contribution of our preprocessing method from the proposed network. As the table II illustrates, 3.6% mAP improvement is solely achieved by better preprocessing the data. The test-set of the "well-aligned" dataset has almost 46% less number of pedestrian instances. Fewer test data and relatively easier scenarios might be the underlying reason behind the higher mAP (i.e. 85.3%) for this class. Considering the number of parameters and what the CFR paper [7] reports, two VGG-16 networks [8] are used as separate feature extractors of each sensor. Hence, knowing that each VGG-16 has around 138 million parameters, one can find out that the number of the parameters in the state- of-the-art network is greater than 276 million. Looking at the table II, it is demonstrated that the RGBT network could outperform the CFR network by using at least 70% less number of parameters. Therefore, not only the mAP is improved in this work, but the complexity of the network is reduced drastically. This will cause a considerably faster inference speed compared to the CFR network. The GAFF, which showed slightly (i.e. 0.6%) better mAP than CFR network, does not report the per class mAPs and the total number of parameters of the complete network. Due to limited resources and long training time, only our proposed network is trained with the original image size. Table II shows that using the full resolution, one can reach 82.9% mAP with 10% performance gain compared to the state-of-the-art network. The results using the original images are compared to other networks in figure 7. 10.5% raise in the mAP due to a higher input resolution, which indicates the importance of the pixel per object value. Working with FLIR dataset [1], one can observe that it contains very small and challenging objects, especially in the pedestrian class. Therefore, when the number of input pixels, which are fed to the network is doubled (i.e. $640\times 640$), the overall outcome elevates significantly. Figure 7: Qualitative comparison of three scenarios: Columns from left to right: RGB, IR, Fusion with CBAM, and Fusion with EBAM. Note that the fusion results are represented using TGB (i.e. thermal, green, blue) instead of RGB colormap. Looking at figure 7, it is to be noted that the detected objects on the sidewalk in the first two scenarios (i.e. first two rows) were not labeled. That means the network is using the extracted features from other instances of the pedestrian class to classify and localize the objects. Since in the training process, detected objects, which do not have a corresponding label are punished by the loss function, the performance of the network in this scene is admirable. The third scenario (i.e. third-row) illustrates a group of heavily occluded pedestrians sitting around a table. The night scenarios are the most challenging scenarios for the RGB cameras and in this case, the target object was neither labeled nor provided a sufficient amount of pixel information. Despite these challenges, the RGBT network having only $320\times 320$ resolution could detect the pedestrians as well as the bicycles in front of them. ## VI Conclusion and Outlook In general, cameras are one of the most informative sensors on the vehicle. However, they are highly sensitive to light and weather conditions. Such disturbances affect their performance severely. In order to introduce robustness to the system, the functionality of an infrared camera as a complementary sensor is investigated. Two groups of researchers [2], [6] and [7] have considered data pre-processing and preparation of FLIR dataset [1] as one of the major challenges and referred to it as an unsolved problem. High- quality data being the core of machine learning approaches motivated us to address this issue with our cross-labeling algorithm. Hence, 34.6% more data is retrieved with respect to the previous approach [7]. Considering the trend in recent publications, the rally was to gain a single additional mAP improvement, whereas thanks to the pre-processing pipeline and the entropy- block attention module, the proposed RGBT network could outperform the state- of-the-art network [2] by 10% mAP. Undoubtedly, the availability of higher quantity and quality labeled data is essential for further investigation in this domain. In our next step, we are planning to combine our RGB-Radar network with the proposed RGBT network to extract depth information from the detected objects. We also have seen that Radar can improve the mAP considerably. At last, the network will be optimized by TensorRT for our target hardware and run on our demonstrator vehicle to perform real-world tests. ## ACKNOWLEDGMENT The authors would like to thank EFS MLOps team, specially Mr. Sebastian Balz for his unceasing support, without which the training and evaluation of the networks would not be possible. ## References * [1] T. F. LLC, “Free flir thermal dataset for algorithm training,” https://www.flir.com/oem/adas/adas-dataset-form/, 2020, accessed: 01.05.2021. * [2] H. Zhang, E. Fromont, S. Lefevre, and B. Avignon, “Guided attentive feature fusion for multispectral pedestrian detection,” in _Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)_ , January 2021, pp. 72–80, accessed: 13.06.2021. * [3] N. N. H. T. S. Administration), “Early estimate of motor vehicle traffic fatalities for the first half (jan–june) of 2021,” https://crashstats.nhtsa.dot.gov/Api/Public/ViewPublication/813199, 2021, accessed: 08.11.2021. * [4] ——, “Fatality analysis reporting system (fars) auxiliary datasets analytical user’s manual 1982-2019,” https://crashstats.nhtsa.dot.gov/Api/Public/ViewPublication/813071, 2021, accessed: 20.08.2021. * [5] S. Hwang, J. Park, N. Kim, Y. Choi, and I. S. Kweon, “Multispectral pedestrian detection: Benchmark dataset and baselines,” in _Proceedings of IEEE Conference on Computer Vision and Pattern Recognition (CVPR)_ , 2015, accessed: 15.05.2021. * [6] R. Yadav, A. Samir, H. Rashed, S. Yogamani, and R. Dahyot, “Cnn based color and thermal image fusion for object detection in automated driving,” 07 2020, accessed: 05.05.2021. * [7] H. Zhang, E. Fromont, S. Lefèvre, and B. Avignon, “Multispectral fusion for object detection with cyclic fuse-and-refine blocks - icip 2020 - ieee international conference on image processing,” Abou Dabi, United Arab Emirates, Oct. 2020, pp. 1–5, accessed: 12.05.2021. [Online]. Available: https://hal.archives-ouvertes.fr/hal-02872132 * [8] K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” _arXiv 1409.1556_ , 09 2014, accessed: 05.05.2021. * [9] S. Ren, K. He, R. B. Girshick, and J. Sun, “Faster R-CNN: towards real-time object detection with region proposal networks,” _CoRR_ , vol. abs/1506.01497, 2015, accessed: 21.05.2021. [Online]. Available: http://arxiv.org/abs/1506.01497 * [10] C. Devaguptapu, N. Akolekar, M. M Sharma, and V. N Balasubramanian, “Borrow from anywhere: Pseudo multi-modal object detection in thermal imagery,” in _The IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops_ , June 2019, accessed: 13.05.2021. * [11] Z. Li and F. Zhou, “FSSD: feature fusion single shot multibox detector,” _CoRR_ , vol. abs/1712.00960, 2017, accessed: 18.06.2021. [Online]. Available: http://arxiv.org/abs/1712.00960 * [12] A. S. Charan, M. Jitesh, M. Chowdhury, and H. Venkataraman, “Abifn: Attention-based bi-modal fusion network for object detection at night time,” _Electronics Letters_ , vol. 56, pp. 1309–1311(2), November 2020, accessed: 13.06.2021. [Online]. Available: https://digital-library.theiet.org/content/journals/10.1049/el.2020.1952 * [13] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” _CoRR_ , vol. abs/1512.03385, 2015, accessed: 05.08.2021. [Online]. Available: http://arxiv.org/abs/1512.03385 * [14] X. Zhang, P. Ye, H. Leung, K. Gong, and G. Xiao, “Object fusion tracking based on visible and infrared images: A comprehensive review,” _Information Fusion_ , vol. 63, pp. 166–187, 2020, accessed: 21.08.2021. [Online]. Available: https://www.sciencedirect.com/science/article/pii/S1566253520302657 * [15] M. Bijelic, T. Gruber, F. Mannan, F. Kraus, W. Ritter, K. Dietmayer, and F. Heide, “Seeing through fog without seeing fog: Deep multimodal sensor fusion in unseen adverse weather,” in _The IEEE Conference on Computer Vision and Pattern Recognition (CVPR)_ , June 2020, accessed: 30.06.2021. * [16] T.-Y. Lin, M. Maire, S. Belongie, L. Bourdev, R. Girshick, J. Hays, P. Perona, D. Ramanan, C. L. Zitnick, and P. Dollár, “Microsoft coco: Common objects in context,” 2014, cite arxiv:1405.0312Comment: 1) updated annotation pipeline description and figures; 2) added new section describing datasets splits; 3) updated author list 4) Accessed: 18.06.2021. [Online]. Available: http://arxiv.org/abs/1405.0312 * [17] C.-Y. Wang, A. Bochkovskiy, and H.-Y. M. Liao, “Scaled-YOLOv4: Scaling cross stage partial network,” in _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)_ , June 2021, pp. 13 029–13 038, accessed: 23.06.2021. * [18] C.-Y. Wang, H.-Y. M. Liao, Y.-H. Wu, P.-Y. Chen, J.-W. Hsieh, and I.-H. Yeh, “Cspnet: A new backbone that can enhance learning capability of cnn,” in _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops_ , June 2020, accessed: 13.06.2021. * [19] S. Woo, J. Park, J.-Y. Lee, and I. S. Kweon, “Cbam: Convolutional block attention module,” in _Proceedings of the European Conference on Computer Vision (ECCV)_ , September 2018, accessed: 11.07.2021.
Measuring Inconsistency over Sequences of Business Rule CasesThis research is part of the research project ”Handling Inconsistencies in Business Process Modeling“, which is funded by the German Research Association (reference number: DE1983/9-1). Carl Corea1 Matthias Thimm2 Patrick Delfmann1 C. Corea et al. Institure for Information Systems Research, University of Koblenz-Landau Institute for Web-Science and Technologies, University of Koblenz-Landau In this report, we investigate (element-based) inconsistency measures for multisets of business rule bases. Currently, related works allow to assess individual rule bases, however, as companies might encounter thousands of such instances daily, studying not only individual rule bases separately, but rather also their interrelations becomes necessary, especially in regard to determining suitable re-modelling strategies. We therefore present an approach to induce multiset-measures from arbitrary (traditional) inconsistency measures, propose new rationality postulates for a multiset use-case, and investigate the complexity of various aspects regarding multi-rule base inconsistency measurement. § INTRODUCTION In the context of Business Process Management, business rules are used as a central artifact to govern the execution of company activities [6]. To this aim, business rules are modelled to capture (legal) regulations as a declarative business logic. Then, given a new process instance (denoted as a case), instance-dependent facts are evaluated against the set of business rules for reasoning at run-time. For example, consider the following set of business rules in Figure <ref> (we will formalize syntax and semantics later) with the intuitive meaning that we have two rules stating that 1) platinum customers are credit worthy, and 2) customers with a mental condition are not credit worthy. Then, given a new customer case, in the example a new loan application, the facts set is evaluated against the rule set and the resulting rule base $\rb_1$ can be used to reason about the customer case. Exemplary business rule base instance $\rb_1$. The observant reader might have noticed, that the shown example yields an inconsistency, i.e., the contradictory conclusions $\mathit{creditWorthy, \neg creditWorthy}$. In fact, this is a current problem for companies, which can result from modelling errors in the business rules, or unexpected (case-dependent) facts. This problem has widely been acknowledged and has been addressed by a series of recent works, cf. e.g. [1, 5, 3]. While existing results allow to handle inconsistencies in a single business rule base instance as shown above, in practice, companies often face thousands of such instances daily. For example, the retailer Zalando reported that 37 million cases were executed in the first quarter of 2020 alone[<https://zln.do/2SFRnfC>]. As we will show in this work, considering not only single rule bases individually, but rather the entirety of all cases and their interrelations, can yield valuable insights, especially in regard to inconsistency resolution. For example, consider the following rule set, and assume there were four customer cases (with respective case-dependent facts), yielding the set of business rule cases $\ms_1$ shown in Figure <ref>: Exemplary rule base instances, constructed over a seq. of case-dependent facts. When auditing such an overview of rule base instances, two main questions are of interest from a business rules management perspective: * How inconsistent in general was the entirety of process executions? * Which specific rules were responsible for these inconsistencies from a global perspective? Towards question 1, recent results studying inconsistency in sets of knowledge bases can easily be adapted to quantify the overall degree of inconsistency (cf. Section 2). While this is a beneficial step for companies, question 2 can however be seen as of much higher importance in the scope of improving business rules. Pin-pointing the culprits of inconsistency is an essential challenge for determining suitable resolution and re-modelling strategies. Here, new methods are needed that support companies in assessing which individual rules are highly problematic from a global perspective. For instance, in Figure <ref>, the rule $a\rightarrow b$ is part of all inconsistencies and can therefore be seen as highly problematic. In this work, we therefore introduce novel means for an element-based assessment of inconsistency over a set of rule base instances by extending results from the field of inconsistency measurement [14]. Here, our contribution is as follows: We present a novel approach for inducing element-based quantitative measures for multisets of business rule bases, allowing to pin-point problematic business rules from a global perspective (Section 3). Here, we also propose postulates that should be satisfied by respective measures for this use-case and analyze the proposed means w.r.t. these postulates. We implement our approach and perform run-time experiments with real-life data-sets, and also examine the complexity of central aspects regarding inconsistency measurement in multisets of business rule bases (Section 4). We present preliminaries in Section 2 and conclude in Section 5. Proofs for technical results are provided in a supplementary document[<https://bit.ly/2V2sIDw>]. § PRELIMINARIES Business Rule Bases. In this work, we consider a basic (monotonic) logic programming language to formalise business rule bases. A (business) rule base is then constructed over a finite set $\atoms$ of atoms, with $\lang$ being the corresponding set of literals, with a rule base $\rb$ being a set of rules $r$ of the form \begin{align}\label{eq:rule} r\,:\quad l_{1},\ldots,l_{m} \rightarrow l_{0}. \end{align} with every $l_i\in\lang$. Let $\allrbs$ denote all such rule bases. Also, we denote $head(r)=l_{0}$ and $body(r)=\{l_{1},\ldots,l_{m} \}$. If $body(r)=\emptyset$, $r$ is called a fact. For a rule base $\rb$, we denote $\F(\rb)\subseteq\rb$ as the facts in $\rb$ and $\R(\rb)\subseteq\rb$ as the rules in $\rb$. We recall the business rule base $\rb_1$. Then we have \begin{align*} \F(\rb_1) & = \{mentalCondition, platinumCustomer\}\\ \R(\rb_1) & = \{platinumCustomer \rightarrow creditWorthy,\\ & \qquad mentalCondition \rightarrow \neg creditWorthy\}. \end{align*} A set of literals $M$ is called closed w.r.t. $\rb$ if it holds that for every rule of the form <ref>: if $l_{1},\ldots,l_{m}\in M$ then $l_{0}\in M$. The minimal model of a rule base $\rb$ is the smallest closed set of literals (w.r.t. set inclusion). A set $M$ of literals is called consistent if it does not contain both $a$ and $\neg a$ for an atom $a$. We say a rule base $\rb$ is consistent if its minimal model is consistent. If $\rb$ is not consistent, we say $\rb$ is inconsistent, denoted as $\rb\models\perp$. As discussed in the introduction, $\rb_1$ is not consistent. To assess inconsistency, the field of inconsistency measurement [7] has evolved, which studies quantitative measures to assess the severity of inconsistency. An inconsistency measure [7, 14] is a function $\inc: \allrbs \rightarrow \posRealInf$, where the semantics of the value are defined such that a higher value reflects a higher degree, or severity, of inconsistency. A basic inconsistency measure is the $\incmi$ inconsistency measure, which counts the number of minimal inconsistent subsets $\MI$ of a rule base $\rb$, defined via \begin{align*} \MI(\rb) = \{M \subseteq \rb \mid M \models \perp, \forall M' \subset M : M' \not\models \perp \}. \end{align*} We recall $\rb_1$. Then we have \begin{align*} \MI(\rb_1) &= \{M_1\}\\ M_{1} &= \{platinumCustomer,\\ & \qquad platinumCustomer \rightarrow creditWorthy,\\ & \qquad mentalCondition,\\ & \qquad mentalCondition \rightarrow \neg creditWorthy\}, \end{align*} consequently, $\incmi(\rb_1)=1$. As the concept of a ”severity“ of inconsistency is not easily characterisable, numerous inconsistency measures have been proposed, see [14] for an overview. To guide the development of inconsistency measures, various rationality postulates have been proposed, cf. [13] for an overview. For example, a widely agreed upon property is that of consistency, which states that an inconsistency measure should return a value of 0 w.r.t. a rule base $\rb$ iff $\rb$ is consistent. As mentioned, various other postulates exist and we will revisit some of them later when introducing culpability measures for multisets of rule bases. Measuring Overall Inconsistency in Multisets of Business Rule Bases. In this work, we are not only interested in measuring inconsistency in single business rule bases, but rather in a series of corresponding business rule base instances. As motivated in the introduction, companies currently apply a set of business rules in order to assess a stream of (case-dependent) fact sets. Therefore, given a stream of fact sets $f = \F_1,...,\F_n$, we consider multisets of business rule bases which are constructed by matching the individual fact sets in $f$ to a shared rule set $\R$. To clarify, a multiset of rule bases is an n-tuple $\ms=(\{\F_1\cup\R\},...,\{\F_n\cup\R\})=(\rb_1, ..., \rb_n)$. Let $\allmss$ denote all such multisets. An initial question for companies is to gain an overview of the overall inconsistency w.r.t. all business rule base instances in $\ms$. For that, we define an inconsistency measure for a multiset of rule bases as follows. An inconsistency measure for a multiset of rule bases is a function $m: \allmss \rightarrow \posRealInf$. In other words, an inconsistency measure for a multiset of rule bases is a function that assigns a non-negative numerical value to an n-tuple of rule bases. Similar to classical inconsistency measures, the intuition is that a higher value reflects a higher degree of inconsistency of the multiset of rule bases. For simplicity, we refer to such measures as multi-rb measures where appropriate. For the intended use-case of gaining insights about the severity of inconsistency regarding the entirety of process instances, existing inconsistency measures can be adapted to induce multi-rb measures via a summation. Given an inconsistency measure $\inc$ and a multiset of rule-bases $\ms$, the $\Sigma$-induced multi-rb measure $m_\inc^\Sigma$ is defined as $m_\inc^\Sigma: \allmss \rightarrow \posRealInf$ with $m_\inc^\Sigma(\ms)=\sum_{B \in \ms} \inc(B)$. We recall the introduced $\MI$-inconsistency measure $\incmi$. Correspondingly, given a multiset of rule bases $\ms$, $\incmi$ can be used to $\Sigma$-induce the multi-rb measure $m_{\incmi}^{\Sigma}(\ms) = \sum_{B \in \ms} \incmi(B)$. Considering again the exemplary multiset $\ms_1$ of business rules from Figure <ref>, with cases $b_1-b_4$, we thus have $m_{\incmi}^{\Sigma}(\ms_1) = \incmi(b_1) + ... + \incmi(b_4) = 1+1+1+2 = 5$. Note that the approach in [12], who—roughly speaking—measures inconsistency in a multiset of knowledge bases by performing a multiset union on all sets and then measuring inconsistency on this union, is not applicable for our use-case, as we are not interested in the disagreement between the individual instances, but rather want to gain an overview of inconsistencies in all instances. While the above discussion showed how existing means can be used to measure the overall degree of inconsistency for a multiset of rule bases, in the following, we develop techniques for an element-based assessment of inconsistency over a multiset of rule bases. § CULPABILITY MEASURES FOR MULTISETS OF BUSINESS RULE BASES In the field of inconsistency measurement, a culpability measure $\culp$ [8] is a function that assigns a non-negative numerical value to elements of a rule base. This quantitative assessment is also referred to as an inconsistency value. Again, the intuition is that a higher inconsistency value reflects a higher blame, that the specific element carries in the context of the overall inconsistency. In this section, we investigate culpability measures that can assess the blame that a rule carries in the context of the overall multiset inconsistency. §.§ Baseline measures and basic properties Given a multiset of business rule bases $\ms$, let $\R(\ms)$ denote the shared rule set of the respective business rule bases in $\ms$. Furthermore, let $\allrs_\allmss$ denote the set of all possible rules that can appear in these shared rule sets. Then, a culpability measure for a multiset of rule bases is defined as follows. A culpability measure for a multiset of rule bases is a function $\culp^m:\allmss \times \allrs_\allmss \rightarrow \posRealInf$. Similar to $\Sigma$-induced inconsistency measures, existing culpability measures can be exploited to entail $\Sigma$-induced culpability measures. Given a culpability measure $\culp$, a multiset of rule-bases $\ms$ and a rule $r\in\R(\ms)$, a $\Sigma$-induced multi-rb culpability measure $m_{\culp}^{\Sigma}$ is defined as $m_{\culp}^{\Sigma}:\allmss \times \allrs_\allmss \rightarrow \posRealInf$ with $m_{\culp}^{\Sigma}(\ms,r)=\sum_ {B\in\ms}\culp(B,r)$. Two baseline measures proposed in [9] are the $\culp_D$ and $\culp_{\#}$ measures. Let a rule base $\rb$ and a rule $r\in\rb$, then * $\culp_D(\rb,r) = \begin{cases} 1 & \text{if } \exists M \in \MI(\rb):r\in M \\ 0 & \, \text{otherwise} \end{cases}$ * $\culp_{\#}(\rb,r) = |\{M \in \MI(\rb) \mid r \in M\}|$ Using $\Sigma$-induction, we can use these baseline culpability measures to entail the multi-rb culpability measures $m_{\culp_D}^{\Sigma}$ and $m_{\culp_{\#}}^{\Sigma}$. We recall the multiset of rule bases $\ms_1$ from Figure <ref>. For the shown rule $a\rightarrow b$, we have that \begin{align*} m_{\culp_D}^{\Sigma}(\ms_1, a\rightarrow b) & = 4\\ m_{\culp_{\#}}^{\Sigma}(\ms_1, a\rightarrow b) & = 5 \end{align*} Regarding multi-rb culpability measures, we propose the following rationality postulates based on an application of postulates for traditional culpability measures [8]. For that, we consider a multiset of business rules $\ms$ and a rule $r\in\R(\ms)$. Also, we define a rule $r\in\R(\ms)$ as a free formula if $r \notin M, \forall M \in \bigcup_{b\in\ms}\MI(b)$. We denote the set of all free formulas of $\R(\ms)$ as $\Free(\ms)$. We then propose the following postulates. Rule Symmetry () $\culp^m(\ms,r)=\culp^m((\rb_{1},...\rb_n),r)$, for any permutation of the order of $\rb_1$ to $\rb_n$. Rule Minimality () if $r\in \Free(\ms)$, then $\culp^m(\ms,r)=0$. The first postulate states that the order of rule bases in the multiset should not affect the inconsistency value of an individual rule. The second postulate states that the inconsistency value of a rule is zero if this rule is a free formula w.r.t. the multiset of business rule bases. $m_{\culp_D}^{\Sigma}$ and $m_{\culp_{\#}}^{\Sigma}$ satisfy and . The second postulate was adapted from a postulate for traditional culpability measures, namely Minimality ()Let a rule $r\in\rb$, if $r\not\in M, \forall M \in \MI(\rb)$, then the inconsistency value of r is zero. As this is a commonly satisfied postulate, this allows for a generalization of the previous proposition. Any $\Sigma$-induced multi-rb culpability measure satisfies . Given a culpability measure $\culp$ satisfying , any multi-rb culpability measure $\Sigma$-induced via $\culp$ satisfies . In the following, given a multiset of business rules $\ms$ and a multi-rb culpability measure $\culp^m$, we consider all rules of $\R(\ms)$ as a vector $(r_1,...r_n)$, and denote $V^{\culp^m}(\ms)$ as the vector of corresponding multi-rb culpability values of all rules in $\R(\ms)$ w.r.t. $\culp^m$, i.e., $V^{\culp^m}(\ms) = (\culp^m(\ms,r_1),...,\culp^m(\ms,r_n))$. Next, let $\hat{V}^{\culp^m}(\ms)=\mathit{max}_{r\in\R(\ms)}(\culp^m(\ms,r))$ denote the largest multi-rb culpability value w.r.t. $\culp^m$ for all rules. Last, we denote adding a rule $r$ to the shared rule set $\R(\ms)$ of a multiset $\ms$ as $\ms\cup \{r\}$ by a slight missuse of notation, i.e., given $\ms = (\rb_1,...,\rb_n), \ms\cup \{r\}= (\rb_1 \cup \{r\},...,\rb_n\cup \{r\})$. This allows to adapt some further desirable properties. Multiset Consistency () $\hat{V}^{\culp^m}(\ms)=0$ iff $\nexists \rb\in\ms:\rb\models\perp$. Multiset Monotony () Let a multiset of business rule bases $\ms$ and a rule r, $\hat{V}^{\culp^m}(\ms \cup \{r\}) \geq \hat{V}^{\culp^m}(\ms)$ Multiset Free formula independence () If a rule r is a free formula of $(\ms\cup\{r\})$, then $\hat{V}^{\culp^m}(\ms \cup r) = \hat{V}^{\culp^m}(\ms)$ The first property states that the largest multi-rb culpability value for a rule can only be zero if all business rule bases of the multiset are consistent. The second property demands that adding a rule to the shared rule set can only increase the culpability values. Similar to this property, the third postulate demands that adding a free formula to the shared rule set does not alter the culpability values. $m_{\culp_D}^{\Sigma}$ and $m_{\culp_{\#}}^{\Sigma}$ satisfy , and . Next to the introduced baseline culpability measures $\culp_D$ and $\culp_{\#}$, various other culpability measures have been proposed (cf. [10]), which could also be used to $\Sigma$-induce multi-rb culpability measures. While an analysis of such measures w.r.t. the introduced postulates could be interesting, we refrain from such a specific analysis and rather show a more generalized approach in the following section, namely how arbitrary inconsistency measures can be used to induce multi-rb culpability measures using Shapley inconsistency values. §.§ (Adjusted) Shapley Inconsistency Values for Multi-RB measures Next to designing specific culpability measures, an important approach in element-based analysis is to decompose the assessment of inconsistency measures (in order to derive corresponding culpability measures) by means of Shapley inconsistency values [8]. Given an inconsistency measure $\inc$ and a rule base $\rb$, the intuition is that the overall blame mass $\inc(\rb)$ is distributed amongst all elements in $\rb$, by applying results from game theory. The advantage of this approach is that arbitrary inconsistency measures can be applied to derive a corresponding element-based assessment. The amount of blame that an individual element is assigned relative to $\inc(\rb)$ is also referred to as the payoff. Let $\inc$ be an inconsistency measure, $\rb$ be a rule base and $\alpha \in \rb$. Then, the Shapley inconsistency value of $\alpha$ w.r.t. $\inc$, denoted $S_\alpha^{\inc}$ is defined via \begin{align*} S_\alpha^{\inc}(\rb) = \sum_{B\subseteq\rb}\frac{(b-1)!(n-b)!}{n!}(\inc(\rb)-\inc(\rb\setminus \alpha)) \end{align*} where $b$ is the cardinality of $B$, and $n$ is the cardinality of $\rb$. Consider the rule base $\rb_2 = \{a,a\rightarrow b,a\rightarrow \neg b\}$. Then, for the Shapley inconsistency values w.r.t. $\incmi$, for all elements $e$ in $\rb$ we have that $S^{\incmi}_{e}(\rb_2) = \frac{1}{12}+\frac{1}{4} = \frac{1}{3}$. As shown in Example <ref>, all elements were assigned an equal payoff. This makes sense w.r.t. $\incmi$ if all elements in the rule base are considered with equal importance. However, in our setting, knowledge contained in rule bases is distinguished into facts and rules. Here, facts have a different veracity than rules, as they are usually provided by a given (non-negotiable) case input and have to be kept "as-is" [6]. In the scope of inconsistency resolution, we are therefore only interested in identifying blamable rules, as these should be considered for re-modelling. Consequently, an element-based assessment for rule bases should only assign a payoff to blamable rules, and not facts. Correspondingly, this has to be considered when applying Shapley's game theoretical approach to distribute a blame mass over all elements. As recently discussed in [3], the Shapley inconsistency value can accordingly be adjusted as follows. For that, let $\Free(B)$ denote the free formula in a rule base $B$, i.e., all $r\in B: r\not\in M, \forall M\in\MI(B)$. Let $\inc$ be a rule-based inconsistency measure, $\rb$ be a rule base and $\alpha \in \rb$. Then, the adjusted Shapley inconsistency value of $\alpha$ w.r.t. $\inc$, denoted $S*_\alpha^{\inc}$ is defined via \begin{align*} \begin{cases} 0 & \text{if } \alpha \in \F(\rb) \\ \sum\limits_{B\subseteq\rb}( \mathit{CoalitionPayoff}_{\alpha,\rb}^{\inc}(B) + \mathit{AdditionalPayoff}_{\alpha,\rb}^{\inc}(B) ) & \text{otherwise} \end{cases}\\ \intertext{with} &\mathit{CoalitionPayoff}_{\alpha,\rb}^{\inc}(B) = \frac{(b-1)!(n-b)!}{n!}(\inc(B)-\inc(B\setminus \alpha))\\ \intertext{being the payoff for an element for any coalition $B\subseteq\rb$, and} = \left\{\begin{array}{ll} 0& \text{if } r \in \Free(B)\\ \frac{\sum_{f\in\F(B)}\mathit{CoalitionPayoff}_{f,\rb}^{\inc}(B)}{|r'\in \R(B) \text{ s.t. } r' \notin \Free(B)|} & \text{otherwise} \end{array}\right. \intertext{being the additional payoff that blamable rules receive, by shifting the blame mass from (given) facts to blamable rules.} \end{align*} Consider the rule bases $\rb_{2} = \{a, a \rightarrow b, a \rightarrow \neg b\}$ and $\rb_{3} = \{a, a \rightarrow b, \neg b\}$. Then for the adjusted Shapley inconsistency values w.r.t. $\incmi$, we have that $S*_a^{\incmi}(\rb_2) = 0, S*_{a\rightarrow b}^{\incmi}(\rb_2) = \frac{1}{3} (+\frac{1}{3}/2) = \frac{1}{2}$, and $S*_{a\rightarrow \neg b}^{\incmi}(\rb_2) = \frac{1}{3} (+\frac{1}{3}/2) = \frac{1}{2}$. Also, we have that $S*_a^{\incmi}(\rb_3) = 0, S*_{\neg b}^{\incmi}(\rb_3) = 0$, and $S*_{a\rightarrow \neg b}^{\incmi}(\rb_3) = \frac{1}{3} (+\frac{2}{3}) = 1$ . For the first rule base $\rb_2$ (containing two rules), the blame is evenly distributed amongst both rules. For the second rule base $\rb_3$, the single rule receives the entire blame. This assessment makes sense in a business rule setting, as the given fact input is evaluated against a set of humanly modelled rules, and any inconsistencies arise due to modelling errors in the set of business rules. The adjusted Shapley inconsistency values can thus be used for pin-pointing problematic rules for re-modelling purposes. This element-based measure can consequently also be used to identify problematic rules over a multiset of cases, by inducing the corresponding multi-rb measure $m^\Sigma_{S*^I}$ (cf. Example <ref>). As mentioned, an advantage of using the adjusted Shapley measure is that arbitrary inconsistency measures can be used to derive an element-based assessment over a set of cases, based on company needs. Furthermore, we can identify the following properties for an adjusted Shapley value $m^\Sigma_{S*^I}$. For this, we assume that any inconsistency measure $\inc$ used to derive adjusted Shapley inconsistency values satisfies the basic properties of consistency', monotony' and free formula independence' as defined in [8][Let a rule base $\rb$ and an inconsistency measure $\inc$. Consistency' states that $\inc(\rb)=0$ iff $\rb$ is consistent. Monotony' states that if $\rb\subseteq \rb'$ then $\inc(\rb)\leq \inc(\rb')$. Free formula independence' states that If $\alpha\in\Free(\rb)$ then $\inc(\rb)=\inc(\rb\setminus\{\alpha\})$.]. The adjusted multi-rb shapley inconsistency value satisfies , and . Also, regarding the relation of inconsisteny measures and the corresponding $\Sigma-$induced Shapley inconsistency values for multi-rb analysis, we propose the following postulates. Distribution ()$\sum_{\alpha\in\R(\ms)}m^\Sigma_{S*^I}(\ms,\alpha) = m^\Sigma_I(\ms)$ Upper Bound () $\hat{V}^{S*^I}(\ms) \leq m^\Sigma_I(\ms)$ The first postulate states that the sum adjusted multi-rb Shapley inconsistency values over all rules is equal to the overall blame mass of the original multi-rb inconsistency measure $\inc$ (used as a parameter to derive the corresponding Shapley values). Also, the second property states that the adjusted multi-rb Shapley inconsistency values for an individual element cannot be greater than the overall assessment of the original multi-rb inconsistency measure $\inc$. As we are only interested in identifying problematic rules (e.g. for re-modelling), it would also be plausible to adapt the property of fact minimality as proposed in [3] for a multi-rb use-case. Fact Minimality ()$m^\Sigma_{S^I}(\ms,\alpha) = 0 \forall \alpha\notin\R(\ms)$. This property states that (non-negotiable) facts should not be assigned any blame value in an element-based multi-rb assessment. The adjusted multi-rb shapley inconsistency value satisfies , and . We conclude with an example illustrating the introduced multi-rb measures. We recall the set of business rule bases $\ms_1$ from Figure <ref> and its shared rule set $\R(\ms_1)=\{r_1,...,r_6\}$. A multi-rb assessment w.r.t. the introduced measures is then as follows. \begin{align*} r_1: a &\rightarrow b & m_{\culp_D}^{\Sigma}(\ms_1, r_1) &= 4 & m_{\culp_{\#}}^{\Sigma}(\ms_1, r_1)&= 5 & m_{S*^{\incmi}}^{\Sigma}(\ms_1, r_1)&= 2\\ r_2: c &\rightarrow \neg b & m_{\culp_D}^{\Sigma}(\ms_1, r_2) &= 3 & m_{\culp_{\#}}^{\Sigma}(\ms_1, r_2)&= 3 & m_{S*^{\incmi}}^{\Sigma}(\ms_1, r_2)&= 1.5\\ r_3: b &\rightarrow x & m_{\culp_D}^{\Sigma}(\ms_1, r_3) &= 2 & m_{\culp_{\#}}^{\Sigma}(\ms_1, r_3)&= 2 & m_{S*^{\incmi}}^{\Sigma}(\ms_1, r_3)&= 0.5\\ r_4: x &\rightarrow z & m_{\culp_D}^{\Sigma}(\ms_1, r_4) &= 2 & m_{\culp_{\#}}^{\Sigma}(\ms_1, r_4)&= 2 & m_{S*^{\incmi}}^{\Sigma}(\ms_1, r_4)&= 0.5\\ r_5: y &\rightarrow \neg z & m_{\culp_D}^{\Sigma}(\ms_1, r_5) &= 2 & m_{\culp_{\#}}^{\Sigma}(\ms_1, r_5)&= 2 & m_{S*^{\incmi}}^{\Sigma}(\ms_1, r_5)&= 0.5\\ \cline{1-8} & & & & && m_{\incmi}^{\Sigma}(\ms_1)&= 5 %& & \footnotesize{\emph{Sum:}} & \incmiQ(\rb_{13})&= 2 & \incpQ(\rb_{13})&= 6 \end{align*} As can be seen, rule $r_1$ is classified as most problematic my all measures. This makes sense, as this rule is a cause of inconsistency over all cases for $\ms_1$. Hence, this rule should be prioritized in the scope of re-modelling and improving the set of business rules. Here, our proposed approach of multi-case inconsistency measurement can support modellers in identifying highly problematic rules from a global perspective, by recommending an order in which rules should be attended to, e.g. $<r1,r2,r3,r4,r5>$ for the shown example. A further important aspect for an application in practice is that the proposed measures can be combined to obtain multivariate metrics for a more fine-grained analysis. For example, as the $m_{\culp_D}^{\Sigma}(r)$ is equivalent to the number of distinct cases in which a rule $r$ is part of an inconsistency, this measure can be used to normalize and explain other measures. For example, a normalization via $m_{\culp_D}^{\Sigma}$ can be used to explain if a high $m_{\culp_{\#}}^{\Sigma}$-value originates from a few highly inconsistent cases (which might be outliers) or if the corresponding rule contributes a smaller amount towards inconsistency but in a vast majority of cases (in which case it might be sensible to re-consider this rule). In this section, we have shown how arbitrary culpability measures can be transformed into multi-rb measures. Also, we have shown that the proposed measures satisfy desirable properties. Our results are summarized in Table <ref>[Proofs can be found in the supplementary document (<https://bit.ly/2V2sIDw>)]. $m^\Sigma_{\culp_D}$ n/a n/a $m^\Sigma_{\culp_{\#}}$ n/a n/a $m^\Sigma_{S^{I}}$ $^{c,i}$ $^{c,i}$ $^{m}$ $^{i}$ c: If $\inc$ satisfies consistency' m: If $\inc$ satisfies monotony' i: If $\inc$ satisfies free formula independence' Compliance with rationality postulates of the investigated measures. § TOOL SUPPORT AND EVALUATION We implemented our approach to assess element-based inconsistency over sequences of business rule cases[<https://gitlab.uni-koblenz.de/fg-bks/multi-rb-inconsistency-measurement/>]. Our implementation takes as input a shared business rule base and a sequence of fact sets, and can then computes the most problematic rules w.r.t. the multiset of rule bases. The $m^\Sigma_{\culp_D}$ and $m^\Sigma_{\culp_{\#}}$ measures can be used out-of-the-box, however, arbitrary culpability measures can be added based on company needs. To evaluate our tool, we then performed run-time experiments and investigated the computational complexity regarding various aspects of measuring inconsistency over sequences of business rule cases. §.§ Run-Time Experiments In the following, we present the results of run-time experiments with real-life and synthetic data-sets. §.§.§ Evaluation with real-life data sets. To evaluate the feasibility of applying our approach in practice, we conducted run-time experiments with real-life data sets of the Business Process Intelligence (BPI) challenge[<https://data.4tu.nl/search?q=bpi+challenge>]. This yearly scientific challenge from the field of process management provides real-life process logs for evaluating approaches in an industrial setting. In a nutshell, we mined a rule set from each event log and then measured inconsistency over all cases of the respective log. Here, we analyzed the data-sets from the last four years, i.e., BPI'17 (log of a loan application process with 31,509 cases), BPI'18 (log of a fund distribution process with 43,809 cases), BPI'19 (log of an application process with 251,734 cases), and BPI'20 (log of a travel expense claim process with 10,500 cases). From these event logs, declarative constraints of the general business rule form in (<ref>) can be mined using the results from [5]. In this way, we were able to mine a rule set from each of the provided data sets. The resulting number of rules for the respective rule sets is provided in Table <ref>. Also, the mined rule sets can be found online[<https://bit.ly/365Vs4C>]. We refer the reader to [5] for further details on the mining technique. Then, for each data set, we analyzed all cases as follows: For each data set, a shared rule base $R$ was mined as described above. Then, for all cases $C_1,...,C_n$, the individual case-dependent fact inputs $F_1,...,F_n$ were extracted from the log. We then constructed a multiset of rule bases $B_1,...,B_n$, where every $B_i=(R,F_i)$. We then applied our implementation to analyze inconsistencies over $B_1,...,B_n$ and measured the run-time. The results of our experiments are shown in Table <ref>. The experiments were run on a machine with 3 GHz Intel Core i7 processor, 16 GB RAM (DDR3) under macOS. Dataset # of Rules # of Cases Runtime # of inconsistent Cases BPI'17 50 31.509 5657s 31.509 (100%) BPI'18 84 43.809 3967s 0 (0%) BPI'19 51 251.734 1610s 434 (0.17%) BPI'20 330 10.500 1329s 323 (3.07%) Runtimes for analyzing all cases for the considered BPI data sets As can be seen, an analysis of all cases was feasible for all data sets. A central assumption of our approach is that a global perspective over all cases should be considered as opposed to viewing cases individually. Interestingly, this was also confirmed by our experiments with the above real-life data sets: For every individual rule base instance, we computed the $\culp_{\#}$ values for all rules and then ranked all rules by this value (rank 1 meaning that this rule is the most problematic element, and so on). If $n$ rules had the same $\culp_{\#}$ value, they were assigned the sum of the occupied ranks divided by $n$ (e.g. if two rules had the highest $\culp_{\#}$ value, they were awarded the rank (1+2)/2 = 1.5, and the next rule had the rank 3). Figure <ref> shows the distribution of all assigned ranks for the rules for the BPI'17 data set over all cases. For readability, rules that did not participate in any inconsistencies are omitted. [col sep=comma]boxPlotData/bpi17.txt boxplot/draw direction = y, x axis line style = opacity=0, axis x line* = bottom, axis y line = left, enlarge y limits, xtick = 1, 2, 3, 4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26, xticklabel style = align=center, font=, rotate=60, xticklabels = r1,r2,r3,,$\cdots$,,,,,,,,,,,,,,,,,,,,,r26, xtick style = draw=none, ylabel = Rank, ytick = 1,10,20,30, in 1,...,26 +[boxplot, draw=black] table[y index=] ; Rank distribution for the individual rules of the BPI'17 rule set over all cases. While there were some rules that had the same rank in all cases (e.g. r2), there were many rules where the respective local rankings had a large variability (e.g. r3). This shows that the global perspective as proposed in this work should be strongly considered in the scope of auditing. In general, we see the above experiments as positive in regard to applying our approach in practice. As the analyzed data-sets were unrelated, no further comparison of run-times can be made. Therefore, we further assess our approach with synthetic data sets. §.§.§ Evaluation with synthetic data sets. We created a generator for synthetic rule base instances. Our generator can produce set of business rule base instances, based on a shared rule set and a sequence of fact sets relative to this rule set. As parameters, our generator takes the desired rule base size and a desired number of cases. Then, the generator constructs a multiset of rule bases as follows: The set of business rule instances is constructed over a (potentially infinite) alphabet $\mathfrak{A}=<a,b,...>$. Then, for a desired number of rules $n_r$, a rule set $R=\{r_1,...,r_{n_r}\}$ is generated, where every $r_i$ is of the form $\mathfrak{A}_i \rightarrow \neg\mathfrak{A}_{i+1}$. For example, for a parameter of $2$ desired rules, the resulting rule set is $R=\{a\rightarrow\neg b, b\rightarrow \neg c\}$. To then generate a desired number of random cases $n_c$, a multiset of fact sets $F=\{F_1,...,F_{n_c}\}$ is initialized. Then, each of the fact sets is populated by adding atoms of the rule base, based on a user-defined probability. For example, for the exemplary rule set $R$ of size $2$, a random fact set relative to $R$ could be any element of $\{\emptyset,a,b,c,ab,ac,bc,abc\}$. In this way, the generator can create a set of random rule base instances $B=\{B_1,...,B_{n_c}\}$, where every $B_i=(R,F_i)$. An advantage of our generator is that the structure of the contained rules is similar in all rule bases, which thus allows for better comparability. We consequently used our generator to analyze multiple sets of business rule cases with different parameters and measured the run-times. As parameter settings, based on the observed sizes from the real-life data-sets, we selected as parameters the rule base size from 10,20,...,100, and the number of cases from 10.000,20.000,...,100.000 and then tested every possible combination. Thus, a total of 100 (10x10) different configurations were tested (see above for hardware). The results of our experiments are shown in Figure <ref>. The smallest setting (10 rules and 10.000 cases) took around 90s, where the largest configuration (100 rules and 100.000 cases) took around 50 minutes. As can be seen, the run-time scales proportionally with the size of the rule base and the number of cases. Thus, we could not identify any of these two factors to be a dominant limiting factor to the run-times in our experiments. scaled y ticks=false [xlabel=Rule base size, ylabel=Number of cases, zlabel=Run-time in seconds, xlabel style=sloped like x axis, ylabel style=sloped,yticklabels=1,2,50T,100T,5,6,7,8,9,10, view/az=35, zmin=0, grid=major] 3[surf, shader=faceted] Run-times for the analysis of 10x10 synthetic sets of rule base instances To summarize, both the evaluation with real-life data sets and with synthetic data sets yielded feasible run-times. To extend this empirical analysis, we continue with an investigation of computational complexity in regard to our proposed approach. §.§ Complexity Analysis[Proofs can be found in the supplementary document (<https://bit.ly/2V2sIDw>)] We assume familiarity with basic concepts of computational complexity and basic complexity classes such as $\textsf{P}$ and $\textsf{NP}$, see [11] for an introduction. We first observe that the satisfiability problem for our formalism of business rules bases is tractable (note that similar observations have been made before on similar formalisms, see e. g., [4]). Let $\rb$ be a rule base. The problem of deciding whether $\rb$ is consistent can be solved in polynomial time. Then, the complexity of deciding whether a certain rule is contributing to the overall inconsistency is as follows. Let $\ms$ be a multiset of rule bases with $\ms=(\rb_1, ..., \rb_n)=(\{\F_1\cup\R\},...,\{\F_n\cup\R\})$ and let $r\in \rb$. The problem of deciding whether there is a $i\in\{1,\ldots,n\}$ and $M\in\MI(\rb_i)$ s. t. $r\in M$ is -complete. The following two results deal with the computational complexity of computing the baseline measure $\culp_{\#}$. Let $\rb$ be a rule base and $M\subseteq \rb$. The problem of deciding whether $M\in\MI(\rb)$ can be solved in polynomial time. For our final result note that $\#\textsf{P}$ is the complexity class of counting problems where the problem of deciding whether a particular element has to be counted is in $\textsf{P}$, cf. [15]. Let $\ms$ be a multiset of rule bases with $\ms=(\rb_1, ..., \rb_n)=(\{\F_1\cup\R\},...,\{\F_n\cup\R\})$ and let $r\in \R$. The problem of determining $|\{M \in \MI(\rb_i) \mid i\in \{1,\ldots,n\} r \in M\}|$ is $\#\textsf{P}$-complete. § CONCLUSION In this work, we have shown how arbitrary culpability measures (for single rule bases) can be automatically transformed into multi-rb measures while maintaining desirable properties. This is highly needed in practice, as companies are often faced with thousands of rule bases daily, and thus need means to assess inconsistency from a global perspective. Here, our proposed measures can be used to gain fine-grained insights into inconsistencies in sequences of business rule bases. For the analyzed (real-life) data-sets, the proposed multi-cases analysis could be performed in a feasible run-time. Intuitively, the number of cases or the size of the rule base affect the run-time of our approach. Here, we plan to develop more efficient algorithms in future work. As a main takeaway, our results indicate that the interrelations of individual cases need to be considered for business rules management, which should be addressed more in future works. [1] Corea, C., Deisen, M., Delfmann, P.: Resolving inconsistencies in declarative process models based on culpability measurement. In: 15. Internationale Tagung Wirtschaftsinformatik, WI 2019 (2019) [2] Corea, C., Thimm, M.: On quasi-inconsistency and its complexity. AI 284 (2020) [3] Corea, C., Thimm, M.: Towards inconsistency measurement in business rule bases. In: Proceedings of the 24th European Conference on Artificial Intelligence (ECAI 2020), Santiago de Compostela, Spain, 2020 (2020) [4] Dantsin, E., Eiter, T., Gottlob, G., Voronkov, A.: Complexity and expressive power of logic programming. In: Proceedings of the 12th Annual IEEE Conference on Computational Complexity (CCC'97). pp. 82–101 (1997) [5] Di Ciccio, C., Maggi, F.M., Montali, M., Mendling, J.: Resolving inconsistencies and redundancies in declarative process models. Inf. Systems 64, 425–446 (2017) [6] Graham, I.: Business rules management and service oriented architecture: a pattern language. John wiley & sons (2007) [7] Grant, J., Martinez, M.V. (eds.): Measuring Inconsistency in Information. College Publications (2018) [8] Hunter, A., Konieczny, S.: On the measure of conflicts: Shapley inconsistency values. Artificial Intelligence 174(14), 1007–1026 (2010) [9] Hunter, A., Konieczny, S., et al.: Measuring inconsistency through minimal inconsistent sets. KR 8, 358–366 (2008) [10] McAreavey, K., Liu, W., Miller, P.: Computational approaches to finding and measuring inconsistency in arbitrary knowledge bases. International Journal of Approximate Reasoning 55(8), 1659–1693 (2014) [11] Papadimitriou, C.: Computational Complexity. Addison-Wesley (1994) [12] Potyka, N.: Measuring disagreement among knowledge bases. In: International Conference on Scalable Uncertainty Management. pp. 212–227. Springer (2018) [13] Thimm, M.: On the compliance of rationality postulates for inconsistency measures: A more or less complete picture. KI 31(1), 31–39 (2017) [14] Thimm, M.: Inconsistency measurement. In: Proceedings of the 13th International Conference on Scalable Uncertainty Management (SUM'19) (2019) [15] Valiant, L.: The complexity of computing the permanent. Theoretical Computer Science 8, 189–201 (1979) § APPENDIX A: PROOFS OF TECHNICAL RESULTS $m_{\culp_D}^{\Sigma}$ and $m_{\culp_{\#}}^{\Sigma}$ satisfy and . We consider $m_{\culp_D}^{\Sigma}$ and $m_{\culp_{\#}}^{\Sigma}$ in turn. For this, let $\ms$ be a multiset of rule bases, $B$ any rule base in $\ms$, and $r$ any rule in a rule base $B$. * We start with the measure $m_{\culp_D}^{\Sigma}$. To show , as $m_{\culp_D}^{\Sigma}(\ms,r)=\sum_ {B\in\ms}\culp_D(B,r)$, we have that $m_{\culp_D}^{\Sigma}(\ms,r)=m_{\culp_D}^{\Sigma}((\rb_{1},...\rb_n),r)$, for any permutation of the order of $\rb_1$ to $\rb_n$ due to commutativity via $\sum_{B_i\in\ms}\culp_D(B_i,r) = \culp_D(B_1,r)+ ... + \culp_D(B_n,r)$, with $n=|\{B\in\ms\}|$. For , recall that a rule $r$ is defined as a free formula in $\ms$ if $r \notin M, \forall M \in \bigcup_{b\in\ms}\MI(b)$. Consequently, if $r\in\Free(\ms)$, then $\sum_ {B\in\ms}\culp_D(B,r) = 0$ per definition. * The proofs for $m_{\culp_{\#}}^{\Sigma}$ are analogous, i.e., $m_{\culp_{\#}}^{\Sigma}(\ms,r)=m_{\culp_{\#}}^{\Sigma}((\rb_{1},...\rb_n),r)$, for any permutation of the order of $\rb_1$ to $\rb_n$ due to commutativity, and $\sum_ {B\in\ms}\culp_{\#}(B,r) = 0$ if $r\in\Free(\ms)$, as $\culp_{\#}(B,r) = |\{M \in \MI(B) \mid r \in M\}|$ for any $\rb\in\ms$. Any $\Sigma$-induced multi-rb culpability measure satisfies . Given a culpability measure $\culp$ satisfying , any multi-rb culpability measure $\Sigma$-induced via $\culp$ satisfies . Let $\ms$ be a multiset of rule bases, $B$ any rule base in $\ms$, and $r$ any rule in a rule base $B$. To show , as $m_{\culp^m}^{\Sigma}(\ms,r)=\sum_ {B\in\ms}\culp^m(B,r)$, $\culp^m(\ms,r)=\culp^m((\rb_{1},...\rb_n),r)$ for any permutation of the order of $\rb_1$ to $\rb_n$ due to commutativity via $\sum_{B_i\in\ms}\culp^m(B_i,r) = \culp^m(B_1,r)+ ... + \culp^m(B_n,r)$, with $n=|\{B\in\ms\}|$. To show , given a culpability measure $\culp$, if we have for a rule $r\in B$: $r\not\in M, \forall M \in \MI(B)$ and $\culp(B,r)=0$ , then $\culp^m = \sum_ {B\in\ms}\culp(B,r) = 0$ per assumption. $m_{\culp_D}^{\Sigma}$ and $m_{\culp_{\#}}^{\Sigma}$ satisfy , and . We consider $m_{\culp_D}^{\Sigma}$ and $m_{\culp_{\#}}^{\Sigma}$ in turn. For this, let $\ms$ be a multiset of rule bases, $B$ any rule base in $\ms$, and $r$ any rule in a rule base $B$. * We start with the measure $m_{\culp_D}^{\Sigma}$. To show , observe that $\culp_D(X,r)$ for a consistent rule base $X$ is $0$ for any rule $r$ per definition, as $\MI(X)=\emptyset$. Thus, if $\nexists B\in\ms:B\models\perp$, then $m_{\culp_D}^{\Sigma}(\ms)=0$. In turn, $\hat{V}^{m_{\culp_D}^{\Sigma}}(\ms)=\mathit{max}_{r\in\R(\ms)}(\culp_D(\ms,r))=0$. For the other direction, assume that $\hat{V}^{m_{\culp_D}^{\Sigma}}(\ms)>0$ for a case where $\nexists \rb\in\ms:\rb\models\perp$. This would mean, that there must exist a rule $r$ in a consistent rule base $X$ of $\ms$, s.t. $\culp_D(X,r)\neq 0$. This contradicts $\culp_D$ by definition. To show , observe that if a rule $r'$ is added, for any rule base $B_i$, we have that $\culp_D(B_i,r')= 0$ or $1$. Thus, $m_{\culp_D}^{\Sigma}(\ms\cup \{r'\})\geq m_{\culp_D}^{\Sigma}(\ms)$. Hence, $\hat{V}^{m_{\culp_D}^{\Sigma}}(\ms \cup \{r'\})=\mathit{max}_{r\in\R(\ms)}(\culp_D(\ms\cup \{r'\},r)) \geq \hat{V}^{m_{\culp_D}^{\Sigma}}(\ms)$. To show , recall that a rule $r'$ is defined as a free formula in $\ms$ if $r' \notin M, \forall M \in \bigcup_{b\in\ms}\MI(b)$. Consequently, if $r'\in \Free(\ms\cup \{r'\})$, then $|\MI(B_i\cup \{r'\})|=|\MI(B_i)|$ for all $B\in\ms$. In turn, for any other rule $r$ in a rule base $B_i\in\ms$, $\culp_D(B_i,r)=\culp_D(B_i\cup\{r'\},r)$. In result, if $r'\in \Free(\ms\cup\{r'\})$, then $\hat{V}^{m_{\culp_D}^{\Sigma}}(\ms \cup \{r'\}) = \hat{V}^{m_{\culp_D}^{\Sigma}}(\ms)$. * The proofs for $m_{\culp_{\#}}^{\Sigma}$ are analogous, i.e., if $\nexists B\in\ms:B\models\perp$ then $m_{\culp_{\#}}^{\Sigma}(\ms)=0$ as $|\MI(B_i)|=0$ for any $B_i\in\ms$, $|\MI(B\cup \{r\})|\geq |\MI(B)|$ for any rule $r$, resp. $|\MI(B\cup \{r\})| = |\MI(B)|$ if $r$ is a free formula in $(B\cup \{r\})$. The adjusted multi-rb shapley inconsistency value satisfies , , (, and ). Let $\ms$ be a multiset of rule bases, $B$ any rule base in $\ms$, and $r$ any rule in a rule base $B$. Also, recall that we assume any inconsistency measure $\inc$ that is used to derived an adjusted multi-rb shapley inconsistency value $m^\Sigma_{S*^{I}}$ satisfies consistency', monotony' and free formula independence'. We then consider the individual properties in turn. To show , observe that for a consistent rule base $B$, we have that $\inc(B)=0$ per assumption of consistency'. It follows that for a consistent rule base $B$, $S*_\alpha^{\inc}(B)=0$ for any $\alpha\in B$, thus $\hat{V}^{m^\Sigma_{S*^{I}}}(\ms)=\mathit{max}_{r\in\R(\ms)}(m^\Sigma_{S*^{I}}(B,r))=0$ if $\nexists \rb\in\ms:\rb\models\perp$. For the only if direction, assume we would have a rule $\alpha$ s.t. $S*_\alpha^{\inc}(B)\neq0$ for a consistent rule base $B$. This would mean $\inc(B)-\inc(B\setminus\alpha)>0$ for a consistent rule base $B$, which contradicts the assumption of consistency'. To show , observe that $\inc(B\cup \{r\}) \geq \inc(B)$ for any rule $r$ if $\inc$ satisfies monotony'. Therefore, $S*_\alpha^{\inc}(B\cup \{r\})\geq S*_\alpha^{\inc}(B)$ for any $\alpha\in B$. It follows that $m^\Sigma_{S*^I}(\ms \cup \{r\},\alpha)\geq m^\Sigma_{S*^I}(\ms,\alpha)$ for any rule $\alpha$, and thus $\hat{V}^{m^\Sigma_{S*^{I}}}(\ms\cup\{r\})\geq \hat{V}^{m^\Sigma_{S*^{I}}}(\ms)$. The proof for is analogous, i.e., $\inc(\MI(B\cup \{r\})) = \inc(\MI(B))$ for any rule $r\in\Free(B\cup\{r\})$ if $\inc$ satisfies free formula independence'. Next, follows from Proposition <ref>. Last, to show , it suffices to show that $S*^{\inc}$ satisfies minimality, which has been shown in [3], i.e., for any fact $f$, $S*_f^{\inc}=0$ per definition, and for any free rule $\alpha$, $\inc(B)-\inc(B\setminus\alpha)$ in the last part of the summand of the coalition payoff will always equate to $0$ due to the consistency' and free formula independence' assumption of the underlying measure $\inc$, thus, for any $r_i\in\R(\rb)$, if $r_i\not\in M, \forall M \in \MI(\rb)$, then $S*_{r_i}^{\inc}=0$. The adjusted multi-rb shapley inconsistency value satisfies , and . Let $\ms$ be a multiset of rule bases, $B$ any rule base in $\ms$, and $r$ any rule in a rule base $B$. We then consider the individual properties in turn. The proof for follows the proof in [3]: We recall the adjusted Shapley inconsistency value \begin{align*} \scriptsize{ \begin{cases} 0 & \text{if } \alpha \in \F(\rb) \\ \sum\limits_{B\subseteq\rb} \mathit{CoalitionPayoff}_{\alpha,\rb}^{\inc}(B) + \sum\limits_{B\subseteq\rb}\mathit{AdditionalPayoff}_{\alpha,\rb}^{\inc}(B) & \text{otherwise} \end{cases} \end{align*} In the following, we abbreviate CoalitionPayoff as CP and AdditionalPayoff as AP for readability. Then, for a set of rule bases $\ms$, we consider the sum of all adjusted Shapley values (for all elements in $\ms$ over all rule bases $\rb\in\ms$). \begin{align*} &\sum_{\alpha \in \rb} \sum_ {B\in\ms} S*_\alpha^{\inc}(\rb)\\ &=\sum_{\alpha \in \rb} \sum_ {B\in\ms} \begin{cases} 0 & \text{if } \alpha \in \F(\rb) \\ \sum\limits_{B\subseteq\rb} \mathit{CP}_{\alpha,\rb}^{\inc}(B) + \sum\limits_{B\subseteq\rb}\mathit{AP}_\alpha^{\inc}(B) & \text{otherwise} \end{cases}\\ &=\sum_{\alpha \in \rb}\sum_ {B\in\ms}\sum\limits_{B\subseteq\rb} \mathit{CP}_{\alpha,\rb}^{\inc}(B) - \sum_{f \in \F(\rb)}\sum_ {B\in\ms}\sum\limits_{B\subseteq\rb} \mathit{CP}_{f,\rb}^{\inc}(B)+\sum_{\alpha \in \R(\rb)}\sum_ {B\in\ms}\sum\limits_{B\subseteq\rb}\mathit{AP}_{\alpha,\rb}^{\inc}(B)\\ \intertext{Following [8], the first summand can be rewritten.} &=\sum_ {B\in\ms}\inc(\rb) - \sum_{f \in \F(\rb)}\sum_ {B\in\ms}\sum\limits_{B\subseteq\rb} \mathit{CP}_{f,\rb}^{\inc}(B)+\sum_{\alpha \in \R(\rb)}\sum_ {B\in\ms}\sum\limits_{B\subseteq\rb}\mathit{AP}_{\alpha,\rb}^{\inc}(B)\\ \intertext{Then} &=\sum_ {B\in\ms}\inc(\rb) - \sum_{f \in \F(\rb)}\sum_ {B\in\ms}\sum\limits_{B\subseteq\rb} \mathit{CP}_{f,\rb}^{\inc}(B)+\sum_{\alpha \in \R(\rb)}\sum_ {B\in\ms}\sum\limits_{B\subseteq\rb} \left\{\begin{array}{ll} 0& \text{if } r \in \Free(B)\\ \frac{\sum_{f\in\F(B)}\mathit{CP}_{f,\rb}^{\inc}(B)}{|r\in \R(B) \text{ s.t. } r \notin \Free(B)|} & \text{otherwise} \end{array}\right. \\ &=\sum_ {B\in\ms}\inc(\rb) - \sum_{f \in \F(\rb)}\sum_ {B\in\ms}\sum\limits_{B\subseteq\rb} \mathit{CP}_{f,\rb}^{\inc}(B)+\sum_{r \in \Free(\R(\rb))}\sum_ {B\in\ms}\sum\limits_{B\subseteq\rb} 0 + \sum_{r \not\in \Free(\R(\rb))}\sum_ {B\in\ms}\sum\limits_{B\subseteq\rb} \frac{\sum_{f\in\F(B)}\mathit{CP}_{f,\rb}^{\inc}(B)}{|r\in \R(B) \text{ s.t. } r \notin \Free(B)|}\\ &=\sum_ {B\in\ms}\inc(\rb) - \sum_{f \in \F(\rb)}\sum_ {B\in\ms}\sum\limits_{B\subseteq\rb} \mathit{CP}_{f,\rb}^{\inc}(B) + \sum_{r \not\in \Free(\R(\rb))}\sum_ {B\in\ms}\sum\limits_{B\subseteq\rb} \frac{\sum_{f\in\F(B)}\mathit{CP}_{f,\rb}^{\inc}(B)}{|r\in \R(B) \text{ s.t. } r \notin \Free(B)|}\\ &=\sum_ {B\in\ms}\inc(\rb) - \sum_{f \in \F(\rb)}\sum_ {B\in\ms}\sum\limits_{B\subseteq\rb} \mathit{CP}_{f,\rb}^{\inc}(B) + \sum\limits_{B\subseteq\rb}\sum_ {B\in\ms} \sum_{f\in\F(B)}\mathit{CP}_{f,\rb}^{\inc}(B)\\ &=\sum_ {B\in\ms}\inc(\rb) - \sum_{f \in \F(\rb)}\sum_ {B\in\ms}\sum\limits_{B\subseteq\rb} \mathit{CP}_{f,\rb}^{\inc}(B) + \sum_{f\in\F(B)}\sum_ {B\in\ms}\sum\limits_{B\subseteq\rb}\mathit{CP}_{f,\rb}^{\inc}(B)\\ &=\sum_ {B\in\ms}\inc(\rb) \end{align*} To show , observe that due to $\sum_{\alpha \in \rb} \sum_ {B\in\ms} S*_\alpha^{\inc}(\rb)=\sum_ {B\in\ms}\inc(\rb)$ via distribution, we have that $\hat{V}^{m_{S*^I}^{\Sigma}}(\ms)=\mathit{max}_{r\in\R(\ms)}(m_{S*^I}(\ms,r)) \leq m^\Sigma_I(\ms)$. Last, to show , observe that $S*^I_f(B)=0$ for any fact $f\in B$ per definition, thus $m^\Sigma_{S^I}(\ms,\alpha) = 0 \forall \alpha\notin\R(\ms)$. Let $\rb$ be a rule base. The problem of deciding whether $\rb$ is consistent can be solved in polynomial time. The minimal model $M$ of $\rb$ can be determined as follows: * $M=\F(\rb)$ * Let $r\in \R(\rb)$ be s. t. $body(r)\subseteq M$ * If there is no such rule, then return $M$ * Otherwise, $M:=M\cup \{head(r)\}$ and continue with 2. I can be seen that $M$ is both closed and minimal and therefore the minimal model of $\rb$. Both, the algorithm above and checking whether $M$ is inconsistent are polynomial, therefore deciding whether $\rb$ is consistent is polynomial. Let $\ms$ be a multiset of rule bases with $\ms=(\rb_1, ..., \rb_n)=(\{\F_1\cup\R\},...,\{\F_n\cup\R\})$ and let $r\in \rb$. The problem of deciding whether there is a $i\in\{1,\ldots,n\}$ and $M\in\MI(\rb_i)$ s. t. $r\in M$ is -complete. For -membership consider the following non-deterministic algorithm: * Guess $i\in\{1,\ldots,n\}$ * Guess a set $M\subseteq \rb_i$ with $r\in M$ * If $M$ is consistent, return False * For each $x\in M$, if $M\setminus\{r\}$ is inconsistent return False * Return True Observe that the above algorithm runs in polynomial non-deterministic time (due to consistency checks being polynomial, cf. Proposition <ref>) and returns True iff $r$ is contained in a minimal inconsistent subset of at least one of $\rb_1, ..., \rb_n$. In order to show -hardness, we reduce the problem 3Sat to the above problem. For that, let $I=\{c_1,\ldots,c_n\}$ be a set of clauses $c_i=\{l_{i,1},l_{i,2},l_{i,2}\}$ where each $l_{i,j}$ is a literal of the form $a$ or $\neg a$ (with an atom a). 3Sat then asks whether there is an assignment $i:A\rightarrow \{\textsc{True},\textsc{False}\}$ that satisfies all clauses of $I$, where $A$ is the set of all atoms appearing in $I$. We introduce new atoms $k_1,\ldots,k_n$ for each of the clauses and a new atom $s$ (indicating satisfiability) and define $\rb_I$ through \begin{align*} \F(\rb_I) & = \{ a, \neg a\mid a\in A\} \cup\{\neg s\}\\ \R(\rb_I) & = \{l_{i,j} \rightarrow k_i \mid j=1,2,3, i=1,\ldots, n\}\cup \{r=k_1,\ldots, k_n\rightarrow s\} \end{align*} We now claim that $I$ is satisfiable iff $r$ is in a minimal inconsistent subset of $\rb_I$ (which is a special case of our problem with $\ms=(\rb_I)$). So assume $I$ is satisfiable and let $i$ be a satisfying assignment. Observe that $M$ defined via \begin{align*} M & = \{a \mid i(a)=\textsc{True} \} \cup \{\neg a \mid i(a)=\textsc{False}\}\cup\{\neg s\}\cup \R(\rb_I) \end{align*} is inconsistent: as $i$ is a satisfying assignment, each $k_i$ ($i=1,\ldots,n$) can be derived in $M$; then $s$ can also be derived, producing a conflict with $\neg s$. On the other hand, note that $M\subseteq \{r\}$ is consistent. It follows that there is a minimal inconsistent set $M'\subseteq M$ with $r\in M'$. Now assume that there is a minimal inconsistent set $M\subseteq \rb_I$ with $r\in M$. First observe that there is no atom $a$ s. t., $a,\neg a\in M$ (otherwise $M\setminus\{r\}$ would still be inconsistent). Let $i:A\rightarrow \{\textsc{True},\textsc{False}\}$ be any assignment with $i(a)=\textsc{True}$ if $a\in M$ and $i(a)=\textsc{False}$ if $\neg a\in M$. It follows that each clause $c_1,\ldots,c_n$ is satisfied by $i$ (as each $k_i$ could be derived in $M$) and so $i$ is a satisfying assignment for $I$. Let $\rb$ be a rule base and $M\subseteq \rb$. The problem of deciding whether $M\in\MI(\rb)$ can be solved in polynomial time. Deciding whether $M$ is inconsistent and $M\setminus \{x\}$ for each $x\in \rb$ is consistent can each be solved in polynomial time due to Proposition <ref>. It follows that deciding $M\in\MI(\rb)$ can be solved in polynomial time. Let $\ms$ be a multiset of rule bases with $\ms=(\rb_1, ..., \rb_n)=(\{\F_1\cup\R\},...,\{\F_n\cup\R\})$ and let $r\in \R$. The problem of determining $|\{M \in \MI(\rb_i) \mid i\in \{1,\ldots,n\} r \in M\}|$ is $\#\textsf{P}$-complete. Membership follows from Proposition <ref> as deciding for a given $M$ whether $M\in \{M \in \MI(\rb_i) \mid i\in \{1,\ldots,n\}, r \in M\}$ is in $\textsf{P}$. The proof of $\#\textsf{P}$-hardness is analogous to the proof of Proposition 5 in [2]. Observe that in the reduction the notion of “issue” coincides with notion of a minimal inconsistent subset containing the rule $\pi\leftarrow \alpha_1,\ldots,\alpha_n,\delta_1,\ldots,\delta_m$ if we add facts $a,\neg a$ for each atom $a$ occurring the in input instance.
# Robust Anthropomorphic Robotic Manipulation through Biomimetic Distributed Compliance Kai Junge 1∗ Josie Hughes1 1CREATE Lab, EPFL, Lausanne, Switzerland $\ast$ Corresponding author. Email<EMAIL_ADDRESS> ###### Abstract The impressive capabilities of humans to robustly perform manipulation relies on compliant interactions, enabled through the structure and materials spatially distributed in our hands. We propose by mimicking this distributed compliance in an anthropomorphic robotic hand, the open-loop manipulation robustness increases and observe the emergence of human-like behaviours. To achieve this, we introduce the ADAPT Hand equipped with tunable compliance throughout the skin, fingers, and the wrist. Through extensive automated pick- and-place tests, we show the grasping robustness closely mirrors an estimated geometric theoretical limit, while ‘stress-testing’ the robot hand to perform 800+ grasps. Finally, 24 items with largely varying geometries are grasped in a constrained environment with a success rate of 93%. We demonstrate the hand- object self-organization behavior underlines this extreme robustness, where the hand automatically exhibits different grasp types depending on object geometries. Furthermore, the robot grasp type mimics a natural human grasp with a direct similarity of 68%. ## 1 Introduction The human ability to robustly pick, place, and manipulate objects in uncertain environments is common place, and we perform it with ease [1]. Whilst our planning and sensory-motor capabilities play a key role, these compliant and robust interactions are also heavily influenced by our hand morphology, kinematics, and dynamics [2, 3]. From softness in the skin creating stable contact, to compliant muscle synergies[4] forming diverse ranges of grasp types[5], this physical intelligence can enable emergent and self-organized behaviours which result in a robust response to uncertainties in the environment[6]. This serves as inspiration for advancing the baseline open- loop capabilities of anthropomorphic robotic manipulators. Such physical intelligence will enable robotic hands with fluidic, human like dynamic behaviours and is essential for fully leveraging the potential of rapidly advancing control and learning strategies [7, 8]. A number of different approaches have been explored for advancing open-loop robustness and versatility of anthropomorphic robotic hands. Incorporating compliance underpins all these methodologies, with many taking a embodied intelligence[9], or physical intelligence perspective[10]. One approach is to develop fully compliant hands such as the RBO Hand I/II/III [11, 12, 13], BCL-26 Hand[14], and the Shorthose Hand[15]. Typically pneumatically operated, they have shown robustness to object geometry in grasping tasks through their continuously compliant structure. In particular, the RBO Hand III has shown significant robustness for dexterous in-hand manipulation of a rubix cube with purely open-loop actions [16]. However, fully soft hands can experience some limitations in terms of the possible force application, repeatability, and agility. An alternative approach is tendon actuated soft-rigid hands, which are inspired by and mimic the human musculoskeletal structure and actuation mechanisms[17]. These can leverage rolling contact condyloid and synovial joints held together by compliant ligaments [18, 19, 20] providing natural, robust behaviour with resistance to impact and joint friction[21, 22, 23, 24]. Compliance embedded within the actuation mechanism is also important for generating open-loop robustness. Although not in an anthropomorphic form factor, a compliant differential and underactuated tendon routing can enable multi-finger grippers which can robustly grasp objects with varying geometry with a single actuator[25, 26, 27]. The theoretical framework for adaptive synergies builds upon this to similarly exploit compliance to achieve diverse behaviour for an anthropomorphic hand[28, 29]. The requirements for compliance extends beyond the hand, with the wrist motion contributing to coordinating the behaviour of passive, i.e. non-actuated hands [30]. This combination of wrist motion and bio-inspired passivity in the structure has been demonstrated to enable tasks such as dynamic and varied piano playing [31] and grasping [32]. In all these varied approaches it is evident that compliance advances open- loop robustness. However, in comparison to human hands, the incorporation of compliance remains limited to distinct spatial regions. Furthermore, we lack metrics or approaches for the systematic assessment of the inclusion of compliance [33]. While there are benchmarks for hands, these are often a high level task based metric which is heavily affected by the control strategy[34, 35, 36], object focused corresponding to a sparse fail/success metric[37], or focus on low level static capabilities such as grasp taxonomies[5], range of motion [38], and hardware features[39]. Figure 1: An anthropomorphic robot hand designed with biomimetic distribution of compliance leads to a emergence of robust and self-organizing behavior. A) The ADAPT Hand is compliant in its skin, finger, and wrist, which operate in the 0.1cm, 1cm, and 10cm displacement ranges. B) Through identical open-loop waypoints (four waypoints describing the full hand-arm motion for this grasp), the hand can grasp objects (a large apple vs three small items) successfully. The distributed compliance allows for the hand to be robust upon unknown environmental interactions, and a self-organization to take place between the hand and objects, resulting in different grasps. We propose that matching the magnitude and distribution of compliance to that of humans is essential for matching their robustness and fluidity in interactions. Specifically, that by incorporating distributed and varying compliance in the skin, fingers, and wrist of an anthropomorphic robot hand, human-like behaviour can be generated with minimal open-loop control. A similar focus has been explored for locomotion[40, 41], where it has been seen that the distribution of compliance across the body and environmental interactions leads to emergent and stable self-organizing gait patterns[42]. By introducing such strategically located compliance into robotic hands, we can explore the resulting hand-environment self-organization to provide robustness through unconscious selection of robust grasp configurations (see Fig.1B). This fundamentally requires a robotic hand where the compliance can be spatially distributed. We introduce our ADAPT (Adaptive Dexterous Anthropomorphic Programmable sTiffness) Hand, shown in Fig.1, which features a soft skin covering the contact surface, series elastic actuated fingers/thumb, and an impedance controlled wrist, all of which are tuned to match their biological counterpart. Each compliant element acts at a different interaction length scale, providing contact stability, pose adaptation, and motion conditioning.Starting from the skin and the fingers, we demonstrate that matching the compliance to that of human hands leads to stable and robust manipulation tasks in comparison to a non-matched rigid configuration. This leverages the tunability of the ADAPT hand, enabling us to compare and contrast bio-inspired distributed compliance and a rigid hand. Extending to the full hand, and using a robotic setup to autonomously perform pick-and- place tasks, we quantify the open-loop robustness for grasping approaches with respect to an estimated theoretical limit based on object and hand geometry. Using the same setup, we also evaluate the robustness of the ADAPT Hand design through an extensive autonomous experiment with over 500 grasps across 9+ hours of uninterrupted pick-and-place operation. The hand showed a success rate of 97%. Finally, by introducing a compliant wrist we show how the distributed compliance leads to self-organization of emergent grasp types, which matches that of a human (direct similarity of 68%), with a success rate of 93% to handle 24 items spanning a few to 100s of millimeters. ## 2 Results ### 2.1 ADAPT Hand Figure 2: A) The ADAPT hand. B) Finger design with two independent actuation and a series spring on the MCP flexor. C) The force-displacement measurement curve for the skin, finger, and wrist for the robot and a human. For the skin and finger, the rigid configuration of the ADAPT Hand is also shown for comparison. The ADAPT Hand is a bio-inspired anthropomorphic compliant robotic hand platform, with joint kinematics that reflect an human adult’s hand (detailed in Section 4.1) and the compliance reflecting a relaxed human adult’s hand. The hand is designed to have independently tunable compliance in the skin, finger, and wrist used to investigate the impact of spatially distributed compliance on the robustness and performance of manipulation tasks. The hand, shown in Fig.2A, has 12 actuators corresponding to 20 degrees of freedom, and leverages tendon driven operation. The defining feature of the robot hand design is the finger (shown in Fig.2B), where two actuators control the flexion-extension action. Two antagonistic tendons actuate the metacarpophalangeal(MCP) joint, while a single tendon is responsible for the flexion motion of the proximal inter-phalangeal(PIP) and distal inter- phalangeal(DIP) joints. The two actuators controls the finger motion independently (see Fig.2B right). The flexor tendon of the MCP joint has a series spring attached, which provides the finger compliance. The abduction- adduction motion of the fingers are coupled (controlled by one actuator) and series elastic. The thumb is designed in a similar way, but has two antagonistic joints at the carpometacarpal(CMC) joint (base of the thumb). Despite being underactuated, the actuated range of motion can achieve all 33 grasp taxonomies[5] and all 10 positions on the Kapandji score[38] (see Fig.S1). Fig.2C shows a force-displacement curve for each type of compliance for the robot and the human. The variability of different curves for the human is resultant of multiple measurement trials. The skin compliance is tuned through the selection of materials to cover the underlying rigid structure. The chosen skin (EcoFlex20) matches that of a human, especially in comparison to a fully rigid skin (10 to 40 times stiffer). The finger stiffness is tuned by swapping the aforementioned series spring in the MCP joint. Compared to the human- matched finger, the average stiffness of a rigid finger (which is realised by removing the series spring) is a factor 30 larger. The plateauing force profile (on the human and tuned finger) from a low stiffness joint means the fingertip exerts near-constant forces under large displacements - suggesting the open-loop force application will be robust to large disturbances. The wrist compliance is achieved by an impedance controlled robot arm. The force profile of the compliance in the vertical direction is shown, where the impedance control parameter is tuned to match a human wrist. Throughout this work, the actuation signals are manually programmed open-loop signals. That is, an operator will record key waypoints for both the finger/thumb joints and the wrist 6dof pose (detailed in Section 4.2). During execution, the recorded waypoints are linearly interpolated and directly replayed. For every task in the work, the motion is programmed to best mimic how a human would perform them. ### 2.2 Skin: Contact stability Figure 3: A) Results from the finger sliding experiment. Schematic (top-left), box plot and raw values for the shear force of rigid and soft skins measured at the midpoint of the interaction (top-right), example time series of the shear force of rigid and soft skins for the two sliding motions (bottom). B) Results from the knob turning experiment. Schematic (top) and turn angles for the soft and rigid skins as the diameter $d$ and resisting torque $\tau$ is varied (bottom). C) Results from the finger gaiting experiment. Schematic (top) and completed gaits for soft and rigid skins as the held block width $w$ is varied (bottom). The skin plays a crucial role in the robot-environment interaction, being the direct interface between the hand and the environment. The cutaneous compliance in human skin offers increased local deformation, assisting with contact stability and shear force generation. To measure the corresponding contact stability offered by matching the skin stiffness we contrast the matched skin to a rigid one made from PLA. The PLA skin is coated in a thin layer of EcoFlex20 to ensure the surface friction properties are consistent and we isolate the effect of compliance. The main impact of the skin compliance is the increased shear forces generated for a given motion. This shear force was analyzed for two sliding motions on a plate (slide front and back motions: refer to Fig.S2) for different relative displacements between the finger and the sliding surface. On the top right of Fig.3A the shear force measured via a loadcell at the midpoint of the interaction motion is plotted. Across a range of motions and surface environments, the shear force is constantly higher with the soft skin in comparison to the rigid case. The bottom of Fig.3A shows two representative shear force profiles for the two sliding motions. The temporal force profile is similar for the soft and rigid, but the generated force has typically twice the force magnitude. #### 2.2.1 Skin compliance on task performance This demonstrated capacity of the soft skin to generate higher shear force under the same action contributes to contact level stability. To measure the resulting performance, we use the task of turning a cylindrical knob with the middle finger and thumb with a predefined finger motion. The turn angle $\theta$ is used as the performance metric, while the environment is varied by independently varying the diameter $d$ and resistance torque $\tau$ of the knob (Fig.3B top). Resultant turn angles for each environmental variation are shown in the bottom of Fig.3B where across all tasks the soft skin is outperforming. In particular, variations in $\tau$ impliy that the soft skin is more robust, shown by a a far greater performance drop for the rigid skin of 37$\deg$ compared to 18$\deg$ for the soft skin. The effect of the softer skin also extends to tasks explicitly requiring contact stability, such as finger gating when holding an object, as the width $w$ of the block is varied from low to high. For a given, pre-determined finger gating pattern, the performance is measured by counting the number of completed gaits until failure. The results given in Figure Fig.3C show that the soft skin leads to greater contact stability, and thus a higher performance. The rigid one shows increasing performance as $w$ increases, as a higher width results in higher holding forces. For the soft skin, there is in fact a reduction in performance with the largest width object, however, this is still almost twice the performance of the rigid setting. In both these experiments, the soft skin creates a larger contact area with the environment which results in a more stable contact, as measured through the higher stabilizing shear forces. This leads to consistently higher task performances and in some cases increased robustness to environmental uncertainty. ### 2.3 Finger: Pose adaptation Figure 4: A) Key frames from two finger sliding motions(left). By overdriving the MCP joint, pseudo force control is possible with good repeatability (right). B) Trajectories of relative changes in the three joint angles for the sliding motions executed by a human, soft robot finger, and a rigid robot finger. C) Schematic for the finger sliding experiment (top). Maximum forces recorded as the sliding plate is displaced in the $z$ and $\theta$ directions for the two motions/soft and rigid fingers (bottom). D) Schematic for the knob turning experiment (top). Tendon waypoints required for the motion(mid) and turn angles as the environment is varied (bottom) for the soft and rigid fingers E) Schematic for the finger gaiting experiment (top). Completed gaits (mid) and average holding forces (bottom) for soft and rigid fingers as the held block width $w$ is varied. F) Success rates for the three cubes rotated in-hand (top). Pictorial sequence of the in-hand manipulation sequence (bottom). The flexion-extension motion of the finger originates from the series elastically actuated MCP joint. In this section we show how the series elastic MCP joint can enable centimeter scale pose adaptation of the finger, resulting in robust behavior. When external contact forces are applied the entire finger will adaptive comply to the environment since the MCP joint is located at the stem of the finger. By combining this series elastic actuated joint with the PIP/DIP actuation, the fingertip can move to follow a surface with consistent contact. For example, Fig.4A shows two sliding motions (slide front and slide back), achieved by flexing or extending the PIP/DIP joints respectively, which is achieved through three waypoint transitions (detailed in Fig.S2). The compliant MCP mechanism also offers pseudo force-control, by overdriving the joint. For instance, if the finger makes contact when the MCP joint is at $q\degree$, instead the position demand can be set to $q+\Delta_{\mathrm{MCP}}\degree$. By doing so, the finger maintains a stable contact with the environment at a similar force magnitude even when the specific shape of the contact surface is unknown. The right side of Fig.4A shows the normal and shear force profile of the slide back motion under $\Delta_{\mathrm{MCP}}=0,\ 7.5,\ 15$mm for two repeats. The force magnitude increases almost linearly with the additional $\Delta_{\mathrm{MCP}}$. This plot also shows the repeatability of the soft finger, as two repeats of the same action are near identical (an average RMSE of $3.6\pm 4.3$g across all experiments). #### 2.3.1 Human motion comparison Through the compliant MCP joint, the ADAPT Hand finger motion shows a kinematic resemblance to that a human finger at the joint displacement level. For the slide back and front motions, a similar motion was performed by a human whilst visually recording its pose. Fig.4B shows the joint evolution of the finger during the two sliding motions for a human, the robot with a soft MCP, and the robot with a rigid MCP. In both motions, the soft finger is most similar to the human motion. For the slide back motion, the difference is most notable in the relative motion of the MCP and PIP joints. The two joint angles are near consistent on the rigid finger, implying that all the motion is through the DIP joint (the image above the plot shows this in effect). For the slide front motion, the soft finger shows even higher similarity with the human while the rigid finger barely moves. The slide front motion inherently relies on the MCP joint to flex as the PIP/DIP joints are extending, which the rigid finger cannot achieve without extra waypoints. Although the results shown through this comparison are only indicative of the similarities between the human and robot finger motions, it motivates the necessity of a compliant MCP joint in the finger structure to provide a natural motion of the PIP/DIP joints while maintaining contact. #### 2.3.2 Measuring robustness through finger compliance In Fig.4A and B we demonstrate human-like sliding motions with minimal planning (only three waypoints) by leveraging the compliant MCP joint. The robustness (behavior invariance to changes in the environment) is measured by comparing task performances between the soft and rigid fingers while varying the environment. Firstly, we consider the consistency of the forces generated in the two sliding motions. While displacing the finger and the surface in the $z$ and $\theta$ directions (see Fig.4C top), the variability in applied forces for the rigid and soft fingers is measured. The scatter plots in Fig.4C compare the normal and shear forces ($F_{V}$, $F_{H}$) for the two sling motions between the soft and rigid fingers for combinations of $\Delta_{z}$ and $\Delta_{\theta}$. The MCP overdrive ($\Delta_{\mathrm{MCP}}$) on the soft finger is chosen to match the maximum force applied to the rigid finger when $\Delta_{z},\ \Delta_{\theta}$=0 as shown by the opaque scatter point. The error bars show the standard deviation of maximum force recorded, where the soft finger is on average 2.4 times lower. The lower variability in interaction forces implies the soft finger is less influenced (i.e.: more robust) by position variability. We can also return to the knob-turning task to evaluate the robustness when actuating the fingers. Here the turn angle of the knob is measured while the size of the knob and robot displacements are varied (see Fig.4D top). For the soft finger, this motion can be is achieved with only three position waypoints reusing the slide back and front motions for the thumb and middle finger respectively. If the same waypoints are executed on the rigid finger, the robot is damaged from high forces. Instead, for the rigid finger, precise motion planning (15 waypoints) is necessary to describe the path specifically around the knob. The waypoints needed for this task are shown in the timeseries graph in Fig.4D, which highlights how the compliance in the finger simplifies control. The performance of the rigid and soft fingers is summarized in the scatter plot (Fig.4D bottom) for three environmental settings: default settings (when the motion was programmed), reduced knob diameter, and displacement of the knob location by 1cm. Overall the soft finger is higher performing. As the rigid finger requires waypoints that follow the contour of the knob, the interaction forces can vary largely compared to the soft finger, leading to inconsistent contact force application and thus lower turn angle. For example, with a smaller knob the turn angle increases in the soft finger (since the same turn distance covers a larger angle for a smaller $d$) by 10$\degree$, the turn angle for the rigid finger reduces by 8$\degree$. Moreover, when the knob is displaced, the rigid finger is damaged from the high reaction forces. In the finger gaiting task, the number of completed gaits until failure and the average holding force $F_{\mathrm{Hold}}$ is recorded while the width of the block $w$ is varied (Fig.4E top). Similar to knob-turning, a separate trajectory for the rigid finger was programmed which carefully tracks the contact with the block. Unlike the knob turning, the number of waypoints remain identical, but a high precision is still required for the MCP joint angle to follow the width of the block. The raw measurements for the completed gaits and $F_{\mathrm{Hold}}$ for the three $w$ settings are shown in Fig.4E as two performance metrics. For both metrics, soft finger has a higher consistency as $w$ changes in comparison to the rigid finger. This is a direct extension of the result in Fig.4C, as the interaction forces are consistent under uncertainty in displacement, resulting in a more robust behavior. #### 2.3.3 In-hand cube re-orientation Throughout Fig.4A-E, there are clear benefits to the motions, but they are simple. While using the same waypoint planning methodology, we can demonstrate a more complex cube re-orientation task (see Video S1). Fig.4F captures key frames of the robot using its fingers, thumb, and palm to continuously re-orient a cube with a total of 12 waypoints (number of waypoints marked at every frame). While the entire motion is complex, it can be formed by combining the simpler motions shown in Fig.4C-E. The same programmed motion is repeated 20 times for cubes with three different sizes, with 100% success rate of the small and medium cubes and 90% for the large cube. ### 2.4 Systematic pick-and-place robustness assessment of the ADAPT Hand Figure 5: A) Robotic setup to conduct large quantities of automatic pick-and- place experiments while controlling the displacement of the object. B) Measured limits on object displacement (orange axis), estimated geometric limits based on object size and hand closure motion. C) Success and failed grasps throughout the two experiments: robustness assessment and uninterrupted pick-and-place totalling 845 grasps. By introducing compliance into the robot skin and finger we have shown low level measurements of improvements in the robustness and stability for interactions that require individual or multiple fingers. Extending to include all five digits, we now explore how this compliance influences a fundamental manipulation task, pick-and-place. Specifically we want to explore the robustness/repeatability of the hardware design through many 100s of picks, and provide quantifiable measurements for the robustness of the open-loop pick-and-place performance. To do this, we evaluate the robots picking success as an object is iteratively displaced further away from the central point of the palm in the x and y direction. We compare this to an estimated theoretical limit defined purely geometrically (see Fig.S3). To systematically perform this large scale evaluation of robustness, a robotic system surrounding the hand is developed to automatically perform pick-and- place tasks. This comprises of a secondary arm with a movable plate and an overhead camera (see Fig.5A). The process begins with Robot 1 introducing an arbitrary offset $\Delta_{M}$ as defined manually. To do this, the system first identifies existing displacement of the object from the center of the place: $\Delta_{O}$. By moving the plate by $\Delta_{M}-\Delta_{O}$, the object is now displaced by $\Delta_{M}$ from the plate center. Robot 2 with the ADAPT Hand can then execute a open-loop pick-and-place motion. This process is outlined in Fig.5A and can be performed continuously for many hours with minimal intervention. #### 2.4.1 Measuring open-loop robustness Using this setup, we experimentally measure the ADAPT Hand’s robustness to displacement as compared to the estimated theoretical upper limit. First, five objects with different geometries were chosen, and used to program a separate grasp for each object mimicking how a human would grasp. Then, five extra unseen objects but with similar geometries were added to the object set. The tested and unseen objects are labeled in Fig.5B. For each objects, $\Delta_{M}$ was varied in the horizontal and vertical directions independently until two consecutive failed grasps were recorded. Fig.5B shows ten pre-grasp poses of the hand overlayed with results of measured and approximate theoretical limits of possible object displacements. The blue shaded area indicate an estimated geometric theoretical limit. This is defined as when the center of mass of each object could be placed within the area spanned by the motion of the fingers without colliding with the fingertips at the start of the grasp (see Fig.S3). The measured limits on the vertical and horizontal directions are indicated by the orange error bars. Surprisingly, the hand is close to or exceeds the theoretical geometric limits of open-loop pick-and-place capabilities, despite the grasps are programmed with approximate motions and half the objects have never been evaluated. When the objects are displaced at the millimeter scale, the contact stability from the skin compliance can maintain a similar grasp configuration (as the programmed one). At larger centimeter scale displacements, the effect of the finger pose adaptation is greater. The same objects under the same motion can be held by different fingers, while still being able to hold the object stably (examples of such grasps are shown in Fig.S4). Disregarding the measurements from the Empty coke can (since it is a clear outlier which skews results), the ratio of the measured over the theoretical limit on the vertical and horizontal axis are 1.2$\pm$0.2 and 0.8$\pm$0.2. On average, the vertical robustness exceeds the theoretical limit because even when objects collide with fingertips, the grasp is successful (an assumption use to derive the theoretical limit). The horizontal robustness is lower than the theoretical limit, since at large deviations the forces from the finger are too focused on one side forcing the object out of the hand. For the coke can, the measured limits far exceeds the theoretical because of the round shape (grasp is still successful even placed at the fingertips), and the extremely lightweight nature (can grasp just at the ends). #### 2.4.2 Extended period of operation Using the continuous test capabilities of the robotic system in Fig.5A the robustness and repeatability of the ADAPT Hand robot design can be evaluated for a large, uninterrupted trial. The robot system completed 500 grasps with $\Delta_{M}$=0 with a total success rate of 97%. During this operation, the robot’s hardware and software system was fully untouched. In fact, the majority of the failures (15/16) are from grasping the tape. This is reasonable in hindsight given the zero position of the tape was already close to the edge of the robustness limits (the point of intersection of the orange error bars in Fig.5B lies near its edge). Fig.5C shows a time series of all experiment performed using the pick-and- place system. Divided by the blue dotted line, the first set of experiments are for assessing the robustness described in Fig.5B. The second set represents the uninterrupted robot operation. In the combined two experiments, the ADAPT Hand was actively used over 16 hours and 845 grasps without any modifications to the hardware/software. The only visible damage to the hardware was the ware/dirt on the silicone fingertips shown in Fig.S5. Video S2 shows a sped up version of this extended trial, while details to the raw video can be found in section S1. ### 2.5 Self-organizing grasps Figure 6: A) Key frames overlayed with wrist motion (blue arrow), finger motion (red arrow), and new contact regions (pink shade) for the human and robot grasping an object off the table. The number of waypoints for the wrist or hand illustrated underneath the frames. B) Objects used for this experiment with failed trials indicated by red circles above the image. C) Four grasps identified on the human and robot. D) Object geometry plotted with the color indicating the grasp type observed. E) Direct comparison between robot and human grasps for every object. In the previous section, the focus was on the skin and finger compliance which offers robustness and stability to local interactions. A compliant wrist enables the scale of these interactions to be expanded, for example grasping while sliding and object off a table (Fig.6A). Through open-loop trajectories of the hand and wrist which mimics human grasping of grasping objects from a table, we explore how robust this is to different objects, and also how different grasp types emerge. When a human accomplishes this task, three behaviors within the trajectory can be identified that are invariant to the object being grasped (top row of Fig.6A shows an example of grasping a lemon with its key frames and the finger/wrist movements). Firstly, the approaching motion is always the hand moving downwards until making contact, either with the table or the object. Secondly, after the initial contact is made the fingers continue to hold its contact during the motion, either with the table or the object. Thirdly, the grasping motion shown by the Flex fingers, Wrist up, and Retract wrist frames is a single continuous motion where the fingers and thumb curls to form a grasp as the wrist moves away from the table. By using a wrist with human-matched compliance in addition to the skin and fingers, the observed human behaviors can be robustly replicated on the robot (as shown in the bottom two rows of Fig.6A). When approaching the table, the compliance allows for a safe and controlled contact of the fingers and/or palm as the wrist presses into the table (first two frames). The grasping motion (the final three frames of Fig.6A) shows a coordinated flexing movement of fingers and thumb as the wrist moves away while constantly maintaining contact with the environment, matching with the second and third observations from the human (last three frames). This motion relies on the distributed compliance, in particular the compliance of the wrist conditions(provides the necessary conditions) the interaction of the fingers with the environment. During the grasping sequence, only the wrist is actively actuated (see the sequence of waypoints executed in Fig.6A bottom). The fingers are only actuated before the final wrist motion to from a grasp, where the movement is constrained by the table. The wrist movement gradually removes the constraint of the table enabling the fingers self-organize and grasp the object, while the compliance of the wrist maintains contact with the surface. Using the same open-loop motion, 24 objects of varying geometry were chosen to be grasped from the table. Fig.6B shows each object, spanning from flat and thin objects such as a pencil, to a large and bulky objects to an apple. Each object was grasped three times with varying placement locations in the vicinity of the robot hand motion path, with an overall success rate of 93% (70/75 tests) failing to grasp only three objects (phone, bolt, and lemon). Video S3 shows this grasping experiment for all items from two view angles. #### 2.5.1 Emergence of discrete grasp types The robustness of the robot to grasp a wide variety of objects through identical commands can be explained through self-organization. Consider the two robot motions given in Fig.6A. In the top row, the lemon is a large object where the robot makes multiple contacts throughout the motion (see pink highlights). On the bottom row, the robot only makes contact with a pencil just when the grasp happens. The resultant grasp of the two objects are also different: the lemon is held with a power grasp while the pencil is held only with the fingertips, which is appropriate if a human were to grasp the two objects. For all the objects, the grasp motions were analyzed for the robot and the human, and categorized to five distinct grasp types shown in Fig.6C. The grasp types are categorized based on the hand posture and which parts of the hand interact with the object which relate with the grasping taxonomies (see section 4.3.3). Video S3 shows the four grasp types identified for the robot. The scatter plot in Fig.6C shows the object geometry (length, width, height) clustered by the observed grasp types of the robot. The clustering shows the robot hand self-organizes to a discrete set of grasp types based on the object’s geometry. For flat or small objects the grasp is using its fingertips (blue cluster). As the height increases, the robot starts to use its thumb (green cluster), slowly transitioning into a form of power grasp (yellow, red clusters). Like the example in Fig.6A, the robot grasp types observed match closely with a human. It is know that a human will vary the grasping strategy based on the object geometry (and other factors too)[43], which is indeed observed when asked to grasp the objects used in this experiment. The matrix in Fig.6C directly compares grasps observed for the robot and human. Categorically the human exhibits one extra grasp type, the finger surface grasp where all fingers a kept straight, which is impossible for the robot as the programmed motion curls the PIP/DIP finger joints. Despite this mismatch of four versus five grasps, the strong correspondence between the two clear by 68% (17/24 objects) lying on the leading diagonal, and all other objects on directly adjacent cells. Assuming the human is choosing the optimally stable grasp through prior knowledge, the robot having a biomimicry of the kinematics and distributed stiffness is able to self-organize to choose the same. ## 3 Discussion In this work, we present an anthropomorphic robot hand, the ADAPT Hand, designed with a biomimetic distribution of compliance across different lengths scales in the skin, finger, and wrist. Starting from low level interactions of the skin and finger levels, we show the presence of a compliant skin and MCP joint on the finger leads to higher performance and robust to environmental uncertainty across three of tasks. Through the ADAPT Hand’s tunable compliance, a direct comparison between a compliant and rigid robotic hardware is made. Expanding the task to include all five digits, the robustness within the hand is measured to lie close to an estimated theoretical limit for a pick-and-place task. Using the same measurement setup, the robustness and repeatability of the hardware platform itself is shown through a damage-less execution of over 800 grasps and 15 hours of operation. Finally, a compliant wrist motion is introduced to grasp 24 objects through an interaction-rich motion across a tabletop surface. This motion which fully combines the distribution of compliance across the robot body, where a wrist motion which conditions the passive pose adaptation of the fingers while contact stability from the skin enhances the grasp. The resultant discrete grasp types are self- organized based on the object geometry, similar to that of a human. Not only, a comparison between human and robot grasp shows 68% of grasps are directly matching. Overall, we demonstrate an physical intelligence approach towards anthropomorphic robot design which considers the interaction and motion at all the lengths scales of robotic manipulation. The extreme robustness given simplest form of actuation control input culminating in the self-organizing grasps. Although the proposed concept realized by the ADAPT Hand show remarkable performance, there are clear limitations to be addressed in the future. At the heart of this work is the distributed compliance, which can negatively affect certain tasks. For instance the soft skin and finger limits the ability to exert high forces in one direction. This means strongly pinching an object or button pressing tasks are challenging. When using the full hand for manipulation, the passive adaptation is simply a result of force equilibrium working in a desired way. This means, when the external forces are too large (e.g.: an object that is too heavy such as the phone in Fig.6B), the execution fails. The same is true for balancing forces between multiple fingers and a held object. Since the fingers will passively adapt, tasks such as controlled object re-orientation is difficult. Sensor feedback and control is one obvious next step to tackle these limitations. This can mean sensing of the joint torques/tendon forces as well as tactile sensing to span the skin surface. However, the control method using the sensor feedback is not straightforward for this robot. This is because the manipulation capabilities of this hand is driven through the emergent behaviours which arise from the interaction between the compliant robot and the environment. Although the actuation signals are position commands, each waypoint is programmed with a higher level intention such as continue to apply force downwards or move roughly in this direction instead of reaching a particular position or force. The sensory-motor control should hence not follow some explicit trajectory, but should rather complement the emergent of behavior. This could be achieved through combining high level intention in the form of the open-loop waypoints and a controller to regulate contact forces, similar to a shared-control scheme used in[44] used to stabilize a noisy EMG signal. Alternatively a bio-inspired sensory-motor control, such as central pattern generators(CPGs) used to transition between walking and swimming gaits for a salamander robot[45], can lead to emergent efficient behaviors. While the concept of CPGs do not directly apply to manipulation, a sensory-motor coordination method that drives emergent behaviours[9] is necessary to maintain and extend the robust interactions. The ADAPT Hand as a hardware platform itself also has room for improvement. One direction is to increase the level of biomimicry of the human hand. The current hand although is bio-inspired, clear differences are present especially in the joint placement and the skin/flesh distribution. For example, there is no material covering the MCP joints. Consequentially, multiple grasp formations do not translate well from a human to the robot. Another limitation is the lack of a wrist just below the hand. While the 7dof robot arm can orient the hand in any angle, in practice the arm must move a large displacement to resemble small wrist motions seen in a human. Likewise to the hand itself, increasing the biomimicry in the arm kinematics can enable more complex human motions to be realised by the robot. Not only could this simplify manual programming or teleoperation by a human, but one could imagine mapping motions learnt from the abundance of videos of humans to be applied to this biomimetic robotic hand-arm structure. ## 4 Methods Figure 7: A) Actuated, underactuated, and passive joints for the ADAPT Hand. B) Finger design and tendon routing. C) Series elastic actuation design for the the MCP joint. D) Soft and rigid skin designs. E) Abduction/adduction linkage mechanism. F) Waypoint recording and replay method used to program the robot motion. ### 4.1 ADAPT Hand Hardware The ADAPT Hand is a custom designed and fabricated anthropomorphic robot hand, with its four fingers and thumb having dimensions similar to an adult human. The entire hand is fabricated from commercially available 3D printed Polylactic acid(PLA) and Thermoplastic Polyurethane(TPU) with shore hardness 98A/65D, and uses tendons (cables) for its actuation. Fig.7A shows the ADAPT Hand without the skin, showing the joint kinematics and actuation scheme. A total of 12 servo motors (Dynamixel XM430-W210-R) located beneath the hand actuates the 20 joints (four joints per finger/thumb). Having less actuators than joints means certain joints are underactuated (shown by subscripts a, b, c in Fig.7A) or passive (P in Fig.7A). Details on the CAD model is presented in section S1. #### 4.1.1 Finger design The finger design shown in Fig.7B is a key feature of the ADAPT Hand, where the same design is used across the five digits. One actuator controls the pin- jointed MCP joint with antagonistic tendons (the thumb has an extra antagonistic joint, but rotate 90 degrees to make the 2DoF CMC joint). The PIP/DIP flexure joints which are coupled by a single flexor tendon is actuated by another motor. The pin joint in the MCP joint allows the routing of the PIP/DIP flexor to pass through the center of rotation, fully decoupling the two axis of actuation. The extension forces for the PIP/DIP come from the combined effect of the TPU flexure joint and the elastic thread, placed on the backside of the finger (shown by a black line in Fig.7B). The compliance at the finger level is generated by a series elastic MCP joint which is achieved by routing the MCP flexor tendon around a pulley connected to an extension spring (shown in Fig.7C). Replacing or removing the series spring can change the finger compliance (as in Fig.2B). #### 4.1.2 MCP abduction-adduction motion A notable feature of the mechanical design of the ADAPT Hand is the series- elastic linkage mechanism driving the abduction-adduction axis of the MCP joints for the index, ring, and little fingers. The mechanism and its open and closed states are shown in Fig.7E. The linkage mechanism connects to the MCP pin joint with a series elastic TPU material, making the MCP joint compliant in 2 axes. Having an actuated spread axis of the MCP joints increases the workspace (thus the capability of the hand) such as holding large objects. #### 4.1.3 Dexterity of the ADAPT Hand Combining all the mechanisms of the fingers, the ADAPT Hand can be actuated to produce dexterous motions. Starting from the zero position (all fingers straight), the hand can be actuated to produce all 33 grasp taxonomies[5]. Likewise, the hand is able to complete all 10 postures on the Kapandji test. The results for both tests are shown in Fig.S1. #### 4.1.4 Skin design A modular skin fully covers one side of the ADAPT Hand. Fig.7D shows the index and middle finger equipped with a rigid and soft skin respectively. The skins are identical in their geometry and is approximately 3mm offset from the bones. The soft skin is fabricated from cast EcoFlex20 and the rigid skin is 3D printed PLA. A thin ($\approx 0.5$mm) layer of EcoFlex20 is glued on the surface of the rigid skin to maintain the surface friction property. ### 4.2 ADAPT Hand motion programming In all experiments, the ADAPT Hand (including the robotic arm) operates through manually programmed open-loop motions. For both the hand and arm, a series of manually determined key waypoints are recorded to then be played back (see Fig.7F). The source code details are presented in section S1, with the software system integration and hardware interfacing which allows the programming and replay of waypoints on the hand and arm described in Fig.S6. #### 4.2.1 Programming the hand The ADAPT Hand is controlled by directly commanding the tendon displacements for each actuator (which is proportional to the motor angle). To simplify the procedure to manually record waypoints, the hand (fingers) is operated using a custom built signal mixer box with 12 linear potentiometers which map to the position of each actuator. As illustrated in the top left of Fig.7F, two waypoints can be simply defined by varying the linear potentiometer positions. Being an interactive device, the mixer box allows for quickly programming motions while having full control over the tenon positions. Once a set of one or more waypoints (slider position) are recorded, the robot can smoothly move between the waypoints as in the top right of Fig.7F. #### 4.2.2 Programming the arm The Franka Research 3 robot arm is controlled using a gravity compensated impedance control introduced in [46], where the end effector 6 dof pose and corresponding stiffness can be commanded. To program the arm, the end effector stiffness is set to zero which allows the arm to be manually moved around. Likewise to the hand, after recording few key poses (see bottom left of Fig.7F), the 6 dof waypoints are interpolated and replayed. For motions which involve both the hand and arm waypoints are replayed sequentially, meaning the hand and arm are not actively actuated at the same time. ### 4.3 Experimental setup and procedure #### 4.3.1 Measuring human compliance The compliance (force displacement characteristics) for a human is measured by recording the reaction force on a loadcell and its Cartesian position as it is moved by a 6-axis robot arm (UR5), while the human is instructed to relax. The measurement setup and movement directions for the skin, finger and wrist is shown in Fig.S7. The forces are recorded for multiple repeats to capture the variation in the muscle activity. #### 4.3.2 Measuring low-level interactions Three tasks: finger sliding, knob turning, and finger gaiting, were conducted to characterise the low level interactions (such as contact forces and kinematics) of the ADAPT Hand skin/finger with the environment. The experimental setup are shown in Fig.8A,B,C for the three tasks respectively. Figure 8: Experimental setup and independent (in blue) and dependent (in orange) variables for the finger sliding experiment (A), knob turning experiment (B), and the finger gaiting experiment (C). In the finger sliding task, a single finger interacts with a wooden plate through two sliding motions generated by combining a flexing motion of the MCP joint and a flexing or extending motion of the PIP/DIP joints (see Fig.S2). Fig.8A shows the experimental setup where the ADAPT Hand (rigid held by a UR5 arm) interacts with the wooden plate mounted above two load cells measuring the vertical ($F_{V}$) and horizontal ($F_{H}$) forces. The finger joint angle data used in Fig.4B were extracted by recording the April tag markers throughout the motion. For the two motions, five independent variables were combinatorialy tested with two repeats which are: pose offset $\Delta\theta$ ($\pm 10\deg$), $\Delta z$ ($\pm 10$mm), soft and rigid configurations for the skin and finger, and overdrive of the MCP tendon $\Delta_{\mathrm{MCP}}$ ($\pm 7.5$mm) (only for the soft finger). In the knob turning task, the middle finger and thumb was used to turn a knob shown in Fig.8B. The knob turn angle $\theta$ is used to assess the performance measured by a position encoder (AMS AS5048B). The environment was varied in three ways: the $x$ and $y$ position offset($\pm 10$mm each), the diameter of the knob $d$($45\pm 15$mm), and the reaction torque of the knob $\tau$ (low: $3.3\pm 0.6$Nmm, High: $18.8\pm 3.9$Nm). The reaction force is modulated by varying the vertical forces applied on the knob which rests on a plastic surface. When the finger is in the soft configuration, the motion is near-identical to the sliding motions introduced in the finger sliding experiment. In the rigid finger configuration, a secondary motion is programmed to replicate the same motion to ensure the robot doesn’t damage itself. In the finger gaiting task, a plastic block is held between the thumb and four fingers shown in Fig.8C. Starting from all four fingers contacting the block, a finger gaiting pattern is executed (shown by 1, 2, 3, 4) in repeat until eventually the block is dropped. Only the width $w$ of the block is varied during this experiment ($\pm 5$mm), while the number of completed gaits and the holding force $F_{\mathrm{grip}}$ is measured by an inbuilt load cell. #### 4.3.3 Grasp type categorization The grasp types shown in Fig.6C for both the robot and human are categorized based on which part of the hand is used to hold/interact with the object and its posture. Each grasp is also related with the grasp taxonomies in[5]. The Finger surface grasp (only present for the human) is achieved by keeping the four fingers straight, and using that as a surface to push the object against by the thumb tip (corresponds with #22:Parallel extension taxonomy). The Fingertip grasp uses only the tips of the fingers and thumb to hold the object (corresponds with #6-8:Prismatic 2-4 finger taxonomy). The Tip and thumb grasp uses the fingertips and the middle and/or proximal phalanges (corresponds with #10:Power disk taxonomy). The Power (small) and Power (large) grasps are both power grasps where one or more phalanges of the fingers/thumb and the palm is used, distinguished based on the diameter of the grasp (corresponds to #2:Small diameter and #1:Large diameter taxonomies). Although only 24 objects are used for the experiment, 25 grasps are recorded because the paper tape generated two distinct grasp types. ## References * [1] David A Rosenbaum et al. “Cognition, action, and object manipulation.” In _Psychological bulletin_ 138.5 American Psychological Association, 2012, pp. 924 * [2] Visakha K Nanayakkara et al. “The role of morphology of the thumb in anthropomorphic grasping: a review” In _Frontiers in mechanical engineering_ 3 Frontiers Media SA, 2017, pp. 5 * [3] Andrew C Murphy et al. “Structure, function, and control of the human musculoskeletal network” In _PLoS biology_ 16.1 Public Library of Science San Francisco, CA USA, 2018, pp. e2002811 * [4] Marco Santello et al. “Hand synergies: Integration of robotics and neuroscience for understanding the control of biological and artificial hands” In _Physics of life reviews_ 17 Elsevier, 2016, pp. 1–23 * [5] Thomas Feix et al. “The grasp taxonomy of human grasp types” In _IEEE Transactions on human-machine systems_ 46.1 IEEE, 2015, pp. 66–77 * [6] Lucia Seminara et al. “A hierarchical sensorimotor control framework for human-in-the-loop robotic hands” In _Science Robotics_ 8.78 American Association for the Advancement of Science, 2023, pp. eadd5434 * [7] C Piazza, G Grioli, MG Catalano and AJAROC Bicchi “A century of robotic hands” In _Annual Review of Control, Robotics, and Autonomous Systems_ 2 Annual Reviews, 2019, pp. 1–32 * [8] Rolf Pfeifer, Fumiya Iida and Gabriel Gomez “Designing Intelligent Robots—On the Implications of Embodiment—” In _Journal of the Robotics Society of Japan_ 24.7 The Robotics Society of Japan, 2006, pp. 783–790 * [9] Rolf Pfeifer, Max Lungarella and Fumiya Iida “Self-organization, embodiment, and biologically inspired robotics” In _science_ 318.5853 American Association for the Advancement of Science, 2007, pp. 1088–1093 * [10] Aslan Miriyev and Mirko Kovač “Skills for physical artificial intelligence” In _Nature Machine Intelligence_ 2.11 Nature Publishing Group UK London, 2020, pp. 658–660 * [11] Raphael Deimel and Oliver Brock “A novel type of compliant and underactuated robotic hand for dexterous grasping” In _The International Journal of Robotics Research_ 35.1-3 SAGE Publications Sage UK: London, England, 2016, pp. 161–185 * [12] Steffen Puhlmann, Jason Harris and Oliver Brock “RBO Hand 3: A platform for soft dexterous manipulation” In _IEEE Transactions on Robotics_ 38.6 IEEE, 2022, pp. 3434–3449 * [13] Raphael Deimel and Oliver Brock “A compliant hand based on a novel pneumatic actuator” In _2013 IEEE International Conference on Robotics and Automation_ , 2013, pp. 2047–2053 IEEE * [14] Jianshu Zhou et al. “A soft-robotic approach to anthropomorphic robotic hand dexterity” In _IEEE Access_ 7 IEEE, 2019, pp. 101483–101495 * [15] Oliver Shorthose, Alessandro Albini, Liang He and Perla Maiolino “Design of a 3D-printed soft robotic hand with integrated distributed tactile sensing” In _IEEE Robotics and Automation Letters_ 7.2 IEEE, 2022, pp. 3945–3952 * [16] Aditya Bhatt, Adrian Sieler, Steffen Puhlmann and Oliver Brock “Surprisingly robust in-hand manipulation: An empirical study” In _arXiv preprint arXiv:2201.11503_ , 2022 * [17] M Vande Weghe, Matthew Rogers, Michael Weissert and Yoky Matsuoka “The ACT hand: Design of the skeletal structure” In _IEEE International Conference on Robotics and Automation, 2004. Proceedings. ICRA’04. 2004_ 4, 2004, pp. 3375–3379 IEEE * [18] Harold M Frost “Perspectives: a vital biomechanical model of synovial joint design” In _The Anatomical Record_ 240.1 Wiley Online Library, 1994, pp. 1–18 * [19] Yasunori Toshimitsu et al. “Getting the Ball Rolling: Learning a Dexterous Policy for a Biomimetic Tendon-Driven Hand with Rolling Contact Joints” In _2023 IEEE-RAS 22nd International Conference on Humanoid Robots (Humanoids)_ , 2023, pp. 1–7 IEEE * [20] Carolyn T Wadsworth “Clinical anatomy and mechanics of the wrist and hand” In _Journal of Orthopaedic & Sports Physical Therapy_ 4.4 JOSPT, Inc. JOSPT, 1033 North Fairfax Street, Suite 304, Alexandria, VA …, 1983, pp. 206–216 * [21] Yong-Jae Kim, Junsuk Yoon and Young-Woo Sim “Fluid lubricated dexterous finger mechanism for human-like impact absorbing capability” In _IEEE Robotics and Automation Letters_ 4.4 IEEE, 2019, pp. 3971–3978 * [22] Utku Çulha and Fumiya Iida “Enhancement of finger motion range with compliant anthropomorphic joint design” In _Bioinspiration & biomimetics_ 11.2 IOP Publishing, 2016, pp. 026001 * [23] James M Bern et al. “Simulation and fabrication of soft robots with embedded skeletons” In _2022 International Conference on Robotics and Automation (ICRA)_ , 2022, pp. 5205–5211 IEEE * [24] Zhe Xu and Emanuel Todorov “Design of a highly biomimetic anthropomorphic robotic hand towards artificial limb regeneration” In _2016 IEEE International Conference on Robotics and Automation (ICRA)_ , 2016, pp. 3485–3492 IEEE * [25] Lael U Odhner et al. “A compliant, underactuated hand for robust manipulation” In _The International Journal of Robotics Research_ 33.5 SAGE Publications Sage UK: London, England, 2014, pp. 736–752 * [26] Mariangela Manti et al. “A bioinspired soft robotic gripper for adaptable and effective grasping” In _Soft Robotics_ 2.3 Mary Ann Liebert, Inc. 140 Huguenot Street, 3rd Floor New Rochelle, NY 10801 USA, 2015, pp. 107–116 * [27] Aaron M Dollar and Robert D Howe “A robust compliant grasper via shape deposition manufacturing” In _IEEE/ASME transactions on mechatronics_ 11.2 IEEE, 2006, pp. 154–161 * [28] Giorgio Grioli et al. “Adaptive synergies: an approach to the design of under-actuated robotic hands” In _2012 IEEE/RSJ International Conference on Intelligent Robots and Systems_ , 2012, pp. 1251–1256 IEEE * [29] Cosimo Della Santina et al. “Toward dexterous manipulation with augmented adaptive synergies: The pisa/iit softhand 2” In _IEEE Transactions on Robotics_ 34.5 IEEE, 2018, pp. 1141–1156 * [30] Kieran Gilday, Josie Hughes and Fumiya Iida “Sensing, actuating, and interacting through passive body dynamics: A framework for soft robotic hand design” In _Soft Robotics_ 10.1 Mary Ann Liebert, Inc., publishers 140 Huguenot Street, 3rd Floor New …, 2023, pp. 159–173 * [31] JAE Hughes, P Maiolino and Fumiya Iida “An anthropomorphic soft skeleton hand exploiting conditional models for piano playing” In _Science Robotics_ 3.25 American Association for the Advancement of Science, 2018, pp. eaau3098 * [32] Kieran Gilday, Josie Hughes and Fumiya Iida “Wrist-driven passive grasping: interaction-based trajectory adaption with a compliant anthropomorphic hand” In _Bioinspiration & Biomimetics_ 16.2 IOP Publishing, 2021, pp. 026024 * [33] Kieran Gilday and Fumiya Iida “Intelligent Soft Hands and Benchmarking towards General-Purpose Robotic Manipulation” In _IOP Conference Series: Materials Science and Engineering_ 1261.1, 2022, pp. 012010 IOP Publishing * [34] Silvia Cruciani et al. “Benchmarking in-hand manipulation” In _IEEE Robotics and Automation Letters_ 5.2 IEEE, 2020, pp. 588–595 * [35] Hussein Mnyusiwalla et al. “A bin-picking benchmark for systematic evaluation of robotic pick-and-place systems” In _IEEE Robotics and Automation Letters_ 5.2 IEEE, 2020, pp. 1389–1396 * [36] Sudeep Dasari et al. “RB2: Robotic Manipulation Benchmarking with a Twist” In _NeurIPS 2021 Datasets and Benchmarks Track_ , 2021 * [37] Berk Calli et al. “The ycb object and model set: Towards common benchmarks for manipulation research” In _2015 international conference on advanced robotics (ICAR)_ , 2015, pp. 510–517 IEEE * [38] A Kapandji “Clinical test of apposition and counter-apposition of the thumb” In _Annales de chirurgie de la main: organe officiel des societes de chirurgie de la main_ 5.1, 1986, pp. 67–73 * [39] Anand Vazhapilli Sureshbabu, Giorgio Metta and Alberto Parmiggiani “A systematic approach to evaluating and benchmarking robotic hands—the ffp index” In _Robotics_ 8.1 MDPI, 2019, pp. 7 * [40] Alexander Badri-Spröwitz, Alborz Aghamaleki Sarvestani, Metin Sitti and Monica A Daley “BirdBot achieves energy-efficient gait with minimal control using avian-inspired leg clutching” In _Science Robotics_ 7.64 American Association for the Advancement of Science, 2022, pp. eabg4055 * [41] Francesco Stella, Mickael Achkar, Cosimo Della Santina and Josie Hughes “PAWS: A Synergy-Based Robotic Quadruped Leveraging Passivity for Robustness and Behavioural Diversity”, 2023 * [42] Yvan Bourquin, Auke Jan Ijspeert and Inman Harvey “Self-organization of locomotion in modular robots” In _Unpublished Diploma Thesis, http://birg. epfl. ch/page53073. html_ , 2004 * [43] Thomas Feix, Ian M Bullock and Aaron M Dollar “Analysis of human grasping behavior: Object characteristics and grasp type” In _IEEE transactions on haptics_ 7.3 IEEE, 2014, pp. 311–323 * [44] Farshad Khadivar et al. “EMG-driven shared human-robot compliant control for in-hand object manipulation in hand prostheses” In _Journal of Neural Engineering_ 19.6 IOP Publishing, 2022, pp. 066024 * [45] Auke Jan Ijspeert, Alessandro Crespi, Dimitri Ryczko and Jean-Marie Cabelguen “From swimming to walking with a salamander robot driven by a spinal cord model” In _science_ 315.5817 American Association for the Advancement of Science, 2007, pp. 1416–1420 * [46] Giovanni Franzese, Anna Mészáros, Luka Peternel and Jens Kober “ILoSA: Interactive learning of stiffness and attractors” In _2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS)_ , 2021, pp. 7778–7785 IEEE ## Author contributions K.J and J.H conceived the idea and designed the research. K.J developed the robotic systems, performed the experiment, and analyzed the data. K.J and J.H wrote the manuscript. ## Competing interests The authors declare no competing interests in this work. ## Supplementary Material ## S1 Links to external resources The uncut video of the extended pick and place operation (section 2.4.2) can be found here. All the code used to operate the ADAPT Hand can be found here. The CAD model of the ADAPT Hand can be found here. All data collected from the experiments alongside any other questions should be forwarded to the corresponding author. Figure S1: The ADAPT Hand showing all 33 grasp taxonomies and the 10 Kapandji score posture. Figure S2: Visual representation of the tendon actuation space for the flexion-extension motion of a single finger. The axis show the two independent motion control for the finger. In the sliding experiment, the slide back and slide front motions are developed by moving between the four waypoints in a clockwise or counterclockwise fashion. $\Delta_{\mathrm{MCP}}$ shows the how much the MCP joint is overdriven. Figure S3: The flowchart used to estimate the geometrical limits for open-loop grasping an object. First, the directions of the finger/thumb motion for the grasp is noted. Then, an approximate area spanned by the finger/thumb in this grasping motion is obtained. Assuming the center of mass of the object must lie within the fingertips, an area where the object would collide with the fingertips is found. Finally, subtracting the two areas gives the estimated geometrical limits of where the center of mass of the object can be placed. Figure S4: Four examples of the ADAPT Hand grasping objects when an artificial offset is given. A: Even with a large offset, the compliance and the motion of the hand self-centers the coke can to be grasped. B, C, D: Examples where the resultant grasp differs from that of $\Delta_{M}=0$. Figure S5: Photograph to show the ware and tare of the silicone fingertips before and after the extended pick and place experiment. Figure S6: A high level software and system integration diagram used to control the ADAPT Hand. The heart of the control is the teaching and playback system described in section 4.2. When teaching (recording) different motions, the hand is actuated using a signal mixer box where the sliders correspond to demand tendon positions. No control input is given to the arm to be moved freely by hand. In the playback mode, the recorded waypoints (a sequence of csv files for the hand or the arm) will be executed sequentially either to the hand or arm. The full system is developed using ROS2 to interface between microcontrollers (micro-ros), custom processes, ADAPT Hand (ROS2 control driver), and the Franka impedance controller[46]. Figure S7: The measurement of human compliance (force- displacement characteristics) using a UR5 robot arm mounted with a 10kg loadcell. The robot arm is moved in different axis to while the human is instructed to relax.
# E Pluribus Unum Ex Machina: Learning from Many Collider Events at Once Benjamin Nachman<EMAIL_ADDRESS>Physics Division, Lawrence Berkeley National Laboratory, Berkeley, CA 94720, USA Berkeley Institute for Data Science, University of California, Berkeley, CA 94720, USA Jesse Thaler <EMAIL_ADDRESS>Center for Theoretical Physics, Massachusetts Institute of Technology, Cambridge, MA 02139, USA The NSF AI Institute for Artificial Intelligence and Fundamental Interactions ###### Abstract There have been a number of recent proposals to enhance the performance of machine learning strategies for collider physics by combining many distinct events into a single ensemble feature. To evaluate the efficacy of these proposals, we study the connection between single-event classifiers and multi- event classifiers under the assumption that collider events are independent and identically distributed (IID). We show how one can build optimal multi- event classifiers from single-event classifiers, and we also show how to construct multi-event classifiers such that they produce optimal single-event classifiers. This is illustrated for a Gaussian example as well as for classification tasks relevant for searches and measurements at the Large Hadron Collider. We extend our discussion to regression tasks by showing how they can be phrased in terms of parametrized classifiers. Empirically, we find that training a single-event (per-instance) classifier is more effective than training a multi-event (per-ensemble) classifier, as least for the cases we studied, and we relate this fact to properties of the loss function gradient in the two cases. While we did not identify a clear benefit from using multi- event classifiers in the collider context, we speculate on the potential value of these methods in cases involving only approximate independence, as relevant for jet substructure studies. ††preprint: MIT-CTP 5271 ###### Contents 1. I Introduction 2. II The Statistics of Per-Ensemble Learning 1. II.1 Review of Per-Instance Learning 2. II.2 Per-Ensemble Binary Classification 3. II.3 Comparing the Loss Gradients 4. II.4 Per-Ensemble Regression 1. II.4.1 Maximum Likelihood 2. II.4.2 Classifier Loss 3. II.4.3 Direct Regression 5. II.5 Beyond Regression 3. III Empirical Studies 1. III.1 Classifiers: Multi-Event from Single-Event 1. III.1.1 Two Gaussian Example 2. III.1.2 Dijet Resonance Search 3. III.1.3 Top Quark Mass Measurement 2. III.2 Classifiers: Single-Event from Multi-Event 3. III.3 Comparison of Regression Strategies 1. III.3.1 Gaussian Mean Example 2. III.3.2 Top Quark Mass Measurement 4. III.4 Beyond Regression Example 4. IV Conclusions 5. A Deriving Maximum Likelihood Classifier Loss ## I Introduction Modern machine learning techniques are being widely applied to enhance or replace existing analysis techniques across collider physics [1, 2, 3, 4, 5, 6]. These approaches hold great promise for new particle searches, for Standard Model measurements, and for high-energy nuclear physics investigations. A subset of these proposals have advocated for a multi-event strategy whereby a machine-learned function acts on multiple collision events at the same time [7, 8, 9, 10, 11, 12, 13, 14]. This multi-event (per- ensemble) strategy contrasts with more typical single-event (per-instance) machine learning methods that process one event at a time, although both strategies make use of many events during the training process. Intuitively, an ensemble approach might seem like a more promising learning strategy because there is more information contained in $N>1$ collision events than in one single event. There is, however, an important distinction between the amount of information contained in a data set and the amount of information needed to encode a machine-learned function. For this reason, there need not be a gain from using multi-event strategies over single-event strategies in the context of machine learning. In this paper, we show that when directly compared on the same task, there is indeed no informational benefit from training a function that processes multiple events simultaneously compared to training a function that processes only a single event at a time. This fact can be easily understood from the statistical structure of collision data. To test for a practical benefit, we perform empirical comparisons of per-ensemble and per-instance methods on benchmark tasks relevant for the Large Hadron Collider (LHC), finding that single-event (per-instance) methods are more effective for the cases we studied. To an excellent approximation, collider events are statistically independent and identically distributed (IID). In simulation, this is exactly true up to deficiencies in random number generators. In data, there are some small time- dependent effects from changing conditions and there are also some correlations between events introduced by detector effects with timescales longer than a typical bunch crossing. These event-to-event correlations, however, are truly negligible when considering the set of events typically used for physics analysis that are selected by triggers. The probability for two events next to each other in time to be saved by the triggers is effectively zero, since triggers save only a tiny fraction of events. The IID nature of collision events therefore ensures that the information content is the same for ensembles of events and for single events drawn from an ensemble. In equations, the probability to observe $N$ events $x_{i}$ is $p(\\{x_{1},\ldots,x_{N}\\}|\theta)=\prod_{i=1}^{N}p(x_{i}|\theta),$ (1) where $\theta$ represents possible parameters of the generative model, such as the physics process being studied or the values of coupling constants. The optimal classifier to distinguish whether events have been generated via $\theta_{A}$ or via $\theta_{B}$ depends only on the per-ensemble likelihood ratio [15]: $\frac{p(\\{x_{1},\ldots,x_{N}\\}|\theta_{A})}{p(\\{x_{1},\ldots,x_{N}\\}|\theta_{B})}=\prod_{i=1}^{N}\frac{p(x_{i}|\theta_{A})}{p(x_{i}|\theta_{B})},$ (2) which by the IID assumption only depends on knowing the per-instance likelihood ratio $p(x_{i}|\theta_{A})/p(x_{i}|\theta_{B})$. This equality explains the informational equivalence of per-ensemble and per-event learning. Given the simplicity of Eq. (2), why are we writing a whole paper on this topic (apart from the opportunity to invoke a gratuitously Latinate paper title that incorporates an aspiration for national unity)? The studies in Refs. [7, 8, 9, 10, 11, 12, 13, 14] find that per-ensemble learning is effective for their respective tasks, in some cases arguing why per-instance learning is deficient. It is certainly true that a set of events $\\{x_{1},\ldots,x_{N}\\}$ contains more information than a single event $x_{i}$ drawn from this set. What we will show in this paper is that if one carefully combines the per-instance information, one can recover the per- ensemble benefit, with the potential for a substantially reduced training cost. We emphasize that our analysis does not contradict the studies in Refs. [7, 8, 9, 10, 11, 12, 13, 14]; rather this work suggests the possibility of achieving the same or better results by replacing per-ensemble learning with per-instance learning. There may be specialized contexts where per-ensemble learning is superior, particularly if the training procedure itself can be made simpler, such as in the linear regression approach of Ref. [12]. This paper also gives us a chance to mention some facts about loss functions that are well known in the statistics literature but might not be as well appreciated in collider physics. Moving away from the IID case, we speculate on the relevance of our analysis for jet substructure tasks where there is a notion of approximate independence of emissions. The remainder of this paper is organized as follows. In Sec. II, we provide the formal statistical basis for building multi-event classifiers from single- event classifiers, and vice versa, under the IID assumption. We also explain how regression tasks can be translated into the language of per-instance parametrized classification. In Sec. III, we present empirical studies that corroborate these analytic results. Our conclusions are given in Sec. IV. ## II The Statistics of Per-Ensemble Learning ### II.1 Review of Per-Instance Learning Suppose that a collider event is represented by features in $\mathbb{E}=\mathbb{R}^{M}$ and we are trying to train a binary classifier to learn a target in $[0,1]$. Let $c:\mathbb{E}\rightarrow[0,1]$ be a function that processes a single event, with the goal of distinguishing events being generated by $\theta_{A}$ ($c\to 1$) versus those generated by $\theta_{B}$ ($c\to 0$). Such a function can be obtained by minimizing an appropriate loss functional, such as the binary cross entropy: $\displaystyle L_{\rm BCE}[c]=-\int dx\,\Big{(}$ $\displaystyle p(x|\theta_{A})\log c(x)$ $\displaystyle~{}+p(x|\theta_{B})\log(1-c(x))\Big{)},$ (3) where $p(x|\theta)$ is the probability density of $x\in\mathbb{E}$ given class $\theta$. Here and throughout this discussion, we consider the infinite statistics limit such that we can replace sums over events by integrals. We have also dropped the prior factors $p(\theta_{i})$, assuming that one has equal numbers of examples from the two hypotheses during training. While this is often true in practice, it is not strictly necessary for our main conclusions, though it does simplify the notation. It is well-known [16, 17] (also in high-energy physics [18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30]) that an optimally trained $c$ will have the following property: $\displaystyle\frac{c(x)}{1-c(x)}=\frac{p(x|\theta_{A})}{p(x|\theta_{B})},$ (4) such that one learns the per-instance likelihood ratio. By the Neyman–Pearson lemma [15], this defines the optimal single-event classifier. Loss Name | $A(f)$ | $B(f)$ | $\operatorname*{argmin}_{f}L[f]$ | Integrand of $-\min_{f}L[f]$ | Related Divergence/Distance ---|---|---|---|---|--- Binary Cross Entropy | $\log f$ | $\log(1-f)$ | $\frac{p_{A}}{p_{A}+p_{B}}$ | $p_{A}\log\frac{p_{A}}{p_{A}+p_{B}}+(A\leftrightarrow B)$ | $2\big{(}\text{Jensen-Shannon}-\log 2\big{)}$ Mean Squared Error | $-(1-f)^{2}$ | $-f^{2}$ | $\frac{p_{A}}{p_{A}+p_{B}}$ | $-\frac{p_{A}p_{B}}{p_{A}+p_{B}}$ | $\frac{1}{2}\big{(}\text{Triangular}-1\big{)}$ Square Root | $\frac{-1}{\sqrt{f}}$ | $-\sqrt{f}$ | $\frac{p_{A}}{p_{B}}$ | $-2\sqrt{p_{A}p_{B}}$ | $2\big{(}\text{Hellinger}^{2}-1\big{)}$ Maximum Likelihood Cl. | $\log f$ | $1-f$ | $\frac{p_{A}}{p_{B}}$ | $p_{A}\log\frac{p_{A}}{p_{B}}$ | Kullback–Leibler Table 1: Examples of loss functionals in the form of Eq. (5), with the associated location and value of the loss minimum, using the shorthand $p_{i}\equiv p(x|\theta_{i})$. We have used the symbol $f$ in all cases to denote the classifier, but some choices require explicit constraints on $f$ to be either non-negative or in the range $[0,1]$. In the last column, we indicate the relation of the loss minimum to statistical divergences and distances, up to an overall scaling and offset. See Ref. [31] for additional relations. There are many loss functionals that satisfy this property. Consider a more general loss functional that depends on a learnable function $f:\mathbb{E}\rightarrow\mathbb{R}$ (which unlike $c$ may or may not map to $[0,1]$) as well as fixed rescaling functions $A:\mathbb{R}\to\mathbb{R}$ and $B:\mathbb{R}\to\mathbb{R}$: $\displaystyle L[f]=-\int dx\,$ $\displaystyle\Big{(}p(x|\theta_{A})\,A(f(x))+p(x|\theta_{B})\ B(f(x))\Big{)}.$ (5) Taking the functional derivative with respect to $f(x)$, the extremum of $L[f]$ satisfies the property: $\displaystyle-\frac{B^{\prime}(f(x))}{A^{\prime}(f(x))}=\frac{p(x|\theta_{A})}{p(x|\theta_{B})}.$ (6) As long as $-B^{\prime}(f)/A^{\prime}(f)$ is a monotonic rescaling of $f$ and the overall loss functional is convex, then the function $f(x)$ learned by minimizing Eq. (5) defines an optimal classifier. In many cases, the minimum value of $L[f]$ itself is interesting in the context of statistical divergences and distances [31], and a few examples are shown in Table 1. To simplify the following discussion, we will focus on the “maximum likelihood classifier” (MLC) loss: $\displaystyle L_{\text{MLC}}[f]=-\int dx\,\Big{(}$ $\displaystyle p(x|\theta_{A})\log f(x)$ $\displaystyle~{}+p(x|\theta_{B})\,(1-f(x))\Big{)}.$ (7) This is of the general form in Eq. (5) with $A(f)=\log f$ and $B(f)=1-f$. To our knowledge, the MLC was first introduced in the collider physics context in Refs. [32, 33], although with an exponential parametrization of $f(x)$. We call Eq. (7) the MLC loss to distinguish it from the related maximum likelihood loss that is often used to fit generative models [34, 35, 36]. Using Eq. (6), the minimum of this loss functional yields directly the likelihood ratio: $\operatorname*{argmin}_{f}L_{\text{MLC}}[f]=\frac{p(x|\theta_{A})}{p(x|\theta_{B})},$ (8) which will be useful to simplify later analyses.111 A variation of Eq. (8) holds for $A(f)=\log C(f)$ and $B(f)=1-C(f)$, where $C(f)$ is any monotonically increasing function with range that covers $(0,\infty)$. In this case, $C(\operatorname*{argmin}_{f}L[f])=p(x|\theta_{A})/p(x|\theta_{B})$. This can be useful in practice if $C(f)$ is everywhere positive, since $f$ can take on negative values and still yield a valid likelihood ratio. See Fig. 10 for an empirical study of $C(f)=\exp f$. The MLC loss functional value at the minimum is $-\min_{f}L_{\text{MLC}}[f]=\int dx\,p(x|\theta_{A})\log\frac{p(x|\theta_{A})}{p(x|\theta_{B})},$ (9) which is the Kullback–Leibler (KL) divergence, also known as the relative entropy from $p(x|\theta_{B})$ to $p(x|\theta_{A})$. See App. A for an intuitive derivation of Eq. (7). ### II.2 Per-Ensemble Binary Classification To move from single-event classification to multi-event classification, we want to learn a classification function $f_{N}$ that can process $N$ events simultaneously. Here, we are using $f_{N}:\mathbb{E}^{N}\rightarrow\mathbb{R}$ instead of $c_{N}:\mathbb{E}^{N}\rightarrow[0,1]$ to avoid algebraic manipulations like Eq. (4). We will use the vector notation $\vec{x}=\\{x_{1},\dots,x_{N}\\}$ (10) to represent an element of $\mathbb{E}^{N}$. Our goal is to distinguish whether $\vec{x}$ is drawn from $p(\vec{x}|\theta_{A})$ ($f_{N}\to\infty$) or from $p(\vec{x}|\theta_{B})$ ($f_{N}\to 0$). Note that we are trying to classify a pure event ensemble as coming from either $\theta_{A}$ or $\theta_{B}$, which is a different question than trying to determine the proportion of events drawn from each class in a mixed event ensemble. For $N=1$, $f_{1}$ is the same as $f$ discussed in Eq. (5). If $f_{N}$ is trained optimally, then the classification performance of $f_{N}$ evaluated on $N>1$ events will be better than the performance of $f_{1}$ evaluated on a single event, as relevant to the discussions in Refs. [7, 8, 9, 10, 11, 12, 13, 14]. The key point of this paper is that one can construct a classifier $f_{1\to N}$ that is built only from $f_{1}$, acts on $N$ events, and has the same asymptotic performance as $f_{N}$. Using the MLC loss in Eq. (7), but now applied to $N$ events, we have $\displaystyle L_{\text{MLC}}[f_{N}]=-\int d^{N}x\,\Big{(}$ $\displaystyle p(\vec{x}|\theta_{A})\,\log f_{N}(\vec{x})$ $\displaystyle~{}+p(\vec{x}|\theta_{B})\,(1-f_{N}(\vec{x}))\Big{)},$ (11) whose minimum is the per-ensemble likelihood ratio: $\operatorname*{argmin}_{f_{N}}L_{\text{MLC}}[f_{N}]=\frac{p(\vec{x}|\theta_{A})}{p(\vec{x}|\theta_{B})}.$ (12) By the Neyman–Pearson lemma, this yields the optimal per-ensemble classifier. On the other hand, once we have trained a single-event classifier $f_{1}$ using Eq. (7), we can build a multi-event classifier $f_{1\to N}$ without any additional training: $f_{1\to N}(\vec{x})\equiv\prod_{i=1}^{N}f_{1}(x_{i})\quad\rightarrow\quad\frac{p(\vec{x}|\theta_{A})}{p(\vec{x}|\theta_{B})},$ (13) where in the last step we have combined the solution found in Eq. (8) with the IID condition in Eq. (2). Whereas minimizing Eq. (11) requires sampling over $\mathbb{E}^{N}$, constructing $f_{1\to N}$ only requires sampling over $\mathbb{E}$, which is a considerable reduction in computational burden for large $N$. The technical details of carrying out this procedure are explained in Sec. III.1. Going in the converse direction, we can learn a single-event classifier $f_{N\to 1}$ starting from a constrained multi-event classifier $\tilde{f}_{N}$. Using weight sharing, we can minimize Eq. (11) subject to the constraint that $\tilde{f}_{N}$ takes the functional form: $\tilde{f}_{N}(\\{x_{1},\dots,x_{N}\\})=\prod_{i=1}^{N}f_{N\to 1}(x_{i}),$ (14) where $f_{N\to 1}(x)$ is a learnable function. Under the IID assumption, $\tilde{f}_{N}$ can still learn the per-ensemble likelihood ratio, but the learned $f_{N\to 1}(x)$ will now be the per-instance likelihood ratio, at least asymptotically.222In the case that the two samples are composed of mixtures of two categories, then the learned $f_{N\to 1}(x)$ will be the ratio of the mixed sample likelihoods, which is monotonically related to the optimal pure sample classifier, as discussed in Ref. [37]. An examination of this converse construction is presented in Sec. III.2. ### II.3 Comparing the Loss Gradients We have shown that the per-ensemble classifier $f_{N}$ and the composite per- event classifier $f_{1\to N}$ have the same asymptotic information content, but one might wonder if there is nevertheless a practical performance gain to be had using per-ensemble learning. Under the IID assumption, the optimal $f_{N}$ takes the form of $\tilde{f}_{N}$ in Eq. (14), and in our empirical studies, we found no benefit to letting $f_{N}$ have more functional freedom. Therefore, to get a sense of the efficacy of per-ensemble versus per-instance training, we can compare the effective loss functions for $f_{N\to 1}$ and $f_{1}$. Since the inputs and outputs of these functions are the same (i.e. $\mathbb{E}\to\mathbb{R}$), we can do an apples-to-apples comparison of their behavior under gradient descent. The following analysis assumes that the neural network training occurs in the vicinity of the global minimum of the loss function. For the per-ensemble case, plugging Eq. (14) into Eq. (11) and using the IID relation in Eq. (1), we find the effective loss functional: $\displaystyle L_{\rm MLC}[f_{N\to 1}]+1$ $\displaystyle=-N\int dx\,p(x|\theta_{A})\,\log f_{N\to 1}(x)$ $\displaystyle\quad+\left(\int dx\,p(x|\theta_{B})f_{N\to 1}(x)\right)^{N}.$ (15) This is to be contrasted with the per-instance loss functional from Eq. (7), repeated for convenience with the $f_{1}$ notation and typeset to be parallel to the above: $\displaystyle L_{\text{MLC}}[f_{1}]+1$ $\displaystyle=-\int dx\,p(x|\theta_{A})\,\log f_{1}(x)$ $\displaystyle\quad+\int dx\,p(x|\theta_{B})\,f_{1}(x).$ (16) To understand the loss gradients, we can Taylor expand the learned functions about the optimal solution: $\displaystyle f_{N\to 1}(x)$ $\displaystyle=\frac{p(x|\theta_{A})}{p(x|\theta_{B})}+\epsilon(x),$ (17) $\displaystyle f_{1}(x)$ $\displaystyle=\frac{p(x|\theta_{A})}{p(x|\theta_{B})}+\epsilon(x).$ (18) Plugging these into their respective loss functionals and looking at the leading-order variations, we have: $\displaystyle\frac{\delta L_{\rm MLC}[f_{N\to 1}]}{N}$ $\displaystyle=\int dx\,\frac{\big{(}p(x|\theta_{B})\,\epsilon(x)\big{)}^{2}}{2\,p(x|\theta_{A})}$ $\displaystyle\quad+\frac{N-1}{2}\left(\int dx\,p(x|\theta_{B})\,\epsilon(x)\right)^{2},$ (19) $\displaystyle\delta L_{\rm MLC}[f_{1}]$ $\displaystyle=\int dx\,\frac{\big{(}p(x|\theta_{B})\,\epsilon(x)\big{)}^{2}}{2\,p(x|\theta_{A})}.$ (20) These expressions are quadratic in $\epsilon(x)$, which means that we are expanding around the correct minimum. The expression for $\delta L_{\rm MLC}[f_{1}]$ involves a single integral over $x$, so under gradient descent, the value of $\epsilon(x)$ can be independently adjusted at each point in phase space to find the minimum. By contrast, $\delta L_{\rm MLC}[f_{N\to 1}]$ has an additional piece involving an integral squared, so even if at a given point in phase space $x_{0}$ we have achieved $\epsilon(x_{0})=0$, gradient descent will tend to push $\epsilon(x_{0})$ away from the correct value until $\epsilon(x)=0$ everywhere. This correlated structure explains the slower convergence of $L_{\rm MLC}[f_{N\to 1}]$ compared to $L_{\rm MLC}[f_{1}]$ in our empirical studies. While we focused on the MLC loss to simplify the algebra, the appearance of these (typically counterproductive) correlations in the loss gradient appears to be a generic feature of per-ensemble learning. ### II.4 Per-Ensemble Regression While the discussion above focused on binary classification, the same basic idea applies to regression problems as well. The goal of regression is to infer parameters $\theta$ from the data $\vec{x}$. There are a variety of approaches that can be used for this task, and each can be connected to parametrized per-instance classification. #### II.4.1 Maximum Likelihood Maximum likelihood is the most common strategy for inference in collider physics. Symbolically, we are trying to find $\theta_{\rm ML}=\operatorname*{argmax}_{\theta}p(\vec{x}|\theta).$ (21) One way to determine $\theta_{\rm ML}$ is with a two-step approach. First, one can train a parametrized classifier $f(x,\theta)$ [26, 38] using, e.g., the per-instance MLC loss: $\displaystyle L_{\rm MLC}[f]=-\int dx\,\Big{(}$ $\displaystyle p(x|\theta)\,p(\theta)\log f(x,\theta)$ $\displaystyle~{}+p(x|\theta_{0})\,p(\theta)\,(1-f(x,\theta))\Big{)}.$ (22) The top line corresponds to a synthetic dataset where every event is generated from $p(x|\theta)$ with different $\theta$ values drawn from the probability density $p(\theta)$. The bottom line corresponds to a synthetic dataset where every event is generated using the same $p(x|\theta_{0})$ for fixed $\theta_{0}$ and then augmented with a value $\theta$ that follows from $p(\theta)$ independently of $x$. Minimizing Eq. (22) with respect to $f(x,\theta)$, the asymptotic solution is the likelihood ratio: $f(x,\theta)=\frac{p(x|\theta)}{p(x|\theta_{0})},$ (23) where the factors of $p(\theta)$ have canceled out. Second, one can estimate $\theta_{\rm ML}$ by using the IID properties of the event ensemble to relate likelihoods to the classifier output $f(x,\theta)$: $\displaystyle\theta_{\rm ML}$ $\displaystyle=\operatorname*{argmin}_{\theta}\left\\{-\sum_{i=1}^{N}\log p(x_{i}|\theta)\right\\}$ $\displaystyle=\operatorname*{argmin}_{\theta}\left\\{-\sum_{i=1}^{N}\log\frac{p(x_{i}|\theta)}{p(x_{i}|\theta_{0})}\right\\}$ $\displaystyle\approx\operatorname*{argmin}_{\theta}\left\\{-\sum_{i=1}^{N}\log f(x_{i},\theta)\right\\}.$ (24) Thus, even though maximum likelihood regression uses information from the full event ensemble, only a parametrized per-instance classifier is required for this procedure. #### II.4.2 Classifier Loss Two recent proposals for parameter estimation are explicitly built on classifiers for regression [18, 19]. For any classifier, one can perform the following optimization:333Note that Ref. [18] used the (non-differentiable) area under the curve instead of the classifier loss, as it is not sensitive to differences in the prior $p(\theta)$ between the two data sets. $\displaystyle\theta_{\rm CL}=\operatorname*{argmax}_{\theta^{\prime}}\left\\{\begin{matrix}\text{Loss of a classifier trained}\cr\text{to distinguish $\theta^{\prime}$ from $\theta_{\rm data}$}\end{matrix}\right\\}.$ (25) Here, we are imagining that the $\theta^{\prime}$ samples come from synthetic data sets. The appearance of a maximum instead of minimum in Eq. (25) is because, as highlighted in Table 1, it is negative loss functions that correspond to statistical divergences and distances. In general, the $\theta_{\rm CL}$ that minimizes the classifier loss will be different from the $\theta_{\rm ML}$ that maximizes the likelihood. For the special case of the MLC loss, though, they are the same in the asymptotic limit if we set $\theta_{A}=\theta_{\rm data}$ and $\theta_{B}=\theta^{\prime}$. To see this, recall from Eq. (9) that after training, the value of the MLC loss is related to the KL divergence: $\displaystyle\operatorname*{argmax}_{\theta^{\prime}}\\{\min_{f}L_{\text{MLC}}[f]\\}$ $\displaystyle\hskip 5.69054pt=\operatorname*{argmax}_{\theta^{\prime}}\left\\{-\int dx\,p(x|\theta_{\rm data})\log\frac{p(x|\theta_{\rm data})}{p(x|\theta^{\prime})}\right\\}$ $\displaystyle\hskip 5.69054pt\approx\operatorname*{argmax}_{\theta^{\prime}}\left\\{\sum_{i=1}^{N}\log\frac{p(x_{i}|\theta^{\prime})}{p(x_{i}|\theta_{\rm data})}\right\\}$ $\displaystyle\hskip 5.69054pt=\operatorname*{argmin}_{\theta^{\prime}}\left\\{-\sum_{i=1}^{N}\log p(x_{i}|\theta^{\prime})\right\\}$ $\displaystyle\hskip 5.69054pt=\theta_{\rm ML}\,,$ (26) where the sum is over data events. #### II.4.3 Direct Regression In terms of information content, a regression model trained in the usual way can be built from a parametrized classification model. Suppose that $\theta\in\mathbb{R}^{Q}$ and $g_{N}:\mathbb{E}^{N}\rightarrow\mathbb{R}^{Q}$ is a regression model trained with the mean squared error loss: $L_{\rm MSE}[g_{N}]=-\int d^{n}x\,p(\vec{x},\theta)\Big{(}g_{N}(\vec{x})-\theta\Big{)}^{2}$ (27) It is well known that the optimally trained $g_{N}$ will be related to the expectation value of $\theta$: $g_{N}(\vec{x})=\mathbb{E}[\theta|\vec{x}]=\int d\theta\,\theta\,p(\theta|\vec{x}).$ (28) Other loss functions approximate other statistics, as discussed in Ref. [39]. For example, the mean absolute error loss approximates the median of $\theta$. Ultimately, all direct regression methods are functionals of $p(\theta|\vec{x})$. We can relate $p(\theta|\vec{x})$ to a parametrized classifier $f_{N}(\vec{x},\theta)$ trained to distinguish $\theta$ from a baseline $\theta_{0}$: $\displaystyle p(\theta|\vec{x})=\frac{p(\vec{x}|\theta)\,p(\theta)}{p(\vec{x})}$ $\displaystyle=\frac{p(\vec{x}|\theta)\,p(\theta)}{\int d\theta^{\prime}\,p(\vec{x}|\theta^{\prime})\,p(\theta^{\prime})}$ $\displaystyle=\frac{\frac{p(\vec{x}|\theta)}{p(\vec{x}|\theta_{0})}\,p(\theta)}{\int d\theta^{\prime}\,\frac{p(\vec{x}|\theta^{\prime})}{p(\vec{x}|\theta_{0})}\,p(\theta^{\prime})}$ $\displaystyle=\frac{f_{N}(\vec{x},\theta)\,p(\theta)}{\int d\theta^{\prime}\,f_{N}(\vec{x},\theta^{\prime})\,p(\theta^{\prime})},$ (29) where $p(\theta)$ is the probability density of $\theta$ used during the training of $g_{N}$. Following the same logic as Sec. II.2, the per-ensemble classifier $f_{N}(\vec{x},\theta)$ can be related to a per-instance classifier $f_{1}(x,\theta)$. Therefore, even though $g_{N}$ acts on $N$ events, it has the same information content as a parametrized classifier that acts on single events. Performing regression via Eqs. (28) and (29) is straightforward but tedious. In practice, one would train a parametrized per-instance classifier $f_{1}(x,\theta)$ as in Eq. (23), multiply it to construct $f_{N}(\vec{x},\theta)=\prod_{i=1}^{N}f_{1}(x_{i},\theta)$, and then sample over values of $\theta$ to approximate the integrals. We show examples of the above regression strategies in Sec. III.3 ### II.5 Beyond Regression In addition to classification and regression, a standard machine learning task is density estimation. While some classical machine learning methods like $k$-nearest neighbors [40, 41] do require multi-instance information at prediction time, many of the standard deep learning solutions to implicit or explicit generative modeling are built on per-instance functions. Such methods include generative adversarial networks [42],444 In the context of adversarial training, it may be beneficial to use per-ensemble information in the discriminator to mitigate mode collapse, as utilized in Ref. [14]. This is also the philosophy behind mini-batch discrimination [43]. variational autoencoders [44], and normalizing flows [45]. One reason for computing explicit densities is to estimate the distance to a reference density. A common set of tools for this task are the $f$-divergences mentioned earlier. As discussed in Ref. [31] and highlighted in Table 1, there is a direct mapping between the loss value of a per-instance classification task and a corresponding $f$-divergence between the underlying probability densities. A related quantity is the mutual information between two random variables $X$ and $Y$: $\displaystyle I(X,Y)=\int dx\,dy\,p(x,y)\log\frac{p(x,y)}{p(x)\,p(y)}.$ (30) For example, $Y$ could be binary (a class label) and then $I(X,Y)$ would encode how much information (in units of nats) is available in $X$ for doing classification. This can be helpful in the context of ranking input features, and was studied in the context of quark/gluon jet classification in Ref. [46]. Naively, Eq. (30) might seem like it requires estimating the densities $p(x)$, $p(y)$, and $p(x,y)$, which in turn may require ensemble information (see e.g. Ref. [47] for a study in the context of HEP). On the other hand, Eq. (30) takes the same form as the KL divergence in Eq. (9). Therefore, this quantity can be estimated using a similar strategy as in earlier sections, by training a classifier to distinguish data following $p(x,y)$ from data following $p(x)\,p(y)$ using the MLC loss. The value of the loss at the minimum will be an estimate of the mutual information. A simple example of this will be studied in Sec. III.4. ## III Empirical Studies We now present empirical studies comparing per-instance and per-ensemble data analysis strategies to highlight the points made in Sec. II. Our analyses are based on three case studies: a simple two Gaussian example, searching for dijet resonances, and measuring the top quark mass. ### III.1 Classifiers: Multi-Event from Single-Event As argued in Sec. II.2, under the IID assumption we can build multi-event classifiers from single-event classifiers. We now demonstrate how to construct $f_{1\rightarrow N}$ defined in Eq. (13), comparing its performance to $f_{N}$. #### III.1.1 Two Gaussian Example (a) (b) Figure 1: Classification in the two Gaussian example. (a) A histogram of the Gaussian random variable $X$, for the “signal” ($x_{0}=0.1$) and background ($x_{0}=-0.1$). (b) ROC curves for various binary classifiers. From the single-event classifier $f_{1}$, we can construct a multi-event classifier $f_{1\to 10}$ that matches the performance of a classifier trained on 10 events simultaneously ($f_{10}$). Our first case study involves one-dimensional Gaussian random variables. As shown in Fig. 1a, we consider two Gaussian distributions $X\sim\mathcal{N}(\pm\epsilon,1)$, with slightly different means ($x_{0}=\pm\epsilon$) but the same variance ($\sigma=1$). Here, the “signal” has positive mean while the “background” has negative mean, and we take $\epsilon=0.1$ for concreteness. Both the per-instance ($f_{1}$) and per-ensemble ($f_{N}$) classifiers are parametrized by neural networks and implemented using Keras [48] with the Tensorflow backend [49] and optimized with Adam [50]. We use the binary cross entropy loss function so Eq. (4) is needed to convert the classifier output to a likelihood ratio. Each classifier consists of two hidden layers with 128 nodes per layer. Rectified Linear Unit (ReLU) activation functions are used for the intermediate layers while sigmoid activation is used for the last layer. The only difference between the per-instance and per-ensemble networks is that the input layer has one input for $f_{1}$ but $N$ inputs for $f_{N}$. We train each network with 50,000 events to minimize the binary cross entropy loss function, and we test the performance with an additional 50,000 events. For each network, we train for up to 1000 epochs with a batch size of 10%, which means that the number of batches per epoch is the same, as is the number of events considered per batch. The training is stopped if the validation loss does not decrease for 20 consecutive epochs (early stopping). For the ensemble network, we take $N=10$. We did not do any detailed hyperparameter optimization for these studies. In Fig. 1b, we show the performance of the resulting classifiers $f_{1}$ and $f_{10}$. We checked that the $f_{1}$ classifier parametrized by a neural network has essentially the same performance as an analytic function derived by taking the ratio of Gaussian probability densities, which means that the neural network $f_{1}$ is nearly optimal. As expected, the per-instance classifier $f_{1}$ has a worse receiver operating characteristic (ROC) curve than the per-ensemble classifier $f_{10}$. This is not a relevant comparison, however, because the two are solving different classification tasks (i.e. classifying individual events as coming from signal or background versus classifying an ensemble of $N=10$ events as all coming from signal or background). With Eq. (13), we can use $f_{1}$ to build a $10$-instance classifier $f_{1\rightarrow 10}$, whose ROC curve is nearly identical to $f_{10}$, if not even slightly better. Thus, as expected from Eq. (2), all of the information in the 10-instance classifier is contained in the per-instance classifier. #### III.1.2 Dijet Resonance Search (a) (b) Figure 2: Classification in the dijet resonance search example. (a,b) Histograms of the four jet features for the signal ($W^{\prime}\to XY$) and background (QCD dijet) processes. (c) ROC curves for various binary classifiers. The multi-event classifier $f_{1\to 3}$ (built from $f_{1}$) outperforms three classifiers trained on triplets of events: $f_{3}^{\text{list}}$ with randomly ordered inputs, $f_{3}^{\text{sort}}$ with sorted inputs, and $f_{3}^{\text{set}}$ based on the deep sets/PFN strategy in Eq. (31) with built-in permutation invariance. (c) We now consider an example from collider physics, motivated by a search for new beyond-the-Standard-Model (BSM) particles in a dijet final state. The simulations used for this study were produced for the LHC Olympics 2020 community challenge [51]. The background process involves generic quantum chromodynamics (QCD) dijet events with a requirement of at least one such jet with transverse momentum $p_{T}>1.3$ TeV. The signal process involves the production of a hypothetical new resonance $W^{\prime}$ with mass $m_{W^{\prime}}=3.5$ TeV, which decays via $W^{\prime}\rightarrow XY$ to two hypothetical particles $X$ and $Y$ of masses 500 GeV and 100 GeV, respectively. Each of the $X$ and $Y$ particles decays promptly into pairs of quarks. Due to the mass hierarchy between the $W^{\prime}$ boson and its decay products, the final state is characterized by two large-radius jets with two- prong substructure. The background and signal are generated using Pythia 8.219 [52, 53]. A detector simulation is performed with Delphes 3.4.1 [54, 55, 56] using the default CMS detector card. Particle flow objects are used as inputs to jet clustering, implemented with FastJet 3.2.1 [57, 58] and the anti-$k_{t}$ algorithm [59] using $R=1.0$ for the radius parameter. Events are required to have a reconstructed dijet mass within the range $m_{JJ}<[3.3,3.7]\,\text{GeV}$. Four features are used to train our classifiers: the invariant mass of the lighter jet, the mass difference of the leading two jets, and the $N$-subjettiess ratios $\tau_{21}$ [60, 61] of the leading two jets. The observable $\tau_{21}$ quantifies the degree to which a jet is characterized by two subjets or one subjet, with smaller values indicating two-prong substructure. The mass features are recorded in units of TeV so that they are numerically $\mathcal{O}(1)$. Histograms of the four features for signal and background are shown in Figs. 2a and 2b. The signal jet masses are localized at the $X$ and $Y$ masses and the $\tau_{21}$ observables are shifted towards lower values, indicating that the jets have two-prong substructure. We train a per-instance classifier ($f_{1}$) and a per-ensemble classifier ($f_{3}$) using the same tools as for the Gaussian example above, again using binary cross entropy for the loss function. Because signal and background are so well separated in this example, we restrict our attention to $N=3$ to avoid saturating the performance. Note that this is an artificially constructed classification problem, since in a more realistic context one would be trying to estimate the signal fraction in an event ensemble, not classify triplets of events as all coming from signal or background. For $f_{1}$, the neural network architecture is the same as Ref. [18] with four hidden layers, each with 64 nodes and ReLU activation, and an output layer with sigmoid activation. For $f_{3}$, the neural network involves $4\times 3=12$ inputs, and the penultimate hidden layer is adjusted to have 128 nodes, yielding a marginal performance gain. In both cases, about 100,000 events are used for testing and training, with roughly balanced classes. All of the networks are trained for up to 1000 epochs with the same early stopping condition as in the Gaussian case and with a batch size of 10%. Following Eq. (13), we construct a tri-event classifier $f_{1\rightarrow 3}$ from $f_{1}$. The ROC curves for $f_{3}$ and $f_{1\to 3}$ are shown in Fig. 2c, with $f_{1}$ also shown for completeness. Interestingly, the $f_{1\rightarrow 3}$ classifier trained on single events significantly outperforms $f_{3}$ trained on multiple events. There are a variety of reasons for this, but one important deficiency of the $f_{3}$ classifier is that it does not respect the permutation symmetry of its inputs. Because events are IID distributed, there is no natural ordering of the events, but the fully connected architecture we are using imposes an artificial ordering. Inspired by Ref. [12], we can break the permutation symmetry of the inputs by imposing a particular order on the events. Specifically, we train a network $f_{3}^{\text{sort}}$ where the triplet of events is sorted by their leading jet mass. Using $f_{3}^{\text{sort}}$ yields a small gain in performance seen in Fig. 2, but not enough to close the gap with $f_{1\rightarrow 3}$. (a) (b) Figure 3: Classification in the top quark mass example. (a) A histogram of $m_{b_{1}\mu\nu}$ for top quark masses of 172.5 GeV and 175 GeV. The “wgt.” curve is explained later in Sec. III.3.2, where we test the performance of a likelihood reweighting. (b) The difference in efficiency for the 172.5 GeV top quark mass sample (true positive) and the 175 GeV top quark mass sample (false positive) as a function of the true positive rate for various binary classifiers. Once again, a multi-event classifier ($f_{1\to 20}$) built from the single-event classifier ($f_{1}$) has the best performance. For the classifiers trained to process 20 events simultaneously, the deep sets/PFN approach ($f_{20}^{\text{set}}$) does better than sorting the inputs ($f_{20}^{\text{sort}}$). A more powerful way to account for the permutation symmetry among events is to explicitly build a permutation-invariant neural network architecture. For this purpose, we use the deep sets approach [62]. In the particle physics context, deep sets were first used to construct particle flow networks (PFNs) [63], where the inputs involve sets of particles. Here, we are interested in sets of events, though we will still use the PFN code from the https://energyflow.network/ package. Following Refs. [62, 63], we decompose our set-based classifier as: $\displaystyle f^{\text{set}}_{N}(\vec{x})=F\left(\sum_{i=1}^{N}\Phi(x_{i})\right),$ (31) where $F:\mathbb{R}^{L}\rightarrow[0,1]$ and $\Phi:\mathbb{E}\rightarrow\mathbb{R}^{L}$ are neural networks that are simultaneously optimized. The network $\Phi$ embeds single events $x_{i}$ into a $L$-dimensional latent space. The sum operator in Eq. (31) guarantees that $f^{\text{set}}_{N}$ is invariant under permutations $x_{\sigma(i)}$ for $\sigma\in S_{N}$, the permutation group acting on $N$ elements. We use the default parameters from the PFN code, with $L=128$, $\Phi$ having two hidden layers with 100 nodes each, and $F$ having three hidden nodes with 100 nodes each. The same learning strategy (up to 1000 epochs, early stopping, 10% batch size) as the other networks is used for the PFN. The performance of $f^{\text{set}}_{3}$ is shown in Fig. 2, which gets much closer to matching the performance of $f_{1\rightarrow 3}$. Part of this improvement is due to enforcing the permutation symmetry, though there is also a potential gain from the fact the PFN we used for $f^{\text{set}}_{3}$ has more trainable weights than the fully connected network for $f^{\text{sort}}_{3}$. All of the $f_{3}$ variants were considerably more difficult to train than $f_{1\to 3}$, likely for the reason discussed in Sec. II.3. Thus, we have empirical evidence for the superiority of single-event training for multi-event classification. #### III.1.3 Top Quark Mass Measurement Our third and final example is motivated by the top quark mass measurement, as recently studied in Refs. [18, 12]. Extracting the top quark mass is really a regression problem, which we investigate in Sec. III.3. Here, we consider a related classification task to distinguish two event samples generated with different top quark masses (172.5 GeV and 175 GeV). This is a realistic hypothesis testing task that requires full event ensemble information, though only per-instance training as we will see. We use the same dataset as Ref. [18]. Top quark pair production is generated using Pythia 8.230 [52, 53] and detector effects are modeled with Delphes 3.4.1 [54, 55, 56] using the default CMS run card. After the production and decay steps $t\bar{t}\to bW^{+}\bar{b}W^{-}$, one of the $W$ bosons is forced to decay to $\mu^{+}\nu$ while the other $W$ boson decays hadronically. Each event is recorded as a variable-length set of objects, consisting of jets, muons, and neutrinos. At simulation-level, the neutrino is replaced with the missing transverse momentum. Generator-level and simulation-level jets are clustered with the anti-$k_{t}$ algorithm using $R=0.4$ and the simulation- level jet is labeled as $b$-tagged if the highest energy parton inside the nearest generator-level jet ($\Delta R<0.5$) is a $b$ quark. Jets are required to have $p_{T}>20$ GeV and they can only be $b$-tagged if $|\eta|<2.5$. Furthermore, jets overlapping with the muon are removed. Events are only saved if they have at least two $b$-tagged jets and at least two additional non $b$-tagged jets. The $b$-jet closest to the muon in rapidity-azimuth is labeled $b_{1}$. Of the remaining $b$-tagged jets, the highest $p_{T}$ one is labeled $b_{2}$. The two highest $p_{T}$ non-$b$-tagged jets are labeled $j_{1}$ and $j_{2}$, and typically come from the $W$ boson. (Imposing the $W$ mass constraint on $j_{1}$ and $j_{2}$ would yield lower efficiency, though without significantly impacting the results.) The four- momentum of the detector-level neutrino ($\nu$) is determined by solving the quadratic equation for the $W$ boson mass; if there is no solution, the mass is set to zero, while if there are two real solutions, the one with the smaller $|p_{z}|$ is selected. Four observables are formed for performing the top quark mass extraction, given by the following invariant masses: $m_{b_{1}\mu\nu}$, $m_{b_{2}\mu\nu}$, $m_{b_{1}j_{1}j_{2}}$, and $m_{b_{2}j_{1}j_{2}}$. A histogram of $m_{b_{1}\mu\nu}$ is shown for illustration in Fig. 3a. We use the same neural network architectures and training procedure as in the BSM example above, with 1.5 million events per fixed-mass sample. The only difference is that the batch size is set to 0.1% in order to keep the number of examples to be $\mathcal{O}(1000)$. For the per-ensemble classifier, we take $N=20$, though of course for a realistic hypothesis testing situation, $N$ would be as large as the number of top quark events recorded in data. To capture the permutation invariance of the inputs, we construct $f_{20}^{\text{set}}$ using the deep sets approach in Eq. (31). We also build a classifier $f_{1\to 20}$ from the per-instance classifier $f_{1}$ using Eq. (13). In Fig. 3b, we see that $f_{1\to 20}$ and $f_{20}^{\text{set}}$ have comparable performance, though $f_{1\to 20}$ is noticeably better. Some of this improvement may be due to differences in the network architecture, but we suspect that most of the gain is due to the more efficient training in the per-instance case. We checked that very poor performance is obtained for a classifier $f_{20}$ lacking permutation invariance, with a ROC curve that was not that much better than $f_{1}$ alone. Explicitly breaking the invariance by sorting the inputs based on $m_{b_{1}\mu\nu}$ does help a little, as indicated by the $f_{20}^{\text{sort}}$ curve in Fig. 3b, but does not reach the set- based approach. Figure 4: Computational performance of single-event versus multi-event training. Shown is the efficiency for the 175 GeV sample (false positive) for a fixed 50% efficiency for the 172.5 GeV sample (true positive), plotted as a function of training epoch. Single-event training ($f_{1\to 20}$) outperforms multi-event training ($f_{20}^{\text{set}}$), where both methods go through the full data set per epoch. Given the similar performance of $f_{1\rightarrow 20}$ and $f_{20}^{\text{set}}$, it is interesting to examine which learning strategy is more computationally efficient. In Fig. 4, we compare the performance as a function of the training epoch, using the difference of the true and false positive rates at a fixed 50% signal efficiency. In each epoch, both $f_{1\rightarrow 20}$ and $f_{20}^{\text{set}}$ see the full ensemble of events, so this is an apples-to-apples comparison as far as data usage is concerned. In particular, we plot this information per epoch instead of per compute time to avoid differences due to the structure of the neural networks. (There is not an easy way to control for possible differences in the training time due to the differences in the network structures, since the underlying tasks are different.) The $f_{1\rightarrow 20}$ classifier trains much faster, in agreement with the analysis in Sec. II.3, even though the ultimate asymptotic performance is similar for both classifiers. Once again, we see better empirical behavior from $f_{1\to 20}$ trained on one event at a time version $f_{20}^{\text{set}}$ trained on multiple events simultaneously.555Away from the asymptotic limit, one could try to improve the empirical per-ensemble performance through data augmentation. Data augmentation is a generic strategy to help neural networks learn symmetries, and the IID structure can be reinforced by showing the network new ensembles built from sampling instances from the existing ensembles. ### III.2 Classifiers: Single-Event from Multi-Event In general, one cannot take a multi-event classifier $f_{N}$ and extract a single-event classifier $f_{1}$. It is, however, possible to construct a special $\tilde{f}_{N}$ network such that one can interpret a subnetwork as a per-event classifier, as discussed in Sec. II.2. When using the MLC loss function, we can use the functional form in Eq. (14), where $\tilde{f}_{N}$ is a product of $f_{N\to 1}$ terms. Training $\tilde{f}_{N}$, where the only trainable weights are contained in $f_{N\to 1}$, we can learn a single-event classifier $f_{N\to 1}$ from multi-event samples. For the binary cross entropy loss used in our case studies, where Eq. (4) is needed to convert the classifier to a likelihood ratio, we have to introduce a slightly different structure than Eq. (14). Let $f_{N}^{\text{set}}$ be a permutation-invariant classifier, as defined in Eq. (31) using the deep sets/PFN strategy. Taking the latent space dimension to be $L=1$, the $\Phi$ network can be interpreted as a single-event classifier. Because the $\Phi$ network outputs are pooled via summation, we can build an optimal multi-event classifier if $\Phi$ learns the _logarithm_ of the likelihood ratio; cf. Eq. (2). With this insight, we can fix the $F$ function to achieve the same asymptotic performance as a trainable $F$ by setting: $\displaystyle F(\vec{x})=\frac{\exp\big{(}\sum_{i=1}^{N}\Phi(x_{i})\big{)}}{1+\exp\big{(}\sum_{i=1}^{N}\Phi(x_{i})\big{)}}\,.$ (32) Using Eq. (4), one can check that this $F$ is monotonically related to the ensemble likelihood ratio. Similarly, $\Phi$ will be monotonically related to the optimal $f_{1}$, which we call $f_{N\to 1}$ for the remainder of this discussion. Figure 5: Revisiting the ROC curves for the two Gaussian example from Fig. 1b. The multi-event classifier $\tilde{f}_{10}$ with the restricted functional form in Eq. (32) has the same performance as $f_{10}$ with no restrictions. Using $\tilde{f}_{10}$, we can construct a single-event classifier $\tilde{f}_{10\to 1}$ with the same performance as $f_{1}$ trained directly. This construction is demonstrated in Fig. 5 for the Gaussian example. We see that the deep sets architecture with the fixed form of Eq. (32) ($\tilde{f}^{\rm set}_{10}$) has the same or better performance as the 10-instance fully-connected classifier with more network capacity ($f_{10}$). Similarly, the $\Phi$ function used as a single-event classifier ($f_{10\rightarrow 1}$) has nearly the same performance as an independently trained single-event classifier ($f_{1}$). Figure 6: Revisiting the ROC curves for the dijet resonance search example in Fig. 2c. The set-based multi-event classifiers $\tilde{f}_{3}^{\rm set}$ and $f_{3}^{\rm set}$ have similar performance, but we can use the former to construct a single-event classifier $f_{3\to 1}$. This construction is not as effective as performing single-event training directly ($f_{1}$). The same conclusion holds for the BSM classification task, shown in Fig. 6. The only difference between the set-based architectures $\tilde{f}_{3}^{\text{set}}$ and $f_{3}^{\text{set}}$ is that the former uses the fixed functional form in Eq. (32). The fact that they achieve nearly the same performance is ensured by the IID relation in Eq. (2). The per-instance $f_{3\rightarrow 1}$ network extracted from $\tilde{f}_{3}^{\text{set}}$ is not quite as powerful as the $f_{1}$ network trained independently on single events, as expected from the gradient issue discussed in Sec. II.3. While we found no benefit to extracting a single-event classifier from a multi-event classifier, it is satisfying to see these IID-derived theoretical predictions borne out in these empirical examples. ### III.3 Comparison of Regression Strategies We now consider the regression methods introduced in Sec. II.4. For classification, the mapping between per-instance and per-ensemble information is relatively straightforward. For regression, though, per-ensemble regression is structurally dissimilar from per-instance regression because of the need to integrate over priors on the regression parameters. Nevertheless, we can perform per-ensemble regression by first mapping the problem to per-instance parametrized classification. We compare three different regression strategies for our empirical studies. The first method is a maximum-likelihood analysis, using the form in Eq. (24) based on the single-event parametrized classifier in Eq. (23). The second method is per-instance direct regression, using the construction in Eqs. (28) and (29) based on the same classifier as above. The third method is per- ensemble direct regression, based on minimizing the mean squared error loss in Eq. (27). #### III.3.1 Gaussian Mean Example Our first regression study is based on the same one-dimensional Gaussian distributions as Sec. III.1.1. The prior distribution for the Gaussian means is taken to be uniform with $\mu\in[-0.5,0.5]$, while the variance is fixed at $\sigma=1$. A training dataset is created from 100 examples each from 10,000 values of the Gaussian mean, for a total of one million training data points. For the reference sample $p(x|\theta_{0})$ needed to build the single-event parametrized classifier $f(x,\mu)$ in Eq. (23), we create a second dataset with one million examples drawn from a standard normal distribution (i.e. $\mu=0$). To implement the $p(\theta)$ term in the second line of Eq. (22), each example $x_{i}$ from the reference dataset is assigned a random mean value picked from the variable-mean dataset. We train a parametrized neural network to distinguish the variable-mean datasets from the reference dataset. This network takes as input two features: one component of $\vec{x}$ and the random mean value $\mu$. The architecture consists of three hidden layers with $(64,128,64)$ nodes per layer and ReLU activation. The output layer has a single node and sigmoid activation. Binary cross entropy is used to train the classifier and Eq. (4) is used to convert it to the likelihood ratio form $f(x,\mu)$. The model is trained for 1000 epochs with early stopping and a batch size of 10% of the training statistics. The same learned function $f(x,\mu)$ is used for both the maximum likelihood analysis and per-instance direct regression. For the maximum-likelihood analysis, the optimization in Eq. (24) is performed over a fixed grid with 20 evenly spaced values in $\mu\in[-0.5,0.5]$. For per-instance direct regression, the function $f_{N}(\vec{x},\mu)$ in Eq. (29) is constructed by taking a product of $f(x,\mu)$ outputs over all 100 examples in a given ensemble data point $\vec{x}$. The integrals in Eqs. (28) and (29) are approximated by evaluating $f_{N}(\vec{x},\mu)$ at 20 evenly spaced $\mu$ values between $-0.5$ and $0.5$ and then adding their values; this is possible because the prior is uniform. The per-ensemble direct regression approach uses a neural network $g_{N}$ that takes as input 100 values (i.e. all of $\vec{x}$) and predicts a single mean value. This network has the same architecture as $f(x,\mu)$, except it directly takes as input $\vec{x}$ and has linear (instead of a sigmoid) activation for the output layer, since the predicted mean can be both positive or negative. It is trained to minimize the mean squared error loss in Eq. (27). Figure 7: Comparison of regression methods with the Gaussian example, with the predicted value of the mean plotted against the true value of the mean. The regression involves analyzing 100 instances drawn from the same Gaussian distribution. Bands are the standard deviation of the predictions over 10,000 generated samples. The per-instance direct regression uses single-event training, yet achieves comparable performance to per-ensemble direct regression that processes 100 events simultaneously. In Fig. 7, we see that all three approaches give nearly the same results in terms of bias and variance. Strictly speaking, maximum likelihood and direct regression are different tasks so their behavior could be different. For per- instance and per-ensemble direct regression, they are constructed to yield the same asymptotic behavior, but there will be differences due to, e.g., the finite approximations to the integrals. Note that maximum likelihood and per- instance direct regression only use neural networks that process per-instance inputs; information about the rest of the events is used only through the training procedure. Thus, we have empirical evidence that per-ensemble regression can be accomplished via per-instance training. #### III.3.2 Top Quark Mass Measurement As a physics example of regression, we consider extracting the top quark mass. Here, the top quark mass is the regression target and the setup is similar to the Gaussian example above. We use the same event generation as Sec. III.1.3, but now with top quark mass parameters sampled uniformly at random in $m_{t}\in[170,180]~{}\text{GeV}$. As with the Gaussian example, a variable- mass dataset is created. In this case, we have 100 events for each of 100,000 sampled top quark mass values. The reference sample uses a top quark mass of 172.5 GeV. Due to event selection effects, the actual number of events for each top quark mass value varies from set-to-set, with a mean of about 40 events. Because this event selection has a slight top quark mass dependence, this yields an effective non-uniform prior on $m_{t}$, which we account for when assigning dummy mass values to the reference sample. The parametrized classifier now takes five inputs: the four mass features from Sec. III.1.3 ($m_{b_{1}\mu\nu}$, $m_{b_{2}\mu\nu}$, $m_{b_{1}j_{1}j_{2}}$, and $m_{b_{2}j_{1}j_{2}}$) plus the top quark mass used for event generation. The neural network has three hidden layers with 50 nodes per layer and ReLU activation, and a single node output layer with sigmoid activation. We train 100 models and take the median as the classifier output, using Eq. (4) to convert it to the likelihood ratio $f(x,m_{t})$. Each model is trained for 1000 epochs with early stopping with a patience of 20 epochs and a batch size of 0.1%. To test the fidelity of the training, we extract the estimated likelihood ratio of $m_{t}=175~{}\text{GeV}$ over $m_{t}=172.5~{}\text{GeV}$ and use it to reweight the $172.5~{}\text{GeV}$ sample. From Fig. 3a, we see that we achieve good reweighting performance despite the relatively limited training data. (a) (b) Figure 8: Regression in the top quark mass example. (a) An estimate of the log likelihood for samples generated with 172.5 and 175 GeV top quark masses. The vertical axis has been shifted such that the minimum value is at zero. Note that the axis represents the average log likelihood which is a factor of $N_{\text{events}}$ different from the total log likelihood. (b) Correlation between the per-instance predicted mass and the per-ensemble predicted mass in the context of direct regression. The per-ensemble mass values are put in bins of 0.1 GeV width, and the bands represent the standard deviation of the per- instance mass values in each bin. The maximum likelihood analysis is performed by scanning the learned log likelihood estimate over a fixed grid with 100 uniformly spaced steps in $m_{t}\in[170,180]~{}\text{GeV}$. In Fig. 8a, we show this scan where the target data comes from the high statistics 172.5 GeV and 175 GeV samples from Sec. III.1.3. As desired, the minimum of the parabolic shapes are near the input top quark masses. For the per-instance direct regression, we follow the same strategy as in the Gaussian case to convert $f(x,m_{t})$ into an estimate of $\mathbb{E}[m_{t}|\vec{x}]$. The integrals in Eqs. (28) and (29) are approximated by sampling 50 random top quark masses per set of 100 following the probability density from the training dataset. Because 40 events are insufficient to make a precision measurement of the top quark mass, we find a noticeable bias between the estimated and true top mass values, which is exacerbated by edge effects at the ends of the training range. For this reason, we do not show a direct analog to Fig. 7, though this bias could be overcome with much larger training datasets with many more than 100 examples per mass value. For the per-ensemble direct regression, we use the deep sets approach in Eq. (31) to handle the permutation-invariance of the inputs. This approach is also well suited to handle the large variation in the number of events in each set due to the event selection effect. We again use PFNs for our practical implementation. We use the default PFN hyperparameters from the https://energyflow.network/ package, except we use linear activation in the output layer and the mean squared error loss function. We found that it was important for the model accuracy to standardize both the inputs and outputs of the network. Note that this is a different per-ensemble direct regression setup than used in Ref. [12], which found excellent performance using linear regression on sorted inputs. In Fig. 8b, we compare the output of per-ensemble direct regression to the output of per-instance direct regression. We find a very strong correlation between these two very different approaches to computing the same quantity $\mathbb{E}[m_{t}|\vec{x}]$. The band in Fig. 8b is the standard deviation over data sets with a true mass in the same one of the 100 bins that are evenly spaced between 170 and 180 GeV. A key advantage of the per-instance approach is that it does not need to be retrained if more events are acquired. By contrast, the per-ensemble approach is only valid for event samples that have the same sizes as were used during training. ### III.4 Beyond Regression Example As remarked in Sec. II.5, the ideas discussed above apply to learning tasks beyond just standard classification and regression. As one simple example to illustrate this, we consider the Gaussian classification task from Sec. III.1.1 and compute the mutual information between the Gaussian feature and the label. This quantifies how much information is available in the feature for classification and can be directly compared with other features and other classification tasks. Figure 9: Mutual information between a Gaussian feature and a label, where the “ signal” ($x_{0}=\epsilon$) and “ background” ($x_{0}=-\epsilon$) have opposite means. The estimate using the MLC loss approach shows good agreement with the exact analytic expression. For this illustration, $10^{5}$ events are generated each from two Gaussian distributions with means $\pm|\epsilon|$ for fixed $\epsilon$. The mutual information is estimated using a per-instance classifier as described in Sec. II.5 and also computed analytically via Eq. (30). For the per-instance classifier, we use a neural network that processes two inputs (label and feature), has two hidden layers with ReLU activation, and has a single node sigmoid output. The classification task is to distinguish the nominal dataset from one where the labels are assigned uniformly at random to the features. The value of the MLC loss yields an estimate of the mutual information. The mutual information results are presented in Fig. 9, as a function of $\epsilon$. As expected, the neural network strategy yields an excellent approximation to the analytic calculation. Note that this strategy does require any binning and naturally extends to high-dimensional data, since the core component is a neural network classifier. We leave an investigation of this approach in the particle physics context to future work. ## IV Conclusions We have demonstrated a connection between classifiers trained on single events and those that process multiple events at the same time. One can take a generic single-event classifier and build an $N$-event classifier using simple arithmetic operations. Such classifiers tend to out-perform generic $N$-event classifiers, since we can enforce the IID assumptions into the learning task. This performance gap can be mostly recovered by deploying a classifier that respects the permutation invariance of the set of $N$ events. We used the deep sets/PFN architecture [62, 63] for this purpose, but other set-based architectures such as graph neural networks [64, 65] would also be appropriate. An amusing feature of the deep sets approach is that we can use it to reverse- engineer a single-event classifier from a multi-event classifier by restricting the latent space to be one-dimensional and fixing a static output function. Even after enforcing these additional structures, though, we found both theoretically and empirically that the loss function gradients are better behaved for single-event classifiers than multi-event classifiers. Going beyond classification, we explained how various regression tasks can be phrased in terms of per-instance parametrized classification, yielding similar performance to per-ensemble direct regression. We also mentioned how to compute distances and divergences between probability densities without requiring explicit density estimation. These results hold for any data sample satisfying the IID property. Ultimately, we did not find any formal or practical advantage for training a multi-event classifier instead of a single-event classifier, as least for the cases we studied. With a carefully selected multi-event architecture, one can achieve similar performance to a scaled-up per-event classifier, but the latter will typically train faster. For direct regression, the per-ensemble strategy might be conceptually simpler than the per-instance method, though the per-instance methods allow for a simpler treatment of variably-sized data sets. Note that there may be situations where a simplifying assumption (e.g. the linear regression model in Ref. [12]) could yield better per-ensemble behavior than indicated by our case studies. At minimum, we hope this paper has demystified aspects of per-ensemble learning and highlighted some interesting features of the MLC loss function. Going beyond the IID assumption, the duality between per-instance classifiers and per-ensemble classifiers could have applications to problems with approximate independence. For example, flavor tagging algorithms have traditionally exploited the approximate independence of individual track features within a jet [66, 67]. Similarly, emissions in the Lund jet plane [68, 69] are approximately independent, with exact independence in the strongly ordered limit of QCD. In both contexts, the instances are particles (or particle-like features) and the ensemble is the jet. A potentially powerful training procedure for these situations might be to first train a per-particle classifier, then build a per-jet classifier using the constructions described in this paper, and finally let the network train further to learn interdependencies between the particles. ## Code and Data The code for this paper can be found at https://github.com/bnachman/EnsembleLearning. The physics datasets are hosted on Zenodo at Ref. [70] for the top quark dataset and Ref. [71] for the BSM dataset. ###### Acknowledgements. We thank Anders Andreassen, Patrick Komiske, and Eric Metodiev for discussions about the MLC loss. We thank Rikab Gambhir and Ian Convy for discussions about mutual information. We thank Adi Suresh for discussions about the regression task with the classifier loss. We thank Katherine Fraiser, Yue Lai, Duff Neill, Bryan Ostdiek, Mateusz Ploskon, Felix Ringer, and Matthew Schwartz for useful comments on our manuscript. BN is supported by the U.S. Department of Energy (DOE), Office of Science under contract DE-AC02-05CH11231. JT is supported by the National Science Foundation under Cooperative Agreement PHY-2019786 (The NSF AI Institute for Artificial Intelligence and Fundamental Interactions, http://iaifi.org/), and by the U.S. DOE Office of High Energy Physics under grant number DE-SC0012567. BN would also like to thank NVIDIA for providing Volta GPUs for neural network training. ## Appendix A Deriving Maximum Likelihood Classifier Loss Beyond just the practical value of learning the likelihood ratio, the MLC loss in Eq. (7) has a nice interpretation in terms of learning probability distributions. Consider trying to learn a function $f(x)$ that is a normalized probability distribution, up to a Jacobian factor $j(x)$: $\int dx\,j(x)f(x)=1.$ (33) We are given samples from a probability distribution $q(x)$, and we want to learn $f(x)$ such that $f(x)\to\frac{q(x)}{j(x)}.$ (34) In other words, we want to learn a function $f(x)$ that reproduces the sampled distribution $q(x)$ after including the Jacobian factor. This problem was studied in Ref. [34], albeit in a context where $f(x)$ had a restricted functional form such that Eq. (33) was automatically enforced. Figure 10: A demonstration of the MLC loss for learning the likelihood ratio directly, using the Gaussian example from Fig. 1a. The linear (lin) and exponential (exp) parametrizations perform similarly. Shown for comparison is likelihood ratio computed using the binary cross entropy (BCE) loss that requires the manipulation in Eq. (4). One strategy to accomplish this is to minimize the cross entropy of $f(x)$ with respect to $q(x)$, since the smallest cross entropy is obtained when $f(x)$ has the same information content as $q(x)$. The associated loss functional is: $L[f]=-\int dx\,q(x)\log f(x)-\lambda\left(1-\int dx\,j(x)f(x)\right),$ (35) where the first term is the cross entropy and $\lambda$ is a Lagrange multiplier to enforce the normalization condition in Eq. (33). Taking the functional derivative of Eq. (35) with respect to $f(x)$ and setting it equal to zero, we find the extremum condition: $-\frac{q(x)}{f(x)}+\lambda\,j(x)=0.$ (36) Multiplying both sides of this equation by $f(x)$ and integrating over $x$ to set the Lagrange multiplier, we find that Eq. (36) is solved for $\lambda=1,\qquad f(x)=\frac{q(x)}{j(x)},$ (37) so $f(x)$ learns the $q(x)/j(x)$ ratio as desired. In the special case that $j(x)$ is itself a normalized probability distribution, we can substitute for the Lagrange multiplier and rewrite Eq. (35) in the following form: $L[f]=-\int dx\,\Big{(}q(x)\log f(x)+j(x)(1-f(x))\Big{)}.$ (38) Identifying $q(x)=p(x|\theta_{A})$ and $j(x)=p(x|\theta_{B})$, this is precisely the MLC loss in Eq. (7). Therefore, we have an intuitive understanding of the MLC loss as trying to maximize the (log) likelihood of $f(x)$ with respect to $p(x|\theta_{A})$, subject to the constraint that $f(x)\,p(x|\theta_{B})$ is a proper probability distribution. In Fig. 10, we plot the learned likelihood ratio between the two Gaussian samples from Fig. 1a, comparing the performance of MLC against binary cross entropy and the exact analytic expression. In all cases, a network is trained with 100 epochs and early stopping with a patience of 10 epochs. We also compare the MLC loss against the $C(f)=\exp f$ variant discussed in footnote 1. We see that both the linear (i.e. $C(f)=f$) and exponential parametrizations perform similarly in the region with ample data. That said, the exponential parametrization has a more robust extrapolation towards the edges, yielding similar behavior to binary cross entropy. Note that the exponential parametrization of the MLC loss was used in Ref. [32]. ## References * Larkoski _et al._ [2020] A. J. Larkoski, I. Moult, and B. Nachman, Jet Substructure at the Large Hadron Collider: A Review of Recent Advances in Theory and Machine Learning, Phys. Rept. 841, 1 (2020), arXiv:1709.04464 [hep-ph] . * Guest _et al._ [2018] D. Guest, K. Cranmer, and D. Whiteson, Deep Learning and its Application to LHC Physics, Ann. Rev. Nucl. Part. Sci. 68, 161 (2018), arXiv:1806.11484 [hep-ex] . * Albertsson _et al._ [2018] K. Albertsson _et al._ , Machine Learning in High Energy Physics Community White Paper, (2018), arXiv:1807.02876 [physics.comp-ph] . * Radovic _et al._ [2018] A. Radovic _et al._ , Machine learning at the energy and intensity frontiers of particle physics, Nature 560, 41 (2018). * Bourilkov [2020] D. Bourilkov, Machine and Deep Learning Applications in Particle Physics, Int. J. Mod. Phys. A 34, 1930019 (2020), arXiv:1912.08245 [physics.data-an] . * [6] HEP ML Community, A Living Review of Machine Learning for Particle Physics. * Lai [2018] Y. S. Lai, Automated Discovery of Jet Substructure Analyses, (2018), arXiv:1810.00835 [nucl-th] . * Khosa _et al._ [2019] C. K. Khosa, V. Sanz, and M. Soughton, Using Machine Learning to disentangle LHC signatures of Dark Matter candidates, (2019), arXiv:1910.06058 [hep-ph] . * Du _et al._ [2020] Y.-L. Du, K. Zhou, J. Steinheimer, L.-G. Pang, A. Motornenko, H.-S. Zong, X.-N. Wang, and H. Stöcker, Identifying the nature of the QCD transition in relativistic collision of heavy nuclei with deep learning, Eur. Phys. J. C 80, 516 (2020), arXiv:1910.11530 [hep-ph] . * Mullin _et al._ [2019] A. Mullin, H. Pacey, M. Parker, M. White, and S. Williams, Does SUSY have friends? A new approach for LHC event analysis, (2019), arXiv:1912.10625 [hep-ph] . * Chang _et al._ [2020] S. Chang, T.-K. Chen, and C.-W. Chiang, Distinguishing $W^{\prime}$ Signals at Hadron Colliders Using Neural Networks, (2020), arXiv:2007.14586 [hep-ph] . * Flesher _et al._ [2020] F. Flesher, K. Fraser, C. Hutchison, B. Ostdiek, and M. D. Schwartz, Parameter Inference from Event Ensembles and the Top-Quark Mass, (2020), arXiv:2011.04666 [hep-ph] . * Lazzarin _et al._ [2020] M. Lazzarin, S. Alioli, and S. Carrazza, MCNNTUNES: tuning Shower Monte Carlo generators with machine learning, (2020), arXiv:2010.02213 [physics.comp-ph] . * Lai _et al._ [2020] Y. S. Lai, D. Neill, M. Płoskoń, and F. Ringer, Explainable machine learning of the underlying physics of high-energy particle collisions, (2020), arXiv:2012.06582 [hep-ph] . * Neyman and Pearson [1933] J. Neyman and E. S. Pearson, On the problem of the most efficient tests of statistical hypotheses, Phil. Trans. R. Soc. Lond. A 231, 289 (1933). * Hastie _et al._ [2001] T. Hastie, R. Tibshirani, and J. Friedman, _The Elements of Statistical Learning_ , Springer Series in Statistics (Springer New York Inc., New York, NY, USA, 2001). * Sugiyama _et al._ [2012] M. Sugiyama, T. Suzuki, and T. Kanamori, _Density Ratio Estimation in Machine Learning_ (Cambridge University Press, 2012). * A. Andreassen, S. Hsu, B. Nachman, N. Suaysom, A. Suresh [2020] A. Andreassen, S. Hsu, B. Nachman, N. Suaysom, A. Suresh, Parameter Estimation using Neural Networks in the Presence of Detector Effects, (2020), arXiv:2010.03569 [hep-ph] . * Andreassen and Nachman [2020] A. Andreassen and B. Nachman, Neural Networks for Full Phase-space Reweighting and Parameter Tuning, Phys. Rev. D 101, 091901(R) (2020), arXiv:1907.08209 [hep-ph] . * Stoye _et al._ [2018] M. Stoye, J. Brehmer, G. Louppe, J. Pavez, and K. Cranmer, Likelihood-free inference with an improved cross-entropy estimator, (2018), arXiv:1808.00973 [stat.ML] . * Hollingsworth and Whiteson [2020] J. Hollingsworth and D. Whiteson, Resonance Searches with Machine Learned Likelihood Ratios, (2020), arXiv:2002.04699 [hep-ph] . * Brehmer _et al._ [2018a] J. Brehmer, K. Cranmer, G. Louppe, and J. Pavez, Constraining Effective Field Theories with Machine Learning, Phys. Rev. Lett. 121, 111801 (2018a), arXiv:1805.00013 [hep-ph] . * Brehmer _et al._ [2018b] J. Brehmer, K. Cranmer, G. Louppe, and J. Pavez, A Guide to Constraining Effective Field Theories with Machine Learning, Phys. Rev. D 98, 052004 (2018b), arXiv:1805.00020 [hep-ph] . * Brehmer _et al._ [2020a] J. Brehmer, F. Kling, I. Espejo, and K. Cranmer, MadMiner: Machine learning-based inference for particle physics, Comput. Softw. Big Sci. 4, 3 (2020a), arXiv:1907.10621 [hep-ph] . * Brehmer _et al._ [2020b] J. Brehmer, G. Louppe, J. Pavez, and K. Cranmer, Mining gold from implicit models to improve likelihood-free inference, Proc. Nat. Acad. Sci. , 201915980 (2020b), arXiv:1805.12244 [stat.ML] . * Cranmer _et al._ [2015] K. Cranmer, J. Pavez, and G. Louppe, Approximating Likelihood Ratios with Calibrated Discriminative Classifiers, (2015), arXiv:1506.02169 [stat.AP] . * Badiali _et al._ [2020] C. Badiali, F. Di Bello, G. Frattari, E. Gross, V. Ippolito, M. Kado, and J. Shlomi, Efficiency Parameterization with Neural Networks, (2020), arXiv:2004.02665 [hep-ex] . * Andreassen _et al._ [2020a] A. Andreassen, B. Nachman, and D. Shih, Simulation Assisted Likelihood-free Anomaly Detection, Phys. Rev. D 101, 095004 (2020a), arXiv:2001.05001 [hep-ph] . * Andreassen _et al._ [2020b] A. Andreassen, P. T. Komiske, E. M. Metodiev, B. Nachman, and J. Thaler, OmniFold: A Method to Simultaneously Unfold All Observables, Phys. Rev. Lett. 124, 182001 (2020b), arXiv:1911.09107 [hep-ph] . * Erdmann _et al._ [2019] M. Erdmann, B. Fischer, D. Noll, Y. Rath, M. Rieger, and D. Schmidt, Adversarial Neural Network-based data-simulation corrections for jet-tagging at CMS, in _Proc. 19th Int. Workshop on Adv. Comp., Anal. Techn. in Phys. Research, ACAT2019_ (2019). * Nguyen _et al._ [2005] X. Nguyen, M. J. Wainwright, and M. I. Jordan, On surrogate loss functions and $f$-divergences, arXiv Mathematics e-prints , math/0510521 (2005), arXiv:math/0510521 [math.ST] . * D’Agnolo and Wulzer [2019] R. T. D’Agnolo and A. Wulzer, Learning New Physics from a Machine, Phys. Rev. D 99, 015014 (2019), arXiv:1806.02350 [hep-ph] . * D’Agnolo _et al._ [2019] R. T. D’Agnolo, G. Grosso, M. Pierini, A. Wulzer, and M. Zanetti, Learning Multivariate New Physics, (2019), arXiv:1912.12155 [hep-ph] . * Andreassen _et al._ [2019] A. Andreassen, I. Feige, C. Frye, and M. D. Schwartz, JUNIPR: a Framework for Unsupervised Machine Learning in Particle Physics, Eur. Phys. J. C 79, 102 (2019), arXiv:1804.09720 [hep-ph] . * Brehmer and Cranmer [2020] J. Brehmer and K. Cranmer, Flows for simultaneous manifold learning and density estimation (2020), arXiv:2003.13913 [stat.ML] . * Nachman and Shih [2020] B. Nachman and D. Shih, Anomaly Detection with Density Estimation, Phys. Rev. D 101, 075042 (2020), arXiv:2001.04990 [hep-ph] . * Metodiev _et al._ [2017] E. M. Metodiev, B. Nachman, and J. Thaler, Classification without labels: Learning from mixed samples in high energy physics, JHEP 10, 174, arXiv:1708.02949 [hep-ph] . * Baldi _et al._ [2016] P. Baldi, K. Cranmer, T. Faucett, P. Sadowski, and D. Whiteson, Parameterized neural networks for high-energy physics, Eur. Phys. J. C 76, 235 (2016), arXiv:1601.07913 [hep-ex] . * S. Cheong, A. Cukierman, B. Nachman, M. Safdari, A. Schwartzman [2020] S. Cheong, A. Cukierman, B. Nachman, M. Safdari, A. Schwartzman, Parametrizing the Detector Response with Neural Networks, JINST 15, P01030, arXiv:1910.03773 [physics.data-an] . * Fix and Hodges Jr. [1951] E. Fix and J. L. Hodges Jr., Discriminatory analysis-nonparametric discrimination: consistency properties, USAF School of Aviation Medicine, Project Number 21-49-004, Report Number 4 (1951). * Cover M. and Hart [1967] T. Cover M. and P. E. Hart, Nearest neighbor pattern classification, IEEE Transactions on Information Theory 13, 21 (1967). * Goodfellow _et al._ [2014] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio, Generative adversarial nets, in _Advances in Neural Information Processing Systems_, Vol. 27, edited by Z. Ghahramani, M. Welling, C. Cortes, N. Lawrence, and K. Q. Weinberger (Curran Associates, Inc., 2014) pp. 2672–2680. * Salimans _et al._ [2016] T. Salimans, I. Goodfellow, W. Zaremba, V. Cheung, A. Radford, and X. Chen, Improved Techniques for Training GANs, arXiv e-prints , arXiv:1606.03498 (2016), arXiv:1606.03498 [cs.LG] . * Kingma and Welling [2014] D. P. Kingma and M. Welling, Auto-encoding variational bayes., in _ICLR_, edited by Y. Bengio and Y. LeCun (2014). * Rezende and Mohamed [2015] D. Rezende and S. Mohamed, Variational inference with normalizing flows, in _Proceedings of the 32nd International Conference on Machine Learning_, Proceedings of Machine Learning Research, Vol. 37, edited by F. Bach and D. Blei (PMLR, Lille, France, 2015) pp. 1530–1538. * Larkoski _et al._ [2014] A. J. Larkoski, J. Thaler, and W. J. Waalewijn, Gaining (Mutual) Information about Quark/Gluon Discrimination, JHEP 11, 129, arXiv:1408.3122 [hep-ph] . * Carrara and Ernst [2019] N. Carrara and J. Ernst, On the estimation of mutual information (2019), arXiv:1910.00365 [physics.data-an] . * Chollet [2017] F. Chollet, Keras, https://github.com/fchollet/keras (2017). * Abadi _et al._ [2016] M. Abadi, P. Barham, J. Chen, Z. Chen, A. Davis, J. Dean, M. Devin, S. Ghemawat, G. Irving, M. Isard, _et al._ , Tensorflow: A system for large-scale machine learning., in _OSDI_ , Vol. 16 (2016) pp. 265–283. * Kingma and Ba [2014] D. Kingma and J. Ba, Adam: A method for stochastic optimization, (2014), arXiv:1412.6980 [cs] . * Kasieczka _et al._ [2019a] G. Kasieczka, B. Nachman, and D. Shih, R&D Dataset for LHC Olympics 2020 Anomaly Detection Challenge, 10.5281/zenodo.2629073 (2019a), https://doi.org/10.5281/zenodo.2629073. * Sjöstrand _et al._ [2006] T. Sjöstrand, S. Mrenna, and P. Z. Skands, PYTHIA 6.4 Physics and Manual, JHEP 05, 026, arXiv:hep-ph/0603175 [hep-ph] . * Sjöstrand _et al._ [2008] T. Sjöstrand, S. Mrenna, and P. Z. Skands, A Brief Introduction to PYTHIA 8.1, Comput. Phys. Commun. 178, 852 (2008), arXiv:0710.3820 [hep-ph] . * de Favereau _et al._ [2014] J. de Favereau, C. Delaere, P. Demin, A. Giammanco, V. Lemaître, A. Mertens, and M. Selvaggi (DELPHES 3), DELPHES 3, A modular framework for fast simulation of a generic collider experiment, JHEP 02, 057, arXiv:1307.6346 [hep-ex] . * Mertens [2015] A. Mertens, New features in Delphes 3, _Proceedings, 16th International workshop on Advanced Computing and Analysis Techniques in physics (ACAT 14): Prague, Czech Republic, September 1-5, 2014_ , J. Phys. Conf. Ser. 608, 012045 (2015). * Selvaggi [2014] M. Selvaggi, DELPHES 3: A modular framework for fast-simulation of generic collider experiments, _Proceedings, 15th International Workshop on Advanced Computing and Analysis Techniques in Physics Research (ACAT 2013): Beijing, China, May 16-21, 2013_ , J. Phys. Conf. Ser. 523, 012033 (2014). * Cacciari _et al._ [2012] M. Cacciari, G. P. Salam, and G. Soyez, FastJet User Manual, Eur. Phys. J. C72, 1896 (2012), arXiv:1111.6097 [hep-ph] . * Cacciari and Salam [2006] M. Cacciari and G. P. Salam, Dispelling the $N^{3}$ myth for the $k_{t}$ jet-finder, Phys. Lett. B641, 57 (2006), arXiv:hep-ph/0512210 [hep-ph] . * Cacciari _et al._ [2008] M. Cacciari, G. P. Salam, and G. Soyez, The anti-$k_{t}$ jet clustering algorithm, JHEP 04, 063, arXiv:0802.1189 [hep-ph] . * Thaler and Van Tilburg [2012] J. Thaler and K. Van Tilburg, Maximizing Boosted Top Identification by Minimizing N-subjettiness, JHEP 02, 093, arXiv:1108.2701 [hep-ph] . * Thaler and Van Tilburg [2011] J. Thaler and K. Van Tilburg, Identifying Boosted Objects with N-subjettiness, JHEP 03, 015, arXiv:1011.2268 [hep-ph] . * Zaheer _et al._ [2017] M. Zaheer, S. Kottur, S. Ravanbhakhsh, B. Póczos, R. Salakhutdinov, and A. J. Smola, Deep sets, in _Proceedings of the 31st International Conference on Neural Information Processing Systems_ , NIPS’17 (Curran Associates Inc., Red Hook, NY, USA, 2017) p. 3394–3404. * Komiske _et al._ [2019] P. T. Komiske, E. M. Metodiev, and J. Thaler, Energy Flow Networks: Deep Sets for Particle Jets, JHEP 01, 121, arXiv:1810.05165 [hep-ph] . * Scarselli _et al._ [2009] F. Scarselli, M. Gori, A. C. Tsoi, M. Hagenbuchner, and G. Monfardini, The graph neural network model, Trans. Neur. Netw. 20, 61–80 (2009). * Shlomi _et al._ [2020] J. Shlomi, P. Battaglia, and J.-R. Vlimant, Graph Neural Networks in Particle Physics 10.1088/2632-2153/abbf9a (2020), arXiv:2007.13681 [hep-ex] . * Aaboud _et al._ [2018] M. Aaboud _et al._ (ATLAS), Measurements of b-jet tagging efficiency with the ATLAS detector using $t\overline{t}$ events at $\sqrt{s}=13$ TeV, JHEP 08, 089, arXiv:1805.01845 [hep-ex] . * Chatrchyan _et al._ [2013] S. Chatrchyan _et al._ (CMS), Identification of b-Quark Jets with the CMS Experiment, JINST 8, P04013, arXiv:1211.4462 [hep-ex] . * Andersson _et al._ [1989] B. Andersson, G. Gustafson, L. Lonnblad, and U. Pettersson, Coherence Effects in Deep Inelastic Scattering, Z. Phys. C 43, 625 (1989). * Dreyer _et al._ [2018] F. A. Dreyer, G. P. Salam, and G. Soyez, The Lund Jet Plane, JHEP 12, 064, arXiv:1807.04758 [hep-ph] . * Andreassen _et al._ [2020c] A. Andreassen, S.-C. Hsu, B. Nachman, N. Suaysom, and A. Suresh, Srgn: Pythia + delphes $pp\rightarrow t\bar{t}$, 10.5281/zenodo.4067673 (2020c). * Kasieczka _et al._ [2019b] G. Kasieczka, B. Nachman, and D. Shih, Official Datasets for LHC Olympics 2020 Anomaly Detection Challenge, 10.5281/zenodo.4287846 (2019b).
Vol.0 (200x) No.0, 000–000 11institutetext: Department of Astronomy, Beijing Normal University, Beijing 100875, China; 22institutetext: Key Laboratory of Optical Astronomy, National Astronomical Observatories, Chinese Academy of Sciences, Beijing 100101, China; 33institutetext: School of Astronomy and Space Science, University of Chinese Academy of Sciences, Beijing 100049, China; 44institutetext: School of Astronomy and Space Science, Nanjing University, Nanjing 210023, Jiangsu, China # Detecting and Monitoring Tidal Dissipation of Hot Jupiters in the Era of SiTian Fan Yang 112233 Wei<EMAIL_ADDRESS>2233 Xing<EMAIL_ADDRESS>11 Hui<EMAIL_ADDRESS>44 Ji-Lin Zhou 44 Su-Su Shan 2233 Jie Zheng 22 Wei-Kai Zong 11 Ming Yang 44 Yu Bai 22 Song Wang 22 Jia-Chen Zheng 11 Yu-Ru Xu 11 Yu-Feng Li 11 You-Jun Lu 2233 Ji-Feng Liu 2233 ###### Abstract Transit Timing Variation (TTV) of hot Jupiters provides direct observational evidence of planet tidal dissipation. Detecting tidal dissipation through TTV needs high precision transit timings and long timing baselines. In this work, we predict and discuss the potential scientific contribution of SiTian Survey in detecting and analyzing exoplanet TTV. We develop a tidal dissipation detection pipeline for SiTian Survey that aims at time-domain astronomy with 72 1-meter optical telescopes. The pipeline includes the modules of light curve deblending, transit timing obtaining, and TTV modeling. SiTian is capable to detect more than 25,000 exoplanets among which we expect $\sim$50 sources showing evidence of tidal dissipation. We present detection and analysis of tidal dissipating targets, based on simulated SiTian light curves of XO-3b and WASP-161b. The transit light curve modeling gives consistent results within 1$\sigma$ to input values of simulated light curves. Also, the parameter uncertainties predicted by Monte-Carlo Markov Chain are consistent with the distribution obtained from simulating and modeling the light curve 1000 times. The timing precision of SiTian observations is $\sim$ 0.5 minutes with one transit visit. We show that differences between TTV origins, e.g., tidal dissipation, apsidal precession, multiple planets, would be significant, considering the timing precision and baseline. The detection rate of tidal dissipating hot Jupiters would answer a crucial question of whether the planet migrates at an early formation stage or random stages due to perturbations, e.g., planet scattering, secular interaction. SiTian identified targets would be constructive given that the sample would extend tenfold. ###### keywords: planets and satellites: gaseous planets, planets and satellites: physical evolution, planets and satellites: dynamical evolution and stability, planets and satellites: detection ## 1 Introduction Tidal migration is one of mechanisms that may explain why hot Jupiters occur at such close orbits, though direct observational evidence of orbit decay is only achieved for a single planet, namely WASP-12b, in the past decades (Patra et al., 2017; Dawson & Johnson, 2018; Turner et al., 2021). The first data release of Transiting Exoplanet Survey Satellite (TESS; Ricker et al., 2015) in 2018, leads to the identification of four new candidates through TTV detection (Dong et al., 2021; Davoudi et al., 2021; Shan et al., 2021; Yang & Chary, 2021, submitted; Yang & Wei, 2021). TTV monitoring depends on the timing precision and baseline length which are key technical specifications for time-domain facilities, e.g., Zwicky Transient Facility (ZTF; Bellm et al., 2019), Large Synoptic Survey Telescope (LSST; Ivezić et al., 2019), Tsinghua University-Ma Huateng Telescopes for Survey (TMTS; Zhang et al., 2020; Lin et al., 2021), Antarctic Survey Telescopes times 3 (AST3; Ma et al., 2018; Zhang et al., 2019), and SiTian (Liu et al., 2021; Zhu et al., 2021). The latter is aiming to tune down the false-positive probability of time- domain signals, by implementing a global network of three-color photometric monitoring of one-quarter of the sky at a cadence of 30min, down to a detection limit of $V\sim 21$ mag (Liu et al., 2021). It may suggest a TTV signal if any constant period ephemeris can not fit well the observed timings. TTV could originate from multiple physical processes, e.g., tidal dissipation, apsidal precession, R$\o$mer effect, and mass loss (Ragozzine & Wolf, 2009; Valsecchi et al., 2015; Patra et al., 2017; Ou et al., 2021). Another TTV generation process is the interaction between planets in a multi-planet system which should cause oscillation in transit timing residual in an observable timescale (Holman et al., 2010; Weiss et al., 2013). However, it has been reported that a planet companion should not be close or massive enough to induce observable TTVs for hot Jupiters (Huang et al., 2016). Distinguishing among the above mentioned various physical origins requires long-term high precision monitoring of transit timings continuously. Combining observations from wide field transit surveys, e.g., Kepler, TESS (Borucki et al., 2010; Ricker et al., 2015) and follow-up observations using focused telescopes e.g., CHEOPS (Benz et al., 2021), has been proved to be feasible in TTV detecting and related researches (Borsato et al., 2021). Sariya et al. (2021) report TTVs of HAT-P-12b in the baseline of ten years, using light curves from small ground-based telescopes. The transit timings of WASP-32b are monitored and no significant TTV is found with the available data (Sun et al., 2015). WASP-43b is reported as a candidate showing period decay (Jiang et al., 2016) which arises wide attention and scientific discussions (Davoudi et al., 2021; Garai et al., 2021). WASP-4b presents a significant TTV and is furtherly explained by the R$\o$mer effect (Bouma et al., 2019, 2020). HAT-P-25 is reported with no significant TTV which sets a limit on the possible planet-planet interactions (Wang et al., 2018). The success in observation in turn motivates the development of planet formation and evolution theory (Dawson & Johnson, 2018; Wang et al., 2019; Liu & Ji, 2020). The sample for tidal dissipation studies would be significantly enlarged in the next decade, especially through the joint survey of ground-based telescopes, e.g., SiTian and the next generation exoplanet space telescopes, e.g., Habitable ExoPlanet Survey (HEPS; Yu et al., 2019, 2020), Earth 2.0 Transit Planet Survey, and shall be studied in details by ARIEL (Tinetti et al., 2018), HABItable Terrestrial planetary ATmospheric Surveyor (HABITATS; Wang et al., 2020). In this work, we discuss the potential contribution of SiTian in detecting and investigating tidal decaying hot Jupiters. We present general-purpose tools for tidal dissipation detection, using the simulated light curves of SiTian. The detection pipeline contains four major steps, i.e., light curve generation, contamination light deblending, transit timing detection, and transit timing modeling. The latter three modules are described given that light curve generation would be integrated into SiTian science processing pipeline (Liu et al., 2021). The paper is organized as follows. In Section 2, we introduce the SiTian exoplanet observation strategy and the generation of simulated SiTian light curves. In Section 3, we present the tidal dissipation identification pipeline and apply it to the simulated light curves of WASP-161b and XO-3b. In Section 4, we discuss the potential contribution of SiTian in planet tidal dissipation. In Section 5, a summary is presented. ## 2 SiTian Exoplanet Observational Strategy and Simulated Transit Light Curves Generation SiTian is an integrated network of 1-meter telescopes, aiming at time-domain astronomy (Liu et al., 2021). It shall produce a huge amount of the light curves, that would contribute to time-domain researches of targets with physical size ranging from galaxy clusters to planets (Liu et al., 2019, 2020; Yang et al., 2022, 2020b, 2020a; Lennon et al., 2021; Wang et al., 2021; Yang et al., 2021; Ngeow et al., 2021). Monitoring the same area of the sky simultaneously in three bands ($u$, $g$, $i$), SiTian will deliver high precision timing measurements, enabling detections of TTVs. The commissioning of SiTian is expected to be before or around 2030, and the first three proto- type equipment may start its operation in the middle of 2022. ### 2.1 SiTian Technical Specification for Exoplanet Research SiTian will eventually accomplish with a worldwide network of 72 telescopes, monitoring a sky area of $\sim$ 30,000 deg2 (Liu et al., 2021). The main monitoring sky coverage is $\sim$ 10,000 deg2 which can be observed by telescopes located in China. The typical exposure time is 1 minute, resulting in a 5 $\sigma$ brightness limit of 21.1 mag in the $g$-band (Liu et al., 2021). The photometric precision is expected to reach 1$\%$ for point sources with $V=16$ mag (Liu et al., 2021), assuming the detector has a QE of 70$\%$, read noise of 7 e-, an optical throughput of 70$\%$, filter transmission fraction of 70$\%$, seeing of 2.5 arcsec, and night-sky brightness of 21.1 mag arcsec-2. The applied CMOS detector GSENSE4040 has a readout noise of 3.7 e- (Liu et al., 2021). A classic good observing site has an average night-sky brightness around or better than 22.0 mag arcsec-2 and a median seeing size better than 1 arcsec, for example, the Lenghu site(Deng et al., 2021), which is the most promising major Chinese site for SiTian. Using these parameters for calculation, the photometric precision is predicted as shown in Table 1. The photometric precision is $\sim$ 700 parts per million (ppm) at $g$-band and 1075 ppm at $i$-band for 12 mag targets (as shown in Figure 1). The photometric uncertainties of sources brighter than 12 mag are dominated by photon noise. In addition, the SiTian mission would occupy at least three 4-meter class spectroscopic telescopes for following-up observation. Figure 1: The simulated and model-predicted light curves. The blue points are the simulated SiTian light curve and the red points are binned data for clarity. The black line shows the best-fit model light curves, while the yellow region presents 1 $\sigma$ significance region. The fitting residuals are shown below the light curves. The top panel refers to WASP-161b and the bottom panel refers to XO-3b. Table 1: The expected limiting magnitude and photometric precision of a 1 minute single exposure for one SiTian telescope, based on the technical specifications from Liu et al. (2021). | $u$ | $g$ | $i$ ---|---|---|--- Limiting magnitudea | 20.2 | 21.1 | 20.4 Photometric Precision at Certain Magnitude Magnitude | $\sigma(u)$ | $\sigma(g)$ | $\sigma(i)$ 16 Mag | 0.85$\%$ | 0.48$\%$ | 0.76$\%$ 12 Magb | 1258 ppm | 708 ppm | 1075 ppm Note. (a) 5 $\sigma$ limit magnitude. (b) The photometric errors of brighter sources or longer exposure time are dominated by Poisson error. The data will be reduced by the SiTian collaboration in two modes, i.e., online and offline (Liu et al., 2021). The quick data reduction, light curve retrieval, and classification will be performed in the online mode immediately after observation with a typical delay of 1 minute. In addition, alerts for transient events and other interesting targets will be triggered. Full data products, e.g. time-domain images, light curves, and target catalogs would be available and released to the public after reducing by the offline pipeline. The capability of an observation to detect transit signal in a planetary system can be represented by the signal to noise ratio (S/R) as: ${\rm S/R}=\sqrt{\frac{t_{obs}}{P}}\frac{R_{p}^{2}}{R_{\ast}^{2}}/\sigma_{total},\\\ $ (1) where $t_{obs}$ is the total integration time for the source, $P$ the orbit period, $R_{p}$ and $R_{\ast}$ the radius of the planet and host star, respectively. $\sigma_{total}$ is the total uncertainty in the duration of transit event, taking into account Poisson photon noise, uncorrected stellar variability, and equipment noise (Borucki et al., 2010). For example, a star with $g$=12 observed by SiTian would have a $\sigma_{total}$ of 36 ppm when the t${}_{obs}\sim$6.5 hr, a typical transit duration of hot Jupiters (Akeson et al., 2013; Yang et al., 2022, 2021). A typical stellar variability level for a low variability star like Sun is $\sim$ 10 ppm on the timescale of a planet transit (Jenkins, 2002). This stellar variability is significantly smaller than Poisson photon noise. The $\sigma_{total}$ is $\sim 239$ ppm for a 16 mag star in the same stacked time. The average observation time tobs can be estimated by multiplying the survey time and the ratio of exposure time to scanning cadence. The total field of view (FOV) of SiTian is 600 deg2. The cadence for the main sky coverage (10,000 deg2) is 30 minutes. Assuming 8-hour observation time per night and 300 observing nights per year, the average annual tobs is 3.3 days. For Neptune-sized planet (3$R_{\oplus}$) orbiting $g$=12 mag stars with orbital periods of 30 days, the average S/N is 8.2 in one year observation, superior to the empirical exoplanet detection threshold S/R = 7.1 (Borucki et al., 2010; Fressin et al., 2013; Christiansen et al., 2015, and references therein). For a $g=16$ Sun-like star, the S/R for an orbiting Jupiter-sized planet is $\sim$ 16 in one-year observation. The S/R is the same for a Neptune orbiting an M-type star. A more detailed discussion of planet detecting rate shall be presented in the following work, describing the pipeline of exoplanet detecting (Yang et al. in preparation). Here, we present a simple and general calculation. The number of planets that can be detected by a survey depends on three factors, i.e., the planet occurrence rate, star counts, and solid angle. The solid angle equals to $R_{p}$/$a$, where $R_{p}$ is planet radius and $a$ is orbital semi-major axis (Borucki et al., 2010). The planet occurrence rate is estimated to be $\sim$ 0.15 for SiTian’s capability, based on the knowledge and lessons gained from the Kepler mission (Howard et al., 2012; Fressin et al., 2013; Christiansen et al., 2015). Taking the advantage of the high spatial resolution of $\sim$ 1 arcsec, and considering the number density of bright stars, SiTian is planned to monitor the galactic plane for exoplanet research. There are about 6.5 million stars brighter than 12 mag in the galactic plane with galactic longitude $b$ between 0∘ and 20∘ (Robin et al., 2003). This leads to an expected planet detection of 25,000 if applying an average solid angle of 2.5$\%$ (Akeson et al., 2013). The expected detected planet would thereby be 250 per year, assuming 1$\%$ SiTian time is allocated for bright sources. We note that the blending of nearby sources is correctable (cf. details in Section 2.2) for bright target transit observation (Yang et al., 2022, 2021, 2021). Thereby, monitoring crowded stars in the Galactic plane is an efficient strategy for SiTian to detect transit planets. The expected number of stars with $g\sim$15-17 mag and with absolute $b$ between 20 and 40 ∘ is $\sim$ 1 million(Robin et al., 2003). The occurrence of Jupiter-sized planets around Sun-like stars is about 0.06 (Fressin et al., 2013), which implies the detection of 1500 Jupiter size exoplanets orbiting Sun-like stars by SiTian. In addition, about 1400 Neptune size exoplanets orbiting M-type star are expected to be detected, considering the small size of M-type stars and their large abundance of $\sim$ 70$\%$ (Chabrier, 2003). The detection of these 2900 exoplanets around faint stars are byproducts of SiTian’s main scientific objectives. We emphasize that with only 1$\%$ observing time on bright stars, SiTian may discover additional 2500 exoplanets in ten-year observation and up to 25000 planets with 10$\%$ time. ### 2.2 SiTian Data for Transit Timing Analysis and Simulated Transit Light Curves The proposed SiTian Mission would significantly extend the sample size of exoplanets. The expected multiple timing measurements of these newly discovered targets from SiTian may already allow a transit timing analysis. Meanwhile, SiTian’s main survey or follow-up discretionary program of the known planets with reported timing observation and TTV evidence will provide opportunities for a full TTV study of these targets. The planet identification package will detect exoplanet candidates and provide their preliminary orbital and planetary parameters (Yang et al., in preparation). For the research of TTV, it is necessary to build a specialized package to obtain more precise transit parameters, especially the transit timings. The potential targets for the pipeline include already-known hot Jupiters and new transiting hot Jupiters detected from the SiTian survey. Below we describe our simulation of SiTian’s capability on the study of TTV and tidal dissipation. The deblending of the light curve is crucial for the transit planet survey programs with poor spatial resolution and is particularly important for studies relying on high precision transit depth measurements (Yang et al., 2022, 2021). We note that the high spatial resolution and sampling are one of the major advantages of SiTian. We model and remove the blending light from the unsolved sources in the vicinity of the target for the TESS image which has a pixel size of 21 arcsec (Yang et al., 2022). Firstly, the correlation between the fraction of blended light and the distance of the contaminating star to the target is built. The step-by-step description of blending-distance correlation is available in our previous work (Yang et al., 2022). With this relation, one can calculate the contamination fraction of every source and their sum is the total blending fraction. The calculation needs external information from the Gaia catalog (Gaia Collaboration et al., 2018) on the flux and position of individual stars inside the aperture of the target star. The deblending method that we have developed delivers TESS brightness highly consistent the Gaia brightness (Yang et al., 2022). The derived transit depths are consistent within 1$\sigma$ with those given by the TESS Pre-Search Data Conditioning (PDC; Smith et al., 2012) modules for the comparison samples (Yang et al., 2022, 2021). The deblending package has been provided as a general-purpose software 111https://github.com/sailoryf/TESS_Deblending/, and will be used for the deblending and correction of the SiTian light curves. For this purpose, we first simulate the “observed” SiTian $g$-band light curves of two hot Jupiters with the host star of 12 mag. The sampling interval is set as 1 minute with the input observational uncertainty of 708 ppm (as shown in Table 1). The input parameters are set as the same as those of WASP-161b and XO-3b, which shows TTV evidence (Yang & Chary, 2021, submitted; Yang & Wei, 2021). For each source, we simulate 1000 transit light curves. The light curves are all set with a total observation time of 8 hours. The simulation light curves are de-trended by a polynomial fit to the several-hour out-of-transit baseline data, which is shown to be a valid approach in (Yang et al., 2022, 2021, 2021). WASP-161b is a hot Jupiter orbiting an F6-type star every 5.41 days (Barkaoui et al., 2019) and is reported to bear significant TTV related to tidal dissipation (Yang & Chary, 2021, submitted). The giant planet has a mass ($M_{p}$) of 2.49$\pm$0.21$M_{J}$ and a radius ($R_{p}$) of $1.14\pm 0.06R_{J}$. The host star has a mass of 1.39$\pm$0.14$M_{\odot}$, a radius of $1.71\pm 0.08R_{\odot}$. Applying a combined-fit to both the radial velocity curve and the TESS 2-minute cadence light curve, Yang & Chary (2021, submitted) has an orbital eccentricity of $e$ = 0.34$\pm$0.03, an inclination of 89.58$\pm$0.28, and a semi-major axis in stellar radii ($a/R_{\ast}$) of 6.57$\pm$0.45 (Yang & Chary, 2021, submitted). In addition, a significant TTV is reported by combining the TESS observation and the archival timing measurements. XO-3b is an another hot Jupiter with significant TTVs (Shan et al., 2021; Yang & Wei, 2021). The planet-star system has an orbit period ($P_{orb}$) of 3.19 days, an $a$ of 4.95$\pm$0.18, an inclination of 84.20$\pm$0.54, and an $e$ of 0.27587${}^{+0.00071}_{-0.00067}$ (Winn et al., 2008; Bonomo et al., 2017; Stassun et al., 2017). The planet has an $M_{p}$ of 11.70$\pm$0.42$M_{J}$ and an $R_{p}$ of 1.217$\pm$0.073$R_{J}$. The host star has an $M_{\ast}$ of 1.213$\pm$0.066$M_{\odot}$ and an $R_{\ast}$ of 1.377$\pm$0.083$R_{\odot}$. When simulating the observed light curve, and performing light curve model fitting, we used the classic planet transit model assuming a Keplerian orbit from Mandel & Agol (2002). The parameters include $R_{p}/R_{\ast}$, $a/R_{\ast}$, transit mid-point $T_{C}$, inclination, the argument of periapsis, the time of periapse passage, the longitude of the ascending node, and quadratic limb darkening coefficients (a1 and a2). The limb darkening coefficients are inserted from limb darkening models, depending on the stellar type of the host star and observational band (Claret, 2000; Claret & Bloemen, 2011; Yang et al., 2021, 2021). We set the transit midpoint as the time zero reference for comparison purposes. Examples of synthetic light curves for WASP-161b and XO-3b are as shown in Figure 1. The blending light from sources nearby is negligible given that no comparable sources within 10′′ are detected and SiTian data has a pixel resolution of 1′′. Figure 2: The posterior distribution obtained from the MCMC fitting to the simulated light curves of WASP-161b (top), XO-3b (bottom). The dashed lines in the diagonal histograms give the 1 $\sigma$ regions. The corner plot uses the routine from Foreman-Mackey (2016). ## 3 The pipeline for Tidal Dissipation Candidates Identification Our pipeline to search for the tidal dissipation candidates is firstly developed for the TESS data, including four main modules: the light curve generation and detrending, the light curve deblending from contamination by additional stars inside the aperture, transit timing determinations, and multi-epoch timing analysis. The light curve generation and detrending will be performed by the SiTian data reduction pipeline (Liu et al., 2021) which should be more specializing for SiTian data. We have described light curve deblending in Section 2.2. Below, we describe timing obtaining, and timing modeling modules in detail. Our pipeline is based on the usage of functions from NumPy (Harris et al., 2020), Astropy (Astropy Collaboration et al., 2013), EXOFAST (Eastman et al., 2013), and PYMC (Patil et al., 2010). NumPy and Astropy provide basic statistics and astronomical calculations. EXOFAST utilize the transit model from Mandel & Agol (2002). PYMC furnishes the Markov Chain Monte-Carlo (MCMC) technique. ### 3.1 Modeling Transit Light Curve and Obtaining Transit Timing The Markov Chain Monte-Carlo technique (Patil et al., 2010; Foreman-Mackey et al., 2013), or alternatively, the multimodal nested sampling algorithms (MULTINEST; Feroz et al., 2009) are widely used for the fitting to transit light curves. These two methods have been compared in previous work and the results obtained are consistent (Yang et al., 2021). In our pipeline, we currently use the MCMC method and may add the MULTINEST method as an option in future versions. The fitting procedure applies the Mandel & Agol (2002) model assuming Keplerian orbit, to generate theoretical transit light curves. All parameters involved in light curve generation are set as free parameters. Uniform priors ranging through the whole reasonable space are used for all these parameters, except for the limb darkening coefficients, which use Gaussian priors with a $\sigma$ of 0.05. The central priors of the limb darkening coefficients are dependent on stellar spectral types and can be inferred via interpolation of the classic limb darkening model (Claret, 2000; Claret & Bloemen, 2011; Yang et al., 2021). The spectral types of the planet host stars can be found from Gaia catalogs. The induced uncertainties in our work can be up to a few percent in radius ratios and are not significant for the timing study (Yang et al., 2021). The MCMC algorithm omits some first steps as burn-in and applies a number of iterations for probability statistics. More iteration should more likely give an effective MCMC fitting. However, the numbers of the omitted steps and the applicable steps linearly relate to the consuming time of MCMC fitting. In practice, we show that 30,000 as burn-in numbers and 50,000 as applicable steps are high enough for the TESS light curve fitting (Yang et al., 2021, 2021). An investigation has been performed for the simulated SiTian light curves, to determine the threshold of the chain steps for securing stable MCMC results. For each of WASP-161b and XO-3b, 100 light curves are generated using their reported parameters as input, respectively. We then fit the light curves and monitor the stability of the MCMC results for the 100 run. The monitoring shows that the MCMC fitting turns stable when 10,000 steps are used as burn-in and 20,000 steps are used as analysis. Considering possible extra fluctuation due to the diversity of planet parameters and light curve quality in real applications, we apply 30,000 as burn-in and 50,000 for a statistic in the pipeline. We note that a tenfold-length chain will be used for high priority sources which already show tidal dissipation evidence, for example, WASP-161 b, and XO-3 b. The probability statistics of their parameters are shown in Figure 2. The MCMC fitting obtains a $T_{C}$ of 0.00027$\pm$0.00031 days (0.39$\pm$0.44 minutes) for WASP-161b and a $T_{C}$ of 0.00010$\pm$0.00022 days (0.14$\pm$0.32 minutes) for XO-3b, consistent with the input $T_{C}$ of 0. We apply further simulations to investigate whether the MCMC fitting is robust. We generate another 1000 light curves for each planet and fit them individually. The resulting distributions are well consistent within 1$\sigma$ to the MCMC derived distributions in fitting one particular light curve. Taking $T_{C}$ as example, the distributions are 0.00000$\pm$0.00038 days (0$\pm$0.54 minutes) for WASP-161b, and 0.00001$\pm$0.00027 days (0.01$\pm$0.39 minutes) for XO-3b, as shown in Figure 3. Therefore, we conclude that the MCMC fitting gives precise timing measurement and the timing precision obtained for the simulated SiTian light curve is $\sim$ 0.5 minutes for a single transit observation. Figure 3: The distribution of the best-fit $T_{C}$ values of the 1000 simulated light curves. Upper panel for WASP-161 b and lower for XO-3 b. The vertical dash lines show the input value of 0. ### 3.2 Transit Timing Modeling and Tidal Dissipation Identification For hot Jupiters, observed TTV signal can be accounted for by various mechanisms, e.g., tidal dissipation, apsidal precession, R$\o$mer effect, multi-planet interaction, and mass loss. Exploring these physical processes is among the main scopes of SiTian for the exoplanet research. The apsidal precession scenario can be identified or rejected by studying the shape of the light curve (Jordán & Bakos, 2008; Antoniciello et al., 2021; Yang & Wei, 2021). For the mass-loss case, the TTV signal is dependent on the mass rate and amplitude, which can be well constrained by orbital parameters. The R$\o$mer effect can be tackled by long-term radial velocity monitoring of the host star, or high-resolution imaging search for stellar-mass companion(Siverd et al., 2018; Yang et al., 2021). In principle, a TTV signal induced by tidal dissipation should persist a constant period derivative, which can be easily spotted out from transit timing data using our developed pipeline (as shown in Figure 4). Figure 4: The archival timings (black points) and the predicted new SiTian timing observations (red diamonds) of WASP-161b (top) and XO-3b (bottom). The black horizontal line gives the best fitted constant period model. The red dashed curve presents tidal dissipation models as described by Yang & Chary (2021, submitted); Yang & Wei (2021) with the yellow region showing 1 $\sigma$ region encompassing 68$\%$ archival timings. The simulated SiTian data are to be taken in 2022, 2027, and 2032, which can effectively distinguish between the linear and the quadratic models. As predicted by the tidal dissipation model, timing variation follows a quadratic function of the transit epochs (Patra et al., 2017; Yang & Chary, 2021, submitted): $t_{\mathrm{tra}}(N)=t_{0}+NP+\frac{1}{2}\frac{dP}{dN}N^{2},\\\ $ (2) where $t_{\mathrm{tra}}(N)$ is the timing of Nth transit, $t_{0}$ is the zero point. It was found in previous works that timings of WASP-161b and XO-3b significantly favor a quadratic function model than linear models (Yang & Chary, 2021, submitted; Yang & Wei, 2021). We present the predicted SiTian timing measurements of WASP-161b and XO-3b in 2022, 2027, and 2031 (as shown in Figure 4). Transit midpoints are generated according to Equation 2. We recalculate BICs to estimate the evidence among different models. The $\Delta$BIC of WASP-161b and XO-3b is 8.8, 383 obtained from archival timings, favoring a quadratic model compared to a linear model. The $\Delta$BIC of WASP-161b and XO-3b between linear and quadratic models would be larger than 5.5$\times$105 and 1.8$\times$104, favoring quadratic models. Adding the future SiTian measurements is shown to be highly capable of distinguishing the difference between the constant period model and the period decaying model. In our tidal dissipation detection module, a target with the BIC difference between the best fit quadratic model and the best fit linear model larger than 5 will be flagged as a candidate planet with tidal dissipation. These candidates shall be further studied and verified by follow- up observations including more transit epochs measuring and radial velocity monitoring. We present timings and following-up strategy of KELT-19Ab as an example. KELT-19Ab is reported to hold a maximum stellar acceleration of 4 m s-1 yr-1 caused by the binarity of its host star (Siverd et al., 2018). Applying the relation between the stellar acceleration and transit period derivative from Bouma et al. (2020), the transit period would present a derivative of 5.32 ms yr-1. Shan et al. (2021) report a possible but not significant period derivative of 112$\pm$94 ms yr-1. Timings obtained from SiTian would potentially distinguish these different scenarios. SiTian timing predictions are obtained following the same process as described above with planet parameters taken from (Siverd et al., 2018). The timing would present a difference at $\sim$ 1.5 minutes in 2022 between a linear and a quadratic model obtained from archival timings available (as shown in Figure 5). The difference would be significant when one has timings in 2027 and 2031. The $\Delta$BIC obtained from archival timings between linear and quadratic models is 2.1, slightly favoring a quadratic model. It would derive a $\Delta$BIC of 1963 until 2031 if the observational timings are as predicted by the quadratic model (seen in Figure 5). Moreover, the stellar acceleration would be detectable with SiTian observation in 2031. It would reveal a 1-minute difference between the constant period and the stellar acceleration models. The period derivative due to acceleration is taken from Shan et al. (2021). Figure 5: Timings of KELT-19Ab with symbols the same to Figure 4. The green line presents the period derivative caused by stellar acceleration calculated by Shan et al. (2021). Moreover, long-term transit observation can also help to discriminate between possible TTV origins. Planets in a multi-planet system are usually not massive and close enough to each other to cause TTVs with an amplitude larger than 10 minutes (Holman et al., 2010; Huang et al., 2016). Therefore, it may rule out the possibility of the multi-planet origin, if either a monotonic TTV or a large TTV amplitude is observed, as shown in Figure 4. For another important arguing possible origin, apsidal precession, the transit duration variation (TDV) should be more significant than TTV (Pál & Kocsis, 2008; Ragozzine & Wolf, 2009). The TDV can be obtained by modeling the transit light curves as well. No evidence of TDV in the archival data is found for WASP-161b and XO-3b (Yang & Chary, 2021, submitted; Yang & Wei, 2021). We expect that the SiTian project may provide additional observation on TDV. ## 4 discussion The observed TTV signal and its interpretation are highly dependent on the data quality of timing observation. Any unexpected factors in deriving the transit timing parameters may change the period derivative (detailed example in Yang & Chary, 2021, submitted; Shan et al., 2021), and thus leads to different explanations. Therefore, long-baseline transit monitoring with high precision is crucial for enhancing understanding of the tidal dissipation process. We have introduced the capacity of SiTian in modeling the tidal dissipation of WASP-161b and XO-3b (as shown in Figure 4). For them, we have scheduled multi- epoch transit observations from 2022 with the SiTian prototype telescopes. In addition, continuous observation of at least one transit per five years is proposed. This observation would improve the accuracy and robustness of the period derivative, which is important for modeling the dissipation process for these two benchmark planets. The migration is believed to be the reason that giant gas planets can exist at such en-close orbit (Dawson & Johnson, 2018). The migration is probably due to one of the two scenarios: tidal migration that loses angular momentum by e.g., planet scattering, stellar companion scattering, and secular interaction (Wu & Lithwick, 2011; Naoz et al., 2011); or migration due to disk friction in the early stage of planet formation (Lin et al., 1996). A high eccentricity system showing period decaying favors the former model (Yang & Chary, 2021, submitted; Yang & Wei, 2021). Also, the tidal dissipation occurrence rate should be higher for the former model. It is likely that both migration mechanisms may contribute separately to different sources, or even act in the same source while in different evolutionary stages. Interestingly, the system parameters of the five reported tidal dissipation exoplanets show obvious diversity (Patra et al., 2017; Dong et al., 2021; Davoudi et al., 2021; Yang & Wei, 2021; Yang & Chary, 2021, submitted). Extending the available sample crucial for further statistic investigations. Using the methods introduced by Dong et al. (2018); Zhu & Dong (2021), we are performing statistic analysis to find possible relationships between the eccentricity, orbital radius or semi-major axis, obliquity, metallicity, and planet-multiplicity with no conclusive results yet due to the limited sample size. A large number of candidate systems with evidence of tidal dissipation may be discovered or identified by SiTian, which will be of great importance for understanding tidal migration. The detection number depends on the real occurrence rate of planets with tidal dissipation, which is one of the most important questions to answer by future observations, e.g. SiTian. The current fraction of such systems is 5 out of more than 4000. We note that planets other than hot Jupiters bearing tidal dissipation should be rarer. We expect to discover $\sim$ 50 new tidal dissipation star-planet systems taking into account at least one order of magnitude larger sample size. Moreover, other targets showing TTV signals but caused by other mechanisms such as planet- planet interactions, apsidal precession, will be identified by our pipeline as byproducts. For targets with high scientific interests, following-up observation would be promptly scheduled with the SiTian 4-meter class spectroscopic following-up telescopes (Liu et al., 2021). The following-up observation would give us information of e.g., host star properties, binarity, and planet multiplicity. In addition, future advanced facilities will provide important opportunities for deep investigations. For example, the HABItable Terrestrial planetary ATmospheric Surveyor (HABITATS), which is a proposed mission for a 4-6 meter space telescope aiming at modeling the atmospheric features of exoplanets down to Earth-like planets(Wang et al., 2020). The high-precision transmission spectra from HABITATS will help us understand the atmosphere properties of tidal dissipation targets, including the composition, thermal structures, inflation, and escape. This information shall yield a comprehensive understanding of the undergoing physical processes in the tidal dissipating sources. ## 5 Summary We describe a pipeline for tidal dissipation detection as software preparation for the ongoing SiTian survey which is expected to have the first light of the three prototype telescopes in the middle of 2022. The SiTian survey will have 72 telescopes in total with full installation till 2030 (Liu et al., 2021). We have estimated SiTian’s capability for detecting exoplanets, based on its technical parameters. SiTian is expected to discovery 5,000 to 25,000 new exoplanets. Assuming a similar tidal dissipation occurrence, there would be $\sim$ 50 hot Jupiters showing tidal dissipation evidence. The pipeline for tidal dissipation detection has modules of light curve deblending, transit light curve modeling, and timing modeling for tidal dissipation detection. The light curve deblending is based on our developed algorithm for TESS light curve deblending (Yang et al., 2022). For each target, a relationship between the contamination fraction and the distance to the target will be built, based on the brightness and position measurements for individual stars from the GAIA catalog (Gaia Collaboration et al., 2018), and then corrected to obtain “uncontaminated” light curves for SiTian. We set the reported period decaying candidates of WASP-161b and XO-3b as examples to describe our pipeline and the contribution of SiTian data. We simulate SiTian light curves for WASP-161b and XO-3b with a photometric precision of $\sim$ 700 ppm in 1-minute cadence, estimated from SiTian’s technical parameters. The MCMC technique is applied when modeling transit light curves. We show that our transit timing measurement has a precision of 0.5 minutes for a single transit observation. We conclude that the inclusion of SiTian data shall provide key evidence to discriminate between various TTV origin models. For WASP-161b and XO-3b, a tidal dissipation origin seems to be the most likely explanation. The proposed future SiTian measurements shall significantly improve the model confidence and robustness. We thank SiTian Collaboration for the support in preparing this work. The work made use of the NASA Exoplanet Archive 222https://exoplanetarchive.ipac.caltech.edu/index.html (Akeson et al., 2013). We would like to thank Jin-Ping Zhu for the careful reading and constructive suggestions. We thank Bo Zhang for the helpful discussion. Fan Yang and Su-Su Shan are supported by funding from the Cultivation Project for LAMOST Scientific Payoff and Research Achievement of CAMS-CAS. Xing Wei acknowledges National Natural Science Foundation of China (NSFC; No.11872246, 12041301), and the Beijing Natural Science Foundation (No. 1202015). Wei Wang is supported by the National Natural Science Foundation of China (NSFC) grants No. 11988101, 42075123, the National Key RD Program of China No. 2019YFA0405102, and the science research grants from the China Manned Space Project with NO. CMS-CSST-2021-B12. ## References * Akeson et al. (2013) Akeson, R. L., Chen, X., Ciardi, D., et al. 2013, PASP, 125, 989 * Antoniciello et al. (2021) Antoniciello, G., Borsato, L., Lacedelli, G., et al. 2021, MNRAS, 505, 1567 * Astropy Collaboration et al. (2013) Astropy Collaboration, Robitaille, T. P., Tollerud, E. J., et al. 2013, A&A, 558, A33 * Barkaoui et al. (2019) Barkaoui, K., Burdanov, A., Hellier, C., et al. 2019, AJ, 157, 43 * Bellm et al. (2019) Bellm, E. C., Kulkarni, S. R., Graham, M. J., et al. 2019, PASP, 131, 018002 * Benz et al. (2021) Benz, W., Broeg, C., Fortier, A., et al. 2021, Experimental Astronomy, 51, 109 * Bonomo et al. (2017) Bonomo, A. S., Desidera, S., Benatti, S., et al. 2017, A&A, 602, A107 * Borsato et al. (2021) Borsato, L., Piotto, G., Gandolfi, D., et al. 2021, MNRAS, 506, 3810 * Borucki et al. (2010) Borucki, W. J., Koch, D., Basri, G., et al. 2010, Science, 327, 977 * Bouma et al. (2020) Bouma, L. G., Winn, J. N., Howard, A. W., et al. 2020, ApJ, 893, L29 * Bouma et al. (2019) Bouma, L. G., Winn, J. N., Baxter, C., et al. 2019, AJ, 157, 217 * Chabrier (2003) Chabrier, G. 2003, PASP, 115, 763 * Christiansen et al. (2015) Christiansen, J. L., Clarke, B. D., Burke, C. J., et al. 2015, ApJ, 810, 95 * Claret (2000) Claret, A. 2000, A&A, 363, 1081 * Claret & Bloemen (2011) Claret, A., & Bloemen, S. 2011, A&A, 529, A75 * Davoudi et al. (2021) Davoudi, F., Baştürk, Ö., Yalçınkaya, S., Esmer, E. M., & Safari, H. 2021, AJ, 162, 210 * Dawson & Johnson (2018) Dawson, R. I., & Johnson, J. A. 2018, ARA&A, 56, 175 * Deng et al. (2021) Deng, L., Yang, F., Chen, X., et al. 2021, Nature, 596, 353 * Dong et al. (2021) Dong, J., Huang, C. X., Zhou, G., et al. 2021, ApJ, 920, L16 * Dong et al. (2018) Dong, S., Xie, J.-W., Zhou, J.-L., Zheng, Z., & Luo, A. 2018, Proceedings of the National Academy of Science, 115, 266 * Eastman et al. (2013) Eastman, J., Gaudi, B. S., & Agol, E. 2013, PASP, 125, 83 * Feroz et al. (2009) Feroz, F., Hobson, M. P., & Bridges, M. 2009, MNRAS, 398, 1601 * Foreman-Mackey (2016) Foreman-Mackey, D. 2016, The Journal of Open Source Software, 1, 24 * Foreman-Mackey et al. (2013) Foreman-Mackey, D., Hogg, D. W., Lang, D., & Goodman, J. 2013, PASP, 125, 306 * Fressin et al. (2013) Fressin, F., Torres, G., Charbonneau, D., et al. 2013, ApJ, 766, 81 * Gaia Collaboration et al. (2018) Gaia Collaboration, Brown, A. G. A., Vallenari, A., et al. 2018, A&A, 616, A1 * Garai et al. (2021) Garai, Z., Pribulla, T., Parviainen, H., et al. 2021, MNRAS, 508, 5514 * Harris et al. (2020) Harris, C. R., Millman, K. J., van der Walt, S. J., et al. 2020, Nature, 585, 357 * Holman et al. (2010) Holman, M. J., Fabrycky, D. C., Ragozzine, D., et al. 2010, Science, 330, 51 * Howard et al. (2012) Howard, A. W., Marcy, G. W., Bryson, S. T., et al. 2012, ApJS, 201, 15 * Huang et al. (2016) Huang, C., Wu, Y., & Triaud, A. H. M. J. 2016, ApJ, 825, 98 * Ivezić et al. (2019) Ivezić, Ž., Kahn, S. M., Tyson, J. A., et al. 2019, ApJ, 873, 111 * Jenkins (2002) Jenkins, J. M. 2002, ApJ, 575, 493 * Jiang et al. (2016) Jiang, I.-G., Lai, C.-Y., Savushkin, A., et al. 2016, AJ, 151, 17 * Jordán & Bakos (2008) Jordán, A., & Bakos, G. Á. 2008, ApJ, 685, 543 * Lennon et al. (2021) Lennon, D. J., Maíz Apellániz, J., Irrgang, A., et al. 2021, A&A, 649, A167 * Lin et al. (1996) Lin, D. N. C., Bodenheimer, P., & Richardson, D. C. 1996, Nature, 380, 606 * Lin et al. (2021) Lin, J., Wang, X., Mo, J., et al. 2021, arXiv:2109.11155 * Liu & Ji (2020) Liu, B., & Ji, J. 2020, Research in Astronomy and Astrophysics, 20, 164 * Liu et al. (2021) Liu, J., Soria, R., Wu, X.-F., Wu, H., & Shang, Z. 2021, An. Acad. Bras. Ciênc. vol.93 supl.1, 93, 20200628 * Liu et al. (2019) Liu, J., Zhang, H., Howard, A. W., et al. 2019, Nature, 575, 618 * Liu et al. (2020) Liu, J., Zheng, Z., Soria, R., et al. 2020, ApJ, 900, 42 * Ma et al. (2018) Ma, B., Shang, Z., Hu, Y., et al. 2018, MNRAS, 479, 111 * Mandel & Agol (2002) Mandel, K., & Agol, E. 2002, ApJ, 580, L171 * Naoz et al. (2011) Naoz, S., Farr, W. M., Lithwick, Y., Rasio, F. A., & Teyssandier, J. 2011, Nature, 473, 187 * Ngeow et al. (2021) Ngeow, C.-C., Liao, S.-H., Bellm, E. C., et al. 2021, AJ, 162, 63 * Ou et al. (2021) Ou, J.-W., Yu, C., Yang, M., et al. 2021, ApJ, 922, 37 * Pál & Kocsis (2008) Pál, A., & Kocsis, B. 2008, MNRAS, 389, 191 * Patil et al. (2010) Patil, A., Huard, D., & Fonnesbeck, C. J. 2010, J. Stat. Softw, 1 * Patra et al. (2017) Patra, K. C., Winn, J. N., Holman, M. J., et al. 2017, AJ, 154, 4 * Ragozzine & Wolf (2009) Ragozzine, D., & Wolf, A. S. 2009, ApJ, 698, 1778 * Ricker et al. (2015) Ricker, G. R., Winn, J. N., Vanderspek, R., et al. 2015, Journal of Astronomical Telescopes, Instruments, and Systems, 1, 014003 * Robin et al. (2003) Robin, A. C., Reylé, C., Derrière, S., & Picaud, S. 2003, A&A, 409, 523 * Sariya et al. (2021) Sariya, D. P., Jiang, I.-G., Su, L.-H., et al. 2021, Research in Astronomy and Astrophysics, 21, 097 * Shan et al. (2021) Shan, S.-S., Yang, F., Lu, Y.-J., et al. 2021, arXiv:2111.06678 * Siverd et al. (2018) Siverd, R. J., Collins, K. A., Zhou, G., et al. 2018, AJ, 155, 35 * Smith et al. (2012) Smith, J. C., Stumpe, M. C., Van Cleve, J. E., et al. 2012, PASP, 124, 1000 * Stassun et al. (2017) Stassun, K. G., Collins, K. A., & Gaudi, B. S. 2017, AJ, 153, 136 * Sun et al. (2015) Sun, L.-L., Gu, S.-H., Wang, X.-B., et al. 2015, Research in Astronomy and Astrophysics, 15, 117 * Tinetti et al. (2018) Tinetti, G., Drossart, P., Eccleston, P., et al. 2018, Experimental Astronomy, 46, 135 * Turner et al. (2021) Turner, J. D., Ridden-Harper, A., & Jayawardhana, R. 2021, AJ, 161, 72 * Valsecchi et al. (2015) Valsecchi, F., Rappaport, S., Rasio, F. A., Marchant, P., & Rogers, L. A. 2015, ApJ, 813, 101 * Wang et al. (2021) Wang, S., Zhang, H., Bai, Z., et al. 2021, arXiv e-prints, arXiv:2109.03149 * Wang et al. (2019) Wang, W.-L., Xu, X.-Q., & Liao, X.-H. 2019, Research in Astronomy and Astrophysics, 19, 130 * Wang et al. (2020) Wang, W., Zhai, M., Zhao, G., et al. 2020, in Society of Photo-Optical Instrumentation Engineers (SPIE) Conference Series, Vol. 11443, Society of Photo-Optical Instrumentation Engineers (SPIE) Conference Series, 114431Z * Wang et al. (2018) Wang, X.-Y., Wang, S., Hinse, T. C., et al. 2018, PASP, 130, 064401 * Weiss et al. (2013) Weiss, L. M., Marcy, G. W., Rowe, J. F., et al. 2013, ApJ, 768, 14 * Winn et al. (2008) Winn, J. N., Holman, M. J., Torres, G., et al. 2008, ApJ, 683, 1076 * Wu & Lithwick (2011) Wu, Y., & Lithwick, Y. 2011, ApJ, 735, 109 * Yang & Chary (2021, submitted) Yang, F., & Chary, R.-R. 2021, submitted, AJ * Yang et al. (2022) Yang, F., Chary, R.-R., & Liu, J.-F. 2022, AJ, 163, 42 * Yang et al. (2021) Yang, F., Shan, S.-S., Guo, R., et al. 2021, Astrophysics and Space Science, 366, 83 * Yang & Wei (2021) Yang, F., & Wei, X. 2021, arXiv:2111.06551 * Yang et al. (2020a) Yang, F., Long, R. J., Shan, S.-S., et al. 2020a, ApJS, 249, 31 * Yang et al. (2020b) Yang, F., Long, R. J., Shan, S.-S., et al. 2020b, MNRAS, 496, L59 * Yang et al. (2021) Yang, F., Long, R. J., Liu, J.-f., et al. 2021, AJ, 161, 294 * Yang et al. (2021) Yang, F., Zhang, B., Long, R. J., et al. 2021, arXiv e-prints, arXiv:2110.07944 * Yu et al. (2020) Yu, Z.-Y., Liu, H.-G., Zhou, J.-L., et al. 2020, Research in Astronomy and Astrophysics, 20, 035 * Yu et al. (2019) Yu, Z.-Y., Liu, H.-G., Zhou, J.-L., et al. 2019, Research in Astronomy and Astrophysics, 19, 004 * Zhang et al. (2019) Zhang, H., Yu, Z., Liang, E., et al. 2019, ApJS, 240, 17 * Zhang et al. (2020) Zhang, J.-C., Wang, X.-F., Mo, J., et al. 2020, PASP, 132, 125001 * Zhu et al. (2021) Zhu, J.-P., Wu, S., Yang, Y.-P., et al. 2021, arXiv e-prints, arXiv:2110.10469 * Zhu & Dong (2021) Zhu, W., & Dong, S. 2021, 59, arXiv:2103.02127
# Early Planet Formation in Embedded Disks (eDisk) VI: Kinematic Structures around the Very Low Mass Protostar IRAS 16253-2429 Yusuke Aso Korea Astronomy and Space Science Institute, 776 Daedeok-daero, Yuseong-gu, Daejeon 34055, Republic of Korea Woojin Kwon Department of Earth Science Education, Seoul National University, 1 Gwanak-ro, Gwanak-gu, Seoul 08826, Republic of Korea SNU Astronomy Research Center, Seoul National University, 1 Gwanak-ro, Gwanak-gu, Seoul 08826, Republic of Korea Nagayoshi Ohashi Academia Sinica Institute of Astronomy and Astrophysics, 11F of Astronomy-Mathematics Building, AS/NTU, No.1, Sec. 4, Roosevelt Rd, Taipei 10617, Taiwan Jes K. Jørgensen Niels Bohr Institute, University of Copenhagen, Øster Voldgade 5–7, DK 1350 Copenhagen K., Denmark John J. Tobin National Radio Astronomy Observatory, 520 Edgemont Rd., Charlottesville, VA 22903 USA Yuri Aikawa Department of Astronomy, Graduate School of Science, The University of Tokyo, 7-3-1 Hongo, Bunkyo-ku, Tokyo 113-0033, Japan Itziar de Gregorio-Monsalvo European Southern Observatory, Alonso de Cordova 3107, Casilla 19, Vitacura, Santiago, Chile Ilseung Han Division of Astronomy and Space Science, University of Science and Technology, 217 Gajeong-ro, Yuseong- gu, Daejeon 34113, Republic of Korea Korea Astronomy and Space Science Institute, 776 Daedeok-daero, Yuseong-gu, Daejeon 34055, Republic of Korea Miyu Kido Department of Physics and Astronomy, Graduate School of Science and Engineering, Kagoshima University, 1-21-35 Korimoto, Kagoshima, Kagoshima 890-0065, Japan Patrick M. Koch Academia Sinica Institute of Astronomy and Astrophysics, 11F of Astronomy-Mathematics Building, AS/NTU, No.1, Sec. 4, Roosevelt Rd, Taipei 10617, Taiwan Shih-Ping Lai Institute of Astronomy, National Tsing Hua University, No. 101, Section 2, Kuang-Fu Road, Hsinchu 30013, Taiwan Center for Informatics and Computation in Astronomy, National Tsing Hua University, No. 101, Section 2, Kuang-Fu Road, Hsinchu 30013, Taiwan Department of Physics, National Tsing Hua University, No. 101, Section 2, Kuang-Fu Road, Hsinchu 30013, Taiwan Academia Sinica Institute of Astronomy and Astrophysics, 11F of Astronomy-Mathematics Building, AS/NTU, No.1, Sec. 4, Roosevelt Rd, Taipei 10617, Taiwan Chang Won Lee Division of Astronomy and Space Science, University of Science and Technology, 217 Gajeong-ro, Yuseong- gu, Daejeon 34113, Republic of Korea Korea Astronomy and Space Science Institute, 776 Daedeok-daero, Yuseong-gu, Daejeon 34055, Republic of Korea Jeong-Eun Lee Department of Physics and Astronomy, Seoul National University, 1 Gwanak-ro, Gwanak-gu, Seoul 08826, Republic of Korea Zhi-Yun Li University of Virginia, 530 McCormick Rd., Charlottesville, Virginia 22904, USA Zhe-Yu Daniel Lin University of Virginia, 530 McCormick Rd., Charlottesville, Virginia 22904, USA Leslie W. Looney Department of Astronomy, University of Illinois, 1002 West Green St, Urbana, IL 61801, USA Suchitra Narayanan Institute for Astronomy, University of Hawai‘i at Mānoa, 2680 Woodlawn Dr., Honolulu, HI 96822, USA Nguyen Thi Phuong Korea Astronomy and Space Science Institute, 776 Daedeok-daero, Yuseong-gu, Daejeon 34055, Republic of Korea Department of Astrophysics, Vietnam National Space Center, Vietnam Academy of Science and Technology, 18 Hoang Quoc Viet, Cau Giay, Hanoi, Vietnam Jinshi Sai (Insa Choi) Academia Sinica Institute of Astronomy and Astrophysics, 11F of Astronomy-Mathematics Building, AS/NTU, No.1, Sec. 4, Roosevelt Rd, Taipei 10617, Taiwan Kazuya Saigo Department of Physics and Astronomy, Graduate School of Science and Engineering, Kagoshima University, 1-21-35 Korimoto, Kagoshima, Kagoshima 890-0065, Japan Alejandro Santamaría-Miranda European Southern Observatory, Alonso de Cordova 3107, Casilla 19, Vitacura, Santiago, Chile Rajeeb Sharma Niels Bohr Institute, University of Copenhagen, Øster Voldgade 5–7, DK 1350 Copenhagen K., Denmark Shigehisa Takakuwa Department of Physics and Astronomy, Graduate School of Science and Engineering, Kagoshima University, 1-21-35 Korimoto, Kagoshima, Kagoshima 890-0065, Japan Academia Sinica Institute of Astronomy and Astrophysics, 11F of Astronomy-Mathematics Building, AS/NTU, No.1, Sec. 4, Roosevelt Rd, Taipei 10617, Taiwan Travis J. Thieme Institute of Astronomy, National Tsing Hua University, No. 101, Section 2, Kuang-Fu Road, Hsinchu 30013, Taiwan Center for Informatics and Computation in Astronomy, National Tsing Hua University, No. 101, Section 2, Kuang-Fu Road, Hsinchu 30013, Taiwan Department of Physics, National Tsing Hua University, No. 101, Section 2, Kuang-Fu Road, Hsinchu 30013, Taiwan Kengo Tomida Astronomical Institute, Graduate School of Science, Tohoku University, Sendai 980-8578, Japan Jonathan P. Williams Institute for Astronomy, University of Hawai‘i at Mānoa, 2680 Woodlawn Dr., Honolulu, HI 96822, USA Hsi-Wei Yen Academia Sinica Institute of Astronomy and Astrophysics, 11F of Astronomy-Mathematics Building, AS/NTU, No.1, Sec. 4, Roosevelt Rd, Taipei 10617, Taiwan ###### Abstract Precise estimates of protostellar masses are crucial to characterize the formation of stars of low masses down to brown-dwarfs (BDs; $M_{*}<0.08~{}M_{\sun}$). The most accurate estimation of protostellar mass uses the Keplerian rotation in the circumstellar disk around the protostar. To apply the Keplerian rotation method to a protostar at the low-mass end, we have observed the Class 0 protostar IRAS 16253-2429 using the Atacama Large Millimeter/submillimeter Array (ALMA) in the 1.3 mm continuum at an angular resolution of $0\farcs 07$ (10 au), and in the 12CO, C18O, 13CO ($J=2-1$), and SO ($J_{N}=6_{5}-5_{4}$) molecular lines, as part of the ALMA Large Program Early Planet Formation in Embedded Disks (eDisk). The continuum emission traces a non-axisymmetric, disk-like structure perpendicular to the associated 12CO outflow. The position-velocity (PV) diagrams in the C18O and 13CO lines can be interpreted as infalling and rotating motions. In contrast, the PV diagram along the major axis of the disk-like structure in the 12CO line allows us to identify Keplerian rotation. The central stellar mass and the disk radius are estimated to be $\sim$0.12-0.17 $M_{\sun}$ and $\sim$13-19 au, respectively. The SO line suggests the existence of an accretion shock at a ring ($r\sim 28$ au) surrounding the disk and a streamer from the eastern side of the envelope. IRAS 16253-2429 is not a proto-BD but has a central stellar mass close to the BD mass regime, and our results provide a typical picture of such very low-mass protostars. Circumstellar disks (235) — Protostars (1302) — Low mass stars (2050) ††facilities: ALMA††software: astropy (Astropy Collaboration et al., 2013, 2018), CASA (McMullin et al., 2007), eDisk script (Tobin et al., 2023), ptemcee (Foreman-Mackey et al., 2013; Vousden et al., 2016), SLAM (Aso & Sai, 2023) ## 1 Introduction The low-mass end of star formation is connected to brown dwarf (BD) formation. Investigating such low-mass regimes is crucial to comprehensively understanding star formation. BDs are characterized by their masses that are not enough to fuse hydrogen ($M_{*}<0.08~{}M_{\sun}$) and are as numerous as hydrogen-burning stars (Chabrier, 2002). The formation process of BDs can be similar to or different from those of low-mass stars (Padoan & Nordlund, 2004). Theoretical studies have suggested various mechanisms for BD formation, such as turbulent fragmentation of a molecular cloud (e.g., Bate, 2012), disk fragmentation (e.g., Stamatellos & Whitworth, 2009), ejection from multiple young stellar systems (e.g., Basu et al., 2012), photo-erosion of a prestellar core by OB stars (e.g., Whitworth & Zinnecker, 2004), and eroding outflows (e.g., Machida et al., 2009). A point of view for testing these scenarios is whether or not physical quantities in star formation and BD formation follow the same scaling laws. Kim et al. (2019) reported that their sample of 15 proto-BD candidates have different scaling laws, than the laws among $\sim 60$ low-mass protostars, between the outflow force versus the luminosity and between the outflow force and the envelope mass. Protostars around the BD threshold i.e., proto-BDs or very low-mass protostars, are then important to determine down to which mass such a scaling law of protostars holds. Recent observational studies have aimed to identify and characterize proto-BDs, as well as pre-BDs (de Gregorio-Monsalvo et al., 2016; Huélamo et al., 2017; Santamaría-Miranda et al., 2021). To study star formation in this very low mass regime, it is, therefore, necessary to establish a method for precisely estimating the central mass of each protostar down to the mass regime around $M_{*}\sim 0.1~{}M_{\sun}$. The most direct method for estimating a central protostellar mass is to identify the Keplerian rotation in a circumstellar disk, as demonstrated in previous observations toward young stellar objects in the typical low mass regime ($M_{*}\gtrsim 0.2~{}M_{\sun}$; Yen et al., 2017). In contrast, previous studies of two representative proto-BD candidates, IC348-SMM2E and L328-IRS, estimated masses in indirect methods, which do not verify that the rotational velocity has the radial profile of the Keplerian rotation but merely assume that the observed gas motion is the Keplerian rotation or use other kinematics. The central mass of IC348-SMM2E is estimated to be $\sim 0.002-0.024~{}M_{\sun}$ from its luminosity, mass accretion rate, and efficiencies of mass accretion (Palau et al., 2014). This range includes a dynamical mass of $\sim 16~{}M_{\rm Jup}$ estimated on the assumption that the velocity gradient in the C18O $J=2-1$ line is due to the Keplerian rotation, based on Submillimeter Array (SMA) observations (Palau et al., 2014). The central mass of L328-IRS is estimated to be $\sim 0.012-0.023~{}M_{\sun}$ from its mass accretion rate converted from its outflow force (Lee et al., 2018). The outflow force is estimated using ALMA observations in the 12CO $J=2-1$ line and converted to the mass accretion rate, assuming an entrainment efficiency of 0.25, a ratio of mass loss and accretion rates of 0.1, and a wind velocity of $15~{}{\rm km~{}s^{-1}}$. Meanwhile, Lee et al. (2018) explained the C18O and 13CO $J=2-1$ emission in L328-IRS by the Keplerian rotation and suggested a central mass, $\sim 0.27$-$0.30~{}M_{\sun}$, significantly larger than the BD mass regime. In addition to the disregard of Keplerian-disk identification, numerical simulations predict that the disk mass can be comparable to the central stellar mass during the early phase with $M_{*}\lesssim 0.1~{}M_{\sun}$ (Machida et al., 2010, 2014). Such a massive disk may cause the rotational velocity to deviate from the Keplerian rotation determined only by $M_{*}$, preventing us from directly estimating $M_{*}$ in observations. It is thus crucial to verify whether the direct method with the Keplerian disk identification can be applied for protostars down to the $M_{*}\sim 0.1$-$M_{\sun}$ regime as is for the typical low mass protostars. ### 1.1 Target IRAS 16253-2429 The Class 0 protostar IRAS 16253-2429 (hereafter I16253) in the Ophiuchus star-forming region is a good target for the verification in that mass regime. I16253 has a bolometric luminosity of $L_{\rm bol}=0.16~{}L_{\sun}$ and a bolometric temperature of $T_{\rm bol}=42$ K (Ohashi et al., 2023). Its internal luminosity is estimated to be $L_{\rm int}\sim 0.08~{}L_{\sun}$ from the infrared luminosity of $L_{\rm IR}=0.046~{}L_{\sun}$ measured in $1.25-70~{}\micron$ and an empirical ratio of $L_{\rm int}/L_{\rm IR}\sim 1.7$ (Dunham et al., 2008). This protostar is thus classified as a very low luminosity object (VeLLO), which suggests that this protostar may eventually evolve into a very-low mass star or a BD. We observed this protostar as a part of the ALMA large program “Early Planet Formation in Embedded Disks” (eDisk) (see the overview paper by Ohashi et al., 2023). The main goal of the program is to reveal signs of planet formation in disks in the course of protostellar evolution. In addition to this main goal, I16253 was observed to investigate the protostellar evolution and the disk formation down to the BD mass regime. The central stellar mass of this target in previous works has been estimated to be a wide range of $\sim 0.02$ to $\sim 0.12~{}M_{\sun}$, depending on methods. Tobin et al. (2012), who observed I16253 in the N2H+ $J,F_{1},F=1,2,2-0,1,1$ line using the Combined Array for Research in Millimeter-wave Astronomy (CARMA) at a resolution of $\sim 9\arcsec$, estimated the stellar mass to be $0.1~{}M_{\sun}$ based on comparisons of a position-velocity (PV) diagram across the associated outflow on a 6000-au scale between the observations and a model of the rotating, collapsing envelope (UCM model; Ulrich, 1976; Cassen & Moosman, 1981). Yen et al. (2017) estimated the stellar mass to be $0.02^{+0.02}_{-0.01}~{}M_{\sun}$ based on C18O $J=2-1$ observations of the protostar using ALMA at an angular resolution of $\sim 1\arcsec$; they reproduced the observed PV diagram along the major and minor axes of the C18O envelope with a model envelope having a rotation conserving its specific angular momentum and a radial free-fall motion. Hsieh et al. (2019) used a similar model of the envelope and estimated the stellar mass to be $\sim 0.028~{}M_{\sun}$ by reproducing PV diagrams of 12CO and C18O $J=2-1$ obtained with ALMA at an angular resolution of $0\farcs 1-0\farcs 4$. Additionally, Hsieh et al. (2019) also suggested the stellar mass of $0.12~{}M_{\sun}$ with the assumption that the 12CO emission arises from a Keplerian disk; they found offsets of the 12CO emission, $\pm 7.8$ au away from the center, at two velocity channels of $V_{\rm LSR}-V_{\rm sys}=\pm 3.4~{}{\rm km~{}s^{-1}}$ by 2D Gaussian fitting. This mass derived on the assumption of Keplerian rotation can be different from the one derived on the assumption of free-fall motion, if the infall velocity is slower than the free-fall velocity, as reported in other protostars (Ohashi et al., 2014; Aso et al., 2015, 2017; Sai et al., 2022). These previous works show controversy about the central stellar mass of I16253 around the BD mass regime (i.e., $M_{*}<0.08~{}M_{\sun}$). Previous observations using the Walraven Photometer estimated the distance of the Ophiuchus star-forming region to be $125\pm 25$ pc (de Geus et al., 1989), while the same observation estimated the distance at a location of $(l,b)=(353.161\arcdeg,15.936\arcdeg)$ closest to I16253 $(353.2\arcdeg,16.5\arcdeg)$ to be 140 pc. This distance is consistent with an estimate using Hipparcos ($120-160$ pc; Knude & Hog, 1998) and a new estimate by Gaia DR2, $139^{+9}_{-10}$ pc, at a location of $(353.2\arcdeg,16.6\arcdeg)$ closest to I16253 (Zucker et al., 2020). In addition to the I16253 distance, the average distance of the whole Ophiuchus region was updated by Gaia DR2 to $144\pm 9$ pc (Zucker et al., 2019). We adopt the Gaia DR2 distance of 139 pc as the distance of I16253 in this paper. This work aims to estimate the central mass of the Class 0 protostar I16253 more accurately and precisely than the previous works and reveal the kinematic structures around the protostar. Our observations have $>2.5$ times better sensitivity in the CO isotopologue lines than that of previous observations in the same lines at similar angular and velocity resolutions to those of our observations. We aim to reveal kinematic structures, such as a rotating disk, an infalling envelope, and an outflow, using the molecular line emission detected at the higher signal to noise ratio. The rest of the paper has the following structure. Section 2 describes key points of the settings and data processing of our I16253 observations. Section 3 shows the observed results in the 1.3-mm dust continuum and the 12CO, 13CO, C18O, and SO lines. We analyze the asymmetry in the continuum image and radial profiles of the rotational velocity in Section 4. The rotational velocity is found to follow the Keplerian rotation, which allows us to directly estimate the central mass of I16253. We discuss kinetic structures around I16253, showing a schematic picture (Figure 18) of those structures, in Section 5. A summary of our results and interpretation is in Section 6. ## 2 Observations The details of the observing strategy, spectral setups, and data reduction process are presented in Ohashi et al. (2023)111The scripts used for reduction, including the self-calibration, can be found at https://github.com/jjtobin/edisk (Tobin et al., 2023).. Here we briefly describe the key points specific to I16253 and summarize the observational parameters in Table 1. We observed I16253 using ALMA in Cycle 8 with the antenna configuration of C-8 on 2021 October 5, 26, 27, and 28 (2019.1.00261.L). The total observing time with C-8 is $\sim 310$ min (5.2 hr), while the on-source observing time is $\sim 60$ min. The number of antennas with C-8 was 46, and the projected baseline length ranges from 52 to 10540 m. Any Gaussian component with FWHM$\gtrsim 2.8\arcsec$ ($\sim 400$ au) was resolved out by $\gtrsim 63\%$ with this shortest baseline length (Wilner & Welch, 1994). The phase center is $(\alpha_{\rm ICRS},\delta_{\rm ICRS})=(16^{\rm h}28^{\rm m}$21$\fs$6, $-24^{\circ}36\arcmin 23\farcs 4)$. To recover the more extended structure, we also observed I16253 using ALMA in Cycle 8 with the antenna configuration of C-5 on 2022 June 14 and 15 (2019.A.00034.S). The total observing time with C-5 is $\sim 170$ min (2.8 hr), while the on-source observing time is $\sim 30$ min. The number of antennas with C-5 was 43, and the projected baseline length ranges from 14 to 1290 m. The resolving out scale is $10\arcsec$ ($\sim 1400$ au) with this shortest baseline length. The phase center is the same as with C-8. The ALMA observations were set up to cover spectral windows including CO isotopologues ($J=2-1$), SO ($J_{N}=6_{5}-5_{4}$), and other molecular lines at Band 6. The spectral window for the 12CO ($J=2-1$) line has 3840 channels covering a 940 MHz bandwidth at an original frequency resolution of 224 kHz. Spectral windows for the 13CO, C18O ($J=2-1$), SO ($J_{N}=6_{5}-5_{4}$), and H2CO ($3_{21}-2_{20}$) lines have 960 channels covering a 59 MHz bandwidth at an original frequency resolution of 61 kHz. Spectral windows for the other lines have 3840 channels covering a 1.9 GHz bandwidth at an original frequency resolution of 488 kHz. Channels were binned to produce the velocity resolutions of 0.32, 0.17, 0.17, $0.17~{}{\rm km~{}s^{-1}}$, respectively, to make maps in the 13CO, C18O, SO, and H2CO lines. The velocity resolution for other lines is $1.34~{}{\rm km~{}s^{-1}}$ (Appendix A). Continuum maps were made using line-free channels of the spectral windows including two wide spectral windows with a $\sim 1.8$ GHz bandwidth centering at $\sim 220$ to $\sim 230$ GHz. The absolute flux density accuracy of ALMA is $\sim 10\%$ in this frequency band. All the imaging procedure was carried out with Common Astronomical Software Applications (CASA) (McMullin et al., 2007) version 6.2.1. The visibilities were Fourier transformed and cleaned with Briggs weighting and a robust parameter of $-2.0$, 0.0, 0.5, or 2.0, using the CASA task tclean at a pixel size of $0\farcs 02$. Continuum images adopt robust$=-2.0$, 0.0, and 2.0 to show the most compact, intermediate, and most extended components, respectively. The line images adopt robust$=0.5$ and 2.0 to show compact components and extended components, respectively. The line images do not have signal-to-noise (S/N) ratios high enough for robust$=-2$. The continuum images with robust$=-2$ and 2 are produced with tapering parameters of 3 and 1 M$\lambda$ ($\sim 0\farcs 06$ and $\sim 0\farcs 18$), respectively. The tapering parameter of 3 M$\lambda$ was selected to focus more on extended components with robust$=2$, while the tapering parameter of 1 M$\lambda$ was selected to increase the S/N ratio of the image with robust$=-2$. All the line images are produced with a tapering parameter of 2 M$\lambda$ ($\sim 0\farcs 09$) to increase the S/N ratio. The resultant angular resolution is $0\farcs 04-0\farcs 29$ for the continuum emission and $0\farcs 17-0\farcs 38$ for the line emission. We also performed self-calibration for the continuum data using tasks in CASA (tclean, gaincal, and applycal). While all the maps are primary- beam corrected in this paper with the primary beam size of $26\arcsec$, the root-mean-square noise levels of the line maps were measured in emission-free channels/areas before the primary-beam correction, which indicates the sensitivity achieved toward the phase center. Table 1: Summary of the parameters of our ALMA observations toward the Class 0 protostar I16253. Date | 2021 October 5, 26, 27, & 28 | 2022 June 14 & 15 ---|---|--- Projected baseline range | 52–10540 m | 14–1290 m Maximum Recoverable Scale | ${0\farcs 62}$ | ${2\farcs 9}$ Bandpass/flux calibrator | J1517-2422 | J1517-2422 Check source | J1650-2943 | J1650-2943 Phase calibrator | J1633-2557 | J1700-2610 | Continuum | 12CO $J=2-1$ | 13CO $J=2-1$ | C18O $J=2-1$ | SO $J_{N}=6_{5}-5_{4}$ Frequency (GHz) | 225 | 230.5380000 | 220.3986842 | 219.5603541 | 219.9494420 Freq./vel. width | $\sim 2$ GHz | $0.32~{}{\rm km~{}s^{-1}}$ | $0.17~{}{\rm km~{}s^{-1}}$ | $0.17~{}{\rm km~{}s^{-1}}$ | $0.17~{}{\rm km~{}s^{-1}}$ Beam (P.A.) | $0\farcs 29\times 0\farcs 24\ (86\arcdeg)^{a}$ | $0\farcs 35\times 0\farcs 25\ (76\arcdeg)^{a}$ | $0\farcs 37\times 0\farcs 26\ (77\arcdeg)^{a}$ | $0\farcs 38\times 0\farcs 27\ (78\arcdeg)^{a}$ | $0\farcs 37\times 0\farcs 26\ (78\arcdeg)^{a}$ noise rms | $23~{}{\rm\mu Jy~{}beam^{-1}}^{a}$ | $1.4~{}{\rm mJy~{}beam^{-1}}^{a}$ | $2.5~{}{\rm mJy~{}beam^{-1}}^{a}$ | $1.9~{}{\rm mJy~{}beam^{-1}}^{a}$ | $2.3~{}{\rm mJy~{}beam^{-1}}^{a}$ Beam (P.A.) | | $0\farcs 073\times 0\farcs 048\ (77\arcdeg)^{c}$ --- $0\farcs 043\times 0\farcs 034\ (73\arcdeg)^{d}$ $0\farcs 17\times 0\farcs 13\ (88\arcdeg)^{b}$ | $0\farcs 17\times 0\farcs 13\ (87\arcdeg)^{b}$ | $0\farcs 18\times 0\farcs 14\ (87\arcdeg)^{b}$ | $0\farcs 18\times 0\farcs 14\ (87\arcdeg)^{b}$ noise rms | | $22~{}{\rm\mu Jy~{}beam^{-1}}^{c}$ --- $100~{}{\rm\mu Jy~{}beam^{-1}}^{d}$ $1.8~{}{\rm mJy~{}beam^{-1}}^{b}$ | $3.0~{}{\rm mJy~{}beam^{-1}}^{b}$ | $2.1~{}{\rm mJy~{}beam^{-1}}^{b}$ | $2.5~{}{\rm mJy~{}beam^{-1}}^{b}$ Note. — The beam and noise rms values are shown for different robust parameters with superscripts of (a) Robust$=2$, (b) 0.5, (c) 0, and (d) $-2$. ## 3 Results ### 3.1 1.3 mm continuum Figure 1 shows continuum images with different robust and tapering parameters. Figure 1(a) shows that the continuum emission with the robust parameter of 0 traces a disk-like structure on a $\sim 40$ au scale. This emission appears more extended to the southeast than to the northwest from the emission peak. This asymmetry is investigated in detail in Section 4.1. Two dimensional (2D) Gaussian fitting to this robust$=0$ continuum image provides the central position of $(\alpha_{\rm ICRS},\ \delta_{\rm ICRS})=(16^{\rm h}28^{\rm m}21\fs 615,\ -24\arcdeg 36\arcmin 24\farcs 33$) with an uncertainty of (0.6 mas, 0.3 mas). The fitting, including the error bar calculation, used the CASA task imfit. The fitting result also provides the deconvolved size of $107\pm 2$ mas $\times$ $40\pm 2$ mas with a major axis in the direction of P.A. of $113\arcdeg\pm 1\arcdeg$. We adopt this direction as the major axis of the protostellar system, whici is perpendicular to the outflow direction, P.A.$=200\arcdeg-205\arcdeg$ (Yen et al., 2017). The aspect ratio corresponds to an inclination angle of ${\rm arccos}(40/107)\sim 68\arcdeg$, assuming a geometrically thin disk. The peak intensity and flux density derived from the Gaussian fitting are $5.10\pm 0.02~{}{\rm mJy~{}beam^{-1}}$ and $11.5\pm 0.2$ mJy, respectively. The flux densities integrated over $0\farcs 4\times 0\farcs 4$ (the compact emission) and $10\arcsec\times 10\arcsec$ (the maximum recoverable scale of our ALMA observations) regions are $12.4\pm 0.3$ and $25\pm 2$ mJy, respectively. The flux density $F_{\nu}$ can be converted to a lower limit of gas mass, assuming that the emission is optically thin: $M_{\rm gas}=d^{2}F_{\nu}/\kappa_{\nu}B_{\nu}(T)$, where $d$, $\kappa_{\nu}$, and $B_{\nu}$ are the target distance, dust mass opacity, and the Planck function with an average temperature $T$, respectively. For this conversion, two different temperatures are adopted: $T=20$ and 27 K. $T=20$ K is derived from the relation between 850-$\micron$ continuum fluxes and model gas masses for 44 young stellar objects (Andrews & Williams, 2005). $T=27$ K is calculated from the empirical relation, $T=43~{}{\rm K}~{}(L_{\rm bol}/1~{}L_{\sun})^{0.25}$, from Tobin et al. (2020). With these temperatures, $d=139$ pc, $\kappa_{\nu}=2.3~{}{\rm cm}^{2}~{}{\rm g}^{-1}$ (Beckwith et al., 1990), and a gas-to-dust mass ratio of 100, $F_{\nu}\simeq 12$ mJy corresponds to a gas mass of $M_{\rm gas}\simeq(1.4-2.1)\times 10^{-3}~{}M_{\sun}$. Figure 1(b) adopts the robust parameter of $-2$ and the 3-M$\lambda$ taper to focus on the central region with a reasonable S/N ratio. The plotted spatial range is half of Figure 1(a). This image shows an extension or a secondary peak to the southeast, separated from the central peak by $\sim 0\farcs 1$ roughly along the major axis. This is consistent with the extension seen in the image with robust$=0$ (Figure 1a). Figure 1(c), adopting the robust parameter of 2 and the 1-M$\lambda$ taper, shows the extended emission over $\sim 600$ au from the continuum emission peak to the northwest, as well as the compact ($\sim 100$ au) emission. This large-scale structure is also reported in Yen et al. (2017). Figure 1: Images of the 1.3 mm continuum emission produced with (a) the robust parameter of 0.0 and no taper, (b) the robust parameter of $-2.0$ and a 3 M$\lambda$ ($\sim 0\farcs 06$) taper, (c) the robust parameter of 2.0 and a 1 M$\lambda$ ($\sim 0\farcs 18$) taper. The brightness temperature $T_{b}$ is calculated from the Rayleigh-Jeans approximation. The contour levels are $3,6,12,24,48,96,192\sigma$, where $1\sigma$ is (a) 0.022, (b) 0.100, (c) $0.023~{}{\rm mJy~{}beam^{-1}}$ (Table 1). The filled ellipses at the bottom- left corners denote the synthesized beams: (a) $0\farcs 073\times 0\farcs 048$ ($77\arcdeg$), (b) $0\farcs 043\times 0\farcs 034$ ($73\arcdeg$), and (c) $0\farcs 29\times 0\farcs 24$ ($86\arcdeg$). The diagonal lines denote the major and minor axes of the emission in the image in panel (a): P.A.$=113\arcdeg$ and $23\arcdeg$. The square in panel (a) shows the plotted range of panel (b). ### 3.2 12CO $J=2-1$ Figure 2(a) shows the integrated intensity (moment 0) and the mean velocity (moment 1) maps in the 12CO emission produced with the robust parameter of 2. Positive and negative intensities are integrated for the moment 0 map, while intensities above the $3\sigma$ level are integrated for the moment 1 map. The noise level of each moment 0 map is calculated by the noise propagation from the noise level of the data cube (Table 1) used to make the moment 0 map: $\sigma_{\rm mom0}=\sigma_{\rm cube}dv\sqrt{N_{\rm ch}}$, where $\sigma_{\rm mom0}$, $\sigma_{\rm cube}$, $dv$, and $N_{\rm ch}$ are the noise level of the moment 0 map, the noise level of the cube data, the velocity width per channel (Table 1), and the number of integrated channels, respectively. The moment 0 and 1 maps of the other lines in this paper are also made in the same method. The 12CO emission traces a clear bipolar outflow whose axis is perpendicular to the major axis of the disk-like structure. An additional eastern component results from a large-scale structure around the systemic velocity, which is velocity resolved out by the interferometric observation. While the emission in the northern outflow lobe is overall blueshifted, redshifted emission can also be found on the northern side near the outflow axis. Similarly, the southern lobe is mostly redshifted but partly blueshifted. This is expected if a given outflow lobe crosses the plane of the sky that contains the central source, with the outflow axis close to the plane of sky (Cabrit, 1989). A part of the lobe in the front of the plane of the sky appears blue-shifted, while the other part of the lobe behind the plane of the sky appears red-shifted. Using a radially-expanding parabolic outflow model, Yen et al. (2017) estimated the outflow shape and inclination by fitting a moment 0 map and a PV diagram along the outflow axis of the 12CO emission observed at an angular resolution of $\sim 1\arcsec$. Their best model provides an inclination angle of $60\arcdeg-65\arcdeg$ ($0\arcdeg$ means pole-on) and an opening angle of $60\arcdeg-70\arcdeg$, calculated from the parabolic shape at the distance from the protostar along the outflow axis of $z\sim 5\arcsec-10\arcsec$. This inclination is consistent with that of the disk-like structure detected in the dust continuum emission (Section 4.1). Thus, their model is consistent with our observations which reveal blue- and red-shifted outflow components to the northeast and southwest, respectively. Figure 2: Moment 0 (integrated intensity) and 1 (mean velocity) maps in the 12CO $J=2-1$ line. The contour maps show the moment 0 map, while the color images show the moment 1 map. (a) Moment maps with the robust parameter of 2.0, integrated from $V_{\rm LSR}=-6.0$ to $14.0~{}{\rm km~{}s^{-1}}$. (b) Zoom-in view of the moment maps with the robust parameter of 0.5, integrated from $V_{\rm LSR}=-6.0$ to $14.0~{}{\rm km~{}s^{-1}}$. The contour levels are in (a) $12\sigma$ steps from $12\sigma$ with $1\sigma=3.2~{}{\rm mJy~{}beam^{-1}}~{}{\rm km~{}s^{-1}}$ and (b) $6\sigma$ steps from $6\sigma$ with $1\sigma=3.5~{}{\rm mJy~{}beam^{-1}}~{}{\rm km~{}s^{-1}}$. The filled ellipse at the bottom-left corner shows the synthesized beam: (a) $0\farcs 35\times 0\farcs 25$ ($76\arcdeg$) and (b) $0\farcs 17\times 0\farcs 13$ ($88\arcdeg$). The diagonal lines are the major and minor axes of the continuum emission, P.A.$=113\arcdeg$ and $23\arcdeg$. The green contour in panel (b) shows the 5$\sigma$ level of the continuum emission (Figure 1a). Figure 2(b) shows a zoom-in view of the moment 0 and 1 maps with a different robust parameter of 0.5. The green contour shows the 5$\sigma$ level of the continuum emission (Figure 1a) for comparison. The moment 0 map shows local peaks within $\sim 0\farcs 4$ from the protostellar position, which is reported in Hsieh et al. (2019) as a quadruple peak. In addition to the velocity gradient along the outflow axis on a $1\arcsec$ scale, this figure also shows a velocity gradient perpendicular to the outflow axis near the midplane on a $0\farcs 3$ scale, i.e., transition from a blueshifted part at P.A.$=113\arcdeg$ to a redshifted part at P.A.$=-67\arcdeg$. This velocity gradient is expected for rotating gas and analyzied in more detail in Section 4.2. Figure 4 shows a position-velocity (PV) diagram of the 12CO emission along the outflow axis from the robust parameter 2 images and a width of $0\farcs 6$ ($\sim 2\times$ beam size). The width allows us to increase the S/N ratio but still focus on the velocity structure on the outflow axis. Because of this width, the noise level of this PV diagram is better than the one in Table 1. This figure also shows that both northern and southern lobes have both blue- and red-shifted parts. The outflow velocity is $|V-V_{\rm sys}|\sim 1.5-2.5~{}{\rm km~{}s^{-1}}$ at any position except for the central $1\arcsec-2\arcsec$ region. The central component appears not to be a part of the outflow because of its $\sim 3$ times larger line widths. The outer ($>2\arcsec$) emission shows several ($\sim 5-9$) local peaks or extensions within $10\arcsec$ ($\sim 1400$ au) in each lobe, based on visual inspection (short lines in Figure 4). This may be suggestive of episodic mass ejection (Lee, 2020). The interval angular scale, $10\arcsec/7=1.4\arcsec$, is significantly larger than the beam size of $0\farcs 35$. With the inclination angle of $65\arcdeg$ (Yen et al., 2017), the deprojected length $l$ and the deprojected velocity $v$ of the outflow can be calculated as $l=l^{\prime}/\cos i$ and $v=v^{\prime}/\sin i$, respectively, where $l^{\prime}$ and $v^{\prime}$ are the projected length and velocity. The interval $t$ of the mass ejection is a time required for material to move from a peak to the next peak. This can be calculated with the number of peaks $n$ within the length: $t=l^{\prime}/n/v^{\prime}$. Then, the values measured above provide a typical interval of $1400~{}{\rm au}/\cos 65\arcdeg/(5\ {\rm to}\ 9)/(1.5\ {\rm to}\ 2.5~{}{\rm km~{}s^{-1}}/\sin 65\arcdeg)=600\ {\rm to}\ 2000$ year. The episodic mass ejection is thought to be caused by episodic accretion bursts. The interval of accretion bursts is investigated in statistical studies using FUor objects. Park et al. (2021) reported accretion- burst intervals of $500-2300$ yr in a statistical study of FUor outbursts based on the fraction of outbursting sample in a set of NEOWISE observations. The fraction of variable sources is similar between VeLLOs like I16253 and more luminous objects in their sample. The uncertainty comes from the number of outbursts ranging from 2 to 9 and the total number of sample protostars ranging from 735 to 1059. This accretion-burst interval is consistent with the mass-ejection interval of I16253, implying that accretion bursts likely occurred in I16253. Figure 3: Figure 4: Position-velocity diagram in the 12CO $J=2-1$ line along the outflow axis (P.A.$=23\arcdeg$) with a width of $0\farcs 6$. The positive offset corresponds to the northern side. The contour levels are in $6\sigma$ steps from $6\sigma$, where $1\sigma$ is $0.85~{}{\rm mJy~{}beam^{-1}}$. The dashed lines denote the systemic velocity ($V_{\rm LSR}=4~{}{\rm km~{}s^{-1}}$) and the protostellar position. The orange short lines denote the position of local peaks or extensions visually identified. This PV diagram shows a strong emission component at offsets of $<-10\arcsec$. This component is not confined around the outflow axis but extended over the redshifted lobe. This may imply that there could be more material on the southern side than on the northern side, although our observation did not target such a large spatial scale. ### 3.3 C18O and 13CO $J=2-1$ Figure 5 shows the moment 0 and 1 maps in the C18O and 13CO lines in robust$=0.5$ and 2. The two lines show an overall structure extended along the major axis (P.A.$=113\arcdeg$) and a velocity gradient primarily along the same direction. The integrated intensity maps with robust$=0.5$ show double peaks with a separation of $\sim 0\farcs 2-0\farcs 3$, and the integrated intensity is stronger at the eastern, blueshifted peaks than at the western, redshifted peaks, by $3\sigma$ in both lines. The integrated intensity maps with robust$=2$ show more clearly that the eastern emission is stronger than the western emission. The detected emission size is larger in the C18O line than in the 13CO line with both robust parameters. This is because the 13CO emission tends to be optically thicker and more extended than the C18O emission, and thus the 13CO emission is resolved out more severely around the systemic velocity. Figure 5: Moment 0 and 1 maps in the C18O $J=2-1$ line emission and the 13CO $J=2-1$ line emission. The contour maps show the moment 0 map, while the color images show the moment 1 map. The contour levels are in $3\sigma$ steps in the robust$=0.5$ images and $6\sigma$ steps in the robust$=2$ images starting from $3\sigma$. The C18O emission is integrated from 0.8 to $7.1~{}{\rm km~{}s^{-1}}$. These moment 0 maps have $1\sigma$ of 2.2 (robust$=0.5$) and $2.0~{}{\rm mJy~{}beam^{-1}}~{}{\rm km~{}s^{-1}}$ (robust$=2$). The 13CO emission is integrated from 0.7 to $7.3~{}{\rm km~{}s^{-1}}$. These moment 0 maps have $1\sigma$ of 2.8 (robust$=0.5$) and $2.7~{}{\rm mJy~{}beam^{-1}}~{}{\rm km~{}s^{-1}}$ (robust$=2$). The beam sizes are (a) $0\farcs 18\times 0\farcs 14$ ($87\arcdeg$), (b) $0\farcs 38\times 0\farcs 27$ ($78\arcdeg$) (c) $0\farcs 17\times 0\farcs 13$ ($87\arcdeg$), and (d) $0\farcs 37\times 0\farcs 26$ ($77\arcdeg$). The diagonal lines are the major and minor axes of the continuum emission, P.A.$=113\arcdeg$ and $23\arcdeg$. ### 3.4 SO $J_{N}=6_{5}-5_{4}$ Figure 6(a) shows the integrated intensity and mean velocity maps in the SO line produced with robust$=0.5$ to focus on the central compact emission. The SO emission shows a double-peaked structure with a separation of $\sim 0\farcs 2-0\farcs 3$ and the velocity gradient along the major axis, like the C18O and 13CO lines. On the other hand, the mean velocity of the SO emission is larger outside the peaks than inside the peaks, unlike the C18O and 13CO lines. Figure 6(b) shows the maps produced with robust$=2$ to focus on extended emission. The central double peaks are not spatially resolved in this map. The velocity structure in the central $1\arcsec$ region is overall the same as that in Figure 6(a): eastern blueshifted emission and western redshifted emission. In addition, this figure shows an extended structure to the east from the central protostar. This structure is extended ($\sim 3\arcsec$) beyond the C18O and 13CO emission ($\sim 1\arcsec$) in Figure 5. An inner part ($\lesssim 1\arcsec$ from the protostellar position) of this extended structure shows velocities blueshifted from the systemic velocity, $V_{\rm LSR}=4~{}{\rm km~{}s^{-1}}$, which is consistent with the C18O and 13CO velocities. In contrast, the outer part ($\gtrsim 1\arcsec$) shows slightly redshifted velocities ($V_{\rm LSR}-V_{\rm sys}\sim 0.3~{}{\rm km~{}s^{-1}}$). This redshifted velocity on the eastern side is different from the velocity gradient seen in the C18O and 13CO lines and also different from the velocity in the eastern cavity wall of the 12CO outflow (Figure 2a). This SO component is discussed in more detail in Section 5.2. Figure 6: Moment 0 and 1 maps in the SO $J_{N}=6_{5}-5_{4}$ line emission. The contour maps show the moment 0 map, while the color images show the moment 1 map. (a) The robust parameter is 0.5. The emission is integrated from 1.7 to $6.3~{}{\rm km~{}s^{-1}}$. The contour levels are in $3\sigma$ steps from $3\sigma$ with $1\sigma=2.3~{}{\rm mJy~{}beam^{-1}}~{}{\rm km~{}s^{-1}}$. The beam size is $0\farcs 18\times 0\farcs 14$ ($87\arcdeg$). (b) The robust parameter is 2. The emission is integrated from 1.7 to $6.3~{}{\rm km~{}s^{-1}}$. The contour levels are in $3\sigma$ steps from $3\sigma$ until $15\sigma$ and then in $10\sigma$ steps with $1\sigma=2.1~{}{\rm mJy~{}beam^{-1}}~{}{\rm km~{}s^{-1}}$. The beam size is $0\farcs 37\times 0\farcs 26$ ($78\arcdeg$). The diagonal lines are the major and minor axes of the continuum emission, P.A.$=113\arcdeg$ and $23\arcdeg$. ## 4 Analysis ### 4.1 Non-axisymmetry of the continuum image The 1.3 mm continuum emission with robust$=0$ (Figure 1a) is more extended in the southeastern side ($\sim 30$ au) than in the northwestern side ($\sim 20$ au). We verify that the extended components are real by comparing images made with various robust parameters. To extract the extended component, an axisymmetric model of the dust continuum emission is constructed in this subsection. The model is made from a radial profile of intensities in the unit of Jy pixel-1, before beam convolution, set with free parameters of $I(r=0),I(dr),I(2dr),...,I(11dr)$. The grid separation $dr$ is half of the beam minor-axis. The largest radius $11dr=0\farcs 26$ covers the entire continuum emission with robust$=0$. Then, the circular 2D intensity distribution is projected by the inclination factor of $\cos i$ in the direction of the minor axis and rotated by the position angle $pa$; these two angles are also free parameters. This elliptical 2D intensity in the unit of Jy pixel-1 is convolved with the observational beam to make the model image in the unit of ${\rm Jy~{}beam^{-1}}$. Hence, this model does not assume any shape, such as a Gaussian function but only the axisymmetry. This modeling method follows the one in Aso et al. (2021). The central position is not a free parameter but fixed at the observed peak position, $(\alpha_{\rm ICRS},\ \delta_{\rm ICRS})=(16^{\rm h}28^{\rm m}21\fs 6153,\ -24\arcdeg 36\arcmin 23\farcs 325)$, which coincides with the Gaussian center derived in Section 3.1. The best-fit parameters are obtained by minimizing $\chi^{2}$ between the observed and model images, divided by the number of pixels in the beam, through the Markov chain Monte Carlo (MCMC) methods with the public Python package ptemcee. The numbers of free parameters, steps, and walkers per parameter are 14, 8000, and 16, respectively. The first half (4000) steps were removed for the burn-in. The derived best-fit angles are $i=64.1\arcdeg\pm 0.5\arcdeg$ and $pa=114.3\arcdeg\pm 0.6\arcdeg$, and Figure 8 shows the posterior distribution of these two angles produced by the MCMC fitting. The uncertainties of these angles are derived as the 16 and 84 percentiles. Figure 7: Figure 8: Corner plot of the MCMC fitting to the continuum image. Two angles, $i$ and $pa$, are plotted among the 14 free parameters. The dashed lines show the 16, 50, and 84 percentiles. Figure 9(a) shows the comparison between the best-fit axisymmetric model and the observed image. The overall structure in the observed image is reproduced, but the model emission is weaker than the observed emission in the southeast, while it is stronger in the northwest, as expected. The contour map in Figure 9(b) shows the residual image after subtracting the model from the observed image with robust$=0$, while the color map shows the observed continuum image with robust$=-2$ (same as Figure 1c). The strong residual is located in the southeast. This residual overlaps with the extension in the robust$=-2$ image, supporting that this excess from the symmetric component is not due to the specific robust parameter. The direction of the residual and extension coincides with the direction of the SO extended structure and the stronger peak on the eastern side in the C18O and 13CO moment 0 maps. Figure 9: (a) Axisymmetric model (red) fitted to the observed 1.3 mm continuum emission (black) with the robust parameter of 0.0. The contour levels are the same as Figure 1(b): 3,6,12,24,48,96,192$\sigma$. (b) Residual (contour map) between the model and observation in panel (a), overlaid on the robust$=-2$ continuum image (color map; Figure 1c). The contour levels are the same: $3,6,12,24\sigma$. ### 4.2 Keplerian rotation in the 12CO emission Previous observational studies have not yet identified any part of gas motion in I16253 having a radial profile of the Keplerian rotation but assumed that the observed gas follows Keplerian rotation. The Keplerian disk is crucial to estimate the central stellar mass in the protostellar phase and verify whether I16253 is a proto-BD or a very low-mass protostar. To tackle this problem, we analyze the position-velocity (PV) diagrams along the major axis. Figure 10(a) shows the PV diagrams in the 12CO (blue contours), C18O (red contours), and 13CO (color) lines. The C18O and 13CO lines show a typical shape of an infalling plus rotating motion: a distorted diamond shape with emission in all the four quadrants, as reported in Hsieh et al. (2019). In contrast, the 12CO emission is concentrated on the first (upper right) and third (lower left) quadrants (the western redshifted emission and the eastern blueshifted emission), showing a clear velocity gradient, in $\pm(0\farcs 1-0\farcs 2)$ at $|V-V_{\rm sys}|\gtrsim 2~{}{\rm km~{}s^{-1}}$. Figure 10(b) shows the PV diagrams in the three lines along the minor axis. The C18O and 13CO emission can be seen in all four quadrants in the minor-axis PV diagram with a diamond shape, which is typical for the infall motion (e.g., Ruíz-Rodríguez et al., 2022). The 12CO emission is mainly concentrated around the center, with additional components in the second and fourth quadrants (the northern blueshifted emission and the southern redshifted emission). While the additional components show a velocity gradient due to the outflow motion, the main component shows no velocity gradient along the minor axis. These PV diagrams suggest that the 12CO line traces a purely rotating, i.e., Keplerian disk, while the C18O and 13CO lines trace the infalling rotating envelope previously identified. This difference among individual molecular lines can be understood by the different strengths of emission and the different missing fluxes in the interferometric observations. The 12CO emission is strong enough to be detected even in the expected compact disk ($r\sim 0\farcs 1-0\farcs 2$ in the PV diagram as well as in the continuum image). The 13CO and C18O emission is not detected sufficiently in this compact region at the high velocities where the 12CO emission is detected. The nondetection in the 13CO and C18O lines suggests the 12CO emission is not optically thick at the high velocities. In contrast to the high velocities, the 12CO emission is more extended and strongly resolved out at the low velocities where the 13CO and C18O emission traces the envelope. Figure 10: Position-velocity diagrams in the 12CO (blue; robust$=0.5$), C18O (red; robust$=2.0$), and 13CO (color; robust$=2.0$) $J=2-1$ lines along the (a) major and (b) minor axes. The positive offset corresponds to the western and northern sides in panels (a) and (b), respectively. The contour levels are in $6\sigma$ steps from $6\sigma$, where $1\sigma$ is $1.4~{}{\rm mJy~{}beam^{-1}}$ for 12CO, $2.1~{}{\rm mJy~{}beam^{-1}}$ for C18O, and $3.0~{}{\rm mJy~{}beam^{-1}}$ for 13CO. The beam sizes in these diagrams are $\sim 0\farcs 15$ for 12CO, $\sim 0\farcs 33$ for C18O, and $\sim 0\farcs 32$ for 13CO (Table 1). In order to verify whether the 12CO line traces the Keplerian rotation, we first find the emission ridge (e.g., Yen et al., 2013; Aso et al., 2015; Sai et al., 2020) and edge (e.g., Seifried et al., 2016; Alves et al., 2017) in the major-axis PV diagram and fit them with power-law functions. The ridge is a center of the emission along the positional axis at each velocity, while the edge is an outer boundary of the emission along the positional axis at each velocity. The analysis and fitting process was performed through the Python open package pvanalysis in Spectral Line Analysis/Modeling (SLAM; Aso & Sai, 2023)222https://github.com/jinshisai/SLAM, and the detail is described in the overview paper of the eDisk project (Ohashi et al., 2023). We thus mention here the settings specific to the case of I16253. The ridge point is defined as the intensity-weighted mean position $x_{m}$ at each velocity $v$: $x_{m}(v)=\int I(x,v)xdx/\int I(x,v)dx$. $x_{m}$ is calculated using the intensities above the $5\sigma$ level along the positional axis at each velocity in the range of $|V-V_{\rm sys}|>2~{}{\rm km~{}s^{-1}}$. This velocity range is selected because the emission is located in the first and third quadrants in this velocity range. Similarly, using the 1D intensity profile, the edge point is defined, at each velocity in the same velocity range, as the outer position where the intensity is at the $5\sigma$ level. In addition, the tool uses velocities higher than or equal to the velocity at which the derived edge/ridge radius is largest so that the relation between radius and velocity can be consistent with spin-up rotation. The ridge and edge tend to under- and overestimate, respectively, the radius at a given velocity (Maret et al., 2020, the ridge and edge here correspond to their “centroid” and “first emission contour”, respectively), and thus we adopt the former and the latter as the lower and upper limits of radius in this paper. Figure 11 shows the estimated ridge and edge points overlaid on the PV diagram in the linear and logarithmic scales. The logarithmic diagram is made by avaraging the emission in the first and third quadrants of the linear diagram. The edge and ridge radii are separately fitted with a power-law relation between either edge or ridge radius $R$ and the velocity $V$: $\displaystyle V=V_{b}\left(\frac{R}{R_{b}}\right)^{-p}+V_{\rm sys},$ (1) $\displaystyle p=p_{\rm in}\ {\rm if}\ R<R_{b}\ {\rm else}\ p_{\rm in}+dp.$ (2) This fitting uses the MCMC method with the tool of emcee, where the numbers of walker per free parameter, burn-in steps, adopted steps are 16, 2000, and 1000, respectively. The error bar of each free parameter is defined as the 16 and 84 percentiles. First, we adopt a single power-law function, where the free parameters are the power-law index $p$ and the radius $R_{b}$ at a fixed middle velocity $V_{b}$. $dp=0$ and $V_{\rm sys}=4.0~{}{\rm km~{}s^{-1}}$ are fixed. The fixed parameter $V_{b}$ is an arbitrary reference velocity where $R_{b}$ is derived from the fitting. The central stellar mass $M_{*}$ is calculated from the radius with the fixed inclination angle $i=65\arcdeg$ as $M_{*}=R(V)(V/\sin i)^{2}/G$, where $R(V)$ is calculated from Equation (1) as a function of $V$. The calculated $M_{*}$ can thus depend on $V$ if $p\neq 0.5$. The uncertainty of $M_{*}$ is calculated through the error propagation. The inclination angle is adopted from an outflow model ($60\arcdeg-65\arcdeg$; Yen et al., 2017) and our continuum analysis ($64.3\arcdeg$ in Section 4.1). Even with $i=60\arcdeg$, the calculated stellar mass is only 10% higher. The best-fit parameters are $p=0.5$ and $M_{*}=0.09~{}M_{\sun}$ with the ridge points, while they are $p=0.67$ and $M_{*}=0.3-0.4~{}M_{\sun}$ with the edge points. The range of edge $M_{*}$, $0.3-0.4~{}M_{\sun}$ is because $p_{\rm in}\neq 0.5$. This range is wider than the statistical uncertainty. The fitting results are summarized in Table 2 including statistical uncertainties. The uncertainty of the distance adds a relative uncertainty of $\sim 7\%$. The best-fit functions are plotted in Figure 11. Figure 11(a) shows that the best- fit functions trace the ridge and edge of the observed PV diagram. Figure 11(b) clearly shows that the ridge and edge indices are close to each other, as well as to that of Keplerian rotation $(p=0.5)$. The ridge and edge points were also fitted with a double power-law function. The best-fit parameters are summarized in Table 2. The double power-law fitting yields similar power-law indices $p=0.5-0.6$ for the most part of the fitted velocity range, except for only the lowest velocity channel. These results indicate that the 12CO line traces the Keplerian disk around I16253. Figure 11: Edge and ridge points, estimated for each velocity channel, overlaid on the 12CO major-axis PV diagram (robust$=0.5$) in the (a) linear and (b) logarithmic scales. The positive radius corresponds to the western side. The velocity is the relative velocity to the systemic velocity of $V_{\rm LSR}=4~{}{\rm km~{}s^{-1}}$. The white curves/lines are the best-fit power-law functions (Section 4.2). The contour levels are in the $3\sigma$ steps from $3\sigma$, where $1\sigma$ is $1.4~{}{\rm mJy~{}beam^{-1}}$. Table 2: Power-law fitting to the edge and ridge points in the major-axis 12CO PV diagram. | Single power ---|--- | $R_{b}$ (au) | $V_{b}$ (${\rm km~{}s^{-1}}$) | $p_{\rm in}$ | $dp$ | $M_{*}(M_{\sun})$ Edge | $24.7\pm 0.7$ | 3.18 (fixed) | $0.67\pm 0.04$ | 0 (fixed) | $0.34\pm 0.01-0.43\pm 0.02$ Ridge | $6.2\pm 0.7$ | 3.20 (fixed) | $0.5\pm 0.1$ | 0 (fixed) | $0.09\pm 0.02$ | Double power | $R_{b}$ (au) | $V_{b}$ (${\rm km~{}s^{-1}}$) | $p_{\rm in}$ | $dp$ | $M_{*}(M_{\sun})$ Edge | $45\pm 10$ | $2.2\pm 0.5$ | $0.63\pm 0.07$ | $2\pm 4$ | $0.3\pm 0.2-0.4\pm 0.2$ Ridge | $13\pm 2$ | $2.3\pm 0.2$ | $0.5\pm 0.1$ | $4\pm 4$ | $0.09\pm 0.04$ Note. — The uncertainties of $R_{b}$ and $M_{b}$ in this table are before the uncertainty of the distance, $\sim 7\%$, is incorpolated. By identifying the Keplerian rotation, we have estimated the central stellar mass to be $M_{*}=0.09-0.34~{}M_{\sun}$ directly (kinematically) for the first time. Even with the lower limit, I16253 already has sufficient mass to evolve into a low-mass star, beyond the brown-dwarf mass regime ($M_{*}>0.08~{}M_{\sun}$). Maret et al. (2020) demonstrated that the central stellar masses derived from the ridge and edge points under- and overestimate the actual stellar mass by $\sim 30\%$ and $\sim 100\%$ in the case where the beam size is a few tens percent larger than the observed disk size. The reason why the ridge underestimates the stellar mass is that a part of the gas inside the Keplerian radius has the same line-of-sight velocity as the gas at the Keplerian radius, and this inner emission shifts the ridge inward. This effect is also discussed quantitatively by Aso et al. (2015). The reason why the edge overestimates the stellar mass is that the observational beam causes emission outside the outermost, i.e., Keplerian radius, and this outer emission shifts the edge outward. The condition in Maret et al. (2020) is applicable to our case (the disk radius of I16253 is discussed in Section 5.1). Then, the central stellar mass is likely within $M_{*}=0.09\times 1.3$ to $0.34/2~{}M_{\sun}=0.12$ to $0.17~{}M_{\sun}$. This mass is consistent with a value derived from the 12CO $J=2-1$ emission at two velocity channels by assuming Keplerian rotation ($\sim 0.12~{}M_{\sun}$; Hsieh et al., 2019). ### 4.3 Linear velocity gradient in the SO emission Figure 13 shows the PV diagram in the SO emission along the major axis. The SO emission is detected within $|V-V_{\rm sys}|\sim\pm 2~{}{\rm km~{}s^{-1}}$, which is similar to the velocity range where the 13CO and C18O emission is present (see Figure 14 for the comparison of the SO and C18O emission). The shape of the SO PV diagram is represented by a linear velocity gradient from the eastern blueshifted emission to the western redshifted emission in the velocity range of $|V-V_{\rm sys}|<1.3~{}{\rm km~{}s^{-1}}$, unlike the CO isotopologues. The SO emission outside this velocity range appears to overlap a component traced by the CO isotopologue emissions (see Figure 14 for the comparison), although it is difficult to discuss it given the limited number of the channels at the common velocity range. The linear velocity gradient can also be seen in the moment 1 map (Figure 6a), where the high velocities are located at the outermost parts along the major axis. To evaluate the linear velocity gradient quantitatively, the ridge points are found at each velocity, as plotted in Figure 13, and fitted with a linear function of radius $r$ passing $(r,v)=(0,0)$, using emcee with the same condition as that in Section 4.2. The best-fit gradient is estimated to be $g=0.056\pm 0.005~{}{\rm km~{}s^{-1}}~{}{\rm au}^{-1}$, after the correction for the inclination angle of $i=65\arcdeg$. This value is slightly smaller than a previous result of $0.082\pm 0.004~{}{\rm km~{}s^{-1}}~{}{\rm au}^{-1}$ ($d=139$ pc) by Yen et al. (2017). This difference is mainly due to the different angular resolutions, $\sim 0\farcs 2$ in the present observations and $\sim 1\arcsec$ in Yen et al. (2017). Their result may thus include SO emission at larger radii and slower velocities. The SO emission is located within the velocities of $|V-V_{\rm sys}|<2~{}{\rm km~{}s^{-1}}$, which are lower than those showing the Keplerian rotation in the 12CO PV diagram (Figure 11), implying that the SO emission traces a different part from the 12CO emission. The SO velocity structure and the velocity gradient are discussed in more detail in Section 5.1 along with the infalling and rotating motions traced in the C18O and 13CO emission. Figure 12: Figure 13: Ridge points, found for each velocity, overlaid on the SO PV diagram (robust$=2$) along the major axis in the linear scale. The positive radius corresponds to the western side. The velocity is relative to the systemic velocity of $V_{\rm LSR}=4~{}{\rm km~{}s^{-1}}$. The white line is the best linear function (Section 4.3). The contour levels are in the $3\sigma$ steps from $3\sigma$, where $1\sigma$ is $2.3~{}{\rm mJy~{}beam^{-1}}$. ## 5 Discussion ### 5.1 Disk, Ring, and Envelope We suggest a picture of the I16253 system based on our results and the previous results. Hsieh et al. (2019) estimated the specific angular momentum of the envelope to be $j\sim 45~{}{\rm km~{}s^{-1}}{\rm au}$ ($d=139~{}pc$), by reproducing the PV diagrams along the major and minor axes in the C18O and 12CO lines with an infalling and rotating envelope model. When the specific angular momentum and the central stellar mass are given, the velocity field of a protostellar envelope can be predicted by the UCM envelope model (Ulrich, 1976; Cassen & Moosman, 1981), where the velocity field consists of ballistic, parabolic flows from an outer boundary in the rigid-body rotation. Figure 14 compares the maximum line-of-sight velocity of the UCM envelope model, as well as that of the Keplerian disk model, at each position and the observed PV diagrams in the C18O, SO, and 12CO lines along the major and minor axes. The model parameters are the specific angular momentum of $j\sim 45~{}{\rm km~{}s^{-1}}{\rm au}$ and the central stellar mass of $M_{*}\sim 0.14~{}M_{\sun}$ (middle of $0.12-0.17~{}M_{\sun}$ in Section 4.2). With the specific angular momentum and the central stellar mass, the centrifugal radius is calculated to be $R_{c}=j^{2}/GM_{*}\sim 16$ au ($0\farcs 12$); $M_{*}=0.12-0.17~{}M_{\sun}$ corresponds to $R_{c}=13-19$ au. $R_{c}=16$ au is used as the disk radius to draw the model curves in Figure 14. The C18O PV diagrams are consistent with the model maximum velocity both in the major and minor axes in the velocity range lower than the disk velocities, $|V-V_{\rm sys}|<2~{}{\rm km~{}s^{-1}}$. The obtained $R_{c}$ is close to the radius of the continuum emission (the red $6\sigma$ contour in Figure 9a), except for the southeastern extension, supporting that this is the disk radius. This is also consistent with a relation suggested by Aso & Machida (2020) that the Gaussian deconvolved radius of the 1.3 mm continuum emission ($\sim 7.4$ au for I16253; Section 3.1) is $\sim 0.5$ as large as the disk radius for an evolutionary phase from $M_{*}\sim 0.1$ to $0.4~{}M_{\sun}$; this relation is based on their synthetic observations of a magnetohydrodynamics simulation of protostellar evolution. Figure 14: Comparison between the predicted maximum line-of-sight velocities (white dashed and dotted curves) and the observed PV diagrams (C18O in red, SO in green, and 12CO in blue). (a) PV diagram along the major axis. The dashed curves represent the UCM envelope model with a central stellar mass of $M_{*}=0.14~{}M_{\sun}$ (Section 4.2) and the specific angular momentum $j=45~{}{\rm km~{}s^{-1}}~{}{\rm au}$ (Hsieh et al., 2019). The envelope maximum velocity is calculated only up to $|V-V_{\rm sys}|=2~{}{\rm km~{}s^{-1}}$. The dotted curves show the maximum line-of-sight velocities of the Keplerian disk with a disk radius of 16 au. (b) PV diagram along the minor axis. The dashed curves represent the same UCM envelope as for the major axis. The SO emission shows a different shape from the envelope and from the disk in the major-axis PV diagram, whereas the detected velocity range is similar to that of the C18O (and 13CO) emission as shown in Figure 14. The linear velocity gradient in the major-axis PV diagram, along with the double peak structure in the moment 0 map at a higher angular resolution (Figure 6a), suggests that the SO emission traces a ring close to edge-on due to accretion shock between the infalling envelope and the disk, as reported in other protostellar systems (e.g., Yen et al., 2014; Ohashi et al., 2014; Sakai et al., 2016). The linear velocity gradient in the SO PV diagram $0.056~{}{\rm km~{}s^{-1}}~{}(r/1~{}{\rm au})$ intersects with the envelope rotation $45~{}{\rm km~{}s^{-1}}~{}(r/1~{}{\rm au})^{-1}$ at $r=28$ au. This radius is close to the outermost radius of the SO ridge points (Figure 13). Recent theoretical studies predict that the accretion shock around the disk occurs at the radius of $\sim 1.5R_{c}$ (Shariff et al., 2022). The inferred radius of the SO ring in I16253 is approximately consistent with this prediction. In conclusion, we suggest that I16253 has the Keplerian disk with $r\sim 16\pm 3$ au traced in the 12CO line, which is surrounded by the shock at $r\sim 28$ au traced in the SO line due to the mass accretion from the envelope traced in the C18O and 13CO lines. ### 5.2 Streamer in the SO emission In addition to the shocked ring, the SO emission also shows an extended structure to the east as shown in Figure 6(b). Its slightly redshifted velocity in the outer ($\gtrsim 1\arcsec$) region cannot be explained by the rotation, the infall motion on the midplane nor the outflow seen in our observations toward I16253, implying a different motion. A possible explanation is a streamer as reported in other protostellar systems on $\lesssim 1000$ au scales (e.g., Yen et al., 2019; Garufi et al., 2022; Thieme et al., 2022) (see also Kido et al., 2023) as well as on larger scales (e.g., Pineda et al., 2020). Hence, we constructed a streamer model by extracting ballistic, parabolic flows from the UCM envelope model used in Section 5.1. The free parameters are the two directional angles $(\theta_{0},\phi_{0})$ to specify the initial polar and azimuthal angle of the streamer trajectory, respectively. The polar angle $\theta_{0}$ is $0\arcdeg$ at the disk northern pole and $90\arcdeg$ at the midplane. The azimuthal angle $\phi_{0}$ is $0\arcdeg$ in the direction of the observer on the midplane and $90\arcdeg$ on the right seen from the observer. The inclination and position angles of the midplane are fixed at $65\arcdeg$ and $113\arcdeg$, respectively. By visual inspection, we found that $(\theta_{0},\phi_{0})=(60\arcdeg,240\arcdeg-270\arcdeg)$ can reasonably reproduce the observed SO distribution and velocity. This suggests that the SO gas is gravitationally bound if it came from these $(\theta_{0},\phi_{0})$ angles, since the UCM envelope model describes how material infalls toward the central gravitational source. Figure 16 shows the distribution of the model streamer projected on the plane-of-sky and the line-of-sight velocity of the streamer model and the rotating ring at a radius of 28 au. The ring rotates in the same velocity as the UCM envelope model. The model ring appears to reproduce the size and line-of-sight velocity of the observed compact component (Figures 6a), while the model streamer appears to reproduce the extended structure to the east and its line-of-sight velocity (Figure 6b). Meanwhile, the observed extended structure appears less confined than the model. This is probably because our model did not consider any beam blurring effect or radiative transfer effect. More sophisticated modeling including such effects will help to verify whether the anisotropy in the SO line is caused by the streamer. Figure 15: Figure 16: Model map of the line-of-sight velocity of a streamer in the UCM envelope ($M_{*}=0.14~{}M_{\sun}$ and $j=45~{}{\rm km~{}s^{-1}}~{}{\rm au}$ are the same as for Figure 14) and a Keplerian ring at 28 au. The streamer comes from $30\arcdeg$ above the midplane (i.e., $\theta_{0}=60\arcdeg$) and $0\arcdeg-20\arcdeg$ away from the left to the front seen from the observer around the rotational axis (i.e., $\phi_{0}=240\arcdeg-270\arcdeg$). Ballistic streamers are simulated in a theoretical study of the cloudlet capture in a protostellar system (Hanawa et al., 2022). The simulation shows that the streamers can be recognized as enhanced molecular line emission. Such an enhancement is seen in the eastern peak ($3\sigma$ higher than the western peak) of the C18O and 13CO moment 0 maps in I16253 (Figure 5), as well as in the SO extended emission. The eastern extension in the continuum emission (Figures 1b and 9b) could also be explained by the enhancement due to the streamer since it is located on the eastern side. Tobin et al. (2010) show the envelope of I16253 in the $8~{}\micron$ extinction observations at a $\sim 2\arcsec$ resolution using InfraRed Array Camera (IRAC) on the Spitzer telescope. The extinction is mainly concentrated in the $\sim 40\arcsec$-long outflow cavity wall, and the northeastern wall shows higher extinction than the other walls. This may suggest the existence of inhomogeneities in the ambient material that could produce the anisotropic streamer observed in our SO result. On the other hand, if the anisotropy in the SO line is due to an inhomogeneous density structure, similar anisotropy could also be seen in the C18O and 13CO lines, unlike our results. This may suggest that the anisotropy in the SO line could be caused by other factors than density, such as temperature, shock, or chemistry. Figure 18 summarizes the structures identified around I16253 in our ALMA observations, along with the large-scale envelope, denoting the radius or length of each structure. Figure 17: Figure 18: Schematic picture of the Keplerian disk, ring, envelope, streamer, outflow identified in the CO isotopologue and SO lines, and the eastern excess in the 1.3 mm continuum emission in the Class 0 protostar IRAS16253-2429. The gray background represents the large-scale envelope identified in 8 $\micron$ extinction (Tobin et al., 2010). The radii or lengths for each of the structures are indicated as well in units of au. ### 5.3 Mass Accretion and Other Quantities in I16253 The central stellar mass of I16253 has been directly estimated to be $0.12-0.17~{}M_{\sun}$ by identifying the Keplerian rotation in its disk in this paper, which ranged from $\sim 0.02$ to $\sim 0.12~{}M_{\sun}$ in previous works (Section 1). In contrast, some previous works estimated the central stellar mass of a protostar or a proto-BD from the mass accretion rate from the disk to the central object $\dot{M}_{\rm acc}$. We discuss the uncertainties for the method using $\dot{M}_{\rm acc}$ along with the accretion rate derived from the updated stellar mass. The central stellar mass of I16253 that we have estimated requires a mass accretion rate of $\dot{M}_{\rm acc}=L_{\rm bol}R_{*}/GM_{*}=(0.9-1.3)\times 10^{-7}~{}M_{\sun}~{}{\rm yr}^{-1}$, where the bolometric luminosity is $L_{\rm bol}=0.16~{}L_{\sun}$, the stellar radius is assumed to be $R_{*}=3~{}R_{\sun}$, $G$ is the gravitational constant, and the stellar mass is $M_{*}=0.12-0.17~{}M_{\sun}$. The stellar radius is predicted in numerical simulations to be within $\sim 30\%$ of $3~{}R_{\sun}$ (e.g., Masunaga & Inutsuka, 2000; Vorobyov & Basu, 2015). In comparison, Hsieh et al. (2016) estimated the mass accretion rate of I16253 from its outflow force. The outflow force, $F_{\rm out}\sim 7\times 10^{-7}~{}M_{\sun}~{}{\rm km~{}s^{-1}}~{}{\rm yr}^{-1}$, was measured through IRAM 30-m and APEX observations in 12CO lines at an angular resolution of $\sim 11\arcsec$. This force was converted to a mass accretion rate of $\dot{M}_{\rm acc}\sim 5\times 10^{-7}~{}M_{\sun}~{}{\rm yr}^{-1}$ (after the distance correction from 125 to 139 pc) as $\dot{M}_{\rm acc}=F_{\rm out}/\epsilon V_{W}f_{\rm ent}$, where the ratio between mass loss and accretion rates $\epsilon=\dot{M}_{W}/\dot{M}_{\rm acc}$ is assumed to be 0.1, the wind (jet) velocity $V_{W}$ is assumed to be 150 ${\rm km~{}s^{-1}}$, and the entrainment efficiency $f_{\rm ent}$ is assumed to be 0.1. $F_{\rm out}$ is estimated from observations in the 12CO $J=2-1$ line using IRAM 30-m and 12CO $J=7-6$ and $J=6-5$ lines using APEX, with the optical depth correction and assumptions that the kinetic temperature is 40 K and the H2 number density is $10^{5}~{}{\rm cm}^{-3}$. This mass accretion rate is $\sim 5$ times higher than the one derived from the updated stellar mass, whereas being similar to that of two proto-BD candidates, IC348-SMM2D and L328-IRS, $\sim 2-9\times 10^{-7}~{}M_{\sun}~{}{\rm yr}^{-1}$. This difference can be explained by large uncertainties in the conversion from the outflow force to the mass accretion rate. The mass loss and accretion ratio $\epsilon=\dot{M}_{\rm W}/\dot{M}_{\rm acc}$, the jet velocity $V_{W}$, and the efficiency $f_{\rm ent}$ can vary from $\sim 0.01$ to $\sim 0.5$ (Ellerbroek et al., 2013; Podio et al., 2021), $\sim 30$ to $\sim 160~{}{\rm km~{}s^{-1}}$ (jets in SiO $5-4$; Podio et al., 2021), and 0.1 to 0.25 (André et al., 1999), respectively, in the protostellar phase. A theoretical model called X-wind suggests that $\epsilon$ and $V_{W}$ are anti-correlated and the factor of $\epsilon V_{W}$ varies only around $\sim 50$ to $\sim 70~{}{\rm km~{}s^{-1}}$ (Najita & Shu, 1994) in the T Tauri phase. However, these velocities are $\gtrsim 4$ times higher than the typical velocity for protostars, $\sim 15~{}{\rm km~{}s^{-1}}$, as adopted in Hsieh et al. (2016), and the anti-correlation is not observationally confirmed in the protostellar phase. Those large uncertainties imply that $\dot{M}_{\rm acc}$ calculated by Hsieh et al. (2016) could be an order of magnitude lower. The abovementioned extreme values provide a range of $4\times 10^{-8}$ to $2\times 10^{-5}~{}M_{\sun}~{}{\rm yr}^{-1}$. For this reason, if the central stellar mass is estimated from the Keplerian rotation, the mass accretion rate could also be estimated better than from the outflow observations. The updated mass accretion rate of I16253, $\sim 1\times 10^{-7}~{}M_{\sun}~{}{\rm yr}^{-1}$, cannot provide $M_{*}=0.14~{}M_{\sun}$ within the lifetime of Class 0 $\lesssim$0.2 Myr. In other words, the updated central stellar mass is large against the luminosity of I16253, as the required time can be estimated to be $M_{*}/\dot{M}_{\rm acc}=GM_{*}^{2}/R_{*}L_{\rm bol}\sim 1$ Myr. This suggests that I16253 likely experienced an accretion burst (or stronger accretion) in the past. The presence of an accretion burst is supported by the episodic mass ejection as seen in the 12CO PV diagram along the outflow axis (Section 3.2). The Class 0 age is also long enough to experience bursts with a typical interval of 2400 years in the Class 0 phase (Hsieh et al., 2019). Once an accretion burst occurs, the luminosity of a protostar increases resulting in higher temperatures in the surrounding envelope at a given radius. This causes CO molecules to sublimate from the icy grain mantles in an extended region of the envelope. After the protostar has returned to its quiescent state and the luminosity decreased again, the molecules remain in the gas-phase for a period of time before freezing out again. This freeze-out time-scale depends on the density at a given radius (Rodgers & Charnley, 2003) but is typically of order $10^{4}$ years where CO sublimation due to an accretion burst occurs around Solar-type protostars (e.g., Jørgensen et al., 2015). From this point of view, the size of the C18O emission observed in I16253 (Figure 5a), $\gtrsim 150$ au, also supports that an accretion burst happened because this size is much larger than expected from the current luminosity of this protostar: The C18O emission radius is predicted to be $\sim 30$ au with the current luminosity of I16253, $0.16~{}L_{\sun}$ (Jørgensen et al., 2015) (see also Lee, 2007). Based on the measured $M_{*}$ and the above discussion about the mass accretion and the luminosity, we conclude that I16253 is not a proto-BD even though its luminosity and mass accretion rate are similar to those of the two proto-BD candidates. The core (or envelope) mass of I16253 is estimated to be $\sim 1~{}M_{\sun}$ from 1.1 mm observations with Bolocam on the CSO telescope at an angular resolution of $31\arcsec$ (Young et al., 2006) and $8~{}\micron$ observations with IRAC on the Spitzer telescope at an angular resolution of $2\arcsec$ (Tobin et al., 2010). If this envelope mass accretes onto I16253 with a typical star formation efficiency ($30\%\pm 10\%$ calculated from the dense core mass function and initial mass function; Alves et al., 2007), a mass of 0.2 to 0.4 $M_{\sun}$ will be added to the central stellar mass. This also supports an idea that I16253 will ultimately obtain mass high enough to fuse hydrogen (i.e., $M_{*}>0.08~{}M_{\sun}$). The Keplerian disk around I16253 has been kinematically identified for the first time. Without any clear identification of the Keplerian disk, previous works have attempted to estimate the central stellar mass using the infalling motion or the outflow force (through the mass accretion rate) in I16253 and suggested this protostar as a proto-BD candidate. Similarly, the central stellar mass of proto-BD candidates was estimated in those methods in some previous works (Section 1), without identifying a Keplerian disk. Our study demonstrates that a proto-BD and a very low mass protostar must be identified through the dynamical central stellar mass derived from the Keplerian rotation of its disk, rather than its infall motion or outflow force. The accurate mass estimation enables us to compare physical quantities in the I16253 system with scaling relations among young stellar objects. For example, its disk radius, $16-19$ au, appears consistent with or slightly lower than the scaling relation between the disk radius and the central stellar mass found with the protostellar sample of Yen et al. (2017). Its mass accretion rate, $(0.9-1.3)\times 10^{-7}~{}M_{\sun}~{}{\rm yr}^{-1}$, and disk mass, $\gtrsim 2\times 10^{-3}~{}M_{\sun}$, are consistent with the scaling relation found with the Class I sample of Fiorellino et al. (2022). More observations around the BD mass threshold with accurate mass estimations will be necessary to determine whether these scaling relations hold down to the BD mass regime and bridge star formation and BD formation. ## 6 Conclusions As a part of the ALMA large project eDisk, we have observed the Class 0 protostar IRAS16253-2429, which has been suggested to be a proto-brown dwarf candidate in previous works, in the 1.3 mm continuum, 12CO $J=2-1$, C18O $J=2-1$, 13CO $J=2-1$, SO $J_{N}=6_{5}-5_{4}$, and other molecular lines at an angular resolution of $0\farcs 07$ ($\sim 10$ au). Our results provide a typical picture of protostars with a very low stellar mass close to the brown dwarf threshold (Figure 18). The main results are summarized below. 1. 1. The continuum emission shows structures from a $\sim 600$-au scale down to a $\sim 15$-au scale. The main component shows a disk-like structure with a radius of $\sim 20$ au. The emission is extended to the southeast along the major axis. These extensions can be interpreted as an enhancement due to a streamer from the east and the near-side (southwestern side) wall of the disk- like structure. 2. 2. The 12CO emission overall traces a clear bipolar outflow up to a $\sim 3000$-au scale. The outflow suggests episodic mass ejections. Furthermore, the 12CO emission on the midplane shows a velocity gradient along the disk’s major axis, implying rotation of the disk. 3. 3. We analyzed the 12CO major-axis position-velocity diagram by the edge and ridge methods and identified a Keplerian disk in IRAS16253-2429 for the first time. From this identification of the Keplerian rotation in both methods, the central stellar mass is estimated to be $0.12-0.17~{}M_{\sun}$. This mass leads us to conclude that IRAS16253-2429 is unlikely a proto-brown dwarf but a very low mass protostar. 4. 4. The C18O and 13CO emissions trace the infalling and rotating envelope as reported in previous observational works. The major- and minor-axis position- velocity diagrams in these lines are consistent with the UCM envelope model with the central stellar mass of $0.14~{}M_{\sun}$ and the disk (centrifugal) radius of $r\sim 16$ au. Their moment 0 maps exhibit stronger emission intensities on the eastern peak, which could result from the streamer mentioned below. 5. 5. The SO emission shows a different velocity structure from the CO isotopologues. Its double peaks in the moment 0 map and linear velocity gradient in the major-axis position-velocity diagram suggest that this emission traces a ring ($r\sim 28$ au) due to the accretion shock between the disk and the envelope. This emission also shows a streamer from the eastern side, which can be explained with a ballistic, parabolic flow at $30\arcdeg$ above the midplane extracted from the UCM envelope model. ## Acknowledgements This paper makes use of the following ALMA data: ADS/JAO.ALMA#2019.1.00261.L and ADS/JAO.ALMA#2019.A.00034.S. ALMA is a partnership of ESO (representing its member states), NSF (USA) and NINS (Japan), together with NRC (Canada), MOST and ASIAA (Taiwan), and KASI (Republic of Korea), in cooperation with the Republic of Chile. The Joint ALMA Observatory is operated by ESO, AUI/NRAO and NAOJ. The National Radio Astronomy Observatory is a facility of the National Science Foundation operated under cooperative agreement by Associated Universities, Inc. W.K. was supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (NRF-2021R1F1A1061794). N.O. acknowledges support from National Science and Technology Council (NSTC) in Taiwan through the grants NSTC 109-2112-M-001-051 and 110-2112-M-001-031. J.K.J. acknowledge support from the Independent Research Fund Denmark (grant No. 0135-00123B). J.J.T. acknowledges support from NASA XRP 80NSSC22K1159. N.O. acknowledges support from National Science and Technology Council (NSTC) in Taiwan through the grants NSTC 109-2112-M-001-051 and 110-2112-M-001-031. Y.A. acknowledges support by NAOJ ALMA Scientific Research Grant code 2019-13B, Grant-in-Aid for Scientific Research (S) 18H05222, and Grant-in-Aid for Transformative Research Areas (A) 20H05844 and 20H05847. IdG acknowledges support from grant PID2020-114461GB-I00, funded by MCIN/AEI/10.13039/501100011033. PMK acknowledges support from NSTC 108-2112- M-001-012, NSTC 109-2112-M-001-022 and NSTC 110-2112-M-001-057. SPL and TJT acknowledge grants from the National Science and Technology Council of Taiwan 106-2119-M-007-021-MY3 and 109-2112-M-007-010-MY3. J.K.J. acknowledges support from the Independent Research Fund Denmark (grant No. 0135-00123B). C.W.L. is supported by the Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education, Science and Technology (NRF- 2019R1A2C1010851), and by the Korea Astronomy and Space Science Institute grant funded by the Korea government (MSIT; Project No. 2022-1-840-05). JEL was supported by the National Research Foundation of Korea (NRF) grant funded by the Korean government (MSIT) (grant number 2021R1A2C1011718). ZYL is supported in part by NASA 80NSSC20K0533 and NSF AST-1910106. ZYDL acknowledges support from NASA 80NSSC18K1095, the Jefferson Scholars Foundation, the NRAO ALMA Student Observing Support (SOS) SOSPA8-003, the Achievements Rewards for College Scientists (ARCS) Foundation Washington Chapter, the Virginia Space Grant Consortium (VSGC), and UVA research computing (RIVANNA). LWL acknowledges support from NSF AST-2108794. S.N. acknowledges support from the National Science Foundation through the Graduate Research Fellowship Program under Grant No. 2236415. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the authors and do not necessarily reflect the views of the National Science Foundation. R.S. acknowledge support from the Independent Research Fund Denmark (grant No. 0135-00123B). S.T. is supported by JSPS KAKENHI grant Nos. 21H00048 and 21H04495, and by NAOJ ALMA Scientific Research grant No. 2022-20A. JPW acknowledges support from NSF AST-2107841. H.-W.Y. acknowledges support from the National Science and Technology Council (NSTC) in Taiwan through the grant NSTC 110-2628-M-001-003-MY3 and from the Academia Sinica Career Development Award (AS-CDA-111-M03). ## Appendix A Other molecular lines Figure 19 shows the moment 0 and 1 maps of the lines observed in the eDisk project toward I16253, except for the 12CO, C18O, 13CO, and SO lines. These lines were observed from the wide spectral windows otherwise intended for continuum measurements. The three H2CO lines have different upper state energies, depending on the rest frequencies (10.5 K; 218.22 GHz, 57.6 K; 218.47 GHz, 57.6 K; 218.76 GHz). The three cyclic C3H2 lines also have different upper state energies (28.2 K; 217.82 GHz, 25.0 K; 217.94 GHz, 24.9 K; 218.16 GHz). The CH3OH, DCN, and SiO lines have upper state energies of 35.0, 10.4 and 20.8 K, respectively. In comparison, the 12CO, C18O, 13CO, and SO lines have upper state energies of 5.5, 5.3, 5.3, and 24.4 K, respectively. All the panels in Figure 19 are made from three channels centered at the systemic velocity ($V_{\rm LSR}=2.66-5.34~{}{\rm km~{}s^{-1}}$) and show the same spatial and velocity ranges. Figure 19: Moment 0 and 1 maps in 5 other molecular lines (9 transitions) observed in the ALMA observations of the eDisk project. These images are made with the robust parameter of 2. The emission is integrated from 2.66 to $5.34~{}{\rm km~{}s^{-1}}$ (3 channels) for all lines. The contour levels are in $3\sigma$ steps from $3\sigma$, where $1\sigma$ is $1.5~{}{\rm mJy~{}beam^{-1}}~{}{\rm km~{}s^{-1}}$. The ellipse at the lower left corner in each panel denotes the synthesized beam, $\sim 0\farcs 38\times 0\farcs 27\ (78\arcdeg)$. The diagonal lines are the major and minor axes of the continuum emission, P.A.$=113\arcdeg$ and $23\arcdeg$. ## References * Alves et al. (2017) Alves, F. O., Girart, J. M., Caselli, P., et al. 2017, A&A, 603, L3. doi:10.1051/0004-6361/201731077 * Alves et al. (2007) Alves, J., Lombardi, M., & Lada, C. J. 2007, A&A, 462, L17. doi:10.1051/0004-6361:20066389 * André et al. (1999) André, P., Motte, F., & Bacmann, A. 1999, ApJ, 513, L57. doi:10.1086/311908 * Andrews & Williams (2005) Andrews, S. M. & Williams, J. P. 2005, ApJ, 631, 1134. doi:10.1086/432712 * Aso & Machida (2020) Aso, Y. & Machida, M. N. 2020, ApJ, 905, 174. doi:10.3847/1538-4357/abc6fc * Aso & Sai (2023) Aso, Y. & Sai, J. 2023, jinshisai/SLAM: First Release of SLAM, v1.0.0, Zenodo, doi:10.5281/zenodo.7783868 * Aso et al. (2017) Aso, Y., Ohashi, N., Aikawa, Y., et al. 2017, ApJ, 849, 56. doi:10.3847/1538-4357/aa8264 * Aso et al. (2015) Aso, Y., Ohashi, N., Saigo, K., et al. 2015, ApJ, 812, 27. doi:10.1088/0004-637X/812/1/27 * Aso et al. (2021) Aso, Y., Kwon, W., Hirano, N., et al. 2021, ApJ, 920, 71. doi:10.3847/1538-4357/ac15f3 * Astropy Collaboration et al. (2018) Astropy Collaboration, Price-Whelan, A. M., Sipőcz, B. M., et al. 2018, AJ, 156, 123. doi:10.3847/1538-3881/aabc4f * Astropy Collaboration et al. (2013) Astropy Collaboration, Robitaille, T. P., Tollerud, E. J., et al. 2013, A&A, 558, A33. doi:10.1051/0004-6361/201322068 * Basu et al. (2012) Basu, S., Vorobyov, E. I., & DeSouza, A. L. 2012, First Stars IV - from Hayashi to the Future -, 1480, 63. doi:10.1063/1.4754329 * Bate (2012) Bate, M. R. 2012, MNRAS, 419, 3115. doi:10.1111/j.1365-2966.2011.19955.x * Beckwith et al. (1990) Beckwith, S. V. W., Sargent, A. I., Chini, R. S., et al. 1990, AJ, 99, 924. doi:10.1086/115385 * Cassen & Moosman (1981) Cassen, P. & Moosman, A. 1981, Icarus, 48, 353. doi:10.1016/0019-1035(81)90051-8 * Chabrier (2002) Chabrier, G. 2002, ApJ, 567, 304. doi:10.1086/324716 * Cabrit (1989) Cabrit, S. 1989, European Southern Observatory Conference and Workshop Proceedings, 33, 119 * de Geus et al. (1989) de Geus, E. J., de Zeeuw, P. T., & Lub, J. 1989, A&A, 216, 44 * de Gregorio-Monsalvo et al. (2016) de Gregorio-Monsalvo, I., Barrado, D., Bouy, H., et al. 2016, A&A, 590, A79. doi:10.1051/0004-6361/201424149 * Dunham et al. (2008) Dunham, M. M., Crapsi, A., Evans, N. J., et al. 2008, ApJS, 179, 249. doi:10.1086/591085 * Ellerbroek et al. (2013) Ellerbroek, L. E., Podio, L., Kaper, L., et al. 2013, A&A, 551, A5. doi:10.1051/0004-6361/201220635 * Fiorellino et al. (2022) Fiorellino, E., Tychoniec, Ł., Manara, C. F., et al. 2022, ApJ, 937, L9. doi:10.3847/2041-8213/ac8fee * Foreman-Mackey et al. (2013) Foreman-Mackey, D., Hogg, D. W., Lang, D., et al. 2013, PASP, 125, 306. doi:10.1086/670067 * Garufi et al. (2022) Garufi, A., Podio, L., Codella, C., et al. 2022, A&A, 658, A104. doi:10.1051/0004-6361/202141264 * Hanawa et al. (2022) Hanawa, T., Sakai, N., & Yamamoto, S. 2022, ApJ, 932, 122. doi:10.3847/1538-4357/ac6e6a * Hsieh et al. (2019) Hsieh, T.-H., Hirano, N., Belloche, A., et al. 2019, ApJ, 871, 100. doi:10.3847/1538-4357/aaf4fe * Hsieh et al. (2019) Hsieh, T.-H., Murillo, N. M., Belloche, A., et al. 2019, ApJ, 884, 149. doi:10.3847/1538-4357/ab425a * Hsieh et al. (2016) Hsieh, T.-H., Lai, S.-P., Belloche, A., et al. 2016, ApJ, 826, 68. doi:10.3847/0004-637X/826/1/68 * Huélamo et al. (2017) Huélamo, N., de Gregorio-Monsalvo, I., Palau, A., et al. 2017, A&A, 597, A17. doi:10.1051/0004-6361/201628510 * Jørgensen et al. (2015) Jørgensen, J. K., Visser, R., Williams, J. P., et al. 2015, A&A, 579, A23. doi:10.1051/0004-6361/201425317 * Kido et al. (2023) Kido, M., Takakuwa, S., Saigo, K., et al. 2023, ApJ, 953, 190. doi:10.3847/1538-4357/acdd7a * Kim et al. (2019) Kim, G., Lee, C. W., Maheswar, G., et al. 2019, ApJS, 240, 18. doi:10.3847/1538-4365/aaf889 * Knude & Hog (1998) Knude, J. & Hog, E. 1998, A&A, 338, 897 * Lee (2020) Lee, C.-F. 2020, A&A Rev., 28, 1. doi:10.1007/s00159-020-0123-7 * Lee et al. (2018) Lee, C. W., Kim, G., Myers, P. C., et al. 2018, ApJ, 865, 131. doi:10.3847/1538-4357/aadcf6 * Lee (2007) Lee, J.-E. 2007, Journal of Korean Astronomical Society, 40, 83. doi:10.5303/JKAS.2007.40.4.083 * Machida et al. (2009) Machida, M. N., Inutsuka, S.-. ichiro ., & Matsumoto, T. 2009, ApJ, 699, L157. doi:10.1088/0004-637X/699/2/L157 * Machida et al. (2010) Machida, M. N., Inutsuka, S.-. ichiro ., & Matsumoto, T. 2010, ApJ, 724, 1006. doi:10.1088/0004-637X/724/2/1006 * Machida et al. (2014) Machida, M. N., Inutsuka, S.-. ichiro ., & Matsumoto, T. 2014, MNRAS, 438, 2278. doi:10.1093/mnras/stt2343 * Maret et al. (2020) Maret, S., Maury, A. J., Belloche, A., et al. 2020, A&A, 635, A15. doi:10.1051/0004-6361/201936798 * Masunaga & Inutsuka (2000) Masunaga, H. & Inutsuka, S.-. ichiro . 2000, ApJ, 531, 350. doi:10.1086/308439 * McMullin et al. (2007) McMullin, J. P., Waters, B., Schiebel, D., et al. 2007, Astronomical Data Analysis Software and Systems XVI, 376, 127 * Najita & Shu (1994) Najita, J. R. & Shu, F. H. 1994, ApJ, 429, 808. doi:10.1086/174365 * Ohashi et al. (2014) Ohashi, N., Saigo, K., Aso, Y., et al. 2014, ApJ, 796, 131. doi:10.1088/0004-637X/796/2/131 * Ohashi et al. (2023) Ohashi, N., Tobin, J. J., Jørgensen, J. K., et al. 2023, ApJ, 951, 8. doi:10.3847/1538-4357/acd384 * Padoan & Nordlund (2004) Padoan, P. & Nordlund, Å. 2004, ApJ, 617, 559. doi:10.1086/345413 * Palau et al. (2014) Palau, A., Zapata, L. A., Rodríguez, L. F., et al. 2014, MNRAS, 444, 833. doi:10.1093/mnras/stu1461 * Park et al. (2021) Park, W., Lee, J.-E., Contreras Peña, C., et al. 2021, ApJ, 920, 132. doi:10.3847/1538-4357/ac1745 * Pineda et al. (2020) Pineda, J. E., Segura-Cox, D., Caselli, P., et al. 2020, Nature Astronomy, 4, 1158. doi:10.1038/s41550-020-1150-z * Podio et al. (2021) Podio, L., Tabone, B., Codella, C., et al. 2021, A&A, 648, A45. doi:10.1051/0004-6361/202038429 * Rodgers & Charnley (2003) Rodgers, S. D. & Charnley, S. B. 2003, ApJ, 585, 355. doi:10.1086/345497 * Ruíz-Rodríguez et al. (2022) Ruíz-Rodríguez, D. A., Williams, J. P., Kastner, J. H., et al. 2022, MNRAS, 515, 2646. doi:10.1093/mnras/stac1879 * Sai et al. (2022) Sai, J., Ohashi, N., Maury, A. J., et al. 2022, ApJ, 925, 12. doi:10.3847/1538-4357/ac341d * Sai et al. (2020) Sai, J., Ohashi, N., Saigo, K., et al. 2020, ApJ, 893, 51. doi:10.3847/1538-4357/ab8065 * Sakai et al. (2016) Sakai, N., Oya, Y., López-Sepulcre, A., et al. 2016, ApJ, 820, L34. doi:10.3847/2041-8205/820/2/L34 * Santamaría-Miranda et al. (2021) Santamaría-Miranda, A., de Gregorio-Monsalvo, I., Plunkett, A. L., et al. 2021, A&A, 646, A10. doi:10.1051/0004-6361/202039419 * Seifried et al. (2016) Seifried, D., Sánchez-Monge, Á., Walch, S., et al. 2016, MNRAS, 459, 1892. doi:10.1093/mnras/stw785 * Shariff et al. (2022) Shariff, K., Gorti, U., & Melon Fuksman, J. D. 2022, MNRAS, 514, 5548. doi:10.1093/mnras/stac1186 * Stamatellos & Whitworth (2009) Stamatellos, D. & Whitworth, A. P. 2009, 15th Cambridge Workshop on Cool Stars, Stellar Systems, and the Sun, 1094, 557. doi:10.1063/1.3099172 * Thieme et al. (2022) Thieme, T. J., Lai, S.-P., Lin, S.-J., et al. 2022, ApJ, 925, 32. doi:10.3847/1538-4357/ac382b * Tobin et al. (2023) Tobin, J. 2023, eDisk data reduction scripts, 1.0.0, Zenodo, doi:10.5281/zenodo.7986682 * Tobin et al. (2012) Tobin, J. J., Hartmann, L., Bergin, E., et al. 2012, ApJ, 748, 16. doi:10.1088/0004-637X/748/1/16 * Tobin et al. (2010) Tobin, J. J., Hartmann, L., Looney, L. W., et al. 2010, ApJ, 712, 1010. doi:10.1088/0004-637X/712/2/1010 * Tobin et al. (2020) Tobin, J. J., Sheehan, P. D., Megeath, S. T., et al. 2020, ApJ, 890, 130. doi:10.3847/1538-4357/ab6f64 * Ulrich (1976) Ulrich, R. K. 1976, ApJ, 210, 377. doi:10.1086/154840 * Vorobyov & Basu (2015) Vorobyov, E. I. & Basu, S. 2015, ApJ, 805, 115. doi:10.1088/0004-637X/805/2/115 * Vousden et al. (2016) Vousden, W. D., Farr, W. M., & Mandel, I. 2016, MNRAS, 455, 1919. doi:10.1093/mnras/stv2422 * Whitworth & Zinnecker (2004) Whitworth, A. P. & Zinnecker, H. 2004, A&A, 427, 299. doi:10.1051/0004-6361:20041131 * Wilner & Welch (1994) Wilner, D. J., & Welch, W. J. 1994, ApJ, 427, 898 * Yen et al. (2019) Yen, H.-W., Gu, P.-G., Hirano, N., et al. 2019, ApJ, 880, 69. doi:10.3847/1538-4357/ab29f8 * Yen et al. (2017) Yen, H.-W., Koch, P. M., Takakuwa, S., et al. 2017, ApJ, 834, 178. doi:10.3847/1538-4357/834/2/178 * Yen et al. (2014) Yen, H.-W., Takakuwa, S., Ohashi, N., et al. 2014, ApJ, 793, 1. doi:10.1088/0004-637X/793/1/1 * Yen et al. (2013) Yen, H.-W., Takakuwa, S., Ohashi, N., et al. 2013, ApJ, 772, 22. doi:10.1088/0004-637X/772/1/22 * Young et al. (2006) Young, K. E., Enoch, M. L., Evans, N. J., et al. 2006, ApJ, 644, 326. doi:10.1086/503327 * Zucker et al. (2020) Zucker, C., Speagle, J. S., Schlafly, E. F., et al. 2020, A&A, 633, A51. doi:10.1051/0004-6361/201936145 * Zucker et al. (2019) Zucker, C., Speagle, J. S., Schlafly, E. F., et al. 2019, ApJ, 879, 125. doi:10.3847/1538-4357/ab2388
# Gravitational Portals with Non-Minimal Couplings Simon Clérya<EMAIL_ADDRESS>Yann Mambrinia,b <EMAIL_ADDRESS>Keith A. Olivec<EMAIL_ADDRESS>Andrey Shkerinc<EMAIL_ADDRESS>Sarunas Vernerc<EMAIL_ADDRESS>a Université Paris- Saclay, CNRS/IN2P3, IJCLab, 91405 Orsay, France b CERN, Theoretical Physics Department, Geneva, Switzerland c William I. Fine Theoretical Physics Institute, School of Physics and Astronomy, University of Minnesota, Minneapolis, MN 55455, USA ###### Abstract We consider the effects of non-minimal couplings to curvature of the form $\xi_{S}S^{2}R$, for three types of scalars: the Higgs boson, the inflaton, and a scalar dark matter candidate. We compute the abundance of dark matter produced by these non-minimal couplings to gravity and compare to similar results with minimal couplings. We also compute the contribution to the radiation bath during reheating. The main effect is a potential augmentation of the maximum temperature during reheating. A model independent limit of $\mathcal{O}(10^{12})$ GeV is obtained. For couplings $\xi_{S}\gtrsim\mathcal{O}(1)$, these dominate over minimal gravitational interactions. ††preprint: UMN–TH–4116/22††preprint: FTPI–MINN–22/07††preprint: CERN- TH-2022-025 ## I Introduction Promoting a field theory Lagrangian from a Lorentz-invariant one to a generally-covariant one necessarily leads to an interaction between the fields of the theory and the gravitational field. In the case of a scalar field, $S$, the natural generalization of this minimal interaction scenario is to introduce a non-minimal coupling term of the form $\propto\xi_{S}S^{2}R\;.$ (1) Here $R$ is the Ricci scalar and $\xi_{S}$ is a non-minimal coupling constant. This non-minimal coupling to gravity proved to be useful in many applications to cosmology. Examples include Higgs inflation Bezrukov:2007ep ; Lebedev:2021xey , where $S$ is associated with the Higgs field degree of freedom $h$ — the only scalar degree of freedom in the Standard Model, preheating Ema:2016dny , where $S$ is associated with the inflaton field $\phi$, and non-perturbative production of dark matter nonminprod , where $S$ represents the scalar dark matter particle $X$. In the general case, when the fields $\phi$, $h$, and $X$ are all different, the question arises as to what extent they must interact with each other in order to successfully reheat the Universe and generate the right amount of dark matter. Recent studies have shown that interactions via gravity alone, to which the fields are coupled minimally, is enough for these purposes. Indeed, the perturbative gravitational production of dark matter through graviton exchange can play a dominant role during reheating with processes involving the inflaton MO ; CMOV ; Barman:2021ugy as well as thermal bath particles CMOV ; Haque:2021mab . Further, the minimal gravitational coupling can lead to the completion of the reheating process for certain types of the inflationary potential, $V(\phi)\sim\phi^{k}$ with $k>2$ CMOV ; Haque:2022kez . Thus, gravity is strong enough to mediate perturbative channels of reheating and dark matter production. The purpose of this work is to study how the inclusion of the non-minimal coupling terms of the form (1) affect the gravitational production of dark matter and radiation during reheating. Note that the presence of these terms is unavoidable: if there were no such couplings at tree level, they would still be generated by quantum corrections Callan:1970ze . We study particle production in the processes $hh\rightarrow XX$, $\phi\phi\rightarrow hh$, and $\phi\phi\rightarrow XX$ which are induced by the non-minimal couplings. Here $\phi$ represents the inflaton background oscillating around its minimum after the end of inflation gravprod . Since the scalar fields couple directly to the curvature scalar $R$, the oscillating background causes the effective masses of the fields to change non-adiabatically and leads to particle production. This regime of particle creation has been considered in several different contexts, including gravitational production of scalar gravscalar ; ema , fermion gravferm , and vector dark matter gravvector . Our main interest is to compare the (dark) matter production channels induced by the non-minimal couplings with the production via the s-channel graviton exchange that sets minimal possible production rates. We will see for which values of the couplings the rates are enhanced, and what are the consequences on the dark matter density or the temperature attained during reheating. Throughout the work we adopt the Starobinsky inflationary potential staro , although our results are largely independent of the particular form of the potential. As for the potentials for the fields $h$ and $X$, we take them to be renormalizable polynomials. We also assume no direct interaction between $\phi$, $h$, and $X$. Working in the perturbative regime implies that the non-minimal couplings must satisfy $|\xi_{S}|\ll M_{P}^{2}/\langle S\rangle^{2}$, where $\langle S\rangle$ is the vacuum expectation value of $S=\phi,h,X$. The value of $\xi_{h}$ is constrained from collider experiments as $|\xi_{h}|\lesssim 10^{15}$ higgscons1 .111Note that in the case of Higgs inflation, $\xi_{h}$ is fixed from CMB measurements Bezrukov:2007ep . Furthermore, the lower bound on $\xi_{h}$ comes from the fact that the Standard Model electroweak vacuum may not be absolutely stable HiggsStab . To prevent the vacuum decay due to quantum fluctuations during inflation fluc , the effective mass of the Higgs field induced by the non-minimal coupling must be large enough; this gives $\xi_{h}\gtrsim 10^{-1}$ higgstree ; higgsloop (see also Markkanen:2018pdo ).222This estimate assumes no new physics interfering the RG running of the Higgs self-coupling constant until inflationary energy scales. The paper is organized as follows: The framework for our computation is presented in Section II. We discuss non-minimal gravitational couplings of the inflaton, the Higgs boson, and a dark matter scalar in detail. We calculate the dark matter production rates either from scattering in the thermal bath or from oscillations in the inflaton condensate. We compare similar processes obtained from the minimal gravitational particle production. We choose the Starobinsky model of inflation and discuss the reheating epoch when the inflaton begins oscillating. In Section III we discuss the resulting abundance of dark matter produced from the thermal bath and directly from scattering of the inflaton condensate. We also compute the effects of the non-minimal couplings on the maximum temperature attained during reheating. We then compare different processes in Section IV, before summarizing our results in Section V. ## II The framework ### II.1 Scalar-gravity Lagrangian The theory we consider comprises 3 scalar fields non-minimally coupled to gravity: the inflaton $\phi$, the Higgs field333We consider the Higgs boson as a surrogate for any additional scalars with Standard Model couplings. $H$, for which we adopt the Unitary gauge, $H=(0,h)^{T}/\sqrt{2}$, and the dark matter candidate $X$. The relevant part of the action takes the form444The metric signature is chosen as $(+,-,-,-)$. $\mathcal{S}\;=\;\int d^{4}x\sqrt{-\tilde{g}}\left[-\frac{M_{P}^{2}}{2}\Omega^{2}\tilde{R}+\mathcal{L}_{\phi}+\mathcal{L}_{h}+\mathcal{L}_{X}\right]$ (2) with the conformal factor $\Omega^{2}$ given by $\Omega^{2}\;\equiv\;1+\frac{\xi_{\phi}\phi^{2}}{M_{P}^{2}}+\frac{\xi_{h}h^{2}}{M_{P}^{2}}+\frac{\xi_{X}X^{2}}{M_{P}^{2}}\,.$ (3) Here $M_{P}=2.4\times 10^{18}\,\rm{GeV}$ is the reduced Planck mass, and the tilde used in Eq. (2) indicates that the theory is considered in the Jordan frame. For the scalar field Lagrangians we have $\mathcal{L}_{S}\;=\frac{1}{2}\tilde{g}^{\mu\nu}\partial_{\mu}S\partial_{\nu}S-V_{S}\,,~{}~{}~{}S=\phi,h,X\,.$ (4) Next, we specify the scalar field potentials. For a model of inflation, we choose the well-motivated Starobinsky model for which staro $V_{\phi}\;=\;\frac{3}{4}m_{\phi}^{2}M_{P}^{2}\left(1-e^{-\sqrt{\frac{2}{3}}\frac{\phi}{M_{P}}}\right)^{2}\,.$ (5) In what follows, we work in the perturbative regime with $\phi\ll M_{P}$, hence the potential is approximated as $V_{\phi}\;\simeq\;\frac{1}{2}m_{\phi}^{2}\phi^{2}\,.$ (6) The inflaton mass, $m_{\phi}$, is fixed by the amplitude of scalar perturbations inferred from CMB measurements Planck ; for the potential (5) this gives $m_{\phi}=3\times 10^{13}$ GeV building . The potential for the Higgs field is taken as follows $V_{h}\;=\;\frac{1}{2}m_{h}^{2}h^{2}+\frac{1}{4}\lambda_{h}h^{4}\,.$ (7) Here $m_{h}$ and $\lambda_{h}$ are the Higgs mass and quartic coupling, correspondingly. Note that both parameters undergo the renormalization group (RG) running. In what follows we take a weak scale mass, which is a good approximation at the time of reheating and our results are insensitive to $\lambda_{h}$. Finally, the dark matter potential is simply given by $V_{X}\;=\;\frac{1}{2}m_{X}^{2}X^{2}\,.$ (8) To study the reheating in the theory (2), it is convenient to remove the non- minimal couplings by performing the redefinition of the metric field. Leaving the details to Appendix A, we write the action (2) in the Einstein frame, $\begin{split}\mathcal{S}\;=\;\int d^{4}x\sqrt{-g}&\left[-\frac{M_{P}^{2}}{2}R+\frac{1}{2}K^{ij}g^{\mu\nu}\partial_{\mu}S_{i}\partial_{\nu}S_{j}\right.\\\ &\qquad\qquad\qquad\left.-\frac{V_{\phi}+V_{h}+V_{X}}{\Omega^{4}}\right]\,.\end{split}$ (9) Here the indices $i,j$ enumerate the fields $\phi,h,X$, and the kinetic function is given by $K^{ij}\;=\;6\frac{\partial\log\Omega}{\partial S_{i}}\frac{\partial\log\Omega}{\partial S_{j}}+\frac{\delta^{ij}}{\Omega^{2}}\,.$ (10) Note that the scalar field kinetic term is not canonical. In general, it is impossible to make a field redefinition that would bring it to the canonical form, unless all three non-minimal couplings vanish.555 Such a redefinition exists if the three-dimensional manifold spanned by the fields $\phi$, $h$ and $X$ is flat. One can show that it is not the case if at least one of the couplings is non-zero. For the theory (9) to be well-defined, the kinetic function (10) must be positive-definite. Computing the eigenvalues, one arrives at the condition $\Omega^{2}>0\,,$ (11) which is satisfied automatically for positive values of the couplings. Note that the negative couplings are also allowed for certain scalar field magnitudes. In what follows, we will be interested in the small-field limit $\frac{|\xi_{\phi}|\phi^{2}}{M_{P}^{2}}\;,~{}~{}\frac{|\xi_{h}|h^{2}}{M_{P}^{2}}\;,~{}~{}\frac{|\xi_{X}|X^{2}}{M_{P}^{2}}\ll 1\,.$ (12) We can expand the kinetic and potential terms in the action (9) in powers of $M_{P}^{-2}$. We obtain a canonical kinetic term for the scalar fields and deduce the leading-order interactions induced by the non-minimal couplings. The latter can be brought to the form $\mathcal{L}_{\rm{non-min.}}\;=\;-\sigma_{hX}^{\xi}h^{2}X^{2}-\sigma_{\phi X}^{\xi}\phi^{2}X^{2}-\sigma_{\phi h}^{\xi}\phi^{2}h^{2}\,,$ (13) where the $\sigma_{ij}^{\xi}$ are functions of the couplings $\xi_{i}$, $\xi_{j}$, the masses $m_{i}$, $m_{j}$, and the Mandelstam variables; see Appendix A for details. The small-field approximation (12) implies the bound $\sqrt{|\xi_{S}|}\lesssim M_{P}/\langle S\rangle$ with $S=\phi,h,X$. Since the inflaton value at the end of inflation is $\phi_{\rm end}\sim M_{P}$ and afterwards $\langle\phi^{2}\rangle\sim a^{-3}$, where $a$ is the cosmological scale factor, then $|\xi_{\phi}|\lesssim(a/a_{\rm end})^{3}$. In particular, at the onset of inflaton oscillations $|\xi_{\phi}|\lesssim 1\,.$ (14) Note that since our calculations involve the effective couplings $\sigma_{\phi X}^{\xi}$ ($\sigma_{\phi h}^{\xi}$), which depend both on $\xi_{\phi}$ and $\xi_{X}$ ($\xi_{h}$), the relatively small value of $|\xi_{\phi}|$ can, in principle, be compensated by a large value of the other couplings. In Fig. 1, we show the scattering processes obtained from the Lagrangian (13). These contribute to reheating (when $h$ is in the final state) and dark matter production (when $X$ is in the final state). Figure 1: Feynman diagram for the 4-point interactions between the inflaton $\phi$, the dark matter scalar candidate $X$, and the Higgs boson $h$, given by the Lagrangian (13). Finally, in evaluating the cosmological parameters, it is important to stay within the validity of the low-energy theory. The cutoff of the theory can be estimated as (see, e.g., higgscons2 ) $\Lambda\sim\frac{M_{P}}{\text{max}_{i}\>|\xi_{i}|}\,.$ (15) In particular, the temperature of reheating must not exceed $\Lambda$. ### II.2 Graviton exchange Let us first consider the case of vanishing $\xi_{\phi,\,h,\,X}$, i.e., the case of the minimal coupling of the scalar fields to gravity MO ; CMOV ; ema ; Garny:2015sjg ; Tang:2017hvq ; Chianese:2020yjo ; Redi:2020ffc . It was argued in MO ; CMOV that the interaction between the dark and visible sectors induced by gravity leads to unavoidable contributions to reheating and dark matter production, in the thermal bath or via the scattering of the inflaton condensate, through the graviton exchange processes shown in Fig. 2. It is therefore important to compare the minimal gravitational particle production to similar processes obtained from the Lagrangian in Eq. (13) with non-minimal couplings. Figure 2: Feynman diagram for the (dark) matter production through the gravitational scattering of the inflaton or the Higgs boson from the thermal bath. To study the universal gravitational interactions in minimally coupled gravity, we expand the space-time metric around flat space using $g_{\mu\nu}\simeq\eta_{\mu\nu}+2h_{\mu\nu}/M_{P}$, where $h_{\mu\nu}$ is the canonically-normalized perturbation. The gravitational interactions are characterized by the following Lagrangian, ${\cal L}_{\rm min.}=-\frac{1}{M_{P}}h_{\mu\nu}\left(T^{\mu\nu}_{h}+T^{\mu\nu}_{\phi}+T^{\mu\nu}_{X}\right)\,,$ (16) where the stress-energy tensor is given by $T^{\mu\nu}_{S}\;=\;\partial^{\mu}S\partial^{\nu}S-g^{\mu\nu}\left[\frac{1}{2}\partial^{\alpha}S\partial_{\alpha}S-V_{S}\right]\,.$ (17) Note that in this work, we consider only the Higgs field in the visible sector. Generalization to the complete spectrum of the Standard Model is straightforward, and we leave it for future work. For models with minimally coupled gravity, the processes $\phi/h(p_{1})+\phi/h(p_{2})\rightarrow{h}/X(p_{3})+{h}/X(p_{4})$ can be parametrized by $\mathcal{M}^{00}\propto M_{\mu\nu}^{0}\Pi^{\mu\nu\rho\sigma}M_{\rho\sigma}^{0}\;,$ (18) where the graviton propagator for the canonically-normalized field $h_{\mu\nu}$ with exchange momentum $k=p_{1}+p_{2}$ is given by $\Pi^{\mu\nu\rho\sigma}(k)=\frac{\eta^{\mu\rho}\eta^{\nu\sigma}+\eta^{\mu\sigma}\eta^{\nu\rho}-\eta^{\mu\nu}\eta^{\rho\sigma}}{2k^{2}}\,,$ (19) and the partial amplitude, $M_{\mu\nu}^{0}$, is given by $\displaystyle M_{\mu\nu}^{0}=\frac{1}{2}\left[p_{1\mu}p_{2\nu}+p_{1\nu}p_{2\mu}-\eta_{\mu\nu}p_{1}\cdot p_{2}-\eta_{\mu\nu}V_{S}^{\prime\prime}\right]\,,$ (20) with analogous expression for the final state in terms of outgoing momenta $p_{3,4}$ and the final state potential. In Fig. 2 we show the s-channel graviton exchange scattering obtained from the Lagrangian (16) for the production of dark matter from either the Higgs field or the inflaton condensate as well as the reheating process (the production of Higgs bosons from the inflaton condensate). ### II.3 Production rates In this work, we consider three processes: 1. A. The production of dark matter from the scattering of thermal Higgs bosons (assuming reheating is produced by inflaton decay). In this case, the dark matter is populated via a freeze-in mechanism throughout the reheating period. 2. B. The production of dark matter from direct excitations of the inflaton condensate. This process, which can be viewed as gravitational inflaton scattering, is independent of the presence of a thermal bath. 3. C. The creation of a radiative bath at the start of reheating arising from the Higgs boson production through gravitational inflaton scattering. Since such a process is unavoidable in minimally coupled gravity, it is interesting to know when such a process becomes dominant in models with non-minimal couplings $\xi_{i}$. The thermal dark matter production rate $R(T)$ for the process $hh\rightarrow XX$ can be calculated from666We include the symmetry factors associated with identical initial and final states in the definition of $|\overline{{\cal{M}}}|^{2}$, and a factor of 2 is explicitly included in the definition of the rate to account for the production of 2 identical particles. gravitino $R(T)=\frac{2\times N_{h}}{1024\pi^{6}}\int f_{1}f_{2}E_{1}\mathop{}\\!\mathrm{d}E_{1}E_{2}\mathop{}\\!\mathrm{d}E_{2}\mathop{}\\!\mathrm{d}\cos\theta_{12}\int|\overline{{\cal M}}|^{2}\mathop{}\\!\mathrm{d}\Omega_{13}\,,$ (21) where $E_{i}$ is the energy of particle $i=1,2$, $\theta_{13}$ and $\theta_{12}$ are the angles formed by momenta ${\bf{p}}_{1,3}$ and ${\bf{p}}_{1,2}$, respectively. $N_{h}=4$ is the number of internal degrees of freedom for 1 complex Higgs doublet, $|\overline{{\cal M}}|^{2}$ is the matrix amplitude squared with all symmetry factors included. This accounts for the explicit factor of 2 in the numerator of Eq. (21). The thermal distribution function of the incoming Higgs particles is given by the Bose-Einstein distribution $f_{i}=\frac{1}{e^{E_{i}/T}-1}\,.$ (22) The rate for minimal gravitational interactions from Eq. (16) was derived in CMOV ; Bernal:2018qlk . The rate we use here differs in two respects. As noted earlier, we only include Higgs scalars in the initial state whereas in CMOV ; Bernal:2018qlk , all Standard Model particle initial states were included. Secondly, we keep terms depending on the dark matter mass which had not previously been taken into account. This allows us to consider dark matter masses approaching the inflaton mass and/or the reheating temperature. For minimal (non-minimal) gravitational interactions, we find that the thermal dark matter production rate can be expressed as $\displaystyle R_{X}^{T,\,(\xi)}(T)$ $\displaystyle=$ $\displaystyle\beta_{1}^{(\xi)}\frac{T^{8}}{M_{P}^{4}}+\beta_{2}^{(\xi)}\frac{m_{X}^{2}T^{6}}{M_{P}^{4}}+\beta_{3}^{(\xi)}\frac{m_{X}^{4}T^{4}}{M_{P}^{4}}\,,$ (23) where the coefficients $\beta_{1,\,2,\,3}^{(\xi)}$ are given in Appendix B by Eqs. (84-86) (Eqs. (80-82)). The ratio of the non-minimal to minimal rate is shown in Fig. 3. However, we note that when $\xi_{i}\sim\mathcal{O}(1)$ both rates are comparable and interference effects become significant. The full coefficients $\beta_{1,\,2,\,3}$ including interference are given by Eqs. (87-89) from Appendix B. We leave the comparison of the effects on dark matter production from the two rates for the next section. Figure 3: Contours of the ratio of the dark matter production rates from the thermal bath based on non-minimal gravitational interactions to those based on minimal interactions. The ratio is displayed in the $(\xi_{h},\xi_{X})$ plane. Note that as discussed in the Introduction, negative values of $\xi_{h}$ may require new physics (such as supersymmetry) to stabilize the Higgs vacuum. The rate for dark matter produced from inflaton oscillations of the inflaton condensate for a potential of the form $V=\lambda\phi^{k}$ were considered in detail in GKMO2 ; CMOV . The time-dependent inflaton can be written as $\phi(t)=\phi_{0}(t)\mathcal{P}(t)$, where $\phi_{0}(t)$ is the time-dependent amplitude that includes the effects of redshift and $\mathcal{P}(t)$ describes the periodicity of the oscillation. The dark matter production rate is calculated by writing the potential in terms of the Fourier modes of the oscillations Ichikawa:2008ne ; Kainulainen:2016vzv ; GKMO2 ; CMOV $V(\phi)=V(\phi_{0})\sum_{n=-\infty}^{\infty}{\cal P}_{n}^{k}e^{-in\omega t}=\rho_{\phi}\sum_{n=-\infty}^{\infty}{\cal P}_{n}^{k}e^{-in\omega t}\,.$ (24) For $k=2$ (the only case considered here), the frequency of oscillation is simply, $\omega=m_{\phi}$. The rate generated by non-minimal couplings can be readily calculated using the Lagrangian (13), which leads to $R_{X}^{\phi,\,\xi}\;=\;\frac{2\times\sigma_{\phi X}^{\xi~{}2}}{\pi}\frac{\rho_{\phi}^{2}}{m_{\phi}^{4}}\Sigma_{0}^{k}\,,$ (25) where $\Sigma_{0}^{k}=\sum_{n=1}^{\infty}|{\cal P}^{k}_{n}|^{2}\sqrt{1-\frac{4m_{X}^{2}}{E_{n}^{2}}}\,,$ (26) and $E_{n}=n\omega$ is the energy of the $n$-th inflaton oscillation mode. For $k=2$, only the second Fourier mode in the sum contributes, with $\sum|\mathcal{P}^{2}_{n}|^{2}=\frac{1}{16}$. Thus, the rate becomes $R_{X}^{\phi,\,\xi}\;=\;\frac{2\times\sigma_{\phi X}^{\xi~{}2}}{16\pi}\frac{\rho_{\phi}^{2}}{m_{\phi}^{4}}\sqrt{1-\frac{m_{X}^{2}}{m_{\phi}^{2}}}\,,$ (27) where $\rho_{\phi}$ is the energy density of the inflaton and the interaction term $\sigma_{\phi X}^{\xi}$ is given in Appendix A by Eq. (75). It was shown in MO that the dark matter production rate through the exchange of a graviton, computed from the partial amplitude (18), is $R^{\phi}_{X}=\frac{2\times\rho_{\phi}^{2}}{256\pi M_{P}^{4}}\left(1+\frac{m_{X}^{2}}{2m^{2}_{\phi}}\right)^{2}\sqrt{1-\frac{m_{X}^{2}}{m_{\phi}^{2}}}\,,$ (28) which can be written in the same form as (27) by defining an effective coupling $\sigma_{\phi X}$ $\sigma_{\phi X}\;=\;-\frac{m_{\phi}^{2}}{4M_{P}^{2}}\left(1+\frac{m_{X}^{2}}{2m_{\phi}^{2}}\right)\,.$ (29) A comparison of the non-minimal to minimal rates for the production of dark matter from inflaton scattering is shown in Fig. 4. Figure 4: Contours of the ratio of the dark matter production rates from oscillations in the inflaton condensate based on non-minimal gravitational interactions to those based on minimal interactions. The ratio is displayed in the $(\xi_{\phi},\xi_{X})$ plane. For the production of Higgs bosons through inflaton condensate scattering, we follow a similar procedure, and from the Lagrangian (13) we find $R_{h}^{\phi,\,\xi}\;\simeq\;N_{h}\frac{2\times\sigma_{\phi h}^{\xi~{}2}}{16\pi}\frac{\rho_{\phi}^{2}}{m_{\phi}^{4}}\,,$ (30) where we assumed that $m_{h}\ll m_{\phi}$, $N_{h}=4$ is the number of internal degrees of freedom for 1 complex Higgs doublet, and $\sigma_{\phi h}^{\xi}$ is given in Appendix A by Eq. (76). On the other hand, it was argued in CMOV that the scattering $\phi\phi\rightarrow hh$ through the graviton exchange can also be parameterized by an effective coupling $\mathcal{L}_{h}=-\sigma_{\phi h}\phi^{2}h^{2}\,,$ (31) with $\sigma_{\phi h}\;=\;-\frac{m_{\phi}^{2}}{4M_{P}^{2}}\,,$ (32) and the rate $R_{h}^{\phi}$ is given by the analogous expression to (30) with $\sigma_{\phi h}^{\xi}$ replaced by $\sigma_{\phi h}$. The full four-point coupling of course is given by the sum $\sigma_{\phi h/X}^{\xi}+\sigma_{\phi h/X}$. However, except for values where the two are similar, which occurs when $12\xi^{2}+5\xi\simeq\frac{1}{2}$ (assuming $m_{X}\ll m_{\phi}$ and taking the $\xi_{i}$ to be equal to $\xi$), either the minimal or the non-minimal contribution dominates. Thus, for the most part, we will consider separately the minimal and non-minimal contributions. Note that for two values of $\xi$ ($\xi\sim-1/2$ and 1/12) destructive interference could occur causing the entire rate to vanish (at the tree level). ## III Particle Production with a Non-Minimal Coupling Given the rates $R_{i}^{j}$ calculated in the previous section, we compute the evolution for the gravitational (minimal and non-minimal) contribution to the reheating processes and the dark matter density for the three reactions outlined above. ### III.1 $h~{}h\rightarrow X~{}X$ The gravitational scattering of thermal Higgs bosons leads to the production of massive scalar dark matter particles $X$. The dark matter number density $n_{X}$ can be calculated from the classical Boltzmann equation $\frac{dn_{X}}{dt}+3Hn_{X}=R^{T}_{X}\,,$ (33) where $H=\frac{\dot{a}}{a}$ is the Hubble parameter and the right-hand side of the equation represents the dark matter production rate. It is more practical to rewrite the above equation in terms of the scale factor $a$ rather than the parameters $t$ or $T$. We proceed by introducing the comoving number density $Y_{X}=na^{3}$ and rewriting the Boltzmann equation as $\frac{dY_{X}}{da}=\frac{a^{2}R^{T}_{X}(a)}{H(a)}\,.$ (34) Since the production rate (23) is a function of the temperature of the thermal bath, it is necessary to determine the relation between $T$ and $a$ in order to solve the Boltzmann equation as a function of the scale factor $a$. For the Starobinsky potential in Eq. (5), at the end of inflation, the inflaton starts oscillating about a quadratic minimum, and we find the following energy conservation equations777For the inflaton scattering with $V(\phi)\sim\phi^{k}$, where $k>2$, see Bernal:2019mhf ; GKMO1 ; GKMO2 ; GKMOV ; Haque:2021mab ; Haque:2022kez ; Ahmed:2021fvt . $\displaystyle\frac{d\rho_{\phi}}{dt}+3H\rho_{\phi}=-\Gamma_{\phi}\rho_{\phi}\,,$ (35) $\displaystyle\frac{d\rho_{R}}{dt}+4H\rho_{R}=\Gamma_{\phi}\rho_{\phi}\,,$ (36) where $\rho_{\phi}$ and $\rho_{R}$ are the energy density of the inflaton and radiation, respectively, $\Gamma_{\phi}$ is the inflaton decay rate, and for a quadratic minimum, we are able to set the equation of state parameter $w_{\phi}=\frac{P_{\phi}}{\rho_{\phi}}\simeq 0$. We will assume that reheating occurs due to an effective inflaton coupling to the Standard Model fermions, given by the interaction Lagrangian $\mathcal{L}_{\phi-SM}^{y}\;=\;-y\phi\bar{f}f\,,$ (37) where $y$ is a Yukawa-like coupling, $f$ is a Standard Model fermion, and the inflaton decay rate is $\Gamma_{\phi}\;=\;\frac{y^{2}}{8\pi}m_{\phi}\,.$ (38) If we solve the Friedmann equations (35, 36), we find GKMO1 ; GKMO2 ; CMOV $\rho_{\phi}(a)=\rho_{\rm end}\left(\frac{a_{\rm end}}{a}\right)^{3}$ (39) and $\rho_{R}(a)=\rho_{\rm RH}\left(\frac{a_{\rm RH}}{a}\right)^{\frac{3}{2}}\frac{1-\left(\frac{a_{\rm end}}{a}\right)^{\frac{5}{2}}}{1-\left(\frac{a_{\rm end}}{a_{\rm RH}}\right)^{\frac{5}{2}}}\,,$ (40) where $a_{\rm{end}}$ is the scale factor at the end of inflation, $\rho_{\rm end}\equiv\rho_{\phi}(a_{\rm{end}})$ is the inflaton energy density at the end of inflation when there is no radiation present, $a_{\rm{RH}}$ is the scale factor at reheating, and $\rho_{\rm RH}\equiv\rho_{R}(a_{\rm{RH}})=\rho_{\phi}(a_{\rm{RH}})$ is the energy density at reheating. We note that these equations are strictly valid for $a_{\rm end}\ll a\ll a_{\rm{RH}}$ and the end of inflation occurs when $\ddot{a}=0$ which corresponds to $\rho_{\rm end}=\frac{3}{2}V(\phi_{\rm{end}})$. For the Starobinsky potential, $\rho_{\rm end}\simeq 0.175m_{\phi}^{2}M_{P}^{2}$ egno5 . The radiation energy density can be parameterized as $\rho_{R}=\frac{g_{T}\pi^{2}}{30}T^{4}\equiv\alpha T^{4}\,,$ (41) where $g_{T}$ is the number of relativistic degrees of freedom at the temperature $T$. The maximum temperature is attained when the radiation energy density reaches its peak at $\rho_{R}(a_{\rm{max}})=\alpha T_{\rm{max}}^{4}$. It was shown in GKMO1 that the ratio of $a_{\rm max}$ to $a_{\rm end}$ is given by $\frac{a_{\rm max}}{a_{\rm end}}=\left(\frac{8}{3}\right)^{\frac{2}{5}}\simeq 1.48\,.$ (42) Using Eq. (40) we can then express the production rate from gravitational scattering of thermal particles (23) as a function of the scale factor $a$ $R_{X}^{T,\,(\xi)}(a)\simeq\beta_{1}^{(\xi)}\frac{\rho_{\rm RH}^{2}}{\alpha^{2}M_{P}^{4}}\left(\frac{a_{\rm RH}}{a}\right)^{3}\left[\frac{1-\left(\frac{a_{\rm end}}{a}\right)^{\frac{5}{2}}}{1-\left(\frac{a_{\rm end}}{a_{\rm RH}}\right)^{\frac{5}{2}}}\right]^{2}\,,$ (43) where we assumed that $m_{X}\ll m_{\phi},\,T$, and thus neglected the terms $\beta_{2,\,3}^{(\xi)}$. If we use $H\simeq\frac{\sqrt{\rho_{\phi}(a)}}{\sqrt{3}M_{P}}$, which is valid for $a\ll a_{\rm RH}$, we can rewrite Eq. (34) as $\frac{dY_{X}^{\xi}}{da}=\frac{\sqrt{3}M_{P}}{\sqrt{\rho_{\rm RH}}}a^{2}\left(\frac{a}{a_{\rm RH}}\right)^{\frac{3}{2}}R_{X}^{T,\,(\xi)}(a)\,.$ (44) We find that the solution to this equation is $\displaystyle~{}~{}n^{T,\,\xi}_{X}(a_{\rm RH})=\frac{2\beta_{1}^{\xi}}{\sqrt{3}\alpha^{2}M_{P}^{3}}\frac{\rho_{\rm RH}^{3/2}}{(1-(a_{\rm end}/a_{\rm RH})^{\frac{5}{2}})^{2}}\times$ $\displaystyle\left(1+3\left(\frac{a_{\rm end}}{a_{\rm RH}}\right)^{\frac{5}{2}}-\frac{25}{7}\left(\frac{a_{\rm end}}{a_{\rm RH}}\right)^{\frac{3}{2}}-\frac{3}{7}\left(\frac{a_{\rm end}}{a_{\rm RH}}\right)^{5}\right)\,,$ (45) where we integrated Eq. (44) in the interval $a_{\rm end}<a<a_{\rm RH}$. The relic abundance is given by book $\Omega_{X}h^{2}=1.6\times 10^{8}\frac{g_{0}}{g_{\rm RH}}\frac{n(T_{\rm RH})}{T_{\rm RH}^{3}}\frac{m_{X}}{1~{}{\rm GeV}}\,,$ (46) and if we combine it with Eq. (III.1), we obtain $\displaystyle\Omega_{X}^{T,\,(\xi)}h^{2}\;=\;\frac{2}{3}\Omega_{k}^{(\xi)}$ $\displaystyle\left[1+3\left(\frac{\rho_{\rm{RH}}}{\rho_{\rm{end}}}\right)^{\frac{5}{6}}-\frac{25}{7}\left(\frac{\rho_{\rm{RH}}}{\rho_{\rm{end}}}\right)^{\frac{1}{2}}\right.$ (47) $\displaystyle-\left.\frac{3}{7}\left(\frac{\rho_{\rm{RH}}}{\rho_{\rm{end}}}\right)^{\frac{5}{3}}\right]\,,$ with $\Omega_{k}^{(\xi)}=1.6\times 10^{8}\frac{g_{0}}{g_{\rm RH}}\frac{\beta_{1}^{(\xi)}\sqrt{3}}{\sqrt{\alpha}}\frac{m_{X}}{1~{}\rm{GeV}}\frac{T_{\rm RH}^{3}}{M_{P}^{3}}\left[1-\left(\frac{\rho_{\rm RH}}{\rho_{\rm end}}\right)^{\frac{5}{6}}\right]^{-2},$ (48) where $g_{0}=43/11$ and we use the Standard Model value $g_{\rm RH}=427/4$. We observe that $\Omega_{X}^{T,\,\xi}\propto\beta_{1}^{\xi}\,T_{\rm{RH}}^{3}$. Therefore large values of the couplings $\xi_{h}$ and $\xi_{X}$ would require a decrease in the reheating temperature. In Section IV we compare the scattering rates and the dark matter abundances with the minimally coupled case. ### III.2 $\phi~{}\phi\rightarrow X~{}X$ Another mode of dark matter production is through the scattering of the inflaton itself. Whereas the graviton exchange channel was treated with care in MO ; CMOV , in the case of non-minimal coupling it suffices to replace $R_{X}^{T,\,\xi}$ in Eq. (44) with the production rate (27), $\frac{dY_{X}^{\xi}}{da}=\frac{\sqrt{3}M_{P}}{\sqrt{\rho_{\rm RH}}}a^{2}\left(\frac{a}{a_{\rm RH}}\right)^{\frac{3}{2}}R_{X}^{\phi,\,\xi}(a)\,,$ (49) and to integrate between $a_{\rm end}$ and $a_{\rm RH}$, which leads to $n_{X}^{\phi,\,\xi}(a_{\rm RH})=\frac{\sigma_{\phi X}^{\xi~{}2}\rho_{\rm{RH}}^{3/2}M_{P}}{4\sqrt{3}\pi m_{\phi}^{4}}\left[\left(\frac{a_{\rm{RH}}}{a_{\rm{end}}}\right)^{\frac{3}{2}}-1\right]\sqrt{1-\frac{m_{X}^{2}}{m_{\phi}^{2}}}\,.$ (50) For $a_{\rm RH}\gg a_{\rm end}$, using Eq. (39) we can express $n_{X}^{\phi,\,\xi}$ as a function of $\rho_{\rm end}$: $n_{X}^{\phi,\,\xi}(a_{\rm RH})\simeq\frac{\sigma_{\phi X}^{\xi~{}2}\rho_{\rm RH}\sqrt{\rho_{\rm end}}M_{P}}{4\sqrt{3}\pi m_{\phi}^{4}}\sqrt{1-\frac{m_{X}^{2}}{m_{\phi}^{2}}}\,,$ (51) and we find $\displaystyle\frac{\Omega_{X}^{\phi,\,\xi}h^{2}}{0.12}\simeq\frac{1.3\times 10^{7}\sigma_{\phi X}^{\xi~{}2}\rho_{\rm{RH}}^{1/4}M_{P}^{2}}{m_{\phi}^{3}}\frac{m_{X}}{1\,\rm{GeV}}\sqrt{1-\frac{m_{X}^{2}}{m_{\phi}^{2}}}\,,$ (52) where we assumed the Starobinsky value for $\rho_{\rm end}$. The analogous expression for models with minimally coupled gravity is found by replacing $\sigma_{\phi X}^{\xi}\rightarrow\sigma_{\phi X}$. Up to this point we have assumed that the radiation is produced via the direct inflaton decay to a fermion pair. In the next subsection we discuss an unavoidable radiation production channel when the inflaton condensate scattering produces Higgs bosons in models with minimal and non-minimal coupling to gravity. ### III.3 $\phi~{}\phi\rightarrow h~{}h$ Gravitational processes that produce dark matter can also populate the thermal bath in the same way. Even if this Planck-suppressed production mechanism does not dominate throughout the entire reheating process, it was shown in CMOV that for $T_{\rm RH}\lesssim 10^{9}$ GeV it is graviton exchange that dominates the production of the thermal bath at the very beginning of the reheating, when $\rho_{\phi}\sim\rho_{\rm end}$. In fact, it was shown that the maximal temperature reached, $T_{\rm max}$, (which can be considered as an absolute lower bound on $T_{\rm max}$) is $T_{\rm max}\sim 10^{12}$ GeV. It is therefore natural to determine the value of the couplings ($\xi_{\phi}$, $\xi_{h}$), for which non-minimal gravitational processes generate the thermal bath at early times, and the maximal temperature which can be attained by these processes. Following the discussion in the previous subsection, to compute the radiation energy density produced by gravitational couplings we implement the rate $R_{h}^{\phi,\,\xi}$ (30) into the Friedmann equation (36) $\frac{d\rho_{R}}{dt}+4H\rho_{R}\;\simeq\;N_{h}\frac{\sigma_{\phi h}^{\xi~{}2}}{8\pi}\frac{\rho_{\phi}^{2}}{m_{\phi}^{3}}\,,$ (53) where we took into account that each scattering corresponds to an energy transfer of $2m_{\phi}$.888Or equivalently that each Higgs quanta carries an energy $m_{\phi}$. The solution to this equation is $\rho_{R}\;=\;N_{h}\frac{\sqrt{3}\sigma_{\phi h}^{\xi~{}2}}{4\pi}\frac{\rho_{\rm end}^{3/2}M_{P}}{m_{\phi}^{3}}\left[\left(\frac{a_{\rm{end}}}{a}\right)^{4}-\left(\frac{a_{\rm{end}}}{a}\right)^{\frac{9}{2}}\right]\,.$ (54) Note that the dependence on the scale factor $a$ is very different from that found in Eq. (40) due to inflaton decay. Indeed, the Higgs bosons produced by gravitational scattering (minimal as well as non-minimal) are redshifted to a greater extent because of the high dependence of the rate on their energy due to the form of the energy-momentum tensor $T_{\mu\nu}^{0}$. Since $\rho_{R}\propto a^{-4}$ in Eq. (54) (at large $a$) and $\rho_{\phi}\propto a^{-3}$ in Eq. (39), reheating through this process does not occur (i.e., $\rho_{R}$ never comes to dominate the total energy at late times) and inflaton decay is necessary.999This conclusion is avoided if the inflaton potential about minimum is approximated by $\phi^{k}$ with a higher power of $k>4$ CMOV ; Haque:2022kez . However, as in the case of the reheating from the inflaton decay, the energy density in Eq. (54) exhibits a maximum when $a=a_{\rm max}=(81/64)a_{\rm end}$. The maximum radiation density is then, $\rho_{\rm{max}}^{\xi}\;\simeq\;N_{h}\frac{\sigma_{\phi h}^{\xi~{}2}}{12\sqrt{3}\pi}\frac{\rho_{\rm end}^{3/2}M_{P}}{m_{\phi}^{3}}\left(\frac{8}{9}\right)^{8}\,,$ (55) and from this expression we find that the maximum temperature produced by gravitational interactions is given by $\displaystyle T_{\rm{max}}^{\xi}$ $\displaystyle\simeq$ $\displaystyle 6.5\times 10^{11}\left(\frac{|\sigma_{\phi h}^{\xi}|}{10^{-11}}\right)^{\frac{1}{2}}{\rm GeV}$ $\displaystyle\simeq$ $\displaystyle 1.8\times 10^{12}\sqrt{|\xi|}\left(|5+12\xi|\right)^{\frac{1}{2}}\left(\frac{m_{\phi}}{3\times 10^{13}\,{\rm GeV}}\right){\rm GeV}\,,$ where we took $\xi_{\phi}=\xi_{h}=\xi$ in the last equality. The analogous expression for models with minimally coupled gravity is found by replacing $\sigma_{\phi h}^{\xi}\rightarrow\sigma_{\phi h}$. To compare the maximum temperature obtained by non-minimal interactions with respect to minimal gravitational interactions, we can rewrite Eq. (III.3) (now including minimal interactions in $T_{\rm max}^{\xi}$) as $T_{\rm{max}}^{\xi}\simeq 1.3\times 10^{12}\left(\frac{|\sigma_{\phi h}^{\xi}+\sigma_{\phi h}|}{\sigma_{\phi h}}\right)^{\frac{1}{2}}{\rm GeV}\,.$ (57) The value of $\xi$ for which the maximum temperature generated by the non- minimal coupling surpasses the one from graviton exchange is shown in Fig. 5 and is determined using $\sqrt{\frac{|\sigma_{\phi h}^{\xi}|}{|\sigma_{\phi h}|}}=\sqrt{2|\xi|}\left(|5+12\xi|\right)^{\frac{1}{2}}>1$ (58) which is satisfied when $\xi>1/12$ or $\xi<-1/2$, as discussed earlier. Figure 5: The maximum temperature during reheating generated separately by minimal and non-minimal gravitational scattering of Higgs bosons in the thermal bath. As noted above and discussed in CMOV , minimal (and non-minimal) gravitational interactions for a quadratic inflaton potential do not lead to the completion of the reheating process, thus requiring additional inflaton interactions for decay. Although radiation density produced in scattering falls off faster than that from decay, at early time, the radiation density may in fact dominate and determine $T_{\rm max}$. To determine when the $\phi~{}\phi\rightarrow h~{}h$ process leads to the maximum temperature, we rewrite Eq. (40) as: $\rho^{y}_{R}\;=\;\frac{\sqrt{3}y^{2}m_{\phi}M_{P}^{3}}{20\pi}\left(\frac{\rho_{\rm end}}{M_{P}^{4}}\right)^{\frac{1}{2}}\left[\left(\frac{a_{\rm end}}{a}\right)^{\frac{3}{2}}-\left(\frac{a_{\rm end}}{a}\right)^{4}\right]\,.$ (59) Using Eq. (42), we find that the maximum radiation density produced by the inflaton decay is given by $\rho_{\rm max}^{y}\;=\;\frac{\sqrt{3}y^{2}m_{\phi}M_{P}^{3}}{32\pi}\left(\frac{\rho_{\rm end}}{M_{P}^{4}}\right)^{\frac{1}{2}}\left(\frac{3}{8}\right)^{\frac{3}{5}}\,.$ (60) The maximum temperature is therefore determined by (non-minimal) gravitational interactions when $y^{2}\lesssim N_{h}\frac{8\rho_{\rm end}\sigma_{\phi h}^{\xi~{}2}}{9m_{\phi}^{4}}\left(\frac{8}{9}\right)^{8}\left(\frac{8}{3}\right)^{\frac{3}{5}}$ (61) or $y\lesssim 1.6~{}\sigma_{\phi h}^{\xi}\sqrt{\frac{\rho_{\rm end}}{m_{\phi}^{4}}}\simeq 5.4\times 10^{4}~{}\sigma_{\phi h}^{\xi}\left(\frac{3\times 10^{13}\,\rm{GeV}}{m_{\phi}}\right)\,.$ (62) This leads to the following reheating temperature: $\displaystyle T_{\rm{RH}}\;$ $\displaystyle\lesssim$ $\displaystyle\;3.1\times 10^{19}\sigma_{\phi h}^{\xi}\left(\frac{m_{\phi}}{3\times 10^{13}\,\rm{GeV}}\right)^{-1/2}{\rm GeV}$ (63) $\displaystyle\lesssim$ $\displaystyle 2.4\times 10^{9}\left(\frac{m_{\phi}}{3\times 10^{13}}\right)^{\frac{3}{2}}\xi(5+12\xi)~{}{\rm GeV}$ where $T_{\rm RH}$ is given by GKMO2 $\rho_{\phi}(a_{\rm RH})=\alpha T_{\rm RH}^{4}=\frac{12}{25}\Gamma_{\phi}^{2}M_{P}^{2}=\frac{3y^{4}m_{\phi}^{2}M_{P}^{2}}{400\pi^{2}}\,,$ (64) when the reheating temperature is determined by inflaton decay. The primary effect of the gravitational scattering processes on reheating is the augmentation of $T_{\rm max}$ for sufficiently small inflaton decay coupling, $y$. This can be seen in Fig. 6 where we show the evolution of the energy density of radiation from scattering and decay as well as the energy density of the inflaton as a function of $a/a_{\rm end}$ for $\sigma_{\phi h}^{\xi}=0$ and $\sigma_{\phi h}^{\xi}/\sigma_{\phi h}=100$, respectively. Figure 6: Evolution of the inflaton density (blue) and the total radiation density (red), with radiation density produced from inflaton decays (dashed orange) and $\phi~{}\phi\rightarrow h~{}h$ scattering processes $\rho_{R}^{\sigma,\,\xi}$ (dotted green) and $\rho_{R}^{\sigma}$ (dash-dotted purple) with $\sigma_{\phi h}^{\xi}/\sigma_{\phi h}=100$ (or $\xi_{\phi}=\xi_{h}=\xi\simeq-2.3~{}{or}~{}1.8$), as a function of $a/a_{\rm end}$ for a Yukawa-like coupling $y=10^{-8}$ and $\rho_{\rm end}\simeq 0.175\,m_{\phi}^{2}M_{P}^{2}$ $\simeq 9\times 10^{62}\,\rm{GeV}^{4}$. The black dashed lines corresponds to the ratios $a_{\rm{int}}/a_{\rm{end}}\simeq 150$ and $6500$, which agrees with Eq. (66). The numerical solutions are obtained from Eqs. (35), (36), and (53). As we saw in Eq. (58), minimal gravitational interactions dominate over non- minimal interactions when $\sigma_{\phi h}^{\xi}<\sigma_{\phi h}$ or when $12\xi_{\phi}\xi_{h}+3\xi_{h}+2\xi_{\phi}<\frac{1}{2}\,,$ (65) when we neglect contributions proportional to the Higgs mass. In this case, the maximum temperature is determined by gravitational interactions when $y\lesssim 2.1\times 10^{-6}$ from Eq. (62) using $\sigma_{\phi h}$ from Eq. (32). The evolution of the energy densities in this case is shown in Fig. 6 with $y=10^{-8}$. However as the energy density of radiation after the maximum falls faster than $\rho_{\phi}$, reheating in the Universe is determined by the inflaton decay. For a sufficiently small coupling $y$, the energy density from the decay dominates the radiation density at $a>a_{\rm int}$, where $\frac{a_{\rm int}}{a_{\rm end}}\simeq\left(\frac{5\sigma_{\phi h}^{2}N_{h}\rho_{\rm end}}{y^{2}m_{\phi}^{4}}\right)^{2/5}\simeq 1.6\left(\frac{\sigma_{\phi h}M_{P}}{ym_{\phi}}\right)^{4/5}\,.$ (66) For $\sigma_{\phi h}=3.8\times 10^{-11}$, $m_{\phi}=3\times 10^{13}$ GeV, and $y=10^{-8}$ we have $a_{\rm int}\approx 160a_{\rm end}$, as seen in the figure. When Eq. (65) is not satisfied, non-minimal interactions may dominate as shown in the bottom panel of Fig. 6, for $\sigma_{\phi h}^{\xi}=100\sigma_{\phi h}$ and $y=10^{-8}$. The cross-over can be determined from Eq. (66) with the replacement $\sigma_{\phi h}\rightarrow\sigma_{\phi h}^{\xi}$. In this example, $a_{\rm int}\approx 6500a_{\rm end}$. ## IV Results We now turn to some general results that may be obtained from the framework described above. Concerning the gravitational production of dark matter from the thermal bath, the difficulty of populating the Universe via the exchange of a graviton was already known Bernal:2018qlk ; CMOV . Summing the minimal and non-minimal contributions in Eq. (47), we find for $\rho_{\rm RH}\ll\rho_{\rm end}$ $\displaystyle\frac{\Omega_{X}^{T}}{0.12}$ $\displaystyle\simeq$ $\displaystyle\left[1+30f(\xi_{h},\xi_{X})\right]\left(\frac{T_{\rm RH}}{10^{14}~{}{\rm GeV}}\right)^{3}\left(\frac{m_{X}}{4.0\times 10^{9}~{}{\rm GeV}}\right)$ (67) $\displaystyle=$ $\displaystyle\left[1+120\xi^{2}(1+6\xi+12\xi^{2})\right]$ $\displaystyle\times\left(\frac{T_{\rm RH}}{10^{14}~{}{\rm GeV}}\right)^{3}\left(\frac{m_{X}}{4.0\times 10^{9}~{}{\rm GeV}}\right)$ with $f(\xi_{h},\xi_{X})=\xi_{h}^{2}+2\xi_{h}\xi_{X}+\xi_{X}^{2}+12\xi_{h}\xi_{X}\left(\xi_{h}+\xi_{X}+4\xi_{h}\xi_{X}\right)$ where we assumed $\xi_{h}=\xi_{X}=\xi$ in the last equality, for simplicity. It is clear that, if we set $\xi=0$, i.e. if we consider only graviton exchange, the reheating temperature necessary to obtain a reasonable density respecting the data Planck is dangerously close to the mass of the inflaton, even for extremely large dark matter masses. This problem had already been raised in Bernal:2018qlk and resolved in MO ; CMOV by considering the dark matter produced from the (minimal) gravitational inflaton scattering. On the other hand, from Eq. (67) we see that there is another solution to this tension if one allows for non-minimal gravitational couplings. Indeed, it is easy to see that for values of $\xi_{i}\gtrsim 0.1$ ($f(\xi_{h},\xi_{X})\gtrsim{\frac{1}{30}}$), non-minimal gravitational production dominates over graviton exchange. In this case, it becomes easier to obtain the correct dark matter density for more reasonable values of $T_{\rm RH}$ and/or $m_{X}$. For example, for a common value $\xi=\xi_{h}=\xi_{X}=1$, a temperature of $T_{\rm RH}\sim 1.2\times 10^{13}$ GeV, thus slightly below the inflaton mass, is sufficient to produce an EeV dark matter candidate, whereas for $\xi=1000$, $T_{\rm RH}\sim 10^{11}$ GeV will saturate the relic density for a 2.6 TeV dark matter mass. We show this result in Fig. 7 where we plot the reheating temperature needed to satisfy the relic density constraint as function of $m_{X}$ for different value of $\xi$. For each value of $\xi$, the relic density exceeds $\Omega_{X}h^{2}=0.12$ above the corresponding curve. As one can see, the line for $\xi=0$ is in the upper corner of the figure at high values of $T_{\rm RH}$ and $m_{X}$ and these drop significantly at higher values of $\xi$. Figure 7: Region of parameter space respecting the relic density constraint $\Omega_{X}h^{2}=0.12$ in the plane ($m_{X}$,$T_{\rm RH}$) for different values of $\xi=\xi_{h}=\xi_{X}$ and $\rho_{\rm end}\simeq 0.175\,m_{\phi}^{2}M_{P}^{2}$ in the case of gravitational production from the thermal bath $h~{}h\rightarrow X~{}X$. Both minimal and non-minimal contributions are taken into account. As was shown in MO ; CMOV , another possibility to avoid the necessity of high reheating temperatures and/or dark matter masses is the production of matter from the oscillations within the inflaton condensate when the energy stored in the condensate is much larger than the reheating temperature. A simple comparison between Eqs. (47) and (52) shows that the production of dark matter via inflaton scattering when $\xi_{i}\neq 0$ generally dominates over the production of dark matter from the thermal bath: $\displaystyle\frac{\Omega_{X}^{\phi,\,\xi}}{\Omega_{X}^{T,\,\xi}}$ $\displaystyle\simeq$ $\displaystyle 34\frac{(\sigma_{\phi X}^{\xi})^{2}}{\beta_{1}^{\xi}}\frac{M_{P}^{5}}{T_{\rm RH}^{2}m_{\phi}^{3}}$ (68) $\displaystyle\simeq$ $\displaystyle 185\frac{M_{P}m_{\phi}}{T_{\rm RH}^{2}}\frac{(5+12\xi)^{2}}{1+6\xi+12\xi^{2}}\gg 1\,,\ $ where we took $\xi=\xi_{\phi}=\xi_{h}=\xi_{X}$ and $m_{X}\ll m_{\phi}$ in the last equality. We are therefore able to state that the relic density of dark matter generated by the non-minimal gravitational scattering of the inflaton is always much more abundant than that produced by the thermal bath. Dark matter production from inflaton scattering via minimal graviton exchange also dominates over minimal gravitational thermal production CMOV . This state of affairs is anything but surprising. Indeed, the energy available in the inflaton condensate at the onset of oscillations is much greater than that available in the thermal bath during the reheating process. As the scattering cross-sections are themselves highly dependent on the energies through the energy-momentum tensor, it is quite normal that inflaton scattering is the dominant process for both minimal and non-minimal gravitational couplings. Since inflaton scattering dominates in both the minimal and non-minimal gravitational interactions we can compare the two. We obtain $\frac{\Omega_{X}^{\phi,\,\xi}}{\Omega_{X}^{\phi}}=\frac{\sigma_{\phi X}^{\xi~{}2}}{\sigma_{\phi X}^{2}}\simeq 4\xi^{2}(5+12\xi)^{2}\,,$ (69) and we see again that non-minimal interactions dominate when $\xi>1/12$ or $<-1/2$. We show in Fig. 8 the region of the parameter space in the ($m_{X}$, $T_{\rm RH}$) plane allowed by the relic density constraint, adding all of the minimal and non minimal gravitational contributions, from inflaton scattering and as well as Higgs scattering from the thermal bath taking $\xi_{\phi}=\xi_{h}=\xi_{X}=\xi$. As expected, for $\xi=0$ we recover the result found in CMOV . As one can see, the difficulty in the gravitational production from the thermal bath is indeed alleviated as a reheating temperature $T_{\rm RH}\simeq 10^{11}$ GeV allows for the production of a PeV scale dark matter candidate. If in addition we introduce the non-minimal couplings $\xi$, the necessary reheating temperature to fit the Planck data may be as low as the electroweak scale for a GeV candidate if $\xi\gtrsim 1000$. Figure 8: Region of parameter space respecting the relic density constraint $\Omega_{X}h^{2}=0.12$ in the plane ($m_{X}$,$T_{\rm RH}$) for different values of $\xi_{\phi}=\xi_{h}=\xi_{X}=\xi$ and $\rho_{\rm end}\simeq 0.175\,m_{\phi}^{2}M_{P}^{2}$ in the case of production from gravitational inflaton scattering $\phi~{}\phi\rightarrow X~{}X$. Both minimal and non- minimal contributions are taken into account. Finally, we note that given the dark matter mass and reheating temperature (if that sector of beyond the Standard Model physics were known), the contours in Fig. 8 allow us to place an upper bound on the non-minimal couplings, $\xi$. We can rewrite Eq. (52) as $\displaystyle\frac{\Omega_{X}h^{2}}{0.12}$ $\displaystyle=$ $\displaystyle 4.1\times 10^{-7}(12\xi^{2}+5\xi+\frac{1}{2})^{2}\left(\frac{T_{\rm RH}}{10^{10}{\rm GeV}}\right)$ (70) $\displaystyle\times\left(\frac{m_{X}}{1{\rm GeV}}\right)\left(\frac{m_{\phi}}{3\times 10^{13}{\rm GeV}}\right)\,,$ when $m_{X}\ll m_{\phi}$ and $\xi=\xi_{\phi}=\xi_{X}$. Then, for example, if $m_{X}=1$ TeV, and $T_{\rm RH}=10^{9}$ GeV, we obtain an upper limit of $|\xi|\lesssim 4$. ## V Conclusions In this paper, we have generalized the minimal gravitational interactions in the early Universe, i.e., the s-channel exchange of a graviton, to include non-minimal couplings of all scalars to the Ricci curvature $R$. We consider a scalar sector $S_{i}$ consisting of the inflaton condensate $\phi$, the Higgs field $H$ and a dark matter candidate $X$, and we have analyzed the impact of couplings of the type $\xi_{i}S_{i}^{2}R$ on the reheating process and dark matter production. The latter can be generated by the thermal Higgs scattering or excitations of the inflaton, both through minimal and non-minimal gravitational couplings. Whereas the Higgs scattering through the exchange of a graviton necessitates a very large reheating temperature and/or dark matter mass in order to fulfill Planck CMB constraints ($T_{\rm RH}\simeq 10^{14}$ GeV with $m_{X}\simeq 10^{9}$ GeV), for $\xi\gtrsim 0.1$, the non-minimal coupling dominates the process and alleviates the tension. For $\xi\simeq 1000$, a dark matter mass of $\sim 1$ PeV with $T_{\rm RH}\simeq 10^{10}$ GeV will satisfy the constraint, see Fig. 7. However, thermal production is not the sole source of dark matter production through gravity. When we include the contribution (necessarily present) of the inflaton scattering, we showed that the energy stored in the condensate at the end of inflation compensates largely the reduced gravitational Planck coupling. These processes yield the correct relic abundance through minimal graviton exchange for a dark matter mass of $\sim 10^{8}$ GeV with $T_{\rm RH}\simeq 10^{10}$ GeV, and the constraint is satisfied for a dark matter mass of $\sim 100$ GeV and $T_{\rm RH}\gtrsim 10^{4}$ GeV if one adds non-minimal couplings of the order $\xi\simeq 100$ as we show in Fig. 8. Gravitational inflaton scattering also affects the reheating process, producing a maximum temperature $\simeq 10^{12}$ GeV with minimal couplings, reaching as large as $T_{\rm max}^{\xi}\simeq 5|\xi|T_{\rm max}\simeq 10^{14}$ GeV for $\xi=100$ as one can see in Fig. 5. This result can be re-expressed as an upper limit to $|\xi|$ given values of $m_{X}$ and $T_{\rm RH}$. We can not over-emphasize that all of our results are unavoidable, in the sense that they are purely gravitational, and do not rely on physics beyond the Strandard Mode. The relic density of dark matter, and maximum temperature of the thermal bath computed here should be considered as lower bounds, that should be implemented in any extension of the Standard Model, whatever is its nature. Note added : During the completion of the manuscript, some overlapping results were presented in Aoki:2022dzd . Acknowledgements. The authors want to thank Emilian Dudas for useful discussions. This work was made possible by with the support of the Institut Pascal at Université Paris-Saclay during the Paris-Saclay Astroparticle Symposium 2021, with the support of the P2IO Laboratory of Excellence (program “Investisse ments d’avenir” ANR-11-IDEX-0003-01 Paris-Saclay and ANR-10-LABX-0038), the P2I axis of the Graduate School Physics of Université Paris-Saclay, as well as IJCLab, CEA, IPhT, APPEC, the IN2P3 master projet UCMN and EuCAPT ANR-11-IDEX-0003-01 Paris-Saclay and ANR-10-LABX-0038). This project has received support from the European Union’s Horizon 2020 research and innovation programme under the Marie Sk $-$ ${\rm l}$ odowska-Curie grant agreement No 860881-HIDDeN. The work of K.A.O. and A.S. was supported in part by DOE grant DE-SC0011842 at the University of Minnesota. ## Appendix ### A PARTICLE PRODUCTION WITH A NON-MINIMAL COUPLING The full Jordan frame action we consider is given by Eq. (2). The conformal transformation to the Einstein frame is given by $g_{\mu\nu}\;=\;\Omega^{2}\tilde{g}_{\mu\nu}\,,$ (71) where $g_{\mu\nu}$ is the Einstein frame spacetime metric and the conformal factor is expressed by Eq. (3). It can readily be shown that the scalar curvature transforms as (see, e.g., Fujii:2003pa ) $\tilde{R}=\Omega^{2}\left[R+6g^{\mu\nu}\nabla_{\mu}\nabla_{\nu}\ln\Omega-6g^{\mu\nu}\left(\nabla_{\mu}\ln\Omega\right)\left(\nabla_{\nu}\ln\Omega\right)\right]\,.$ (72) After eliminating the total divergence term, we find the Einstein frame action (9). To find the effective interaction terms we assume the small field limit (12) and expand the conformal factors in the Einstein frame action. We find the following effective interaction Lagrangian: $\displaystyle\mathcal{L}_{\rm{eff}}\;$ $\displaystyle=$ $\displaystyle\;-\frac{1}{2}\left(\frac{\xi_{\phi}\phi^{2}}{M_{P}^{2}}+\frac{\xi_{X}X^{2}}{M_{P}^{2}}\right)\partial^{\mu}h\partial_{\mu}h-\frac{1}{2}\left(\frac{\xi_{h}h^{2}}{M_{P}^{2}}+\frac{\xi_{X}X^{2}}{M_{P}^{2}}\right)\partial^{\mu}\phi\partial_{\mu}\phi-\frac{1}{2}\left(\frac{\xi_{\phi}\phi^{2}}{M_{P}^{2}}+\frac{\xi_{h}h^{2}}{M_{P}^{2}}\right)\partial^{\mu}X\partial_{\mu}X$ (73) $\displaystyle+$ $\displaystyle\frac{6\xi_{h}\xi_{X}hX}{M_{P}^{2}}\partial^{\mu}h\partial_{\mu}X+\frac{6\xi_{h}\xi_{\phi}h\phi}{M_{P}^{2}}\partial^{\mu}h\partial_{\mu}\phi+\frac{6\xi_{\phi}\xi_{X}\phi X}{M_{P}^{2}}\partial^{\mu}\phi\partial_{\mu}X+m_{X}^{2}X^{2}\left(\frac{\xi_{\phi}\phi^{2}}{M_{P}^{2}}+\frac{\xi_{h}h^{2}}{M_{P}^{2}}\right)$ $\displaystyle+$ $\displaystyle m_{\phi}^{2}\phi^{2}M_{P}^{2}\left(\frac{\xi_{X}X^{2}}{M_{P}^{2}}+\frac{\xi_{h}h^{2}}{M_{P}^{2}}\right)+m_{h}^{2}h^{2}\left(\frac{\xi_{\phi}\phi^{2}}{M_{P}^{2}}+\frac{\xi_{X}X^{2}}{M_{P}^{2}}\right)\,,$ and we can rewrite the above Lagrangian in terms of the effective couplings as Eq. (13), with $\displaystyle\sigma_{hX}^{\xi}\;$ $\displaystyle=$ $\displaystyle\;\frac{1}{4M_{P}^{2}}\left[\xi_{h}(2m_{X}^{2}+s)+\xi_{X}(2m_{h}^{2}+s)\right.$ (74) $\displaystyle+$ $\displaystyle\left.\left(12\xi_{X}\xi_{h}(m_{h}^{2}+m_{X}^{2}-t)\right)\right]\,,$ $\sigma_{\phi X}^{\xi}\;=\;\frac{1}{2M_{P}^{2}}\left[\xi_{\phi}m_{X}^{2}+12\xi_{\phi}\xi_{X}m_{\phi}^{2}+3\xi_{X}m_{\phi}^{2}+2\xi_{\phi}m_{\phi}^{2}\right]\,,$ (75) $\sigma_{\phi h}^{\xi}\;=\;\frac{1}{2M_{P}^{2}}\left[\xi_{\phi}m_{h}^{2}+12\xi_{\phi}\xi_{h}m_{\phi}^{2}+3\xi_{h}m_{\phi}^{2}+2\xi_{\phi}m_{\phi}^{2}\right]\,,$ (76) where $s,t$ are the Mandelstam variables. The latter couplings assume an inflaton condensate in the initial state rather than a thermal Higgs in the initial state accounting for the lack of symmetry in the three couplings. ### B THERMAL PRODUCTION In this appendix we calculate the thermal dark matter production rate $R_{X}^{T,\,\xi}(T)$ arising from the effective four-point interaction $\sigma_{hX}h^{2}X^{2}$, where $\sigma_{hX}$ is given by Eq. (74). We also calculate the production rate $R_{X}^{T}(T)$ for the thermal scattering processes mediated by gravity alone, ${\rm{SM}}~{}{\rm{SM}}\rightarrow X~{}X$, that are unavoidable in models with a minimal coupling to gravity ($\xi_{\phi,h,X}=0$) Bernal:2018qlk ; CMOV , and compare the two results. The production rate $R_{X}^{T,\,\xi}(T)$ can be computed from Eq. (21). The matrix element squared is given by $|\overline{{\cal M}}^{hX,\,\xi}|^{2}\;=\;4\sigma_{hX}^{\xi~{}2}\,,$ (77) where in the limit where the Higgs boson mass is neglected, the Mandelstam variables $s$ and $t$ are given by $\displaystyle t\,=$ $\displaystyle\,\dfrac{s}{2}\left(\sqrt{1-\frac{4m_{X}^{2}}{s}}\cos\theta_{13}-1\right)+m_{X}^{2}\,,$ (78) $\displaystyle s\,=$ $\displaystyle\,2E_{1}E_{2}\left(1-\cos\theta_{12}\right)\,.$ (79) We find the following coefficients for Eq. (23) $\displaystyle\beta_{1}^{\xi}$ $\displaystyle\;=\;\frac{\pi^{3}}{2700}\left[\xi_{h}^{2}+2\xi_{h}\xi_{X}+\xi_{X}^{2}+12\xi_{h}\xi_{X}\left(\xi_{h}+\xi_{X}+4\xi_{h}\xi_{X}\right)\right]\,,$ (80) $\displaystyle\beta_{2}^{\xi}$ $\displaystyle\;=\;\frac{\zeta(3)^{2}\xi_{h}}{2\pi^{5}}\left[\xi_{h}+\xi_{X}+6\xi_{h}\xi_{X}-12\xi_{h}\xi_{X}^{2}\right]\,,$ (81) $\displaystyle\beta_{3}^{\xi}$ $\displaystyle\;=\;\frac{\xi_{h}^{2}}{576\pi}\,.$ (82) Similarly, using Eqs. (18)-(20), we find the matrix element squared for minimally coupled gravity: $|\overline{{\cal M}}^{hX}|^{2}\;=\;\frac{1}{4M_{P}^{4}}\frac{\left(t(s+t)-2m_{X}^{2}t+m_{X}^{4}\right)^{2}}{s^{2}},$ (83) where we have neglected the Higgs field mass. We find the coefficients: $\displaystyle\beta_{1}$ $\displaystyle\;=\;\frac{\pi^{3}}{81000}\,,$ (84) $\displaystyle\beta_{2}$ $\displaystyle\;=\;-\frac{\zeta(3)^{2}}{30\pi^{5}}\,,$ (85) $\displaystyle\beta_{3}$ $\displaystyle\;=\;\frac{1}{4320\pi}\,.$ (86) Note that when both contributions are kept, and we neglect $m_{h}\ll m_{X}$, the full coefficients (including interference) are given by $\displaystyle\beta_{1}^{\xi}$ $\displaystyle=$ $\displaystyle\frac{\pi^{3}}{81000}\left[30\xi_{h}^{2}\left(12\xi_{X}(4\xi_{X}+1)+1\right)\right.$ (87) $\displaystyle\left.+10\xi_{h}(6\xi_{X}+1)^{2}+10\xi_{X}(3\xi_{X}+1)+1\right]\,,$ $\displaystyle\beta_{2}^{\xi}$ $\displaystyle=$ $\displaystyle-\frac{\zeta(3)^{2}}{60\pi^{5}}\left[2+10\xi_{X}\right.$ (88) $\displaystyle\left.+5\xi_{h}\left(1+6\xi_{X}+6\xi_{h}\left(6\xi_{X}(2\xi_{X}-1)-1\right)\right)\right]\,,$ $\displaystyle\beta_{3}^{\xi}$ $\displaystyle=$ $\displaystyle\frac{1}{8640\pi}\left[2+5\xi_{h}\left(32\xi_{h}-2\right)\right]\,.$ (89) which reduces to Eqs. (84-86) when all $\xi_{i}=0$. ## References * (1) F. L. Bezrukov and M. Shaposhnikov, Phys. Lett. B 659 (2008), 703-706 [arXiv:0710.3755 [hep-th]]. * (2) O. Lebedev, Prog. Part. Nucl. Phys. 120 (2021), 103881 doi:10.1016/j.ppnp.2021.103881 [arXiv:2104.03342 [hep-ph]]. * (3) Y. Ema, R. Jinno, K. Mukaida and K. Nakayama, JCAP 02 (2017), 045 [arXiv:1609.05209 [hep-ph]]. * (4) B. A. Bassett and S. Liberati, Phys. Rev. D 58, 021302 (1998) [erratum: Phys. Rev. D 60, 049902 (1999)] [arXiv:hep-ph/9709417 [hep-ph]]; S. Tsujikawa, K. i. Maeda and T. Torii, Phys. Rev. D 60, 063515 (1999) [arXiv:hep-ph/9901306 [hep-ph]]; T. Markkanen and S. Nurmi, JCAP 02, 008 (2017) [arXiv:1512.07288 [astro-ph.CO]]. * (5) Y. Mambrini and K. A. Olive, Phys. Rev. D 103 (2021) no.11, 115009 [arXiv:2102.06214 [hep-ph]]. * (6) S. Clery, Y. Mambrini, K. A. Olive and S. Verner, [arXiv:2112.15214 [hep-ph]]; * (7) B. Barman and N. Bernal, JCAP 06 (2021), 011 doi:10.1088/1475-7516/2021/06/011 [arXiv:2104.10699 [hep-ph]]; B. Barman, N. Bernal, Y. Xu and Ó. Zapata, [arXiv:2202.12906 [hep-ph]]. * (8) M. R. Haque and D. Maity, [arXiv:2112.14668 [hep-ph]]. * (9) M. R. Haque and D. Maity, [arXiv:2201.02348 [hep-ph]]. * (10) C. G. Callan, Jr., S. R. Coleman and R. Jackiw, Annals Phys. 59, 42-73 (1970) doi:10.1016/0003-4916(70)90394-5 * (11) L. Parker, Phys. Rev. Lett. 21, 562-564 (1968); L. Ford, Phys. Rev. D 35, 2955 (1987). * (12) D. J. H. Chung, E. W. Kolb and A. Riotto, Phys. Rev. Lett. 81, 4048-4051 (1998) [arXiv:hep-ph/9805473 [hep-ph]]; S. Hashiba and J. Yokoyama, JCAP 01, 028 (2019) [arXiv:1809.05410 [gr-qc]]; N. Bernal, A. Chatterjee and A. Paul, JCAP 12, 020 (2018) [arXiv:1809.02338 [hep-ph]]; L. Li, T. Nakama, C. M. Sou, Y. Wang and S. Zhou, JHEP 07, 067 (2019) [arXiv:1903.08842 [astro-ph.CO]]; J. A. R. Cembranos, L. J. Garay and J. M. Sánchez Velázquez, JHEP 06, 084 (2020) [arXiv:1910.13937 [hep-ph]]; N. Herring, D. Boyanovsky and A. R. Zentner, Phys. Rev. D 101, no.8, 083516 (2020) [arXiv:1912.10859 [gr-qc]]; E. Babichev, D. Gorbunov, S. Ramazanov and L. Reverberi, JCAP 09 (2020), 059 [arXiv:2006.02225 [hep-ph]]; S. Ling and A. J. Long, Phys. Rev. D 103, no.10, 103532 (2021) [arXiv:2101.11621 [astro-ph.CO]]. * (13) Y. Ema, R. Jinno, K. Mukaida and K. Nakayama, JCAP 05, 038 (2015) [arXiv:1502.02475 [hep-ph]]; Y. Ema, R. Jinno, K. Mukaida and K. Nakayama, Phys. Rev. D 94, no.6, 063517 (2016) [arXiv:1604.08898 [hep-ph]]; Y. Ema, K. Nakayama and Y. Tang, JHEP 09, 135 (2018) [arXiv:1804.07471 [hep-ph]]; Y. Ema, K. Mukaida and K. Nakayama, [arXiv:2112.12774 [hep-ph]]. * (14) Y. Ema, K. Nakayama and Y. Tang, JHEP 07, 060 (2019) [arXiv:1903.10973 [hep-ph]]. N. Herring and D. Boyanovsky, Phys. Rev. D 101, no.12, 123522 (2020) [arXiv:2005.00391 [astro-ph.CO]]; * (15) A. Ahmed, B. Grzadkowski and A. Socha, JHEP 08, 059 (2020) [arXiv:2005.01766 [hep-ph]]; C. Gross, S. Karamitsos, G. Landini and A. Strumia, JHEP 03, 174 (2021) [arXiv:2012.12087 [hep-ph]]; B. Barman, N. Bernal, A. Das and R. Roshan, JCAP 01, no.01, 047 (2022) [arXiv:2108.13447 [hep-ph]]. * (16) A. A. Starobinsky, Phys. Lett. B 91, 99 (1980). * (17) M. Atkins and X. Calmet, Phys. Rev. Lett. 110, no.5, 051301 (2013) [arXiv:1211.0281 [hep-ph]]; Z. Z. Xianyu, J. Ren and H. J. He, Phys. Rev. D 88, 096013 (2013) [arXiv:1305.0251 [hep-ph]]. * (18) R. A. Flores and M. Sher, Phys. Rev. D 27, 1679 (1983); M. Sher, Phys. Rept. 179, 273-418 (1989); G. Isidori, G. Ridolfi and A. Strumia, Nucl. Phys. B 609, 387-409 (2001) [arXiv:hep-ph/0104016 [hep-ph]]; F. Bezrukov, M. Y. Kalmykov, B. A. Kniehl and M. Shaposhnikov, JHEP 10, 140 (2012) [arXiv:1205.2893 [hep-ph]]; G. Degrassi, S. Di Vita, J. Elias-Miro, J. R. Espinosa, G. F. Giudice, G. Isidori and A. Strumia, JHEP 08, 098 (2012) [arXiv:1205.6497 [hep-ph]]; D. Buttazzo, G. Degrassi, P. P. Giardino, G. F. Giudice, F. Sala, A. Salvio and A. Strumia, JHEP 12, 089 (2013) [arXiv:1307.3536 [hep-ph]]; A. V. Bednyakov, B. A. Kniehl, A. F. Pikelner and O. L. Veretin, Phys. Rev. Lett. 115, no.20, 201802 (2015) [arXiv:1507.08833 [hep-ph]]. * (19) A. Vilenkin, Phys. Lett. 115B, 91 (1982); A. Vilenkin and L. H. Ford, Phys. Rev. D 26, 1231 (1982); A. D. Linde, Phys. Lett. 116B, 335 (1982); A. A. Starobinsky, Phys. Lett. 117B, 175 (1982); A. Vilenkin, Nucl. Phys. B 226, 527 (1983); A. D. Linde, Phys. Lett. 160B, 243 (1985); K. Enqvist, K. W. Ng and K. A. Olive, Nucl. Phys. B 303, 713 (1988). * (20) J. R. Espinosa, G. F. Giudice and A. Riotto, JCAP 05, 002 (2008) [arXiv:0710.2484 [hep-ph]]. * (21) M. Herranen, T. Markkanen, S. Nurmi and A. Rajantie, Phys. Rev. Lett. 113, no.21, 211102 (2014) [arXiv:1407.3141 [hep-ph]]. T. Markkanen, S. Nurmi, A. Rajantie and S. Stopyra, JHEP 06, 040 (2018) [arXiv:1804.02020 [hep-ph]]. * (22) T. Markkanen, A. Rajantie and S. Stopyra, Front. Astron. Space Sci. 5, 40 (2018) [arXiv:1809.06923 [astro-ph.CO]]. * (23) N. Aghanim et al. [Planck], Astron. Astrophys. 641, A6 (2020) [arXiv:1807.06209 [astro-ph.CO]]. * (24) See e.g., J. Ellis, M. A. G. García, N. Nagata, D. V. Nanopoulos, K. A. Olive and S. Verner, Int. J. Mod. Phys. D 29, no.16, 2030011 (2020) [arXiv:2009.01709 [hep-ph]]. * (25) F. Bezrukov, A. Magnin, M. Shaposhnikov and S. Sibiryakov, JHEP 01, 016 (2011) [arXiv:1008.5157 [hep-ph]]. * (26) M. Garny, M. Sandora and M. S. Sloth, Phys. Rev. Lett. 116, no.10, 101302 (2016) [arXiv:1511.03278 [hep-ph]]; M. Garny, A. Palessandro, M. Sandora and M. S. Sloth, JCAP 02, 027 (2018) [arXiv:1709.09688 [hep-ph]]. * (27) Y. Tang and Y. L. Wu, Phys. Lett. B 774, 676-681 (2017) [arXiv:1708.05138 [hep-ph]]. * (28) M. Chianese, B. Fu and S. F. King, JCAP 06, 019 (2020) [arXiv:2003.07366 [hep-ph]]; M. Chianese, B. Fu and S. F. King, JCAP 01, 034 (2021) [arXiv:2009.01847 [hep-ph]]. * (29) M. Redi, A. Tesi and H. Tillim, [arXiv:2011.10565 [hep-ph]]. * (30) K. Benakli, Y. Chen, E. Dudas and Y. Mambrini, Phys. Rev. D 95 (2017) no.9, 095002 [arXiv:1701.06574 [hep-ph]]; E. Dudas, Y. Mambrini and K. Olive, Phys. Rev. Lett. 119 (2017) no.5, 051801 [arXiv:1704.03008 [hep-ph]]. * (31) N. Bernal, M. Dutra, Y. Mambrini, K. Olive, M. Peloso and M. Pierre, Phys. Rev. D 97 (2018) no.11, 115020 [arXiv:1803.01866 [hep-ph]]. * (32) M. A. G. Garcia, K. Kaneta, Y. Mambrini and K. A. Olive, JCAP 04, 012 (2021) [arXiv:2012.10756 [hep-ph]]. * (33) K. Ichikawa, T. Suyama, T. Takahashi and M. Yamaguchi, Phys. Rev. D 78 (2008), 063545 [arXiv:0807.3988 [astro-ph]]. * (34) K. Kainulainen, S. Nurmi, T. Tenkanen, K. Tuominen and V. Vaskonen, JCAP 06 (2016), 022 [arXiv:1601.07733 [astro-ph.CO]]. * (35) N. Bernal, F. Elahi, C. Maldonado and J. Unwin, JCAP 11 (2019), 026 [arXiv:1909.07992 [hep-ph]]. * (36) M. A. G. Garcia, K. Kaneta, Y. Mambrini and K. A. Olive, Phys. Rev. D 101 (2020) no.12, 123507 [arXiv:2004.08404 [hep-ph]]. * (37) M. A. G. Garcia, K. Kaneta, Y. Mambrini, K. A. Olive and S. Verner, [arXiv:2109.13280 [hep-ph]]. * (38) A. Ahmed, B. Grządkowski and A. Socha, [arXiv:2111.06065 [hep-ph]]. * (39) J. Ellis, M. A. G. Garcia, D. V. Nanopoulos and K. A. Olive, JCAP 07, 050 (2015) [arXiv:1505.06986 [hep-ph]]. * (40) Y. Mambrini, “Particles in the dark Universe", Springer Ed., ISBN 978-3-030-78139-2 * (41) S. Aoki, H. M. Lee, A. G. Menkara and K. Yamashita, [arXiv:2202.13063 [hep-ph]]. * (42) Y. Fujii and K. Maeda, “The scalar-tensor theory of gravitation,” Cambridge University Press, 2003.
f_{\alpha}\cdot f_{\beta}\bigr{)}\cdot_{g}f_{I}\Bigr{)}$ $\displaystyle=\Bigl{(}\sum_{\alpha<\beta}g(\nabla_{e_{i}}^{L.C.,g}\tau_{g_{0},g}e_{\alpha},\tau_{g_{0},g}e_{\beta})\cdot e_{\alpha}\cdot e_{\beta}\Bigr{)}\cdot e_{I}.$ We already know that $g(\nabla_{e_{i}}^{L.C.,g}e_{\alpha},e_{\beta})={}^{g}\Gamma_{i\alpha}^{\beta}$ depends smoothly on $g$. Thus, the assignment $g\mapsto{p\mathcal{G}_{g,g_{0}}}_{\ast}\nabla^{\mathfrak{S}_{g}}-\nabla^{\mathfrak{S}_{g_{0}}}$ is a smooth map $\mathrm{Riem}(M)\rightarrow\Gamma\bigl{(}M;\mathrm{Hom}(\mathfrak{S}_{g_{0}},T^{\vee}M\otimes\mathfrak{S}_{g_{0}})\bigr{)}$ and hence it is also a smooth map $\mathrm{Riem}(M)\rightarrow\Psi\mathrm{DO}^{0}(\mathfrak{S}_{g},T^{\vee}M\otimes\mathfrak{S}_{g}).$ This implies that $g\mapsto{\not{\mathfrak{D}}}_{g}$ is smooth as it is a composition of smooth maps. ∎ The next result is classical and follows immediately from the theory developed in [LawsonMichelsonSpin]*Chapter III.5. ###### Proposition 4.1.33. Let $M$ be a closed manifold and let $P\in\Psi\mathrm{Dir}(M)$ be a pseudo Dirac operator with underlying Riemannian metric $g\in\mathrm{Riem}(M)$. Then $P$ is a self-adjoint, unbounded Fredholm operator on $L^{2}(\mathfrak{S}_{g})$ with domain $H^{1}(\mathfrak{S}_{g})$. ###### Proof. By definition, $P$ is symmetric, so $P-{\not{\mathfrak{D}}}_{g}$ is a symmetric pseudo differential operator of order zero on $\mathfrak{S}_{g}$. Thus the difference extends to a bounded, symmetric, and hence self-adjoint, operator on $L^{2}(\mathfrak{S}_{g})$. Since the sum of a self-adjoint operator with a self-adjoint bounded operator is again self-adjoint [higson2000analytic]*Exercise 1.9.20, $P={\not{\mathfrak{D}}}_{g}+(P-{\not{\mathfrak{D}}}_{g})$ is self-adjoint. Gårdings inequality, see [LawsonMichelsonSpin]*Thm III.5.2, implies that the minimal domain of $P$ agrees with $H^{1}(\mathfrak{S}_{g})$. It is classical that elliptic pseudo differential operators on vector bundles over closed manifolds are unbounded Fredholm operators, see [LawsonMichelsonSpin]*Thm III.5.2. ∎ ##### Relations to K-theory Our interests in pseudo Dirac operators arise from their connection to $KO$-theory, which we will outline in this section. All presented results have already appeared in [ebert2017indexdiff] in a different language and we do not claim originality for the results presented here; rather we translate his results into our language, fill some gaps, and give more details. Needless to say, the presentation here closely follows [ebert2017indexdiff]. Throughout this section, we assume that $M^{d}$ is a _closed_ spin manifold of positive dimension. We start with recalling the connection between $KO$-theory and Clifford-linear Fredholm operators. Let $H$ be a separable real Hilbert space with $\mathord{\mathbb{Z}}_{2}$-grading $\iota$ and graded $Cl_{d,0}$-right action $\mathbf{r}$, which we assume to be a $\ast$-homomorphism. For brevity, we simply call it a $Cl_{d,0}$-Hilbert space. Call $\omega_{d,0}\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=\iota\mathbf{r}(e_{1}\cdots e_{d})$ the _Chirality element_. Recall that $Cl_{d,0}$ has, up to isomorphism, exactly one irreducible representation if $d\not\equiv 0\mod 4$, and exactly two irreducible representations if $d\equiv-1\mod 4$. In the latter case, the irreducible representations are distinguished as different eigenspaces of the Chirality element. The triple $(H,\iota,\mathbf{r})$ is called _ample_ if $H$ contains all irreducible representations with infinite multiplicity. ###### Definition 4.1.34. Let $(H,\iota,\mathbf{r})$ be an ample $Cl_{d,0}$-Hilbert space. We call an operator $F$ on $H$ an $Cl_{d,0}$_-Fredholm operator_ if it is Clifford- linear, odd, and self-adjoint. If $d\equiv-1\mod 4$, we further require that $\omega_{p,0}F\iota$ is neither essentially positive nor essentially negative, that means, there exist infinite dimensional, closed subspaces that are orthogonal to each other and on those $F$ restricts to a positive or negative operator, respectively. Let $\mathrm{Fred}^{d,0}(H)$ be the space of all $Cl_{d,0}$-Fredholm operators on $H$ equipped with the norm topology. By a classical result of Atiyah and Singer, these spaces represent real $K$-theory. ###### Theorem 4.1.35 (Theorem [atiyah1969indexskew]). If $H$ is ample, then $\mathrm{Fred}^{d,0}(H)$ is a classifying space for $KO^{-d}$. That means there is a natural bijection $\mathrm{ind}\colon[X,\mathrm{Fred}^{d,0}(H)]\rightarrow KO^{-d}(X)$ for every compact $CW$-complex $X$. A detailed construction of this map was carried out in the master’s thesis of Jonathan Glöckle [gloeckle2019master]*Theorem 2.17. For more details on this space like Bott-periodicity or Morita equivalences, we refer the reader to [ebert2017indexdiff]. To identify the space of invertible pseudo Dirac operators as a classifying space for $KO$-theory, we first need to introduce auxiliary spaces. ###### Definition 4.1.36. Let $\Psi\mathrm{Dir}^{\times}(M)$ be the subspace of $\Psi\mathrm{Dir}(M)$ consisting of all invertible pseudo Dirac operators and denote $[-1,1]$ by $I$. Provided $\Psi\mathrm{Dir}^{\times}(M)_{g_{0}}$ is non-empty for a fixed metric $g_{0}$, we define, for a fixed choice of base point $B\in\Psi\mathrm{Dir}^{\times}(M)$, the spaces ${\mathcal{X}}\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=\\{P\colon I\rightarrow\Psi\mathrm{Dir}(M)\,:\,P\text{ smooth, }P(-1)=B,\ P(1)\in\Psi\mathrm{Dir}^{\times}(M)\\}$ and ${\mathcal{X}}_{g_{0}}\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=\\{P\colon I\rightarrow\Psi\mathrm{Dir}(M)_{g_{0}}\,:\,P\text{ smooth, }P(-1)=B,\ P(1)\in\Psi\mathrm{Dir}^{\times}(M)_{g_{0}}\\}.$ ###### Lemma 4.1.37. If $\Psi\mathrm{Dir}^{\times}(M)_{g_{0}}$ is non-empty, then the evaluation map $\mathrm{ev}_{1}\colon{\mathcal{X}}\rightarrow\Psi\mathrm{Dir}^{\times}(M)$ is a weak homotopy equivalence that also restricts to a weak homotopy equivalence between ${\mathcal{X}}_{g_{0}}$ and $\Psi\mathrm{Dir}^{\times}(M)_{g_{0}}$. ###### Proof. We only prove the first statement; for the second statement, one simply needs to add the subscript ${}_{g_{0}}$ to $\Psi\mathrm{Dir}^{\times}(M)$ and the space $\mathcal{Y}$ defined below. Let $\mathcal{Y}$ be the set of all _continuous_ paths $P\colon I\rightarrow\Psi\mathrm{Dir}(M)$ having the same boundary conditions as elements in ${\mathcal{X}}$. Smoothing theory yields that the inclusion of ${\mathcal{X}}$ into $\mathcal{Y}$ is a weak homotopy equivalence. The map $\mathrm{ev}_{1}\colon\mathcal{Y}\rightarrow\Psi\mathrm{Dir}^{\times}(M)$ is a fibration. Indeed, given a commutative square $\textstyle{A\times 0\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{F}$$\textstyle{\mathcal{Y}\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\mathrm{ev}_{1}}$$\textstyle{A\times I\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{h}$$\textstyle{\Psi\mathrm{Dir}^{\times}(M)}$ we define the lift $\mathcal{H}\colon A\times I\rightarrow\mathcal{Y}$ by $\mathcal{H}(a,s)(t)\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=\frac{(1-t)}{2}\cdot F(a)+\frac{(1+t)}{2}\cdot h(a,s).$ This is a continuous map because the affine structure on $\Psi\mathrm{Dir}(M)$ is (jointly) continuous with respect to the Fréchet structure. It is also a lift for $\mathrm{ev}_{1}(\mathcal{H})=h$, showing that $\mathrm{ev}_{1}$ is a fibration. The homotopy fibre of $\mathrm{ev}_{1}$ is therefore the preimage of $B$, which is the set of all closed curves in $\Psi\mathrm{Dir}(M)$ that start (and end) at $B$. This subspace is clearly convex and therefore contractible, implying that $\mathcal{Y}$ and $\Psi\mathrm{Dir}^{\times}(M)$ are weak homotopy equivalent. ∎ The auxiliary path spaces ${\mathcal{X}}$ and ${\mathcal{X}}_{g_{0}}$ help to identify $\Psi\mathrm{Dir}^{\times}(M)$ as classifying space for $KO$-theory. We remark that, for a closed manifold $M^{d}$ of positive dimension $d$, the $Cl_{d,0}$-Hilbert space $H=L^{2}(M,\mathfrak{S}_{g_{0}})$ is always ample. Indeed, the right action of the Chirality element decomposes $\mathfrak{S}_{g_{0}}$ into non-zero subbundles $\mathfrak{S}_{g_{0}}^{\pm}$ on which the Chirality element acts as $\pm\mathrm{id}$, hence $L^{2}(M,\mathfrak{S}_{g_{0}})\cong L^{2}(M,\mathfrak{S}_{g_{0}}^{+})\oplus L^{2}(M,\mathfrak{S}_{g_{0}}^{-})$ is a decomposition into infinite dimensional eigenspaces of the eigenvalues $\pm 1$ of the Chirality element. The following theorem is a collection of results proved by Johannes Ebert in [ebert2017indexdiff]*Section 4.1: ###### Theorem 4.1.38. Assume that $\Psi\mathrm{Dir}^{\times}(M)_{g_{0}}$ is non-empty. Then the diagram of continuous maps $\textstyle{{\mathcal{X}}\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{\Omega\mathrm{Fred}^{d,0}\bigl{(}L^{2}(M,\mathfrak{S}_{g_{0}})\bigr{)}\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{{\mathcal{X}}_{g_{0}}\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\simeq}$$\textstyle{\Omega\mathrm{Fred}^{d,0}\bigl{(}L^{2}(M,\mathfrak{S}_{g_{0}})\bigr{)}}$ commutes. The horizontal maps are given by $P_{t}\mapsto\frac{\mathcal{G}_{g_{t},g_{0}}P_{t}\mathcal{G}_{g_{t},g_{0}}^{-1}}{(1+(\mathcal{G}_{g_{t},g_{0}}P_{t}\mathcal{G}_{g_{t},g_{0}}^{-1})^{2})^{1/2}}$ and the lower map is a weak homotopy equivalence. Here $\Omega\mathrm{Fred}^{d,0}(L^{2}(\mathfrak{S}_{g_{0}}))$ denotes the space of paths into $\mathrm{Fred}^{d,0}(L^{2}(\mathfrak{S}_{g_{0}}))$ whose start and end-points are invertible operators. ###### Proof. For fixed $t$, these maps certainly take values in the space of self-adjoint, odd, Clifford-linear Fredholm operators. From the discussion in [ebert2017indexdiff]*Section 4.1 it is not clear why the horizontal map takes values in the subspace with the additional condition that $\omega F\iota$ is neither essentially positive nor essentially negative if $d\equiv-1\mod 4$. In order to show this, first recall from [LawsonMichelsonSpin]*Theorem III.5.8 that the spectrum of every self-adjoint elliptic pseudo differential operator $P$ of positive order on a bundle over a closed manifold consists of countably many eigenvalues only, that the eigenspaces are finite dimensional, and that the eigenvalues go to infinity. The operator $P/\sqrt{1+P^{2}}$ is the unique operator that acts as $\lambda/\sqrt{1+\lambda^{2}}\cdot\mathrm{id}$ on the $\lambda$-eigenspace of $P$. The Chirality element satisfies $\omega_{d,0}^{2}=(-1)^{\frac{d(d+1)}{2}}\quad\text{ and }\quad\omega_{d,0}^{\ast}=(-1)^{\frac{d(d+1)}{2}}\omega_{d,0}.$ Thus, it is a self-adjoint, odd involution if $d\equiv-1\mod 4$. Since $P$, and therefore $P/\sqrt{1+P^{2}}$ too, commute with $\omega_{d,0}$, we can simultaneously diagonalise the two operators. The grading $\mathrm{e/o}$ anti-commutes with $P$, so the grading sends the $\lambda$-eigenspace of $P$ to the $-\lambda$-eigenspaces of $P$. The corresponding statement holds true for the $\pm 1$-eigenspaces of $\omega_{d,0}$ for the same reason. Abbreviate $P/\sqrt{1+P^{2}}$ to $F$. Since the grading $\mathrm{e/o}$ commutes with $\omega_{d,0}F$, the operator $\omega_{d,0}F\mathrm{e/o}$ is positive definite on $H^{+}\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=\bigoplus_{\lambda>0}\mathrm{Eig}(P;\lambda)\cap\mathrm{Eig}(\omega_{d,0};1)\oplus\mathrm{Eig}(P;-\lambda)\cap\mathrm{Eig}(\omega_{d,0};-1)$ and negative definite on $H^{-}\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=\bigoplus_{\lambda>0}\mathrm{Eig}(P;\lambda)\cap\mathrm{Eig}(\omega_{d,0};-1)\oplus\mathrm{Eig}(P;-\lambda)\cap\mathrm{Eig}(\omega_{d,0};1).$ By the discussion above, the two subspaces are infinite dimensional. Thus, $\omega_{d,0}F\mathrm{e/o}$ cannot be essentially positive or essentially negative. The upper horizontal map is continuous because conjugation with invertible functions is a continuous operation as $\Psi\mathrm{DO}^{\bullet}(\mathfrak{S}_{g_{0}})$ is a Fréchet algebra, and bounded functional calculus is continuous by [ebert2017indexdiff]*Proposition 3.7. The weak equivalence statement is [ebert2017indexdiff]*Proposition 4.3. ∎ The space $\Psi\mathrm{Dir}^{\times}(M)_{g_{0}}$ is unsuitable for our purposes. Because we want to compare different metrics, we need the bigger space $\Psi\mathrm{Dir}^{\times}(M)$. Luckily, the canonical inclusion is a weak homotopy equivalence. This is the main result of [ebert2017indexdiff]*Section 4.2 formulated in our language. ###### Theorem 4.1.39. The inclusion $\Psi\mathrm{Dir}^{\times}(M)_{g_{0}}\hookrightarrow\Psi\mathrm{Dir}^{\times}(M)$ is a weak homotopy equivalence. The identification between $\Psi\mathrm{Dir}^{\times}(M)$ and $\Omega\mathrm{Fred}^{d,0}(L^{2}(\mathfrak{S}_{g_{0}}))$ is the space-level version of the index difference of Hitchin [hitchin1974harmonic]. ###### Definition 4.1.40. Fix base points $g_{0}\in\mathrm{Riem}(M)$ and $B\in\Psi\mathrm{Dir}^{\times}(M)$. For any choice of section $s\colon\Psi\mathrm{Dir}^{\times}(M)\rightarrow\mathcal{X}$ of $\mathrm{ev}_{1}$, we call the composition $\textstyle{\mathrm{inddif}(\text{--},B)\colon\Psi\mathrm{Dir}^{\times}(M)\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{s}$$\textstyle{\mathcal{X}\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\simeq}$$\scriptstyle{\mathrm{ev}_{1}}$$\textstyle{\Omega\mathrm{Fred}^{d,0}(L^{2}(\mathfrak{S}_{g_{0}}))}$ the _(operator) index difference_. If $g_{0}$ is a psc-metric then we call the pre-composition with the Dirac operator $\textstyle{\alpha(\text{--},g_{0})\colon\mathrm{Riem}^{+}(M)\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{{\not{\mathfrak{D}}}_{(\text{--})}}$$\textstyle{\Psi\mathrm{Dir}^{\times}(M)\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\mathrm{inddif}(\text{--},{\not{\mathfrak{D}}}_{g_{0}})}$$\textstyle{\Omega\mathrm{Fred}^{d,0}(L^{2}(\mathfrak{S}_{g_{0}}))}$ the _index difference_. We summarise our achievement of this subsection in the following theorem. ###### Theorem 4.1.41. Let $M^{d}$ be a closed spin manifold of positive dimension. If the space of all invertible pseudo Dirac operators $\Psi\mathrm{Dir}^{\times}(M)$ is non- empty, then it is a classifying space for $KO^{-(d+1)}(\mathrm{pt})$. The natural transformation is given by the _Hitchin index difference_ $\textstyle{[X;\Psi\mathrm{Dir}^{\times}(M)]\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\mathrm{inddif}(\text{--},B)_{\ast}}$$\textstyle{[X;\Omega\mathrm{Fred}^{d,0}(L^{2}(\mathfrak{S}_{g_{0}}))]\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\mathrm{adjunction}}$$\scriptstyle{\cong}$$\textstyle{KO^{-(d+1)}(X)}$$\textstyle{KO^{-d}(\Sigma X)\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\cong}$$\textstyle{[\Sigma X;\mathrm{Fred}^{d,0}(L^{2}(\mathfrak{S}_{g_{0}}))]\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\mathrm{ind}}$$\scriptstyle{\cong}$ Our primary goal is to show that the index difference factors through the concordance set. To this end, we will construct another model for real $K$-theory that is built out of $\Psi\mathrm{Dir}^{\times}(M)$ in the same way as $\widetilde{\mathcal{R}^{+}_{\bullet}}(M)$ is built out of $\mathrm{Riem}^{+}(M)$. We will pursue this construction in Section 4.4 by “blockifying” the objects of this section. In the next section, we lay the necessary analytical groundwork. ### 4.2 Block Operators In this section, we discuss the foundations of _block pseudo differential operators._ These are operators on $M\times\mathord{\mathbb{R}}^{n}$ that can be formally thought of as pseudo differentials operators with the property that if we think of $\mathord{\mathbb{R}}^{n}$ as a cube of infinite length, then they decompose into a product operator near each face. We will discuss the required combinatorial language to deal with those operators, present a construction of a partition of unity that respects the decomposition of a block pseudo differential operator near infinity, and discuss the application of these partitions to the support of these block operators. Finally, we define the operator suspension of a family of pseudo Dirac operator and show that it is a block operator. We assume that the reader is familiar with the concept of pseudo differential operators and their extension of Atiyah and Singer, otherwise he or she may consult Appendix B and the references therein. #### Block Bundles and Basics of Block Operators The seemingly harmless space $\mathord{\mathbb{R}}^{n}$ can be decomposed into products $\mathord{\mathbb{R}}^{p}\times\mathord{\mathbb{R}}^{n-p}$ in various ways. It will be important to keep track of these decompositions, which requires notation that seems quite heavy (and maybe unnecessary) at the beginning. For ${\mathrm{dom}\,}{\bm{\varepsilon}}\subseteq\mathbf{n}=\\{1,\dots,n\\}$ and ${\bm{\varepsilon}}\colon{\mathrm{dom}\,}{\bm{\varepsilon}}\rightarrow\mathord{\mathbb{Z}}_{2}$ we define, for a given $M$ the sets $\displaystyle U_{R}({\bm{\varepsilon}})\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=M\times\\{x\in\mathord{\mathbb{R}}^{n}\,:\,{\bm{\varepsilon}}(i)x_{i}>R\text{ for }i\in{\mathrm{dom}\,}{\bm{\varepsilon}}\\},$ $\displaystyle\mathord{\mathbb{R}}^{n}({\bm{\varepsilon}})\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=\mathrm{Map}(\mathbf{n}\setminus{\mathrm{dom}\,}{\bm{\varepsilon}},\mathord{\mathbb{R}})\cong\\{x\in\mathord{\mathbb{R}}^{n}\,:\,{\bm{\varepsilon}}(i)x_{i}=R\\}$ $\displaystyle\mathord{\mathbb{R}}^{{\bm{\varepsilon}}}_{R}\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=\\{x\in\mathord{\mathbb{R}}^{|{\mathrm{dom}\,}{\bm{\varepsilon}}|}\,:\,{\bm{\varepsilon}}(i_{j})x_{j}>R,\,i_{j}\in{\mathrm{dom}\,}{\bm{\varepsilon}},\,i_{j}<i_{j+1}\\}.$ We would like to emphasise that the identification between $\mathrm{Map}(\mathbf{n}\setminus{\mathrm{dom}\,}{\bm{\varepsilon}},\mathord{\mathbb{R}})$ and $\\{x\in\mathord{\mathbb{R}}^{n}\,:\,{\bm{\varepsilon}}(i)x_{i}=R\\}$ is only canonical once $R$ is fixed. For such an ${\bm{\varepsilon}}$ let $c({\bm{\varepsilon}})^{-1}\colon\mathbf{n}\rightarrow\mathbf{n}\setminus{\mathrm{dom}\,}{\bm{\varepsilon}}\sqcup{\mathrm{dom}\,}{\bm{\varepsilon}}$ be the unique map that partitions $\mathbf{n}$ into $\mathbf{n}\setminus{\mathrm{dom}\,}{\bm{\varepsilon}}$ and ${\mathrm{dom}\,}{\bm{\varepsilon}}$ in an order preserving manner. We write ${\bm{\eta}}\geq{\bm{\varepsilon}}$ if ${\mathrm{dom}\,}{\bm{\eta}}\supseteq{\mathrm{dom}\,}{\bm{\varepsilon}}$ and ${\bm{\eta}}|_{{\mathrm{dom}\,}{\bm{\varepsilon}}}={\bm{\varepsilon}}$. For all ${\bm{\eta}}\geq{\bm{\varepsilon}}$ let $c({\bm{\eta}},{\bm{\varepsilon}})^{-1}$ be the unique map that partitions ${\mathrm{dom}\,}{\bm{\eta}}^{c}$ into ${\mathrm{dom}\,}{\bm{\eta}}\setminus{\mathrm{dom}\,}{\bm{\varepsilon}}$ and ${\mathrm{dom}\,}{\bm{\varepsilon}}^{c}$ in an order preserving manner and let $\mathrm{per}({\bm{\eta}},{\bm{\varepsilon}})^{-1}$ be the unique map that partitions ${\mathrm{dom}\,}{\bm{\eta}}$ into ${\mathrm{dom}\,}{\bm{\eta}}-{\bm{\varepsilon}}$ and ${\mathrm{dom}\,}{\bm{\varepsilon}}$ in an order preserving manner. Here, ${\bm{\eta}}-{\bm{\varepsilon}}$ denotes ${\bm{\eta}}$ restricted to ${\mathrm{dom}\,}{\bm{\eta}}\setminus{\mathrm{dom}\,}{\bm{\varepsilon}}$. For each ${\bm{\eta}}\geq{\bm{\varepsilon}}$ we have a commutative diagram $\textstyle{\mathbf{n}\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{c({\bm{\eta}})^{-1}}$$\textstyle{{\mathrm{dom}\,}{\bm{\eta}}^{c}\sqcup{\mathrm{dom}\,}{\bm{\eta}}\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\mathrm{id}\sqcup\mathrm{per}({\bm{\eta}},{\bm{\varepsilon}})^{-1}}$$\textstyle{{\mathrm{dom}\,}{\bm{\eta}}^{c}\sqcup{\mathrm{dom}\,}{\bm{\eta}}-{\bm{\varepsilon}}\sqcup{\mathrm{dom}\,}{\bm{\varepsilon}}}$$\textstyle{\mathbf{n}\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{c({\bm{\varepsilon}})^{-1}}$$\textstyle{{\mathrm{dom}\,}{\bm{\varepsilon}}^{c}\sqcup{\mathrm{dom}\,}{\bm{\varepsilon}}\ignorespaces\ignorespaces\ignorespaces\ignorespaces.}$$\scriptstyle{c({\bm{\eta}},{\bm{\varepsilon}})^{-1}\sqcup\mathrm{id}}$ These maps induce linear maps on Euclidean spaces by pulling back their inverse. For example, $\displaystyle c({\bm{\varepsilon}})\colon\mathord{\mathbb{R}}^{n}$ $\displaystyle\rightarrow\mathord{\mathbb{R}}^{n}({\bm{\varepsilon}})\times\mathord{\mathbb{R}}^{{\bm{\varepsilon}}}$ $\displaystyle(x_{1},\dots,x_{n})$ $\displaystyle\mapsto(x_{c({\bm{\varepsilon}})^{-1}(1)},\dots,x_{c({\bm{\varepsilon}})^{-1}(|{\mathrm{dom}\,}{\bm{\varepsilon}}^{c}|)};x_{c({\bm{\varepsilon}})^{-1}(|{\mathrm{dom}\,}{\bm{\varepsilon}}^{c}|+1)},\dots,x_{c({\bm{\varepsilon}})^{-1}(n)}).$ We make the following notational convention: For a block metric $g$ on $M\times\mathord{\mathbb{R}}^{n}$, we denote by $g({\bm{\varepsilon}})$ the restriction to $M\times\mathord{\mathbb{R}}^{n}({\bm{\varepsilon}})$ if we interpret $\mathord{\mathbb{R}}^{n}({\bm{\varepsilon}})$ as $\\{x\in\mathord{\mathbb{R}}^{n}\,:\,{\bm{\varepsilon}}(i)x_{i}=R\\}$ and by $\partial^{{\bm{\varepsilon}}}g$ the restriction to $M\times\mathord{\mathbb{R}}^{n}({\bm{\varepsilon}})$ if we interpret $\mathord{\mathbb{R}}^{n}({\bm{\varepsilon}})$ as $\mathord{\mathbb{R}}^{|{\mathrm{dom}\,}({\bm{\varepsilon}})^{c}|}$. The identification between these two is the pull-back with ${\delta^{R{\bm{\varepsilon}}}}$. The pull-back of $c({\bm{\varepsilon}})$ coordinates yields an isometry $c({\bm{\varepsilon}})\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=\mathrm{id}_{M}\times c({\bm{\varepsilon}})\colon\left(U_{R}({\bm{\varepsilon}}),g|_{U({\bm{\varepsilon}})}\right)\rightarrow\bigl{(}M\times\mathord{\mathbb{R}}^{n}({\bm{\varepsilon}})\times\mathord{\mathbb{R}}_{R}^{\bm{\varepsilon}},\underbrace{g({\bm{\varepsilon}})\oplus\langle\cdot,\cdot\rangle}_{=c({\bm{\varepsilon}})_{\ast}g|_{U_{R}({\bm{\varepsilon}})}}\bigr{)}.$ This fits into the following diagram of isometries: $\textstyle{U_{R}({\bm{\eta}})\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{c({\bm{\eta}})}$$\textstyle{(M\times\mathord{\mathbb{R}}^{n})({\bm{\eta}})\times\mathord{\mathbb{R}}_{R}^{{\bm{\eta}}}\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\mathrm{id}\times\mathrm{per}({\bm{\eta}},{\bm{\varepsilon}})}$$\textstyle{(M\times\mathord{\mathbb{R}}^{n})({\bm{\eta}})\times\mathord{\mathbb{R}}_{R}^{{\bm{\eta}}-{\bm{\varepsilon}}}\times\mathord{\mathbb{R}}_{R}^{\bm{\varepsilon}}\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{c({\bm{\eta}},{\bm{\varepsilon}})}$$\textstyle{U_{R}({\bm{\varepsilon}})\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{c({\bm{\varepsilon}})}$$\textstyle{(M\times\mathord{\mathbb{R}}^{n})({\bm{\varepsilon}})\times\mathord{\mathbb{R}}_{R}^{\bm{\varepsilon}}.}$ Since all of these maps are induced by permutations on Euclidean spaces, they are Pin-structure preserving isometries because each Euclidean space has, up to equivalence, only one $\mathrm{Pin}^{-}$-structure. Functoriality of the spinor bundle construction, see Lemma 4.1.17, yields a splitting on the associated spinor bundles: $\Phi_{\bm{\varepsilon}}=\mathfrak{S}(c({\bm{\varepsilon}}))\colon\mathfrak{S}_{U_{R}({\bm{\varepsilon}})}\rightarrow\mathfrak{S}_{g({\bm{\varepsilon}})}\boxtimes Cl_{{\bm{\varepsilon}},0}=\mathfrak{S}_{g({\bm{\varepsilon}})\oplus\langle\cdot,\cdot\rangle},$ where $Cl_{{\bm{\varepsilon}},0}$ denotes $Cl_{|{\bm{\varepsilon}}|,0}$. This gives rise to morphisms between spinor bundles: $\textstyle{\mathfrak{S}|_{M\times U_{R}({\bm{\eta}})}\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\Phi_{{\bm{\eta}}}}$$\textstyle{\mathfrak{S}_{g({\bm{\eta}})}\boxtimes Cl_{{\bm{\eta}},0}\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\mathrm{id}\times\mathfrak{S}(\mathrm{per}({\bm{\varepsilon}},{\bm{\eta}}))}$$\textstyle{\mathfrak{S}_{g({\bm{\eta}})}\boxtimes Cl_{{\bm{\eta}}-{\bm{\varepsilon}},0}\boxtimes Cl_{{\bm{\varepsilon}},0}\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\mathcal{C}_{{\bm{\eta}},{\bm{\varepsilon}}}}$$\textstyle{\mathfrak{S}|_{M\times U_{R}({\bm{\varepsilon}})}\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\Phi_{{\bm{\varepsilon}}}}$$\textstyle{\mathfrak{S}_{g({\bm{\varepsilon}})}\boxtimes Cl_{{\bm{\varepsilon}},0}.}$ ###### Definition 4.2.1. Let $g$ be a block metric on $M\times\mathord{\mathbb{R}}^{n}$ and let $A\in\\{\mathrm{id},\mathrm{e/o}\\}$. A pseudo differential operator $P\in\overline{\Psi\mathrm{DO}}^{m}(\mathfrak{S}_{g})$ is of _block form_ (_of type_ $A$) if there are $R>r>0$ such that: * (1) The block metric $g$ decomposes outside of $M\times rI^{n}$. * (2) If $\sigma\in\Gamma_{c}(\mathfrak{S}_{g})$ is supported within $M\times(RI^{n})^{\circ}$, then $P\sigma$ is supported within $M\times(RI^{n})^{\circ}$. * (3) The operator $P$ restricts555This means that if a section is supported within $U_{r}({\bm{\varepsilon}})$, then $P\sigma$ is also supported within this set. to $M\times U_{r}({\bm{\varepsilon}})$ and there is a $P({\bm{\varepsilon}})\in\overline{\Psi\mathrm{DO}}^{m}(\mathfrak{S}_{g({\bm{\varepsilon}})})$ such that $(\Phi_{\bm{\varepsilon}})_{\ast}P|_{M\times U_{r}({\bm{\varepsilon}})}=P({\bm{\varepsilon}})\boxtimes\mathrm{id}+A\boxtimes D({\bm{\varepsilon}}),$ where $P({\bm{\varepsilon}})\in\overline{\Psi\mathrm{DO}}^{m}(\mathfrak{S}_{g({\bm{\varepsilon}})})$ and $D({\bm{\varepsilon}})$ is a differential operator of order $m$ on the trivial bundle $Cl_{{\bm{\varepsilon}},0}\rightarrow\mathord{\mathbb{R}}^{\bm{\varepsilon}}_{r}$ with constant coefficients. We call $M\times RI^{n}$ _the core_ of $P$. The constants $R$ and $r$ are not unique. We will see in Lemma 4.2.6 below that property (2) and (3) hold for all larger constants $R^{\prime}>R$ and $r^{\prime}>r$ as well. ###### Lemma 4.2.2. Let $P\in\overline{\Psi\mathrm{DO}}^{m}(\mathfrak{S}_{g})$ be a block operator of type $A$. Then the operators $P({\bm{\varepsilon}})$ and $D({\bm{\varepsilon}})$ are uniquely determined by $P$. ###### Proof. We prove the statement by induction over $|{\mathrm{dom}\,}{\bm{\varepsilon}}|$. On ${\mathrm{dom}\,}{\bm{\varepsilon}}=\emptyset$, there is nothing to prove. For the induction step, it suffices to prove the following statement: Every operator of the form $P=Q\boxtimes\mathrm{id}+A\boxtimes D\in\overline{\Psi\mathrm{DO}}^{m}(\mathfrak{S}_{g_{M}}\boxtimes Cl_{n,0})$ on a _not necessarily closed_ spin manifolds $M$ with product metrics $g=g_{M}\oplus\langle\cdot,\cdot\rangle$ uniquely deterimes $Q$ and $D$. We will prove this statement by expressing $Q$ and $D$ through $P$. Let $x^{1},\dots,x^{n}$ be the standard coordinates of $\mathord{\mathbb{R}}^{n}$ and $\partial_{j}$ the derivation into the $j$-th direction. The operator $D$ can be written as $D=\sum_{|{\bm{\alpha}}|\leq m}a_{\bm{\alpha}}\partial^{\bm{\alpha}},$ for some constants $a_{\bm{\alpha}}\in\mathrm{End}(Cl_{n,0})$. For a point $p\in\mathord{\mathbb{R}}^{n}$, choose smooth functions $f_{1},\dots,f_{n}$ with compact support satisfying $f_{j}(x)=(x_{j}-p_{j})$ in some neighbourhood $V$ of $p$. Choose further $u\in\Gamma_{c}(M,\mathfrak{S}_{g_{M}})$ and $v\in\Gamma_{c}(V,Cl_{n,0})$. Finally, for each multiindex ${\bm{\alpha}}$ with $|{\bm{\alpha}}|\leq m$, define the multiplication operator $\mu_{\bm{\alpha}}$ by $\displaystyle\mu_{\bm{\alpha}}\colon\Gamma_{c}(M\times\mathord{\mathbb{R}}^{n},\mathfrak{S}_{g_{M}}\boxtimes Cl_{n,0})$ $\displaystyle\rightarrow\Gamma_{c}(M\times\mathord{\mathbb{R}}^{n},\mathfrak{S}_{g_{M}}\boxtimes Cl_{n,0})$ $\displaystyle\sigma$ $\displaystyle\mapsto\prod_{j=1}^{n}f_{j}^{\alpha_{j}}\cdot\sigma.$ The Leibniz rule implies $[P,\mu_{\bm{\alpha}}](u\otimes v)(m,p)=A(u)(m)\otimes a_{\bm{\alpha}}v(p),$ so $D$ is uniquely determined by $P$. It remains to determine $Q$ from $P$ and $D$. Since $P-A\boxtimes D$ is $\mathcal{C}^{\infty}(\mathord{\mathbb{R}}^{n})$-linear on $\Gamma_{c}(M\times\mathord{\mathbb{R}}^{n},\mathfrak{S}_{g_{M}}\boxtimes Cl_{n,0})$ we can define a linear map $\tilde{Q}\colon\Gamma_{c}(M,\mathfrak{S}_{g_{M}})\rightarrow\Gamma(M,\mathfrak{S}_{g_{M}})$ as follows: Fix a point $p\in\mathord{\mathbb{R}}^{n}$ and a function $f\in\mathcal{C}^{\infty}(\mathord{\mathbb{R}}^{n},\mathord{\mathbb{R}})$ that is identically $1$ near $x$ and compactly supported. Set then $\tilde{Q}(\sigma)(m)\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=(P-A\boxtimes D)(\sigma\otimes f\cdot 1)(m,p).$ Plugging in the decomposition gives $\displaystyle\tilde{Q}(\sigma)(m)$ $\displaystyle=(P-A\boxtimes D)(\sigma)(m,p)$ $\displaystyle=Q\boxtimes\mathrm{id}(\sigma\otimes f\cdot 1)(m,p)$ $\displaystyle=Q(\sigma)(m)\otimes f(p)=Q(\sigma)(m)\otimes 1,$ which can be identified with $Q(\sigma)$ under the inclusion $\mathfrak{S}_{g_{M}}\subseteq\mathfrak{S}_{g_{M}}\boxtimes Cl_{n,0}$. ∎ ###### Corollary 4.2.3. If $P$ is a block operator on $\mathfrak{S}_{g}$, then $P({\bm{\varepsilon}})$ is a block operator on $\mathfrak{S}_{g({\bm{\varepsilon}})}$. Equivalently, if ${\delta^{2\rho{\bm{\varepsilon}}}}\colon\mathord{\mathbb{R}}^{n}({\bm{\varepsilon}})\xrightarrow{\cong}\\{x\in\mathord{\mathbb{R}}^{n}\,:\,{\bm{\varepsilon}}(i)x_{i}=2\rho\\}$ denotes the isometric inclusion that plugs $2{\bm{\varepsilon}}(j)\rho$ into the $j$-th coordinate for all $j\in{\mathrm{dom}\,}{\bm{\varepsilon}}$, then ${\delta^{2\rho{\bm{\varepsilon}}}}^{\ast}P({\bm{\varepsilon}})=({\delta^{2\rho{\bm{\varepsilon}}}})^{-1}_{\ast}P({\bm{\varepsilon}})$ is a block operator on $\mathfrak{S}_{\partial^{{\bm{\varepsilon}}}g}$ for all $\rho>R$, where $R$ denotes the size of the core of $P$. These block operators satisfy conditions (1)-(3) for the same constants as $P$. ###### Proof. Since $g$ decomposes outside of $rI^{n}$, it also decomposes outside of $\rho I^{n}$. If we identify $\mathord{\mathbb{R}}^{n}({\bm{\varepsilon}})$ with the affine subspace $\\{x\in\mathord{\mathbb{R}}^{n}\,:\,{\bm{\varepsilon}}(j)x_{j}=\rho\\}$, then the restricted metric $g({\bm{\varepsilon}})=g\mathord{\upharpoonleft}_{\mathord{\mathbb{R}}^{n}({\bm{\varepsilon}})}$ is again a block metric that decomposes outside of $rI^{n}\cap\mathord{\mathbb{R}}^{n}{\bm{\varepsilon}})\cong rI^{n-|{\mathrm{dom}\,}{\bm{\varepsilon}}|}$. Set $m\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=n-|{\mathrm{dom}\,}{\bm{\varepsilon}}|$. We verify axiom $(2)$ for ${{\delta^{2r{\bm{\varepsilon}}}}}^{\ast}P({\bm{\varepsilon}})$ by contra- position. Assume there is a section $\sigma\in\Gamma_{c}(\mathfrak{S}_{\partial^{\varepsilon}_{{\bm{\varepsilon}}}g({\bm{\varepsilon}})})$ with support in $M\times(RI^{m})^{\circ}$ such that ${{\delta^{2r{\bm{\varepsilon}}}}}^{\ast}P({\bm{\varepsilon}})(\sigma)$ is not supported within $M\times(RI^{m})^{\circ}$. Pick a smooth function $f\colon\mathord{\mathbb{R}}\rightarrow[0,1]$ that is supported within $(r,R)$ and define on $R_{r}^{\bm{\varepsilon}}$ the real valued function $f^{\bm{\varepsilon}}(x)=\prod_{j=1}^{|{\mathrm{dom}\,}{\bm{\varepsilon}}|}f(x_{j}).$ Then the section $\sigma\otimes f^{\bm{\varepsilon}}$ is supported within the interior of $M\times RI^{m}\times RI^{|{\mathrm{dom}\,}{\bm{\varepsilon}}|}$ but, by assumption, the support of $\bigl{(}P({\bm{\varepsilon}})\boxtimes\mathrm{id}+A\boxtimes D\bigr{)}(\sigma\otimes f^{\bm{\varepsilon}})$ is not supported within $M\times RI^{m}\times RI^{|{\mathrm{dom}\,}{\bm{\varepsilon}}|}$. Thus $\tau\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt={\Phi_{\bm{\varepsilon}}}^{-1}_{\ast}(\sigma\otimes f^{\bm{\varepsilon}})$ is a section of $\mathfrak{S}_{g}$ that is supported within the interior of $M\times RI^{n}$. But the support of $P(\tau)$ does not lie in $M\times RI^{n}$, contradicting axiom $(2)$ of $P$. We first give an informal proof for (3). We know that $\Phi_{{\delta^{{\bm{\varepsilon}}}}({\bm{\eta}})\,\ast}P|_{U_{r}({\delta^{{\bm{\varepsilon}}}}({\bm{\eta}}))}=P\bigl{(}{\delta^{{\bm{\varepsilon}}}}({\bm{\eta}})))\boxtimes\mathrm{id}+\mathrm{e/o}\boxtimes D({\delta^{{\bm{\varepsilon}}}}({\bm{\eta}}))\bigr{)}.$ If we permute the (fixed) coordinates parametrised by ${\mathrm{dom}\,}\,{\bm{\varepsilon}}$ “to the right”, using $\mathrm{per}({\delta^{{\bm{\varepsilon}}}}({\bm{\eta}})),{\bm{\varepsilon}})$ we can identify $\mathord{\mathbb{R}}^{{\delta^{{\bm{\varepsilon}}}}({\bm{\eta}}))}_{r}$ with $\mathord{\mathbb{R}}^{{\delta^{{\bm{\varepsilon}}}}({\bm{\eta}}))-{\bm{\varepsilon}}}_{r}\times\mathord{\mathbb{R}}^{{\bm{\varepsilon}}}_{r}$. The uniqueness statement of Lemma 4.2.2 implies (with abuse of notation) $P({\bm{\varepsilon}})=P\bigl{(}{\delta^{{\bm{\varepsilon}}}}({\bm{\eta}})\bigr{)}\otimes\mathrm{id}+\mathrm{per}({\delta^{{\bm{\varepsilon}}}}({\bm{\eta}}),{\bm{\varepsilon}})_{\ast}D\bigl{(}{\delta^{{\bm{\varepsilon}}}}({\bm{\eta}})\bigr{)}.$ Now the formal proof of (3): The inclusion ${\delta^{2r{\bm{\varepsilon}}}}$ maps $U^{m}_{r}({\bm{\eta}})$ into $U^{n}_{r}({\delta^{{\bm{\varepsilon}}}}({\bm{\eta}}))$. Furthermore, the following diagram commutes: $\textstyle{M\times U^{m}_{r}({\bm{\eta}})\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{c^{m}({\bm{\eta}})}$$\scriptstyle{{\delta^{2r{\bm{\varepsilon}}}}}$$\textstyle{M\times\mathord{\mathbb{R}}^{m}({\bm{\eta}})\times\mathord{\mathbb{R}}^{{\bm{\eta}}}_{r}\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\mathrm{id}\times{\delta^{{\bm{\varepsilon}}}}}$$\textstyle{M\times U^{n}_{r}({\delta^{{\bm{\varepsilon}}}}({\bm{\eta}}))\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{c({\delta^{{\bm{\varepsilon}}}}({\bm{\eta}}))}$$\textstyle{M\times\mathord{\mathbb{R}}^{n}({\delta^{{\bm{\varepsilon}}}}({\bm{\eta}}))\times\mathord{\mathbb{R}}_{r}^{{\delta^{{\bm{\varepsilon}}}}({\bm{\eta}})}\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\mathrm{per}({\delta^{{\bm{\varepsilon}}}}({\bm{\eta}}),{\bm{\varepsilon}})}$$\textstyle{M\times\mathord{\mathbb{R}}^{n}({\delta^{{\bm{\varepsilon}}}}({\bm{\eta}}))\times\mathord{\mathbb{R}}^{{\delta^{{\bm{\varepsilon}}}}({\bm{\eta}})-{\bm{\varepsilon}}}_{r}\times\mathord{\mathbb{R}}_{r}^{{\bm{\varepsilon}}}}$ This implies $(\Phi_{\bm{\eta}})_{\ast}({{\delta^{2r{\bm{\varepsilon}}}}}^{\ast}P({\bm{\varepsilon}})))=(\mathrm{id}\times{\delta^{{\bm{\varepsilon}}}})^{\ast}\bigl{(}P({\delta^{{\bm{\varepsilon}}}}({\bm{\eta}}))\otimes\mathrm{id}+A\otimes\mathrm{per}({\delta^{{\bm{\varepsilon}}}}({\bm{\eta}}),{\bm{\varepsilon}})_{\ast}D({\delta^{{\bm{\varepsilon}}}}({\bm{\eta}}))\bigr{)},$ so ${{\delta^{2r{\bm{\varepsilon}}}}}^{\ast}P({\bm{\varepsilon}})$ is indeed of block form. ∎ The converse of this lemma is also true. ###### Lemma 4.2.4. Let $m>0$ and $P\in\overline{\Psi\mathrm{DO}}^{m}(\mathfrak{S}_{g})$ be of block form of type $A$. For each differential operator $D$ of order $m$ with constant coefficients on $\mathcal{C}_{c}^{\infty}(\mathord{\mathbb{R}},Cl_{1,0})$, the operator $Q\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=P\boxtimes\mathrm{id}+A\boxtimes D\in\overline{\Psi\mathrm{DO}}^{m}(\mathfrak{S}_{g}\boxtimes Cl_{1,0})=\overline{\Psi\mathrm{DO}}^{m}(\mathfrak{S}_{g\oplus\mathrm{d}x_{n+1}^{2}})$ is also block operator of type $A$. ###### Proof. If $m>0$, then $Q$ is an element of $\overline{\Psi\mathrm{DO}}^{m}(\mathfrak{S}_{g\oplus\mathrm{d}x_{n+1}^{2}})$ by Theorem B.0.28. Axiom (1) is trivially satisfied. Axiom (2) also holds because $A\boxtimes D$ acts as a differential operator in the additional dimension. Axiom (3) requires some definitions. Let ${\bm{\varepsilon}}\colon\mathbf{n+1}\rightarrow\mathord{\mathbb{Z}}_{2}$ be given. If $n+1\in{\mathrm{dom}\,}{\bm{\varepsilon}}$, then $U_{r}({\bm{\varepsilon}})=U_{r}(p_{n+1}({\bm{\varepsilon}}))\times\\{{\bm{\varepsilon}}(n+1)x_{n+1}>r\\}$ and $\Phi_{{\bm{\varepsilon}}}=\Phi_{p_{n+1}({\bm{\varepsilon}})}\times\mathrm{id}$. In this case, we have $P({\bm{\varepsilon}})=Q(p_{n+1}({\bm{\varepsilon}}))\boxtimes\mathrm{id}_{\mathord{\mathbb{R}}}\text{ and }D^{Q}({\bm{\varepsilon}})=D^{P}(p_{n+1}({\bm{\varepsilon}}))\boxtimes\mathrm{id}.$ If $n+1\notin{\mathrm{dom}\,}{\bm{\varepsilon}}$, then $U_{r}({\bm{\varepsilon}})=U_{r}^{n}({\bm{\varepsilon}})\times\mathord{\mathbb{R}}$, where $U_{r}^{n}({\bm{\varepsilon}})\subseteq\mathord{\mathbb{R}}^{n}$, and $\Phi_{\bm{\varepsilon}}$ is given by the decomposition $\textstyle{\mathfrak{S}|_{M\times U_{r}({\bm{\varepsilon}})}=\mathfrak{S}\mathord{\upharpoonleft}_{M\times U_{r}^{n}({\bm{\varepsilon}})}\boxtimes Cl_{1,0}\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\Phi_{\bm{\varepsilon}}\boxtimes\mathrm{id}}$$\textstyle{\mathfrak{S}_{g({\bm{\varepsilon}})}\boxtimes Cl_{{\bm{\varepsilon}},0}\boxtimes Cl_{1,0}\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\mathrm{swap}}$$\scriptstyle{\mathrm{id}\boxtimes}$$\textstyle{\mathfrak{S}_{g({\bm{\varepsilon}})}\boxtimes Cl_{1,0}\boxtimes Cl_{{\bm{\varepsilon}},0}.}$ In this case, we have $Q({\bm{\varepsilon}})=P({\bm{\varepsilon}})\boxtimes\mathrm{id}+A\boxtimes D\quad\text{ and }\quad D^{Q}({\bm{\varepsilon}})=D^{P}({\bm{\varepsilon}}).$ inline]Figure out if $\mathrm{Cycl}(1:1+|{\mathrm{dom}\,}{\bm{\varepsilon}}|$ or $\mathrm{Cycl}(1:1+|{\mathrm{dom}\,}{\bm{\varepsilon}}|^{-1}$. Define the correct map in an ad-hoc manner as the conceptual notation will be used later. With these definitions the equation ${\Phi_{\bm{\varepsilon}}}_{\ast}Q|_{M\times U_{r}({\bm{\varepsilon}})}=Q({\bm{\varepsilon}})\boxtimes\mathrm{id}+A\boxtimes D^{Q}({\bm{\varepsilon}})$ is easily verified so that $Q$ is a block operator. ∎ Pseudo differential operators are not local by nature. Block operators on the other hand are local in certain directions on special classes of open sets. In order to use local-to-global principles for block operators, we need partition of unities that are adapted to the specific block form of a block operator. ###### Definition 4.2.5. Let $\chi\colon\mathord{\mathbb{R}}\rightarrow[0,1]$ be a compactly supported, smooth, and even function that is identically $1$ near the origin. We denote the extensions by zero of $\chi^{-1}\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=(1-\chi)|_{\mathord{\mathbb{R}}_{\leq 0}},\quad\chi^{0}\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=\chi,\quad\text{ and }\quad\chi^{1}\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=(1-\chi)|_{\mathord{\mathbb{R}}_{\geq 0}}$ with the same symbols. More generally, we define on $\mathord{\mathbb{R}}^{n}$ or $M\times\mathord{\mathbb{R}}^{n}$ the partition of unity $\biggl{\\{}\chi^{\bm{\alpha}}(m,x)\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=\prod_{j=1}^{n}\chi^{{\bm{\alpha}}(j)}(x_{j})\biggr{\\}}.$ parameterised ${\alpha\colon\mathbf{n}\rightarrow\mathord{\mathbb{Z}}_{3}}$. As an application of this partition, we prove that the constants in the definition of a block operator are actually lower bounds. ###### Lemma 4.2.6. Let $P\in\overline{\Psi\mathrm{DO}}^{m}(\mathfrak{S}_{g})$ be a block operator and let $R>r$ be constants such that the conditions (1)-(3) of $P$ hold. Then these conditions also hold for all $R^{\prime}>R$ and $r^{\prime}>r$ with $R^{\prime}>r^{\prime}$, ###### Proof. Condition (1) is clearly satisfied for all $r^{\prime}>r$. We continue with the proof of condition (3). Let $\sigma\in\Gamma_{c}(\mathfrak{S}_{g})$ be a section with $\mathrm{supp}\,\,\sigma\subseteq U_{r^{\prime}}({\bm{\varepsilon}})$. Since the support is compact, we find a slightly bigger $r^{\prime\prime}>r^{\prime}$ such that $\mathrm{supp}\,\,\sigma\subseteq U_{r^{\prime\prime}}({\bm{\varepsilon}})$. Let $\chi\colon\mathord{\mathbb{R}}\rightarrow[0,1]$ be a smooth function that is identically $1$ near $\mathord{\mathbb{R}}_{\geq r^{\prime\prime}}$ and identically zero near $\mathord{\mathbb{R}}_{\leq r^{\prime}}$. For ${\bm{\varepsilon}}\colon{\mathrm{dom}\,}{\bm{\varepsilon}}\subseteq\mathbf{n}\rightarrow\mathord{\mathbb{Z}}_{2}$ define $\chi^{\bm{\varepsilon}}(x)\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=\prod_{j\in{\mathrm{dom}\,}{\bm{\varepsilon}}}\chi({\bm{\varepsilon}}(j)x_{j}).$ This function is supported within $U_{r^{\prime}}({\bm{\varepsilon}})$. If we do not notationally distinguish $\chi^{\bm{\varepsilon}}$ and $(\Phi_{\bm{\varepsilon}})_{\ast}\chi^{\bm{\varepsilon}}=\chi^{\bm{\varepsilon}}\circ\Phi_{\bm{\varepsilon}}^{-1}$, we have $\displaystyle(\Phi_{\bm{\varepsilon}})_{\ast}(P\chi^{\bm{\varepsilon}}\sigma)$ $\displaystyle=P({\bm{\varepsilon}})\boxtimes\mathrm{id}(\chi^{\bm{\varepsilon}}\sigma)+A\boxtimes D(\chi^{\bm{\varepsilon}}\sigma)$ $\displaystyle=\chi^{\bm{\varepsilon}}\cdot(P({\bm{\varepsilon}})\boxtimes\mathrm{id})(\sigma)+D(\chi^{\bm{\varepsilon}})\cdot(A\boxtimes\mathrm{id})(\sigma)+\chi^{\bm{\varepsilon}}\cdot(A\boxtimes D)(\sigma),$ which is supported within $U_{r^{\prime}}({\bm{\varepsilon}})$, too. Thus, $P$ also restricts to $U_{r^{\prime}}({\bm{\varepsilon}})$ and necessarily decomposes there as in (3). Condition (2) uses the construction of Definition 4.2.5. We will prove condition (2) inductively over $n$, the dimension of the Euclidean factor of $M\times\mathord{\mathbb{R}}^{n}$. For $n=0$ there is nothing to prove, and we can continuue with $n\geq 1$. Let $\chi\colon\mathord{\mathbb{R}}\rightarrow[0,1]$ be a smooth and even function that is identically $1$ near $[-r,r]$ and that vanishes near $\mathord{\mathbb{R}}\setminus(-R,R)$. Let $\\{\chi^{\bm{\alpha}}\\}_{{\bm{\alpha}}\in\mathord{\mathbb{Z}}_{3}^{n}}$ be the corresponding partition of unity on $M\times\mathord{\mathbb{R}}^{n}$ from Definition 4.2.5. We clearly have $P(\sigma)=P(\chi^{\mathbf{0}}\sigma)+\sum_{{\bm{\alpha}}\in\mathord{\mathbb{Z}}_{3}^{n}\setminus\mathbf{0}}P(\chi^{\bm{\alpha}}\sigma).$ The summand $\chi^{\mathbf{0}}\sigma$ is supported within $M\times(RI^{n})^{\circ}$, so, by condition (2) of $P$, also $P(\chi^{\mathbf{0}}\sigma)$ is supported within $M\times(RI^{n})^{\circ}$. Recall the notation $\hat{{\bm{\alpha}}}\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt={\bm{\alpha}}|_{\mathrm{supp}\,{\bm{\alpha}}}\colon\mathrm{supp}\,\,{\bm{\alpha}}\rightarrow\mathord{\mathbb{Z}}_{2}$. The section $\chi^{\bm{\alpha}}\sigma$ is supported within $U_{r}(\hat{{\bm{\alpha}}})$. If ${\bm{\alpha}}\neq\mathbf{0}$, then $\hat{{\bm{\alpha}}}$ is not the empty map so that $P(\chi^{\bm{\alpha}}\sigma)$ is also supported within $U_{r}(\hat{{\bm{\alpha}}})$ by the proof of condition (3) above. Since $\mathrm{supp}\,\sigma\subseteq M\times(R^{\prime}I^{n})^{\circ}$ we can find a smooth and even function $\psi\colon\mathord{\mathbb{R}}\rightarrow[0,1]$ that vanishes near $\mathord{\mathbb{R}}\setminus(-R^{\prime},R^{\prime})$ and such that $\psi^{\mathbf{0}}(x)\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=\prod_{j=1}^{n}\psi(x_{j})$ is identically $1$ on $\mathrm{supp}\,\,\sigma$. The block form of $P$ implies $\displaystyle P(\chi^{\mathbf{{\bm{\alpha}}}}\sigma)$ $\displaystyle=P(\chi^{\mathbf{{\bm{\alpha}}}}\psi^{\mathbf{0}}\sigma)$ $\displaystyle\begin{split}&=\prod_{j\in\mathrm{supp}\,{\bm{\alpha}}}\psi\circ\mathord{\mathrm{pr}}_{j}\cdot P\left(\Bigl{(}\prod_{j\in\mathrm{Null}({\bm{\alpha}})}\psi\circ\mathord{\mathrm{pr}}_{j}\Bigr{)}\cdot\chi^{\bm{\alpha}}\sigma\right)\\\ &\qquad\quad+D(\hat{{\bm{\alpha}}})\left(\prod_{j\in\mathrm{supp}\,{\bm{\alpha}}}\psi\circ\mathord{\mathrm{pr}}_{j}\right)\cdot P\left(\Bigl{(}\prod_{j\in\mathrm{Null}\,{\bm{\alpha}}}\psi\circ\mathord{\mathrm{pr}}_{j}\Bigr{)}\cdot\chi^{\bm{\alpha}}\sigma\right).\end{split}$ By Corollary 4.2.3, $P({\bm{\alpha}}|_{\mathrm{supp}\,{\bm{\alpha}}})$ is a block operator on $\mathfrak{S}_{g({\bm{\alpha}}|_{\mathrm{supp}\,{\bm{\alpha}}})}$ that satisfy conditions (1) - (3) for the same constants as $P$ and $A\boxtimes D({\bm{\alpha}}|_{\mathrm{supp}\,{\bm{\alpha}}})$ is a differential operator and hence only support-decreasing. This and the induction assumption, implies that the section $P\left(\Bigl{(}\prod_{j\in\mathrm{Null}({\bm{\alpha}})}\psi\circ\mathord{\mathrm{pr}}_{j}\Bigr{)}\cdot\chi^{\bm{\alpha}}\sigma\right)$ must be supported within $\\{|x_{j}|<R^{\prime}\,:\,j\in\mathrm{Null}\,{\bm{\alpha}}\\}\cap U_{r}({\bm{\alpha}}|_{\mathrm{supp}\,{\bm{\alpha}}})$. It follows that $\prod_{j\in\mathrm{supp}\,{\bm{\alpha}}}\psi\circ\mathord{\mathrm{pr}}_{j}\cdot P\left(\prod_{j\in\mathrm{Null}({\bm{\alpha}})}\psi\circ\mathord{\mathrm{pr}}_{j}\chi^{\bm{\alpha}}\sigma\right)$ and $D(\hat{{\bm{\alpha}}})\left(\prod_{j\in\mathrm{supp}\,{\bm{\alpha}}}\psi\circ\mathord{\mathrm{pr}}_{j}\right)\cdot P\left(\prod_{j\in\mathrm{Null}\,{\bm{\alpha}}}\psi\circ\mathord{\mathrm{pr}}_{j}\chi^{\bm{\alpha}}\sigma\right)$ are supported within $M\times(R^{\prime}I^{n})^{\circ}\cap U_{r}({\bm{\alpha}}|_{\mathrm{supp}\,\,{\bm{\alpha}}})$, which finishes the proof of (2). ∎ ###### Corollary 4.2.7. Let $P\in\overline{\Psi\mathrm{DO}}^{1}(\mathfrak{S}_{g})$ be a block operator of type $A$ whose core is contained in $M\times RI^{n}$. For ${\bm{\alpha}}\colon\mathbf{n}\rightarrow\mathord{\mathbb{Z}}_{3}$ and $\rho^{\prime}>\rho>R$ define $V_{\rho<\rho^{\prime}}({\bm{\alpha}})\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=M\times\\{x\in\mathord{\mathbb{R}}^{n}\,:\,x_{j}\in(-\rho^{\prime},\rho^{\prime})\text{ if }j\in\mathrm{Null}({\bm{\alpha}}),\alpha_{j}x_{j}>\rho\text{ otherwise}\\}.$ If $\sigma\in\Gamma_{c}(\mathfrak{S}_{g})$ is supported within $V_{\rho<\rho^{\prime}}({\bm{\alpha}})$, then $P\sigma$ is also supported there. ###### Proof. If ${\bm{\alpha}}=\mathbf{0}$, then $V_{\rho<\rho^{\prime}}(0)=M\times\rho^{\prime}I^{n}$ and the claim follows from Lemma 4.2.6. Assume now that ${\bm{\alpha}}\neq\mathbf{0}$, so that $V_{\rho<\rho^{\prime}}({\bm{\alpha}})\subseteq U_{\rho}(\hat{{\bm{\alpha}}})$. For each $u\in\Gamma_{c}(\mathfrak{S}_{g})$ supported with in $V_{\rho<\rho^{\prime}}({\bm{\alpha}})$ we deduce with abuse of notation $\displaystyle Pu=(P(\hat{{\bm{\alpha}}})\boxtimes\mathrm{id})(\chi^{\bm{\alpha}}u)+\mathrm{e/o}\boxtimes D(\hat{{\bm{\alpha}}})(\chi^{\bm{\alpha}})(\chi^{\bm{\alpha}}u)$ The second summand is a differential operator and hence support-decreasing. We know from Corollary 4.2.3 that $P(\hat{{\bm{\alpha}}})$ is again a block operator whose core is contained in $M\times\rho^{\prime}I^{|\mathrm{Null}({\bm{\alpha}})|}$. Since $V_{\rho<\rho^{\prime}}({\bm{\alpha}})\cong\rho^{\prime}I^{|\mathrm{Null}({\bm{\alpha}})|}\times\mathord{\mathbb{R}}_{\rho}^{\hat{{\bm{\alpha}}}}$, it follows from the previous Lemma that $\mathrm{supp}\,\,P({\bm{\alpha}})(u(\text{--},y))\subseteq\rho^{\prime}I^{|\mathrm{Null}({\bm{\alpha}})|}$ and hence $\mathrm{supp}\,\,P({\bm{\alpha}})(u)\subseteq\rho^{\prime}I^{|\mathrm{Null}({\bm{\alpha}})|}\times\mathord{\mathbb{R}}_{\rho}^{\hat{{\bm{\alpha}}}}=V_{\rho<\rho^{\prime}}({\bm{\alpha}}).$ ∎ The technique used in the previous proofs can be used to prove that block operators are local in the following limited sense. ###### Corollary 4.2.8. Let $P$ be a block operator whose core is contained in $M\times\rho I^{n}$. If $\rho^{\prime}>\rho$ and $V\subseteq U_{\rho^{\prime}}({\bm{\varepsilon}})$ is an open set that corresponds under $U_{\rho^{\prime}}({\bm{\varepsilon}})\cong M\times\mathord{\mathbb{R}}^{n}({\bm{\varepsilon}})\times\mathord{\mathbb{R}}^{{\bm{\varepsilon}}}_{\rho^{\prime}}$ to $M\times\mathord{\mathbb{R}}^{n}({\bm{\varepsilon}})\times V^{\bm{\varepsilon}}$, then $(Pu)|_{V}$ only depends on (the germ of) $u|_{\bar{V}}$. ###### Proof. It suffices to show that if $u$ vanishes identically near $\bar{V}$, then $Pu$ vanishes identically on $V$. Let $\phi\colon M\times\mathord{\mathbb{R}}^{n}\rightarrow\mathord{\mathbb{R}}_{\geq 0}$ be a smooth function such that $\phi^{-1}(\mathord{\mathbb{R}}_{>0})=V$ and such that $\phi$ only depends on variables indexed by ${\mathrm{dom}\,}({\bm{\varepsilon}})$. Let $\chi\colon\mathord{\mathbb{R}}\rightarrow[0,1]$ be a smooth even function such that $\chi$ is identically $1$ on $[-\rho,\rho]$ and whose support lies in $(-\rho^{\prime},\rho^{\prime})$. Denote by $\chi^{\bm{\alpha}}$ the partition of unity from Definition 4.2.5. Since $\mathrm{supp}\,\phi\subseteq U_{\rho^{\prime}}({\bm{\varepsilon}})$, we conclude that $\mathrm{supp}\,\,\phi$ and $V_{\rho<\rho^{\prime}}(\hat{{\bm{\alpha}}})$ have non-empty intersection only if $\hat{{\bm{\alpha}}}\geq{\bm{\varepsilon}}$. We know from Corollary 4.2.7 that $\mathrm{supp}\,(P\chi^{\bm{\alpha}}u)\subseteq V_{\rho<\rho^{\prime}}({\bm{\alpha}})$, so we conclude $\phi\cdot P\chi^{\bm{\alpha}}u=0$ if $\hat{{\bm{\alpha}}}\geq{\bm{\varepsilon}}$ is false. On the other hand, if $\hat{{\bm{\alpha}}}\geq{\bm{\varepsilon}}$, then $V_{\rho<\rho^{\prime}}({\bm{\alpha}})\subseteq U_{\rho}({\bm{\varepsilon}})$ and the block decomposition of $P$ implies $\displaystyle\phi\cdot P\chi^{\bm{\alpha}}u$ $\displaystyle=P\phi\chi^{\bm{\alpha}}u+[\phi,P]\chi^{\bm{\alpha}}u=P\chi^{\bm{\alpha}}\underbrace{\phi u}_{=0}-i^{-1}\underbrace{\mathrm{symb}_{1}(P)(\text{--},\mathrm{d}\phi)\chi^{\bm{\alpha}}u}_{=0\text{ on }V}$ $\displaystyle=0\qquad\text{ on }V$ because $u|_{V}=0$. Since $\\{\chi^{\bm{\alpha}}\\}_{{\bm{\alpha}}\in\mathord{\mathbb{Z}}_{3}^{n}}$ is a partition of unity, the previous calculations imply $\phi Pu|_{V}=0$ from which we deduce $Pu|_{V}=0$. ∎ ###### Definition 4.2.9. A block operator $P\in\overline{\Psi\mathrm{DO}}^{1}(\mathfrak{S}_{g})$ is called _block operator of Dirac-type_ , if $A=\mathrm{e/o}$ and $D({\bm{\varepsilon}})={\not{\mathit{D}}}_{\mathord{\mathbb{R}}^{\bm{\varepsilon}}}$ is the Dirac operator of the Euclidean metric on $\mathord{\mathbb{R}}^{\bm{\varepsilon}}$ for all ${\bm{\varepsilon}}\colon{\mathrm{dom}\,}{\bm{\varepsilon}}\rightarrow\mathord{\mathbb{Z}}_{2}$, and if the principal symbol satisfies $\mathrm{symb}_{1}(P)=\mathrm{symb}_{1}({\not{\mathfrak{D}}}_{g})$. The motivating examples for this definition are the Dirac operator of a block metric and the adjoint operator of a block map of pseudo Dirac operators. ###### Example 4.2.10. We apply the discussion of Example 4.1.18 to show that ${\not{\mathfrak{D}}}_{g}$ is a block operator (of Dirac type), if $g$ is a block metric on $M\times\mathord{\mathbb{R}}^{n}$. Assume that the block metric decomposes outside of $M\times rI^{n}$. Then, since differential operators are only support decreasing, we can choose any $R>r$ so that condition (1) and (2) are trivially satisfied. It remains to show that ${\not{\mathfrak{D}}}_{g}$ decomposes appropriately on $U_{r}({\bm{\varepsilon}})$. Since the defining sequence of the Dirac operator (4.1.10) is natural with respect to Pin-structure preserving, isometric diffeomorphisms, we have $\displaystyle(\Phi_{{\bm{\varepsilon}}})_{\ast}{\not{\mathfrak{D}}}_{g}|_{M\times U_{r}({\bm{\varepsilon}})}$ $\displaystyle={\not{\mathfrak{D}}}_{c({\bm{\varepsilon}})_{\ast}g}={\not{\mathfrak{D}}}_{g({\bm{\varepsilon}})+\langle\cdot,\cdot\rangle_{\mathord{\mathbb{R}}^{|{\mathrm{dom}\,}{\bm{\varepsilon}}|}}}={\not{\mathfrak{D}}}_{g({\bm{\varepsilon}})}\boxtimes\mathrm{id}+\mathrm{e/o}\boxtimes{\not{\mathit{D}}}_{\mathord{\mathbb{R}}^{|{\mathrm{dom}\,}{\bm{\varepsilon}}|}}.$ The decomposition of the spinor bundle in Example 4.1.18 has a weaker version. If $U\subseteq\mathord{\mathbb{R}}^{n}$ is open and $g\colon U\rightarrow\mathrm{Riem}(M)$ is a smooth map, then $TU$ is still an orthogonal complement of $TM$ in $T(M\times U)$ with respect to $\mathrm{susp}(g)$. This induces a splitting of the spinor bundle $\mathfrak{S}_{\mathrm{susp}(g)}\cong\mathfrak{S}_{g}\otimes Cl_{n,0}$. ###### Definition 4.2.11. Let $P\colon U\rightarrow\Psi\mathrm{DO}^{m}(\mathfrak{S})$ be a smooth map, where the target carries the (subspace topology of the) amplitude topology. It comes with a smooth map $g\colon U\rightarrow\mathrm{Riem}(M)$. Define on $\mathfrak{S}_{\mathrm{susp}(g)}$ the operators that are given under the isomorphism $\mathfrak{S}_{\mathrm{susp}(g)}\cong\mathfrak{S}_{g}\otimes Cl_{n,0}$ by $\displaystyle P^{\mathrm{ex}}(\sigma\otimes e_{I})(m,t)\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=\left(P_{t}\sigma(\cdot,t)\right)\otimes e_{I},$ $\displaystyle\mathrm{susp}(P)\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=P^{\mathrm{ex}}+\sum_{k=1}^{n}\partial_{t_{k}}\cdot\nabla_{\partial_{t_{k}}}^{\mathfrak{S}_{\mathrm{susp}\,g}}.$ We call $\mathrm{susp}(P)$ the _operator suspension of_ $P$. By the parametrised version of the Atiyah-Singer exterior tensor theorem B.0.29 we have $\mathrm{susp}(P)\in\overline{\Psi\mathrm{DO}}^{m}(\mathfrak{S}_{\mathrm{susp}(g)})$ if $m>0$. We are mostly interested in the case where $U=\mathord{\mathbb{R}}^{n}$ and $P$ is a smooth block map. Note that this implies that the underlying map of Riemannian metrics is also a smooth block map. However, in some proofs, we need the operator suspension on more general open subsets. ###### Lemma 4.2.12. Let $m>0$ and let $P\colon\mathord{\mathbb{R}}^{n}\rightarrow\Psi\mathrm{DO}^{m}(\mathfrak{S})$ be a smooth block map, where the target carries the (subspace topology of the) amplitude topology and underlying block map $g$. Then $P^{\mathrm{ex}}$ and $\mathrm{susp}(P)$ are a block operators of type $\mathrm{id}$ and $\mathrm{e/o}$, respectively. If $P$ takes values in $\Psi\mathrm{Dir}(M)$, then $\mathrm{susp}(P)$ is a block operator of Dirac type. ###### Proof. We will only prove the statement for $\mathrm{susp}(P)$ in the case that $P$ takes values in $\Psi\mathrm{Dir}(M)$, for the other cases are similar. Since $P$ and $g$ are block maps, there is an $r>0$ such that $P$ and $g$ are independent of $x_{i}$ if $|x_{i}|>r$. Then $\mathrm{susp}(g)$ decomposes outside of $M\times rI^{n}$. Since $\mathrm{susp}(P)-\mathrm{e/o}\otimes{\not{\mathit{D}}}_{\mathord{\mathbb{R}}^{n}}$ is $\mathcal{C}^{\infty}(\mathord{\mathbb{R}}^{n})$-linear, the operator restricts to open subsets of the form $M\times U$ with $U\subseteq\mathord{\mathbb{R}}^{n}$ open, and thus $P$ does it, too. It remains to show that $P$ decomposes appropriately on $U_{r}({\bm{\varepsilon}})$. On this subset, a permutation of coordinates $c({\bm{\varepsilon}})\colon M\times U_{r}({\bm{\varepsilon}})\rightarrow M\times\mathord{\mathbb{R}}^{n}({\bm{\varepsilon}})\times\mathord{\mathbb{R}}^{\bm{\varepsilon}}_{r}$ is an isometry if the domain carries the metric $g|_{U_{r}({\bm{\varepsilon}})}$ and the right hand side has $\mathrm{susp}(g)({\bm{\varepsilon}})\oplus\langle\cdot,\cdot\rangle$. Thus the spinor bundle decomposes into $\mathfrak{S}_{\mathrm{susp}(g)({\bm{\varepsilon}})}\boxtimes Cl_{{\bm{\varepsilon}},0}\cong\mathfrak{S}_{g}\otimes Cl_{n-|{\bm{\varepsilon}}|,0}\boxtimes Cl_{{\bm{\varepsilon}},0}$ over $(M\times\mathord{\mathbb{R}}^{n}({\bm{\varepsilon}}))\times\mathord{\mathbb{R}}^{\bm{\varepsilon}}_{r}$ via $\Phi_{\bm{\varepsilon}}=\mathfrak{S}(c({\bm{\varepsilon}}))$. Since $P|_{U_{r}({\bm{\varepsilon}})}$ is independent of the coordinates indexed by ${\mathrm{dom}\,}{\bm{\varepsilon}}$, we derive the equation $\displaystyle(\Phi_{\bm{\varepsilon}})_{\ast}(\mathrm{susp}(P)|_{M\times U_{r}({\bm{\varepsilon}})})$ $\displaystyle=(\Phi_{\bm{\varepsilon}})_{\ast}\bigl{(}P\otimes\mathrm{id}\bigr{)}+(\Phi_{\bm{\varepsilon}})_{\ast}\bigl{(}\mathrm{e/o}_{\mathfrak{S}_{g}}\otimes{\not{\mathit{D}}}_{\mathord{\mathbb{R}}^{n}}\bigr{)}$ $\displaystyle\begin{split}&=P_{c({\bm{\varepsilon}})^{-1}(\text{--})}\otimes\mathrm{id}+(\mathrm{e/o}_{\mathfrak{S}_{g}}\otimes{\not{\mathit{D}}}_{\mathord{\mathbb{R}}^{n}({\bm{\varepsilon}})})\boxtimes\mathrm{id}_{Cl_{{\bm{\varepsilon}},0}}\\\ &\qquad+\mathrm{e/o}_{\mathfrak{S}_{g}}\otimes\mathrm{e/o}_{Cl_{n-|{\bm{\varepsilon}}|,0}}\boxtimes{\not{\mathit{D}}}_{\mathord{\mathbb{R}}^{\bm{\varepsilon}}_{r}}\end{split}$ $\displaystyle\begin{split}&=\Bigl{(}(P|_{M\times\mathord{\mathbb{R}}^{n}({\bm{\varepsilon}})})_{c({\bm{\varepsilon}})^{-1}(\text{--})}\otimes\mathrm{id}_{Cl_{n-|{\bm{\varepsilon}}|,0}}+\mathrm{e/o}_{\mathfrak{S}_{g}}\otimes{\not{\mathit{D}}}_{\mathord{\mathbb{R}}^{n}({\bm{\varepsilon}})}\Bigr{)}\boxtimes\mathrm{id}\\\ &\qquad+\mathrm{e/o}_{\mathfrak{S}_{\mathrm{susp}(g)({\bm{\varepsilon}})}}\boxtimes{\not{\mathit{D}}}_{\mathord{\mathbb{R}}^{\bm{\varepsilon}}},\end{split}$ which shows that $\mathrm{susp}(P)$ decomposes as required. ∎ So far, we have only studied block operators on the level of vector bundles over manifolds. Before we are going to construct the block version of $\Psi\mathrm{Dir}^{\times}(M)$, we will study its (functional) analytic properties in the next section. ### 4.3 Analytical Properties of Block Operators The purpose of this section is to derive the needed analytical properties of block operators. We will show that each block operator realises to a bounded operator on positive Sobolev spaces and that the realisations of block operators of Dirac type satisfy analogous versions of the fundamental elliptic estimate. Moreover, we will show that if they are self-adjoint and invertible near infinity, then they are also Fredholm operators. Many of these properties rely on the structure of a block metric near infinity because we are working on non-compact manifolds. In the following, we only consider pseudo differential operators of order $m\in\\{0,1\\}$. ###### Lemma 4.3.1. Let $g$ be a block metric on $M\times\mathord{\mathbb{R}}^{n}$ and let $P\in\overline{\Psi\mathrm{DO}}^{m}(\mathfrak{S}_{g})$ be a block operator. Then $P$ extends to a bounded linear operator $P\colon H^{s}(\mathfrak{S}_{g})\rightarrow H^{s-m}(\mathfrak{S}_{g})$ for all $s\in\mathord{\mathbb{Z}}$. ###### Proof. We prove this lemma by induction over $n$. For $n=0$, this is classical because $M$ is closed, see Theorem B.0.18. For $n>0$, we pick appropriate $R>r>0$ such that $P$ restricts to $M\times(RI^{n})^{\circ}$ and decomposes on $\\{\varepsilon x_{i}>r\\}$ for all $i\in\\{1,\dots,n\\}$ and all $\varepsilon\in\mathord{\mathbb{Z}}_{2}$. Let $\chi\colon\mathord{\mathbb{R}}\rightarrow[0,1]$ be a smooth, symmetric function that is identically $1$ on $(-r,r)$ and vanishes near $\mathord{\mathbb{R}}\setminus(-R,R)$. Let $\\{\chi^{\bm{\alpha}}\\}_{{\bm{\alpha}}\in\mathord{\mathbb{Z}}_{3}^{n}}$ be the partition of unity obtained from Definition 4.2.5. By the triangle inequality, it suffices to show that $P\cdot\chi^{\bm{\alpha}}$ extends to a bounded operator. If ${\bm{\alpha}}=\mathbf{0}$, then $P\chi^{\mathbf{0}}$ restricts to an operator on the relatively compact set $M\times(RI^{n})^{\circ}$. Thus, by Theorem B.0.18, $P\chi^{\mathbf{0}}$ extends to a bounded operator $H^{s}(\mathfrak{S}_{g})\rightarrow H^{s-1}(\mathfrak{S}_{g})$. If ${\bm{\alpha}}\neq\mathbf{0}$, then there is a tupel $(i,\varepsilon)\in\\{1,\dots,n\\}\times\mathord{\mathbb{Z}}_{2}$ such that $\mathrm{supp}\,\chi^{\bm{\alpha}}$ is contained in $U_{r}((i,\varepsilon))$. The induction hypothesis and [palais1965seminar]*Corollary XIV.1 applied to the discrete Hilbert chains $H^{s}(M\times\mathord{\mathbb{R}}^{n-1};\mathfrak{S}_{g(i,\varepsilon)})$ and $H^{s}(\mathord{\mathbb{R}};Cl_{1,0})$ imply $P(i,\varepsilon)\boxtimes\mathrm{id}+A\boxtimes D(i,\varepsilon)\in Op^{1}(H^{s}(\mathfrak{S}_{g(i,\varepsilon)}\boxtimes Cl_{1,0})),$ which means that this operator induces a bounded linear map $H^{s}(\mathfrak{S}_{g(i,\varepsilon)}\boxtimes Cl_{1,0})\rightarrow H^{s-1}(\mathfrak{S}_{g(i,\varepsilon)}\boxtimes Cl_{1,0})$ for all $s\in\mathord{\mathbb{Z}}$. Since $P$ decomposes into $P(i,\varepsilon)\boxtimes\mathrm{id}+A\boxtimes D(i,\varepsilon)$ on $\mathfrak{S}_{g}|_{U_{r}((i,\varepsilon))}\cong\mathfrak{S}_{g(i,\varepsilon)}\boxtimes Cl_{1,0}|_{\varepsilon\mathord{\mathbb{R}}_{>r}}$, it follows $P\chi^{\bm{\alpha}}\in Op^{1}(\mathfrak{S}_{g})$. ∎ Recall that a sequence of operators $P_{n}$ converges to $P$ in the Atiyah- Singer topology if the $P_{i}$ converge uniformly on compact subsets to $P$. Thus, it is not clear that the fundamental ellitpic estimate carries over to elements in the Atiyah-Singer closure. However, this is the case, if the operator $P$ increases the support in a controlled manner. ###### Lemma 4.3.2. Let $M$ be a not necessarily closed manifold and $E,F\rightarrow M$ be vector bundles over $M$. Let $P\in\overline{\Psi\mathrm{DO}}^{1}(E,F)$ be an elliptic pseudo differential operator whose principal symbol is the principal symbol of an actual pseudo differential operator (i.e. smooth) and that has the following property: For all compact neighbourhoods $K\subseteq M$ there is a relatively compact open subset $W$ such that if $\mathrm{supp}\,u\subseteq K$ then $\mathrm{supp}\,Pu\subseteq W$. Then, for all compact $K\subseteq M$, there is a compact $L\subseteq M$ such that $P$ induces a bounded map $H_{K}^{s}(E)\rightarrow H_{L}^{s-1}(F)$, where $H_{K}^{s}(E)$ denotes the subspace of all section with support in $K$. The operator $P$ furthermore satisfies the elliptic estimate: $||u||_{s}\leq C\left(||u||_{s-1}+||Pu||_{s-1}\right).$ ###### Proof. The subspace $\Gamma_{c}(E)$ is dense in $H^{s}(E)$ for all $s\in\mathord{\mathbb{Z}}$ and the multiplication with a smooth compactly supported function induces a continuous endomorphism on all Sobolev spaces. Thus, if $\chi$ is a smooth compactly supported function that is identically $1$ on $K$, then we can approximate each element of $H_{K}^{s}(E)$ by an element of $\chi\cdot\Gamma_{c}(E)$. If $L$ is the compact subset from the assumption assigned to $\mathrm{supp}\,\,\chi$, then a simple approximation argument shows that $P\colon H_{K}^{s}(E)\rightarrow H^{s-1}_{L}(F)$ is well defined and continuous, because $H^{s-1}_{L}(F)$ is a closed subspace. To prove the elliptic estimate we argue as follows. Pick smooth functions $\rho_{0}$ and $\rho_{1}$ such that $\rho_{0}$ is identically $1$ on $K$ and supported within $L^{\circ}$ and $\rho_{1}$ is identically $1$ on $\mathrm{supp}\,\rho_{0}$ and supported within $L^{\circ}$. Pick further a pseudo-inverse $R$ for $P$, i.e., an actual pseudo differential operator of order $-1$ whose principal symbol is the inverse of $\mathrm{symb}_{1}(P)$. Then $\rho_{0}-\rho_{0}RP\rho_{1}\in\overline{\Psi\mathrm{DO}}^{0}_{L}(E)$ and its principal symbol vanishes. Since $\Psi\mathrm{DO}^{-1}(E)$ lies dense in $\ker\mathrm{symb}_{0}$, we can approximate it by actual pseudo differential operators of order $-1$, see Lemma B.0.24, so there is a $Q\in\Psi\mathrm{DO}^{-1}_{L}(E)$ such that $||\rho_{0}-\rho_{0}RP\rho_{1}-Q||_{s,s}<1/2$. With that knowledge we can derive the required elliptic estimate from the elliptic estimate of actual pseudo differential operators. For any $u\in H_{K}^{s}(E)$ we have $\rho_{j}u=u$ and thus $\displaystyle||u||_{s}$ $\displaystyle=||(\rho_{0}-\rho_{0}RP\rho_{1})u+\rho_{0}RP\rho_{1}u||_{s}$ $\displaystyle\leq||(\rho_{0}-\rho_{0}RP\rho_{1})u||_{s}+||\rho_{0}RPu||_{s}$ $\displaystyle\leq||(\rho-\rho_{0}RP\rho_{1})u-Qu||_{s}+||Qu||_{s}+||\rho_{0}R||_{s-1,s}||Pu||_{s-1}$ $\displaystyle\leq 1/2||u||_{s}+||Q||_{s-1,s}||u||_{s-1}+||\rho_{0}R||_{s-1,s}||Pu||_{s-1},$ where we used the fact that $\rho_{0}R$ is an actual pseudo differential operators of order $-1$ that maps sections to sections that are supported within $\mathrm{supp}\,\rho_{0}$, to deduce that $\rho_{0}R$ induces bounded maps between Sobolev spaces. The chain of inequalities can be rewritten as $\displaystyle||u||_{s}$ $\displaystyle\leq 2\left(||Q||_{s-1,s}||u||_{s-1}+||\rho_{0}R||_{s-1,s}||Pu||_{s-1}\right)$ $\displaystyle\leq 2\max\\{||Q||_{s-1,s},||\rho_{0}R||_{s-1,s}\\}\left(||u||_{s-1}+||Pu||_{s-1}\right),$ which gives the second claim. ∎ Since the constants in the fundamental elliptic estimate may depend on the compact subsets, the result does not immediately carry over to block operators. To carry the result over, we use the following partition of unity which is due to Bunke [bunke2009index], who introduced it in special local coordinates for manifolds with corners. ###### Lemma 4.3.3. Let $P\in\overline{\Psi\mathrm{DO}}^{1}(\mathfrak{S}_{g})$ be a block operator of Dirac type whose core is contained in $M\times(-\rho,\rho)^{n}$ and that decomposes on all $U_{\rho-1/2}({\bm{\varepsilon}})$. For ${\bm{\alpha}}\in\mathord{\mathbb{Z}}_{3}^{n}$, set $V_{\rho}({\bm{\alpha}})=\bigl{\\{}(m,x)\in M\times\mathord{\mathbb{R}}^{n}\,:\,\alpha_{j}x_{j}>\rho-1/2\text{ if }\alpha_{j}\neq 0,\,x_{j}\in(-\rho,\rho)\text{ otherwise}\bigr{\\}}.$ There is a continuous partition of unity $(\phi_{0},\dots,\phi_{2n})$ that satisfies the following three properties: 1. (i) Each $\phi_{l}$ is almost everywhere differentiable, has bounded derivative that is zero at infinity, which means $\lim_{s\to\infty}\mathrm{sup}\\{|\mathrm{d}\phi_{l}|(x)\,|\,x\in M\times\mathord{\mathbb{R}}^{n}\setminus(-s,s)^{n}\\}=0$. 2. (ii) If $u$ is supported within $V_{\rho}({\bm{\alpha}})$ with ${\bm{\alpha}}\neq 0$, then $[P,\phi_{l}]u=\mathbf{c}(\mathrm{grad}(\phi_{l}))(u)$. 3. (iii) $\phi_{0}$ is compactly supported and for each $l>0$ there is a unique tupel $(i,\varepsilon)\in\\{1,\dots,n\\}\times\mathord{\mathbb{Z}}_{2}$ such that $\phi_{l}$ is supported within $\\{\varepsilon x_{i}>\rho\\}$. ###### Proof. Fix $0<a<(2\sqrt{n})^{-1}$. Let $\chi\colon\mathord{\mathbb{R}}_{\geq 0}\rightarrow[0,1]$ be a smooth function whose zero set is $\mathord{\mathbb{R}}_{\leq a}$ and that is identically $1$ near $\mathord{\mathbb{R}}_{\geq 2a}$. We denote its mirror symmetric extension to $\mathord{\mathbb{R}}$ again with $\chi$. For $i\in\\{1,\dots,n\\}$, we define $\widetilde{\chi}_{i}\colon S^{n-1}\rightarrow[0,1]$ via $\widetilde{\chi}_{i}(x)=\chi(\mathord{\mathrm{pr}}_{i}(x))$ and set $\chi_{i}\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=\frac{\widetilde{\chi_{i}}}{\sum_{j=1}^{n}\widetilde{\chi}_{j}}.$ The denominator is always non-zero because it vanishes at $x$ if and only if all summands vanish at $x$, which implies $||x||^{2}<n\cdot a<1$. The $n$-tuple of functions $(\chi_{1},\ldots,\chi_{n})$ has the following three evident properties: 1. 1. $\chi_{i}(x)=0$ if $|x_{i}|<a$, 2. 2. $\sigma^{\ast}\chi_{i}=\chi_{\sigma^{-1}(i)}$ for each permutation $\sigma\in\Sigma_{n}$, 3. 3. $\sum_{j}\chi_{j}=1$. For each ${\bm{\alpha}}=(\alpha_{1},\ldots,\alpha_{n})\in\mathord{\mathbb{Z}}_{3}^{n}$, we define the set $\displaystyle A_{\bm{\alpha}}$ $\displaystyle\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=\\{(m,x)\in M\times\mathord{\mathbb{R}}^{n}\,:\,x_{j}\in[-\rho,\rho]\text{ if }\alpha_{j}=0,\ \alpha_{j}x_{j}\geq\rho\\}$ $\displaystyle\cong M\times[-\rho,\rho]^{|\mathrm{Null}({\bm{\alpha}})|}\times\prod_{j\in\mathrm{supp}\,{\bm{\alpha}}}\alpha_{j}\mathord{\mathbb{R}}_{\geq\rho}.$ The complement of $M\times(-\rho,\rho)^{n}$ is the union of all $A_{\bm{\alpha}}$ with ${\bm{\alpha}}\neq 0$. For $(i,\varepsilon)\in\\{1,\ldots,n\\}\times\mathord{\mathbb{Z}}_{2}$, we define $\hat{\phi}^{\varepsilon}_{i}|_{A_{\bm{\alpha}}}(m,x)=\begin{cases}\chi_{i}\left(\frac{\hat{x}}{||\hat{x}||}\right),&\text{if }\alpha_{i}\neq 0,\\\ 0,&\text{if }\alpha_{i}=0,\end{cases}$ where $\hat{x}=(|\alpha_{1}|x_{1}-\alpha_{1}\rho,\dots,|\alpha_{n}|x_{n}-\alpha_{n}\rho)$. The assignments agree on the overlaps $A_{\bm{\alpha}}\cap A_{\bm{\beta}}$ yielding well defined functions on the complement of $M\times(-\rho,\rho)^{n}$. Each $\hat{\phi}^{\varepsilon}_{i}$ is continuous because it is continuous on all $A_{\bm{\alpha}}$. For $0<\delta<1$, we pick a smooth function $\xi\colon\mathord{\mathbb{R}}_{\geq 0}\rightarrow[0,1]$ whose zero set is $[0,\rho+\delta]$ and which is identically $1$ near $\mathord{\mathbb{R}}_{\geq\rho+1}$. Its mirror-symmetric extension to $\mathord{\mathbb{R}}$ is again denoted by $\xi$. Set $\xi_{i}(x):=\xi(\mathord{\mathrm{pr}}_{i}(x))$ and extend $\hat{\phi_{i}^{\varepsilon}}$ to $M\times\mathord{\mathbb{R}}^{n}$ via $\phi_{i}^{\varepsilon}\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=\xi_{i}\hat{\phi}_{i}^{\varepsilon}$. By construction, all $\phi_{i}^{\varepsilon}$ sum up to $1$ on the complement of $M\times(-\rho+1,\rho+1)^{n}$ and the sum is bounded from above by 1 everywhere. Each $\phi^{\varepsilon}_{i}$ is supported within $\\{\varepsilon x_{i}\geq\rho+\delta\\}$. Thus, by setting $\phi_{0}\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=1-\sum_{(i,\varepsilon)}\phi_{i}^{\varepsilon}$, we have constructed a continuous partition of unity. The function $\phi_{0}$ is supported within $M\times(-\rho-1,\rho+1)^{n}$, which proves property (iii). The points at which the partition of unity might not be differentiable lie in $\bigcup_{{\bm{\alpha}}\neq{\bm{\beta}}}A_{\bm{\alpha}}\cap A_{{\bm{\beta}}}$, which is a zero-set. Outside this set, the differential is given by $\mathrm{d}\phi_{i}^{\varepsilon}=\hat{\phi}_{i}^{\varepsilon}(x)\xi^{\prime}(x_{i})\mathrm{d}x_{i}+\xi(x_{i})\mathrm{d}\chi_{i}\cdot\left(\frac{\hat{x}\cdot\hat{x}^{T}-||\hat{x}||^{2}\mathrm{id}}{||\hat{x}||^{3}}\right)\mathrm{diag}(|\alpha_{1}|,\dots,|\alpha_{n}|).$ Since $\mathrm{d}\chi_{i}$ is bounded, we conclude $|\mathrm{d}\phi_{i}^{\varepsilon}|\in L^{\infty}(M\times\mathord{\mathbb{R}}^{n})\text{ and }\lim_{s\to\infty}\mathrm{sup}\\{|\mathrm{d}\phi_{i}^{\varepsilon}|(x)\,|\,x\in M\times\mathord{\mathbb{R}}^{n}\setminus(-s,s)^{n}\\}=0,$ showing property (i). By construction, the restriction of each $\phi^{\varepsilon}_{i}$ to $V_{\rho}({\bm{\alpha}})$ with ${\bm{\alpha}}\neq 0$ depends only on the coordinates $x_{j}$ indexed by $j\in\mathrm{supp}\,{\bm{\alpha}}$. If $u\in\Gamma_{c}(\mathfrak{S}_{g})$ is supported within $V_{\rho}({\bm{\alpha}})$ with ${\bm{\alpha}}\neq\mathbf{0}$, then the block form of $P$ implies $\displaystyle[P,\phi_{i}^{\varepsilon}]u$ $\displaystyle=\left(P({\bm{\alpha}})+\sum_{j\in\mathrm{supp}\,{\bm{\alpha}}}e_{j}\cdot\partial_{j}\right)\phi_{i}^{\varepsilon}u-\phi_{i}^{\varepsilon}\cdot\left(P({\bm{\alpha}})+\sum_{j\in\mathrm{supp}\,{\bm{\alpha}}}e_{j}\cdot\partial_{j}\right)u$ $\displaystyle=\phi^{\varepsilon}_{i}(P({\bm{\alpha}})-P({\bm{\alpha}}))u+\sum_{j\in\mathrm{supp}\,{\bm{\alpha}}}[e_{j}\cdot\partial_{x_{j}},\phi_{i}^{\varepsilon}]u$ $\displaystyle=\mathrm{grad}(\phi_{i}^{\varepsilon})\cdot u=\mathbf{c}(\mathrm{grad}(\phi_{i}^{\varepsilon}))(u).$ This shows property (ii). ∎ We use this partition of unity to derive the fundamental elliptic estimates for block operators of Dirac type. The idea of the following two propositions is borrowed from [ebert2017indexdiff]*Prop 3.3. ###### Proposition 4.3.4. Let $P\in\overline{\Psi\mathrm{DO}}^{1}(\mathfrak{S}_{g})$ be an elliptic block operator of Dirac type. Then $P$ satisfies the fundamental elliptic inequality, which means that there is a $C>0$ such that $||u||_{1}\leq C\left(||u||_{0}+||Pu||_{0}\right)$ for all $u\in H^{1}(\mathfrak{S}_{g})$ . ###### Proof. The proof will be carried out by induction over $n$, the dimension of the Euclidean part of our block manifolds $M\times\mathord{\mathbb{R}}^{n}$. For $n=0$ our base manifold is compact and the theorem follows from Lemma 4.3.2. Assume the statement for $n-1$ and pick an elliptic block operator $P\in\overline{\Psi\mathrm{DO}}^{1}(\mathfrak{S}_{g})$ over a block metric $g\in\widetilde{\mathcal{R}_{n}}(M)$. Pick $R>r>0$ such that $g$ decomposes outside $M\times(-r,r)^{n}$ and $M\times(-R,R)^{n}$ is a core for $P$. Since $P$ is of block form the classical elliptic estimate gives us $||u||_{1}\leq C_{0,R}\left(||u||_{0}+||Pu||_{0}\right)$ for some $C_{0,R}$ and all sections $u$ with support in $M\times(-R,R)^{n}$. If $u$ is supported within $U_{r}(i,\varepsilon)=\\{\varepsilon x_{i}>r\\}$, then we can identify $\mathfrak{S}_{g}|_{U_{r}(i,\varepsilon)}$ with $\mathfrak{S}_{g(i,\varepsilon)}$ and $P$ with $P(i,\varepsilon)\boxtimes\mathrm{id}+\mathrm{e/o}\boxtimes{\not{\mathit{D}}}$. Since $g$ is a product metric here, we can apply Fubini’s theorem to get: $\displaystyle||u||_{1}^{2}$ $\displaystyle=\int_{\varepsilon\mathord{\mathbb{R}}_{>r}}||u(t)||_{1,g(i,\varepsilon)}^{2}\mathrm{d}t+\int_{\varepsilon\mathord{\mathbb{R}}_{>r}}||\partial_{t}u(t)||_{0,g(i,\varepsilon)}^{2}\mathrm{d}t$ $\displaystyle\leq\int C^{2}\left(||u(t)||^{2}_{0,g(i,\varepsilon)}+||P(i,\varepsilon)u||^{2}_{0,g(i,\varepsilon)}\right)+||\partial_{t}u(t)||_{0,g(i,\varepsilon)}^{2}\mathrm{d}t$ $\displaystyle\leq C^{2}\int||u(t)||^{2}_{0,g(i,\varepsilon)}+||P(i,\varepsilon)u||^{2}_{0,g(i,\varepsilon)}+||\partial_{t}u(t)||_{0,g(i,\varepsilon)}^{2}\mathrm{d}t$ $\displaystyle=C^{2}\int||u(t)||^{2}_{0,g(i,\varepsilon)}+\langle P(i,\varepsilon)^{2}u,u\rangle_{0,g(i,\varepsilon)}-\langle\partial^{2}_{t}u(t),u(t)\rangle_{0,g(i,\varepsilon)}\mathrm{d}t$ $\displaystyle=C^{2}\int||u(t)||^{2}_{0,g(i,\varepsilon)}+\langle P(i,\varepsilon)^{2}u-\partial^{2}_{t}u(t),u\rangle_{0,g(i,\varepsilon)}\mathrm{d}t$ $\displaystyle=C^{2}\int||u(t)||_{0,g(i,\varepsilon)}^{2}+\langle(P(i,\varepsilon)\boxtimes\mathrm{id}+\mathrm{e/o}\boxtimes{\not{\mathit{D}}})^{2}u(t),u(t)\rangle\mathrm{d}t$ $\displaystyle=C^{2}\int||u(t)||^{2}_{0,g(i,\varepsilon)}+||(Pu)(t)||_{0,g(i,\varepsilon)}^{2}\mathrm{d}t$ $\displaystyle=C^{2}(||u||_{0}^{2}+||Pu||_{0}^{2}),$ from which we immediately get the fundamental elliptic inequality for this case. Finally, we need to patch these inequalities together. We define the set $L\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=\\{0\\}\cup\\{1,\dots,n\\}\times\mathord{\mathbb{Z}}_{2}$ and let $(\phi_{l})_{l\in L}$ be the partition of unity from Lemma 4.3.3 adapted to $P$. We use this partition of unity to patch the inequalities together: $\displaystyle||u||_{1}$ $\displaystyle=||\sum_{L}\phi_{l}u||_{1}\leq\sum_{L}||\phi_{l}u||_{1}$ $\displaystyle\leq C_{0}(||\phi_{0}u||_{0}+||P\phi_{0}u||_{0})+\sum_{L\setminus\\{0\\}}C_{l}(||\phi_{l}u||_{0}+||P\phi_{l}u||_{0})$ $\displaystyle\leq(2n+1)\max\\{C_{l}\\}||u||_{0}+\sum_{L}C_{l}(||\phi_{l}Pu||_{0}+||[P,\phi_{l}\cdot]u||_{0})$ $\displaystyle\leq(2n+1)\max\\{C_{l},||[P,\phi_{l}\cdot]||_{0,0}\\}\left(||u||_{0}+||Pu||_{0}\right).$ This concludes the induction step. ∎ ###### Definition 4.3.5. A block operator $P\in\overline{\Psi\mathrm{DO}}^{1}(\mathfrak{S}_{g})$ of Dirac type is _invertible_ if its induced operator $H^{1}(\mathfrak{S}_{g})\rightarrow H^{0}(\mathfrak{S}_{g})$ is an isomorphism. ###### Remark 4.3.6. By Proposition 4.3.4 the defining condition is equivalent to require that $P$ is an invertible unbounded operator $H^{0}(\mathfrak{S}_{g_{P}})$. Recall that this means that the extension $P\colon\overline{\Gamma_{c}(\mathfrak{S}_{g_{P}})}^{||\text{--}||_{P}}\rightarrow H^{0}(\mathfrak{S}_{g})$ is invertible, where the domain is the completion of $\Gamma_{c}(\mathfrak{S}_{g_{P}})$ with respect to the graph norm $||\text{--}||_{P}=||\text{--}||+||P(\text{--})||$. If the block operator $P$ has invertible faces, then the previous proposition can be improved. ###### Proposition 4.3.7. Let $P$ be a block operator of Dirac type whose faces $P(i,\varepsilon)$ are invertible. Then $P$ satisfies the elliptic estimate $||u||_{1}\leq C\left(||Pu||_{0}+||\phi_{0}u||_{0}+\sum_{l=1}^{2n}||[P,\phi_{l}\cdot]u||_{0}\right),$ where $\\{\phi_{l}\\}$ is the partition of unity from Lemma 4.3.3. Furthermore, $P$ has closed image and finite dimensional kernel. If $P$ is self-adjoint, then $P$ is an (unbounded) Fredholm operator with domain $H^{1}(M\times\mathord{\mathbb{R}}^{n};\mathfrak{S}_{g})$. The proof of this proposition requires the main Fredholm Lemma. ###### Lemma 4.3.8 ([salamon2018functional]*Lemma 4.3.9). Let $X,Y$ be Banach spaces and $Z$ be a normed vector space. Let $A\colon X\rightarrow Y$ be a bounded linear map and let $K\colon X\rightarrow Z$ be a compact operator. If there is a constant $C$ such that for all $x\in X$ the inequality $||x||_{X}\leq C(||Ax||_{Y}+||Kx||_{Z})$ holds, then $A$ has closed image and finite dimensional kernel. ###### Proof of Proposition 4.3.7. Pick $R>r$ such that $M\times(-R,R)^{n}$ is a core of $P$ and that the block operator $P$ decomposes on $U_{r}(i,\varepsilon)$ under the identification $U_{r}(i,\varepsilon)\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=\\{(m,x)\in M\times\mathord{\mathbb{R}}^{n}\,:\,\varepsilon x_{i}>r\\}\cong M\times\mathord{\mathbb{R}}^{n}(i,\varepsilon)\times\\{\varepsilon x_{i}>r\\}$ into the operator $P(i,\varepsilon)\boxtimes\mathrm{id}+\mathrm{e/o}\boxtimes{\not{\mathit{D}}}$. For each $u\in\Gamma_{c}(M\times\mathord{\mathbb{R}}^{n},\mathfrak{S}_{g})$ with support in $U_{r}(i,\varepsilon)$, we have $\displaystyle||Pu||_{0}^{2}$ $\displaystyle=\int_{M\times\mathord{\mathbb{R}}^{n}}|Pu|^{2}=\int_{M\times\mathord{\mathbb{R}}^{n}}|P(i,\varepsilon)u+\partial_{i}\cdot\partial_{i}u|^{2}$ $\displaystyle=\int_{M\times\mathord{\mathbb{R}}^{n}}\langle u,(P(i,\varepsilon)+\partial_{i}\cdot\partial_{i})^{2}u\rangle=\int_{M\times\mathord{\mathbb{R}}^{n}}\langle u,P(i,\varepsilon)^{2}u-\partial_{i}^{2}u\rangle$ $\displaystyle=\int_{M\times\mathord{\mathbb{R}}^{n}}|P(i,\varepsilon)u|^{2}+|\partial_{i}u|^{2}$ $\displaystyle=\int_{\varepsilon\cdot\mathord{\mathbb{R}}_{\geq r}}||(P(i,\varepsilon)u)(x)||^{2}_{L^{2}(M\times\mathord{\mathbb{R}}^{n}(i,\varepsilon),\mathfrak{S}_{g})}+\int_{M\times\mathord{\mathbb{R}}^{n}}|\partial_{i}u|^{2}$ $\displaystyle\geq\int_{\varepsilon\cdot\mathord{\mathbb{R}}_{\geq r}}c^{-2}_{(i,\varepsilon)}||u(x)||^{2}_{H^{1}(M\times\mathord{\mathbb{R}}^{n}(i,\varepsilon),\mathfrak{S}_{g})}+\int_{M\times\mathord{\mathbb{R}}^{n}}|\nabla_{\partial_{i}}u|^{2}$ $\displaystyle\geq\mathrm{min}\\{c^{-2}_{(i,\varepsilon)},1\\}||u||^{2}_{H^{1}(M\times\mathord{\mathbb{R}}^{n},\mathfrak{S}_{g})}.$ In this chain of equations, we used in the second line that $P$ is symmetric and that $P(i,\varepsilon)$ anti-commutes with $e_{i}$. In the fourth line, we applied Fubini’s theorem, while in the fifth line, we used the invertibility of $P(i,\varepsilon)$ and that $\nabla_{\partial_{i}}=\partial_{i}$. In the last line we estimated against the minimum and identified the expression as the Sobolev-1-norm. On the other hand, if $u$ is supported within $M\times(-R,R)^{n}$, then $Pu$ is again supported within this relatively compact subset and the classical elliptic estimate implies $||u||_{1}\leq C(||u||_{0}+||Pu||_{0}).$ We now use the special adapted partition of unity from Lemma 4.3.3 with $\rho=R$ and the notation from there to patch the inequalities together. Using that $\phi_{l}$ is contained in some $U_{r}(i,\varepsilon)$ for $l\neq 0$, the previous estimates yield $\displaystyle||u||_{1}$ $\displaystyle=||\sum_{L}\phi_{l}u||_{1}\leq\sum_{L}||\phi_{l}u||_{1}$ $\displaystyle\leq C_{0}(||\phi_{0}u||_{0}+||P\phi_{0}u||_{0})+\sum_{L\setminus\\{0\\}}C_{l}||P\phi_{l}u||_{0}$ $\displaystyle\leq C_{0}||\phi_{0}u||_{0}+\sum_{L}C_{l}(||\phi_{l}Pu||_{0}+||[P,\phi_{l}\cdot]u||_{0})$ $\displaystyle\leq C_{0}||\phi_{0}u||_{0}+\max\\{C_{l}\cdot||\phi_{l}||_{0,0},1\\}\cdot\left(\sum_{L}||Pu||_{0}+||[P,\phi_{l}\cdot]u||_{0}\right)$ $\displaystyle\leq C\left(||Pu||_{0}+||\phi_{0}u||_{0}+\sum_{L}||[P,\phi_{l}\cdot]u||_{0}\right),$ where $C_{l}$ is the inverse of $\min\\{c_{(i,\varepsilon)^{-1}}\\}$ for the pair $(i,\varepsilon)$ corresponding to $l$. It turns out that $\phi_{0}$ and all $[P,\phi_{l}\cdot]$ are compact operators. Indeed, $\phi_{0}$ is compactly supported, so $\phi_{0}\cdot$ is a compact operator by the classical Rellich-Lemma. To see that $T_{l}\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=[P,\phi_{l}\cdot]$ is compact, note first that it is an element of $\overline{\Psi\mathrm{DO}}^{1}(\mathfrak{S}_{g})$. Symbol calculus shows that its principal symbol vanishes. Pick a partition of unity as in Definition 4.2.5 such that $\chi^{\mathbf{0}}$ is identically $1$ on $M\times[-R,R]^{n}$. We decompose it into $T_{l}=\sum_{{\bm{\alpha}}\in\mathord{\mathbb{Z}}_{3}^{n}}T_{l}\chi^{\bm{\alpha}}$. By Lemma 4.2.6 the summand $T_{l}\chi^{\mathbf{0}}$ is supported within $\mathrm{supp}\,\chi^{\mathbf{0}}$ in the sense that $\mathrm{supp}\,T_{l}\chi^{\mathbf{0}}u\subseteq\mathrm{supp}\,\chi^{\mathbf{0}}$ and $T_{l}\chi^{\bm{\alpha}}u=0$ if $\mathrm{supp}\,\chi^{\bm{\alpha}}\cap\mathrm{supp}\,u=\emptyset$. Since the principal symbols of $T_{l}\chi^{\mathbf{0}}$ is zero, the Atiyah-Singer exact sequence, see Lemma B.0.24, implies that $T_{l}\chi^{\mathbf{0}}$ is compact. If ${\bm{\alpha}}\neq 0$, then $\hat{{\bm{\alpha}}}\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt={\bm{\alpha}}|_{\mathrm{supp}\,{\bm{\alpha}}}$ is not the empty map and $\chi^{\bm{\alpha}}u$ is supported within $U_{R}(\hat{{\bm{\alpha}}})$. Thus, property (ii) of the partition of unity $\\{\phi_{l}\\}$ yields $T_{l}\chi^{\bm{\alpha}}u=\mathrm{grad}\phi_{l}\cdot\chi^{\bm{\alpha}}u.$ By Lemma 4.3.3 (i) and the generalised Rellich lemma for complete manifolds, see [bunke2009index]*p.50, we conclude that $T_{l}\chi_{\bm{\alpha}}$ is also compact. A finite sum of compact operators is still compact, so $T_{l}$ is a compact operator. Lemma 4.3.8 now yields that $P$ has finite kernel and closed image. If $P$ is a self-adjoint (unbounded) operator on $L^{2}(M\times\mathord{\mathbb{R}}^{n},\mathfrak{S}_{g})$, then the maximal domain agrees with the minimal domain. The elliptic estimate implies that the graph norm of $P$ is equivalent to $||\text{--}||_{1}$, so the (minimal) domain of $P$ agrees with $H^{1}(M\times\mathord{\mathbb{R}}^{n},\mathfrak{S}_{g})$. The kernel and the cokernel of a self-adjoint operator are isomorphic, so the cokernel is also finite dimensional and $P$ is a Fredholm operator. ∎ Usually, it is not easy to show that a (pseudo) differential operator is self- adjoint. On complete manifolds, there is, however, a general criterion, the boundness of the propagation speed. ###### Proposition 4.3.9 ([higson2000analytic]*Prop. 10.2.11). Let $N$ be a complete Riemannian manifold and let $D$ be a symmetric differential operator on $M$. If the _propagation speed_ $c_{D}\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=\sup\\{||\mathrm{symb}_{1}(D)(x,\xi)||\,:\,x\in M,\xi\in T^{\vee}_{x}M,||\xi||=1\\}$ is finite, then $D$ is (essentially) self-adjoint, i.e. its closure is a self- adjoint operator. ###### Corollary 4.3.10. Let $g\in\widetilde{\mathcal{R}_{n}}(M)$ be a block metric on $M\times\mathord{\mathbb{R}}^{n}$. Then ${\not{\mathfrak{D}}}_{g}$ is a self- adjoint operator. More generally, if $P\in\overline{\Psi\mathrm{DO}}^{1}(\mathfrak{S}_{g})$ is a symmetric, block operator of Dirac type, such that $P-{\not{\mathfrak{D}}}_{g}$ extends to a bounded operator on $H^{0}(\mathfrak{S}_{g})=L^{2}(\mathfrak{S}_{g})$, then $P$ is a self-adjoint operator. ###### Proof. A block metric is complete by Proposition 3.1.5 and the propagation speed obviously satisfies $c_{{\not{\mathfrak{D}}}_{g}}=1$. By the preceding proposition, ${\not{\mathfrak{D}}}_{g}$ is therefore self-adjoint. The second statement follows from Exercise 1.9.21 of [higson2000analytic]. ∎ A common strategy to verify that a self-adjoint operator is invertible is to verify that it squares to a uniformly positive operator. We recall this concept. ###### Definition 4.3.11. An unbounded operator $T$ on a Hilbert space $H$ is called _non-negative_ , if $\langle Tx,x\rangle\geq 0$ for all $x\in{\mathrm{dom}\,}T$. It is called _uniformly positive_ , if there is a constant $c>0$ such that $\langle Tx,x\rangle\geq c||x||^{2}$. An unbounded operator is _bounded from below_ if there is a positive constant $c>0$ such that $||Tx||\geq c||x||$ for all $x\in{\mathrm{dom}\,}T$. Clearly, a closed unbounded operator $T$ is bounded from below if and only if $T^{\ast}T$ is uniformly positive. The next theorem is an immediate consequence from unbounded functional calculus, see [werner2006funktionalanalysis]. ###### Lemma 4.3.12. Let $T$ be a self-adjoint unbounded operator on a separable Hilbert space. 1. (i) If $T$ is uniformly positive, then $T$ is invertible. 2. (ii) If $T$ is closed and invertible, then $T^{\ast}T$ is uniformly positive. Being uniformly positive is an open condition for block operators of Dirac type in the space $\mathrm{Hom}(H^{1},H^{0})$, where $H^{j}=H^{j}(\mathfrak{S}_{g})$. ###### Lemma 4.3.13. Let $P,Q\in\mathrm{Hom}(H^{1},H^{0})$ be bounded operators. Let $P$ be bounded from below and satisfy a fundamental elliptic estimate. If $Q$ is sufficiently close to $P$ in the operator norm of $\mathrm{Hom}(H^{1},H^{0})$, then $Q$ is also bounded from below. ###### Proof. Given $c\in(0,1)$ such that $||Px||_{0}>c||x||_{0}$ and a constant $C=C(P)$ making the fundamental elliptic estimate valid for $P$. We claim that all operators $Q$ satisfying the inequality $||P-Q||_{1,0}<\frac{c}{(c+1)C}$ are bounded from below. Indeed, the fundamental elliptic estimate applied to $P$ and the assumed inequality yield $\displaystyle||Qu||_{0}$ $\displaystyle\geq\bigl{|}||Pu||_{0}-||(Q-P)u||_{0}\bigr{|}$ $\displaystyle\geq||Pu||_{0}-||P-Q||_{1,0}C(||u||_{0}+||Pu||_{0})$ $\displaystyle=(1-||P-Q||_{1,0}\cdot C)||Pu||_{0}-||P-Q||_{1,0}C||u||_{0}.$ Using uniform positivity and the assumed inequality a second time, we continue the estimation with $\displaystyle||Qu||_{0}$ $\displaystyle\geq c(1-C)||P-Q||_{1,0}||u||_{0}-C||P-Q||_{1,0}||u||_{0}$ $\displaystyle=\underbrace{\bigl{(}c-(c+1)C||P-Q||_{1,0}\bigr{)}}_{>0}||u||_{0}$ This proves the claim. ∎ ##### Gluing Techniques In the forthcoming sections we will carry out several cut and paste arguments for block operators. Although we will start and end with a block operator on a complete manifold, during the procedures we will consider operators on non- complete manifolds. Being self-adjoint or invertible are global properties of an operator that are usually lost when we restrict the operator to an open subset, so we cannot expect them to hold through the entire cut and paste procedure. On the other hand, being symmetric and bounded from below are properties that are preserved under such restrictions and the purpose of this section is to show that they are also preserved under certain gluing constructions. The first gluing technique for block operator roughly says that being symmetric is a sheaf-like property, provided the operators are suitably local. ###### Lemma 4.3.14. Let $E\rightarrow M$ be a vector bundle over a not necessarily compact manifold $M$. Let $\\{U_{\alpha}\\}_{\alpha\in A}$ be a locally finite open cover and let $\\{P_{\alpha}\\}_{\alpha\in A}$ be a family of operators $P_{\alpha}\colon\Gamma_{c}(U_{\alpha},E|_{U_{\alpha}})\rightarrow\Gamma_{c}(U_{\alpha},E|_{U_{\alpha}})\quad\text{ and }\quad P_{\alpha}\in\overline{\Psi\mathrm{DO}}^{1}(E|_{U_{\alpha}})$ such that $P_{\alpha}|_{U_{\alpha}\cap U_{\beta}}=P_{\beta}|_{U_{\alpha}\cap U_{\beta}}\colon\Gamma_{c}(U_{\alpha\beta},E_{U_{\alpha\beta}})\rightarrow\Gamma_{c}(U_{\alpha\beta},E_{U_{\alpha\beta}})$ for all $\alpha,\beta\in A$. Assume further that there is a partition of unity $\\{\chi_{\alpha}\\}_{\alpha\in A}$ that is subordinated to the cover $\\{U_{\alpha}\\}_{\alpha\in A}$ such that the operators acts as derivation on the partition of unity, that is $[P_{\alpha},\chi_{\beta}\cdot]u=i^{-1}\mathrm{symb}_{1}(P)(\text{--},\mathrm{d}\chi_{\beta})\cdot u.$ Then each $P_{\alpha}$ is symmetric if and only if the operator $P\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=\sum_{\alpha\in A}P_{\alpha}\circ\chi_{\alpha}$ is symmetric. ###### Proof. Assume that $P$ is symmetric, then also $P_{a}$ is symmetric for all $a\in A$ because for all $u\in\Gamma_{c}(U_{a},E|_{U_{a}})$ we have $Pu=\sum_{\alpha\in A}P_{\alpha}\chi_{\alpha}u=\sum_{\\{\alpha:U_{\alpha}\cap U_{a}\neq\emptyset\\}}P_{a}\chi_{\alpha}u=P_{a}u.$ For the converse, let $u,v\in\Gamma_{c}(M,E)$ be two sections. Set $u_{\alpha}\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=\chi_{\alpha}u$ and $v_{\beta}\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=\chi_{\beta}v$. By bilinearity, it suffices to show that $(P_{\alpha}u_{\alpha},v_{\beta})=(u_{\alpha},P_{\beta}v_{\beta})$ for all $\alpha,\beta\in A$. To verify this, we rely on the following two facts that are immediate consequences from the derivation property: * (I) $\mathrm{supp}\,\,P_{\beta}u_{\alpha}\subseteq\mathrm{supp}\,\,\chi_{\alpha}$ for all $\alpha,\beta\in A$. * (II) If $u|_{\\{\chi_{\alpha}\neq 0\\}}=0$, then $(P_{\beta}u)|_{\\{\chi_{\alpha}\neq 0\\}}=0$. Indeed, property (I) follows from $\displaystyle P_{\beta}u_{\alpha}$ $\displaystyle=P_{\beta}\chi_{\alpha}u=\chi_{\alpha}P_{\beta}u+[P_{\beta},\chi_{\alpha}]u=\chi_{\alpha}u+i^{-1}\mathrm{symb}_{1}(P_{\beta})(\text{--},\mathrm{d}\chi_{\alpha})u$ and that $\mathrm{symb}_{1}(P_{\beta})(\text{--},\mathrm{d}\chi_{\alpha})=\mathrm{symb}_{1}(P_{\beta})(\text{--},0)=0$ on $\\{\chi_{\alpha}=0\\}$. Property (II) follows from $\displaystyle(\chi_{\alpha}P_{\beta}u)|_{\chi_{\alpha}\neq 0}=P_{\beta}\underbrace{\chi_{\alpha}u}_{=0}-i^{-1}\mathrm{symb}_{1}(P_{\beta})(\text{--},\mathrm{d}\chi_{\alpha})\underbrace{u|_{\chi_{\alpha}\neq 0}}_{=0}.$ Now, for each $\alpha\in A$, we pick a smooth function $\psi_{\alpha}$ that is identically $1$ near $\mathrm{supp}\,\,\chi_{\alpha}$ and supported within $U_{\alpha}$. Symmetry of $P$ now follows from the calculation $\displaystyle(P_{\alpha}u_{\alpha},v_{\beta})$ $\displaystyle\overset{(\mathrm{I})}{=}(\psi_{\alpha}P_{\alpha}u_{\alpha},v_{\beta})=(P_{\alpha}u_{\alpha},\psi_{\alpha}v_{\beta})$ $\displaystyle=(u_{\alpha},P_{\alpha}\psi_{\alpha}v_{\beta})$ $\displaystyle=(u_{\alpha},P_{\beta}\psi_{\alpha}v_{\beta})\qquad\text{because }P_{\alpha}|_{U_{\alpha\beta}}=P_{\beta}|_{U_{\alpha\beta}}$ $\displaystyle=(u_{\alpha},(P_{\beta}\psi_{\alpha}v_{\beta})|_{\mathrm{supp}\,\,\chi_{\alpha}})$ $\displaystyle=(u_{\alpha},P_{\beta}v_{\beta})$ in which we use (II) and linearity of $P_{\beta}$ to deduce that $(P_{\beta}\psi_{\alpha}v_{\beta})|_{\mathrm{supp}\,\,\chi_{\alpha}}$ only depends on (the germ of) $\psi_{\alpha}v_{\beta}|_{\mathrm{supp}\,\,\chi_{\alpha}}=v_{\beta}|_{\mathrm{supp}\,\,\chi_{\alpha}}$. ∎ The second gluing technique yields that two invertible block operators with matching faces can be glued together to an invertible operator, provided the “gluing strip” is sufficiently large. We carry out the gluing construction in a more general setup as we allow our underlying Riemannian manifolds to be non-complete. This forces us, to work with the following weaker notion that is equivalent to invertibility if the operator in question is self-adjoint. ###### Definition 4.3.15. Let $P$ be a densely defined operator on a Hilbert space $H$. We call $P$ _(uniformly) bounded from below_ , if there is a $c>0$ such that $||Pv||\geq c||v||$ for all $v\in\mathrm{dom}(P)$. More generally, an operator $P\in\overline{\Psi\mathrm{DO}}^{1}(E)$ is bounded from below, if $P\colon\Gamma_{c}(M,E)\rightarrow\Gamma(M,E)\subset L^{2}_{loc}(M,E)$ is bounded from below (meaning we also allow $||P\sigma||_{0}=\infty$). We work in the following setup. ###### Setup 4.3.16. Let $H\subseteq M$ be a separating hypersurface that is closed as a subset but not necessarily compact. Set $M\setminus H=M^{-}\sqcup M^{+}$ and let $\Phi\colon H\times(-a,a)\hookrightarrow M$ be an open embedding that restricts to the inclusion on $H\times\\{0\\}=H$ and satisfies $\Phi(H\times(-a,0))\subseteq M^{-}$ and $\Phi(H\times(0,a))\subseteq M^{+}$. We further denote $M_{\pm a}\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=M^{\pm}\setminus\Phi(H\times(-a,a))$ and, for all $t,t_{1},t_{2}\in(-a,a)$, we set $\displaystyle M_{<t}\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=M_{-a}\,\cup\,\Phi(H\times(-a,t)),\qquad M_{>t}\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=\Phi(H\times(t,a))\,\cup\,M_{a},\quad\text{and}$ $\displaystyle M_{(t_{1},t_{2})}\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=\Phi(H\times(t_{1},t_{2})).$ ###### Definition 4.3.17. Let $\mathcal{A}\subseteq\mathcal{C}^{\infty}(M,\mathord{\mathbb{R}})$ be a sub-algebra of smooth functions on $M$. An operator $P\in\overline{\Psi\mathrm{DO}}^{1}(E)$ is an $\mathcal{A}$_-derivation_ , if $[P,f\cdot]=i^{-1}\mathrm{symb}_{1}(P)(\text{--},\mathrm{d}f)$ for all $f\in\mathcal{A}$. We are interested in the algebra $\mathcal{A}\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=(\mathord{\mathrm{pr}}_{2}\circ\Phi^{-1})^{\ast}\mathcal{C}_{c}^{\infty}((-a,a),\mathord{\mathbb{R}}))$ considered as a sub-algebra of $\mathcal{C}^{\infty}(M,\mathord{\mathbb{R}})$ (via extension by zero). ###### Lemma 4.3.18. Let $E\rightarrow M$ be a vector bundle over a manifold as in setup 4.3.16. Assume that $P\in\overline{\Psi\mathrm{DO}}^{1}(E)$ restricts to $M_{>b}$ and $M_{<-b}$ for some $0<b<a$ and that $P$ is an $\mathcal{A}$-derivation. If $\sigma\in\Gamma_{c}(M,E)$ is supported within $M_{(t_{1},t_{2})}$, then $P\sigma$ is also supported within $M_{(t_{1},t_{2})}$ for all $-a<t_{1}<t_{2}<a$. In particular, $(P\sigma)|_{M_{(t_{1},t_{2})}}$ only depends on the germ of $\sigma|_{\overline{M}_{(t_{1},t_{2})}}$. Analogous statements hold for $M_{>t}$ and $M_{<t}$ for all $t\in(-b,b)$. ###### Proof. We only prove the statement for $M_{(t_{1},t_{2})}$. The other proofs are quite similar. Assume $\sigma$ is supported within this set. Since the support is compact, there are $t_{1}<t_{1}^{\prime}<t_{2}^{\prime}<t_{2}$ such that $\sigma$ is also supported within $M_{(t_{1}^{\prime}t_{2}^{\prime})}$. Let $\chi\colon(t_{1},t_{2})\rightarrow[0,1]$ be identically $1$ on $(t_{1}^{\prime},t_{2}^{\prime})$ and compactly supported. To ease the notation, we denote $\chi\circ\mathord{\mathrm{pr}}_{2}\circ\Phi^{-1}$ (and its extension by zero) again with $\chi$. We have $\displaystyle P\sigma$ $\displaystyle=P\chi\sigma=\chi P\sigma+[P,\chi\cdot]\sigma$ $\displaystyle=\chi P\sigma+i^{-1}\mathrm{symb}_{1}(P)(\text{--},\chi^{\prime})\sigma$ $\displaystyle=\chi P\sigma,$ because $\chi\equiv 1$ on $\mathrm{supp}\,\sigma$. Thus, $P\sigma$ is contained in $\mathrm{supp}\,\chi\subseteq M_{(t_{1},t_{2})}$. To prove that $P\sigma|_{M_{(t_{1},t_{2})}}$ only depends on the germ of $\sigma|_{\overline{M_{(t_{1},t_{2})}}}$, it suffices to show that if $\sigma$ is identically zero near $\overline{M_{(t_{1},t_{2})}}$, then $P\sigma$ vanishes identically on $M_{(t_{1},t_{2})}$. But if $\sigma$ vanishes near $\overline{M_{(t_{1},t_{2})}}$, then, by compactness of $\mathrm{supp}\,\sigma$, there are $t_{1}^{\prime}<t_{1}$ and $t_{2}^{\prime}>t_{2}$ such that $\mathrm{supp}\,(\sigma)\cap M_{(t_{1}^{\prime},t_{2}^{\prime})}=\emptyset$. Thus $\mathrm{supp}\,(\sigma)$ is supported within $M_{<t_{1}^{\prime}}$ and $M_{>t_{2}^{\prime}}$, so $P\sigma$ must be supported there, too. ∎ ###### Definition 4.3.19. Let $R<b<a$ and let $P_{1}$, $P_{2}$ be operators on $M_{<a}$ and $M_{>-a}$, respectively. Assume that $P_{1}$ and $P_{2}$ restrict to $M_{<-b}$ and $M_{>b}$, respectively, and that they are $\mathcal{A}$-derivations. By the previous Lemma, the two operators restrict to $M_{(-R,R)}$ and we assume that the restrictions agree. This allows us to patch the two operators together: Let $\chi\colon\mathord{\mathbb{R}}\rightarrow[0,1]$ is a smooth function that is identically $1$ near $\mathord{\mathbb{R}}_{\leq-R}$ and identically $0$ near $\mathord{\mathbb{R}}_{\geq R}$. We extend he function via $0$ and $1$ to $M$ and denote the result again with $\chi$. We define $P_{1}\cup P_{2}\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=P_{1}(1-\chi)+P_{2}\chi.$ Since $P_{1}$ and $P_{2}$ agree on $M_{(-R,R)}$ and act as $\mathcal{A}$-derivations, the result does not depend on the choice of $\chi$. An application of Lemma 4.3.14 shows that $\cup$ preserves symmetry. ###### Corollary 4.3.20. The operator $P_{1}\cup P_{2}$ is symmetric if and only if the operators $P_{1}$ and $P_{2}$ are symmetric. ###### Proof. The only if direction is easy because restrictions of symmetric operators are symmetric and $P_{1}$ agrees with $P_{2}$ on $M_{(-R,R)}$. For the other direction, we will prove that $M_{<R}$ and $M_{>-R}$ together with $\chi$ and $1-\chi$ from Definition 4.3.19 satisfy the condition of Lemma 4.3.14 for $P_{1}$ and $P_{2}$. To this end, fix a function $\eta\colon(-a,a)\rightarrow[0,1]$ with compact support and $\eta|_{(-b,b)}\equiv 1$. Let $\eta^{-1}$, $\eta^{0}$, and $\eta^{1}$ be the functions obtained from the construction in Definition 4.2.5 but considered as function on $M$. They form a partition of unity. For all compactly supported sections $\sigma\in\Gamma_{c}(M_{<R},E)$, the section $\eta^{-1}\sigma$ is supported within $M_{<-b}$, so that $P_{1}\eta^{-1}\sigma$ is also supported there by assumption. This implies $\displaystyle P_{1}\chi\sigma$ $\displaystyle=P_{1}\chi\eta^{-1}\sigma+P_{1}\chi\eta^{0}\sigma=P_{1}\eta^{-1}+P_{1}\chi\eta^{0}\sigma$ $\displaystyle=\chi P_{1}\eta^{-1}\sigma+\chi P_{1}\eta^{0}\sigma+[P_{1},\chi]\eta^{0}\sigma$ $\displaystyle=\chi P_{1}(\eta^{-1}+\eta^{0})\sigma+i^{-1}\mathrm{symb}_{1}(P_{1})(\text{--},\chi^{\prime})\cdot\eta^{0}\sigma$ $\displaystyle=\chi P_{1}\sigma+i^{-1}\mathrm{symb}_{1}(P_{1})(\text{--},\chi^{\prime})\sigma.$ Note that the last equation follows from $\chi^{\prime}\equiv 0$ on $\mathrm{supp}\,\eta^{-1}$. Of course, this implies that $P_{1}$ acts on $1-\chi$ as a derivation. The corresponding statement for $P_{2}$ and $(1-\chi)$ can be derived in an analogous manner. ∎ ###### Lemma 4.3.21. Let $P_{1}$ and $P_{2}$ be operators on $M_{<a}$ and $M_{>-a}$ as in setup 4.3.19. Assume that $||\mathrm{symb}_{1}(P_{1})(\text{--},\mathrm{d}t)||_{\infty,M_{(-R,R)}}\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=\\{||\mathrm{symb}_{1}(P_{1})(x,\mathrm{d}t)||_{\mathrm{op}}\,:\,x\in M_{(-R,R)}\\}\leq C.$ If there is a section $u\in\Gamma_{c}(M,E)$ such that $||(P_{1}\cup P_{2})u||_{0}<\varepsilon||u||_{0}$, then there is a $v\in\Gamma_{c}(M,E)$ that is supported within $M_{<R}$ that satisfies $||P_{1}v||_{0}^{2}<4\left(\varepsilon^{2}+\frac{C^{2}}{R^{2}}\right)||v||_{0}^{2}$ or is supported within $M_{>-R}$ and satisfies an analogous inequality with $P_{1}$ replaced by $P_{2}$. ###### Proof. Abbreviate $P_{1}\cup P_{2}$ to $Q$. Let $u\in\Gamma_{c}(M,E)$ with $||Qu||_{0}^{2}<\varepsilon^{2}||u||_{0}^{2}$. Assume without loss of generality, that $||u||_{0,M_{\leq 0}}^{2}\geq 1/2||u||_{0}^{2}$. Let $\eta\colon\mathord{\mathbb{R}}\rightarrow[0,1]$ be a smooth function that satisfies $\eta\equiv 1$ on $\mathord{\mathbb{R}}_{\leq 0}$, $\eta\equiv 0$ on $\mathord{\mathbb{R}}_{\geq 1}$, and $|\eta^{\prime}|\leq 1.2$. Set $\eta_{R}(t)=\eta(R^{-1}t)$ and extend this function as described above to $M$. We denote the result again with $\eta_{R}$. The norm splits into $\displaystyle||P_{1}\eta_{R}u||_{0}^{2}=||P_{1}\eta_{R}u||_{0,M_{\leq-R}}^{2}+||P_{1}u||_{0,M_{(-R,R)}}^{2}$ and, by Lemma 4.3.18, the summands only depend on the germs of the restrictions to these domains. For example, $||P_{1}\eta_{R}u||_{0,M_{\leq-R}}^{2}=||P_{1}\eta_{R}u|_{M_{\leq-R}}||_{0,M_{\leq-R}}^{2}=||P_{1}u||_{0,M_{\leq-R}}^{2}.$ Since also $Q$ satisfies the assumptions of Lemma 4.3.18, we have $(P_{1}\eta_{R}u)|_{M\leq-R}=P_{1}(\eta_{R}u|_{M\leq-R})=(P_{1}u)|_{M\leq-R}=(Qu)|_{M\leq-R}$ and because $P_{1}$ and $P_{2}$ agree on $M_{(-R,R)}$ we have $\displaystyle(Q\eta_{R}u)|_{M_{(-R,R)}}$ $\displaystyle=Q\eta_{R}(u|_{M_{(-R,R)}})=P_{1}\eta_{R}(u|_{M_{(-R,R)}})$ $\displaystyle=\eta_{R}P_{1}(u|_{M_{(-R,R)}})+[P,\eta_{R}\cdot](u|_{M_{(-R,R)}})$ $\displaystyle=(\eta_{R}Qu)|_{M_{(-R,R)}}+i^{-1}\mathrm{symb}_{1}(P_{1})(\text{--},\eta_{R}^{\prime}\mathrm{d}t)(u)|_{M_{(-R,R)}}.$ This implies $\displaystyle||P_{1}\eta_{R}u||_{0}^{2}$ $\displaystyle=||Qu||^{2}_{0,M_{\leq-R}}+||\eta_{R}Qu+\mathrm{symb}_{1}(P_{1})(\text{--},\eta_{R}^{\prime}\mathrm{d}t)(u)||_{0,M_{(-R,R)}}^{2}$ $\displaystyle\leq||Qu||^{2}_{0,M_{\leq-R}}+2||\eta_{R}Qu||^{2}_{0,M_{(-R,R)}}+2||\mathrm{symb}_{1}(P_{1})(\text{--},\eta_{R}^{\prime}\mathrm{d}t)(u)||_{0,M_{(-R,R)}}^{2}$ $\displaystyle\leq 2||Qu||^{2}_{0,M_{<R}}+2||\mathrm{symb}_{1}(P_{1})(\text{--},\eta_{R}^{\prime}\mathrm{d}t)(u)||_{0,M_{(-R,R)}}^{2}$ $\displaystyle<2\varepsilon^{2}||u||^{2}_{0}+2\frac{C^{2}}{R^{2}}||u||_{0}^{2},$ where we used in the last inequality that $||Qu||_{0}^{2}<\varepsilon^{2}||u||^{2}_{0}$ and that $\displaystyle\quad\ ||\mathrm{symb}_{1}(P_{1})(\text{--},\eta_{R}^{\prime}\mathrm{d}t)(u)||_{0;M_{(-R,R)}}$ $\displaystyle\leq||\mathrm{symb}_{1}(P_{1})(\text{--},\eta_{R}^{\prime}\mathrm{d}t)||_{0,0;M_{(-R,R)}}||u||_{0,M_{(-R,R)}}$ $\displaystyle\leq||\mathrm{symb}_{1}(P_{1})(\text{--},\eta_{R}^{\prime}\mathrm{d}t)||_{\infty;M_{(-R,R)}}||u||_{0}$ $\displaystyle\leq\frac{|\eta^{\prime}|_{\infty}}{R}||\mathrm{symb}_{1}(P_{1})(\text{--},\mathrm{d}t)||_{\infty;M_{(-R,R)}}||u||_{0}$ $\displaystyle\leq\frac{C}{R}||u||_{0}$ From $1/2||u||_{0}^{2}\leq||u||_{0,M_{\leq 0}}^{2}$ it follows that $||u||_{0}^{2}\leq 2||\eta_{R}u||_{0}^{2}$. Putting all inequalities together, we get $||P_{1}\eta_{R}u||_{0}^{2}<4\left(\varepsilon^{2}+C^{2}/R^{2}\right)||\eta_{R}u||_{0}^{2},$ so the claim follows for $v\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=\eta_{R}u$. ∎ The following implication is an indispensable tool in the following sections because it allows us to glue invertible block operators of Dirac type together. Note that, in this set up, we are allowed to choose the length of the “gluing-strip” to be arbitrarily large without increasing the propagation speed of the operator. ###### Corollary 4.3.22. Let $P_{1}$ and $P_{2}$ be as in the previous lemma. If they are bounded from below with lower bounds $c_{1}$ and $c_{2}$ respectively and if $R>\max\\{C^{2}/c_{1}^{2},C^{2}/c_{2}^{2}\\}$, then $P_{1}\cup P_{2}$ is bounded from below with lower bound $c_{P_{1}\cup P_{2}}^{2}=\frac{\min\left\\{{c_{1}^{2}-C^{2}/R^{2}},c_{2}^{2}-C^{2}/R^{2}\right\\}}{4}.$ ###### Proof. By assumption on $R$, the constant $c_{P_{1}\cup P_{2}}$ is positive. Assume that there is a section $u\in\Gamma_{c}(M,E)$ such that $||(P_{1}\cup P_{2})(u)||_{0}<c_{P_{1}\cup P_{2}}||u||_{0}$, then the previous Lemma implies the existence of $v\in\Gamma_{c}(M,E)$ that is, without loss of generality, supported within $M_{<R}$ and satisfies $\displaystyle||P_{1}v||_{0}^{2}$ $\displaystyle<4\left(c_{P_{1}\cup P_{2}}^{2}+C^{2}/R^{2}\right)||v||_{0}^{2}<c_{1}^{2}||v||_{0}^{2},$ which contradicts that $P_{1}$ is bounded from below by $c_{1}$. ∎ ### 4.4 Foundations of the Operator Concordance Set With the theory developed in the previous section we are finally in the position to define the operator concordance set $\widetilde{\Psi\mathrm{Dir}}^{\times}_{\bullet}(M)$ as the cubical subset of all block Dirac operators. But, first we give a combinatorial model for the classifying space for real $K$-theory, by considering the cubical set of block maps into $\Psi\mathrm{Dir}(M)$ and $\Psi\mathrm{Dir}^{\times}(M)$ instead of the spaces themselves. ###### Definition 4.4.1. Let $\Psi\mathrm{Dir}_{\bullet}(M)$ be the cubical set whose $n$-cubes consists of smooth block maps $P\colon\mathord{\mathbb{R}}^{n}\rightarrow\Psi\mathrm{Dir}(M)$. The connecting maps are given by $\partial^{\varepsilon}_{i}P=\lim_{R\to\infty}P\circ{\delta^{R\varepsilon}_{i}}\qquad\text{ and }\qquad\sigma_{i}P=P\circ p_{i}.$ Let $\Psi\mathrm{Dir}^{\times}_{\bullet}(M)$ be the cubical subset consisting of all smooth block maps with values in $\Psi\mathrm{Dir}^{\times}(M)$. Recall that $\Psi\mathrm{Dir}(M)$ is an affine bundle over $\mathrm{Riem}(M)$ with fibre $\Psi\mathrm{Dir}(M)_{g}$, the set of all self-adjoint, odd, Clifford-linear pseudo differential operators of order $1$ whose principal symbol agree with the principal symbol of the Dirac operator ${\not{\mathfrak{D}}}_{g}$. The map $\Psi\mathrm{Dir}(M)\rightarrow\mathrm{Riem}(M)$ assigns to an operator $P$ its underlying metric $g$. This map induces a map of cubical sets $\Psi\mathrm{Dir}_{\bullet}(M)\rightarrow\mathcal{R}^{+}_{\bullet}(M)$. Now, we are going to construct the combinatorial reference space. ###### Definition 4.4.2. Let $g$ be a block metric on $M\times\mathord{\mathbb{R}}^{n}$. A block operator $P\in\overline{\Psi\mathrm{DO}}^{1}(\mathfrak{S}_{g})$ of Dirac type is called a _block Dirac operator_ if it is symmetric and $P-{\not{\mathfrak{D}}}_{g}$ extends to a bounded endomorphism on $H^{0}(\mathfrak{S}_{g})=L^{2}(\mathfrak{S}_{g})$. An immediate consequence of Corollary 4.3.10 is that a block Dirac operator is automatically essentially self-adjoint. For the definition of the operator concordance set it will be important to trace permutation of coordinates. A permutation $\sigma\in\mathrm{Aut}(\\{1,\dots,n\\})=S_{n}$ induces a linear map $\sigma\colon(x_{1},\dots,x_{n})\mapsto(x_{\sigma^{-1}(1)},\dots,x_{\sigma^{-1}(n)}).$ If $g$ is a block metric on $M\times\mathord{\mathbb{R}}^{n}$, then $\sigma^{\ast}g$ is also a block metric on $M\times\mathord{\mathbb{R}}^{n}$. Since $\mathord{\mathbb{R}}^{n}$ carries up to equivalence only one $\mathrm{Pin}^{-}$-structure, permutations are Pin structure preserving. Functoriality of the spinor bundle construction yields a map, which we denote with $\mathfrak{S}(\sigma)\colon\mathfrak{S}_{\sigma^{\ast}g}\rightarrow\mathfrak{S}_{g}\quad\text{ or equivalently }\quad\mathfrak{S}(\sigma)\colon\mathfrak{S}_{g}\rightarrow\mathfrak{S}_{\sigma_{\ast}g}.$ We are particularly interested in cyclic permutations. ###### Definition 4.4.3. For positive integers $a<b$ let $\mathrm{cycl}(a,b)$ be the cyclic permutation given by $a\mapsto a+1$ and $b\mapsto a$ and let $\mathrm{cycl}(b,a)=\mathrm{cycl}(a,b)^{-1}$. We denote the induced maps on spinor bundles with $\mathrm{Cycl}(a,b)$ and $\mathrm{Cycl}(b,a)$, respectively. ###### Definition 4.4.4. Let $\widetilde{\Psi\mathrm{Dir}}_{\bullet}(M)$ be the cubical set whose set of $n$-cubes is given by $\widetilde{\Psi\mathrm{Dir}}_{n}(M)\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=\\{P\in\overline{\Psi\mathrm{DO}}^{1}(\mathfrak{S}_{g})\,:\,g\in\widetilde{\mathcal{R}_{n}}(M),\,P\text{ block Dirac operator}\\}$ and whose connecting maps are given by $\displaystyle\partial^{\varepsilon}_{i}P$ $\displaystyle\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=\lim_{R\to\infty}{{\delta^{R\varepsilon}_{i}}}^{\ast}P(i,\varepsilon)\in\overline{\Psi\mathrm{DO}}^{1}(\mathfrak{S}_{\partial^{\varepsilon}_{i}g}),$ $\displaystyle\sigma_{i}P$ $\displaystyle\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=\mathrm{Cycl}(i,n+1)_{\ast}(P\boxtimes\mathrm{id}+\mathrm{e/o}\boxtimes{\not{\mathit{D}}})\in\overline{\Psi\mathrm{DO}}^{1}(\mathfrak{S}_{\sigma_{i}g})$ ###### Remark 4.4.5. A comment to the (subtle) notation. Recall that if a block operator $P$ with underlying block metric $g$ decomposes on $M\times U_{r}(i,\varepsilon)=\\{(m,x)\,:\,\varepsilon x_{i}>r\\}$ into $P(i,\varepsilon)\boxtimes\mathrm{id}+\mathrm{e/o}\boxtimes{\not{\mathit{D}}}$ under the isometry $U_{r}(i,\varepsilon)\cong M\times\\{\varepsilon x_{i}=r\\}\times\varepsilon\mathord{\mathbb{R}}_{>r}$ given by permutation, then the operator $P(i,\varepsilon)$ restricts to a block operator on $\mathfrak{S}_{g(i,\varepsilon)}\rightarrow M\times\\{\varepsilon x_{i}=R\\}$ for all sufficiently large $R$ and is independent of $R$. The map ${\delta^{R\varepsilon}_{i}}\colon(M\times\mathord{\mathbb{R}}^{n-1},\partial^{\varepsilon}_{i}g)\rightarrow\bigl{(}M\times\\{\varepsilon x_{i}=R\\},g(i,\varepsilon)\bigr{)}$ is a Pin-structure preserving isometry and thus provides a bundle isometry $\mathfrak{S}({\delta^{R\varepsilon}_{i}})$ between the corresponding spinor bundles that induces an isomorphism between the corresponding sections with compact support. We use this isomorphism to pull $P(i,\varepsilon)$, an operator on $\Gamma_{c}(\mathfrak{S}_{g(i,\varepsilon)})$, back to ${{\delta^{R\varepsilon}_{i}}}^{\ast}P(i,\varepsilon)$, an operator on $\Gamma_{c}(\mathfrak{S}_{\partial^{\varepsilon}_{i}g})$. For the degeneracy map, we use that the cyclic permutation $\mathrm{cycl}(i,n+1)\colon\left(M\times\\{\varepsilon x_{i}>r\\},\sigma_{i}g\right)\rightarrow\left(M\times\mathord{\mathbb{R}}^{n}\times\varepsilon\mathord{\mathbb{R}}_{>r},g\oplus\mathrm{d}x_{n+1}^{2}\right)$ is an isometry and the natural isomorphism $\mathfrak{S}_{g}\boxtimes Cl_{1,0}\cong\mathfrak{S}_{g\oplus\mathrm{d}x_{n+1}^{2}}$. ###### Lemma 4.4.6. The connecting maps of $\widetilde{\Psi\mathrm{Dir}}_{\bullet}(M)$ are well- defined and satisfy the cubical identities. The proof requires the following simple functional analytic result. ###### Lemma 4.4.7. Let $H_{1}$, $H_{2}$ be Hilbert spaces and $A\colon H_{1}\rightarrow H_{1}$ be an unbounded operator. Then $A\otimes\mathrm{id}$ is a bounded operator on $H_{1}\otimes H_{2}$ if and only if $A$ is bounded on $H_{1}$. In this case, the operator norms satisfy $||A\otimes\mathrm{id}||_{\mathrm{op}}=||A||_{\mathrm{op}}$. ###### Proof. Since the algebraic tensor product $H_{1}\odot H_{2}$ lies dense in $H_{1}\otimes H_{2}$, so does ${\mathrm{dom}\,}(A\otimes\mathrm{id})\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt={\mathrm{dom}\,}A\odot H_{2}$. If $A$ is bounded, then we can continuously extend it to an operator on $H_{1}\otimes H_{2}$ and the operator norm of this extension necessarily agrees with $||A||_{\mathrm{op}}$. Conversely, assume that $A\otimes\mathrm{id}$ is bounded. Let $v\in H_{2}$ be a unit-length vector. The map $\iota_{v}\colon H_{1}\rightarrow H_{1}\otimes H_{2}$ given by $x\mapsto x\otimes v$ is linear and bounded. The map $x\otimes y\mapsto\langle y,v\rangle\cdot x$ extends from $H_{1}\odot H_{2}$ to a bounded linear map $p_{v}\colon H_{1}\otimes H_{2}\rightarrow H_{1}$. From $A=p_{v}\circ A\otimes\mathrm{id}\circ\iota_{v}$ we deduce that $A$ is a bounded operator. ∎ ###### Proof of Lemma 4.4.6. By Corollary 4.2.3, $\partial^{\varepsilon}_{i}P$ is a again a block operator. Symbol calculus shows that the principal symbol of $\partial^{\varepsilon}_{i}P$ agrees with the principal symbol of ${\not{\mathfrak{D}}}_{\partial^{\varepsilon}_{i}g}$, so $\partial^{\varepsilon}_{i}P$ is a block operator of Dirac type. In a similar fashion, but using Lemma 4.2.4 instead, we see that $\sigma_{i}P$ is a block operator of Dirac type. We need to show that $\partial^{\varepsilon}_{i}P$ and $\sigma_{i}P$ are symmetric operators if $P$ is symmetric. We start with $\sigma_{i}P=\mathrm{Cycl}(i,n+1)_{\ast}(P\boxtimes\mathrm{id}+\mathrm{e/o}\boxtimes{\not{\mathit{D}}})$. Since $\mathrm{Cycl}(i,n+1)$ induces an isometry between the corresponding Hilbert spaces of square integrable sections, the push-forward $\mathrm{Cycl}(i,n+1)_{\ast}$ maps symmetric operators to symmetric operators. Thus, it suffices to prove that $P\boxtimes\mathrm{id}+\mathrm{e/o}\boxtimes{\not{\mathit{D}}}$ is a symmetric operator on $L^{2}(M\times\mathord{\mathbb{R}}^{n+1};\mathfrak{S}_{g\oplus\mathrm{d}x_{n+1}^{2}})\cong L^{2}(M\times\mathord{\mathbb{R}}^{n};\mathfrak{S}_{g})\otimes L^{2}(\mathord{\mathbb{R}},Cl_{1,0}),$ which is true because the tensor product of two symmetric operators is again symmetric and all four operators in question $P$, $\mathrm{id}$, $\mathrm{e/o}$ and ${\not{\mathit{D}}}$ are symmetric. To argue that $\partial^{\varepsilon}_{i}P$ is symmetric, we proceed as follows: On $U(i,\varepsilon)\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=U_{R}(i,\varepsilon)$, for some sufficiently large $R>0$, we can identify the symmetric operator $P|_{U(i,\varepsilon)}$ with $P(i,\varepsilon)\boxtimes\mathrm{id}+\mathrm{e/o}\boxtimes{\not{\mathit{D}}}_{\mathord{\mathbb{R}}}$ via an isometric bundle morphism. The second summand is symmetric, so $P(i,\varepsilon)\boxtimes\mathrm{id}$ must be a symmetric operator on $\mathfrak{S}_{g(i,\varepsilon)\oplus\mathrm{d}x_{n}^{2}}\rightarrow(M\times\mathord{\mathbb{R}}^{n})(i,\varepsilon)\times\varepsilon\mathord{\mathbb{R}}_{>R}$. This implies that $P(i,\varepsilon)$ is symmetric on $\mathfrak{S}_{g(i,\varepsilon)}$ and consequently that $\partial^{\varepsilon}_{i}P={\delta^{R\varepsilon}_{i}}^{\ast}P(i,\varepsilon)$ is symmetric because ${\delta^{R\varepsilon}_{i}}\colon(M\times\mathord{\mathbb{R}}^{n-1},\partial^{\varepsilon}_{i}g)\rightarrow(M\times\mathord{\mathbb{R}}^{n}(i,\varepsilon),g(i,\varepsilon))$ is an isometry. From $\displaystyle\quad\ \sigma_{i}P-{\not{\mathfrak{D}}}_{\sigma_{i}g}$ $\displaystyle=\mathrm{Cycl}(i,n+1)_{\ast}\left(P\boxtimes\mathrm{id}+\mathrm{e/o}\boxtimes{\not{\mathit{D}}}\right)-\mathrm{Cycl}(i,n+1)_{\ast}\left({\not{\mathfrak{D}}}_{g}\boxtimes\mathrm{id}+\mathrm{e/o}\boxtimes{\not{\mathit{D}}}\right)$ $\displaystyle=\mathrm{Cycl}(i,n+1)_{\ast}((P-{\not{\mathfrak{D}}}_{g})\boxtimes\mathrm{id})$ and Lemma 4.4.7 it follows that the difference $\sigma_{i}P-{\not{\mathfrak{D}}}_{\sigma_{i}g}$ extends to a bounded operator on $H^{0}(\mathfrak{S}_{\sigma_{i}g})$ with operator norm $||P-{\not{\mathfrak{D}}}_{g}||_{0,0}$. To see that $\partial^{\varepsilon}_{i}P-{\not{\mathfrak{D}}}_{\partial^{\varepsilon}_{i}g}$ extends to a bounded operator on $H^{0}(\mathfrak{S}_{\partial^{\varepsilon}_{i}g})$ we argue as follows. For $R>0$ sufficiently large the operator $P-{\not{\mathfrak{D}}}_{g}$ restricts to $\Gamma_{c}(U_{R}(i,\varepsilon),\mathfrak{S}_{g})$. Since $P-{\not{\mathfrak{D}}}_{g}$ extends to a bounded operator on $H^{0}(\mathfrak{S}_{g})$, the restriction extends to a bounded operator on $H^{0}(U_{R}(i,\varepsilon),\mathfrak{S}_{g})$, which is isometric equivalent to $H^{0}(M\times\\{\varepsilon x_{i}=R\\},\mathfrak{S}_{g(i,\varepsilon)})\otimes L^{2}(\varepsilon\mathord{\mathbb{R}}_{>R},Cl_{1,0})$. Under this identification, the restriction of $P-{\not{\mathfrak{D}}}_{g}$ agrees with $\left(P(i,\varepsilon)-{\not{\mathfrak{D}}}_{g(i,\varepsilon)}\right)\boxtimes\mathrm{id}$. By Lemma 4.4.7, this implies that $P(i,\varepsilon)-{\not{\mathfrak{D}}}_{g(i,\varepsilon)}$ is bounded. Thus, $\partial^{\varepsilon}_{i}P-{\not{\mathfrak{D}}}_{\partial^{\varepsilon}_{i}g}={\delta^{R\varepsilon\,\ast}_{i}}\left(P(i,\varepsilon)-{\not{\mathfrak{D}}}_{g(i,\varepsilon)}\right)$ is a bounded operator on $H^{0}(\mathfrak{S}_{\partial^{\varepsilon}_{i}g})$. Lastly, we need to show, that the connecting maps satisfy the cubical identities. For $i<j$ and $R>0$ sufficiently large, we have $\displaystyle\partial^{\varepsilon}_{i}\partial^{\omega}_{j}P$ $\displaystyle=\partial^{\varepsilon}_{i}\left({\delta^{\omega R}_{j}}^{\ast}P(j,\omega)\right)$ $\displaystyle={\delta^{\varepsilon R}_{i}}^{\ast}\left({\delta^{\omega R}_{j}}^{\ast}\left(P(j,\omega)\right)(i,\varepsilon)\right)$ $\displaystyle=\left({\delta^{\omega R}_{j}}{\delta^{\varepsilon R}_{i}}\right)^{\ast}P(i,j,\varepsilon,\omega)$ $\displaystyle=\left({\delta^{\varepsilon R}_{i}}{\delta^{\omega R}_{j-1}}\right)^{\ast}P(i,j,\varepsilon,\omega)$ $\displaystyle={\delta^{\omega R}_{j-1}}^{\ast}\left({\delta^{\varepsilon R}_{i}}^{\ast}\left(P(i,\varepsilon)\right)(j-1,\omega)\right)$ $\displaystyle=\partial^{\omega}_{j-1}\partial^{\varepsilon}_{i}P.$ For $i<j$ and a sufficiently large $R>0$, we calculate $\displaystyle\partial^{\varepsilon}_{i}\sigma_{j}(P)$ $\displaystyle={\delta^{R\varepsilon}_{i}}^{\ast}(\sigma_{j}(P)(i,\varepsilon))$ $\displaystyle={\delta^{R\varepsilon}_{i}}^{\ast}\mathrm{Cycl}(j,n+1)_{\ast}\left(P\boxtimes\mathrm{id}+\mathrm{e/o}\boxtimes{\not{\mathit{D}}}\right)(\mathrm{cycl}(j,n+1)^{-1}(i),\varepsilon)$ $\displaystyle={\delta^{R\varepsilon}_{i}}^{\ast}\mathrm{Cycl}(j,n+1)_{\ast}(P(i,\varepsilon)\boxtimes\mathrm{id}+\mathrm{e/o}\boxtimes{\not{\mathit{D}}})\quad\quad\quad\text{ for }i<j$ $\displaystyle=\left({\delta^{R\varepsilon}_{i}}^{-1}\right)_{\ast}\mathrm{Cycl}(j,n+1)_{\ast}(P(i,\varepsilon)\boxtimes\mathrm{id}+\mathrm{e/o}\boxtimes{\not{\mathit{D}}})$ $\displaystyle=\mathrm{Cycl}(j-1,n)_{\ast}\left({\delta^{R\varepsilon}_{i}}^{-1}\right)_{\ast}\left(P(i,\varepsilon)\boxtimes\mathrm{id}+\mathrm{e/o}_{\mathfrak{S}_{g(i,\varepsilon)}}\boxtimes{\not{\mathit{D}}}\right)$ $\displaystyle=\mathrm{Cycl}(j-1,n)_{\ast}(\partial^{\varepsilon}_{i}P\boxtimes\mathrm{id}+\mathrm{e/o}_{\mathfrak{S}_{\partial^{\varepsilon}_{i}g}}\boxtimes{\not{\mathit{D}}})$ $\displaystyle=\sigma_{n-1}\partial^{\varepsilon}_{i}P,$ and analogously $\partial^{\varepsilon}_{i}\sigma_{j}P=\sigma_{j}\partial^{\varepsilon}_{i-1}P$ for $i>j$. Finally, with a slight abuse of notation in the last lines, we calculate $\displaystyle\partial^{\varepsilon}_{i}\sigma_{i}P$ $\displaystyle={\delta^{R\varepsilon}_{i}}^{\ast}\left((\sigma_{i}P)(i,\varepsilon)\right)$ $\displaystyle={\delta^{R\varepsilon}_{i}}^{\ast}\left(\mathrm{Cycl}(i,n+1)_{\ast}(P\boxtimes\mathrm{id}+\mathrm{e/o}\boxtimes{\not{\mathit{D}}})(i,\varepsilon)\right)$ $\displaystyle={\delta^{R\varepsilon}_{i}}^{-1}_{\ast}\mathrm{Cycl}(i,n+1)_{\ast}\left((P\boxtimes\mathrm{id}+\mathrm{e/o}\boxtimes{\not{\mathit{D}}})(\mathrm{cycl}(i,n+1)^{-1}(i),\varepsilon)\right)$ $\displaystyle={\delta^{R\varepsilon}_{i}}^{-1}_{\ast}\mathrm{Cycl}(i,n+1)_{\ast}\left((P\boxtimes\mathrm{id}+\mathrm{e/o}\boxtimes{\not{\mathit{D}}})(n+1,\varepsilon)\right)$ $\displaystyle={\delta^{R\varepsilon}_{n+1}}^{-1}_{\ast}\left((P\boxtimes\mathrm{id}+\mathrm{e/o}\boxtimes{\not{\mathit{D}}})(n+1,\varepsilon)\right)=P.$ For $i\leq j$, we use the identity $\mathrm{cycl}(i,n+2)\circ\mathrm{cycl}(j,n+1)=\mathrm{cycl}(j+1,n+2)\circ\mathrm{cycl}(i,n+1)\circ\tau_{n+1,n+2}$ in which $\tau_{n+1,n+2}$ denotes the transposition that interchanges $n+1$ and $n+2$, to conclude $\displaystyle\sigma_{i}\sigma_{j}(P)$ $\displaystyle=\sigma_{i}(\mathrm{Cycl}(j,n+1)_{\ast}(P\boxtimes\mathrm{id}+\mathrm{e/o}\boxtimes{\not{\mathit{D}}}))$ $\displaystyle\begin{split}&=\mathrm{Cycl}(i,n+2)_{\ast}\left(\mathrm{Cycl}(j,n+1)_{\ast}(P\boxtimes\mathrm{id}+\mathrm{e/o}_{\mathfrak{S}_{g}}\boxtimes{\not{\mathit{D}}})\boxtimes\mathrm{id}\right.\\\ &\quad\left.+\mathrm{e/o}_{\mathfrak{S}_{\sigma_{j}(g)}}\boxtimes{\not{\mathit{D}}}\right)\end{split}$ $\displaystyle\begin{split}&=\mathrm{Cycl}(i,n+2)_{\ast}\circ\mathrm{Cycl}(j,n+1)_{\ast}\biggl{(}P\boxtimes\mathrm{id}\boxtimes\mathrm{id}\\\ &\quad+\left.\mathrm{e/o}_{\mathfrak{S}_{g}}\boxtimes{\not{\mathit{D}}}\boxtimes\mathrm{id}+\mathrm{e/o}_{\mathfrak{S}_{\sigma_{n+1}(g)}}\boxtimes{\not{\mathit{D}}}\right)\end{split}$ $\displaystyle\begin{split}&=\mathrm{Cycl}(i,n+2)_{\ast}\circ\mathrm{Cycl}(j,n+1)_{\ast}\biggl{(}P\boxtimes\mathrm{id}\boxtimes\mathrm{id}\\\ &\quad\left.+\mathrm{e/o}_{\mathfrak{S}_{g}}\boxtimes{\not{\mathit{D}}}\boxtimes\mathrm{id}+\mathrm{e/o}_{\mathfrak{S}_{g}}\boxtimes\mathrm{e/o}\boxtimes{\not{\mathit{D}}}\right)\end{split}$ $\displaystyle\begin{split}&=\mathrm{Cycl}(j+1,n+2)_{\ast}\circ\mathrm{Cycl}(i,n+1)_{\ast}\circ{\tau_{n+1,n+2}}_{\ast}\biggl{(}P\boxtimes\mathrm{id}\boxtimes\mathrm{id}\\\ &\quad\left.+\mathrm{e/o}_{\mathfrak{S}_{g}}\boxtimes{\not{\mathit{D}}}\boxtimes\mathrm{id}+\mathrm{e/o}_{\mathfrak{S}_{g}}\boxtimes\mathrm{e/o}\boxtimes{\not{\mathit{D}}}\right)\end{split}$ $\displaystyle\begin{split}&=\mathrm{Cycl}(j+1,n+2)_{\ast}\circ\mathrm{Cycl}(i,n+1)_{\ast}\biggl{(}P\boxtimes\mathrm{id}\boxtimes\mathrm{id}\\\ &\quad\left.+\mathrm{e/o}_{\mathfrak{S}_{g}}\boxtimes{\not{\mathit{D}}}\boxtimes\mathrm{id}+\mathrm{e/o}_{\mathfrak{S}_{g}}\boxtimes\mathrm{e/o}\boxtimes{\not{\mathit{D}}}\right)\end{split}$ $\displaystyle=...=\sigma_{j+1}\sigma_{i}(P).$ In the firth line, we can drop ${\tau_{n+1,n+2}}_{\ast}$ in the second factor to go to the sixth line because ${\not{\mathit{D}}}\boxtimes\mathrm{id}+\mathrm{e/o}\boxtimes{\not{\mathit{D}}}=\partial_{x_{n+1}}\cdot\frac{\partial}{\partial{x_{n+1}}}+\partial_{x_{n+2}}\cdot\frac{\partial}{\partial{x_{n+2}}}$ is symmetric in the $(n+1)$-th and $(n+2)$-th coordinate. ∎ We come now to one of the most essential definitions of this thesis. ###### Definition 4.4.8. Let $\widetilde{\Psi\mathrm{Dir}}^{\times}_{\bullet}(M)$ be the sequence of subsets of $\widetilde{\Psi\mathrm{Dir}}_{\bullet}(M)$, where $\widetilde{\Psi\mathrm{Dir}}^{\times}_{n}(M)$ consists of all block Dirac operator $P\in\widetilde{\Psi\mathrm{Dir}}_{n}(M)$ that extend to an invertible operator $P\colon H^{1}(\mathfrak{S}_{g_{P}})\rightarrow H^{0}(\mathfrak{S}_{g_{P}})$. The proof that the face maps of $\widetilde{\Psi\mathrm{Dir}}_{\bullet}(M)$ restricts to $\widetilde{\Psi\mathrm{Dir}}^{\times}_{\bullet}(M)$ requires the following lemma that roughly says that positivity descends to hypersurfaces. ###### Lemma 4.4.9. Let $\mathfrak{S}_{g}\rightarrow M$ be a spinor bundle over a Riemannian manifold $(M,g)$. For $Q\in\overline{\Psi\mathrm{DO}}^{1}(E)$ symmetric and odd, define $P\mathrel{\vbox{\offinterlineskip\hbox{.}\vskip-3.44444pt\hbox{.}}}\joinrel\hskip 1.0pt=Q\boxtimes\mathrm{id}+\mathrm{e/o}\boxtimes{\not{\mathit{D}}}$ on $\mathfrak{S}\boxtimes Cl_{1,0}\rightarrow M\times(-a,a)$ for all $0<a\leq\infty$. If $||Pu||_{0}\geq c||u||_{0}$ for all $u\in\Gamma_{c}(M\times(-a,a),\mathfrak{S}_{g}\boxtimes Cl_{1,0})$, then $||Qv||_{0}^{2}\geq\max\\{0,c-4/a^{2}\\}||v||_{0}^{2}$ for all $v\in\Gamma_{c}(M,\mathfrak{S}_{g})$. ###### Proof. Let $v\in\Gamma_{c}(M,\mathfrak{S}_{g})$ and $f\in\Gamma_{c}((-a,a),Cl_{1,0})$. The operator $P$ is again symmetric by the proof of Lemma 4.4.6 and odd. Using Fubini’s theorem and that $P$ is bounded from below, we get $\displaystyle||Pv\otimes f||_{0}^{2}$ $\displaystyle=\langle P^{2}(v\otimes f),v\otimes f\rangle$ $\displaystyle=\langle Q^{2}(v)\otimes f-v\otimes\Delta_{\mathord{\mathbb{R}}}(f),v\otimes f\rangle$ $\displaystyle=||Q(v)\otimes f||_{0}^{2}+||v\otimes f^{\prime}||_{0}^{2}$ $\displaystyle=||Q(v)||_{0}^{2}||f||_{0}^{2}+||v||_{0}^{2}||f^{\prime}||_{0}^{2}$ $\displaystyle\geq c||v||_{0}^{2}||f||_{0}^{2}=c||v\otimes f||_{0}^{2}.$ The function $\varphi\colon(-1,1)\rightarrow\mathord{\mathbb{R}}\quad\text{ given by }\quad\varphi(t)=\begin{cases}1+t,&\text{if }t\leq 0,\\\ 1-t,&\text{if }t\geq 0,\end{cases}$ satisfies $||\varphi||_{0}^{2}=2/3$ while $||\varphi^{\prime}||_{0}^{2}=2$. Pick a compactly supported function $f$ that is sufficiently $||\text{--}||_{1}$-close to $\varphi$ such that $||f^{\prime}||_{0}^{2}/||f||_{0}^{2}\leq 4$. By the transformation formula, the assignment $f\mapsto a^{-1/2}f(a^{-1}\cdot)$ is an isometry $L^{2}(-1,1)\rightarrow L^{2}(-a,a)$. Thus, $||f_{a}^{\prime}||_{0}^{2}/||f_{a}||_{0}^{2}=a^{-2}||f^{\prime}||_{0}^{2}/||f||_{0}^{2}\leq 4/a^{2}$. In conclusion, if we plug into $f_{a}$ in the previous inequality, we obtain $\displaystyle||Qv||_{0}^{2}$ $\displaystyle\geq\left(c-||f_{a}^{\prime}||_{0}^{2}/||f_{a}||_{0}^{2}\right)||v||_{0}^{2}$ $\displaystyle\geq\left(c-4/a^{2}\right)||v||_{0}^{2},$ which is positive if $c>4/a^{2}$. ∎ ###### Lemma 4.4.10. The connecting maps of $\widetilde{\Psi\mathrm{Dir}}_{\bullet}(M)$ send invertible elements to invertible elements. Thus, $\widetilde{\Psi\mathrm{Dir}}^{\times}_{\bullet}(M)$ is a cubical subset. ###### Proof. Fix a sufficiently large $R>0$ such that $g=g_{P}$ decomposes into $g(i,\varepsilon)\oplus\mathrm{d}x_{i}^{2}$ on $U_{R}{(i,\varepsilon)}$ and, by abusing notation, $P$ decomposes into $P(i,\varepsilon)\boxtimes\mathrm{id}+\mathrm{e/o}\boxtimes{\not{\mathit{D}}}_{\mathord{\mathbb{R}}}$. Since $P$ is bounded from below, its restriction $P|_{U_{R}(i,\varepsilon)}$ is also bounded from below. Lemma 4.4.9 implies that $P(i,\varepsilon)$, and hence $\partial^{\varepsilon}_{i}P$ too, is bounded from below. Since $\partial^{\varepsilon}_{i}P$ is self-adjoint, it must be invertible. To see that $\sigma_{i}(P)$ is invertible, it suffices to check that $P\boxtimes\mathrm{id}+\mathrm{e/o}\boxtimes{\not{\mathit{D}}}$ is invertible. Since $P\boxtimes\mathrm{id}$ and $\mathrm{e/o}\boxtimes{\not{\mathit{D}}}$ anti-commute we have $\left(P\boxtimes\mathrm{id}+\mathrm{e/o}\boxtimes{\not{\mathit{D}}}\right)^{2}=P^{2}\boxtimes\mathrm{id}-\mathrm{id}\boxtimes\Delta.$ Both summands are non-negative operators. Since $P$ is invertible, $P^{2}$ is positive and therefore the sum is positive and therefore invertible. ∎ We would like to know whether $\widetilde{\Psi\mathrm{Dir}}_{\bullet}(M)$ and $\widetilde{\Psi\mathrm{Dir}}^{\times}_{\bullet}(M)$ are Kan sets. For $\widetilde{\Psi\mathrm{Dir}}_{\bullet}(M)$ even more is true. ###### Proposition 4.4.11. The cubical set $\widetilde{\Psi\mathrm{Dir}}_{\bullet}(M)$ is combinatorially contractible. The proposition follows immediately from the following elementary lemma. ###### Lemma 4.4.12. Let $P_{(j,\omega)}$ and $P_{(k,\eta)}$ be two block Dirac operators on $M\times\mathord{\mathbb{R}}^{n-1}$. Assume that $j\leq k$, that $\partial^{\omega}_{j}P_{(k,\eta)}=\partial^{\eta}_{k-1}P_{(j,\omega)}$ and that the two operators decompose outside of $M\times\rho I^{n}$. Then the operators $\sigma_{j}P_{(j,\omega)}$ and $\sigma_{k}P_{(k,\eta)}$ agree on the set $\\{\omega x_{j}>\rho,\eta x_{k}>\rho\\}$. ###### Proof. Abbreviate $\\{\omega x_{j}>\rho,\eta x_{k}>\rho\\}$ to $U$. If $j=k$ there is nothing to prove for $U$ is either empty or $(j,\omega)=(k,\eta)$. In the other cases, it follows from Lemma 3.1.10 that the underlying metrics agree on $U$. The proof of Lemma 4.2.4 shows that $\sigma_{j}P_{(j,\omega)}$ restricts to $U$. Furthermore, for $j<k$, the calculation $\displaystyle\sigma_{j}P_{(j,\omega)}|_{U}$
# Remarks on existence and uniqueness of the solution for stochastic partial differential equations Benny Avelin Uppsala University, Department of Mathematics, Sweden <EMAIL_ADDRESS>and Lauri Viitasaari Aalto University School of Business, Department of Information and Service Management, Finland <EMAIL_ADDRESS> ###### Abstract. In this article we consider existence and uniqueness of the solutions to a large class of stochastic partial differential of form $\partial_{t}u=L_{x}u+b(t,u)+\sigma(t,u)\dot{W}$, driven by a Gaussian noise $\dot{W}$, white in time and spatial correlations given by a generic covariance $\gamma$. We provide natural conditions under which classical Picard iteration procedure provides a unique solution. We illustrate the applicability of our general result by providing several interesting particular choices for the operator $L_{x}$ under which our existence and uniqueness result hold. In particular, we show that Dalang condition given in [5] is sufficient in the case of many parabolic and hypoelliptic operators $L_{x}$. Mathematics Subject Classifications (2010): 60H15, 60G15, 35C15, 35K58, 35S10. Keywords: Stochastic partial differential equations, existence and uniqueness, mild solution, semilinear parabolic equations, hypoelliptic equations. ## 1\. Introduction In this article we consider the stochastic partial differential equation (SPDE) of form $\frac{\partial u}{\partial t}(t,x)=(L_{x}u)(t,x)+b(t,u(t,x))+\sigma(t,u(t,x))\dot{W}(t,x),\hskip 14.22636ptt\geq 0,x\in\mathbb{R}^{d}$ (1.1) with initial condition $u(0,x)=u_{0}(x)$. Here $b$ and $\sigma$ are assumed to be Lipschitz continuous functions and bounded on compacts in the spatial variable $x$, and uniformly in $t$. The stochastic force $\dot{W}$ is assumed to be Gaussian with correlation structure that is white in the time variable $t$ and given by a generic covariance $\gamma$ in the spatial variable $x$. SPDEs have been a subject of active research in the literature for recent years, and the basic theory is already rather well-established. Especially, initiated by the seminal paper by Dalang [5], stochastic heat equations, in which case $L_{x}=\Delta$ is the Laplace operator, with Gaussian noise that is white in time have received a lot of attention. In this case one obtains the existence and uniqueness of the solution provided that the following so-called Dalang’s condition $\int_{\mathbb{R}^{d}}\frac{1}{\beta+2|\xi|^{2}}\hat{\gamma}(d\xi)<\infty$ (1.2) holds for some $\beta_{0}>0$ (in which case it holds for all $\beta>\beta_{0}$). Here $\hat{\gamma}$ denotes the non-negative measure arising as the Fourier transform of the spatial covariance $\gamma$ of the noise $\dot{W}$. One also observes that here the term $|\xi|^{2}$ corresponds to the Fourier multiplier arising from the Laplace operator $\Delta$. This leads to a natural extension to the case where $L_{x}$ is the $L^{2}$-generator of a Levy process. In this case 1.2 is replaced by $\int_{\mathbb{R}^{d}}\frac{1}{\beta+2\text{Re}\Psi(\xi)}\hat{\gamma}(d\xi)<\infty,$ (1.3) where $\Psi(\xi)$ arises from the characteristic exponent of the associated Levy process. Here $\text{Re}\Psi(\xi)$ is again a non-negative function, and if 1.3 holds for some $\beta>0$, then it holds for all $\beta>0$. As a particular interesting example, this covers the case of the stochastic fractional heat equation where $L_{x}$ is given by the fractional Laplace operator $-(-\Delta)^{\alpha}$, $\alpha\in(0,1]$. Indeed, then $\Psi(\xi)=|\xi|^{2\alpha}$ and one recovers the classical heat equation and condition 1.2 by plugging in $\alpha=1$. This generalisation is studied, among others, in [10] where the existence and uniqueness result was given under suitable assumptions on the coefficients $b,\sigma$. A standard technique to prove existence and uniqueness of the solution to 1.1 is based on the so-called fundamental solution $G_{t}(x)$ (or the Green kernel) associated to the equation $\partial_{t}u=L_{x}u$. Then one obtains a candidate for the solution through convolutions with the kernel $G_{t}(x)$. In particular, in the settings mentioned above one obtains that $G_{t}$ is the density of the underlying Levy process and forms a semigroup. After that, using the positivity of $G_{t}$ as well, one obtains the solution through Picard iteration. In this article we provide a general existence and uniqueness result for the solution to 1.1. Our condition, given by Equation 2.6 below, is similar to 1.3. However, in our results we do not require non-negativity of the fundamental solutions $G_{t}$ as we only consider upper bounds for $|G_{t}|$. Furthermore, we do not require the associated function, given by $\Psi(\xi)$ in 1.3, to be non-negative. This fact allows to consider a larger class of operators $L_{x}$ in 1.1, in the case where we do not have a Gaussian upper bound on the fundamental solution. Our main contribution is that, by a careful analysis on the essential requirements in the classical arguments, we get a remarkably strong generalization. As such, we are able to cover a host of equations which was previously not known in the literature, including the important Kolmogorov equations. Finally, we stress that similar considerations can be applied to equations of form 1.1, where $\partial_{t}$ is replaced with more general operator $L_{t}$, see Section 4. The rest of the paper is organised as follows. In Section 2 we present our assumptions and provide the existence and uniqueness result, Theorem 2.2. In Section 2.1, we illustrate the applicability of our result by providing a detailed discussion on several interesting examples. The proof of Theorem 2.2 is postponed to Section 3 where we also recall some basic facts on stochastic calculus. We end the paper with some concluding remarks. ## 2\. General existence and uniqueness result We consider the stochastic partial differential equation $\frac{\partial u}{\partial t}(t,x)=(L_{x}u)(t,x)+b(t,u(t,x))+\sigma(t,u(t,x))\dot{W}(t,x),\hskip 14.22636ptt\geq 0,x\in\mathbb{R}^{d}$ (2.1) where $L_{x}$ is a suitable differential operator (acting on the variable $x$), and coefficients $b$ and $\sigma$ are assumed to be spatially Lipschitz continuous functions, uniformly in $s$ over compacts. That is, for any $T>0$ we have for all $t\in[0,T]$ that $|b(t,x)-b(t,y)|\leq L_{b,T}|x-y|$ and $|\sigma(t,x)-\sigma(t,y)|\leq L_{\sigma,T}|x-y|.$ Note that then the functions $b$ and $\sigma$ are spatially locally bounded, uniformly in $s$ over compacts. That is, $\sup_{s\in[0,T],y\in K}\max(|\sigma(s,y)|,b(s,y)|)<\infty$ for all compact sets $K\subset\mathbb{R}^{d}$ and all finite $T>0$. For the (centered) Gaussian noise $\dot{W}$ we assume that the covariance is given by $\mathbb{E}\left[\dot{W}(t,x)\dot{W}(s,y)\right]=\delta_{0}(t-s)\gamma(x-y),$ where $\delta_{0}$ denotes the Dirac delta and $\gamma$ are non-negative and non-negative definite measures. That is, the noise can be described by a centered Gaussian family with covariance $\mathbb{E}[W(\phi)W(\psi)]=\int_{0}^{\infty}\int_{\mathbb{R}^{d}}\mathcal{F}\phi(\cdot,s)(\xi)\overline{\mathcal{F}(\psi)(\cdot,s)}(\xi)\hat{\gamma}(d\xi)ds,$ (2.2) where $\hat{\gamma}$ is the spectral measure of $\gamma$, $\mathcal{F}$ denotes the Fourier transform, and $\phi,\psi$ are suitable functions. We provide rigorous treatment on the construction of the Gaussian noise $\dot{W}$ in Section 3.1 that contains a brief introduction to Gaussian analysis and stochastic integration required for our analysis. We denote by $G_{t}(x)$ the fundamental solution (or the Green kernel) associated to the operator $M=\partial_{t}-L_{x}$, in the sense that for any "nice enough" intial data $u_{0}$ we have that $\displaystyle u(x,t)=\int G_{t}(x-y)u_{0}(y)dy$ satisfies $Mu=0$ and $u\to u_{0}$ as $t\to 0$. Consider now our original equation 2.1. For the initial condition $u(0,x)=u_{0}(x)$, we assume that $u_{0}(x)$ is deterministic and satisfies, for every $T>0$, $\sup_{t\in(0,T],x\in\mathbb{R}^{d}}\int_{\mathbb{R}^{d}}|G_{t}(x-y)||u_{0}(y)|dy<\infty.$ (2.3) We prove that under certain conditions, equation 2.1 admits a unique mild solution in the following sense. ###### Definition 2.1. We say that a random field $u(t,x)$, adapted to the filtration generated by $\dot{W}$, is a mild solution to 2.1 if for all $(t,x)\in\mathbb{R}_{+}\times\mathbb{R}^{d}$, the process $(s,y)\rightarrow G_{t-s}(x-y)\sigma(u(s,y))\textbf{1}_{[0,t]}(s)$ is integrable and we have $\displaystyle u(t,x)$ $\displaystyle=$ $\displaystyle\int_{\mathbb{R}^{d}}G_{t}(x-y)u_{0}(y)dy+\int_{0}^{t}\int_{\mathbb{R}^{d}}G_{t-s}(x-y)b(s,u(s,y))dyds.$ (2.4) $\displaystyle+$ $\displaystyle\int_{0}^{t}\int_{\mathbb{R}^{d}}G_{t-s}(x-y)\sigma(s,u(s,y))W(ds,dy).$ Here the stochastic integral exists in the sense of Dalang-Walsh, see Section 3.1. The following existence and uniqueness result is the main result of this paper. The proof follows standard arguments and is presented in Section 3. ###### Theorem 2.2. Let $G_{s}$ be the fundamental solution to 2.1 and assume that there exists an integrable function $g_{s}(x)=\mathcal{F}^{-1}(e^{-s\hat{f}})$, $\hat{f}\geq-C$ such that $|G_{s}(x)|\leq g_{s}(x)$ and there exists $\beta_{0}>2C$ such that, $\int_{0}^{\infty}e^{-\beta_{0}s}\|g_{s}\|_{L^{1}(\mathbb{R}^{d})}ds<\infty$ (2.5) and $\int_{\mathbb{R}^{d}}\frac{1}{\beta_{0}+2\hat{f}(\xi)}\hat{\gamma}(d\xi)<\infty$ (2.6) Then 2.1 admits a unique mild solution. ###### Remark 1. Observe that the condition 2.5 is independent of the chosen covariance $\gamma$, and is rather mild. Indeed, in many examples $g_{s}$ is given by a density of some generating stochastic process at time $s$, and hence we have $\|g_{s}\|_{L^{1}(\mathbb{R}^{d})}=1$ for all $s>0$. Moreover, by carefully examining our proof one observes that 2.5 can be replaced with a weaker condition $\int_{0}^{T}\|g_{s}\|_{L^{1}(\mathbb{R}^{d})}ds<\infty$ for all $T>0$ finite. Finally, we note that 2.5 can be omitted in the case $b\equiv 0$. ###### Remark 2. The proof of Theorem 2.2 actually gives more, however for purposes of exposition we chose to present the simplified form 2.6. For a discussion about further cases (including for instance the wave equation) see, Section 4. ###### Remark 3. Note that if 2.6 is satisfied for some $\beta_{0}>0$, then it is automatically valid for all $\beta>\beta_{0}$ as well. ### 2.1. Examples In this section we outline some examples of quite general linear equations for which Theorem 2.2 can be applied. #### 2.1.1. Linear operators with Gaussian upper bounds Many parabolic and hypoelliptic evolution operators satisfy a Gaussian upper bound for their fundamental solution, namely an estimate of the following type $\displaystyle G_{s}(x)\leq g(t)e^{-\frac{C|x|_{\mathbb{G}}^{2}}{t}}$ (2.7) where $\mathbb{G}$ is a homogeneous Lie group. The reason for writing it as above is that if we consider $|x|_{\mathbb{G}}=|x|$, being the standard Euclidean norm, we can cover a big class of parabolic operators. But for hypoelliptic equations, we get a nontrivial norm induced by the corresponding Lie-group. Let us now consider some interesting examples satisfying 2.7 with the Euclidean norm. Let $L$ in $\mathbb{R}^{n}$ be given as $\displaystyle L=\sum_{i,j=1}^{n}a_{ij}(x,t)\partial_{ij}+\sum_{j=1}^{n}b_{i}(x,t)\partial_{i}+c(x,t)$ then if the matrix $a_{ij}(x,t)$ is uniformly elliptic, the functions $a,b,c$ are bounded and uniformly Hölder continuous with exponent $\alpha$, then there exists positive constants $c_{1},c_{2}$ such that $\displaystyle|G_{t}(x)|\leq c_{1}t^{-n/2}e^{-c_{2}\frac{|x|^{2}}{t}}$ (2.8) where $G_{t}(x)$ is the fundamental solution to $\partial_{t}-L$, see [8, 9]. As such Theorem 2.2 is applicable, and since the Fourier transform of a Gaussian is a Gaussian we get that if $\displaystyle\int_{\mathbb{R}^{d}}\frac{1}{\beta_{0}+2C(c_{1})|\xi|^{2}}\hat{\gamma}(d\xi)<\infty$ (2.9) then there exists a mild solution to 2.1. That is, the condition reduces to the standard one for the heat equation, [5]. For instance, if $\gamma$ is given by the Riesz kernel, i.e. $\hat{\gamma}=|\xi|^{\lambda-d}d\xi$, then the above is verified as long as $\lambda<2$. In the case when $\gamma$ is in $W^{k,1}(\mathbb{R}^{n})$ then $\displaystyle|\hat{\gamma}|\leq\frac{C}{(1+|\xi|)^{k}},$ from this we see that 2.9 is verified if $k>n-2$. Finally we remark that if $\gamma=\delta_{0}$ (white noise), then 2.9 only holds in dimension 1. In fact the above can be extended as follows, if we instead consider the divergence form operator $\displaystyle L=\sum_{i,j=1}^{n}\partial_{i}(a_{ij}(x,t)\partial_{j}\cdot),$ with $a_{ij}$ still an elliptic matrix but now only bounded and measurable, then $\partial_{t}-L$, Aronson [1] proved that 2.8 holds also in this case. The case for non-divergence form equations with rough coefficients was treated in [7]. Let us now consider the hypoelliptic setting and consider some relevant examples. Let us begin with some notation. We consider the hypoelliptic evolution operator $\displaystyle M=\sum_{i=1}^{m}X_{i}^{2}+X_{0}-\partial_{t},$ where $X_{i}$ are smooth vector fields in $\mathbb{R}^{n}$ for $i=0,\ldots,m$, and usually $m<n$. Such an equation induces an interesting geometry. Namely, let us denote $\displaystyle Y=X_{0}-\partial_{t},\quad\text{and}\quad\lambda\cdot X=\lambda_{1}X_{1}+\ldots\lambda_{m}X_{m}$ then a curve in $\gamma:[0,R]\to\mathbb{R}^{n}\times[0,T]$ is $M$-admissible if it is absolutely continuous and $\displaystyle\gamma^{\prime}(s)=\lambda(s)\cdot X(\gamma(s))+Y(\gamma(s)),\quad\text{a.e. in [0,R]}.$ That is, the curve has a direction at each point in the tangent-bundle described by the vector-fields $X,Y$. If we assume that the vector fields are such that * • every two points $(x,t),(\xi,\tau)\in\mathbb{R}^{n}\times[0,T]$ can be connected with an $M$ admissible curve, * • there exists a homogeneous Lie group $\mathbb{G}=(\mathbb{R}^{n}\times[0,T],\circ,\delta_{\lambda})$ for which $X,Y$ are left translation invariant. Furthermore, that $X$ is $\delta_{\lambda}$-homogeneous of degree 1 and $Y$ is $\delta_{\lambda}$-homogeneous of degree 2. Then the operator $M$ is hypoelliptic (distributional solutions are smooth), see [3, 11]. This allows us to write $\mathbb{R}^{n}=V_{1}\oplus\ldots\oplus V_{l}$, i.e. the vector field $X$ stratifies $\mathbb{R}^{n}$ into a direct sum of subspaces, such that if we write $x=x^{(1)}+\ldots+x^{(l)}$, where $x^{(k)}\in V_{k}$, then the dilations becomes simple multiplication in the sense that $\displaystyle\delta_{\lambda}(x,t)=(\lambda x^{(1)}+\ldots+\lambda^{l}x^{(l)},\lambda^{2}t).$ Furthermore if we introduce the $\delta_{\lambda}$ homogeneous norm as $\displaystyle|x|_{\mathbb{G}}=\max\left\\{|x_{i}^{(k)}|^{\frac{1}{k}},k=1,\ldots,l,i=1,\ldots,m_{k}\right\\}$ Then, for such operators $M$ we have that there exists (see [11]) a positive constant $C$ such that $\displaystyle|G_{t}(x)|\leq\frac{C}{t^{\frac{Q-2}{2}}}e^{-\frac{|x|_{\mathbb{G}}^{2}}{Ct}}.$ From this we see that we can apply Theorem 2.2 again in this context, however now $\gamma$ needs to be adapted to the geometry of $\mathbb{G}$. Furthermore, for certain hypoelliptic operators we even have positivity of the fundamental solution, [2]. A prototypical example of a hypoelliptic operator is the following (well known Kolmogorov operator) $\displaystyle K=\sum_{j=1}^{n}\partial_{x_{j}}^{2}+\sum_{j=1}^{n}x_{j}\partial_{y_{j}}-\partial_{t}$ where we are working in $\mathbb{R}^{2n}\times[0,T]$ and we denote the first $n$ coordinates as $x_{i}$ and the other $n$ as $y_{i}$. The operator $K$ is hypoelliptic in the above sense [3]. Furthermore, the fundamental solution for $n=1$ is given by $\displaystyle G_{t}(x,y)=\frac{\sqrt{3}}{2\pi t^{2}}\exp\left(-\frac{x^{2}}{t}-\frac{3xy}{t^{2}}-\frac{3y^{2}}{t^{3}}\right)$ Considering the above "heat type" kernel in the finite interval $[0,T]$, it is clear that we can bound it from above by $g_{t}(x)=\frac{C}{t^{2}}\exp\left(-\frac{1}{C}\frac{x^{2}+y^{2}}{t}\right)$ for a positive constant $C$. As such the same conclusion as in Section 2.1.1 holds. Operators of this type often appear in the context of SDE’s where the noise is only in some directions, like the kinetic equations, for which the density satisfies the kinetic Fokker-Planck equation. The kinetic Fokker-Planck is hypoelliptic, see [13] and the references therein. #### 2.1.2. Fractional heat equation Consider the fractional evolution equation defined as $\displaystyle\partial_{t}u=-(-\Delta)^{s}u+Cu,\quad 0<s\leq 1,$ where $C=0$ corresponds to the fractional heat equation. The fourier transform of said equation is $\displaystyle\partial_{t}\hat{u}=-|\xi|^{2s}\hat{u}+C\hat{u}.$ As such, the fundamental solution is given by $\hat{G}_{t}(\xi)=e^{-(|\xi|^{2s}-C)t}$, and we can thus apply Theorem 2.2 with $b\equiv 0$ and $\hat{f}\geq-C$, together with Remark 1. #### 2.1.3. Mixture operators We formally consider an equation of the following type $\displaystyle\partial_{t}u=L_{1}u+L_{2}u.$ If $G_{1},G_{2}$ are the fundamental solutions of $\displaystyle\partial_{t}u=L_{1}u,\quad\partial_{t}u=L_{2}u,$ respectively, then $G=G_{1}\ast G_{2}$ is a fundamental solution of the above, (under some integrability assumptions). Indeed, we have $\displaystyle\partial_{t}(G_{1}\ast G_{2})$ $\displaystyle=(\partial_{t}G_{1}\ast G_{2})+(G_{1}\ast\partial_{t}G_{2})=(L_{1}G_{1})\ast G_{2}+G_{1}\ast(L_{2}G_{2})$ $\displaystyle=(L_{1}+L_{2})(G_{1}\ast G_{2}).$ Thus for instance we can combine Sections 2.1.1 and 2.1.2 and apply Theorem 2.2 to get the existence of a mild solution to 2.1. ## 3\. Proof of Theorem 2.2 ### 3.1. Preliminaries on stochastic calculus In this section we introduce stochastic analysis with respect to the noise $\dot{W}$. Denote by $C_{c}^{\infty}\left([0,\infty)\times\mathbb{R}^{d}\right)$ the class of $C^{\infty}$ functions on $[0,\infty)\times\mathbb{R}^{d}$ with compact support. We consider a Gaussian family of centered random variables $\left(W(\varphi),\varphi\in C^{\infty}_{c}\left([0,\infty)\times\mathbb{R}^{d}\right)\right)$ on some complete probability space $\left(\Omega,\mathcal{F},P\right)$ such that $\displaystyle\mathbb{E}[W(\varphi)W(\psi)]$ (3.1) $\displaystyle=$ $\displaystyle\int_{0}^{\infty}\int_{\mathbb{R}^{d}}\int_{\mathbb{R}^{d}}\varphi(s,y)\psi(s,y^{\prime})\gamma(y-y^{\prime})dydy^{\prime}ds:=\langle\varphi,\psi\rangle_{\mathfrak{H}}.$ By taking the Fourier transform, this can equivalently be written as 2.2. Note that here $\gamma$ is not a function, and hence 3.1 should be understood as $\mathbb{E}[W(\varphi)W(\psi)]=\int_{0}^{\infty}\int_{\mathbb{R}^{d}}\varphi(s,y)\left[\psi(s,\cdot)\ast\eta\right](y)dyds,$ (3.2) where $\ast$ denotes the convolution. For the simplicity of our presentation, we use notation $\gamma(y-y^{\prime})dydy^{\prime}$ in Sections 3.2 and 3.3 from which change of variable transformations are easier to follow. We denote by $\mathfrak{H}$ the Hilbert space defined as the closure of $C_{c}^{\infty}\left([0,\infty)\times\mathbb{R}^{d}\right)$ with respect to the inner product (3.1). As a result, we obtain an isonormal process $(W(\varphi),\varphi\in\mathfrak{H})$, which consists of a Gaussian family of centered random variable such that, for every $\varphi,\psi\in\mathfrak{H}$, $\mathbb{E}[W(\varphi)W(\Psi)]=\langle\varphi,\psi\rangle_{\mathfrak{H}}.$ The filtration $\mathbb{F}=(\mathcal{F}_{t})_{t\geq 0}$ associated to the random noise $W$ is generated by random variables $W(\varphi)$ for which $\varphi\in\mathfrak{H}$ has support contained in $[0,t]\times\mathbb{R}^{d}$. Let us now define the stochastic integral with respect to $\dot{W}$. For every random field $\\{X(s,y),s\geq 0,y\in\mathbb{R}^{d}\\}$ such that $\mathbb{E}\|X\|_{\mathfrak{H}}^{2}=\mathbb{E}\int_{0}^{\infty}\int_{\mathbb{R}^{d}}X(s,y)X(s,y^{\prime})\gamma(y-y^{\prime})dydy^{\prime}<\infty,$ we can define the stochastic integral $\int_{0}^{\infty}\int_{\mathbb{R}^{d}}X(s,y)W(ds,dy)$ in the sense of Dalang-Walsh (see, e.g. [5, 14]). It follows that we have the Isometry $\mathbb{E}\left[\int_{0}^{\infty}\int_{\mathbb{R}^{d}}X_{n}(s,y)W(ds,dy)\right]^{2}=\mathbb{E}\|X_{n}\|_{\mathfrak{H}}.$ (3.3) Moreover, we have the following version of the Burkholder-Davis-Gundy inequality: for any $t\geq 0$ and $p\geq 2$, $\displaystyle\left|\left|\int_{0}^{\infty}\int_{\mathbb{R}^{d}}X(s,y)W(ds,dy)\right|\right|_{p}^{2}$ $\displaystyle\leq c_{p}\int_{0}^{t}\int_{\mathbb{R}^{d}}\int_{\mathbb{R}^{d}}\|X(s,y)X(s,y^{\prime})\|_{\frac{p}{2}}\gamma(y-y^{\prime})dydy^{\prime}ds.$ (3.4) ### 3.2. Auxiliary results Set $I(s)=\|g_{s}\|_{L^{1}(\mathbb{R}^{d})}+\int_{\mathbb{R}^{d}}\int_{\mathbb{R}^{d}}g_{s}(y)g_{s}(y^{\prime})\gamma(y-y^{\prime})dydy^{\prime}.$ (3.5) By taking the Fourier transform, we have $I(s)=\|g_{s}\|_{L^{1}(\mathbb{R}^{d})}+\int_{\mathbb{R}^{d}}|\hat{g}_{s}(\xi)|^{2}\hat{\gamma}(d\xi)=\|g_{s}\|_{L^{1}(\mathbb{R}^{d})}+\int_{\mathbb{R}^{d}}e^{-2s\hat{f}(\xi)}\widehat{\gamma}(d\xi).$ ###### Lemma 3.1. Suppose that 2.5-2.6 hold for some $\beta_{0}$ and set $\Upsilon(\beta)=\int_{0}^{\infty}e^{-\beta s}I(s)ds.$ Then $\Upsilon:(\beta_{0},\infty)\rightarrow(0,\infty)$ is well-defined and non-increasing in $\beta$. Moreover, $\lim_{\beta\to\infty}\Upsilon(\beta)=0$. ###### Proof. By Tonelli’s theorem, we have $\int_{0}^{\infty}e^{-\beta s}\int_{\mathbb{R}^{d}}e^{-2s\hat{f}(\xi)}\widehat{\gamma}(d\xi)ds=\int_{\mathbb{R}^{d}}\int_{0}^{\infty}e^{-\beta s-2s\hat{f}(\xi)}ds\hat{\gamma}(d\xi)=\int_{\mathbb{R}^{d}}\frac{1}{\beta+2\hat{f}(\xi)}\hat{\gamma}(d\xi)$ for every $\beta$ such that $\beta+2\hat{f}(\xi)>0$. Since $I(s)$ is non- negative, it is clear that $\Upsilon(\beta)$ is non-increasing which concludes the proof. ∎ ###### Remark 4. We remark that for our purposes, it would actually suffice to consider $\Upsilon_{T}(\beta)=\int_{0}^{T}e^{-\beta s}I(s)ds$ for each fixed $T<\infty$. Hence we could replace the condition 2.5 with $\int_{0}^{T}\|g_{s}\|_{L^{1}(\mathbb{R}^{d})}ds<\infty$, cf. Remark 1. The following Proposition is the main technical ingredient. The result follows directly from [5, Lemma 15] adapted to our context. ###### Proposition 3.2. Let $I(s)$ be given by 3.5 and suppose that 2.5-2.6 hold. Let $\iota>0$, $\beta>\beta_{0}$, and $T\in(0,\infty)$ be fixed, and let $h_{n}$ be a sequence of non-negative functions such that $\sup_{t\in[0,T]}h_{0}(t)<\infty$ and, for $n\geq 1$, we have $h_{n}(t)\leq\iota\int_{0}^{t}h_{n-1}(s)e^{-\beta(t-s)}I(t-s)ds.$ (3.6) Then the series $H(\iota,p,t):=\sum_{n\geq 0}h_{n}(t)$ (3.7) converges uniformly in $t\in[0,T]$. ### 3.3. Proof of Theorem 2.2 ###### Proof of Theorem 2.2. Let $T>0$ be fixed and finite. We consider the standard Picard iterations by setting $u_{0}(t,x)=\int_{\mathbb{R}^{d}}G_{t}(x-y)u_{0}(y)dy$ and, for $n\geq 1$ and a given $\beta>\beta_{0}$, $\begin{split}e^{-\beta t}u_{n+1}(t,x)&=e^{-\beta t}u_{0}(t,x)+\int_{0}^{t}e^{-\beta t}\int_{\mathbb{R}^{d}}G_{t-s}(x-y)b(s,u_{n}(s,y))dyds\\\ &+\int_{0}^{t}e^{-\beta t}\int_{\mathbb{R}^{d}}G_{t-s}(x-y)\sigma(s,u_{n}(s,y))W(ds,dy),\hskip 14.22636ptt\geq 0,x\in\mathbb{R}^{d}.\end{split}$ We first prove that, for each $n\geq 1$, $u_{n}(t,x)$ is well-defined and satisfies $\sup_{t\in(0,T]}\sup_{x\in\mathbb{R}^{d}}\mathbb{E}\left[e^{-p\beta t}\left|u_{n}(t,x)\right|^{p}\right]<\infty.$ (3.8) We first note that if 3.8 holds for some $n\geq 0$, it follows from the Lipschitz continuity of $\sigma$ that $\begin{split}\mathbb{E}\left[e^{-p\beta s}|\sigma(s,u_{n}(s,y))|^{p}\right]&\leq C\left[\mathbb{E}\left[e^{-p\beta s}|\sigma(s,u_{0}(s,y))|^{p}\right]\right.\\\ &+\left.\mathbb{E}\left[e^{-p\beta s}|u_{n}(s,y))-u_{0}(s,y)|^{p}\right]\right].\end{split}$ By the boundedness of $\sigma$ on compacts and 2.3, we get $\sup_{s\in[0,T],y\in\mathbb{R}^{d}}\mathbb{E}\left[e^{-p\beta s}|\sigma(s,u_{0}(s,y))|^{p}\right]=\sup_{s\in[0,T],y\in\mathbb{R}^{d}}e^{-p\beta s}|\sigma(s,u_{0}(s,y))|^{p}<\infty$ and consequently, $\mathbb{E}\left[e^{-p\beta s}|\sigma(s,u_{n}(s,y))|^{p}\right]\leq C\left(1+\sup_{s\in[0,T],y\in\mathbb{R}^{d}}\mathbb{E}\left[e^{-p\beta s}|u_{n}(s,y)|^{p}\right]\right)<\infty.$ This in turn gives us, by Hölder’s inequality, that $\sup_{s\in[0,T]}\sup_{y,y^{\prime}\in\mathbb{R}^{d}}e^{-\beta s}\left|\left|\sigma(s,u_{n}(s,y))\sigma(s,u_{n}(s,y^{\prime}))\right|\right|_{\frac{p}{2}}<\infty.$ Exactly the same way, we obtain $\sup_{s\in[0,T]}\sup_{y\in\mathbb{R}^{d}}\mathbb{E}\left[e^{-p\beta s}|b(s,u_{n}(s,y))|^{p}\right]<\infty.$ In view of 3.3 and 3.4 together with the boundedness of $u_{0}(t,x)$ and $|G_{t}(y)|\leq g_{t}(y)$, applying the Minkowski integral inequality leads to $\displaystyle\mathbb{E}\left[e^{-p\beta t}\left|u_{n+1}(t,x)\right|^{p}\right]$ $\displaystyle\leq$ $\displaystyle C\left(\left[e^{-\beta t}u_{0}(t,x)\right]^{p}+\Big{\|}\int_{0}^{t}e^{-\beta t}\int_{\mathbb{R}^{d}}G_{t-s}(x-y)b(s,u_{n}(s,y))dyds\Big{\|}_{p}^{p}\right.$ $\displaystyle+$ $\displaystyle\left.\Big{\|}\int_{0}^{t}e^{-\beta t}\int_{\mathbb{R}^{d}}\int_{\mathbb{R}^{d}}G_{t-s}(x-y)G_{t-s}(x-y^{\prime})\right.$ $\displaystyle\phantom{kukkuu}\left.\times\sigma(s,u_{n}(s,y))\sigma(s,u_{n}(s,y^{\prime}))\gamma(y-y^{\prime})dy^{\prime}dyds\Big{\|}_{\frac{p}{2}}^{\frac{p}{2}}\right)$ $\displaystyle\leq$ $\displaystyle C\left[1+\left(\int_{0}^{t}e^{-\beta(t-s)}\int_{\mathbb{R}^{d}}g_{t-s}(x-y)e^{-\beta s}\|b(s,u_{n}(s,y))\|_{p}dyds\right)^{p}\right.$ $\displaystyle+$ $\displaystyle\left.\left(\int_{0}^{t}e^{-\beta(t-s)}\int_{\mathbb{R}^{d}}\int_{\mathbb{R}^{d}}g_{t-s}(x-y)g_{t-s}(x-y^{\prime})\right.\right.$ $\displaystyle\phantom{kukkuu}\left.\left.\times e^{-\beta s}\left|\left|\sigma(s,u_{n}(s,y))\sigma(s,u_{n}(s,y^{\prime}))\right|\right|_{\frac{p}{2}}\gamma(y-y^{\prime})dy^{\prime}dyds\right)^{\frac{p}{2}}\right]$ $\displaystyle\leq$ $\displaystyle C\left[1+\left(\int_{0}^{t}e^{-\beta(t-s)}I(t-s)ds\right)^{p}+\left(\int_{0}^{t}e^{-\beta(t-s)}I(t-s)ds\right)^{\frac{p}{2}}\right]$ $\displaystyle\leq$ $\displaystyle C\left(1+\Upsilon^{p}(\beta)+\Upsilon^{\frac{p}{2}}(\beta)\right).$ which is finite for $\beta>\beta_{0}$. Since $u_{0}(t,x)$ is uniformly bounded on $[0,T]\times\mathbb{R}^{d}$, 3.8 for all $n\geq 0$ thus follows from induction and hence $u_{n+1}$ is well-defined. The same arguments applied to $H_{n}(t):=\sup_{x\in\mathbb{R}^{d}}\left[e^{-\beta pt}\mathbb{E}\left[\left|u_{n+1}(t,x)-u_{n}(t,x)\right|^{p}\right]\right]^{\frac{1}{p}}$ gives us $\displaystyle H_{n}(t)$ $\displaystyle\leq C\left[\int_{0}^{t}e^{-\beta(t-s)}\int_{\mathbb{R}^{d}}g_{t-s}(x-y)H_{n-1}(s)dyds\right.$ $\displaystyle+\left(\int_{0}^{t}\int_{\mathbb{R}^{d}}\int_{\mathbb{R}^{d}}e^{-2\beta t}g_{t-s}(x-y)g_{t-s}(x-y^{\prime})\left|\left|\sigma(s,u_{n}(s,y))-\sigma(s,u_{n-1}(s,y))\right|\right|_{p}\right.$ $\displaystyle\phantom{kukkuu}\times\left.\left.\left|\left|\sigma(s,u_{n}(s,y^{\prime}))-\sigma(s,u_{n-1}(s,y^{\prime}))\right|\right|_{p}\gamma(y-y^{\prime})dy^{\prime}dyds\right)^{\frac{1}{2}}\right]$ $\displaystyle\leq C\left[\int_{0}^{t}e^{-\beta(t-s)}I(t-s)H_{n-1}(s)ds+\left(\int_{0}^{t}e^{-2\beta(t-s)}I(t-s)H^{2}_{n-1}(s)ds\right)^{\frac{1}{2}}\right].$ Here Hölder’s inequality gives $\left(\int_{0}^{t}e^{-2\beta(t-s)}I(t-s)H^{2}_{n-1}(s)ds\right)^{\frac{1}{2}}\leq C\int_{0}^{t}e^{-\beta(t-s)}I(t-s)H_{n-1}(s)ds$ leading to $H_{n}(t)\leq C\int_{0}^{t}e^{-\beta(t-s)}I(t-s)H_{n-1}(s)ds.$ By noting that $b(s,u_{0}(s,y))$ and $\sigma(s,u_{0}(s,y))$ are uniformly bounded over $[0,T]\times\mathbb{R}^{d}$, we obtain from the above computations that $\sup_{s\in[0,T]}H_{0}(s)<\infty$. Consequently, it follows from Proposition 3.2 that $\sum_{n\geq 1}H_{n}(t)$ converges uniformly on $[0,T]$, and hence the sequence $u_{n}$ converges in $L^{p}(\Omega)$ to some process $u\in L^{p}(\Omega)$, uniformly on $[0,T]\times\mathbb{R}^{d}$. From the uniform convergence one can deduce further that $u$ satisfies 2.4, and thus we have obtained the existence of the mild solution. Proving the uniqueness in a similar fashion and noting that $T>0$ was arbitrary concludes the proof. ∎ ## 4\. Concluding remarks In this article we have considered general SPDEs 2.1 and showed that one can obtain existence and uniqueness of the mild solution by classical arguments, initiated by Dalang [5], for a very large class of operators $L_{x}$. Our condition 2.6 is similar to the Dalang condition 1.2 for the stochastic heat equation and actually, our examples in Section 2.1 reveals that 1.2 is indeed sufficient for a very large class of operators $L_{x}$. By carefully examining the above proofs, we observe that even more is true. Indeed, condition 2.6 could be replaced with a condition $|G_{s}(x)|\leq g_{s}(x)$ and, for all $T>0$, $\int_{0}^{T}\int_{\mathbb{R}^{d}}e^{-\beta s}[\hat{g}_{s}(\xi)]^{2}\hat{\gamma}(d\xi)ds<\infty.$ This formulation allows to consider even more general operators $L_{t}$ in the time variable instead of considering merely $L_{t}=\partial_{t}$. In particular, with this formulation one can easily recover the existence and uniqueness result of [5] for the stochastic wave equation (in one or two dimensions) with $L_{t}=\partial_{tt}$. As another example, we can obtain the existence and uniqueness result in the case of the fractional power of the full heat operator $(\partial_{t}-\Delta)^{s}$, studied for instance in [12]. ## References * [1] T. Aronson. Bounds for the fundamental solution of a parabolic equation.Bulletin of the American Mathematical society 73.6 (1967): 890-896. * [2] B. Avelin, L. Capogna, G. Citti and K. Nyström. Harnack estimates for degenerate parabolic equations modeled on the subelliptic p-Laplacian. Adv. Math. 257 (2014), 25–65. * [3] A. Bonfiglioli, E. Lanconelli, and F. Uguzzoni. Stratified Lie groups and potential theory for their sub-Laplacians. Springer Science & Business Media, 2007. * [4] L. Chen and R. Dalang. Moments, intermittency and growth indices for nonlinear stochastic fractional heat equation. Stoch. Partial Differ. Equ. Anal. Comput., 3(3): 360–397, 2015. * [5] R. Dalang. Extending the Martingale Measure Stochastic Integral With Applications to Spatially Homogeneous S.P.D.E.’s. Electron. J. Probab. Volume 4, paper no. 6, 29 pp., 1999. * [6] L. Debbi and M. Dozzi. On the solutions of nonlinear stochastic fractional partial differential equations in one spatial dimension. Stoch. Proc. Appl., 115: 1761–1781, 2005. * [7] L. Escauriaza. Bounds for the fundamental solutions of elliptic and parabolic equations: In memory of eugene fabes. Communications in Partial Differential Equations, 25.5-6 (2000): 821–845. * [8] A. Friedman. Partial differential equations of parabolic type. Courier Dover Publications, 2008. * [9] A. Friedman. Stochastic differential equations and applications. Vol. 1. Probability and Mathematical Statistics, Vol. 28. Academic Press [Harcourt Brace Jovanovich, Publishers], New York-London, 1975. * [10] M. Foondun and D. Khoshnevisan. On the stochastic heat equation with spatially-colored random forcing. Transactions of the American Mathematical Society, 365: 409–458, 2013. * [11] A.E. Kogoj and E. Lanconelli. An invariant Harnack inequality for a class of hypoelliptic ultraparabolic equations. Mediterranean Journal of Mathematics 1.1 (2004): 51–80. * [12] K. Nyström and O. Sande. Extension properties and boundary estimates for a fractional heat operator. Nonlinear Analysis 140 (2016): 29–37. * [13] C. Villani, Hypocoercivity. No. 949-951. American Mathematical Soc., 2009. * [14] J. B. Walsh. An Introduction to Stochastic Partial Differential Equations. In: École d’été de probabilités de Saint-Flour, XIV—1984, 265-439. Lecture Notes in Math. 1180, Springer, Berlin, 1986.
# Resolution Complete In-Place Object Retrieval given Known Object Models Daniel Nakhimovich, Yinglong Miao, and Kostas E. Bekris The authors are with the Dept. of Computer Science, Rutgers, New Brunswick, NJ. Email: {d.nak, <EMAIL_ADDRESS>The work is partially supported by NSF awards 1934924 and 2021628. The opinions expressed here are those of the authors and do not reflect the positions of the sponsor. ###### Abstract This work proposes a robot task planning framework for retrieving a target object in a confined workspace among multiple stacked objects that obstruct the target. The robot can use prehensile picking and in-workspace placing actions. The method assumes access to 3D models for the visible objects in the scene. The key contribution is in achieving desirable properties, i.e., to provide (a) safety, by avoiding collisions with sensed obstacles, objects, and occluded regions, and (b) resolution completeness ($\tt RC$) - or probabilistic completeness ($\tt PC$) depending on implementation - which indicates a solution will be eventually found (if it exists) as the resolution of algorithmic parameters increases. A heuristic variant of the basic $\tt RC$ algorithm is also proposed to solve the task more efficiently while retaining the desirable properties. Simulation results compare using random picking and placing operations against the basic $\tt RC$ algorithm that reasons about object dependency as well as its heuristic variant. The success rate is higher for the $\tt RC$ approaches given the same amount of time. The heuristic variant is able to solve the problem even more efficiently than the basic approach. The integration of the $\tt RC$ algorithm with perception, where an RGB-D sensor detects the objects as they are being moved, enables real robot demonstrations of safely retrieving target objects from a cluttered shelf. ## I INTRODUCTION Robotic manipulation has the potential of being integrated into daily lives of people, such as in household service areas [1, 2]. A useful skill for such household settings involves the retrieval of a target object from a confined and cluttered workspace, such as a fridge or a shelf, which may also require the rearrangement of other objects in the process. In this context, it is important to consider how to safely retrieve objects while minimizing the time spent or the amount of pick and place operations, so as to assist humans efficiently. Figure 1: (Top) Setup for the real demonstration using an RGB-D sensor, robotiq gripper, and Yaskawa Motoman robot to retrieve the target bottle. (Bottom Left) The camera view in which objects are occluded. (Bottom Right) The corresponding voxel map. One of the challenging aspects of these problems that requires explicit reasoning relates to heavy occlusions in the scene, as the sensor is often mounted on the robot and has limited visibility. These visibility constraints complicate the task planning process, as rearranging one object can limit placements for others and can introduce new occlusions. Moreover, real-world scenes in household setups are often unstructured and involve objects with complex spatial relationships, such as objects stacked on each other. Many previous efforts on object retrieval have focused on cases where blocking objects are extracted from the workspace [3, 4], which simplifies the challenge as it does not require identifying temporary placement locations for the objects within the confined space. In-place rearrangement has been considered in some prior efforts [5]. While this prior method is efficient, it is not complete as it limits the reasoning on the largest object in the scene to analyze object traversability [4]. Alternatives use machine learning to guide the decision making [6, 7, 8], which is an exciting direction but does not easily allow for performance guarantees, such as resolution completeness. Setups where object stacking arise have received less attention and most solutions that do consider stacking are dependent on machine learning for reasoning [9, 10, 11]. Some works have proposed testbeds [12] that can help evaluate solutions in this domain. This work focuses on object retrieval in clutter where occlusions arise and objects may be initially stacked under the assumption of known object models (e.g. Figure 2). It aims at a theoretical understanding to show the algorithm has safety and $\tt RC$ guarantees. A heuristic variant improves the practical efficiency. Key features of the proposed $\tt RC$ framework are the following: * $\bullet$ it employs an adaptive dependency graph data structure inspired by solutions in object rearrangement with performance guarantees [13] that express a larger variety of object relationships than previously considered (namely occlusion dependencies); * $\bullet$ it computes the occlusion volume of detected objects as a heuristic to inform the planning process; * $\bullet$ it reasons about the collision-free placement of objects in the confined workspace efficiently by utilizing a voxelized representation of the space; * $\bullet$ it achieves $\tt RC$ (or $\tt PC$) depending on the implementation of the underlying sampling subroutines; * $\bullet$ it provides an early termination criterion when a solution cannot be found for the given resolution. Simulation results, using a model of a Yaskawa Motoman manipulator for rearranging objects on a tabletop as shown in Fig. 1, evaluate the proposed $\tt RC$ framework against a baseline using random picking and placing operations. Both variants of the $\tt RC$ reason about object dependencies. One doesn’t use heuristics and one is heuristically guided while retaining $\tt RC$. Both of the $\tt RC$ approaches outperform the baseline. The heuristically guided solution is able to solve the problem more efficiently than the basic $\tt RC$ solution. The integration of the proposed approach with perception, where an RGB-D sensor detects the objects as they are being moved, provides real robot demonstrations of safe object retrieval from a cluttered shelf. ## II RELATED WORK Some works on object retrieval rely on geometric analysis of object occlusion [3, 4]. They provide theoretical insights but frequently do not limit actions to in-place rearrangement of blocking objects. Specifically, one method constructs a dependency graph taking into account objects that jointly occlude a region and objects that block others [3]. The occlusion volume is used to estimate belief regarding the target object position and helps to construct an optimal A* algorithm. An alternative constructs a Traversability graph (T-graph) [4], where the edges encode if the largest object in the scene can be moved between two poses. It then constructs an algorithm to extract the target object, but is limited as the traversability edges are too constraining. The POMDP formulation is popular for the task [14, 15], which allows the application of general POMDP solvers. The POMDP formulation was also adopted by the work that formalizes object retrieval in unstructured scenes as ”mechanical search” [16]. Alternatives rely on learning-based methods to solve such challenges, such as reinforcement learning [6] or target belief prediction [7, 9, 17]. They report good performance but do not provide theoretical guarantees given the black-box nature of the solutions. In particular, a reinforcement learning solution [6] uses the rendered top-down projections of the scene to predict the target poses. A recent follow-up effort [7] on previous work [17] estimates the 1D position belief of the target object on the shelf via machine learning. It then constructs a policy based on the distribution change after applying pushing and suction actions. It incorporates stacking and unstacking actions, where object stacking is represented by a tree structure. Other works such as [18] utilize learning for planning grasps to greedily empty bins of complex and novel objects. A related work [19] proposes a complete framework to safely reconstruct all objects in the scene amidst object occlusions. Nevertheless, object retrieval may not require reconstructing all objects and requires a search procedure that is more task-driven for efficiency. There are also previous works [20, 19] that construct a voxelization of the environment to model object occlusions, similar to the current work. This representation is used to compute an object’s occlusion volume, which provides heuristic guidance. Object spatial relationships are often represented by scene graphs [10, 11], or implicitly in machine learning solutions [16, 21, 22]. What stands out in this work is that it proposes a general template for a $\tt RC$ or $\tt PC$ approach to task retrieval in occluded environments that only relies on basic motion and perception primitives. This modular nature allows for quick sim-to-real transfer and passive performance improvement as the primitives are improved over time. Furthermore, an efficient implementation is demonstrated utilizing a voxelized representation of the environment for quick collision filtering of object placements as well as providing an effective heuristic to rank object manipulations. Thus, the framework enables effective in-workspace manipulation. ## III PROBLEM STATEMENT Consider an environment with a set $\mathbb{S}=\\{o_{1},...,o_{n}\\}\subset\mathbb{O}$ of $n$ objects for which there are available 3D models. The objects are stably resting on a support surface. Objects are allowed to be initially stacked and occlude each other from the camera view. The robot has one fixed RGB-D sensor at its disposal. Discovered objects are those recognized given the observation history. An objects is assumed to be recognized once an image segmentation process recognizes it as an individual object in the observed image. Similarly, a perception method for detecting the target object once observed is assumed. The region of the workspace occluded by object $o_{i}$ at pose $s_{i}$ is denoted as $O_{i}(s_{i})$. Similarly the space uniquely occluded by object $o_{i}$ at pose $s_{i}$, called the direct occlusion space, is denoted as $\tilde{O}_{i}(s_{i})$. The proposed algorithms gradually removes occlusions and recognizes objects. A motion planner is used to plan pick-and-place actions. While objects can start out stacked, they are not re-stacked and are only placed on the ground surface during actions. While objects can start out stacked, no reasoning about stability and ability to re-stack objects is considered. Thus, once an action to pick up a stacked object is taken, that object will only be placed on the ground surface. For further assumptions and required properties of the motion planner see section V. The objective for the object retrieval task is to determine a sequence of pick and place actions in order to discover and subsequently retrieve the target object; the target need not be directly visible or pickable from the robot’s sensors. The corresponding solution should provide desirable guarantees: (a) safety, by avoiding collisions with sensed obstacles and objects as well as occluded regions, and (b) resolution completeness (RC) - or alternatively probabilistic completeness, depending on the implementation of the underlying motion planner, grasping process and object placement sampling. The optimization objective is to minimize the number of performed actions until the target object is retrieved. ## IV METHOD The proposed pipeline is detailed in Algorithm 1. First a voxelized representation of the scene and a dependency graph are computed (lines 3,4,5), which are detailed in subsection IV-A. The dependency graph contains a belief state of the current scene based on visibility and reachability constraints. All the sinks of this directed graph represent likely pickable objects. The ranks are described later (see subsection IV-C). If the target object is pickable then it is retrieved and the pipeline is terminated line 6. Otherwise, a placement is planned (see subsection IV-B) one at a time for each object in a shuffled order biased by the ranks (TryMoveOne line 7). The first successful plan is executed. If no placement is found for any of the pickable objects, then a fallback procedure is called to try and pick one object and move it temporarily out of view of the camera; the first successful plan is executed, the scene is re- sensed, and a new placement is sampled for the object or the object is simply put back at the same spot (MoveOrPlaceback line 10). At this stage, the pipeline could restart if a new object was discovered (lines 12-15). Otherwise the same set of pickable objects are tested for new placements in the scene (line 16). If these two operations of “moving an object to look behind it”(MoveOrPlaceback) followed by retrying to “move one of the pickable objects to a new spot”(TryMoveOne) fail for all objects, then the pipeline can return and report failure for the current resolution (line 22). If the sequence of operations succeeds, then the pipeline can restart (line $19\rightarrow 2$). Algorithm 1 $\tt RC\\_Pipeline$($\tt target$) 1:$\tt failure\leftarrow false$ 2:while $\tt failure=false$ do 3: $\tt space\leftarrow UpdateVoxelsFromImage()$ 4: $\tt dg\leftarrow{\tt DepGraph(space)}$ 5: $\tt sinks,ranks\leftarrow RankSinks(target,dg)$ 6: if $\tt target\in sinks$ then $\tt break$ 7: if $\tt{\tt TryMoveOne(sinks,ranks)}=false$ then 8: $\tt failure\leftarrow true$ 9: for $\tt sink\in sinks$ do 10: if $\tt{\tt MoveOrPlaceback(sink)}=false$ then 11: $\tt continue$ 12: $\tt space\leftarrow UpdateVoxelsFromImage()$ 13: if $\tt DidDiscoverObject(space)$ then 14: $\tt failure\leftarrow false$ 15: $\tt break$ 16: if $\tt{\tt TryMoveOne(sinks,\emptyset)}=false$ then 17: $\tt continue$ 18: $\tt failure\leftarrow false$ 19: $\tt break$ 20:if $\tt not~{}failure$ then 21: $\tt Retrieve(target)$ 22:$\tt return~{}failure$ \begin{overpic}[width=145.69568pt,trim=0.0pt 140.525pt 40.15pt 22.08249pt,clip]{figures/sim_exp/step1_ws.png} \put(3.0,10.0){\huge$\displaystyle 1)$} \end{overpic} \begin{overpic}[width=145.69568pt,trim=0.0pt 140.525pt 40.15pt 22.08249pt,clip]{figures/sim_exp/step2_ws.png} \put(3.0,10.0){\huge$\displaystyle 2)$} \end{overpic} \begin{overpic}[width=145.69568pt,trim=0.0pt 140.525pt 40.15pt 22.08249pt,clip]{figures/sim_exp/step3_ws.png} \put(3.0,10.0){\huge$\displaystyle 3)$} \end{overpic} \begin{overpic}[width=145.69568pt,trim=0.0pt 140.525pt 40.15pt 22.08249pt,clip]{figures/sim_exp/step4_ws.png} \put(3.0,10.0){\huge$\displaystyle 4)$} \end{overpic} \begin{overpic}[width=145.69568pt,trim=0.0pt 140.525pt 40.15pt 22.08249pt,clip]{figures/sim_exp/step5_ws.png} \put(3.0,10.0){\huge$\displaystyle 5)$} \end{overpic} \begin{overpic}[width=145.69568pt,trim=0.0pt 140.525pt 40.15pt 22.08249pt,clip]{figures/sim_exp/step6_ws.png} \put(3.0,10.0){\huge$\displaystyle 6)$} \end{overpic} \begin{overpic}[width=106.23477pt]{figures/sim_exp/step1_dg_c.png} \put(-1.0,5.0){\huge a)} \end{overpic} \begin{overpic}[width=106.23477pt]{figures/sim_exp/step2_dg.png} \put(-1.0,5.0){\huge b)} \end{overpic} \begin{overpic}[width=106.23477pt]{figures/sim_exp/step3_dg.png} \put(-1.0,5.0){\huge c)} \end{overpic} \begin{overpic}[width=106.23477pt]{figures/sim_exp/step4_dg.png} \put(-1.0,5.0){\huge d)} \end{overpic} Figure 2: Images (1)-(6) show a simulated experiment from initial configuration to final one action at a time with corresponding camera views in the top left. The corresponding generated dependency graphs transitioning between the images (1)-(2), (2)-(3), (3)-(4), and (4)-(5) are shown in images (a)-(d). The colors of nodes corresponds with the objects in the scene and the red object (labeled ‘T’) is the target object for the trial. The last graph between images 5-6 is not shown since it is trivial having no dependencies between any objects. ### IV-A Voxel Map and Dependency Graph A 3D occlusion voxel grid within the workspace is constructed from RGB-D images. First the point cloud (in world frame) is generated using the RGB-D image and the inverse of the camera projection. These points are down-sampled into a voxelization of the scene. Given segmentation image information, each object is associated with a portion of the voxel grid. Object geometry is used to label voxels as occupied and remaining associated voxels as occluded. The occluded regions of objects may intersect when jointly occluded. The dependency graph is a directed graph where each node represents a visible (or target) object and a labeled edge $(o_{i},o_{j},r)$ represents a relation $r$ from object $o_{i}$ to $o_{j}$ that necessitates $o_{j}$ to be picked and placed before $o_{i}$ could be picked. Valid relations in this work include “below”, “grasp blocked by”, and – for the prediction of the target object – “hidden by”. See Figure 2 for a sequence of such dependency graphs generated during an example experiment. Object x is defined to be “below” object y ($x\xrightarrow{below}y$) if object x touches object y and the z-coordinate of the center of mass of x is less than that of y. Note that this isn’t guaranteed to capture all intuitive cases of one object being below another for non-convex objects. This relation is computed using object models and poses given by the perception system. Object x has its “grasp blocked by” y ($x\xrightarrow{blocked}y$) if there are no collision free grasp poses for object x and the arm is in collision with y for one or more valid grasp poses. (Grasp poses are sampled and tested by inverse kinematics (IK) for discovered objects) Grasp poses are sampled using inverse kinematics (IK) to discovered objects. (or, if the grasping pose collides with objects, an edge to each collided object is added) If there exists a collision free grasp for an object, no blocking edges are added; otherwise, an edge to each object that has a collision with the arm is added. Note that although this relation guarantees that the source object blocks the target, it doesn’t capture all such reachability dependencies. This however is not an issue for completeness as Algorithm 1 will eventually try to grasp all objects in the case of motion planning failure. The target object t is possibly “hidden by” x ($t\xrightarrow{hidden}x$) if the target isn’t sensed in the scene and object x is touching the table. This relation is used to keep track of the belief state of where the target is. Each edge is assigned a probability based on the volume of the occluded space behind object x (see subsection IV-C). ### IV-B Placement Sampling A valid placement is one that doesn’t collide with another object or any undiscovered area of the workspace. Instead of randomly sampling x,y coordinates for an object and checking for collision at that point, we create a grid matching the horizontal extents of the workspace and add a collision mask which is the shadow of the object occupancy and occlusion voxel grid looking from a birds eye view. This mask is then convolved with the shadow of the object that is to be placed. The occupied pixel indices indicate collision-free placements and are converted to world coordinates. Object orientations can be enumerated by rotating the object shadow. ### IV-C Target Object Prediction Intelligent object location prediction is achieved by applying a heuristic which ranks the pickable objects determined by the dependency graph (line 4 in Algorithm 1). This is done by augmenting the dependency graph edges with a weight $p\in(0,1]$ estimating the probability for the relation $o_{i}\overset{r}{\longrightarrow}o_{j}$ to be true. To rank a pickable object, the sum of products of edge weights of all simple paths between the target and the object is computed. When sampling from the list of pickable objects, this rank is used as a probability weight. For the “below” relation $p=1$ since object segmentation is assumed to be reliable. For the “grasp blocked by” relation p is equal to the fraction of total sampled grasps for which that object is in collision with the arm. Note, the weights of grasp blocking edges coming out of any object need not add to one (hence don’t truly represent probabilities) since the arm could be colliding with multiple objects for any single grasp. For the “hidden by” relation, the goal is to encourage knowledge gain of the environment. This is done by normalizing the volume of the direct occlusion region of each stack of objects and assigning the inverse as the probability estimate that the target is hidden behind each stack. This heuristic biases the pipeline towards discovering large volumes of occluded workspace. From an algorithmic point of view, there is technically no reason to normalize the output of the heuristic, however, representing the heuristics as probabilities is insightful since - without prior knowledge - the probability the hidden object to be in a larger volume is larger than the probability of it being in a smaller volume. Furthermore, modeling the dependency graph edges with probabilities as opposed to non-normalized weights is conducive for exploring future work which might seek to combine the probabilities based on the proposed volumetric-heuristics with priors based on the semantics of the objects involved or from an additional human instruction (see section VII work). ## V RESOLUTION COMPLETENESS Given a (formally) complete motion planner (finds solution in finite time if exists), a continuous space grasp sampler, and continuous placement sampler, the proposed pipeline would be $\tt PC$ Given a $\tt RC$ motion planner, a discrete space grasp sampler, and discrete placement sampler, the proposed pipeline would be $\tt RC$. To show the $\tt PC$ or $\tt RC$ of the algorithm proposed in this work, a simpler version of the algorithm is analysed first. Without loss of generality, the actual proposed algorithm will be likewise proven complete. Consider a much simpler algorithm that, at every iteration, tries to pick an object at random and, if is not the target object, subsequently place it randomly in the explored region of the workspace; call this RAND-ACT. ###### Lemma V.1 RAND-ACT is $\tt PC$ or $\tt RC$ (depending on the planning and sampling subroutines). ###### Proof: At every iteration, RAND-ACT attempts to perform a random action. Consequently, RAND-ACT executes a random walk on the space of all actions. Since pick and place actions are reversible, if a sequence of such valid actions exists, this algorithm will eventually perform it (or an augmented version of the sequence - i.e. placing an object back to where it was picked from) in the limit (or finite time for RC subroutines). ∎ ###### Corollary V.2 Algorithm 1 is $\tt PC$ or $\tt RC$ (depending on the planning and sampling subroutines)). ###### Proof: Indeed Algorithm 1 is really a fancy implementation of RAND-ACT. At every iteration, the dependency graph is used to identify (and heuristically rank) the currently pickable objects. One of these is chosen randomly for a pick and place action via the $\tt TryMoveOne$ subroutine. The $\tt MoveOrPlaceback$ subroutine acts as a fallback in case there are no new discovered placements found; it delays placement sampling till after the environment is re-sensed with the picked object moved out of the way. Notice further, that even if the pickable objects are sampled weighted according to their ranking rather than uniformly, the action space is still explored entirely because each action has a positive probability of being sampled. Thus, w.l.o.g. Algorithm 1 is $\tt PC$ or $\tt RC$ as well. ∎ Failure Detection: The implementation in this work uses the $\tt RC$ approach. In addition to $\tt RC$, Algorithm 1 takes a step closer towards achieving general completeness by actually detecting certain unsolvable cases within the completeness constraints of the motion and sampling subroutines. The detectable unsolvable instances for which the algorithm will return failure in finite time are as follows. * $\bullet$ No object can be grasped. This could happen if two objects each block the grasp of the other. * $\bullet$ No objects can be placed anywhere (except its current spot). This could happen in a highly cluttered scene where the only valid placement for each object is to just put it back where it was. Thus, Algorithm 1 has stronger guarantees than $\tt RC$ but is not formally complete since it may run forever by juggling two objects between two placements. A Caveat: A fundamental assumption in the argument presented is that actions are reversible. This is not always true in practice depending on the implementation of the sampling subroutines. And indeed, the implementation of the placement sampling process proposed in subsection IV-B implies irreversible actions for scenes with stacked objects because it does not consider the possibility of re-stacking objects. Thus, the proposed pipeline (as implemented) is only resolution complete on the any sub-task where the objects are no longer stacked. Implementation of stacking actions is planned for future work. ## VI EXPERIMENTS Figure 3: Execution on the real robot : (a) Initial scene where the red bottle is hidden. (b) The robot moves the yellow bottle, which occludes the most space. (c) The robot moves the second yellow bottle, revealing the red bottle. (d and e) The robot moves the green and the blue bottles to reach the red bottle. (f) Target is now reachable. Simulated experiments and the real demonstrations are performed with the Yaskawa Motoman sda10f, with a robotiq 85 gripper attached on the right arm. The simulated trials are randomly generated by picking random objects, dimensions, and collision-free placements within the specified workspace. 20 scenes with each of 6, 8, 10, 12, and 14 objects were used, all of which contain objects occluded from the camera. Each of the 100 trials is a unique scene. The target object in each scene was selected to be the hidden object with the most objects above it, if any. All tested algorithms were given 20 minutes to run before being terminated. A trial run is considered successful only if the target object was retrieved within the time limit. Discovering but failing to pick up the object was still considered a failure. Comparisons of the success rate, number of actions for solved trials, total run-time for solved trials, and number of timed out trials are shown in Figure 4 for 3 algorithms. The algorithms compared are the baseline random action approach (blue in figure), the proposed resolution complete pipeline without the object ranking heuristic (orange) and with it (green). For the resolution complete approaches timing-out is not the only failure mode since they could detect certain infeasible problems (section V: Failure Detection) The success rate of the resolution complete approaches is higher than that of the random baseline. Although its not directly apparent from the plotted results in Figure 4, the resolution complete approaches (as expected) always found a solution when the random baseline found a solution; however in one such trial the resolution complete approach without heuristic exceeded the 20 min time-limit. It is also clear that the heuristic approach has better success than the non heuristic approach even though they are both complete. Looking at the data for timed-out experiments, it becomes clear that the increased success of the heuristic approach is due to timing out less frequently. This also coincides with the data showing that the heuristic approach overwhelmingly finds solutions faster and with fewer object manipulations. In fact, while the non-heuristic RC approach started timing out linearly with increase in the number of objects, the heuristic approach had virtually no issue until the scenes got very cluttered with 14 objects. It is clear that for all methods, success rate starts dropping off significantly at around 14 objects. This marks the difficulty level for the given industrial Motoman robot and the workspace. A more compact robot with a streamlined end-effector (such as the “bluction” tool [7]) could scale to more cluttered scenes. Figure 4: On top are shown the graphs of the: success rate (left) and number of timed out trials (right). On the bottom are the number of actions and the total runtime for the subset of trials in which all algorithms were successful. ### VI-A Integration with Perception & Real Robot Demonstration The pipeline is directly transferable to scenarios on the real robot to retrieve a target red bottle from a cluttered shelf. Due to time constraints, a simple implementation of a perception system is used which only segments and detects colored cylinders without stacking. Despite the simplifications, a scene with significant object occlusion is still demonstrated with a successful retrieval. The proposed pipeline (with heuristic) is run online and communicates with the robot controller and the RGBD camera for execution and sensing. The camera extrinsic matrix is estimated by a classical robot-camera calibration procedure using ArUco markers [23]. For object recognition, the perception component is implemented via plane fitting, DBSCAN segmentation [24], and cylinder fitting using Open3D [25]. The plane fitting algorithm extracts the boundaries of the workspace, which is used to construct the collision geometries in MoveIt [26]. The inliers of each segmented cylinder are used to produce the segmentation mask for the RGBD image, which is used to label occlusion correspondence for each object. To ensure safety, additional cubic collision geometries are added to the planning scene to avoid collisions between the robot and the camera. 111Videos can be found at https://sites.google.com/scarletmail.rutgers.edu/occluded-obj-retrieval. Extensive experiments of the proposed pipeline were not performed on the real robot but the demonstration presented was performed a few times and the pipeline was observed to have qualitatively similar performance as in simulated experiments; however, calibration and perception issues were observed to lead to pipeline failure. ## VII DISCUSSION It’s worth mentioning that the physical execution accounts for over 60% of time used for the trials. This shows that there could be room for performance improvement by performing scene perception asynchronously, since a lot can still be sensed while the robot is moving. Further performance improvement can be found by parallelizing the planning of picks and placements for multiple objects as well. While this work applies heuristics for selecting objects based on the occlusion volume, additional information regarding effective placements can also improve practical performance. In order to solve a larger variety of problems it would be useful to adapt the placement primitive to allow placing objects on top of others when there is limited space on the workspace surface. Another direction is to integrate the task planner with human instructions. For instance, it would be helpful to use human language to identify the target as well as influence the search at some regions over others. Additional heuristics can also be obtained from semantic reasoning of the scene when objects of the same category tend to be placed closer [27]. Since current experiments only include simple geometries, such as cylinders and rectangular prisms, future work can investigate more complex objects where state-of-the- art perception algorithms are necessary. This would also be necessary for realistic human-robot integration. ## References * [1] L. L. Wong, L. P. Kaelbling, and T. Lozano-Pérez, “Manipulation-based active search for occluded objects,” in 2013 IEEE International Conference on Robotics and Automation, pp. 2814–2819, IEEE, 2013. * [2] C. R. Garrett, C. Paxton, T. Lozano-Pérez, L. P. Kaelbling, and D. Fox, “Online replanning in belief space for partially observable task and motion problems,” in 2020 IEEE International Conference on Robotics and Automation (ICRA), pp. 5678–5684, IEEE, 2020. * [3] M. R. Dogar, M. C. Koval, A. Tallavajhula, and S. S. Srinivasa, “Object search by manipulation,” Autonomous Robots, vol. 36, no. 1, pp. 153–167, 2014\. * [4] C. Nam, S. H. Cheong, J. Lee, D. H. Kim, and C. Kim, “Fast and resilient manipulation planning for object retrieval in cluttered and confined environments,” IEEE Transactions on Robotics, vol. 37, no. 5, pp. 1539–1552, 2021. * [5] J. Ahn, J. Lee, S. H. Cheong, C. Kim, and C. Nam, “An integrated approach for determining objects to be relocated and their goal positions inside clutter for object retrieval,” in 2021 IEEE International Conference on Robotics and Automation (ICRA), pp. 6408–6414, IEEE, 2021. * [6] W. Bejjani, W. C. Agboh, M. R. Dogar, and M. Leonetti, “Occlusion-aware search for object retrieval in clutter,” in 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 4678–4685, IEEE, 2021\. * [7] H. Huang, M. Danielczuk, C. M. Kim, L. Fu, Z. Tam, J. Ichnowski, A. Angelova, B. Ichter, and K. Goldberg, “Mechanical search on shelves using a novel ”bluction” tool,” IEEE International Conference on Robotics and Automation (ICRA), 2022. * [8] A. Price, L. Jin, and D. Berenson, “Inferring occluded geometry improves performance when retrieving an object from dense clutter,” in The International Symposium of Robotics Research, pp. 376–392, Springer, 2019. * [9] H. Huang, L. Fu, M. Danielczuk, C. M. Kim, Z. Tam, J. Ichnowski, A. Angelova, B. Ichter, and K. Goldberg, “Mechanical search on shelves with efficient stacking and destacking of objects,” arXiv preprint arXiv:2207.02347, 2022\. * [10] Y. Zhu, J. Tremblay, S. Birchfield, and Y. Zhu, “Hierarchical planning for long-horizon manipulation with geometric and symbolic scene graphs,” in 2021 IEEE International Conference on Robotics and Automation (ICRA), pp. 6541–6548, IEEE, 2021. * [11] K. N. Kumar, I. Essa, and S. Ha, “Graph-based cluttered scene generation and interactive exploration using deep reinforcement learning,” in 2022 International Conference on Robotics and Automation (ICRA), pp. 7521–7527, IEEE, 2022. * [12] Z. Liu, W. Liu, Y. Qin, F. Xiang, M. Gou, S. Xin, M. A. Roa, B. Calli, H. Su, Y. Sun, et al., “Ocrtoc: A cloud-based competition and benchmark for robotic grasping and manipulation,” IEEE Robotics and Automation Letters, vol. 7, no. 1, pp. 486–493, 2021. * [13] R. Wang, K. Gao, D. Nakhimovich, J. Yu, and K. E. Bekris, “Uniform Object Rearrangement: From Complete Monotone Primitives to Efficient Non-Monotone Informed Search,” in 2021 IEEE International Conference on Robotics and Automation (ICRA), pp. 6621–6627, May 2021\. ISSN: 2577-087X. * [14] W. Zhao and W. Chen, “Hierarchical pomdp planning for object manipulation in clutter,” Robotics and Autonomous Systems, vol. 139, p. 103736, 2021. * [15] Y. Xiao, S. Katt, A. ten Pas, S. Chen, and C. Amato, “Online planning for target object search in clutter under partial observability,” in 2019 International Conference on Robotics and Automation (ICRA), pp. 8241–8247, IEEE, 2019. * [16] M. Danielczuk, A. Kurenkov, A. Balakrishna, M. Matl, D. Wang, R. Martín-Martín, A. Garg, S. Savarese, and K. Goldberg, “Mechanical search: Multi-step retrieval of a target object occluded by clutter,” in 2019 International Conference on Robotics and Automation (ICRA), pp. 1614–1621, IEEE, 2019. * [17] H. Huang, M. Dominguez-Kuhne, V. Satish, M. Danielczuk, K. Sanders, J. Ichnowski, A. Lee, A. Angelova, V. Vanhoucke, and K. Goldberg, “Mechanical search on shelves using lateral access x-ray,” in 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 2045–2052, IEEE, 2021. * [18] A. Zeng, S. Song, K.-T. Yu, E. Donlon, F. R. Hogan, M. Bauza, D. Ma, O. Taylor, M. Liu, E. Romo, et al., “Robotic pick-and-place of novel objects in clutter with multi-affordance grasping and cross-domain image matching,” The International Journal of Robotics Research, vol. 41, no. 7, pp. 690–705, 2022. * [19] Y. Miao, R. Wang, and K. Bekris, “Safe, occlusion-aware manipulation for online object reconstruction in confined spaces,” in International Symposium on Robotics Research (ISRR), 2022. * [20] M. Gupta, T. Rühr, M. Beetz, and G. S. Sukhatme, “Interactive environment exploration in clutter,” in 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems, pp. 5265–5272, IEEE, 2013. * [21] J. Poon, Y. Cui, J. Ooga, A. Ogawa, and T. Matsubara, “Probabilistic active filtering for object search in clutter,” in 2019 International Conference on Robotics and Automation (ICRA), pp. 7256–7261, IEEE, 2019. * [22] T. Novkovic, R. Pautrat, F. Furrer, M. Breyer, R. Siegwart, and J. Nieto, “Object finding in cluttered scenes using interactive perception,” in 2020 IEEE International Conference on Robotics and Automation (ICRA), pp. 8338–8344, IEEE, 2020. * [23] R. Y. Tsai, R. K. Lenz, et al., “A new technique for fully autonomous and efficient 3 d robotics hand/eye calibration,” IEEE Transactions on robotics and automation, vol. 5, no. 3, pp. 345–358, 1989. * [24] M. Ester, H.-P. Kriegel, J. Sander, X. Xu, et al., “A density-based algorithm for discovering clusters in large spatial databases with noise.,” in kdd, vol. 96, pp. 226–231, 1996. * [25] Q.-Y. Zhou, J. Park, and V. Koltun, “Open3d: A modern library for 3d data processing,” arXiv preprint arXiv:1801.09847, 2018. * [26] S. Chitta, “Moveit!: an introduction,” in Robot Operating System (ROS), pp. 3–27, Springer, 2016. * [27] J. K. Li, D. Hsu, and W. S. Lee, “Act to see and see to act: Pomdp planning for objects search in clutter,” in 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 5701–5707, IEEE, 2016\.
# Disorder induced transition from type-I to type-II superconductivity in the Dirac semimetal PdTe2 M. V. Salis<EMAIL_ADDRESS>J. P. Lorenz Y. K. Huang A. de Visser <EMAIL_ADDRESS>Van der Waals - Zeeman Institute, University of Amsterdam, Science Park 904, 1098 XH Amsterdam, The Netherlands ###### Abstract We report a doping study directed to intentionally induce disorder in PdTe2 by the isoelectronic substitution of Pt. Two single-crystalline batches Pd1-xPtxTe2 have been prepared with nominal doping concentrations $x=0.05$ and $x=0.10$. Sample characterization by energy dispersive x-ray spectroscopy (EDX) revealed Pt did not dissolve homogeneously in the crystals. For the nominal value $x=0.10$ small single crystals cut from the batch appeared to have $x=0.09$, as well as the non stoichiometric composition Pd0.97Pt<0.004Te2.03. Magnetic and heat capacity measurements demonstrate a transition from type-I to type-II superconducting behavior upon increasing disorder. From transport measurements we calculate a residual resistivity $\rho_{0}=1.4~{}\mu\Omega$cm suffices to turn PdTe2 into a superconductor of the second kind. ## I introduction Recently, interest in transition metal dichalcogenides has increased significantly due to their extraordinary electronic properties. Notably, the opportunity to realize novel quantum states arising from the topologically non-trivial band structure, as found by density functional theory Soluyanov _et al._ (2015); Huang _et al._ (2016); Yan _et al._ (2017); Bahramy _et al._ (2018), attracts much attention. The formation of both type-I and type-II bulk Dirac cones has been predicted Bahramy _et al._ (2018). Of special interest in this family is the semimetal PdTe2 since it undergoes a superconducting transition at $T_{c}\sim 1.7$ K Guggenheim _et al._ (1961). Furthermore, PdTe2 is classified as a type-II Dirac semimetal, as uncovered by angle-resolved photoemission spectroscopy and ab initio electronic structure calculations Bahramy _et al._ (2018); Yan _et al._ (2015); Fei _et al._ (2017); Noh _et al._ (2017); Clark _et al._ (2018). A type-II Dirac semimetal is characterized by a Dirac cone with a tilt parameter $k>1$ leading to broken Lorentz invariance Soluyanov _et al._ (2015). It is predicted that for Dirac semimetals with $k\approx 1$, meaning close to the topological transition at $k=1$, superconductivity is generally of the second type (type- II) Rosenstein _et al._ (2018). For $k>1$, superconductivity becomes of the first kind (type-I). Interestingly, PdTe2 Leng _et al._ (2017); Salis _et al._ (2021) is a type-I superconductor and based on its $T_{c}$ Shapiro et al. Shapiro _et al._ (2018) estimated $k\approx 2$. In view of the effect topology has on superconductivity in these systems, it is of interest to investigate whether the superconductivity type can be altered by, for instance, doping. Superconductivity in PdTe2 has been explored in great detail. Type-I superconductivity was uncovered with help of magnetic and transport measurements on single crystals Leng _et al._ (2017). The intermediate state, a hallmark of type-I behavior, was observed through the dc magnetization curves and the differential paramagnetic effect in the ac susceptibility data. Here, a bulk critical field $B_{c}(0)=13.6$ mT was determined in conjunction with a surface critical field $B_{c}^{S}(0)=34.9$ mT. Moreover, the temperature dependence of the surface superconductivity did not follow the Saint-James $-$ de Gennes model Saint-James and de Gennes (1963). Peculiarly, from resistance measurements a critical field $B_{c}^{R}(0)=0.32$ T was deduced. Weak-coupling conventional superconductivity in PdTe2 was demonstrated via measurements of the heat capacity Amit and Singh (2018); Salis _et al._ (2021), penetration depth Teknowijoyo _et al._ (2018); Salis _et al._ (2018), scanning tunneling microscopy and spectroscopy (STM/STS) Clark _et al._ (2018); Das _et al._ (2018); Sirohi _et al._ (2019), and side junction tunneling spectroscopy Voerman _et al._ (2019). Superconductivity is partly attributed to a van Hove singularity situated at $\sim 30$ meV above the Fermi level Kim _et al._ (2018); van Heumen _et al._ (2019). On the other hand, a mixed type-I and type-II superconducting state was concluded from STM/STS Das _et al._ (2018); Sirohi _et al._ (2019) and point contact spectroscopy (PCS) Le _et al._ (2019) measurements. In a magnetic field a range of critical fields was observed at the surface, which was explained by spatially separated type-I and type-II regions. However, later muon spin rotation measurements Leng _et al._ (2019a) and scanning squid magnetometry Garcia-Campos _et al._ (2021) provide solid evidence for bulk type-I superconductivity probed on the microscopic and macroscopic scale, respectively. Finally, evidence for bulk type-I superconductivity was attained through heat capacity measurements by demonstrating the presence of latent heat Salis _et al._ (2021). Measurements under hydrostatic pressure show that superconductivity is still present at 5.5 GPa Furue _et al._ (2021) and remains of the first kind at least till 2.5 GPa Leng _et al._ (2019b). Substitution or doping studies using PdTe2 are scarce. Kudo et al. Kudo _et al._ (2016) examined Pd substitution in AuTe2 by preparing a series of Au1-xPdxTe2 samples. Bulk superconductivity emerges at $x\approx 0.55$ with $T_{c}\approx 4.0$ K as evidenced by heat capacity measurements. At lower $x$-values the Te-Te dimer connections stabilize a monoclinic crystal structure in which superconductivity is absent Kudo _et al._ (2016). The strong-coupled nature of superconductivity near $x\approx 0.55$ is attributed to a large density of states (DOS) at the Fermi level. Further increasing the Pd content results in weak coupling superconductivity with lower transition temperatures, as expected from approaching the stoichiometric end compound PdTe2. Ryu investigated Cu doping in PdTe2 by preparing a series of CuxPdTe2 samples Ryu (2015). Optimal doping was found near $x=0.05$ with bulk superconductivity at $T_{c}\approx 2.6$ K Ryu (2015); Hooda and Yadav (2018). The increase of $T_{c}$ is attributed to an increase in the DOS at the Fermi level due to the hybridization of Te-p and Cu-d orbitals along the $c$-axis, effectively reducing the 2D-nature of this layered material. This is in-line with the Cu atoms being intercalated in the Van der Waals gaps. STM/STS measurements provide evidence that Cu0.05PdTe2 is a homogeneous type-II superconductor Vasdev _et al._ (2019). This change, compared to the STM/STS data on PdTe2 Das _et al._ (2018); Sirohi _et al._ (2019) that revealed a mixed type-I/II behavior, is explained by Cu intercalation inducing disorder. This effectively reduces the electron mean free path $l_{e}$ and the coherence length $\xi$, thus increasing the Ginzburg Landau (GL) parameter $\kappa=\frac{\lambda}{\xi}$ to larger than the $1/\sqrt{2}$ threshold for type-I behavior. Here we report the results of a doping study, directed to intentionally induce disorder in PdTe2 by substituting Pd by iso-electronic Pt. We have prepared Pd1-xPtxTe2 crystals with nominal doping concentrations $x=0.05$ and $x=0.10$. Sample characterization by energy dispersive x-ray spectroscopy (EDX) revealed that Pt did not dissolve homogeneously in the crystals. Notably, small crystals cut from the nominal $x=0.10$ batch appeared to have $x=0.09$, or the non stoichiometric composition Pd0.97Pt<0.004Te2.03. Transport, magnetic and heat capacity measurements demonstrate a transition from type-I to type-II superconducting behavior upon increasing disorder. ## II Experimental PdTe2 crystallizes in the trigonal CdI2 structure (space group $P\bar{3}m1$). Two single-crystalline batches Pd1-xPtxTe2 were prepared with $x=0.05$ and $x=0.10$ using a modified Bridgman technique Lyons _et al._ (1976). The same technique was previously used to prepare PdTe2 single crystals Leng _et al._ (2017). Small flat crystals were cut from the prepared batches by a scalpel. The crystals have an area of $2\times 3$ mm2 and a thickness of about 0.3 mm. Scanning electron microscopy with energy dispersive x-ray spectroscopy (SEM/EDX) was carried out with help of a Hitachi table top microscope TM3000. For details of the SEM/EDX results we refer to the Supplemental Material file Sup . SEM micrographs taken on cut crystals and other sample pieces revealed the final composition can deviate from the nominal one and that Pt did not dissolve in the same amount in all pieces. In fact for the cut crystal with a nominal Pt content of 5 at.% no Pt was detected. This crystal has a stoichiometric composition with a Pd:Te ratio of 1:2 (the error in these numbers is 1%). Transport, ac susceptibility and heat capacity measurements were carried out on this sample, which we labeled #ptnom5. For the experiments on the 10 at.%Pt concentration we used two crystals. One sample had a composition close to the nominal $x=0.10$ composition Pd0.91Pt0.09Te2. This sample, labeled #ptnom10res, was used for transport experiments only. EDX on the second sample showed a small Te excess and a very small Pt content ($<$ 0.4%). Its composition is Pd0.97Pt<0.004Te2.03. This sample was used for transport, ac susceptibility and heat capacity measurements and it is labeled #ptnom10. We remark that the EDX determined compositions above each yield the average over a large part of the sample surface and are thus representative for the specific sample. The experimental results on the doped samples are compared with previous resistance, ac susceptibility and heat capacity data taken on a crystal with the stoichiometric 1:2 composition to within 0.5% as determined by EDX Leng _et al._ (2017); Salis _et al._ (2021). In the following this sample is labelled #pdte2. Resistance measurements were performed using the standard four point method in a Quantum Design Physical Property Measurement System (PPMS) down to 2.0 K. Data at lower temperatures were collected in a 3-He refrigerator (Heliox, Oxford Instruments) down to 0.3 K using a low frequency (16 Hz) ac-resistance bridge (Linear Research LR700). The ac susceptibility was measured in the Heliox with a custom-made coil set. Data were also taken with the LR700 bridge, operated at a driving field of 0.026 mT. The heat capacity was measured using the dual slope thermal relaxation calorimetry technique Stewart (1983), using a home-built set-up Salis _et al._ (2021), where each data point is the average of four dual slope measurements. The increase in temperature $\Delta T$ in the measurement of the heat capacity is always in between 1$\%$ and 1.6$\%$ of the bath temperature of the particular measurement. In the ac susceptibility and specific heat experiment the dc magnetic field was applied in the $ab$-plane. The demagnetization factor of the crystals is $N\simeq 0.1$, which implies the intermediate state is formed between $(1-N)H_{c}\simeq 0.9H_{c}$ and $H_{c}$ in the case of type-I superconductivity. The resistance and ac susceptibility measurements in field have been carried out by applying the field above $T_{c}$ and subsequently cooling in field, while the specific heat data in field were taken after zero field cooling and then applying the field. ## III Results Figure 1: Temperature dependence of the resistivity of crystals #ptnom5 (red circles), #ptnom10 (blue circles) and #ptnom10res (green circles). The data for #pdte2 (black circles) are taken from Ref. Leng _et al._ (2017). The resistivity of samples #ptnom5, #ptnom10 and #ptnom10res in the temperature range 2-300 K is shown in figure 1, where we have also traced the data for crystal #pdte2 reported in Ref. Leng _et al._ (2017). The curves for #ptnom5 and #pdte2 are very similar with a residual resistivity value, $\rho_{0}$, taken at 2 K, of 0.75 and 0.76 $\mu\Omega$cm, respectively. This is in agreement with both samples having the same stoichiometric 1:2 composition. The residual resistance ratio, RRR= $\rho(300$K$)/\rho_{0}$, amounts to 40 and 30, respectively. For the non-stoichiometric sample #ptnom10 $\rho_{0}$ has increased to 3.6 $\mu\Omega$cm and RRR = 12. The $\rho_{0}$-value of the substituted sample #ptnom10res is considerably higher as expected, and equals 16.3 $\mu\Omega$cm. Its RRR is 3. Figure 2: Resistance as a function of temperature around the superconducting transition for crystal #ptnom5 (panel (b)) and #ptnom10res (panel (c)) in zero field (black curves) and small applied fields, $\mu_{0}H_{a}$, as indicated. The data in panel (a) for #pdte2 are taken from Ref. Leng _et al._ (2017). Figure 3: Ac susceptibility of crystals #ptnom5 (panel (b)) and #ptnom10 (panel (c)) measured in zero field (black curves) and small applied dc fields as indicated. The field is applied in the $ab$-plane. The data of #pdte2 are taken from Ref. Leng _et al._ (2017). Note the ac driving field applied to take the data in panels (b) and (c) is a factor 10 smaller than in panel (a). The resistance as a function of temperature around the superconducting transition in zero field and applied magnetic fields of crystals #ptnom5 and #ptnom10res is depicted in figure 2. Again, the data for #pdte2, shown in panel (a), are taken from Ref. Leng _et al._ (2017). The critical temperature in zero field, $T_{c}(0)$, here defined by the onset of the transition, is 1.87 K and 1.56 K for the stoichiometric samples #ptnom5 and #pdte2, respectively. Surprisingly, the higher $T_{c}$ and RRR for #ptnom5 indicate it has a somewhat higher purity than sample #pdte2. For the substituted sample the superconducting transition shows several steps and $T_{c}$ is lower. It ranges from 1.44 to 1.12 K. In a magnetic field superconductivity is rapidly suppressed. The data in panels (b) and (c) of figure 2 show these crystals also have superconducting resistance paths in fields above the critical field $B_{c}(0)$ determined by ac susceptibility and heat capacity (see below and figure 5). The $B_{c}^{R}(0)$ values that can be deduced are however not as large as the value $B_{c}^{R}(0)\approx$ 0.3 T for $H$ $\|~{}c$ reported for PdTe2 (see figure S6 in the Supplemental Material file of Ref. Leng _et al._ (2017)). In figure 3 we show the in-phase component of the ac susceptibility, $\chi_{ac}^{\prime}$, in arbitrary units measured on crystals #ptnom5 and #ptnom10 in the temperature range 0.3-2.0 K. Again the data are compared with those of #pdte2 (data in S.I. units taken from Ref. Leng _et al._ (2017)). The onset $T_{c}$ values are 1.64 K and 1.85 K for #pdte2 and #ptnom5 and compare well to the values determined above from the resistivity. The onset $T_{c}$ value for #ptnom10 is 1.91 K, but the transition is rather broad (the width is 0.3 K) with a slow decrease below $T_{c}$. The resistance of this sample was only measured in the PPMS down to 2.0 K. The RRR-value of 12 tells us the disorder is enhanced, which is also reflected in the broad transition. The $\chi_{ac}^{\prime}(T)$ data measured in applied magnetic fields for #pdte2 and #ptnom5 show pronounced peaks below $T_{c}$ that are due to the differential paramagnetic effect (DPE). The DPE is due to the positive $dM/dH$ ($M$ is the magnetization) in the intermediate state Hein and Falge (1961). The intermediate phase is due to the sample shape and is present when the demagnetization factor, $N$, is finite. Observation of red a DPE that largely exceeds the Meissner signal can therefore be used as solid proof for type-I superconductivity. Most importantly, the DPE is absent for crystal #ptnom10, which provides the first piece of evidence it is a type-II superconductor. Figure 4: Electronic specific heat, $C_{el}$, of crystals #ptnom5 (panel (b)) and #ptnom10 (panel (c)) measured in zero field (black curves) and small applied dc fields as indicated. The field is directed in the $ab$-plane. The data of #pdte2 are taken from Ref. Salis _et al._ (2021). In figure 4 we show the electronic specific heat, $C_{el}$, of crystals #ptnom5 and #ptnom10 in the temperature range 0.3-2.0 K. The $C_{el}(T)$-curves are obtained by subtracting the phononic contribution from the measured $C$ in the standard way, i.e. by using the relation $C=\gamma T+\beta T^{3}$, where $\gamma$ is the Sommerfeld coefficient and $\beta$ the phononic coefficient. The data are compared with $C_{el}$ of PdTe2 reported in Ref. Salis _et al._ (2021) (panel (a) of figure 4). This PdTe2 crystal was cut from the same batch as the samples studied in Ref. Leng _et al._ (2017) and we also label it #pdte2. The onset $T_{c}$ values of crystals #pdte2 and #ptnom5 are 1.62 K and 1.75 K and compare well to the values determined above. The onset $T_{c}=1.60$ K for #ptnom10 is however lower than the value 1.91 K determined by $\chi_{ac}^{\prime}(T)$. The $\gamma$-values of the three crystals in panel (a), (b) and (c) of figure 4 amount to 4.4, 4.5 and 4.7 mJ/molK2 and the $\beta$-values are 0.7, 1.1 and 1.0 mJ/molK4, respectively. These $\gamma$-values are very similar, which indicates the density of states near the Fermi level is not affected much by doping. The $\beta$-values do show some variation, which is not correlated with the amount of disorder, and likely related to an experimental uncertainty because of the small temperature interval in which $\beta$ is obtained. To examine the strength of the electron-phonon coupling, the step size $\Delta C|_{T_{c}}$ is analysed using the BCS relation $\Delta C|_{T_{c}}/\gamma T_{c}=1.43$, where $T_{c}$ is the superconducting transition temperature, here taken as the onset of superconductivity. For crystal #pdte2 a ratio $\Delta C|_{T_{c}}/\gamma T_{c}=1.42$ is found Salis _et al._ (2021), which is close to the textbook value of 1.43 for a weakly coupled BCS superconductor. For crystal #ptnom5 a ratio of 1.41 is found, which presents a minute change from the textbook value. However, for crystal #ptnom10 we determine a ratio of 1.48, suggesting that superconductivity is slightly more than weakly coupled. Next we discuss the electronic specific heat measured in applied magnetic fields (figure 4). Distinguishing between type-I and type-II superconductivity via heat capacity can be achieved by observing the presence or absence of latent heat. The extra energy necessary to facilitate a first order phase transition is reflected in the heat capacity as an increased value of $C$ at the transition. A type-I superconductor has a second order phase transition in zero field, but a first order one in field. While for a type-II superconductor the transition remains second order in an applied field. The excess $C_{el}$ above the standard BCS heat capacity in panel (a) provided solid thermodynamic evidence PdTe2 is a type-I superconductor Salis _et al._ (2021). Surprisingly, for crystal #ptnom5 (panel b) the excess specific heat becomes more pronounced as illustrated by the sharp peaks below $T_{c}(B)$. Thus the contribution of the latent heat to $C_{el}$ is much larger, which indicates the transition has a stronger first order character than observed for crystal #pdte2. On the other hand, for crystal #ptnom10 the data in panel (c) show latent heat is absent, which provides the second piece of evidence of type-II superconductivity, in-line with the $\chi_{ac}^{\prime}$-data. Finally, we trace the temperature variation of the critical field, $B_{c}(T)$, extracted from the ac susceptibility (figure 3) and specific heat data (figure 4). The $B-T$ phase diagram is reported in figure 5. For crystals #pdte2 and #ptnom5 we identify $T_{c}(B)$ by the onset in $C_{el}$ and the onset of the DPE in $\chi_{ac}^{\prime}(T)$. $B_{c}(T)$ follows the standard BCS quadratic temperature variation $B_{c}(T)=B_{c}(0)[1-(T/T_{c})^{2}]$, with $B_{c}(0)=14.2$ mT and $T_{c}$ = 1.63 K for #ptpde2 Leng _et al._ (2017); Salis _et al._ (2021), and $B_{c}(0)=15.9$ mT and $T_{c}$ = 1.77 K for #ptnom5. For crystal #ptnom10 the transition in $\chi_{ac}^{\prime}(T)$ is rather broad. Here we identify $T_{c}$ by the onset temperature in $C_{el}$, which corresponds to the temperature where the magnetic transition is complete in $\chi_{ac}^{\prime}(T)$. The $B-T$ phase-line provides further evidence for type-II superconductivity. It compares well to the Werthamer-Helfand-Hohenberg (WHH) model curve Werthamer _et al._ (1966) for an orbital-limited weak- coupling spin-singlet superconductor with an upper critical field $B_{c2}(0)=21.8$ mT. Figure 5: Critical field $B_{c}(T)$ of crystal #pdte2 and #ptnom5 and upper critical field $B_{c2}(T)$ of crystal #ptnom10 extracted from the specific heat (closed squares) and ac susceptibility (closed circles) data. ## IV Discussion From the sample preparation side our goal was to prepare Pd1-xPtxTe2 crystals with $x=0.05$ and $x=0.10$. The SEM/EDX micrographs showed that Pt did not dissolve as expected in these crystals and that the single-crystalline batches are inhomogeneous. Crystals cut from the nominal $x=0.05$ batch appeared to be undoped and have the 1:2 stoichiometry. From the nominal $x=0.10$ batch we managed to obtain a crystal with $x=0.09$, and a non-stoichiometric crystal Pd0.97Pt<0.004Te2.03. Specific heat and ac-susceptibility measurements on this last crystal #ptnom10 demonstrated we could make a doping-induced transition to type-II superconductivity. To observe type-II superconductivity the disorder should be large enough such that the threshold $\kappa=1/\sqrt{2}$ can be overcome. The effect of controlled non-magnetic disorder on the normal and superconducting properties of PdTe2 was recently studied by electron irradiation by Timmons et al. Timmons _et al._ (2020). The residual resistivity was found to increase from a pristine crystal value of 0.6 $\mu\Omega$cm to 2.4 $\mu\Omega$cm for a irradiation dose of 2.4 C/cm2, while at the same time $T_{c}$ decreased from 1.76 K to 1.65 K as identified by reaching the zero resistance state $R=0$. Assuming a linear relation between $\rho_{0}$ and $T_{c}$, $T_{c}$ decreases at a rate of 0.046 K/$\mu\Omega$cm. With this rate we estimate for crystal #ptnom10res ($\Delta\rho_{0}=15.5~{}\mu\Omega$cm) $T_{c}=0.9$ K, which compares favorably to the measured $T_{c}=1.1$ K ($R=0$), given the crude approximation. In this electron irradiation work no discussion was made whether disorder is strong enough to induce type-II behavior. For crystal #ptnom10 the coherence length $\xi$ can be calculated from the relation $B_{c2}(0)=\Phi_{0}/2\pi\xi^{2}$, where $\Phi_{0}$ is the flux quantum. From figure 5 we determine $B_{c2}(0)=21.8$ mT and obtain $\xi=123$ nm. The coherence length can be related to the electron mean free path, $l_{e}$, via Pippard’s relation $1/\xi=1/\xi_{0}+1/l_{e}$, where $\xi_{0}$ is the intrinsic coherence length given by the BCS value Tinkham (1996). With $\xi_{0}=1.8~{}\mu$m Salis _et al._ (2018) and $\xi=~{}123$ nm we obtain $\l_{e}=132$ nm. As expected, this value is reduced compared to $\l_{e}=531$ nm calculated from the residual resistivity value $\rho_{0}=0.76~{}\mu\Omega$cm Salis _et al._ (2018) of nominally pure PdTe2. Reversely, using the experimental value $\rho_{0}=3.6~{}\mu\Omega$cm (figure 1) we calculate $\l_{e}=112$ nm for crystal #ptnom10, which is close to the value $\l_{e}=132$ nm derived from Pippard’s relation. Next we calculate $\kappa=\lambda/\xi$ of crystal #ptnom10. In their controlled disorder study Timmons et al. Timmons _et al._ (2020) measured the penetration depth and found that upon increasing the disorder $\lambda$ stays nearly constant Timmons _et al._ (2020) at a value of 220 nm. This is in-line with the minute change in the $\gamma$-value reported above. With $\xi=123$ nm we calculate $\kappa=1.8$, which is in agreement with superconductivity being of the second kind. For crystals #pdte2 and #ptnom5 we calculate $\kappa\simeq$ 0.5-0.6 Leng _et al._ (2017). Here $\xi\simeq$ 440-370 nm is estimated from the GL relation $\xi=\Phi_{0}/(2\sqrt{2}\pi B_{c}\lambda_{L})$ Tinkham (1996), where $\lambda_{L}\propto(m_{e}/n_{s})^{1/2}$ is the London penetration depth with $m_{e}$ the effective electron mass and $n_{s}$ the superfluid density. Another way to provide an estimate of $\kappa$ of crystal #ptnom10 is from the GL relation $\kappa=B_{c2}/\sqrt{2}B_{c}$. The thermodynamic critical field, $B_{c}$, can be determined from the specific heat by the relation $\Delta C|_{T_{c}}=4B_{c}(0)^{2}/\mu_{0}T_{c}$ Poole _et al._ (2007), where $C$ is in units of J/m3. From $\Delta C|_{T_{c}}$ in figure 4 (panel (c)) we calculate $B_{c}(0)=11.1$ mT. We remark this value is close to the calculated value $B_{c}(0)=12.6$ mT reported for PdTe2 Leng _et al._ (2017). Using $B_{c}(0)=11.1$ mT and $B_{c2}=21.8$ mT in the expression above, we calculate $\kappa=1.4$, which is similar to the value of 1.8 directly estimated from the ratio $\lambda/\xi$. We remark that for Type-I superconductivity $B_{c}(0)$ can also be obtained from the latent heat with help of the Clausius-Clapeyron relation. We calculate $B_{c}(0)$ = 11.2 mT and 11.1 mT for #pdte2 and #ptnom5, respectively, in good agreement with the values obtained from $\Delta C|_{T_{c}}$ in zero field Sup . Our results are of relevance for the observation of a mixed type-I and type-II superconducting state in PdTe2 probed by surface sensitive techniques Das _et al._ (2018); Sirohi _et al._ (2019); Le _et al._ (2019). Our doping study shows that nominal pure PdTe2 crystals can already be close to the type-I/II border. Using the value $\lambda=230$ nm Timmons _et al._ (2020), we calculate $\xi=310$ nm at the threshold value $\kappa=1/\sqrt{2}$. This implies $l_{e}$ should be smaller than 375 nm for type-II superconductivity, or $\rho_{0}>1.4~{}\mu\Omega$cm. From the resistivity graph reported in Ref. Das _et al._ (2018) we deduce $\rho_{0}\simeq 1~{}\mu\Omega$cm, which indeed is not far from the type-I/II border. Thus it is plausible inhomogeneities give rise to the mixed phase observation reported in Refs. Das _et al._ (2018); Sirohi _et al._ (2019); Le _et al._ (2019). An unsolved aspect of superconductivity in PdTe2 is the observation of surface superconductivity detected in the screening signal $\chi_{ac}^{\prime}(T)$ measured in small applied dc fields Leng _et al._ (2017, 2019b). The extracted surface critical field $B_{c}^{S}(0)=34.9$ mT exceeds the value predicted by the Saint-James $-$ de Gennes model Saint-James and de Gennes (1963) $B_{c3}=2.39\times\kappa B_{c}=16.3$ mT. Recently, the GL model at the superconducting-insulator boundary was revisited Samoilenka and Babaev (2021) and it was shown that $T_{c}$ and the third critical field $B_{c3}$ can be enhanced to exceed the Saint-James $-$ de Gennes value, which is worthy to explore further. On the other hand, it is tempting to attribute the surface superconductivity in PdTe2 to superconductivity of the topological surface state detected by ARPES Bahramy _et al._ (2018); Yan _et al._ (2015); Fei _et al._ (2017); Noh _et al._ (2017); Clark _et al._ (2018). We remark that the $\chi_{ac}^{\prime}(T)$ data for the doped crystals, reported in figure 3 panel (b) and (c), also show superconducting screening signals above the $B_{c}(0)$ and $B_{c2}(0)$-values reported in figure 5. Likewise, the resistance traces in figure 2 reveal $B_{c}^{R}(0)$ is similarly enhanced. These screening signals of enhanced superconductivity are however not as pronounced as reported for PdTe2 in Ref. Leng _et al._ (2017). Nonetheless, the robustness of superconducting screening signals above $B_{c}(0)$ or $B_{c2}(0)$ upon doping, as well as under high pressure Leng _et al._ (2019b), calls for further experiments. ## V Conclusion The Dirac semimetal PdTe2 is a type-I superconductor with $T_{c}=1.7$ K. We have carried out a doping study directed to intentionally increase the disorder and induce type-II superconductivity. Two single-crystalline batches Pd1-xPtxTe2 have been prepared with nominal doping concentrations $x=0.05$ and $x=0.10$. Sample characterization by energy dispersive x-ray spectroscopy (EDX) on small crystals cut from the batches revealed that Pt did not dissolve homogeneously in the crystals. In fact the nominal $x=0.05$ crystal appeared to be undoped and have the stoichiometric 1:2 composition. From the nominal $x=0.10$ batch we obtained a small single crystal with $x=0.09$, as well as a crystal with the non stoichiometric composition Pd0.97Pt<0.004Te2.03. The presence of the differential paramagnetic effect in the ac susceptibility and latent heat in the heat capacity demonstrate the nominal $x=0.05$ crystal is a type-I superconductor, just like PdTe2. The absence of these effects for Pd0.97Pt<0.004Te2.03 revealed it is a type-II superconductor with an upper critical field $B_{c2}=21.8$ mT. The analysis of $B_{c2}$ and resistance data using Pippard’s model convincingly show PdTe2 can be turned into a superconductor of the second kind when the residual resistivity $\rho_{0}>1.4~{}\mu\Omega$cm. Acknowledgement: This work is part of the Projectruimte programme with project number 680-91-109, which is financed by the Dutch Research Council (NWO). ## References * Soluyanov _et al._ (2015) A. A. Soluyanov, D. Gresch, Z. Wang, Q. Wu, M. Troyer, X. Dai, and B. Bernevig, Nature 257, 495 (2015). * Huang _et al._ (2016) H. Huang, S. Zhou, and W. Duan, Phys. Rev. B 94, 121117 (2016). * Yan _et al._ (2017) M. Yan, H. Huang, K. Zhang, E. Wang, W. Yao, K. Deng, G. Wan, H. Zhang, M. Arita, H. Yang, Z. Sun, H. Yao, Y. Wu, S. Fan, W. Duan, and S. Zhou, Nature Comm. 8, 257 (2017). * Bahramy _et al._ (2018) M. S. Bahramy, O. J. Clark, B.-J. Yang, J. Feng, L. Bawden, J. M. Riley, I. Markovic, F. Mazzola, V. Sunko, D. Biswas, S. P. Cooil, M. Jorge, J. W. Wells, M. Leandersson, T. Balasubramanian, J. Fujii, I. Vobornik, J. E. Rault, T. K. Kim, M. Hoesch, K. Okawa, M. Asakawa, T. Sasagawa, T. Eknapakul, W. Meevasana, and P. D. C. King, Nature Mat. 17, 21 (2018). * Guggenheim _et al._ (1961) J. Guggenheim, F. Hulliger, and J. Müller, Helv. Phys. Acta 34, 408 (1961). * Yan _et al._ (2015) L. Yan, Z. Jian-Zhou, Y. Li, L. Cheng-Tian, L. Ai-Ji, H. Cheng, D. Ying, X. Yu, H. Shao-Long, Z. Lin, L. Guo-Dong, D. Xiao-Li, Z. Jun, C. Chuang-Tian, X. Zu-Yan, W. Hong-Ming, D. Xi, F. Zhong, and Z. Xing-Jiang, Chin. Phys. Lett. 32, 067303 (2015). * Fei _et al._ (2017) F. Fei, X. Bo, R. Wang, B. Wu, J. Jiang, D. Fu, M. Gao, H. Zheng, Y. Chen, X. Wang, H. Bu, F. Song, X. Wan, B. Wang, and G. Wang, Phys. Rev. B 96, 041201 (2017). * Noh _et al._ (2017) H.-J. Noh, J. Jeong, E.-J. Cho, K. Kim, B. I. Min, and B.-G. Park, Phys. Rev. Lett. 119, 016401 (2017). * Clark _et al._ (2018) O. J. Clark, M. J. Neat, K. Okawa, L. Bawden, I. Marković, F. Mazzola, J. Feng, V. Sunko, J. M. Riley, W. Meevasana, J. Fujii, I. Vobornik, T. K. Kim, M. Hoesch, T. Sasagawa, P. Wahl, M. S. Bahramy, and P. D. C. King, Phys. Rev. Lett. 120, 156401 (2018). * Rosenstein _et al._ (2018) B. Rosenstein, B. Y. Shapiro, D. Li, and I. Shapiro, Phys. Rev. B 97, 144510 (2018). * Leng _et al._ (2017) H. Leng, C. Paulsen, Y. K. Huang, and A. de Visser, Phys. Rev. B 96, 220506 (2017). * Salis _et al._ (2021) M. V. Salis, Y. K. Huang, and A. de Visser, Phys. Rev. B 103, 104502 (2021). * Shapiro _et al._ (2018) B. Y. Shapiro, I. Shapiro, D. Li, and B. Rosenstein, Journal of Physics: Condensed Matter 30, 335403 (2018). * Saint-James and de Gennes (1963) D. Saint-James and P. G. de Gennes, Phys. Lett. 7, 306 (1963). * Amit and Singh (2018) Amit and Y. Singh, Phys. Rev. B 97, 054515 (2018). * Teknowijoyo _et al._ (2018) S. Teknowijoyo, N. H. Jo, M. S. Scheurer, M. A. Tanatar, K. Cho, S. L. Bud’ko, P. P. Orth, P. C. Canfield, and R. Prozorov, Phys. Rev. B 98, 024508 (2018). * Salis _et al._ (2018) M. V. Salis, P. Rodière, H. Leng, Y. K. Huang, and A. de Visser, Journal of Physics: Condensed Matter 30, 505602 (2018). * Das _et al._ (2018) S. Das, Amit, A. Sirohi, L. Yadav, S. Gayen, Y. Singh, and G. Sheet, Phys. Rev. B 97, 014523 (2018). * Sirohi _et al._ (2019) A. Sirohi, S. Das, P. Adhikary, R. R. Chowdhury, A. Vashist, Y. Singh, S. Gayen, T. Das, and G. Sheet, Journal of Physics: Condensed Matter 31, 085701 (2019). * Voerman _et al._ (2019) J. A. Voerman, J. C. de Boer, T. Hashimoto, Y. Huang, C. Li, and A. Brinkman, Phys. Rev. B 99, 014510 (2019). * Kim _et al._ (2018) K. Kim, S. Kim, J. S. Kim, H. Kim, J.-H. Park, and B. I. Min, Phys. Rev. B 97, 165102 (2018). * van Heumen _et al._ (2019) E. van Heumen, M. Berben, L. Neubrand, and Y. Huang, Phys. Rev. Materials 3, 114202 (2019). * Le _et al._ (2019) T. Le, L. Yin, Z. Feng, Q. Huang, L. Che, J. Li, Y. Shi, and X. Lu, Phys. Rev. B 99, 180504 (2019). * Leng _et al._ (2019a) H. Leng, J.-C. Orain, A. Amato, Y. K. Huang, and A. de Visser, Phys. Rev. B 100, 224501 (2019a). * Garcia-Campos _et al._ (2021) P. Garcia-Campos, Y. K. Huang, A. de Visser, and K. Hasselbach, Phys. Rev. B 103, 104510 (2021). * Furue _et al._ (2021) Y. Furue, T. Fujino, M. V. Salis, H. Leng, F. Ishikawa, T. Naka, S. Nakano, Y. Huang, A. de Visser, and A. Ohmura, Phys. Rev. B 104, 144510 (2021). * Leng _et al._ (2019b) H. Leng, A. Ohmura, L. N. Anh, F. Ishikawa, T. Naka, Y. K. Huang, and A. de Visser, Journal of Physics: Condensed Matter 32, 025603 (2019b). * Kudo _et al._ (2016) K. Kudo, H. Ishii, and M. Nohara, Phys. Rev. B 93, 140505 (2016). * Ryu (2015) G. Ryu, Journal of Superconductivity and Novel Magnetism 28, 3275 (2015). * Hooda and Yadav (2018) M. K. Hooda and C. S. Yadav, Europhysics Letters 121, 17001 (2018). * Vasdev _et al._ (2019) A. Vasdev, A. Sirohi, M. K. Hooda, C. S. Yadav, and G. Sheet, Journal of Physics: Condensed Matter 32, 125701 (2019). * Lyons _et al._ (1976) A. Lyons, D. Schleich, and A. Wold, Mat. Res. Bull. 11, 1155 (1976). * (33) See Supplemental Material https:… . * Stewart (1983) G. R. Stewart, Review of Scientific Instruments 54, 1 (1983). * Hein and Falge (1961) R. A. Hein and R. L. Falge, Phys. Rev. 123, 407 (1961). * Werthamer _et al._ (1966) N. R. Werthamer, E. Helfand, and P. C. Hohenberg, Phys. Rev. 147, 295 (1966). * Timmons _et al._ (2020) E. I. Timmons, S. Teknowijoyo, M. Kończykowski, O. Cavani, M. A. Tanatar, S. Ghimire, K. Cho, Y. Lee, L. Ke, N. H. Jo, S. L. Bud’ko, P. C. Canfield, P. P. Orth, M. S. Scheurer, and R. Prozorov, Phys. Rev. Research 2, 023140 (2020). * Tinkham (1996) M. Tinkham, _Introduction to Superconductivity_ (McGraw-Hill Inc., New York, 1996). * Poole _et al._ (2007) C. P. Poole, H. A. Farach, R. J. Creswick, and R. Prozorov, _Superconductivity_ , 2nd ed. (Elsevier, Amsterdam, 2007). * Samoilenka and Babaev (2021) A. Samoilenka and E. Babaev, Phys. Rev. B 103, 224516 (2021). Supplemental Material: Disorder induced transition from type-I to type-II superconductivity in the Dirac semimetal PdTe2 Content 1\. SEM/EDX mapping experimental and results Fig.S1 SEM/EDX mapping of crystal #ptnom5 Fig.S2 SEM/EDX mapping of crystal #ptnom10 Fig.S3 SEM/EDX mapping of crystal #ptnom10res 2\. Latent heat and Clausius-Clapeyron relation 1\. SEM/EDX mapping: experimental Scanning electron microscopy with energy dispersive x-ray spectroscopy (SEM/EDX) was carried out with help of a Hitachi table top microscope TM3000. The acceleration voltage in all measurements is 15 kV. Of each single- crystalline boule prepared with a certain nominal Pt content small thin crystals were isolated with typical size 2 $\times$ 3 mm2. On each of these crystals we have investigated the composition by EDX in several areas of typically 200 $\times$ 200 $\mu$m2. Overall the composition in the sleected and measured crystals was found to be homogenous. The table below gives the labels of the doped crystals, the nominal Pt content and the EDX determined composition. Crystal | Nominal Pt cotent | EDX compostion ---|---|--- #ptnom5 | 5 at. $\%$ | PdTe2 #ptnom10 | 10 at. $\%$ | Pd0.97Pt<0.004Te2.03 #ptnom10res | 10 at.$\%$ | Pd0.91Pt0.09Te2 Typical SEM/EDX results for crystals #ptnom5, #ptnom10 and #ptnom10res are given in Fig. S1, Fig. S2 and Fig. S3, respectively. Shown are: Top panel: The SEM spectrum in cps/eV (counts per second per electron-volt) with Pd, Pt and Te peaks labelled. Middle panel: Table with quantitative results of the composition analysis. Lower left panel: SEM picture of the crystal with scanned area for the composition analysis indicated. Lower right panel: Pd, Pt and Te element distribution in the scanned area. Figure S1 SEM/EDX mapping of crystal #ptnom5 Figure S2 SEM/EDX mapping of crystal #ptnom10 Figure S3 SEM/EDX mapping of crystal #ptnom10res 2\. Latent heat and Clausius-Clapeyron relation As mentioned in the Discussion section in the manuscript an estimate for the thermodynamic critical field $B_{c}(0)$ can be obtained by evaluating the relation $\Delta C=4B_{c}(0)^{2}/\mu_{0}T_{c}$, where $\Delta C$ is the step at $T_{c}$ in the specific heat at $B=0$. Alternatively, for a type-I superconductor in a magnetic field $B_{c}(0)$ can be calculated from the latent heat, $L$. Using the Clausius-Clapeyron equation the entropy change is given by $\Delta S=L(T)/T=-2\mu_{0}B_{c}(T)\frac{dB_{c}(T)}{dT}$, with $B_{c}(T)=B_{c}(0)(1-(T/T_{c})^{2})$. $\Delta S$ was calculated by graphically integrating the $C/T$ data in field and subtracting the data taken at $B=0$. This procedure is carried out for all applied field values, and from $\Delta S(B)$ we obtain $B_{c}(0)$ as a fit parameter. For crystals #pdte2 and #ptnom5, that exhibit type-I superconductivity, we obtain $B_{c}(0)=11.2$ mT and 11.1 mT, respectively. We remark these values are close to the $B_{c}(0)$ values evaluated at $B=0$. On the other hand, the $B_{c}(0)$-values derived from $\Delta C$ and $\Delta S$ are smaller than the value derived directly from the experiment. In the table below we compare the $B_{c}(0)$ values obtained in different ways. Crystal | $B_{c}(0)$ from $\Delta C$ | $B_{c}(0)$ from $\Delta S$ | $B_{c}(0)$ experiment ---|---|---|--- #pdte2 | 10.9 mT | 11.2 mT | 14.2 mT #ptnom5 | 11.8 mT | 11.1 mT | 15.9 mT #ptnom10 | 11.1 mT | - | -
$\mathcal{S}:\mathbb{T}_{p}\rightarrow\mathbb{Z}[p^{-1}][\mathrm{X}^{\ast}(\hat{\mathrm{T}})]^{\mathrm{W}_{\mathbf{G}}}$ where $\mathrm{W}_{\mathbf{G}}$ is the absolute Weyl group of $\mathbf{G}$. We fix a datum $\mathbf{T}\subset\mathbf{B}\subset\mathbf{G}$ of the diagonal torus contained in the upper triangular Borel subgroup as in the global case. This determines a positive Weyl chamber $\mathrm{P}^{+}$ in $\mathrm{X}_{\ast}(\mathrm{T})\cong\mathrm{X}^{\ast}(\hat{\mathrm{T}})$. An element $\nu\in\mathrm{X}^{\ast}(\hat{\mathrm{T}})$ determines an irreducible representation $V_{\nu}$ of the dual group $\hat{\mathrm{G}}=\mathrm{GSpin}(5)$ whose character is denoted by $\chi_{\nu}\in\mathrm{R}(\hat{\mathrm{G}})=\mathbb{Z}[\mathrm{X}^{\ast}(\hat{\mathrm{T}})]^{\mathrm{W}_{\mathbf{G}}}$. It is well-known that $\mathbb{T}_{p}$ is isomorphic to $\mathbb{Z}[\mathrm{T}^{\pm}_{p,0},\mathrm{T}_{p,1},\mathrm{T}_{p,2}]$ where $\begin{split}&\mathrm{T}_{p,0}=c_{\nu_{0}}=\mathrm{char}(\mathbf{G}(\mathbb{Z}_{p})\begin{pmatrix}p&&&\\\ &p&&\\\ &&p&\\\ &&&p\\\ \end{pmatrix}\mathbf{G}(\mathbb{Z}_{p}))\text{ for }\nu_{0}=(1,1,1,1)\\\ &\mathrm{T}_{p,2}=c_{\nu_{2}}=\mathrm{char}(\mathbf{G}(\mathbb{Z}_{p})\begin{pmatrix}p&&&\\\ &p&&\\\ &&1&\\\ &&&1\\\ \end{pmatrix}\mathbf{G}(\mathbb{Z}_{p}))\text{ for }\nu_{2}=(1,1,0,0)\\\ &\mathrm{T}_{p,1}=c_{\nu_{1}}=\mathrm{char}(\mathbf{G}(\mathbb{Z}_{p})\begin{pmatrix}p^{2}&&&\\\ &p&&\\\ &&p&\\\ &&&1\\\ \end{pmatrix}\mathbf{G}(\mathbb{Z}_{p}))\text{ for }\nu_{1}=(2,1,1,0)\\\ \end{split}$ and where $\mathrm{char}(\cdot)$ is the characteristic function. The following spherical Hecke operator will appear naturally in later computations $\mathrm{T}_{p^{2},2}=c_{2\nu_{2}}=\mathrm{char}(\mathbf{G}(\mathbb{Z}_{p})\begin{pmatrix}p^{2}&&&\\\ &p^{2}&&\\\ &&1&\\\ &&&1\\\ \end{pmatrix}\mathbf{G}(\mathbb{Z}_{p}))\text{ for }2\nu_{2}=(2,2,0,0).$ We will also need the following two Hecke operators in $\mathbb{Z}[\mathrm{K}_{\\{2\\}}\backslash\mathbf{G}(\mathbb{Q}_{p})/\mathrm{K}_{\\{0\\}}]$ $\displaystyle\mathrm{T}_{02}=\mathrm{char}(\mathrm{K}_{\\{2\\}}\begin{pmatrix}p&&&\\\ &p&&\\\ &&1&\\\ &&&1\\\ \end{pmatrix}\mathrm{K}_{\\{0\\}})$ $\displaystyle\mathrm{T}_{20}=\mathrm{char}(\mathrm{K}_{\\{0\\}}\begin{pmatrix}p^{-1}&&&\\\ &p^{-1}&&\\\ &&1&\\\ &&&1\\\ \end{pmatrix}\mathrm{K}_{\\{2\\}}).$ In the computations below, we will follow mostly the notations and conventions in [Gross] and use results from there freely. We have the following useful formula [Gross, 3.12] $p^{\langle\nu,\rho\rangle}\chi_{\nu}=\mathcal{S}(c_{\nu})+\sum_{\xi<\nu}d_{\nu}(\xi)\mathcal{S}(c_{\xi}).$ The integers $d_{\nu}(\xi)$ are given by $\mathrm{P}_{\xi,\nu}(p)$ for the Kazhdan-Lusztig polynomial $\mathrm{P}_{\xi,\nu}$ evaluated at $p$ and $2\rho$ is the sum of the positive roots. Since the weight $\nu_{2}$ is minuscule, we immediately have $p^{\frac{3}{2}}\chi_{\nu_{2}}=\mathcal{S}(c_{\nu_{2}})=\mathcal{S}(\mathrm{T}_{p,2}).$ On the other hand for the non-minuscule weight $\nu_{1}$, we have $p^{2}\chi_{\nu_{1}}=\mathcal{S}(c_{\nu_{1}})+\mathcal{S}(c_{0})=\mathcal{S}(\mathrm{T}_{p,1})+1.$ The following identity is the main result of this subsection. It can perhaps be proved by a massive double coset computation whereas our proof uses crucially the Satake isomorphism to simplify the computation. This identity seems not to exist already in the literature and therefore we give full details which we think are of independent interest. ###### Proposition 6.9. The following identity holds in $\mathbb{T}_{p}$ $\mathrm{T}_{p^{2},2}=\mathrm{T}^{2}_{p,2}-(p+1)\mathrm{T}_{p,1}-(p+1)(p^{2}+1).$ ###### Proof. The representation $V_{\nu_{1}}$ is the $5$-dimensional orthogonal representation and $V_{\nu_{2}}$ is the standard representation. Therefore $\wedge^{2}V_{\nu_{2}}=V_{\nu_{1}}\oplus\mathbb{C}$. We also have $\wedge^{2}V_{\nu_{1}}=V_{2\nu_{2}}$ is the adjoint representation of dimension $10$. Thus we have $\displaystyle V^{\otimes 2}_{\nu_{2}}$ $\displaystyle=V_{2\nu_{2}}+\wedge^{2}V_{\nu_{2}}$ $\displaystyle=V_{2\nu_{2}}+V_{\nu_{1}}+\mathbb{C}$ Taking characters, we have $\chi^{2}_{\nu_{2}}=\chi_{2\nu_{2}}+\chi_{\nu_{1}}+1$ and hence $p^{3}\chi^{2}_{\nu_{2}}=p^{3}\chi_{2\nu_{2}}+p^{3}\chi_{\nu_{1}}+p^{3}$. Therefore $\displaystyle p^{3}\chi_{2\nu_{2}}$ $\displaystyle=p^{3}\chi^{2}_{\nu_{2}}-p^{3}\chi_{\nu_{1}}-p^{3}$ $\displaystyle=\mathcal{S}(\mathrm{T}^{2}_{p,2})-p\mathcal{S}(\mathrm{T}_{p,1})-(p+p^{3}).$ On the other hand, we also have $\displaystyle p^{3}\chi_{2\nu_{2}}$ $\displaystyle=\mathcal{S}(c_{2\nu_{2}})+d_{2\nu_{2}}(\nu_{1})S(c_{\nu_{1}})+d_{2\nu_{2}}(0)$ $\displaystyle=\mathcal{S}(\mathrm{T}_{p^{2},2})+d_{2\nu_{2}}(\nu_{1})S(\mathrm{T}_{p,1})+d_{2\nu_{2}}(0).$ * • $d_{2\nu_{2}}(\nu_{1})=1$; This follows from the fact that $\mathrm{P}_{\nu_{1},2\nu_{2}}$ has degree strictly less than $\langle 2\nu_{2}-\nu_{1},\rho\rangle=1$ and the constant coefficient is $1$. * • $d_{2\nu_{2}}(0)=1+p^{2}$; This follows from the fact that $2\nu_{2}$ is the highest weight of the adjoint representation, which implies that $\mathrm{P}_{0,2\nu_{2}}(p)=\sum_{i=1}p^{m_{i}-1}$ where $m_{i}$ are the exponents of the Weyl group of type $\mathrm{C}_{2}$ which are $1$ and $3$, see [Gross, 4.6]. Equating the two expressions of $p^{3}\chi_{2\nu_{2}}$ gives the desired result. ∎ ### 6.2. Computing the level raising matrix We now calculate the entries of the level raising matrix $\mathcal{T}_{\mathrm{lr}}$ term by term. The main result is the following proposition. ###### Proposition 6.10. The level raising matrix $\mathcal{T}_{\mathrm{lr}}$ is given by $\mathcal{T}_{\mathrm{lr}}=-2\begin{pmatrix}&\mathrm{T}^{-1}_{p,0}\mathrm{T}_{p,1}+(p+1)(p^{2}+1)&(p+1)\mathrm{T}_{02}\\\ &(p+1)\mathrm{T}_{20}&\mathrm{T}^{-1}_{p,0}\mathrm{T}_{p,1}+(p+1)(p^{2}+1)\\\ \end{pmatrix}.$ ###### Proof. To begin the proof, note that we can identify both $\bigoplus\limits_{\sigma\in\mathrm{Z}_{\\{1\\}}(\overline{\mathrm{B}})}\mathrm{H}^{3}_{\\{\sigma\\}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}(1)))$ and $\bigoplus\limits_{\sigma\in\mathrm{Z}_{\\{1\\}}(\overline{\mathrm{B}})}\mathrm{R}^{3}\Phi_{\sigma}(\mathcal{O}_{\lambda}(2))$ with $\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{Pa}}(\overline{\mathrm{B}})]$ ignoring the Galois actions. The level raising diagram (7.2) reduces to (6.11) ${\mathcal{O}_{\lambda}[\mathrm{Z}_{\\{0\\}}(\overline{\mathrm{B}})]}$${\mathcal{O}_{\lambda}[\mathrm{Z}_{\\{2\\}}(\overline{\mathrm{B}})]}$${\mathrm{H}^{2}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(1))}$${\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{Pa}}(\overline{\mathrm{B}})]}$${\mathrm{H}^{4}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(2))}$${\mathcal{O}_{\lambda}[\mathrm{Z}_{\\{0\\}}(\overline{\mathrm{B}})]}$${\mathcal{O}_{\lambda}[\mathrm{Z}_{\\{2\\}}(\overline{\mathrm{B}})]}$$\scriptstyle{\mathrm{inc}^{\\{0\\}}_{!}}$$\scriptstyle{\mathrm{inc}^{\\{2\\}}_{!}}$$\scriptstyle{\beta}$$\scriptstyle{\alpha}$$\scriptstyle{\mathrm{inc}^{\ast}_{\\{0\\}}}$$\scriptstyle{\mathrm{inc}^{\ast}_{\\{2\\}}}$ By Proposition 5.8, the map $\alpha$ can be realized as a Gysin map and $\beta$ can be realized as a restriction map. The situation is similar to that of [LTXZZ, Proposition 5.8.8]. We proceed to analyze each entry. #### 6.2.1. $\mathcal{T}_{\mathrm{lr},\\{00\\}}$ Since $\mathrm{Z}_{\\{i\\}}(\overline{\mathrm{B}})$ corresponds to the set of vertex lattices of type $i$ for $i\in\\{0,1,2\\}$. Therefore to understand the map $\mathcal{T}_{\mathrm{lr},\\{00\\}}=\mathrm{inc}^{\ast}_{\\{0\\}}\circ\alpha\circ\mathrm{N}^{-1}_{\Sigma}\circ\beta\circ\mathrm{inc}^{\\{0\\}}_{!}$, we need to understand the possible relative position of vertex lattices $\Lambda_{0}$, $\Lambda_{\mathrm{Pa}}$ and $\Lambda^{\prime}_{0}$ of type $0$, $1$ and $0$. It not hard to see that the only possible situation is that $\Lambda_{\mathrm{Pa}}\subset^{1}\Lambda_{0}$ and $\Lambda_{\mathrm{Pa}}\subset^{1}\Lambda^{\prime}_{0}$. There are two possible cases. Case $(1)$: $\Lambda_{0}=\Lambda^{\prime}_{0}$. Then we have $p\Lambda^{\vee}_{0}\subset^{1}p\Lambda^{\vee}_{\mathrm{Pa}}\subset^{2}\Lambda_{\mathrm{Pa}}\subset^{1}\Lambda_{0}$. Giving such a $\Lambda_{\mathrm{Pa}}$ is equivalent to giving a complete flag in $\Lambda_{0}/p\Lambda_{0}$. The number of such flags is given by $[\mathbf{G}(\mathbb{Z}_{p}):\mathrm{Kl}]=(p+1)(p^{2}+1)$ where $\mathrm{Kl}$ is the Klingen parahoric. Case $(2)$: $\Lambda_{0}\neq\Lambda^{\prime}_{0}$. Then we have $\Lambda_{0}\cap\Lambda^{\prime}_{0}=\Lambda_{\mathrm{Pa}}$ and the relative position between $\Lambda_{0}$ and $\Lambda^{\prime}_{0}$ is given by $\mathrm{K}_{\\{0\\}}\begin{pmatrix}p&&&\\\ &1&&\\\ &&1&\\\ &&&p^{-1}\\\ \end{pmatrix}\mathrm{K}_{\\{0\\}}.$ This operator is the same as $\mathrm{T}^{-1}_{p,0}\mathrm{T}_{p,1}$ by identifying $\mathrm{K}_{\\{0\\}}$ with the hyperspecial group $\mathbf{G}(\mathbb{Z}_{p})$. Therefore the entry $\mathcal{T}_{\mathrm{lr},\\{00\\}}$ is given by $\mathcal{T}_{\mathrm{lr},\\{00\\}}=-2(\mathrm{T}^{-1}_{p,0}\mathrm{T}_{p,1}+(p+1)(p^{2}+1)).$ Here the coefficient $-2$ comes from the self-intersection number of the class $\eta_{\sigma}$ as in Lemma 5.7. #### 6.2.2. $\mathcal{T}_{\mathrm{lr},\\{22\\}}$ To understand the map $\mathcal{T}_{\mathrm{lr},\\{22\\}}=\mathrm{inc}^{\ast}_{\\{2\\}}\circ\alpha\circ\mathrm{N}^{-1}_{\Sigma}\circ\beta\circ\mathrm{inc}^{\\{2\\}}_{!}$, we need to understand the possible relative position of vertex lattices $\Lambda_{2}$, $\Lambda_{\mathrm{Pa}}$ and $\Lambda^{\prime}_{2}$ of type $2$, $1$ and $2$. It not hard to see that the only possible way is that $\Lambda_{2}\subset^{1}\Lambda_{\mathrm{Pa}}$ and $\Lambda^{\prime}_{2}\subset^{1}\Lambda_{\mathrm{Pa}}$. The rest of the analysis is completely the same as in the case of $\mathcal{T}_{\mathrm{lr},\\{00\\}}$. Once we identify $\mathrm{K}_{\\{2\\}}$ with the hyperspecial group $\mathbf{G}(\mathbb{Z}_{p})$, the entry $\mathcal{T}_{\mathrm{lr},\\{22\\}}$ is given by $\mathcal{T}_{\mathrm{lr},\\{22\\}}=-2(\mathrm{T}^{-1}_{p,0}\mathrm{T}_{p,1}+(p+1)(p^{2}+1)).$ Here the coefficient $-2$ comes from the self-intersection number of the class $\eta_{\sigma}$ as in Lemma 5.7. #### 6.2.3. $\mathcal{T}_{\mathrm{lr},\\{02\\}}$ To understand the map $\mathcal{T}_{\mathrm{lr},\\{02\\}}=\mathrm{inc}^{\ast}_{\\{2\\}}\circ\alpha\circ\mathrm{N}^{-1}_{\Sigma}\circ\beta\circ\mathrm{inc}^{\\{0\\}}_{!}$, we need to understand the possible relative position of vertex lattices $\Lambda_{0}$, $\Lambda_{\mathrm{Pa}}$ and $\Lambda^{\prime}_{2}$ of type $0$, $1$ and $2$. We find that the only possible situation is that $\Lambda_{2}\subset^{1}\Lambda_{\mathrm{Pa}}\subset^{1}\Lambda_{0}=\Lambda^{\vee}_{0}\subset^{1}\Lambda^{\vee}_{\mathrm{Pa}}\subset^{1}\Lambda^{\vee}_{2}.$ It follows each such $\Lambda_{\mathrm{Pa}}$ corresponds to a line in the two dimensional space $\Lambda_{0}/\Lambda_{2}$. Therefore the entry $\mathcal{T}_{\mathrm{lr},\\{02\\}}$ is given by $\mathcal{T}_{\mathrm{lr},\\{02\\}}=-2(p+1)\mathrm{char}(\mathrm{K}_{\\{0\\}}\begin{pmatrix}p&&&\\\ &p&&\\\ &&1&\\\ &&&1\end{pmatrix}\mathrm{K}_{\\{2\\}})=-2(p+1)\mathrm{T}_{02}.\ $ #### 6.2.4. $\mathcal{T}_{\mathrm{lr},\\{20\\}}$ To understand the map $\mathcal{T}_{\mathrm{lr},\\{20\\}}=\mathrm{inc}^{\ast}_{\\{0\\}}\circ\alpha\circ\mathrm{N}^{-1}_{\Sigma}\circ\beta\circ\mathrm{inc}^{\\{2\\}}_{!}$, we need to understand the possible relative positions of the vertex lattices $\Lambda_{0}$, $\Lambda_{\mathrm{Pa}}$ and $\Lambda^{\prime}_{2}$ of type $0$, $1$ and $2$. We find that the only possible way is that $\Lambda_{2}\subset^{1}\Lambda_{\mathrm{Pa}}\subset^{1}\Lambda_{0}=\Lambda^{\vee}_{0}\subset^{1}\Lambda^{\vee}_{\mathrm{Pa}}\subset^{1}\Lambda^{\vee}_{2}.$ Each $\Lambda_{\mathrm{Pa}}$ corresponds to a line in the two dimensional space $\Lambda_{0}/\Lambda_{2}$. Therefore the entry $\mathcal{T}_{\mathrm{lr},\\{20\\}}$ is given by $\mathcal{T}_{\mathrm{lr},\\{20\\}}=-2(p+1)\mathrm{char}(\mathrm{K}_{\\{0\\}}\begin{pmatrix}p^{-1}&&&\\\ &p^{-1}&&\\\ &&1&\\\ &&&1\end{pmatrix}\mathrm{K}_{\\{2\\}})=-2(p+1)\mathrm{T}_{02}.\ $ This finishes the proof of Proposition 6.10. ∎ We will compute the determinant of the level raising matrix and the supersingular matrix that we will define in the next section. For this reason, we will next compute the composite $\mathrm{T}_{20}\circ\mathrm{T}_{02}$ and $\mathrm{T}_{02}\circ\mathrm{T}_{20}$ in terms of elements in the spherical Hecke algebra. ###### Lemma 6.12. The composite $\mathrm{T}_{20}\circ\mathrm{T}_{02}$ and $\mathrm{T}_{02}\circ\mathrm{T}_{20}$ are both given by $\mathrm{T}^{-1}_{p,0}\mathrm{T}_{p^{2},2}+(p+1)\mathrm{T}^{-1}_{p,0}\mathrm{T}_{p,1}+(p^{2}+1)(p+1).$ As a result, if $\mathrm{T}_{p,0}$ acts trivially, then $\mathrm{T}_{20}\circ\mathrm{T}_{02}=\mathrm{T}_{02}\circ\mathrm{T}_{20}=\mathrm{T}^{2}_{p,2}$ ###### Proof. The proof of this lemma is similar to that of the above proposition. #### 6.2.5. $\mathrm{T}_{20}\circ\mathrm{T}_{02}$ Let $\Lambda_{0}$ be a vertex lattice of type $0$. Then $\mathrm{T}_{20}\circ\mathrm{T}_{02}(\Lambda_{0})$ classifies a pair of vertex lattices $(\Lambda_{2},\Lambda^{\prime}_{0})$ of type $2$ and $0$. We need to understand all the possible relative position of the vertex lattices $\Lambda_{0}$, $\Lambda_{2}$ and $\Lambda^{\prime}_{0}$. The only possible situation is given below $\displaystyle p\Lambda^{\vee}_{0}\subset^{2}\Lambda_{2}\subset^{2}\Lambda_{0}=\Lambda^{\vee}_{0}\subset^{2}\Lambda^{\vee}_{2}$ $\displaystyle p\Lambda^{\prime\vee}_{0}\subset^{2}\Lambda_{2}\subset^{2}\Lambda^{\prime}_{0}=\Lambda^{\prime\vee}_{0}\subset^{2}\Lambda^{\vee}_{2}.$ Note that $\Lambda_{2}$ determines an isotropic subspace of the symplectic space $\Lambda_{0}\cap\Lambda^{\prime}_{0}/p\Lambda^{\vee}_{0}+p\Lambda^{\prime\vee}_{0}$ of possible dimensions in the set $\\{0,2,4\\}$. When $\dim_{\mathbb{F}}\Lambda_{0}\cap\Lambda^{\prime}_{0}/p\Lambda^{\vee}_{0}+p\Lambda^{\prime\vee}_{0}=0$, then $\displaystyle\Lambda_{0}\cap\Lambda^{\prime}_{0}\subset^{2}\Lambda_{0}$ $\displaystyle\Lambda_{0}\cap\Lambda^{\prime}_{0}\subset^{2}\Lambda^{\prime}_{0}.$ Hence this case contributes to the $\mathrm{T}_{20}\circ\mathrm{T}_{02}$ by the double coset operator $\mathrm{K}_{\\{0\\}}\begin{pmatrix}p&&&\\\ &p&&\\\ &&p^{-1}&\\\ &&&p^{-1}\end{pmatrix}\mathrm{K}_{\\{0\\}}.$ Once we identify $\mathrm{K}_{\\{0\\}}$ with $\mathbf{G}(\mathbb{Z}_{p})$, we see this is the same as $\mathrm{T}^{-1}_{0,p}\mathrm{T}_{p^{2},2}$. When $\dim_{\mathbb{F}}\Lambda_{0}\cap\Lambda^{\prime}_{0}/p\Lambda^{\vee}_{0}+p\Lambda^{\prime\vee}_{0}=2$, there are $p+1$ choices of $\Lambda_{2}$. And we have $\displaystyle\Lambda_{0}\cap\Lambda^{\prime}_{0}\subset^{1}\Lambda_{0}$ $\displaystyle\Lambda_{0}\cap\Lambda^{\prime}_{1}\subset^{1}\Lambda^{\prime}_{0}.$ Hence this case contributes to the $\mathrm{T}_{20}\circ\mathrm{T}_{02}$ by the double coset operator $(p+1)\mathrm{K}_{\\{0\\}}\begin{pmatrix}p&&&\\\ &1&&\\\ &&1&\\\ &&&p^{-1}\end{pmatrix}\mathrm{K}_{\\{0\\}}.$ Once we identify $\mathrm{K}_{\\{0\\}}$ with $\mathbf{G}(\mathbb{Z}_{p})$, we see this is the same as $(p+1)\mathrm{T}^{-1}_{0,p}\mathrm{T}_{p,1}$. When $\dim_{\mathbb{F}}\Lambda_{0}\cap\Lambda^{\prime}_{0}/p\Lambda^{\vee}_{0}+p\Lambda^{\prime\vee}_{0}=4$, there are $[\mathbf{G}(\mathbb{Z}_{p}):\mathrm{Sie})]=(p+1)(p^{2}+1)$ choices of $\Lambda_{2}$. And we have $\displaystyle\Lambda_{0}\cap\Lambda^{\prime}_{0}\subset^{0}\Lambda_{0}$ $\displaystyle\Lambda_{0}\cap\Lambda^{\prime}_{0}\subset^{0}\Lambda^{\prime}_{0}.$ Hence this case contributes to the $\mathrm{T}_{20}\circ\mathrm{T}_{02}$ by the constant $(p^{2}+1)(p+1)$. All in all, we obtain the following formula (6.13) $\mathrm{T}_{20}\circ\mathrm{T}_{02}=\mathrm{T}^{-1}_{p,0}\mathrm{T}_{p^{2},2}+(p+1)\mathrm{T}^{-1}_{p,0}\mathrm{T}_{p,1}+(p^{2}+1)(p+1).$ #### 6.2.6. $\mathrm{T}_{02}\circ\mathrm{T}_{20}$ The computation is almost the same to the above. Let $\Lambda_{2}$ be a vertex lattice of type $2$. Then $\mathrm{T}_{02}\circ\mathrm{T}_{20}(\Lambda_{2})$ classifies pair of vertex lattices $(\Lambda_{0},\Lambda^{\prime}_{2})$ of type $0$ and $2$. We need to understand the relative position of vertex lattices $\Lambda_{0}$, $\Lambda_{2}$ and $\Lambda^{\prime}_{2}$. The only possible situation is given below $\displaystyle p\Lambda^{\vee}_{0}\subset^{2}\Lambda_{2}\subset^{2}\Lambda_{0}=\Lambda^{\vee}_{0}\subset^{2}\Lambda^{\vee}_{2};$ $\displaystyle p\Lambda^{\vee}_{0}\subset^{2}\Lambda^{\prime}_{2}\subset^{2}\Lambda_{0}=\Lambda^{\vee}_{0}\subset^{2}\Lambda^{\prime\vee}_{2}.$ Note that $\Lambda_{0}$ is determined by an isotropic subspace of the symplectic space $\Lambda_{2}\cap\Lambda^{\prime}_{2}/\Lambda^{\vee}_{2}+\Lambda^{\prime\vee}_{2}$ of possible dimensions $\\{0,2,4\\}$. When $\dim_{\mathbb{F}}\Lambda_{2}\cap\Lambda^{\prime}_{2}/\Lambda^{\vee}_{2}+\Lambda^{\prime\vee}_{2}=0$, then in this case we have $\displaystyle\Lambda_{2}\subset^{2}\Lambda_{2}+\Lambda^{\prime}_{2}$ $\displaystyle\Lambda^{\prime}_{2}\subset^{2}\Lambda_{2}+\Lambda^{\prime}_{2}.$ Hence this case contributes to the $\mathrm{T}_{02}\circ\mathrm{T}_{20}$ by the double coset operator $\mathrm{K}_{\\{2\\}}\begin{pmatrix}p^{-1}&&&\\\ &p^{-1}&&\\\ &&p&\\\ &&&p\end{pmatrix}\mathrm{K}_{\\{2\\}}.$ Once we identify $\mathrm{K}_{\\{2\\}}$ with $\mathbf{G}(\mathbb{Z}_{p})$, we see this is the same as $\mathrm{T}^{-1}_{0,p}\mathrm{T}_{p^{2},2}$. When $\dim_{\mathbb{F}}\Lambda^{\vee}_{2}\cap\Lambda^{\prime\vee}_{2}/\Lambda_{2}+\Lambda^{\prime}_{2}=2$, there are $p+1$ choices of $\Lambda_{0}$. And in this case we have $\displaystyle\Lambda_{2}\subset^{1}\Lambda_{2}+\Lambda^{\prime}_{2}$ $\displaystyle\Lambda^{\prime}_{2}\subset^{1}\Lambda_{2}+\Lambda^{\prime}_{2}.$ Hence this case contributes to the $\mathrm{T}_{20}\circ\mathrm{T}_{02}$ by the Hecke operator $(p+1)\mathrm{K}_{\\{2\\}}\begin{pmatrix}p^{-1}&&&\\\ &1&&\\\ &&1&\\\ &&&p\end{pmatrix}\mathrm{K}_{\\{2\\}}.$ Once we identify $\mathrm{K}_{\\{2\\}}$ with the hyperspecial subgroup $\mathbf{G}(\mathbb{Z}_{p})$, we see this is the same as $(p+1)\mathrm{T}^{-1}_{0,p}\mathrm{T}_{p,1}$. When $\dim_{\mathbb{F}}\Lambda^{\vee}_{2}\cap\Lambda^{\prime\vee}_{2}/\Lambda_{2}+\Lambda^{\prime}_{2}=4$, there are $[\mathbf{G}(\mathbb{Z}_{p}):\mathrm{Sie})]=(p+1)(p^{2}+1)$ choices of $\Lambda_{0}$. And in this case we have $\displaystyle\Lambda_{2}\subset^{0}\Lambda_{2}+\Lambda^{\prime}_{2}$ $\displaystyle\Lambda^{\prime}_{2}\subset^{0}\Lambda_{2}+\Lambda^{\prime}_{2}.$ Hence this case contributes to the $\mathrm{T}_{20}\circ\mathrm{T}_{02}$ by the constant $(p^{2}+1)(p+1)$. All in all, we obtain the following formula (6.14) $\mathrm{T}_{20}\circ\mathrm{T}_{02}=\mathrm{T}^{-1}_{0,p}\mathrm{T}_{p^{2},2}+(p+1)\mathrm{T}^{-1}_{0,p}\mathrm{T}_{p,1}+(p^{2}+1)(p+1).$ The last claim in the lemma follows immediately from Proposition 6.9. This finishes the proof of the lemma. ∎ The final result of this section is the computation of the determinant of the level raising matrix modulo $\mathfrak{m}$. ###### Proposition 6.15. Let $(\pi,\Sigma,\mathfrak{m})$ be the datum considered as in the beginning of the section. Suppose the maximal ideal $\mathfrak{m}$ appears in the support of $\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]$. Let $[\alpha_{p},\beta_{p},p^{3}\beta^{-1},p^{3}\alpha^{-1}_{p}]$ be the Hecke parameter of $\pi$ at $p$. Then the determinant of the level raising matrix modulo $\mathfrak{m}$ is given by $\mathrm{det}\mathcal{T}_{\mathrm{lr}/\mathfrak{m}}=4p^{-2}\prod_{\mathrm{u}=\pm 1}(\alpha_{p}+p^{3}\alpha^{-1}_{p}-\mathrm{u}p(p+1))(\beta_{p}+p^{3}\beta^{-1}_{p}-\mathrm{u}p(p+1)).$ ###### Proof. Since $\pi$ has trivial central character, $\mathrm{T}_{p,0}$ acts trivially on $\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]/\mathfrak{m}$. We therefore have $\mathrm{T}_{20}\circ\mathrm{T}_{02}=\mathrm{T}_{02}\circ\mathrm{T}_{20}=\mathrm{T}_{p^{2},2}+(p+1)\mathrm{T}_{p,1}+(p^{2}+1)(p+1)=\mathrm{T}^{2}_{p,2}$ by Lemma 6.13. Then $\displaystyle\mathrm{det}\mathcal{T}_{\mathrm{lr}/\mathfrak{m}}$ $\displaystyle=4((\mathrm{T}_{p,1}+p(p+1))^{2}-(p+1)^{2}\mathrm{T}_{20}\mathrm{T}_{02})$ $\displaystyle=4(\mathrm{T}_{p,1}+p(p+1)-(p+1)\mathrm{T}_{p,2})(\mathrm{T}_{p,1}+p(p+1)+(p+1)\mathrm{T}_{p,2})$ $\displaystyle=4p^{-2}$ $\displaystyle[(\alpha_{p}+p^{3}\alpha^{-1}_{p})(\beta_{p}+p^{3}\beta^{-1}_{p})-p(p+1)(\alpha_{p}+\beta_{p}+p^{3}\alpha^{-1}_{p}+p^{3}\beta^{-1}_{p}+p^{2}(p+1)^{2})]$ $\displaystyle[(\alpha_{p}+p^{3}\alpha^{-1}_{p})(\beta_{p}+p^{3}\beta^{-1}_{p})+p(p+1)(\alpha_{p}+\beta_{p}+p^{3}\alpha^{-1}_{p}+p^{3}\beta^{-1}_{p}+p^{2}(p+1)^{2})]$ $\displaystyle=4p^{-2}\prod_{\mathrm{u}=\pm 1}(\alpha_{p}+p^{3}\alpha^{-1}_{p}-\mathrm{u}p(p+1))(\beta_{p}+p^{3}\beta^{-1}_{p}-\mathrm{u}p(p+1)).$ This finishes the proof. ∎ ## 7\. Tate cycles on the quaternionic unitary Shimura variety In this section, our goal is to establish the following priniciple: the localized cohomology group $\mathrm{H}^{2}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(1))_{\mathfrak{m}}$ is generated by cycles coming from the supersingular locus. On the dual side, similar result also holds for $\mathrm{H}^{4}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(2))_{\mathfrak{m}}$, however we need to show this cohomology group contains no $\mathcal{O}_{\lambda}$-torsion first. We will establish the full result simultaneously with our main theorem on arithmetic level raising in Corollary 9.8. Let $(\pi,\Sigma,\mathfrak{m})$ be the datum considered as in the beginning of the last section. ###### Construction 7.1. Consider the supersingular locus $\overline{\mathrm{X}}^{\mathrm{ss}}_{\mathrm{Pa}}(\mathrm{B})$, we have the following restriction morphism $\mathrm{H}^{2}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(1))\rightarrow\mathrm{H}^{2}_{\mathrm{c}}(\overline{\mathrm{X}}^{\mathrm{ss}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(1))$ and we denote the composite of this restriction morphism with the natural map $\mathrm{H}^{2}_{\mathrm{c}}(\overline{\mathrm{X}}^{\mathrm{ss}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(1))\rightarrow\mathrm{H}^{2}(\overline{\mathrm{X}}^{\mathrm{ss}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(1))$ by $\mathrm{inc}^{\ast}_{\\{\mathrm{ss}\\}}:\mathrm{H}^{2}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(1))\rightarrow\mathrm{H}^{2}(\overline{\mathrm{X}}^{\mathrm{ss}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(1)).$ We also have the natural Gysin map $\mathrm{inc}^{\\{\mathrm{ss}\\}}_{!}:\mathrm{H}^{2}(\overline{\mathrm{X}}^{\mathrm{ss}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(1))\rightarrow\mathrm{H}^{4}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(2)).$ We can fit the two maps $(\mathrm{inc}^{\ast}_{\\{\mathrm{ss}\\}},\mathrm{inc}^{\\{\mathrm{ss}\\}}_{!})$ in the above construction in the diagram below. (7.2) ${\mathcal{O}_{\lambda}[\mathrm{Z}_{\\{0\\}}(\overline{\mathrm{B}})]}$${\mathcal{O}_{\lambda}[\mathrm{Z}_{\\{2\\}}(\overline{\mathrm{B}})]}$${\mathrm{H}^{2}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(1))}$${\mathrm{H}^{2}(\overline{\mathrm{X}}^{\mathrm{ss}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(1))}$${\mathrm{H}^{4}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(2))}$${\mathcal{O}_{\lambda}[\mathrm{Z}_{\\{0\\}}(\overline{\mathrm{B}})]}$${\mathcal{O}_{\lambda}[\mathrm{Z}_{\\{2\\}}(\overline{\mathrm{B}})]}$$\scriptstyle{\mathrm{inc}^{\\{0\\}}_{!}}$$\scriptstyle{\mathrm{inc}^{\\{2\\}}_{!}}$$\scriptstyle{\mathrm{inc}^{\ast}_{\\{\mathrm{ss}\\}}}$$\scriptstyle{\mathrm{inc}^{\\{\mathrm{ss}\\}}_{!}}$$\scriptstyle{\mathrm{inc}^{\ast}_{\\{0\\}}}$$\scriptstyle{\mathrm{inc}^{\ast}_{\\{2\\}}}$ The above diagram sets up the intersection matrix for the supersingular locus which we will elaborate below. ###### Construction 7.3. We obtain naturally the following four maps from the above diagram $\displaystyle\mathcal{T}_{\mathrm{ss},\\{00\\}}=\mathrm{inc}^{\ast}_{\\{0\\}}\circ\mathrm{inc}^{\\{\mathrm{ss}\\}}_{!}\circ\mathrm{inc}^{\ast}_{\\{\mathrm{ss}\\}}\circ\mathrm{inc}^{\\{0\\}}_{!}$ $\displaystyle\mathcal{T}_{\mathrm{ss},\\{02\\}}=\mathrm{inc}^{\ast}_{\\{2\\}}\circ\mathrm{inc}^{\\{\mathrm{ss}\\}}_{!}\circ\mathrm{inc}^{\ast}_{\\{\mathrm{ss}\\}}\circ\circ\mathrm{inc}^{\\{0\\}}_{!}$ $\displaystyle\mathcal{T}_{\mathrm{ss},\\{20\\}}=\mathrm{inc}^{\ast}_{\\{0\\}}\circ\mathrm{inc}^{\\{\mathrm{ss}\\}}_{!}\circ\mathrm{inc}^{\ast}_{\\{\mathrm{ss}\\}}\circ\mathrm{inc}^{\\{2\\}}_{!}$ $\displaystyle\mathcal{T}_{\mathrm{ss},\\{22\\}}=\mathrm{inc}^{\ast}_{\\{2\\}}\circ\mathrm{inc}^{\\{\mathrm{ss}\\}}_{!}\circ\mathrm{inc}^{\ast}_{\\{\mathrm{ss}\\}}\circ\circ\mathrm{inc}^{\\{2\\}}_{!}.$ The resulting matrix $\mathcal{T}_{\mathrm{ss}}=\begin{pmatrix}&\mathcal{T}_{\mathrm{ss},\\{00\\}}&\mathcal{T}_{\mathrm{ss},\\{02\\}}\\\ &\mathcal{T}_{\mathrm{ss},\\{20\\}}&\mathcal{T}_{\mathrm{ss},\\{22\\}}\\\ \end{pmatrix}$ will be referred to as the supersingular matrix. We can localize the above diagram at the maximal ideal $\mathfrak{m}$ and write the resulting matrix as $\mathcal{T}_{\mathrm{ss},\mathfrak{m}}$, however all the matrix entries will still be denoted by symbols without referring to $\mathfrak{m}$. If we need to consider this matrix modulo $\mathfrak{m}$, then we will denote it by $\mathcal{T}_{\mathrm{ss}/\mathfrak{m}}$. The entries of this matrix will be denoted without referring to $\mathfrak{m}$. ### 7.1. Computing the supersingular matrix We compute the entries of the supersingular matrix using the same procedure as we did for computing the level raising matrix. ###### Proposition 7.4. The supersingular matrix $\mathcal{T}_{\mathrm{ss}}$ is given by $\mathcal{T}_{\mathrm{ss}}=\begin{pmatrix}&4p^{2}(p+1)^{2}+\mathrm{T}_{20}\circ\mathrm{T}_{02}&-4p(p+1)\mathrm{T}_{02}\\\ &-4p(p+1)\mathrm{T}_{20}&4p^{2}(p+1)^{2}+\mathrm{T}_{02}\circ\mathrm{T}_{20}\\\ \end{pmatrix}.$ ###### Proof. We will prove the proposition in an entry by entry manner. #### 7.1.1. $\mathcal{T}_{\mathrm{ss},\\{00\\}}$ Recall $\mathcal{T}_{\mathrm{ss},\\{00\\}}=\mathrm{inc}^{\ast}_{\\{0\\}}\circ\mathrm{inc}^{\\{\mathrm{ss}\\}}_{!}\circ\mathrm{inc}^{\ast}_{\\{\mathrm{ss}\\}}\circ\mathrm{inc}^{\\{0\\}}_{!}$. If the maps $\mathrm{H}^{2}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(1))\xrightarrow{\mathrm{inc}^{\ast}_{\\{\mathrm{ss}\\}}}\mathrm{H}^{2}(\overline{\mathrm{X}}^{\mathrm{ss}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(1))\xrightarrow{\mathrm{inc}^{\\{\mathrm{ss}\\}}_{!}}\mathrm{H}^{4}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(2))$ factor through $\mathrm{H}^{2}(\overline{\mathrm{X}}^{\mathrm{ss}}_{\mathrm{Pa},\\{0\\}}(\mathrm{B}),\mathcal{O}_{\lambda}(1))$. Then the contribution is $4p^{2}(p+1)^{2}$. This follows from the fact that for each $\Lambda_{0}\in\mathcal{L}_{\\{0\\}}$ the degree of the variety $\mathrm{DL}(\Lambda_{0})$ has degree $(p+1)$ and the normal bundle has degree $-2p$ by Lemma 2.10. If the maps $\mathrm{H}^{2}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(1))\xrightarrow{\mathrm{inc}^{\ast}_{\\{\mathrm{ss}\\}}}\mathrm{H}^{2}(\overline{\mathrm{X}}^{\mathrm{ss}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(1))\xrightarrow{\mathrm{inc}^{\\{\mathrm{ss}\\}}_{!}}\mathrm{H}^{4}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(2))$ factor through $\mathrm{H}^{2}(\overline{\mathrm{X}}^{\mathrm{ss}}_{\mathrm{Pa},\\{2\\}}(\mathrm{B}),\mathcal{O}_{\lambda}(1))$, then as we have seen before we need to understand the relative position of the vertex lattices $\Lambda_{0}$, $\Lambda_{2}$ and $\Lambda^{\prime}_{0}$ of type $0$, $2$ and $0$. The only possible situation is given by $\displaystyle p\Lambda^{\vee}_{0}\subset^{2}\Lambda_{2}\subset^{2}\Lambda_{0}=\Lambda^{\vee}_{0}\subset^{2}\Lambda^{\vee}_{2}$ $\displaystyle p\Lambda^{\prime\vee}_{0}\subset^{2}\Lambda_{2}\subset^{2}\Lambda^{\prime}_{0}=\Lambda^{\prime\vee}_{0}\subset^{2}\Lambda^{\vee}_{2}.$ As we have seen, this is parametrized by $\mathrm{T}_{20}\circ\mathrm{T}_{02}$. Therefore this entry is given by $\mathcal{T}_{\mathrm{ss},\\{00\\}}=(p+1)^{2}+\mathrm{T}_{20}\circ\mathrm{T}_{02}.$ #### 7.1.2. $\mathcal{T}_{\mathrm{ss},\\{02\\}}$ If the maps $\mathrm{H}^{2}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(1))\xrightarrow{\mathrm{inc}^{\ast}_{\\{\mathrm{ss}\\}}}\mathrm{H}^{2}(\overline{\mathrm{X}}^{\mathrm{ss}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(1))\xrightarrow{\mathrm{inc}^{\\{\mathrm{ss}\\}}_{!}}\mathrm{H}^{4}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(2))$ factor through $\mathrm{H}^{2}(\overline{\mathrm{X}}^{\mathrm{ss}}_{\mathrm{Pa},\\{0\\}}(\mathrm{B}),\mathcal{O}_{\lambda}(1))$. This case contributes by $-2p(p+1)\mathrm{T}_{02}$. If the maps factor through $\mathrm{H}^{2}(\overline{\mathrm{X}}^{\mathrm{ss}}_{\mathrm{Pa},\\{2\\}}(\mathrm{B}),\mathcal{O}_{\lambda}(1))$. This case contributes by $-2p(p+1)\mathrm{T}_{02}$. Therefore the total contribution gives the entry $\mathcal{T}_{\mathrm{ss},\\{02\\}}=-4p(p+1)\mathrm{T}_{02}.$ #### 7.1.3. $\mathcal{T}_{\mathrm{ss},\\{20\\}}$ The computation is the same as in the case for $\mathcal{T}_{\mathrm{ss},\\{02\\}}$. Thus this entry is given by $\mathcal{T}_{\mathrm{ss},\\{20\\}}=-4p(p+1)\mathrm{T}_{20}.$ #### 7.1.4. $\mathcal{T}_{\mathrm{ss},\\{22\\}}$ The computation is exactly the same as in the case for $\mathcal{T}_{\mathrm{ss},\\{00\\}}$. Thus this entry is given by $\mathcal{T}_{\mathrm{ss},\\{22\\}}=4p^{2}(p+1)^{2}+\mathrm{T}_{02}\circ\mathrm{T}_{20}.$ This finishes the proof of Proposition 7.4. ∎ ###### Proposition 7.5. Let $(\pi,\Sigma,\mathfrak{m})$ be the datum considered as in the beginning of the section. Suppose the maximal ideal $\mathfrak{m}$ appears in the support of $\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]$. Suppose that $[\alpha_{p},\beta_{p},p^{3}\beta^{-1},p^{3}\alpha^{-1}_{p}]$ is the Hecke parameter of $\pi$ at $p$. Then determinant of the supersingular matrix modulo $\mathfrak{m}$ is given by $\mathrm{det}\phantom{.}\mathcal{T}_{\mathrm{ss}/\mathfrak{m}}=\prod_{\mathrm{u}=\pm 1}(\alpha_{p}+\beta_{p}+p^{3}\alpha^{-1}_{p}+p^{3}\beta^{-1}_{p}-2\mathrm{u}p(p+1))^{2}.$ ###### Proof. Since $\pi$ has trivial central character, then $\mathrm{T}_{p,0}$ acts trivially on $\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]/\mathfrak{m}$. We therefore have $\mathrm{T}_{20}\circ\mathrm{T}_{02}=\mathrm{T}_{02}\circ\mathrm{T}_{20}=\mathrm{T}_{p^{2},2}+(p+1)\mathrm{T}_{p,1}+(p^{2}+1)(p+1)=\mathrm{T}^{2}_{p,2}$ by Lemma 6.13. Then we have $\displaystyle\mathrm{det}\phantom{.}\mathcal{T}_{\mathrm{ss}/\mathfrak{m}}$ $\displaystyle=(\mathrm{T}^{2}_{p,2}+4p^{2}(p+1))^{2}-16p^{2}(p+1)^{2}\mathrm{T}^{2}_{p,2}$ $\displaystyle=(\mathrm{T}^{2}_{p,2}+4p^{2}(p+1)-4p(p+1)\mathrm{T}_{p,2})$ $\displaystyle\phantom{abc}(\mathrm{T}^{2}_{p,2}+4p^{2}(p+1)+4p(p+1)\mathrm{T}_{p,2})$ $\displaystyle=(\alpha_{p}+\beta_{p}+p^{3}\beta^{-1}_{p}+p^{3}\alpha^{-1}_{p}-2p(p+1))^{2}$ $\displaystyle\phantom{abc}(\alpha_{p}+\beta_{p}+p^{3}\beta^{-1}_{p}+p^{3}\alpha^{-1}_{p}+2p(p+1))^{2}$ $\displaystyle=\prod_{\mathrm{u}=\pm 1}(\alpha_{p}+\beta_{p}+p^{3}\beta^{-1}_{p}+p^{3}\alpha^{-1}_{p}-2\mathrm{u}p(p+1))^{2}.$ ∎ ### 7.2. Tate classes in the quaternionic unitary Shimura variety We fix the datum $(\pi,\Sigma,\mathfrak{m})$ as in the beginning of this section. Let $p$ be a level raising special prime for $\pi$ of length $m$. We make the following assumption regarding to the vanishing of the nearby cycle cohomology of the quaternionic unitary Shimura variety. ###### Assumption 7.6. Let $\mathrm{H}^{d}_{(\mathrm{c})}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}))_{\mathfrak{m}}$ be the usual or compact support nearby cycle cohomology of the special fiber $\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B})$. We assume the maximal ideal $\mathfrak{m}$ satisfies the following properties 1. (1) The natural morphism $\mathrm{H}^{d}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}))_{\mathfrak{m}}\xrightarrow{\sim}\mathrm{H}^{d}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}))_{\mathfrak{m}}$ is an isomorphism for all $d$. 2. (2) The cohomology $\mathrm{H}^{d}_{(\mathrm{c})}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}))_{\mathfrak{m}}=0$ for $d\neq 3$, and that the middle degree cohomology $\mathrm{H}^{3}_{(\mathrm{c})}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}))_{\mathfrak{m}}$ is a finite free $\mathcal{O}_{\lambda}$ module. ###### Remark 7.7. The first assumption is usually known in literature as that of $\mathfrak{m}$ being a non-Eisenstein ideal. It is satisfied if $\mathfrak{m}$ is associated to a cuspidal automorphic representation of general type. The second assumption is more serious. This is reasonable assumption in light of the recent work of Caraiani-Scholze [CS17, CS19] and recent advances by Koshikawa [Kos19, Kos21]. Although they deal with certain unitary Shimura varieties, the methods are general enough to treat at least our cases: if one assumes that $\pi$ has an unramified type $\mathrm{I}$ component which is generic modulo $l$, then combining the proof of Koshikawa [Kos21] with a semi-perversity result of the pushforward along the Hodge-Tate period map [San22], one can prove that $\mathfrak{m}$ satisfies $(2)$. If we make the assumption that $\mathrm{X}_{\mathrm{Pa}}(\mathrm{B})$ has good reduction at $l$ and that $\pi$ is ordinary at $l$, then the method of [MT02] could be used to show that the Assumption 7.6 is fulfilled. We also conjecture that the product of Shimura curves for $\mathrm{B}$ embedded in $\mathrm{Sh}(\mathrm{B},\mathrm{K}_{\mathrm{Pa}})$ has an affine complement. Then the above assumption can be proved if one assumes that the residual Galois representation has large image. This is the analogue of the fact that the Igusa divisor on the classical Siegel threefold has an affine complement. ###### Definition 7.8. If $\mathfrak{m}$ satisfies either of the two conditions below, we call $\mathfrak{m}$ a generic maximal ideal. 1. (1) We say $\mathfrak{m}$ is generic level raising at $p$ if it is level raising special at $p$ and Assumption 7.6 is satisfied. 2. (2) We say $\mathfrak{m}$ is generic non-level raising at $p$ if the Hecke parameters for $\pi$ satisfies $\displaystyle\alpha_{p}+p^{3}\alpha^{-1}_{p}\not\equiv\pm(p+p^{2})\mod\lambda$ $\displaystyle\beta_{p}+p^{3}\beta^{-1}_{p}\not\equiv\pm(p+p^{2})\mod\lambda$ and Assumption 7.6 is satisfied. ###### Theorem 7.9. Let $\mathfrak{m}$ be a generic maximal ideal as in the above definition, we have the following statements. 1. (1) The map from Construction 6.4 $(\mathrm{inc}^{\\{0\\}}_{!,\mathfrak{m}}+\mathrm{inc}^{\\{2\\}}_{!,\mathfrak{m}}):\mathcal{O}_{\lambda}[\mathrm{Z}_{\\{0\\}}(\overline{\mathrm{B}})]_{\mathfrak{m}}\oplus\mathcal{O}_{\lambda}[\mathrm{Z}_{\\{2\\}}(\overline{\mathrm{B}})]_{\mathfrak{m}}\rightarrow\mathrm{H}^{2}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(1))_{\mathfrak{m}}$ is an isomorphism. 2. (2) The map from Construction 6.4 $(\mathrm{inc}^{\ast}_{\\{0\\},\mathfrak{m}},\mathrm{inc}^{\ast}_{\\{2\\},\mathfrak{m}}):\mathrm{H}^{4}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(2))_{\mathfrak{m}}\rightarrow\mathcal{O}_{\lambda}[\mathrm{Z}_{\\{0\\}}(\overline{\mathrm{B}})]_{\mathfrak{m}}\oplus\mathcal{O}_{\lambda}[\mathrm{Z}_{\\{2\\}}(\overline{\mathrm{B}})]_{\mathfrak{m}}$ is surjective and whose kernel is the torsion part of $\mathrm{H}^{4}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(2))_{\mathfrak{m}}$. ###### Proof. Suppose for the moment that $\mathfrak{m}$ is generic level raising at $p$. Then the determinant of the supersingular matrix $\mathrm{det}\mathcal{T}_{\mathrm{ss}/\mathfrak{m}}$ is given by $\prod_{\mathrm{u}=\pm 1}(\alpha_{p}+\beta_{p}+p^{3}\beta^{-1}_{p}+p^{3}\alpha^{-1}_{p}-2\mathrm{u}p(p+1))^{2}.$ This is non-zero by our definition of the level raising condition. This immediately implies that we have an injection $\mathcal{O}_{\lambda}[\mathrm{Z}_{\\{0\\}}(\overline{\mathrm{B}})]_{\mathfrak{m}}\oplus\mathcal{O}_{\lambda}[\mathrm{Z}_{\\{2\\}}(\overline{\mathrm{B}})]_{\mathfrak{m}}\hookrightarrow\mathrm{H}^{2}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(1))_{\mathfrak{m}}$ by the Nakayama’s lemma. Note that $\mathrm{H}^{2}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(1))_{\mathfrak{m}}$ is a free $\mathcal{O}_{\lambda}$-module as it injects into $\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda})(1))_{\mathfrak{m}}$ which is torsion free by the genericness of $\mathfrak{m}$. Therefore to finish the proof in this case, we need to show that $2\mathrm{rank}_{\mathcal{O}_{\lambda}}\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]_{\mathfrak{m}}\geq\mathrm{rank}_{\mathcal{O}_{\lambda}}\mathrm{H}^{2}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(1))_{\mathfrak{m}}.$ For this purpose, let $\vec{\pi}$ be an automorphic representation of $\mathbf{G}(\mathbb{A}_{\mathbb{Q}})$ of general type. It suffices to show that $2\dim_{\overline{\mathbb{Q}}_{l}}\overline{\mathbb{Q}}_{l}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})][\iota_{l}\vec{\pi}^{\infty pq}]\geq\dim_{\overline{\mathbb{Q}}_{l}}\mathrm{H}^{2}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\overline{\mathbb{Q}}_{l}(1))[\iota_{l}\vec{\pi}^{\infty pq}].$ Consider the co-specialization exact sequence $\displaystyle 0$ $\displaystyle\rightarrow\mathrm{H}^{2}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\overline{\mathbb{Q}}_{l}(1))[\iota_{l}\vec{\pi}^{\infty pq}]\rightarrow\bigoplus_{\sigma\in\mathrm{Z}_{\\{1\\}}(\overline{\mathrm{B}})}\mathrm{H}^{3}_{\\{\sigma\\}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\overline{\mathbb{Q}}_{l})(1))[\iota_{l}\vec{\pi}^{\infty p}]$ $\displaystyle\rightarrow\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\overline{\mathbb{Q}}_{l})(1))[\iota_{l}\vec{\pi}^{\infty p}]\rightarrow\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\overline{\mathbb{Q}}_{l}(1))[\iota_{l}\vec{\pi}^{\infty p}]\rightarrow 0$ We can canonically identify the space $\bigoplus\limits_{\sigma\in\mathrm{Z}_{\\{1\\}}(\overline{\mathrm{B}})}\mathrm{H}^{3}_{\\{\sigma\\}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\overline{\mathbb{Q}}_{l})[\iota_{l}\vec{\pi}^{\infty p}]$ with the space $\overline{\mathbb{Q}}_{l}[\mathrm{Z}_{\mathrm{Pa}}(\overline{\mathrm{B}})][\iota_{l}\vec{\pi}^{\infty p}]$. Note that to complete $\vec{\pi}^{p\infty}$ to a representation $\vec{\pi}$ of $\mathbf{G}(\mathbb{A}_{\mathbb{Q}})$ occuring in $\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\overline{\mathbb{Q}}_{l})(1))$, $\vec{\pi}_{p}$ must be of type $\mathrm{II}\mathrm{a}$ as a representation of $\mathbf{G}(\mathbb{Q}_{p})$ by the proof of Theorem 4.5. On the other hand, those $\vec{\pi}_{p}$ completing $\vec{\pi}^{\infty p}$ in $\overline{\mathbb{Q}}_{l}[\mathrm{Z}_{\mathrm{Pa}}(\overline{\mathrm{B}})][\iota_{l}\pi^{\infty p}]$ can be either the unramified principal series or of type $\mathrm{II}\mathrm{a}$. In the proof of Theorem 4.5, we have seen the the Jacquet–Langlands correspondence preserves the automorphic multiplicity. Thus by the uniqueness of local newforms [BS, Theorem 5.6.1] or [Sch05, Theorem 2.3.1], those $\vec{\pi}_{p}$ completing $\vec{\pi}^{\infty p}$ in $\mathrm{H}^{2}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\overline{\mathbb{Q}}_{l}(1))[\iota_{l}\vec{\pi}^{\infty p}]$ can only be unramified unless the map $\bigoplus_{\sigma\in\mathrm{Z}_{\\{1\\}}(\overline{\mathrm{B}})}\mathrm{H}^{3}_{\\{\sigma\\}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\overline{\mathbb{Q}}_{l})(1))[\iota_{l}\vec{\pi}^{\infty p}]\rightarrow\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\overline{\mathbb{Q}}_{l})(1))[\iota_{l}\vec{\pi}^{\infty p}]\ $ is zero. Howerver this is impossible as otherwise, it would imply the isomorphism $\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\overline{\mathbb{Q}}_{l})(1))[\iota_{l}\vec{\pi}^{\infty p}]\rightarrow\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\overline{\mathbb{Q}}_{l}(1))[\iota_{l}\vec{\pi}^{\infty p}]$ which is absurd by considering the inertia action. By the oldforms principle [BS, Theorem 5.6.1] or [Sch05, Theorem 2.3.1], we have the desired $2\dim_{\overline{\mathbb{Q}}_{l}}\overline{\mathbb{Q}}_{l}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})][\iota_{l}\vec{\pi}^{\infty p}]\geq\dim_{\overline{\mathbb{Q}}_{l}}\mathrm{H}^{2}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\overline{\mathbb{Q}}_{l}(1))[\iota_{l}\vec{\pi}^{\infty p}].$ Next we show the same result under the assumption that $\mathfrak{m}$ is generic non-level raising. Suppose that $\mathfrak{m}$ is not level raising special. Then the determinant of the level raising matrix $\mathrm{det}\mathcal{T}_{\mathrm{lr}/\mathfrak{m}}$ is given by $\mathrm{det}\mathcal{T}_{\mathrm{lr}/\mathfrak{m}}=4p^{-2}\prod_{\mathrm{u}=\pm 1}(\alpha_{p}+p^{3}\alpha^{-1}_{p}-\mathrm{u}p(p+1))(\beta_{p}+p^{3}\beta^{-1}_{p}-\mathrm{u}p(p+1)).$ This is non-zero by our assumption. This immediately implies that we have an injection $\mathcal{O}_{\lambda}[\mathrm{Z}_{\\{0\\}}(\overline{\mathrm{B}})]_{\mathfrak{m}}\oplus\mathcal{O}_{\lambda}[\mathrm{Z}_{\\{2\\}}(\overline{\mathrm{B}})]_{\mathfrak{m}}\hookrightarrow\mathrm{H}^{2}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(1))_{\mathfrak{m}}$ by the Nakayama’s lemma. The rest of the argument is the same as in the previous case. Finally, the second part of the proposition is dual to the first part. In particular, the map $(\mathrm{inc}^{\ast}_{\\{0\\},\mathfrak{m}},\mathrm{inc}^{\ast}_{\\{2\\},\mathfrak{m}}):\mathrm{H}^{4}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(2))_{\mathfrak{m}}\rightarrow\mathcal{O}_{\lambda}[\mathrm{Z}_{\\{0\\}}(\overline{\mathrm{B}})]_{\mathfrak{m}}\oplus\mathcal{O}_{\lambda}[\mathrm{Z}_{\\{2\\}}(\overline{\mathrm{B}})]_{\mathfrak{m}}$ is surjective. Using the same argument as in the first part, we see that $(\mathrm{inc}^{\ast}_{\\{0\\},\mathfrak{m}},\mathrm{inc}^{\ast}_{\\{2\\},\mathfrak{m}})$ is an isomorphism up to the torsion of $\mathrm{H}^{4}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(2))_{\mathfrak{m}}$. The second part now follows ∎ We finish this section with several remarks about this theorem. First of all, this result can be understood as the statement that the cohomology of the special fiber $\mathrm{H}^{2}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(1))_{\mathfrak{m}}$ is generated by Tate cycles coming from the supersingular locus. This result along with [LTXZZ, Proposition 6.3.1] can be seen as an analogue of the main result of [XZ] in the bad reduction case for low degree cohomology groups localized at generic maximal ideals. It seems that these cohomology groups always correspond to oldforms. It would be of great interest to formulate some predictions in a similar manner of [XZ]. Secondly, we have established the following analogue of the Ihara’s lemma in the setting of definite quaternionic unitary groups. ###### Corollary 7.10. There is an injection $\mathbf{Ih}_{\mathfrak{m}}:\mathcal{O}_{\lambda}[\mathrm{Z}_{\\{0\\}}(\overline{\mathrm{B}})]_{\mathfrak{m}}\oplus\mathcal{O}_{\lambda}[\mathrm{Z}_{\\{2\\}}(\overline{\mathrm{B}})]_{\mathfrak{m}}\xhookrightarrow{\phantom{aaaa}}\mathcal{O}_{\lambda}[\mathrm{Z}_{\\{1\\}}(\overline{\mathrm{B}})]_{\mathfrak{m}}.$ ###### Proof. We have shown in the above proof that the composite of the two maps $\displaystyle\mathcal{O}_{\lambda}[\mathrm{Z}_{\\{0\\}}(\overline{\mathrm{B}})]_{\mathfrak{m}}\oplus\mathcal{O}_{\lambda}[\mathrm{Z}_{\\{2\\}}(\overline{\mathrm{B}})]_{\mathfrak{m}}\xrightarrow{(\mathrm{inc}^{\\{0\\}}_{!,\mathfrak{m}}+\mathrm{inc}^{\\{2\\}}_{!,\mathfrak{m}})}\mathrm{H}^{2}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(1))_{\mathfrak{m}}$ $\displaystyle\mathrm{H}^{2}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(1))_{\mathfrak{m}}\xhookrightarrow{\phantom{aa}\alpha\phantom{aa}}\bigoplus_{\sigma\in Z_{\\{1\\}}(\overline{\mathrm{B}})}\mathrm{H}^{3}_{\\{\sigma\\}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda})(1))=\mathcal{O}_{\lambda}[\mathrm{Z}_{\\{1\\}}(\overline{\mathrm{B}})]_{\mathfrak{m}}$ gives rise to an injection which we call $\mathbf{Ih}_{\mathfrak{m}}:\mathcal{O}_{\lambda}[\mathrm{Z}_{\\{0\\}}(\overline{\mathrm{B}})]_{\mathfrak{m}}\oplus\mathcal{O}_{\lambda}[\mathrm{Z}_{\\{2\\}}(\overline{\mathrm{B}})]_{\mathfrak{m}}\xhookrightarrow{\phantom{aaaa}}\mathcal{O}_{\lambda}[\mathrm{Z}_{\\{1\\}}(\overline{\mathrm{B}})]_{\mathfrak{m}}.$ ∎ Note that there is no natural degeneracy map from neither $\mathcal{O}_{\lambda}[\mathrm{Z}_{\\{0\\}}(\overline{\mathrm{B}})]_{\mathfrak{m}}$ nor $\mathcal{O}_{\lambda}[\mathrm{Z}_{\\{2\\}}(\overline{\mathrm{B}})]_{\mathfrak{m}}$ to $\mathcal{O}_{\lambda}[\mathrm{Z}_{\\{1\\}}(\overline{\mathrm{B}})]_{\mathfrak{m}}$. Instead, these maps should be the analogues of the level raising operators in the sense of [BS]. Finally suppose that $\mathfrak{m}$ is generic level raising at $p$, we will show that $\mathrm{H}^{4}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(2))_{\mathfrak{m}}$ is torsion free following the method of [LTXZZ] using Galois deformation argument. ## 8\. Galois deformation rings and cohomology of Shimura varieties In this section, we recall the main results in [Wang22a] where we use the method of [LTXZZa] to show that the cohomology of the quaternionic unitary Shimura varieties are free over certain universal deformation rings. These results are necessary as we lack of torsion vanishing results on the cohomology of the special fibers of these quaternionic unitary Shimura varieties. ### 8.1. Galois representation for $\mathrm{GSp}_{4}$ First we recall some results on associating Galois representations to cuspidal automorphic representations of $\mathrm{GSp}_{4}(\mathbb{A}_{\mathbb{Q}})$ and the local-global compatibility of Langlands correspondence. Let $\pi$ be a cuspidal automorphic representation of $\mathrm{GSp}_{4}(\mathbb{A}_{\mathbb{Q}})$ with trivial central character. We assume that $\pi$ has weight $(k_{1},k_{2})$ and trivial central character. In this article, we will only consider discrete automorphic representation of $\mathrm{GSp}_{4}(\mathbb{A}_{\mathbb{Q}})$ that is of general type. This means there is a cuspidal automorphic representation $\Pi$ of $\mathrm{GL}_{4}(\mathbb{A}_{\mathbb{Q}})$ of symplectic type such that for each place $v$ of $\mathbb{Q}$, the $L$-parameter obtained from $\mathrm{rec}_{\mathrm{GT}}(\pi_{v})$ by composing with the embedding $\mathrm{GSp}_{4}\hookrightarrow\mathrm{GL}_{4}$ is precisely $\mathrm{rec}_{\mathrm{GL}_{4}}(\Pi_{v})$ which is the Langlands parameter attached to $\Pi_{v}$. Here $\Pi$ is of symplectic type if the partial $L$-function $L^{\mathrm{S}}(s,\Pi,\wedge^{2})$ has a pole at $s=1$ for any finite set $S$ of places of $\mathbb{Q}$. In this case, we say $\Pi$ is the transfer of $\pi$. We now recall some general results of Mok and Sorensen on the existence of Galois representations attached to $\pi$. This theorem is very general and we temporarily drop the assumption that $\pi$ is of general type. ###### Theorem 8.1. Suppose $\pi$ is a cuspidal automorphic representation of $\mathrm{GSp}_{4}(\mathbb{A}_{\mathbb{Q}})$ of weight $(k_{1},k_{2})$ where $k_{1}\geq k_{2}\geq 3$ and $k_{1}\equiv k_{2}\mod 2$. Suppose that $\pi$ has trivial central character. Let $l\geq 5$ be a fixed prime.Then there is a continuous semisimple representation $\rho_{\pi,\iota_{l}}:\mathrm{G}_{\mathbb{Q}}\rightarrow\mathrm{GSp}_{4}(\overline{\mathbb{Q}}_{l})$ satisfying the following properties. 1. (1) $\mathrm{c}\circ\rho_{\pi,\iota_{l}}=\epsilon^{-3}_{l}$; 2. (2) For each finite place $v$, we have $\mathrm{WD}(\rho_{\pi,\iota_{l}}|_{\mathrm{G}_{\mathbb{Q}_{v}}})^{\mathrm{F}-\mathrm{ss}}\cong\mathrm{rec}_{\mathrm{GT}}(\pi_{v}\otimes|\mathrm{c}|^{-3/2})^{\mathrm{ss}};$ 3. (3) The local representation $\rho_{\pi,\iota_{l}}|_{\mathrm{G}_{\mathbb{Q}_{l}}}$ is de Rham with Hodge–Tate weights $(2-\frac{k_{1}+k_{2}}{2},-\frac{k_{1}-k_{2}}{2},1+\frac{k_{1}-k_{2}}{2},-1+\frac{k_{1}+k_{2}}{2}).$ 4. (4) If $\pi$ is unramified at $l$, then $\rho_{\pi,\iota_{l}}|_{\mathrm{G}_{\mathbb{Q}_{v}}}$ is moreover crystalline at $l$. 5. (5) If $\rho_{\pi,\iota_{l}}$ is irreducible, then for all finite place $v$, $\rho_{\pi,\iota_{l}}|_{\mathrm{G}_{\mathbb{Q}_{v}}}$ is pure. ###### Remark 8.2. Several remarks about this theorem are in order. The construction of the Galois representation in this case is given in [Lau97, Lau05, Wei05, Tay93] using the Langlands–Kottwitz method. The statements about local-global compatibilities are mostly proved in [Sor10] and completed in [Mok05]. In fact the authors use the strong transfer of cuspidal automorphic representation from $\mathrm{GSp}_{4}$ to $\mathrm{GL}_{4}$ to construct the desired Galois representation in the totally real field case. The Harish-Chandra parameter $(\mu_{1},\mu_{2})$ of $\pi$ used more often has the following relation with the weight of $\pi$ used in this theorem: $k_{1}=\mu_{1}+1$ and $k_{2}=\mu_{2}+2$. In this article, we will be interested in the case $\pi$ is of general type and has trivial central character and weights $(3,3)$. In our normalization, this is the case when these representation appear in the cohomology of quaternionic unitary Shimura varieties with trivial coeffficient. ###### Definition 8.3. Let $\pi$ be a cuspidal automorphic representation of $\mathrm{GSp}_{4}(\mathbb{A}_{\mathbb{Q}})$ as in the previous remark. We say a number field $E\subset\mathbb{C}$ is a strong coefficient field of $\pi$ if for every prime $\lambda$ of $E$ there exits a continuous homomorphism $\rho_{\pi,\lambda}:\mathrm{G}_{\mathbb{Q}}\rightarrow\mathrm{GSp}_{4}(E_{\lambda})$ up to conjugation, such that for every isomorphism $\iota_{l}:\mathbb{C}\xrightarrow{\sim}\overline{\mathbb{Q}}_{l}$ inducing the prime $\lambda$, the representations $\rho_{\pi,\lambda}\otimes_{E_{\lambda}}\overline{\mathbb{Q}}_{l}$ and $\rho_{\pi,\iota_{l}}$ are conjugate. Let $p$ be a place at which $\pi$ is an unramified type $\mathrm{I}$ representation. Then the Frobenius eigenvalues of $\rho_{\pi,\lambda}(\mathrm{Frob}_{p})$ agree with the Hecke parameter $[\alpha_{p},\beta_{p},\gamma_{p},\delta_{p}]$ of $\pi_{p}$ by Theorem 8.1 $(2)$. We will always make the following assumption on $\rho_{\pi,\lambda}$. ###### Assumption 8.4. The Galois representation $\rho_{\pi,\lambda}:\mathrm{G}_{\mathbb{Q}}\rightarrow\mathrm{GSp}_{4}(E_{\lambda})$ is residually absolutely irreducible. The above assumption allows us to define the residual Galois representation (8.5) $\overline{\rho}_{\pi,\lambda}:\mathrm{G}_{\mathbb{Q}}\rightarrow\mathrm{GSp}_{4}(k)$ which is unique up to conjugation. Note it also follows that $\pi$ is necessarily of general type if assumption 8.4 is effective. ### 8.2. Galois deformation rings We summarize the main results in [Wang22a]. Suppose we are given a Galois representation $\overline{\rho}:\mathrm{G}_{\mathbb{Q}}\rightarrow\mathrm{GSp}_{4}(k)$. We will write the restriction $\overline{\rho}|_{\mathrm{G}_{\mathbb{Q}_{v}}}:\mathrm{G}_{\mathbb{Q}_{v}}\rightarrow\mathrm{GSp}(k)$ by $\overline{\rho}_{v}$. Let $\Sigma_{\mathrm{min}}$ and $\Sigma_{\mathrm{lr}}$ be two sets of non- archimedean places of $\mathbb{Q}$ away from $l$ and such that * • $q$ is contained in $\Sigma_{\mathrm{lr}}$; * • $\Sigma_{\mathrm{min}}$, $\Sigma_{\mathrm{lr}}$ and $\\{p\\}$ are mutually disjoint; * • for every $v\in\Sigma_{\mathrm{lr}}\cup\\{p\\}$, the prime $v$ satisfies $l\nmid(v^{2}-1)$. ###### Definition 8.6. For a Galois representation $\overline{\rho}:\mathrm{G}_{\mathbb{Q}}\rightarrow\mathrm{GSp}(k)$ and $\psi=\epsilon^{-3}_{l}$, we say the pair $\overline{\rho}$ is rigid for $(\Sigma_{\mathrm{min}},\Sigma_{\mathrm{lr}})$ if the followings are satisfied. * • For every $v\in\Sigma_{\mathrm{min}}$, every lifting of $\overline{\rho}_{v}$ is minimally ramified in the sense of [Wang22a, Definition 3.4]; * • For every $v\in\Sigma_{\mathrm{lr}}$, the generalized eigenvalues of $\overline{\rho}_{v}(\phi_{v})$ contain the pair $\\{v^{-1},v^{-2}\\}$ exactly once; * • At $v=l$, $\overline{\rho}_{v}$ is regular Fontaine-Laffaille crystalline as in [Wang22a, Definition 3.10]; * • For $v\not\in\Sigma_{\mathrm{min}}\cup\Sigma_{\mathrm{lr}}\cup\\{l\\}$, the representation $\overline{\rho}_{v}$ is unramified. Suppose that $\overline{\rho}$ is rigid for $(\Sigma_{\mathrm{min}},\Sigma_{\mathrm{lr}})$. We consider a global deformation problem in the sense of [Wang22a, Definition 2.2] of the form $\ \mathcal{S}^{\ast}=(\overline{\rho},\psi,\Sigma_{\mathrm{min}}\cup\Sigma_{\mathrm{lr}}\cup\\{p\\}\cup\\{l\\},\\{\mathcal{D}_{v}\\}_{v\in\Sigma_{\mathrm{min}}\cup\Sigma_{\mathrm{lr}}\cup\\{p\\}\cup\\{l\\}})$ where $\ast=\\{\mathrm{ram},\mathrm{unr},\mathrm{mix}\\}$. * • For $v\in\Sigma_{\mathrm{min}}$, $\mathcal{D}_{v}$ is the local deformation problem $\mathcal{D}^{\mathrm{min}}_{v}$ classifying all the minimal ramified liftings defined in [Wang22a, Defintion 3.4]; * • For $v\in\Sigma_{\mathrm{lr}}\cup\\{q\\}$, $\mathcal{D}_{v}$ is the local deformation problem $\mathcal{D}^{\mathrm{ram}}_{v}$ defined as in [Wang22a, Definition 3.6 (2)] classifying certain ramified liftings; * • For $v=l$, $\mathcal{D}_{v}$ is the local deformation problem $\mathcal{D}^{\mathrm{FL}}_{v}$ defined as in [Wang22a, Definition 3.1.1] classifying regular Fontaine–Laffaille crystalline liftings; * • For $v=p$, depending on $\ast\in\\{\mathrm{ram},\mathrm{unr},\mathrm{mix}\\}$, we have * – $\mathcal{D}_{v}$ is the local deformation problem $\mathcal{D}^{\mathrm{unr}}_{v}$ which is given by [Wang22a, Definition 3.6 (2)] when $\ast=\mathrm{unr}$; * – $\mathcal{D}_{v}$ is the local deformation problem $\mathcal{D}^{\mathrm{ram}}_{v}$ which is given by [Wang22a, Definition 3.6 (3)] when $\ast=\mathrm{ram}$; * – $\mathcal{D}_{v}$ is the local deformation problem $\mathcal{D}^{\mathrm{mix}}_{v}$ defined in [Wang22a, Definition 3.6 (1)] when $\ast=\mathrm{mix}$; * • For each global deformation problem $\mathcal{S}^{\ast}$ with $\ast=\\{\mathrm{ram},\mathrm{unr},\mathrm{mix}\\}$, we have its corresponding universal deformation ring denoted by $\mathrm{R}^{\ast}=\mathrm{R}^{\mathrm{univ}}_{\mathcal{S}^{\ast}}$ for $\ast=\\{\mathrm{ram},\mathrm{unr},\mathrm{mix}\\}$. ###### Construction 8.7. When $\mathrm{B}$ is indefinite, we choose the level $\mathrm{K}=\mathrm{K}_{\mathrm{Pa}}$ for the Shimura variety $\mathrm{Sh}(\mathrm{B},\mathrm{K}_{\mathrm{Pa}})$ in the following way. * • For $v\not\in\Sigma_{\mathrm{lr}}\cup\Sigma_{\mathrm{min}}$ or $v=l$, then $\mathrm{K}_{v}$ is hyperspecial; * • For $v\in\Sigma_{\mathrm{lr}}-\\{p,q\\}$, then $\mathrm{K}_{v}$ is the paramodular subgroup of $\mathrm{GSp}(\mathbb{Z}_{v})$; * • For $v\in\Sigma_{\mathrm{min}}$, then $\mathrm{K}_{v}$ is contained in the pro-$v$ Iwahori subgroup $\mathrm{Iw}_{1}(v)$; * • For $v=p,q$, then $\mathrm{K}_{v}$ is the paramodular subgroup $\mathbf{G}(\mathrm{B})(\mathbb{Z}_{v})$ of $\mathrm{GU}_{2}(\mathrm{D})$ and $\mathrm{GU}_{2}(\mathrm{D}^{\prime})$ respectively; Let $\Sigma_{\mathrm{Pa}}=\Sigma_{\mathrm{min}}\cup\Sigma_{\mathrm{lr}}\cup\\{p,q\\}$. We denote by $\mathbf{T}^{\mathrm{ram}}$ the image of $\mathbb{T}^{\Sigma_{\mathrm{Pa}}}$ in $\operatorname{End}_{\mathcal{O}_{\lambda}}(\mathrm{H}^{3}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}))).$ ###### Construction 8.8. We choose the level $\overline{\mathrm{K}}=\mathrm{K}_{\mathrm{H}}$ for the Shimura variety $\mathrm{Sh}(\overline{\mathrm{B}},\mathrm{K}_{\mathrm{H}})$ when $\overline{\mathrm{B}}$ is definite in such the following way. * • For $v\not\in\Sigma_{\mathrm{lr}}\cup\Sigma_{\mathrm{min}}$ or $v=l$ or $v=p$, then $\overline{\mathrm{K}}_{v}$ is hyperspecial; * • For $v\in\Sigma_{\mathrm{lr}}$, then $\overline{\mathrm{K}}_{v}$ is the paramodular subgroup of $\mathrm{GSp}(\mathbb{Z}_{v})$; * • For $v\in\Sigma_{\mathrm{min}}$, then $\overline{\mathrm{K}}_{v}$ is contained in the pro-$v$ Iwahori subgroup $\mathrm{Iw}_{1}(v)$; * • For $v=q$, then $\overline{\mathrm{K}}_{v}$ is the paramodular subgroup $\mathbf{G}(\overline{\mathrm{B}})(\mathbb{Z}_{v})$ of $\mathrm{GU}_{2}(\mathrm{D}^{\prime})$. Let $\Sigma_{\mathrm{H}}=\Sigma_{\mathrm{min}}\cup\Sigma_{\mathrm{lr}}\cup\\{q\\}$. We denote by $\mathbf{T}^{\mathrm{unr}}$ the image of $\mathbb{T}^{\Sigma_{\mathrm{H}}}$ in $\operatorname{End}_{\mathcal{O}_{\lambda}}(\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})])$. Let $\overline{\rho}$ be a Galois representation $\overline{\rho}:\mathrm{G}_{\mathbb{Q}}\rightarrow\mathrm{GSp}_{4}(k)$ and let $\mathfrak{m}$ be the maximal ideal corresponding to $\overline{\rho}$ in $\mathbb{T}^{\Sigma}$ for $\Sigma\in\\{\Sigma_{\mathrm{Pa}},\Sigma_{\mathrm{H}}\\}$, in the sense that the characteristic polynomial $\mathrm{det}(\mathrm{X}-\overline{\rho}(\mathrm{Frob}_{v}))$ is congruent to the Hecke polynomial $\mathrm{Q}_{v}(\mathrm{X})$ given by $\mathrm{X}^{4}-\mathrm{T}_{v,2}\mathrm{X}^{3}+(v\mathrm{T}_{v,1}+(v^{3}+v)\mathrm{T}_{v,0})\mathrm{X}^{2}-v^{3}\mathrm{T}_{v,0}\mathrm{T}_{v,2}\mathrm{X}+v^{6}\mathrm{T}^{2}_{v,0}$ modulo $\mathfrak{m}$ for each $v$ not in $\Sigma$. We will also assume that the image of $\overline{\rho}$ contains $\mathrm{GSp}_{4}(\mathbb{F}_{l})$ and that $\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}))_{\mathfrak{m}}=\mathrm{H}^{3}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}))_{\mathfrak{m}}.$ The following theorem is the main result of [Wang22a] proved using the Taylor- Wiles method. ###### Theorem 8.9. Let $\overline{\rho}$ and $\mathfrak{m}$ be given as above. Suppose the following assumptions hold. * (D1) the image of $\overline{\rho}(\mathrm{G}_{\mathbb{Q}})$ contains $\mathrm{GSp}_{4}(k)$; * (D2) $\overline{\rho}$ is rigid for $(\Sigma_{\mathrm{min}},\Sigma_{\mathrm{lr}})$; * (D3) For every finite set $\Sigma^{\prime}$ of nonarchimedean places containing $\Sigma$ and every open compact subgroup $\mathrm{K}^{\prime}$ of $\mathrm{K}$ satisfying $\mathrm{K}^{\prime}_{v}=\mathrm{K}_{v}$ for all $v\not\in\Sigma^{\prime}$, we have $\mathrm{H}^{d}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(k))_{\mathfrak{m}^{\prime}}=0$ for $d\neq 3$ where $\mathfrak{m}^{\prime}=\mathfrak{m}\cap\mathbb{T}^{\Sigma^{\prime}}$. Then the following holds true. 1. (1) If $\mathbf{T}^{\mathrm{unr}}_{\mathfrak{m}}$ is non-zero, then we have an isomorphim of complete intersection rings: $\mathrm{R}^{\mathrm{unr}}=\mathbf{T}^{\mathrm{unr}}_{\mathfrak{m}}$ and $\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]_{\mathfrak{m}}$ is a finite free $\mathbf{T}^{\mathrm{unr}}_{\mathfrak{m}}$-module. 2. (2) If $\mathbf{T}^{\mathrm{ram}}_{\mathfrak{m}}$ is non-zero, then we have an isomorphim of complete intersection rings: $\mathrm{R}^{\mathrm{ram}}=\mathbf{T}^{\mathrm{ram}}_{\mathfrak{m}}$ and $\mathrm{H}^{3}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}))_{\mathfrak{m}}$ is a finite free $\mathbf{T}^{\mathrm{ram}}_{\mathfrak{m}}$-module. ###### Remark 8.10. Let $\pi$ be a cuspidal automorphic representation of general type. Assume that $\overline{\rho}_{\pi,\lambda}(\mathrm{G}_{\mathbb{Q}})$ contains $\mathrm{GSp}_{4}(\mathbb{F}_{l})$ for sufficiently large $l$, then we have shown that $\overline{\rho}_{\pi,\lambda}$ is indeed rigid, see [Wang22a, Theorem 5.5]. ## 9\. Arithmetic level raising and applications Now we are ready to prove our main result on arithmetic level raising on the quaternionic Shimura varieties studied in the previous sections. First we recall the setting we are in. Let $\pi$ be a cuspidal automorphic representation of $\mathrm{GSp}_{4}(\mathbb{A}_{\mathbb{Q}})$ with weight $(3,3)$ and trivial central character. Suppose that $\pi$ is of general type and $E$ be the coefficient field of $\pi$. Let $\Sigma_{\pi}$ be the minimal set of finite places outside of which $\pi$ is unramified. We fix an isomorphism $\iota_{l}:\mathbb{C}\xrightarrow{\sim}\overline{\mathbb{Q}}_{l}$ which induces a place $\lambda$ in $E$ over $l$. We fix the following datum. * • A finite set $\Sigma_{\mathrm{min}}$ of places of $\mathbb{Q}$ contained in $\Sigma_{\pi}$; * • A finite set $\Sigma_{\mathrm{lr}}$ of places of $\mathbb{Q}$ which is disjoint from $\Sigma_{\mathrm{min}}$ and contained in $\Sigma_{\mathrm{lr}}$; * • A finite set of places $\Sigma$ containing $\Sigma_{\mathrm{min}}\cup\Sigma_{\mathrm{lr}}$ which is away from $l$. * • We have, by Construction 3.3 $(2)$, a homomorphism $\phi_{\pi}:\mathbb{T}^{\Sigma}\rightarrow\mathcal{O}_{E}.$ and its $\lambda$-adic avatar $\phi_{\pi,\lambda}:\mathbb{T}^{\Sigma}\rightarrow\mathcal{O}_{\lambda}$ for the valuation ring $\mathcal{O}_{\lambda}\subset E_{\lambda}$. * • We can associate to it a Galois representation $\rho_{\pi,\lambda}:\mathrm{G}_{\mathbb{Q}}\rightarrow\mathrm{GSp}_{4}(E_{\lambda})$ which we assume is residually absolutely irreducible as in Assumption 8.4. * • Let $p$ be a prime which is level raising special of length $m$ for $\pi$ in the sense of Definition 3.2. * • Let $q$ be a prime such that $l\nmid q^{2}-1$. Suppose that $\pi$ is of type $\mathrm{II}\mathrm{a}$ and $q$ is contained in $\Sigma_{\mathrm{lr}}$. In this case $\pi_{q}$ admits Jacquet-Langlands transfer to a representation of the group $\mathrm{GU}_{2}(\mathrm{D}^{\prime})$ where $\mathrm{D}^{\prime}$ which is the quaternion division algebra over $\mathbb{Q}_{q}$. * • We choose the level $\mathrm{K}_{\mathrm{Pa}}$ and $\Sigma_{\mathrm{Pa}}$ as in Construction 8.7 for the Shimura varieity $\mathrm{Sh}(\mathrm{B},\mathrm{K}_{\mathrm{Pa}})$ whose integral model is $\mathrm{X}_{\mathrm{Pa}}(\mathrm{B})$ and whose special fiber is $\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B})$; and choose the level $\mathrm{K}_{\mathrm{H}}$ and $\Sigma_{\mathrm{H}}$ as in Construction 8.8 for the Shimura set $\mathrm{Sh}(\overline{\mathrm{B}},\mathrm{K}_{\mathrm{H}})$ which is also denoted by $\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})$. * • We introduced two ideals (9.1) $\displaystyle\mathfrak{m}=\mathbb{T}^{\Sigma\cup\\{p\\}}\cap\ker(\mathbb{T}^{\Sigma}\xrightarrow{\phi_{\pi,\lambda}}\mathcal{O}_{\lambda}\rightarrow\mathcal{O}_{\lambda}/\lambda)$ $\displaystyle\mathfrak{n}=\mathbb{T}^{\Sigma\cup\\{p\\}}\cap\ker(\mathbb{T}^{\Sigma}\xrightarrow{\phi_{\pi,\lambda}}\mathcal{O}_{\lambda}\rightarrow\mathcal{O}_{\lambda}/\lambda^{m})$ as in Construction 3.3 $(3)$ and we assume that $\mathfrak{m}$ is generic level raising at $p$ in the sense of 7.8. ###### Construction 9.2. We define the following map $\nabla:\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]\oplus\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]\rightarrow\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]$ by sending $(x,y)\rightarrow(\mathrm{T}^{-1}_{p,0}\mathrm{T}_{p,1}+(p+1)(p^{2}+1))x-(p+1)\mathrm{T}_{20}y$ for $(x,y)\in\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]\oplus\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]$. We also write $\nabla_{\mathfrak{m}}:\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]_{\mathfrak{m}}\oplus\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]_{\mathfrak{m}}\rightarrow\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]_{\mathfrak{m}}$ for the localization of $\nabla$ at $\mathfrak{m}$ Note that the composition of the level raising matrix $\mathcal{T}_{\mathrm{lr},\mathfrak{m}}:\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]_{\mathfrak{m}}\oplus\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]_{\mathfrak{m}}\rightarrow\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]_{\mathfrak{m}}\oplus\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]_{\mathfrak{m}}$ with the map $\nabla_{\mathfrak{m}}$ is given by sending $(x,y)\rightarrow\mathrm{det}\phantom{.}\mathcal{T}_{\mathrm{lr},\mathfrak{m}}\cdot x$ for $(x,y)\in\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]\oplus\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]$ where we have $\mathrm{det}\phantom{.}\mathcal{T}_{\mathrm{lr},\mathfrak{m}}=(\mathrm{T}^{-1}_{p,0}\mathrm{T}_{p,1}+(p+1)(p^{2}+1))^{2}-(p+1)^{2}\mathrm{T}_{20}\circ\mathrm{T}_{02}$ by Proposition 6.10. ###### Proposition 9.3. Let $p$ be a level raising special prime for $\pi$ and $\mathfrak{m}$ be the maximal ideal as above. We have a surjective morphism $\mathrm{F}_{-1}\mathrm{H}^{1}(\mathrm{I}_{\mathbb{Q}_{p^{2}}},\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(1))_{\mathfrak{m}})\twoheadrightarrow\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\mathrm{B})]_{\mathfrak{m}}/\mathrm{det}\phantom{.}\mathcal{T}_{\mathrm{lr},\mathfrak{m}}$ whose kernel can be identified with the torsion part of $\mathrm{H}^{4}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(2))_{\mathfrak{m}}$. ###### Proof. Note that, by Proposition 5.4, we have an isomorphism $\mathrm{F}_{-1}\mathrm{H}^{1}(\mathrm{I}_{\mathbb{Q}_{p^{2}}},\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda})(1))_{\mathfrak{m}})\xrightarrow{\sim}\mathrm{Coker}(\alpha\circ\mathrm{N}^{-1}_{\Sigma}\circ\beta)_{\mathfrak{m}}.$ On the other hand, by Proposition 7.9 we have an isomorphism $(\mathrm{inc}^{\\{0\\}}_{!,\mathfrak{m}}+\mathrm{inc}^{\\{2\\}}_{!,\mathfrak{m}}):\mathcal{O}_{\lambda}[\mathrm{Z}_{\\{0\\}}(\overline{\mathrm{B}})]_{\mathfrak{m}}\oplus\mathcal{O}_{\lambda}[\mathrm{Z}_{\\{2\\}}(\overline{\mathrm{B}})]_{\mathfrak{m}}\rightarrow\mathrm{H}^{2}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(1))_{\mathfrak{m}}$ and a surjection $(\mathrm{inc}^{\ast}_{\\{0\\},\mathfrak{m}},\mathrm{inc}^{\ast}_{\\{2\\},\mathfrak{m}}):\mathrm{H}^{4}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(2))_{\mathfrak{m}}\rightarrow\mathcal{O}_{\lambda}[\mathrm{Z}_{\\{0\\}}(\overline{\mathrm{B}})]_{\mathfrak{m}}\oplus\mathcal{O}_{\lambda}[\mathrm{Z}_{\\{2\\}}(\overline{\mathrm{B}})]_{\mathfrak{m}}$ whose kernel is given by the torsion part of $\mathrm{H}^{4}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(2))_{\mathfrak{m}}$. It follows that we have a surjection $\mathrm{F}_{-1}\mathrm{H}^{1}(\mathrm{I}_{\mathbb{Q}_{p^{2}}},\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda})(1))_{\mathfrak{m}})\rightarrow\frac{\mathcal{O}_{\lambda}[\mathrm{Z}_{\\{0\\}}(\overline{\mathrm{B}})]_{\mathfrak{m}}\oplus\mathcal{O}_{\lambda}[\mathrm{Z}_{\\{2\\}}(\overline{\mathrm{B}})]_{\mathfrak{m}}}{\mathcal{T}_{\mathrm{lr},\mathfrak{m}}(\mathcal{O}_{\lambda}[\mathrm{Z}_{\\{0\\}}(\overline{\mathrm{B}})]_{\mathfrak{m}}\oplus\mathcal{O}_{\lambda}[\mathrm{Z}_{\\{2\\}}(\overline{\mathrm{B}})]_{\mathfrak{m}})}.$ The kernel of this map is given by $\frac{\mathrm{H}^{4}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(2))^{\mathrm{tor}}_{\mathfrak{m}}}{\mathrm{Im}(\alpha\circ\mathrm{N}^{-1}_{\Sigma}\circ\beta)_{\mathfrak{m}}\cap\mathrm{H}^{4}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(2))^{\mathrm{tor}}_{\mathfrak{m}}}.$ Composing this with the map $\nabla_{\mathfrak{m}}:\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]_{\mathfrak{m}}\oplus\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]_{\mathfrak{m}}\rightarrow\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]_{\mathfrak{m}}$, we therefore obtain a map $\mathrm{F}_{-1}\mathrm{H}^{1}(\mathrm{I}_{\mathbb{Q}_{p^{2}}},\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda})(1))_{\mathfrak{m}})\rightarrow\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]_{\mathfrak{m}}/\mathrm{det}\phantom{.}\mathcal{T}_{\mathrm{lr},\mathfrak{m}}.$ which is surjective by Nakayama’s lemma. It remains to show that $\mathrm{Im}(\alpha\circ\mathrm{N}^{-1}_{\Sigma}\circ\beta)_{\mathfrak{m}}\cap\mathrm{H}^{4}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(2))^{\mathrm{tor}}_{\mathfrak{m}}$ is zero. By [Nek07, Proposition 4.2.2(1)], the group $\mathrm{F}_{-1}\mathrm{H}^{1}(\mathrm{I}_{\mathbb{Q}_{p^{2}}},\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(E_{\lambda})(1))_{\mathfrak{m}})$ vanishes. It follows that the $\mathcal{O}_{\lambda}$-rank of $\mathrm{Im}(\alpha\circ\mathrm{N}^{-1}_{\Sigma}\circ\beta)_{\mathfrak{m}}$ is the same as the $\mathcal{O}_{\lambda}$-rank of $\mathrm{H}^{4}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(2))_{\mathfrak{m}}$. Next, using the excision exact sequences, we have isomorphisms (9.4) $\displaystyle\mathrm{H}^{4}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),E_{\lambda}(2))\xrightarrow{\sim}\mathrm{H}^{4}(\overline{\mathrm{X}}^{\mathrm{sm}}_{\mathrm{Pa}}(\mathrm{B}),E_{\lambda}(2))$ $\displaystyle\mathrm{H}^{2}_{\mathrm{c}}(\overline{\mathrm{X}}^{\mathrm{sm}}_{\mathrm{Pa}}(\mathrm{B}),E_{\lambda}(1))\xrightarrow{\sim}\mathrm{H}^{2}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),E_{\lambda}(1))$ where $\overline{\mathrm{X}}^{\mathrm{sm}}_{\mathrm{Pa}}(\mathrm{B})=\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B})-\overline{\mathrm{X}}^{\mathrm{sing}}_{\mathrm{Pa}}(\mathrm{B})$, using the fact that the singular locus $\overline{\mathrm{X}}^{\mathrm{sing}}_{\mathrm{Pa}}(\mathrm{B})$ is zero dimensional. Therefore Poincare duality for $\overline{\mathrm{X}}^{\mathrm{sm}}_{\mathrm{Pa}}(\mathrm{B})$ implies that we have $\dim_{E_{\lambda}}\mathrm{H}^{4}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),E_{\lambda}(2))_{\mathfrak{m}}=\dim_{E_{\lambda}}\mathrm{H}^{2}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),E_{\lambda}(1))_{\mathfrak{m}}.$ Since the source $\mathrm{H}^{2}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(1))_{\mathfrak{m}}$ of $\mathrm{Im}(\alpha\circ\mathrm{N}^{-1}_{\Sigma}\circ\beta)_{\mathfrak{m}}$ is free over $\mathcal{O}_{\lambda}$, the intersection $\mathrm{Im}(\alpha\circ\mathrm{N}^{-1}_{\Sigma}\circ\beta)_{\mathfrak{m}}\cap\mathrm{H}^{4}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(2))^{\mathrm{tor}}_{\mathfrak{m}}$ is trivial. The proposition is proved. ∎ Now we are ready to state and prove our main result on arithmetic level raising for the quaternionic Shimura variety studied in this article. We first recall a list of running assumptions below. * • We assume that $\mathrm{H}^{i}_{(\mathrm{c})}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}))_{\mathfrak{m}}=0$ for $i\neq 3$, and that $\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}))_{\mathfrak{m}}\xrightarrow{\sim}\mathrm{H}^{3}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}))_{\mathfrak{m}}$ is a finite free $\mathcal{O}_{\lambda}$ module. This is Assumption 7.6. * • We assume that for $\pi$, the associated Galois representation $\rho_{\pi,\lambda}:\mathrm{G}_{\mathbb{Q}}\rightarrow\mathrm{GSp}_{4}(E_{\lambda})$ is residually absolutely irreducible. This is Assumption 8.4. ###### Theorem 9.5 (Arithmetic level raising for $\mathrm{GSp}_{4}$). Suppose that $\pi$ is a cuspidal automorphic representation of $\mathrm{GSp}_{4}(\mathbb{A}_{\mathbb{Q}})$ of general type. We assume that the assumptions in 7.6, 8.4 recalled above are effective. Let $p$ be a level raising special prime for $\pi$ of depth $m$. We assume further that 1. (1) $l\geq 3$ and $l\nmid p^{2}-1$; 2. (2) $\overline{\rho}_{\pi,\lambda}$ is rigid for $(\Sigma_{\mathrm{min}},\Sigma_{\mathrm{lr}})$ as in Definition 8.6; 3. (3) The image of $\overline{\rho}_{\pi,\lambda}(\mathrm{G}_{\mathbb{Q}})$ contains $\mathrm{GSp}_{4}(\mathbb{F}_{l})$; 4. (4) $\mathfrak{m}$ is in the support of $\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]$ and is generic level raising. Then following holds. 1. (1) There is an isomorphism $\mathrm{H}^{1}_{\mathrm{sing}}(\mathbb{Q}_{p^{2}},\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}(1)))_{\mathfrak{m}})\xrightarrow{\sim}\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]_{\mathfrak{m}}/\mathrm{det}\phantom{.}\mathcal{T}_{\mathrm{lr}}.$ 2. (2) The above isomorphism induces an isomorphism $\mathrm{H}^{1}_{\mathrm{sing}}(\mathbb{Q}_{p^{2}},\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}/\lambda^{m}(1)))_{\mathfrak{m}})\xrightarrow{\sim}\mathcal{O}_{\lambda}/\lambda^{m}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]_{\mathfrak{m}}.$ ### 9.1. Proof of Theorem 9.5 Consider the universal deformation problem $\mathcal{S}^{\ast}=(\overline{\rho}_{\pi,\lambda},\psi,\Sigma_{\mathrm{min}}\cup\Sigma_{\mathrm{lr}}\cup\\{p\\}\cup\\{l\\},\\{\mathcal{D}_{v}\\}_{v\in\Sigma_{\mathrm{min}}\cup\Sigma_{\mathrm{lr}}\cup\\{p\\}\cup\\{l\\}})$ where $\ast=\\{\mathrm{ram},\mathrm{unr},\mathrm{mix}\\}$. Recall that $\mathrm{Spf}\phantom{.}\mathrm{R}^{\mathrm{ram}}$ and $\mathrm{Spf}\phantom{.}\mathrm{R}^{\mathrm{unr}}$ are subspaces of $\mathrm{Spf}\phantom{.}\mathrm{R}^{\mathrm{mix}}$. Thus we have surjections $\displaystyle\mathrm{R}^{\mathrm{mix}}\twoheadrightarrow\mathrm{R}^{\mathrm{ram}}$ $\displaystyle\mathrm{R}^{\mathrm{mix}}\twoheadrightarrow\mathrm{R}^{\mathrm{unr}}.$ we define $\mathrm{R}^{\mathrm{cong}}=\mathrm{R}^{\mathrm{unr}}\otimes_{\mathrm{R}^{\mathrm{mix}}}\mathrm{R}^{\mathrm{ram}}$. We have a universal lifting $\rho^{\mathrm{mix}}:\mathrm{G}_{\mathbb{Q}}\rightarrow\mathrm{GSp}_{4}(\mathrm{R}^{\mathrm{mix}}).$ It also induces two representations $\displaystyle\rho^{\mathrm{ram}}:\mathrm{G}_{\mathbb{Q}}\rightarrow\mathrm{GSp}_{4}(\mathrm{R}^{\mathrm{ram}})$ $\displaystyle\rho^{\mathrm{unr}}:\mathrm{G}_{\mathbb{Q}}\rightarrow\mathrm{GSp}_{4}(\mathrm{R}^{\mathrm{unr}})$ by the above surjections. Denote by $\mathrm{P}_{\mathbb{Q}_{p}}$ be the maximal closed subgroup of the inertia subgroup $\mathrm{I}_{\mathbb{Q}_{p}}\subset G_{\mathbb{Q}_{p}}$ of pro-order coprime to $l$. Then $G_{\mathbb{Q}_{p}}/\mathrm{P}_{\mathbb{Q}_{p}}\cong t^{\mathbb{Z}_{l}}\rtimes\phi^{\widehat{\mathbb{Z}}}_{p}$ is a $p$-tame group. By definition, $\rho^{\mathrm{mix}}$ is trivial on $\mathrm{P}_{\mathbb{Q}_{p}}$. Let $\overline{\mathrm{v}}_{0}$ and $\overline{\mathrm{v}}_{1}$ be eigenvectors in $k^{4}=(\mathcal{O}_{\lambda}/\lambda)^{4}$ for $\rho^{\mathrm{mix}}(\phi^{2}_{p})$ with eigenvalues $p^{-2}$ and $p^{-4}$, respectively. By Hensel’s lemma, they lift to $\mathrm{v}_{0}$ and $\mathrm{v}_{1}$ in $(\mathrm{R}^{\mathrm{mix}})^{\oplus 4}$ for $\rho^{\mathrm{mix}}(\phi^{2}_{p})$ with eigenvalues $\mathrm{s}_{0}$ and $\mathrm{s}_{1}$ lifting $p^{-2}$ and $p^{-4}$. Let $\mathrm{x}\in\mathrm{R}^{\mathrm{mix}}$ be the unique element such that $\rho^{\mathrm{mix}}(t)\mathrm{v}_{1}=\mathrm{x}\mathrm{v}_{0}+\mathrm{v}_{1}$. It follows from the relation $\rho^{\mathrm{mix}}(\phi^{2}_{p})\rho^{\mathrm{mix}}(t)\rho^{\mathrm{mix}}(\phi^{-2}_{p})=\rho^{\mathrm{mix}}(t)^{p^{2}}$ that $\mathrm{x}(\mathrm{s}-p^{-2})=0$. By the definitions of $\mathrm{R}^{\ast}$ for $\ast=\\{\mathrm{ram},\mathrm{unr},\mathrm{mix}\\}$, we have $\displaystyle\mathrm{R}^{\mathrm{unr}}=\mathrm{R}^{\mathrm{mix}}/(\mathrm{x})$ $\displaystyle\mathrm{R}^{\mathrm{ram}}=\mathrm{R}^{\mathrm{mix}}/(\mathrm{s}_{0}-p^{-2})$ $\displaystyle\mathrm{R}^{\mathrm{cong}}=\mathrm{R}^{\mathrm{mix}}/(\mathrm{x},\mathrm{s}_{0}-p^{-2}).$ By Theorem 8.9, $\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]_{\mathfrak{m}}$ is a free over $\mathbf{T}^{\mathrm{unr}}_{\mathfrak{m}}$ and we have an isomorphism $\mathrm{R}^{\mathrm{unr}}\cong\mathbf{T}_{\mathfrak{m}}$. Thus $\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]_{\mathfrak{m}}$ is a free $\mathrm{R}^{\mathrm{unr}}$-module of rank $d_{\mathrm{unr}}$. Note that the characteristic polynomial of $\rho^{\mathrm{mix}}(\phi^{2}_{p})$ can written as $(\mathrm{X}-\mathrm{s}_{0})(\mathrm{X}-p^{-6}\mathrm{s}^{-1}_{0})\mathrm{Q}(\mathrm{X})$ where $\mathrm{Q}(\mathrm{X})$ is a polynomial in $\mathrm{R}^{\mathrm{unr}}[\mathrm{X}]$ whose reduction in $\mathcal{O}_{\lambda}/\lambda[\mathrm{X}]=k[\mathrm{X}]$ does not contain $p^{-2}$ and $p^{-4}$ as its roots. By Proposition 6.15 and the equations $\displaystyle(\alpha_{p}+p^{3}\alpha^{-1}_{p}-p(p+1))(\alpha_{p}+p^{3}\alpha^{-1}_{p}+p(p+1))=\alpha^{2}_{p}+p^{6}\alpha^{-2}_{p}-p^{2}(p^{2}+1),$ $\displaystyle(\beta_{p}+p^{3}\beta^{-1}_{p}-p(p+1))(\beta_{p}+p^{3}\beta^{-1}_{p}+p(p+1))=\beta^{2}_{p}+p^{6}\beta^{-2}_{p}-p^{2}(p^{2}+1),$ we have $(\mathrm{det}\phantom{.}\mathcal{T}_{\mathrm{lr},\mathfrak{m}})\cdot\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]_{\mathfrak{m}}=(\mathrm{s}_{0}-p^{-2})\cdot\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]_{\mathfrak{m}}.$ Thus we have $\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]_{\mathfrak{m}}/\mathrm{det}\phantom{.}\mathcal{T}_{\mathrm{lr},\mathfrak{m}}\cong\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]_{\mathfrak{m}}\otimes_{\mathrm{R}^{\mathrm{unr}}}\mathrm{R}^{\mathrm{cong}}$ and it follows that $\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]_{\mathfrak{m}}/\mathrm{det}\phantom{.}\mathcal{T}_{\mathrm{lr},\mathfrak{m}}$ is a free $\mathrm{R}^{\mathrm{cong}}$-module of rank $d_{\mathrm{unr}}$. Since we have a surjection $\mathrm{F}_{-1}\mathrm{H}^{1}(\mathrm{I}_{\mathbb{Q}_{p^{2}}},\mathrm{H}^{3}_{c}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}(1)))_{\mathfrak{m}})\twoheadrightarrow\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\mathrm{B})]_{\mathfrak{m}}/\mathrm{det}\phantom{.}\mathcal{T}_{\mathrm{lr},\mathfrak{m}}$ by Proposition 9.3, we have $\mathbf{T}^{\mathrm{ram}}_{\mathfrak{m}}$ is non- zero. Moreover $\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}(1)))_{\mathfrak{m}}$ is a free $\mathrm{R}^{\mathrm{ram}}$-module by Theorem 8.9 $(2)$, say of rank $d_{\mathrm{ram}}$. Consider the $\mathrm{R}^{\mathrm{ram}}$-module $\mathrm{H}:=\operatorname{Hom}_{\mathrm{G}_{\mathbb{Q}}}((\mathrm{R}^{\mathrm{ram}})^{\oplus 4},\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}(1)))_{\mathfrak{m}})$ where $\mathrm{G}_{\mathbb{Q}}$ acts on $(\mathrm{R}^{\mathrm{ram}})^{\oplus 4}$ via $\rho^{\mathrm{ram}}$ which is free of rank $d_{\mathrm{ram}}$. It follows this that we have an isomorphism $\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}))_{\mathfrak{m}}\xrightarrow{\sim}\mathrm{H}\otimes_{\mathrm{R}^{\mathrm{ram}}}(\mathrm{R}^{\mathrm{ram}})^{\oplus 4}.$ It follows from that $\mathrm{H}^{1}_{\mathrm{sing}}(\mathbb{Q}_{p^{2}},\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}(1)))_{\mathfrak{m}})\xrightarrow{\sim}\mathrm{H}\otimes_{\mathrm{R}^{\mathrm{ram}}}\mathrm{H}^{1}_{\mathrm{sing}}(\mathbb{Q}_{p^{2}},(\mathrm{R}^{\mathrm{ram}})^{\oplus 4}(1))$ We still denote by $\mathrm{v}_{0}$ and $\mathrm{v}_{1}$ the projections of $\mathrm{v}_{0}$ and $\mathrm{v}_{1}$ in $(\mathrm{R}^{\mathrm{ram}})^{\oplus 4}$ of $\mathrm{G}_{\mathbb{Q}}$-modules. Then a straightforward computation shows that $\mathrm{H}^{1}_{\mathrm{sing}}(\mathbb{Q}_{p^{2}},(\mathrm{R}^{\mathrm{ram}})^{\oplus 4}(1))\xrightarrow{\sim}\mathrm{R}^{\mathrm{ram}}\mathrm{v}_{0}/\mathrm{x}\mathrm{v}_{0}\cong\mathrm{R}^{\mathrm{cong}}.$ Hence we have $\mathrm{H}^{1}_{\mathrm{sing}}(\mathbb{Q}_{p^{2}},\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}(1)))_{\mathfrak{m}})$ is a free $\mathrm{R}^{\mathrm{cong}}$-module of rank $d_{\mathrm{ram}}$. ###### Lemma 9.6. We have an equality $d_{\mathrm{unr}}=d_{\mathrm{ram}}.$ ###### Proof. Let $\eta^{\mathrm{unr}}\in\mathrm{Spec}\phantom{.}\mathrm{R}^{\mathrm{unr}}[1/l](\overline{\mathbb{Q}}_{l})$ be in the support of $\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]_{\mathfrak{m}}$ which gives rise to an automorphic representation $\pi^{\mathrm{unr}}$ of $\mathrm{GSp}_{4}(\mathbb{A}_{\mathbb{Q}})$ such that $\overline{\rho}_{\pi,\lambda}\otimes_{k}\overline{\mathbb{F}}_{l}$ and $\rho_{\pi^{\mathrm{unr}},\iota_{l}}$ are residually isomorphic. Then we have $d_{\mathrm{unr}}=\dim_{\overline{\mathbb{Q}}_{l}}\overline{\mathbb{Q}}_{l}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})][\iota_{l}\phi_{\pi^{\mathrm{unr}}}].$ Similarly, let $\eta^{\mathrm{ram}}\in\mathrm{Spec}\phantom{.}\mathrm{R}^{\mathrm{ram}}[1/l](\overline{\mathbb{Q}}_{l})$ be in the support of $\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}(1)))_{\mathfrak{m}}$ which gives rise to an automorphic representation $\pi^{\mathrm{ram}}$ of $\mathrm{GSp}_{4}(\mathbb{A}_{\mathbb{Q}})$ such that $\rho_{\pi^{\mathrm{ram}},\iota_{l}}$ is residually isomorphic to $\overline{\rho}_{\pi,\lambda}\otimes_{k}\overline{\mathbb{F}}_{l}$. Then we have $4d_{\mathrm{ram}}=\dim_{\overline{\mathbb{Q}}_{l}}\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\overline{\mathbb{Q}}_{l}))[\iota_{l}\phi_{\pi^{\mathrm{ram}}}].$ Note that $\pi^{\mathrm{unr}}$ and $\pi^{\mathrm{ram}}$ are necessarily of general type. By considering the transfer of $\pi^{\mathrm{unr}}$ and $\pi^{\mathrm{ram}}$ to $\mathrm{GL}_{4}(\mathbb{A}_{\mathbb{Q}})$, we have $d_{\mathrm{ram}}=d_{\mathrm{unr}}$ by [LTXZZ, Lemma 6.4.2]. ∎ _Proof of Theorem 9.5._ Now we can prove our main theorem on arithmetic level raising. We have proved that we have a surjection $\mathrm{F}_{-1}\mathrm{H}^{1}(\mathrm{I}_{\mathbb{Q}_{p^{2}}},\mathrm{H}^{3}_{c}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}(1)))_{\mathfrak{m}})\twoheadrightarrow\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\mathrm{B})]_{\mathfrak{m}}/\mathrm{det}\phantom{.}\mathcal{T}_{\mathrm{lr},\mathfrak{m}}$ where the target is free of rank $d_{\mathrm{unr}}$ over $\mathrm{R}^{\mathrm{cong}}$. On the other hand, we have an injection $\mathrm{F}_{-1}\mathrm{H}^{1}(\mathrm{I}_{\mathbb{Q}_{p^{2}}},\mathrm{H}^{3}_{c}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}(1)))_{\mathfrak{m}})\hookrightarrow\mathrm{H}^{1}_{\mathrm{sing}}(\mathbb{Q}_{p^{2}},\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}(1))_{\mathfrak{m}})$ and the target is free of rank $d_{\mathrm{ram}}$ over $\mathrm{R}^{\mathrm{cong}}$. Since $d_{\mathrm{unr}}=d_{\mathrm{ram}}$, we have the desired isomorphism (9.7) $\mathrm{H}^{1}_{\mathrm{sing}}(\mathbb{Q}_{p^{2}},\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}(1)))_{\mathfrak{m}})\xrightarrow{\sim}\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]_{\mathfrak{m}}/\mathrm{det}\phantom{.}\mathcal{T}_{\mathrm{lr}}.$ This finishes the proof of the first part of the theorem. For the second part, by the freeness of $\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}(1)))_{\mathfrak{m}}$ as a $\mathbf{T}^{\mathrm{ram}}_{\mathfrak{m}}$-module. It follows that the natural map $\mathrm{H}^{1}(\mathrm{I}_{\mathbb{Q}_{p^{2}}},\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}(1)))_{\mathfrak{m}})/\mathfrak{n}\xrightarrow{\sim}\mathrm{H}^{1}(\mathrm{I}_{\mathbb{Q}_{p^{2}}},\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}(1)))_{\mathfrak{m}}/\mathfrak{n})$ is an isomorphism. We also have the short exact sequence as in Proposition 6.1 $\displaystyle 0$ $\displaystyle\rightarrow\mathrm{F}_{-1}\mathrm{H}^{1}(\mathrm{I}_{\mathbb{Q}_{p^{2}}},\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}(1)))_{\mathfrak{m}})\rightarrow\mathrm{H}^{1}(\mathrm{I}_{\mathbb{Q}_{p^{2}}},\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda})(1))_{\mathfrak{m}})$ $\displaystyle\rightarrow\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}(1)))_{\mathfrak{m}}/\mathrm{F}_{-1}\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}(1)))_{\mathfrak{m}}\rightarrow 0$ which is split by the $\mathrm{G}_{\mathbb{F}_{p^{2}}}$-action. And it follows that $\mathrm{F}_{-1}\mathrm{H}^{1}(\mathrm{I}_{\mathbb{Q}_{p^{2}}},\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}(1)))_{\mathfrak{m}})/\mathfrak{n}\xrightarrow{\sim}\mathrm{F}_{-1}\mathrm{H}^{1}(\mathrm{I}_{\mathbb{Q}_{p^{2}}},\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}(1)))_{\mathfrak{m}}/\mathfrak{n})$ is an isomorphism. Since we have shown above that there is an isomorphism $\mathrm{F}_{-1}\mathrm{H}^{1}(\mathrm{I}_{\mathbb{Q}_{p^{2}}},\mathrm{H}^{3}_{c}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}(1)))_{\mathfrak{m}})\xrightarrow{\sim}\mathrm{H}^{1}_{\mathrm{sing}}(\mathbb{Q}_{p^{2}},\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}(1))_{\mathfrak{m}}),$ it follows that $\mathrm{H}^{1}_{\mathrm{sing}}({\mathbb{Q}_{p^{2}}},\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}(1)))_{\mathfrak{m}})/\mathfrak{n}\xrightarrow{\sim}\mathrm{H}^{1}_{\mathrm{sing}}({\mathbb{Q}_{p^{2}}},\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}(1)))_{\mathfrak{m}}/\mathfrak{n}).$ On the other hand, we have $\mathbf{T}^{\mathrm{ram}}/\mathfrak{n}\cong\mathcal{O}_{\lambda}/\lambda^{m}$ and thus $\mathrm{H}^{1}_{\mathrm{sing}}({\mathbb{Q}_{p^{2}}},\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}(1)))_{\mathfrak{m}}/\mathfrak{n})\cong\mathrm{H}^{1}_{\mathrm{sing}}({\mathbb{Q}_{p^{2}}},\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}/\lambda^{m}(1)))_{\mathfrak{m}}$ by the freeness of $\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}(1)))_{\mathfrak{m}}$ as $\mathbf{T}^{\mathrm{ram}}_{\mathfrak{m}}$-module. Finally the second claim follows from the observation that $\displaystyle\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]_{\mathfrak{m}}/\mathrm{det}\phantom{.}\mathcal{T}_{\mathrm{lr}}$ $\displaystyle\cong\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]_{\mathfrak{m}}\otimes_{\mathrm{R}^{\mathrm{unr}}}\mathrm{R}^{\mathrm{cong}}$ $\displaystyle\cong\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]_{\mathfrak{m}}\otimes_{\mathrm{R}^{\mathrm{unr}}}\mathbf{T}^{\mathrm{ram}}_{\mathfrak{m}}/\mathfrak{n}$ $\displaystyle\cong\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]_{\mathfrak{m}}\otimes\mathcal{O}_{\lambda}/\lambda^{m}$ and the isomorphism 9.7. $\square$ ###### Corollary 9.8. We maintain the assumptions in Theorem 9.5. 1. (1) $\mathrm{H}^{4}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda})_{\mathfrak{m}}$ is a torsion free $\mathcal{O}_{\lambda}$-module. In particular, the map from Construction 6.4 $(\mathrm{inc}^{\ast}_{\\{0\\},\mathfrak{m}},\mathrm{inc}^{\ast}_{\\{2\\},\mathfrak{m}}):\mathrm{H}^{4}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(2))_{\mathfrak{m}}\rightarrow\mathcal{O}_{\lambda}[\mathrm{Z}_{\\{0\\}}(\overline{\mathrm{B}})]_{\mathfrak{m}}\oplus\mathcal{O}_{\lambda}[\mathrm{Z}_{\\{2\\}}(\overline{\mathrm{B}})]_{\mathfrak{m}}$ is an isomorphism. 2. (2) $\mathrm{H}^{3}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda})_{\mathfrak{m}}$ and $\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda})_{\mathfrak{m}}$ are both torsion free $\mathcal{O}_{\lambda}$-modules. ###### Proof. Recall we have proved that the kernel of the natural map $\mathrm{F}_{-1}\mathrm{H}^{1}(\mathrm{I}_{\mathbb{Q}_{p^{2}}},\mathrm{H}^{n}_{c}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(1))_{\mathfrak{m}})\twoheadrightarrow\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\mathrm{B})]_{\mathfrak{m}}/\mathrm{det}\phantom{.}\mathcal{T}_{\mathrm{lr},\mathfrak{m}}$ can be identified with the torsion part of $\mathrm{H}^{4}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(2))_{\mathfrak{m}}$. We have shown in the above proof that this surjection is an isomorphism and the result follows. For the second part, we have an injection $\mathrm{H}^{3}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(2))_{\mathfrak{m}}\hookrightarrow\mathrm{H}^{3}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}(2)))_{\mathfrak{m}}$ and $\mathrm{H}^{3}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}(2)))_{\mathfrak{m}}$ is a torsion free $\mathcal{O}_{\lambda}$-module by our assumption. Thus $\mathrm{H}^{3}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(2))_{\mathfrak{m}}$ is torsion free. Next, we show the cohomology group $\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(1))_{\mathfrak{m}}$ is torsion free. We consider the long exact sequence $\cdots\rightarrow\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(1))_{\mathfrak{m}}\rightarrow\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(1))_{\mathfrak{m}}\rightarrow\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}/\lambda(1))_{\mathfrak{m}}\rightarrow\cdots$ induced by the short exact sequence $0\rightarrow\mathcal{O}_{\lambda}\rightarrow\mathcal{O}_{\lambda}\rightarrow\mathcal{O}_{\lambda}/\lambda\rightarrow 0.$ We need to show the injection $\mathrm{H}^{2}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(1))_{\mathfrak{m}}/\lambda\hookrightarrow\mathrm{H}^{2}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}/\lambda(1))_{\mathfrak{m}}$ is in fact surjective. For this, note that we have established that $\mathrm{H}^{2}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(1))_{\mathfrak{m}}$ is torsion free and thus $\dim_{k}\mathrm{H}^{2}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(1))_{\mathfrak{m}}/\lambda=\dim_{E_{\lambda}}\mathrm{H}^{2}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),E_{\lambda}(1))_{\mathfrak{m}}$. But we have $\displaystyle\dim_{k}\mathrm{H}^{2}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}/\lambda(1))_{\mathfrak{m}}=\dim_{k}\mathrm{H}^{4}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}/\lambda(2))_{\mathfrak{m}};$ $\displaystyle\dim_{E_{\lambda}}\mathrm{H}^{2}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),E_{\lambda}(1))_{\mathfrak{m}}=\dim_{E_{\lambda}}\mathrm{H}^{4}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),E_{\lambda}(2))_{\mathfrak{m}}.$ by the Poincare duality and (9.4). By the considering the same long exact sequence and the reasoning as above but applied to $\mathrm{H}^{4}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(2))_{\mathfrak{m}}$, we have $\dim_{k}\mathrm{H}^{4}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}/\lambda(2))_{\mathfrak{m}}=\dim_{E_{\lambda}}\mathrm{H}^{4}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),E_{\lambda}(2))_{\mathfrak{m}}$ and hence $\dim_{k}\mathrm{H}^{2}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}/\lambda(1))_{\mathfrak{m}}=\dim_{E_{\lambda}}\mathrm{H}^{2}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),E_{\lambda}(1))_{\mathfrak{m}}.$ It follows that $\mathrm{H}^{2}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}(1))_{\mathfrak{m}}/\lambda\hookrightarrow\mathrm{H}^{2}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathcal{O}_{\lambda}/\lambda(1))_{\mathfrak{m}}$ is in fact surjective and we are done. ∎ ### 9.2. Level raising for $\mathrm{GSp}_{4}$ Let $\pi$ be a cuspidal automorphic representation of $\mathrm{GSp}_{4}(\mathbb{A}_{\mathbb{Q}})$ of general type with weight $(3,3)$ and trivial central character. Let $p$ be a level raising special prime for $\pi$ of length $m$. From the arithmetic level raising theorem proved in the last subsection, we can deduce a level raising theorem for $\pi$ which can be applied to deduce level raising theorems in [Wang22b]. This is the content of this subsection. Suppose the component $\pi_{q}$ of $\pi$ at the prime $q$ is of type $\mathrm{I}\mathrm{I}\mathrm{a}$. Then $\pi$ admits a global Jacquet–Langlands transfer to an automorphic representation of $\mathbf{G}(\overline{\mathrm{B}})(\mathbb{A}_{\mathbb{Q}})$. In particular the maximal ideal $\mathfrak{m}$ associated to $\pi$ in the Construction 3.3 $(3)$ lies in the support of $\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]$ as a $\mathbb{T}^{\Sigma\cup\\{p\\}}$-module. ###### Theorem 9.9. Let $\pi$ be an automorphic representation of $\mathrm{GSp}_{4}(\mathbb{A}_{\mathbb{Q}})$ as above. Suppose that $\pi$ satisfy all the assumptions in Theorem 9.5. Let $p$ be a level raising special prime for $\pi$ of length $m$. Then there exists an automorphic representation $\pi^{\prime}$ of $\mathrm{GSp}_{4}(\mathbb{A}_{\mathbb{Q}})$ of general type with weight $(3,3)$ and trivial central character such that 1. (1) the component $\pi^{\prime}_{p}$ at $p$ is of type $\mathrm{I}\mathrm{I}\mathrm{a}$; 2. (2) we have an isomorphism of the residual Galois representation $\overline{\rho}_{\pi,\lambda}\cong\overline{\rho}_{\pi^{\prime},\lambda}.$ In this case we will say $\pi^{\prime}$ is a level raising of $\pi$. ###### Proof. As remark in the above, the localization $\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]_{\mathfrak{m}}$ is non-zero. By Theorem 9.5, we have an isomorphism $\mathrm{H}^{1}_{\mathrm{sing}}(\mathbb{Q}_{p^{2}},\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}(1)))_{\mathfrak{m}})\xrightarrow{\sim}\mathcal{O}_{\lambda}[\mathrm{Z}_{\mathrm{H}}(\overline{\mathrm{B}})]_{\mathfrak{m}}/\mathrm{det}\phantom{.}\mathcal{T}_{\mathrm{lr}}.$ In particular $\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}(1)))_{\mathfrak{m}}$ is non-zero. Since $\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}(1)))_{\mathfrak{m}}$ is a free $\mathrm{R}^{\mathrm{ram}}$-module, we can find a point $\eta^{\mathrm{ram}}\in\mathrm{Spec}\phantom{.}\mathrm{R}^{\mathrm{ram}}[1/l]$ in the support of $\mathrm{H}^{3}_{\mathrm{c}}(\overline{\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B}),\mathrm{R}\Psi(\mathcal{O}_{\lambda}(1)))_{\mathfrak{m}}.$ Then we can find an automorphic representation $\pi^{\prime}$ of $\mathrm{GSp}_{4}(\mathbb{A}_{\mathbb{Q}})$ which satisfy the isomorphism $\overline{\rho}_{\pi,\lambda}\cong\overline{\rho}_{\pi^{\prime},\lambda}.$ In particular $\pi^{\prime}$ is of general type as $\overline{\rho}_{\pi^{\prime},\lambda}$ is absolutely irreducible. By the Picard-Lefschetz formula applied to the Shimura variety ${\mathrm{X}}_{\mathrm{Pa}}(\mathrm{B})$ , we see that the monodromy of the associated Weil-Deligne representation of $\rho_{\pi^{\prime},\lambda}$ is necessarily is of the form $\begin{pmatrix}0&&&\\\ &0&1&\\\ &&0&\\\ &&&0\\\ \end{pmatrix}$ up to conjugation. Since $\pi^{\prime}$ is of general type and hence tempered at $p$, it follows that $\pi^{\prime}$ is of type $\mathrm{I}\mathrm{I}\mathrm{a}$ by the local Langlands correspondence for non-supercuspidal representations, see for example [Sch05, Table 2]. The theorem is proved. ∎ ## References * [Art01] J. Arthur, A stable trace formula. I. General expansions, J. Inst. Math. Jussieu 1 (2002), no.2, 175–277. * [Art02] J. Arthur, A stable trace formula .II. Global descent, Invent. Math. 143 (2001), no.1, 157–220. * [Art03] J. Arthur, A stable trace formula .III. Proof of the main theorem, Ann. of Math. (2) 158 (2003), 769–837. * [Art04] J. Arthur, Automorphic representations of $\mathrm{GSp}(4)$ in Contributions to automorphic forms, geometry, and number theory, Johns Hopkins Univ. Press, Baltimore, MD, (2004), 65–81. * [BD05] M. Bertolini and H. Darmon, Iwasawa’s main conjecture for elliptic curves over anticyclotomic $\mathbb{Z}_{p}$-extensions, Ann. of Math. (2) 162, (2005), 1–64. * [BS] R. Brooks and R. Schmidt, Local newforms for GSp(4), Lecture Notes in Mathematics, 1918 (2007), Springer, Berlin, viii+307. * [Clo] L. Clozel, Ribet’s theorem for U(3), Amer. J. Math., 122 (2000), 1265–1287. * [CHT] L. Clozel, M. Harris and R. Taylor Automorphy for some $l$-adic lifts of automorphic mod $l$ Galois representations, Publ. Math. Inst. Hautes Études Sci. 63 (1991), 281–332. * [CG15] P-S. Chan and W-T. Gan, The local Langlands conjecture for $\rm GSp(4)$ III: Stability and twisted endoscopy, J. Number Theory 146 (2015), 69–133. * [CS17] A. Caraiani and P. Scholze, On the generic part of the cohomology of compact unitary Shimura varieties, Ann. of Math. (2), 186 (2017), no.3, 649–766. * [CS19] A. Caraiani and P. Scholze, On the generic part of the cohomology of non-compact unitary Shimura varieties, arXiv:1909.01898v1, preprint. * [DL76] P. Deligne and G. Lusztig, Representations of reductive groups over finite fields, Ann. of Math. (2) 103 (1976), no. 1, 103–161. * [Gee11] T. Gee, Automorphic lifts of prescribed types, Math. Ann. 350 (1), (2011), 107–144. * [Gross] B. Gross, On the Satake isomorphism, Galois representations in arithmetic algebraic geometry (Durham, 1996), London Math. Soc. Lecture Note Ser. Vol. 254, Cambridge Univ. Press, Cambridge, 1998. * [Hal97] T. Hales, The fundamental lemma for Sp(4), Proc. American. Math. Soc. 125 (1) (1997), 301–308. * [II10] A. Ichino and T. Ikeda, On the Periods of Automorphic Forms on Special Orthogonal Groups and the Gross–Prasad Conjecture, Geom. Funct. Anal. 19 (2010), 1378–1425. * [Ill02] L. Illusie, Sur la Formule de Picard-Lefschetz, Algebraic Geometry 2000, Azumino, Adv. Stud. Pure Math.36, Math. Soc. Japan, Tokyo (2002), 249–268. * [Kos19] T. Koshikawa, Vanishing theorems for the mod p cohomology of some simple Shimura varieties, arXiv:1910.03147, preprint, (2019). * [Kos21] T. Koshikawa, On the generic part of the cohomology of local and global Shimura varieties, arXiv:2106.10602, preprint, (2021). * [Ko83] R. Kottwitz, Sign changes in harmonic analysis on reductive groups, Trans. Amer. Math. Soc. 278 (1983), 289–297. * [KR00] S. Kudla and M. Rapoport, Cycles on Siegel threefolds and derivatives of Eisenstein series, Ann. Sci. École Norm. Sup. (4) 33 (2000), no. 5, 695–756. * [Lau97] G. Laumon, Sur la cohomologie à supports compacts des variétés de Shimura pour ${\rm GSp}(4)_{\bf Q}$, Compositio Math. 105 (1997), no. 3, 267–359. * [Lau05] G. Laumon, Fonctions zêtas des variétés de Siegel de dimension trois, Formes automorphes. II. Le cas du groupe $\rm GSp(4)$ Astérisque. 302 (2005), 1–66. * [Liu] Y. Liu, Bounding cubic-triple product Selmer groups of elliptic curves, J. Eur. Math. Soc. (JEMS). 21 (2017) no.5 1411–1508. * [LS13] K-W. Lan and J. Suh, Vanishing theorems for torsion automorphic sheaves on general PEL-type Shimura varieties, Adv. Math. 242 (2013), 228–286. * [LS18b] K-W. Lan and B. Stroh, Nearby cycles of automorphic étale sheaves, Compos. Math. 154 (2018), no. 1, 80–119. * [LS18a] K-W. Lan and B. Stroh, Nearby cycles of automorphic étale sheaves II, Cohomology of arithmetic groups, Springer Proc. Math. Stat. 245 (2018), Springer, Cham, 83–106. * [LTXZZ] Y. Liu, Y. Tian, L. Xiao, W. Zhang and X. Zhu, On the Beilinson-Bloch-Kato conjecture for Rankin-Selberg motives, arXiv:1912.11942, online first article in Invent. Math. (2022). * [LTXZZa] Y. Liu, Y. Tian, L. Xiao, W. Zhang and X. Zhu, Deformation of rigid conjugate self-dual Galois representations, arXiv:2108.06998, preprint,(2021). * [MT02] A. Mokrane and J. Tilouine, Cohomology of Siegel varieties with p-adic integral coefficients and applications, Cohomology of Siegel varieties, Asterisque no. 280 (2002), 1–95. * [Mok05] C. Mok, Galois representations attached to automorphic forms on ${\rm GL}_{2}$ over CM fields, Compos. Math. 150 (2014), no. 4, 523–567. * [Nek07] J. Nekovář, On the parity of ranks of Selmer groups. III., Doc. Math., 12:243–274, 2007. * [Oki22] Y. Oki, On supersingular loci of Shimura varieties for quaternionic unitary groups of degree 2, Manuscripta Math. 167 (2022), 263–343. * [Rib84] K. Ribet, Congruence relations between modular forms, Proceedings of the International Congress of Mathematicians, Vol. 1, 2 (Warsaw, 1983), 503–514, PWN, Warsaw, 1984. * [She79] D. Shelstad, Characters and inner forms of a quasi-split group over $\mathbb{R}$, Compositio. Math. 39 (1979), 11–45. * [RW21] M. Rosner and R. Weissauer, Global liftings between inner forms GSp4, arXiv:2103.14715, preprint, 2021. * [RZ96] M. Rapoport and T. Zink, Period spaces for $p$-divisible groups, Annals of Mathematics Studies 141, Princeton University Press, Princeton, NJ (1996), xxii+324. * [San22] M. Santos, Upcoming thesis at LSNTG. * [SGA7 II] P. Deligne and N. Katz, Groupes de monodromie en géométrie algébrique. II, Séminaire de Géométrie Algébrique du Bois-Marie 1967–1969 (SGA 7, II), Lect. Notes Math. Vol. 340, Springer-Verlag, Berlin, 1973. * [Sch05] R. Schmidt, Iwahori-spherical representations of GSp(4) and Siegel modular forms of degree 2 with square-free level, J. Math. Soc. Japan, 57 (2005), no.1, 259–293. * [Sch18] R. Schmidt, Packet structure and paramodular forms, Trans. Amer. Math. Soc., 370 (2018) no.5, 3085–3112. * [She79] D. Shelstad, Characters and inner forms of a quasi-split group over $\mathbb{R}$, Compositio. Math. 39 (1979), 11–45. * [ST93] P. Sally and M. Tadić, Induced representations and classifications for ${\rm GSp}(2,F)$ and ${\rm Sp}(2,F)$, Mém. Soc. Math. France (N.S.) 52 (1993), 75–133. * [Sor09] C. Sorensen, Level-raising for Saito–Kurokawa forms, Compositio Math. 145 (2010), no. 4, 915–953. * [Sor10] C. Sorensen, Galois representations attached to Hilbert-Siegel modular forms, Doc. Math. 15 (2010), 623–670. * [Tay91] R. Taylor, Galois representations associated to Siegel modular forms of low weight, Duke Math. J. 63 (1991), 281–332. * [Tay93] R. Taylor, On the $l$-adic cohomology of Siegel threefolds, Invent. Math. 114 (1993), 289–310. * [Wang19a] H. Wang, On the Bruhat–Tits stratification of a quaternionic unitary Shimura variety, Math. Ann. 376 (2020), 1107–1144. * [Wal97] J. Waldspurger, Le lemme fondamental implique le transfert, Compositio. Math. 105 (1997), 153–236. * [Wang19b] H. Wang, On quaternionic unitary Rapoport–Zink spaces with parahoric level structures, arXiv:1909.12263, online first article in IMRN (2020). * [Wang21] H. Wang, Arithmetic level raising on triple product of Shimura curves and Gross–Schoen Diagonal cycles I: Ramified case, arXiv:2004.00555, to appear in Algebra Number Theory. * [Wang22a] H. Wang, Deformation of rigid Galois representations and cohomology of certain quaternionic unitary Shimura varieties, preprint. * [Wang22b] H. Wang, Level lowering on Siegel modular threefold of paramodular level, preprint. * [Wei05] R. Weissauer, Four dimensional Galois representations, Formes automorphes. II. Le cas du groupe $\rm GSp(4)$, Astérisque, 302, 67–150. * [XZ] L. Xiao and X. Zhu, Cycles on Shimura varieties via geometric Satake, arXiv:1707.05700, preprint.
# An Attention Module for Convolutional Neural Networks Zhu Baozhou1, Peter Hofstee12, Jinho Lee3, Zaid Al-Ars1 ###### Abstract Attention mechanism has been regarded as an advanced technique to capture long-range feature interactions and to boost the representation capability for convolutional neural networks. However, we found two ignored problems in current attentional activations-based models: the approximation problem and the insufficient capacity problem of the attention maps. To solve the two problems together, we initially propose an attention module for convolutional neural networks by developing an AW-convolution, where the shape of attention maps matches that of the weights rather than the activations. Our proposed attention module is a complementary method to previous attention-based schemes, such as those that apply the attention mechanism to explore the relationship between channel-wise and spatial features. Experiments on several datasets for image classification and object detection tasks show the effectiveness of our proposed attention module. In particular, our proposed attention module achieves $1.00\%$ Top-1 accuracy improvement on ImageNet classification over a ResNet101 baseline and $0.63$ COCO-style Average Precision improvement on the COCO object detection on top of a Faster R-CNN baseline with the backbone of ResNet101-FPN. When integrating with the previous attentional activations-based models, our proposed attention module can further increase their Top-1 accuracy on ImageNet classification by up to $0.57\%$ and COCO-style Average Precision on the COCO object detection by up to $0.45$. Code and pre-trained models will be publicly available. ## Introduction Convolutional neural networks have demonstrated to be the gold-standard to solving various problems in the field of computer vision, including image classification (He et al. 2016a), object detection (Liu et al. 2016; Ren et al. 2015), and segmentation (Chen et al. 2017). To improve their performance, many researchers explored various aspects of CNN design and implementation (Gu et al. 2018). To enrich the representation power of CNNs, we can build substantially deeper convolutional neural networks. For example, VGGNet (Simonyan and Zisserman 2015) stacks very small $3\times 3$ convolutional layers, while ResNet (He et al. 2016a) stacks residual blocks with skip connections. GoogLeNet (Szegedy et al. 2015) uses multi-scales of processing to capture spatial correlation. Wide ResNet (Zagoruyko and Komodakis 2016) shows that the width increase of residual networks can enlarge the representation capability and reuse the features better. Xception (Chollet 2017) and ResNeXt (Xie et al. 2017) expose new dimensions to increase cardinalities. Besides, recent literature (Jaderberg et al. 2015; Gregor et al. 2015; Xu et al. 2015) have investigated the attention mechanism since it can improve not only the representation power but also the representation of interests. Convolutional neural networks can extract informative features by blending cross-channel and spatial information (Hu et al. 2018). Attention modules (Woo et al. 2018; Linsley et al. 2019) can learn ”where” and ”what” to attend in channel and space axes, respectively, by focusing on important features and suppressing unnecessary ones of the activations. Dynamic Filter Networks (Jia et al. 2016; Li et al. 2019b) generate the filters conditioned on the input and show the flexibility power of such filters because of their adaptive nature, which has become popular in prediction (Klein, Wolf, and Afek 2015) and Natural Language Processing (Wu et al. 2019). Both Dynamic Filter Networks and attention-based models are adaptive based on the inputs, but there are significant differences between them. Attention-based models (Hu et al. 2018; Woo et al. 2018) produce attention maps using the attention mechanism to operate on the activations of convolution. On the contrary, Dynamic Filer Networks (Su et al. 2019; Liu et al. 2019) generate input information-specific kernels, such as position- specific kernels (Su et al. 2019), semantic label map-specific kernels (Liu et al. 2019), and few-shot learning setting-specific kernels (Zhao et al. 2018), which work as the weights of convolution. Our proposed attention module leverages the attention mechanism to compute the attention maps for attending the activations of convolution, so it is clear to categorized the models applied with our proposed attention module as attention-based models instead of Dynamic Filter Networks. Based on our analysis, the approximation problem and the insufficient capacity problem of the attention maps are ignored in current attentional activations- based models. Motivated by solving the two problems together, we develop an attention module and inspect the complementary relationship between our proposed attention module and previously published attention-based models, such as the attention augmented models and the attentional activations-based models. Our contributions are summarized as follows. * • We point out and analyze two ignored problems of the current attentional activations-based models: the approximation problem and the insufficient capacity problem of the attention maps. To address the two problems together, we originally propose an attention module by developing an AW-convolution, where the shape of the attention maps matches that of the weights instead of the activations. * • Our proposed attention module is a complementary method to previous attention mechanism-based modules, such as Attention Augmented (AA) convolution (Bello et al. 2019), the SE (Hu, Shen, and Sun 2018) and CBAM (Woo et al. 2018) modules in the attentional activations-based models. Integrating with our proposed attention module, the accuracy of AA-Net, SE-Net, and CBAM-Net will be improved further. * • We use image classification and object detection tasks to demonstrate the effectiveness of our proposed attention module. With negligible computational complexity increase, our proposed attention module can boost the image classification and object detection task performance, and it can achieve better accuracy when integrating with other attention-based models. ## Related work In this section, we discuss the recent developments of network engineering and attention mechanism. ### Network engineering ”Network engineering” has been one of the most active research areas since it targets building powerful convolutional neural networks on image classification, which are the backbones of various computer vision tasks and ensure their remarkable performance (Kornblith, Shlens, and Le 2019). Increasing the depth of convolutional neural networks has been regarded as an intuitive way to boost performance, which is the philosophy of VGGNet (Simonyan and Zisserman 2015) and ResNet (He et al. 2016a). In addition, since the skip connection from ResNet shows a strong ability to assist the gradient flow, WideResNet (Zagoruyko and Komodakis 2016), PyramidNet (Han, Kim, and Kim 2017), Inception-ResNet (Szegedy et al. 2017), and ResNeXt (He et al. 2016b; Xie et al. 2017) are ResNet-based versions proposed to explore further the influence of the width, the increase of the width, the multi-scale and the cardinality of convolution, respectively. In terms of efficiency, DenseNet (Huang et al. 2017) reuses the feature maps by concatenating the feature maps from different layers. In particular, MobileNet (Howard et al. 2017, 2019) and ShuffleNet (Ma et al. 2018) series present the advantage of depthwise convolution and the shuffle operation between various group convolutions, respectively. Another design approach uses automated neural architecture search, which achieves state-of-the-art performance regarding both accuracy and efficiency across a range of computer vision tasks (Tan et al. 2019). ### Attention mechanism The attention mechanism plays an important role in the human vision perceptron since it can allocate the available resources to selectively focus on processing the salient part instead of the whole scene (Rensink 2000; Corbetta and Shulman 2002). Multiple attention mechanisms are used to address a known weakness in convolution (Chen et al. 2018; Szegedy et al. 2015; Ke, Maire, and Yu 2017; Chen et al. 2019; Hu, Shen, and Sun 2018; Linsley et al. 2019), by capturing long-range information interactions (Bello et al. 2018; Vinyals, Fortunato, and Jaitly 2015). The Inception family of architectures (Szegedy et al. 2015, 2017), Multigrid Neural Architectures (Ke, Maire, and Yu 2017), and Octave Convolution (Chen et al. 2019) aggregate the scale-space information, while Squeeze-and-Excitation Networks (Hu, Shen, and Sun 2018) and Gather- Excite (Hu et al. 2018) adaptively recalibrate channel-wise response by modeling interdependency between channels. GALA (Linsley et al. 2019), CBAM (Woo et al. 2018), and BAM (Park et al. 2018) refine the feature maps separately in the channel and spatial dimensions. Attention Modules (Wang et al. 2017) and self-attention (Vaswani et al. 2017; Bello et al. 2019) can be used to exploit global context information. Precisely, non-local networks (Wang et al. 2018) deploy self-attention as a generalized global operator to capture the relationship between all pairwise convolutional feature maps interactions. Except for applying the attention mechanism to computer vision tasks (Li et al. 2019a), it has been a widespread adoption to modeling sequences in Natural Language Processing (Yang et al. 2019). ## Proposed attention module In this section, we analyze the two ignored problems in current attentional activations-based models: the approximation problem and the insufficient capacity problem of the attention maps. To address the two problems together, we develop an attention module that mainly refers to the AW-convolution, where the shape of attention maps matches that of the weights rather than the activations. Besides, we refine the branch of calculating the attention maps to achieve a better trade-off between efficiency and accuracy. Last but not least, we integrate our proposed attention module with other attention-based models to enlarge their representational capability. ### Motivation First, we define basic notations in a traditional convolutional layer. In a traditional convolutional layer, the input activations, weights, and output activations are denoted as $I$, $K$, and $O$, respectively. For the input activations $I\in R^{N\times C_{1}\times H\times W}$, $N$, $C_{1}$, $H$, and $W$ refer to the batch size, the number of input channels, the height, and width of the input feature maps, respectively. For the weights $K\in R^{C_{2}\times C_{1}\times h\times w}$, $C_{2}$, $h$ and $w$ refer to the number of output channels, the height and width of the weights, respectively. For the output activations $O\in R^{N\times C_{2}\times H\times W}$, it is computed as the convolution between the input activations $I$ and the weights $K$. In particular, every individual value of the output activations ${O_{[l,p,m,n]}}$ is calculated as follows. $\displaystyle{O_{[l,p,m,n]}}=\text{Convolution}(I,K)$ (1) $\displaystyle=\sum\limits_{o=1}^{{C_{1}}}{\sum\limits_{j=1}^{h-1}{\sum\limits_{k=1}^{w-1}{{I_{[l,o,m^{\prime}+j,n^{\prime}+k]}}\times{K_{[p,o,j,k]}}}}}$ where $l=0,...,N-1$, $m=0,...,H-1$, $n=0,...,W-1$, $o=0,...,C_{1}-1$, $p=0,...,C_{2}-1$, $m^{\prime}=m-\frac{h-1}{2}$, $n^{\prime}=n-\frac{w-1}{2}$. To apply the attention mechanism on the input activations $I$, previous attentional activations-based models produce the channel attention maps $A_{c}\in R^{N\times C_{1}\times 1\times 1}$ and spatial attention maps $A_{s}\in R^{N\times 1\times H\times W}$ separately. For example, applying the channel attention maps $A_{c}$ on the input activations $I$ is presented as $O=\text{Convolution}((I\odot{A_{c}}),K)$, where $\odot$ refers to the Hadamard product and broadcasting during element-wise multiplication is omitted. #### Approximation problem of the attention maps Instead of directly computing the three-dimensional attention map (N is omitted, otherwise the attention maps are of four dimensions.), all the current attentional activations-based models produce the attention maps separately into the channel attention maps $A_{c}$ and spatial attention maps $A_{s}$, which leads to the approximation problem of attention maps. However, to thoroughly attend the input activations $I$, we need to compute the attention maps $A_{f}\in R^{N\times C_{1}\times H\times W}$ and apply it as $O=\text{Convolution}((I\odot{A_{f}}),K)$, which requires too much computational and parameter overhead. Inspired by convolution, we adopt local connection and attention maps sharing to reduce the size of the attention maps. We compute the attention maps $A_{a}\in R^{N\times C_{1}\times h\times w}$ as follows, where $\otimes$ is a special element-wise multiplication since it only works associated with convolution. $\displaystyle{O_{[l,p,m,n]}}=\text{Convolution}(I\otimes A_{a},K)$ (2) $\displaystyle=\sum\limits_{o=1}^{{C_{1}}}{\sum\limits_{j=1}^{h-1}{\sum\limits_{k=1}^{w-1}{({{I_{[l,o,m^{\prime}+j,n^{\prime}+k]}}}\times{{A_{a}}_{[l,o,j,k]}})\times{K_{[p,o,j,k]}}}}}$ ##### Insufficient capacity problem of the attention maps To compute different channels of the output activations of the convolution, the input activations are constrained to be recalibrated by the same attention maps, which indicates the insufficient capacity of the attention maps. As each channel of the feature maps is considered as a feature detector, different channels of the output activations of the convolution expect the input activations to be adapted by different attention maps. Take two channels of output activations of a convolutional layer as an example, the two channels are responsible for recognizing rectangle shape and triangle shape, respectively. Thus, it is reasonable for the two channels to expect that there are different attention maps for attending the input activations of the convolution (i.e., the attention maps to compute the channel of recognizing the rectangle shape should be different from the attention maps to compute the channel of recognizing the triangle shape). To meet this expectation, we need to compute the attention maps $A_{ic}\in R^{N\times C_{2}\times C_{1}\times 1\times 1}$ and apply it on the input activations as follows. $\displaystyle{O_{[l,p,m,n]}}=\text{Convolution}(I\odot{{A_{ic}}_{[l,p,:,:,:]}},K)$ (3) $\displaystyle=\sum\limits_{o=1}^{{C_{1}}}{\sum\limits_{j=1}^{h-1}{\sum\limits_{k=1}^{w-1}{({{I_{[l,o,m^{\prime}+j,n^{\prime}+k]}}}\times{{A_{ic}}_{[l,p,o,0,0]}})\times{K_{[p,o,j,k]}}}}}$ To solve the approximation problem and the insufficient capacity problem of the attention maps together (i.e., combining the solution of Equation.2 and the solution of Equation. 3), we introduce our proposed attention module by developing the AW-convolution. Specifically, we propose to compute the attention maps $A\in R^{N\times C_{2}\times C_{1}\times h\times w}$ and apply it as follows where the attention maps ${A_{[l,:,:,:,:]}}$ has the same shape as that of the weights instead of the input activations. In this paper, ”Attentional weights” refers to the element-wise multiplication result between the attention maps and the weights. Similarly, ”Attentional activations” refers to the element-wise multiplication result between the attention maps and the activations in previous attentional activations-based models. Thus, $I\otimes A$ and ${A_{[l,:,:,:,:]}}\odot K$ represent the attentional activations and attentional weights, respectively. To reduces half the number of element-wise multiplications, we calculate attentional weights instead of attentional activations as follows. $\displaystyle{O_{[l,p,m,n]}}=\text{Convolution}(I\otimes A,K)$ (4) $\displaystyle=\sum\limits_{o=1}^{{C_{1}}}{\sum\limits_{j=1}^{h-1}{\sum\limits_{k=1}^{w-1}{{{I_{[l,o,m^{\prime}+j,n^{\prime}+k]}}}\times({A_{[l,p,o,j,k]}}\times{K_{[p,o,j,k]}}})}}$ $\displaystyle=\text{Convolution}(I,{A_{[l,:,:,:,:]}}\odot K)$ $\displaystyle=\text{AW-Convolution}(I,A\odot K)$ (a) The AW-convolution architecture. (b) The architecture of calculating attention maps $A$. Figure 1: The architecture of our proposed attention module. ### AW-convolution in proposed attention module The AW-convolution in our proposed attention module is presented in Figure 1(a). In this figure, the attention maps $A$ has five dimensions, which is computed from the input activations $I$ as $A=F_{1}(I)$. $F_{1}$ is a function to calculate the attention maps $A$ given the input activations $I$. Then, the attentional weights $AK\in R^{N\times C_{2}\times C_{1}\times h\times w}$ is calculated as $AK=F_{2}(A,K)=K+A\odot K$. $F_{2}$ is a function to calculate the attentional weights $AK$ given the weights $K$ and the attention maps $A$. Finally, the output activations $O$ is calculated from the input activations $I$ and the attentional weights $AK$ as follows. $\displaystyle{O_{[l,p,m,n]}}={F_{3}}(I,AK)$ (5) $\displaystyle=\text{AW- Convolution}(I,AK)$ $\displaystyle=\sum\limits_{i=1}^{{C_{1}}}{\sum\limits_{j=1}^{h-1}{\sum\limits_{k=1}^{w-1}{{I_{[l,o,m^{\prime}+j,n^{\prime}+k]}}\times A{K_{[l,p,o,j,k]}}}}}$ $\displaystyle=\text{Convolution}(I,A{K_{[l,:,:,:,:]}})$ where $F_{3}$ is a function to calculate the output activations $O$ given the input activations $I$ and the attentional weights $AK$. Compared with the traditional convolution, the attentional weights $AK$ of the AW-convolution in our proposed attention module has five dimensions rather than four dimensions, which are different from each other for every individual sample of the input activations batch to convolute. It is also worth explaining the definition of the function $F_{2}$. $AK=K+A\odot K$ instead of $AK=A\odot K$ is used to describe the function $F_{2}$ since it can be regarded as a residual design as follows. $\displaystyle O={F_{3}}(I,AK)$ (6) $\displaystyle=\text{AW- Convolution}(I,{F_{2}}(A,K))$ $\displaystyle=\text{Convolution}(I,K)+\text{AW- Convolution}(I,A\odot K)$ ### Calculating the attention maps $A$ As shown in Figure 1(b), the architecture to compute the attention maps $A$ (i.e., the definition of the function $F_{1}$) is presented, which can be expressed as follows. Avgpool2d aggregates feature responses from the whole spatial extent and embeds them into $A_{0}$, and Pointconv1 and Pointconv2 followed by Relu redistribute the pooled information to capture the dynamic and no-linear dependencies between channels and spatial spaces. $\displaystyle A={F_{1}}(I)=\text{Expand}{{}_{{C_{1}}}}({A_{2}})$ (7) $\displaystyle=\text{Expand}{{}_{{C_{1}}}}(\text{Pointconv2}({A_{1}}))$ $\displaystyle=\text{Expand}{{}_{{C_{1}}}}(\text{Pointconv2}(\text{Pointconv1}({A_{0}})))$ $\displaystyle=\text{Expand}{{}_{{C_{1}}}}(\text{Pointconv2}(\text{Pointconv1}(\text{Avgpool2d}(I))))$ where Pointconv1 and Pointconv2 are pointwise convolutions. We add Batch Normalization and Relu layers after Pointconv1, while adding Batch Normalization and Sigmoid layers after Pointconv2, and they are omitted here to provide a clear expression. In Figure 1(b), Expand function along $C_{1}$ dimension, denoted as $\text{Expand}_{C_{1}}$, is used as an example, and Expand function can be also executed along $N$, $C_{2}$, $h$, and $w$ dimensions in a similar way. $\text{Expand}{{}_{C_{1}}}$ function is used to expand the tensor $A_{2}\in R^{N\times(C_{2}C_{1}/r_{C_{1}})\times h\times w}$ into the attention maps $A\in R^{N\times C_{2}\times C_{1}\times h\times w}$ with the reduction ratio $r_{C_{1}}$, including necessary squeeze, reshape, and expand operations. $\text{Expand}_{C_{1}}$ can be expressed as follows. $\displaystyle A=\text{Expand}{{}_{{C_{1}}}}({A_{2}})$ (8) $\displaystyle={A_{2}}.\text{reshape}(N,{C_{2}},{C_{1}}/{r_{C_{1}}},h,w)$ $\displaystyle.\text{unsqueeze(dim=3)}\text{}$ $\displaystyle\text{.expand}(N,{C_{2}},{C_{1}}/{r_{C_{1}}},{r_{C_{1}}},h,w)\text{}$ $\displaystyle\text{.reshape}(N,{C_{2}},{C_{1}},h,w)$ Calculating the five-dimension attention maps $A$ is not an easy computational task without careful design. Thus, we analyze the additional computational complexity of an AW-convolution compared with a traditional convolution as a reference to refine this design. Considering the trade-off between computational complexity and accuracy, all the experiments in the remainder of this paper use the same settings for the architecture of calculating the attention maps $A$ in our proposed attention module, including $r_{C1}={C_{1}}$, $r_{C_{2}}$ = $r_{hw}$ = 1, $r$ = 16, used in all the stages, and $AK=K+A\odot K$ as the definition for the function $F_{2}$. The details of refining the architecture of calculating the attention maps $A$ are in Section C of Supplementary Material. ### Integrating with other attention-based modules (a) Integrating with SE-ResNet/CBAM-ResNet. (b) Integrating with AA-Wide-ResNet. Figure 2: The schema of bottlenecks and blocks when integrating with our proposed attention module. In this section, we show how to integrate our proposed attention module with the previous attention-based convolutional neural networks to demonstrate the complementary relationship between our proposed attention module and other attention-based modules. Since applying our proposed attention module is using the AW-convolution to replace the traditional convolution, we can easily integrate our proposed attention module with any convolutional neural networks consisting of traditional convolution, including all the recently developed attention-based models (Hu, Shen, and Sun 2018; Woo et al. 2018; Park et al. 2018; Linsley et al. 2019; Bello et al. 2019). We choose the recent attentional activations-based models, i.e., SE-Net and CBAM-Net, as examples to show how to integrate our proposed attention module with other attention-based models. Here we use the popular ResNet (He et al. 2016a) as the backbone to apply the attention mechanism. As shown in Figure 2(a), the left side is the structure of a primary bottleneck in ResNet. The middle one is the structure of a bottleneck with SE/CBAM modules in SE- ResNet/CBAM-ResNet. Integrating the central bottleneck with our proposed attention module is completed by replacing its $3\times 3$ convolution with a $3\times 3$ AW-convolution, and its final structure in AW-SE-ResNet/AW-CBAM- ResNet is shown on the right side. In summary, our proposed attention module is a general module to be integrated seamlessly with any CNNs architectures, including previous attention-based CNNs. In Figure 2(b), we integrate our proposed attention module with Attention Augmented (AA) convolutional networks and describe the architecture of their possible AW-AA-blocks. Experiments of AW-AA-Wide-ResNet (Zagoruyko and Komodakis 2016; Bello et al. 2019) on CIFAR-100 image classification (Krizhevsky and Hinton 2009), as shown in Table 2, suggest that integrating our proposed attention module with attention-based models should be explored carefully since different integration architectures achieve different accuracy. In some cases, an improper integration architecture leads to a small accuracy drop. With a careful design for integration, our proposed attention module is complementary to AA convolution and improves the accuracy of AA-Net further. In particular, we augment Wide-ResNet-28-10 by replacing the first convolution of all the residual blocks with Attention Augmented convolution, which is the AA-Wide-ResNet baseline. Here we set $N_{h}=8$ heads, $k=2v=0.2$, and a minimum of $20$ dimensions per head for the keys as in (Bello et al. 2019). In this figure, we can develop four possible architectures to integrate our proposed attention module with AA-Wide-ResNet. On the left side of this figure, we can build AA-Wide-ResNet-0 by replacing the second $3\times 3$ convolution with our AW-convolution in an AA-block. On the right side, there are three traditional convolutions in an AA convolution, including conv1, conv2, conv3. The conv1 is parallel to the multi-head attention, while the conv2 and conv3 are used to calculate QKV (i.e., queries, keys, and values) and to output attention maps, respectively. On the medial side, we can replace one traditional convolution from conv1, conv2, or conv3 with our AW-conv1, AW- conv2, or AW-conv3 to construct AW-AA-Wide-ResNet-1, AW-AA-Wide-ResNet-2, or AW-AA-Wide-ResNet-3, respectively. The Top-1 accuracy of AW-AA-Wide-ResNet-0 is $1.87\%$ higher than the AA-Wide-ResNet baseline, while AW-AA-Wide-ResNet-3 shows worse performance by $0.10\%$ drop. Model | Top-1 Error | Top-5 Error | GFLOPs | Parameters (M) ---|---|---|---|--- ResNet50 Baseline (He et al. 2016a) | $24.56\%$(+$0.00\%$) | $7.50\%$ | $3.86$ | $25.56$ AW-ResNet50 | $23.38\%$(+$1.18\%$) | $6.79\%$ | $3.87$ | $25.72$ SE-ResNet50 (Hu, Shen, and Sun 2018) | $23.14\%$(+$1.42\%$) | $6.70\%$ | $3.87$ | $28.09$ AW-SE-ResNet50 | $22.72\%$(+$1.84\%$) | $6.47\%$ | $3.88$ | $28.25$ AW-CBAM-ResNet50 (MaxPool) | $22.82\%$(+$1.74\%$) | $6.41\%$ | $3.89$ | $28.25$ AW-CBAM-ResNet50 (Spatial) | $23.20\%$(+$1.36\%$) | $6.58\%$ | $3.90$ | $28.25$ ResNet101 Baseline (He et al. 2016a) | $23.38\%$(+$0.00\%$) | $6.88\%$ | $7.57$ | $44.55$ AW-ResNet101 | $22.38\%$(+$1.00\%$) | $6.21\%$ | $7.58$ | $44.95$ SE-ResNet101 (Hu, Shen, and Sun 2018) | $22.35\%$(+$1.03\%$) | $6.19\%$ | $7.58$ | $49.33$ AW-SE-ResNet101 | $21.78\%$(+$1.60\%$) | $5.74\%$ | $7.59$ | $49.73$ AW-CBAM-ResNet101 (MaxPool) | $21.64\%$(+$1.74\%$) | $5.76\%$ | $7.60$ | $49.73$ AW-CBAM-ResNet101 (Spatial) | $22.32\%$(+$1.06\%$) | $6.18\%$ | $7.61$ | $49.73$ MobileNet Baseline (Howard et al. 2017) | $31.39\%$(+$0.00\%$) | $11.51\%$ | $0.569$ | $4.23$ SE-MobileNet (Hu, Shen, and Sun 2018) | $29.97\%$(+$1.42\%$) | $10.63\%$ | $0.581$ | $5.07$ AW-SE-MobileNet | $29.41\%$(+$1.98\%$) | $10.59\%$ | $0.623$ | $5.52$ CBAM-MobileNet (Woo et al. 2018) | $29.01\%$(+$2.38\%$) | $9.99\%$ | $0.611$ | $5.07$ AW-CBAM-MobileNet (Spatial) | $28.82\%$(+$2.57\%$) | $9.98\%$ | $0.652$ | $5.52$ Table 1: Comparisons of attention-based models on ImageNet classfication. ## Experimental results In this section, we use extensive experiments to demonstrate the effectiveness of our proposed attention module. We use ResNet (He et al. 2016a), MobileNet (Howard et al. 2017), SSD300 (Liu et al. 2016), and Faster R-CNN (Ren et al. 2015) as the baseline models, and various variants of these models are developed, including using our proposed attention module for these baseline models and integrating our proposed attention module with their attentional activations-based models. The datasets to train these models include CIFAR-100 classification (Krizhevsky and Hinton 2009), ImageNet classification (Deng et al. 2009), VOC object detection datasets (Everingham et al. 2015) (The experimental results are included in Section E of Supplementary Material.), and COCO object detection datasets (Lin et al. 2014). To have a better interpretation of our proposed attention module, its feature visualizations using Grad-CAM (Selvaraju et al. 2017) can be found in Section F of Supplementary Material. All the data augmentation and training settings can be found in Section A of Supplementary Material. ### ImageNet image classification To investigate the performance of our proposed attention module on high- resolution images, we train ResNet50 and ResNet101 (He et al. 2016a) and their attention-based variants on the ImageNet classification dataset (Deng et al. 2009). According to the results shown in Table 1, our proposed attention module is complementary to other attentional activations-based models. AW- ResNet50 achieves a $1.18\%$ Top-1 error reduction compared with the ResNet50 baseline. Integrating with our proposed attention module, SE-ResNet50 (Hu, Shen, and Sun 2018) can improve further by $0.42\%$ Top-1 accuracy. The Top-1 accuracy of our AW-SE-ResNet101 is $1.60\%$ and $0.57\%$ higher than that of ResNet101 and SE-ResNet101, respectively. To integrate with CBAM-ResNet (Woo et al. 2018) more carefully, we define CBAM-ResNet (MaxPool) and CBAM-ResNet (Spatial) separately. In CBAM-ResNet (MaxPool), we do not deploy the spatial attention maps, while we do not use max-pooled features in CBAM-ResNet (Spatial). The Top-1 accuracy of AW-CBAM-ResNet50 (MaxPool) and AW-CBAM- ResNet50 (Spatial) are better than AW-ResNet50 by $0.56\%$ and $0.18\%$, respectively, but worse than AW-SE-ResNet50. The number of additional parameters for our proposed attention module is $0.16$ M, which is much smaller than $2.83$ M (i.e., one-sixteenth) of SE and CBAM modules. Moreover, it takes only $0.01$ GFLOPs to apply our proposed attention module on the ResNet50 model on ImageNet classification, which is comparable with $0.01$ GFLOPs and $0.04$ to adopt the SE and CBAM modules and is negligible in terms of FLOPs to implement the baseline model. The computational complexity analysis introduced by the attention mechanism can be found in Section B of Supplementary Material. Model | Top-1 Error | Top-5 Error | GFLOPs | Parameters (M) ---|---|---|---|--- AA-Wide-ResNet Baseline (Bello et al. 2019) | $28.01\%$($+0.00\%$) | $7.92\%$ | $3.89$ | $8.43$ AW-AA-Wide-ResNet-0 | $26.14\%$($+1.87\%$) | $7.43\%$ | $3.90$ | $8.50$ AW-AA-Wide-ResNet-1 | $27.17\%$($+0.84\%$) | $7.49\%$ | $3.89$ | $8.48$ AW-AA-Wide-ResNet-2 | $27.09\%$($+0.92\%$) | $8.00\%$ | $3.89$ | $8.45$ AW-AA-Wide-ResNet-3 | $28.11\%$($-0.10\%$) | $8.24\%$ | $3.89$ | $8.44$ ResNet50 Baseline (He et al. 2016a) | $22.33\%$($+0.00\%$) | $5.83\%$ | $1.22$ | $23.71$ AW-ResNet50 | $19.87\%$($+2.46\%$) | $4.76\%$ | $1.23$ | $23.87$ SE-ResNet50 (Hu, Shen, and Sun 2018) | $20.43\%$($+1.90\%$) | $5.01\%$ | $1.23$ | $26.24$ AW-SE-ResNet50 | $19.00\%$($+3.33\%$) | $4.51\%$ | $1.24$ | $26.40$ CBAM-ResNet50 (Woo et al. 2018) | $19.46\%$($+2.87\%$) | $4.56\%$ | $1.24$ | $26.24$ AW-CBAM-ResNet50 | $18.94\%$($+3.39\%$) | $4.76\%$ | $1.25$ | $26.40$ Table 2: Comparisons of attention-based models on CIFAR-100 classfication. Backbone | Detector |<EMAIL_ADDRESS>0.95] |<EMAIL_ADDRESS>|<EMAIL_ADDRESS> ---|---|---|---|--- ResNet101-FPN (Lin et al. 2017) | Faster R-CNN | $37.13$($+0.00\%$) | $58.28$ | $40.29$ ResNet101-AW-FPN | Faster R-CNN | $37.76$($+0.63\%$) | $59.17$ | $40.91$ ResNet101-SE-FPN (Hu, Shen, and Sun 2018) | Faster R-CNN | $38.11$($+0.98\%$) | $59.41$ | $41.33$ ResNet101-AW-SE-FPN | Faster R-CNN | $38.45$($+1.32\%$) | $59.70$ | $41.86$ ResNet101-CBAM-FPN (Woo et al. 2018) | Faster R-CNN | $37.74$($+0.61\%$) | $58.84$ | $40.77$ ResNet101-AW-CBAM-FPN | Faster R-CNN | $38.19$($+1.06\%$) | $59.52$ | $41.43$ Table 3: Comparisons of attention-based Faster R-CNN on COCO. #### Resource-constrained architecture Driven by demand for mobile applications, many depthwise convolution-based models are developed to take care of the trade-off between accuracy and efficiency. To inspect the generalization of our proposed attention module in this resource-constrained scenario, we conduct the ImageNet classification (Deng et al. 2009) with the MobileNet architecture (Howard et al. 2017). Since our proposed attention module is efficient in terms of both the storage and computational complexity, integrating it into the light-weight architecture is worth exploring. We apply our proposed attention module to pointwise convolution instead of depthwise convolution in every two depthwise separable convolutions. When integrating with the CBAM models (Woo et al. 2018), we remove the max-pooled features and keep spatial attention maps. As shown in Table 1, AW-SE-MobileNet and AW-CBAM-MobileNet achieve $0.56\%$ and $0.19\%$ Top-1 accuracy improvements compared with SE-MobileNet (Hu, Shen, and Sun 2018) and CBAM-MobileNet, respectively. It is an impressive result that the Top-1 accuracy of AW-CBAM-MobileNet is $2.57\%$ better than that of the MobileNet baseline. For the MobileNet model, our proposed attention module increases the computation by $0.041$ GFLOPs, while SE and CBAM modules need $0.012$ and $0.041$ GFLOPs, respectively. Also, the required parameters for our proposed attention module are $0.45$ M, which is much less than $0.84$ M for SE and CBAM modules. ### CIFAR-100 image classification According to the results shown in Table 2, we conclude that our proposed attention module can boost the CIFAR-100 (Krizhevsky and Hinton 2009) accuracy of both the ResNet50 baseline model (He et al. 2016a) and their attentional activations-based models with negligible additional computational complexity. AW-ResNet50 achieves a $2.46\%$ Top-1 error reduction compared with the ResNet50 baseline. Integrating with our proposed attention module, SE-ResNet50 (Hu, Shen, and Sun 2018) and CBAM-ResNet50 (Woo et al. 2018) can increase Top-1 accuracy by $1.43\%$ and $1.52\%$, respectively. In terms of the computational complexity, our proposed attention module requires $0.01$ GFLOPs, which is acceptable compared with $0.01$ GFLOPs for the SE module, $0.02$ GFLOPs for the CBAM module, and $1.22$ GFLOPs for the baseline model. Besides, we only introduce $0.16$ M parameters for our proposed attention module, which is less than $2.53$ M parameters for the SE and CBAM modules. We train ResNet with various depths on CIRAR-100 image classification to show that our proposed attention module works for CNNs with different depths, which are in Section D of Supplementary Material. ### Object Detection on COCO To show the generalization of our proposed attention module, we apply it to object detection tasks. We evaluate our proposed attention module further on the COCO dataset (Lin et al. 2014), which contains $118K$ images (i.e., train2017) for training and $5K$ images (i.e., val2017) for validating. We use Faster R-CNN (Ren et al. 2015) as the detection method with the ResNet101-FPN backbone (Lin et al. 2017). Here we intend to evaluate the benefits of applying our proposed attention module on the ResNet101-FPN backbone (Lin et al. 2017), where all the lateral and output convolutions of the FPN adopt our AW-convolution. The SE and CBAM modules are placed right before the lateral and output convolutions. As shown in Table 3, applying our proposed attention module on ResNet101-FPN boosts<EMAIL_ADDRESS>0.95] by $0.63$ for the Faster R-CNN baseline. Integrating with attentional activations-based models, Faster R-CNNs with the backbones of ResNet101-AW-SE-FPN and ResNet101-AW-CBAM-FPN outperform Faster R-CNNs with the backbones of ResNet101-SE-FPN and ResNet101-CBAM-FPN by $0.34$ and $0.45$ on COCO’s standard metric AP. ## Conclusion In this paper, We analyze the two ignored problems in attentional activations- based models: the approximation problem and the insufficient capacity problem of the attention maps. To address the two problems together, we propose an attention module by developing the AW-convolution, where the shape of the attention maps matches that of the weights rather than the activations, and integrate it with attention-based models as a complementary method to enlarge their attentional capability. We have implemented extensive experiments to demonstrate the effectiveness of our proposed attention module, both on image classification and object detection tasks. Our proposed attention module alone shows noticeable accuracy improvement compared with baseline models. More importantly, integrating our proposed module with previous attention-based models, such as AA-Net (Bello et al. 2019), SE-Net (Hu, Shen, and Sun 2018), and CBAM-Net (Woo et al. 2018), will further boost their performance. ## References * Bello et al. (2018) Bello, I.; Kulkarni, S.; Jain, S.; Boutilier, C.; Chi, E. H.; Eban, E.; Luo, X.; Mackey, A.; and Meshi, O. 2018. Seq2Slate: Re-ranking and Slate Optimization with RNNs. _CoRR_ abs/1810.02019. URL http://arxiv.org/abs/1810.02019. * Bello et al. (2019) Bello, I.; Zoph, B.; Vaswani, A.; Shlens, J.; and Le, Q. V. 2019. Attention Augmented Convolutional Networks. In _The IEEE International Conference on Computer Vision (ICCV)_. * Chen et al. (2017) Chen, L.-C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; and Yuille, A. L. 2017. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. _IEEE transactions on pattern analysis and machine intelligence_ 40(4): 834–848. * Chen et al. (2019) Chen, Y.; Fan, H.; Xu, B.; Yan, Z.; Kalantidis, Y.; Rohrbach, M.; Yan, S.; and Feng, J. 2019. Drop an Octave: Reducing Spatial Redundancy in Convolutional Neural Networks With Octave Convolution. In _The IEEE International Conference on Computer Vision (ICCV)_. * Chen et al. (2018) Chen, Y.; Kalantidis, Y.; Li, J.; Yan, S.; and Feng, J. 2018. A^ 2-Nets: Double Attention Networks. In _Advances in Neural Information Processing Systems_ , 352–361. * Chollet (2017) Chollet, F. 2017. Xception: Deep learning with depthwise separable convolutions. In _Proceedings of the IEEE conference on computer vision and pattern recognition_ , 1251–1258. * Corbetta and Shulman (2002) Corbetta, M.; and Shulman, G. L. 2002. Control of goal-directed and stimulus-driven attention in the brain. _Nature reviews neuroscience_ 3(3): 201. * Deng et al. (2009) Deng, J.; Dong, W.; Socher, R.; Li, L.-J.; Li, K.; and Fei-Fei, L. 2009. Imagenet: A large-scale hierarchical image database. In _2009 IEEE conference on computer vision and pattern recognition_ , 248–255. Ieee. * Everingham et al. (2015) Everingham, M.; Eslami, S. A.; Van Gool, L.; Williams, C. K.; Winn, J.; and Zisserman, A. 2015. The pascal visual object classes challenge: A retrospective. _International journal of computer vision_ 111(1): 98–136. * Gregor et al. (2015) Gregor, K.; Danihelka, I.; Graves, A.; Rezende, D.; and Wierstra, D. 2015. DRAW: A Recurrent Neural Network For Image Generation. In _International Conference on Machine Learning_ , 1462–1471. * Gu et al. (2018) Gu, J.; Wang, Z.; Kuen, J.; Ma, L.; Shahroudy, A.; Shuai, B.; Liu, T.; Wang, X.; Wang, G.; Cai, J.; et al. 2018. Recent advances in convolutional neural networks. _Pattern Recognition_ 77: 354–377. * Han, Kim, and Kim (2017) Han, D.; Kim, J.; and Kim, J. 2017. Deep pyramidal residual networks. In _Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition_ , 5927–5935. * He et al. (2016a) He, K.; Zhang, X.; Ren, S.; and Sun, J. 2016a. Deep residual learning for image recognition. In _Proceedings of the IEEE conference on computer vision and pattern recognition_ , 770–778. * He et al. (2016b) He, K.; Zhang, X.; Ren, S.; and Sun, J. 2016b. Identity mappings in deep residual networks. In _European conference on computer vision_ , 630–645. Springer. * Howard et al. (2019) Howard, A.; Sandler, M.; Chu, G.; Chen, L.-C.; Chen, B.; Tan, M.; Wang, W.; Zhu, Y.; Pang, R.; Vasudevan, V.; Le, Q. V.; and Adam, H. 2019. Searching for MobileNetV3. In _The IEEE International Conference on Computer Vision (ICCV)_. * Howard et al. (2017) Howard, A. G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; and Adam, H. 2017. MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications. _CoRR_ abs/1704.04861. URL http://arxiv.org/abs/1704.04861. * Hu et al. (2018) Hu, J.; Shen, L.; Albanie, S.; Sun, G.; and Vedaldi, A. 2018. Gather-excite: Exploiting feature context in convolutional neural networks. In _Advances in Neural Information Processing Systems_ , 9401–9411. * Hu, Shen, and Sun (2018) Hu, J.; Shen, L.; and Sun, G. 2018. Squeeze-and-excitation networks. In _Proceedings of the IEEE conference on computer vision and pattern recognition_ , 7132–7141. * Huang et al. (2017) Huang, G.; Liu, Z.; Van Der Maaten, L.; and Weinberger, K. Q. 2017. Densely connected convolutional networks. In _Proceedings of the IEEE conference on computer vision and pattern recognition_ , 4700–4708. * Jaderberg et al. (2015) Jaderberg, M.; Simonyan, K.; Zisserman, A.; et al. 2015. Spatial transformer networks. In _Advances in neural information processing systems_ , 2017–2025. * Jia et al. (2016) Jia, X.; De Brabandere, B.; Tuytelaars, T.; and Gool, L. V. 2016. Dynamic filter networks. In _Advances in Neural Information Processing Systems_ , 667–675. * Ke, Maire, and Yu (2017) Ke, T.-W.; Maire, M.; and Yu, S. X. 2017. Multigrid neural architectures. In _Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition_ , 6665–6673. * Klein, Wolf, and Afek (2015) Klein, B.; Wolf, L.; and Afek, Y. 2015. A dynamic convolutional layer for short range weather prediction. In _Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition_ , 4840–4848. * Kornblith, Shlens, and Le (2019) Kornblith, S.; Shlens, J.; and Le, Q. V. 2019. Do better imagenet models transfer better? In _Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition_ , 2661–2671. * Krizhevsky and Hinton (2009) Krizhevsky, A.; and Hinton, G. 2009. Learning multiple layers of features from tiny images. _Master’s thesis, Department of Computer Science, University of Toronto_ . * Li et al. (2019a) Li, H.; Liu, Y.; Ouyang, W.; and Wang, X. 2019a. Zoom out-and-in network with map attention decision for region proposal and object detection. _International Journal of Computer Vision_ 127(3): 225–238. * Li et al. (2019b) Li, X.; Wang, W.; Hu, X.; and Yang, J. 2019b. Selective kernel networks. In _Proceedings of the IEEE conference on computer vision and pattern recognition_ , 510–519. * Lin et al. (2017) Lin, T.-Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; and Belongie, S. 2017. Feature pyramid networks for object detection. In _Proceedings of the IEEE conference on computer vision and pattern recognition_ , 2117–2125. * Lin et al. (2014) Lin, T.-Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; and Zitnick, C. L. 2014. Microsoft coco: Common objects in context. In _European conference on computer vision_ , 740–755. Springer. * Linsley et al. (2019) Linsley, D.; Shiebler, D.; Eberhardt, S.; and Serre, T. 2019. Learning what and where to attend with humans in the loop. In _International Conference on Learning Representations_. URL https://openreview.net/forum?id=BJgLg3R9KQ. * Liu et al. (2016) Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.-Y.; and Berg, A. C. 2016. Ssd: Single shot multibox detector. In _European conference on computer vision_ , 21–37. Springer. * Liu et al. (2019) Liu, X.; Yin, G.; Shao, J.; Wang, X.; et al. 2019. Learning to predict layout-to-image conditional convolutions for semantic image synthesis. In _Advances in Neural Information Processing Systems_ , 568–578. * Ma et al. (2018) Ma, N.; Zhang, X.; Zheng, H.-T.; and Sun, J. 2018. Shufflenet v2: Practical guidelines for efficient cnn architecture design. In _Proceedings of the European Conference on Computer Vision (ECCV)_ , 116–131. * Park et al. (2018) Park, J.; Woo, S.; Lee, J.-Y.; and Kweon, I.-S. 2018. BAM: Bottleneck Attention Module. In _British Machine Vision Conference (BMVC)_. British Machine Vision Association (BMVA). * Ren et al. (2015) Ren, S.; He, K.; Girshick, R.; and Sun, J. 2015. Faster r-cnn: Towards real-time object detection with region proposal networks. In _Advances in neural information processing systems_ , 91–99. * Rensink (2000) Rensink, R. A. 2000. The dynamic representation of scenes. _Visual cognition_ 7(1-3): 17–42. * Selvaraju et al. (2017) Selvaraju, R. R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; and Batra, D. 2017. Grad-cam: Visual explanations from deep networks via gradient-based localization. In _Proceedings of the IEEE international conference on computer vision_ , 618–626. * Simonyan and Zisserman (2015) Simonyan, K.; and Zisserman, A. 2015. Very Deep Convolutional Networks for Large-Scale Image Recognition. In Bengio, Y.; and LeCun, Y., eds., _3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Conference Track Proceedings_. URL http://arxiv.org/abs/1409.1556. * Su et al. (2019) Su, H.; Jampani, V.; Sun, D.; Gallo, O.; Learned-Miller, E.; and Kautz, J. 2019\. Pixel-adaptive convolutional neural networks. In _Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition_ , 11166–11175. * Szegedy et al. (2017) Szegedy, C.; Ioffe, S.; Vanhoucke, V.; and Alemi, A. A. 2017. Inception-v4, inception-resnet and the impact of residual connections on learning. In _Thirty-First AAAI Conference on Artificial Intelligence_. * Szegedy et al. (2015) Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; and Rabinovich, A. 2015. Going deeper with convolutions. In _Proceedings of the IEEE conference on computer vision and pattern recognition_ , 1–9. * Tan et al. (2019) Tan, M.; Chen, B.; Pang, R.; Vasudevan, V.; Sandler, M.; Howard, A.; and Le, Q. V. 2019. Mnasnet: Platform-aware neural architecture search for mobile. In _Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition_ , 2820–2828. * Vaswani et al. (2017) Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A. N.; Kaiser, Ł.; and Polosukhin, I. 2017. Attention is all you need. In _Advances in neural information processing systems_ , 5998–6008. * Vinyals, Fortunato, and Jaitly (2015) Vinyals, O.; Fortunato, M.; and Jaitly, N. 2015. Pointer Networks. In Cortes, C.; Lawrence, N. D.; Lee, D. D.; Sugiyama, M.; and Garnett, R., eds., _Advances in Neural Information Processing Systems 28_ , 2692–2700. Curran Associates, Inc. URL http://papers.nips.cc/paper/5866-pointer-networks.pdf. * Wang et al. (2017) Wang, F.; Jiang, M.; Qian, C.; Yang, S.; Li, C.; Zhang, H.; Wang, X.; and Tang, X. 2017. Residual attention network for image classification. In _Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition_ , 3156–3164. * Wang et al. (2018) Wang, X.; Girshick, R.; Gupta, A.; and He, K. 2018. Non-local neural networks. In _Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition_ , 7794–7803. * Woo et al. (2018) Woo, S.; Park, J.; Lee, J.-Y.; and So Kweon, I. 2018. Cbam: Convolutional block attention module. In _Proceedings of the European Conference on Computer Vision (ECCV)_ , 3–19. * Wu et al. (2019) Wu, F.; Fan, A.; Baevski, A.; Dauphin, Y.; and Auli, M. 2019. Pay Less Attention with Lightweight and Dynamic Convolutions. In _International Conference on Learning Representations_. URL https://openreview.net/forum?id=SkVhlh09tX. * Xie et al. (2017) Xie, S.; Girshick, R.; Dollár, P.; Tu, Z.; and He, K. 2017. Aggregated residual transformations for deep neural networks. In _Proceedings of the IEEE conference on computer vision and pattern recognition_ , 1492–1500. * Xu et al. (2015) Xu, K.; Ba, J.; Kiros, R.; Cho, K.; Courville, A.; Salakhudinov, R.; Zemel, R.; and Bengio, Y. 2015. Show, attend and tell: Neural image caption generation with visual attention. In _International conference on machine learning_ , 2048–2057. * Yang et al. (2019) Yang, B.; Wang, L.; Wong, D. F.; Chao, L. S.; and Tu, Z. 2019. Convolutional Self-Attention Networks. In _Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers)_ , 4040–4045. * Zagoruyko and Komodakis (2016) Zagoruyko, S.; and Komodakis, N. 2016. Wide Residual Networks. In Richard C. Wilson, E. R. H.; and Smith, W. A. P., eds., _Proceedings of the British Machine Vision Conference (BMVC)_ , 87.1–87.12. BMVA Press. ISBN 1-901725-59-6. doi:10.5244/C.30.87. URL https://dx.doi.org/10.5244/C.30.87. * Zhao et al. (2018) Zhao, F.; Zhao, J.; Yan, S.; and Feng, J. 2018. Dynamic conditional networks for few-shot learning. In _Proceedings of the European Conference on Computer Vision (ECCV)_ , 19–35.
Vortex Creep Heating in Neutron Stars Motoko Fujiwaraa,b*** E-mail address<EMAIL_ADDRESS>Koichi Hamaguchia,c††† E-mail address<EMAIL_ADDRESS>Natsumi Nagataa‡‡‡ E-mail address<EMAIL_ADDRESS> and Maura E. Ramirez-Quezadaa,d§§§ E-mail address: me.quezada@hep- th.phys.s.u-tokyo.ac.jp, aDepartment of Physics, University of Tokyo, Bunkyo-ku, Tokyo 113–0033, Japan b Physik-Department, Technische Universität, München, James-Franck-Straße, 85748 Garching, Germany cKavli IPMU (WPI), University of Tokyo, Kashiwa, Chiba 277–8583, Japan dDual CP Institute of High Energy Physics, C.P. 28045, Colima, México Recent observations of old warm neutron stars suggest the presence of a heating source in these stars, requiring a paradigm beyond the standard neutron-star cooling theory. In this work, we study the scenario where this heating is caused by the friction associated with the creep motion of neutron superfluid vortex lines in the crust. As it turns out, the heating luminosity in this scenario is proportional to the time derivative of the angular velocity of the pulsar rotation, and the proportional constant $J$ has an approximately universal value for all neutron stars. This $J$ parameter can be determined from the temperature observation of old neutron stars because the heating luminosity is balanced with the photon emission at late times. We study the latest data of neutron star temperature observation and find that these data indeed give similar values of $J$, in favor of the assumption that the frictional motion of vortex lines heats these neutron stars. These values turn out to be consistent with the theoretical calculations of the vortex- nuclear interaction. ## 1 Introduction A neutron star (NS) is incredibly dense and exists under extreme conditions of pressure and temperature that cannot be found in other places in the universe. While the internal structure of NSs remains elusive, indirect evidence suggests the existence of a neutron superfluid in the inner crust. The first indication of superfluidity in NSs came from the observation of non-zero pairing energy associated with attractive forces, leading to the formation of an energy gap and hence superfluidity [1]. This scenario was predicted to occur in stars with neutron cores [2]. After the discovery of pulsars, the energy gap in NS matter has been further studied; see, e.g., Refs. [3, 4, 5] for a recent review on superfluidity in NSs. In a rotating NS, the irrotational property of a superfluid requires the formation of vortex lines, whose distribution determines the angular velocity of the superfluid component. In the inner crust region, these vortex lines are fixed at certain positions by the interactions with nuclei and cannot move freely. This preserves the rotational speed of the superfluid component and prevents it from following the slowdown of the pulsar rotation, giving rise to the deviation in the rotational speed between the superfluid and other components. This deviation increases until the vortex lines are forced to move by the Magnus force, which increases as the difference in the rotational speed increases. This vortex-line dynamics leads to some observational consequences. A well-known example is the pulsar glitch phenomenon, namely, sudden changes in the rotational frequency of NSs,111See Refs. [6, 7, 8, 9] for a recent review on pulsar glitches. which could be attributed to an avalanche of unpinning of superfluid vortex lines [10, 11]. Another phenomenon is the heating effect caused by the friction associated with the creep motion of vortex lines [12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22], which is the subject of the present paper. Our motivation to revisit this heating mechanism is provided by the recent observations of old warm NSs [23, 24, 25, 26, 27, 28, 29], whose observed temperature is considerably higher than that predicted in the standard NS cooling scenario [30, 31, 32, 33, 34, 35]. This work aims to study whether these observations can be explained by the vortex-creep heating effect. With this objective in mind, we focus on the following characteristic property of the vortex-creep heating. As we see below in detail, the heating luminosity in this heating mechanism is proportional to the time derivative of the angular velocity of the pulsar rotation, and the proportional constant is determined only by the NS structure and the vortex-nuclear interactions. As a result, the value of this proportional constant, denoted by $J$ in this paper, is almost universal over NSs. In addition, we can obtain the value of $J$ for an old NS by observing its temperature and its pulsar motion since, at late times, the heating luminosity balances with the luminosity of photon emission, which is determined by the surface temperature of the NS. As it turns out, the present data of old warm NSs indeed show similar values of $J$ for these stars, in agreement with the prediction of the vortex-creep heating scenario. We also find that these values are compatible with the $J$ parameter evaluated from the calculations of the nuclear pinning force available in the literature. The remainder of this paper is structured as follows. In Section 2, we provide a brief overview of the thermal evolution of NSs, focusing on the isothermal phase. In Section 3, we describe the vortex creep heating mechanism for NSs and explain how we can relate a universal parameter with the late-time temperature prediction. In Section 4, we summarize the numerical evaluation of the pinning force, from which we calculate the parameter $J$. In Section 5, we study the recent observations of old and warm NSs to assess the current status of the vortex creep heating hypothesis. Finally, we summarize our findings and conclude our discussion in Section 6. ## 2 Thermal evolution of neutron star This section reviews the surface temperature prediction of NSs based on their thermal evolution and sources of heating and cooling. The temperature distribution within NSs is characterized by a local core temperature $T$, which exhibits a temperature gradient only during the early stages of the NS, typically within the first 10–100 years of its existence [36, 37]. The high thermal conductivity of the highly degenerate electron gas causes the core to become isothermal over time, reaching thermal equilibrium. Hence, the red- shifted temperature of NSs, $T^{\infty}(\bar{r},t)=T(\bar{r},t)e^{\phi(\bar{r})}$, where $\phi(\bar{r})$ specifies the gravitational redshift, reaches a constant value $T^{\infty}(\bar{r},t)\simeq T^{\infty}(t)$ and only the outermost layers exhibit an appreciable temperature gradient. The relativistic equation describing the thermal evolution of NSs after thermal relaxation is given by [38, 39] $\displaystyle C(T^{\infty})\frac{dT^{\infty}}{dt}=-L_{\nu}^{\infty}(T^{\infty})-L_{\gamma}^{\infty}(T^{\infty})+L_{\rm H}^{\infty}.$ (2.1) Here, $C$ represents the total heat capacity of the NS and is temperature- dependent. The right-hand side of the equation expresses the red-shifted luminosity for three different processes, namely, neutrino cooling $L_{\nu}^{\infty}$, photon cooling $L_{\gamma}^{\infty}$, and heating source $L_{\rm H}^{\infty}$. At temperatures below a few $\times~{}10^{9}~{}\mathrm{K}$, the NS becomes transparent to neutrinos, allowing them to escape without interacting with the stellar matter and carrying away energy. Therefore, the cooling process during the early stage of the star’s life is dominated by neutrino emission. At later times, typically for $t\gtrsim 10^{5}~{}\mathrm{yrs}$,222The dominance of photon emission might be delayed for massive NSs, in which the rapid neutrino emission via the direct Urca process could occur; we do not consider this possibility in what follows. photon emission dominates neutrino emission. The thermal photon emission follows a blackbody spectrum. The photon luminosity $L_{\gamma}$ can be described by the Stefan-Boltzmann law and related to surface temperature $T_{\rm s}$ as $\displaystyle L_{\gamma}=4\pi R_{\rm NS}^{2}\sigma_{\rm SB}T_{\rm s}^{4},$ (2.2) in the local reference frame of the NS, where $\sigma_{\rm SB}$ is the Stephan-Boltzmann constant and $R_{\rm NS}$ the NS radius. To relate the internal temperature $T$ of a NS to its surface temperature $T_{\rm s}$, we use the heat envelope model proposed by Potekhin et al. in 1997 [40]. According to this model, the observed thermal emission from old isolated NSs can be explained by the heat trapped in a thin envelope surrounding the star’s crust. If we have an internal heating source, the heating luminosity will balance with the photon luminosity at a sufficiently late time, $\displaystyle L^{\infty}_{\rm H}$ $\displaystyle\simeq L^{\infty}_{\gamma},$ (2.3) which determines the surface temperature. It is pointed out that NSs have some internal heating mechanisms, such as vortex creep heating, rotochemical heating, and magnetic field decay. These heating effects may become visible at late times and can operate even for isolated NSs. These internal mechanisms are comprehensively compared with the observed surface temperature in Refs. [20, 22], which is recently revisited by Ref. [41]. We note in passing that the balance equation (2.3) generically holds with good accuracy for old NSs that are older than $10^{5}~{}\mathrm{yrs}$. This can be seen by estimating the typical timescale $\tau_{\rm eq}$ for the relaxation into the equilibrium state: $\displaystyle\tau_{\rm eq}$ $\displaystyle\simeq\frac{CT^{\infty}}{4\pi R_{\rm NS}^{2}\sigma_{\rm SB}T_{\rm s}^{4}}$ (2.4) $\displaystyle\sim 3\times 10^{4}~{}\mathrm{yrs}\left(\frac{C}{10^{35}~{}\mathrm{erg/K}}\right)\left(\frac{R_{\rm NS}}{11.43~{}{\rm km}}\right)^{-2}\left(\frac{T_{\rm s}}{10^{5}~{}\mathrm{K}}\right)^{-4}\left(\frac{T^{\infty}}{10^{6}~{}\mathrm{K}}\right)~{},$ (2.5) where we have used $T^{\infty}\sim 10^{6}~{}\mathrm{K}$ corresponding to the surface temperature, at the equilibrium phase, of $T_{\rm s}\sim 10^{5}~{}\mathrm{K}$.333To derive the typical scale of $\tau_{\rm eq}$, we fitted the relation between the NS internal temperature $T$ and the NS surface temperature $T_{\rm s}$ for $t_{\rm age}\lesssim 10^{6}~{}\mathrm{yrs}$ [40] by assuming $T\propto T_{\rm s}^{2}$. It is found that this timescale is shorter than the age of old NSs, assuring the equilibrium condition (2.3). ## 3 Review of vortex creep heating In this section, we will review vortex creep heating, where the presence of a superfluid in the inner crust of a NS plays a key role. In this region, vortex lines are thought to exist as a consequence of the NS rotation. In Sec. 3.1, we will derive the equation of motion for this rotational motion by introducing two different angular velocities for the inner crust superfluid and the other part of the star. In Sec. 3.2, we will consider the dynamics of a vortex line and evaluate its radial velocity. Finally, in Sec. 3.3, we will assess the effect of vortex creep on the late-time temperature prediction of NSs. ### 3.1 Equations of motion for crust and superfluid To describe a rotating NS, let us divide it into two components depending on how external torque exerts on it, the crust component and the superfluid component [42]. The crust component comprises a rigid crust, a lattice of nuclei, and charged particles tightly coupled to electromagnetic field lines [42]. This component is directly affected by external torque provided by pulsar magnetic radiation. The superfluid component refers to ${}^{1}S_{0}$ superfluid of neutrons.444The neutron triplet (${}^{3}P_{2}$) superfluid in the core region is classified into the crust component in the two-component model [42]. This is because the neutron superfluid in this region is expected to coexist with the proton superconductor and be tightly coupled to the crust component [43, 44]. This superfluid phase is believed to appear in the inner crust region based on pairing gap evaluations [3, 4, 5]. This component is just indirectly affected by the external torque through the interaction with the crust component. Figure 1: The structure of a NS and vortex lines. Left: The grey, blue, and brown regions represent the outer crust, inner crust, and core regions, respectively. The red lines represent vortex lines. Right: A single vortex line in the inner crust. The vortex line in the neutron superfluid is attached to the outer crust and has two boundaries. On the left side of Fig. 1, we show a schematic diagram of a NS. The grey region is the outer crust, composed of ions and electrons. The blue layer represents the inner crust region and has an approximately $\sim 1~{}\mathrm{km}$ thickness. In this region, nuclei, electrons, and neutrons exist, and neutrons are expected to be in the form of the neutron singlet superfluid. The brown region is the NS core, whose internal structure remains uncertain and a subject of ongoing research and debate. In the two-component model, the neutron superfluid in the blue layer is classified into the superfluid component, while all the other parts are classified into the crust component. Based on this two-component model, we derive the equations of motion and describe the rotation. For later convenience, we divide the system into a thin disc and introduce the cylindrical coordinate $(r,\varphi,z)$. Since we treat the crust component as a rigid body, its angular velocity $\Omega_{\rm c}(t)$ is independent of $r$. On the other hand, the superfluid angular velocity varies with $r$, and we denote it by $\Omega_{\rm s}(t,r)$. The equation of motion for the crust component is $\displaystyle I_{\rm c}\dot{\Omega}_{\rm c}(t)=N_{\rm ext}(t)+N_{\rm int}(t),$ (3.6) where $I_{\rm c}$ represents the moment of inertia of the crust component, and the dot represents the derivative with respect to time $t$. On the right-hand side, we divide the torque into two parts: The first term $N_{\rm ext}(t)$ represents the external torque acting on the system. The second term $N_{\rm int}(t)$ corresponds to the effects of internal torques in the system $\displaystyle N_{\rm int}(t)$ $\displaystyle=-\int dI_{p}(r)\frac{\partial\Omega_{\rm s}}{\partial t}(t,r),$ (3.7) where $dI_{p}$ represents the differential inertial momenta. The integral is taken over the region where the crust component interacts with the superfluid component, called the pinning region. These two components connect through the pinning force between the vortex line and the nuclei-like object in the inner crust, as we will discuss in Sec. 3.2. The equations of the superfluid component are obtained by introducing two fundamental measures of rotation in the fluid: vorticity and circulation. The vorticity vector $\bm{\omega}$ characterizes the microscopic features of rotation and is a locally determined value defined as the curl of the fluid velocity $\bm{v}$, $\displaystyle\bm{\omega}\equiv\nabla\times\bm{v}.$ (3.8) In contrast, the circulation $\Gamma$ measures the macroscopic rotation and is defined over a finite region. It is given by the line integral of the fluid velocity $\bm{v}$ around a closed path $C$, which can be expressed as a surface integral over any surface $S$ with the boundary $C$, $\displaystyle\Gamma\equiv\oint_{C}\bm{v}\cdot d\bm{\ell}=\iint_{S}\bm{\omega}\cdot d\bm{S},$ (3.9) where $d\bm{\ell}$ and $d\bm{S}$ denote the line and surface elements, respectively. We used Stokes’ theorem to obtain the last expression. Superfluid motion obeys the potential flow condition, $\displaystyle\nabla\times\bm{v}{{}_{\rm s}}=\bm{0},$ (3.10) where $\bm{v}{{}_{\rm s}}$ denotes the superfluid velocity. This condition, implying the absence of the vorticity in the superfluid, holds because the superfluid velocity is proportional to the gradient of the phase of the condensate wave-function of the superfluid. Nevertheless, we still have a nonzero circulation if there exists a singular object known as a vortex line. In Fig. 1, we show a schematic picture of a single vortex line in the NS inner crust. The vortex line is a string-like configuration with a thickness of the order of femtometers (red curve). The circulation for each vortex line is quantized in units of $\displaystyle\kappa$ $\displaystyle\equiv\frac{h}{2m_{n}},$ (3.11) where $h$ is the Planck constant and $m_{n}$ is the neutron mass. This quantization follows from the condition that the wave-function of the condensate is single-valued, and thus the change in its phase must be $2\pi k$ with $k$ as an integer. Since a vortex line with $k=1$ is energetically favored and stable, the system with larger angular velocity contains a larger number of vortex lines [45]. The number of vortex lines will be saturated if the total circulation reaches that of the rigid rotation as a whole system, which is expected in NSs. The vortex lines in the inner crust have boundaries corresponding to the normal matter in the outer crust. Under this circumstance, the circulation for the contour $C$ in Fig. 1 is uniquely determined and, because of the potential flow condition (3.10), can be regarded as topological as it remains unchanged under deformations of $C$ (unless it passes through another vortex line). We may express the superfluid velocity on average in the same form as the normal fluid,555This relation is confirmed by observing the shape of the free surface for rotating superfluid in liquid He system [46, 47, 48]. $\displaystyle\Braket{\bm{v}_{\rm s}}=\bm{\Omega}_{\rm s}\times\bm{r}~{},$ (3.12) where $\bm{r}$ denotes the position vector from the center of the NS. The total circulation of a superfluid system is equal to the sum of the circulation of each vortex line. This means that the number of vortex lines is directly related to the superfluid angular velocity $\bm{\Omega}_{\rm s}$. By substituting Eq. (3.12) in Eq. (3.9), we obtain $\displaystyle\Gamma_{\rm superfluid}$ $\displaystyle=\int_{C}d\bm{\ell}\cdot\bigl{(}\bm{\Omega}_{\rm s}(t,r)\times\bm{r}\bigr{)}=\int_{0}^{r}dr^{\prime}~{}2\pi r^{\prime}\kappa n(t,r^{\prime}),$ (3.13) where we choose the integral path $C$ around the edge of the disc with radius $r$, and $n(t,r)$ is the number density of the vortex lines per unit area. Noting that only the radial motion of vortex lines changes $n$ due to the axial symmetry around the rotation axis, we obtain the following conservation law, $\displaystyle\frac{\partial n}{\partial t}+\nabla\cdot(nv_{r}\bm{e}_{r})=0,$ (3.14) where $v_{r}$ is the vortex velocity in the radial direction $\bm{e}_{r}$, which we call the creep rate. Combining Eqs. (3.13) and (3.14), we obtain the equation of motion for the superfluid component: $\displaystyle\frac{\partial\Omega_{\rm s}}{\partial t}$ $\displaystyle=-\left(2\Omega_{\rm s}+r\frac{\partial\Omega_{\rm s}}{\partial r}\right)\frac{v_{r}}{r}.$ (3.15) The NS rotation is described by Eqs. (3.6) and (3.15) coupled through a nonzero $v_{r}$. In other words, by switching off the radial motion of the vortex line, we have $\partial\Omega_{\rm s}/\partial t=0$. In this case, $N_{\rm int}$ turns out to be zero, and the two equations of motion are decoupled. ### 3.2 Dynamics of a vortex line In the inner crust, a vortex line feels two forces, the pinning force and the Magnus force. The pinning force arises from the interaction between a vortex line and a nuclear-like object within the inner crust, resulting in the pinning of the vortex line to the lattice of nuclei where the energy is minimized [10]. As long as the pinning force is dominant, the vortex lines remain attached to the crust and move at the same velocity as the crust component: $\displaystyle\bm{v}_{\rm VL}(t,r)=\bm{\Omega}_{\rm c}(t)\times\bm{r},$ (3.16) where $\bm{v}_{\rm VL}(t,r)$ denotes the velocity of a vortex line. One way to quantify the pinning force is to compare the energies associated with different configurations of a vortex line. Figure 2: The nuclear pinning and interstitial pinning for the vortex line pinning configurations. Red cylinders and black spheres represent vortex lines and nuclei, respectively. In Fig. 2, we show two possibilities for the pinning configurations, the nuclear pinning and the interstitial pinning. The difference in the energies between these two configurations defines the pinning energy, $\displaystyle E_{\rm pin}\equiv E_{\rm NP}-E_{\rm IP},$ (3.17) where $E_{\rm NP}$ ($E_{\rm IP}$) denotes the energy of the nuclear (interstitial) pinning configuration. A nuclear pinning configuration occurs when the pinning energy is negative, and the vortex line is directly attached to the nuclei lattice. Conversely, when the pinning force is repulsive, the vortex line is pinned in the interstitial regions. For a nuclear pinning configuration, a rather crude estimate of the pinning force per unit length is then given by $\displaystyle f_{\rm pin}|_{\rm NP}$ $\displaystyle\simeq\frac{|E_{\rm pin}|}{\Delta r\Delta L},$ (3.18) where $\Delta r$ is the distance between the nuclear and interstitial pinning positions and $\Delta L$ is the distance between the successive pinning sites along a vortex line. These two quantities are expected to be of the order of the Wigner-Seitz radius $R_{\rm WS}$, the radius of an imaginary sphere whose volume is equal to the average volume per nucleus in each region. Due to this pinning effect, the relative velocity between the superfluid and the vortex lines is developed, $\displaystyle\delta\bm{v}\equiv\bm{v}_{\rm s}-\bm{v}_{\rm VL}=\delta\bm{\Omega}\times\bm{r},$ (3.19) where we use Eqs. (3.12) and (3.16) to obtain the last expression and introduce the relative angular velocity, $\displaystyle\delta\bm{\Omega}$ $\displaystyle\equiv\bm{\Omega}_{\rm s}-\bm{\Omega}_{\rm c}.$ (3.20) This velocity difference induces the Magnus force per unit length of a vortex line, $\displaystyle\bm{f}_{\rm Mag}$ $\displaystyle=\rho\,(\delta\bm{v})\times\bm{\kappa},$ (3.21) where $\rho$ is the superfluid density and $\bm{\kappa}$ is the vorticity vector, which is parallel to the vortex line (hence, parallel to the rotational axis of the NS) and has the absolute value $|\bm{\kappa}|\equiv\kappa$. Figure 3: The Magnus force acting on a vortex line. We define the direction of $\bm{\kappa}$ as the direction of the right-hand screw. We see that the Magnus force always acts in the outward direction, as illustrated in Fig. 3, since the crust component rotates slower than the superfluid component due to the deceleration by the external torque. Vortex lines overcome the trapping through thermal fluctuations or quantum tunneling [49, 50], and start to creep outward. If this creep rate is rapid enough, the superfluid rotation can smoothly follow the change in the crust rotation, and the system can reach a steady state. To see if this is the case for the NSs of our interest, let us briefly review the evaluation of the vortex creep rate $v_{r}$ following Ref. [50]. In this analysis, the pinning force is modeled by a periodic potential with a period equal to the span of the nuclei lattice $(\sim R_{\rm WS})$ and a height equal to the pinning energy $(\sim|E_{\rm pin}|)$. The Magnus force is considered as a bias that tilts the periodic potential. Let us introduce the transition rate for a vortex line to move from a local minimum into the next local minimum as ${\cal R}_{\rm VC}$. The zero-point frequency in the vicinity of local minima of the potential $\omega_{0}$ controls ${\cal R}_{\rm VC}$ and is obtained through the quantization of the vortex system [49, 50]. If the vortex tension is negligible compared to the pinning force, we obtain666In Ref. [50], the case where the vortex tension dominates the pinning force is also studied, and the conclusion of the steady state turns out to remain unchanged. $\displaystyle\omega_{0}\simeq\frac{\pi\kappa\Lambda}{4R_{\rm WS}^{2}}\simeq 1.2\times 10^{20}\,{\rm s}^{-1}\left(\frac{R_{\rm WS}}{50~{}{\rm fm}}\right)^{-2}\left(\frac{\Lambda}{2}\right),$ (3.22) where $\Lambda$ characterizes the vortex tension, $T_{v}=\rho\kappa^{2}\Lambda/(4\pi)$, and is evaluated as $2\lesssim\Lambda\lesssim 10$ in the inner crust [51]. The transition rate is then given by [49, 50] $\displaystyle{\cal R}_{\rm VC}\simeq\frac{\omega_{0}}{2\pi}\,e^{-\frac{|E_{\rm pin}|}{k_{\rm B}T_{\rm eff}}},$ (3.23) where $T_{\rm eff}$ is defined by $\displaystyle k_{\rm B}T_{\rm eff}\equiv\frac{\hbar\omega_{0}}{2}\coth\left(\frac{T_{\rm q}}{T}\right)\sim\begin{cases}k_{\rm B}T&(T\gg T_{\rm q})\\\ \frac{\hbar\omega_{0}}{2}&(T\ll T_{\rm q})\end{cases},$ (3.24) with $T$ being the temperature of the inner crust and $\displaystyle T_{\rm q}\equiv\frac{\hbar\omega_{0}}{2k_{\rm B}}\simeq 3.8\times 10^{8}~{}\mathrm{K}~{}\left(\frac{\omega_{0}}{10^{20}~{}\mathrm{s}^{-1}}\right).$ (3.25) As can be seen from these expressions, the unpinning occurs predominantly through thermal activation (quantum tunneling) for $T\gg T_{\rm q}$ ($T\ll T_{\rm q}$). In particular, for NSs as old as those considered in the following analysis, their internal temperature is $\lesssim 10^{8}$ K. Therefore, the vortex creep motion in these NSs is triggered by quantum tunneling. By using the transition rate in Eq. (3.23), we evaluate the creep rate as $v_{r}\simeq{\cal R}_{\rm VC}\cdot R_{\rm WS}$. With the vortex creep rate obtained above, we can estimate the typical distance at which a vortex line travels through the creep motion during the lifetime of the NS: $\displaystyle v_{r}\times t_{\rm age}$ $\displaystyle\simeq R_{\rm WS}\cdot\frac{\omega_{0}}{2\pi}\cdot e^{-\frac{2|E_{\rm pin}|}{\hbar\omega_{0}}}\times t_{\rm age}$ $\displaystyle\simeq 160~{}{\rm km}\times\left(\frac{\omega_{0}}{10^{20}~{}\rm s^{-1}}\right)\left(\frac{R_{\rm WS}}{50~{}\rm fm}\right)\left(\frac{t_{\rm age}}{10^{5}~{}\rm yr}\right),$ (3.26) where we set $|E_{\rm pin}|=1$ MeV as a representative value. We find that this distance is considerably larger than the crust thickness ($\simeq 1$ km), implying that the vortex creep motion would reach a steady state in old NSs. Once the system enters the steady phase, the crust and superfluid components decelerate at the same rate, i.e., $\displaystyle\frac{\partial\Omega_{\rm s}(t,r)}{\partial t}=\dot{\Omega}_{\rm c}\equiv\dot{\Omega}_{\infty}~{}.$ (3.27) Note that $\Omega_{\rm c}$ and $|\dot{\Omega}_{\infty}|$ are identified as the current observed values of the angular velocity and deceleration rate of NSs, respectively. As a consequence, the relative angular velocity between the crust and superfluid components becomes independent of time, and this is found to be fairly close to the critical angular velocity determined by the condition $f_{\rm pin}=f_{\rm Mag}$ [50]: $\displaystyle\delta\Omega_{\infty}\simeq\delta\Omega_{\rm cr}~{},$ (3.28) with $\displaystyle\delta\Omega_{\rm cr}$ $\displaystyle\equiv|\delta\bm{\Omega}|_{f_{\rm pin}=f_{\rm Mag}}=\frac{f_{\rm pin}}{\rho\kappa r}.$ (3.29) ### 3.3 Prediction of surface temperature As vortices move outward, the rotational energy of the superfluid is dissipated through their frictional interaction with the normal components of the inner crust, which heats the NS. This heating luminosity is computed as [12, 15] $\displaystyle L_{\mathrm{H}}$ $\displaystyle=N_{\rm ext}\Omega_{\rm c}(t)-\frac{d}{dt}\left[\frac{1}{2}I_{\rm c}\Omega_{\rm c}^{2}(t)+\frac{1}{2}\int dI_{\rm p}\Omega_{\rm s}^{2}(t,r)\right]$ $\displaystyle=\int dI_{\rm p}\left[\Omega_{\rm c}(t)-\Omega_{\rm s}(t,r)\right]\frac{\partial\Omega_{\rm s}(t,r)}{\partial t}$ $\displaystyle=\int dI_{\rm p}\,\delta\Omega\left|\frac{\partial\Omega_{\rm s}(t,r)}{\partial t}\right|~{},$ (3.30) where we use Eqs. (3.6) and (3.7), and the inertial momenta $I_{\rm p}$ is integrated over the region where the pinning process efficiently occurs. This expression is further simplified in the steady state with the condition (3.27), $\displaystyle L_{\mathrm{H}}$ $\displaystyle=J|\dot{\Omega}_{\infty}|~{},$ (3.31) where we define, $\displaystyle J\equiv\int dI_{\rm p}\delta\Omega_{\infty}~{},$ (3.32) and $\delta\Omega_{\infty}$ denotes the steady-state value of the relative angular velocity. As we see, the heating luminosity is proportional to the current deceleration rate of the pulsar. Note that the value of $J$ can be estimated from Eqs. (3.27), (3.29), and (3.32) by specifying the value of $f_{\rm pin}$ and the region of pinning. We will evaluate the proportional coefficient $J$ in Sec. 4.2. In particular, if this heating luminosity balances with the photon luminosity, which we expect to occur for old NSs as we discussed in Sec. 2, the surface temperature $T_{\rm s}$ can be estimated using Eq. (2.3) as $\displaystyle T_{\rm s}^{\rm eq}$ $\displaystyle\equiv\left(\frac{J|\dot{\Omega}_{\infty}|}{4\pi R_{\rm NS}^{2}\sigma_{\rm SB}}\right)^{\frac{1}{4}}$ (3.33) $\displaystyle\simeq 1.0\times 10^{5}~{}\mathrm{K}\left(\frac{J}{10^{43}~{}\mathrm{erg~{}s}}\right)^{\frac{1}{4}}\left(\frac{|\dot{\Omega}_{\infty}|}{10^{-14}~{}\mathrm{s}^{-2}}\right)^{\frac{1}{4}}\left(\frac{R_{\rm NS}}{11.43~{}\mathrm{km}}\right)^{-\frac{1}{2}}.$ (3.34) In the steady vortex creep scenario, therefore, the surface temperature is predicted as a function of $J$, $|\dot{\Omega}_{\infty}|$, and $R_{\rm NS}$, free from the uncertainty of the initial condition and the subsequent temperature evolution. We will examine this prediction against observation in Sec. 5. ## 4 Theoretical approaches for the vortex pinning To compute the energy dissipation due to vortex creep, we evaluate the parameter $J$ in Eq. (3.32). We first review the calculation of the pinning force $f_{\rm pin}$ available in the literature in Sec. 4.1; the values of $f_{\rm pin}$ which our analysis is based on are summarized in Appendix A. Then, in Sec. 4.2, we estimate possible ranges of $J$ using the results in Sec. 4.1, which will be compared with observation in the subsequent section. ### 4.1 Evaluation of pinning force To evaluate the pinning force, we need to analyze a nucleon many-body system at high densities, which generically suffers from technical difficulties due to less-known properties of nuclear interactions. A traditional method of treating nuclear interactions is to model a form of the interaction and fit it to experimental data, such as nucleon-nucleon scattering [52, 53, 54]. For example, the Argonne interaction [54], which we consider in the following analysis, is a two-body nucleon potential fitted to nucleon scattering data and deuteron properties. This sort of bare interaction does not include in- medium effects. The many-body calculation based on bare interaction is necessary to obtain the properties of nucleon systems. At the same time, it is still challenging to perform it in general due to, e.g., the strong repulsive core. An alternative method is to use an effective interaction incorporating in-medium effect phenomenologically. Skyrme-type interactions [55, 56] are well-known examples, which consist of contact (zero-range) interactions with momentum-dependent coefficients. The parameters of the interactions are determined by fitting to the experimental data of binding energies and radii of several nuclei. There are many sets of fitting parameters used in the literature [57, 58, 59], such as SLy4 [60] and SkM* [61]. There are also finite-range interactions, such as the Gogny interaction [62]. There are several approaches to analyzing the nuclear matter, and the following two are often used for the calculation of the pinning energy: * • Quantum approach A standard method to analyze a quantum multi-body system is to calculate the energy levels of a single particle in the mean field of a self-consistent potential by solving the corresponding Schrödinger equation. A neutron pairing interaction is then considered to determine the pairing field as in the BCS theory. These solutions are obtained via an iterative process such that they satisfy the self-consistent conditions. This method is called the Hartree- Fock-Bogoliubov (HFB) method and adopted in Refs. [63, 64, 65, 66]. * • Semi-classical approach The quantum approach based on the HFB method often requires a high computational cost. To evade this, a semi-classical approach based on the Thomas-Fermi approximation is also frequently used, where nuclear matter is regarded as a many-body system of nucleons subject to the Pauli exclusion principle and moving independently from each other in a mean-field potential. The energy of the system for a given chemical potential is obtained by the variational principle.777Strictly speaking, we minimize the modified Hamiltonian defined by $H^{\prime}=H-\mu N$, where $\mu$ and $N$ denote the chemical potential and the number of particles, respectively. We also note that for NSs, the temperature $T$ can be regarded as zero; thus, the free and internal energy are equivalent. This approach is used in Refs [67, 68, 69, 70, 71, 72]. We can then estimate the pinning force from the pinning energy obtained above. There are two approaches for this calculation: * • Microscopic calculation We may estimate the pinning force through Eq. (3.18) for the nuclear pinning configuration. We refer to this estimation as the microscopic approach since, as illustrated in the most right window in Fig. 4, it focuses on the microscopic scale of $\mathcal{O}(R_{\rm WS})$. This approach considers the interaction between a vortex and the single nucleus in the Wigner-Seitz cell, and thus the interaction of the vortex with other distant nuclei is neglected. We obtain the pinning force per unit length by just multiplying the pinning force per nucleus with the number of nuclei in the unit length along the vortex line ($\simeq 1/R_{\mathrm{WS}}$). * • Mesoscopic calculation Vortex lines are much longer than the lattice spacing; therefore, each vortex line pins onto a large number of nuclei in reality. Such a vortex line does not align to the crystal axis over its total length in general. The mesoscopic approach considers this realistic configuration by taking the average of the force exerted on a vortex over the possible directions of the vortex line with respect to the crystal lattice. This calculation focuses on the length-scale $L\sim(10^{2}$–$10^{3})\times R_{\rm WS}$, for which the vortex line can be regarded as a straight line, as illustrated in the middle window in Fig. 4—we call this scale the mesoscopic scale. The derived pinning force thus tends to be smaller than those obtained with the microscopic calculation. Figure 4: The landscape of a vortex-line configuration at different length-scales. | Semi-classical | Quantum ---|---|--- Microscopic | Mean field: | Pairing: | Hartree-Fock: | Pairing: Woods-Saxon | Argonne | SLy4 | contact force $f_{\mathrm{pin}}=(1-7)\times 10^{-3}$ [71] | $f_{\mathrm{pin}}=(3-4)\times 10^{-4}$ [65] Mesoscopic | Mean field: | Pairing: | Hartree-Fock: | Pairing: Woods-Saxon | Argonne | SLy4 | contact force $f_{\mathrm{pin}}=8\times 10^{-7}-4\times 10^{-4}$ [72] | $f_{\mathrm{pin}}=5\times 10^{-7}-8\times 10^{-5}$ [66] Table 1: Calculations of the vortex pinning force considered in this work. The first row in each column lists the potentials for the mean-field and pairing interactions used in the evaluations; the second row shows the range of the pinning force in the inner crust in units of $\mathrm{MeV}\cdot\mathrm{fm}^{-2}$. All in all, we have $2\times 2=4$ combinations for the prescription of the pinning force calculation. In the following discussion, we consider a representative calculation with a specific choice of nuclear interactions for each combination, as summarized in Table 1. For the microscopic semi-classical approach, we consider the calculation in Ref. [71] with a Woods-Saxon potential for the mean-field potential and the Argonne interaction for the neutron-neutron pairing interaction. It is found that the nuclear pinning configuration (see Fig. 2) occurs only in high- density regions. The pinning force per nuclear pinning site is estimated by $E_{\mathrm{pin}}/R_{\mathrm{WS}}$ for this configuration in Ref. [71]. To convert this into the pinning force per unit length $f_{\mathrm{pin}}$, we multiply this by a factor of $1/(2R_{\mathrm{WS}})$, as cubic cells with the side length $2R_{\mathrm{WS}}$ are used in the calculation of Ref. [71]. As a result, we obtain the values of $f_{\mathrm{pin}}=\text{($1$--$7$)}\times 10^{-3}~{}\mathrm{MeV}\cdot\mathrm{fm}^{-2}$, depending on the position in the inner crust. We list quantities relevant to this calculation in Table 3 in Appendix A.888To get some ideas about the dependence of this calculation on the choice of potentials, we note that Ref. [71] also considers the case where the Gogny interaction is used instead of the Argonne interaction. In this case, the nuclear pinning occurs in higher density regions, and the pinning forces tend to be larger by a factor of a few, compared with the calculation with the Argonne interaction; see Table 4 in Appendix A for this calculation. Notice that the values of the pinning force shown here should be regarded as a ballpark estimate. The lower limit ($f_{\mathrm{pin}}=1\times 10^{-3}~{}\mathrm{MeV}\cdot\mathrm{fm}^{-2}$) could have been overestimated because of the discretization of the positions at which the pinning energy is estimated; as seen in Table 3, there is a density region of $\rho\sim 10^{13}~{}\mathrm{g}\cdot\mathrm{cm}^{-3}$ around which $E_{\mathrm{pin}}\simeq 0$, leading to a very small pinning force if we use Eq. (3.18). The upper limit could also be underestimated since the pinning force obtained by this equation corresponds to the average taken over the distance between the nuclear center and the interstitial position. For the microscopic quantum approach, we consider the results given in Ref. [65], where the SLy4 Skyrme interaction is used for the Hartree-Fock calculation and a density-dependent contact interaction is used for the neutron-neutron pairing interaction. The parameters of the contact interaction are discussed in Ref. [73]. Table 5 summarises the relevant quantities for this calculation. In contrast to the semi-classical analysis [71], nuclear pinning occurs in lower-density regions in this case. However, it also occurs in the highest-density regions (see Table 5 in Appendix A).999The qualitative feature described here does depend on the choice of the nuclear interactions. As shown in Ref. [65], if we use the SkM* interaction instead of SLy4, the nuclear pinning never occurs at high densities. The values of the pinning force estimated from the pinning energy and the Wigner-Seitz radius as in the semi-classical calculation are $f_{\mathrm{pin}}=\text{($3$--$4$)}\times 10^{-4}~{}\mathrm{MeV}\cdot\mathrm{fm}^{-2}$. See Ref. [65] for detailed discussions regarding the difference between the semi-classical and quantum results. The semi-classical mesoscopic calculation is given in Ref. [72], where the nuclear potentials are taken to be the same as in the semi-classical microscopic calculation in Table 1. The resultant values of the pinning force are found to be $f_{\mathrm{pin}}=8\times 10^{-7}$ – $4\times 10^{-4}~{}\mathrm{MeV}\cdot\mathrm{fm}^{-2}$, which are summarized in Table 6. We see that these values are considerably smaller than those for the semi- classical microscopic calculation in Ref. [71] due to the averaging over the vortex-line directions. For the quantum mesoscopic calculation, we consider the result given in Ref. [66], where the SLy4 interaction and a contact interaction are used for the Hartree-Fock calculation and pairing interactions, respectively, as in the quantum microscopic calculation in Table 1. As summarized in Table 7, the pinning force is found to be in the range $f_{\mathrm{pin}}=5\times 10^{-7}-8\times 10^{-5}~{}\mathrm{MeV}\cdot\mathrm{fm}^{-2}$.101010If we use the SkM* potential instead of SLy4, we obtain slightly smaller values of $f_{\mathrm{pin}}$, as also shown in Table 7. We again find that these values are much smaller than those obtained with the quantum microscopic approach. Before concluding this section, we note that we can also calculate the pinning force with a three-dimensional dynamical simulation of a vortex [74, 75, 76, 77].111111See also Ref. [78] for a kinetic approach. Such a calculation tends to be costly; thus, a certain degree of simplification is usually required for the moment. Besides, the current evaluation is limited to a few benchmark values of density. The estimated values of the pinning force are consistent with the above estimates. ### 4.2 Theoretical evaluation of $J$ For the evaluation of $J$ in Eq. (3.32), it is convenient to change the coordinate from cylindrical coordinates $(r,\varphi,z)$ to spherical coordinates $(R,\theta,\phi)$: $\displaystyle J$ $\displaystyle=\int_{\rm pin}dr\,dz\,d\varphi\,\rho r^{3}\cdot\frac{f_{\mathrm{pin}}}{\rho\kappa r}\simeq\int_{R_{\rm in}}^{R_{\rm out}}dR\,d\theta\,d\phi\,R^{3}\sin^{2}\theta\cdot\frac{f_{\rm pin}}{\kappa}\,,$ (4.35) where we approximate $\delta\Omega_{\infty}$ in Eq. (3.32) by $\delta\Omega_{\mathrm{cr}}$ in Eq. (3.29), which holds with good accuracy in the situation of our interest [50] as we mentioned in Sec. 3.2. We perform the integral over the range $[R_{\rm in},R_{\rm out}]$ where the pinning force is evaluated. We use the Akmal-Pandharipande-Ravenhall (APR) [79] equation of state to determine the NS core size and the equation of state tabulated in Crust_EOS_Cat_HZD-NV.dat in NSCool [80] based on Refs. [81, 82] to determine the density distribution in the crust. For the evaluation method of pinning force, we focus on the mesoscopic approach shown in Table 1, since for the microscopic calculation, the pinning force is obtained only in a limited region in the crust, as can be seen in Table 3–5. Considering that the evaluation of the pinning force suffers from large uncertainty depending on calculation methods, we make the following crude approximation in the calculation of the above integral—we neglect the density dependence of $f_{\mathrm{pin}}$ and fix it to a value in the range shown in Table 1. We thus obtain a range of $J$ accordingly, which we regard as the uncertainty of this pinning force estimation. As a result, we obtain $J=3.9\times 10^{40}-1.9\times 10^{43}~{}\mathrm{erg}\cdot\mathrm{s}$ for the semi-classic mesoscopic calculation and $J=1.7\times 10^{40}-2.7\times 10^{42}~{}\mathrm{erg}\cdot\mathrm{s}$ for the quantum mesoscopic calculation. ## 5 Vortex creep heating vs. observation No. | Type | Name | $\log_{10}t_{\rm sd}$ | $\log_{10}t_{\rm kin}$ | $\log_{10}|\dot{\Omega}|$ | $\log_{10}T_{\rm s}$ | $\log_{10}J_{\rm obs}$ ---|---|---|---|---|---|---|--- | | | [yr] | [yr] | [s-2] | [K] | [erg s] 1. | [Y] | PSR B1706-44 | $4.2$ | — | $-10.3$ | $5.68$–$6.34$ [83] | $41.9$–$44.6$ 2. | [O] | PSR J1740+1000 | $5.1$ | — | $-11.2$ | $5.89$ [84] | $43.8$ 3. | [Y] | PSR B2334+61 | $4.6$ | — | $-11.3$ | $5.76$ [85] | $43.3$ 4. | [O] | PSR B0656+14 | $5.0$ | — | $-11.6$ | $5.87$ [86] | $44.1$ 5. | [O] | PSR J0633+1746 | $5.5$ | — | $-11.9$ | $5.71$ [87] | $43.7$ 6. | [Y] | PSR J0538+2817 | $5.8$ | $4.60^{+0.18}_{-0.30}$[88] | $-11.9$ | $6.02$ [88] | $45.0$ 7. | [O] | PSR B1055-52 | $5.7$ | — | $-12.0$ | $5.81$ [89] | $45.1$ 8. | [X] | RX J1605.3+3249 | $4.5$ | $5.66^{+0.04}_{-0.07}$ [90] | $-12.1$ | $5.86$ [91] | $44.5$ 9. | [O] | PSR J2043+2740 | $6.1$ | — | $-12.1$ | $<5.95$ [92] | $<44.8$ 10. | [O] | PSR J1741-2054 | $5.6$ | — | $-12.2$ | $5.85$ [93] | $44.6$ 11. | [O] | PSR J0357+3205 | $5.7$ | — | $-12.4$ | $5.62$ [94] | $43.8$ 12. | [O] | PSR B0950+08 | $7.2$ | — | $-13.6$ | $4.78$–$5.08$ [29] | $41.7$–$42.9$ 13. | [X] | RX J0420.0-5022 | $6.3$ | — | $-13.8$ | $5.74$ [95] | $45.8$ 14. | [M] | PSR J0437-4715 | $9.20$ | — | $-14.0$ | $5.54$ [25] | $45.1$ 15. | [X] | RX J1308.6+2127 | $6.2$ | $5.74^{+0.16}_{-0.26}$ [96] | $-14.2$ | $6.08$ [97] | $47.5$ 16. | [X] | RX J0720.4-3125 | $6.3$ | $5.93^{+0.07}_{-0.26}$ [98] | $-14.2$ | $6.02$ [99] | $47.3$ 17. | [M] | PSR J2124-3358 | $9.58$ | — | $-14.3$ | $4.70$–$5.32$ [26] | $42.0$–$44.5$ 18. | [X] | RX J1856.5-3754 | $6.6$ | $5.66^{+0.04}_{-0.05}$[98] | $-14.4$ | $5.65$ [100] | $46.0$ 19. | [X] | RX J2143.0+0654 | $6.6$ | — | $-14.5$ | $5.67$–$6.06$ [101, 102] | $46.1$–$47.8$ 20. | [X] | RX J0806.4-4123 | $6.5$ | — | $-14.6$ | $6.01$ [103] | $47.6$ 21. | [O] | PSR J0108-1431 | $8.3$ | — | $-15.1$ | $4.43$–$4.74$ [28] | $41.8$–$43.1$ 22. | [O] | PSR J2144-3933 | $8.4$ | — | $-16.4$ | $<4.62$ [104] | $<43.8$ Table 2: The data of the NSs considered in this paper. We classify them into four types—ordinary pulsars younger than $10^{5}~{}\mathrm{yrs}$ [Y], ordinary pulsars older than $10^{5}~{}\mathrm{yrs}$ [O], XDINSs [X], and millisecond pulsars [M]. The values of $t_{\rm sd}$ and $\dot{\Omega}$ without references are computed from the data given in the ATNF pulsar catalogue [105, 106]. The value of $J_{\rm obs}$ is evaluated as in Eq. (5.36) with $R_{\rm NS}=11.43~{}\mathrm{km}$. We now compare the prediction of the vortex-creep heating mechanism with observation. For this purpose, it is useful to calculate the following quantity for each NS:121212We neglect the gravitational redshift factor since its effect is within the $\mathcal{O}(1)$ uncertainty of $J$ discussed below. $\displaystyle J_{\rm obs}$ $\displaystyle\equiv\frac{4\pi R_{\rm NS}^{2}\sigma_{\rm SB}T_{\rm s}^{4}}{|\dot{\Omega}|}~{}.$ (5.36) As evident from Eqs.(2.2), (2.3), and (3.31), this corresponds to the $J$ parameter inferred from the observation of each NS. This inference assumes the steady creeping of vortices (discussed in Sec.3.2) and the balance between vortex-creep heating luminosity and photon cooling luminosity, which we expect to hold if the NS is older than $\sim 10^{5}$ years. Since NSs are comparable in size and mass, we expect that $J_{\rm obs}$ is also roughly equal (up to a factor of $\mathcal{O}(1)$) for every NS. We test this expectation by using the data of $J_{\rm obs}$ for old NSs. We also compare the values of $J_{\mathrm{obs}}$ with the theoretical computations given in Sec. 4.2. In Table 2, we list the values of $J_{\mathrm{obs}}$ for the NSs we consider in this paper. We select isolated NSs older than $10^{4}$ yrs. In evaluating $J_{\mathrm{obs}}$, we have just assumed $R_{\rm NS}=11.43~{}\mathrm{km}$ for all NSs, as the radius is poorly known for most NSs; we keep in mind that this may introduce an $\mathcal{O}(1)$ error in the determination of $J_{\mathrm{obs}}$. We also show the age, surface temperature, and $\dot{\Omega}=2\pi\dot{P}/P^{2}$ of the NSs, ($P$ and $\dot{P}$ are the period and its time derivative, respectively). Regarding the NS age, we use the kinetic age $t_{\rm kin}$ if available. Otherwise, we use the spin-down age $t_{\rm sd}=P/(2\dot{P})$. We calculate $t_{\rm sd}$ and $\dot{\Omega}$ from $P$ and $\dot{P}$ given in the Australia Telescope National Facility (ATNF) pulsar catalogue [105, 106]. Notice that the surface temperatures of some of the old NSs in this table are much higher than the predicted temperature in the standard NS cooling scenario [30, 31, 32, 33, 34, 35]. It is important to note that not all NSs listed in Table 2 are useful for testing the vortex-creep heating. As we have discussed in Sec. 2, photon emission becomes the dominant cooling source for NSs older than $\sim 10^{5}$ years. For younger NSs, this may not be the case, so $L^{\infty}_{\gamma}\lesssim L^{\infty}_{\rm H}$ instead of (2.3), for which the values of $J_{\mathrm{obs}}$ in Table 2 may be underestimated. To distinguish such young NSs from others, we indicate them by the type [Y] in the table. Another class of NSs that are inappropriate for our test is the X-ray Dim Isolated NSs (XDINSs). These NSs are considered to be descendants of magnetars [103, 95] that experienced the decay of magnetic fields before. This may make these NSs hotter than ordinary NSs of the same age [107, 95, 108, 109], resulting in an overestimate of $J_{\mathrm{obs}}$. We denote these NSs by the type [X]. The rest of the NSs, which we use for the test of the vortex- creep heating, are classified into old ordinary pulsars [O] and millisecond pulsars [M]. The uncertainty in the determination of $J_{\mathrm{obs}}$ stems mainly from that in the surface temperature, which is significant due to its quartic dependence on $T_{\rm s}$. Generically speaking, it is very difficult to identify all of the sources of uncertainties in the measurement of the NS surface temperature, and it is often the case that the error shown in the literature is only a part of them, such as those from the spectrum fitting, the determination of the distance and/or radius of the NS, and so on. At present, it is fair to say that the NS temperature measurement typically suffers from $\mathcal{O}(1)$ uncertainty, as can be seen in, e.g., Ref. [110]. Motivated by this, we include a factor of two uncertainty in $T_{\rm s}$ for the stars in Table 2 for which only the central value is presented. For the other stars, we describe our prescription for the error estimation in Appendix B. We have checked that the errors thus obtained are similar to or more conservative than those adopted in Ref. [110]. Figure 5: The values of the $J$ parameter obtained from the observation. The grey triangles, blue circles, green inverse triangles, and orange stars correspond to the young ordinary pulsars ([Y]), the old ordinary pulsars ([O]), the XDINSs ([X]), and the millisecond pulsars ([M]), respectively. The points with an arrow indicate upper limits. The red shaded region shows observationally favored range, $J\simeq 10^{42.9\text{--}43.8}~{}\mathrm{erg}\cdot\mathrm{s}$. For comparison, we also show the values of $J$ estimated with the mesoscopic calculations by black bars. In Fig. 5, we show the range of $J_{\mathrm{obs}}$ estimated as described above for each NS listed in Table 2. The grey triangles, green inverse triangles, blue circles, and orange stars represent the young ordinary pulsars ([Y]), the XDINSs ([X]), the old ordinary pulsars ([O]), and the millisecond pulsars ([M]), respectively. The points with an arrow indicate that we only have an upper limit on $J_{\rm obs}$ for those NSs. Recall that we are concerned only with the NSs represented by the blue [O] and orange [M] points. It is found that the estimated values of $J_{\mathrm{obs}}$ for these NSs are in the same ballpark, $J\sim 10^{43}~{}\mathrm{erg}\cdot\mathrm{s}$, even though their $|\dot{\Omega}|$’s distribute over orders of magnitude. This is in good agreement with the prediction of the vortex-creep heating mechanism. On the other hand, $J_{\mathrm{obs}}$ for the green points (XDINSs [X]) tend to be larger than this, as expected. We also show the theoretical estimations given in Sec. 4.2 in the upper panel of Fig. 5. We see that the semi-classical mesoscopic calculation is consistent with the observation, given that this theoretical estimation suffers from a NS-dependent uncertainty of $\mathcal{O}(1)$ coming from the integration in Eq. (4.35), in addition to that from the estimation of $f_{\mathrm{pin}}$. The quantum mesoscopic calculation can explain some of the points with a small $J_{\mathrm{obs}}$, but they are not large enough to explain, e.g., that of J0437-4715. However, we note that this theoretical estimation is still allowed by the observations since it just results in a lower heating luminosity than the observed one. If this is the case, the vortex-creep heating may operate but there exists another heating mechanism that dominates the vortex-creep heating, such as the rotochemical heating [111, 112, 113, 114, 115, 116, 117, 118, 119]. It is premature to establish the existence of the vortex-creep heating, as well as to conclude if an extra heating mechanism is required to be present. To that end, we need to accumulate more data on the surface temperature of old NSs with high accuracy, which we anticipate to be provided by future optical, UV, and X-ray observations.131313See, for instance, Ref. [120]. Nevertheless, obtaining a current compilation of the value of $J$ suggested by the observation is intriguing. Considering intrinsic $\mathcal{O}(1)$ uncertainty in $J$, we determine its rough range by requiring that it covers the range suggested by B0950+08, which favors the smallest value, and satisfies the upper limit set by J2144-3944 based on non-observation of thermal flux. This yields $\displaystyle J\simeq 10^{42.9-43.8}~{}\mathrm{erg}\cdot\mathrm{s}~{},$ (5.37) which we show as the red band in Fig. 5. (a) Ordinary pulsars (b) Millisecond pulsars Figure 6: The evolution of NS surface temperature with (without) the vortex creep heating effect in the red band (black dashed line). The band corresponds to the range of $J$ in Eq. (5.37). The dots with error bars show the observed temperatures, presented in Table 2, with the same colors as in Fig. 5. Finally, in Fig. 6, we show the evolution of NS surface temperature with (without) the vortex creep heating effect in the red band (black dashed line).141414In these plots, we use the APR equation of state [79] for a NS mass of $1.4M_{\odot}$ to calculate the NS structure. For Cooper pairing gap models, we use the SFB model [121] for the neutron singlet pairing, the model “b” in Ref. [33] for the neutron triplet pairing, and the CCDK model [122] for the proton singlet pairing. The temperature evolution at late times scarcely depends on the choice of these models. The band corresponds to the range of $J$ in Eq. (5.37). The dots with error bars show the observed temperatures in Table 2, with the same colors as in Fig. 5. In Fig. 6(a), we take $P\dot{P}=10^{-15}$ s and the initial period $P_{0}=10~{}\mathrm{ms}$ to calculate $|\dot{\Omega}(t)|$, where we assume that the external torque is dominated by magnetic dipole radiation.151515In this case, we have $\dot{\Omega}\propto-\Omega^{3}$, i.e., $P\dot{P}=\text{constant}$, and by solving this we obtain $\displaystyle|\dot{\Omega}|(t)=\frac{\pi}{\sqrt{2P\dot{P}}}\left[t+t_{\mathrm{sd},0}\right]^{-3/2},$ with $t_{\mathrm{sd},0}\equiv P_{0}^{2}/(2P\dot{P})$. For the choice of parameters in Fig. 6(a) and Fig. 6(b), $t_{\mathrm{sd},0}\simeq 2\times 10^{3}$ and $5\times 10^{7}$ years, respectively. The value of $P\dot{P}$ is related to the surface magnetic flux density $B_{s}$. In the ATNF pulsar catalogue [105, 106], $B_{s}=3.2\times 10^{19}(P\dot{P})^{1/2}$ G is used for this relation, with which we have $B_{s}\simeq 1.0\times 10^{12}$ G and $5.8\times 10^{8}$ G in Fig. 6(a) and Fig. 6(b), respectively. These values are typical for ordinary pulsars. Nevertheless, we note that $|\dot{\Omega}(t)|$ obtained with these parameters do not exactly agree with the observed values of $|\dot{\Omega}|$ in Table 2, so the data points shown in this figure should be regarded as just an eye guide. In Fig. 6(b), we set $P\dot{P}=3.3\times 10^{-22}$ s, which is the observed value for PSR J2124-3358, and $P_{0}=1~{}\mathrm{ms}$. As we see in these plots, the predicted temperature with the vortex creep heating starts to deviate from that in the standard cooling scenario at $t\sim 10^{5}$ years and remains high enough at later times to be compatible with the observed data. ## 6 Conclusion and discussion We have revisited the vortex-creep heating mechanism in light of recent observations of old warm NSs. As we have seen, this heating mechanism gives a characteristic prediction that the heating luminosity is proportional to $|\dot{\Omega}|$, with the proportional constant $J$ having an almost universal value over NSs since the NS structure and the vortex-nuclear interactions determine it. We have found that this prediction agrees with the observational data of old NSs, with the favored range of $J$ in the same ballpark as the theoretical calculations. Notice that the scenario where vortex creep heating dominates all NSs can readily be overturned if we discover a NS having $J$ much smaller than those presented in Fig. 5. On the other hand, if we find a NS with a larger $J$, we can disfavor our scenario only after excluding the existence of other heating sources specific to this NS, such as accretion from its environment. It is possible that other heating mechanisms also work in old NSs. Indeed, we have already considered potential heating caused by the decay of magnetic fields in XDINSs (see Sec. 5), and we have not used these NSs in our test of the vortex-creep heating mechanism for this reason. Another heating mechanism that may operate without relying on exotic phenomena is provided by the out- of-equilibrium beta processes, which is dubbed rotochemical heating [111, 112, 113, 114, 115, 116, 117, 118, 119]. It is known that this rotochemical heating mechanism can increase the surface temperature of old NSs up to $\sim 10^{6}$ K. Thus, its heating luminosity can be comparable to or even dominate the vortex-creep heating. It would be worthwhile to study the vortex creep heating in the presence of rotochemical heating and compare its prediction with the temperature observations of old warm NSs. The NS heating caused by the accretion of dark matter particles is also widely discussed in the literature [123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155]. In this case, the surface temperature at late times is predicted to be $\text{a few}\times 10^{3}$ K, and thus this heating effect could be concealed by the vortex heating mechanism, making it improbable to observe the dark matter signature through the temperature observation of old NSs. A detailed study of this issue will be given in the forthcoming paper [156]. ## Acknowledgments MF thanks Kazuyuki Sekizawa and Tomoya Naito for the fruitful discussion about the current situation and the possible future directions in the pinning force evaluation. This work is supported in part by the Collaborative Research Center SFB1258 and by the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation) under Germany’s Excellence Strategy - EXC-2094 - 390783311 [MF], JSPS Core-to-Core Program (No. JPJSCCA 20200002 [MF]), the Grant-in-Aid for Innovative Areas (No.19H05810 [KH and MRQ], No.19H05802 [KH], No.18H05542 [MF and NN]), Scientific Research B (No.20H01897 [KH and NN]), Young Scientists (No.21K13916 [NN]). ## Appendix A Pinning force In this appendix, we show the density dependence of the quantities relevant to the pinning force calculations discussed in Sec. 4.1. Tables 3 and 4 are for the microscopic semi-classical approach in Ref. [71], where the Argonne and Gogny interactions are used for the nuclear pairing interaction, respectively. The element corresponding to the cell nuclear composition and Wigner-Seitz radius $R_{\rm WS}$ are derived for each baryon density $\rho$ in Ref. [81]. In Ref. [71], the pinning force is evaluated only for the nuclear pinning configuration, and thus we show the values of $E_{\mathrm{pin}}$ and $f_{\mathrm{pin}}=|E_{\mathrm{pin}}|/(2R_{\mathrm{WS}}^{2})$ only for this case. The labels of NP and IP in the last column indicate the nuclear and interstitial pinnings, respectively. zone | Element | $\rho$ | $R_{\rm WS}$ | $\xi$ | $E_{\rm pin}$ | $f_{\rm pin}$ | config ---|---|---|---|---|---|---|--- | | [g cm-3] | [fm] | [fm] | [MeV] | [MeV fm-2] | 1 | ${}^{320}_{40}\mathrm{Zr}$ | $1.5\times 10^{12}$ | $44.0$ | $7.02$ | — | — | IP 2 | ${}^{1100}_{50}\mathrm{Sn}$ | $9.6\times 10^{12}$ | $35.5$ | $4.34$ | — | — | IP 3 | ${}^{1800}_{50}\mathrm{Sn}$ | $3.4\times 10^{13}$ | $27.0$ | $8.54$ | $-5.2$ | $0.0036$ | NP 4 | ${}^{1500}_{40}\mathrm{Zr}$ | $7.8\times 10^{13}$ | $19.4$ | $11.71$ | $-5.1$ | $0.0068$ | NP 5 | ${}^{982}_{32}\mathrm{Ge}$ | $1.3\times 10^{14}$ | $13.8$ | $8.62$ | $-0.4$ | $0.0011$ | NP Table 3: Quantities relevant to the pinning force calculation obtained with the microscopic semi-classical approach in Ref. [71], where the Argonne potential is used for the nuclear pairing interaction. $\rho$, $R_{\mathrm{WS}}$, and $\xi$ are the mass density, Wigner-Seitz radius, and coherence length, respectively. zone | Element | $\rho$ | $R_{\rm WS}$ | $\xi$ | $E_{\rm pin}$ | $f_{\rm pin}$ | config ---|---|---|---|---|---|---|--- | | [g cm-3] | [fm] | [fm] | [MeV] | [MeV fm-2] | 1 | ${}^{320}_{40}\mathrm{Zr}$ | $1.5\times 10^{12}$ | $44.0$ | $7.76$ | — | — | IP 2 | ${}^{1100}_{50}\mathrm{Sn}$ | $9.6\times 10^{12}$ | $35.5$ | $4.07$ | — | — | IP 3 | ${}^{1800}_{50}\mathrm{Sn}$ | $3.4\times 10^{13}$ | $27.0$ | $3.93$ | — | — | IP 4 | ${}^{1500}_{40}\mathrm{Zr}$ | $7.8\times 10^{13}$ | $19.4$ | $7.78$ | $-7.5$ | $0.010$ | NP 5 | ${}^{982}_{32}\mathrm{Ge}$ | $1.3\times 10^{14}$ | $13.8$ | $8.62$ | $-5.9$ | $0.015$ | NP Table 4: Quantities relevant to the pinning force calculation obtained with the microscopic semi-classical approach in Ref. [71], where the Gogny potential is used for the nuclear pairing interaction. zone | Element | $\rho$ | $n$ | $R_{\rm WS}$ | $\xi$ | $E_{\rm pin}$ | $f_{\rm pin}$ | config ---|---|---|---|---|---|---|---|--- | | [g cm-3] | [fm-3] | [fm] | [fm] | [MeV] | [MeV fm-2] | 1a | ${}^{320}_{40}\mathrm{Zr}$ | $1.7\times 10^{12}$ | $0.001$ | $43.3$ | $4.43$ | $-1.08$ | $0.00029$ | NP 1b | ${}^{320}_{40}\mathrm{Zr}$ | $3.4\times 10^{12}$ | $0.002$ | $40.0$ | $4.21$ | $-1.20$ | $0.00038$ | NP 1c | ${}^{320}_{40}\mathrm{Zr}$ | $6.7\times 10^{12}$ | $0.004$ | $36.9$ | $3.93$ | — | — | IP 2a | ${}^{1100}_{50}\mathrm{Sn}$ | $1.3\times 10^{13}$ | $0.008$ | $33.0$ | $4.04$ | — | — | IP 2b | ${}^{1100}_{50}\mathrm{Sn}$ | $1.8\times 10^{13}$ | $0.011$ | $31.0$ | $4.12$ | — | — | IP 2c | ${}^{1100}_{50}\mathrm{Sn}$ | $2.8\times 10^{13}$ | $0.017$ | $28.0$ | $4.70$ | — | — | IP 3a | ${}^{1800}_{50}\mathrm{Sn}$ | $4.3\times 10^{13}$ | $0.026$ | $24.5$ | $6.05$ | — | — | IP 3b | ${}^{1800}_{50}\mathrm{Sn}$ | $6.2\times 10^{13}$ | $0.037$ | $21.4$ | $8.75$ | $-0.41$ | $0.00045$ | NP Table 5: Relevant quantities for the pinning force calculation obtained with the microscopic quantum approach in Ref. [65]. The results for the microscopic quantum approach in Ref. [65] are summarized in Table 5, where the SLy4 Skyrme interaction is used for the mean-field interaction. The Wigner-Seitz radius $R_{\rm WS}$ shown in this table is interpolated from the plot in Ref. [81]. For the evaluation of $f_{\mathrm{pin}}$, we again use the formula $f_{\mathrm{pin}}=|E_{\mathrm{pin}}|/(2R_{\mathrm{WS}}^{2})$ and show the values only for the nuclear pinning, just for easy comparison with the semi- classical calculations. zone | | | $f_{\rm pin}$ [MeV fm-2] | | ---|---|---|---|---|--- $L=100R_{\rm WS}$ | $L=500R_{\rm WS}$ | $L=1000R_{\rm WS}$ | $L=2500R_{\rm WS}$ | $L=5000R_{\rm WS}$ 1 | $7.63\times 10^{-6}$ | $2.29\times 10^{-6}$ | $1.49\times 10^{-6}$ | $9.30\times 10^{-7}$ | $7.68\times 10^{-7}$ 2 | $2.12\times 10^{-5}$ | $6.34\times 10^{-6}$ | $4.06\times 10^{-6}$ | $2.62\times 10^{-6}$ | $2.12\times 10^{-6}$ 3 | $1.43\times 10^{-4}$ | $4.49\times 10^{-5}$ | $2.74\times 10^{-5}$ | $1.54\times 10^{-5}$ | $1.14\times 10^{-5}$ 4 | $3.84\times 10^{-4}$ | $1.10\times 10^{-4}$ | $6.89\times 10^{-5}$ | $4.23\times 10^{-5}$ | $3.32\times 10^{-5}$ 5 | $7.85\times 10^{-5}$ | $2.71\times 10^{-5}$ | $1.89\times 10^{-5}$ | $1.36\times 10^{-5}$ | $1.12\times 10^{-5}$ Table 6: The pinning force obtained in the semi-classical mesoscopic approach for different values of $L$ over which the forces exerted on a vortex are integrated [72]. The zones correspond to those in Table 3. In Table 6, we show the pinning force obtained in the semi-classical mesoscopic approach for different values of $L$ over which the forces exerted on a vortex are integrated [72]. The zone numbers in the first column correspond to those in Table 3. We show the calculation in which the reduction of the pairing gap due to the polarization effects in the nuclear matter is not included, corresponding to the choice of the reduction factor $\beta=1$ introduced in Ref. [72]. Because of the averaging procedure, we find that a larger $L$ results in a smaller value of $f_{\mathrm{pin}}$. model | zone | $E_{\rm pin}$ | config | | $f_{\rm pin}$ [MeV fm-2] | ---|---|---|---|---|---|--- | | [MeV] | | $L=1000R_{\rm WS}$ | $L=2500R_{\rm WS}$ | $L=5000R_{\rm WS}$ SLy4 | 1a | $-0.72$ | NP | $1.39\times 10^{-6}$ | $7.38\times 10^{-7}$ | $5.40\times 10^{-7}$ 1b | $-0.91$ | NP | $1.97\times 10^{-6}$ | $1.08\times 10^{-6}$ | $8.00\times 10^{-7}$ 1c | $-0.89$ | NP | $2.20\times 10^{-6}$ | $1.19\times 10^{-6}$ | $8.74\times 10^{-7}$ 2a | $2.73$ | IP | $5.61\times 10^{-6}$ | $3.72\times 10^{-6}$ | $2.95\times 10^{-6}$ 2b | $3.01$ | IP | $7.52\times 10^{-6}$ | $5.03\times 10^{-6}$ | $4.00\times 10^{-6}$ 2c | $10.00$ | IP | $1.47\times 10^{-5}$ | $1.01\times 10^{-5}$ | $8.08\times 10^{-6}$ 3a | $11.78$ | IP | $3.25\times 10^{-5}$ | $2.31\times 10^{-5}$ | $1.88\times 10^{-5}$ 3b | $9.85$ | IP | $8.47\times 10^{-5}$ | $6.41\times 10^{-5}$ | $5.31\times 10^{-5}$ SkM* | 1a | $-0.72$ | NP | $3.61\times 10^{-7}$ | $1.60\times 10^{-7}$ | $9.50\times 10^{-8}$ 1b | $-0.91$ | NP | $2.20\times 10^{-7}$ | $8.72\times 10^{-8}$ | $4.08\times 10^{-8}$ 1c | $-0.89$ | NP | $2.83\times 10^{-6}$ | $1.82\times 10^{-6}$ | $1.44\times 10^{-6}$ 2a | $2.73$ | IP | $5.68\times 10^{-6}$ | $3.73\times 10^{-6}$ | $2.93\times 10^{-6}$ 2b | $3.01$ | IP | $7.58\times 10^{-6}$ | $5.01\times 10^{-6}$ | $4.00\times 10^{-6}$ 2c | $10.00$ | IP | $1.25\times 10^{-5}$ | $8.50\times 10^{-6}$ | $6.76\times 10^{-6}$ 3a | $11.78$ | IP | $2.54\times 10^{-5}$ | $1.80\times 10^{-5}$ | $1.45\times 10^{-5}$ 3b | $9.85$ | IP | $8.00\times 10^{-5}$ | $5.93\times 10^{-5}$ | $4.79\times 10^{-5}$ Table 7: The pinning force obtained in the quantum mesoscopic calculation where the Skytme interactions, SLy4 and SkM*, are used for the mean-field potential [66]. The zones correspond to those in Table 5. Table 7 shows the pinning force obtained in the quantum mesoscopic calculation where the SLy4 and SkM* Skytme interactions are used for the Hartree-Fock calculation [66]. The zones correspond to those in Table 5. We again show the results obtained without including the polarization effect, i.e., $\beta=1$ as in the previous case. Figure 7: The values of $f_{\rm pin}$ given in the tables in this appendix against the density $\rho$. The filled and opened markers correspond to the nuclear and interstitial pinnings. Finally, we plot the values of the pinning force for each density region in Fig. 7. The filled and opened markers correspond to the nuclear and interstitial pinnings. As we see, the values of $f_{\rm pin}$ are distributed in the range $10^{-8}$–$10^{-2}$ MeV $\cdot$ fm-2, depending on the evaluation scheme and the selected nuclear potential. ## Appendix B Selection criteria of NS data We explain how we choose the range of uncertainty of $T_{\rm s}$ for each NS shown in Table 2. * • No. 1, PSR B1706-4: In Ref. [83], the X-ray data of PSR B1706-44 obtained in XMM-Newton is fitted by the BB (blackbody), BB+PL (power law), and atmosphere+PL models, and only BB+PL and atmosphere+PL models result in acceptable $\chi^{2}$ values. The atmosphere model includes the light-element NS atmosphere (e.g. dominated by Hydrogen) and shows large Wien excesses in the high-energy region. Therefore, atmosphere+PL models tend to favor lower temperature and larger radius for the emitting area. We selected the minimum and maximum among the BB+PL and atmosphere+PL models, $T^{\infty}=(0.48-2.2)\times 10^{6}~{}\mathrm{K}$, to include the uncertainty coming from the choice of fitting models. * • No. 9, PSR J2043+2740: Ref. [157] studied the XMM-Newton data of PSR J2043+2740. Using the BB + PL model, the upper bound is derived as $T^{\infty}_{s}<6.27\times 10^{5}~{}\mathrm{K}$, where $R_{\rm NS}=10~{}\mathrm{km}$ is assumed for the emission radius. On the other hand, Ref. [158] also fitted the X-ray data of the XMM-Newton and obtained an even higher BB temperature, $T^{\infty}_{s}\simeq 9\times 10^{5}~{}\mathrm{K}$ with the radiation radius $R^{\infty}\simeq 2~{}\mathrm{km}~{}\text{\cite[cite]{[\@@bibref{Number}{Zavlin:2004wt}{}{}]}}$. Although the fitted radius is smaller than the expected NS radius, it is too large to be interpreted as the magnetic cap radius. Thus, we can not exclude the possibility that this BB temperature corresponds to the emission from the NS surface. To evaluate $T_{\rm s}$ conservatively, we chose the highest value of the BB temperature as an upper bound. * • No. 12, PSR B0950+08: In Ref. [29], the optical-UV flux of PSR B0950+08 obtained in the Hubble Space Telescope (HST) far-UV (FUV) detector is analyzed. The best-fit temperature is obtained as $T_{\rm s}=(6-12)\times 10^{4}~{}\mathrm{K}$, and we decided to use the proposed value. Note that the conservative upper bound is also derived as $T_{\rm s}<1.7\times 10^{5}~{}\mathrm{K}$ by varying the parameter, such as the ratio of NS radius and distance. * • No. 17, PSR J2124-3358: Ref. [26] analyzed the optical data from the J2124-3358. The BB+PL model gives the following possible range $T_{\rm s}\in[0.5,2.1]\times 10^{5}~{}\mathrm{K}$ with the uncertainty from the distance. We decided to select the original range, which is almost the same uncertainty added by hand in the way we described in Sec. 5. * • No. 19, RX J2143.0+0654: In Ref. [101], the X-ray data in XMM-Newton is fitted using the BB absorption model, and the authors obtain the BB temperature $k_{\rm B}T^{\infty}=104.0\pm 0.4~{}\mathrm{eV}$ with the BB radius as $R^{\infty}=(3.10\pm 0.04)~{}\mathrm{km}$, where distance is fixed to be $d=500~{}\mathrm{pc}$. This value is smaller than the typical NS radius, which implies that this BB temperature is not from the surface but from the small areas around the magnetic caps. In Ref. [102], the authors fitted the data from Large Binocular Telescope (optical) by combining with the X-ray data in XMM-Newton. Using the BB absorption model, $k_{\rm B}T^{\infty}=105.1\pm 0.9~{}\mathrm{eV}$ is obtained, which is consistent with the result in Ref. [101]. They also perform fitting using the two-component BB model and the hotter (cooler) component is obtained as $k_{\rm B}T=104~{}\mathrm{eV}$ ($k_{\rm B}T=40~{}\mathrm{eV}$). It is impossible to eliminate uncertainty from the model selection to fit the data from this situation, and thus we choose all the possible ranges of the temperature, $k_{\rm B}T_{\rm s}=40$–$106~{}\mathrm{eV}$. * • No. 20, PSR J0108-1431: The X-ray data from the direction of J1080-1431 observed in XMM-Newton is fitted by the BB+PL model [159] with $k_{\rm B}T=110^{+30}_{-10}~{}\mathrm{eV}$ and $R_{\rm NS}=43^{+16}_{-9}~{}\mathrm{m}$. This small emission radius implies that this BB component is not the cool surface temperature but the hot magnetic pole component. We can interpret this result as the surface temperature is much cooler than the magnetic pole, and thus, the hot component dominates the observed flux. The latest analysis [28] analyses both the XMM-Newton and optical data (HST, VLT). In particular, HST F140LP detected thermal emission, and they put the conservative upper bound on the surface temperature as $T_{\rm s}<5.9\times 10^{4}~{}\mathrm{K}$. To derive this conservative bound, they included uncertainty from the parallax distance [160]. Furthermore, they obtain the value of $T_{\rm s}$ by assuming the FUV flux is dominated by a thermal component, $T_{\rm s}=27000-55000~{}\mathrm{K}$. We selected this range to represent the uncertainty. * • No. 22, PSR J2144-3933: The upper bound is obtained for the surface temperature of J2144-3933 using XMM-Newton data (combining with the optical data of Very Large Telescope (VLT)) [161] as $T_{\rm s}<2.3\times 10^{5}~{}\mathrm{K}$. The latest analysis [104] used deep optical and FUV observation data by HST and derived the upper bound on the surface temperature of J2144-3933. The conservative upper bound on the surface temperature is derived based on the non-detection, $\displaystyle T_{\rm s}<4.2\times 10^{4}~{}\mathrm{K},$ (B.38) where a range of NS radius $R_{\rm NS}=[11,13]~{}\mathrm{km}$ and parallax distance $d=172^{+20}_{-15}~{}\mathrm{pc}$ is considered to estimate the uncertainty [160]. In this analysis, NS mass is fixed as $M_{\rm NS}=1.4~{}M_{\odot}$. Let us also comment on the rejected observational data from our list. * $-$ PSR B1929+10: The BB+PL fit is performed for the X-ray data [162]. However, the magnetic pole component is reported to dominate the temperature because the fitted radiation radius is much smaller than the NS radius. We conclude this data is not appropriate to test the vortex creep heating. * $-$ XMMU J1732-344: We also omit XMMU J1732-344 from our list because the observed value of $|\dot{\Omega}|$ is not determined. Once its pulsation data is fixed, it is worth studying whether the vortex creep heating can explain this data; its thermal emission is expected to exceed the value expected from minimal cooling [163, 164] with its kinetic time information [163]. ## References * [1] A. Bohr, B. R. Mottelson, and D. Pines, Possible Analogy between the Excitation Spectra of Nuclei and Those of the Superconducting Metallic State, Phys. Rev. 110 (1958) 936–938. * [2] A. B. Migdal, Superfluidity and the moments of inertia of nuclei, Nuclear Physics 13 (1959) 655–674. * [3] D. Page, J. M. Lattimer, M. Prakash, and A. W. Steiner, Stellar Superfluids, arXiv:1302.6626 (2013). * [4] B. Haskell and A. Sedrakian, Superfluidity and Superconductivity in Neutron Stars, Astrophys. Space Sci. Libr. 457 (2018) 401–454 [arXiv:1709.10340]. * [5] A. Sedrakian and J. W. Clark, Superfluidity in nuclear systems and neutron stars, Eur. Phys. J. A 55 (2019) 167 [arXiv:1802.00017]. * [6] B. Haskell and A. Melatos, Models of Pulsar Glitches, Int. J. Mod. Phys. D 24 (2015) 1530008 [arXiv:1502.07062]. * [7] J. R. Fuentes, et al., The glitch activity of neutron stars, Astron. Astrophys. 608 (2017) A131 [arXiv:1710.00952]. * [8] D. Antonopoulou, B. Haskell, and C. M. Espinoza, Pulsar glitches: observations and physical interpretation, Rept. Prog. Phys. 85 (2022) 126901. * [9] S. Zhou, E. Gügercinoğlu, J. Yuan, M. Ge, and C. Yu, Pulsar Glitches: A Review, Universe 8 (2022) 641 [arXiv:2211.13885]. * [10] P. W. Anderson and N. Itoh, Pulsar glitches and restlessness as a hard superfluidity phenomenon, Nature 256 (1975) 25–27. * [11] M. A. Alpar, Pinning and Threading of Quantized Vortices in the Pulsar Crust Superfluid, Astrophys. J. 213 (1977) 527–530. * [12] M. A. Alpar, D. Pines, P. W. Anderson, and J. Shaham, Vortex creep and the internal temperature of neutron stars. I - General theory, Astrophys. J. 276 (1984) 325–334. * [13] M. A. Alpar, P. W. Anderson, D. Pines, and J. Shaham, Vortex creep and the internal temperature of neutron stars. II. VELA pulsar., Astrophys. J. 278 (1984) 791–805. * [14] M. A. Alpar, R. Nandkumar, and D. Pines, Vortex creep and the internal temperature of neutron stars : the Crabpulsar and PSR 0525+21., Astrophys. J. 288 (1985) 191–195. * [15] N. Shibazaki and F. K. Lamb, Neutron Star Evolution with Internal Heating, Astrophys. J. 346 (1989) 808. * [16] M. A. Alpar, K. S. Cheng, and D. Pines, Vortex Creep and the Internal Temperature of Neutron Stars: Linear and Nonlinear Response to a Glitch, Astrophys. J. 346 (1989) 823. * [17] K. A. van Riper, R. I. Epstein, and G. S. Miller, Soft X-Ray Pulses from Neutron Star Glitches, Astrophys. J. Lett. 381 (1991) L47. * [18] H. Umeda, N. Shibazaki, K. Nomoto, and S. Tsuruta, Thermal Evolution of Neutron Stars with Internal Frictional Heating, Astrophys. J. 408 (1993) 186. * [19] K. A. van Riper, B. Link, and R. I. Epstein, Frictional Heating and Neutron Star Thermal Evolution, Astrophys. J. 448 (1995) 294 [astro-ph/9404060]. * [20] C. Schaab, A. Sedrakian, F. Weber, and M. K. Weigel, Impact of internal heating on the thermal evolution of neutron stars, Astron. Astrophys. 346 (1999) 465–480 [astro-ph/9904127]. * [21] M. B. Larson and B. Link, Superfluid friction and late-time thermal evolution of neutron stars, Astrophys. J. 521 (1999) 271 [astro-ph/9810441]. * [22] D. Gonzalez and A. Reisenegger, Internal Heating of Old Neutron Stars: Contrasting Different Mechanisms, Astron. Astrophys. 522 (2010) A16 [arXiv:1005.5699]. * [23] O. Kargaltsev, G. G. Pavlov, and R. W. Romani, Ultraviolet emission from the millisecond pulsar j0437-4715, Astrophys. J. 602 (2004) 327–335 [astro-ph/0310854]. * [24] R. P. Mignani, G. G. Pavlov, and O. Kargaltsev, A possible optical counterpart to the old nearby pulsar J0108-1431, Astron. Astrophys. 488 (2008) 1027 [arXiv:0805.2586]. * [25] M. Durant, et al., The spectrum of the recycled PSR J0437-4715 and its white dwarf companion, Astrophys. J. 746 (2012) 6 [arXiv:1111.2346]. * [26] B. Rangelov, et al., Hubble Space Telescope Detection of the Millisecond Pulsar J2124$-$3358 and its Far-ultraviolet Bow Shock Nebula, Astrophys. J. 835 (2017) 264 [arXiv:1701.00002]. * [27] G. G. Pavlov, et al., Old but still warm: Far-UV detection of PSR B0950+08, Astrophys. J. 850 (2017) 79 [arXiv:1710.06448]. * [28] V. Abramkin, Y. Shibanov, R. P. Mignani, and G. G. Pavlov, Hubble Space Telescope Observations of the Old Pulsar PSR J0108–1431, Astrophys. J. 911 (2021) 1 [arXiv:2103.00332]. * [29] V. Abramkin, G. G. Pavlov, Y. Shibanov, and O. Kargaltsev, Thermal and Nonthermal Emission in the Optical-UV Spectrum of PSR B0950+08*, Astrophys. J. 924 (2022) 128 [arXiv:2111.08801]. * [30] D. G. Yakovlev, K. P. Levenfish, and Y. A. Shibanov, Cooling neutron stars and superfluidity in their interiors, Phys. Usp. 42 (1999) 737–778 [astro-ph/9906456]. * [31] D. G. Yakovlev, A. D. Kaminker, O. Y. Gnedin, and P. Haensel, Neutrino emission from neutron stars, Phys. Rept. 354 (2001) 1 [astro-ph/0012122]. * [32] D. G. Yakovlev and C. J. Pethick, Neutron star cooling, Ann. Rev. Astron. Astrophys. 42 (2004) 169–210 [astro-ph/0402143]. * [33] D. Page, J. M. Lattimer, M. Prakash, and A. W. Steiner, Minimal cooling of neutron stars: A New paradigm, Astrophys. J. Suppl. 155 (2004) 623–650 [astro-ph/0403657]. * [34] M. E. Gusakov, A. D. Kaminker, D. G. Yakovlev, and O. Y. Gnedin, Enhanced cooling of neutron stars via Cooper-pairing neutrino emission, Astron. Astrophys. 423 (2004) 1063–1072 [astro-ph/0404002]. * [35] D. Page, J. M. Lattimer, M. Prakash, and A. W. Steiner, Neutrino Emission from Cooper Pairs and Minimal Cooling of Neutron Stars, Astrophys. J. 707 (2009) 1131–1140 [arXiv:0906.1621]. * [36] J. M. Lattimer, K. A. van Riper, M. Prakash, and M. Prakash, Rapid Cooling and the Structure of Neutron Stars, Astrophys. J. 425 (1994) 802. * [37] O. Y. Gnedin, D. G. Yakovlev, and A. Y. Potekhin, Thermal relaxation in young neutron stars, Mon. Not. Roy. Astron. Soc. 324 (2001) 725 [astro-ph/0012306]. * [38] K. S. Thorne, The relativistic equations of stellar structure and evolution., Astrophys. J. 212 (1977) 825–831. * [39] G. Glen and P. Sutherland, On the cooling of neutron stars, Astrophys. J. 239 (1980) 671–684. * [40] A. Y. Potekhin, G. Chabrier, and D. G. Yakovlev, Internal temperatures and cooling of neutron stars with accreted envelopes, Astron. Astrophys. 323 (1997) 415 [astro-ph/9706148]. * [41] F. Köpp, J. E. Horvath, D. Hadjimichef, C. A. Z. Vasconcellos, and P. O. Hess, Internal heating mechanisms in neutron stars, Int. J. Mod. Phys. D 32 (2023) 2350046 [arXiv:2208.07770]. * [42] G. Baym, C. Pethick, D. Pines, and M. Ruderman, Spin Up in Neutron Stars : The Future of the Vela Pulsar, Nature, 224 (1969) 872–874. * [43] J. A. Sauls, D. L. Stein, and J. W. Serene, Magnetic vortices in a rotating ${}^{3}P_{2}$ neutron superfluid, Phys. Rev. D 25 (1982) 967–975. * [44] M. A. Alpar, S. A. Langer, and J. A. Sauls, Rapid postglitch spin-up of the superfluid core in pulsars., Astrophys. J. 282 (1984) 533–541. * [45] E. M. Lifshitz and L. P. Pitaevskii, Statistical Physics: Theory of the Condensed State (Part 2), $\S~{}29$, (1980). * [46] G. A. Williams and R. E. Packard, Photographs of Quantized Vortex Lines in Rotating He II, Phys. Rev. Lett. 33 (1974) 280–283. * [47] D. V. Osborne, The Rotation of Liquid Helium II, Proceedings of the Physical Society. Section A 63 (1950) 909. * [48] A. J. Manninen, et al., First optical observations of superfluid ${}^{3}\mathrm{He}$, Phys. Rev. Lett. 69 (1992) 2392–2395. * [49] G. Baym, R. I. Epstein, and B. Link, Dynamics of vortices in neutron stars, Physica B Condensed Matter 178 (1992) 1–12. * [50] B. Link, R. I. Epstein, and G. Baym, Superfluid Vortex Creep and Rotational Dynamics of Neutron Stars, Astrophys. J. 403 (1993) 285. * [51] B. K. Link and R. I. Epstein, Mechanics and Energetics of Vortex Unpinning in Neutron Stars, Astrophys. J. 373 (1991) 592. * [52] T. Hamada and I. D. Johnston, A potential model representation of two-nucleon data below $315\,\mathrm{MeV}$, Nucl. Phys. 34 (1962) 382. * [53] R. Machleidt, The Meson Theory of Nuclear Forces and Nuclear Structure, in Advances in Nuclear Physics, Volume 19, J. W. Negele, et al., eds., ch. 2, p. 189. Springer US, Boston, 1989. * [54] R. B. Wiringa, V. G. J. Stoks, and R. Schiavilla, An Accurate nucleon-nucleon potential with charge independence breaking, Phys. Rev. C 51 (1995) 38–51 [nucl-th/9408016]. * [55] T. H. R. Skyrme, CVII. The nuclear surface, Phil. Mag. 1 (1956) 1043–1054. * [56] T. H. R. Skyrme, The effective nuclear potential, Nucl. Phys. 9 (1958) 615. * [57] E. Chabanat, P. Bonche, P. Haensel, J. Meyer, and R. Schaeffer, A Skyrme parametrization from subnuclear to neutron star densities, Nuclear Physics A 627 (1997) 710–746. * [58] M. Bender, P.-H. Heenen, and P.-G. Reinhard, Self-consistent mean-field models for nuclear structure, Rev. Mod. Phys. 75 (2003) 121–180. * [59] M. Dutra, et al., Skyrme interaction and nuclear matter constraints, Phys. Rev. C 85 (2012) 035201. * [60] E. Chabanat, P. Bonche, P. Haensel, J. Meyer, and R. Schaeffer, A Skyrme parametrization from subnuclear to neutron star densities. 2. Nuclei far from stablities, Nucl. Phys. A 635 (1998) 231–256. [Erratum: Nucl.Phys.A 643, 441–441 (1998)]. * [61] J. Bartel, P. Quentin, M. Brack, C. Guet, and H. B. Håkansson, Towards a better parametrisation of Skyrme-like effective forces: A critical study of the SkM force, Nuclear Physics A 386 (1982) 79–100. * [62] J. Dechargé and D. Gogny, Hartree-Fock-Bogolyubov calculations with the $D1$ effective interaction on spherical nuclei, Phys. Rev. C 21 (1980) 1568–1593. * [63] P. Avogadro, F. Barranco, R. A. Broglia, and E. Vigezzi, Quantum calculation of vortices in the inner crust of neutron stars, Phys. Rev. C 75 (2007) 012805 [nucl-th/0602028]. * [64] P. Avogadro, F. Barranco, R. A. Broglia, and E. Vigezzi, Microscopic quantal calculation of the vortex-nucleus interaction in the inner crust of neutron stars, Nucl. Phys. A 788 (2007) 130–135. * [65] P. Avogadro, F. Barranco, R. A. Broglia, and E. Vigezzi, Vortex-nucleus interaction in the inner crust of neutron stars, Nucl. Phys. A 811 (2008) 378–412 [arXiv:0804.1765]. * [66] P. Klausner, F. Barranco, P. M. Pizzochero, X. Roca-Maza, and E. Vigezzi, Microscopic calculation of the pinning energy of a vortex in the inner crust of a neutron star, arXiv:2303.18151 (2023). * [67] R. I. Epstein and G. Baym, Vortex Pinning in Neutron Stars, Astrophys. J. 328 (1988) 680. * [68] P. M. Pizzochero, L. Viverit, and R. A. Broglia, Vortex nucleus interaction and pinning forces in neutron stars, Phys. Rev. Lett. 79 (1997) 3347–3350 [astro-ph/9709060]. * [69] P. Donati and P. M. Pizzochero, Is there Nuclear Pinning of Vortices in Superfluid Pulsars?, Phys. Rev. Lett. 90 (2003) 211101. * [70] P. Donati and P. M. Pizzochero, Realistic energies for vortex pinning in intermediate-density neutron star matter, Phys. Lett. B 640 (2006) 74–81. * [71] P. Donati and P. M. Pizzochero, Fully consistent semi-classical treatment of vortex–nucleus interaction in rotating neutron stars, Nucl. Phys. A 742 (2004) 363–379. * [72] S. Seveso, P. M. Pizzochero, F. Grill, and B. Haskell, Mesoscopic pinning forces in neutron star crusts, Monthly Notices of the Royal Astronomical Society 455 (2015) 3952–3967. * [73] E. Garrido, P. Sarriguren, E. Moya de Guerra, and P. Schuck, Effective density dependent pairing forces in the T = 1 and T = 0 channels, Phys. Rev. C 60 (1999) 064312 [nucl-th/9909026]. * [74] B. Link, Dynamics of Quantum Vorticity in a Random Potential, Phys. Rev. Lett. 102 (2009) 131101 [arXiv:0807.1945]. * [75] A. Bulgac, M. M. Forbes, and R. Sharma, Strength of the Vortex-Pinning Interaction from Real-Time Dynamics, Phys. Rev. Lett. 110 (2013) 241102 [arXiv:1302.2172]. * [76] G. Wlazłowski, K. Sekizawa, P. Magierski, A. Bulgac, and M. M. Forbes, Vortex pinning and dynamics in the neutron star crust, Phys. Rev. Lett. 117 (2016) 232701 [arXiv:1606.04847]. * [77] B. Link and Y. Levin, Vortex Pinning in Neutron Stars, Slipstick Dynamics, and the Origin of Spin Glitches, Astrophys. J. 941 (2022) 148 [arXiv:2208.11494]. * [78] M. Antonelli and B. Haskell, Superfluid vortex-mediated mutual friction in non-homogeneous neutron star interiors, Mon. Not. Roy. Astron. Soc. 499 (2020) 3690–3705 [arXiv:2007.11720]. * [79] A. Akmal, V. R. Pandharipande, and D. G. Ravenhall, The Equation of state of nucleon matter and neutron star structure, Phys. Rev. C 58 (1998) 1804–1828 [nucl-th/9804027]. * [80] NSCool, http://www.astroscu.unam.mx/neutrones/NSCool/. * [81] J. W. Negele and D. Vautherin, Neutron star matter at subnuclear densities, Nucl. Phys. A 207 (1973) 298–320. * [82] P. Haensel, J. L. Zdunik, and J. Dobaczewski, Composition and equation of state of cold catalyzed matter below neutron drip, Astron. Astrophys. 222 (1989) 353–357. * [83] K. E. McGowan, et al., XMM - Newton observations of PSR B1706-44, Astrophys. J. 600 (2004) 343–350 [astro-ph/0309445]. * [84] O. Kargaltsev, M. Durant, Z. Misanovic, and G. Pavlov, Absorption Features in the X-ray Spectrum of an Ordinary Radio Pulsar, Science 337 (2012) 946–949 [arXiv:1208.5400]. * [85] K. E. McGowan, et al., Evidence for surface cooling emission in the xmm-newton spectrum of the x-ray pulsar psr b2334+61, Astrophys. J. 639 (2006) 377–381 [astro-ph/0508439]. * [86] P. Arumugasamy, O. Kargaltsev, B. Posselt, G. G. Pavlov, and J. Hare, Possible phase-dependent absorption feature in the x-ray spectrum of the middle-aged PSR J0659+1414, Astrophys. J. 869 (2018) 97 [arXiv:1810.11814]. * [87] K. Mori et al., A broadband X-ray study of the Geminga pulsar with NuSTAR and XMM-Newton, Astrophys. J. 793 (2014) 88 [arXiv:1408.1644]. * [88] C.-Y. Ng, R. W. Romani, W. F. Brisken, S. Chatterjee, and M. Kramer, The Origin and Motion of PSR J0538+2817 in S147, Astrophys. J. 654 (2006) 487–493 [astro-ph/0611068]. * [89] A. De Luca, P. A. Caraveo, S. Mereghetti, M. Negroni, and G. F. Bignami, On the polar caps of the 3 Musketeers, Astrophys. J. 623 (2005) 1051 [astro-ph/0412662]. * [90] N. Tetzlaff, J. G. Schmidt, M. M. Hohle, and R. Neuhaeuser, Neutron stars from young nearby associations the origin of RXJ1605.3+3249, Publ. Astron. Soc. Austral. 29 (2012) 98 [arXiv:1202.1388]. * [91] C. Malacaria et al., A joint NICER and XMM-Newton view of the ”Magnificent” thermally emitting X-ray Isolated Neutron Star RX J1605.3+3249, Astrophys. J. 880 (2019) 74 [arXiv:1906.02806]. * [92] S. Beloin, S. Han, A. W. Steiner, and D. Page, Constraining Superfluidity in Dense Matter from the Cooling of Isolated Neutron Stars, Phys. Rev. C 97 (2018) 015804 [arXiv:1612.04289]. * [93] K. Auchettl et al., X-ray analysis of the proper motion and pulsar wind nebula for PSR J1741-2054, Astrophys. J. 802 (2015) 68 [arXiv:1501.03225]. * [94] A. Kirichenko, et al., Deep optical observations of the $\gamma$-ray pulsar J0357+3205, Astron. Astrophys. 564 (2014) A81 [arXiv:1402.2246]. * [95] D. L. Kaplan and M. H. van Kerkwijk, A Coherent Timing Solution for the Nearby, Thermally Emitting Isolated Neutron Star RX J0420.0-5022, Astrophys. J. Lett. 740 (2011) L30 [arXiv:1109.2105]. * [96] C. Motch, A. M. Pires, F. Haberl, A. Schwope, and V. E. Zavlin, Proper motions of thermally emitting isolated neutron stars measured with Chandra, Astron. Astrophys. 497 (2009) 423 [arXiv:0901.1006]. * [97] A. D. Schwope, V. Hambaryan, F. Haberl, and C. Motch, The complex X-ray spectrum of the isolated neutron star RBS1223, Astrophys. Space Sci. 308 (2007) 619–623 [astro-ph/0609705]. * [98] N. Tetzlaff, T. Eisenbeiss, R. Neuhaeuser, and M. M. Hohle, The origin of RXJ1856.5-3754 and RXJ0720.4-3125 – updated using new parallax measurements, Mon. Not. Roy. Astron. Soc. 417 (2011) 617 [arXiv:1107.1673]. * [99] V. Hambaryan, et al., The compactness of the isolated neutron star RX J0720.4$-$3125, Astron. Astrophys. 601 (2017) A108 [arXiv:1702.07635]. * [100] N. Sartore, et al., Spectral monitoring of RX J1856.5-3754 with XMM-Newton. Analysis of EPIC-pn data, Astron. Astrophys. 541 (2012) A66 [arXiv:1202.2121]. * [101] D. L. Kaplan and M. H. van Kerkwijk, Constraining the Spin-down of the Nearby Isolated Neutron Star RX J2143.0+0654, Astrophys. J. Lett. 692 (2009) L62–L66 [arXiv:0901.4133]. * [102] A. Schwope et al., The isolated neutron star RBS1774 revisited. Revised XMM-Newton X-ray parameters and an optical counterpart from deep LBT-observations, Astron. Astrophys. 499 (2009) 267 [arXiv:0902.4110]. * [103] D. L. Kaplan and M. H. van Kerkwijk, Constraining the Spin-down of the Nearby Isolated Neutron Star RX J0806.4-4123, and Implications for the Population of Nearby Neutron Stars, Astrophys. J. 705 (2009) 798–808 [arXiv:0909.5218]. * [104] S. Guillot, et al., Hubble Space Telescope Nondetection of PSR J2144–3933: The Coldest Known Neutron Star, Astrophys. J. 874 (2019) 175 [arXiv:1901.07998]. * [105] R. N. Manchester, G. B. Hobbs, A. Teoh, and M. Hobbs, The Australia Telescope National Facility pulsar catalogue, Astron. J. 129 (2005) 1993 [astro-ph/0412641]. * [106] ATNF Pulsar Catalogue, https://www.atnf.csiro.au/research/pulsar/psrcat. * [107] J. A. Pons, J. A. Miralles, and U. Geppert, Magneto–thermal evolution of neutron stars, Astron. Astrophys. 496 (2009) 207–216 [arXiv:0812.3018]. * [108] S. Dall’Osso, J. Granot, and T. Piran, Magnetic field decay in neutron stars: from soft gamma repeaters to ’weak-field magnetars’, Mon. Not. Roy. Astron. Soc. 422 (2012) 2878–2903 [arXiv:1110.2498]. * [109] D. Viganò, et al., Unifying the observational diversity of isolated neutron stars via magneto-thermal evolution models, Mon. Not. Roy. Astron. Soc. 434 (2013) 123 [arXiv:1306.2156]. * [110] A. Y. Potekhin, D. A. Zyuzin, D. G. Yakovlev, M. V. Beznogov, and Y. A. Shibanov, Thermal luminosities of cooling neutron stars, Mon. Not. Roy. Astron. Soc. 496 (2020) 5052–5071 [arXiv:2006.15004]. * [111] A. Reisenegger, Deviations from chemical equilibrium due to spindown as an internal heat source in neutron stars, Astrophys. J. 442 (1995) 749 [astro-ph/9410035]. * [112] P. Haensel, Non-equilibrium neutrino emissivities and opacities of neutron star matter, Astron. Astrophys. 262 (1992) 131–137. * [113] E. Gourgoulhon and P. Haensel, Upper bounds on the neutrino burst from collapse of a neutron star into a black hole, Astron. Astrophys. 271 (1993) 187. * [114] R. Fernandez and A. Reisenegger, Rotochemical heating in millisecond pulsars. Formalism and non-superfluid case, Astrophys. J. 625 (2005) 291–306 [astro-ph/0502116]. * [115] L. Villain and P. Haensel, Non-equilibrium beta processes in superfluid neutron star cores, Astron. Astrophys. 444 (2005) 539 [astro-ph/0504572]. * [116] C. Petrovich and A. Reisenegger, Rotochemical heating in millisecond pulsars: modified Urca reactions with uniform Cooper pairing gaps, Astron. Astrophys. 521 (2010) A77 [arXiv:0912.2564]. * [117] C.-M. Pi, X.-P. Zheng, and S.-H. Yang, Neutrino Emissivity of Non-equilibrium beta processes With Nucleon Superfluidity, Phys. Rev. C 81 (2010) 045802 [arXiv:0912.2884]. * [118] N. González-Jiménez, C. Petrovich, and A. Reisenegger, Rotochemical heating of millisecond and classical pulsars with anisotropic and density-dependent superfluid gap models, Mon. Not. Roy. Astron. Soc. 447 (2015) 2073 [arXiv:1411.6500]. * [119] K. Yanagi, N. Nagata, and K. Hamaguchi, Cooling Theory Faced with Old Warm Neutron Stars: Role of Non-Equilibrium Processes with Proton and Neutron Gaps, Mon. Not. Roy. Astron. Soc. 492 (2020) 5508–5523 [arXiv:1904.04667]. * [120] D. Toyouchi, K. Hotokezaka, and M. Takada, Hunting wide-area optical surveys for high proper motion isolated neutron stars, Mon. Not. Roy. Astron. Soc. 510 (2021) 611–619 [arXiv:2106.04846]. * [121] A. Schwenk, B. Friman, and G. E. Brown, Renormalization group approach to neutron matter: Quasiparticle interactions, superfluid gaps and the equation of state, Nucl. Phys. A 713 (2003) 191–216 [nucl-th/0207004]. * [122] J. M. C. Chen, J. W. Clark, R. D. Davé, and V. V. Khodel, Pairing gaps in nucleonic superfluids, Nucl. Phys. A 555 (1993) 59–89. * [123] C. Kouvaris, WIMP Annihilation and Cooling of Neutron Stars, Phys. Rev. D 77 (2008) 023006 [arXiv:0708.2362]. * [124] G. Bertone and M. Fairbairn, Compact Stars as Dark Matter Probes, Phys. Rev. D 77 (2008) 043515 [arXiv:0709.1485]. * [125] C. Kouvaris and P. Tinyakov, Can Neutron stars constrain Dark Matter?, Phys. Rev. D 82 (2010) 063531 [arXiv:1004.0586]. * [126] A. de Lavallaz and M. Fairbairn, Neutron Stars as Dark Matter Probes, Phys. Rev. D 81 (2010) 123521 [arXiv:1004.0629]. * [127] J. Bramante, A. Delgado, and A. Martin, Multiscatter stellar capture of dark matter, Phys. Rev. D 96 (2017) 063002 [arXiv:1703.04043]. * [128] M. Baryakhtar, J. Bramante, S. W. Li, T. Linden, and N. Raj, Dark Kinetic Heating of Neutron Stars and An Infrared Window On WIMPs, SIMPs, and Pure Higgsinos, Phys. Rev. Lett. 119 (2017) 131801 [arXiv:1704.01577]. * [129] N. Raj, P. Tanedo, and H.-B. Yu, Neutron stars at the dark matter direct detection frontier, Phys. Rev. D 97 (2018) 043006 [arXiv:1707.09442]. * [130] C.-S. Chen and Y.-H. Lin, Reheating neutron stars with the annihilation of self-interacting dark matter, JHEP 08 (2018) 069 [arXiv:1804.03409]. * [131] N. F. Bell, G. Busoni, and S. Robles, Heating up Neutron Stars with Inelastic Dark Matter, JCAP 09 (2018) 018 [arXiv:1807.02840]. * [132] R. Garani, Y. Genolini, and T. Hambye, New Analysis of Neutron Star Constraints on Asymmetric Dark Matter, JCAP 05 (2019) 035 [arXiv:1812.08773]. * [133] D. A. Camargo, F. S. Queiroz, and R. Sturani, Detecting Dark Matter with Neutron Star Spectroscopy, JCAP 09 (2019) 051 [arXiv:1901.05474]. * [134] N. F. Bell, G. Busoni, and S. Robles, Capture of Leptophilic Dark Matter in Neutron Stars, JCAP 06 (2019) 054 [arXiv:1904.09803]. * [135] K. Hamaguchi, N. Nagata, and K. Yanagi, Dark Matter Heating vs. Rotochemical Heating in Old Neutron Stars, Phys. Lett. B 795 (2019) 484–489 [arXiv:1905.02991]. * [136] R. Garani and J. Heeck, Dark matter interactions with muons in neutron stars, Phys. Rev. D 100 (2019) 035039 [arXiv:1906.10145]. * [137] J. F. Acevedo, J. Bramante, R. K. Leane, and N. Raj, Warming Nuclear Pasta with Dark Matter: Kinetic and Annihilation Heating of Neutron Star Crusts, JCAP 03 (2020) 038 [arXiv:1911.06334]. * [138] A. Joglekar, N. Raj, P. Tanedo, and H.-B. Yu, Relativistic capture of dark matter by electrons in neutron stars, Phys. Lett. B (2020) 135767 [arXiv:1911.13293]. * [139] W.-Y. Keung, D. Marfatia, and P.-Y. Tseng, Heating neutron stars with GeV dark matter, JHEP 07 (2020) 181 [arXiv:2001.09140]. * [140] K. Yanagi, Thermal Evolution of Neutron Stars as a Probe of Physics beyond the Standard Model, arXiv:2003.08199 (2020). * [141] A. Joglekar, N. Raj, P. Tanedo, and H.-B. Yu, Dark kinetic heating of neutron stars from contact interactions with relativistic targets, Phys. Rev. D 102 (2020) 123002 [arXiv:2004.09539]. * [142] N. F. Bell, G. Busoni, S. Robles, and M. Virgato, Improved Treatment of Dark Matter Capture in Neutron Stars, JCAP 09 (2020) 028 [arXiv:2004.14888]. * [143] N. F. Bell, G. Busoni, S. Robles, and M. Virgato, Improved Treatment of Dark Matter Capture in Neutron Stars II: Leptonic Targets, JCAP 03 (2021) 086 [arXiv:2010.13257]. * [144] F. Anzuini, et al., Improved treatment of dark matter capture in neutron stars III: nucleon and exotic targets, JCAP 11 (2021) 056 [arXiv:2108.02525]. * [145] Y.-P. Zeng, X. Xiao, and W. Wang, Constraints on Pseudo-Nambu-Goldstone dark matter from direct detection experiment and neutron star reheating temperature, Phys. Lett. B 824 (2022) 136822 [arXiv:2108.11381]. * [146] J. Bramante, B. J. Kavanagh, and N. Raj, Scattering Searches for Dark Matter in Subhalos: Neutron Stars, Cosmic Rays, and Old Rocks, Phys. Rev. Lett. 128 (2022) 231801 [arXiv:2109.04582]. * [147] P. Tinyakov, M. Pshirkov, and S. Popov, Astroparticle Physics with Compact Objects, Universe 7 (2021) 401 [arXiv:2110.12298]. * [148] T. N. Maity and F. S. Queiroz, Detecting bosonic dark matter with neutron stars, Phys. Rev. D 104 (2021) 083019 [arXiv:2104.02700]. * [149] M. Fujiwara, K. Hamaguchi, N. Nagata, and J. Zheng, Capture of electroweak multiplet dark matter in neutron stars, Phys. Rev. D 106 (2022) 055031 [arXiv:2204.02238]. * [150] K. Hamaguchi, N. Nagata, and M. E. Ramirez-Quezada, Neutron star heating in dark matter models for the muon g $-$ 2 discrepancy, JHEP 10 (2022) 088 [arXiv:2204.02413]. * [151] S. Chatterjee, et al., Faint light of old neutron stars and detectability at the James Webb Space Telescope, Phys. Rev. D 108 (2023) L021301 [arXiv:2205.05048]. * [152] J. Coffey, D. McKeen, D. E. Morrissey, and N. Raj, Neutron star observations of pseudoscalar-mediated dark matter, Phys. Rev. D 106 (2022) 115019 [arXiv:2207.02221]. * [153] J. T. Acuña, P. Stengel, and P. Ullio, Neutron Star Heating in Dark Matter Models for Muon g-2 with Scalar Lepton Partners up to the TeV Scale, arXiv:2209.12552 (2022). * [154] G. Alvarez, A. Joglekar, M. Phoroutan-Mehr, and H.-B. Yu, Heating neutron stars with inelastic dark matter and relativistic targets, Phys. Rev. D 107 (2023) 103024 [arXiv:2301.08767]. * [155] J. Bramante and N. Raj, Dark matter in compact stars, arXiv:2307.14435 (2023). * [156] M. Fujiwara, K. Hamaguchi, N. Nagata, and E. M. Ramirez-Quezada, to be published,. * [157] W. Becker, et al., Revealing the x-ray emission processes of old rotation-powered pulsars: XMM-Newton observations of PSR B0950+08, PSR B0823+26 and PSR J2043+2740, Astrophys. J. 615 (2004) 908–920 [astro-ph/0405180]. * [158] V. E. Zavlin and G. G. Pavlov, X-ray emission from the old pulsar B0950+08, Astrophys. J. 616 (2004) 452–462 [astro-ph/0405212]. * [159] B. Posselt, et al., XMM-Newton observation of the very old pulsar J0108-1431, Astrophys. J. 761 (2012) 117 [arXiv:1210.7179]. * [160] J. P. W. Verbiest, D. R. Lorimer, and M. A. McLaughlin, Lutz-Kelker bias in pulsar parallax measurements, Mon. Not. Roy. Astron. Soc. 405 (2010) 564–572 [arXiv:1002.1213]. * [161] A. Tiengo, et al., The first deep X-ray and optical observations of the closest isolated radio pulsar, AIP Conf. Proc. 1357 (2011) 269–272 [arXiv:1103.4538]. * [162] Z. Misanovic, G. Pavlov, and G. Garmire, Chandra observations of the pulsar PSR B1929+10 and its environment, Astrophys. J. 685 (2008) 1129–1142 [arXiv:0711.4171]. * [163] W. W. Tian, D. A. Leahy, M. Haverkorn, and B. Jiang, Discovery of the counterpart of TeV Gamma-ray source HESS J1731-347: a new SNR G353.6-0.7 with radio and X-ray images, Astrophys. J. Lett. 679 (2008) L85 [arXiv:0801.3254]. * [164] D. Klochkov, et al., The neutron star in HESSJ1731-347: Central compact objects as laboratories to study the equation of state of superdense matter, Astron. Astrophys. 573 (2015) A53 [arXiv:1410.1055].
1.2 0.95 # We both think you did wrong - How agreement shapes and is shaped by indirect reciprocity - DRAFT Marcus Krellner1,⋆ and The Anh Han1 1 School Computing, Engineering and Digital Technologies, Teesside University ⋆ Corresponding authors<EMAIL_ADDRESS> ## Abstract Humans judge each other’s actions, which at least partly functions to detect and deter cheating and to enable helpfulness in an indirect reciprocity fashion. However, most forms of judging do not only concern the action itself, but also the moral status of the receiving individual (to deter cheating it must be morally acceptable to withhold help from cheaters). This is a problem, when not everybody agrees who is good and who is bad. Although it has been widely acknowledged that disagreement may exist and that it can be detrimental for indirect reciprocity, the details of this crucial feature of moral judgments have never been studied in depth. We show, that even when everybody assesses individually (aka privately), some moral judgement systems (aka norms) can lead to high levels of agreement. We give a detailed account of the mechanisms which cause it and we show how to predict agreement analytically without requiring agent-based simulations, and for any observation rate. Finally, we show that agreement may increase or decrease reputations and therefore how much helpfulness (aka cooperation) occurs. Keywords:cooperation, evolutionary game theory, indirect reciprocity, donation game, private assessments, analytical predictions ## 1 Introduction From simplest replicators, evolution has worked its wonders and today we marvel at the success, diversity and complexity of life. Many wonders of life were enabled by cooperation. Cells work together to form multicellular organisms, former bacteria serve as the powerplants of cells in exchange for eternal protection from an outside world, and individual organism work together in communities of thousands and millions to become the most widespread forms of life there are, ants/wasps and, in similar but different way, humans. Indirect Reciprocity (IR) is one of the few other mechanisms to facilitate cooperation among self-interested agents (Rand and Nowak,, 2013, Nowak and Sigmund,, 2005, perc2017statistical). You help me, without receiving any direct returns (that is, in that moment you behave altruistically). But you will be rewarded by others in the future. Because others have observed your behavior and will like you for it. I.e. they judge your action and form an opinion about you. And because they like you, they decide to help you. Indirect reciprocity allows cooperation (aka continued help amongst helpers) in large groups of unrelated individuals, even if they do not interact enough to establish direct reciprocity (Schmid et al.,, 2022). Hence, indirect reciprocity has been the focus of many influential studies of evolutionary game theory (Nowak and Sigmund, 1998a , Ohtsuki and Iwasa, (2004), see also Sigmund, (2016)) and is even considered as an important foundation of human morality (Nowak and Sigmund,, 2005). Traditionally, studies simplified the mechanism of IR, assuming that judgments are unanimous (also called public assessments). More realistically, everybody would make their own judgements and therefore have their own opinions (aka private assessments). When the public assessment assumption is removed, dynamics and outcomes of IR models change (Brandt and Sigmund,, 2004, Okada, 2020b, ), often to the worse for the evolution of cooperation (Uchida and Sasaki,, 2013, Hilbe et al.,, 2018). The main reason is the emergence and spreading of disagreements about someone’s reputation, which had not been possible with public assessments. Disagreements can significantly disturb IR strategies that are stable under the public assessment regime, because they hinder a universal principle: if you withhold help towards somebody bad, this should not be judged as bad 111at least if done by somebody good, which can be seen as in general judging it neutral or even good (Okada, 2020b, , Krellner and Han,, 2022). (Panchanathan and Boyd,, 2003, Ohtsuki and Iwasa,, 2006). It is problematic to apply this rule, if potential helpers and observers may disagree. You withhold help to me, since you do not like me. But some observers like me, so they will disapprove of your action. So, if there was disagreement about my reputation, your reputation can get tainted. In addition, a disagreement can cause further disagreements. If two observers disagree about me, they may also judge you differently for withholding help. That is, they now disagree about another person’s reputation. Even a single disagreement can cascade through the population, even if no further errors occur, so that all opinions become bad or essentially random (Hilbe et al.,, 2018). Private assessment and disagreements are not only detrimental for IR strategies, but also lead to new challenges for IR research. Whereas for public assessment, there had long been exhaustive investigations of hundreds of strategies (Ohtsuki and Iwasa,, 2004)[santos] using analytical models, something similar for private assessment was only achieved in recent years (Okada, 2020b, , Perret et al.,, 2021). And these models still require restricting assumptions. Namely that, in an infinite population, only a finite number of players observe. That is, the probability to observe $\psi$ and hence also the fraction of opinions changed by a single interaction are required to be negligible ($\psi\rightarrow 0$). Analytical models and most models of IR in general (Okada, 2020a, ) consider opinions as binary. That is, an individual is judged as either good (1) or bad (0). To understand the state of opinions in a population, it is important to know the average reputation $r$: the probability of a random opinion being good, as well as the average probability of an agreement $a$: the probability that two randomly selected players have the same opinion about another random player. The models mentioned above only considered $r$ (hence we shall call them R-models for short). They assume that, when the observation probability is very small ($\psi\to 0$), the probability of agreement $a$ only depends on $r$ and is given by $\widehat{a}=r^{2}+(1-r)^{2}$ (i.e. probability that two random opinions are either both good or both bad) (Okada et al.,, 2018). Figure 1: a) Two dimensions of global opinion state. Graph shows agreement-reputation space, with the solid lines $a=1$ and $a=\widehat{a}$ indicating outer borders of possible agreement values in grey. (Subsequently, values above $a=\widehat{a}+0.05$ will be considered as significant additional agreement, since values very close to $\widehat{a}$ could conceivably be the result of noise or finite properties of the simulations; or be just too insignificant to make a difference). Case 1 has minimal agreement for the specific global average reputation $r$. Case 3 has maximal agreement for the same $r$ and case 2 an intermediate one for the same $r$ still. b) Model of simplified image matrices. We simplify the image matrices for the three cases by removing all information about whose opinions it is and what specific player it is about. We only consider whether the opinion is about a player of group 1 or group 2 . The result can be imagined as two containers (left and right, separated by the double line) holding good opinions (dark blue areas). We may call the containers or groups ”lucky” on the left and ”unlucky” on the right. In case 1, both containers are filled to the same height, representing that the reputations of lucky and unlucky group, $r_{L}$ and $r_{U}$, are equal to the global average. In case 2, some good opinions from the unlucky container were ’poured’ into the lucky container, keeping $r$ unchanged since we do not add or remove any good opinions. In case 3, all remaining of these good opinions were poured as well, filling the lucky container exactly to the brim. It is apparent that in case 3, all players of the lucky group have only good opinions about them, therefore agreement about a player of this group is always $100\%$. The same applies to the unlucky group, which has only bad opinions about them. c) Size of the groups and precise state of agreement d. The complete transition from case 1 to case 3 can only be achieved under a specific condition. The size of the groups (the width of the containers) must be $r$ and $1-r$, respectively. With this, the areas $A_{L}$ and $A_{U}$ are of equal size. That means, for minimal agreement as in case 1, there are exactly as many good opinions about the unlucky group as there are bad opinions about the lucky one. Therefore, they can be entirely exchanged. The central feature of the analytical model introduced in this paper, is the parameter $d$, which describes the percentage of good opinions from area $A_{U}$ exchanged with bad opinions from area $A_{L}$. This parameter is independent of $r$ and can obtain any value in the interval $[0,1]$ (note that both areas disappear for the extreme cases $r=0$ and $r=1$, which is why $d$ is arbitrary in these cases). For details see equations 2 and 4. In theory, $a$ can be much higher. The population could for example consist of two groups, one with perfect reputation, the other with an abysmal one, see case 3 in Figure 1. The average reputation is still $r=0.6$. But all opinions about a player of the unlucky group are bad. Which means that the opinions of any two players about such a player will always be the same. This is similarly true for the lucky group in which all opinions are good. In case 3, average agreement is therefore 100%. There are of course also intermediate states, such as case 2. The left side of the figure shows the entire agreement- reputation space and how agreement can range anywhere between $\widehat{a}$ and 1. Note that $0.5\leq\widehat{a}\leq 1$ (since $0\leq r\leq 1$). There are obvious mechanisms that cause disagreement and push state of the population towards $\widehat{a}$, as discussed above. However, there could be other mechanisms that increase agreement. For example, if you donated your help, observers may not care about their current opinion about the recipient and like you either way (which is the case for image scoring (Nowak and Sigmund, 1998b, ) or standing (Sudgen,, 1986, Leimar and Hammerstein,, 2001)). Hence, aside from a few who miss-perceive your action due to noise, all observers will have the same opinion about you. The existence of additional agreement, i.e. states where agreement is significantly above $\widehat{a}$, was already implicitly shown by the work of Fujimoto and Ohtsuki, (2022), who investigated the exact distributions of reputations for four strategies. They found that even infinite populations sometimes have two or even up to an infinite number of distinct reputation states, that a player could be in. It was not stated by the authors but any split of players into groups with different reputation states entails that the opinions of this population have additional agreement (see a proof sketch in Supplementary Information (SI)). Nevertheless, their seminal work included an analytical model to predict distributions of strategies exactly, which will likely be foundational for future analytical models of IR. Yet, two problems remain. The number of strategies it can be applied to is small and the model is limited to the condition of full observation (i.e., $\psi=1$). In conclusion, in the literature so far, there have been only two types of analytical models of private assessment, and they are all limited to either $\psi=1$ or $\psi\to 0$. Both cases are extreme ends of possible natural conditions. Full observation seems impossible due to physical constraints. Even if all interactions between people were public, there are still limitations to the attention observers can pay. Nominal observation rates, on the other hand, make IR less effective in the short-term, and may cause it to be replaced by direct reciprocity (Schmid et al., 2021a, ). As Fujimoto and Ohtsuki, (2022) state themselves, a model for intermediate observation rates is necessary, especially to study IR under private assessment in an exhaustive fashion, like the defining work of Ohtsuki and Iwasa, (2004). Overall, our investigation fills two major gaps in the literature. We provide the first the exhaustive investigation of agreement in reputation dynamics under private assessment. We test whether additional agreement is common for high or intermediate observation rates, but also if the assumption of the R-models always holds for $\psi\to 0$. We test this with simulations of a vast number of strategies and conditions. Second, we define the first A-R-model, that moves beyond the one-dimensional approach of the R-models, to capture the two-dimensional space of agreement and reputation. It is capable of predicting long-term behavior of said simulations. Both contributions advance the understanding of indirect reciprocity under private assessment and lay important groundwork for exhaustive evolutionary investigations. ## 2 Methods ### 2.1 Simulation We examine if the average agreement $a$ is greater than assumed in the R-model, $\widehat{a}=r^{2}+(1-r)^{2}$. For that, we run simulations with different strategies and conditions. To limit complexity, following the example of Fujimoto and Ohtsuki, (2022), we consider homogeneous populations, i.e. within a simulation all agents follow the same strategy. We have chosen a framework to model many different strategies. The first characteristic of a strategy is its assessment rules, given by vector $\alpha$. It defines how an observer judges the action, i.e. cooperation $C$ or defection $D$, depending on its opinion about the recipient, i.e. good $G$ or bad $B$. An assessment can be one of three forms: the action is approved of (+1), disapproved of (-1) or ignored (0). If the player approves of an action, their opinion of the donor will be good afterwards (which is the same as to say that it has increased within the limit of the two opinion values, 0 and 1, aka bad and good). If the action is disapproved of, the opinion will be bad. And if the action is ignored, the opinion does not change. The second characteristic of the strategy is given by its action rules $\beta$. They define what players will do, when they meet a recipient they like or dislike. These actions are cooperate (1) or defect (0). In the following, if we refer to $\alpha$ and $\beta$ as a whole, in the order given below. But we will sometimes refer to specific values or pairs of values, such $\alpha_{CG}$: cooperation towards good. With this framework we can model 324 different strategies. For example the commonly studied strategy “staying” (Okada et al.,, 2017, Sasaki et al.,, 2017) is described as follows: $\begin{matrix}staying:&\begin{matrix}\alpha=\\{\alpha_{CG},\alpha_{DG},\alpha_{CB},\alpha_{DB}\\}=(1,-1,0,0)\\\ \beta=\\{\beta_{G},\beta_{B}\\}=(1,0)\end{matrix}\par\end{matrix}$ (1) Of all the possible strategies of our framework, we study 171 strategies with unique behavior. Each of the possible strategies has a mirror image, which behaves equivalently in regard to cooperation. If we were to label opinion states as blue and red (instead of good and bad), an mirror image would cooperate in the exact same way to the same partners, but would think of them as blue instead of red or vice versa. We obtain such a mirror image of a strategy by exchanging the symbols +1 and -1 in the assessment rules and flipping them for good and bad recipients ($\alpha_{m}=(x_{1},x_{2},x_{3},x_{4})$ to $\alpha=(-x_{3},-x_{4},-x_{1},-x_{2})$) as well as flipping values of the action rules ($\beta=(xy)$ to $\beta_{m}=(yx)$). Note, that some strategies are their own mirror strategy. This focus on unique strategies conveniently leaves us with only three possible action rules: (1,1): unconditional cooperators or AllC, (1,0): conditional cooperators or Coco222sometimes, somewhat unfortunately, called discriminators and (0,0): unconditional defectors or AllD (see SI for examples of mirror images). For our simulations, we consider a well-mixed population of size $N=100$. Every agent can have either a good (1) or a bad (0) opinion about each other agent. Hence, the state of the population can be described by the $N\times N$ image matrix $M(t)$ of the population at time $t$ (Uchida,, 2010). Initially all entries are filled by a fair coin toss. Other studies have reported, that other initial conditions almost never change the outcomes (Hilbe et al.,, 2018). Time is discrete. In total, each time step consists of three parts. First, a donor $do$ and a recipient $re$ are drawn at random from the population. The donor then decides whether to cooperate. Note, these two steps are usually referred to as the donation game. To apply IR, donors base their decision on their action rule and their current opinion about the recipient. In the third part, opinions about the donor are updated due to observations. These observations and updating can be broken down further. First, for each player (except the donor and recipient) it is decided if they will observe the interaction. For each player, who observes, it is determined if it observes accurately or if the observation is altered by a perception error, i.e. with a probability $\epsilon$ they will perceive the opposite of what the donor is actually doing. Next, individually perceived action and individual opinion of the observer about the recipient are combined for the private assessment of the donor. The opinion is updated if assessment is 1 or -1, but left as is if assessment is 0. Finally, each observer (whether they made an assessment or not) may change their opinion to a random value if they commit a cognitive error with probability $\mu$. We study the behavior of the simulation in the long term and extract precise values for average reputation $r$ and average agreement $a$. In order to ensure accurate and reliable results, the simulations are run until we are reach an objective level of confidence about the measured average values (for details see SI.) Note, in our version of the donation game, the donor and recipient do not update their opinion. I.e. players ignore the information of interactions they are themselves involved. This means we study pure indirect reciprocity, without any direct reciprocity (Schmid et al., 2021a, ). Because of that, the diagonal of the image matrix is never updated nor used. We therefore exclude it from all computations of averages. For a detailed motivation of this design see SI. ### 2.2 Predictive A-R-Model How could we define a new model which can represent both $r$ and $a$, but is otherwise as simple as possible? It has to be able to represent a continuous transition of $a$, from the baseline of the former prediction $\widehat{a}$ to its maximum 1 (see transition from 1 to 3 in figure 1). The transition must be possible for any $0\leq r\leq 1$ while keeping $r$ constant. As mentioned, we could first nominally divide the population into two groups, lucky and unlucky. To increase agreement a minimal step, we could take away a random good opinion about an unlucky player and exchange with a bad opinion about a lucky player. I.e. we decrease the average opinion of unlucky players and increase the average opinion about lucky players. We could continue to do so until lucky players have the average reputation $r_{L}=1$ and unlucky players have the average reputation $r_{U}=0$, hence both agreement about lucky players $a_{L}$ and agreement about unlucky players $a_{U}$ is 1. We can do this only if the number of good opinions about unlucky players $n_{GU}$ is equal to the number of bad opinions about the lucky players $n_{BL}$. This is the case, if the size of the lucky group (or the percentage of lucky players) is equal to $r$, hence the size of the unlucky group is $1-r$. With this assumption, we may define a single value $d$, which is both the percentage of good opinions about unlucky players stolen and the percentage of bad opinions about lucky players replaced (see figure 1c). $r_{L}=r+d(1-r)\quad\&\quad r_{U}=r-dr,\quad 0\leq d\leq 1.$ (2) With this model we define the two dimensional space with just two variables. We assume a specific distribution of reputations, namely that there are exactly two groups (in general there could be up to $N$ groups in a population of $N$ players). The goal of our model is not to represent the exact distribution. Just to represent $a$ as well as $r$, for which the two-group scenario provides a minimal model. It allows us to to model all agreement states in all reputation states with a single parameter $d$. Both $r$ and $d$ can range between 0 and 1 (whereas the minimum of $a$ would depend on current $r$). Note that $d$ is undefined for $r=0$ and $r=1$ (see equation 3), but it does not need to be. In the case of $r=0$, $d$ vanishes from $r_{U}$ and the size of the lucky group is zero, hence $n_{GU}$ is zero for any $d$. The opposite is true for $r=1$ respectively. Besides these extreme cases, $d$ is given by $d=\frac{2^{1/2}(r(1-r)(-2r^{2}+2r+a-1))^{1/2}}{2r(1-r)}=\left(1-\frac{1-a}{2r(1-r)}\right)^{1/2},\ 0<r<1.$ (3) This close form of $d$ is derived by solving the equation (4) below for $d$, by replacing $r_{L}$ and $r_{U}$ with the equations in (2). The global average agreement $a$ is given by the sum of agreements about members in each group discounted by the group size, i.e. $a=ra_{L}+(1-r)a_{U}$. The agreement for a group is derived by its average reputation, that is $a_{L}=r_{L}^{2}+(1-r_{L})^{2}$ and $a_{U}=r_{U}^{2}+(1-r_{U})^{2}$. Thus, $a=r(r_{L}^{2}+(1-r_{L})^{2})+(1-r)(r_{U}^{2}+(1-r_{U})^{2}).$ (4) We will now use $r$, $r_{L}$ and $r_{U}$ to predict the probabilities of events of our simulation, assuming an infinite population. To model a time step of the simulation, we need to distinguish between onetime events and repeated events. There are three events that happen only once per time step: picking a donor, picking a recipient and the decision of the donor. For these we need to know three kinds of probabilities for later computation. First, a pair of probabilities $q$ about the donor’s group affiliation: $q_{L}=r$ (lucky group) and $q_{U}=1-r$ (unlucky group). Second, four probabilities $p$ representing the combinations of donor’s choice and recipient’s group. For example, $p_{CL}$, the probability that the donor cooperates and the recipient is in the lucky group (where $\beta$ is the set of the donor’s action rules and $1-r_{L}$ is the chance the donor has a bad opinion of the recipient from the lucky group), is given as follows $p_{CL}=r\Big{(}r_{L}\beta_{G}+(1-r_{L})\beta_{B}\Big{)}$ (5) or in general $\begin{matrix}p_{jk}=q_{k}\Big{(}r_{k}c_{jG}+(1-r_{k})c_{jB}\Big{)},\\\ \text{where }j\in\\{C,D\\},\ k\in\\{L,U\\}\text{ and }c_{Cx}=\beta_{x},c_{Dx}=1-\beta_{x}.\end{matrix}$ (6) Third, a 2-by-4 matrix $G$ showing how likely the opinion of the donor is good after assessment. It depends on the previous reputation of the donor’s group $r_{i}$ and the four assessment rules $\alpha$ (see equation 1). For example, since staying judges cooperation with bad recipient as neutral, $\alpha_{CB}=0$, the probability to be considered good is equal to the previous reputation, which is $r_{L}$ if the donor belongs to the lucky group $G_{L,CB}=r_{L},$ (7) or in general $\begin{matrix}G_{i,mn}=\begin{matrix}1,&\text{if}&\alpha_{mn}=1\\\ r_{i},&\text{if}&\alpha_{mn}=0\\\ 0,&\text{if}&\alpha_{mn}=-1\end{matrix},\\\ \\\ \text{where }i\in\\{L,U\\},m\in\\{C,D\\},n\in\\{G,B\\}.\end{matrix}$ (8) These are the probabilities that depend onetime events. There are also probabilities of repeated events, namely observations, since there can be many observers. Each observation includes the action the observer perceives and what opinion it had about the recipient. Hence there can be four kinds of observations, and we compute their probability $O$ for each of the four combinations of onetime events $p$ described in equation 6. For example, in the event that the donor cooperated with a lucky recipient, we can compute the probability $O_{CL,CG}$, that the observer has a good opinion about this lucky recipient and that the observer perceives the cooperation without a perception error $\epsilon$. $O_{CL,CG}=r_{L}(1-\epsilon),$ (9) or in general $\begin{matrix}O_{jk,mn}=r_{i}^{*}e,\\\ \\\ \begin{matrix}r_{x}^{*}=r_{x}&\text{if}&n=G\\\ r_{x}^{*}=1-r_{x}&\text{if}&n=B\\\ e=\epsilon&\text{if}&m\neq j\\\ e=1-\epsilon&\text{if}&m=j\\\ \end{matrix}\\\ \\\ \text{where }j,m\in\\{C,D\\},k\in\\{L,U\\},n\in\\{G,B\\}.\end{matrix}$ (10) We then take into account observation rates $\psi$ and cognitive errors $\mu$, which are also repeated events. Each observation might not take place, leaving opinions unchanged. When it takes place, the resulting opinion may be altered by a cognitive error and its value would be reversed. We therefore adjust $G$ to $G^{*}$. $G^{*}_{i,mn}=(1-\psi)r_{i}+\psi(G_{i,mn}(1-\mu)+(1-G_{i,mn})\mu).$ (11) With these probabilities we can now compute the probability of the eight possible interactions, that is, the combination of all three onetime events, as follows $\Pi_{ijk}=q_{i}p_{jk},$ (12) and the expected reputation of the donor for each of these combinations $\Gamma_{ijk}=\sum_{m,n}G^{*}_{i,mn}O_{jk,mn}.$ (13) Finally, we can now compute the expected change of reputation $\Delta_{r}=\Big{(}\sum_{i,j,k}\Gamma_{ijk}\Pi_{ijk}\Big{)}-r$ (14) and the expected change of agreement $\Delta_{a}=\Big{(}\sum_{i,j,k}((\Gamma_{ijk})^{2}+(1-\Gamma_{ijk})^{2})\Pi_{ijk}\Big{)}-a.$ (15) Note, as it was when computing agreement in the simulation, the instances of interaction have to be computed separately. Each of the eight possible interactions predicts a specific agreement about a donor’s image. We compute each instance and only then compute the expected value of $\Delta_{a}$. Note also, that we do not assume that the donor can be attributed to the lucky or unlucky group after assessment. These groups are simplifications of the real image matrix that we use to describe its state with only two parameters. Information about the exact expected reputations of the donor and hence the agreement about it for the eight possible interactions is lost. Our model does not give exact predictions. To validate our analytical model, we create a numerical algorithm to find the stable equilibria points in the $r\times a$ space (see SI for a detailed description). We use it to find the equilibria for all cases which we simulated and compare the results. For additional comparison we also use an analytical approach that models agreement always as $\widehat{a}$, aka a R-model (Perret et al.,, 2021), and compare the fit of both predictive models. ## 3 Results (a) Simulations. Agreement and reputation averaged across runs for homogeneous populations with 100 players. Not shown are cases where range of averaged reputations between runs exceeded 0.1 (2.9% of cases for Cocos). For Cocos, $25.2\%$ have significant additional agreement (above the broken line). (b) Predictions. Unique stable states of the model for infinitely large homogeneous populations. Not shown are cases where multiple equilibria were found which ranged more than 0.01 in total distance in A-R space (for Cocos, 27 cases spread over only 3 strategies. See SI for some examples.) Figure 2: Comparing results of simulations and analytical predictions. Values for 171 unique strategies in 15 conditions (namely, $\epsilon\in\\{0,0.001,0.01,0.05,0.1,0.2,0.4\\}$ and $\mu\in\\{0.001,0.01,0.05\\}$). Reputation $r$ (x-axis) can range between $0$ and $1$. Agreement $a$ can range between $a=\widehat{a}$ and $a=1$ (solid black lines). Values above $a_{th}=\widehat{a}+0.05$ (broken black line) are considered as significant additional agreement. Figure 3: Results for Cocos with other observation rates $\psi$. On the left are the simulation results, and on the right, the predictions of the A-R-model. Values for all 81 unique conditional cooperators, with conditions and excluding criteria the same as those given in figure 2. As described above, we studied 171 strategies with unique behavior. Each homogeneous population of 100 players is first tested for the observation rate $\psi=1$ in 15 conditions. The conditions span an exhaustive range of perception errors $\epsilon\in\\{0,0.001,0.01,0.05,0.1,0.2,0.4\\}$ and a reasonable range of cognitive error $\mu\in\\{0.001,0.01,0.05\\}$. Note, this full observation is the natural opposite to limiting assumption of the R-model where $\psi\to 0$. Figure 2(a) shows the results of our simulations as reddish dots. They are depicted separately for the three kinds as action rules: Coco, AllC and AllD. For AllC and AllD, agreement is always virtually minimal. But for Coco strategies, many points lay far above. Sometimes agreement is almost maximal ($a\to 1$), while $\widehat{a}$ predicted it to be the smallest possible value ($a=0.5$ at $r=0.5$). For Cocos, 25.2% of the measured agreements were at least 0.05 higher than expected (indicated by the broken black line). Note that we removed cases in which results differed significantly across runs. Averaging them could give a false impression of average agreement. For example, two runs with $r_{1}=0.3$ and $r_{2}=0.7$ and each minimal $a_{i}\to\widehat{a}=0.58$ would average to $r=0.5$ and $a=0.58$. For this averaged reputation, the minimal agreement would be $\widehat{a}=0.5$, so the average measurement would be $0.08$ higher than that. In other words, although the actual agreement in each run was minimal, the averaged agreement would indicate additional agreement above the threshold. We exclude cases in which the range of $r$ in a single condition is above $0.1$ ($2.8\%$ of cases in condition $\psi=1$, see SI for examples of excluded cases and alternative figure with cases included). It is clear that our findings are not caused by such artifacts and that for $\psi=1$ results differed substantially from the prediction $\widehat{a}$ with more agreement than expected for many conditional cooperators. Next, we show that these results can be replicated for $\psi=1$ by our A-R- model, see blue dots in figure 2(b). Excluded are again cases for with multiple equilibria (given the much higher precision of the analytical approach we excluded all cases in which the total distance in the agreement- reputation space, similar to equation 16, exceeded 0.01, see SI for examples of excluded cases). The general patterns of the simulation could be replicated. AllCs and AllDs have virtually no additional agreement whereas Cocos do. The qualitative fit between simulation and model is striking. Before we quantitatively describe the fit, we expand our investigation to other observation rates. We show only results for Cocos, since these are the most interesting cases, in figure 3. We studied three intermediate observation rates $\psi\in\\{.75,.5,.25\\}$ as well as a special case, where every round would have exactly one observer. A single observer is the smallest amount of meaningful observation possible. No observation would simply not change the image matrix and have no impact on reputation or agreement. It is the closest one can get to the limiting assumption of the R-Model $\psi\to 0$.333Even taking a very small probabilistic observation rate such as $\psi=0.0001$ would have actually increased the number of observers. All rounds in which no observation takes place are simply ignored. The rest may have one or more observers. In addition, simulating many rounds just to be ignored would increase the computational demand unnecessarily for no benefit. Both simulation and prediction show a steady decline in agreement, until $a$ is virtually minimal when there is only a single observer. This is a qualitative fit of what the R-model assumes. Adjusting the observation rate in the A-R-model replicates it. For all observation rates, qualitative fit is still high. And, similarly to full observation, intermediate observation rates still show significant additional agreement. We quantified the fit of simulation and predictions in two ways. Our A-R-model makes predictions about agreement $a_{p}$ as well as reputation $r_{p}$. We therefore compute the absolute distance to simulated agreement $a_{s}$ and reputation $r_{s}$ by $\Delta=\sqrt{(a_{p}-a_{s})^{2}+(r_{p}-r_{s})^{2}}.$ (16) We excluded values not shown in figures 2 and 3 (if either simulation or prediction had to be excluded, we excluded the entire pair). Results for all observation rates are shown in the left of table 1. For both maximal and intermediate observation range, the fits are excellent. The median of absolute distance between the simulations and the model is 0.003. For 92.7-95.7% of predictions, the deviation is smaller than 0.01. For the minimal observation rate, fits are slightly worse. The median 0.005 and only 72.3% of predictions are less than 0.01 of target. We also compared the predictions of our A-R-model with predictions of the R-model. Since the R-model can only predict reputation $r_{o}$, we compared the two models in that regard. The difference in deviation is shown in the right of table 1. Keep in mind that the R-model was designed only for $\psi\to 0$, which is most closely met in the single observer condition. The new A-R- model is not often better as the R-model in this case. And at least for one case, it is much worse, as indicated by the range. However, the A-R-model is superior for all high and intermediate observation rates $\psi$. Its predictions are better in $4.6\%$ to $11.5\%$ of all cases (AllD, AllC and Cocos), and the accuracy is increased by as much as $0.5$, which is half of the maximum range of $r$. Higher accuracy is especially prevalent for Cocos, to which all evolutionary successful strategies belong, such as image scoring (Nowak and Sigmund, 1998a, ) and the leading-eight (Ohtsuki and Iwasa,, 2006). Here, for $\psi=1$, the A-R-model is better in $24\%$ of cases. Considering an independently changing average agreement seems to have increased the prediction of average reputation substantially. Table 1: Deviations between Model Predictions and Simulation | A-R-model | | | Difference A-R-model minus R-model | | ---|---|---|---|---|---|--- | absolute deviation | | | deviation in reputation | | | median | $<.01$ | $<.05$ | $<-.01$ | | (A-R better) | $>.01$ | | | | | (A-R worse) | range | | | | | $\psi=1$ | 0.003 | 92.7% | 99% | 11.5% | 0% | -0.498:0.006 $\psi=0.75$ | 0.003 | 95.7% | 99% | 9.7% | 0.1% | -0.497:0.051 $\psi=0.5$ | 0.003 | 95.4% | 98.9% | 7.8% | 0.2% | -0.495:0.053 $\psi=0.25$ | 0.003 | 94.7% | 98.4% | 4.6% | 0.1% | -0.491:0.014 single observer | 0.005 | 72.3% | 86% | 1.2% | 0.4% | -0.034:0.334 ## 4 Discussion We now summarize our findings and their immediate implications, before highlighting remaining issues and limitations. We further compare our A-R- model with the recent model of Fujimoto and Ohtsuki, (2022) to highlight the differences and common ground. We then discuss some important potential extensions of the A-R-model, which would enable it to implement recently proposed enhancements to IR strategies, such as generous assessment (Schmid et al., 2021a, ) and pleasing (Krellner and Han,, 2021). We will close with possible real world implications of the discovered additional agreement. In the first part of our report, we showed that additional agreement can emerge in indirect reciprocity under private assessment. For high or intermediate observation rates, opinions about a person were shared much more often than mere chance. Therefore, the assumption of the minimal agreement $\widehat{a}$, on which most previous models (R-models) relied, cannot be generalized to these circumstances. We could on the other hand confirm that the assumption is reasonable for very low observation rates, such as a single observer for each interaction. But we showed, that results about evolutionary success with solitary observers (Okada et al.,, 2018, Okada, 2020b, , Perret et al.,, 2021) cannot yet be generalized to other observation rates. The second part of our report makes an important step towards that goal. Previous R-models can only model average reputation and therefore must rely on the assumption of minimal agreement. Our A-R-model is able to represent average levels of agreement and reputation independently. It predicts both with astounding accuracy and does so for any intermediate or high observation rates. It outperforms the predictions of R-models in these circumstances by a large margin. Making precise measurements or predictions of reputation in particular is the basis for studies on the evolutionary stability of any reputation-based IR strategies (Okada, 2020a, ). An individual’s reputation directly corresponds to how much cooperation they receive, hence it determines their payoffs and even the payoffs of others (from which the may receive cooperation, hence cause costs to them). Parallel to this work, other researchers have discovered another way to predict opinion dynamics more accurately than by average reputation alone. Fujimoto and Ohtsuki, (2022) were able to predict precise distributions of reputations, e.g. 80% of players who would have a reputation of about 0.8, 15% of 0.2, and 5% of 0.25. This was another hugely important step for analytical models of IR under private assessment and is, of course, closely related to the current paper. Our model assumes a simplified distribution of reputations. We treat the populations as if there were at most two groups, and as if their size was given by the average reputation. Fujimoto and Ohtsuki, (2022) show that this is often not the case, that there can be more groups or other compositions. Their analytical approach is exact and uses no simplification. However, as the fit of our model shows, this level of detail may not be necessary. And the level of detail in their model seems to come at a cost as their analysis is limited to only four strategies. It seems reasonable however that their framework could model all strategies, that care about action as well as reputation of the recipient, but apply only binary assessment rules (i.e. assess each action as either good or bad, but not as neutral). This would cover 64 strategies (mirror images included) compared to the 324 of our study.444Indeed, in a preprint, Fujimoto2023 modelled 16 strategies, focusing on a single action rule, instead of four. If they had looked at all actions rules and not exclude mirror images, they could have modelled 64 strategies. However, since they were only concerned with levels of cooperation, the limitation to Cocos ($\beta=(1,0)$) was perfectly justified. AllD and AllC with different assessment rules may held different reputations, yet the all cooperate the same. Their model also covers only a single observation rate $\psi=1$. The authors themselves state that being able to model arbitrary observation rates would be a very important extension of their approach. Our A-R-model is already able to do that. And, it seems at least possible, that their approach is not capable of dealing with intermediate observation rates, i.e. $\psi<1$. Because their model currently relies on the fact that the new reputation of the donor is entirely independent of their current reputation. If some players do not observe, their new opinions depend (entirely) on their current opinions, since their opinions are just kept as they are. Their model could no longer be based on one dimension (the current reputation of recipients) but would have to incorporate entirely new dimension (the current reputation of the donor). This increase in complexity might be a serious problem. These problems for arbitrary observation rates also concern the modelling of more complex, so-called 3rd-order strategies (Santos et al.,, 2021), which also use the current reputation of the donor in their assessment. In contrast, in our A-R-model, the current reputation of the donor is already incorporated. It will be straightforward to extend our model to consider all 2048 3rd-order strategies (mirror images included) (Ohtsuki,, 2004). We envisage the following additional extensions of our model. First, we can include strategies that do not use deterministic rules but probabilistic ones (Schmid et al., 2021a, , Schmid et al., 2021b, ). Instead of assessing defection always as bad, observers may only do so 80% of the time (in other words they are sometimes generous in the judgment). The same could be applied to their action rules. They may want to cooperate even with bad individuals about 10% of the time. Second, we can include pleasing (Krellner and Han,, 2020, 2021). Instead of granting or refusing help only based on the donor’s own opinion (which can be easily disturbed by perception error), the donor pools some of the opinions of others and act like the majority would decide. Some of these extensions are as easy as replacing -1, 0 or 1 in the action and assessment rules of this paper with -0.8 and 0.1. Figure 4: Detailed comparison between models for staying norm (see equation 1). Parameters are $\psi=1$, $\epsilon_{p}=0.1$ and $\mu=0.01$. Bottom graph shows the two-dimensional space of the A-R-model. Dark blue areas indicate likely changes in either agreement or reputation or both, whereas light yellow areas indicate little to no change. Arrows indicate the direction of the change. The precise predictions of the model is indicated by a blue x. Simulations are shown in red, the averages of single runs are shown as red circles and their total average as a red +. Top graph shows one-dimensional R-model in comparison. Change in reputation and its direction is depicted on the y-axis in addition to the arrows. The stable point (green x) is where the x-axis is crossed with a downward slope. The white line in the A-R-model indicates a shift from reputation decrease (below) to reputation increase (above). This paper is the first to study agreement in such detail. Agreement is the key feature that distinguishes previous research on IR under public assessment from the more recent research on private assessment. Public assessment fixes agreement at its maximum (all agree all the time), but private assessment does not fix it at the opposite state i.e. its minimum $\widehat{a}$. Public and private are not opposite sides of the same coin. Rather private assessment allows agreement to vary, opening a new dimension of complexity to the dynamics of reputation-based IR. Our A-R-model allows one to study agreement and reputation in even more detail than was reported within the results of this paper. We focused on the stable states of the population. But one could study the direction and likelihood of change in each state, as seen in figure 4 for the example of the staying strategy (see also SI for other important strategies, such as image scoring (Nowak and Sigmund, 1998a, ) and the leading-eight (Ohtsuki and Iwasa,, 2006)). One important insight this grants is about the stability of other regions. For staying, there exist somewhat stable areas with reputations lower than $r=0.3$ and minimal agreement. The existence of additional relatively stable regions has large implications for the short-term behavior of the population. Seeing the direction of change in the entire space also allows us to make educated guesses about another dynamic. The white line in the A-R-model of figure 4 indicates a shift from reputation decrease (below) to reputation increase (above). This highlights how the trend in reputation depends on the state of agreement. It seems reasonable, that increasing agreement could increase reputation as well. Increase in agreement could be done by a form of opinion synchronization, for example by some players gossiping about their observation or opinions to make other opinions fall into line with their own. Hilbe et al., (2018) suggested that any IR strategy would profit of, or indeed rely on, some sort of opinion synchronization to maintain stable cooperation under private assessment. With the A-R-model, we can visualize which strategies can actually profit from such mechanisms. Some are not affected, e.g. image scoring, some actually show the opposite pattern, such as ”GKGB” (Okada, 2020b, ), which seems only stable under private assessment, but not under public one (see SI for examples). In the future, we can seek a way to incorporate a probability for opinion synchronization in the A-R-model, to predict new stable points under various synchronization regimes. Related to that is a last alteration of the model. In the current investigation, we considered private assessment in accordance with most literature (Okada, 2020a, ). Every individual or player in the population observes independently. But we can imagine a situation, where only a few players observe, who then share their assessment with many others. Consider for example the extreme case of a single observer that shares with the entire population. This would correspond to public assessment (Ohtsuki and Iwasa,, 2004), where no disagreement is possible ($a=1$). In a more realistic scenario, multiple observers are at least possible, and they may judge the interaction differently due to individual errors or different previously held opinions. The information of their judgments may also fail to reach some players. In such scenarios, disagreement can exist. However, all players who got the news from the same observer will most likely have the same opinion (even if information transfer is noisy). It is a form of built-in opinion synchronization. We can therefore expect agreement to be higher than in the scenario of the current paper. The amount of agreement reported here might only be the lower end for private assessment scenarios. Lastly, we look at the reasons additional agreement emerges. As discussed above, there can be no additional agreement if every player has the same reputation. Additional agreement requires at least two groups exist (but up to $N$-many), with different average reputations. Such groups can only emerge, if players with the same strategy face different fates. For example, a donor of strategy $\alpha=(1,-1,1,-1)$ and $\beta=(1,0)$ (image scoring) can get lucky by meeting a recipient they believe to be good. Hence the donor cooperates and earns a high reputation (everybody who observes, and does not make an error, believes the donor to be good). But another donor with the same strategy might get unlucky, because they believe the recipient to be bad. Hence they defect and most observers will now think this donor is bad. In general, groups can emerge if different onetime events cause different expected reputations. Such onetime event are which recipient is met or which decision is made. In the image scoring example, if errors happen with probability $0.1$, we expect the reputation of a donor to be either $0.1$ if they defected or $0.9$ if they cooperated. Different decisions as donor seem indeed to be the most important factor for additional agreement. That is why only conditional cooperators (Coco) show additional agreement, but unconditional cooperators (AllC) or defectors (AllD) show none. Because some Cocos cooperate and some defect, they can form groups with different reputations and therefore have additional agreement. Is additional agreement a good thing? Our results for strategies such as staying show that it can increase reputations. The reputations studied here correspond to how often a player finds another player of the same strategy worth of cooperation (and Cocos also act on it). More agreement can increase reputations, hence increase cooperation rates within the strategy, hence increase the stability of that strategy. This could keep defectors at bay and increase cooperation in general. Evolutionary investigations need to confirm these assumptions, but in that regard, agreement might be a good thing. However, we also showed another consequence of agreement. It is connected with differences in reputations. Such differences inevitably lead to short-term inequality between players. Over the long run, each player of the same strategy will alternate between being lucky and being unlucky, so each will earn the same pay-offs. But some players may earn much less than others if only a few consecutive interactions are considered. This can be a problem, if the game (or life) would require a player to earn a minimal amount to sustain themselves. It may also be a problem if such inequality itself has bad consequences for individuals or society. Research on IR seems to continue to converge (Okada, 2020a, , Krellner and Han,, 2022). However, there seem to be still a lot of potential to deepen our understanding of the processes, as demonstrated in Fujimoto and Ohtsuki, (2022) and the current paper. The understanding of agreement is central for any analytical model of IR under private assessment. It is central in understanding what benefits or problems that opinion synchronization or different observation rates may bring. And it is as central to all research on reputation-based IR under private assessment, since agreement can change independently of reputation and can significantly alter the latter as well. Our A-R-models provides the means to study the evolutionary dynamics of indirect reciprocity under private assessment for yet the largest strategy space and widest range of conditions. ## References * Brandt and Sigmund, (2004) Brandt, H. and Sigmund, K. (2004). The logic of reprobation: Assessment and action rules for indirect reciprocation. Journal of Theoretical Biology, 231(4):475–486. * Fujimoto and Ohtsuki, (2022) Fujimoto, Y. and Ohtsuki, H. (2022). Reputation structure in indirect reciprocity under noisy and private assessment. Scientific Reports, 12(1):1–13. * Hilbe et al., (2018) Hilbe, C., Schmid, L., Tkadlec, J., Chatterjee, K., and Nowak, M. A. (2018). Indirect reciprocity with private, noisy, and incomplete information. Proceedings of the National Academy of Sciences, page 201810565\. * Krellner and Han, (2020) Krellner, M. and Han, T. A. (2020). Putting oneself in everybody’s shoes - Pleasing enables indirect reciprocity under private assessments. In The 2020 Conference on Artificial Life, pages 402–410, Cambridge, MA. MIT Press. * Krellner and Han, (2021) Krellner, M. and Han, T. A. (2021). Pleasing Enhances Indirect Reciprocity-Based Cooperation Under Private Assessment. Artificial Life, pages 1–31. * Krellner and Han, (2022) Krellner, M. and Han, T. A. (2022). The Last One Standing? - Recent Findings on the Feasibility of Indirect Reciprocity under Private Assessment. In The 2022 Conference on Artificial Life, volume 1, Cambridge, MA. MIT Press. * Leimar and Hammerstein, (2001) Leimar, O. and Hammerstein, P. (2001). Evolution of cooperation through indirect reciprocity. * (8) Nowak, M. A. and Sigmund, K. (1998a). Evolution of indirect reciprocity by image scoring. Nature, 393(6685):573–577. * (9) Nowak, M. A. and Sigmund, K. (1998b). The dynamics of indirect reciprocity. Journal of Theoretical Biology, 194(4):561–574. * Nowak and Sigmund, (2005) Nowak, M. A. and Sigmund, K. (2005). Evolution of indirect reciprocity. Nature, 437(7063):1291–1298. * Ohtsuki, (2004) Ohtsuki, H. (2004). Reactive strategies in indirect reciprocity. Journal of Theoretical Biology, 227(3):299–314. * Ohtsuki and Iwasa, (2004) Ohtsuki, H. and Iwasa, Y. (2004). How should we define goodness? - Reputation dynamics in indirect reciprocity. Journal of Theoretical Biology, 231(1):107–120. * Ohtsuki and Iwasa, (2006) Ohtsuki, H. and Iwasa, Y. (2006). The leading eight: Social norms that can maintain cooperation by indirect reciprocity. Journal of Theoretical Biology, 239(4):435–444. * (14) Okada, I. (2020a). A Review of Theoretical Studies on Indirect Reciprocity. Games, 11(3):27. * (15) Okada, I. (2020b). Two ways to overcome the three social dilemmas of indirect reciprocity. Scientific Reports, 10(1). * Okada et al., (2017) Okada, I., Sasaki, T., and Nakai, Y. (2017). Tolerant indirect reciprocity can boost social welfare through solidarity with unconditional cooperators in private monitoring. Scientific Reports, 7(1):9737. * Okada et al., (2018) Okada, I., Sasaki, T., and Nakai, Y. (2018). A solution for private assessment in indirect reciprocity using solitary observation. Journal of Theoretical Biology, 455:7–15. * Panchanathan and Boyd, (2003) Panchanathan, K. and Boyd, R. (2003). A tale of two defectors: The importance of standing for evolution of indirect reciprocity. Journal of Theoretical Biology, 224(1):115–126. * Perret et al., (2021) Perret, C., Krellner, M., and Han, T. A. (2021). The evolution of moral rules in a model of indirect reciprocity with private assessment - DRAFT. Scientific Reports, 11(1). * Rand and Nowak, (2013) Rand, D. G. and Nowak, M. A. (2013). Human cooperation. Trends in Cognitive Sciences, 17(8):413–425. * Santos et al., (2021) Santos, F. P., Pacheco, J. M., and Santos, F. C. (2021). The complexity of human cooperation under indirect reciprocity. Philosophical Transactions of the Royal Society B: Biological Sciences, 376(1838). * Sasaki et al., (2017) Sasaki, T., Okada, I., and Nakai, Y. (2017). The evolution of conditional moral assessment in indirect reciprocity. Scientific Reports, 7(1):41870. * (23) Schmid, L., Chatterjee, K., Hilbe, C., and Nowak, M. A. (2021a). A unified framework of direct and indirect reciprocity. Nature Human Behaviour, 5(10):1292–1302. * Schmid et al., (2022) Schmid, L., Hilbe, C., Chatterjee, K., and Nowak, M. A. (2022). Direct reciprocity between individuals that use different strategy spaces. PLOS Computational Biology, 18(6):e1010149. * (25) Schmid, L., Shati, P., Hilbe, C., and Chatterjee, K. (2021b). The evolution of indirect reciprocity under action and assessment generosity. Scientific Reports, 11(1):1–14. * Sigmund, (2016) Sigmund, K. (2016). The calculus of selflessness. Princeton University Press. * Sudgen, (1986) Sudgen, R. (1986). The Economics of Rights, Cooperation and Welfare. Basic Blackwell. * Uchida, (2010) Uchida, S. (2010). Effect of private information on indirect reciprocity. Physical Review E, 82(3):036111. * Uchida and Sasaki, (2013) Uchida, S. and Sasaki, T. (2013). Effect of assessment error and private information on stern-judging in indirect reciprocity. Chaos, Solitons & Fractals, 56:175–180.
# LLMs Still Can’t Avoid Instanceof: An Investigation Into GPT-3.5, GPT-4 and Bard’s Capacity to Handle Object-Oriented Programming Assignments Bruno Pereira Cipriano<EMAIL_ADDRESS>0000-0002-2017-7511 Lusófona University, COPELABSCampo Grande, 376LisbonPortugal1700-097 and Pedro Alves 0000-0003-4054-0792<EMAIL_ADDRESS>Lusófona University, COPELABSCampo Grande, 376LisbonPortugal1700-097 (2024) ###### Abstract. Large Language Models (LLMs) have emerged as promising tools to assist students while solving programming assignments. However, object-oriented programming (OOP), with its inherent complexity involving the identification of entities, relationships, and responsibilities, is not yet mastered by these tools. Contrary to introductory programming exercises, there exists a research gap with regard to the behavior of LLMs in OOP contexts. In this study, we experimented with three prominent LLMs - GPT-3.5, GPT-4, and Bard - to solve real-world OOP exercises used in educational settings, subsequently validating their solutions using an Automatic Assessment Tool (AAT). The findings revealed that while the models frequently achieved mostly working solutions to the exercises, they often overlooked the best practices of OOP. GPT-4 stood out as the most proficient, followed by GPT-3.5, with Bard trailing last. We advocate for a renewed emphasis on code quality when employing these models and explore the potential of pairing LLMs with AATs in pedagogical settings. In conclusion, while GPT-4 showcases promise, the deployment of these models in OOP education still mandates supervision. programming assignments, teaching, object-oriented programming, object- oriented design, OOP best practices, large language models, gpt-3, gpt-4, bard ††journalyear: 2024††copyright: rightsretained††conference: 46th International Conference on Software Engineering: : Software Engineering Education and Training; April 14–20, 2024; Lisbon, Portugal††booktitle: 46th International Conference on Software Engineering: : Software Engineering Education and Training (ICSE-SEET ’24), April 14–20, 2024, Lisbon, Portugal††doi: 10.1145/3639474.3640052††isbn: 979-8-4007-0498-7/24/04††ccs: Applied computing Computer-assisted instruction ## 1\. Introduction In the last few years, researchers introduced us to Large Language Models (LLMs), which are tools that can predict the next word in a sequence, after being trained on large amounts of data, and taking into consideration a large number of parameters. Popular implementations of LLMs are ChatGPT, a chatbot developed by OpenAI on top of their Generative Pre-trained Model (GPT), and GitHub Copilot111https://github.com/features/copilot, a developer support tool that can generate code from annotations and is also based on GPT. As CS educators, we are interested in determining how these tools can be used to support students in their programming learning process, both inside and outside of the classroom. We are specifically interested in the possibilities of using these tools for teaching and learning object-oriented design and programming. Previous research has shown that LLMs have the capacity to generate computer program code from natural language descriptions (Xu et al., 2022). Furthermore, GPT-based models, such as Codex, have been shown to be able to solve most types of exercises that are used in introductory programming classes (Finnie-Ansley et al., 2022). Additionally, researchers have reported that GPT-3.5 can handle object-oriented programming assignments, managing to reach decent to good scores, with the caveat that the generated code is not up-to-par with the industry’s best practices of that paradigm (Cipriano and Alves, 2023a). However, both the availability and the capabilities of these tools are increasing rapidly. Updated models, such as GPT-4222Launched on March 14, 2023, promise even better performance than their older versions (OpenAI, 2023; Bubeck et al., 2023), while new tools, such as Bard333Launched in the US & UK on Mar 21, 2023; in EU & Brazil on Jul 13, 2023., a chatbot developed by Google on top of their Language Model for Dialogue Applications (LaMDA) (Sundar, 2023; Thoppilan et al., 2022), are becoming available to the general public. These new models justify performing further research on these topics, to confirm and/or update the findings that resulted from the previous works. Thus, we decided to explore these newer tools to learn their capabilities and limitations, as well as identify opportunities to integrate them into our courses, similarly to what is being done by other instructors (Lau and Guo, 2023; Denny et al., 2023; Daun and Brings, 2023; Leinonen et al., 2023; Liffiton et al., 2023). As such, we performed an evaluation of some of the most recently available LLMs, GPT-4, and Bard, following previous research focused on object-oriented design, programming and best-practices using GPT-3.5 (Cipriano and Alves, 2023a). This evaluation was based on assignments that are automatically graded by an Automatic Assessment Tool (AAT). This paper makes the following contributions: * • Presents and compares the performance (as measured by an AAT) of GPT-3.5, GPT-4, and Bard using 6 real-world assignments that have been used to grade students in an Object-Oriented Programming university course; * • Identifies and categorizes the errors found in the code generated by the tested LLMs; * • Provides full logs of the interactions used to get each model to solve one of the assignments, in an annotated prompt/reply format; * • Presents a set of recommendations for CS Educators wishing to adapt their classes to the availability of these technologies. ## 2\. Related work In the last couple of years, a relevant body of research related with measuring LLMs’ ability to handle introductory programming exercises has been published (Chen et al., 2021; Finnie-Ansley et al., 2022, 2023; OpenAI, 2023; Reeves et al., 2023; Wermelinger, 2023; Destefanis et al., 2023). Those introductory exercises are usually solved with a single function that receives a set of parameters and transforms them into an expected output. For example, in (Chen et al., 2021), researchers managed to solve 70% of 164 Python programming exercises using code generated by Codex, a GPT-based language model trained on code publicly available in GitHub. In (Finnie-Ansley et al., 2022), Codex was evaluated using 23 introductory programming assignments, and managed to make the 17th position when ranked alongside 71 students. The authors of (Reeves et al., 2023) focused on Codex’s capacity to handle Python- based Parsons problems (Parsons and Haden, 2006) and revealed that the model solved 50% of the cases when indentation errors were considered, and 80% when those errors were ignored. Finally, a study (Destefanis et al., 2023) based on single-function Java exercises from CodingBat444https://codingbat.com, reported that GPT-3.5 solved 90.6% of the functions, while Bard solved 53.1%. Some papers have also delved into GPT’s capacity to handle OOP assignments, which typically require the implementation of multiple classes that interact with each other and have mutable states. In one of these studies (Savelka et al., 2023b), researchers assessed the performance of two GPT-3.5 models (“text-davinci-002” and “text-davinci-003”) on various introductory and intermediate Python programming exercises. Some of the intermediate exercises involved OOP, and the models’ solutions were evaluated using an AAT. The first model achieved a success rate of 52.9% in the tests, while the second model scored 70.6%. However, due to the unavailability of the exercise statements, we could not ascertain whether the type of OOP exercises was equivalent to those presented in this research paper. The same authors have updated their research by performing the evaluations using GPT-4, which reached an improved score of 82.4% (Savelka et al., 2023a). In another study (Ouh et al., 2023), researchers attempted to utilize ChatGPT (versions 3.5 and 4) for solving OOP exercises. Some of these exercises involved introductory-level object usage without inheritance or polymorphism, and the chatbot managed to provide partially correct solutions. Other exercises introduced inheritance, but a UML class diagram with the solution was provided to the students, who only had to implement the corresponding code. As it was not feasible to supply the UML diagram to the model, the generated solution remained incomplete. Despite encompassing OOP exercises, this study’s tasks are either very elementary or highly guided. Another study (Cai et al., 2023), proposed the utilization of LLMs as low-code tools. While directly using ChatGPT led to solutions with weak object-oriented design, they were able to achieve good results by pre- guiding the model on OOP best practices. Finally, the authors of (Cipriano and Alves, 2023a), evaluated GPT-3.5 (“text-davinci-003”) using both functional as well as code quality criteria and found that, although GPT-3.5 was able to pass an average of 77.63% of unit tests, it only passed an average of 50% of the code quality evaluations. As far as we know, no research focused on Bard’s (or LaMDA’s) ability to handle object-oriented assignments has been published. ## 3\. The assignments This research was based on assignments used for student evaluation in a Computer Engineering degree. The course occurs on the 2nd curricular year of the degree and focus mostly on teaching Object-Oriented Design and Programming using the Java programming language. In this course, students are expected to learn how to implement object-oriented software solutions following the paradigm’s best practices (Wegner, 1990), with a strong focus on issues like code readability, modularity and extensibility. For example, students are expected to understand the drawbacks of using type testing, such as those permitted by ‘instanceof’, to decide program behaviors, since that technique tends to make programs harder to modify. The course’s assignments (e.g., tests and projects, among others) are evaluated using an open-source AAT, called Drop Project (Cipriano et al., 2022). It evaluates if the student’s code is respecting the assignment’s requirements using teacher-defined unit tests. It is also capable of verifying if the student’s code follows the expected code quality rules and guidelines. In this course, some assignments are configured to warn the teacher about the use of certain keywords that allow the program to work while disregarding some of the aforementioned best practices. An example of this is the usage of the ‘instanceof’ keyword or the ‘getClass()’ function. Due to the limitations of the plugin used by the AAT for the code quality validations (Checkstyle (Ivanov et al., 2023)), some quality validations are implemented using unit tests. A case in point is verifying whether a certain class has been declared as abstract. To evaluate the LLMs, a set of assignments that have been used in this course as mini-tests focused on inheritance and polymorphism was used. Each assignment has a different business domain, as well as different requirements. All the assignments have the following goals: identify and implement entity relationships (both composition and inheritance), implement some getters and setters, implement a non-trivial ‘toString()’ function, and, implement some functions that have to create, query and/or manipulate objects of several classes. Note that the relationships aren’t directly provided to students, who must infer which classes are the super and sub-classes. This approach sets our study apart from others where assignments give more explicit directions (Ouh et al., 2023). ⬇ We want to implement a software to help an IT company that provides ITConsultant services with the management of their employees. In this challenge, there exist two types of employees: those who pertain to the human resources management area (HRWorker), and the IT (information technology) experts, who pertain to ITConsultant. The classes Employee, HRWorker and ITConsultant must be created (...) however it should not be possible to instantiate an object using the super class’s constructor. A company is characterized by its name. A company can have multiple employees, but each employee only the belongs to a single company. A employee is identified by its id (int), name (String) and monthly salary (int). (...) Add, to the appropriate classes, the following methods: - A public int getId() method, which returns the employee’s ID. - A public int getSalary() method, which returns the employee’s salary. - A public int getValue() method, which returns the IT consultant’s hourly rate. (...) (Within the Company class) add a public String toString() method, which must be implemented to return the following: - "The company <name> does not have employees." - if the company does not have any employee; - "The company <name> has <number_employees> employees:" - if the company has at least one employee. Besides that, the information for each employee should be display, considering the respective type. Listing 1: Partial instructions for the ITCompany assignment Moreover, each assignment has behaviors based on the object type to assess whether students can devise solutions without resorting to explicit type checks like those allowed by the ‘instanceof’ keyword and ‘getClass()’ function. Finally, in some cases, we ask for the student ID to be used as the value of an object’s attribute, to guarantee unique solutions among students. Listing 1 presents a partial example of an assignment’s text. Six such assignments were used in this study. The selected assignments were originally used to grade students in different school years (2018/19, 2021/22, and 2022/23). ## 4\. Methodology To interface with GPT-3.5, openAI’s Playground555available at https://platform.openai.com/playground (free account) was used. The selected model was “text-davinci-003”. The “temperature” parameter, which controls randomness, was left at the default value of 0.7, and the “maximum length” parameter was left at the default of 256. GPT-4 was studied using OpenAI’s ChatGPT Plus666available at https://chat.openai.com/ (paid account). Finally, to study Bard, we used the respective chat interface 777available at https://bard.google.com/. Note that neither ChatGPT Plus nor Bard allows changing the models’ parameters. The experiments were done in different time periods, since we did them when the models became available to us: the original experiment with GPT-3.5 was done in December 2022, the experiment with GPT-4 was done in May 2023, and finally, the experiment with Bard was done in July 2023. To evaluate the models’ output, version 0.9.7 of the Drop Project AAT was used. For each model and assignment, the following algorithm was performed: 1. (1) paste the assignment’s text into the LLM’s text input area 2. (2) submit it 3. (3) examine the output to determine whether all the mandatory classes and functions are present 1. (a) if they are, move to step 4 2. (b) otherwise, prompt for the missing code and repeat step 3 4. (4) inspect the code for syntactic, logic, and output format errors 5. (5) manually fix any syntactic / compilation errors 6. (6) submit the LLM’s generated code to the AAT 7. (7) analyze the AAT’s output ## 5\. It works but it’s not OO - how LLM(s) tried to solve the IT Company assignment Figure 1. Simplified UML class diagram for the “IT Company” assignment, excluding non-essential attributes, methods, and constructors. Students only receive textual instructions, not this diagram. This section presents a high level overview of our attempts to use each model to solve one of the 6 assignments: the “IT Company” assignment. In this assignment, students are expected to design the object model of an IT Company with the following concepts: ‘Company’, ‘Employee’, ‘ITConsultant’, and ‘HRWorker’. The concepts and their attributes are explained to the students by text. However, the text does not explicitly indicate the relations that exist between the classes: it is up to the student to infer that there exists a composition relationship — between the ‘Company’ and the ‘Employee’ classes —, and an inheritance relationship where the ‘Employee’ class is the super-class and the ‘ITConsultant’ and ‘HRWorker’ classes are the sub-classes. Students are also expected to understand that the ‘Employee’ class should be abstract, since the text mentions that it must not be possible to instantiate that class. Finally, students must implement several mandatory functions, identifying the class of the hierarchy where each function belongs. For example, there is an attribute that only makes sense for the ‘ITConsultant’ class: the hourly rate. As such, only that class should have that attribute and the respective getter, which must be called ‘getValue()’. Also mandatory is the creation of a function called ‘getEmployeeHighestHourlyRate()‘ in the ‘Company’ class. This function must return the ‘Employee’ with the highest hourly rate. However, since the hourly rate concept only applies to the ‘ITConsultant’ class, the student must find the proper Object-Oriented design to implement this challenge. As such, if a student uses ‘instanceof’ or ‘getClass()’, a penalty will be applied to the respective grade. Figure 1 presents an overview of this assignment’s required classes, as well as the most relevant design expectations. Listing 1 presents part of this assignment’s instructions. A full version of the assignment, including both instructions as well as unit tests is available online.888https://github.com/drop-project-edu/itCompanyAssignment Refer to (Anonymous, 2023) for full logs of these interactions. ### 5.1. Main analysis GPT-3.5 required 18 interactions to generate a solution that included all mandatory elements. Issues found in that solution: * • Function ‘getEmployeeHighestHourlyRate()’ uses ‘instanceof’. * • Fails 1 test because the ‘Employee’ class is not abstract. * • Fails 2 tests because the package name was incorrectly included in ‘ITConsultant.toString()’ function’s return value. * • Fails the test for the ‘Main.myCompany()’ function due to using an uninformed number in the solution.999This problem was expected, since we did not input the student ID to the models. * • Failed to import the ‘ArrayList’ class. Bard required 11 interactions to generate a solution that included all mandatory elements. Issues found in that solution: * • Function ‘getEmployeeHighestHourlyRate()’ uses ‘instanceof’. * • Fails 1 test because the ‘Employee’ class is not abstract. * • Fails 2 tests due to incorrect usage of the downcast operator in the ‘getEmployeeHighestHourlyRate()’. * • Fails 2 tests due to not respecting the expected format of the ‘HRWorker.toString()’ and ‘Company.toString()’ functions. * • Fails the same test related to the ‘Main.myCompany()’ function that the other models failed, with a difference in the implementation: instead of guessing a number, Bard’s code attempts to obtain the student ID by using ‘System.getEnv()’, resulting in a ‘NullPointerException’.101010See footnote 9 * • Failed to import the ‘ArrayList’ class. * • Declared 3 attributes of the ‘Employee’ class as private, and tried to access them directly in the ‘ITConsultant’ sub-class. GPT-4 required 4 interactions to generate a solution that included all mandatory elements. Issues found in that solution: * • Function ‘getEmployeeHighestHourlyRate()’ uses ‘instanceof’. * • Fails 1 unit test because it did not fully respect the format of the ‘Company.toString()’ function (a new line was missing). * • Fails the test related to the ‘Main.myCompany()’ function due to using an uninformed number in the code.111111See footnote 9 In summary… These experiments show that these 3 LLMs were able to partially solve the ‘IT Company’ assignment, although with some errors, both in terms of basic syntax (e.g. missing library importations), as well as more complex object-oriented programming errors (e.g. not declaring a class as abstract, or using explicit type testing to decide program behaviors). This is similar to what happened in the other 5 assignments, as described in Section 6. ### 5.2. Attempts to improve the LLM(s)’ solutions After obtaining each model’s solutions for the “IT Company” assignment and verifying that all of them were breaking one of our code quality rules — the restriction not to use ‘instanceof’ —, we attempted to guide the models toward better solutions, similar to how we expect a student would do after seeing the AAT’s feedback, and to get a better grade. Note that these extra prompts are not part of our main methodology and are not represented in Table 1. These attempts to improve the models’ solutions were done using an ad-hoc methodology where we identified problems with each model’s solution, and attempted to give it a direct indication of the problem, to see what it would do. The first extra prompt was the same for all models: ‘Change the code to not use instanceof’. Afterward, we re-prompted based on what each model gave us. The prompts for GPT-3.5 and Bard ended up being similar because the respective intermediate solutions had similar problems from the object- oriented perspective: they both were assuming that all sub-classes of ‘Employee’ have the ‘getValue()’ function, which is not coherent with the rest of the respective solutions. Besides the initial extra prompt, GPT-3.5 required two more prompts to reach an acceptable solution: as a reply to the third prompt, it suggested adding the ‘getValue()’ function to both the super-class and the sub-class with an implementation that would return 0. In this case, 0 is being used as a flag to indicate that the object should be ignored, which is coherent with the model’s implementation of ‘getEmployeeHighestHourlyRate()’. Although that solution is more OO, it is somewhat risky: such a design would be problematic if a function ‘getEmployeeLowestHourlyRate()’ is needed at some point in the future. Bard had a behavior similar to that of GPT-3.5, although it required a fourth prompt to reach a solution acceptable from the OOP perspective. Bard’s fourth solution was equivalent to GPT-3.5’s third solution, thus having the equivalent problems. Note that, although the OO solution is acceptable, Bard implemented the ‘getValue()’ function incorrectly: it is returning the value of the salary instead of the value of the hourly rate. Since we were focused on getting Bard to solve the OO issue, we ignored this implementation detail. Finally, GPT-4 only needed the extra initial prompt to reach an acceptable OO solution. The model suggested adding an abstract version of ‘getHourlyRate()’ to the ‘Employee’ class, as well as adding a concrete version of that function (returning -1, which is much safer than returning 0) to the ‘HRWorker’ class. It also gave us the code for those changes, as well as a version of the ‘getEmployeeHighestHourlyRate()’ function that only considers objects for which ‘getValue()’ returns more than 0. In summary… Our experiments show that it was possible and easy to guide GPT-4 towards a fully working solution that respects the object-oriented best practice of avoiding explicit type testing. Also, while the other models also managed to reach acceptable, although risky, OO solutions, they required more effort (GPT-3.5: 3 prompts; Bard: 4 prompts) and generated some minor errors (e.g. Bard’s incorrect return of the salary attribute in the ‘getValue()’ function). This extra effort is coherent with the findings from other researchers (Wermelinger, 2023). ## 6\. Overall Findings This section presents the 3 models’ performances across the 6 assignments. As Table 1 shows, all models needed multiple prompts and no model passed all the tests. GPT-4 generally had the best performance, requiring fewer interactions than the other models, generating fewer compilation errors and, in general, having better results in the unit tests. Bard always had equal or worse unit test results than GPT-3.5, except in the “Home Banking” assignment, where it surpassed both GPT models. In terms of the Code Quality validation, we observed that GPT-4 yields the best results: while GPT-3.5 and Bard respected the code quality rules in 3 out of the 6 assignments, GPT-4 did it in 4 out of 6 assignments. Listing 2 shows an example of the type of code that these models tend to generate: it (mostly) works, but tends to decide behaviors using explicit type testing, via ‘instanceof’ and/or ‘getClass()’ . In summary, similarly to what previous research found for GPT-3.5 (Cipriano and Alves, 2023a), the two newer models still can’t avoid using ‘instanceof’. It should be noted that, in the “Condominium Mgt.” assignment, GPT-4 managed to pass the code quality, as well as 13/14 tests. Considering that the single test failure was expected121212See footnote 9, we can consider that GPT successfully solved one of the OOP assignments. Table 1. Evaluation of the 3 models’ solutions for each assignment. G-3.5 is GPT-3.5. The values are related to the first solution that contained all mandatory classes and functions. For example, GPT-3.5 needed 5 prompts to generate all classes and functions for the first assignment, that solution had 4 compilation errors, the code quality failed and it passed 8/13 unit tests. | Nr. of prompts | Nr. of Compilation errors | Code quality Ok? | Tests passed ---|---|---|---|--- Assignment | G-3.5 | GPT-4 | Bard | G-3.5 | GPT-4 | Bard | G-3.5 | GPT-4 | Bard | G-3.5 | GPT-4 | Bard Realtor Agency | 5 | 2 | 2 | 4 | 0 | 8 | No | Yes | Yes | | 8/13 --- (61.54%) | 12/13 --- (92.31%) | 7/13 --- (53.85%) IT Company | 18 | 4 | 11 | 1 | 0 | 4 | No | No | No | | 9/13 --- (69.23%) | 11/13 --- (84.62%) | 7/13 --- (53.85%) Home Banking | 7 | 2 | 10 | 1 | 0 | 14 | Yes | Yes | Yes | | 7/13 --- (53.85%) | 7/13 --- (53.85%) | 9/13 --- (69.23%) Condominium Mgt. | 5 | 2 | 5 | 12 | 0 | 3 | Yes | Yes | Yes | | 11/14 --- (78.57%) | 13/14 --- (92.86%) | 10/14 --- (71.43%) Home Cinema A/V | 5 | 2 | 6 | 3 | 0 | 2 | Yes | Yes | No | | 8/10 --- (80%) | 9/10 --- (90%) | 8/10 --- (80%) Railway Co. Vehicles | 11 | 6 | 8 | 4 | 3 | 10 | No | No | No | | 11/16 --- (68.75%) | 12/16 --- (75%) | 10/16 --- (62.5%) Average | 8,5 | 3 | 7 | 4,17 | 0,5 | 6,83 | 50% | 66.67% | 50% | 68,66% | 81,44% | 65,14% Listing 2: A function generated by Bard for the “Condominium Mgt.” assignment that fails our Code Quality validations. Besides using explicit type-testing (i.e. uses instanceof) Bard also failed to implement the proper business rules since the formula for the ‘Garage’ should multiply the area by 3 instead of 4 (see line 3). ⬇ 1public int calculateCondominiumPayment() { 2 int value = baseValue; 3 value += area * 4; 4 if (this instanceof Apartment) { 5 value += ((Apartment) this).getFloorNumber() * 3; 6 value += ((Apartment) this).getNrRooms() * 2; 7 } 8 else if (this instanceof Store) { 9 value += ((Store) this).getNrFronts() * 2; 10 value += ((Store) this).getNrDoors() * 2; 11 } 12 else if (this instanceof Garage) { 13 value += Math.abs(((Garage) this).getNrFloor()) * 2; 14 } 15 return value; 16} ### 6.1. Problem categorization This section presents a brief categorization of the problems observed across assignments and models. After each problem, we indicate which models displayed it at least in one assignment. Issues related with syntax * • Failed to include library imports. [GPT-3.5, Bard] * • Created minor compilation errors (e.g. missing “}”). [GPT-3.5, Bard] * • Failed to declare a required package. [GPT-3.5] * • Generated incoherent code (e.g. called a function with less arguments than it receives). [Bard] Issues related with OOP concepts * • Used instanceof or getClass() to decide program behaviours. [all] * • Failed to declare a class as abstract. [all] * • Failed to correctly identify and implement business rules. [all] * • Did not respect the ‘toString()’ function’s required format. [all] * • Failed to apply inheritance best practices. [all] * • Suggested solutions with code duplication. [all] * • Accessed a super-class’s private fields from its sub-classes. [GPT-3.5, Bard] * • Declared constructors with problems (e.g. added extra arguments). [GPT-3.5] * • Failed to properly apply the downcast operator. [Bard] * • Declared abstract methods in the super-class but failed to implement them in the sub-classes. [GPT-4] ## 7\. Recommendations for CS Educators Students are likely to misuse tools like GPT-3.5 and Bard, which offer decent- to-good OOP code generation for free. Superior models like GPT-4 are currently available at a cost. With the evident progress from GPT-3.5 to GPT-4, observed both in other research (Savelka et al., 2023a) as well as in our study, we believe that educators should adapt to this emerging landscape. ### 7.1. Give more weight to code quality evaluations For educators focusing on object-oriented programming, given our observations, as well as previous research (Keuning et al., 2023), we recommend putting more weight on evaluating items such as code quality, design patterns and other similar aspects. The course’s focus should change from just producing “functional code” to producing “functional and high-quality code”. The assessment work can be scaled using an AAT with the capacity to evaluate functional and quality requirements. ### 7.2. Use LLMs in your classes Consider embracing LLMs in your classes. One option is the adoption of in- class exercises where students have to 1) interact with LLMs to generate code, and then, 2) evaluate the respective solutions. This process would be supervised by the teacher, who would help the students analyze and critique the LLMs’ output. This has the advantage of showing students that these models should not be trusted blindly, and that they should inspect and test the generated code, possibly improving their critical thinking, similarly to other mistake-finding exercises (Naumova, 2023). This exercise can be enhanced by having students validate their findings through an AAT using teacher-defined tests or by creating their own unit tests. ### 7.3. Adopt project-based learning If you currently rely solely on using small assignments for assessments, we recommend that you consider including also project-based evaluations. Projects will require more complex code and interactions to be implemented, which will possibly require more complex interactions with the LLMs to obtain a fully working solution that respects the object-oriented best practices. The projects can be incremental, in order to allow students to grow and apply their knowledge incrementally. For example: consider having a first project delivery where the students do not need to use inheritance, followed by another delivery where inheritance is needed, followed by another delivery where some of the requirements change, and so on. Although LLMs’ performance over larger assignments has not yet been measured, there are known benefits to project-based learning (Mills et al., 2003; Shin, 2018). ## 8\. On the importance of citing your sources One of the interesting features of Bard is its ability to display sources for parts of the generated content. In the case of the “IT Company” assignment, Bard displayed a single source: a GitHub repository with a Java course.131313https:github.com/camilaabrantes/CursoJava We performed a brief analysis of the repository and found some OOP examples that are incorrect. For example, the repository contains a class called ‘Product’ with a ‘getPrice()’ method, which is extended by the sub-class ‘ImportedProduct’ that redefines the price calculation formula to take into account a customs fee. Instead of overriding the ‘getPrice()’, ‘ImportedProduct’ declares a new method ‘totalPrice()’. This is clearly wrong, in light of OOP best practices. Although we are unsure of how much of this repository actually contributed to Bard’s reply, we suspect that it is the use of data sources that have not been curated that results in the generation of code that does not follow the OOP best practices. Note that GPT-3.5 and GPT-4 do not disclose any information regarding their sources. Since the only publicly available information are high-level descriptions of the training datasets (e.g. a filtered version of Common Crawl 141414https://commoncrawl.org/the-data, English Wikipedia, among others) (Brown et al., 2020), it was not possible to perform this analysis for those models. ## 9\. Limitations The 3 experiments were done using different interaction interfaces. This may have some kind of impact on the model’s replies, namely in terms of the number of prompts required to obtain a solution. These models’ output is not deterministic: repeating the experiments for each assignment might have led us to different results. We did not employ any Prompt Engineering (PE) techniques. The models were initially inputted with the original assignments’ text. The prompts used to obtain missing elements or to guide the models’ toward better solutions consisted of very literal and direct requests. It is possible that the models would have given better if some PE techniques were applied. ## 10\. Conclusions and Future Work In this work, we present the performance of 3 LLMs when solving several OOP assignments. From our observations, GPT-4 demonstrated better performance than GPT-3.5, needing fewer interactions and producing code with fewer issues. Additionally, GPT-4 consistently passed more unit tests than GPT-3.5. In comparison, Bard lagged behind both GPT versions in number of issues, passed tests, and the number of interactions needed to guide it toward a reasonable solution. If the 3 models were students, we would say that GPT-3.5 is a student that has learned the basics of programming but struggles with object-oriented design and programming concepts; GPT-4 is a more experienced student who, with minimal guidance, is able to achieve good solutions; and Bard is a student of a level slightly below GPT-3.5, with less autonomy than the others and more issues with basic programming elements. In the end, all these students struggled somehow and made some mistakes. But can we blame students for making mistakes when they had a bad teacher? Perhaps the challenges encountered by LLMs in adhering to OOP best practices are not due to inherent reasoning limitations often associated with this technology but rather attributed to suboptimal choices in their information resources. We have verified the poor quality of the source referenced by Bard, and while GPT-3.5 and GPT-4 do not disclose their sources, it is plausible that they too might have relied on inadequate references. The significance of LLMs disclosing their sources, and the more general issue of explaining their reasoning has been a topic of repeated discussion (Arrieta et al., 2020). We not only endorse such explanations but also emphasize the necessity of curating high quality sources for training these models. As for future work, we plan to investigate how these models handle larger OOP assignments (e.g. 10-15 classes) following our recommendation to switch to project-based learning. We also plan on experimenting new ways of presenting OOP exercises to the students, with the goal of creating barriers to naive ‘copy-and-prompt’ approaches. One of the possible approaches is the creation of diagram-based and video-based assignments. ## 11\. Data Availability Since this work is based on real assignments that we use for student assessment in our course, publishing the full assignment and/or interactions dataset would require us to fully recreate all the assignments, ensuring we use unpublished materials when evaluating our students. As such, we have decided to release just one of the assignments: the IT Company Assignment. This was achieved by creating the respective GitHub repository 151515https://github.com/drop-project-edu/itCompanyAssignment, as well as publishing the respective interaction logs, for each LLM, in Zenodo (Cipriano and Alves, 2023b). ###### Acknowledgements. This research was funded by the Fundação para a Ciência e a Tecnologia under Grant No.: UIDB/04111/2020 (COPELABS). ## References * (1) * Anonymous (2023) Anonymous. 2023\. How GPT-3.5, GPT-4 and Bard handled an Object Oriented Programming Assignment - Full Interaction Logs. https://doi.org/10.5281/zenodo.8246165 This is the anonymized version to support peer review. * Arrieta et al. (2020) Alejandro Barredo Arrieta, Natalia Díaz-Rodríguez, Javier Del Ser, Adrien Bennetot, Siham Tabik, Alberto Barbado, Salvador García, Sergio Gil-López, Daniel Molina, Richard Benjamins, et al. 2020\. Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI. _Information fusion_ 58 (2020), 82–115. * Brown et al. (2020) Tom Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared D Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, et al. 2020\. Language Models are Few-Shot Learners. _Advances in neural information processing systems_ 33 (2020), 1877–1901. * Bubeck et al. (2023) Sébastien Bubeck, Varun Chandrasekaran, Ronen Eldan, Johannes Gehrke, Eric Horvitz, Ece Kamar, Peter Lee, Yin Tat Lee, Yuanzhi Li, Scott Lundberg, et al. 2023\. Sparks of Artificial General Intelligence: Early experiments with GPT-4. _arXiv preprint arXiv:2303.12712_ (2023). * Cai et al. (2023) Yuzhe Cai, Shaoguang Mao, Wenshan Wu, Zehua Wang, Yaobo Liang, Tao Ge, Chenfei Wu, Wang You, Ting Song, Yan Xia, et al. 2023\. Low-code LLM: Visual Programming over LLMs. _arXiv preprint arXiv:2304.08103_ (2023). * Chen et al. (2021) Mark Chen, Jerry Tworek, Heewoo Jun, Qiming Yuan, Henrique Ponde de Oliveira Pinto, Jared Kaplan, Harri Edwards, Yuri Burda, Nicholas Joseph, Greg Brockman, et al. 2021\. Evaluating Large Language Models Trained on Code. _arXiv preprint arXiv:2107.03374_ (2021). * Cipriano and Alves (2023a) Bruno Pereira Cipriano and Pedro Alves. 2023a. GPT-3 vs Object Oriented Programming Assignments: An Experience Report. In _Proceedings of the 2023 Conference on Innovation and Technology in Computer Science Education V. 1_ (Turku, Finland) _(ITiCSE 2023)_. Association for Computing Machinery, New York, NY, USA, 61–67. https://doi.org/10.1145/3587102.3588814 * Cipriano and Alves (2023b) Bruno Pereira Cipriano and Pedro Alves. 2023b. How GPT-3.5, GPT-4 and Bard handled an Object Oriented Programming Assignment - Full Interaction Logs. https://doi.org/10.5281/zenodo.8246165 * Cipriano et al. (2022) Bruno Pereira Cipriano, Nuno Fachada, and Pedro Alves. 2022\. Drop Project: An automatic assessment tool for programming assignments. _SoftwareX_ 18 (2022), 101079. * Daun and Brings (2023) Marian Daun and Jennifer Brings. 2023. How ChatGPT Will Change Software Engineering Education. In _Proceedings of the 2023 Conference on Innovation and Technology in Computer Science Education V. 1_. 110–116. * Denny et al. (2023) Paul Denny, Juho Leinonen, James Prather, Andrew Luxton-Reilly, Thezyrie Amarouche, Brett A Becker, and Brent N Reeves. 2023\. Promptly: Using Prompt Problems to Teach Learners How to Effectively Utilize AI Code Generators. _arXiv preprint arXiv:2307.16364_ (2023). * Destefanis et al. (2023) Giuseppe Destefanis, Silvia Bartolucci, and Marco Ortu. 2023\. A Preliminary Analysis on the Code Generation Capabilities of GPT-3.5 and Bard AI Models for Java Functions. _arXiv preprint arXiv:2305.09402_ (2023). * Finnie-Ansley et al. (2022) James Finnie-Ansley, Paul Denny, Brett A Becker, Andrew Luxton-Reilly, and James Prather. 2022\. The Robots Are Coming: Exploring the Implications of OpenAI Codex on Introductory Programming. In _Proceedings of the 24th Australasian Computing Education Conference_. 10–19. * Finnie-Ansley et al. (2023) James Finnie-Ansley, Paul Denny, Andrew Luxton-Reilly, Eddie Antonio Santos, James Prather, and Brett A Becker. 2023. My AI Wants to Know if This Will Be on the Exam: Testing OpenAI’s Codex on CS2 Programming Exercises. In _Proceedings of the 25th Australasian Computing Education Conference_. 97–104. * Ivanov et al. (2023) Roman Ivanov et al. 2023\. Checkstyle. https://checkstyle.org/. [Online; last accessed 20-January-2023]. * Keuning et al. (2023) Hieke Keuning, Johan Jeuring, and Bastiaan Heeren. 2023\. A Systematic Mapping Study of Code Quality in Education. In _Proceedings of the 2023 Conference on Innovation and Technology in Computer Science Education V. 1_. 5–11. * Lau and Guo (2023) Sam Lau and Philip J Guo. 2023. From” Ban It Till We Understand It” to” Resistance is Futile”: How University Programming Instructors Plan to Adapt as More Students Use AI Code Generation and Explanation Tools such as ChatGPT and GitHub Copilot. (2023). * Leinonen et al. (2023) Juho Leinonen, Paul Denny, Stephen MacNeil, Sami Sarsa, Seth Bernstein, Joanne Kim, Andrew Tran, and Arto Hellas. 2023\. Comparing Code Explanations Created by Students and Large Language Models. _arXiv preprint arXiv:2304.03938_ (2023). * Liffiton et al. (2023) Mark Liffiton, Brad Sheese, Jaromir Savelka, and Paul Denny. 2023. CodeHelp: Using Large Language Models with Guardrails for Scalable Support in Programming Classes. _arXiv preprint arXiv:2308.06921_ (2023). * Mills et al. (2003) Julie E Mills, David F Treagust, et al. 2003\. Engineering education—Is problem-based or project-based learning the answer. _Australasian journal of engineering education_ 3, 2 (2003), 2–16. * Naumova (2023) Elena N Naumova. 2023\. A mistake-find exercise: a teacher’s tool to engage with information innovations, ChatGPT, and their analogs. _Journal of Public Health Policy_ 44, 2 (2023), 173–178. * OpenAI (2023) OpenAI. 2023. GPT-4 Technical Report. https://doi.org/10.48550/arXiv.2303.08774 arXiv:2303.08774 [cs]. * Ouh et al. (2023) Eng Lieh Ouh, Benjamin Kok Siew Gan, Kyong Jin Shim, and Swavek Wlodkowski. 2023. ChatGPT, Can You Generate Solutions for my Coding Exercises? An Evaluation on its Effectiveness in an undergraduate Java Programming Course. _arXiv preprint arXiv:2305.13680_ (2023). * Parsons and Haden (2006) Dale Parsons and Patricia Haden. 2006. Parson’s Programming Puzzles: A Fun and Effective Learning Tool for First Programming Courses. In _Proceedings of the 8th Australasian Conference on Computing Education-Volume 52_. 157–163. * Reeves et al. (2023) Brent Reeves, Sami Sarsa, James Prather, Paul Denny, Brett A Becker, Arto Hellas, Bailey Kimmel, Garrett Powell, and Juho Leinonen. 2023. Evaluating the Performance of Code Generation Models for Solving Parsons Problems With Small Prompt Variations. In _Proceedings of the 2023 Conference on Innovation and Technology in Computer Science Education V. 1_. 299–305. * Savelka et al. (2023a) Jaromir Savelka, Arav Agarwal, Marshall An, Chris Bogart, and Majd Sakr. 2023a. Thrilled by Your Progress! Large Language Models (GPT-4) No Longer Struggle to Pass Assessments in Higher Education Programming Courses. _arXiv preprint arXiv:2306.10073_ (2023). * Savelka et al. (2023b) Jaromir Savelka, Arav Agarwal, Christopher Bogart, Yifan Song, and Majd Sakr. 2023b. Can Generative Pre-trained Transformers (GPT) Pass Assessments in Higher Education Programming Courses? _arXiv preprint arXiv:2303.09325_ (2023). * Shin (2018) Myeong-Hee Shin. 2018\. Effects of Project-Based Learning on Students’ Motivation and Self-Efficacy. _English Teaching_ 73, 1 (2018), 95–114. * Sundar (2023) Pichai Sundar. 2023\. An important next step on our AI journey. https://blog.google/technology/ai/bard-google-ai-search-updates/. [Online; last accessed 10-August-2023]. * Thoppilan et al. (2022) Romal Thoppilan, Daniel De Freitas, Jamie Hall, Noam Shazeer, Apoorv Kulshreshtha, Heng-Tze Cheng, Alicia Jin, Taylor Bos, Leslie Baker, Yu Du, et al. 2022\. Lamda: Language Models for Dialog Applications. _arXiv preprint arXiv:2201.08239_ (2022). * Wegner (1990) Peter Wegner. 1990\. Concepts and Paradigms of Object-Oriented Programming. _ACM Sigplan Oops Messenger_ 1, 1 (1990), 7–87. * Wermelinger (2023) Michel Wermelinger. 2023\. Using GitHub Copilot to Solve Simple Programming Problems. In _Proceedings of the 54th ACM Technical Symposium on Computer Science Education V. 1_. 172–178. * Xu et al. (2022) Frank F Xu, Uri Alon, Graham Neubig, and Vincent Josua Hellendoorn. 2022. A Systematic Evaluation of Large Language Models of Code. In _Proceedings of the 6th ACM SIGPLAN International Symposium on Machine Programming_. 1–10.
# Atomic transitions for adaptive optics Rui Yang School of Physics and Astronomy, Yunnan University, South Section, East Outer Ring Road, Chenggong District, Kunming, 650500, China Department of Physics and Astronomy, University of British Columbia, 6224 Agriculture Road, Vancouver BC, V6T 1Z1,Canada Joschua Hellemeier Department of Physics and Astronomy, University of British Columbia, 6224 Agriculture Road, Vancouver BC, V6T 1Z1,Canada Paul Hickson Department of Physics and Astronomy, University of British Columbia, 6224 Agriculture Road, Vancouver BC, V6T 1Z1,Canada National Astronomical Observatories, Chinese Academy of Sciences, A20 Datun Road, Chaoyang District, Beijing, China Corresponding author<EMAIL_ADDRESS> ###### Abstract Adaptive optics systems using sodium laser guide stars are widely employed at major astronomical observatories. It is natural to ask whether other atomic species might offer advantages. In this paper we review all abundant atoms and ions in the upper atmosphere, including Na, Fe, Mg+, Si+, Ca+, K and also the non-metallic species N, N+, O, H, considering their potential for adaptive optics. Return fluxes for all transitions that can be excited using either one or two wavelengths were computed. We also considered multi-wavelength emission, comparing the performance of different transitions for polychromatic laser guide star (PLGS) adaptive optics. We find that of all the mesospheric metals, Na is the most suitable for both monochromatic and polychromatic laser guide stars, providing about six times more return flux than the best transitions in Fe. For high-altitude observatories, excitation at 330 nm in Na should give the highest PLGS performance. Atomic O, N and N+ have strong transitions and very high abundances in the mesosphere. This makes them potential candidates for the generation of intense laser guide stars by amplified spontaneous emission, if a suitable excitation process can be demonstrated. Direct excitation by CW lasers is impractical as all transitions from the ground state are beyond the atmospheric cutoff. Nevertheless, it may be possible using high-power pulsed lasers. ††journal: josab ## 1 Introduction Adaptive optics (AO) systems are employed by many ground-based telescopes to compensate atmospheric turbulence, thereby greatly improving the resolution of astronomical images [1]. Many of these systems employ laser guide star (LGS) beacons, which allow sensing of atmospheric turbulence in any desired direction, increasing sky coverage [2]. Large telescopes typically employ resonant LGS systems in which transitions in mesospheric atom are excited, typically the $D_{2}$ transition of neutral atomic sodium. Although other chemical species have been considered [3, 4], sodium is preferred due to its relatively-high abundance and strong resonance lines. Despite the impressive success of AO systems employing sodium LGS, there are several reasons to reconsider other atoms and ions. The performance of such AO systems is limited by the return flux of the LGS, due to photon shot noise. Advances have been made both in increasing laser power, and improving the efficiency of sodium LGS [5, 6, 7, 8, 9]. For example, simulations indicate that repumping atoms from the $D_{2b}$ ground state, in combination with circular polarization, can increase the return flux by a factor of $3-4$ [6]. Nevertheless, any technique that might further increase the return flux would be valuable. Several new techniques that might potentially achieve high-intensity LGS are discussed in [10] . One possible approach is amplified spontaneous emission (ASE) [11]. In this idea, laser pumping of an abundant atmospheric atomic or molecular species could produce a population inversion that would amplify radiation generated by spontaneous emission. A critical requirement for ASE is a suitable transition for which the optical depth $\tau$, defined as the integral of the absorption coefficient along the propagation path, is negative and has an absolute value that exceeds unity. This is a consequence of the equation of radiative transfer, from which it follows that the gain in intensity is proportional to $\exp(-\tau)$. Another motivation for considering alternatives to sodium is the problem of atmospheric tilt indeterminacy [12]. Backscattered photons follow the same path as the transmitted beam, so the apparent position of the LGS is independent of wavefront tilt induced by the atmosphere. A number of techniques have been proposed to break this degeneracy (see, for example, [13, 14] and references therein). A promising approach is that of polychromatic laser guide stars (PLGS) [15, 16]. In this technique, a laser is used to generate emission at two or more widely-separated wavelengths. Because of differential refraction, the backscattered light propagates along different paths through the atmosphere. Thus one need not rely on natural guide stars to sense the atmospheric tilt. The differential positions of the PLGS at different wavelengths is small and must be measured with high precision. Creating sufficiently-intense polychromatic emission is a challenge. There is also a related problem of focus indeterminacy, as the sodium centroid altitude varies on short timescales [17]. This makes it difficult to separate atmospheric focus variations from other focus errors. This problem primarily affects very-large telescopes as the wavefront error due to focus variations increases in proportion to the square of the telescope aperture diameter [18]. The present paper examines atoms and ions in the upper atmosphere that might be useful for AO. We consider the abundance and distribution of all relevant species and compute the relative return flux for transitions that can be excited by ground-based lasers. Our aim is to identify potential transitions for monochromatic and PLGS, and to assess the feasibility of ASE. ## 2 Mesospheric atoms and parameters of interest Meteoric ablation is the primary source of metallic species in the mesosphere and lower thermosphere (MLT) region of the atmosphere, at altitudes between 80 and 120 km. In this altitude range, the atmospheric density decreases from $10^{-5}$ to $10^{-8}$ kg m-3 and collision rates are in the range $10^{3}-10^{5}$ s-1. This is also the coldest region of the atmosphere. At mid-latitude sites, the temperature at the mesopause ranges from $178-192$ K, at an altitude of $85-102$ km, depending on the season [19]. Typical values of the column density $N$, altitude of the centroid of the density distribution, and the RMS width (vertical extent) for each species are listed in Table 1. For sodium, we use recent data by Gardner [20] to update earlier values listed by Happer [3]. The centroid altitude and RMS width for non-metallic species were computed from the MSIS-90 atmospheric model [21]. These were evaluated at midnight, January 1 for latitude $+30^{\circ}$ and longitude $0^{\circ}$. The major meteoric species are Na, Fe, Mg and Si. There are also two less- abundant species, Ca and K. These metals exist as layers of atoms between about 80 and 105 km and occur primarily as atomic ions at higher altitudes. Below 85 km they form compounds: oxides, hydroxides, and carbonates, which polymerize into nanometer-sized meteoric smoke particles. Above 85 km, photolysis of N2, O2 and H2O by extreme ultraviolet (EUV) radiation leads to high concentrations of nitrogen, oxygen and hydrogen atoms. These atoms attack metallic compounds such as hydroxides and oxides, reducing them and thereby returning the metals to their atomic phase. Table 1: Parameters of mesospheric atoms and ions Species | $N$ | Altitude | RMS width | Sources ---|---|---|---|--- | m-2 | km | km | Na | $4.0$$\times 10^{13}$ | $91.5$ | $4.6$ | [20] Fe | $10.2$$\times 10^{13}$ | $88.3$ | $4.5$ | [20] Mg | $1$$\times 10^{13}$ | $89.4$ | $4.2$ | [23, 22] Mg+ | $8$$\times 10^{13}$ | $94.6$ | $7.0$ | [23, 22] Si+ | $4$$\times 10^{13}$ | $111.0$ | $10.1$ | [24] Ca | $3.4$$\times 10^{11}$ | $90.5$ | $3.5$ | [20] Ca+ | $7.2$$\times 10^{11}$ | $95.0$ | $3.6$ | [20] K | $4.5$$\times 10^{11}$ | $91.0$ | $4.7$ | [20] N | $1$$\times 10^{18}$ | $223.0$ | $60.2$ | [25, 21] N+ | $1$$\times 10^{18}$ | - | - | [25] O | $6.5$$\times 10^{21}$ | $110.3$ | $28.7$ | [26, 21] H | $2.3$$\times 10^{18}$ | $95.6$ | $38.9$ | [27, 21] In a typical AO application, a laser is used to excite atoms from the ground state to an excited state. The atom then returns to the ground state via spontaneous and stimulated emission. Only spontaneous emission contributes to backscattered radiation since stimulated emission propagates in the same direction as the laser radiation. The LGS return flux can be computed from the equation of radiative transfer. For an optically-thin medium (optical depth $\tau\ll 1$) the specific intensity $I_{\nu}$ (W m-2 Hz-1 sr-1) of the backscattered radiation is given by [10] $I_{\nu}(\nu)=\frac{h\nu}{4\pi}NA_{21}x\varphi(\nu)=\frac{2h\nu}{\lambda^{2}}N\sigma_{21}(\nu)x.$ (1) Here $\lambda$ is the wavelength, $\nu=c/\lambda$ is the frequency, $c$ is the speed of light in vacuum and $N$ is the column density of atoms of this species. $A_{21}$ is the Einstein A coefficient for the transition and the function $\varphi(\nu)$ is the mean line profile, normalized so that its integral over frequency $\nu$ is unity. The dimensionless parameter $x$ is the occupation fraction of the upper state, i.e. the equilibrium fraction of atoms found in this state under continuous excitation by a laser having specific energy density $\rho_{\nu}$, and $\sigma_{21}(\nu)$ is the cross section for stimulated emission. The second equality follows from the relation between stimulated and spontaneous emission, $B_{21}$ = $\lambda^{3}A_{21}/8\pi h$, and the definition $\sigma_{21}(\nu)\equiv\frac{h\nu}{c}B_{21}\varphi_{\nu}(\nu)=\frac{\lambda^{2}}{8\pi}A_{21}\varphi(\nu).$ (2) Here the subscript 1 denotes the lower (ground) state, 2 denotes the upper state, and $B_{21}$ is the Einstein B coefficient for stimulated emission. For a CW laser having beam area $\mathcal{A}$ in the MLT, the return flux $\Phi$, in units of photons s-1 m-2 is related to the intensity by $\Phi=\frac{\mathcal{A}}{z^{2}}\int_{0}^{\infty}\frac{I_{\nu}(\nu)}{h\nu}d\nu,$ (3) where $z$ is the line-of-sight distance to the LGS. In the mesosphere, Doppler broadening is the dominant line-broadening process. As the temperature varies within the MLT, the line profile is more precisely written as $\varphi(\nu)=\frac{1}{N}\int_{0}^{\infty}n(z)\varphi(\nu,z)dz,$ (4) where $n(z)$ is the number density of the atomic species under consideration at distance $z$ along the line of sight. In practice, the function $\varphi(\nu)$ is well-represented by a Gaussian $\varphi(\nu)=\frac{2}{w}\sqrt{\frac{\ln 2}{\pi}}e^{-4\ln 2(\nu-\nu_{0})^{2}/w^{2}},$ (5) where $w$ is the line width at half maximum intensity. It depends on the characteristic temperature $T$ in the MLT and the mass $m$ of the atomic species, $w=\frac{1}{\lambda_{0}}\left(\frac{8k_{B}T\ln 2}{m}\right)^{1/2}$ (6) where $k_{B}$ is Boltzmann’s constant. Adopting $T=180$ K results in a width for the sodium D line of 1.02 GHz. The absorption cross section $\sigma_{12}(\nu)$ is given by $\sigma_{12}(\nu)=\frac{g_{2}}{g_{1}}\sigma_{21}(\nu),$ (7) where $g_{1}$ and $g_{2}$ are the statistical weights of the lower and upper state, respectively. Tabulated values of cross sections $\sigma_{21}\equiv\sigma_{21}(\nu_{0})$ are computed at the line center $\nu_{0}$. Substituting Eqs. (1) and (2) into Eq. (3), we obtain $\displaystyle\Phi$ $\displaystyle=\frac{\mathcal{A}}{4\pi z^{2}}NA_{21}\int_{0}^{\infty}x(\nu)\varphi(\nu)d\nu$ $\displaystyle\simeq\frac{\mathcal{A}}{4\pi z^{2}}NA_{21}x(\nu_{0})\varphi(\nu_{0})\Delta\nu,$ (8) where $\Delta\nu$ is the laser line width. The second equality follows because the line width for a CW laser, and therefore the frequency range over which $x$ is nonzero, is typically much smaller than the Doppler width. From this we see that the atomic quantity that best characterizes the LGS return flux, is $\varepsilon=NA_{21}x(\nu_{0}).$ (9) We shall call $\varepsilon$ the _return flux coefficient_ for the transition. It has units of photons s-1 m-2. Also of interest is the dimensionless product $N\sigma_{21}$, which determines the maximum possible optical depth. Specifically, $\tau$ is bounded by $-N\sigma_{21}\leq\tau\leq N\sigma_{12}$ [10]. Therefore, ASE is not possible if $N\sigma_{21}<1$. Several other physical processes also play a role in the response of atoms to laser excitation. Atomic collisions effectively reset atoms that have been excited, either returning them to the ground state, or causing a transition to a different excited state. The latter process makes the atom unavailable for excitation by the laser. At the same time, collisions impart a velocity to the atom, generally pushing it out of the velocity interval that can be excited by the laser (the “velocity class”). In equilibrium this is balanced by collisions that move atoms into the velocity class. Those atoms are in the ground state as the collision energies, typically $\sim 0.02$ eV, are not sufficient to excite the atom from the ground state. Therefore any individual atom is effectively “reset” to the ground state on a time scale equal to the mean time between collisions. Collision rates vary by species and altitude. Rates for the species of interest are calculated in Appendix A, for a range of altitudes. In order to be useful for AO, a transition should have a high spontaneous transition rate to the lower state (Einstein $A$ coefficient). Also, as most atoms will be in the ground state, there should be a path to the upper state from the ground state, with a high transition probability (Einstein $B$ coefficient). As the Einstein coefficients are related, it is sufficient to consider transitions that have a large $A$ coefficient. A second consideration is that the upper state should not have many transitions to metastable states, which have low $A$ coefficients. In such a situation, atoms can become trapped in those states, reducing the number of atoms available for the desired transition. Ultimately, atoms trapped in metastable states will be reset at the collision rate. The performance of PLGS will critically depend on the wavelengths and return flux of the available transitions. For amplitude $\theta$ of a Zernike mode (except for piston), the ratio of the difference in amplitude $\Delta\theta$ at different wavelengths, is related to the ratio of refracting indexes by [15] $p\equiv\frac{\theta}{\Delta\theta}=\frac{n_{\lambda_{\text{obs}}}-1}{n_{\lambda_{1}}-n_{\lambda_{2}}},$ (10) where $n_{\lambda_{1}}$ and $n_{\lambda_{2}}$ are the refractive indexes at wavelengths of the measurement and $n_{\lambda_{obs}}$ is the refractive index of the wavelength at which the amplitude of the Zernike mode is determined. The dimensionless ratio $p$ is called the _penalty factor_. Smaller values of $p$ correspond to greater PLGS performance. We take $\lambda_{\text{obs}}=500$ nm as the reference wavelength. The impact of return flux and penalty factor can be captured in a single merit function. The signal-to-noise ratio for the measurement of the atmospheric wavefront tilt is proportional to the square root of the return flux and inversely proportional to the penalty factor [15]. Therefore, $\varepsilon^{1/2}/p$ is a measure of relative PLGS performance. To make this dimensionless, we divide by the same quantity evaluated for a reference system. We thus define the PLGS merit function $q\equiv\frac{p_{0}}{p}\left(\frac{\varepsilon}{\varepsilon_{0}}\right)^{1/2},$ (11) where $\varepsilon_{0}$ and $p_{0}$ are reference values. For these we use the 330.298/2208.370 nm transitions in Na I, which is the highest-performance system employing single-photon excitation. In all cases, the return flux coefficient used is that of the weaker line. When discussing atomic transitions we use spectroscopic notation and refer to Fe as Fe I, Fe+ as Fe II, etc. We refer to transitions having wavelengths greater than 300 nm as “visible” transitions. Shorter-wavelengths are greatly attenuated by the atmosphere so we refer to those as “vacuum-ultraviolet” transitions. We refer to transitions having an Einstein $A$ coefficient greater than $10^{6}$ s-1 as “strong transitions”, and “weak transitions” otherwise. This distinction is somewhat arbitrary but serves to generally separate permitted lines from forbidden lines. An excitation scheme that uses two different transitions to reach an excited state will be called a “two- step” process. One that uses two photons to excite a single transition, using a virtual intermediate state, will be called a “two-photon” process. Transition wavelengths described in the text and listed in the tables in the following sections are in air for wavelengths greater than 185 nm and in vacuum otherwise. ## 3 Analysis and Results The data set employed for this analysis is the atomic database of the National Institute of Standards and Technology (NIST) [28]. Complete data for all transitions were downloaded and a computer program was written to apply various criteria to identify transitions of interest for AO. We examined all excited states that could be reached from the ground state by at most two successive excitations using wavelengths greater than 300 nm. For species in which no such transitions are possible, two-photon excitation was considered. For every such state, all possible downward transitions were examined, and the excitation fraction $x$ of the excited state was computed according to the methods presented in Appendix B. A specific energy density of $\rho_{\nu}=10^{-12}$ J m-1 Hz-1 was assumed for the laser excitation. This is at the high end of range of energy densities that can be achieved with current AO lasers. A complete listing of all such transitions that have $A\geq 10^{7}$ s-1 is available from the authors upon request. From this and the data of Table 1, the return flux coefficient was computed. For PLGS, the penalty factor $p$ and merit function $q$ were computed for all downward transitions. Our results are shown in Tables 2 – 9. Table 2 lists all transitions, for all species, that have a return flux coefficient $\varepsilon>10^{18}$ photons s-1 m-2. In this, and subsequent tables, $\lambda_{\text{ex}}$ refers to the excitation wavelength(s) employed to reach the excited state, and $\lambda$ is the wavelength of the downward transition. Unless otherwise specified, values of $p$ are computed with respect to the smallest wavelength. We now discuss individual atomic species in more detail. ### 3.1 Sodium Sodium is most-often employed for the generation of LGS. Neutral sodium is used as it is more abundant than ionized forms. Averaged over all seasons and latitudes, the observed ratio of Na+ to Na is between 0.2 and 0.25 [29]. An energy level diagram of neutral sodium, indicating the lifetimes and wavelengths for selected transitions is shown in Fig. 1. For simplicity, fine- structure splitting of the levels is not shown. At visible wavelengths, Na I has four strong transitions from the ground state, at 330.237, 330.298, 588.995 and 589.592 nm. The latter two form the sodium D doublet. A fifth transition, at 388.390 nm, has a very-small rate ($A=6.95\times 10^{-4}$ s-1). The transition between the 3S1/2 ground state and the 3P3/2 excited state, the D2 line, has been of interest for adaptive optics since 1985 [2]. Laser- induced excitation followed by spontaneous emission produces the backscattered photons of the LGS. The LGS return flux is limited by the low abundance and small absorption cross section, $N\sigma_{12}\simeq 3$%. In practice, the D2b line, which involves only transitions between the $F=2$ lower and $F=1,2$ and 3 upper sub-levels, is used for AO. A CW laser can excite 3/4 of the available atoms to the upper state, corresponding to the ratio $15/20$ of the number of magnetic substates in the upper levels to the total. Current AO lasers are sufficiently powerful that the excitation fraction closely approaches this limit. However, only a fraction of the total number of sodium atoms are available for excitation because of limited spectral overlap as the laser linewidth is typically much smaller than the sodium linewidth. This limitation might be overcome with further development of efficient broad-band modeless pulsed lasers [30]. Figure 1: Low-lying energy levels of the sodium atom and transitions of interest. Table 2. Transitions for a monochromatic LGS having $\lambda\geq 300$ nm, $\lambda_{\text{ex}}\geq 150$ nm and $\log(\varepsilon)>18.3$. --- Species | $\lambda_{\text{ex}}$ | $\lambda$ | $\log(\varepsilon)$ | $\log(N\sigma_{21})$ | (nm) | (nm) | | Na I | $588.995$ | $588.995$ | $21.20$ | $-1.50$ | $589.592$ | $589.592$ | $21.07$ | $-1.50$ | $330.237$ | $2205.647$ | $20.00$ | $-0.75$ | $330.237$ | $588.995$ | $19.83$ | $-1.50$ | $330.298$ | $2208.370$ | $19.82$ | $-0.75$ | $330.237$ | $1140.377$ | $19.82$ | $-1.19$ | $330.298$ | $588.995$ | $19.65$ | $-1.50$ | $330.298$ | $1140.377$ | $19.65$ | $-1.19$ | $330.237$ | $330.237$ | $19.61$ | $-3.61$ | $330.237$ | $589.592$ | $19.52$ | $-1.50$ | $330.237$ | $1138.144$ | $19.52$ | $-1.49$ | $330.298$ | $330.298$ | $19.44$ | $-3.61$ | $330.298$ | $589.592$ | $19.37$ | $-1.50$ | $330.298$ | $1138.144$ | $19.35$ | $-1.49$ | $285.281$ | $5426.880$ | $19.20$ | $-0.25$ | $285.281$ | $588.995$ | $19.18$ | $-1.50$ | $285.301$ | $5434.170$ | $18.98$ | $-0.25$ | $285.301$ | $588.995$ | $18.95$ | $-1.50$ | $285.281$ | $1140.377$ | $18.93$ | $-1.19$ | $285.281$ | $1074.645$ | $18.91$ | $-2.65$ | $285.281$ | $589.592$ | $18.87$ | $-1.50$ | $285.281$ | $616.075$ | $18.78$ | $-2.54$ | $268.034$ | $588.995$ | $18.71$ | $-1.50$ | $285.301$ | $1140.377$ | $18.71$ | $-1.19$ | $285.301$ | $1074.930$ | $18.68$ | $-2.65$ | $285.301$ | $589.592$ | $18.67$ | $-1.50$ | $285.281$ | $3414.261$ | $18.64$ | $-0.45$ | $285.281$ | $1138.144$ | $18.63$ | $-1.49$ | $268.034$ | $10\,807.960$ | $18.61$ | $0.14$ | $285.301$ | $616.075$ | $18.56$ | $-2.54$ | $285.281$ | $2205.647$ | $18.51$ | $-0.75$ | $285.281$ | $615.423$ | $18.49$ | $-2.84$ | $268.043$ | $588.995$ | $18.45$ | $-1.50$ | $285.301$ | $3414.261$ | $18.42$ | $-0.45$ | $268.034$ | $1140.377$ | $18.41$ | $-1.19$ | $285.301$ | $1138.144$ | $18.41$ | $-1.49$ | $268.034$ | $589.592$ | $18.40$ | $-1.50$ | $268.043$ | $10\,823.050$ | $18.36$ | $0.14$ | $259.387$ | $588.995$ | $18.34$ | $-1.50$ | $285.281$ | $3407.755$ | $18.34$ | $-0.75$ | $330.237$ | $9090.560$ | $18.33$ | $-0.57$ | $330.237$ | $819.482$ | $18.33$ | $-1.15$ | $268.034$ | $2440.182$ | $18.32$ | $-2.08$ | $268.034$ | $864.993$ | $18.31$ | $-3.44$ Fe I | $371.993$ | $371.993$ | $20.40$ | $-2.08$ | $385.991$ | $385.991$ | $18.72$ | $-2.26$ | $344.061$ | $344.061$ | $18.49$ | $-2.16$ | $367.991$ | $373.713$ | $18.18$ | $-2.14$ | $296.690$ | $373.486$ | $18.17$ | $-1.33$ | $216.677$ | $349.710$ | $18.13$ | $-2.42$ | $302.064$ | $302.064$ | $18.13$ | $-1.68$ | $261.871$ | $381.584$ | $18.08$ | $-1.21$ | $302.064$ | $382.042$ | $18.07$ | $-1.43$ | $344.061$ | $349.057$ | $18.05$ | $-2.59$ | $213.202$ | $340.746$ | $18.04$ | $-1.62$ | $437.593$ | $646.271$ | $18.04$ | $-3.59$ | $319.166$ | $516.749$ | $18.03$ | $-2.43$ Ca I | $422.673$ | $422.673$ | $19.72$ | $-3.34$ K I | $766.490$ | $766.490$ | $19.05$ | $-3.21$ | $769.896$ | $769.896$ | $18.92$ | $-3.21$ Mg I | $202.582$ | $1710.866$ | $18.15$ | $-1.55$ | $202.582$ | $1182.819$ | $18.15$ | $-1.63$ Quantitative results for single-photon excitation in Na I are shown in Table 2. In Table 3 we list all transitions in Na I that are available with two-step excitation and have a return flux coefficient $\varepsilon>10^{20}$ photons s-1 m-2. There are many, but even the strongest is two times weaker that the sodium D2 line. Of particular interest for PLGS are transitions from the 4P3/2 state [16]. This state may be populated directly from the ground state by absorption of a 330 nm photon. De-excitation occurs through two pathways. Approximately 29% decay directly to the ground state, emitting a 330 nm photon and 70% decay to the 4S1/2 state, emitting a 2206 nm photon. That state can decay to the ground state via two pathways. About 67% decay via the 3P3/2 state, emitting 1140 and 589 nm photons. The remaining decay via the 3P1/2 state emitting 1138 and 590 nm photons. About 1.5% of the atoms in the 4P3/2 state return to the ground state via the 3D5/2 state, emitting 9093, 819 and 589 nm photons. Table 3. Na I transitions for a monochromatic LGS employing two-step excitation having $\lambda\geq 300$ nm and $\log(\varepsilon)>20$. --- $\lambda$ | $\lambda_{\text{ex}}$ | $\log(\varepsilon)$ | $\log(N\sigma_{21})$ (nm) | (nm) | | $819.482$ | 588.995+819.482 | $21.00$ | $-1.15$ $818.326$ | 589.592+818.326 | $20.89$ | $-1.23$ $818.326$ | 588.995+819.479 | $20.68$ | $-1.23$ $589.592$ | 588.995+819.479 | $20.68$ | $-1.50$ $1140.377$ | 589.592+1138.144 | $20.31$ | $-1.19$ $588.995$ | 589.592+1138.144 | $20.31$ | $-1.50$ $1140.377$ | 588.995+1140.377 | $20.22$ | $-1.19$ $568.820$ | 588.995+568.820 | $20.19$ | $-2.26$ $819.479$ | 589.592+818.326 | $20.19$ | $-1.93$ $588.995$ | 589.592+818.326 | $20.19$ | $-1.50$ $588.995$ | 330.237+2337.915 | $20.18$ | $-1.50$ $568.263$ | 589.592+568.263 | $20.10$ | $-2.34$ $568.820$ | 330.237+2337.915 | $20.08$ | $-2.26$ $1138.144$ | 589.592+1138.144 | $20.01$ | $-1.49$ It is also possible to use a 330 nm photon to excite directly to the 4P1/2 state, rather than the 4P3/2 state, with a comparable transition rate. That state decays directly to the ground state with a 29% probability, to the 4S1/2 state with a 70% probability, and to the 3D3/2 state with a 1.6% probability, producing a 9140 nm photon. The 3D3/2 state decays to ground state via the 3P1/2 state, producing 818 and 590 nm photons, or via the 3P3/2 state, producing 819 and 589 nm photons. We searched for all possible PLGS line combinations employing single-photon excitation in Na I, with the condition that they have a return flux greater than $10^{17}$ photons s-1 m-2 (four orders of magnitude weaker than the Na D2 lines) and a penalty factor less than 50. The results are shown in Table 4. The only possibilities are the two already mentioned. Table 4. Na I transitions for a PLGS. All transitions having $q>0.4$ and $\lambda_{\text{ex}}\geq 300$ nm are listed. --- $\lambda_{\text{ex}}$ | $q$ | $p$ | $\lambda_{1}$ | $\log(\varepsilon_{1})$ | $\lambda_{2}$ | $\log(\varepsilon_{2})$ (nm) | | | (nm) | | (nm) | $330.237$ | $1.00$ | $18.59$ | $330.237$ | $19.61$ | $2205.647$ | $20.00$ | $0.94$ | $19.74$ | $330.237$ | $19.61$ | $1140.377$ | $19.82$ | $0.84$ | $19.74$ | $330.237$ | $19.61$ | $1138.144$ | $19.52$ | $0.72$ | $25.86$ | $330.237$ | $19.61$ | $588.995$ | $19.83$ | $0.65$ | $25.84$ | $330.237$ | $19.61$ | $589.592$ | $19.52$ $330.298$ | $0.82$ | $18.60$ | $330.298$ | $19.44$ | $2208.370$ | $19.82$ | $0.77$ | $19.75$ | $330.298$ | $19.44$ | $1140.377$ | $19.65$ | $0.69$ | $19.75$ | $330.298$ | $19.44$ | $1138.144$ | $19.35$ | $0.59$ | $25.87$ | $330.298$ | $19.44$ | $588.995$ | $19.65$ | $0.54$ | $25.85$ | $330.298$ | $19.44$ | $589.592$ | $19.37$ Several transitions involving two-step excitation have been discussed in the literature. The processes of 3S${}_{1/2}\to$ 3P${}_{3/2}\to$ 4D5/2 ($589+569$ nm) [15], 3S${}_{1/2}\to$ 3P${}_{3/2}\to$ 4S1/2 ($589+1140$ nm) [31] and 3S${}_{1/2}\to$ 3P${}_{3/2}\to$ 3D5/2 ($589+819.7$ nm) [32]. Excitation via the $3P_{1/2}$ state is also possible but with somewhat lower transition rates. Atoms excited to the 3P levels can be further excited to the 4D, 4S or 3D levels by absorption of visible or infrared light, provided that this is done within a few ns. Sodium can also be excited from the ground state to the 4D5/2 level by absorption of two 578 nm photons. This is a non-linear process involving a virtual intermediate state, requiring picosecond pulses and a laser irradiance on the order of $10^{10}$ W m-2 [31]. All possible combinations employing two-step excitation in Na I are listed in Table 5. The selection criteria is $q>0.4$. The strongest of these is $589+569$ nm excitation producing lines at 569 and 2334 nm. These transitions would provide a SNR that is 0.79 times the single-photon excitation transitions shown in Table 4. That is roughly equivalent to a decrease to 65% in LGS return flux. Table 5. Na I transitions for a PLGS employing two-step excitation. All transitions that have $q>0.4$ and $\lambda_{\text{ex}}\geq 300$ nm are listed. The last line for each excitation scheme shows the return flux coefficient for the first excitation wavelength. --- $\lambda_{\text{ex}}$ | $q$ | $p$ | $\lambda_{1}$ | $\log(\varepsilon_{1})$ | $\lambda_{2}$ | $\log(\varepsilon_{2})$ (nm) | | | (nm) | | (nm) | 588.995 | $0.79$ | $18.55$ | $330.237$ | $19.41$ | $2337.915$ | $19.95$ \+ 568.820 | $0.79$ | $18.59$ | $330.237$ | $19.41$ | $2205.647$ | $19.79$ | $0.75$ | $19.74$ | $330.237$ | $19.41$ | $1140.377$ | $19.62$ | $0.67$ | $19.74$ | $330.237$ | $19.41$ | $1138.144$ | $19.32$ | $0.55$ | $26.69$ | $330.237$ | $19.41$ | $568.820$ | $20.19$ | $0.51$ | $25.84$ | $330.237$ | $19.41$ | $589.592$ | $19.32$ | $0.45$ | $60.81$ | $568.820$ | $20.19$ | $2337.915$ | $19.95$ | | | $588.995$ | $20.78$ | | 589.592 | $0.71$ | $18.56$ | $330.298$ | $19.32$ | $2334.843$ | $19.86$ \+ 568.263 | $0.71$ | $18.60$ | $330.298$ | $19.32$ | $2208.370$ | $19.70$ | $0.67$ | $19.75$ | $330.298$ | $19.32$ | $1140.377$ | $19.60$ | $0.66$ | $19.75$ | $330.298$ | $19.32$ | $1138.144$ | $19.30$ | $0.59$ | $18.56$ | $330.298$ | $19.32$ | $2337.896$ | $19.16$ | $0.51$ | $25.87$ | $330.298$ | $19.32$ | $588.995$ | $19.82$ | $0.49$ | $18.60$ | $330.298$ | $19.32$ | $2205.647$ | $19.00$ | $0.49$ | $26.71$ | $330.298$ | $19.32$ | $568.819$ | $19.40$ | $0.49$ | $26.73$ | $330.298$ | $19.32$ | $568.263$ | $20.10$ | $0.41$ | $60.69$ | $568.263$ | $20.10$ | $2334.843$ | $19.86$ | | | $589.592$ | $20.65$ | | 588.995 | $0.57$ | $18.56$ | $330.298$ | $19.13$ | $2334.843$ | $19.67$ \+ 568.819 | $0.57$ | $18.60$ | $330.298$ | $19.13$ | $2208.370$ | $19.51$ | $0.54$ | $19.75$ | $330.298$ | $19.13$ | $1140.377$ | $19.41$ | $0.53$ | $19.75$ | $330.298$ | $19.13$ | $1138.144$ | $19.11$ | $0.48$ | $18.56$ | $330.298$ | $19.13$ | $2337.896$ | $18.97$ | $0.41$ | $25.85$ | $330.298$ | $19.13$ | $589.592$ | $19.98$ | $0.40$ | $18.60$ | $330.298$ | $19.13$ | $2205.647$ | $18.81$ | $0.40$ | $26.71$ | $330.298$ | $19.13$ | $568.819$ | $19.21$ | $0.40$ | $26.73$ | $330.298$ | $19.13$ | $568.263$ | $19.91$ | | | $588.995$ | $20.96$ | | 330.237 | $0.51$ | $18.56$ | $330.298$ | $19.03$ | $2334.843$ | $19.57$ \+ 2337.896 | $0.51$ | $18.60$ | $330.298$ | $19.03$ | $2208.370$ | $19.41$ | $0.51$ | $18.60$ | $330.298$ | $19.03$ | $2205.647$ | $19.64$ | $0.48$ | $19.75$ | $330.298$ | $19.03$ | $1140.377$ | $19.66$ | $0.48$ | $19.75$ | $330.298$ | $19.03$ | $1138.144$ | $19.36$ | $0.43$ | $18.56$ | $330.298$ | $19.03$ | $2337.896$ | $18.87$ | | | $330.237$ | $19.25$ | | 588.995 | $0.50$ | $126.58$ | $589.592$ | $20.68$ | $818.326$ | $20.68$ \+ 819.479 | $0.23$ | $126.21$ | $589.592$ | $20.68$ | $819.479$ | $19.99$ | | | $588.995$ | $20.91$ | | 589.592 | $0.50$ | $83.39$ | $588.995$ | $20.31$ | $1140.377$ | $20.31$ \+ 1138.144 | $0.35$ | $83.50$ | $588.995$ | $20.31$ | $1138.144$ | $20.01$ | | | $589.592$ | $20.87$ | | 588.995 | $0.49$ | $18.27$ | $330.237$ | $18.99$ | $5426.880$ | $19.01$ \+ 498.281 | $0.49$ | $18.28$ | $330.237$ | $18.99$ | $5013.920$ | $19.30$ | $0.49$ | $18.59$ | $330.237$ | $18.99$ | $2205.647$ | $19.37$ | $0.47$ | $19.09$ | $330.237$ | $18.99$ | $1477.974$ | $19.49$ | $0.46$ | $19.74$ | $330.237$ | $18.99$ | $1140.377$ | $19.30$ | $0.46$ | $19.74$ | $330.237$ | $18.99$ | $1138.144$ | $18.99$ | | | $588.995$ | $20.78$ | | 589.592 | $0.47$ | $18.37$ | $330.237$ | $18.94$ | $3414.261$ | $19.49$ \+ 615.423 | $0.47$ | $18.37$ | $330.237$ | $18.94$ | $3407.755$ | $19.19$ | $0.46$ | $18.59$ | $330.237$ | $18.94$ | $2208.370$ | $19.03$ | $0.46$ | $18.59$ | $330.237$ | $18.94$ | $2205.647$ | $19.33$ | $0.44$ | $19.74$ | $330.237$ | $18.94$ | $1140.377$ | $19.33$ | $0.44$ | $19.74$ | $330.237$ | $18.94$ | $1138.144$ | $19.03$ | | | $589.592$ | $20.86$ | | 588.995 | $0.44$ | $18.37$ | $330.237$ | $18.88$ | $3414.261$ | $19.42$ \+ 616.075 | $0.44$ | $18.37$ | $330.237$ | $18.88$ | $3407.755$ | $19.12$ | $0.43$ | $18.59$ | $330.237$ | $18.88$ | $2208.370$ | $18.96$ | $0.43$ | $18.59$ | $330.237$ | $18.88$ | $2205.647$ | $19.26$ | $0.41$ | $19.74$ | $330.237$ | $18.88$ | $1140.377$ | $19.26$ | $0.40$ | $19.74$ | $330.237$ | $18.88$ | $1138.144$ | $18.96$ | | | $588.995$ | $21.03$ | | 589.592 | $0.44$ | $18.28$ | $330.298$ | $18.88$ | $5434.170$ | $18.92$ \+ 497.854 | $0.44$ | $18.29$ | $330.298$ | $18.88$ | $5007.670$ | $19.21$ | $0.43$ | $18.60$ | $330.298$ | $18.88$ | $2208.370$ | $19.26$ | $0.42$ | $19.10$ | $330.298$ | $18.88$ | $1476.749$ | $19.40$ | $0.40$ | $19.75$ | $330.298$ | $18.88$ | $1140.377$ | $19.28$ | $0.40$ | $19.75$ | $330.298$ | $18.88$ | $1138.144$ | $18.98$ | | | $589.592$ | $20.66$ | | 330.237 | $0.40$ | $60.81$ | $568.820$ | $20.08$ | $2337.915$ | $19.85$ \+ 2337.915 | | | $330.237$ | $19.26$ | | ### 3.2 Iron The average iron abundances near the density peak at 90 km is two orders of magnitude higher than that of most other mesospheric metals, and twice that of sodium. Although its cross section is small, the key product $N\sigma_{21}$ is still larger than that of all other metallic elements except sodium. The abundance ratio Fe${}^{+}/$Fe is approximately 0.2 above 90 km [33]. Atomic iron has the largest number of spectral lines of all mesospheric species. The iron-group elements have complex transitions with thousands of lines from the vacuum ultraviolet to the infrared. An energy-level diagram for Fe I is shown in Fig. 2. The ground state is split into three fine-structure levels. Only the lowest-energy a5D4 state is populated in thermal equilibrium in the mesosphere. There are five strong visible-wavelength transitions from this ground state, at 344.061, 367.991, 371.993, 382.444 and 385.991 nm. Most of these can decay to the other two ground states. The z5D3 state, reached by the 372 nm transition from the ground state, has a moderately-strong ($A=1.15\times 10^{6}$ s-1) downward transition to the metastable a5F4 state. Figure 2: Selected states and transitions of Fe I. The strongest transition in Fe I using single-photon excitation has a return flux coefficient of $2.51\times 10^{20}$ photons s-1 m-2 (Table 2). This is six times weaker than the sodium D2 line. Two-step excitation (Table 6) provides more possibilities, but these are all weaker by at least an order of magnitude. Fe I has combinations of transitions which have promising wavelengths and penalty factors for PLGS. However, for one-step and two-step excitation, $q$ does not exceed $q>0.18$ for any combination of wavelengths (Tables 7 and 8). Fe II has no transitions from the ground state for wavelengths greater than 260 nm. This rules out excitation by ground-based lasers unless two-photon excitation is used. That would require pulsed lasers with very high irradiance. Table 6. Fe I transitions for a monochromatic LGS employing two-step excitation. All transitions that have $\log(\varepsilon)>19$ and $\lambda_{\text{ex}}\geq 300$ nm are listed. --- $\lambda$ | $\lambda_{\text{ex}}$ | $\log(\varepsilon)$ | $\log(N\sigma_{21})$ (nm) | (nm) | | 355.492 | 437.593+355.492 | $19.32$ | $-1.21$ 314.399 | 385.991+314.399 | $19.31$ | $-1.73$ 381.764 | 371.993+381.764 | $19.29$ | $-2.37$ 380.198 | 371.993+380.198 | $19.08$ | $-2.70$ Table 7. Fe I transitions for a PLGS employing one-step excitation. All transitions that have $q>0.1$ and $\lambda_{\text{ex}}\geq 300$ nm are listed. --- $\lambda_{\text{ex}}$ | $q$ | $p$ | $\lambda_{1}$ | $\log(\varepsilon_{1})$ | $\lambda_{2}$ | $\log(\varepsilon_{2})$ (nm) | | | (nm) | | (nm) | 319.323 | 0.10 | 26.39 | 319.323 | $17.90$ | 517.160 | $17.90$ ### 3.3 Magnesium Magnesium is one of the more abundant meteoritic constituents, 9.6% by mass, so that meteoric ablation should inject large quantities of this metal into the MLT region [22]. Compared to other meteoric metals, magnesium has the largest ionization fraction, with a ratio Mg+/Mg in the range 4 to 12 [22]. Mg I has six strong transitions from the ground state, all in the vacuum ultraviolet. The strongest of these goes to the 31P1 state, with a wavelength of 285.213 nm and a rate ($A_{21}=4.91\times 10^{8}$ s-1) that is an order of magnitude greater than that of the sodium D lines. The only other downward transitions from this upper state are forbidden lines that have negligible rates. In vacuum, this would be a strong candidate for LGS, however, the wavelength is beyond the atmospheric cutoff. Table 8. Fe I transitions for a PLGS employing two-step excitation. All transitions that have a de-excitation $q>0.1$ and $\lambda_{\text{ex}}\geq 300$ nm are listed. In brackets in the last line for each excitation scheme the return flux coefficient for the first excitation wavelength is shown. --- $\lambda_{\text{ex}}$ | $q$ | $p$ | $\lambda_{1}$ | $\log(\varepsilon_{1})$ | $\lambda_{2}$ | $\log(\varepsilon_{2})$ (nm) | | | (nm) | | (nm) | 437.593 | $0.12$ | $25.41$ | $361.207$ | $18.17$ | $973.857$ | $18.01$ \+ 361.207 | $0.11$ | $27.33$ | $371.993$ | $18.85$ | $973.857$ | $18.01$ | | | $437.593$ | $16.06$ | | 385.991 | $0.12$ | $40.78$ | $371.993$ | $18.49$ | $561.564$ | $18.49$ \+ 532.418 | $0.10$ | $44.57$ | $371.993$ | $18.49$ | $532.418$ | $18.38$ | | | $385.991$ | $18.18$ | | Figure 3: Selected states and transitions of Mg II. Mg II has two transitions from the ground state that have rates greater than $10^{8}$ s-1. These go to the 32P1/2 and 32P3/2 states, with wavelengths of 280.271 and 279.553 nm respectively. Both of these upper states return directly to the ground state, with no other significant transitions. Given the high abundance of Mg+ (twice that of Na), these would be very good candidates for LGS, but they are beyond the atmospheric cutoff. Mg II does have strong emission lines in the visible region [34] (Fig.3). However, populating these levels from the ground state requires photons having wavelengths of 124 nm or less. This probably rules out Mg+ for ground-based adaptive optics. ### 3.4 Silicon Silicon is one of the most abundant elements in cosmic dust, comprising around 11% by mass, and meteoric ablation injects a significant amount of Si into the atmosphere [24]. However, neutral Si is oxidized to SiO very rapidly by reaction with O2 at the altitude of maximum ablation, around 90 km. Si+, SiO, and Si(OH)4 are predicted to be the major silicon species above 80 km. Below 97 km the dominant sink is Si(OH)4. The column density of Si+ varies seasonally by an order of magnitude and peaks at $4.0\times 10^{9}$ cm-2 during the summer [24]. Figure 4: Selected states and transitions of Si II. Singly-ionized Silicon has several strong visible-light transitions, as shown in Fig. 4 [35]. However, excitation from the ground state requires UV photons. The ground state is a doublet (32P1/2, 32P3/2), separated by 0.0356 eV. This is comparable to the mean particle energy of $3kT/2\simeq 0.023$ eV, so both states will be populated. Transitions can occur from the lower ground state to the 4P1/2 state, followed by a 637 nm visible-light transition, but this requires a 123-nm photon. The lowest-energy transitions from either ground state to the 4P levels requires a 123-nm photon. There is a transition from the ground state to the 32P3/2 state, but this is quite weak ($A_{21}=2.54\times 10^{6}$ s-1). This likely rules out SI II as a viable candidate for AO, and indeed no transitions met our thresholds for inclusion. ### 3.5 Calcium Lidar measurements of calcium showed a peak density of $2\times 10^{7}$ atoms m-3, about 200 times lower than the typical sodium atom density and 400 times lower than that of iron [36]. The annual average Ca+/Ca column density ratio is 2.4, the second largest ratio after magnesium. In comparison, the ratios of Na+/Na and Fe+/Fe are only about 0.2 [37]. Ca+ is the dominant form above 90 km altitude, peaking near 105 km. Ca I has a strong transition from the ground state to the 31P1 state with a rate of $2.18\times 10^{8}$ s-1 and a wavelength of 422.673 nm. The upper state returns directly to the ground state, with no other transitions. The only problem here is the low abundance. As can be seen in Table 2, the return flux coefficient is about 30 times smaller than that of the sodium D2 line. Table 9. Ca I transitions for a monochromatic LGS employing two-step excitation having wavelength $\lambda\geq 300$ nm and return-flux coefficient $\log(\varepsilon)>19$. --- $\lambda$ | $\lambda_{\text{ex}}$ | $\log(\varepsilon)$ | $\log(N\sigma_{21})$ (nm) | (nm) | | $585.745$ | 422.673+585.745 | 19.07 | -3.43 $430.774$ | 657.278+430.774 | 19.01 | -3.35 Fig. 5 shows the relevant energy levels and transitions of Ca+ [38]. Excitation from the ground state to the 4P1/2 and 4P3/2 levels is possible using 397 nm and 393 nm photons, respectively. These states decay quickly to the ground state and also, with 7% probability, to the 3D3/2 and 3D5/2 states, emitting 866 nm, 850 nm and 854 nm photons. These states are metastable, so most of the atoms excited from the ground state using a UV laser will become trapped in these states. No lines of Ca II met the criteria for inclusion in Table 2. Figure 5: Selected states and transitions of Ca II. ### 3.6 Potassium Potassium has relatively low abundance in the upper atmosphere, having a column density that is about two orders of magnitude less than that of Na. For neutral potassium atoms, many transitions covering the visible light range have high transition probabilities. The relevant energy levels and transitions are shown as Fig. 6. Its persistent lines are 769.896, 766.490, 404.414 and 404.721 nm. Other strong lines having slightly smaller transition probabilities occur at higher energy levels. Direct transitions to and from the ground state are possible at 766.490 and 769.896 nm. These have cross sections that are comparable to that of the sodium D2 line. However, the relatively low column density of potassium results in return flux coefficients that are two orders of magnitude smaller. Figure 6: Selected states and transitions of K I. Figure 7: Energy level diagram of N I (left) and N II (right). Potassium can also be excited directly from the ground state to the 5P1/2 and 5P3/2 levels. These have a path back to the ground state via the 3D states and 4P states, producing strong lines at 3139.265, 3160.163, 1169.024 and 1176.967 nm, in addition to the 766.490 and 769.896 nm doublet. This would be an attractive system for PLGS if it was not for the relatively low abundance. ### 3.7 Nitrogen In the mesosphere, nitrogen molecules are ionized and dissociated by energetic electrons and protons, producing N${}_{2}^{+}$ by direct ionization and N+ by dissociative ionization. Secondary electrons associated with ionization produce N in both the ground and 2D states. At 80 km atomic nitrogen has a peak density of $2\times 10^{8}$ cm-3, falling to $10^{6}$ cm-3 at 70 km [25]. The estimated column density of atomic nitrogen is $\sim 1\times 10^{14}$ cm-2. The most relevant energy levels and transitions of atomic NI and singly- ionized NII are shown in Fig.7. It can be seen that NII has many more strong emission lines in the visible range than does NI. Furthermore, nitrogen- discharge experiments [39] show that the optical emission spectrum in the visible and UV regions, is quite complex with multiple-peaks and many blended lines. Direct transitions from the ground state require vacuum-ultraviolet photons, even using two-photon excitation. This likely rules out nitrogen for ground- based AO systems. ### 3.8 Oxygen Atomic oxygen (O) is a fundamental component in chemical aeronomy of Earth’s MLT region extending from approximately 50 km to over 100 km altitude. Primarily, O is generated through photolysis of molecular oxygen by UV radiation. The peak atomic oxygen concentrations, $\sim 6.0-6.5\times 10^{17}$ m-3, are found at an altitude of approximately 95 km [26]. Assuming an atomic oxygen column width of 10 km in the mesosphere, the estimated column density of atomic oxygen is $\sim 6.0-6.5\times 10^{21}$m-2. Energy level diagrams for O and O+ are shown in Fig. 8. The triplet at $\sim 130$ nm has been observed in the cometary spectra from above the Earth’s atmosphere [40]. These lines arise when O atoms in the ground state are excited to the 5S1 state by solar photons. In the visible spectrum region, several astrophysically-important forbidden transitions exist. For O, these are the red doublet at 630.030 nm and 636.377 nm, arising from transitions between the 2D2 and the 2P2 and the 2P1 levels respectively, and the green line at 557.734 nm from the 2D2 and 2S0 levels. These are magnetic dipole and electric quadrupole transitions respectively. These forbidden oxygen emission lines likely arise from atoms produced directly in the excited 2S0 or 2D2 states by photo-dissociation of parent molecules such as H2O, CO and CO2. In O+, the doublet at 372.602 and 372.882 nm arises from transitions from the 2D3/2 and 2D5/2 levels to the 2S3/2 ground state. As with nitrogen, $N\sigma_{21}$ is high for permitted transitions in oxygen, but low for the forbidden transitions. In O, strong visible lines arise from transitions between the 3P and 3S levels, and also the 5S1 level. Visible lines can also occur from states above the 5S1 energy level. Each line is a doublet or triplet due to hyperfine structure. However, all these lines require vacuum-ultraviolet photons for excitation from the ground state. In O+, strong visible lines result from transitions from the 3P and 3D levels to the 3P levels, and also the 3S1/2 and 3P3/2 levels. In this case there are no direct transitions from the ground state to the upper states involved, and transitions to the lower states require photons of wavelength 54 nm or less. Figure 8: Strong resonance transitions in O I (left) and O II (right). ### 3.9 Hydrogen Atomic hydrogen (H) is also a fundamental component in the photochemistry and energy balance of the mesopause region between approximately 80 and 100 km altitude. H is generated primarily by photolysis of water vapor and participates in a highly-exothermic reaction with ozone. During the day, atomic hydrogen concentrations peak in excess of $2.25\times 10^{14}$ m-3 at higher latitudes near 85 km [27]. The estimated column density of atomic hydrogen is $\sim 2\times 10^{18}$ m-2. Hydrogen and hydrogen-like ions have familiar Lyman, Balmer, and Paschen spectral series. Hydrogen emission spectral lines in the visible range include Balmer lines at 410.173, 434.047, 486.134, and 656.283 nm wavelengths, corresponding to transitions to the $n=2$ level. All energy levels for a given principal quantum number $n$ are degenerate. Transitions from the ground state require vacuum-ultraviolet photons having wavelengths of 122 nm or less. ### 3.10 Carbon Meteoric ablation injects a significant amount of carbon into the mesosphere. However, atomic carbon is a very short-lived species. It is stabilized in various multi-atomic structures having different molecular configurations. The carbon monoxide (CO) mixing ratio increases steeply with altitude in the mesosphere and in the lower part of the thermosphere is a result of the photolysis of carbon dioxide (CO2) [41]. Additionally, as the chemical lifetime of CO is much longer than the characteristic dynamical times, this gas is an excellent tracer of dynamics due to its strong 4.7 $\mu$m transition. The most common oxidation state of carbon in inorganic compounds is $+4$, while $+2$ is found in carbon monoxide. The four outer electrons of carbon are valence electrons, so its ionization energies are much higher than those of most other elements. Short wavelength UV photons are needed to excite the atoms from the ground state, making excitation using a ground-based laser unfeasible. Also, there is little information about the detailed distribution and abundance of carbon and their ions in mesosphere, so carbon and related species are not considered further here. ## 4 Discussion It can be seen from Table 2, that the sodium $D_{2}$ line provides the highest return flux of all possible transitions, including two-step excitation schemes. This is due to its relatively-large natural abundance and large cross section compared to other metal atoms. Mg+ and Si+ have stronger lines, but lack an effective scheme for excitation from the ground state. Although the column density of Fe is twice that of sodium, its lower cross-section results in a return flux that is one to two orders of magnitude smaller. In contrast, the Mg+ lines at 280.271 and 279.553 nm have transition rates that are an order of magnitude greater than that of the sodium D lines, and the column density of Mg+ is twice that of sodium. These would be very strong lines if they were not beyond the atmospheric cutoff. Similarly, Si+ has strong transitions, and a column density comparable to that of Na, but vacuum- ultraviolet excitation is required. Potassium, like Ca+, has a column density that is smaller than that of sodium by two orders of magnitude, so its return flux is small. As with monochromatic LGS, Na yields the highest performance for PLGS from both one-step and two-step excitation. According to our results, the next- ranked element, Fe, provides at most 10% of the SNR of the best-suited transitions in Na. If atmospheric absorption can be compensated by sufficiently-high laser power the one-step excitation at 330 nm yields a higher SNR than the two-step excitation at $569+589$ nm. At an altitude of 4200 m, the atmospheric transmission at 330 nm is about 60% of the transmission at 600 nm. The most-promising combinations of transitions for PLGS include de-excitations at 330 nm, however atmospheric absorption will decrease the return flux at 330 nm. For two-step excitation, an alternative approach could be to combine the return flux from the D1 and D2 line at 589 nm. For an excitation at $589+569$ nm, the photons at 589 nm can only be used if it was ensured that the photons were emitted from fully-coincident laser spots in the mesosphere. Even though the two-step excitation adds wavelengths at about 3410 and 2338 nm, the PLGS performance does not exceed that of the one-step excitation, since the penalty factor does not decrease significantly for infrared wavelengths. Our analysis used statistical weights provided by NIST, which give the number of hyperfine structure states. Magnetic sub-states arising from Zeeman splitting in an external magnetic field are not considered. This is a limitation to our approach. Excitation with optimized laser polarization, which can provide a greater return flux, can be assessed by employing a more- detailed treatment of single transitions using Bloch-equations. Nevertheless, we believe that the simple approach that we employ is useful for a first assessment of potential transitions. For a species in which essentially all atoms are in the ground state, the optical depth at the line center is equal to the product of column density, $\tau_{0}=N\sigma_{12}(\nu_{0})=g_{2}N\sigma_{21}(\nu_{0})/g_{1}$. From the values listed in Table C1, we see that all transitions of mesospheric metals have optical depth less than 0.2, so the atmosphere is optically thin ($\tau<1$). In contrast, the abundant species of N I and O I have $\tau_{0}\gg 1$ for most permitted transitions, which means that the atmosphere is opaque at those wavelengths. The optical depth for stimulated emission is closely related to the optical depth for absorption. In a fully-excited medium, $\tau\sim-\sigma_{21}N$. Significant amplification by stimulated emission can only occur in a medium for which $-\tau\gg 1$. To achieve an optical depth of order unity, $N_{2}A_{21}\geq 10^{23}$ m-2s-1 is required. For a typical transition rate, $A_{21}\sim 10^{7}$ s-1, the column density of the medium must therefore exceed $10^{16}$ m-2 to achieve ASE. This would appear to rule out all metallic species in the mesosphere. An interesting new development is the use of high-powered lasers to create a plasma in the atmosphere. Rairoux et al. [42] have demonstrated ASE in backscattered fluorescence of nitrogen from filaments generated by intense ultra-fast Ti:sapphire laser pulses propagating over a distance of up to several km. As it propagates along the filament, the backscattered radiation is amplified by a population inversion resulting from direct ionization. The laser excitation leads to plasma formation and the consequence is a white- light laser pulse. Backscattered fluorescence from N2 molecules and ions shows the exponential increase with increasing filament length expected for ASE. It has the potential to generate very bright LGS, but a terawatt femtosecond Ti:sapphire laser system is required to produce the $10^{17}$ W/m2 intensity that is necessary to generate non-linear effects and plasma. Nevertheless, the potential of a white-light N2 ASE for adaptive optics might well be realized as laser technology continues to improve. Other factors, in addition to LGS return flux, also affect AO performance. With large-aperture telescopes, LGS appear elongated for subapertures of the pupil that are offset from the axis of the laser launch telescope. This increases noise and reduces the wavefront-sensing accuracy. In this regard, an atomic species that has a high centroid altitude and small vertical extent would be favored. In practice, the altitude is limited by the rapidly decreasing atmospheric density and there is not a large difference between the atomic species listed in Table 1. If an LGS can be created by a high-power pulsed laser, an important consideration will be to limit the extent of the emitting region along the line of sight. ## 5 Summary and conclusions We have reviewed transitions of atoms and ions in the upper atmosphere of interest for AO. Besides Na, which is used extensively by current systems, Fe, Mg, Si, Ca, K all have potential uses. Iron has the highest abundance and the largest number of transitions. However, its return flux is less than that of sodium due to lower cross sections. Our results confirm the general presumption of the LGS community that Fe is not as well-suited as Na for PLGS. However, we find that one-step excitation at 330 nm may be more suitable than two-step excitation at $569+589$ nm. Detailed simulations, using Bloch equations and including the effects of Zeeman splitting, would be a useful next step to evaluate the potential of these excitation schemes. Mg+ and Si+ have high abundance and several very-strong visible-light transitions. However they require vacuum ultraviolet photons for excitation that appear to be implausible even with two-photon excitation. Amplified spontaneous emission using metallic species appears to be unfeasible because of the small optical depths. A high abundance is required for ASE, so nitrogen and oxygen are likely the only atoms that could be used. ## Appendix A Atomic collision rates in the mesosphere and lower thermosphere We wish to estimate the collision rates for various metallic atoms and ions in the mesosphere and lower thermosphere (MLT) region. The atmospheric composition in the MLT is essentially the same as in the troposphere, consisting almost entirely of three species, N2, O2 and Ar. Their fractions, by volume, are approximately 78%, 21% and 1% respectively. The species of interest to us are Na, Fe, Ca, Si, Mg, K, N, O, H and their ions. The collision rate $R$, for a given atom, is given in terms of the collision cross-section Q and the number densities $n$ for the dominant atomic and molecular species. For atom $i$, interacting with field atoms or molecules $j$, $R_{i}=\sum_{j}n_{j}Q_{ij}v_{ij},$ (A1) where $v_{ij}$ is the relative velocity. For a Maxwell-Boltzmann distribution at temperature $T$, the RMS relative velocity of an atom or molecule of mass $m_{i}$ and one of mass $m_{j}$ is, $v_{ij}=\sqrt{v_{i}^{2}+v_{j}^{2}}=\sqrt{\frac{8kT}{\pi\mu_{ij}}},$ (A2) where $\mu_{ij}=m_{i}m_{j}/(m_{i}+m_{j})$ is the reduced mass. The low-energy collision cross section between two atoms is well-approximated by the hard- sphere model, $Q_{ij}=\pi(r_{i}+r_{j})^{2},$ (A3) where $r_{i}$ and $r_{j}$ are the atomic radii. For a diatomic molecule $j$, the situation is more complicated. Modelling the molecule by two spheres of radius $r_{j}$ in contact, the appropriate geometrical cross section is the projected area $A$ of two spheres of radius $r=r_{i}+r_{j}$, having centres separated by a distance $d$. The separation depends on the angle $\theta$ between the axis of the molecule and the velocity vector, $d=2r_{j}\sin\theta$. To find the mean cross-section we average over this angle, assuming an isotropic velocity distribution. From plane geometry, $A=2\pi r^{2}-2r^{2}\arccos\left(\frac{d}{2r}\right)+\frac{d}{2}\sqrt{4r^{2}-d^{2}}.$ (A4) The frequency with which a particular angle $\theta$ appears is proportional to $\sin\theta$, so $\displaystyle Q_{ij}$ $\displaystyle=\frac{4}{\pi}\int_{0}^{\pi/2}A\sin^{2}\theta d\theta,$ $\displaystyle=2\pi r^{2}+\frac{r^{4}+2r^{2}r_{j}^{2}-3r_{j}^{4}}{\pi r_{j}^{2}}\text{arcsinh}\left(\frac{r_{j}}{\sqrt{r^{2}-r_{j}^{2}}}\right)$ $\displaystyle\quad-\frac{r^{3}-3rr_{j}^{2}}{\pi r_{j}}+\frac{32r_{j}r}{3\pi\sqrt{2\pi}}{{}_{3}F_{2}}\left[2,\frac{1}{2},\frac{1}{2};\frac{5}{2},\frac{3}{2};\left(\frac{r_{j}}{r}\right)^{2}\right],$ (A5) where ${}_{3}F_{2}$ is a generalized hypergeometric function. Atomic radii are taken from [43]. For the atmospheric density and temperature the MSIS-E-90 atmospheric model was used [21]. These were evaluated, for a latitude of $30^{\circ}$ ond longitude of $0^{\circ}$, at midnight on the first day of each month, and the average over a full year was computed. The resulting mean cross sections for the atoms and ions of interest, colliding with N2, O2 and Ar, are listed in Table A1. The computed collision rates are listed in Tables A2 and A3. For comparison, Holzlöhner et al. [6] estimated a Na – O2 collision rate of 1/(35 $\mu$s) at an altitude of 92 km, which is within a factor of two of our estimate. Table A1. Mean collision cross sections. --- Atom | mass | radius | cross section | (amu) | (nm) | (nm2) | | | N2 | O2 | Ar Na | 22.990 | 0.191 | 0.565 | 0.550 | 0.277 Fe | 55.845 | 0.127 | 0.315 | 0.303 | 0.171 Mg | 24.305 | 0.162 | 0.443 | 0.429 | 0.226 Si | 28.085 | 0.109 | 0.257 | 0.247 | 0.145 Ca | 40.078 | 0.197 | 0.593 | 0.577 | 0.288 K | 39.098 | 0.237 | 0.790 | 0.771 | 0.370 N | 14.007 | 0.070 | 0.152 | 0.144 | 0.097 O | 15.999 | 0.066 | 0.143 | 0.135 | 0.093 Ar | 39.948 | 0.106 | 0.248 | 0.238 | 0.141 H | 1.008 | 0.070 | 0.152 | 0.144 | 0.097 Table A2. Collision rates, in units of $10^{2}$ s-1, for metallic species. --- altitude | Ca | Fe | K | Mg | Na | Si (km) | | | | | | $75$ | 2449 | 1219 | 3283 | 2062 | 2675 | 1150 $80$ | 1108 | 551.7 | 1486 | 933.3 | 1211 | 520.7 $85$ | 484.1 | 241.0 | 649.1 | 407.7 | 528.8 | 227.5 $90$ | 203.7 | 101.4 | 273.1 | 171.5 | 222.5 | 95.70 $95$ | 82.66 | 41.15 | 110.8 | 69.60 | 90.28 | 38.84 $100$ | 32.65 | 16.26 | 43.77 | 27.49 | 35.65 | 15.34 $105$ | 13.08 | 6.515 | 17.54 | 11.01 | 14.28 | 6.147 $110$ | 5.662 | 2.821 | 7.590 | 4.766 | 6.181 | 2.661 $115$ | 2.740 | 1.365 | 3.673 | 2.306 | 2.991 | 1.288 $120$ | 1.489 | 0.742 | 1.996 | 1.253 | 1.625 | 0.700 Table A3. Collision rates, in units of $10^{2}$ s-1, for non-metallic species. --- altitude | Ar | H | N | O (km) | | | | $75$ | 1023 | 2592 | 833.3 | 748.7 $80$ | 463.0 | 1173 | 377.1 | 338.8 $85$ | 202.3 | 512.3 | 164.7 | 148.0 $90$ | 85.10 | 215.5 | 69.30 | 62.27 $95$ | 34.54 | 87.46 | 28.13 | 25.27 $100$ | 13.64 | 34.54 | 11.11 | 9.982 $105$ | 5.468 | 13.83 | 4.451 | 4.000 $110$ | 2.367 | 5.984 | 1.926 | 1.731 $115$ | 1.146 | 2.895 | 0.932 | 0.838 $120$ | 0.623 | 1.573 | 0.507 | 0.455 ## Appendix B Transition rates in a multi-level atom In a multi-level atom, with states $j$, $j=0,1,\cdots,n-1$, the fraction of atoms $x_{j}$ for which the electron is in state $j$ is determined by $n$ rate equations $\dot{x}_{j}=\Gamma_{jk}x_{k}.$ (B1) Here $\Gamma_{jk}$ specifies the net transition rate between states $j$ and $k$. These equations are not linearly independent because the occupations fractions are related by the constraint $\sum x_{k}=1.$ (B2) This condition can be used to eliminate one of the variables, $x_{0}$ say, which results in a set of $n-1$ linearly-independent inhomogeneous equations $\dot{x}_{j}=M_{jk}x_{k}+b_{j},\quad j=1,2,\cdots,n-1.$ (B3) The quantities $M_{jk}$ and $b_{j}$ depend on transition rates and the radiation energy densities at the transition wavelengths. If the radiation energy densities can be regarded as independent of time, these quantities are constant and the equations are readily solved using standard techniques. The general solution is $\boldsymbol{x}=\sum c_{k}\boldsymbol{a}_{k}e^{\lambda_{k}t}-M^{-1}\boldsymbol{b},$ (B4) where $\lambda_{k}$ and $c_{k}$ are the eigenvalues and eigenvectors of $M$. The steady-state solution is found by equating the left hand side of Eqn (B3) to zero, $\boldsymbol{x}=-M^{-1}\boldsymbol{b},$ (B5) Funding. Yunnan Provincial Department of Education; Natural Sciences and Engineering Research Council of Canada (RGPIN-2019-04369); Chinese Academy of Sciences, CAS President’s International Fellowship Initiative (2017VMA0013). Acknowledgement. We thank Profs. D. Budker and R. Holzlöhner for comments on an earlier version of the manuscript. PH thanks the National Astronomical Observatories, Chinese Academy of Sciences for hospitality during a sabbatical visit. Disclosures. The authors declare no conflict of interests. Data availability. Data underlying the results presented in this paper are available in Refs. [21] and [28]. Selected data for strong transitions in the atomic species discussed here are available from the authors upon request. ## References * [1] J. M. Beckers, Ann. Rev. Astron. Astrophys., 31, 13 (1993). * [2] R. Foy and A. Labeyrie, Astron. Astrophys. 152, L29 (1985). * [3] W. Happer, G. J. MacDonald, C. E. Max and F. J. Dyson, J. Opt. Soc. Am. A, 11, 263 (1994). * [4] J. M. Telle, Opt. Soc. Am. Tech. Dig. Ser. 13, 100 (1996). * [5] T. W. Fan, T. H. Zhou, and Y. Feng, Sci. Rep. 6,19859 (2016). * [6] a R. Holzlöhner, S. M. Rochester, D. B. Calia, D. Budker, J. M. Higbie, and W. Hackenberg, Astron. Astrophys. 510, A20 (2010). * [7] P. W. Milonni, H. Fearn, J. M. Tell and R. Q. Fugate, J. Opt. Soc. Am. A 16, 11 (1999). * [8] R. Rampy, D. Gavel, S. M. Rochester and R. Holzlöhner, J. Opt. Soc. Am. B 32, 2425 (2015). * [9] F. P. Bustos, R. Holzlöhner, S. Rochester, D. B. Calia, J. Hellemeier and D. Budker, J. Opt. Soc. Am. B 37, 1208 (2020). * [10] P. Hickson, J. Hellemeier and R. Yang, Opt. Lett., 46, 1792 (2021). * [11] F. P. Bustos, A. Akulshin, R. Holzlöhner, et al., Proc. SPIE 10703, 107030R-1 (2018). * [12] F. Rigaut and E. Gendron, Astron. Astrophys. 261, 677 (1992). * [13] S. Esposito, Proc. SPIE. 3353, 468 (1998). * [14] R. Ragazzoni, “Laser Guide Star Adaptive Optics for Astronomy,” Ed. N. Ageorges and C. Dainty, NATO ASI Series C, 551, 125 (2000). * [15] R. Foy, A. Migus, F. Biraben, G. Grynberg, P. R. McCullough, and M. Tallon, Astron. Astrophys. Suppl. 111, 569 (1995). * [16] J. P. Pique, I. C. Moldovan, and V. Fesquet, J. Opt. Soc. Am. A 23, 2817 (2006). * [17] Pfrommer, T. and Hickson, P., JOSA A, 27, A97 (2010) * [18] D. S. Davis, P. Hickson, G. Herriot, and C-Y She, Optics Let., 31, 3369 (2006). * [19] C. Y. She, S. Chen, Z. Hu, J. Sherman, J. D. Vance, V. Vasoli, M. A. White, J. Yu, and D. A. Krueger, Geophys. Res. Let. 27, 3289 (2000). * [20] C. S. Gardner, App. Opt. 43, 4941 (2004). * [21] A. E. Hedin, J. Geophys. Res. 96, 1159 (1991), available online at https://ccmc.gsfc.nasa.gov/modelweb * [22] C. L. Whalley, J. C. Gomez-Martin, T. G. Wright, et al., Phys. Chem. Chem. Phys., 13, 6352 (2011). * [23] J. M. C. Plane, Air Force Research Laboratory Report AFRL-AFOSR-UK-TR-2010-0005, 22 pp, (2010). * [24] J. M. C. Plane, J. C. Gomez-Martin, W. Feng, and D. Janches, J. Geophys. Res. Atmos. 121, 3718 (2016). * [25] A. C. Aikin and H. J. P. Smith, Phys. Chem. Earth (C) 25, 203 (2000). * [26] M. G. Mlynczak, L. A. Hunt, J. C. Mast, et al., J. Geophys. Res. Atmos. 118, 5724 (2013). * [27] M. G. Mlynczak, L. A. Hunt, B. T. Marshall, et al., J. Geophys. Res. Atmos. 119, 3516 (2014). * [28] A. Kramida, Y. Ralchenko, J. Reader, J. and NIST ASD Team (2020). NIST Atomic Spectra Database (ver. 5.8), [Online]. Available: https://physics.nist.gov/asd [2020, November 17]. National Institute of Standards and Technology, Gaithersburg, MD. DOI: https://doi.org/10.18434/T4W30F * [29] D. R. Marsh, D. Janches, W. H Feng, and J. M. C. Plane, J. Geophys. Res. Atmos. 118,11442 (2013). * [30] J-P Pique and S. Farinotti, JOSA B 20, 2093 (2003). * [31] J. Biegert and J. C. Diels, Phys. Rev. A 67, 043403 (2003). * [32] C. Xu, C. Guo, H-B. Yu, et al., Appl. Phys. B 123, 94 (2017). * [33] W. H. Feng, D. R. Marsh, M. P. Chipperfield, et al., J. Geophys. Res. Atmos. 118, 9456 (2013). * [34] D. E. Kelleher and L. I. Podobedova J. Phys. Chem. Ref. Data 37, 267 (2008). * [35] D. E. Kelleher and L. I. Podobedova, J. Phys. Chem. Ref. Data 37, 1285 (2008). * [36] J. M. C. Plane, Nature Chem. 3, 900 (2011). * [37] E. Kopp, J. Geophys. Res. 102, 9667 (1997). * [38] C. Maurer, C. Becher, C. Russo, J. Eshner, and R. Blatt, New J. Phys. 6, 94 (2004). * [39] S. Conti, P. I. Porshnev, A. Fridman, et al., Exp. Therm. Fluid Sci. 14, 79 (2001). * [40] A. L. Cochran and W. D. Cochran, Icarus 154, 381 (2001). * [41] K. U. Grossmann, O. Gusev, and P. Knieling, J. Atmos. Solar-Terr. Phys. 68, 1764 (2006). * [42] P. Rairoux, H. Schillinger, S. Niedermeier, et al., Appl. Phys. B 71, 573 (2000). * [43] A. N. Cox, “Allen’s Astrophysical Quantities,” Springer, NY (1999).
# Knowledge Prompts: Injecting World Knowledge into Language Models through Soft Prompts Cicero Nogueira dos Santos, Zhe Dong, Daniel Cer, John Nham, Siamak Shakeri, Jianmo Ni, Yun-hsuan Sung Google Research $\\{$cicerons, zhedong, cer, jnham, siamaks, jianmon<EMAIL_ADDRESS> ###### Abstract Soft prompts have been recently proposed as a tool for adapting large frozen language models (LMs) to new tasks. In this work, we repurpose soft prompts to the task of injecting world knowledge into LMs. We introduce a method to train soft prompts via self-supervised learning on data from knowledge bases. The resulting soft knowledge prompts (KPs) are task independent and work as an external memory of the LMs. We perform qualitative and quantitative experiments and demonstrate that: (1) KPs can effectively model the structure of the training data; (2) KPs can be used to improve the performance of LMs in different knowledge intensive tasks. ## 1 Introduction Very large neural language models (LMs) are known to perform well on knowledge intensive natural language understanding (NLU) tasks, because they memorize a significant amount of world knowledge from the training data. The larger the LM, the more facts it can memorize at the training time, and the better the results at the inference time Roberts et al. (2020). Despite their success, these models also present some important drawbacks such as: the parametric memory of these models have a fixed size and cannot grow (or shrink) over time without fully retraining the model; there is no control in terms of which part of the memory stores data about what; facts that do not co-occur frequently in the training data are not well represented in the model; very large models are needed to memorize enough data in order to perform well on knowledge intensive tasks such as generative question answering; and at last, but not the least, the memorized knowledge gets obsolete overtime, and requires re-training the model for refreshness. In this work, we employ soft prompts to overcome some of these issues of LMs. _Soft prompts_ Lester et al. (2021); Li and Liang (2021); Hambardzumyan et al. (2021) have been recently proposed as a tool for adapting large frozen LMs to new tasks. Nevertheless, we repurpose soft prompts to the task of injecting world knowledge into LMs. The goal is to train an external memory that is composed of a large set of soft prompts that encode world knowledge. We introduce a method to train knowledge driven soft prompts via self-supervised learning on data from knowledge bases. The resulting soft prompts, which we call _knowledge prompts_ (KPs), function as an auxiliary memory of the LMs that is activated when solving knowledge intensive tasks. Different from regular applications of soft prompts that concatenate a fixed small set of embeddings to every input, our approach learns a very large set of KPs, which are sparsely activated depending on the input. We focus on entity-centric KPs, which means that each prompt primarily encodes information about one entity from a knowledge base. We use Wikidata Vrandečić and Krötzsch (2014) triples as our training data and train KPs for the top 1.1M entities, based on the number of triples. We present a qualitative analysis of KPs using t-SNE plots and k-nearest neighbors approaches. In terms of quantitative analysis, we show experimental results for three knowledge intensive tasks: question answering, fact checking and relation classification. For all datasets, the use of KPs improves the performance of the T5 baseline. Our experimental results demonstrate that KPs are an effective way to expand the memory of frozen LMs. Figure 1: Training of Knowledge Prompts: given a serialized KB triple where one of the entities has been masked out, the frozen LM has to predict the masked entity given the input and the knowledge prompt of the non-masked entity, which is _Michelle Obama_ in the example. The cross-entropy loss is computed and the error is back-propagated through the frozen LM in order to update the KP. The main contributions of this work are the following: * • we propose a self-supervised approach to train knowledge driven soft prompts that can be used to inject world knowledge into LMs. * • we demonstrate that knowledge prompts can effectively model the structure of the training data and can also improve the performance of LMs on knowledge intensive tasks. * • this work sheds light on the usability of soft prompts for storing data rather than storing instructions on how to solve specific tasks. ## 2 Methods ### 2.1 Soft Prompts Different approaches have been recently proposed to train soft prompts Lester et al. (2021); Li and Liang (2021); Hambardzumyan et al. (2021). One of the most popular methods, and probably the simplest one, consists of the following steps (Lester et al., 2021): * (1) for a task in the dataset, prepend a fixed number of embeddings (soft prompts) to the word embeddings of every input; * (2) during finetuning, update the soft prompt while keeping all the other parameters of the LM frozen. Despite its simplicity, this approach has demonstrated to be very effective when used with large language models. ### 2.2 Soft Knowledge Prompts We are interested in training _soft knowledge prompts_ (KPs) to encode world knowledge, which could work as an external memory for LMs. In this work, we focus on the training of _entity-centric KPs_ , each of which stores the knowledge related to a specific entity from a knowledge base (KB). In other words, the KP of an entity encodes information from the KB triples that mention the entity either as a subject or an object. We adopt KB triples from Wikidata Vrandečić and Krötzsch (2014), as a simple and trustworthy source of world knowledge. #### 2.2.1 KP Training We train KPs with a masked language modeling (MLM) objective (Devlin et al., 2019; Taylor, 1953), where the goal is to generate the object entity of a KB triple given the subject entity and relation, and vice versa. As an example, the input / target pair "Germany capital <MASK>" / "Berlin" will be used to update the KP for Germany, while the pair "<MASK> capital Berlin" / "Germany" will be used to update the KP for Berlin. The KPs are randomly initialized, and are updated only when the corresponding entities appear (not masked) in the input. This makes the training of KPs sparse and parallelizable. Given an input triple with the object entity being masked, a training iteration has the following steps: * (1) retrieve the KP of the subject entity, which is a simple lookup operation; * (2) concatenate the KP to the sequence of word embeddings of the input text; * (3) predict the object entity name and compute the cross-entropy loss; * (4) back-propagate the error through the frozen LM to the KP, and update the KP using stochastic gradient descent; Figure 1 illustrates our proposed method for training knowledge prompts. Notice that this method is general, and can be used with any textual input as long as the entities of interest are already identified/linked. #### 2.2.2 Using KPs in Downstream Tasks Using KPs during the finetuning of the LM is straightforward. Given the input sequence, e.g. a question, the relevant KPs are retrieved and concatenated to the word embeddings of the input, to generate or predict the answer. At the finetuning time, we freeze the KPs and only update the parameters of the LM, as the KPs are used as pre-trained external knowledge. One can also use task specific soft prompts instead of finetuning the parameters of the LM as in Lester et al. (2021), however in this work we focused on finetuning the LM as we use relatively small LMs. Retrieving KPs that are relevant to the input sequence is crucial for good performance in the downstream task. KPs are useful only if they contain the knowledge that is helpful to solve the input at hand. In this work we employ entity linking as a way to retrieve relevant KPs during training/inference for downstream tasks. Given an input, we first perform entity linking to identify the entities mentioned in the input. Then, we simply do a lookup operation to retrieve the KPs of the identified entities. #### 2.2.3 Why Knowledge Prompts? Some advantages of the proposed approach include: * • It allows a better control of what information is stored, by choosing what examples are used to train the KPs. * • KPs are trained independently, therefore the training can be massively parallelized. * • As the LM is kept frozen during the training of KPs, we do not mess up with the language generation/understanding capabilities of the LM. * • KPs can increase the capacity of small LMs in a dynamic way. We can add/remove KPs at any time. Moreover, if information about a single entity changes, we can update that entity’s KP without changing other KPs. This addresses the freshness issue of the LMs. ## 3 Related Work Most works on soft prompting focus on the problem of learning a set of embeddings that can _reprogram_ a frozen LM to solve a specific task Lester et al. (2021); Li and Liang (2021); Hambardzumyan et al. (2021). Some key aspects that distinct KPs from these approaches include: (1) in KPs, the goal is to create an external memory that stores world knowledge. (2) while traditional soft prompt approaches learn a small set of embeddings our approach learns a very large set of KPs, which are sparsely accessed by the LM. (3) different from regular soft prompts, KPs are not task-specific. Our work is related to the recent body of work on memory augmented LMs. The most related approaches are the Entities as Experts (EaE) model proposed by Févry et al. (2020) and the Fact Injected Language Model (FILM) proposed by Verga et al. (2021). In both papers the authors present methods to train entity embeddings during the pretraining of the LM. During inference time, the model uses a separate classification head to identify entities, which are then used to retrieve entity embeddings that are merged to the current entity representation. de Jong et al. (2022) proposed the TOME model, which employs embeddings in a similar way to EaE but with a much larger granularity. Instead of modeling entities, their method embeds entity mentions, which results in a very large external memory. In their experiments, they use up to 150M entries in the external memory while EaE, FILM and our method use about 1M entries only. Compared to these three models, our proposed approach trains an external LM memory in a quite different way. Our method uses frozen pretrained LMs while those methods train the external memory and the LMs together. Our method concatenates the relevant memory entry to the input as if they were additional word embeddings, while the other three approaches merge the relevant memory entries to the contextual embedding of the identified entities. Additionally, while EaE, FILM and TOME adapt the memory for each new task, our approach uses the exact same external memory for different tasks. This work is also related to multiple recent papers on integrating knowledge bases into LMs Zhang et al. (2019); Peters et al. (2019); Poerner et al. (2020); Sun et al. (2020); Wang et al. (2021); Agarwal et al. (2021). On of the key differences between KPs and all these methods is the use of soft prompts to integrate knowledge bases to frozen LMs. ## 4 Experimental Setup ### 4.1 KP Training Data We adopt Wikidata triples Vrandečić and Krötzsch (2014) as our source of data to train KPs. We start with the set of 45M triples that was previously preprocessed by Agarwal et al. (2021). Next, we filter out triples whose subject entity appears less than 12 times as subject entity in the dataset. This results in a set of 23M triples containing 1.1M distinct subject entities, which form our entity vocabulary and, respectively, the number of KPs in our experiments. ### 4.2 KP Training Setup We adopt the T5.1.1 model family Raffel et al. (2020) and perform experiments with three model sizes: small, base and large, which contain 60M, 220M and 770M parameters, respectively. We use the T5.1.1 checkpoints that Lester et al. (2021) adapted from the original T5.1.1 checkpoints by running an additional 100K training steps using the “LM” objective discussed in Raffel et al. (2020). Just like reported by Lester et al. (2021), we also noticed that these adapted checkpoints make the training of soft prompts easier. Although we use an encoder-decoder LM, our approach is not limited to this type of architecture and can be used with encoder only models like BERT Devlin et al. (2019) or decoder only models like GPT2 Radford et al. (2019). The input length for training KPs is normally short because our examples are masked serialized triples (concatenation of Subject/Object entity and a relation). Therefore, we set the input length to 64, which allows us to use very large batch sizes: between 4K and 8K, depending on the model size. Note that the objective when training KPs is to memorize the training data. Hence, we let KP training run for up to 200 epochs. In the beginning of the training, KPs are initialized by randomly sampling from the word embeddings matrix. This allows KPs to start from a region that is known by the LM, which makes the training smoother and less sensitive to hyperparameters. After training, KPs are kept frozen during LM finetuning for downstream tasks. Therefore, for each model size, the exact same set of KPs is used in our experiments with the different downstream tasks and datasets. ### 4.3 Entity Linking In all experiments where KPs are used, we first preprocess the input text using Google Cloud Natural Language API 111https://cloud.google.com/natural- language/docs/analyzing-entities to perform entity linking. ### 4.4 Downstream Tasks We perform experiments with three different knowledge intensive tasks: (1) question answering (QA), (2) fact checking and (3) relation classification. In terms of datasets, for question Answering experiments we use Entity Questions Sciavolino et al. (2021) and TriviaQA Joshi et al. (2017) datasets. For fact checking we use the FEVER dataset Thorne et al. (2018). For relation classification, we use the TACRED dataset Zhang et al. (2017). In the question answering experiments we follow the closed-book QA (CBQA) setup of Roberts et al. (2020). In this setup, the model has no access to external text, which means that there is no retrieval step and the model has to solve the task using the world knowledge it acquired from pretraining and finetuning data only. During training, we try to use the default hyperparameters as much as possible except for the learning rate, which is finetuned on the development sets. Following previous works we use exact matching (EM) as the evaluation metric in CBQA. ## 5 Results and Discussion Figure 2: t-SNE visualization of 100K KPs trained with T5-Base model on Wikidata. KPs form well separated clusters whose member entities are very similar in terms of their properties and the relations they belong to. . ### 5.1 Qualitative Results We perform a qualitative assessment of KPs through different experiments including t-SNE visualizations, analysis of entity similarity in KP space and evaluation of KPs for QA when golden entity linking is provided. #### 5.1.1 t-SNE visualization of KPs One of the main goals in our qualitative assessment of KPs is to check whether the learned KPs can model the structure of the training data. An approach that can give us some clue about the data structure learning aspect are t-SNE visualizations van der Maaten and Hinton (2008). In Fig. 2, we show a t-SNE visualization of 100K randomly selected KPs that were trained using T5-BASE model on Wikidata triples. We can see in Fig. 2 that KPs form multiple well separated clusters. Zooming in into these clusters we can notice that they are very coherent. There are clusters about companies, celestial objects, movies, locations, etc. This is a good indication that, although trained independently, KPs encapsulate a notion of similarity between entities that aligns well with the KB structure. #### 5.1.2 k-Nearest Neighbors in KP Space We investigate further the quality of the entity similarity captured by KPs using cosine similarity in the KP space to retrieve the k-nearest neighbors of different entities. In Table 1 we show the top four neighbours of four different entities. We present results for T5 models of three different sizes. The top two entities (_Barack Obama_ and _Roger Waters_) are cherry picked popular entities to make it easier for the reader to grasp the quality of the results. The bottom two entities (_Fairmont station_ and _Iacobeni , Sibiu_) were randomly selected. The KP space learned by the three model sizes can produce high quality nearest neighbors for the four different entities. For instance, in the case of the search entity _Fairmont station_ , which is a streetcar stop in Salt Lake City - Utah - USA, all the retrieve entities are also streetcar stops in Salt Lake City. Similar results can be seen for the other entities, where the retrieved neighbors share multiple properties (e.g. same occupation and nationality) with the search entity. Search Entity | Top 4 neighbors in KP space ---|--- | Small | Base | Large Barack Obama | Donald Trump | Bill Clinton | Richard Nixon George W. Bush | Donald Trump | Donald Trump Michelle Obama | Ronald Reagan | Michelle Obama | Jimmy Carter | Jimmy Carter | Harrison Ford Roger Waters | George Harrison | David Gilmour | Tom Waits Pink Floyd | Syd Barrett | Freddie Mercury Kris Kristofferson | Eddie Vedder | David Gilmour | James Hetfield | Brian Wilson | Cliff Edwards Fairmont station | 500 East station | 700 East station | 700 East station 300 East station | 300 East station | Historic Gardner station Sugarmont station | 500 East station | 300 East station | 700 East station | Fairpark station | Decker Lake station Iacobeni , Sibiu | Pălatca | Mica , Cluj | Sic , Cluj Racovita , Timis | Gârbău , Cluj | Gârbău , Cluj Movila , Ialomita | Aiton , Cluj | Mihăileni , Sibiu | Mosna , Sibiu | Râsca , Cluj | Mica , Cluj Table 1: Nearest neighbors in the knowledge prompt space for different search entities and model sizes. The top two entities were cherry picked while the two bottom ones were randomly selected. Model | Simple Questions ---|--- | Zero-shot Learning | Finetuning | no KPs | KPs | no KPs | KPs T5-SMALL | 0.1 | 4.3 | 32.9 | 54.1 T5-BASE | 0.0 | 8.8 | 35.1 | 58.9 T5-LARGE | 0.0 | 8.8 | 36.6 | 58.3 Table 2: Comparing LMs performances on Simple Questions Bordes et al. (2015) with and without KPs as a knowledge source, measured by exact match (EM) score (%). The KPs are retrieved using golden entity linking information. Most performant results for each setup is marked in bold. #### 5.1.3 KPs as a Knowledge Source for LMs In order to assess in a controlled manner whether KPs can be used as a knowledge source for LMs, we perform an experiment on closed-book QA using the Simple Questions dataset Bordes et al. (2015). This dataset is a good fit for our purpose because it contains golden information about the entity involved in the question (golden entity linking). We use the Simple Questions version that was processed by Diefenbach et al. (2017) to align the original Freebase entities to Wikidata entities. We further preprocessed the dataset to contain only questions involving one of the 1.1M entities for which we trained KPs. In Table 2, we present two sets of experiments for models of different sizes. In the first experiment we check whether the use of KPs can improve the performance of the models for zero-shot learning. In this scenario, we can see that T5 models without KPs performs very poorly and achieve exact match (EM) score of 0 percent. The use of KPs boosts the performance of all model sizes, with the base and large models achieving EM of 8.8. In the finetuning scenario, the use of KPs also brings a significant boost in performance. In particular, for T5-Base the improvement is of $\sim$24 points in EM. We know the improvement is actually larger than this because for some questions there are multiple good answers (e.g. songwriter and singer are valid occupations for _John Lennon_), but the Simple Questions dataset list a single answer only. These experimental results indicate that KPs are an effective approach to store information about entities in a way that can be readily used by the LM without any adaptation of the KPs for the downstream QA task. ### 5.2 Quantitative Results Model | Entity Questions | TriviaQA ---|---|--- | Dev | Test | Dev | Test | no KPs | KPs | no KP | KPs | no KPs | KPs | no KP | KPs T5-SMALL | 23.3 | 30.6 | 23.6 | 30.8 | 17.4 | 19.4 | 21.4 | 23.8 T5-BASE | 24.5 | 33.3 | 25.4 | 33.1 | 22.7 | 24.9 | 27.1 | 28.1 T5-LARGE | 26.8 | 33.8 | 26.7 | 33.9 | 28.3 | 28.9 | 32.0 | 32.5 Table 3: Comparing LMs performances on EntityQuestions (Sciavolino et al., 2021) and TriviaQA (Joshi et al., 2017) with and without KPs as a knowledge source, measured by exact match (EM) score (%). #### 5.2.1 Closed-book Question Answering Table 3 presents experimental results for two closed-book QA (CBQA) datasets and different T5 model sizes. KPs provide a significant performance improvement on the Entity Questions dataset, which is an entity-centric QA task. For instance, the improvement for T5-Base is of 7.7 points in EM. Interestingly, T5-SMALL + KPs outperforms T5-LARGE model by a large margin, 30.8 vs 26.7, respectively. Although to a smaller extent, KPs also bring performance improvements for all model sizes on TriviaQA dataset. In Table 4 we compare the performance of T5-Base/Large + KPs with other recently published results on CBQA for TriviaQA. To the best of our knowledge, there is no previous work that reports CBQA results for the Entity Questions dataset. T5 + KPs model does not perform as well as the other CBQA approaches on TriviaQA. We believe this is mainly due to the following factors: (1) in EaE, TOME and T5-11B+SSM, the full LM is heavily pretrained on Wikipedia text using entity focused objective functions, which is known to make the model very effective to QA Guu et al. (2020). In our method, we use an entity focused objective to train KPs only while the LM is frozen. Note in Table 3, no KPs column, that our initial baseline is poor. (2) Models like EaE and TOME update their external memory component during finetuning for a new task. We keep KPs frozen when finetuning for a new task. (3) Our model generates the answer token by token, which is more prone to errors compared to the entity retrieval approaches used in EaE and TOME. (4) We conjecture that the additional pretraining steps performed by Lester et al. (2021) might hurt the performance for CBQA. Model | TriviaQA ---|--- | Dev | Test T5-BASE + KPs | 24.9 | 28.1 T5-LARGE + KPs | 28.9 | 32.5 BERT-base Févry et al. (2020) | 28.9 | EaE Févry et al. (2020) | 43.2 | 53.4 T5-11B+SSM Roberts et al. (2020) | 51.0 | 60.5 TOME-2 de Jong et al. (2022) | 54.6 | 65.8 Table 4: Comparison with other closed-book QA approaches for TriviaQA. #### 5.2.2 Fact Checking We present experimental results for the fact checking task in Table 5. The use of KPs brings significant improvements for the three model sizes on the FEVER dataset. Compared to recent works that use LMs with external memory, our T5-Base + KPs model has similar performance to EaE (Févry et al., 2020), and T5-LARGE + KPs achieves results competitive to TOME-2 (de Jong et al., 2022) model. TOME-2 achieves better results than EaE and T5 + KPs because of the granularity of its memory. While TOME-2 has an external memory with 150M entries that store fine-grained information about entities, both EaE and our model have a memory with about 1M entries only. Our KP training method allows to increase the granularity of KPs in a straightforward manner. For instance, one can use multiple KPs per entity, where each KP is trained using a subset of the triples that mention the entity. We leave the investigation on multiple KPs per entity as a future work. Model | FEVER ---|--- | no KPs | KPs T5-SMALL | 60.2 | 61.4 T5-BASE | 61.3 | 63.4 T5-LARGE | 63.0 | 65.2 EaE (Févry et al., 2020) | 63.6 TOME-2 (de Jong et al., 2022) | 68.1 Table 5: Comparing LMs performances on fact checking dataset, FEVER (test split) (Thorne et al., 2018), with and without KPs as a knowledge source, measured by accuracy (%). Two baselines, Entities-as-Experts (EaE, Févry et al. (2020)) and MentionMemory (TOME-2, de Jong et al. (2022)), are included. #### 5.2.3 Relation Classification Table 6 presents experimental results for the relation classification task using the original TACRED dataset. Following previous papers, we use F1 as the metric and report results for the test set. Similar to the other two tasks, in relation classification KPs also provide performance improvements for all three model sizes. Interestingly, T5-Base + KPs outperform T5-LARGE by almost one point in F1. Model | TACRED ---|--- | no KPs | KPs T5-SMALL | 64.3 | 66.1 T5-BASE | 68.3 | 70.0 T5-LARGE | 69.1 | 69.8 EaE Févry et al. (2020) | 70.2 KnowBERT Peters et al. (2019) | 71.5 Table 6: Experimental results on relation classification. Compared to previous papers that use knowledge augmented approaches, T5-Base + KPs achieves performance similar to EaE and is competitive with KnowBERT. It is important to note that KnowBERT uses entity types as input while both EaE and our method do not use that additional information. ### 5.3 Ablation Experiments #### 5.3.1 KP$\rightarrow$Encoder vs. KP$\rightarrow$Decoder The use of Encoder-Decoder LM gives us the flexibility to introduce KPs at either encoder or decoder. In all experiments presented so far in the paper, KPs are concatenated to the word embeddings of the input sequence and given as input to the encoder (KP$\rightarrow$Encoder). However, one can instead concatenate KPs to the output of the encoder, which makes KPs accessible only by the decoder via cross-attention (KP$\rightarrow$Decoder). In Table 7 we present a comparison of the results of KP$\rightarrow$Encoder and KP$\rightarrow$Decoder for the QA datasets. In both cases, KPs were trained using T5-Base model and we report results for the dev set. We believe KP$\rightarrow$Encoder achieves better performance because it allows interaction (self-attention) between input and KPs in the encoder, which gives the model more opportunity to select and extract the correct information from KPs. On the other hand, the advantage of KP$\rightarrow$Decoder is that its training is faster because it is a simpler optimization problem as the error does not have to be back-propagated through the frozen encoder and KPs are used via cross-attention in the decoder only. KP$\rightarrow$Decoder requires 3x less training iterations to converge compared to KP$\rightarrow$Enc. Dataset | KP$\rightarrow$Enc | KP$\rightarrow$Dec ---|---|--- Simple Questions | 58.9 | 57.7 Entity Questions | 33.3 | 32.8 TriviaQA | 24.9 | 24.2 Table 7: Comparison of models that inject KPs into encoder (KP$\rightarrow$Enc) vs models that inject KPs into decoder (KP$\rightarrow$Dec). #### 5.3.2 Entity Linking vs. Searching in KP Space Retrieving relevant KPs given an input is a fundamental task that has direct impact on the usefulness of KPs. Beyond entity linking, another approach to retrieve KPs is to transform the input into a single dense vector, then search for the most similar vectors in the KP space. We have experimented with this strategy by training an external encoder that creates a vector representation of the input. The external encoder has the same architecture and size of the respective T5 model. We use the TEKGEN dataset Agarwal et al. (2021), which contains Wikipedia sentences mapped to Wikidata triples, as a source of noisy- labeled data to train the encoder via contrastive loss. KPs are kept frozen during the training of the input encoder. Table 8 presents a comparison between T5 + KP models that use either entity linking or search in the KP space. The results were computed on TriviaQA. We can see in the results that entity linking performs better for both model sizes. We conjecture that searching in KP space does not work well because KPs are not optimized to be used in search. KPs are trained to memorize knowledge in a way that can later be extracted by the LM. Model | Ent. Linking | KP Search ---|---|--- T5-SMALL+KPs | 19.4 | 17.9 T5-BASE+KPs | 24.9 | 23.2 Table 8: Evaluation of the LM performances on TriviaQA with KPs retrieved with Entity Linking and Searching in KP space. ## 6 Conclusions We present a new method for training soft prompts that can be used to extend the world knowledge of LMs. We demonstrate the generality and usefulness of the resulting KPs by employing the same set of KP to improve the LM performance in three different tasks: question answering, fact checking, and relation classification. Although in this work we focused on the use of KPs for injecting knowledge into LMs, we believe that entity-centric KPs can be seen as a general purpose knowledge base embedding approach. We leave the investigation of KPs as a general KB embebedding approach for future work. ## References * Agarwal et al. (2021) Oshin Agarwal, Heming Ge, Siamak Shakeri, and Rami Al-Rfou. 2021. Knowledge graph based synthetic corpus generation for knowledge-enhanced language model pre-training. In _Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies_ , pages 3554–3565, Online. Association for Computational Linguistics. * Bordes et al. (2015) Antoine Bordes, Nicolas Usunier, Sumit Chopra, and Jason Weston. 2015. Large-scale simple question answering with memory networks. _CoRR_ , abs/1506.02075. * de Jong et al. (2022) Michiel de Jong, Yury Zemlyanskiy, Nicholas FitzGerald, Fei Sha, and William W. Cohen. 2022. Mention memory: incorporating textual knowledge into transformers through entity mention attention. In _International Conference on Learning Representations_. * Devlin et al. (2019) Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of deep bidirectional transformers for language understanding. In _Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers)_ , pages 4171–4186, Minneapolis, Minnesota. Association for Computational Linguistics. * Diefenbach et al. (2017) Dennis Diefenbach, Thomas Pellissier Tanon, Kamal Deep Singh, and Pierre Maret. 2017\. Question answering benchmarks for wikidata. In _Proceedings of the ISWC 2017 Posters & Demonstrations and Industry Tracks co-located with 16th International Semantic Web Conference (ISWC 2017), Vienna, Austria, October 23rd - to - 25th, 2017._ * Févry et al. (2020) Thibault Févry, Livio Baldini Soares, Nicholas FitzGerald, Eunsol Choi, and Tom Kwiatkowski. 2020. Entities as experts: Sparse memory access with entity supervision. In _Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)_ , pages 4937–4951, Online. Association for Computational Linguistics. * Guu et al. (2020) Kelvin Guu, Kenton Lee, Zora Tung, Panupong Pasupat, and Ming-Wei Chang. 2020. REALM: Retrieval-augmented language model pre-training. _arXiv preprint arXiv:2002.08909_. * Hambardzumyan et al. (2021) Karen Hambardzumyan, Hrant Khachatrian, and Jonathan May. 2021. WARP: Word-level Adversarial ReProgramming. In _Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers)_ , pages 4921–4933, Online. Association for Computational Linguistics. * Joshi et al. (2017) Mandar Joshi, Eunsol Choi, Daniel Weld, and Luke Zettlemoyer. 2017. TriviaQA: A large scale distantly supervised challenge dataset for reading comprehension. In _Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)_ , pages 1601–1611, Vancouver, Canada. Association for Computational Linguistics. * Lester et al. (2021) Brian Lester, Rami Al-Rfou, and Noah Constant. 2021. The power of scale for parameter-efficient prompt tuning. In _Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing_ , pages 3045–3059, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics. * Li and Liang (2021) Xiang Lisa Li and Percy Liang. 2021. Prefix-tuning: Optimizing continuous prompts for generation. In _Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers)_ , pages 4582–4597, Online. Association for Computational Linguistics. * Peters et al. (2019) Matthew E. Peters, Mark Neumann, Robert Logan, Roy Schwartz, Vidur Joshi, Sameer Singh, and Noah A. Smith. 2019. Knowledge enhanced contextual word representations. In _Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)_ , pages 43–54, Hong Kong, China. Association for Computational Linguistics. * Poerner et al. (2020) Nina Poerner, Ulli Waltinger, and Hinrich Schütze. 2020. E-BERT: Efficient-yet-effective entity embeddings for BERT. In _Findings of the Association for Computational Linguistics: EMNLP 2020_ , pages 803–818, Online. Association for Computational Linguistics. * Radford et al. (2019) Alec Radford, Jeff Wu, Rewon Child, David Luan, Dario Amodei, and Ilya Sutskever. 2019. Language models are unsupervised multitask learners. * Raffel et al. (2020) Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, and Peter J. Liu. 2020. Exploring the limits of transfer learning with a unified text-to-text transformer. _Journal of Machine Learning Research_ , 21(140):1–67. * Roberts et al. (2020) Adam Roberts, Colin Raffel, and Noam Shazeer. 2020. How much knowledge can you pack into the parameters of a language model? In _Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)_ , pages 5418–5426, Online. Association for Computational Linguistics. * Sciavolino et al. (2021) Christopher Sciavolino, Zexuan Zhong, Jinhyuk Lee, and Danqi Chen. 2021. Simple entity-centric questions challenge dense retrievers. In _Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing_ , pages 6138–6148, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics. * Sun et al. (2020) Tianxiang Sun, Yunfan Shao, Xipeng Qiu, Qipeng Guo, Yaru Hu, Xuanjing Huang, and Zheng Zhang. 2020. CoLAKE: Contextualized language and knowledge embedding. In _Proceedings of the 28th International Conference on Computational Linguistics_ , pages 3660–3670, Barcelona, Spain (Online). International Committee on Computational Linguistics. * Taylor (1953) Wilson L. Taylor. 1953. “Cloze Procedure”: A new tool for measuring readability. _Journalism & Mass Communication Quarterly_, 30:415 – 433. * Thorne et al. (2018) James Thorne, Andreas Vlachos, Christos Christodoulopoulos, and Arpit Mittal. 2018\. FEVER: a large-scale dataset for fact extraction and VERification. In _Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers)_ , pages 809–819, New Orleans, Louisiana. Association for Computational Linguistics. * van der Maaten and Hinton (2008) Laurens van der Maaten and Geoffrey Hinton. 2008. Visualizing data using t-sne. _Journal of Machine Learning Research_ , 9(86):2579–2605. * Verga et al. (2021) Pat Verga, Haitian Sun, Livio Baldini Soares, and William Cohen. 2021. Adaptable and interpretable neural MemoryOver symbolic knowledge. In _Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies_ , pages 3678–3691, Online. Association for Computational Linguistics. * Vrandečić and Krötzsch (2014) Denny Vrandečić and Markus Krötzsch. 2014. Wikidata: A free collaborative knowledgebase. _Commun. ACM_ , 57(10):78–85. * Wang et al. (2021) Xiaozhi Wang, Tianyu Gao, Zhaocheng Zhu, Zhengyan Zhang, Zhiyuan Liu, Juanzi Li, and Jian Tang. 2021. KEPLER: A unified model for knowledge embedding and pre-trained language representation. _Transactions of the Association for Computational Linguistics_ , 9:176–194. * Zhang et al. (2017) Yuhao Zhang, Victor Zhong, Danqi Chen, Gabor Angeli, and Christopher D. Manning. 2017. Position-aware attention and supervised data improve slot filling. In _Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing_ , pages 35–45, Copenhagen, Denmark. Association for Computational Linguistics. * Zhang et al. (2019) Zhengyan Zhang, Xu Han, Zhiyuan Liu, Xin Jiang, Maosong Sun, and Qun Liu. 2019. ERNIE: Enhanced language representation with informative entities. In _Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics_ , pages 1441–1451, Florence, Italy. Association for Computational Linguistics.
§ INTRODUCTION Verbal autopsy (VA) is a well established approach to ascertain cause-of-death when medical certification and full autopsies are not feasible or practical [Garenne, 2014, Taylor et al., 1983]. After a death is identified, a specially-trained fieldworker interviews the caregivers (usually family members) of the decedent. A typical VA interview includes a set of structured questions with categorical or quantitative responses and a narrative section that records the `story' of the death from the respondent's point of view [World Health Organization and others, 2012]. Currently, there are multiple commonly used questionnaires with overlapping, but not identical questions. The process of inferring a cause from VA data consists of two components. First, there must be some external information about the relationship between causes and symptoms. This is similar to the role of training data in supervised learning tasks. In the context of VA, training dataset of cases with cause of death assigned by in person autopsy is rare, and collecting such data is extremely time and resource-intensive, and requires strong assumptions about the generalizability of deaths in the training set to the population of interest. A more common practice is to obtain training data using clinically trained, experienced physicians read a fraction of the interviews and determine causes. To address the fact that physicians frequently do not agree on causes, VA interviews are often read by two physicians, and sometimes three, and the final causes are determined through a consensus mechanism <cit.>. Another means of obtaining this information is directly through expert opinion. A common practice is to ask groups of physicians to rate the likelihood of a symptoms occurring given a particular cause of death, which can be converted into a set of probabilities of observing a symptom given a particular cause. Such expert knowledge can then be used for modeling. The second component is the algorithmic or statistical method that actually assigns the cause of death by combining the expert or learned cause-symptom relationship with the symptoms observed in the VA interviews from the target population. In all, therefore, there are three key pieces required to analyze VA data: (i) VA survey data itself, (ii) inputs that give information about the association between symptoms and causes, and (iii) a statistical or algorithmic method for assigning a likely cause. The current state of VA literature usually does not distinguish between these three, in part because existing software for algorithmic and statistical methods require a specific set of inputs and survey format. This restriction prevents robust comparison between methods and contexts. A health agency in one region may, for example, want to analyze VA data using the same VA algorithm as a neighboring region to ensure estimates are comparable. Unless the agencies used the same survey format, however, this is not possible with existing tools. The openVA package [Li et al., 2021] addresses this issue through an open-source toolkit that (i) performs data processing and conversion between existing survey formats, (ii) implements multiple currently available algorithmic and statistical methods, and (iii) provides visualization and tools for interpreting §.§ The openVA package The openVA comprises a suite of R [R Core Team, 2020] packages for the analysis of verbal autopsy data. The goal of this package is to provide researchers with an open-sourced tool that supports flexible data input format and allows easier data manipulation and model fitting. The openVA family consists of four core required packages that are on CRAN, InterVA4, InterVA5, InSilicoVA, and Tariff, and an optional package nbc4va. Each of these packages implements one coding algorithm. Three survey formats are supported currently: the WHO 2012 and WHO 2016 instrument after standard dichotomization procedures into binary variables [World Health Organization and others, 2012, 26], the Institute for Health Metrics and Evaluation (IHME) questionnaire in the format of the Population Health Medical Research Consortium (PHMRC) dataset [Murray et al., 2011], and customized binary data with training The main focus of this paper is to provide a general introduction to the implementation details of the included algorithms both in terms of the underlying methodology, and through a series of case studies. For each of the algorithms discussed, there is a standalone R package available on CRAN, and three of them, InterVA-4, InterVA-5 and Tariff, are also available in compiled software program distributed by the original authors. The openVA package has four major contributions: * It provides a standardized and user-friendly interface for analysts to fit and evaluate each method on different types of data input using standard syntax. Previously, most of the methods are designed to be used specifically with their own input formats and are usually incompatible with others. The openVA package closes this gap and allows easier and fair model comparison of multiple algorithms on the same data. This significantly facilitates further research on VA algorithms. * It provides a series of functionalities to summarize and visualize results from multiple algorithms, which is helpful for analysts not familiar with data manipulation and visualization in R. * It does not directly implement any algorithms for coding VA data[A special case is when we extend the InterVA-4 algorithm to the scenario where symptoms and causes are not in the pre-defined set provided by the original InterVA software, the extension is included in the openVA package instead of the InterVA4 and InterVA-5 packages for a faithful replication of the InterVA methods themselves.], so that it is possible for a research group to maintain their own algorithm implementations callable from the openVA package, while also make it available to the general users as a standalone piece of software. For example, the nbc4va was developed and maintained independently by researchers at the Center for Global Health Research in Toronto, but is designed so that it can be seamlessly integrated into the openVA package. * It is fully open-sourced, and can be run on multiple platforms. The open-sourced nature of openVA significantly expands its potential for methodological research and its suitability for integration within a larger data analysis pipeline. Compared to the alternative implementations, the InterVA-4 and InterVA-5 software are distributed as compiled software that can only be run on Windows system. They provide the source codes as an additional code script, which are difficult to modify and re-compile. Tariff, as implemented through the SmartVA-Analyze application [Serina et al., 2015], is also primarily distributed as a compiled application that can only be run on Windows system [Institute for Health Metrics and Evaluation, 2021]. Their source codes was recently made available under the open source MIT License on GitHub [Institute for Health Metrics and Evaluation, 2021]. The rest of this paper is organized as follows: In Section <ref> we briefly introduce the main component packages and the underlying algorithms. We demonstrate model fitting with the openVA package for different input data formats in Section <ref> to <ref>. We first show in Section <ref> how to prepare the input data. We then demonstrate fitting different VA models in Section <ref>, and functionalities to summarize results in Section <ref>. We then discuss how additional information can be incorporated into modeling VA data in Section <ref>. Section <ref> briefly surveys additional packages and software developments built around the openVA package. We end in Section <ref> with a discussion of remaining issues and limitations of the existing automated VA algorithms and proposes new functionalities to be included in openVA package. § STRUCTURE OF THE OPENVA SUITE The openVA suite of packages currently consists of four standalone packages that are available on the CRAN and one optional package hosted on Github. In this section, we first provide a brief introduction to these five packages, and we discuss the mathematical details behind each algorithm in the next subsection. * InterVA4 [Li et al., 2014, Li et al., 2019] is an R package that implements the InterVA-4 model [Byass et al., 2012]. It provides replication of both InterVA software version 4.02 and the later release of version 4.03 update [Byass, 2015]. The standard input of InterVA4 is in the form of a pre-defined set of indicators, based on the 2012 WHO VA instrument (see The default InterVA-4 algorithm cannot be applied to other data input format because its internal built-in prior information is specific to a fixed set of indicators and causes. The same restriction is also maintained in InterVA4 package. However, the mathematical formulation of InterVA-4 model is completely generalizable to other binary input format. The generalized algorithm is described in Section <ref>, and also implemented in the openVA * InterVA5 [Thomas et al., 2021] is an R package that implements the InterVA-5 model developed by Peter Byass The InterVA-5 model updates the previous version in several ways. First, the input data must adhere to the format of the 2016 WHO VA instrument [26]. Second, changes have been made to the data processing steps, which are described in Section <ref>. It is also worth noting that the model outputs have expanded by the inclusion of the most likely Circumstances Of Mortality CATegory, or COMCAT, among the results – the categories include: culture, emergency, health systems, inevitable, knowledge, resources, or an indeterminant combination of multiple factors <cit.>. Despite these changes, the mathematical formulation of InterVA-5 is identical to that of InterVA-4. * InSilicoVA [Li et al., 2021] is an R package that implements the InSilicoVA algorithm, a Bayesian hierarchical framework for cause-of-death assignment and cause-specific mortality fraction estimation proposed in McCormick et al., 2016. It is originally designed to work with the WHO VA instrument, i.e., the same input data as in InterVA software, but is also generalizable to other data input format. It is a fully probabilistic algorithm and could incorporate multiple sources of information, such as known sub-population in the dataset, and physician coding when such data is available. The MCMC sampler is implemented in Java for improved speed. * Tariff [Li et al., 2018] is an R package that implements the Tariff algorithm [James et al., 2011]. It most closely reflects the description of Tariff 2.0 method [Serina et al., 2015]. The Tariff algorithm is developed by the Institute for Health Metrics and Evaluation (IHME) and officially implemented in the SmartVA-Analyze software [Institute for Health Metrics and Evaluation, 2021]. However, as the developers of this R package are not affiliated with the authors of the original algorithm, there are some discrepancies in implementation. The source code of the two versions of Tariff was not publicly available at the time when the Tariff package was created, so the package was developed based solely on the descriptions in the published work. Despite the difference in implementation, Tariff is able to achieve comparable results as the published work as demonstrated in McCormick et al., 2016. More detailed descriptions of the Tariff implementations are also discussed in the supplement of McCormick et al., 2016. The later released Python source codes of SmartVA-Analyze has been incorporated in the web application extension of the openVA package, which we briefly discuss in Section <ref>. * nbc4va [Miasnikof et al., 2015, Wen et al., 2018] is an R package that implements the Naive Bayes Classifier for VA encoding. It calculates the conditional probabilities of symptoms given causes of death from training dataset instead of using physician provided values. nbc4va is developed and maintained by researchers at the Center for Global Health Research in Toronto, but is designed so that it can be seamlessly integrated into openVA. Currently the nbc4va package is hosted on Github and is an optional package that users can choose to load separately. The openVA package is hosted on CRAN and can be installed with the following commands. For the analysis in this paper, we also install the nbc4va package separately from Github. The versions of the supporting packages can be checked in R using the openVA_status() function. R> library(openVA) R> library(devtools) R> install_github("rrwen/nbc4va") R> library(nbc4va) R> openVA_status() §.§ Overview of VA cause-of-death assignment methods The general modeling framework for model VA data consist of first converting the surveys into a series of binary responses to questions about each death, and then the two main goals of a typical VA analysis are to estimate the population cause-specific mortality fractions (CSMF), and the probability distribution or rankings of cause-of-death (COD) for each individual death. In this section, we formally compare the modeling approaches utilized by each algorithm. We adopt the following notations. Consider \(N\) deaths, each with \(S\) binary indicators of symptoms. Let \(s_{ij}\) denote the indicator for the presence of \(j\)-th symptom in the \(i\)-th death, which can take values from 0, 1, or NA (for missing data). We consider a pre-defined set of causes is of size \(C\). For the \(i\)-th death, denote the COD by \(y_i \in \{1, ..., C\}\) and the probability of dying from cause \(k\) is denoted by \(P_{ik}\). For the population, the CSMF of cause \(k\) is denoted as \(\pi_k\), with \(\sum_{k=1}^C \pi_k = 1\). * InterVA4 [Byass et al., 2012] and InterVA5 [Byass, 2018] algorithm calculate the probability of each COD given the observed symptoms using the Bayes rule, so that \[ P_{ik} = \frac{\pi_{k}^{(0)} \prod_{j=1}^S P(s_{ij}=1|y_{i}=k) \mathbf{1}_{s_{ij} = 1}} {\sum_{k' = 1}^C \pi_{k'}^{(0)} \prod_{j=1}^S P(s_{ij}=1|y_{i}=k') \mathbf{1}_{s_{ij} = 1}} \] where both the prior distribution of each causes, $\pi_{k}^{(0)}$, and the conditional probability of each symptom given each cause, $P(s_{ij}=1|y_{i}=k)$, are fixed values provided in the algorithm. The conditional probabilities, $P(s_{ij}=1|y_{i}=k)$, used in InterVA algorithms are represented as rankings with letter grades instead of numerical value [Byass et al., 2012]. For example, $P(s_{ij}=1|y_{i}=k) = A+$ is translated into $P(s_{ij}=1|y_{i}=k) = 0.8$, etc. The standard InterVA software only supports the fixed set of symptoms and causes where such prior information is provided. For a different data input format, this formulation can be easily generalized if training data is available. We include in the openVA package an extension of the algorithm that calculates $\hat P(s_{ij}=1|y_{i}=k)$ from the empirical distribution in the training data and then maps to letter grades with different truncation rules. Details of the extended InterVA algorithm can be found in McCormick et al., 2016. After the individual COD distributions are calculated, InterVA-4 utilizes a series of pre-defined rules to decide up to top three most likely COD assignments and truncates the probabilities for the rest of the CODs to 0 and adds an 'undetermined' category so that the probabilities sum up to 1 (See the user guide of Byass, 2015). Then the population-level CSMFs are calculated as the aggregation of individual COD distribution, such that \[ \pi_k = \sum_{i=1}^N P^*_{ik} \] where $P^*_{ik}$ denotes the individual COD distribution after introducing the undetermined category. * Naive Bayes Classifier [Miasnikof et al., 2015] is very similar to the InterVA algorithm with two major differences. First, instead of considering only symptoms that present, NBC algorithm also considers symptoms that are absent. Second, the conditional probabilities of symptoms given causes are calculated from training data instead of given by physicians, which is similar to our extension of InterVA discussed above. Similar to InterVA, the NBC method can be written as \[ P_{ik} = \frac{\pi_{k}^{(0)} \prod_{j=1}^S (P(s_{ij}=1|y_{i}=k) \mathbf{1}_{s_{ij} = 1} + P(s_{ij} \neq 1|y_{i}=k) \mathbf{1}_{s_{ij} \neq 1})} {\sum_{k' = 1}^C \pi_{k'}^{(0)} \prod_{j=1}^S (P(s_{ij}=1|y_{i}=k') \mathbf{1}_{s_{ij} = 1}+ P(s_{ij} \neq 1|y_{i}=k') \mathbf{1}_{s_{ij} \neq 1})} \] and the CSMFs are calculated by $\pi_k = \sum_{i=1}^N P_{ik}$. * InSilicoVA algorithm [McCormick et al., 2016] assumes a generative model that characterizes both CSMF at the population level, and the COD distributions at the individual level. In short, the core generative process assumes \begin{eqnarray} \nonumber s_{ij} | y_i = k &\propto& \mbox{Bernoulli}(P(s_{ij} | y_i = k)) \\\nonumber y_i | \pi_1, ..., \pi_C &\propto& \mbox{Multinomial}(\pi_1, ..., \pi_C) \\\nonumber \pi_k &=& \exp \theta_k / \sum_{k=1}^C \exp \theta_k \\\nonumber \theta_k &\propto& \mbox{Normal}(\mu, \sigma^2) \end{eqnarray} and hyperpriors are also placed on $P(s_{ij} | y_i = k)$, $\mu$, and $\sigma^2$. The priors for $P(s_{ij} | y_i = k)$ are set by the rankings used in InterVA-4 if the data is prepared into InterVA format, or learned from training data if otherwise. Parameter estimation is performed using Markov Chain Monte Carlo (MCMC), so that a sample of posterior distribution of $\pi_k$ can be obtained after the sampler converges. Additional heuristics of data processing steps introduced since the original publication of the method is discussed in Section <ref> of the Appendix. * Tariff algorithm [James et al., 2011] differs from all other three methods in that it does not calculate an explicit probability distribution of COD for each death. Instead, for each death $i$, a Tariff score is calculated for each COD $k$ so that \[ Score_{ik} = \sum_{j = 1}^{S} \mbox{Tariff}_{kj}\mathbf{1}_{s_{ij}=1} \] where the symptom-specific Tariff score $\mbox{Tariff}_{kj}$ is defined as \[ \mbox{Tariff}_{kj} = \frac{n_{kj} - median(n_{1j}, n_{2j}, ..., n_{Cj})} {IQR(n_{1j}, n_{2j}, ..., n_{Cj})} \] where $n_{kj}$ is the count of how many deaths from cause $k$ contain symptom $j$ in the training data. The scores calculated are then turned into rankings by comparing to a reference distribution of scores calculated from re-sampling the training dataset to have a uniform COD distribution. It is worth noting that the Tariff algorithm produces the COD distribution for each death in terms of their rankings instead of the probability distributions. And thus the CSMF for each cause $k$ is calculated by the fraction of deaths with cause $k$ being the highest ranked cause, i.e., \[ \pi_k = \frac{\sum_{i=1}^N\mathbf{1}_{y_i = k}}{N} \] In addition to the different model specifications underlying each algorithm, two major distinctions are most significant in understanding the four methods for practitioners to interpret the model results. First, the missing indicators are assumed to be equivalent to `absence' in InterVA, NBC, and Tariff, but strictly as `missing' in InSilicoVA. Second, the CSMFs are calculated in three different ways. Tariff calculates CSMF as the proportion of the most likely COD assignments in the dataset. InterVA-4 calculates CSMF as the aggregated distribution of up to the top three most likely CODs. And InSilicoVA estimates CSMF directly from the parametric model using MCMC. Some further feature comparisons are summarized in Table <ref>. Feature InterVA Tariff NBC InSilicoVA Exact replication in the current openVA package Yes No Yes Yes Implementable without training dataset Yes No No Yes Can produce instantaneous results for single death Yes Yes Yes No Accounts for absence of symptoms No No Yes Yes Accounts for missing symptoms No No No Yes Provides individual COD distribution Yes No Yes Yes Direct estimation of CSMF and its uncertainty No No No Yes Comparison the features of the four VA methods implemented in the openVA package. § DATA PREPARATION In the openVA package, we consider two main forms of standardized questionnaire: the WHO instrument and the IHME questionnaire. In this section, we focus on describing these two data formats and tools to clean and convert data into standardized forms for VA algorithms. Pre-processing the raw data collected from the Open Data Toolkit is usually performed with additional packages and software outside of the analysis pipeline in R. We briefly mention software for data preprocessing in Section <ref>. §.§ Standardizing format: WHO For users familiar with InterVA software and the associated data processing steps, the standard input format from the WHO 2012 and 2016 instrument is usually well understood. For the 2012 instrument, the data expected by the InterVA-4 software is organized into a data frame where each row represents one death, and contains \(246\) fields, starting from the first item being the ID of the death. The \(245\) items following the ID each represent one binary variable of symptom/indicator, where `presence' is coded by `Y', and `absence' is coded by an empty cell. The details of this format, as well as the translation from WHO 2012 instrument, could be found at To accommodate updates for the WHO 2016 instrument [26], InterVA-5 software accepts a data frame with \(354\) columns that include \(353\) columns of symptom/indicators followed by an additional column for the record ID. It should be noted that the R package InterVA5 retains the format with the record ID residing in the first column. Another important update with InterVA-5 is that it acknowledge the difference between both “Yes” and “No” (or “Y/y” and “N/n”, which is different from the coding scheme in InterVA-4) are processed as relevant responses, while all other responses are treated as missing values and ignored. With respect to the list of causes of death, InterVA-5 utilizes the WHO 2016 COD categories, which is nearly identical to the WHO 2012 COD categories (used by InterVA-4) except that haemorrhagic fever and dengue fever are two separate categories in the 2016 COD categories. The same input format is inherited by the openVA package, except for one modification. We further distinguish `missing' and `absent' in the input data frame explicitly. This is conceptually easy to understand: knowing that a symptom does not exist provides some information to the possible cause assignment, while a missing symptom does not. Missing data could arise from different stages of the data collection process. Although in theory, most of the VA algorithms could benefit from distinguishing `missing' from `absent', InSilicoVA is the only algorithm that has implemented with missing data. We highly recommend users to pre-process all the raw data so that a `missing' value in the data spreadsheet is coded by a `.' (following the stata practice), and an `absent' is coded by a empty cell, as in the standard InterVA-4 software. For WHO 2016 data, both `.' and `-' (following the default coding scheme of InterVA-5 software) are interpreted as missing values. For methods other than InSilicoVA, the `missing' and `absent' will be considered the same internally and thus will not introduce compatibility problem. §.§ Standardizing format: PHMRC For studies that need to compare with existing work using PHMRC gold standard dataset [Murray et al., 2011], a practical challenge is that although the data is publicly accessible, the pre-processing steps described from the relevant publications are not clear enough nor easy to implement. The openVA package provides functions to clean up the PHMRC gold standard data in the best way we could replicate from First, we allow users to download part or all of the PHMRC gold standard data directly from its on-line repository: R> PHMRC_first1000 <- read.csv(getPHMRC_url("adult"), nrows = 1000) The data can then be cleaned up into a set of dichotomized symptoms, as described in Section 2 of the supplement material of McCormick et al., 2016. This dichotomization process was originally described in Murray et al., 2011. However, following the procedures described in the original paper, we are not able to reproduce the thresholds provided with the on-line supplement materials. Therefore we have included both the thresholds used in Murray et al., 2011 and the thresholds calculated as in their description based on the data that user inputs. See the following codes for the summary of the two different transformation results. R> convert.default <- ConvertData.phmrc(PHMRC_first1000, phmrc.type = "adult", + cutoff = "default", cause = "va34") The first column is site, assign IDs to each row by default 1000 deaths in input. Format: adult 168 binary symptoms generated Number of Yes 21023 Number of No 124644 Number of Not known 22333 R> convert.adapt <- ConvertData.phmrc(PHMRC_first1000, phmrc.type = "adult", + cutoff = "adapt", cause = "va34") The first column is site, assign IDs to each row by default 1000 deaths in input. Format: adult 168 binary symptoms generated Number of Yes 21711 Number of No 123956 Number of Not known 22333 Notice that the original PHMRC data are useful for comparing and validating new methods, as well as for using as training data, but the cleaning functions require only the columns to be exactly as the PHMRC gold standard dataset on-line, so they could also be used for new data that are pre-processed into the same format[However, since calculating the thresholds adaptively requires the knowledge of underlying cause-of-deaths, the latter approach could not be applied to unlabeled datasets.]. §.§ Other formats In addition to the standard formats discussed previously, researchers might also be interested in using customized dichotomous symptoms data in their analysis. The openVA package also supports customized input as long as they are dichotomous. In such case, neither the built-in conditional probability matrix of InterVA nor the PHMRC gold standard dataset could be used to learn the relationship between training and testing data, thus some training data with known causes of death is necessary for all three algorithm. The ConvertData() function can be used to convert data in a different coding scheme into the format used by the openVA package. This is demonstrated in the following small example. R> toyid <- c("d1", "d2") R> toycause <- c("A", "B") R> toyData <- matrix(c("Yes", "No", "Don't know", + "Yes", "Refused to answer", "No"), + byrow = TRUE, nrow = 2, ncol = 3) R> toyData <- cbind(toyid, toycause, toyData) R> colnames(toyData) <- c("ID", "Cause", "S1", "S2", "S3") R> toyData ID Cause S1 S2 S3 [1,] "d1" "A" "Yes" "No" "Don't know" [2,] "d2" "B" "Yes" "Refused to answer" "No" R> toyDataNew <- ConvertData(toyData, yesLabel = "Yes", noLabel = "No", + missLabel = c("Don't know", "Refused to answer")) R> toyDataNew ID Cause S1 S2 S3 1 d1 A Y . 2 d2 B Y . §.§ Convert data between standardized formats The openVA package supports customized set of symptoms to be used as input for all methods, thus it is usually less important to convert data from one format to another, since the conversion inevitably creates loss of information. We recognize the exact mapping of symptoms between different format can be useful for some applications. For example, a full mapping of PHMRC dataset into the InterVA format enables the use of physician provided conditional probabilities included in the InterVA software. This remains as an important feature to be added to the package in the future. § FITTING VA CAUSE-OF-DEATH ASSIGNMENT MODELS In this section, we demonstrate model fitting process in the openVA package using two types of datasets: (1) a random sample of \(1,000\) deaths from ALPHA network without gold standard labels collected with the WHO 2012 instrument, and (2) the adult VA records in the PHMRC gold standard data, with the \(1,554\) records from Andhra Pradesh, India used as testing set and the rest as training dataset. In the first case without gold standard training data, only InterVA and InSilicoVA can be fitted. All four methods can be fitted on the second §.§ Modeling data collected with WHO 2012 questionnaire The randomly sampled VA records from ALPHA network sites are already included in the openVA package and can be loaded directly. R> data(RandomVA1) R> dim(RandomVA1) [1] 1000 246 R> head(RandomVA1[, 1:10]) ID elder midage adult child under5 infant neonate male female 1 d1 Y Y 2 d2 Y Y 3 d3 Y Y 4 d4 Y Y 5 d5 Y Y 6 d6 Y Y The codeVA() function provides a standardized syntax to fit different VA models. Internally, the codeVA() function organizes the input data into standardized formats, checks for incompatibility of the data and specified model, and calls the specified model fitting functions. It returns a classed object of the specified model class. In this example, we use the version 4.03 of InterVA software, which is the latest release of the original software compatible with the WHO 2012 instrument. Any additional model-specific parameters can be passed through the arguments of codeVA(). Here we specify the HIV and malaria prevalence levels required by the InterVA model to be `high'. R> fit_inter_who <- codeVA(data = RandomVA1, data.type = "WHO2012", + model = "InterVA", version = "4.03", + HIV = "h", Malaria = "h") R> summary(fit_inter_who) InterVA-4 fitted on 1000 deaths CSMF calculated using reported causes by InterVA-4 only The remaining probabilities are assigned to 'Undetermined' Top 5 CSMFs: cause likelihood Undetermined 0.156 HIV/AIDS related death 0.122 Stroke 0.072 Reproductive neoplasms MF 0.058 Pulmonary tuberculosis 0.055 We can implement InSilicoVA method with the similar syntax. We use the default parameters and run the MCMC for \(10,000\) iterations. Setting the auto.length argument to FALSE specifies that the algorithm does not automatically increase the length of the chain when convergence failed. In addition to this simple interface, the algorithm implementation in the InSilicoVA packages allows many more types of model customization and extensions that cater to specific practical modeling needs. We will briefly discuss some unique features in Section R> fit_ins_who <- codeVA(RandomVA1, data.type = "WHO2012", model = "InSilicoVA", + Nsim = 10000, auto.length = FALSE) R> summary(fit_ins_who) InSilicoVA Call: 1000 death processed 10000 iterations performed, with first 5000 iterations discarded 250 iterations saved after thinning Fitted with re-estimated conditional probability level table Data consistency check performed as in InterVA4 Top 10 CSMFs: Mean Std.Error Lower Median Upper Other and unspecified infect dis 0.266 0.0168 0.235 0.265 0.301 HIV/AIDS related death 0.102 0.0091 0.085 0.102 0.119 Renal failure 0.101 0.0108 0.084 0.101 0.123 Other and unspecified neoplasms 0.062 0.0089 0.046 0.061 0.080 Other and unspecified cardiac dis 0.058 0.0076 0.044 0.058 0.075 Digestive neoplasms 0.050 0.0077 0.033 0.050 0.065 Acute resp infect incl pneumonia 0.048 0.0073 0.034 0.049 0.063 Pulmonary tuberculosis 0.039 0.0068 0.025 0.039 0.054 Stroke 0.038 0.0061 0.027 0.038 0.052 Other and unspecified NCD 0.034 0.0089 0.018 0.034 0.052 §.§ Modeling the PHMRC data For the PHMRC gold standard dataset, we first load the complete dataset from its on-line repository, and then organize them into training and testing splits. R> PHMRC_all <- read.csv(getPHMRC_url("adult")) R> is.test <- which(PHMRC_all$site == "AP") R> test <- PHMRC_all[is.test, ] R> train <- PHMRC_all[-is.test, ] R> dim(test) \end{CodeInput} \begin{CodeOutput} [1] 1554 946 \end{CodeOutput} \begin{CodeInput} R> dim(train) \end{CodeInput} \begin{CodeOutput} [1] 6287 946 \end{CodeOutput} \end{CodeChunk} In order to fit the models on the PHMRC data, we specify \code{data.type = "PHMRC"}, and the column of the causes of death label in the training data. The rest of the syntax is similar to the previous example. When the input consist of both training and testing data, the InterVA and InSilicoVA algorithm requires a small modification that map the conditional probabilities of symptoms given causes in the training dataset to a letter grade system assumed by the model. In such case, the \code{version} argument for the InterVA algorithm is no longer needed. There are several ways to do this conversion, specified by the \code{convert.type} argument. The \code{convert.type = "quantile"} performs the mapping so that the percentile of each rank stays the same as the original \(P_{s|c}\) matrix in InterVA software. Alternatively we can also use the original fixed values of translation, and assign letter grades closest to each entry in \(\hat{P}_{s|c}\). This conversion is specified by \code{convert.type = "fixed"}, and is more closely aligned to the original InterVA and InSilicoVA setting. Finally, we can also directly use the values in the \(\hat{P}_{s|c}\) without converting them to ranks and re-estimating the values associated with each rank. This can be specified by \code{convert.type = "empirical"}. In this demonstration, we assume the fixed value conversion. \begin{CodeChunk} \begin{CodeInput} R> fit_inter <- codeVA(data = test, data.type = "PHMRC", model = "InterVA", + data.train = train, causes.train = "gs_text34", + phmrc.type = "adult", convert.type = "fixed") \end{CodeInput} \end{CodeChunk} \begin{CodeChunk} \begin{CodeInput} R> fit_ins <- codeVA(data = test, data.type = "PHMRC", model = "InSilicoVA", + data.train = train, causes.train = "gs_text34", + phmrc.type = "adult", convert.type = "fixed", + Nsim=10000, auto.length = FALSE) \end{CodeInput} \end{CodeChunk} The NBC and Tariff method, on the other hand, does not need to perform such conversion, and only need to change the \code{model} argument. \begin{CodeChunk} \begin{CodeInput} R> fit_nbc <- codeVA(data = test, data.type = "PHMRC", model = "NBC", + data.train = train, causes.train = "gs_text34", + phmrc.type = "adult") \end{CodeInput} \end{CodeChunk} \begin{CodeChunk} \begin{CodeInput} R> fit_tariff <- codeVA(data = test, data.type = "PHMRC", model = "Tariff", + data.train = train, causes.train = "gs_text34", + phmrc.type = "adult") \end{CodeInput} \end{CodeChunk} Notice that we do not need to transform the PHMRC data manually as described in the previous section. Data transformations are performed automatically within the \code{codeVA()} function. The arguments in \code{ConvertData.phmrc()} can also be passed into \code{codeVA()}. \section{Summarizing results} \label{sec:results} In this section we demonstrate how to summarize results, extract output, and visualize and compare fitted results. All the fitted object returned by \code{codeVA()} are \code{S3} objects, where readable summary of model results can be obtained with the \code{summary()} function as shown in the previous section. In addition, several other metrics are commonly used to evaluate and compare VA algorithms at either the population or individual levels. In the rest of this section, we show how to easily calculate and visualize some of these metrics with the \pkg{openVA} package. \subsection{CSMF Accuracy} We can extract the CSMFs directly using the \code{getCSMF()} function. The function returns a vector of the point estimates of the CSMFs, or a matrix of posterior summaries of the CSMF for the InSilicoVA algorithm. \begin{CodeChunk} \begin{CodeInput} R> csmf_inter <- getCSMF(fit_inter) R> csmf_ins <- getCSMF(fit_ins) R> csmf_nbc <- getCSMF(fit_nbc) R> csmf_tariff <- getCSMF(fit_tariff) \end{CodeInput} \end{CodeChunk} One commonly used metric to evaluate the CSMF estimates is the so-called CSMF accuracy, defined as \[ CSMF_{acc} = 1 - \frac{\sum_i^C CSMF_i - CSMF_i^{(true)}}{2(1 - \min CSMF^{(true)})} \] We use the empirical distribution in the test data to calculate the true CSMF distribution and evaluate the CSMF accuracy using the \code{getCSMF_accuracy()} function. As discussed previously, the default CSMF calculation is slightly different for different methods. For example, the InterVA algorithm creates the additional category of \texttt{Undetermined} by default, which is not in the true CSMF categories and needs to be specified. The creation of the undetermined category can also be suppressed by \code{interVA.rule = FALSE} in the \code{getCSMF()} function call. For the InSilicoVA algorithm, we will use the posterior mean to calculate the point estimates of the CSMF \begin{CodeChunk} \begin{CodeInput} R> csmf_true <- table(c(test$gs_text34, unique(PHMRC_all$gs_text34))) - 1 R> csmf_true <- csmf_true / sum(csmf_true) R> c(getCSMF_accuracy(csmf_inter, csmf_true, undet = "Undetermined"), + getCSMF_accuracy(csmf_ins[, "Mean"], csmf_true), + getCSMF_accuracy(csmf_nbc, csmf_true), + getCSMF_accuracy(csmf_tariff, csmf_true)) \end{CodeInput} \begin{CodeOutput} [1] 0.58 0.74 0.77 0.68 \end{CodeOutput} \end{CodeChunk} \subsection{Individual COD summary} At each individual level, we can extract the most likely cause-of-death assignment from the fitted object using the \code{getTopCOD()} function. \begin{CodeChunk} \begin{CodeInput} R> cod_inter <- getTopCOD(fit_inter) R> cod_ins <- getTopCOD(fit_ins) R> cod_nbc <- getTopCOD(fit_nbc) R> cod_tariff <- getTopCOD(fit_tariff) \end{CodeInput} \end{CodeChunk} With the most likely COD assignment, other types of metrics based on individual COD assignment accuracy can be similarly constructed by users. The summary methods can also be called for each death ID. For example, using Tariff method, we can extract the fitted rankings of causes for the death with ID 6288 by \begin{CodeChunk} \begin{CodeInput} R> summary(fit_inter, id = "6288") \end{CodeInput} \begin{CodeOutput} InterVA-4 fitted top 5 causes for death ID: 6288 Cause Likelihood Stroke 0.509 Pneumonia 0.318 COPD 0.081 Other Infectious Diseases 0.064 Renal Failure 0.013 \end{CodeOutput} \end{CodeChunk} Notice that the direct call of summary for InSilcoVA does not provide uncertainty estimates for individual COD assignments. This is because the calculation of individual posterior probabilities of COD distribution is relatively time-consuming and memory-intense. To obtain individual-level uncertainty measurement, we can either run the MCMC chain with the additional argument \code{indiv.CI = 0.95} when calling \code{codeVA()}, or update the fitted object directly with the saved posterior draws. \begin{CodeChunk} \begin{CodeInput} R> fit_ins <- updateIndiv(fit_ins, CI = 0.95) R> summary(fit_ins, id = "6288") \end{CodeInput} \begin{CodeOutput} InSilicoVA fitted top causes for death ID: 6288 Credible intervals shown: 95% Mean Lower Median Upper Stroke 0.5043 0.3485 0.5083 0.6361 Pneumonia 0.4116 0.2615 0.4083 0.5834 Other Infectious Diseases 0.0660 0.0411 0.0642 0.0966 Epilepsy 0.0099 0.0064 0.0097 0.0142 COPD 0.0053 0.0031 0.0052 0.0079 Malaria 0.0007 0.0005 0.0007 0.0011 Diabetes 0.0005 0.0003 0.0005 0.0009 Acute Myocardial Infarction 0.0004 0.0003 0.0004 0.0006 Falls 0.0004 0.0001 0.0004 0.0013 Renal Failure 0.0004 0.0002 0.0003 0.0005 \end{CodeOutput} \end{CodeChunk} For \(N\) deaths, \(C\) causes, the posterior mean of individual COD distributions returned by the InSilicoVA model, along with median and with credible intervals can be represented by a \((N \times C \times 4)\)-dimensional array. The function \code{getIndivProb()} extracts this summary in the form of a list of \(4\) matrices of dimension \(N\) by \(C\), which can then be saved to other formats to facilitate further analysis. For other methods, the point estimates of individual COD distribution are returned as the \(N\) by \(C\) matrix. \begin{CodeChunk} \begin{CodeInput} R> fit_prob <- getIndivProb(fit_inter) R> dim(fit_prob) \end{CodeInput} \begin{CodeOutput} [1] 1554 34 \end{CodeOutput} \end{CodeChunk} \subsection{Visualization} The previous sections discuss how results could be extracted and examined in \proglang{R}. In this subsection, we show some visualization tools provided in the \pkg{openVA} package for presenting these results. The fitted CSMFs for the top causes can be easily visualized by the \code{plotVA()} function. The default graph components are specific to each algorithm and individual package implementations, with options for further customization. \begin{CodeChunk} \begin{CodeInput} R> plotVA(fit_inter, title = "InterVA") R> plotVA(fit_ins, title = "InSilicoVA", bw = TRUE) R> plotVA(fit_nbc, title = "NBC") R> plotVA(fit_tariff, title = "Tariff") \end{CodeInput} \begin{figure}[ht] {\centering \includegraphics[width=0.49\linewidth,]{openVA-JSS_files/figure-latex/vis-1-1} \includegraphics[width=0.49\linewidth,]{openVA-JSS_files/figure-latex/vis-1-2} \includegraphics[width=0.49\linewidth,]{openVA-JSS_files/figure-latex/vis-1-3} \includegraphics[width=0.49\linewidth,]{openVA-JSS_files/figure-latex/vis-1-4} \caption[Plot of top 10 CSMFs estimated by InterVA, InSilicoVA, NBC, and Tariff]{Plot of top 10 CSMFs estimated by InterVA, InSilicoVA, NBC, and Tariff.}\label{fig:vis-1} \end{figure} \end{CodeChunk} The CSMFs can also be aggregated for easier visualization of groups of causes. For InterVA-4 cause list, we included an example grouping built into the package, so the aggregated CSMFs can be compared directly. In practice, the grouping of causes of deaths often needs to be determined according to context and the research question of interest. Changing the grouping can be easily achieved by modifying the \code{grouping} argument in \code{stackplotVA()} function. For example, to facilitate the new category of \texttt{Undetermined} returned by InterVA, we first modify the grouping matrix to include it as a new cause: \begin{CodeChunk} \begin{CodeInput} R> data(SampleCategory) R> grouping <- SampleCategory R> grouping[,1] <- as.character(grouping[,1]) R> grouping <- rbind(grouping, c("Undetermined", "Undetermined")) R> compare <- list(InterVA4 = fit_inter_who, + InSilicoVA = fit_ins_who) R> stackplotVA(compare, xlab = "", angle = 0, grouping = grouping) \end{CodeInput} \begin{figure}[H] {\centering \includegraphics[width=.7\linewidth,]{openVA-JSS_files/figure-latex/vis-2-1} \caption[Comparing aggregated CSMF for InterVA-4 and InSilicoVA, adding undetermined category]{Comparing aggregated CSMF for InterVA-4 and InSilicoVA, adding undetermined category.}\label{fig:vis-2} \end{figure} \end{CodeChunk} The ordering of the stacked bars can also be changed to reflect the structures within the aggregated causes. \begin{CodeChunk} \begin{CodeInput} R> order.group <- c("TB/AIDS", "Communicable", "NCD", "External", "Maternal", + "causes specific to infancy", "Undetermined") R> stackplotVA(compare, xlab = "", angle = 0, grouping = grouping, + order.group = order.group) \end{CodeInput} \begin{figure}[H] {\centering \includegraphics[width=.7\linewidth,]{openVA-JSS_files/figure-latex/vis-3-1} \caption[Comparing aggregated CSMF for InterVA-4 and InSilicoVA, with the causes reordered]{Comparing aggregated CSMF for InterVA-4 and InSilicoVA, with the causes reordered.}\label{fig:vis-3} \end{figure} \end{CodeChunk} \section{Incorporating additional information} \label{sec:insilico} Among the VA methods discussed in this paper, the InSilicoVA algorithm [McCormick \emph{et~al.}, 2016] allows more flexible modifications to the Bayesian hierarchical model structure when additional information is available. In this section, we illustrate two features unique to the InSilicoVA method: jointly estimating CSMFs from multiple population, and incorporating partial and potentially noisy physician codings in the \subsection{Sub-population specific CSMFs} In practice researchers may want to estimate and compare CSMFs for different regions, time periods, or demographic groups in the population. Running separate models on subsets of data can be inefficient and does not allow each parameter estimation to borrow information across different groups. The generative framework adopted by InSilicoVA allows the specification of sub-population in analyzing VA data. Consider an input dataset with \(G\) different sub-population, we can estimate different CSMFs \(\pi^{(g)}\) for \(g = 1, ..., G\) for each sub-population, while assuming the same conditional probability matrix, \(P_{s|c}\) and other hyperpriors. As an example, we show how to estimate different CSMFs for sub-populations specified by sex and age groups, using a randomly sampled ALPHA dataset with additional columns specifying the sub-population each death belongs to. \begin{CodeChunk} \begin{CodeInput} R> data(RandomVA2) R> head(RandomVA2[, 244:248]) \end{CodeInput} \begin{CodeOutput} stradm smobph scosts sex age 1 . . . Men 60+ 2 . . . Women 60- 3 . . . Women 60- 4 . . . Women 60+ 5 . . . Women 60- 6 . . . Women 60- \end{CodeOutput} \end{CodeChunk} Then we can fit the model with one or multiple additional columns specifying sub-population membership for each observation. \begin{CodeChunk} \begin{CodeInput} R> fit_sub <- codeVA(RandomVA2, model = "InSilicoVA", + subpop = list("sex", "age"), indiv.CI = 0.95, + Nsim = 10000, auto.length = FALSE) \end{CodeInput} \end{CodeChunk} Functions discussed in the previous sections works in the same way for the fitted object with multiple sub-populations. Additional visualization tools are also available. By specify \code{type = "compare"}, we can plot the CSMFs for two sub-populations on the same plot. \begin{CodeChunk} \begin{CodeInput} R> plotVA(fit_sub, type = "compare", title = "Comparing CSMFs", top = 3) \end{CodeInput} \begin{figure}[H] {\centering \includegraphics[width=0.9\linewidth,]{openVA-JSS_files/figure-latex/ins3-1} \caption[Comparing CSMF for different sub-population]{Comparing CSMF for different sub-population.}\label{fig:ins3} \end{figure} \end{CodeChunk} By default, the comparison plots will select all the CODs that are included in the top \(K\) for each of the sub-populations. We can also plot only subsets of them by specifying with causes are of interest. \begin{CodeChunk} \begin{CodeInput} R> plotVA(fit_sub, type = "compare", title = "Comparing CSMFs", + causelist = c("HIV/AIDS related death", + "Pulmonary tuberculosis", + "Other and unspecified infect dis", + "Other and unspecified NCD")) \end{CodeInput} \begin{figure}[H] {\centering \includegraphics[width=0.8\linewidth,]{openVA-JSS_files/figure-latex/ins4-1} \caption[ Comparing the fraction of deaths due to selected CODs for different sub-populations]{ Comparing the fraction of deaths due to selected CODs for different sub-populations.}\label{fig:ins4} \end{figure} \end{CodeChunk} If a single sub-population is selected by the \code{which.sub} argument, the top CSMFs of the chosen sub-population will be plotted in the same way as before. \begin{CodeChunk} \begin{CodeInput} R> plotVA(fit_sub, which.sub = "Women 60-", title = "Women 60-") \end{CodeInput} \begin{figure}[H] {\centering \includegraphics[width=0.8\linewidth,]{openVA-JSS_files/figure-latex/ins5-1} \caption[Top 10 CSMF for a specified sub-population]{Top 10 CSMF for a specified sub-population.}\label{fig:ins5} \end{figure} \end{CodeChunk} The \code{stackplot()} function can also be used to compare different sub-populations in aggregated cause groups. \begin{CodeChunk} \begin{CodeInput} R> stackplot(fit_sub) R> stackplot(fit_sub, type = "dodge") \end{CodeInput} \begin{figure}[H] {\centering \includegraphics[width=0.49\linewidth,]{openVA-JSS_files/figure-latex/ins6-1} \includegraphics[width=0.49\linewidth,]{openVA-JSS_files/figure-latex/ins6-2} \caption[Comparing aggregated CSMF for four different sub-population]{Comparing aggregated CSMF for four different sub-population.}\label{fig:ins6} \end{figure} \end{CodeChunk} \subsection{Physician coding} When physician coded cause of death is available for all or a subset of the death, we can incorporate such information in the InSilicoVA model. The physician coded causes can be either the same as the CODs used for the final algorithm, or a higher level aggregation of them. When there are more than one physician codes for each death and the physician identity is known, we can first de-bias the multiple codes provided from different physicians using the process described in McCormick \emph{et~al.}, 2016. For the purpose of implementation, we only need to specify which columns are physician IDs, and which are their coded causes respectively. \begin{CodeChunk} \begin{CodeInput} R> data(SampleCategory) R> data(RandomPhysician) R> head(RandomPhysician[, 245:250]) \end{CodeInput} \begin{CodeOutput} smobph scosts code1 rev1 code2 rev2 1 . . NCD doc9 NCD doc6 2 . . NCD doc4 NCD doc3 3 . . NCD doc1 NCD doc5 4 . . TB/AIDS doc4 TB/AIDS doc7 5 . . TB/AIDS doc5 TB/AIDS doc9 6 . . Communicable doc9 Communicable <NA> \end{CodeOutput} \end{CodeChunk} \begin{CodeChunk} \begin{CodeInput} R> doctors <- paste0("doc", c(1:15)) R> causelist <- c("Communicable", "TB/AIDS", "Maternal", + "NCD", "External", "Unknown") R> phydebias <- physician_debias(RandomPhysician, + phy.id = c("rev1", "rev2"), phy.code = c("code1", "code2"), + phylist = doctors, causelist = causelist, tol = 0.0001, max.itr = 100) \end{CodeInput} \end{CodeChunk} The de-biased step essentially creates a prior probability distribution for each death over the broader categories of causes. Then to run InSilicoVA with the de-biased physician coding, we can simply pass the fitted object from the previous step to the model. Additional arguments are needed to specify the external cause category, since they are handled by separated heuristics, and the unknown category, which is equivalent to an uniform probability distribution over all other categories, i.e., the same as the case where no physician coding exist. \begin{CodeChunk} \begin{CodeInput} R> fit_ins_phy <- codeVA(RandomVA1, model = "InSilicoVA", + phy.debias = phydebias, phy.cat = SampleCategory, + phy.external = "External", phy.unknown = "Unknown", + Nsim = 10000, auto.length = FALSE) \end{CodeInput} \end{CodeChunk} This can be compared with the previous results without including physicians codes. \begin{CodeChunk} \begin{CodeInput} R> plotVA(fit_ins_who, title = "Without physician coding") R> plotVA(fit_ins_phy, title = "With physician coding") \end{CodeInput} \begin{figure}[!h] {\centering \includegraphics[width=0.7\linewidth,]{openVA-JSS_files/figure-latex/phy3-1} \includegraphics[width=0.7\linewidth,]{openVA-JSS_files/figure-latex/phy3-2} \caption[Comparing fitted CSMF with and without physicians]{Comparing fitted CSMF with and without physicians}\label{fig:phy3} \end{figure} \end{CodeChunk} \section{Other related software packages} \label{sec:other} Since the release of the \pkg{openVA} package on CRAN, there have been many new developments in both methodology and software implementations that build on the \pkg{openVA} suite of packages and further extended its functionalities. Here we briefly survey some of the related methods, software packages, and ongoing work in the VA community that are related to \pkg{openVA}. First, the ability to easily fit and compare existing methods using the \pkg{openVA} package facilitated the development of several new VA methods in the last several years. Most of the development focuses on building statistical models to relax the conditional independence assumption of symptoms given a cause of death \citep[e.g.,][]{li2017mix, tsuyoshi2017, moran2021bayesian}. These methods tend to be computationally more demanding compared to the algorithms currently included in the \pkg{openVA} package, but usually provide improved inference. It is future work to include some of these latest developments to the \pkg{openVA} package for routine use. Most of these methods have publicly available implementations, such as the \pkg{farva} package [Moran, 2020] on Github. In another direction of research, a class of transfer learning methods focuses on building models to correct for bias in existing methods when predicting out of domain. These methods take the predicted cause of death assignments and distributions obtained with the \pkg{openVA} package and learn an ensemble of the predictions calibrated to the target population [Datta \emph{et~al.}, 2020, Fiksel \emph{et~al.}, 2021]. The \pkg{calibratedVA} package [Fiksel and Datta, 2018] is available to implement these models. Outside of research community that develops new VA algorithms, \pkg{openVA} has also been used extensively by governments and health care organizations, particularly in locations that lack a strong vital registration system and use VA data to identify the leading causes of death. To facilitate the work of these users, \pkg{openVA} has been wrapped into a web application, the openVA App [Thomas, 2021], using the \pkg{shiny} package. The open source openVA App is available on GitHub and provides an intuitive interface to \pkg{openVA} that does not require one to learn \proglang{R}, but provides visual and tabular output produced by the different VA algorithms. It also runs the official Tariff algorithm by calling the \proglang{Python} source code of SmartVA-Analyze [Institute for Health Metrics and Evaluation}, 2021] and processing the output to be consistent with the other algorithms. The \pkg{openVA} \proglang{R} package has also been a key component in a larger data analysis pipeline that pulls VA data from an Open Data Kit (ODK) Collect server and deposits the assigned causes of death to another server running the District Health Information Software 2 (DHIS2), which is a common information management system used in low and middle-income countries. This open source software is implemented as a \proglang{Python} package, \pkg{openva-pipeline}, and is available on GitHub and the Python Package Index [Thomas \emph{et~al.}, 2021]. Finally, the \proglang{R} package \pkg{CrossVA} [Thomas \emph{et~al.}, 2020] and the \proglang{Python} package \pkg{pyCrossVA} [Choi and Thomas, 2021] provide additional toolkit to convert raw VA data from its original format from the ODK server to the standardized formats discussed before. Both packages are open source and available on GitHub. The \pkg{pyCrossVA} package is also available on the Python Package Index. \section{Conclusion} \label{sec:conclusion} In this paper, we introduce the \pkg{openVA} package. This is the first open-sourced software that implements and compares the major VA methods. The \pkg{openVA} package allows researchers to easily access the most recent tools that are previously difficult to work with or unavailable. It also enables the compatibility of multiple data input formats and significantly reduces the tedious work to pre-process different data formats specific to each algorithm. The software framework of the \pkg{openVA} package allows the integration of new methods in the future. The \pkg{openVA} package makes all the steps involved in analyzing VA data, i.e., data processing, model tunning and fitting, result summarization, and evaluation metrics, transparent and reproducible. This contributes significantly to the public health community using VA. Finally, we note that several features that will be helpful for future development. First, many users of the \pkg{openVA} package may not be familiar with the command line tools or does not have access to \proglang{R} on their local machines. A well designed graphical user interface can be very useful in such settings. The work on the shiny web application, openVA app, is a first step towards making the package more accessible. The authors intend to extend it to a better front end hosted on secure centralized servers. Second, although in this paper, we aim to provide users with all the available methods for assigning causes of death, without comparing the accuracy and robustness between them, much future work is needed to systematically assess, compare, and combine these methods in a better analysis framework. Finally, the development of VA algorithms is still an active area of research and it would be possible to extend the \pkg{openVA} suite to incorporate better VA algorithms and new types of data such as free-text narratives. \section*{Acknowledgement} This work was supported by grants K01HD078452, R01HD086227, and R21HD095451 from the Eunice Kennedy Shriver National Institute of Child Health and Human Development (NICHD). \begin{thebibliography}{35} \newcommand{\enquote}[1]{``#1''} \providecommand{\natexlab}[1]{#1} \providecommand{\url}[1]{\texttt{#1}} \providecommand{\urlprefix}{URL } \expandafter\ifx\csname urlstyle\endcsname\relax \providecommand{\doi}[1]{doi:\discretionary{}{}{}#1}\else \providecommand{\doi}{doi:\discretionary{}{}{}\begingroup \urlstyle{rm}\Url}\fi \providecommand{\eprint}[2][]{\url{#2}} [Byass, 2015] Byass P (2015). \newblock \enquote{{InterVA-4 Software [Windows Executable]}.} \newblock \emph{www.interva.net}. [Byass, 2018] Byass P (2018). \newblock \enquote{{InterVA-5 Software [Windows Executable]}.} \newblock \emph{www.interva.net}. [Byass \emph{et~al.}, 2012] Byass P, Chandramohan D, Clark SJ, D'Ambruoso L, Fottrell E, Graham WJ, Herbst AJ, Hodgson A, Hounton S, Kahn K, \emph{et~al.} (2012). \newblock \enquote{Strengthening Standardised Interpretation of Verbal Autopsy Data: The New {InterVA-4} Tool.} \newblock \emph{Global Health Action}, \textbf{5}. [Choi and Thomas, 2021] Choi E, Thomas J (2021). \newblock \enquote{Prepare data from WHO and PHRMC instruments for verbal autopsy algorithms.} \newblock \url{https://pypi.org/project/pycrossva/}; Accessed July 21, 2021. [5] D'Ambruoso L, Boerma T, Byass P, Fottrell E, Herbst K, Kallander K, Mullan Z \newblock \enquote{{The case for verbal autopsy in health systems \newblock \emph{LANCET GLOBAL HEALTH}, \textbf{{5}}({1}), {E20--E21}. \newblock ISSN {2214-109X}. [Datta \emph{et~al.}, 2020] Datta A, Fiksel J, Amouzou A, Zeger SL (2020). \newblock \enquote{Regularized Bayesian transfer learning for population-level etiological distributions.} \newblock \emph{Biostatistics}, p. ISSN 1465–4644. [Fiksel and Datta, 2018] Fiksel J, Datta A (2018). \newblock \emph{calibratedVA: Locally calibrated cause specific mortality fractions using verbal autopsy data}. \newblock \urlprefix\url{https://github.com/jfiksel/CalibratedVA}. [Fiksel \emph{et~al.}, 2021] Fiksel J, Datta A, Amouzou A, Zeger S (2021). \newblock \enquote{Generalized {Bayesian} Quantification Learning.} \newblock \emph{Journal of the American Statistical Association}, pp. 1--19. [Garenne, 2014] Garenne M (2014). \newblock \enquote{Prospects for automated diagnosis of verbal autopsies.} \newblock \emph{BMC Medicine}, \textbf{12}(1), 18. [Institute for Health Metrics and Evaluation}, 2021] {Institute for Health Metrics and Evaluation} (2021{\natexlab{a}}). \newblock \enquote{{SmartVA-Analyze} Desktop Application.} \newblock \url{https://github.com/ihmeuw/SmartVA-Analyze}; Accessed July 21, [Institute for Health Metrics and Evaluation}, 2021] {Institute for Health Metrics and Evaluation} (2021{\natexlab{b}}). \newblock \enquote{Verbal Autopsy tools.} \newblock \url{http://www.healthdata.org/verbal-autopsy/tools}; Accessed July 21, 2021. [James \emph{et~al.}, 2011] James SL, Flaxman AD, Murray CJ, {Consortium Population Health Metrics Research} (2011). \newblock \enquote{Performance of the Tariff Method: validation of a simple additive algorithm for analysis of verbal autopsies.} \newblock \emph{Population Health Metrics}, \textbf{9}(31). [Kahn \emph{et~al.}, 2012] Kahn K, Collinson MA, Gómez-Olivé FX, Mokoena O, Twine R, Mee P, Afolabi SA, Clark BD, Kabudula CW, Khosa A, Khoza S, Shabangu MG, Silaule B, Tibane JB, Wagner RG, Garenne ML, Clark SJ, Tollman SM (2012). \newblock \enquote{Profile: {Agincourt} Health and Socio-Demographic Surveillance System.} \newblock \emph{International Journal of Epidemiology}, \textbf{41}(4), [Kunihama \emph{et~al.}, 2020] Kunihama T, Li ZR, Clark SJ, McCormick TH (2020). \newblock \enquote{{Bayesian} factor models for probabilistic cause of death assessment with verbal autopsies.} \newblock \emph{The Annals of Applied Statistics}, \textbf{14}(1), 241--256. [Li \emph{et~al.}, 2021] Li ZR, McCormick T, Clark S (2021{\natexlab{a}}). \newblock \emph{\pkg{openVA}: Automated Method for Verbal Autopsy}. \newblock R package version 1.0.12, \urlprefix\url{https://cran.r-project.org/web/packages/openVA/index.html}. [Li \emph{et~al.}, 2014] Li ZR, McCormick TH, Clark SJ (2014). \newblock \enquote{InterVA4: An {R} package to analyze verbal autopsy data.} \newblock \emph{Center for Statistics and the Social Sciences Working Paper, [Li \emph{et~al.}, 2018] Li ZR, McCormick TH, Clark SJ (2018). \newblock \emph{\pkg{Tariff}: Replicate Tariff Method for Verbal Autopsy}. \newblock \proglang{R} package version 1.0.5, \urlprefix\url{http://CRAN.R-project.org/package=Tariff}. [Li \emph{et~al.}, 2020] Li ZR, McCormick TH, Clark SJ (2020). \newblock \enquote{Using {Bayesian} latent {Gaussian} graphical models to infer symptom associations in verbal autopsies.} \newblock \emph{Bayesian Analysis}, \textbf{15}(3), 781. [Li \emph{et~al.}, 2021] Li ZR, McCormick TH, Clark SJ (2021{\natexlab{b}}). \newblock \emph{\pkg{InSilicoVA}: Probabilistic Verbal Autopsy Coding with 'InSilicoVA' Algorithm}. \newblock \proglang{R} package version 1.3.4, \urlprefix\url{http://CRAN.R-project.org/package=InSilicoVA}. [Li \emph{et~al.}, 2019] Li ZR, McCormick TH, Clark SJ, Byass P (2019). \newblock \emph{\pkg{InterVA4}: Replicate and Analyse 'InterVA4'}. \newblock \proglang{R} package version 1.7.6, \urlprefix\url{http://CRAN.R-project.org/package=InterVA4}. [McCormick \emph{et~al.}, 2016] McCormick TH, Li ZR, Calvert C, Crampin AC, Kahn K, Clark SJ (2016). \newblock \enquote{Probabilistic Cause-of-Death Assignment Using Verbal \newblock \emph{Journal of the American Statistical Association}, \textbf{111}(515), 1036--1049. [Miasnikof \emph{et~al.}, 2015] Miasnikof P, Giannakeas V, Gomes M, Aleksandrowicz L, Shestopaloff AY, Alam D, Tollman S, Samarikhalaj A, Jha P (2015). \newblock \enquote{Naive Bayes classifiers for verbal autopsies: comparison to physician-based classification for 21,000 child and adult deaths.} \newblock \emph{BMC Medicine}, \textbf{13}(1), 1. [Moran, 2020] Moran KR (2020). \newblock \enquote{FActor Regression for Verbal Autopsy.} \newblock \url{https://github.com/kelrenmor/farva}; Accessed July 21, 2021. [Moran \emph{et~al.}, 2021] Moran KR, Turner EL, Dunson D, Herring AH (2021). \newblock \enquote{Bayesian hierarchical factor regression models to infer cause of death from verbal autopsy data.} \newblock \emph{Journal of the Royal Statistical Society: Series C (Applied [Murray \emph{et~al.}, 2011] Murray CJ, Lopez AD, Black R, Ahuja R, Ali SM, Baqui A, Dandona L, Dantzer E, Das V, Dhingra U, \emph{et~al.} (2011). \newblock \enquote{Population Health Metrics Research Consortium gold standard verbal autopsy validation study: design, implementation, and development of analysis datasets.} \newblock \emph{Population Health Metrics}, \textbf{9}(1), 27. [26] Nichols EK, Byass P, Chandramohan D, Clark SJ, Flaxman AD, Jakob R, Leitao J, Maire N, Rao C, Riley I, Setel PW, Grp WVAW ({2018}). \newblock \enquote{{The WHO 2016 verbal autopsy instrument: An international standard suitable for automated analysis by InterVA, InSilicoVA, and Tariff \newblock \emph{PLOS Medicine}, \textbf{{15}}({1}). \newblock ISSN {1549-1676}. [R Core Team}, 2020] {R Core Team} (2020). \newblock \emph{R: A Language and Environment for Statistical Computing}. \newblock R Foundation for Statistical Computing, Vienna, Austria. \newblock \urlprefix\url{https://www.R-project.org/}. [Serina \emph{et~al.}, 2015] Serina P, Riley I, Stewart A, James SL, Flaxman AD, Lozano R, Hernandez B, Mooney MD, Luning R, Black R, \emph{et~al.} (2015). \newblock \enquote{Improving performance of the Tariff Method for assigning causes of death to verbal autopsies.} \newblock \emph{BMC Medicine}, \textbf{13}(1), 1. [Taylor \emph{et~al.}, 1983] Taylor CE, Parker R, Reinke W, Faruquee R, \emph{et~al.} (1983). \newblock \emph{Child and maternal health services in rural India. The Narangwal experiment. 2. Integrated family planning and health care.} \newblock Johns Hopkins University Press. [Thomas, 2021] Thomas J (2021). \newblock \enquote{Software for automating the processing of verbal autopsy \newblock \url{https://github.com/verbal-autopsy-software/openVA_App}; Accessed July 21, 2021. [Thomas \emph{et~al.}, 2020] Thomas J, Choi E, Li Z, Maire N, McCormick T, Byass P, Clark S (2020). \newblock \emph{\pkg{CrossVA}: Verbal Autopsy Data Transformation for {InSilicoVA} and {InterVA5} Algorithms}. \newblock R package version 1.0.0, \urlprefix\url{https://CRAN.R-project.org/package=CrossVA}. [Thomas \emph{et~al.}, 2021] Thomas J, Clark SJ, Bratschi MW (2021{\natexlab{a}}). \newblock \enquote{Software for automating the processing of verbal autopsy \newblock \url{http://openva-pipeline.readthedocs.io/}; Accessed July 21, 2021. [Thomas \emph{et~al.}, 2021] Thomas J, Li Z, Byass P, McCormick T, Boyas M, Clark S (2021{\natexlab{b}}). \newblock \emph{{InterVA5}: Replicate and Analyse {'InterVA5'}}. \newblock R package version 1.1.3, \urlprefix\url{https://CRAN.R-project.org/package=InterVA5}. [Wen \emph{et~al.}, 2018] Wen R, Miasnikof P, Giannakeas V, Gomes M (2018). \newblock \emph{\pkg{nbc4va}: Bayes Classifier for Verbal Autopsy Data}. \newblock \proglang{R} package version 1.1, \urlprefix\url{https://github.com/rrwen/nbc4va}. [World Health Organization and others}, 2012] {World Health Organization and others} (2012). \newblock \enquote{Verbal Autopsy Standards: The 2012 WHO Verbal Autopsy Instrument Release Candidate 1.} \newblock \emph{Technical report}, World Health Organization (WHO), Geneva. \end{thebibliography} \appendix \section{New changes to the InSilicoVA algorithm} \label{app:insilico} Since the original publication of McCormick \emph{et~al.}, 2016, several major changes have been introduced in the InSilicoVA algorithm implementation to improve cause-of-death assignments. Two major changes are summarized \subsection{Removal of physically impossible causes}\label{sec:impossible} The originally proposed InSilicoVA assumes all causes of death are possible for each observation. The impact form such assumption is mild when data are abundant, but could be problematic when either sample size is small or the proportion of missing data is high. In both cases, physically impossible causes might get assigned with non-ignorable posterior mass. Since the version 1.1.5 of \pkg{InSilicoVA}, the algorithm automatically checks and removes impossible causes before fitting the model. The \(k\)-th cause is defined as physically impossible for the \(i\)-th death if \(P(s_{ij}=1|y_{i}=k) = 0\) for any \(j\) representing either sex or age group that presents. We then consider a cause to be physically impossible to the underlying population if it is impossible for all the observations of the input data. For example, with the new implementation, CSMF for pregnancy-related causes will not be estimated if the input data consist of only male deaths. \subsection{Structured symptom dependence}\label{sec:dependence} Another change in InSilicoVA algorithm is the implementation of the data consistency check. Inherited from the InterVA-4 algorithm, InSilicoVA performs a data consistency check before fitting the algorithm as described in Algorithm \ref{alg:check}. This consistency check consists of two parts. First, it removes indicators whose presence contradicts the existence of a lower-level indicator. For example, if the question \textit{pregnant at the time of death} is answered with an `Yes', then the algorithm makes sure that the item \textit{child} is set to `No'. The second part of the algorithm checks the existence of each higher-level indicator when lower-level indicators that implied by it exists. Again, if the question \textit{pregnant at the time of death} is answered with an `Yes', then the algorithm makes sure that the item \textit{female} is set to `Yes' as well. \begin{algorithm} \caption{Data Check} \label{alg:check} \begin{algorithmic}[1] \Statex \Require \Statex $\mathbf{s}_j$ = vector of $j=1 \mbox{ to } 245$ indicators for one death \vspace{3pt} \Statex $\mbox{notask}(s)$: higher-level indicator(s) whose presence ensures indicator $s$ does not exist \vspace{3pt} \Statex $\mbox{anc}(s)$: higher-level indicator(s) that must exist if indicator $s$ exists \vspace{3pt} \Statex \Ensure $\mathbf{s}^*$ = modified vector of $j=1 \mbox{ to } 245$ indicators for one death \vspace{3pt} \Statex \For{each indicator $j = 1 \mbox{ to } 245$} \State Set $ \mathbf{s}^*_j \leftarrow \mathbf{s}_j$ \Comment{Initialize} \EndFor \For{each indicator $j = 1 \mbox{ to } 245$} \label{line:start} \If{there exists $\mathbf{s}^*_{\ell}$ in $\mbox{notask}(\mathbf{s}^*_j)$ and $\mathbf{s}^*_{\ell}$ is `yes'}\label{line:firstif} \State $\mathbf{s}^*_j \leftarrow$ `no' \Comment{This indicator is nonsensical, set to `No'} \label{line:changetomissing} \EndIf \If{there exists $\mathbf{s}^*_{\ell}$ in $\mbox{anc}(\mathbf{s}_j)$ and $\mathbf{s}^*_j$ is `yes'}\label{line:secondif} \State $\mathbf{s}^*_\ell \leftarrow$ `yes' \Comment{Set more general version of this indicator to `yes'} \EndIf \EndFor \label{line:end} \State Repeat loop in lines \ref{line:start} to line \ref{line:end} again \Comment{Processes 2-level hierarchies} \Statex \end{algorithmic} \end{algorithm} This data check procedure ensures the collected data are internal consistent, yet it might introduce additional complications under the common assumption that symptoms are conditionally independent given any cause of death. This can be illustrated with a simple example below. In calculating the conditional probability of observing two symptoms, e.g., `infant', and `recent abortion', we can decompose \begin{eqnarray}\nonumber \mbox{Pr}(\mbox{infant} = Y \;\&\; \mbox{recent abortion} = N \;|\; \mbox{some cause}) \mbox{Pr}(\mbox{infant} = Y \;|\; \mbox{some cause})\\\nonumber \mbox{Pr}(\mbox{recent abortion} = N \;|\; \mbox{some cause}). \end{eqnarray} However, such independence does not hold since the two symptoms are mutually exclusive, i.e., the presence of symptom `infant' implies no recent abortion. In fact, the structure of the WHO instrument assures that the lower-level question will not be asked at all if the interview is about an infant death. Thus in practice, the above probability calculation yields an underestimate of the target value and it should be calculated simply with \[ \mbox{Pr}(\mbox{infant} = Y \;\&\; \mbox{recent abortion} = N \;|\; \mbox{some cause}) \mbox{Pr}(\mbox{infant} = Y \;|\; \mbox{some cause}) \] To reflect the different joint probability decompositions when the known symptom hierarchy is available, it turns out we can easily modify the data check algorithm by imputing missing instead of absence to the lower level symptoms in line \ref{line:changetomissing}. We have found that such a change usually yield more reasonable CSMF estimates, especially for child and neonate deaths, in practice. Also it should be noticed that this problem stems from the conditional independence assumption that has been adopted in all other methods besides InSilicoVA. The computational trick here only eliminates the bias from mutually exclusive symptoms known from the hierarchical structure of VA questionnaires. More systematic ways to deal with symptom correlations are needed to further improve the estimation. With the update of the WHO VA questionnaire to the 2016 format, three major changes were made to the data check procedure implemented by InterVA-5. First, similar to InSilicoVA, InterVA-5 also imputes missing instead of absence to the lower level symptoms in line \ref{line:changetomissing}. Second, the hierarchical structure of the 2016 VA questionnaire makes use of both ``Yes'' and ``No'' (or ``Y'' and ``N'') as substantive values that trigger the asking of lower-level questions. For example, if the indicator/question ``Did the baby ever cry?'' has the response ``No'', then the following indicator/question, ``Did the baby cry immediately after birth, even if only a little bit?'' should not be asked -- as implemented with the notask(\(s\)) function in Algorithm \ref{alg:check}. Thus, additional steps are needed in the InterVA-5 data check which test the value of each indicator with the appropriate substantive response that implies subsequent action (i.e., asking or not asking a different indicator/question) in line \ref{line:firstif} and \ref{line:secondif}. The last major change is that an additional check is made for indicators/questions that only apply to neonates. If a record for a non-neonate contains a value for an indicator/question that only applies to neonates, then that value is cleared in the working copy of the data file. The same check rules are used by InSilicoVA for WHO 2016 data as well. The implementation is slightly different between InSilicoVA and InterVA-5. InterVA-5 updates \(s_j^*\) to be missing in line \ref{line:changetomissing} only when it is recorded to be the corresponding substantive value, since the non-substantive value is considered the same as missing values in the calculation. InSilicoVA, on the other hand, make the update regardless of the value of \(s_j^*\), since the non-substantive value and missing are treated differently in the algorithm. \end{document}
# Servicifying zk-SNARKs Execution for Verifiable Off-chain Computations ††thanks: Alvaro Alonso Domenech, Jonathan Heiss, Stefan Tai Information System Engineering TU Berlin Berlin, Germany <EMAIL_ADDRESS> ###### Abstract Zk-SNARKs help scale blockchains with Verifiable Off-chain Computations (VOC). zk-SNARK DSL toolkits are key when designing arithmetic circuits but fall short of automating the subsequent proof-generation step in an automated manner. We emphasize the need for portability, interoperability, and manageability in VOC-based solutions and introduce a Proving Service that is designed to provide a scalable and reusable solution for generating zk-SNARK proofs leveraging clouds. ###### Index Terms: Blockchain, Zero-knowledge Proofs, zk-SNARKs, ZoKrates, Proving, Service Architecture, Cloud Computing ## I Introduction The succinctness property and the short verification times of zk-SNARK s come at the cost of large computational complexity, static execution models, and memory overheads during the proof generation process [1]. These limitations represent a problem in VOC applications which require handling large and varying workloads, like rollups. This type of application would benefit from scalable, interoperable, and manageable system environments like clouds. However, DSL toolkits like ZoKrates [2] or Circom [3] concentrate on circuit development and put little emphasis on how to integrate these circuits into production systems. Bridging the gap between advanced cryptography and modern systems engineering also helps to create standardized benchmarks for proving systems and tools. ## II Proving Service We introduce a service-oriented approach for VOC that facilitates the use of the cryptographic procedures of zk-SNARK s within cloud system architectures. Our system allows for executing arithmetic circuits as encapsulated application logic in containers that are deployable to different machines realizing scalability, provide interoperability with other services, and are better manageable. For that, we give an overview of the service-oriented system architecture and describe the internals of the proving service. ### II-A Service Oriented Verifiable Off-chain Computation The problems above can be fulfilled through a service-oriented approach as depicted in Figure 1. Starting from a higher-level system’s perspective, we treat the proving service as a black box which, upon a request, returns the proof together with the computation’s output. Following the VOC model [2], we distinguish between the blockchain infrastructure hosting the verifier contract and a cloud-based off-chain infrastructure that runs outside the consensus protocol and hosts the consumer and the proving service. Figure 1: Proving Service Model The consumer service is responsible for managing the proving service’s inputs and outputs and interacting with the relying verifier contract. It receives data from external sources and translates them into a proof request. Upon a request through the prover client, the proving service executes the VOC and returns the ZKP attesting to the computational integrity of the VOC. The consumer service then submits the ZKP to the verifier contract through its verifier client. On submission, the verifier contract verifies the proof computed by the proving service. Figure 2: Proving Service Architecture Figure 3: Proving Time and Memory Consumption ### II-B Service Architecture The Proving Service depicted in Figure 2 is an application service that runs stand-alone on the prover’s off-chain infrastructure. It serves proof requests by exposing the proving-related operations through an Application Program Interface (API) as minimal functionality to the consumer service. The procedure can be summarized in four simple steps: First, the proving service receives the Proof Request containing the public ($x$) and private ($x^{\prime}$) proof arguments and the identifier of the addressed Executable Constraint System ($ecs$)($id$). Second, using the $id$ the API fetches the corresponding $ecs$ and private key ($pk$) from the Proof Registry which is the persistent storage component containing these large, recurrently requested files needed for proving. The proof registry manages different reusable pairs of $ecs$ and $pk$, each addressable through a unique $id$. Third, the Proving Instance is executed in two stages: the witness computations and the proof generation. Fourth, the ZKP is returned to the consumer service through the API upon successful execution. ## III ZoKrates-API The previous service-oriented architecture serves as a technology-agnostic blueprint for building proving services for DSL circuits. For evaluation, we technically instantiate the proving service for ZoKrates [2] and present the ZoKrates-API111https://github.com/ZK-Plus/zokrates-api as a ready-to-use open- source software. We servicify ZoKrates by wrapping an API around the ZoKrates interpreter, the central component of the ZoKrates software that previously has only been addressable through a Command Line Interface and a Javascript library. The ZoKrates-API exposes the methods of the ZoKrates interpreter through HTTP endpoints. We containerized the ZoKrates-API using Docker, making the services easily deployable among a wide range of machines and allowing us to further leverage cloud-native tools like Kubernetes for horizontal scalability, manageability, and observability. Furthermore, the ZoKrates-API supports multi-threading so a single instance can compute multiple proofs in parallel. ## IV Evaluation To test the presented implementation, we deployed the containerized ZoKrates- API on a Kubernetes cluster. As a workload for our experiments, we generated a large number of EdDSA signatures[4] which amounts for more than $2^{29}$ of circuit constraints similar to [5]. We conducted three experiments (see Figure 3) to measure the average proving time per signature in [sec] and the memory consumption in [gb] using various cluster configurations. Figure 3A) shows a 33% improvement in proving time for a single machine when choosing an appropriate machine size. An argument in favour of vertical escalability. For a single machine as well, Figure 3B) shows that the proving time can be brought down drastically when few parallel threads are enabled, though the gains plateaued rapidly due to the increasing resources needed. Running the same experiments in parallel VMs instead of threads, Figure 3C) demonstrate a better approach to scaling proving as the computational burden is distributed over several machines, proportional increasing the processing time. ## V conclusion As the experiments demonstrate, significant performance improvements can be gained from horizontal (more nodes) and vertical (larger nodes) scalability of an arbitrary zk-SNARK proof. We facilitate this process by leveraging modern cloud-native architectures such as Docker and Kubernetes. ## References * [1] E. Ben-Sasson, A. Chiesa, E. Tromer, and M. Virza, “Scalable zero knowledge via cycles of elliptic curves,” _Algorithmica_ , vol. 79, pp. 1102–1160, 2017. * [2] J. Eberhardt and S. Tai, “Zokrates - scalable privacy-preserving off-chain computations,” in _IEEE International Conference on Blockchain_. IEEE, 2018. * [3] M. Bellés-Muñoz, M. Isabel, J. L. Muñoz-Tapia, A. Rubio, and J. Baylina, “Circom: A circuit description language for building zero-knowledge applications,” _IEEE Transactions on Dependable and Secure Computing_ , 2022. * [4] J. Heiss, A. Busse, and S. Tai, “Trustworthy pre-processing of sensor data in data on-chaining workflows for blockchain-based iot applications,” in _Service-Oriented Computing: 19th International Conference, ICSOC 2021, Virtual Event, November 22–25, 2021, Proceedings 19_. Springer, 2021, pp. 133–149. * [5] A. Chiesa, R. Lehmkuhl, P. Mishra, and Y. Zhang, “Eos: Efficient private delegation of zksnark provers,” in _USENIX Security Symposium. USENIX Association_ , 2023.
11institutetext: Tecnológico de Monterrey, School of Engineering and Sciences, Mexico. 22institutetext: Universidad de las Americas Puebla, Department of Chemical, Food and Environmental Engineering, Puebla, 72810, Mexico. 33institutetext: Universitat Politècnica de Catalunya. EEBE, Eduard Maristany 16, 08019 Barcelona. Catalonia, Spain. # Comparing Machine Learning based Segmentation Models on Jet Fire Radiation Zones††thanks: Currently under review for the Mexican Conference on AI (MICAI 2021) Carmina Pérez-Guerrero 11 Adriana Palacios 22 Gilberto Ochoa-Ruiz 11 Christian Mata 33 Miguel Gonzalez-Mendoza 11 Luis Eduardo Falcón-Morales 11 ###### Abstract Risk assessment is relevant in any workplace, however there is a degree of unpredictability when dealing with flammable or hazardous materials so that detection of fire accidents by itself may not be enough. An example of this is the impingement of jet fires, where the heat fluxes of the flame could reach nearby equipment and dramatically increase the probability of a domino effect with catastrophic results. Because of this, the characterization of such fire accidents is important from a risk management point of view. One such characterization would be the segmentation of different radiation zones within the flame, so this paper presents an exploratory research regarding several traditional computer vision and Deep Learning segmentation approaches to solve this specific problem. A data set of propane jet fires is used to train and evaluate the different approaches and given the difference in the distribution of the zones and background of the images, different loss functions, that seek to alleviate data imbalance, are also explored. Additionally, different metrics are correlated to a manual ranking performed by experts to make an evaluation that closely resembles the expert’s criteria. The Hausdorff Distance and Adjusted Random Index were the metrics with the highest correlation and the best results were obtained from the UNet architecture with a Weighted Cross-Entropy Loss. These results can be used in future research to extract more geometric information from the segmentation masks or could even be implemented on other types of fire accidents. ###### Keywords: semantic segmentation deep learning computer vision jet fires. ## 1 Introduction In the industrial environment there are certain activities such as the storage of fuels or the transportation of hazardous material, that can be involved in severe accidents that affect the industrial plant or activity border, as well as external factors like human health, environmental damage, property damage, and others. An overall knowledge of the features and characteristics of major accidents is required to prevent and manage them or, in the worst case scenario, take action to reduce and control their severity and aftermath. There are some accidents that are relatively well known and researched, however, there are other accidents that have not been broadly explored. Sometimes the detection of fire, in an industrial setting, is not enough to make the correct decisions when managing certain fire accidents. An example of this, is the impingement of jet flames on nearby pipes, depending on the heat fluxes of the flame, the pipe wall could heat up quickly and reach dangerous temperatures that lead to severe domino effect sequences. These heat fluxes can be defined as three main areas of interest within the flame [6], and the definition of their geometric characteristics and localization becomes valuable information in risk management. Semantic segmentation could be an useful approach for this kind of flame characterization, and the same procedure could even be employed in other fire related accidents. To explore this proposal, different segmentation methodologies are evaluated on a set of images from real jet fires of propane to accurately segment the radiation zones within the flames. This different zones are illustrated in Figure 1 and are defined across this paper as the Central Zone, the Middle Zone and the Outer Zone. The rest of this paper is organized as follows. Section 2 describes the previous work present in the literature related to the problem and different semantic segmentation methods. Section 3 describes the approaches used to perform the exploration research. Section 4 explains the data set used for the experiments and the pre-processing methods applied. Section 5 describes the different evaluation metrics and loss functions explored during the experiments. Section 6 contains the training protocol used for the experiments with Deep Learning architectures. Section 7 explains the testing procedure applied to the segmentation methods. Section 8 presents the results of the exploration research. This section also offers a discussion of the future work that can stem from the knowledge obtained. Finally, Section 9 summarizes the major findings of the presented work. (a) (b) Figure 1: Image (a) is an infrared visualization of an horizontal propane jet flame. Image (b) is the corresponding ground truth segmentation, with the segment names indicated. Modified from [6]. ## 2 State of the Art There has been previous research work regarding the use of Computer Vision and image processing for the detection and monitoring of flares in the context of industrial security. For example, Rodrigues and Yan [13] used imaging sensors combined with digital image processing to determine the size, shape and region of the flare, successfully characterizing the dynamism of the fire. Another example is the work done in Janssen and Sepasian [7], where, by separating the flare form the background using temperature thresholding, and adding false colors to represent different temperature regions, a system was created, capable of tracking the flare size for automated event signaling. ### 2.1 Deep Learning Architectures Deep learning algorithms, such as Convolutional Neural Networks, have shown outstanding performance in many complex tasks, such as image recognition, object detection, and semantic and instance segmentation [9]. Advancements on these methods have been increasing rapidly, with constant research being published regarding better and more robust algorithms. Some important architectures present in the literature are summarized in Table 1 with their advantages and disadvantages described. Table 1: Summary of the selected architectures. Source | Architecture | Pros | Cons ---|---|---|--- [4] | DeepLabv3 | | Recovers detailed structures --- lost due to spatial invariance. Efficient approximations via probabilistic inference. Wider receptive fields. | Low accuracy on small --- scaled objects. Has trouble capturing delicate boundaries of objects. [2] | SegNet | | Efficient inference in terms of memory --- and computational time. Small number of trainable parameters. Improved boundary delineation. | Input image must be fixed. --- Performance drops with a large number of classes and is sensitive to class imbalance. [14] | UNet | | Can handle inputs of arbitrary sizes. --- Smaller model weight size. Precise localization of regions. | Size of network comparable --- to the size of features. Significant amount of time to train. High GPU memory footprint in larger images. [10] | | Attention --- UNet | Avoids the use of multiple similar --- feature maps. Focuses on the most informative features without additional supervision. Enhances the results of the UNet architecture. | Adds more weight parameters --- The time for training increases, especially for long sequences In general, these architectures were selected due to their efficiency during segmentation and their capacity to accurately portray the shape of such dynamic figures obtained from the flames. ### 2.2 Traditional Computer Vision Methods Different traditional segmentation methods are included in the analysis as a baseline for the results given by the Deep Learning architectures, these methods are: Gaussian Mixture Model (GMM), K-means clustering, Thresholding, and Chan-Vese segmentation. * • GMM is one of the main methods applied to fire and smoke image segmentation. It offers a clear definition of their dynamic shape, but sometimes missed pixels from the inner parts of the fire and smoke [1]. * • K-means clustering provides shape-based image segmentation [18] and has been previously used for fire segmentation [15] [1]. However, this method does not guarantee continuous areas [18] * • Thresholding is the simplest method of image segmentation and has a fast operation speed. It is most commonly used in region-based segmentation, but it is sensitive to noise and gray scale unevenness [17]. * • The Chan-Vese segmentation [3] belongs to the group of active contour models, which present some advantages for infrared image segmentation, since the edges obtained are smooth and are represented with closed curves. However, this kind of segmentation is very sensitive to noise and depends heavily on the location of the initial contour, so it needs to be manually placed near the image of interest [19]. ## 3 Proposed Approach To explore the semantic segmentation of radiation zones within the flames as characterization of fire incidents, a group of 4 traditional segmentation methods and 4 deep learning architectures are explored. These segmentation methods are trained and tested using a data set of jet fire images obtained from videos of an experiment performed in an open field. To properly evaluate and compare the different segmentation approaches, several metrics with different evaluation methods are correlated to manual rankings performed by two experts in the field, this is to make sure that the evaluation is the most representative of a fire engineer’s perception of good segmentation. The best model found from this exploratory research will then be used in future work, not included in this paper, to extract other geometric characteristics from the resulting segmentation masks. ## 4 Data set Investigators from Universitat Politècnica de Catalunya performed an experiment to produce horizontal jet fires at subsonic and sonic gas exit rates. The experiment was filmed using an infrared thermographic camera, more specifically, an AGEMA 570 by Flir Systems. The video was saved in four frames per second, resulting in a total of 201 images with a resolution of 640 x 480 pixels. After obtaining the infrared visualizations of the flames, a segmentation was performed of the three radiation zones within the fire, the results were validated by experts in the field and the result was a ground- truth for each of the infrared images. The images were saved as Matlab files that contain a temperature matrix corresponding to the temperature values detected by the camera for the infrared images, and the a label matrix corresponding to the segments for the ground-truth segmentations. These files were then exported as PNG files to be used by the different segmentation algorithms described in Section 2.1. ### 4.1 Image Processing To enhance the characteristics of the jet fires represented in the infrared images, and to reduce their variance, a process of image normalization was employed, which also helps in the convergence of the Deep Learning methods. The ground truth images were transformed into labeled images, where a label id was used instead of the original RGB values. Given the small number of samples of the data set, and to increase the variability of the input during the training of the models, data augmentation techniques were applied. These processing of the images can increase the performance of the models, can avoid the over-fitting to the training samples, and can help the models to also perform well for instances that may not be present in the original data set. Horizontal flipping, random cropping, and random scaling were applied in parallel of the training workflow, so for each iteration of training, a different augmented image was inputted. The probability of horizontal flipping was set to 50%, and random scaling had values that ranged between 0.7 and 2.0. ## 5 Metrics and Loss Functions ### 5.1 Metrics An analysis was performed to select evaluation metrics that were more representative of an expert’s evaluation of the segmentation. The metrics analysed are separated into groups that describe their evaluation method. This diversity is important because the consideration of particular properties could prevent the discovery of other particular errors or lead to over or underestimating them. To compare the metrics enlisted in Table 2, a group of images were evaluated with the metrics and two manual rankings performed by two experts in the field. A Pearson pairwise correlation is used to perform this comparison at segmentation level. Table 2: Summary of the metrics analysed in this paper. The ”Group” column describes the method group that the metric belongs to. Based on [16]. Metric | Group ---|--- Jaccard Index | Spatial Overlap Based F-measure | Spatial Overlap Based Adjusted Rand Index | Pair Counting Based Mutual Information | Information Theoretic Based Cohen’s Kappa | Probabilistic Based Hausdorff Distance | Spatial Distance Based Mean Absolute Error | Performance Based Mean Square Error | Performance Based Peak Signal to Noise Ratio | Performance Based ### 5.2 Loss Functions The proportion of the different radiation zones within each flame is different. The Outer Zone tends to be the largest segment and the central zone is usually smaller than the Middle Zone, this differences could affect the overall segmentation obtained during training, which is why the loss functions employed in this research work were focused on dealing with this class imbalance. The implemented loss functions are summarized in Table 3. Table 3: Summary of the loss functions implemented in this paper. Source | Loss Function | Description ---|---|--- [12] | | Weighted Cross-Entropy --- Loss | Combines Log Softmax and Negative Log Likelihood. --- Useful for multiple classes that are unbalanced. [8] | Focal Loss | | Addresses foreground-background class imbalance. --- Training is focused on a sparse set of hard examples. [5] | | Generalized Wasserstein --- Dice Loss (GWDL) | Semantically-informed generalization of the Dice score. --- Based on the Wasserstein distance on the probabilistic label space. ## 6 Training The PyTorch framework [12] was used for the implementation of the Deep Learning architectures. To maintain the weights of the Convolutional Neural Networks as small as possible, a weight decay strategy was used with L2 regularization. The learning rate had an initial value of 0.0001 and used an ADAM optimizer during training. The class weights used by the Weighted Cross- Entropy and Focal losses were computed according to the ENet custom class weighing scheme [11] and are defined as the following: 1.59 for background, 10.61 for Outer zone, 17.13 for Middle zone, and 22.25 for Central zone. The class distances used for the loss function of GWDL is defined as 1 between the background and the radiation zones, and as 0.5 between the zones themselves. The training was performed using an Nvidia DGX workstation that has 8 Nvidia GPUs and allows a batch size of 4 as maximum, which is the batch size used for the models. The data was split in 80% for training and 20% for testing and validation, resulting in a total of 161 images for training, 20 images for testing, and 20 images for validation. The models were trained for up to 5000 epochs with an Early Stopping strategy to avoid over fitting. The resulting loss values for the best models can be visualized in Fig. 2. (a) Loss values for the DeepLabv3 model. Early Stopping took place at epoch 701. (b) Loss values for the SegNet model. Early Stopping took place at epoch 1274. (c) Loss values for the UNet model. Early Stopping took place at epoch 1460. (d) Loss values for the Attention UNet model. Early Stopping took place at epoch 1560. Figure 2: The Weighted Cross-Entropy loss values for the best models obtained for each architecture explored. ## 7 Testing Testing and Validation was performed on 20% of the data set, which represent a total of 40 images. The results are compared using the metrics with the highest correlation to the manual ranking of experts. The results for each Deep Learning architecture are first compared across all 3 different loss functions mentioned in Section 5.2. The best performing combination, for each architecture, is then compared to the results obtained from the other 4 traditional segmentation models. The time each method takes to segment the whole data set of 201 images is also taken into account. The goal of the comparison is to find the best overall model for the segmentation of the radiation zones within the flames. ## 8 Results and Discussion ### 8.1 Selected Metrics The results of the correlation between the metrics, mentioned in Section 5, and the manual rankings performed by experts can be observed in Fig. 3. Figure 3: Heatmap representing the Pearson Correlation Coefficient for the metrics in Section 5 and the manual rankings Rank1 and Rank2 done by experts in the field of the problem. The highest Pearson correlation value was given by the Hausdorff Distance, with a value of -0.32 to the first manual ranking, this negative relationship is because a smaller Hausdorff Distance is preferable and the manual ranking assigned higher values as the segmentation improved. The second highest correlations were given by the Adjusted Rand Index in both manual rankings, with a value of 0.29 for the first one and 0.27 for the second one, this positive relationship is because the values of the Adjusted Rand Index go from 0 to 1, with 1 being the best result, and showing a similar behaviour to the manual ranking that assigns higher values to better segmentations. ### 8.2 Best Loss Function Each combination of Deep Learning architecture and Loss Function was evaluated using the Hausdorff Distance as main metric. The best results across all architectures were obtained when using the Weighted Cross-Entropy Loss, this can be observed in Fig 4. In general the Focal Loss also showed good results, being close to the Weighted Cross-Entropy results and most of the times surpassing the GWDL results, with the only exception being the Attention UNet model, where the mean Hausdorff Distance with Focal Loss is larger than with GWDL, but has a much closer distribution. The most dramatic difference happened with the SegNet architecture, which showed very poor results with GWDL. (a) Hausdorff Distance for all the DeepLabv3 models. (b) Hausdorff Distance for all the SegNet models. (c) Hausdorff Distance for all the UNet models. (d) Hausdorff Distance for all the Attention UNet models. Figure 4: The Hausdorff Distance for all the model and loss function combinations across the validation set. For each model the order of loss functions is Weighted Cross-Entropy, Focal and GWDL from left to right. ### 8.3 Traditional and Deep Learning Segmentation A comparison was done between the traditional Computer Vision methods mentioned in Section 2.2 and the Deep Learning models mentioned in Section 2.1,using the Weighted Cross-Entropy loss. The Hausdorff Distance and the Adjusted Rand Index are used to evaluate the models across the testing set and the time each method takes to perform the segmentation on all the images from the data set is also taken into account. These results are summarized in Table 4 and the distribution of the results can be visualized in Fig 5. Overall the best performing models are observed to be the UNet and Attention UNet models. Even if Attention UNet achieved a slightly better Adjusted Rand Index score, UNet obtained a better Hausdorff distance and segmentation time, therefore we can say that the best model is the one that uses the UNet architecture. Table 4: Mean Hausdorff Distance and Adjusted Rand Index values for all the segmentation models across the testing set, as well as the time in seconds that each method takes to segment the whole data set. The best results are in bold. Method | Hausdorff Distance | Adjusted Rand Index | Time (s) ---|---|---|--- GMM | 1288.10 | 0.9156 | 2723.8 K-means | 1000.63 | 0.8855 | 3035.1 Thresholding | 1029.08 | 0.9152 | 30.7 Chan-Vese | 1031.90 | 0.8568 | 18177.5 DeepLabv3 | 784.86 | 0.9514 | 17.1 SegNet | 692.73 | 0.9381 | 16.4 UNet | 586.46 | 0.9504 | 15.7 Attention UNet | 601.05 | 0.9592 | 17.7 (a) Hausdorff distance for all segmentation models. (b) Adjusted Rand Index for all segmentation models. Figure 5: The Hausdorff Distance and Adjusted Rand Index values for all the segmentation models across the testing set. The difference in the segmentation of each method can also be visualized in Fig 6. It can be observed that the shape of the Outer and Middle zones are generally well represented in most of the segmentation models, however the most important differences are found in the Central zone, where the Deep Learning architectures defined more clearly it’s shape. The similar results of the UNet and Attention UNet architectures are also observed in this sample segmentation masks, with a really close resemblance to the ground truth. (a) GMM. (b) K-means. (c) Thresholding. (d) Chan-Vese. (e) DeepLabv3. (f) SegNet. (g) UNet. (h) Attention UNet. Figure 6: Sample segmentations of all the models. ### 8.4 Discussion Overall, the Deep Leaning algorithms greatly outperformed the traditional Computer Vision methods and the best proposed model would be a UNet architecture with a Weighted Cross-Entropy loss function. The segmentation masks obtained from this model could be used in the future to further extract other geometric characteristics of the flame, such as length, area and lift- off distance. This additional information would improve greatly the decision making process involved in the risk assessment and management of fire related accidents that can take place in an industrial setting. Furthermore, the metrics of Hausdorff Distance and Adjusted Rand Index can be used to evaluate other segmentation approaches that may try to solve similar problems in the future, having the certainty that the evaluation would be a close representation of an expert’s opinion. ## 9 Conclusions The semantic segmentation of radiation zones within the flames can be used to characterize fire accidents, such as jet fires, and the information obtained from the segmentation can prove to be critical when dealing with risk management in industrial settings. The exploratory research presented in this paper continued to show that Deep Learning architectures greatly outperform other traditional Computer Vision approaches. It was also found that for this specific problem, the best loss function to train a Deep Learning model is a Weighted Cross-Entropy Loss, the best architecture to be used is UNet, and the best evaluation metrics are both the Hausdorff Distance and the Adjusted Rand Index. All this knowledge can be later used in future research focused on extracting even more geometric information from the segmentation masks. This could bring about a more complete characterization analysis of jet fires and the methods applied to this types of fire accidents could then be used on other fire scenarios. #### Acknowledgements This research is supported in part by the Mexican National Council of Science and Technology (CONACYT). This research is part of the project 7817-2019 funded by the Jalisco State Council of Science and Technology (COECYTJAL). ## References * [1] Ajith, M., Martínez-Ramón, M.: Unsupervised segmentation of fire and smoke from infra-red videos. IEEE Access 7, 182381–182394 (2019). https://doi.org/10.1109/ACCESS.2019.2960209 * [2] Badrinarayanan, V., Kendall, A., Cipolla, R.: Segnet: A deep convolutional encoder-decoder architecture for image segmentation (2016) * [3] Chan, T., Vese, L.: Active contours without edges. IEEE Transactions on Image Processing 10(2), 266–277 (2001). https://doi.org/10.1109/83.902291 * [4] Chen, L.C., Papandreou, G., Schroff, F., Adam, H.: Rethinking atrous convolution for semantic image segmentation (2017) * [5] Fidon, L., Li, W., Garcia-Peraza-Herrera, L.C., Ekanayake, J., Kitchen, N., Ourselin, S., Vercauteren, T.: Generalised Wasserstein dice score for imbalanced multi-class segmentation using holistic convolutional networks. In: International MICCAI Brain Lesion Workshop. pp. 64–76. Springer (2017). https://doi.org/10.1007/978-3-319-75238-9_6 * [6] Foroughi, V., Palacios, A., Barraza, C., Àgueda, A., Mata, C., Pastor, E., Casal, J.: Thermal effects of a sonic jet fire impingement on a pipe. Journal of Loss Prevention in the Process Industries 71, 104449 (2021). https://doi.org/10.1016/j.jlp.2021.104449 * [7] Janssen, R., Sepasian, N.: Automatic Flare-Stack Monitoring. SPE Production & Operations 34(01), 18–23 (09 2018). https://doi.org/10.2118/187257-PA * [8] Lin, T., Goyal, P., Girshick, R., He, K., Dollar, P.: Focal loss for dense object detection. IEEE Transactions on Pattern Analysis & Machine Intelligence 42(02), 318–327 (2020). https://doi.org/10.1109/TPAMI.2018.2858826 * [9] Litjens, G., Kooi, T., Bejnordi, B.E., Setio, A.A.A., Ciompi, F., Ghafoorian, M., van der Laak, J.A., van Ginneken, B., Sánchez, C.I.: A survey on deep learning in medical image analysis. Medical Image Analysis 42, 60–88 (Dec 2017). https://doi.org/10.1016/j.media.2017.07.005 * [10] Oktay, O., Schlemper, J., Folgoc, L.L., Lee, M., Heinrich, M., Misawa, K., Mori, K., McDonagh, S., Hammerla, N.Y., Kainz, B., Glocker, B., Rueckert, D.: Attention u-net: Learning where to look for the pancreas (2018) * [11] Paszke, A., Chaurasia, A., Kim, S., Culurciello, E.: Enet: A deep neural network architecture for real-time semantic segmentation. CoRR 1606.02147 (2016) * [12] Paszke, A., Gross, S., Massa, F., Lerer, A., Bradbury, J., Chanan, G., Killeen, T., Lin, Z., Gimelshein, N., Antiga, L., Desmaison, A., Kopf, A., Yang, E., DeVito, Z., Raison, M., Tejani, A., Chilamkurthy, S., Steiner, B., Fang, L., Bai, J., Chintala, S.: Pytorch: An imperative style, high-performance deep learning library. In: Wallach, H., Larochelle, H., Beygelzimer, A., d'Alché-Buc, F., Fox, E., Garnett, R. (eds.) Advances in Neural Information Processing Systems 32, pp. 8024–8035. Curran Associates, Inc. (2019), http://papers.neurips.cc/paper/9015-pytorch-an-imperative-style-high-performance-deep-learning-library.pdf * [13] Rodrigues, S.J., Yan, Y.: Application of digital imaging techniques to flare monitoring. Journal of Physics: Conference Series 307, 012048 (aug 2011). https://doi.org/10.1088/1742-6596/307/1/012048 * [14] Ronneberger, O., Fischer, P., Brox, T.: U-net: Convolutional networks for biomedical image segmentation (2015) * [15] Rudz, S., Chetehouna, K., Hafiane, A., Laurent, H., Séro-Guillaume, O.: Investigation of a novel image segmentation method dedicated to forest fire applications*. Measurement Science and Technology 24, 075403 (06 2013). https://doi.org/10.1088/0957-0233/24/7/075403 * [16] Taha, A.A., Hanbury, A.: Metrics for evaluating 3d medical image segmentation: Analysis, selection, and tool. BMC Medical Imaging 15(29) (08 2015). https://doi.org/10.1186/s12880-015-0068-x * [17] Yuheng, S., Hao, Y.: Image segmentation algorithms overview (2017) * [18] Zaitoun, N., Aqel, M.: Survey on image segmentation techniques. Procedia Computer Science 65, 797–806 (12 2015). https://doi.org/10.1016/j.procs.2015.09.027 * [19] Zhang, R., Zhu, S., Zhou, Q.: A novel gradient vector flow snake model based on convex function for infrared image segmentation. Sensors 16, 1756 (10 2016). https://doi.org/10.3390/s16101756
# Long time solutions for 1D cubic dispersive equations, Part II: the focusing case Mihaela Ifrim Department of Mathematics, University of Wisconsin, Madison <EMAIL_ADDRESS>and Daniel Tataru Department of Mathematics, University of California at Berkeley<EMAIL_ADDRESS> ###### Abstract. This article is concerned with one dimensional dispersive flows with cubic nonlinearities on the real line. In a very recent work, the authors have introduced a broad conjecture for such flows, asserting that in the defocusing case, small initial data yields global, scattering solutions. Then this conjecture was proved in the case of a Schrödinger dispersion relation. In terms of scattering, our global solutions were proved to satisfy both global $L^{6}$ Strichartz estimates and bilinear $L^{2}$ bounds. Notably, no localization assumption is made on the initial data. In this article we consider the focusing scenario. There potentially one may have small solitons, so one cannot hope to have global scattering solutions in general. Instead, we look for long time solutions, and ask what is the time- scale on which the solutions exist and satisfy good dispersive estimates. Our main result, which also applies in the case of the Schrödinger dispersion relation, asserts that for initial data of size $\epsilon$, the solutions exist on the time-scale $\epsilon^{-8}$, and satisfy the desired $L^{6}$ Strichartz estimates and bilinear $L^{2}$ bounds on the time-scale $\epsilon^{-6}$. To the best of our knowledge, this is the first result to reach such a threshold. ###### Contents 1. 1 Introduction 2. 2 Density-flux and interaction Morawetz identities 3. 3 Strichartz and bilinear $L^{2}$ bounds 4. 4 Long time energy estimates 5. 5 Optimality remarks ## 1\. Introduction The question of obtaining long time solutions for one dimensional dispersive flows with quadratic/cubic nonlinearities has attracted a lot of attention in recent years. One can distinguish two different but closely related types of results that have emerged, as well as several successful approaches. On one hand, _normal form methods_ have been developed in order to extend the lifespan of solutions, beginning with [28] in the late ’80’s. Somewhat later, around 2000, the _I-method_ , introduced in [5] brought forth the idea of constructing better almost conserved quantities. These two ideas serve well in the study of semilinear flows, where it was later understood that they are connected [3]. Neither of these techniques can be directly applied to quasilinear problems. Addressing this problem, it was discovered in the work of the authors and collaborators [16], [18] that one can adapt the normal form method to quasilinear problems by constructing energies which simultaneously capture both the quasilinear and the normal form structures. This idea was called the _modified energy method_ , and can also be seen in some way as a quasilinear adaptation of the I-method. An alternate approach, also in the quasilinear setting, is provided by the _flow method_ of [15], where a better normal form transformation is constructed using a well chosen auxiliary flow. On the other hand, the further goal of obtaining scattering, global in time solutions for one dimensional dispersive flows with quadratic/cubic nonlinearities has also been extensively studied in the last two decades for a number of models, under the assumption that the initial data is both _small_ and _localized_ ; without being exhaustive, see for instance [13, 14, 25, 23, 17]. The nonlinearities in these models are primarily cubic, though the analysis has also been extended via normal form and modified energy methods to problems which also have nonresonant quadratic interactions; several such examples are [1, 18, 10, 19, 24], see also further references therein, as well as the authors’ expository paper [22]. If instead one considers initial data which is just _small_ , without any localization assumption, then the problem becomes much more difficult, because this allows for far stronger nonlinear interactions over long time-scales. One also needs to distinguish between the focusing and the defocusing problems. In a recent paper [21], the authors have introduced a broad global well-posedness (GWP) conjecture, which applies to both semilinear and quasilinear problems: ###### Conjecture 1 (Non-localized data defocusing GWP conjecture). One dimensional dispersive flows on the real line with cubic defocusing nonlinearities and small initial data have global in time, scattering solutions. The main result of [21] asserts that this conjecture is true under suitable assumptions, most notably that the dispersion relation is the Schrödinger dispersion relation. That was the first global in time well-posedness result of this type. Notably, scattering here is interpreted in a weak sense, to mean that the solution satisfies global $L^{6}$ Strichartz estimates and bilinear $L^{2}$ bounds. This is because of the strong nonlinear effects, which preclude any kind of classical scattering. The precise result is stated later in Theorem 3. Our interest in this article is instead in the focusing case of the same problem. Since 1D focusing dispersive problems typically admit small solitons, a global result as stated in the above conjecture simply cannot hold. Even if global solutions exist (as it is the case for instance for the cubic focusing NLS problem) the presence of solitons will defeat any kind of global decay estimates. For this reason, in the focusing case we will rethink the problem as a question about the lifespan of solutions with small initial data. Precisely, if the initial data has size $\epsilon$ when measured in a suitable $H^{s}$ Sobolev norm, what can be said about the lifespan of the solutions as a function of $\epsilon$ ? Following the lead of our earlier paper, we begin by formulating the focusing counterpart of the previous conjecture. Then we will prove that the conjecture is true under suitable assumptions. Our main conjecture is as follows: ###### Conjecture 2 (Non-localized data focusing conjecture). One dimensional dispersive flows on the real line with cubic nonlinearities and small initial data of size $\epsilon$ have solutions which remain of comparable size at least on an $\epsilon^{-8}$ time-scale. The main result of this paper, see Theorem 1 below, asserts that this conjecture is valid under the additional assumption that the dispersion relation is of Schrödinger type. As part of this result, we also prove that our long time solutions satisfy both $L^{6}$ Strichartz estimates and bilinear $L^{2}$ bounds on suitable time-scales. This is akin to our earlier work on the defocusing case, but with the difference that in the defocusing case such estimates were proved globally in time. For reference purposes, we note some intermediate lifespan thresholds which can be reached with methods which were developed earlier: * • A cubic lifespan $\epsilon^{-2}$ can be reached using direct energy estimates, using only the fact that the nonlinearity is cubic. * • A quintic lifespan $\epsilon^{-4}$ can be reached by more accurate energy estimates using a quartic energy correction. This requires the cubic nonlinearity to be conservative, as defined later in the introduction. * • In the case of a perturbative nonlinearity, a quintic lifespan can also be obtained by directly using Strichartz estimates. Heuristically, if it were possible to directly combine the last two ideas above, that would lead to the $\epsilon^{-8}$ threshold in the present paper. However, the price to achieve that would be very steep, as one would need to assume both a perturbative nonlinearity and high regularity. By comparison, our new result provides a much more robust approach, which is both of a nonperturbative nature and far more efficient in terms of the regularity requirements. We show here how dispersive and normal form tools can be combined very efficiently in order to make substantial gains. In the present paper we aim for a reasonably simple setting, where our model problem is borderline semilinear, and where we prove results we expect to be optimal. This should also serve as a baseline for further developments. In particular, we believe that our methods can be also applied in non- perturbative, quasilinear settings. ### 1.1. Cubic NLS problems in one space dimension The fundamental model for one-dimensional dispersive flows with cubic nonlinearity in one space dimension is the cubic nonlinear Schrödinger (NLS) flow, (1.1) $\left\\{\begin{aligned} &iu_{t}+u_{xx}=\pm 2u|u|^{2}\\\ &u(0)=u_{0},\end{aligned}\right.$ with $u:\mathbb{R}\times\mathbb{R}\rightarrow\mathbb{C}$. This comes in a defocusing (+) and a focusing (-) version. The above cubic NLS flows are globally well-posed in $L^{2}$ both in the focusing and in the defocusing case, though the global behavior differs in the two cases. Both of these model problems are completely integrable, and one may study their global behavior using inverse scattering tools [9], [2]. In the defocusing case, the inverse scattering approach allows one to treat the case of localized data, and show that global solutions scatter at infinity, see for instance [9]. This can also be proved in a more robust way, without using inverse scattering, under the assumption that the initial data is small and localized, see [17] and references therein. Much less is known in terms of scattering for nonlocalized $L^{2}$ data. However, if more regularity is assumed for the data, then we have the following estimate due to Planchon- Vega [26], see also the work of Colliander-Grillakis-Tzirakis [4]: (1.2) $\|u\|_{L^{6}_{t,x}}^{6}+\|\partial_{x}|u|^{2}\|_{L^{2}_{t,x}}^{2}\lesssim\|u_{0}\|_{L^{2}_{x}}^{3}\|u_{0}\|_{H^{1}_{x}}.$ This allows one to estimate the $L^{6}$ Strichartz norm of the solution, i.e. to prove some type of scattering or dispersive decay. This estimate was improved and extended to $L^{2}$ solutions as a corollary of the results in our previous paper [21]. Precisely, we have (1.3) $\|u\|_{L^{6}_{t,x}}^{2}+\|\partial_{x}|u|^{2}\|_{L^{2}_{t}(\dot{H}^{-\frac{1}{2}}_{x}+cL^{2}_{x})}\lesssim\|u_{0}\|_{L^{2}_{x}}^{2},\qquad c=\|u_{0}\|_{L^{2}_{x}}.$ On the other hand, the focusing problem admits small solitons, so the solutions cannot in general scatter at infinity. If in addition the initial data is localized, then one expects the solution to resolve into a superposition of (finitely many) solitons, and a dispersive part; this is called _the soliton resolution conjecture_ , and is known to hold in a restrictive setting, via the method of inverse scattering, see e.g. [2]. Our interest here is in focusing problems, but without any integrability assumptions, and even without assuming any conservation laws. The model we consider is similar to the one in [21], namely (1.4) $\left\\{\begin{aligned} &iu_{t}+u_{xx}=C(u,\bar{u},u)\\\ &u(0)=u_{0},\end{aligned}\right.$ where $u$ is a complex valued function, $u:\mathbb{R}\times\mathbb{R}\rightarrow\mathbb{C}$. Here $C$ is a trilinear translation invariant form, whose symbol $c(\xi_{1},\xi_{2},\xi_{3})$ can always be assumed to be symmetric in $\xi_{1},\xi_{3}$; see [21] for an expanded discussion of multilinear forms. The arguments $u,\bar{u}$ and $u$ of $C$ are chosen so that our equation (1.4) has the phase rotation symmetry, $u\to ue^{i\theta}$, as it is the case in many examples of interest. The symbol $c(\xi_{1},\xi_{2},\xi_{3})$ will be required to satisfy the following set of assumptions, which are similar to [21]: 1. (H1) Bounded and regular: (1.5) $|\partial_{\xi}^{\alpha}c(\xi_{1},\xi_{2},\xi_{3})|\leq c_{\alpha},\qquad\xi_{1},\xi_{2},\xi_{3}\in\mathbb{R},\,\mbox{ for every multi-index $\alpha$}.$ 2. (H2) Conservative: (1.6) $\Im c(\xi,\xi,\eta)=0,\qquad\xi,\eta\in\mathbb{R},\mbox{ where }\Im z=\mbox{imaginary part of }z\in\mathbb{C}.$ In addition to these two conditions, in [21] we have also employed a defocusing assumption, namely 1. (H3) Defocusing: (1.7) $c(\xi,\xi,\xi)\geq c_{0}>0,\qquad\xi\in\mathbb{R}\mbox{ and }c_{0}\in\mathbb{R^{+}}.$ Here one might think that we should require the opposite, namely 1. (H4) Focusing: (1.8) $-c(\xi,\xi,\xi)\geq c_{0}>0,\qquad\xi\in\mathbb{R}\mbox{ and }c_{0}\in\mathbb{R^{+}}.$ But as it turns out, no such assumption is needed here, as the result of this paper applies equally regardless of any sign condition; so we will simply drop it. Using the same assumptions (H1), (H2) as in [21] is convenient here because it will allow us to reuse a good part of the analysis there, up to the point where the defocusing assumption is needed. Repeating a similar comment in [21], one should view both our choices of the Schrödinger dispersion relation and the uniform bounds in (H1) not as fundamental, but rather as a balance between the generality of the result on one hand, and a streamlined exposition on the other hand. This choice places our model in the semilinear class, but just barely so. The simplest example of such a trilinear form $C$ is of course $C=\pm 1$, which corresponds to the classical one-dimensional cubic NLS problem. But this problem has too much structure, in particular it is completely integrable, and also globally well-posed in $L^{2}$. At the other end, both our use of the linear Schrödinger operator and the boundedness condition (H1) are non-optimal, and we hope to relax both of these restrictions in subsequent work. ### 1.2. The main result Our main result asserts that long time well-posedness holds for our problem for small $L^{2}$ data. In addition, our solutions not only satisfy uniform $L^{2}$, but also space-time $L^{6}$ Strichartz estimates, as well as bilinear $L^{2}$ bounds, on appropriate time-scales: ###### Theorem 1. Consider the problem (1.4) where the cubic nonlinearity $C$ satisfies the assumptions (H1) and (H2). Assume that the initial data $u_{0}$ is small, $\|u_{0}\|_{L^{2}_{x}}\leq\epsilon\ll 1,$ Then the solution $u$ exists on on a time interval $I_{\epsilon}=[0,c\epsilon^{-8}]$ and has the following properties for every interval $I\subset I_{\epsilon}$ of size $|I|\leq\epsilon^{-6}$: 1. (i) Uniform $L^{2}$ bound: (1.9) $\|u\|_{L^{\infty}_{t}(I_{\epsilon};L^{2}_{x})}\lesssim\epsilon.$ 2. (ii) Strichartz bound: (1.10) $\|u\|_{L^{6}_{t,x}(I\times\mathbb{R})}\lesssim\epsilon^{\frac{2}{3}}.$ 3. (iii) Bilinear Strichartz bound: (1.11) $\|\partial_{x}(u\bar{u}(\cdot+x_{0}))\|_{L^{2}_{t}(I;H_{x}^{-\frac{1}{2}})}\lesssim\epsilon^{2},\qquad x_{0}\in\mathbb{R}.$ The local well-posedness in $L^{2}$ for the problem (1.4) was already proved in [21], so the emphasis here and later in the proof is on the lifespan bound and the long time estimates in the theorem. We remark that the intermediate time-scale $\epsilon^{-6}$ does not have an intrinsic meaning from a scaling perspective, but is instead connected to the unit frequency-scale which is implicit in (H1), and which motivates the frequency decomposition on the unit frequency-scale which is used in the proof of both the present result and the earlier result in [21]. One could also use a smaller frequency-scale for this decomposition, which in turn corresponds to a smaller size for $|I|$. This is however not needed in the proof of the $\epsilon^{-8}$ result, so, in order to avoid cluttering the theorem, we omit the details. But the interested reader should see Remark 1.1 below. A natural question to ask is whether this result is optimal. On one hand, the bounds (1.10) and (1.11) are sharp for the cubic NLS, and likely for any focusing flow (i.e. which satisfies (H4) at least in some region); this is discussed in Section 5. But the $\epsilon^{-8}$ lifespan bound is not optimal for any flow satisfying our hypotheses. Indeed, if for instance the $L^{2}$ norm is conserved (as is the case for the focusing NLS) then global well- posedness follows. However, we conjecture that ###### Conjecture 3. The result in Theorem 1 is sharp for generic focusing flows satisfying our hypotheses. It it also interesting to see how our theorem applies to the focusing cubic NLS problem. There we can also consider large data simply by scaling. Global well-posedness in $L^{2}$ is relatively straightforward there, but some of the estimates we prove are new: ###### Theorem 2. Consider the focusing 1-d cubic NLS problem (1.1)(-) with $L^{2}$ initial data $u_{0}$. Then the global solution $u$ satisfies the following bounds in all intervals $I$ with size $|I|>\|u_{0}\|_{L^{2}_{x}}^{-4}$: 1. (i) Uniform $L^{2}$ bound: (1.12) $\|u\|_{L^{\infty}_{t}L^{2}_{x}}\lesssim\|u_{0}\|_{L^{2}_{x}}.$ 2. (ii) Strichartz bound: (1.13) $\|u\|_{L^{6}_{t,x}(I\times\mathbb{R})}\lesssim\|u_{0}\|_{L^{2}_{x}}(I\|u_{0}\|_{L^{2}_{x}}^{4})^{\frac{1}{6}}.$ 3. (iii) Bilinear Strichartz bound: (1.14) $\|\partial_{x}|u|^{2}\|_{L^{2}_{t}(I;\dot{H}_{x}^{-\frac{1}{2}}+cL^{2}_{x})}\lesssim\|u_{0}\|_{L^{2}_{x}}^{2},\qquad c^{2}=\|u_{0}\|_{L^{2}_{x}}^{2}(|I|\|u_{0}\|_{L^{2}_{x}}^{4}).$ Here the Strichartz estimates are fairly easy to obtain directly, but the bilinear Strichartz bounds are new. Returning to our discussion after Theorem 1, the numerology in this application helps clarify the earlier comment about the choice of the time-scales in Theorem 1: ###### Remark 1.1. In the context of Theorem 1, a scaling argument shows that the bounds (1.13) and (1.14) hold for all intervals $I$ so that $\epsilon^{-4}\leq|I|\leq\epsilon^{-6}$. One may gain further insights into our result for focusing problems by comparing it with our earlier result in [21]. ###### Theorem 3 ([21]). Under the above assumptions (H1), (H2) and (H3) on the symbol of the cubic form $C$, small initial data $\|u_{0}\|_{L^{2}_{x}}\leq\epsilon\ll 1,$ yields a unique global solution $u$ for (1.4), which satisfies the following bounds: 1. (i) Uniform $L^{2}$ bound: (1.15) $\|u\|_{L^{\infty}_{t}L^{2}_{x}}\lesssim\epsilon.$ 2. (ii) Strichartz bound: (1.16) $\|u\|_{L^{6}_{t,x}}\lesssim\epsilon^{\frac{2}{3}}.$ 3. (iii) Bilinear Strichartz bound: (1.17) $\|\partial_{x}(u\bar{u}(\cdot+x_{0}))\|_{L^{2}_{t}H_{x}^{-\frac{1}{2}}}\lesssim\epsilon^{2},\qquad x_{0}\in\mathbb{R}.$ One may observe here that the estimates are similar in the two cases, and the only difference is the time-scale on which the estimates hold: in the defocusing case this is global, while in the focusing case it is finite and depends on the solution size. For this reason, the proofs of Theorem 3 and (1) are closely related, and we will take advantage of this within the proof. For the convenience of the reader, we also recall the main ideas in the proof of the last theorem in [21], which are equally employed here: _1\. Energy estimates via density flux identities._ This is a classical idea in pde’s, and particularly in the study of conservation laws. The novelty in [21] is that this analysis is carried out in a nonlocal setting, where both the densities and the fluxes involve translation invariant multilinear forms. The densities and the fluxes are not uniquely determined here, so careful choices need to be made. _2\. The use of energy corrections._ This is an idea originally developed in the context of the so called I-method [5] or more precisely the second generation I-method [8], whose aim was to construct more accurate almost conserved quantities. In [21] this idea is instead implemented at the level of density-flux identities. _3\. Interaction Morawetz bounds._ These were originally developed in the context of the three-dimensional NLS problems by Colliander-Keel-Stafillani- Takaoka-Tao in [6], and have played a fundamental role in the study of many nonlinear Schrödinger flows, see e,g. [7, 27], and also for one-dimensional quintic flows in the work of Dodson [11, 12]. Our take on this is somewhat closer to the one-dimensional approach of Planchon-Vega [26], though recast in the setting and language of nonlocal multilinear forms. _4\. Tao’s frequency envelope method._ This is used as a way to accurately track the evolution of the energy distribution across frequencies. Unlike the classical implementation relative to dyadic Littlewood-Paley decompositions, in [21] we adapt and refine this notion for lattice decompositions instead. This is also very convenient as a bootstrap tool, see e.g. Tao [29], [30] but with the added twist of also bootstrapping bilinear Strichartz bounds, as in the authors’ paper [20]. ### 1.3. An outline of the paper To a large extent, the proof of our main result mirrors the proof of the global result in the defocusing case in [21]. The primary difference is in how the $L^{6}$ Strichartz norms are handled, both globally and in a frequency localized setting. Section 2 reviews two of the main ideas in [21], namely the construction of modified density-flux identities for the mass/momentum in a frequency localized setting, as well as the interaction Morawetz identities associated to those density-flux relations. The proof of the main $L^{6}$ Strichartz bounds and the bilinear $L^{2}$ estimates is done in a frequency localized setting, using a bootstrap argument based on a frequency decomposition on the unit scale, where the components are measured using a maximal frequency envelope, another notion introduced in [21]. This is described in Section 3, and leads to dispersive bounds on the $\epsilon^{-6}$ time-scale. It is within this argument where the $L^{6}$ norms are treated differently from the defocusing case. In order to advance from the $\epsilon^{-6}$ to the $\epsilon^{-8}$ time-scale it suffices to propagate the $L^{2}$ bound (i.e. the mass) on the larger time- scale. However the mass is not a conserved quantity, so instead it is better to propagate the bounds for the modified mass. This analysis is carried out in Section 4. Finally, in the last section of the paper we discuss the optimality of our result, or rather the optimality of the $L^{6}$ and the bilinear $L^{2}$ estimates on the $\epsilon^{-6}$ time-scale. This is done by considering the obvious enemies, namely the solitons, in the focusing NLS context. ### 1.4. Acknowledgements The first author was supported by the Sloan Foundation, and by an NSF CAREER grant DMS-1845037. The second author was supported by the NSF grant DMS-2054975 as well as by a Simons Investigator grant from the Simons Foundation. ## 2\. Density-flux and interaction Morawetz identities A key role in the proof of the results in both [21] and in the present paper is played by the approximate conservation laws for the mass and the momentum. Rather than considering them directly, we instead consider several improvements: * • the conservation laws are written in density-flux form, rather than as integral identities, where both the densities and the fluxes are multilinear forms. * • we improve the accuracy of these conservation laws by using well chosen quartic corrections for both the densities and the fluxes, with $6$-linear errors. * • we use these densities and fluxes not only globally in frequency, but also in a frequency localized setting. The aim of this section is to provide an overview of these density-flux identities, following the set-up of [21]. We conclude the section with an overview of the interaction Morawetz identities obtained in [21] from the above density flux identities. ### 2.1. Resonances and multilinear forms A key role in our analysis is played by four wave resonances. Given three input frequencies $(\xi_{1},\xi_{2},\xi_{3})$ in the cubic nonlinearity $C$, the output is at frequency $\xi_{4}=\xi_{1}-\xi_{2}+\xi_{3}.$ The three wave interaction is resonant if $\xi_{4}^{2}=\xi_{1}^{2}-\xi_{2}^{2}+\xi_{3}^{2}.$ To rewrite these relations in a symmetric fashion we use the notations $\Delta^{4}\xi=\xi_{1}-\xi_{2}+\xi_{3}-\xi_{4},\qquad\Delta^{4}\xi^{2}=\xi_{1}^{2}-\xi_{2}^{2}+\xi_{3}^{2}-\xi_{4}^{2},$ The first expression is Galilean invariant but not the second, which is why we also use the adjusted, Galilean invariant expression ${\tilde{\Delta}}^{4}\xi^{2}=\Delta^{4}\xi^{2}-2\xi_{avg}\Delta^{4}\xi,$ where $\xi_{avg}$ represents the average of the four frequencies. With these notations, the resonant set is described as $\mathcal{R}=\\{[\xi]=(\xi_{1},\xi_{2},\xi_{3},\xi_{4});\ \Delta^{4}\xi=0,\Delta^{4}\xi^{2}=0\\}$ which can be explicitly characterized as $\mathcal{R}=\\{[\xi];(\xi_{1},\xi_{3})=(\xi_{2},\xi_{4})\\}$ Quadruples in the resonant set can be described by two parameters, namely (2.1) $\displaystyle{\delta\xi^{\text{hi}}}=$ $\displaystyle\ \max\\{|\xi_{1}-\xi_{2}|+|\xi_{3}-\xi_{4}|,|\xi_{1}-\xi_{4}|+|\xi_{3}-\xi_{2}|\\},$ $\displaystyle{\delta\xi^{\text{med}}}=$ $\displaystyle\ \min\\{|\xi_{1}-\xi_{2}|+|\xi_{3}-\xi_{4}|,|\xi_{1}-\xi_{4}|+|\xi_{3}-\xi_{2}|\\},$ These distance parameters are carefully defined so that they can also be used outside the resonant set to characterize frequency quadruples. This is very useful in the density-flux relations later on. ### 2.2. The conservation of mass The starting point of the analysis in [21] is to consider energy estimates for our flow from a density-flux perspective. In the simplest case, we start with the mass density $M(u,\bar{u})=|u|^{2},$ whose linear flux is given by the momentum $P(u,\bar{u})=2i\Im(u\partial_{x}\bar{u}).$ These can be viewed as translation invariant bilinear forms with symbols $m(\xi,\eta)=1,\qquad p(\xi,\eta)=\xi+\eta.$ Integrating the densities we obtain the familiar mass and momentum, $\mathbf{M}(u)=\int_{\mathbb{R}}M(u,\bar{u})\,dx,\quad\mathbf{P}(u)=\int_{\mathbb{R}}M(u,\bar{u})\,dx.$ At the nonlinear level, we have the density flux relation $\partial_{t}M(u,\bar{u})=\partial_{x}P(u,\bar{u})+C^{4}_{m}(u,\bar{u},u,\bar{u}),$ where $C^{4}_{m}$ is a symmetric translation invariant real multilinear form which depends on our cubic nonlinearity $C$. A key observation in [21] is that, under the conservative assumption (H2) on the nonlinearity, the mass density admits a quartic correction which is accurate to sixth order. Precisely, the correction has the form (2.2) $M^{\sharp}(u)=M(u)+B^{4}_{m}(u,\bar{u},u,\bar{u}),$ and the associated density flux relation has the form (2.3) $\partial_{t}M^{\sharp}(u)=\partial_{x}(P(u)+R^{4}_{m}(u,\bar{u},u,\bar{u}))+R^{6}_{m}(u,\bar{u},u,\bar{u},u,\bar{u}).$ with suitable translation invariant multilinear forms $R^{4}_{m}$ and $R^{6}_{m}$. The corresponding integral corrected mass is $\mathbf{M}^{\sharp}(u)=\int_{\mathbb{R}}M^{\sharp}(u)\,dx.$ The choice of the symbols $b^{4}_{m}$ and $r^{4}_{m}$ above depends on the behavior of $c^{4}_{m}$ near the resonant set $\mathcal{R}$, precisely they have to solve the division problem (2.4) $c^{4}_{m}+i\Delta^{4}\xi^{2}\,b^{4}_{m}=i\Delta^{4}\xi\,r^{4}_{m}.$ This is possible due to our condition (H2), which implies that $c^{4}_{m}=0$ on $\mathcal{R}$. But the choice is not uniquely determined, so it is important to make a good one, i.e. which insures good symbol bounds. To achieve this, in [21] we decompose the phase space for frequency quadruples into three overlapping regions which can be separated using cutoff functions which are smooth on the unit scale: 1. i) The full division region, $\Omega_{1}=\\{{\delta\xi^{\text{med}}}\lesssim 1\\},$ which represents a full unit size neighbourhood of the resonant set $\mathcal{R}$. 2. ii) The region $\Omega_{2}=\\{1+|\Delta^{4}\xi|\ll{\delta\xi^{\text{med}}}\\},$ where ${\tilde{\Delta}}^{4}\xi^{2}$ must be elliptic, ${\tilde{\Delta}}^{4}\xi^{2}\approx{\delta\xi^{\text{hi}}}{\delta\xi^{\text{med}}}$, and thus we will favor division by the symbol ${\tilde{\Delta}}^{4}\xi^{2}$. 3. iii) The region $\Omega_{3}=\\{1\ll{\delta\xi^{\text{med}}}\lesssim|\Delta^{4}\xi|\\},$ we will instead divide by $\Delta^{4}\xi$; this is compensated by the relatively small size of this region. Since we will also need this in the present paper, we state the result in the following ###### Proposition 2.1 ([21]). Assume that the nonlinearity $C$ satisfies the conditions (H1), (H2). Then there exist multilinear forms $B^{4}_{m}$ $R^{4}_{m}$ and $R^{6}_{m}$ so that the relation (2.3) holds for solutions $u$ to (1.4), and so that the symbols $c^{4}_{m}$ and $r^{4}_{m}$ satisfy the bounds 1. i) Size (2.5) $\displaystyle|\partial^{\alpha}r^{4}_{m}|\lesssim$ $\displaystyle\ \frac{1}{\langle{\delta\xi^{\text{med}}}\rangle},$ $\displaystyle|\partial^{\alpha}b^{4}_{m}|\lesssim$ $\displaystyle\ \frac{1}{\langle{\delta\xi^{\text{hi}}}\rangle\langle{\delta\xi^{\text{med}}}\rangle}.$ 2. ii) Support: $b^{4}$ is supported in $\Omega_{1}\cup\Omega_{2}$ and $\tilde{r}^{4}$ is supported in $\Omega_{1}\cup\Omega_{3}$. In addition, we have the fixed time bound (2.6) $\left|\int_{\mathbb{R}}B^{4}_{m}(u,u,u,u)\,dx\right|\lesssim\|u\|_{L^{2}_{x}}^{4}.$ We make several remarks concerning this result: * • No bound for the symbol $r^{6}_{m}$ is provided in the proposition. This is because $r^{6}_{m}$ is obtained directly as the contribution of $C$ to the time derivative of $b^{4}_{m}$. * • This proposition is a consequence of Lemma 4.1 and Lemma 7.1 in [21]. * • The defocusing hypothesis (H3), which is used for the final result in [21], plays no role here. * • The estimate (2.6) shows that the mass correction is perturbative for as long as the solution $u$ remains small in $L^{2}$. * • A similar analysis applies for the momentum conservation law. But the counterpart of the above Proposition for the momentum is less useful directly, and instead it is used in [21] only in a frequency localized context. ### 2.3. Frequency localized density-flux identities Instead of relying on the more standard Littlewood-Paley decomposition, the analysis in [21] uses a frequency decomposition on the unit scale in frequency. Given any integer $j$, we will use localized versions of the mass in a unit size region around $j$. More generally, for an interval $A\subset{\mathbb{Z}}$, we use a symbol $a_{0}$ which is frequency localized in a unit neighbourhood of $A$. At the level of bilinear forms, we will use the symbol $a(\xi,\eta)=a_{0}(\xi)a_{0}(\eta).$ Corresponding to such $a$ we define quadratic localized mass, momentum and energy densities by the symbols $m_{a}(\xi,\eta)=a(\xi,\eta),\qquad p_{a}(\xi,\eta)=(\xi+\eta)a(\xi,\eta),\qquad e_{a}(\xi,\eta)=(\xi+\eta)^{2}a(\xi,\eta).$ The associated bilinear forms are denoted by $M_{a}$, $P_{a}$, respectively $E_{a}$. If $A=\\{j\\}$ then we simply replace the subscript $a$ with $j$. It is shown in [21], again under the assumptions (H1) and (H2), that one may find quartic corrections $M^{\sharp}_{a}$ and $P^{\sharp}_{a}$ of the form (2.7) $M^{\sharp}_{a}(u)=M_{a}(u)+B^{4}_{m,a}(u,\bar{u},u,\bar{u}),$ (2.8) $P^{\sharp}_{a}(u)=P_{a}(u)+B^{4}_{p,a}(u,\bar{u},u,\bar{u}),$ for which we obtain density-flux identities akin to (2.3), namely (2.9) $\partial_{t}M^{\sharp}_{a}(u)=\partial_{x}(P_{a}(u)+R^{4}_{m,a}(u))+R^{6}_{m,a}(u),$ and (2.10) $\partial_{t}P^{\sharp}_{a}(u)=\partial_{x}(E_{a}(u)+R^{4}_{p,a}(u))+R^{6}_{p,a}(u).$ We will consider these relations together with their Galilean shifts obtaining relations of the form (2.11) $(\partial_{t}-2\xi_{0}\partial_{x})M^{\sharp}_{a}(u)=\partial_{x}(P_{a,\xi_{0}}(u)+R^{4}_{m,a,\xi_{0}}(u))+R^{6}_{m,a,\xi_{0}}(u),$ respectively (2.12) $(\partial_{t}-2\xi_{0}\partial_{x})P^{\sharp}_{a,\xi_{0}}(u)=\partial_{x}(E_{a,\xi_{0}}(u)+R^{4}_{p,a,\xi_{0}}(u))+R^{6}_{p,a,\xi_{0}}(u).$ These correspond to the algebraic division relations (2.13) $c^{4}_{m,a}+i\Delta^{4}(\xi-\xi_{0})^{2}b^{4}_{m,a}=i\Delta^{4}\xi r^{4}_{m,a,\xi_{0}},$ respectively (2.14) $c^{4}_{p,a,\xi_{0}}+i\Delta^{4}(\xi-\xi_{0})^{2}b^{4}_{p,a,\xi_{0}}=i\Delta^{4}\xi r^{4}_{p,a,\xi_{0}},$ where $c^{4}_{m,a}$ and $c^{4}_{p,a,\xi_{0}}$ are the density-flux sources corresponding to uncorrected mass, respectively momentum. The symbols above are connected in the obvious way. Precisely, we have (2.15) $r^{4}_{m,a,\xi_{0}}=r^{4}_{m,a}-2\xi_{0}b^{4}_{m,a},\qquad$ and (2.16) $P^{\sharp}_{a,\xi_{0}}=P^{\sharp}_{a}-2\xi_{0}M^{\sharp}_{a},\qquad b^{4}_{p,a,\xi_{0}}=b^{4}_{p,a}-2\xi_{0}b^{4}_{m,a},$ and finally (2.17) $r^{4}_{p,a,\xi_{0}}=r^{4}_{p,a}-2\xi_{0}b^{4}_{p,a}-2\xi_{0}r^{4}_{m,a,\xi_{0}}.$ To use these density flux relations we need to have appropriate bounds for our symbols: ###### Proposition 2.2. Let $J\subset\mathbb{R}$ be an interval of length $r$, and $d(\xi_{0},J)\lesssim r$. Assume that $a$ is supported in $J\times J$, with bounded and uniformly smooth symbol. Then the relations (2.11) and (2.12) hold with symbols $b^{4}_{m,a}$, $b^{4}_{p,a,\xi_{0}}$, $r^{4}_{m,a,\xi_{0}}$ and $r^{4}_{p,a,\xi_{0}}$ which can be chosen to have the following properties: 1. i) Support: they are all supported in the region where at least one of the frequencies is in $J$. 2. ii) Size: (2.18) $|b^{4}_{m,a}|\lesssim\frac{1}{\langle{\delta\xi^{\text{hi}}}\rangle\langle{\delta\xi^{\text{med}}}\rangle},\qquad|b^{4}_{p,a,\xi_{0}}|\lesssim\frac{r}{\langle{\delta\xi^{\text{hi}}}\rangle\langle{\delta\xi^{\text{med}}}\rangle},$ (2.19) $\displaystyle|r^{4}_{m,a,\xi_{0}}|\lesssim$ $\displaystyle\ \frac{1}{\langle{\delta\xi^{\text{med}}}\rangle}1_{\Omega_{1}\cup\Omega_{2}}+\frac{r}{\langle{\delta\xi^{\text{hi}}}\rangle\langle{\delta\xi^{\text{med}}}\rangle}1_{\Omega_{1}\cup\Omega_{3}},$ $\displaystyle|R^{4}_{p,a,\xi_{0}}|\lesssim$ $\displaystyle\ \frac{r}{\langle{\delta\xi^{\text{med}}}\rangle}1_{\Omega_{1}\cup\Omega_{2}}+\frac{r^{2}}{\langle{\delta\xi^{\text{hi}}}\rangle\langle{\delta\xi^{\text{med}}}\rangle}1_{\Omega_{1}\cup\Omega_{3}}.$ 3. iii) Regularity: similar bounds hold for all derivatives. This is Proposition 4.3 in [21]. ### 2.4. Interaction Morawetz identities One way the desity flux relations above are used is to obtain more accurate bounds for the mass propagation. However, another way to use them is via interaction Morawetz identities, which yield bilinear estimates for the interaction of different frequency portions of the solutions, or even the self-interaction of unit frequency portions of solutions. Given two frequency intervals $A$ and $B$ and corresponding mass/momentum modified density associated to these intervals for two solutions $u,v$ to (1.4), in [21] we define the associated _interaction Morawetz functional_ by (2.20) $\mathbf{I}_{AB}=\iint_{x>y}M^{\sharp}_{a}(u)(x)P^{\sharp}_{b,\xi_{0}}(v)(y)-P^{\sharp}_{a,\xi_{0}}(u)(x)M^{\sharp}_{b}(v)(y)\,dxdy,$ where we add several remarks: * • In applications, for the second solution $v$ we will simply choose a spatial translation of the solution $u$. * • A velocity parameter $\xi_{0}$ is introduced on the right, but the functional does not depend on $\xi_{0}$. This parameter plays a role, however, in estimating the time derivative of $\mathbf{I}_{AB}$, and will be chosen to be close to the sets $A$ and $B$. * • Such interaction functionals will be used in two settings: 1. (i) the separated case $1\ll|A|\approx|B|\approx dist(A,B)$, and 2. (ii) the self-interaction case $|A|+|B|+dist(A,B)\lesssim 1$. The time derivative of the interaction Morawetz functional is computed using the frequency localized mass density-flux (2.11) and the corresponding momentum density-flux (2.12). This yields a localized _interaction Morawetz identity_ , (2.21) $\frac{d}{dt}\mathbf{I}_{AB}=\mathbf{J}^{4}_{AB}+\mathbf{J}^{6}_{AB}+\mathbf{J}^{8}_{AB}+\mathbf{K}^{8}_{AB},$ where the terms on the right are described as follows: 1. a) the quartic contribution $\mathbf{J}^{4}_{a}$ is $\mathbf{J}^{4}_{AB}=\int_{\mathbb{R}}M_{a}(u)(x)E_{b,\xi_{0}}(v)(x)+M_{b}(v)(x)E_{a,\xi_{0}}(u)(x)-2P_{a,\xi_{0}}(u)(x)P_{b,\xi_{0}}(v)(x)\,dx,$ and is used to capture bilinear $L^{2}$ bounds. 2. b) The sixth order term $\mathbf{J}^{6}_{a}$ has the form $\mathbf{J}^{6}_{AB}=\int_{\mathbb{R}}-(P_{a,\xi_{0}}B^{4}_{p,b,\xi_{0}}+P_{b,\xi_{0}}R^{4}_{m,a,\xi_{0}})+(M_{a}R^{4}_{p,b,\xi_{0}}+E_{b,\xi_{0}}B^{4}_{m,a,\xi_{0}})-\text{symmetric}\,dx,$ where in the symmetric part we interchange both the indices $a,b$ and the functions $u,v$. In the defocusing case its symbol has a favorable sign on the diagonal, and is used to capture the $L^{6}$ bound in the self-interaction case. Here, it is estimated perturbatively. 3. c) The eight-linear term $\mathbf{J}^{8}_{AB}=\int_{\mathbb{R}}-R^{4}_{m,a,\xi_{0}}B^{4}_{p,b,\xi_{0}}+B^{4}_{m,a,\xi_{0}}R^{4}_{p,b,\xi_{0}}-\text{symmetric}\,dx.$ plays a perturbative role. 4. d) The $8^{+}$-linear term $\mathbf{K}^{8}_{AB}$ has the form $\mathbf{K}^{8}_{AB}=\iint_{x>y}M^{\sharp}_{a}(x)R^{6}_{p,a,\xi_{0}}+P^{\sharp}_{a,\xi_{0}}R^{6}_{m,a,\xi_{0}}-\text{symmetric}\,dxdy.$ This is a double integral, which also includes a $10$-linear term. It is also estimated perturbatively. ## 3\. Strichartz and bilinear $L^{2}$ bounds To obtain estimates for $L^{2}$ solutions $u$ to (1.4), a unit scale frequency decomposition is needed, $u=\sum_{k\in{\mathbb{Z}}}u_{k},\qquad u_{k}:=P_{k}u,$ where $P_{k}$ are multipliers with smooth symbols localized in a unit neighbourhood of the integer frequency $k$. To measure the components $u_{k}$ we use a frequency envelope $\\{c_{k}\\}\in\ell^{2}$ in order to transfer bounds from the initial data to the solutions. These frequency envelopes are chosen to satisfy an adapted version of the slowly varying property, originally introduced by Tao [29] in the context of dyadic decompositions. Such a property is needed in order to account for the nonlinear leakage of energy between nearby frequencies. We recall the frequency envelope set-up in [21], associated to lattice decompositions: ###### Definition 3.1. A lattice frequency envelope $\\{c_{k}\\}$ is said to have the maximal property if (3.1) $Mc\leq Cc,$ where $Mc$ represents the maximal function of $c$. Here $C$ is a universal constant. Frequency envelopes that have this property will be called _admissible_. The proof of the $L^{6}$ Strichartz and the bilinear $L^{2}$ bounds will be phrased as a bootstrap argument relative to an admissible frequency envelope for the initial data. ###### Theorem 4. Let $u\in C[0,T;L^{2}]$ be a solution for the equation (1.4) with initial data $u_{0}$ which has $L^{2}$ size at most $\epsilon$. Let $\\{c_{k}\\}$ be a maximal frequency envelope for the initial data in $L^{2}$, also of size $\epsilon$, $\|u_{0k}\|_{L^{2}}\lesssim\epsilon c_{k}.$ Assume that (3.2) $T\ll\epsilon^{-6}.$ Then the solution $u$ satisfies the following bounds in $[0,T]$: 1. (i) Uniform frequency envelope bound: (3.3) $\|u_{k}\|_{L^{\infty}_{t}L^{2}_{x}}\lesssim\epsilon c_{k},$ 2. (ii) Localized Strichartz bound: (3.4) $\|u_{k}\|_{L_{t,x}^{6}}\lesssim(\epsilon c_{k})^{\frac{2}{3}},$ 3. (iii) Localized Interaction Morawetz: (3.5) $\|\partial_{x}|u_{k}|^{2}\|_{L^{2}_{t,x}}\lesssim\epsilon^{2}c_{k}^{2},$ 4. (iv) Transversal bilinear $L^{2}$ bound: (3.6) $\|\partial_{x}(u_{A}{\bar{u}}_{B}(\cdot+x_{0}))\|_{L^{2}_{t,x}}\lesssim\epsilon^{2}c_{A}c_{B}\,\langle dist(A,B)\rangle^{\frac{1}{2}},$ for all $x_{0}\in\mathbb{R}$ whenever $|A|+|B|\lesssim\langle dist(A,B)\rangle$. This proposition mirrors a similar result in Section 7 of [21], with two key differences. On one hand we drop the defocusing assumption (H3), and on the other hand we limit the size of the time interval in (3.2). The $L^{6}$ Strichartz estimates (1.10) and the bilinear $L^{2}$ bounds in (1.11) follow from the estimates in the above proposition, by the same arguments as those in Section 8 of [21]. To prove this theorem, we make a bootstrap assumption where we assume the same bounds but with a worse constant $C$, as follows: 1. (i) Uniform frequency envelope bound, (3.7) $\|u_{k}\|_{L^{\infty}_{t}L^{2}_{x}}\lesssim C\epsilon c_{k},$ 2. (ii) Localized Strichartz bound, (3.8) $\|u_{k}\|_{L^{6}_{t,x}}\lesssim C(\epsilon c_{k})^{\frac{2}{3}},$ 3. (iii) Localized Interaction Morawetz, (3.9) $\|\partial_{x}|u_{k}|^{2}\|_{L^{2}_{t,x}}\lesssim C\epsilon^{2}c_{k}^{2},$ 4. (iv) Transversal Interaction Morawetz, (3.10) $\|\partial_{x}(u_{k_{1}}\bar{u}_{k_{2}}(\cdot+x_{0}))\|_{L^{2}_{t,x}}\lesssim C\epsilon^{2}c_{k_{1}}c_{k_{2}}\langle k_{1}-k_{2}\rangle^{\frac{1}{2}}$ uniformly for all $x_{0}\in\mathbb{R}$. Then we seek to improve the constant in these bounds. The gain will come from the fact that the $C$’s will always come paired either with extra $\epsilon$ factors, or with $T\epsilon^{6}$ factors. To a large extent the proof largely repeats the proof of the corresponding result in [21], so we review the steps and expand the portion where the argument differs here. _STEP 1:_ The proof of the energy bound (3.3). This is done by integrating the density flux relation (2.9) for the localized mass $\mathbf{M}_{k}(u)$. The argument in [21] applies unchanged. _STEP 2:_ The proof of the $L^{6}$ Strichartz bound (3.3). In [21] this is proved together with (3.5) by integrating the interaction Morawetz identity applied to the pair $(u_{k},u_{k})$. However, in the focusing case the sign of the $u_{k}^{6}$ contribution changes, and the same argument no longer applies. Instead, here we will estimate the $L^{6}$ norm directly using an interpolation argument. Precisely, for the function $v_{k}=|u_{k}|^{2}$, using (3.7) and (3.9), we have the bounds $\|v_{k}\|_{L^{\infty}_{t}L^{1}_{x}}\lesssim C^{2}\epsilon^{2}c_{k}^{2},$ respectively $\|v_{k}\|_{L^{2}_{t}\dot{H}^{1}_{x}}\lesssim C^{2}\epsilon^{2}c_{k}^{2}.$ We interpolate between the two estimates in homogeneous Sobolev spaces, with weights $5/9$ and $4/9$. We obtain $\|v_{k}\|_{L^{\frac{9}{2}}_{t}\dot{W}^{\frac{4}{9},\frac{9}{7}}_{x}}\lesssim\|v_{k}\|_{L^{\infty}_{t}L^{1}_{x}}^{\frac{5}{9}}\|v_{k}\|_{L^{2}_{t}\dot{H}^{1}_{x}}^{\frac{4}{9}}\lesssim C^{2}\epsilon^{2}c_{k}^{2}.$ By Sobolev embeddings $\dot{W}^{\frac{4}{9},\frac{9}{7}}$ embeds in $L^{3}$ so, using also Hölder’s inequality with respect to time, we obtain $\|v_{k}\|_{L^{3}_{t}L_{x}^{3}}\lesssim T^{\frac{1}{9}}\|v_{k}\|_{L^{\frac{9}{2}}L^{3}}\lesssim T^{\frac{1}{9}}\|v_{k}\|_{L^{\frac{9}{2}}\dot{W}^{\frac{4}{9},\frac{9}{7}}}\lesssim C^{2}T^{\frac{1}{9}}\epsilon^{2}c_{k}^{2}=C^{2}(T\epsilon^{6})^{\frac{1}{9}}\epsilon^{\frac{4}{3}}c_{k}^{2}.$ This implies the desired Strichartz bound (3.4) under the time constraint (3.2). _STEP 3:_ The proof of the bilinear $L^{2}$ bound (3.6). This is again exactly as in [21], by applying the interaction Morawetz identity to the functions $(u_{A},u_{B}(\cdot+x_{0}))$. We note that the $L^{6}$ bound is used as an input in this proof, and the defocusing assumption (H2) is not needed. Here we view (3.5) as a special case of (3.6), and no longer in conjunction with (3.3). ## 4\. Long time energy estimates The frequency envelope bounds in Theorem 4 provide us with uniform energy bounds on the $\epsilon^{-6}$ time-scale, and so they do not suffice in order to prove our main result in Theorem 1, which is on the $\epsilon^{-8}$ time- scale. To fill in this gap, we will prove a direct energy estimate on the $\epsilon^{-8}$ time-scale. Precisely, we will show the following: ###### Proposition 4.1. Let $u$ be an $L^{2}$ solution for (1.4) in a time interval $[0,T]$. Assume that the initial data for (1.4) satisfies (4.1) $\|u_{0}\|_{L_{x}^{2}}\leq\epsilon\ll 1,$ and that $T\ll\epsilon^{-8}$. Then the solution $u$ satisfies (4.2) $\|u\|_{L_{t}^{\infty}(0,T;L_{x}^{2})}\leq 4\epsilon.$ Once we have this proposition, a continuity argument based on the local well- posedness for (1.4) in $L^{2}$ implies Theorem 1. ###### Proof. It suffices to prove that the conclusion holds assuming that we have the bootstrap assumption (4.3) $\|u\|_{L^{\infty}(0,T;L^{2})}\leq 8\epsilon.$ Instead of tracking directly the mass $\mathbf{M}(u)=\|u\|_{L_{x}^{2}}^{2}$, it is more efficient to work with the modified mass $\mathbf{M}^{\sharp}(u)=\int_{\mathbb{R}}M^{\sharp}(u)\,dx.$ In view of the bound (2.6), we have $\mathbf{M}(u)=\mathbf{M}^{\sharp}(u)+O(\epsilon^{4}).$ Since $\epsilon\ll 1$, we have (4.4) $\mathbf{M}(u)(0)\leq 2\epsilon^{2},$ and it suffices to show that (4.5) $\mathbf{M}(u)(t)\leq 4\epsilon^{2},\qquad t\in[0,T].$ In view of Proposition 2.3, the time evolution of $\mathbf{M}^{\sharp}$ is given by (4.6) $\frac{d}{dt}\mathbf{M}^{\sharp}(u)=\int_{\mathbb{R}}R^{6}_{m}(u,{\bar{u}},u,{\bar{u}},u,{\bar{u}})\,dx.$ To bound its growth, we use the following ###### Lemma 4.2. Assume that the bounds (3.3)-(3.6) for $u$ hold in a time interval $[0,T]$. Then we have (4.7) $\|R^{6}_{m}\|_{L^{1}_{t,x}([0,T]\times\mathbb{R})}\lesssim\epsilon^{4}.$ This is Lemma 7.3 in [21]. We apply this lemma on time intervals of size $\epsilon^{-6}$, where the bounds (3.3)-(3.6) hold in view of Theorem 4, and then add up the results. Then for $T>\epsilon^{-6}$ we get (4.8) $\|R^{6}_{m}\|_{L^{1}_{t,x}([0,T]\times\mathbb{R})}\lesssim\epsilon^{4}(T\epsilon^{6}).$ Hence for $T\ll\epsilon^{-8}$ we arrive at (4.9) $\|R^{6}_{m}\|_{L^{1}_{t,x}([0,T]\times\mathbb{R})}\ll\epsilon^{2}.$ This allows us to obtain (4.5) from (4.4), thereby concluding the proof of the proposition. ∎ ## 5\. Optimality remarks As noted earlier, a key obstruction to global dispersive estimates in the focusing case is given by the potential existence of small solitons. Here we will restrict our attention to the simplest model, namely the focusing cubic NLS, and test the optimality of our estimates on solitons for this model. It is not so difficult to show that small solitons exist for our model whenever the focusing assumption (H4) is satisfied in some frequency region. All cubic NLS solitons are equivalent modulo scaling and Galilean transformations. Our bounds are Galilean invariant, so we set the soliton velocity to zero and we focus on scaling. Then the unit scale soliton has the form $u(x,t)=e^{it}Q(x),\qquad Q(x)=\operatorname{sech}(x).$ Rescaled to the frequency scale $\lambda$, this yields the solitons $u_{\lambda}(x,t)=e^{it\lambda^{2}}Q_{\lambda}(x),\qquad Q_{\lambda}(x)=\lambda Q(\lambda x),$ which has initial data size $\|u_{0}\|_{L_{x}^{2}}^{2}=\|Q_{\lambda}\|_{L^{2}_{x}}^{2}=\int_{\mathbb{R}}\lambda^{2}Q^{2}(\lambda x)\,dx\approx\lambda,$ so for Theorem 1 we will choose $\lambda=\epsilon^{2}$. On the other hand, $\|u_{\lambda}\|_{L^{6}_{t,x}(0,T,\mathbb{R})}^{6}=T\int_{\mathbb{R}}\lambda^{6}Q^{6}(\lambda x)\,dx\approx T\lambda^{5}.$ Then it is easily seen that we have approximate equality in (1.10) and (1.13). Similarly, we compute the bilinear Strichartz norm. Due to the bound from below on $|I|$, we have $c\gtrsim\lambda^{-\frac{1}{2}}$ whereas $u$ is concentrated at frequency $\lesssim\lambda$. Then $\|\partial_{x}|u_{\lambda}|^{2}\|_{L^{2}_{t}(0,T;\dot{H}^{\frac{1}{2}}_{x}+cL^{2}_{x})}^{2}\approx c^{-2}\|\partial_{x}|u_{\lambda}|^{2}\|_{L^{2}_{t,x}(0,T;L^{2})}^{2}=\lambda(T\lambda^{2})^{-1}T\int_{\mathbb{R}}\lambda^{6}\partial_{x}Q_{x}^{2}(\lambda x)\,dx\approx\lambda^{2}.$ This corresponds to having equality in (1.11) and (1.14). ## References * [1] Thomas Alazard and Jean-Marc Delort. Global solutions and asymptotic behavior for two dimensional gravity water waves. Ann. Sci. Éc. Norm. Supér. (4), 48(5):1149–1238, 2015. * [2] Michael Borghese, Robert Jenkins, and Kenneth D. T.-R. McLaughlin. Long time asymptotic behavior of the focusing nonlinear Schrödinger equation. Ann. Inst. H. Poincaré C Anal. Non Linéaire, 35(4):887–920, 2018. * [3] Jean Bourgain. A remark on normal forms and the “$I$-method” for periodic NLS. J. Anal. Math., 94:125–157, 2004. * [4] J. Colliander, M. Grillakis, and N. Tzirakis. Tensor products and correlation estimates with applications to nonlinear Schrödinger equations. Comm. Pure Appl. Math., 62(7):920–968, 2009. * [5] J. Colliander, M. Keel, G. Staffilani, H. Takaoka, and T. Tao. Almost conservation laws and global rough solutions to a nonlinear Schrödinger equation. Math. Res. Lett., 9(5-6):659–682, 2002. * [6] J. Colliander, M. Keel, G. Staffilani, H. Takaoka, and T. Tao. Global existence and scattering for rough solutions of a nonlinear Schrödinger equation on $\mathbb{R}^{3}$. Comm. Pure Appl. Math., 57(8):987–1014, 2004. * [7] J. Colliander, M. Keel, G. Staffilani, H. Takaoka, and T. Tao. Global well-posedness and scattering for the energy-critical nonlinear Schrödinger equation in $\mathbb{R}^{3}$. Ann. of Math. (2), 167(3):767–865, 2008. * [8] J. Colliander, M. Keel, G. Staffilani, H. Takaoka, and T. Tao. Resonant decompositions and the $I$-method for the cubic nonlinear Schrödinger equation on $\mathbb{R}^{2}$. Discrete Contin. Dyn. Syst., 21(3):665–686, 2008. * [9] Percy Deift and Xin Zhou. Long-time asymptotics for solutions of the NLS equation with initial data in a weighted Sobolev space. Comm. Pure Appl. Math., 56(8):1029–1077, 2003. Dedicated to the memory of Jürgen K. Moser. * [10] Jean-Marc Delort. Semiclassical microlocal normal forms and global solutions of modified one-dimensional KG equations. Ann. Inst. Fourier (Grenoble), 66(4):1451–1528, 2016. * [11] Benjamin Dodson. Global well-posedness and scattering for the defocusing, $L^{2}$ critical, nonlinear Schrödinger equation when $d=1$. Amer. J. Math., 138(2):531–569, 2016. * [12] Benjamin Dodson. Global well-posedness and scattering for the defocusing, mass-critical generalized KdV equation. Ann. PDE, 3(1):Paper No. 5, 35, 2017. * [13] Nakao Hayashi and Pavel I. Naumkin. Asymptotics for large time of solutions to the nonlinear Schrödinger and Hartree equations. Amer. J. Math., 120(2):369–389, 1998. * [14] Nakao Hayashi and Pavel I. Naumkin. Large time asymptotics for the fractional nonlinear Schrödinger equation. Adv. Differential Equations, 25(1-2):31–80, 2020. * [15] John K. Hunter and Mihaela Ifrim. Enhanced life span of smooth solutions of a Burgers-Hilbert equation. SIAM J. Math. Anal., 44(3):2039–2052, 2012. * [16] John K. Hunter, Mihaela Ifrim, Daniel Tataru, and Tak Kwong Wong. Long time solutions for a Burgers-Hilbert equation via a modified energy method. Proc. Amer. Math. Soc., 143(8):3407–3412, 2015. * [17] Mihaela Ifrim and Daniel Tataru. Global bounds for the cubic nonlinear Schrödinger equation (NLS) in one space dimension. Nonlinearity, 28(8):2661–2675, 2015. * [18] Mihaela Ifrim and Daniel Tataru. Two dimensional water waves in holomorphic coordinates II: Global solutions. Bull. Soc. Math. France, 144(2):369–394, 2016. * [19] Mihaela Ifrim and Daniel Tataru. The lifespan of small data solutions in two dimensional capillary water waves. Arch. Ration. Mech. Anal., 225(3):1279–1346, 2017. * [20] Mihaela Ifrim and Daniel Tataru. Well-posedness and dispersive decay of small data solutions for the Benjamin-Ono equation. Ann. Sci. Éc. Norm. Supér. (4), 52(2):297–335, 2019. * [21] Mihaela Ifrim and Daniel Tataru. Global solutions for 1D cubic defocusing dispersive equations: Part I. arXiv e-prints, page arXiv:2205.12212, May 2022. * [22] Mihaela Ifrim and Daniel Tataru. Testing by wave packets and modified scattering in nonlinear dispersive pde’s. arXiv e-prints, page arXiv:2204.13285, April 2022. * [23] Jun Kato and Fabio Pusateri. A new proof of long-range scattering for critical nonlinear Schrödinger equations. Differential Integral Equations, 24(9-10):923–940, 2011. * [24] Hans Lindblad, Jonas Lührmann, and Avy Soffer. Asymptotics for 1D Klein-Gordon equations with variable coefficient quadratic nonlinearities. Arch. Ration. Mech. Anal., 241(3):1459–1527, 2021. * [25] Hans Lindblad and Avy Soffer. Scattering and small data completeness for the critical nonlinear Schrödinger equation. Nonlinearity, 19(2):345–353, 2006. * [26] Fabrice Planchon and Luis Vega. Bilinear virial identities and applications. Ann. Sci. Éc. Norm. Supér. (4), 42(2):261–290, 2009. * [27] E. Ryckman and M. Visan. Global well-posedness and scattering for the defocusing energy-critical nonlinear Schrödinger equation in $\mathbb{R}^{1+4}$. Amer. J. Math., 129(1):1–60, 2007. * [28] Jalal Shatah. Normal forms and quadratic nonlinear Klein-Gordon equations. Comm. Pure Appl. Math., 38(5):685–696, 1985. * [29] Terence Tao. Global regularity of wave maps. II. Small energy in two dimensions. Comm. Math. Phys., 224(2):443–544, 2001. * [30] Terence Tao. Global well-posedness of the Benjamin-Ono equation in $H^{1}({\bf R})$. J. Hyperbolic Differ. Equ., 1(1):27–49, 2004.
# Salt-bearing disk candidates around high-mass young stellar objects Adam Ginsburg Department of Astronomy, University of Florida, P.O. Box 112055, Gainesville, FL, USA Brett A. McGuire Department of Chemistry, Massachusetts Institute of Technology, Cambridge, MA 02139 National Radio Astronomy Observatory, Charlottesville, VA 22903 Patricio Sanhueza National Astronomical Observatory of Japan, National Institutes of Natural Sciences, 2-21-1 Osawa, Mitaka, Tokyo 181-8588, Japan Department of Astronomical Science, SOKENDAI (The Graduate University for Advanced Studies), 2-21-1 Osawa, Mitaka, Tokyo 181-8588, Japan Fernando Olguin Institute of Astronomy and Department of Physics, National Tsing Hua University, Hsinchu 30013, Taiwan Luke T. Maud ESO Headquarters, Karl-Schwarzchild-Str 2 85748 Garching, Germany Kei E. I. Tanaka Center for Astrophysics and Space Astronomy, University of Colorado Boulder, Boulder, CO 80309, USA National Astronomical Observatory of Japan, National Institutes of Natural Sciences, 2-21-1 Osawa, Mitaka, Tokyo 181-8588, Japan Yichen Zhang Department of Astronomy, University of Virginia, Charlottesville, VA 22904, USA RIKEN Cluster for Pioneering Research, Wako, Saitama 351-0198, Japan Henrik Beuther Max Planck Institute for Astronomy, Königstuhl 17, 69117, Heidelberg, Germany Nick Indriolo AURA for the European Space Agency (ESA), Space Telescope Science Institute, 3700 San Martin Drive, Baltimore, MD 21218, USA ###### Abstract Molecular lines tracing the orbital motion of gas in a well-defined disk are valuable tools for inferring both the properties of the disk and the star it surrounds. Lines that arise only from a disk, and not also from the surrounding molecular cloud core that birthed the star or from the outflow it drives, are rare. Several such emission lines have recently been discovered in one example case, those from NaCl and KCl salt molecules. We studied a sample of 23 candidate high-mass young stellar objects (HMYSOs) in 17 high-mass star- forming regions to determine how frequently emission from these species is detected. We present 5 new detections of water, NaCl, KCl, PN, and SiS from the innermost regions around the objects, bringing the total number of known briny disk candidates to 9. Their kinematic structure is generally disk-like, though we are unable to determine whether they arise from a disk or outflow in the sources with new detections. We demonstrate that these species are spatially coincident in a few resolved cases and show that they are generally detected together, suggesting a common origin or excitation mechanism. We also show that several disks around HMYSOs clearly do not exhibit emission in these species. Salty disks are therefore neither particularly rare in high-mass disks, nor are they ubiquitous. ††software: The source code underlying this work are available from github at https://github.com/keflavich/saltmining/releases/tag/accepted-2022-11-04. This work used CARTA (Comrie et al., 2021), Jupyter notebooks (Kluyver et al., 2016). numpy (van der Walt et al., 2011; Harris et al., 2020), scipy (Virtanen et al., 2020), astropy (Astropy Collaboration et al., 2013, 2018; The Astropy Collaboration et al., 2022), spectral-cube (Ginsburg et al., 2019b), radio- beam (Koch et al., 2021), and CASA-6 (https://casa.nrao.edu/casadocs/casa-5.6.0/introduction/casa6-installation- and-usage). matplotlib (Hunter, 2007), astroquery (Ginsburg et al., 2019c), and pyspeckit (Ginsburg et al., 2022). doublespace ## 1 Introduction Circumstellar accretion disks develop around forming new stars. While the presence of disks around low-mass stars has been clear for decades, we have only definitively demonstrated that accretion disks exist around high-mass young stellar objects (HYMSOs) in the last decade (e.g., Beltrán & de Wit, 2016; Maud et al., 2017; Ilee et al., 2018; Motogi et al., 2019; Johnston et al., 2020; Sanna et al., 2021; Moscadelli et al., 2021). One limiting factor in the detection and subsequent characterization of HYMSOs has been the lack of molecular emission lines that arise from the disk, but are not confused with or absorbed by the surrounding molecular cloud. A select few lines have recently been discovered that are uniquely produced in the disks of some HMYSOs. Emission from salt molecules has been detected in the surroundings of four stars in three star-forming regions: Orion Source I (Ginsburg et al., 2019a), G17.64+0.16 (Maud et al., 2019), and a pair in IRAS 16547-4247 (Tanaka et al., 2020) (hereafter, SrcI, G17, and I16547, respectively). Of these, SrcI and G17 are confirmed disks, while the I16547 pair remain candidates, and in all cases the salt emission comes from zones within $\lesssim 100$ au of the central source. Each of these sources also exhibits H2O emission from both the (candidate) disk and a slightly more extended region, and thus we dub these objects ‘brinaries.’ Salts are detected in the atmospheres of AGB and post-AGB stars, giving some clues as to the physical conditions needed to produce them. Salts have been detected in CRL2688 (Highberger et al., 2003), IRAS+10216 (Cernicharo & Guelin, 1987), IK Tauri, VY Canis Majoris (Milam et al., 2007; Decin et al., 2016), VX Sgr (mentioned in passing in Danilovich et al., 2021), and OH231.8+44.2 (Sánchez Contreras et al., 2022). Sánchez Contreras et al. (2022) discovered a brinary disk surrounding the post-AGB mass transfer system OH231.8+4.2, highlighting the similarity between birth and death among moderately massive stars. However, there are also some non-detections in well- surveyed sources, including the S-type AGB stars W Aql and $\pi^{1}$ Gru (Homan et al., 2020; Danilovich et al., 2021). The development of line lists in the infrared, and more complete lists in the radio, has been driven in part by interest in salts as constituents of the atmospheres of hot planets, in which these species are predicted to be important components of upper cloud layers (Barton et al., 2014). There is therefore some motivation to understand where salts occur in disks and in or on dust grains. Studies of (post)-AGB stars and models of planetary atmospheres provide some clues about the physical conditions required to produce gas-phase salts. NaCl is expected to change states (from solid to gas or vice-versa) around $\sim$500–600 K at planetary atmospheric pressure (i.e., 1 bar; Woitke et al., 2018). Decin et al. (2016) suggest that it comes off of grains at 100–300 K based on the detection locations in supergiant stars. These studies provide a first hint about where NaCl may come from if it precipitates out of cooling gas, though it remains unclear if the same mechanisms apply in and around YSO disks. Motivated by the detection of salts in a few HMYSOs with ALMA in recent years, we present a first search for salt-bearing disks in the ALMA archives. In Section 2, we describe the ALMA observations we analyze. In Section 3, we describe the analysis approach (3.1.1) and detections (3.3). We discuss the chemical correlations observed in the sample and possible reasons for (non)detections in Section 4, then conclude. ## 2 Data & Sample Selection We utilize archival and new data from several projects. We select data sets that have high angular resolution ($\lesssim 0.1$ ″) targeting high-mass star- forming regions. Most of the sources in our sample come from the Digging into the Interior of Hot Cores with ALMA (DIHCA; PI: Sanhueza) program, which is surveying $\sim 30$ candidate disks. The main aims of the DIHCA survey are to study the interior of massive hot cores to determine whether they form high- mass stars collapsing monolithically or by fragmenting into binary (multiple) systems and to search for accretion disks around high-mass stars. The DIHCA targets were selected from the literature as regions with previous interferometric (e.g., SMA) observations and having an expected flux of $>$0.1 Jy at 230 GHz. All clumps follow the empirical threshold for high-mass star formation suggested by Kauffmann & Pillai (2010). We only examine a small subset of the DIHCA sample here because not all data were available as of March 2022. Because our sample is not uniformly selected, we can say little about completeness; we can only search for general trends. Nevertheless, the trends we find are interesting and suggest that observing a more uniform sample in the future would be productive. DIHCA observations of G335, G333.23, NGC6334, IRAS16562, G34.43mm1, G29.96, and G351.77 were obtained during July 2019, and observations of G5.89, G11.92, IRAS18089, and W33A were taken both in September 2017 and July 2019. The observations were reduced using CASA (v5.4.0-70; McMullin et al., 2007). The data were then phase self-calibrated in three steps with decreasing solution intervals and the continuum subtracted following the procedure of Olguin et al. (2021). Dirty cubes were produced with a Briggs weighting robust parameter of 0.5 using the CASA tclean task. We use dirty image cubes for expediency, so it is possible significant improved images of these objects could be obtained, though we note that the targeted lines are generally faint and would not be affected by cleaning with typical clean parameters. Table 1: Observation Summary Field | Source Name | $\theta_{\rm maj}$ | $\theta_{\rm maj}$ | $\theta_{\rm min}$ | PA | $\sigma$ | $f(>5\sigma)$ | $N_{\rm beams}$ | $\sigma_{\rm avg}$ | $v$ ref. line | Distance ---|---|---|---|---|---|---|---|---|---|---|--- | | $\mathrm{AU}$ | ′′ | ′′ | ∘ | $\mathrm{K}$ | | | $\mathrm{K}$ | | $\mathrm{kpc}$ Orion | SrcI | 20 | 0.050 | 0.039 | 72.7 | 19.3 | 0.05 | 43 | 2.9 | NaCl J=18-17 v=0 | 0.4 S255IR | SMA1 | 40 | 0.036 | 0.027 | 5.6 | 46.9 | 0.04 | 19 | 10.6 | no clear disk | 1.6k G351.77 | mm12 | 50 | 0.027 | 0.022 | -89.3 | 6.0 | 0.08 | 24 | 1.2 | H2O | 2.2c G351.77 | mm2 | 50 | 0.027 | 0.022 | -89.3 | 6.0 | 0.05 | 21 | 1.3 | H2O | 2.2c G351.77 | mm1 | 50 | 0.027 | 0.022 | -89.3 | 6.0 | 0.14 | 41 | 0.9 | H2O | 2.2c G17 | G17 | 50 | 0.038 | 0.022 | 44.4 | 10.0 | 0.01 | 75 | 1.2 | H2O | 2.2c NGC6334I | mm1d | 50 | 0.074 | 0.041 | 62.8 | 19.6 | 0.10 | 20 | 4.3 | H2O | 1.3d NGC6334I | mm1b | 50 | 0.074 | 0.041 | 62.8 | 19.6 | 0.23 | 27 | 3.7 | H2O | 1.3d NGC6334I | mm2b | 50 | 0.074 | 0.041 | 62.8 | 19.6 | 0.07 | 7 | 7.0 | H2O | 1.3d NGC6334IN | SMA1b/d | 50 | 0.074 | 0.042 | 63.4 | 17.5 | 0.16 | 278 | 1.0 | H2O | 1.3d NGC6334IN | SMA6 | 50 | 0.074 | 0.042 | 63.4 | 17.5 | 0.06 | 22 | 3.7 | H2O | 1.3d IRAS18162 | GGD27 | 90 | 0.099 | 0.066 | -89.0 | 6.9 | 0.01 | 36 | 1.1 | SO $6_{5}-5_{4}$ | 1.3a IRAS18089 | I18089-1732 | 100 | 0.064 | 0.045 | 65.8 | 18.4 | 0.33 | 57 | 2.4 | CH3OH | 2.3f G34.43 | mm1 | 110 | 0.105 | 0.069 | 59.5 | 9.2 | 0.35 | 53 | 1.3 | no clear disk | 1.6l IRAS16562 | G345.4938+01.4677 | 120 | 0.106 | 0.053 | 81.6 | 10.5 | 0.04 | 25 | 2.1 | H30$\alpha$ | 2.3g I16547 | A | 130 | 0.065 | 0.043 | 35.7 | 23.2 | 0.19 | 7 | 8.4 | H2O | 2.9c I16547 | B | 130 | 0.065 | 0.043 | 35.7 | 23.2 | 0.18 | 5 | 9.5 | H2O | 2.9c G5.89 | mm15 | 130 | 0.063 | 0.043 | 66.3 | 19.2 | 0.00 | 85 | 2.1 | H2O | 3.0b G335 | ALMA1 | 140 | 0.066 | 0.041 | 47.3 | 20.4 | 0.29 | 118 | 1.9 | no clear disk | 3.3i W33A | mm1-main | 170 | 0.103 | 0.067 | -86.4 | 6.5 | 0.17 | 51 | 0.9 | H2O | 2.6k G333.23 | mm1 | 220 | 0.069 | 0.041 | 54.1 | 18.4 | 0.02 | 70 | 2.2 | SO $6_{5}-5_{4}$ | 5.3h G333.23 | mm2 | 220 | 0.069 | 0.041 | 54.1 | 18.4 | 0.08 | 45 | 2.7 | SO $6_{5}-5_{4}$ | 5.3h G11.92 | mm1 | 220 | 0.101 | 0.066 | -86.9 | 6.6 | 0.16 | 103 | 0.7 | SO $6_{5}-5_{4}$ | 3.3e G29.96 | submm1 | 530 | 0.099 | 0.071 | 65.0 | 9.6 | 0.22 | 369 | 0.5 | no clear disk | 7.4j Observation properties. The ‘Field’ name indicates the region of the ALMA pointing. The ‘Source Name’ is the identifier of the disk candidate examined. $\theta$ gives the beam parameters, with $\theta_{maj}$ [au] providing the physical size using the adopted distance. $\sigma$ is the average noise level of the field, which is averaged down by $N_{beams}^{1/2}$ to give $\sigma_{\rm avg}$, the noise level in the stacked spectrum. $f(>5\sigma)$ is the fraction of the stacked spectrum that is above five times $\sigma_{avg}$; it is used as a diagnostic of the line crowding in the spectrum covering 219.2–220.8 GHz, which is high for complex-molecule-rich regions. The $v$ reference line is the line used to create a velocity map to produce stacked spectra. Section 3.1.4 provides additional details. Distances come from the following sources: a Añez-López et al. (2020), b Sato et al. (2014); Fernández-López et al. (2021), c Beuther et al. (2017), d Chibueze et al. (2014), e Sato et al. (2014), f Xu et al. (2011), g Guzmán et al. (2020), h Whitaker et al. (2017), i Peretto et al. (2013), j Kalcheva et al. (2018), k Reid et al. (2014), l Kurayama et al. (2011), m Ilee et al. (2018), We use the SrcI data from Ginsburg et al. (2018, 2019a). We use G17 data from Maud et al. (2019). We use G351.77 data from both DIHCA and Beuther et al. (2019). We use I16547 data from both DIHCA and Tanaka et al. (2020). For each of these data sets, we refer the reader to the cited papers for the data reduction description. We include summary statistics of these observations in Table 1. We give additional details about the physical resolution, distance to the targets, and the line used as a velocity guide (see §3.1) in Table 1. Finally, we use proprietary data toward Sh 255-IR SMA1 (hereafter S255IR) from project 2019.1.00492.S (PI Ginsburg). We use the archive-produced data products, which were cleaned with the ALMA pipeline, and imaged them with CASA 6.4.3.4. The images were cleaned to a depth of 10 mJy using Briggs robust=0 weighting. ## 3 Results and Analysis We search for lines of NaCl, KCl, H2O, SiS, and H30$\alpha$ in each of the target pointings (see Table 2). Since each target was selected for having a high luminosity or a strong HMYSO disk candidate beforehand, we used the literature identification of existing sources as our starting point. We cut out cubes centered on the brightest continuum source in the ALMA images (except G5.89; see §B.4). We also searched fainter continuum sources, selecting small sub-regions around each of the compact continuum peaks that could plausibly contain disks. Because the source selection is based on a by- eye examination of the data over a limited field of view (in most cases, only the inner 5–10 arcseconds of the ALMA field of view was imaged), the sample presented here has unknown completeness - the conclusions we draw will therefore be only suggestive, not conclusive. Most of our line detections come from stacked spectra of resolved disk-like objects. We describe in Section 3.1 the line stacking approach used to obtain higher signal-to-noise ratio spectra that represent average values over the candidate disk. Section 3.3 describes the detections in individual sources and shows some of the extracted images and spectra. Additional images and spectra are displayed in the appendices. The main result is the detection of the ‘brinary’ lines toward the 9 sources (of which 3 are tentative detections) shown in Figures 1 and 2. These initial figures show moment maps of the NaCl lines as described in Section 3.1.3. Figure 1: Moment-0 (integrated intensity) images of the resolved sources in NaCl lines. For SrcI (top-left), this is the integrated intensity of the NaCl $J$=18-17 $v$=0 line. For the remainder, G17 (top right), G351 mm1 (bottom left), and NGC 6334I mm1b (bottom right), these are the average of the $J$=18-17 and $J$=17-16 transitions of both the $v$=0 and $v$=1 states. The coordinates are given in RA/Dec offset from the central position specified under the abscissa. The scalebars show physical sizes as labeled. The ellipses in the corners show the full-width half-maximum beam ellipse. Figure 2: Integrated intensity (moment-0) images of the NaCl stacked lines toward G351mm2 (top left), G351mm12 (top middle), NGC 6334I mm2b - which is only a tentative detection (top right), I16547A (bottom left) and B (bottom center), and W33A (bottom right). These disk candidates are either unresolved or marginally resolved. ### 3.1 Line stacking extraction From source candidate identification from the continuum data, the analysis forks down two different paths. We start by searching by eye for emission associated with the H2O line, which is quite bright in SrcI, in §3.1.1, which we then use as a kinematic reference. If water is not detected, we use a different line as our kinematic reference as described in §3.1.2. #### 3.1.1 Water-driven analysis If the water line is detected, we use it to create a ‘velocity map’ following this procedure: 1. 1. Cut out a cube containing the region around the estimated central $v_{\rm LSR}\pm 20\>\textrm{km~{}s}^{-1}$ in velocity and encompassing only the candidate disk region in space. 2. 2. Create a peak intensity map of the line in that region. 3. 3. Create a threshold mask including only the (by-eye) estimated significant emission in the peak intensity map. Use one iteration of binary erosion and three to seven iterations of binary dilation to remove isolated bright noise pixels and fill back in the mask.111Binary dilation refers to mathematical morphology operations in which any pixel having value False and a neighbor with the value True is set to True. Binary erosion is the inverse operation. 4. 4. Create a volumetric threshold mask including only pixels above the estimated noise level. Then, use one iteration of binary erosion followed by one to three iterations of binary dilation to fill in the mask. As in the previous step, this step is to eliminate isolated bright pixels, but because it is in 3D, a different threshold can be adopted. 5. 5. Create a moment-1 (intensity-weighted velocity) map of the spatially and volumetrically masked data cube. The erosion and dilation steps are performed to exclude isolated bright pixels and to maximally include all pixels associated with source emission. We then use the velocity map to stack the spectra obtained across the disk candidate. Each spectrum from each spatial pixel in the cube that has a measured velocity is shifted such that the line peak is moved to 0 km s-1. The spectra are then averaged to produce the stacked spectrum. This stacking process assumes that all spectra through the candidate disk will have similar peak intensity and width but different central velocities; this assumption held well in the SrcI spectrum (Ginsburg et al., 2018). We assume that the kinematics of our selected line are the same as the lines of interest; this assumption is justified by position-velocity diagrams in §3.3.2. We demonstrate the advantage of this line stacking approach in Figure 3. The signal-to-noise in targeted lines is substantially increased, and faint lines appear that are otherwise missed. In the example figure, the most obvious case is KCl $v$=4 $J$=31-30, which is not apparent in either the peak intensity or average spectrum, but is strongly evident in the stacked spectrum. Those sources that are best-resolved - in our sample, G351 and G17 - show the most improvement. Figure 3: Demonstration of the utility of stacking analysis toward G17 in one spectral window. The black spectrum is stacked by shifting each individual spectrum to match the location of the peak intensity of the H2O line; it is the same in both panels. The top panel shows the peak intensity spectrum taken over the same area as the stacked spectrum, and the bottom shows the mean spectrum in blue. It is clear that the stacked spectrum has clearer features and better signal-to-noise ratios than either the peak or average spectrum. We label the resulting averaged spectra with known NaCl and KCl transitions and selected other lines, including those of SiO, SiS, H2O, H30$\alpha$, and several prominent other molecules (see Table 2). We then use these plots to populate the detection table, Table 3. We regard the lines as firm detections only if they: * • Are prominent, bright, and relatively isolated (e.g., H2O, as in Figure 4) * • Exhibit an appropriately broad linewidth (H30$\alpha$ is expected to have $\sigma_{v}\gtrsim 5$ km s-1 because it comes from hot plasma ($T\sim 10^{4}$ K), while molecular species are expected to be narrower, coming from gas at $T<1000$ K). For species for which we label the vibrationally excited transitions, we consider only the $v$=0, $v$=1, and $v$=2 levels, as higher levels are expected to be weaker if present. There are several cases where a compelling detection of one line of a species (NaCl or KCl) is detected, but an adjacent state is not detected. For example, the $J$=18-17 $v$=1 line is detected, but the $J$=16-17 $v$=1 line is not. In these cases, we regard the detection as tentative and note it in Table 3 with an asterisk. These were cases in which the line may still be present, but may be hidden by either blending with neighboring species or absorption by a line in the surrounding core. We report detections qualitatively rather than quantitatively because the detections are generally obvious and high signal-to-noise. When a detection is ambiguous, it is not because of high noise but because of confusion with neighboring lines (e.g., 41KCl $J$=31-30 is confused with NaCl $v=1$ $J=18-17$) or absorption by non-disk material (e.g., KCl $v$=4 $J$=31-30). #### 3.1.2 Water nondetection-driven analysis The path for water non-detections is less linear. If we are unable to clearly identify emission in the H2O line, we search for other lines to use as the basis for stacking. We first create a simple averaged spectrum over the selected disk candidate’s emitting region by examining several different lines in the cube. We then search for other plausible guiding lines, focusing on those that exhibit gradients in the direction expected given known outflows (i.e., we look at lines rotating perpendicular to outflows). We look at SiS, SO, and, when truly desperate, $\textrm{CH}_{3}\textrm{OH}$ lines. If we are able to identify a reasonably disk-like line from among these lines, we use it to produce a velocity map as above (§3.1.1). If, after this search, we are still unable to find a line that traces disk-like kinematics, we remove the source from further consideration. For both water- and non-water-driven stacking, we report the achieved noise level in Table 1. #### 3.1.3 Cube stacking When NaCl is detected, it is generally seen in multiple transitions. The NaCl $v$=0 and $v$=1 $J$=18-17 and $v$=1 and $v$=2 $J$=17-16 transitions are present in the DIHCA observational setups and are not too badly contaminated by neighboring lines. We therefore create ‘stacked’ NaCl cubes by cutting out cubes centered on each of those transitions, smoothing them to a common beam, regridding them to a common spectral resolution, and then averaging the cubes. We use these stacked cubes for further analysis of the NaCl lines, producing both moment-0 images and position-velocity diagrams (except for SrcI, for which the signal-to-noise ratio in individual lines was high enough to produce moment maps without stacking). The stacked cubes have higher signal-to-noise than the individual cubes and de-emphasize contaminant lines that are adjacent to the target lines, since the contaminants arise at different relative velocities for each NaCl line (for example, while the 41KCl 31-30 line is 13 km s-1 to the red of the NaCl $v=1$ $J=18-17$ line, there is no line 13 km s-1 to the red of the three other NaCl lines included here). This stacking approach is not strictly necessary for use or analysis of the NaCl lines, but it aesthetically improves the resulting images and makes visual inspection and comparison more straightforward. We created these cubes whether or not we first noted an NaCl detection in the spectrum. In cases where no detection was apparent in the stacked spetrum, we nevertheless checked the stacked NaCl cubes to see if extended emission at the expected velocity was apparent. No additional detections were obtained through this approach. #### 3.1.4 COMs We identify the presence of complex organic molecules (COMs) in the spectrum in a very broad-strokes manner. We do not identify specific species, though we note that $\textrm{CH}_{3}\textrm{OH}$ and CH3CN are commonly detected, but instead characterize the spectra by the richness of the ‘line forest.’ For each observation, in the spectral range 219.2-220.8 GHz (which is COM-rich and includes the CH3CN $J=12-11$ ladder), we measure the per-pixel noise ($\sigma_{\rm cube}$) by obtaining the standard deviation over the full field of view over all pixels; this approach slightly overestimates the noise because it includes signal in the noise estimate. For image cubes that were not already continuum-subtracted, we estimate, and then subtract, the continuum by performing pixel-by-pixel sigma clipping to 3-$\sigma$, then taking the median across the spectral axis (i.e., as in Sánchez-Monge et al., 2018). Then, for each extracted region around a candidate disk, we average the spectra within that region, then determine what fraction of the spectrum exceeds five times the expected noise level, where the expected noise level is $\sigma_{\rm cube}n_{\rm beams}^{-1/2}$, where $n_{\rm beams}$ is the number of beams included in the averaging area. Note that we only search for COMs in emission; absorption by C- and O-bearing species is seen toward most sources, but is not directly associated with the candidate disk, instead, it likely comes from the surrounding envelope or molecular cloud. Table 1 gives a summary of these statistics in addition to general properties of the data. ### 3.2 Line Identification We briefly discuss the key lines used for identification of NaCl, KCl, PN, SiS, and SiO in this section. The summary of lines considered is in Table 2. The NaCl $v$=0 and $v$=1 $J$=18-17 and $v$=1 and $v$=2 $J$=17-16 transitions are present in the DIHCA observational setups. The $v=1$, $J=18-17$ line is very close to the 41KCl 31-30 line, but the latter can be ruled out as a contaminant because the 41KCl 29-28 is also included in the observations. In Orion SrcI, the peak intensity ratio was NaCl $v=1$ $J=18-17$ $\approx 5\times$ 41KCl 29-28. Table 2: Summary of spectroscopic lines used in this analysis Line Name | Frequency | EU ---|---|--- | $\mathrm{GHz}$ | $\mathrm{K}$ SiS v=1 12-11 | 216.757603 | 1138.75 SiO v=0 5-4 | 217.104980 | 31.26 KCl v=4 29-28 | 217.228912 | 1733.21 41KCl 29-28 | 217.543178 | 156.71 SiS v=0 12-11 | 217.817644 | 67.95 NaCl v=2 17-16 | 217.979967 | 1128.38 H2CO $3_{0,3}-2_{0,2}$ | 218.222192 | 20.96 HC3N 24-23 | 218.324788 | 1084.99 CH3OH $4_{2,2}-3_{1,2}$ | 218.440063 | 1422.49 KCl v=3 29-28 | 218.579708 | 1345.06 C18O 2-1 | 219.560354 | 15.81 NaCl v=1 17-16 | 219.614936 | 614.51 HNCO $10_{10}-9_{9}$ | 219.798274 | 58.02 KCl v=2 29-28 | 219.936113 | 671.38 SO $6_{5}-5_{4}$ | 219.949440 | 34.98 K37Cl v=2 30-29 | 221.078543 | 948.25 41KCl v=1 31-30 | 231.088150 | 572.25 K37Cl 31-30 | 231.218839 | 177.68 H30$\alpha$ | 231.900928 | - C30$\alpha$ | 232.016632 | - KCl v=4 31-30 | 232.163002 | 1755.14 41KCl 31-30 | 232.499840 | 178.67 NaCl v=1 18-17 | 232.509950 | 625.67 Si33S 13-12 | 232.628545 | 78.16 H2O v2=1 $5_{5,0}-6_{4,3}$ | 232.686700 | 3461.91 K37Cl v=4 32-31 | 232.907553 | 1739.18 PN v=1 J=5-4 | 233.271800 | 1937.30 KCl v=3 31-30 | 233.605698 | 1367.12 NaCl v=0 18-17 | 234.251912 | 106.85 SiS v=1 13-12 | 234.812968 | 865.40 PN J=5-4 | 234.935663 | 33.83 KCl v=2 31-30 | 235.055578 | 687.16 K37Cl v=2 31-30 | 235.768235 | 970.53 SiS v=0 13-12 | 235.961363 | 79.28 Lines covered by one or more observations in this work. The frequency and upper state energy levels are pulled from Splatalogue and refer either to CDMS, SLAIM, or JPL values. For KCl v=3 and v=4 values, EU is drawn from the modified version of the Barton et al. (2014) line list used in Ginsburg et al. (2019a). ### 3.3 Salt detections We detect salts in 9 sources (of which 3 are tentative) in 6 regions. Figure 1 gives an overview of those objects that are spatially resolved. In the following sections, we describe the salt-bearing sources in more detail: G17 (§3.3.1), G351 mm1, mm2, and mm12 (§3.3.2), W33A (§3.3.4), NGC6334I mm1b and mm2b (§3.3.3), and I16547 A and B (§3.3.5). The sixth region and disk, Orion SrcI, is not discussed in detail here because the same analysis was already done in Ginsburg et al. (2019a), but it is included in the discussion section below. Table 3: Summary of detections, tentative detections, and non-detections of target species in the source sample Source | disk | H2O | NaCl | KCl | SiO | RRL | COMs | SiS | SO | PN ---|---|---|---|---|---|---|---|---|---|--- Orion SrcI | yes | yes | yes | yes | yes | no | no | yes | yes | ? G17 | yes | yes | yes | yes | yes | yes | no | yes | yes* | no I16547A | yes-c | yes | yes* | no | yes | no | yes | yes | yes | yes I16547B | yes-c | yes | yes* | no | yes | no | yes | yes | yes | yes G351.77mm1 | yes-c | yes | yes | no | yes | no* | no | yes | no | yes G351.77mm2 | unres | yes | yes* | no | no | yes | no | no | no | no* G351.77mm12 | unres | yes | yes* | no | yes | no | no | yes | no | yes W33A mm1-main | unres | yes* | yes* | no | yes | yes* | yes | yes* | yes | yes NGC6334Imm1b | yes-c | yes | yes | yes | yes | no | yes* | yes | no* | yes* G5.89 mm15 | cont | no | no | no | ? | no | no | no | no | no IRAS18162 GGD27 | yes | no | no | no | ? | no | yes | no | yes | no NGC6334IN SMA6 | no | yes* | no | no | no | no | yes | no | no | yes* NGC6334IN SMA1b/d | no | no* | no | no | no | no | yes | no | no | no* G11.92mm1 | yes | no | no | no | yes | no | yes | yes* | yes | yes* IRAS18089 I18089-1732 | yes-c | no | no | no | yes | no | yes | yes* | yes | no* IRAS16562 G345.4938+01.4677 | no | no | no | no | yes | yes | ext | no | ext | no G333.23mm1 | no | no | no | no | no* | no | yes | no | yes | yes* G333.23mm2 | yes* | no | no | no | no* | no | yes | no | yes | yes* G335 ALMA1 | no | no | no | no | yes | no | yes | no | no* | no G29.96 submm1 | no | no | no | no | yes | no | yes | no* | no* | no* G34.43mm1 | no | no | no | no | no* | no* | yes | no | yes | no* S255IR SMA1 | no* | no* | no | no | yes | yes | yes | no | yes | no NGC6334Imm1d | yes* | yes* | no | no | no | no | no | no | no | yes* NGC6334Imm2b | unres | yes | no* | no | yes | no | yes | yes | yes | yes* We use ‘?’ to indicate “not observed”, ‘yes’ for a definitive detection, ‘no’ for definitive non-detections, ‘yes*’ for tentative detections, and ‘no*’ for tentative nondetections, where this uncertainty can either be from line confusion or low signal-to-noise, and ‘ext’ for those associated with the envelope but not the disk. These all refer to detections in emission; COMs are seen in absorption toward many sources, but we do not consider these. For SiO, we do not attempt to distinguish between SiO from the outflow and from the disk; in Orion SrcI, we know that SiO is present in both. For the ‘disk’ column, we either answer ‘yes’ for a clear disk detection from literature kinematic characterization, ‘yes-c’ for a candidate disk toward which a kinematic signature consisten with rotation has been observed, but for which the kinematics have not yet been confirmed to be Keplerian, ‘continuum’ if a disk-like (linear, $\lesssim 300$ au long) feature is seen in the continuum, ‘unresolved’ if the targeted source is too small for us to say, and ‘no’ if neither of the above hold; ‘no’ does not indicate that no disk is present, merely that we did not identify one. In many cases, we suspect a disk must be present because there is an outflow, but we say ‘no’ if we can’t see it. #### 3.3.1 G17 The G17 disk is the best resolved source in our sample after SrcI (Fig 1; Maud et al., 2019). It is a confirmed Keplerian disk (Maud et al., 2019). Because of its high signal-to-noise and well-resolved structure, we investigate it in somewhat more detail than the other sources in the sample. Figure 4 shows the stacked spectrum based on the water line as described in Section 3.1.1. Figure 4: Stacked spectra toward G17 from the Maud et al. (2019) data set. The stacking was based on the H2O line. Line IDs are shown. Different colors are used for targeted species with multiple transitions in-band: orange for SiS, blue for KCl, red for NaCl, magenta for H30$\alpha$, purple for PN, and green for H2O. The remaining species, with only one transition marked, are shown in black. The salt detections toward G17 resemble that toward SrcI, with both NaCl and KCl tracing the same rotational structure. Figure 5 shows position-velocity diagrams perpendicular to the outflow axis measured by Maud et al. 2018 at position angle $\theta\approx 135^{\circ}$ based on the large-scale CO outflow. Maud et al. (2019) measured the disk position angle to be 25.9∘, which traces the direction of maximum gradient in the H2O disk and through the emission peak of the continuum structure. This angle is nearly perpendicular to the large-scale outflow; we adopt this angle as the disk PA. Maud et al. (2019) measured the disk inclination to be $i=40\pm 4^{\circ}$ from the axis ratio of the continuum image, so we adopt that inclination when overplotting Keplerian curves. As with SrcI, both the water and salt lines trace out orbits consistent with Keplerian rotation around a high-mass star. Figure 6 shows integrated intensity (moment-0) and intensity-weighted velocity (moment-1) maps of the G17 disk. Figure 5: Position-velocity diagrams extracted across the G17 disk. (top left) KCl grayscale with NaCl contours (top right) NaCl grayscale with Keplerian rotation curves drawn for a 15, 30, 40 $M_{\odot}$ star with a $i=40^{\circ}$ disk (Maud et al., 2019) in orange, green, and blue, respectively. (bottom left) Water (H2O) in grayscale with thin NaCl contours. In all cases, the velocity gradient matches that of the disk identified in Maud et al. (2019). (bottom right) Position-velocity diagram of the H30$\alpha$ line toward G17, with NaCl contours overlaid. The H30$\alpha$ emission is clearly confined to within the NaCl disk. While there is a hint of a velocity gradient in the H30$\alpha$ line, it is unclear whether this gradient traces rotation. Figure 6: Moment-0 (integrated intensity) and moment-1 (intensity-weighted velocity) images of stacked NaCl (left) and H2O (right) for G17. The red and blue arrows indicate the outflow direction from Maud et al. (2017). The dashed gray line shows the direction the position-velocity diagram was extracted from (Figure 5). G17 chemically resembles SrcI in several ways: the only lines seen directly toward the source are H2O, NaCl, KCl, SiS, and SiO. There is little ‘contamination’ in the spectrum from COMs. As in SrcI, KCl lines are detected at about half the peak brightness of NaCl lines with similar $E_{\rm U}$; no $v$=0, $v$=1, or $v$=2 transitions of KCl are covered by the observations (except KCl v=2 J=29-28, which is blended with SO $6_{5}-5_{4}$ and cannot be clearly identified). We measure enough lines to produce both rotational and vibrational diagrams for KCl, but only a rotational diagram for NaCl (Figure 7). These plots provide temperature and column density measurements of the target molecules if the observed transitions are in local thermodynamic equilibrium (LTE). As in SrcI, the rotational temperatures are much cooler ($T_{rot}\sim 35-60$ K) than the vibrational temperatures ($T_{vib}\sim 900$ K), indicating that non-LTE effects are important. We have yet to determine the underlying mechanism, but several possibilities are discussed in Ginsburg et al. (2019a). We defer further discussion of the excitation to a future work in which we will integrate additional transitions. Figure 7: Rotation diagrams from the fitted lines toward G17. (left) KCl rotation diagram (right) NaCl rotation diagram A key difference between SrcI and G17 is that G17 exhibits radio recombination line (RRL) emission, while SrcI does not. Figure 5 shows the H30$\alpha$ RRL in grayscale with NaCl contours on top, showing that the RRL comes from a smaller region contained within the NaCl-bearing disk. Both SrcI and G17 have central ‘holes’ in which there is no NaCl emission. In SrcI, Ginsburg et al. (2019a) inferred the presence of this hole from the gas kinematics, as the _apparent_ hole seen in Figure 1a is an observational effect caused by high optical depth in the edge-on dust disk. In G17, which is less inclined, the hole is directly observed (Figure 1b). Since the radiation field is more energetic in G17 than in SrcI, it is possible that the salt hole is related to the gas temperature. #### 3.3.2 G351.77-0.54 Beuther et al. (2017) and Beuther et al. (2019) published high-resolution observations of the G351.77-0.54 region, which we use here over the lower- resolution DIHCA data. We focus on the two brightest mm sources, mm1 and mm2, which both exhibit H2O and salt line emission. G351mm2 is a substantially fainter source, but similarly exhibits brine lines. Spectra of the G351 sources and subsequent sources are presented in the Appendices. We identify several lines in mm1, including NaCl $v$=1 and $v$=2 $J$=18-17 and SiS $v$=0 and $v$=1 $J$=13-12 (these latter were clearly detected, and used for disk kinematic study, in Beuther et al. (2019), but were listed as ‘unidentified’). Because of the slightly different spectral configuration adopted in these observations as compared to the DIHCA observations, both the $J$=12-11 and $J$=13-12 lines of SiS $v$=0 are covered and detected. The morphology and range covered by NaCl and SiS are very similar (Appendix A.1), though the gap in the center for SiS is more pronounced than for NaCl. These molecules arise in similar, but not identical, regions. The velocity gradients in NaCl, H2O, and SiS in mm1 appear to trace a bipolar outflow. Beuther et al. (2019) discussed the SiS emission lines in detail, comparing the velocity gradient observed in this line to that seen in SiO. The SiO $J$=5-4 and CO $J$=6-5 lines both appear in an extended redshifted lobe to the northwest of the source. The outflow is asymmetric and truncates at the position of mm1, suggesting that mm1 is the source (Beuther et al., 2017). Since the redshifted lobe occurs on the redshifted side of the observed SiS velocity gradient, Beuther et al. (2019) interpreted the lines as part of an outflow rather than a disk. While there is a blueshifted component opposite the redshifted flow, it is detected in only two channels and is only seen in the beam adjacent to the central source, so its direction cannot be determined independent from the elongated red lobe. They also noted the presence of a velocity gradient perpendicular to the outflow direction in CH3CN, suggesting that CH3CN traces the disk in a disk-outflow system. The observed velocity gradient in briny lines is perpendicular to the CH3CN disk, indicating either that the emission comes from outflow or that the disk changes orientation with scale. We re-examine the kinematics of the now-identified lines here. Figure 8 shows moment-0 and moment-1 maps of both NaCl (stacked) and H2O. The kinematics resemble a disk, with kinematics consistent with orbital motion around a $\sim 20/\sin i$ $M_{\odot}$ central potential, However, the extended SiO / CO $J$=6-5 outflow feature is parallel, rather than perpendicular, to the axis of the velocity gradient. Assuming the SiO structure is tracing an outflow, the velocity structure seen in Figure 8 is the base of the outflow. The lack of any gradient perpendicular to the outflow direction suggests that the briny lines are not tracing a disk wind in G351mm1, as they are in SrcI (Hirota et al., 2017), since they would retain that rotation signature for at least some distance above the disk. However, the lines are marginally resolved and extended in the direction perpendicular to the outflow, suggesting that the emission arises from an area at least comparable to be beam size ($\sim 50$ AU). The extended launching region is difficult to reconcile with the lack of disklike kinematics. We are not able to definitively conclude on the nature of the velocity gradient in mm1 and suggest that it should be studied further at high resolution, particularly to trace the kinematics of the disk-outflow system(s). Nevertheless, we note that the briny emission is limited to a region $<100$ AU across. Figure 8: Moment-0 (integrated intensity) and moment-1 (intensity-weighted velocity) images of stacked NaCl (left) and H2O (right) for G351 mm1. The position-velocity diagrams in Figure 14 are taken along position angle $130^{\circ}$, parallel to the velocity gradient seen in the lower panels here. The red arrow shows the direction of the outflow from Beuther et al. (2019) that extends $\gtrsim 0.2$ pc to the northwest. A compact blueshifted feature was also weakly detected bin SiO opposite the redshifted flow, but it is unresolved. The blue side of the NaCl, H2O, and (in Appendix A.1) SiS and PN corresponds with the SiO blueshifted lobe seen in Fig. 7 of Beuther et al. (2019), but since its extent is limited to the $\lesssim 100$ AU scale shown in this figure, we cannot confirm whether it is comprised of outflowing material. By contrast to mm1, mm2 has H2O emission (Appendix A.2), but only tentative NaCl emission (the $v$=1 line is detected at $\sim{\mathrm{few}}-\sigma$, but the $v$=2 line is below the noise). No SiS emission is detected. However, H30$\alpha$ emission is fairly clearly detected. There is a velocity gradient along the NW-SE axis in the H2O line (Appendix A.2). Curiously, this is perpendicular to the direction of the gradient shown in Figure 9 of Beuther et al. (2019), which shows the unidentified 231.986 GHz line, suggesting that there may be perpendicular gradients here tracing outflow and disk. However, no outflow is known toward this source, and we do not see clear signs of outflow in any of the lines studied here, including SiO. While SiO is detected in emission and absorption toward mm1, it is detected only in absorption toward mm2. Despite the presence of many KCl transitions in band, there are no detections - but our limits on these lines are relatively weak, as all of the $v$=0 and $v$=1 transitions tend to land in confused regions or come from doubly-rare isotopologues. A third source, G351mm12, also has salt detected. Appendix A.3 shows the standard suite of figures for this detection. This source is barely resolved. Its moment maps indicate a hint of velocity gradient, but the gradient is at the limit of our sensitivity and may be spurious. #### 3.3.3 NGC6334I There are three sources within the NGC 6334I region that may exhibit brine emission. Sources mm1b and mm2b are both notable for being faint continuum sources adjacent to bright sources identified in lower-resolution data (the ‘b’ designation indicates that there are sources mm1a and mm2a that are brighter). mm1b exhibits clear signatures of H2O, multiple NaCl lines, SiS, and several likely KCl detections (Appendix A.4). The H2O line shows hints of rotation perpendicular to the outflow axis (Fig. 8 of Brogan et al. (2018) shows the outflow, which is aligned to $\mathrm{PA}\approx 0^{\circ}$ to $-5^{\circ}$, close to straight north-south), though the other lines do not as clearly exhibit this signature. Brogan et al. (2018) note several other less prominent outflows centered on this source, however, which hints that this object cannot be interpreted as a single disk-outflow system. In the PV diagram of NaCl (Fig. 9), we show curves at 10, 20, and 30 $M_{\odot}$ for an edge-on Keplerian disk. Since the emission is confined to lower velocities than the 10 $M_{\odot}$ curve, it appears that this source is $<10$ $M_{\odot}$. However, it may also be significantly inclined to the line of sight, in which case it may be more massive. KCl is tentatively detected toward mm1b, with reasonably strong peaks appearing in the KCl $v$=2 $J$=29-28 and K^37Cl $v$=0 $J$=31-30 lines. Some others that might be expected to be bright, e.g., KCl $v$=3 $J$=29-28 and $v$=3 $J$=31-30, are ambiguous or blended. Figure 9: Position-velocity diagram of NaCl in the NGC 6334I mm1b disk. Keplerian velocity curves for edge-on 10, 20, and 30 $M_{\odot}$ central potentials are overplotted; the central potential here appears to have $M\sin i<10$ $M_{\odot}$. The other sources are less clear detections and are therefore considered only candidate brinary sources. Similar to mm1b, mm2b has a reasonably clear H2O detection and strong signs of SiS $v$=0 $J$=12-11 emission, but it has no clear detection of any salt line (Appendix A.4). No clear outflow is present. The SiS line profile is broad and somewhat different from that of water, so it is not obvious that they trace the same kinematics. mm1d has only has marginal SiS and H2O detections. We do not include it in the figures or detection statistics. It is a strong candidate for follow-up observations. #### 3.3.4 W33A We marginally detect H2O, H30$\alpha$, and NaCl toward the bright source at the center of W33A (Appendix A.5). At the current resolution, the emitting region is unresolved. The detection of any of these lines individually is tentative because we have only one firm line detection for each of these molecules and they are potentially blended with other emission lines. The PN 5-4, H2O, and NaCl v=1 18-17 lines are the most prominently detected. The NaCl v=1 and v=2 J=17-16 lines are too weak to confirm the NaCl detection. While we detect only one PN line, it is isolated enough that confusion is unlikely to affect it, so it is a reasonably firm detection. This source is a prime candidate for further followup. #### 3.3.5 I16547 I16547A and B were reported to have salt, water, and SiS emission in Tanaka et al. (2020). We confirm their detections both with their original data and with coarser-resolution observations from the DIHCA program. Despite the clear detection of those three briny species (see spectra in Appendix A.6 and A.7), there is no sign of KCl in the data. The non-detection is in part driven by confusion, in that many of the lower-$J$ transitions of KCl lay atop transitions from other molecular species that are spatially extended and not filtered out. Line stacking was not very helpful for these two targets because they are only marginally resolved (Fig. 2). Nevertheless, velocity gradients consistent with rotation are apparent in position-velocity diagrams extracted along the direction of the maximum gradient (Appendices A.6 and A.7), which is perpendicular to the outflow direction (Tanaka et al., 2020), suggesting that these are both likely disks around high-mass YSOs. ### 3.4 Unsalted sources The remainder of our targets do not have salt or water detections. We describe them in slightly more detail in Appendix B. We note here that several of these, i.e., G11.92, GGD27, and I18089, have clear extended disks that are detected in other molecules (e.g., CH3CN) but not in brinary lines. ## 4 Discussion ### 4.1 Are briny lines disk-only tracers? Many of our targets are only candidate disk sources, in that no resolved Keplerian rotation curve has been observed. We therefore consider the question: Could the briny emission be from outflows? Under the assumption that outflows are driven by either a wind from a disk or from accretion from a disk onto a star, the presence of an outflow still indicates the presence of a disk, but that does not mean the lines we observe necessarily arise within that disk. In SrcI, it is clear that the water is partly outflowing, but it arose from a disk wind and had a small observable scale height ($h<40$ au; see Fig. 10 of Ginsburg et al., 2018). Even in that case, the water line was dominated by disk kinematics, not outflow. In most of the observed sources, a velocity gradient across the lines of interest was observed. Such a velocity gradient can be produced by either outflow ejection or disk rotation. In the best-resolved cases, G17 and G351, the emission forms a complete ring, which is expected of a disk or disk wind. The circular extent of the briny emission shows that it is not tracing a collimated jet feature. However, in G351mm1, the direction of the gradient is along the known outflow, suggesting that the briny lines do not trace a disk or a disk wind; this source remains difficult to interpret. While we cannot definitively determine the general origin of briny emission, we observe here that it is restricted to radii $<300$ au in our full sample, such that it is always consistent with arising in either the disk or the very inner portion of the outflow. ### 4.2 Chemical similarities between the salt disks We examine the general chemical properties of the brinary disks. Each has several properties in common, but several differences. Table 3 lists the detections and non-detections toward each examined source. We observe that NaCl, H2O, and SiS lines are often detected toward the same sources, and generally if one is absent, all are. In other words, they exhibit comparable brightness when they are observed. This correlation hints that they come from similar regions within disks or outflows, either because of excitation or chemical (formation/destruction) conditions. Figure 10: Cross-correlation plot made from Table 3. We cross-correlate purely on the boolean value; those with a ‘yes’ are marked True, and those without a ‘yes’ are marked False. The rows are sorted in order of correlation with NaCl. Figure 10 shows the cross-correlation of the boolean value (“yes” or “no”) encoded in table 3; the tentative detections marked with ‘*’s are assigned to their corresponding ‘yes’ or ‘no’ prefix. It shows that the presence of NaCl, SiS, and H2O are strongly correlated. The presence of COMs in the spectra is moderately anti-correlated with these ‘brinary’ species. RRLs are strongly anti-correlated with COMs. The correlations among the other molecules (and RRLs) are less pronounced. We note that our data are extremely incomplete, as we explicitly select against COM-bearing targets by selecting H2O-bearing disks. There are many fainter disk candidates in the fields of view of our observations that we did not include in our sample or in this plot. Most of these sources do exhibit COM emission and do not exhibit RRL emission, though, so adding them to this plot would generally strengthen the trends shown. Because of the incompleteness of our sample, we caution against over- interpretation. A repeat of this analysis with a consistently-selected sample will be needed to draw firm conclusions. However, the correlations in the top- left of the plot, between H2O, NaCl, and SiS appear firm - these species coexist in this sample of disk candidates. #### 4.2.1 The low detection rate of KCl is an observational effect The KCl detections are not perfectly correlated with NaCl detections, but we argue here that this is an observational selection effect. Most likely, KCl is better-correlated with NaCl than is apparent in Figure 10. In the targeted band, there are no $v$=1 or $v$=0 transitions of the most common isotopologue (39K-35Cl) of KCl, such that all detections are of higher- excitation, lower-brightness transitions. The targeted KCl lines are also partly hidden by confusion with other lines. The KCl lines are strongly anti- correlated with the presence of COMs, which is primarily (or entirely) because of confusion: the KCl lines in this band are weaker than NaCl when they are detected, so they are more difficult to distinguish from the molecular line forest present in dense spectra. They land inconveniently near bright COM lines in the 215-235 GHz range more often than the NaCl lines. The comparison to SrcI illustrates some of these effects. In SrcI, the peak brightness temperature of the NaCl and KCl lines with E${}_{U}<1000$ K were similar, with $T_{\rm B,max}\sim 100-200$ K. The NaCl $v$=0 lines were no more than twice as bright as the KCl $v$=0 lines. The SrcI data set was targeted on outflow-tracing lines, including 12CO $J$=2-1 and SiO $v$=1 $J$=5-4, which both have nearby $v$=0 and $v$=1 KCl transitions, while the DIHCA observations and others presented here chose to target the CH3CN ladder at 220.4 GHz and therefore did not cover these low-$v$ transitions. #### 4.2.2 PN in the brine The PN J=5-4 line appears to be detected in several of our sources. The PN v=1 J=5-4 transition, at 233.27182 GHz, is not detected. For the line-poor brinaries, there is little confusion around this line, and its velocity lines up perfectly with that of salts, so this identification is reasonably certain. For the rest of the sample, the case is less clear; while there are some tentative detections with clear lines at this velocity, the spectra are so rich that we cannot definitively identify PN as the carrier species. The presence of PN in the same regions as the highly-excited salt lines may suggest that PN occupies a similar location within and binding energy to dust grains. Rivilla et al. (2020) suggest that PN is present in the cavity walls of an outflow toward AFGL 5142, but that it is released to the gas phase as PH3, and the PN molecules are subsequently formed under the influence of the star’s UV radiation. Given the lack of UV photons in these sources, as indicated by the lack of correlation between PN and RRLs in Figure 10, our data may indicate that PN is present in the grains and not formed in the gas phase. Figure 11: The integrated intensity (moment-0) maps of NaCl shown in Fig. 1 and 2, but now with each image scaled to the same physical resolution. The 100 AU scalebar, shown on the W33A mm1-main panel, applies to all panels. The intensity scales are arbitrary, as the intent is only to show the disk candidate structure. #### 4.2.3 RRLs Hydrogen recombination lines are likely to be produced in the ionized regions surrounding accreting high-mass young stars once they have contracted onto the main sequence. In our sample, few RRLs are detected. Their presence, or absence, is only weakly anti-correlated with the presence of COMs and PN, but has little correlation with other molecules. While we might expect RRL emission to become detectable from accreting HMYSOs toward the end of their accretion phase, as the ionization rate is able to overtake the accretion rate of fresh neutral material, our data provide little evidence for this process. ### 4.3 Excitation At least in the best-resolved cases, G17 and G351, vibrationally excited states of NaCl are detected ($v$=1, and 2). This feature is in common with SrcI, where states up to $v$=6 were convincingly detected. The $v$=2 detections in particular, with $E_{\rm U}\gtrsim 1000$ K, suggest that an excitation pattern similar to that in SrcI, in which $T_{\rm vib}>T_{\rm rot}$, is common. This general feature is also common in evolved stars that exhibit salt emission, suggesting that these salt lines only appear in regions with strong radiative backgrounds in the infrared. Figure 7 highlights the high excitation, though we defer deeper analysis of the excitation properties to a future work. ### 4.4 Spatial Resolution Our observed sample has non-uniform spatial resolution (Table 1), which helps explain several of the non-detections. We show a version of Figure 1 with all images resized to the same physical scale shown in Figure 11, highlighting that the detected NaCl disks are small. We did not detect NaCl toward any sources observed with beam size $>300$ au, which included four of our targeted fields. Five of the fifteen targeted fields included detections. We performed a logistic regression of the detection of salt against resolution and found that the likelihood of salt detection in our sample is $>50\%$ for resolution $<120$ au, peaking at $\sim 80\%$ at infinite resolution. While our haphazard sample selection prevents drawing strong conclusions, these results suggest that salty regions are limited to small scales ($\lesssim 300$ au) and are challenging to observe if they are not well-resolved. This observational limitation highlights the need for more extensive extremely high-resolution observations with ALMA’s longest-baseline configurations to obtain dynamical mass measurements for high-mass YSOs. It is possible that more sensitive observations with poor physical resolution may detect brinary lines now that we know to look for them. However, the main difficulty in detecting these lines is not raw sensitivity, but confusion with other lines. It will be beneficial to search other parts of the spectrum for more isolated brinary lines, which might be expected to be more common at lower frequencies. It may be possible to determine the presence or absence of brinary lines by stacking the various transitions across species to average down the ‘contaminant’ noise provided by other lines; we leave investigation of this possibility to future work. ## 5 Conclusions * • We have substantially increased the number of high-mass protostellar objects with detection of salt, hot water, and SiS, increasing the number from three to nine published detections across six regions. * • Salt, water, and SiS tend to coexist. PN may also coexist with these species. When any of these molecules is detected in a HMYSO disk, all of them are likely to be. * • Brinary sources are line-poor compared to typical hot cores. The lack of COMs in the briny regions suggests that the chemistry of these regions is different from hot cores, even when the central objects are surrounded by hot cores. * • These emission lines do not come from the same volume as ionized hydrogen. While some of the HMYSO candidates targeted exhibit both RRL and brine emission, the presence of an RRL is a poor predictor of whether NaCl and H2O are detected. The resolved case of G17 shows that the ionized gas comes from a smaller radius than the NaCl. * • With the nine brinaries presented here, we demonstrate that salt emission is not rare. The primary reason it is not often detected is resolution: the emission comes exclusively from small ($\lesssim 100$ au) size scales, confirming that either chemistry or excitation restricts the millimeter lines described here to disk-sized regions. Line confusion limits our ability to detect these lines even when they are present, though confusion can be alleviated with high spatial resolution. * • However, salt and water emission is also not ubiquitous in HMYSO disks. Some of the most compelling Keplerian disks around HMYSOs in the literature, such as GGD27 and G11.92mm1, show no sign of salt emission despite their similarities to other disks and superior data quality. There is clearly substantial future work to do with these data and expanded samples of brinaries. Some of the more obvious questions about brinary lines include: * • Are they correlated with source luminosity or stellar mass? * • What disks, or outflows, produce them? * • Do they occur around low-mass YSOs? * • Why are highly vibrationally excited lines ($v>3$) detected? Many of these questions will require establishing less biased, more systematic samples of YSO candidates. Others will require more detailed, multi-line studies toward a limited sample. Acknowledgements We thank the anonymous referee for a thorough and constructive report that significantly improved the paper, particularly in terms of structure and readability. AG acknowledges support from NSF AAG 2008101 and NSF CAREER 2142300. HB acknowledges support from the European Research Council under the European Community’s Horizon 2020 frame-work program (2014-2020) via the ERC Consolidator Grant ‘From Cloud to Star Formation (CSF)’ (project number 648505). HB also acknowledges support from the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation) – Project-ID 138713538 – SFB 881 (“The Milky Way System”, subproject B01). P.S. was partially supported by JSPS KAKENHI grant Nos. JP18H01259 and JP22H01271. K.E.I.T. acknowledges support by JSPS KAKENHI grant Nos. JP19K14760, JP19H05080, JP21H00058, and JP21H01145. This paper makes use of the following ALMA data: ADS/JAO.ALMA#2016.1.01036.S, ADS/JAO.ALMA#2017.1.00237.S, ADS/JAO.ALMA#2016.1.00550.S, ADS/JAO.ALMA#2019.1.00492.S, ADS/JAO.ALMA#2017.1.00098.S, ADS/JAO.ALMA#2018.1.01656.S, and ADS/JAO.ALMA#2013.1.00260.S. ALMA is a partnership of ESO (representing its member states), NSF (USA) and NINS (Japan), together with NRC (Canada), MOST and ASIAA (Taiwan), and KASI (Republic of Korea), in cooperation with the Republic of Chile. The Joint ALMA Observatory is operated by ESO, AUI/NRAO and NAOJ. In addition, publications from NA authors must include the standard NRAO acknowledgement: The National Radio Astronomy Observatory is a facility of the National Science Foundation operated under cooperative agreement by Associated Universities, Inc. ## References * Añez-López et al. (2020) Añez-López, N., Osorio, M., Busquet, G., et al. 2020, ApJ, 888, 41, doi: 10.3847/1538-4357/ab5dbc * Astropy Collaboration et al. (2013) Astropy Collaboration, Robitaille, T. P., Tollerud, E. J., et al. 2013, A&A, 558, A33, doi: 10.1051/0004-6361/201322068 * Astropy Collaboration et al. (2018) Astropy Collaboration, Price-Whelan, A. M., Sipőcz, B. M., et al. 2018, AJ, 156, 123, doi: 10.3847/1538-3881/aabc4f * Barton et al. (2014) Barton, E. J., Chiu, C., Golpayegani, S., et al. 2014, MNRAS, 442, 1821, doi: 10.1093/mnras/stu944 * Beltrán et al. (2011) Beltrán, M. T., Cesaroni, R., Neri, R., & Codella, C. 2011, A&A, 525, A151, doi: 10.1051/0004-6361/201015049 * Beltrán & de Wit (2016) Beltrán, M. T., & de Wit, W. J. 2016, A&A Rev., 24, 6, doi: 10.1007/s00159-015-0089-z * Beuther et al. (2017) Beuther, H., Walsh, A. J., Johnston, K. G., et al. 2017, A&A, 603, A10, doi: 10.1051/0004-6361/201630126 * Beuther et al. (2007) Beuther, H., Zhang, Q., Bergin, E. A., et al. 2007, A&A, 468, 1045, doi: 10.1051/0004-6361:20066954 * Beuther et al. (2005) Beuther, H., Zhang, Q., Sridharan, T. K., & Chen, Y. 2005, ApJ, 628, 800, doi: 10.1086/431277 * Beuther et al. (2004) Beuther, H., Hunter, T. R., Zhang, Q., et al. 2004, ApJ, 616, L23, doi: 10.1086/383570 * Beuther et al. (2019) Beuther, H., Ahmadi, A., Mottram, J. C., et al. 2019, A&A, 621, A122, doi: 10.1051/0004-6361/201834064 * Brogan et al. (2018) Brogan, C. L., Hunter, T. R., Cyganowski, C. J., et al. 2018, ApJ, 866, 87, doi: 10.3847/1538-4357/aae151 * Cernicharo & Guelin (1987) Cernicharo, J., & Guelin, M. 1987, A&A, 183, L10 * Chibueze et al. (2014) Chibueze, J. O., Omodaka, T., Handa, T., et al. 2014, ApJ, 784, 114, doi: 10.1088/0004-637X/784/2/114 * Comrie et al. (2021) Comrie, A., Wang, K.-S., Hsu, S.-C., et al. 2021, CARTA: The Cube Analysis and Rendering Tool for Astronomy, 2.0.0, Zenodo, Zenodo, doi: 10.5281/zenodo.3377984 * Danilovich et al. (2021) Danilovich, T., Van de Sande, M., Plane, J. M. C., et al. 2021, A&A, 655, A80, doi: 10.1051/0004-6361/202141757 * Decin et al. (2016) Decin, L., Richards, A. M. S., Millar, T. J., et al. 2016, A&A, 592, A76, doi: 10.1051/0004-6361/201527934 * Fernández-López et al. (2021) Fernández-López, M., Sanhueza, P., Zapata, L. A., et al. 2021, ApJ, 913, 29, doi: 10.3847/1538-4357/abf2b6 * Ginsburg et al. (2018) Ginsburg, A., Bally, J., Goddi, C., Plambeck, R., & Wright, M. 2018, ApJ, 860, 119, doi: 10.3847/1538-4357/aac205 * Ginsburg et al. (2019a) Ginsburg, A., McGuire, B., Plambeck, R., et al. 2019a, ApJ, 872, 54, doi: 10.3847/1538-4357/aafb71 * Ginsburg et al. (2022) Ginsburg, A., Sokolov, V., de Val-Borro, M., et al. 2022, AJ, 163, 291, doi: 10.3847/1538-3881/ac695a * Ginsburg et al. (2019b) Ginsburg, A., Koch, E., Robitaille, T., et al. 2019b, radio-astro-tools/spectral-cube: Release v0.4.5, v0.4.5, Zenodo, Zenodo, doi: 10.5281/zenodo.3558614 * Ginsburg et al. (2019c) Ginsburg, A., Sipőcz, B. M., Brasseur, C. E., et al. 2019c, AJ, 157, 98, doi: 10.3847/1538-3881/aafc33 * Girart et al. (2017) Girart, J. M., Estalella, R., Fernández-López, M., et al. 2017, ApJ, 847, 58, doi: 10.3847/1538-4357/aa81c9 * Girart et al. (2018) Girart, J. M., Fernández-López, M., Li, Z. Y., et al. 2018, ApJ, 856, L27, doi: 10.3847/2041-8213/aab76b * Guzmán et al. (2020) Guzmán, A. E., Sanhueza, P., Zapata, L., Garay, G., & Rodríguez, L. F. 2020, ApJ, 904, 77, doi: 10.3847/1538-4357/abbe09 * Harris et al. (2020) Harris, C. R., Millman, K. J., van der Walt, S. J., et al. 2020, Nature, 585, 357, doi: 10.1038/s41586-020-2649-2 * Highberger et al. (2003) Highberger, J. L., Thomson, K. J., Young, P. A., Arnett, D., & Ziurys, L. M. 2003, ApJ, 593, 393, doi: 10.1086/376446 * Hirota et al. (2017) Hirota, T., Machida, M. N., Matsushita, Y., et al. 2017, Nature Astronomy, 1, 0146, doi: 10.1038/s41550-017-0146 * Homan et al. (2020) Homan, W., Montargès, M., Pimpanuwat, B., et al. 2020, A&A, 644, A61, doi: 10.1051/0004-6361/202039185 * Hunter (2007) Hunter, J. D. 2007, Computing in Science and Engineering, 9, 90, doi: 10.1109/MCSE.2007.55 * Ilee et al. (2018) Ilee, J. D., Cyganowski, C. J., Brogan, C. L., et al. 2018, ApJ, 869, L24, doi: 10.3847/2041-8213/aaeffc * Johnston et al. (2020) Johnston, K. G., Hoare, M. G., Beuther, H., et al. 2020, ApJ, 896, 35, doi: 10.3847/1538-4357/ab8adc * Kalcheva et al. (2018) Kalcheva, I. E., Hoare, M. G., Urquhart, J. S., et al. 2018, A&A, 615, A103, doi: 10.1051/0004-6361/201832734 * Kauffmann & Pillai (2010) Kauffmann, J., & Pillai, T. 2010, ApJ, 723, L7, doi: 10.1088/2041-8205/723/1/L7 * Kluyver et al. (2016) Kluyver, T., Ragan-Kelley, B., Pérez, F., et al. 2016, in IOS Press, 87–90, doi: 10.3233/978-1-61499-649-1-87 * Koch et al. (2021) Koch, E., Ginsburg, A., AKL, et al. 2021, radio-astro-tools/radio-beam: v0.3.3, v0.3.3, Zenodo, Zenodo, doi: 10.5281/zenodo.4623788 * Kurayama et al. (2011) Kurayama, T., Nakagawa, A., Sawada-Satoh, S., et al. 2011, PASJ, 63, 513, doi: 10.1093/pasj/63.3.513 * Liu et al. (2020a) Liu, H.-L., Sanhueza, P., Liu, T., et al. 2020a, ApJ, 901, 31, doi: 10.3847/1538-4357/abadfe * Liu et al. (2020b) Liu, S.-Y., Su, Y.-N., Zinchenko, I., et al. 2020b, ApJ, 904, 181, doi: 10.3847/1538-4357/abc0ec * Maud et al. (2017) Maud, L. T., Hoare, M. G., Galván-Madrid, R., et al. 2017, MNRAS, 467, L120, doi: 10.1093/mnrasl/slx010 * Maud et al. (2018) Maud, L. T., Cesaroni, R., Kumar, M. S. N., et al. 2018, A&A, 620, A31, doi: 10.1051/0004-6361/201833908 * Maud et al. (2019) —. 2019, A&A, 627, L6, doi: 10.1051/0004-6361/201935633 * McMullin et al. (2007) McMullin, J. P., Waters, B., Schiebel, D., Young, W., & Golap, K. 2007, in Astronomical Society of the Pacific Conference Series, Vol. 376, Astronomical Data Analysis Software and Systems XVI, ed. R. A. Shaw, F. Hill, & D. J. Bell, 127 * Milam et al. (2007) Milam, S. N., Apponi, A. J., Woolf, N. J., & Ziurys, L. M. 2007, ApJ, 668, L131, doi: 10.1086/522928 * Moscadelli et al. (2021) Moscadelli, L., Cesaroni, R., Beltrán, M. T., & Rivilla, V. M. 2021, A&A, 650, A142, doi: 10.1051/0004-6361/202140829 * Motogi et al. (2019) Motogi, K., Hirota, T., Machida, M. N., et al. 2019, ApJ, 877, L25, doi: 10.3847/2041-8213/ab212f * Olguin et al. (2022) Olguin, F. A., Sanhueza, P., Ginsburg, A., et al. 2022, ApJ, 929, 68, doi: 10.3847/1538-4357/ac5bd8 * Olguin et al. (2021) Olguin, F. A., Sanhueza, P., Guzmán, A. E., et al. 2021, ApJ, 909, 199, doi: 10.3847/1538-4357/abde3f * Peretto et al. (2013) Peretto, N., Fuller, G. A., Duarte-Cabral, A., et al. 2013, A&A, 555, A112, doi: 10.1051/0004-6361/201321318 * Qiu et al. (2019) Qiu, K., Wyrowski, F., Menten, K., Zhang, Q., & Güsten, R. 2019, ApJ, 871, 141, doi: 10.3847/1538-4357/aaf728 * Reid et al. (2014) Reid, M. J., Menten, K. M., Brunthaler, A., et al. 2014, ApJ, 783, 130, doi: 10.1088/0004-637X/783/2/130 * Rivilla et al. (2020) Rivilla, V. M., Drozdovskaya, M. N., Altwegg, K., et al. 2020, MNRAS, 492, 1180, doi: 10.1093/mnras/stz3336 * Rodríguez-Kamenetzky et al. (2017) Rodríguez-Kamenetzky, A., Carrasco-González, C., Araudo, A., et al. 2017, ApJ, 851, 16, doi: 10.3847/1538-4357/aa9895 * Sadaghiani et al. (2020) Sadaghiani, M., Sánchez-Monge, Á., Schilke, P., et al. 2020, A&A, 635, A2, doi: 10.1051/0004-6361/201935699 * Sánchez Contreras et al. (2022) Sánchez Contreras, C., Alcolea, J., Rodríguez Cardoso, R., et al. 2022, arXiv e-prints, arXiv:2206.12185. https://arxiv.org/abs/2206.12185 * Sánchez-Monge et al. (2018) Sánchez-Monge, Á., Schilke, P., Ginsburg, A., Cesaroni, R., & Schmiedeke, A. 2018, A&A, 609, A101, doi: 10.1051/0004-6361/201730425 * Sanhueza et al. (2010) Sanhueza, P., Garay, G., Bronfman, L., et al. 2010, ApJ, 715, 18, doi: 10.1088/0004-637X/715/1/18 * Sanhueza et al. (2012) Sanhueza, P., Jackson, J. M., Foster, J. B., et al. 2012, ApJ, 756, 60, doi: 10.1088/0004-637X/756/1/60 * Sanhueza et al. (2021) Sanhueza, P., Girart, J. M., Padovani, M., et al. 2021, ApJ, 915, L10, doi: 10.3847/2041-8213/ac081c * Sanna et al. (2021) Sanna, A., Giannetti, A., Bonfand, M., et al. 2021, A&A, 655, A72, doi: 10.1051/0004-6361/202040000 * Sato et al. (2014) Sato, M., Wu, Y. W., Immer, K., et al. 2014, ApJ, 793, 72, doi: 10.1088/0004-637X/793/2/72 * Stephens et al. (2015) Stephens, I. W., Jackson, J. M., Sanhueza, P., et al. 2015, ApJ, 802, 6, doi: 10.1088/0004-637X/802/1/6 * Tachibana et al. (2019) Tachibana, S., Kamizuka, T., Hirota, T., et al. 2019, ApJ, 875, L29, doi: 10.3847/2041-8213/ab1653 * Tanaka et al. (2020) Tanaka, K. E. I., Zhang, Y., Hirota, T., et al. 2020, ApJ, 900, L2, doi: 10.3847/2041-8213/abadfc * The Astropy Collaboration et al. (2022) The Astropy Collaboration, Price-Whelan, A. M., Lian Lim, P., et al. 2022, arXiv e-prints, arXiv:2206.14220. https://arxiv.org/abs/2206.14220 * van der Walt et al. (2011) van der Walt, S., Colbert, S. C., & Varoquaux, G. 2011, Computing in Science and Engineering, 13, 22, doi: 10.1109/MCSE.2011.37 * Virtanen et al. (2020) Virtanen, P., Gommers, R., Oliphant, T. E., et al. 2020, Nature Methods, 17, 261, doi: 10.1038/s41592-019-0686-2 * Whitaker et al. (2017) Whitaker, J. S., Jackson, J. M., Rathborne, J. M., et al. 2017, AJ, 154, 140, doi: 10.3847/1538-3881/aa86ad * Woitke et al. (2018) Woitke, P., Helling, C., Hunter, G. H., et al. 2018, A&A, 614, A1, doi: 10.1051/0004-6361/201732193 * Xu et al. (2011) Xu, Y., Moscadelli, L., Reid, M. J., et al. 2011, ApJ, 733, 25, doi: 10.1088/0004-637X/733/1/25 * Zinchenko et al. (2015) Zinchenko, I., Liu, S. Y., Su, Y. N., et al. 2015, ApJ, 810, 10, doi: 10.1088/0004-637X/810/1/10 * Zinchenko et al. (2020) Zinchenko, I. I., Liu, S.-Y., Su, Y.-N., Wang, K.-S., & Wang, Y. 2020, ApJ, 889, 43, doi: 10.3847/1538-4357/ab5c18 ## Appendix A Additional figures for salted sources For the salted sources, we present moment maps and spectra with lines identified in the following Appendices. ### A.1 G351.77mm1 We show additional moment maps, spectra, and position-velocity diagrams from G351.77mm1 in this section. This object is one of the most compelling, high signal-to-noise objects in the sample, yet its kinematics are perplexing, exhibiting disk-like rotation in the same direction as the larger-scale outflow rather than rotating perpendicular to the outflow, as we would expect. Figure 12 shows the stacked spectra. Figure 8 shows moment-0 and moment-1 maps of NaCl and H2O. Figure 13 shows the same for SiS and PN. All four molecules exhibit similar morphology and kinematics, including a prominent central hole, which is strongly suggestive of a disk. Figure 14 shows moment-0 maps and position-velocity diagrams, and Figure 15 more position-velocity diagrams, illustrating that rotation is a plausible explanation for the observed kinematics. Figure 12: Stacked spectra from G351.77mm1 from the Beuther et al. (2019) data set. The stacking was based on the H2O line. Line IDs are shown; no KCl detections are clear. Different colors are used for targeted species with multiple transitions in-band: orange for SiS, blue for KCl, red for NaCl, magenta for H30$\alpha$, purple for PN, and green for H2O. The remaining species, with only one transition marked, are shown in black. The PN line identification should be taken with a grain of salt since it is the only transition we observe from PN. Figure 13: Moment-0 and 1 images as in Figure 8, but for the SiS v=0 J=13-12 and PN 5-4 lines of G351 mm1. Figure 14: Comparison of NaCl and SiS moment-0 and position-velocity maps of G351.77mm1. (top left) NaCl stack moment 0 image, as seen in Figure 1. (top right) SiS 13-12 moment 0 image, showing similar morphology. (bottom left) NaCl stack cube position-velocity diagram extracted along the direction of maximum gradient. (bottom right) SiS 13-12 position-velocity diagram extracted along the direction of maximum gradient. Keplerian rotation curves assuming an edge-on central source with the listed mass are shown in colored lines; these curves are not fits to the data and do not account for inclination, they are just provided to guide the eye. Furthermore, as discussed in Section 3.3.2, the velocity gradient shown here may trace an outflow rather than a disk. Figure 15: Two more position-velocity diagrams of G351 mm1, showing the 235 GHz PN 5-4 line (newly identified here) and the H2O line. The common structure seen in these diagrams, and in Figure 14, justifies our assumption that these have common kinematics. Keplerian rotation curves assuming an edge-on central source with the listed mass are shown in colored lines; these curves are not fits to the data and do not account for inclination, they are just provided to guide the eye. Furthermore, the direction of maximum gradient, along which these diagrams are extracted, points in the direction of the outflow. ### A.2 G351.77mm2 Figure 16 shows the stacked spectra of G351 mm2. Figure 17 shows the moment-0 and moment-1 maps of NaCl and H2O, which are only marginally resolved. Figure 16: The G351 mm2 disk candidate stacked spectrum. Like mm1 (Fig. 12), the stacking was done on the H2O line. However, other lines are at most weakly detected; the NaCl v=1 J=18-17 line is evident, but no other clear detections are present in this or other bands. In all spectra, but particularly in spectral window 0, much of the spectrum is absorbed by material unassociated with the disk; we cut off the absorption features to emphasize emission features here. Figure 17: Moment-0 (integrated intensity) and moment-1 (intensity-weighted velocity) images of stacked NaCl (left) and H2O (right) for G351 mm2. ### A.3 G351.77mm12 Figure 18 shows the stacked spectra of G351 mm12. This source is unresolved, as shown in Fig 2. Figure 19 shows the moment-0 and moment-1 maps of NaCl and H2O, which are only marginally resolved. Figure 18: The G351 mm12 disk candidate stacked spectrum. See Figure 12 for additional description. Figure 19: Moment-0 (integrated intensity) and moment-1 (intensity-weighted velocity) images of stacked NaCl (left) and H2O (right) for G351 mm12. ### A.4 NGC6334I We show additional figures of NGC6334I, including spectra (Fig 20 and moment maps (Fig 21. For mm2b, we show only the stacked spectrum (Fig 22), since the source is unresolved Figure 20: The NGC6334I mm1b disk stacked spectrum. See Figure 12 for additional description. Figure 21: Moment-0 and 1 images as in Figure 8, but for NGC 6334I mm1b. The outflow noted by Brogan et al. (2018) is shown at PA=-5∘ with green arrows, as it appears to be primarily in the plane of the sky. The solid gray line shows the orientation from which the position- velocity diagram (Figure 9) is extracted. Figure 22: The NGC6334I mm2b disk stacked spectrum. See Figure 12 for additional description. ### A.5 W33A The labeled, stacked spectrum from W33A is shown in Figure 23. The four-panel moment map is not shown for this source because it is unresolved and shows no structure; it is consistent with a point source. Figure 23: The W33A disk stacked spectrum. See Figure 12 for additional description. ### A.6 I16547A The labeled, stacked spectrum from I16547A is shown in Figure 24. Moment-0 and moment-1 images are shown in Figure 25. A position-velocity diagram, with overlaid Keplerian curves for an edge-on orbit with masses labeled, is shown in Figure 28. We show these to provide order-of-magnitude mass estimates, but note that we have no constraint on the disk inclination and have not attempted to model the extent of the disk emission. No outflow is observed toward A (Tanaka et al., 2020). Figure 24: The I16547A disk stacked spectrum. See Figure 12 for additional description. Figure 25: Moment-0 and 1 images as in Figure 8, but for IRAS 16547A The solid gray line shows the orientation from which the position- velocity diagram (Figure 28) is extracted, based on the angle determined in Tanaka et al. (2020). The radio jet at PA=-16∘ is shown with arrows (Tanaka et al., 2020). ### A.7 I16547B The labeled, stacked spectrum from I16547B is shown in Figure 26. Moment-0 and moment-1 images are shown in Figure 27. A position-velocity diagram, with overlaid Keplerian curves for an edge-on orbit with masses labeled, is shown in Figure 28. We show these to provide order-of-magnitude mass estimates, but note that we have no constraint on the disk inclination and have not attempted to model the extent of the disk emission. An SiO outflow is observed toward B, perpendicular to the gradient in the PV diagram (Tanaka et al., 2020). Figure 26: The I16547B disk stacked spectrum. See Figure 12 for additional description. Figure 27: Moment-0 and 1 images as in Figure 8, but for IRAS 16547B The solid gray line shows the orientation from which the position- velocity diagram (Figure 28) is extracted, based on the angle determined in Tanaka et al. (2020). The red and blue arrows indicate the direction of the SiO outflow noted in Tanaka et al. (2020). Figure 28: Position-velocity diagrams of the I16547 disks in the stacked NaCl lines. The overplotted curves show Keplerian rotation around central point sources with masses indicated in the caption, assuming an edge-on inclination. As noted in Tanaka et al. (2020), the disks appear to be counter-rotating along similar position angles. ### A.8 Continuum We show Figures 1 and 2 again, but this time with continuum contours overlaid, in Figures 29 and 30. Figure 29: Moment-0 (integrated intensity) images of the resolved sources in NaCl lines as described in Figure 1. The red ellipse shows the continuum beam corresponding to the contours, while the black ellipse shows the line image beam. The contours are: SrcI $\sigma=0.21$ mJy/beam, contours at 25, 50, 75, 100, 125, 150 $\sigma$, G17 $\sigma=0.12$ mJy/beam, contours at 50, 100, 150, 200 $\sigma$, G351mm1 $\sigma=0.08$ mJy/beam, contours at 50, 75, 100, 125 $\sigma$, NGC6334Imm1b $\sigma=1$ mJy/beam, contours at 10, 15, 20, 25 $\sigma$. Figure 30: As in Figure 2, but with continuum contours overlaid. The red ellipse shows the continuum beam corresponding to the contours, while the black ellipse shows the line image beam. The contours are: G351mm2 $\sigma=0.08$ mJy/beam, contours at 50, 75, 100, 125 $\sigma$, G351mm12 $\sigma=0.08$ mJy/beam, contours at 10, 20, 30 $\sigma$, NGC6334Imm2 $\sigma=1$ mJy/beam, contours at 2, 3, 4 $\sigma$, I16547A $\sigma=0.08$ mJy/beam, contours at 100, 150, 200, 250, 300 $\sigma$, I16547B $\sigma=0.08$ mJy/beam, contours at 75, 100, 125, 150 $\sigma$, W33A $\sigma=0.3$ mJy/beam, contours at 20,40,60,80 $\sigma$. ## Appendix B Unsalted Sources ### B.1 I16562 IRAS16562-3959 is also known as G345.4938+01.4677. This source shows SO in absorption against a continuum disk. It exhibits RRL emission that shows a very tiny gradient; Guzmán et al. (2020) reported the detection of an ionized disk in this system based on RRL emission. No other emission lines are associated with the RRL possible disk. There is extensive CH3CN emission around this source with kinematics more complicated than a simple Keplerian disk. ### B.2 G333 We adopt naming from Stephens et al. (2015), though their resolution was only $\sim 2.5\arcsec$. G333.23mm1 is the bright central region. Morphologically, it looks like the hub at the center of several converging filaments. There is no obvious sign of a disk in any of the lines we examined. There is no hint of brinary lines. G333.23mm2 shows some sign of a line gradient in $\textrm{CH}_{3}\textrm{OH}$, hinting that a disk is present. However, no brinary lines are seen. ### B.3 G335 Olguin et al. (2022) performed an extensive study of this system using the data presented here. G335-ALMA1 is the bright central source that shows signs of rotation, but no clear disk. Instead, it appears dominated by inflowing accretion filaments. The spectrum is extremely rich, and we adopt the same $\textrm{CH}_{3}\textrm{OH}$ line as in Olguin et al. (2022) as our velocity reference. No brinary lines are detected. ### B.4 G5.89 The G5.89 image is dominated by an extended HII region. The only disklike source in the imaged field of view is mm15, so our cutout centered on that source. It apparently exhibits _no_ line emission at all in the present data set. We examined the other millimeter peaks in the region, but found no obvious signatures of disks or brinary lines. ### B.5 GGD27 GGD27mm1 (otherwise referred to here as GGD27) is the driving source of HH80/81 (Girart et al., 2017, 2018; Añez-López et al., 2020). It is also known as IRAS 18162-2048. It has a clear, well-defined, massive ($M\sim 5M_{\odot}$) disk orbiting a $\sim 20$ $M_{\odot}$ star (Añez-López et al., 2020). There is no sign of NaCl, H2O, KCl, or SiS in the spectrum of the disk in our data or previous observations (Girart et al., 2017). It is bright in SO $6_{5}-5_{4}$, which we used to stack spectra to perform a deeper search. This source is the archetype of non-salt-bearing disks. Its well-defined Keplerian line profiles, high central mass, and line-poor spectrum demonstrate that not all HMYSO disks exhibit salt emission. The system is very similar to SrcI in terms of its central stellar mass, disk size (though GGD27’s is 2-4$\times$ larger in radius), and luminosity, suggesting that none of these features are critical for releasing salt into the gas phase. As the driver of the HH80/81 outflow, it is also clear that the simple presence of an outflow does not determine whether brinary lines are produced. However, one notable difference is that the HH80/81 jet is highly collimated (Rodríguez-Kamenetzky et al., 2017; Qiu et al., 2019), while SrcI drives a broader disk wind (Hirota et al., 2017; Tachibana et al., 2019), hinting that the driving mechanism of the outflow may have a role in determining when salts are detectable. The comparison of this source to SrcI and others in this sample will be useful for future understanding of the origin of salts. ### B.6 IRAS18089 Sanhueza et al. (2021) observed this source at high resolution, though they focused on the magnetic field. Previously, Beuther et al. (2004, 2005) showed that this source was line-rich with SMA observations. We used a $\textrm{CH}_{3}\textrm{OH}$ as the stacking line because the kinematics were similar across all bright lines, including the SO $6_{5}-5_{4}$ line, but the SO line was affected by strong absorption toward the inner disk. The line kinematics were reasonably disk-like, but not consistent with a single Keplerian disk; the velocity structure in this source requires more sophisticated modeling. Nevertheless, there is no sign of brinary lines in the averaged or stacked spectra. ### B.7 G34.43mm1 There is a structure toward the center of G34.43mm1 with a clear line gradient, but it does not trace disk-like kinematics and is quite extended. On the larger scale, G34.43mm1 drives a powerful outflow (Sanhueza et al., 2010). While this is region is quite nearby (1.56 kpc Xu et al., 2011), it is not among the salt-bearing candidates. Despite its relatively near distance, the observed spatial resolution is only $\sim 400$ au, which is much larger than the detected disks. The overall appearance of the inner region suggests that streamers are feeding in to a central region. The chemically rich inflowing material results in a very spectrally dense spectrum (e.g., Sanhueza et al., 2012; Liu et al., 2020a), which likely prevents detection of brinary features even if they are present. This is a good candidate for followup at higher angular resolution. ### B.8 G29.96 Beuther et al. (2007) and Beltrán et al. (2011) studied this region at $\sim$arcsecond resolution. The target region is centered on the brightest few sources at the center. The distance to this object is unclear, with Beltrán et al. (2011) reporting 3.5 kpc and Kalcheva et al. (2018) reporting 7.4 kpc. There is no clear signature of any of the lines of interest, nor is there a clear signature of rotation. There is a tentative detection of SiS v=0 12-11, but neither of the SiS v=1 lines (13-12 or 12-11) appear, so this detection is uncertain. ### B.9 S255IR Sh 2-255 IR SMA1 (S255IR) was the recent ($<10$ years) site of a major accretion outburst. As such it is a strong candidate for being actively heated over its ‘normal’ level. No disk is obvious in the data cube from rotational signatures. While several authors (Zinchenko et al., 2015, 2020; Liu et al., 2020b) have noted bulk rotation in the molecular gas around S255IR, these signatures are not evident on the smaller ($\sim 70$ au) scales probed here. There is, however, a hint of both H2O and H30$\alpha$ emission along the innermost part of the outflow, within about one resolution element of the central source. There is also a hint of SiS here. There is clearly SiO emission with a velocity gradient along the outflow axis, but curiously it is not very bright along the continuum jet. This source warrants followup at higher resolution and sensitivity to try to identify the location of the actual disk, though it is not yet clear which lines to use for this search. ### B.10 NGC6334IN The NGC6334IN region contains several bright sources. The brightest in our observations are SMA6 and SMA1 b/d (names from Sadaghiani et al., 2020). Neither source shows clear signs of rotation in any line. SMA1b/d contains several sources that may exhibit some hint of H2O emission, so we use that line as the basis for stacking We do not detect any other lines. Notably, both SO and SiO appear absent toward these sources. SMA6 also shows a hint of a water line, but similarly shows no sign of any of the other targeted lines. ### B.11 G11.92mm1 Ilee et al. (2018) reported on the disk G11.92mm1, showing a much larger ($r\sim 800$ au) size than most of the sources in our sample. We use SO $6_{5}-5_{4}$ as the guide line for stacking, since it shows clear disk kinematics. No brinary lines are detected.
2010 Vol. 10 No. XX, 000–000 11institutetext: LPTA, Université Montpellier 2 - CNRS/IN2P3, 34095 Montpellier, France 22institutetext: Department of Astronomy, Nanjing University, Nanjing 210093, P. R. China 33institutetext: European Southern Observatory, Alonso de Córdova 3107, Santiago, Chile 44institutetext: Department of Astronomy and Astrophysics, P. Universidad Catolica de Chile, Casilla 306, Santiago, Chile Received [year] [month] [day]; accepted [year] [month] [day] # Low-ionization galaxies and evolution in a pilot survey up to z = 1 00footnotetext: Based on observations obtained in service mode at the European Southern Observatory at Paranal E. Giraud 11 Q.-S. Gu 22 J. Melnick 33 H. Quintana 44 F. Selman 33 I. Toledo 44 P. Zelaya 44 ###### Abstract We present galaxy spectroscopic data on a pencil beam of $10.75^{\prime}\times 7.5^{\prime}$ centered on the X-ray cluster RXJ0054.0-2823 at $z=0.29$. We study the spectral evolution of galaxies from $z=1$ down to the cluster redshift in a magnitude-limited sample at $\rm R\leq 23$, for which the statistical properties of the sample are well understood. We divide emission- line galaxies in star-forming galaxies, LINERs, and Seyferts by using emission-line ratios of [OII], $\rm H\beta$, and [OIII], and derive stellar fractions from population synthesis models. We focus our analysis on absorption and low-ionization galaxies. For absorption-line galaxies we recover the well known result that these galaxies have had no detectable evolution since $z\sim 0.6-0.7$, but we also find that in the range $z=0.65-1$ at least 50% of the stars in bright absorption systems are younger than 2.5Gyr. Faint absorption-line galaxies in the cluster at $z=0.29$ also had significant star formation during the previous 2-3Gyr, while their brighter counterparts seem to be composed only of old stars. At $z\sim 0.8$, our dynamically young cluster had a truncated red-sequence. This result seems to be consistent with a scenario where the final assembly of E/S0 took place at $z<1$. In the volume-limited range $0.35\leq z\leq 0.65$ we find that 23% of the early-type galaxies have LINER-like spectra with $\rm H\beta$ in absorption and a significant component of A stars. The vast majority of LINERs in our sample have significant populations of young and intermediate-aged stars and are thus not related to AGN, but to the population of ‘retired galaxies’ recently identified by Cid-Fernandes et al. (2010) in the SDSS. Early-type LINERs with various fractions of A stars, and E+A galaxies appear to play an important role in the formation of the red sequence. ###### keywords: cosmology: observations – galaxies: evolution - large scale structures - evolution – RX J0054.0-2823 ## 1 Introduction In the course of an investigation of the diffuse intergalactic light in X-ray emitting clusters at intermediate redshifts (Melnick et al., 1999), we detected a puzzling S-shaped arc-like structure in the ROSAT cluster RX J0054.0-2823 (Faure et al., 2007), which we tentatively identified as the gravitationally lensed image of a background galaxy at a redshift between z=0.5 and z=1.0. The cluster, however, is characterized by having three dominant D or cD galaxies in the center, two of which are clearly interacting. We designed an observing strategy that allowed us at the same time to observe the arc, the diffuse Intra-Cluster Light (ICL), and a magnitude limited sample of individual galaxies in the field taking advantage of the multi-object spectroscopic mode of the FORS2 instrument on Paranal. By optimizing the mask design (see below) we were able to obtain: (a) very deep observations of the arc; (b) very deep long-slit observations of the ICL; and (c) redshifts and flux distributions for 654 galaxies of which 550 are in the pencil beam and at $0.275\leq z\leq 1.05$. Our pencil beam sample covers a redshift range up to z = 1 (with some galaxies up to z = 1.7). In standard cosmology with $H_{o}=75$ $\rm km~{}s^{-1}~{}Mpc^{-1}$, $\Omega_{0,m}=0.30$, and $\Omega_{0,\Lambda}=0.70$, this range provides a large leverage of about 3000 Mpc or 7 Gyr, which should be sufficient to extract some of the most conspicuous characteristics on galaxy evolution at $z\leq 1$. About half of all stars seem to be still forming, mostly in disks, in this redshift range (Dickinson et al., 2003; Hammer et al., 2005). Our spectroscopy provides a 50-60% complete sample of the galaxies in a pencil beam of $\sim 10^{\prime}\times 10^{\prime}$, centered on the cluster, uniformly down to R=23. Our sample compares in size with the DEEP1 spectroscopic pilot survey (Weiner et al., 2005) but is smaller than large surveys such as DEEP2 (e.q. Lin et al. 2008; Yan et al. 2009), VVDS (e.q. Franzetti et al. 2007; Garilli et al. 2008), GOODS (e.q. Bell et al. 2005; Weiner et al. 2006). The advantage of a pilot survey is that it can be handled rather easily by a single (or a few) researcher(s) to test new methods, new ideas before applying these new methods to large samples. The vast majority of our individual spectra reduced to zero redshift have S/N ratios per $\rm 2.6\AA$ pixel larger than 3 at $\rm 4200\AA$. This resolution is very well adapted to the detection of small equivalent width [OII] emission, which is expected to be found in bulge dominated galaxies with small disks, in some LINERs, in “mixed” mergers between E/S0 and star-forming objects, and perhaps in some post-starbursts galaxies. The line of sight of our field crosses three main structures: a dynamically young cluster at $z=0.29$, an over-dense region with layers at $z=0.4-0.5$, and a mixed region of field and possible layers from $z=0.6$ to $z=1$. According to morphology- density relations (Dressler, 1980; Dressler et al., 1997; Melnick & Sargent, 1977; Smith et al., 2005; Postman et al., 2005; Cooper et al., 2006; Scoville et al., 2007), we expect that over-dense regions will provide a rather large number of red objects available to our study. Therefore red objects with or without star formation, or with low photo-ionization is the subject which we will focus on, having in mind the possible roles of E+A galaxies (Dressler & Gunn, 1983; Norton et al., 2001; Blake et al., 2004; Goto, 2007; Yang et al., 2008, and references therein) and of LINERs (Yan et al., 2006) in the building-up of the red sequence. We focus on galaxies with either low star-formation or low ionization which appear at $z\leq 0.6$. We use line ratio diagnosis based upon [OII], $\rm H\beta$, and [OIII], from Yan et al. (2006), to classify galaxies in LINERs, star-forming galaxies, and Seyferts. This method, combined with visible morphology, allow us to isolate a significant population of early-type LINERs, and galaxies with diluted star-formation in later morphological types at $z=0.35-0.6$. Several studies suggest that the bulk of stars in early-type cluster galaxies had a formation redshift of $z\geq 3$, while those in lower density environments may have formed later, but still at $z\geq 1.5-2$ (for reviews see Renzini, 2006, 2007). This may be in contradiction with the rise in the number of massive red galaxies found by Faber et al. (2007) who concluded that most early types galaxies reached their final form below $z=1$. Our data include a clear red sequence at $z=0.29$ and a quite large number of absorption systems up to $z\sim 1$ which we fit with population synthesis models in order to search for age variations with $z$ and luminosity. The paper is structured as follows. Section 2 presents details of the observations and the data reduction procedures. Section 3 is on the resulting redshift catalog. Section 4 presents an overview of variations in spectral energy distribution with redshift for absorption and emission systems. Section 5 is dedicated to population variations with $z$ and luminosity in absorption systems. Low-ionization galaxies are in 5.3. In Section 5.4 we suggest a scenario in which early-type LINERs will become E/S0 galaxies once the A stars die, and photo-ionization disappear. Summary and Conclusions are in Section 6. ## 2 Observations and data reduction The observations (ESO program 078A-0456(A) were obtained with the FORS2 instrument (fors:2005, 2005) on the Cassegrain focus of the VLT UT1 telescope in multi-object spectroscopy mode with the exchangeable mask unit (MXU). They were acquired in service observing and were spread over two periods 78 and 80 to satisfy our observing conditions. FORS2 was equipped with two $\rm 2k\times 4k$ MIT CCDs with $15\rm\mu m$ pixels. These CCDs have high efficiency in the red combined with very low fringe amplitudes. We used the grisms 300V and 600RI, both with the order sorting filter GG435. With this filter, the 300V grism has a central wavelength at 5950 Å and covers a wavelength between $4450-8700$ Å at a resolution of 112 Å$~{}{\rm mm}^{-1}$. The 600RI grism has a central wavelength of 6780 Å and covers the 5120-8450 Å region at a resolution of 55 Å ${\rm mm}^{-1}$. Combined with a detector used in binned mode, the 300V grism has a pixel resolution of 3.36 Å pixel-1. The grisms were used with a slit width of 1′′. In order to match the major and minor axis of the ICL and the prominent arc-like feature rotation angles of $-343^{o}$, $-85^{o}$, and $-55^{o}$ were applied. The slit lenghts used for the ICL spectra are 56.5′′, 32.5′′, and 24.5′′, while those of typical galaxies vary between 7′′ and 12′′. The ICL was located either on the master CCD or the second one, resulting in a combined pencil beam field of $\rm 10.75^{\prime}\times 7.5^{\prime}$ (Figure 1). A total of 30 hours of observing time including field acquisition, mask positioning, and integration time were dedicated to our pencil beam redshift survey of the J0054.0-2823 field. Each mask was filled with 39-49 slitlets in addition to the ICL slits. In order to trace some of the apparent structures connected to J0054.0-2823, and to reach beyond its Virial radius, we also obtained MXU exposures of 8 FORS2 fields of $\rm 7^{\prime}\times 5^{\prime}$ adjacent to the pencil beam, so in total we obtained spectra of 730 individual sources. Figure 1: The central (pencil beam) field from R images obtained with the wide field camera at the 2.2m telescope in La Silla ### 2.1 Mask preparation Tables for preparing the masks and instrument setups were obtained with the FORS Instrumental Mask Simulator111 http://www.eso.org/sci/observing/phase2/FORS/FIMS.html (FIMS, 2006). The selection of the objects for the preparation of the slit masks of the pencil beam field was done by using a photometric catalog in V and I which we had derived from deep images obtained in a previous NTT run (Faure et al., 2007), and pre-images in R from the VLT. The selection of the objects in the fields adjacent to the pencil beam were obtained by using images taken with the WIde Field Imager (WIFI; $34^{\prime}\times 33^{\prime}$) at the 2.2m telescope on La Silla. Photometry in V and R from the WFI images are used throughout the paper. The allocated time was divided in observing blocks (OBs) to be executed in service mode. A typical OB of 1h execution time had a science integration time of 2900s in two exposures of 1450s. We estimated exposure times for E to Sb galaxies in the range z = 0.3 - 0.8. Using the exposure time calculator of FORS, we obtained magnitude limits, the major steps of which are given in Table LABEL:maglimit, which we used to optimize the distribution of slitlets in the masks. After isolating bright objects which did not require long exposure time, we prepared a grid with an exposure time step $\rm 2\times 1450~{}s$ which we filled with galaxies having V magnitudes such that the expected S/N ratio would be better than 2.8 (1 pixel along the dispersion). After receiving VLT pre-images in the red band, we did a similar grid in R and adjusted the two grids. The masks were prepared interactively with the FIMS tool and the R pre-images. We started to fill masks with objects that require an exposure time $\rm\leq 2\times 1450~{}s$, then moved to $\rm\leq 4,~{}6,~{}and~{}8\times 1450~{}s$. Because we prepared sets of masks with slits in very different directions (those of the ICL long and short axis in particular), objects that could not be targeted with a mask in a given direction (i.e. such as any mask with running name ICL-s in Table 2) were targeted in a perpendicular one (i.e. masks with running name ICL-L), an approach which made the mask filling quite efficient, in particular in over-dense areas and field edges. Objects which were close to a predicted S/N of 2.8 in an OB, were selected to be also observed in another OB as often as possible. Some objects with expected good S/N in an OB, were re-observed in another OB when there was no other target in the corresponding slit strip. They provides a set of high S/N $(\sim 20)$ ratio spectra. A total of 973 slitlets were selected, 621 in 14 different masks in the pencil beam field, and 352 in 8 masks in the adjacent fields. Thirty five percent of the sources of the pencil beam field were observed through different masks, whereas the slitlets of the adjacent fields are all for different sources. Table 1: Table used for preparing MXU plates of multiple Observing Blocks Number of OBs of 1h | Integration time | Magnitude limit in V | S/N for S0-Sb at $0.3\leq z\leq 0.8$ ---|---|---|--- 1 | 2900s | 24.4 - 24.8 | 2.8 - 5.2 2 | 5800s | 24.8 - 25.2 | 2.8 - 5.2 4 | 11600s | 25.2 - 25.6 | 2.8 - 5.2 The resulting list of masks and OBs, and the journal of observations are given in Table 2. Spectra of the pencil beam field were obtained through masks with running names Bright, ICL-L, ICL-s, and arc. ICL-L and ICL-s were obtained with rotator angle $-343^{o}$ and $-85^{o}$ respectively, and arc with a rotation of $-55^{o}$. Masks with names SE, E, NE, N, NW, W, SW1 & SW2 are on adjacent fields. The observations were obtained during clear nights, with seeing between $0.7^{\prime\prime}$ and $1.5^{\prime\prime}$ and dark sky. Table 2: Journal of the MXU Observations Name | OB ID | Date | Exp. time (s) | # slitlets | Grism ---|---|---|---|---|--- Bright1 | 255728 | 20 Oct. 06 | $3\times 550$ | 34 | 600RI Bright2 | 255726 | 23 Oct. 06 | $3\times 550$ | 38 | 600RI SW1 | 255710 | 18 Oct. 06 | $3\times 550$ | 45 | 300V SW2 | 255708 | 15 Oct. 06 | $3\times 550$ | 40 | 300V W | 255712 | 19 Oct. 06 | $3\times 550$ | 49 | 300V SE | 255706 | 3 Oct. 07 | $3\times 550$ | 42 | 300V N | 255716 | 5 Oct. 07 | $3\times 550$ | 42 | 300V NW | 255714 | 5 Oct. 07 | $3\times 550$ | 48 | 300V NE | 255718 | 14 Oct. 07 | $3\times 550$ | 47 | 300V E | 255704 | 15 Oct. 06 | $3\times 710$ | 39 | 600RI ICL-s1 | 255750 | 12 Dec. 06 | $2\times 1450$ | 46 | 300V ICL-s2 | 255748 | 15 Nov. 06 | $2\times 1450$ | 48 | 300V ICL-L1 | 255761 | 12 Dec. 07 | $2\times 1450$ | 41 | 300V ICL-L2 | 255763 | 9 Jan. 07 | $2\times 1450$ | 39 | 300V arc2 | 255734 | 24 Nov. 06 | $2\times 1450$ | 48 | 300V arc1 | 255736, 38 | 9 Jan. 07, 11 Sept. 07 | $4\times 1450$ | 43 | 300V ICL-s3 | 255744, 46, 47 | 27 Oct. 06, 9 Nov. 06 | $6\times 1450$ | 47 | 300V ICL-L3 | 255752, 59, 60 | 21 Sept. 07, 31 Oct. 07 | $6\times 1450$ | 49 | 300V arc3 | 255730, 32, 33 | 17 Aug. 07 | $6\times 1450$ | 46 | 300V ICL-s4 | 255739, 41, 42, 43 | 23 Oct. 06, 13 Nov. 06 | $8\times 1450$ | 46 | 300V ICL-L4 | 255754, 56, 57, 58 | 15, 17 & 20 Nov. 06 | $8\times 1450$ | 49 | 300V RI | 255720, 22, 23, 24, 25 | 13 Nov. 06, 12 & 14 Sept. 07, | | | | | & 3 Oct. 07 | $10\times 1450$ | 47 | 600RI ### 2.2 Spectral extraction The data were reduced by the ESO quality control group who provided us with science products (i.e. sky subtracted, flat fielded and wavelength calibrated spectra of our objects), together with calibration data: master bias (bias and dark levels, read-out noise), master screen flats (high spatial frequency flat, slit function), wavelength calibration spectra from He-Ar lamps, and a set of spectrophotometric standards, which were routinely observed. The sky subtracted and wavelength calibrated 2D spectra allowed a very efficient extraction of about 60 % of the spectra. Nevertheless the pipeline lost a significant fraction of objects, in particular when they were located on the edges of the slitlets. To increase the efficiency of the spectral extraction we performed a new reduction starting from frames that were dark subtracted, flat-fielded and wavelength calibrated, but not sky subtracted, using a list of commands taken from the LONG context of the MIDAS package. For each slitlet, the position of the object spectrum was estimated by averaging 500 columns in the dispersion direction between the brightest sky lines and measuring the maximum on the resulting profile. The sky background was estimated on one side of the object, or on both, depending on each case. Spatial distortion with respect to the columns was measured on the sky line at 5577 Å and used to build a 2D sky which was subtracted to the 2D spectrum. Multiple exposures where then aligned and median averaged. The 1D spectra of objects were extracted from 2D medians by using the optimal extraction method in MIDAS. ### 2.3 Redshift identification The identification of lines for determining the redshifts was done independently by two methods and three of the authors. The 2D spectrum was visually scanned to search for a break in the continuum, or an emission-line candidate (e.g. [OII] $\lambda$3728.2 Å). A plot of the 1D spectrum was displayed in the corresponding wavelength region to search for [OII], the Ca H & K lines, and/or Balmer lines H$\epsilon$, H9 $\lambda$3835.4 Å, H8 $\lambda$3889.1 Å, H10 $\lambda$3797.9 Å, and H$\delta$. The redshift was then confirmed by searching for the [OIII] doublet $\lambda$4958.9 & 5006.8 Å, and H$\beta$ in emission if [OII] had been detected, or G and the Mgb band, if the 4000 Å break and (or) the H and K lines had been identified. The MgII $\lambda$2799 Å line in absorption and, in some cases AlII $\lambda$ 3584 Å, were searched to confirm a potential redshift $z\geq 0.65$, while in the cases of low redshift candidates we searched for $\rm H\beta$, the NaD doublet $\lambda$5890 & 5896 Å, and in a few cases H$\alpha$. The H$\gamma$ line, the E (FeI+CaI $\lambda$5270 Å) absorption feature and, in some bright galaxies the Fe $\lambda$4383 Å, Ca $\lambda$4455 Å, Fe $\lambda$4531 Å absorption lines, were used to improve the redshift value. The resulting identification ratio of galaxy redshifts is of the order of 90%. The 10% of so-called unidentified include stars, objects with absorption lines which were not understood, a few objects with low signal, and defects. Six QSO’s were also found. An example of good spectrum of red galaxy, with its main absorption lines identified, is shown in Figure 2. Figure 2: Example of a spectrum of a red and bright galaxy with [OII] and the main absorption lines identified A second independent visual identification was performed using Starlink’s Spectral Analysis Tool (SPLAT-VO), matching an SDSS reference table of emission and absorption lines222http://www.sdss.org/dr5/algorithms/linestable.html to the spectra. After a first estimate of the redshift a cross-correlation was performed using the FXCOR task on the RV package of IRAF333IRAF is distributed by the National Optical Astronomy Observatory, which is operated by the Association of Universities for Research in Astronomy, Inc., under cooperative agreement with the National Science Foundation.. Due to the large span of redshifts, two sets of templates were used. The first one consisting of 3 template spectra of galaxies ($\lambda=3500-9000$Å with emission and absorption lines and a dispersion of 3Å/pix) from the SDSS survey444http://www.sdss.org/dr2/algorithms/spectemplates/index.html with continuum subtraction using a spline3 order 5 fitting function. The second set of templates were two average composite spectra of early type and intermediate type galaxies ($\lambda=2000-7000$Å with only absorption lines and a dispersion of 2Å/pix) from the K20 survey555http://www.arcetri.astro.it/$\sim$k20/spe_release_dec04/index.html using a spline3 order 7 function for continuum subtraction. An interactive selection of the wavelength range used in the cross correlation was done on each spectrum avoiding contamination by sky lines. The spectra were re-binned to the template dispersion (smaller for 300V spectra and larger for 600RI spectra) , which gave the best results. Velocity errors were determined from the quality of the cross-correlation, by using standard R value of Tonry & Davies (Tonry:1979p5170, 1979). Here we used $R_{T}$ to differentiate it from the R band magnitudes symbol. These values are provided in the IRAF task FXCOR and explained in the reference quoted. In brief, $R_{T}$ is proportional to the ratio of the fitted peak height and the antisymmetric noise as defined by Tonry & Davies (1979). The redshifts, $R_{T}$ values, and velocity errors are given in Table 5, which also includes the list of visually identified lines. A third independent visual inspection was carried out when a discrepancy was seen between the previous two sets of measurements, and also in the very few cases were no redshift could be measured. For these spectra we first tried to detect emission or absorption lines and then used Gaussian fits to establish the line centroids and their errors and shifts. The redshift of each line was measured independently and the galaxy redshift was obtained from the weighted average of all lines. This third inspection resolved nearly all the few remaining discrepancies so we have retained the cross-correlation values whenever possible. We note that Xcorr failed in two instances: 1) for $z>0.8$ galaxies with low S/N and few weak absorption lines, and, 2) when no absorption lines, but 1, 2 or 3 clear emission lines were present. In these cases we used the visual line identifications and assigned a conservative error of 300 km/s. Spectra from more than one mask were obtained for 94 objects. Their final velocities and velocity errors were calculated as error-weighted means from multiple observations, although no significant disagreements were found. These repeated observations serve as a check on the internal errors. Figure 3 presents the differences between the cross-correlation velocity measurements for all galaxies with multiple observations. The representative full width half maximum (FWHM) error is 200 km/s. In Figure 4 we have plotted the relation between velocity errors and the Tonry $R_{T}$ value obtained in our cross correlations. Most errors are $<300$ km/s even for $4>R_{T}>2$ and the typical error is of order 80 km/s with the vast majority of the radial velocities have errors below 200 km/s. We have only discarded a few values with $R_{T}<1$ when there were no measurable emission lines. Figure 3: Radial velocity differences for galaxies with multiple observations. The objects with velocity discrepancies larger than $400$ km/s are broad line QSO’s and one high-z galaxy. Figure 4: Relation between radial velocity errors ($V_{err}$) and the Tonry $R_{T}$ parameter (Tonry:1979p5170, 1979) in redshifts obtained by cross-correlation. The points away from the general trend (5 points with $R_{T}>10$ and 5 with $V_{err}>500~{}\rm km~{}s^{-}1$) are 5 QSO’s and distant weak spectrum galaxies with emission lines. Objects with $V_{err}>500~{}\rm km~{}s^{-}1$, marked in red, were not used in combined spectra. ### 2.4 Flux Calibrations The 1D spectra were divided by the response curve of the detector, which had been determined from 4 spectrophotometric standard stars observed along the runs, and reduced by the same method (bias, flat field, wavelength calibration, and extraction) as the galaxy spectra. The thick absorption telluric band of O2 centered at 7621 Å (unresolved line series) was not removed from the observation response curve and was considered as a feature of the global wavelength dependent efficiency. The relative fluxes per wavelength of the corrected spectra can be compared with stellar population models, in arbitrary unit, but are not calibrated in flux. The spectra were re-binned to the z = 0 rest frame with relative flux conservation. Because a significant fraction of spectra have a too low S/N ratio for a meaningful comparison with population synthesis models, one may either select the brightest objects or combine spectra of similar types. The spectra taken at different locations of the MXU masks have different lengths along the dispersion direction. In order to merge them the spectra were normalized to have the same flux in the region 4050-4250 Å (see below). ### 2.5 Quality of the spectra The final S/N ratio of the extracted spectra, corrected for the response curve, and re-binned to zero redshift depends on a number of parameters: seeing, night sky transparency and background, magnitude of the object and integration time, wavelength of the S/N measurement, and redshift. To give an idea of the final products we present in Table 3 a representative set of 28 spectra at various z, magnitudes, number of OBs and resulting S/N ratio measured on zero redshift spectra in the wavelength range 4150-4250 Å which corresponds well to the location where we will measure the main indexes of this work. S/N ratios of spectra re-binned to zero redshift are for a pixel element of $\rm 2.6~{}\AA$ throughout the paper. Table 3 gives also the names of the OBs. Table 3: Signal-to-noise ratio of representative spectra. The columns indicate respectively: the redshift $(z)$ of a selected object, its V and R Petrosian magnitudes, the number of observing blocks, N(OB), from which its spectrum is extracted, the S/N ratio measured in the wavelength range 4150-4250Å of the spectrum rebinned to zero redshift, the name of observing blocks from Table 2, and the grism used. Spectra from OB’s with running name “arc” have on the average higher S/N ratio than those with name “ICL” as illustrated by the two objects marked (*). $z$ | V | R | N(OB) | S/N | Name of OBs | Grism ---|---|---|---|---|---|--- 0.2923 | 19.2 | 18.6 | 2 | 14 | arc1 | 300V 0.2932 | 20.3 | 19.3 | 2 | 18 | ICL-L1 & L2 | 300V 0.2928 | 21.6 | 20.5 | 3 | 17 | arc2 & ICL-L1 | 300V 0.2905 | 22.8 | 22.1 | 2 | 10 | ICL-L1 & L2 | 300V 0.2910 | 23.5 | 22.8 | 3 | 9 | arc1 & 2 | 300V 0.4486 | 21.3 | 20.0 | 1 | 6 | Bright2 | 600RI 0.4477 | 22.3 | 21.3 | 2 | 20 | arc1 | 300V 0.4148 | 23.0 | 22.3 | 2 | 7 | ICL-s1 & s2 | 300V 0.4538 | 23.1 | 22.0 | 5 | 14 | ICL-L4 & arc2 | 300V 0.5355 | 22.3 | 20.9 | 1 | 8 | arc2 | 300V 0.6309 | 22.7 | 21.4 | 4 | 11 | arc2 & 3 | 300V 0.6553 | 22.3 | 21.5 | 4 | 9 | ICL-s4 & arc2 | 300V 0.6282 | 23.5 | 22.5 | 5 | 9 | arc1 & 3 | 300V 0.6267 | 23.9 | 22.9 | 4 | 10 | arc3 | 300V 0.6864 | 22.6 | 21.9 | 1 | 4 | ICL-s2 | 300V 0.6886 | 23.0 | 22.0 | 7 | 13 | ICL-L3 & L4 | 300V 0.6861 | 23.0 | 22.1 | 4 | 8 | ICL-s4 | 300V 0.6864 | 23.5 | 22.3 | 4 | 10 | ICL-s3 & arc2 | 300V 0.6879 | 23.8 | 22.8 | 4 | 7 | ICL-s4 | 300V 0.8222 | 20.7 | 20.0 | 1 | 10 | ICL-s1 | 300V 0.8287 | 22.7 | 22.4 | 5 | 10 | ICL-s4 & arc2 | 300V 0.8249 | 23.2 | 22.6 | 3 | 8 | arc3 | 300V 0.8823 | 23.8 | 23.4 | 4 | 3.5 | ICL-s4 | 300V 0.9792 | 23.3 | 22.7 | 3 | 8 | arc3 (*) | 300V 0.9626 | 23.2 | 22.7 | 4 | 5 | ICL-L4 (*) | 300V 0.9637 | 23.4 | 23.2 | 3 | 6 | ICL-s3 | 300V 0.9809 | 23.8 | 23.7 | 5 | 6 | RI | 600RI 1.0220 | 24.1 | 23.3 | 4 | 3 | ICL-s4 | 300V ### 2.6 Spectral indexes The 4000 Å break amplitude definition used in the present paper is the ‘narrow’ 4000 Å break defined by Balogh et al. (1999) as the flux ratio in the range 4000-4100Å over 3850-3950Å (e.g. Kauffmann et al., 2003). The error in D(4000) is calculated from the spectral noise in the two passbands. The equivalent widths of [OII] and of $\rm H\delta$ were measured by using the MIDAS context ALICE as follows: the continuum was obtained by linear interpolation through two passbands each side of the line, a Gaussian was fitted to the emission or absorption line, and an integration was done over the resulting Gaussian profile above or below the continuum. The continuum and line fits, and the integration were done interactively on a graphic window in which the spectral region of the line was displayed. Table LABEL:integr lists the wavelength ranges of the sidebands used to define the fluxes and continua. Table 4: Wavelength bands used in the measurement of 4000 Å break amplitude, and in the determination of the continua of the [OII] and $\rm H\delta$ indexes (equivalent widths). Index | Blue band | Red band ---|---|--- D(4000) | 3850 - 3950 Å | 4000 - 4100 Å EQW([OII]) | 3650 - 3700 Å | 3750 - 3780 Å $\rm EQW(H\delta)$ | 4030 - 4070 Å | 4130 - 4180 Å Uncertainties in equivalent widths were deduced from simple Monte Carlo: the values of the equivalent widths are the average of 20 continuum determinations and best Gaussian fits to the absorption or emission lines, and the errors in equivalent widths are deduced from the Monte Carlo dispersion. The largest index errors are for spectra in which $\rm H\delta$ is both in absorption and in emission. In such cases the emission line was removed after fitting the spectrum of an A star onto all Balmer lines to estimate the depth of $\rm H\delta$ in absorption, and this step was added to the Monte Carlo. The errors on indexes given in Tables of combined spectra throughout the paper are those which were measured on combined spectra. They do not take into account the astrophysical dispersions in the distributions of individual galaxies which were used to build combined spectra. Those astrophysical dispersions are given in relevant Tables concerning spectral variations. Full observational measurement errors on indexes of individual spectra were obtained by measuring $\rm D(4000)$ and $\rm EQW([OII])$ on spectra with multiple observations. Thus $17\%$ of the spectra have typical errors of $4\%$ in D(4000) and $10\%$ in EQW([OII]); $54\%$ have typical errors of $8\%$ in D(4000) and $20\%$ in EQW([OII]); and $14\%$ have poorer spectra with typical errors of $16\%$ in D(4000) and $40\%$ in EQW([OII]). ### 2.7 Stellar Population Analysis In order to study the stellar population quantitatively, we applied a modified version of the spectral population synthesis code, starlight666http://www.starlight.ufsc.br/ (Cid Fernandes et al., 2004; Gu et al., 2006) to fit the observed and combined spectra. The code does a search for the best-fitting linear combination of 45 simple stellar populations (SSPs), 15 ages, and 3 metallicities ($0.2\,Z_{\odot}$, $1\,Z_{\odot}$, $2.5\,Z_{\odot}$) provided by Bruzual & Charlot (2003) to match a given observed spectrum $O_{\lambda}$. The model spectrum $M_{\lambda}$ is: $M_{\lambda}(x,M_{\lambda_{0}},A_{V},v_{\star},\sigma_{\star})=M_{\lambda_{0}}\left[\sum_{j=1}^{N_{\star}}x_{j}b_{j,\lambda}r_{\lambda}\right]\otimes G(v_{\star},\sigma_{\star})$ (1) where $b_{j,\lambda}=L_{\lambda}^{SSP}(t_{j},Z_{j})/L_{\lambda_{0}}^{SSP}(t_{j},Z_{j})$ is the spectrum of the $j^{\rm th}$ SSP normalized at $\lambda_{0}$, $r_{\lambda}=10^{-0.4(A_{\lambda}-A_{\lambda_{0}})}$ is the reddening term, $x$ is the population vector, $M_{\lambda_{0}}$ is the synthetic flux at the normalization wavelength, and $G(v_{\star},\sigma_{\star})$ is the line-of- sight stellar velocity distribution modeled as a Gaussian centered at velocity $v_{\star}$ and broadened by $\sigma_{\star}$. The match between model and observed spectra is calculated as $\chi^{2}(x,M_{\lambda_{0}},A_{V},v_{\star},\sigma_{\star})=\sum_{\lambda=1}^{N_{\lambda}}\left[\left(O_{\lambda}-M_{\lambda}\right)w_{\lambda}\right]^{2}$, where the weight spectrum $w_{\lambda}$ is defined as the inverse of the noise in $O_{\lambda}$. The code yields a table with input and ouput parameters for each component. Input parameters include individual stellar masses, ages, metallicities, L/M, … and ouput parameters include luminosity fractions, mass fractions, fit parameters of individual components …, and global parameters such as velocity dispersion and extinction. For more details we refer to the paper by Cid Fernandes et al. (2005). In the present work we use the standard luminosity fraction in the rest frame of normalized spectra at $\rm 4050~{}\AA$, which we compare in different redshift bins. Figure 5: Spectral fitting results with SSP models for the redshift $<z>=0.29$ bin. (a): Observed (thin black line), model (red line) and residuals for the absorption spectrum. Points indicate bad pixels and emission-line windows that were masked out during fitting. (b): Emission-line spectrum; (c): Total spectrum. Figure 5 shows an example of the fit for the averaged spectrum at $<z>=0.29$. Panels (a), (b), and (c) correspond to absorption-line, emission-line, and all spectra respectively. After fitting the spectra, we rebin the 45 SSPs into 5 components according to their age: I ($10^{6}\leq t<10^{8}$ yr), II ($10^{8}\leq t<5\times 10^{8}$ yr), III ($5\times 10^{8}\leq t<10^{9}$ yr), IV ($10^{9}\leq t<2.5\times 10^{9}$ yr), and V ($t\geq 2.5\times 10^{9}$ yr). Components with the same age and different metallicities are combined together. ## 3 The Catalog of Galaxies and Large Scale Structures in the Line of Sight in the Pencil Beam Table 5 presents positions, redshifts, Petrossian R-magnitudes ($m_{R}$), and line identifications for the full sample of 654 galaxies observed in our program. The radial velocities and the corresponding measurement errors are also given. The full Catalogue from which Table 5 is extracted will be sent as a public database to CDS. The rough data are presently in the public domain at ESO. Table 5: Properties of galaxies in the field of RX J0054.0-2823 obj | RA ($\alpha$) | DEC ($\delta)$ | z | $m_{R}$ | V | Verr | $R_{T}$ | Nobs | lines ---|---|---|---|---|---|---|---|---|--- | J2000 | J2000 | | | km/s | km/s | | | 23 | 13.598707 | -28.434965 | 0.79304 | 22.78 | 237912 | 161 | 4.7 | 1 | K–H 26 | 13.590379 | -28.416917 | 0.77636 | 22.26 | 232908 | 77 | 8.1 | 1 | [OII]–H10–H9–H 27 | 13.584442 | -28.394515 | 0.41463 | 22.29 | 124389 | 22 | 17.3 | 1 | [OII]–H9–H–H$\beta$–[OIII] 28 | 13.586628 | -28.438063 | 0.44877 | 22.02 | 134631 | 73 | 6.7 | 1 | K–H–G 30 | 13.580301 | -28.435437 | 0.29032 | 21.08 | 87096 | 68 | 11.3 | 1 | H9–K–H–H$\delta$–H$\alpha$ 31 | 13.572009 | -28.380385 | 0.63267 | 21.32 | 189801 | 49 | 11.5 | 1 | [OII]–K–H 32 | 13.579012 | -28.439414 | 0.45335 | — | 136005 | 22 | 19.5 | 1 | [OII]–H$\gamma$–H$\beta$–[OIII] 33 | 13.574997 | -28.439377 | 0.44741 | 19.21 | 134223 | 80 | 11.4 | 1 | K–H–G–H$\beta$ 34 | 13.573913 | -28.442855 | 0.63013 | 20.63 | 189039 | 87 | 7.9 | 1 | K–H–H$\delta$–G 35 | 13.571781 | -28.435423 | 0.44862 | 20.26 | 134586 | 73 | 9.8 | 1 | H9–K–H–G–H$\beta$ Figure 6 shows the R-magnitude histogram of the galaxies with measured redshifts superimposed on the magnitude histogram of all galaxies in our pencil-beam indicating that our observations sample uniformly at a rate of 50-60% the population of galaxies down to $\rm R=22.5$. The sampling seems fairly representative in the magnitude bin $\rm R=22.5-23.0$, and sparse at $\rm R>23$. The apparent increase in incompleteness toward brighter magnitudes is due to a selection bias in the observations, which were designed to avoid bright galaxies at redshifts $z\leq 0.25$. Figure 6: R-magnitude histogram of galaxies with measured redshift in the central beam. Figure 7 presents the magnitude redshift relation and the cone diagrams for the full sample. The points are color coded according to the presence or absence of emission lines. Figure 7: (a) Magnitude redshift relation for the full sample. The three lines overploted over the measured points correspond to absolute R magnitudes of -22.5, -20.5, and -18.5. The distances have been calculated using a cosmology with $\Omega_{0,\Lambda}=0.70$, $\Omega_{0,m}=0.30$, $w=-1$, and $\rm H_{0}=75km~{}s^{-1}~{}Mpc^{-1}$ (h = $\rm H_{0}/75km~{}s^{-1}~{}Mpc^{-1}$). Red dots are galaxies with no emission lines and blue dots are galaxies with emission lines. (b) Cone diagrams in Dec for all the galaxies measured in the field of RX J0054.0-2823. The scales is in Mpc calculated using the angular distance for the standard cosmology. The detection threshold for emission-lines is $\rm EQW([OII])\sim 2-3$Å. (c) Same as (b) but for RA. A cursory inspection of Figure 7 reveals the presence of several conspicuous structures - walls of objects spanning almost the entire field of view - over the full range of redshifts covered by our observations. Ignoring objects with $z<0.28$, we see structures centered at $z=0.29$ (our prime target); two distinct structures at $z\sim 0.4$, which we will denote $z=0.415$ and $z=0.447$; a rather complex structure at $z\sim 0.6$, with two main over- densities at $z=0.58-0.63$, and $z=0.68$; a single rather sparsely populated layer at $z=0.82$. In what follows, we will refer to these groups (including the main cluster at $z=0.29$) as our pencil beam structures. Making bins centered on the peaks of the redshift distribution maximizes the number of objects in each bin and minimizes its redshift dispersion. So using the apparent structures rather than a blind slicing appears well adapted to our sample. If the structures are real, the objects of a given structure may have a common history and this may also help to reduce the cosmic scatter. The numbers of spectra observed in each structure are given in Table LABEL:zdist. The redshift bins given in column (2) are chosen a posteriori to fit the structures. The numbers of redshifts measured in each bin are given in column (3), with the respective numbers of absorption and emission systems in columns (4) & (5). We have determined the median distance to the nearest object $\triangle\delta$ in each of the apparent structures in column (6), and the median velocity dispersion, $\rm\sigma(V)\equiv c\sigma(z)/(1+z)$ in column (7). We give a rough morphology of the structures in column (1). The cluster at $z=0.293$ appears to have small projected separation and velocity dispersion. Layers or filaments have a comoving velocity dispersion (dynamical and cosmological) less than $\rm\sim 1500~{}km~{}s^{-1}$; clouds have $\rm\sigma(V)>2000~{}km~{}s^{-1}$. The structures marked “filaments” are the arc layers seen in Figure 7. Projected on the sky they seem to be filamentary, but the median distances $\triangle\delta$ to the nearest object are approximately 2/3 those expected for uniform distributions, so they are not clearly different from 2D layers. The names of the bins that are used to combine spectra are given in column (8). Large scale arc structures, as seen in cone diagrams, are expected to be formed by infall of galaxies on gravitational potentials: galaxies which are on the far side have a negative infall velocity, while those on the nearby side have a positive infall component, which when superimposed on the Hubble flow reduces the velocity dispersion. This is presumably what we observe in the two filaments or layers with low velocity dispersion at z = 0.4. We combined the spectra in each structure using the median. This results in a slightly lower total S/N (by $\sqrt{2}$), but allows to eliminate spurious features. Table 6: Apparent structures in the field of RX J0054.0-2823 Apparent Structure | z range | N | N(abs) | N(em) | $\triangle\delta~{}(kpc)$ | $c\sigma(z)/(1+z)$ | Composite name ---|---|---|---|---|---|---|--- (1) | (2) | (3) | (4) | (5) | (6) | (7) | (8) | 0.275 - 0.285 | 0 | 0 | 0 | | | Cluster | 0.285 - 0.298 | 91 | 60 | 31 | 165 | 527 | SPEC029 | 0.298 - 0.320 | 5 | 1 | 4 | | | | 0.320 - 0.330 | 12 | 7 | 5 | | | | 0.330 - 0.390 | 28 | 7 | 21 | | | | 0.390 - 0.430 | 35 | 6 | 29 | 490 | 3250 | SPEC0415 Filament (layer) | 0.432 - 0.440 | 29 | 5 | 24 | 500 | 350 | SPEC0415 | 0.440 - 0.444 | 1 | 1 | 0 | | | Filament (layer) | 0.444 - 0.456 | 46 | 21 | 25 | 470 | 450 | SPEC0447 | 0.456 - 0.465 | 1 | 1 | 0 | | | Cloud | 0.465 - 0.550 | 53 | 15 | 38 | 480 | 3370 | Cloud | 0.550 - 0.620 | 56 | 15 | 41 | 520 | 3550 | SPEC063 Filament (layer) | 0.620 - 0.657 | 48 | 12 | 36 | 580 | 1450 | SPEC063 | 0.657 - 0.673 | 1 | 0 | 1 | | | Filament (layer) | 0.673 - 0.696 | 43 | 12 | 31 | 650 | 1010 | SPEC068 | 0.710 - 0.790 | 22 | 4 | 18 | | | Filament & cloud | 0.790 - 0.850 | 33 | 6 | 27 | 820 | 2050 | SPEC082 | 0.850 - 0.880 | 0 | 0 | 0 | | | Cloud | 0.880 - 0.930 | 11 | 1 | 10 | | | | 0.930 - 0.946 | 0 | 0 | 0 | | | Cloud | 0.946 - 1.046 | 25 | 4 | 21 | 1080 | 4110 | SPEC099 ### 3.1 Magnitudes The R-band average magnitudes of galaxies in each redshift bin are given in Table LABEL:magnitudes separately for absorption, “red” and “blue” emission- line galaxies, together with the adopted distance moduli. The partition “red” versus “blue” is defined by the median spectral slope in each redshift bin. In a study on emission line galaxies (Giraud et al., 2010) we divided the sample of emission-line galaxies in two halves: those with continuum slopes bluer than the average and those with continuum slopes redder than the average in each redshift bin. This was done interactively by displaying reduced 1D spectra and using MIDAS. While a median partition is not necessarily a physical partition, we showed that, in the present case, it divides ”young” galaxies, for which the evolution is dominated by on-going star formation from ”old” galaxies where the evolution is dominated by changes in the older stellar populations. Table 7: Average R-band magnitudes of absorption systems (abs), and red and blue emission-line galaxies. The adopted distance moduli $(m-M)_{0}$ and the 4150-4250Å fluxes $f$ normalized to the blue galaxies at $z=0.9$ are also tabulated. $<z>$ | R(abs) | R(red) | R(blue) | $(m-M)_{0}$ | $f$(abs) | $f$(red) | $f$(blue) ---|---|---|---|---|---|---|--- 0.29 | 19.80 | 20.12 | 20.97 | 40.18 | 0.74 | 0.72 | 0.48 $0.43$ | 20.24 | 20.59 | 20.95 | 40.86 | 1.08 | 0.92 | 0.75 $0.65$ | 21.50 | 21.60 | 21.94 | 41.51 | 1.42 | 1.28 | 0.86 $0.9$ | 22.45 | 22.13 | 22.35 | 41.98 | 2.08 | 1.70 | 1 We used the R-band photometry to calibrate individual spectra by convolving each spectrum with a box filter 1290 Å wide, centered at $\lambda=6460$ Å. Once the spectra were calibrated in the observer R-band, we measured the average fluxes in the wavelength range 4150-4250Å of the galaxies, which we normalized to the flux of blue emission galaxies at $<z>=0.9$ to compute the luminosity index $f$. Thus $f$ (that is equal to 1 for blue galaxies at $<z>=0.9$) is an indicator of AB(4200) that allows us to compare the luminosities of red and blue galaxies at a given redshift and to investigate luminosity variations with $z$. Thus Table LABEL:magnitudes clearly shows that in each redshift bin, absorption-line and red emission-line galaxies are more luminous than blue galaxies. ## 4 Composite spectra Each galaxy spectrum was wavelength calibrated, corrected for instrument response, re-binned to zero redshift, and normalized to have the same flux in the wavelength range $\rm\triangle\lambda=4050-4250\AA$. Normalizing spectra gives the same weight to all galaxies. As a consequence stellar fractions must be understood as average stellar fractions per galaxy. We have truncated the sample at $z=1.05$ and assembled the spectra in bins centered on (pseudo) structures at 0.29, 0.41, 0.45, 0.63, 0.68, 0.82, and 0.99 to build high S/N composite spectra for each bin. In order to compensate (or at least alleviate) for Malmquist bias we rejected objects fainter than $\rm M_{R}=-18.8$ mostly at $z\leq 0.45$ (Figure 7a). A sample completely free of Malmquist bias would require a cutoff at $\rm M_{R}\sim-20.5$. For clarity of the figures, we often combined the mean spectra at $z=0.41$ & $0.45$ into a single bin at $<z>=0.43$, the spectra at $z=0.63$ & $0.68$ into a bin at $<z>=0.65$, and in some cases the spectra at $z=0.82$ & $0.99$ into a bin at $<z>=0.9$. The spectra of galaxies in these four bins are presented in Figure 8 where we show the spectra of absorption systems (top) and emission line galaxies (bottom) separately. The corresponding $4000\AA$ break amplitudes are given in Table LABEL:D4000 Figure 8: Composite spectra of absorption systems (top); and emission line galaxies (bottom) normalized in the wavelength range $\rm\triangle\lambda=4050-4250$ Å. All individual galaxies are brighter than $\rm M_{R}=-18.8$ Table 8: 4000Å break amplitudes for absorption (abs) and emission (em) galaxies, and equivalent width of H$\delta$ for absorption galaxies with measurement errors. The S/N ratios of the combined spectra were measured in the interval 4050Å–4250Å. The magnitude cutoff is $\rm M_{R}=-18.8$ for all redshift bins. | Absorption systems | Emission systems ---|---|--- $<z>$ | D(4000) | $\rm EQW(H\delta)$ | S/N | D(4000) | S/N 0.29 | $1.67\pm 0.065$ | $-1.5\pm 0.2$ | 23 | $1.22\pm 0.02$ | 32 $0.43$ | $1.70\pm 0.06$ | $-1.5\pm 0.2$ | 22 | $1.22\pm 0.01$ | 52 $0.65$ | $1.60\pm 0.055$ | $-1.8\pm 0.2$ | 24 | $1.14\pm 0.01$ | 35 0.82 | $1.57\pm 0.06$ | $-2.4\pm 0.5$ | 18 | $1.07\pm 0.02$ | 28 0.99 | $1.43\pm 0.05$ | $-2.9\pm 0.3$ | 23 | $1.08\pm 0.02$ | 25 The most conspicuous spectral change with redshift is a decrease in flux redward of the G-band from $<z>=0.29$ and $<z>=0.43$ to higher $z$ coupled to an increase to the blue of [OII] from $<z>=0.65$ to $<z>=0.82$ and higher $z$ in emission-line galaxies. This systematic change of the continuum implies that the galaxy population varies as a function of redshift: more star forming galaxies at higher $z$ and more galaxies with old stars at lower $z$. This spectral change, which is known, will not be studied further in this paper except to quantify (in 5.3.1) the impact of LINER-like galaxies at $z=0.4-- 0.9$. In the following section we concentrate on absorption systems and low- ionization galaxies. ## 5 Absorption line systems The spectral resolution of the 300V grism allows us to detect [OII] emission down to $\rm EQW([OII])\sim 2-3~{}\AA$. We will call absorption-line galaxies those for which any mechanism of ionization is low enough to preclude [OII] detection at our detection level. Thus, our pure absorption-line sample comprises mostly E, E+A, and S0 galaxies with no on-going star formation, nuclear activity, or other mechanism of ionization. ### 5.1 Absorption systems as function of redshift The normalized and combined spectra of absorption line systems presented in Figure 8 (top) do not show any obvious change in their continuum and 4000Å break amplitude up to $z\approx 0.6$ (Table LABEL:D4000). There is a moderate decrease in the 4000Å break at $z\geq 0.65$ ranging from $5\%$ at $z\sim 0.65$ to $7\%$ at $z\sim 0.82$ and up to $15\%$ at $z\sim 1$, while the $\rm H\delta$ absorption line becomes stronger at $z\geq 0.65$ (Table LABEL:D4000), suggesting the presence of increasing numbers of A stars at higher redshifts. The indexes suggest that these galaxies had the bulk of their star formation at $z\geq 1$, while some of the systems at $z>0.8$ still had clearly detectable star formation about 1 Gyr ago. We have compared our spectral indexes at $z\sim 0.82$ with those measured by Tran et al. (2007) in the rich cluster MS 1054-03 at z = 0.83 using the same index definitions from Kauffmann et al. (2003). The average break amplitude and $\rm H\delta$ index of absorption systems in MS 1054-03 are respectively $\rm D(4000)(abs)=1.67\pm 0.00$ and $\rm EQW(H\delta)(abs)=-1.7\pm 0.0$ (Tran et al., 2007, Table 4). Our absorption systems at $z\sim 0.82$ appear to have younger stellar populations as indicated both by $\rm D(4000)$ and $\rm EQW(H\delta)$ (Table LABEL:D4000). Therefore our absorption systems contain A stars, but clearly less than composite field E+A galaxies at $<z>=0.6$ for which $\rm D(4000)(abs)=1.36\pm 0.02$ and $\rm EQW(H\delta)(abs)=-4.6\pm 0.2$ (quoted in Tran et al. (2007, Table 4) from data in Tran et al. (2004)). Consequently our average spectrum at $z\sim 0.82$ is intermediate between pure E and pure E + A. In fact, our SSP models (Table LABEL:poptable) indicate that absorption-line systems at $z\geq 0.65$ contain on average more than $50\%$ of stars younger than 2.5Gyr per galaxy, while those at $z\geq 0.8$ had significant star formation as recently as one Gyr ago (Table LABEL:poptable). Table 9: Stellar population properties of normalized average absorption (abs) spectra in each redshift bin. The magnitude cutoff is at $\rm M_{R}=-18.8$, except for the 10 faintest absorption systems at $z=0.29$ where we used all the observed objects. The fractions indicated in all SSP Tables are standard luminosity fractions at $\rm 4050\AA$, as in Cid-Fernandes et al. (2010, and references therein) $<z>$ | log(Age): | $<8$ | $8-8.7$ | $8.7-9$ | $9-9.4$ | $>9.4$ ---|---|---|---|---|---|--- $0.29$ | abs | 0.0% | 0.0% | 30.1% | 0.1% | 69.8% | abs (10 brightest) | 0 | 0 | 17.4 | 0 | 82.4 | abs (10 faintest) | 0 | 0 | 12.0 | 66.4 | 21.7 $0.43$ | abs | 0.0 | 0.7 | 11.7 | 6.9 | 80.7 $0.65$ | abs | 0.0 | 0.0 | 18.2 | 38.5 | 43.3 $0.82$ | abs | 0.0 | 0.0 | 86.8 | 3.3 | 9.8 $0.99$ | abs | 0.0 | 0.0 | 42.3 | 0.0 | 57.7 Post-starburst E+A galaxies are thought to be in a transition phase between a star-forming period and a passively evolving period. Being close to the phase of shutdown or quenching of star formation, they probably play an important role in the build-up of early-type systems (e.g. Wild et al, 2009; Yan et al., 2009). Studies of intermediate redshift clusters at $0.3\leq z\leq 0.6$ have found either a higher fraction of post-starburst galaxies in clusters than in the field (Dressler et al., 1999; Tran et al., 2003, 2004), or a similar fraction (Balogh et al., 1999). In fact, there is a strong variation in the E+A fraction between the SSDS low redshift survey at $z\sim 0.07-0.09$, and high $z$ surveys at $z\approx 0.5-1$ (VVDS, Wild et al, 2009), or $z\approx 0.7-0.9$ (DEEP2, Yan et al., 2009). In order to search for E+A galaxies in our sample we built template spectra by combining a pure E spectrum from our sample with various fractions of an A stellar template. We then compared our models with absorption-line systems in the range $\rm 1.2\leq D(4000)\leq 1.5$ assuming, by definition, that E+A galaxies contain at least $\rm 25\%$ A stars. Using this (standard) definition we searched our sample at $0.35\leq z\leq 1$ and found only 6 bona-fide E+A galaxies. In fact all the objects were found at $0.68\leq z\leq 1$, which makes our small number consistent with the VVDS and the DDEP2 surveys within a factor of 2. The median of the normalized spectra of these 6 (as far as we can judge from our images) elliptical galaxies is presented in Figure 9 (a). We were surprised to find no E+As at $z\sim 0.4$, but we did find 4 objects with early-type morphology and very small $\rm<EQW([OII])>\approx 3.5\AA$, which probably would have been classified as E+As on lower resolution spectra. The average spectrum of these 4 objects is shown in Figure 9 (b). Their $\rm R\sim 22$ magnitudes place them at the faint end of absorption line systems at the corresponding redshifts. Figure 9: (Top) Average spectrum of 6 E+A galaxies at $0.68\leq z\leq 1$. The Balmer series $\rm H\delta$, $\rm H+H\epsilon$, H8, H9, H10, and H11 is very promintent and $\rm<D(4000)>=1.40$. (Botton) Average spectrum of 4 galaxies in the intermediate redshift range $0.4\leq z\leq 0.5$ with E/S0 morphological type, showing a poststarburst E+A spectrum with still some star formation. $\rm<EQW([OII])>\approx 3.5\AA$ and $\rm<D(4000)>=1.41$. Probably this spectrum would have been classified as E+A at lower resolution. ### 5.2 Absorption-line galaxies as function of luminosity at $z=0.29$ Having tested bright absorption galaxies at various redshifts (with cut-off at $\rm M_{R}=-18.8$), we now turn to faint absorption galaxies in the cluster at $<z>=0.29$ by combining the spectra of the 10 faintest galaxies without emission lines. Their average R-band magnitude is $\rm R=22$, which at a distance modulus of 40.18 corresponds to $\rm M_{R}=-18.2$, and the faintest object has $\rm M_{R}=-17.44$. Their mean indexes, $\rm D(4000)=1.55\pm 0.01$; $\rm H\delta=-2.27\pm 0.04$, measured on the spectrum shown in Figure 10, are consistent with a younger age than absorption-line galaxies with $\rm M_{R}\leq-18.8$ (Table LABEL:D4000) in the same cluster. This is in agreement with the well known evidence that the stellar populations in absorption systems tend to be younger in low mass galaxies than in the more massive ones (e.g. Renzini, 2006). The index values are in fact very close to those of our absorption systems at z = 0.8 (Table LABEL:D4000) which, by selection effects, are bright (Table LABEL:magnitudes and Figure 7). The SSP models indicate that on average about 80% of the stars in the 10 faintest galaxies are younger than 2.5 Gyr (Table LABEL:poptable), i.e. were born at $z<1$. In comparison, 80% of the stars contributing to the spectrum of the brightest absorption galaxies in the cluster are older than 2.5 Gyr (Table LABEL:poptable). To illustrate the spectral differences between bright and faint systems at $z=0.29$, and the striking similarity between the spectra of faint galaxies at $z=0.29$ and those of bright galaxies at $z=0.8$, we have plotted in Figure 10 the average spectra of the 10 brightest and the 10 faintest absorption systems at $z=0.29$, and the average spectrum of absorption galaxies at $z=0.82$. The effect of downsizing, (in the present case the so-called ’archeological dowsizing’) where star formation shifts from high mass galaxies at high redshifts, to low mass galaxies at low redshifts is clearly exemplified in this figure. Figure 10: Normalized spectra of the 10 brightest (in red) and the 10 faintest (in blue) absorption-line galaxies in the cluster at $z=0.29$, and the full sample of absorption-line systems at $z=0.82$ (in cyan). At a redshift of $z\sim 0.8$ (i.e. $\sim 4$ Gy earlier), the red-sequence of our unrelaxed (merging central system; elongated intra-cluster light and galaxy distribution) cluster at z=$0.29$ was already in place, but was truncated at brighter magnitudes because the faint absorption-line galaxies were still copiously forming stars. This seems consistent with the observation that some clusters at $z\simeq 1$ have red sequences truncated at faint limits (Kodama et al., 2004; Koyama et al., 2007), and supports the picture of an environmental dependence of red-sequence truncation presented by Tanaka et al. (2005). This is also in agreement with scenarios where the final assembling of the red-sequence can be observed well below $z=1$ (Faber et al., 2007). As discussed above, the strict definition of E+A galaxy requires a mix of an E-type spectrum with at least 25% A stars and no traces of star formation, which in our sample implies no emission lines with equivalent widths larger than 2-3Å. With this definition our $z=0.29$ cluster contains only one E+A galaxy while the dense layers at $z\sim 0.4$, where the red-sequence is already in place (layer in Figure 7), contains none. However, both in the cluster and in the intermediate redshift layers we find plenty of galaxies with early type morphologies, A stars, and very weak emission lines. In the next section we present a closer look at these low-ionization emission line galaxies. ### 5.3 Galaxy evolution and low-ionization emission-line galaxies (LINERs). In an extensive work based on the SDSS survey, Yan et al. (2006) determined the extent to which [OII] emission produced by mechanisms other than recent star formation introduces biases in galaxy evolution studies based upon [OII] only. They showed that the $\rm[OII]/H\beta$ ratio separates LINERs from star- forming galaxies, while $\rm[OIII]/[OII]$ and $\rm[OIII]/H\beta$ separate Seyferts from LINERs and star-forming galaxies. Using the classification scheme of Yan et al. (2006) we divided our spectra in 3 main classes: LINERs, with clearly detected [OII], but no ($3\sigma$) detection of [OIII] and $\rm H\beta$ in emission after subtracting an E+A profile; Seyferts, with $\rm[OIII]/H\beta\geq 3$; and star-forming galaxies, which are the objects with clearly detected [OII] that are neither Seyferts nor LINERs after subtracting an E+A profile. Typical spectra of low-ionization objects, star- forming galaxies and Seyferts are shown in Figure 11. Figure 11: Typical average spectra of low-ionization objects, star-forming galaxies and Seyferts from the sample in the $<z>=0.415$ layer. #### 5.3.1 The impact of LINERs in our previous results on red emission-line galaxies Because the spectral coverage in a rather large fraction of our objects at $<z>=0.68$ and higher is truncated below $\rm 5000\AA$ in the rest frame, we applied our classification scheme only to objects in the range $0.29\leq z\leq 0.65$. To extract $\rm H\beta$ in emission we built a series of E+A models, combining an observed E spectrum with different fractions of an A stellar template, ranging form $0.05\%$ to $80\%$ of the total luminosity. To determine the best-fit model we minimized the continuum slope of the difference between the spectrum and the E+A model. Thus, in the range $0.35-0.55$ our sample contains 23% LINERs, 51% star-forming galaxies, 8% Seyferts, and 14% uncertain types. The layer at $z=0.63$ has 18% LINERs, 50% star-forming galaxies, 7% Seyferts, 13% uncertain types and 11% of truncated spectra. Altogether, the fraction of LINERs among emission-line galaxies up to $z=0.65$ in our pencil beam is $\rm\approx 22\%$. With an average $\rm<D(4000)>=1.39\pm 0.18$. LINERs at $z\leq 0.65$ have a potentially significant impact on the conclusions of Giraud et al. (2010) about the evolution of red emission-line galaxies. To quantify this impact, we have subtracted all LINERs from the sample of emission-line spectra in the $z=0.43$ bin, determined the new blue-to-red partition (as in Giraud et al. 2010; section 5.1), and computed a new average spectrum for the red galaxies. This (also cleaned of rare red Seyferts) is shown in Figure 12 where it is compared with the mean red spectrum at $z=0.9$. We find that the differences in continuum slope and D(4000) between $<z>=0.43$ and $<z>=0.9$ is reduced by a factor of 2/3. The main difference between red galaxies with LINERs and those without is the presence of young stellar population. Figure 12: Average spectra of red emission-line galaxies after subtracting early-type LINERs and galaxies with diluted star formation (and rare Seyferts) from the sample in the $<z>=0.43$ bin and recalculating the median blue-to-red partition, and of the red half of emission-line galaxies at $<z>=0.9$. The spectra at $<z>=0.9$ were not classified because $\rm H\beta$ and [OIII] are missing in most cases. #### 5.3.2 Early-type LINERs The fraction of nearby early-type galaxies hosting bona-fide (i.e. nuclear) LINERs in the Palomar survey (Filippenko & Sargent, 1985; Ho et al. 1997a, ) was found to be $\sim 30\%$ (Ho et al. 1997b, ), but LINER-like emission line ratios are also observed in extended regions (Phillips et al., 1986; Goudfrooij et al., 1994; Zeilinger et al., 1996; Sarzi et al., 2006, and references therein). A similar fraction of LINER-like ratios is found in the SDSS at $0.05\leq z\leq 0.1$ in color-selected red galaxies (Yan et al., 2006). Because it is very difficult to disentangle early-type LINERs from spirals with extended and diluted star formation by using only [OII] and $\rm H\beta$, we make use of morphology to distinguish compact objects with low ellipticity and profiles consistent with early type galaxies, from other morphologies: apparent disks, high ellipticity, and irregular or distorted morphologies. Images of early-type galaxies with low ionization spectra are shown in Figure 13. Figure 13: Examples of early-type galaxies having low-ionization spectra, and indicated redshifts. Our visual early-type morphologies are the same as ZEST type T=1 (Scarlata et al., 2007, Figure 4 (b), (c), (d)). In the $<z>=0.43$ bin we find that 92% of the galaxies classified as star-forming objects have morphologies inconsistent with early-types. At $<z>=0.43$ and in the $z=0.63$ layer, about half of the LINERs have compact morphology while the other half are mainly bulge-dominated disk galaxies, or “early disks” of ZEST type T=2.0 ((Bundy et al, 2009, Figure 4)). At $z=0.29$ all LINERs have disks. The spectra of galaxies with apparent disks have an extended [OII] emission suggesting that they do have extended star-formation. Average spectra of 11 early-type galaxies (E) and 10 later types (hereafter S) resulting from our morphological classification are shown in Figure 14. Figure 14: Median spectra of 11 early-type galaxies and of 10 galaxies with later type morphology (S) with low ionization at $z=0.4-0.5$. The absence of $\rm H\beta$ in the S sample suggests that $\rm H\beta$ in emission resulting from star formation is diluted in $\rm H\beta$ in absorption from A and older stars. The closest spectral comparison in the atlas of galaxy spectra (Kennicutt, 1992) is with an Sb galaxy. The rather strong $\rm H\beta$ in absorption in early-types (E) combined with [OII] suggests either a low fraction of young stars or a mechanism of photo- ionization other than young stars as in (Fillipenko, 2003; Ho, 2004, and references therein). In fact, the recent work by the SEAGAL collaboration (Cid Fernandes et al., 2010, and references therein) has shown that the majority of galaxies with LINER spectra in the SDSS can be explained as retired galaxies, that is, galaxies that have stopped forming stars but still contain appreciable amounts of gas that is being photoionized by intermediate-aged post-AGB stars. In fact, the SEAGAL models with no young stars, but with significant populations of 100Myr-1Gyr stars resemble remarkably well our average LINER spectrum shown in Figure 11. We calculated population synthesis models for our average spectra of LINERs with early-type and late-type morphologies. The results, shown in Figure 15 and Table LABEL:earlylinerpop, indicate that both early-type and late-type LINERs have significant populations of young and intermediate age stars, but late-type (S) LINERs have much younger populations. In fact, the residuals of the S-LINER fit show $\rm H\beta$ in emission stronger than [OIII], consistent with the idea that they are red spirals with diluted star formation. Table 10: Stellar population properties of an average of LINERs with early-type morphology and with morphology of later types Type | log Age | $\rm\chi^{2}$ ---|---|--- | $<8$ | $8-8.7$ | $8.7-9$ | $9-9.4$ | $>9.4$ | Early | 18.2 | 0 | 57.6 | 0 | 24.2 | 1.4 Later-type | 32.7 | 0 | 53.4 | 0 | 13.9 | 1.3 Figure 15: Spectral fitting with SSP models for an average spectrum of LINERs with early-type morphology (left), and with morphology of later type (right). Thus, our results are consistent with the interpretation that most early-type LINERs at intermediate redshifts are in fact post-starburst galaxies, as postulated by the SEAGAL collaboration for lower redshift objects. These results indicate that LINERs and E+As depict the quenching phase in the evolution of galaxies massive enough to retain significant amounts of gas after the stellar-wind and supernova phases of the most massive stars. ### 5.4 The red limit of emission-line galaxies At each $z$ we have selected galaxies with the reddest continuum (the reddest quartile at each redshift bin) to construct the combined spectra of the red envelope or red limit of emission line galaxies. Since we are working with small numbers of galaxies, typically 5-10, it was necessary to combine the samples at $z=0.82$ and $z=0.99$ to improve statistics. Nevertheless, because our red emission-line galaxies are rather luminous, the combined spectra still have high continuum S/N ratios (Table LABEL:redest). The common parts of the red envelopes of spectra at $<z>=0.29$, $<z>=0.43$, $<z>=0.65$ are similar, while the red limit at $<z>=0.9$ has noticeably stronger UV continuum. The spectra in different bins are shown in Figure 7 of Giraud et al. (2010). Both the continuum and the indexes of the red limit at $z\leq 0.65$ (i.e. $\rm D(4000)\sim 1.35-1.45;EQW([OII])\sim 4-8$), are typical of nearby spirals with prominent bulges and low star formation (Kennicutt, 1992; Kinney et al., 1996; Balogh et al., 1999), or early-type LINERs. Up to $z\sim 0.7$ the populations of red spirals and early-types can be well separated by their morphology. The higher UV continuum and lower $\rm D(4000)$ of the limit spectrum at $<z>=0.9$ indicate that such red objects become rare at $z\geq 0.68$ in our sample. Absorption systems have (by definition) already lost enough gas to suppress any detected star formation by the time they first appear in our sample at $z\simeq 1$. At $z=0.8-1$ emission-line galaxies in our sample are found to have very strong star formation, which declines at lower $z$, the reddest quartile being bluer than at lower $z$. Therefore the evolutionary paths of bright absorption and emission systems might have been more separated at $z\simeq 1$ than at lower redshift suggesting two different physical processes of different time scales. In one we have early-type LINERs and E+A galaxies that define the ”entrance gate” to the red sequence of passively evolving galaxies. In the other we have red spirals with diluted star formation, that are in a final phase of smooth star formation, possibly of a “main sequence” (Noeske et al., 2007). In Section 5.3 we found a large fraction of LINERs in layers at intermediate $z$. More precisely, in the volume-limited range $0.35\leq z\leq 0.65$, we find, gathering the counts of Section 5.3, that LINERs are $\rm 23\%$ of all early-type galaxies with measured redshifts. Table 11: Equivalent width of [OII], 4000 Å break amplitude, $\rm H\delta$ index, and the G-step of the red envelope of emission-line galaxies. The continuum S/N ratios are given in the last column. $<z>$ | EQW([OII]) | D(4000) | $\rm EQW(H\delta)$ | G step | $\rm S/N$ ---|---|---|---|---|--- 0.29 | $4.2\pm 0.3$ | $1.35\pm 0.07$ | $-1.8\pm 0.3$ | $1.224\pm 0.017$ | $\rm 19$ $0.43$ | $8.5\pm 0.2$ | $1.39\pm 0.06$ | $-2.5\pm 0.2$ | $1.268\pm 0.014$ | $\rm 24$ $0.65$ | $8.3\pm 0.2$ | $1.44\pm 0.06$ | $-3.0\pm 0.2$ | $1.273\pm 0.012$ | $\rm 29$ $0.9$ | $9.0\pm 0.3$ | $1.30\pm 0.07$ | $-3.9\pm 0.2$ | - | $\rm 18$ ## 6 Summary and Conclusions We have presented a catalogue of galaxy spectra in a pencil beam survey of $\sim 10.75^{\prime}\times 7.5^{\prime}$, and used these data to make an analysis of the spectral energy distribution of a magnitude limited sample up to $z\sim 1$, concentrating on absorption and low ionization emission-line systems. The redshift range has been divided in bins centered on the structures that were detected in the (RA, Dec, $z$) pseudo-volume, and corresponding to cosmic time slices of $\rm\sim 1Gyr$. Our sample is reasonably complete for galaxies brighter than $\rm M_{R}=-18.8$ up to $z\approx 0.5$; at $z\geq 0.75$ the cutoff is at -20.5. From this analysis we reach the following conclusions: 1. 1. We confirm in our pencil-beam sample the well known result Hamilton (1985) that absorption-line galaxies do not show significant variations in their continuum energy distributions up to $z=0.6$, and a moderate decrease of the 4000 Å break amplitude of 5% at $z\sim 0.65$, 7% at $z\sim 0.82$, and up to 15% at $z\sim 1$. Using stellar population synthesis models we find that absorption-line galaxies at $z\geq 0.65$ show more than 50% of stars younger than 2.5Gyrs, while those at $z\geq 0.8$ had star formation as recently as 1Gyr ago. This suggests that the red sequence is still in a buildup phase at $z\leq 1$. The faint absorption-line galaxies in our dynamically young cluster at $z=0.29$ have indexes similar to those of bright absorption-line systems at $z=0.8$, suggesting that faint galaxies without emission lines tend to be younger than more massive galaxies with similar spectra. Our population synthesis models indicate that about 50% of the stars contributing to the luminosity of faint absorption-line galaxies at z = 0.29 were formed at $z<1$. This is consistent with cases of truncated red sequences observed in some high-$z$ clusters and suggests that clusters with truncated red-sequences may be dynamically young. 2. 2. Combining simple emission-line diagnostics with galaxy morphology we identify a significant population of early-type LINERs at $0.35\leq z\leq 0.65$. In that redshift range early-type LINERs constitute about 23% of all early-types galaxies, a much larger fraction than E+A post-starburst galaxies. However, our population synthesis models show that early-type LINERs contain substantial populations of intermediate age stars that can easily explain the observed line emission, as recently proposed by Cid-Fernandes et al. (2010). This led us to conclude that most LINERs in our sample are in fact post- starburst galaxies. 3. 3. The red limit in the spectral energy distribution of emission-line galaxies at $z\leq 0.65$ is typical of bulge-dominated spirals with moderate star formation, and of early-type LINERs. Thus, early-type LINERs and E+As define the “entrance gate” of the red sequence of passively evolving galaxies, while bulge-dominated spirals have diluted star formation. ###### Acknowledgements. EG thanks the hospitality of ESO and Universidad Catolica in Santiago during the initial phase of this work. JME thanks the hospitality of Nanjing University during the initial phase of this research. QGU would like to acknowledge the financial support from the China Scholarship Council (CSC), the National Natural Science Foundation of China under grants 10878010, 10221001, and 10633040, and the National Basic Research Program (973 program No. 2007CB815405). HQU thanks partial support from FONDAP “Centro de Astrofísica”. PZE acknowledge a studentship from CONICYT. We thank S. di Serego Alighieri for reading a preliminary version of the manuscript and for his suggestions, and R. Cid-Fernandes for fruitful discussions. ## References * Adelman-McCarthy et al (2006) Adelman-McCarthy, J.K., et al, 2006, ApJS 162, 38 * Baldwin et al. (1981) Baldwin, J.A., Phillips, M.M., & Terlevitch, R., 1981, PASP 93, 5 * Balogh et al. (1999) Balogh, M. L., Morris, S. L., Yee, H. K. C., Carlberg, R. G., & Ellingson, E., 1999, ApJ 527, 54 * Baugh et al. (1996) Baugh, C.M.,Cole, S., & Frenk, C.S., 1996, MNRAS.283, 1361 * Bell et al. (2005) Bell, E. F., et al, 2005, ApJ 625, 23 * Blake et al. (2004) Blake, C., et al., 2004, MNRAS 355, 713 * Bundy et al (2009) Bundy, K., et al., 2009, astro-ph 0912.1077v1 * Bruzual & Charlot (2003) Bruzual G., & Charlot S., 2003, MNRAS, 344, 1000 * Cid Fernandes et al. (2004) Cid Fernandes R. , et al., 2004, MNRAS, 355, 273 * Cid Fernandes et al. (2005) Cid Fernandes R., Mateus A., Sodre L., Stasinska G., Gomes J., 2005, MNRAS 358, 363 * Cid Fernandes et al. (2010) Cid Fernandes R., et al., 2010, MNRAS 403, 1036 * Cooper et al. (2006) Cooper, M. C. et al, 2006, MNRAS 370, 198 * Cucciati et al. (2006) Cucciati, O., et al (the VVDS Team), 2006, A&A 458, 39 * Dickinson et al. (2003) Dickinson, M., Papovich, C., Fergusson, H.C., Budavari, T., 2003, ApJ 587, 25 * Dressler et al. (1999) Dressler, A., et al, 1999, ApJS 122, 51 * Dressler et al. (1997) Dressler, A., et al, 1997, ApJ 490, 577 * Dressler & Gunn (1983) Dressler A., & Gunn J.E., 1983 ApJ 270 7 * Dressler (1980) Dressler, A., 1980, ApJ 236, 351 * Faber et al. (2007) Faber, S. et al, 2007, ApJ 665, 265 * Faure et al. (2007) Faure, C., Giraud, E., Melnick, J., Quintana, H., Selman, F., & Wambganss, J., 2007, A&A, 463, 833 * Filippenko & Sargent (1985) Filippenko, A.V., & Sargent, W.L., 1985, ApJS 57, 503 * Fillipenko (2003) Filippenko, A.V., 2003 in ASP Conf. Ser. 290 Active Galactic Nuclei: From Central Engine to Host Galaxy, 369 * (23) FORS1+2 User Manual VLT-MAN-ESO-13100-1543 Issue 4, 2005 * FIMS (2006) FORS1+2 FIMS Manual VLT-MAN-ESO-13100-2308 Issue 78, 2006 * Franzetti et al. (2007) Franzetti, P., et al (the VVDS Team), 2007, A&A 465, 711 * Garilli et al. (2008) Garilli, B., et al (the VVDS Team), 2008, astro-ph 0804.4568 * Giraud et al. (2010) Giraud E., Melnick J., Gu Q.S., et al., 2010, in press * Goto (2007) Goto, T., 2007, MNRAS 377, 1222 * Goudfrooij et al. (1994) Goudfrooij, P., et al., 1994, A & AS 105, 341 * Gu et al. (2006) Gu Q. et al. 2006, MNRAS, 366, 480 * Hamilton (1985) Hamilton, D., 1985 ApJ, 297, 371 * Hammer et al. (2005) Hammer, F., et al, 2005, A&A 430, 115 * (33) Ho, L.C.W., et al., 1997a, ApJS 112, 315 * (34) Ho, L.C.W., et al., 1997b, ApJ 487, 568 * Ho (2004) Ho, L.C.W,. 2004, in Coevolution of Black Holes and Galaxies, 293 * Huang & Gu (2009) Huang, S., & Gu Q., 2009, MNRAS 398, 165 * Kauffmann et al. (2003) Kauffmann, G., et al, 2003, MNRAS 341, 33 * Kauffmann et al. (2003) Kauffmann, G., et al, 2003, MNRAS 346, 1055 * Kennicutt (1992) Kennicutt, R.C.., 1992, ApJS 79, 255 * Kinney et al. (1996) Kinney, A.L., Calzetti, D., Bohlin, R.C., McQuade, K., Storchi-Bergmann, T., & Schmitt, H.R., 1996, ApJ 467, 38 * Kodama et al. (2004) Kodama, T., et al, 2004, MNRAS 350, 1005 * Koyama et al. (2007) Koyama, Y., Kodama, T., Tanaka, M, Shimasaku, K., & Okamura, S., 2007, MNRAS 382, 1719 * Lin et al. (2008) Lin, L., et al., 2008, ApJ 681, 232 * Melnick & Sargent (1977) Melnick, J., & Sargent, L. W. L., 1977, ApJ 215, 401 * Melnick et al. (1999) Melnick, J., Selman, F., & Quintana, H., 1999, PASP 111, 1444 * Neinstein et al. (2006) Neinstein, E., van den Bosch, F.C., & Dekel, A., 2006, MNRAS 372, 933 * Noeske et al. (2007) Noeske, K.G., et al., 2007, ApJ 660, L43 * Norton et al. (2001) Norton, S.A., Gebhardt, K., Zabludoff, A.I., & Zaritsky, D., 2001, ApJ 557, 150 * Pei & Fall (1995) Pei, Y. C., & Fall, S. M., 1995, ApJ 454, 69 * Phillips et al. (1986) Phillips, M.M., et al., 1986, AJ 91, 1062 * (51) p2pp, 2006: http://www.eso.org/observing/p2pp/ * Postman et al. (2005) Postman, M., et al, 2005, ApJ 623, 721 * Renzini (2006) Renzini, A., 2006, ARA&A 44, 141 * Renzini (2007) Renzini, A., 2007, ASPC 390, 309 * Sarzi et al. (2006) Sarzi, M., et al., 2006, MNRAS 366, 1151 * Scarlata et al. (2007) Scarlata, C., et al., 2007, ApJS 172, 406 * Scoville et al. (2007) Scoville, N., et al, 2007a, ApJS 172, 1 * Scoville et al. (2007) Scoville, N., et al, 2007b, ApJS 172, 150 * Smith et al. (2005) Smith, G. P., Treu, T., Ellis, R. S., Moran, S. M., & Dressler, A., 2005, ApJ 620, 78 * Tanaka et al. (2005) Tanaka, M., et al, 2005, MNRAS 362, 268 * Toledo et al. (2010) Toledo, I., et al. 2010 in preparation * (62) Tonry, J., & Davis, M., 1979, AJ 84, 1511. * Tran et al. (2007) Tran, K.-V. H., et al, 2007, ApJ 661, 750 * Tran et al. (2004) Tran, K.-V. H., et al, 2004, ApJ 609, 683 * Tran et al. (2003) Tran, K.-V. H,. et al, 2003, ApJ 599, 865 * Veilleux & Osterbrock (1987) Veilleux, S., & Osterbrock, D.E., 1987, ApJS 63, 295 * Weiner et al. (2005) Weiner, B.J., et al, 2005, ApJ 620, 595 * Weiner et al. (2006) Weiner, B.J., et al, 2006, ApJ 653, 1027 * Wild et al (2009) Wild V. et al. 2009 MNRAS 395 144 * Yan et al. (2009) Yan, R., et al., 2009, MNRAS 398, 735 * Yan et al. (2006) Yan, R., Newman, J.A., Faber, S., Konidaris, N., Koo, D., & Davis, M., 2006 ApJ 648 281 * Yang et al. (2008) Yang Y., Zabludoff, A.I., Zaritsky, D., & Mihos, J.C., 2008, ApJ 688, 945 * Zeilinger et al. (1996) Zeilinger, W.W., et al., 1996, A & AS 120, 257
# Image Inpainting Models are Effective Tools for Instruction-guided Image Editing Solution for GenAI Media Generation Challenge Workshop @ CVPR Xuan Ju1,2, Junhao Zhuang1, Zhaoyang Zhang1∗, Yuxuan Bian1,2, Qiang Xu2, Ying Shan1 1ARC Lab, Tencent PCG, 2The Chinese University of Hong Kong ∗Project Lead https://github.com/TencentARC/BrushNet/tree/main/InstructionGuidedEditing ###### Abstract This is the technique report for the winning solution of the CVPR2024 GenAI Media Generation Challenge Workshop’s Instruction-guided Image Editing track. Instruction-guided image editing has been largely studied in recent years. The most advanced methods, such as SmartEdit and MGIE, usually combine large language models with diffusion models through joint training, where the former provides text understanding ability, and the latter provides image generation ability. However, in our experiments, we find that simply connecting large language models and image generation models through intermediary guidance such as masks instead of joint fine-tuning leads to a better editing performance and success rate. We use a 4-step process IIIE (Inpainting-based Instruction- guided Image Editing): editing category classification, main editing object identification, editing mask acquisition, and image inpainting. Results show that through proper combinations of language models and image inpainting models, our pipeline can reach a high success rate with satisfying visual quality. ## 1 Introduction With the rapid development of diffusion models, the field of text-guided image generation [7, 19, 13, 15] has seen unprecedented progress in creating images with superior quality [18], diversity [3], and adherence to text guidance [14]. However, in image editing tasks, which provide a source image and an editing instruction as input and expect a target image as output, we do not observe such success. This implies that the language understanding ability and image generation ability are not fully explored in editing tasks. In the aim of applying image generation capabilities to image editing, previous methods have attempted two strategies: (1) collecting paired “source image-instruction-target image” data and fine-tuning diffusion models for editing tasks (e.g., InstructPix2Pix [1] and InstructDiffusion [5]), and (2) jointly fine-tuning Large Language Models (LLMs) and diffusion models to endow the diffusion models with a stronger understanding of text (e.g., SmartEdit [8] and MGIE [4]). For the former strategy, due to the difficulty of collecting paired manually edited data, the training data are usually generated by LLMs and inference-based image editing methods (e.g., Prompt-to- Prompt [6] and Masactrl [2]). Due to the low success rate and unstable editing quality of these inference-based image editing methods [11], the collected dataset is usually noisy and unreliable, which lead to an unsatisfying performance of the trained image editing model. For the latter strategy, a joint training of the LLMs and diffusion model usually make it hard to fully use the text understanding capability of LLMs. Although SmartEdit and MGIE have achieved better performance than previous solutions, we find that they still do not fully optimize the potential of LLMs and diffusion. This technique report provides a different solution, connecting large language models and image generation models simply through intermediary guidance (e.g., edit objects and masks), and is the winning solution for The GenAI Media Generation Challenge (MAGIC). By detracting two models apart in an agent architecture, we find it easier to fully leverage the capabilities of both. Specifically, we use a 4-step process IIIE (Inpainting-based Instruction- guided Image Editing): (1) editing category classification, (2) main editing object identification, (3) editing mask acquisition, and (4) image inpainting. Step (1)-(3) use LLMs and detection model to determine the editing type, edit object, edit masks, and target prompt. Then, step (4) perform image editing in the way of image inpainting, which fully use generation ability. In this way, step (1)-(3) use LLMs to extract information in instruction and summarize them to intermediary guidance that can be used by diffusion models. Visual results show that the proposed IIIE substantially surpass previous instruction-guided image editing methods and other solutions in MAGIC considering visual quality and instruction faithfulness. ## 2 MAGIC The MAGIC hosts two challenge tracks: (1) text-to-image generation and (2) text-guided image editing. This technique report presents the solution for the second track, text-guided image editing. We list the instructions here: Guidelines For text-guided image editing, we test the capacity of the model to change a given image’s contents based on some text instructions. The specific type of instructions that we test for are the following: • Addition: Adding new objects within the images. • Remove: Removing objects • Local: Replace local parts of an object and later the object’s attributes, i.e., make it smile • Global: Edit the entire image, i.e., let’s see it in winter • Inpaint: Alter an object’s visual appearance without affecting its structure • Background: Change the scene’s background Evaluation Protocol We leverage both human and automated evaluations. In human-based evaluations, we use human annotators. We mainly evaluate the following aspects: • Edit Faithfulness - whether the edited image follows the editing instruction • Content Preservation - whether the edited image preserves the regions of the original image that should not be changed • Overall Instruction Following - considering both edit faithfulness and content preservation, whether the edited image is artifact-free, keeping the core visual features of the original image, etc On automatic evaluation, similar to the text to image track, we will leverage existing methods that have developed automatic metrics to help in assessing the outputs of the image based on the prompt and instruction. To determine winners, we use automatic evaluation to help prune the total number of entries to 10 finalists. At 10, we would use human annotation and evaluation to determine the final winners. More information about the detailed workshop information can be found at the official website page: https://gamgc.github.io/. Figure 1: The pipeline of our proposed 4-step image editing process ITIE. ## 3 Method Previous instruction-guided image editing methods include two categories: (1) diffusion model finetuning [1, 5], and (2) jointly LLMs and diffusion model [8, 4]. These methods either not include LLMs in the model, or use joint fine- tuning of LLMs and diffusion models. Both of these strategies can not fully unleash the capabilities of LLMs, leading to a weak understanding of instructions. Consequently, these methods show low success rates and unsatisfying results. Contrary to these methods, we find that a simple tool- based combination of LLMs and diffusion models can lead to a much better visual results, coming from a full utilization of the language understanding ability of LLMs. In this competition, we propose a 4-step process for instruction-based image editing, IIIE (Inpainting-based Instruction-guided Image Editing), as shown in Fig. 1. Firstly, we use GPT4-o to categorize the current editing instructions into one of the editing categories: Local Edit, Background Edit, Global Edit, Addition, and Remove. Local Edit includes local changes such as replacing object (e.g., change a cat to a dog) and change the attribute of an object (e.g., change color or texture). For example, Fig. LABEL:fig:teaser row 1 show two examples of Local Edit. Background Edit means changing the background and remain the main object unchanged. For example, Fig. LABEL:fig:teaser row 2 column 1 and row 4 column 2 show two examples of Background Edit. Global Edit transfer the overall style of an image. For example, Fig. LABEL:fig:teaser row 4 column 1 show one example of Global Edit. Addition and Remove separately add and remove object from an image. For example, Fig. LABEL:fig:teaser row 3 show two examples of adding object, and row 2 column 2 shows one example of removing object. We classify editing category to these 4 categories since they can cover most editing instruction types, and requires different operations in editing. Then, in step 2, we find the main editing object by making further conversations with GPT4-o based on the editing category. For example, editing instruction “Make the horse into a unicorn” has a main object of “horse”. For editing categories of global edit and addition that do not contain an editing object, we leave the editing object blank. After that, we use Grounded-SAM [17] combined with GPT4-o to obtain the editing mask in step 3. Specifically, we generate image background as mask for Background Edit. For Local Edit and Removing, we use the main object of step 2 as the input of Grounded-SAM to locate the editing mask. For addition editing category, we use the visual ability of GPT4-o and get a possible location of adding the object. For Global editing, the mask is default as the whole image. Finally, in step 4, we use image generation model to perform image editing based on a target prompt generated by GPT4-o. For Background Edit, Local Edit, Remove, and Addition, we use image inpainting model BrushNet [10] combined with PowerPaint [21] to inpaint the masked region based on the target prompt. For Global Edit, we use InfEdit [20] to change the global style. We find that a direct utilization of the diffusion model by giving conditions of mask and text prompt can fully leverage the capabilities of these models, thus leads to high-quality generation results. ## 4 Experiment Results Method | Rank | Edit | Content | Overall Instruction ---|---|---|---|--- Faithfulness | Preservation | Following IIIE | 1 | 0.51 | 0.78 | 0.80 LEdits++ | 2 | 0.46 | 0.64 | 0.74 Tasvir | 3 | 0.40 | 0.49 | 0.62 Table 1: Comparison of IIIE and the other two winning solutions. The score is calculated with an average of 1.2k images in MAGIC benchmark. A bigger score means a better user preference. Bold stands for the best results. Visualization comparison in Fig. LABEL:fig:teaser show a higher success rate and visual quality of IIIE compared to previous methods. In the MAGIC workshop, a benchmark of 1.2k images is used for evaluation and 3 winning solutions are announced. The benchmark measure the quality of different editing methods using user study on instruct following, edit fidelity, and content preservation. 3 raters are involved in each job and the majority vote is taken as the final results. Each image is rated with 1 or 0 for each metric. Results of the three winning solution of this competition is shown in Tab. 1. IIIE show better score on all three metrics compared to the other two winning solutions. We have made our code and edited results publicly available in the hope of our findings can offer some insights for relevant field [12, 9, 16]. In conclusion, in this technique report presents the winning solution for the CVPR2024 GenAI Media Generation Challenge’s Instruction-guided Image Editing track. We show that without cumbersome fine-tuning or training, a simple combination of LLMs and text-to-image diffusion model can lead to a image editing results with better performance and higher success rates than previous methods. ## References * [1] Tim Brooks, Aleksander Holynski, and Alexei A Efros. Instructpix2pix: Learning to follow image editing instructions. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 18392–18402, 2023. * [2] Mingdeng Cao, Xintao Wang, Zhongang Qi, Ying Shan, Xiaohu Qie, and Yinqiang Zheng. Masactrl: Tuning-free mutual self-attention control for consistent image synthesis and editing. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 22560–22570, 2023. * [3] Xiaoliang Dai, Ji Hou, Chih-Yao Ma, Sam Tsai, Jialiang Wang, Rui Wang, Peizhao Zhang, Simon Vandenhende, Xiaofang Wang, Abhimanyu Dubey, et al. Emu: Enhancing image generation models using photogenic needles in a haystack. arXiv preprint arXiv:2309.15807, 2023. * [4] Tsu-Jui Fu, Wenze Hu, Xianzhi Du, William Yang Wang, Yinfei Yang, and Zhe Gan. Guiding instruction-based image editing via multimodal large language models. arXiv preprint arXiv:2309.17102, 2023. * [5] Zigang Geng, Binxin Yang, Tiankai Hang, Chen Li, Shuyang Gu, Ting Zhang, Jianmin Bao, Zheng Zhang, Houqiang Li, Han Hu, et al. Instructdiffusion: A generalist modeling interface for vision tasks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 12709–12720, 2024. * [6] Amir Hertz, Ron Mokady, Jay Tenenbaum, Kfir Aberman, Yael Pritch, and Daniel Cohen-Or. Prompt-to-prompt image editing with cross attention control. arXiv preprint arXiv:2208.01626, 2022. * [7] Jonathan Ho, Ajay Jain, and Pieter Abbeel. Denoising diffusion probabilistic models. Advances in Neural Information Processing Systems (NIPS), 33:6840–6851, 2020. * [8] Yuzhou Huang, Liangbin Xie, Xintao Wang, Ziyang Yuan, Xiaodong Cun, Yixiao Ge, Jiantao Zhou, Chao Dong, Rui Huang, Ruimao Zhang, et al. Smartedit: Exploring complex instruction-based image editing with multimodal large language models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 8362–8371, 2024. * [9] Xuan Ju, Yiming Gao, Zhaoyang Zhang, Ziyang Yuan, Xintao Wang, Ailing Zeng, Yu Xiong, Qiang Xu, and Ying Shan. Miradata: A large-scale video dataset with long durations and structured captions. arXiv preprint arXiv:2407.06358, 2024. * [10] Xuan Ju, Xian Liu, Xintao Wang, Yuxuan Bian, Ying Shan, and Qiang Xu. Brushnet: A plug-and-play image inpainting model with decomposed dual-branch diffusion. arXiv preprint arXiv:2403.06976, 2024. * [11] Xuan Ju, Ailing Zeng, Yuxuan Bian, Shaoteng Liu, and Qiang Xu. Direct inversion: Boosting diffusion-based editing with 3 lines of code. arXiv preprint arXiv:2310.01506, 2023. * [12] Xuan Ju, Ailing Zeng, Jianan Wang, Qiang Xu, and Lei Zhang. Human-art: A versatile human-centric dataset bridging natural and artificial scenes. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 618–629, 2023. * [13] Xuan Ju, Ailing Zeng, Chenchen Zhao, Jianan Wang, Lei Zhang, and Qiang Xu. Humansd: A native skeleton-guided diffusion model for human image generation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 15988–15998, 2023. * [14] Yanyu Li, Xian Liu, Anil Kag, Ju Hu, Yerlan Idelbayev, Dhritiman Sagar, Yanzhi Wang, Sergey Tulyakov, and Jian Ren. Textcraftor: Your text encoder can be image quality controller. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 7985–7995, 2024. * [15] Xian Liu, Jian Ren, Aliaksandr Siarohin, Ivan Skorokhodov, Yanyu Li, Dahua Lin, Xihui Liu, Ziwei Liu, and Sergey Tulyakov. Hyperhuman: Hyper-realistic human generation with latent structural diffusion. arXiv preprint arXiv:2310.08579, 2023. * [16] Xian Liu, Yinghao Xu, Qianyi Wu, Hang Zhou, Wayne Wu, and Bolei Zhou. Semantic-aware implicit neural audio-driven video portrait generation. In European conference on computer vision, pages 106–125. Springer, 2022. * [17] Tianhe Ren, Shilong Liu, Ailing Zeng, Jing Lin, Kunchang Li, He Cao, Jiayu Chen, Xinyu Huang, Yukang Chen, Feng Yan, et al. Grounded sam: Assembling open-world models for diverse visual tasks. arXiv preprint arXiv:2401.14159, 2024. * [18] Robin Rombach, Andreas Blattmann, Dominik Lorenz, Patrick Esser, and Björn Ommer. High-resolution image synthesis with latent diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 10684–10695, June 2022. * [19] Jiaming Song, Chenlin Meng, and Stefano Ermon. Denoising diffusion implicit models. arXiv preprint arXiv:2010.02502, 2020. * [20] Sihan Xu, Yidong Huang, Jiayi Pan, Ziqiao Ma, and Joyce Chai. Inversion-free image editing with natural language. arXiv preprint arXiv:2312.04965, 2023. * [21] Junhao Zhuang, Yanhong Zeng, Wenran Liu, Chun Yuan, and Kai Chen. A task is worth one word: Learning with task prompts for high-quality versatile image inpainting. arXiv preprint arXiv:2312.03594, 2023.
# Catenary and Mercator projection Mikhail A. Akhukov Huawei Nizhny Novgorod Research Center, Nizhny Novgorod, Russia <EMAIL_ADDRESS> & Vasiliy A. Es’kin Department of Radiophysics, University of Nizhny Novgorod, Nizhny Novgorod, Russia, 603950 and Huawei Nizhny Novgorod Research Center, Nizhny Novgorod, Russia <EMAIL_ADDRESS> & Mikhail E. Smorkalov Skolkovo Institute of Science and Technology, Moscow, Russia and Huawei Nizhny Novgorod Research Center, Nizhny Novgorod, Russia <EMAIL_ADDRESS>Corresponding author: Mikhail Aleksandrovich Akhukov <EMAIL_ADDRESS> ###### Abstract The Mercator projection is sometimes confused with another mapping technique, specifically the central cylindrical projection, which projects the Earth’s surface onto a cylinder tangent to the equator, as if a light source is at the Earth’s center. Accidentally, this misconception is rather close to a truth. The only operation that the map needs is a free bending in a uniform gravitational field if the map’s material is dense and soft enough to produce a catenary profile. The north and south edges of the map should be parallel and placed in the same plane at the appropriate distance. In this case, the bent map been projected onto this plane gives the Mercator projection. This property is rather curious, since it allows to make such a sophisticated one- to-one mapping as the Mercator projection using simple tools available in the workroom. _Keywords_ Mercator projection, Central cylindrical projection, Gudermannian function, Catenary ## 1 Introduction A catenary is a curve that appears in the profile of an idealized freely hanging heavy chain under its own weight, supported only at its ends in a uniform gravitational field. The exact shape of the curve is described by the hyperbolic cosine function. The Mercator projection is a cylindrical map projection. It is a conformal projection, which means that it preserves the angles. The formal equation for the Mercator projection can be expressed in a number of ways. Within them, there is also a variant that includes a hyperbolic cosine function (Osborne, 2016). This fact will be discussed in more details further. The map in Mercator projection was published in 1569 by the Flemish philosopher and mathematician Gerhardus Mercator. The map was originally designed for navigation, as it maps loxodromes into the straight lines. The loxodrome or rhumb line is a line on the sphere of a ship’s course in the constant compass direction. This allows sailors to plot a straight course between two points by following a constant compass bearing. The exact calculation of Mercator projection requires logarithms and modern calculus that had not yet been invented at that time. Figure 1: To calculate the Mercator projection of a given point $N$ with latitude $\varphi$, we first project the point $N$ to the point $B$ on the vertical line $AB$, then move it vertically downward to the point $Q$. The exact vertical coordinate of the point $Q$ is $\psi$, which is expressed by the equation (1). ## 2 Mercator projection The Mercator projection is derived from a differential equation, which is based on the conception of equal scales and requires the integration of the secant function. Having a latitude $\varphi$ (see figure 1), in case of a unit sphere, the formal equation of the vertical coordinate of the Mercator map projection is as follow: $\psi:=\int\limits_{0}^{\varphi}\frac{dt}{\cos{t}}=\ln\left(\tan\left(\frac{\varphi}{2}+\frac{\pi}{4}\right)\right).$ (1) This equation can be expressed in a number of ways using the Gudermannian function ((Weisstein, 1999) page 778, (Yanpolskiy, 1960) page 47), which is popular in calculus, because it relates circular ($\varphi$) and hyperbolic ($\psi$) angle measures as $\varphi=gd(\psi)$, and allows to establish a number of relations between trigonometric and hyperbolic function directly without the imaginary unit. The theory of the Gudermannian function introduces a concept of hyperbolic amplitude ((Yanpolskiy, 1960) page 47), that is a special angle constructed for an arbitrary point belonging to a unit hyperbola. It is also known as the gudermannian of the hyperbolic measure $\psi$ because it is expressed in terms of the Gudermannian function: $\varphi=\operatorname{gd}(\psi)$. Figure 2: The hyperbolic amplitude $\varphi$ is constructed for an arbitrary point $M$ belonging to the unit hyperbola (blue curve), as shown here. Also $\varphi$ is the latitude of the point $N$ figured out in the Mercator projection (see figure 1 for comparison). The theory of the Gudermannian function allows to establish a number of relations (see figure 2): $L:=|AB|=\tan(\varphi)=\sinh(\psi)=|MP|,$ (2) $|OB|=\frac{1}{\cos(\varphi)}=\cosh(\psi)=|OP|.$ (3) Using elementary trigonometry in case of $0\leq\varphi<\pi/2$ we can write: $\tan\left(\frac{\varphi+\pi/2}{2}\right)=\frac{1+\sin{\varphi}}{\cos{\varphi}}=\frac{1}{\cos{\varphi}}+\sqrt{\frac{1}{\cos^{2}{\varphi}}-1}.$ (4) Equation (4) allows to rewrite the right-hand side of equation (1) in the next form: $\psi=\ln\left(\tan\left(\frac{\varphi}{2}+\frac{\pi}{4}\right)\right)=\operatorname{arcosh}\left(\frac{1}{\cos(\varphi)}\right).$ (5) See also equation (3) for comparison. For more details on the integration of the secant function and its relation to the history of the Mercator projection see Vis (2018). Equation (5), via area hyperbolic cosine, implicitly introduces a catenary oriented in the horizontal direction, when vertexes of the unit hyperbola and area hyperbolic cosine touch each other, see figure 3. Figure 3: After introduction of the hyperbolic amplitude $\varphi$ and disclose its relation to the hyperbolic measure $\psi$ via the Gudermannian function, the length of the catenary (solid magenta curve $AC$) introduced implicitly via the area hyperbolic cosine as described by equation (5) is equal to the length $L$ (solid green line $AB$) that is vertical component of the central cylindrical map projection. A key point in this work is next. We need to calculate the length of the area hyperbolic cosine introduced in such a way as in the equation (5), as a function of the hyperbolic measure $\psi$, which is the vertical component of the Mercator map projection: $L_{catenary}(\psi)=\\\ \int\limits_{0}^{\psi}\sqrt{1+\left(\frac{d\cosh(t)}{dt}\right)^{2}}dt=\\\ \int\limits_{0}^{\psi}\sqrt{1+\sinh^{2}(t)}dt=\\\ \int\limits_{0}^{\psi}\cosh(t)dt=\sinh(\psi).$ (6) Using equation (2) we see that $L_{catenary}(\psi)=\tan(\varphi)=L$, where $L$ — is vertical component of the central cylindrical map projection. The equality of two these lengths, the length of line segment $AB$ and the length of catenary segment $AC$ allows to construct the Mercator map projection from the central cylindrical map projection by bending the former on the catenary profile and then project the bent map onto the plane. See figure 3 111All figures are generated by the scripts in Python programming language (Van Rossum and Drake Jr, 1995) using the matplotlib library (Hunter, 2007). ### 2.1 Bending parameter Finally, we calculate the exact parameter of the bending for the map in the central cylindrical projection to produce the map in the Mercator projection. If the map in the central cylindrical projection has the plots of a number of loxodromes, then the straightening of the loxodromes has to be observed from high enough point of view, that will be an unavoidable sign of the Mercator map projection. Suppose, we have a map in the central cylindrical projection constructed between $\alpha$ degrees of north and south latitudes and physical map’s height is $H$. The only parameter we need is a distance $D$ between the north and south edges of the map been placed parallel and in the same plane. Using equation (2) in case of the unit sphere the distance $D_{unit~{}sphere}$ is exactly $2\psi$: $D_{unit~{}sphere}=2\operatorname{arsinh}(\tan(\alpha))$. In case of sphere with arbitrary radius $R$: $R=\frac{H}{2\tan{\alpha}},$ (7) we re-scale $D_{unit~{}sphere}$ by $R$, that gives the value of distance $D$: $D=\frac{H}{\tan{\alpha}}\operatorname{arsinh}\left(\tan{\alpha}\right).$ (8) ## 3 Conclusion We show that the map in the Mercator projection can be constructed from the map in the central cylindrical projection as if some one has such map been printed on dense and soft enough material allows it to hang freely to make the catenary profile and then project the bent map onto the plane. Interestingly speaking, if such map has a number of loxodromes, then a person standing high enough can see that the loxodromes become straight lines when original map is bent when, for example, the map is carried by two other persons if they hold the map at the north and south edges. This hypothetical situation could happen in real life in some kind of workroom and serve as a basis for further study of this property — the straightening of loxodromes printed on the map in the central cylindrical projection when this map hangs freely as described above. ## References * Osborne [2016] Peter Osborne. _The Mercator Projections_. Zenodo, February 2016. doi:10.5281/zenodo.35392. URL https://doi.org/10.5281/zenodo.35392. * Weisstein [1999] Eric W Weisstein. _CRC concise encyclopedia of mathematics_. CRC press, 1999. * Yanpolskiy [1960] Avraam Ruvimovich Yanpolskiy. _Giperbolicheskie funkcii_. Ripol Classic, 1960. * Vis [2018] MJP Vis. History of the mercator projection. B.S. thesis, Utrecht University, The Netherlands, 2018. * Van Rossum and Drake Jr [1995] Guido Van Rossum and Fred L Drake Jr. _Python tutorial_ , volume 620. Centrum voor Wiskunde en Informatica Amsterdam, The Netherlands, 1995\. * Hunter [2007] J. D. Hunter. Matplotlib: A 2d graphics environment. _Computing in Science & Engineering_, 9(3):90–95, 2007.
# Approximation schemes for bounded distance problems on fractionally treewidth-fragile graphs Zdeněk Dvořák Supported by the ERC-CZ project LL2005 (Algorithms and complexity within and beyond bounded expansion) of the Ministry of Education of Czech Republic. email<EMAIL_ADDRESS>Charles University, Prague, Czech Republic Abhiruk Lahiri Supported by ISF grant 822/18 and Ariel University Post-doctoral fellowship. email<EMAIL_ADDRESS>Ariel University ###### Abstract We give polynomial-time approximation schemes for monotone maximization problems expressible in terms of distances (up to a fixed upper bound) and efficiently solvable in graphs of bounded treewidth. These schemes apply in all fractionally treewidth-fragile graph classes, a property which is true for many natural graph classes with sublinear separators. We also provide quasipolynomial-time approximation schemes for these problems in all classes with sublinear separators. ## 1 Introduction In this paper, we consider optimization problems such as: * • Maximum $r$-Independent Set, $r\in\mathbb{Z}^{+}$: Given a graph $G$, the objective is to find a largest subset $X\subseteq V(G)$ such that distance in $G$ between any two vertices in $X$ is at least $r$. * • Maximum weight induced forest: Given a graph $G$ and an assignment $w:V(G)\to\mathbb{Z}_{0}^{+}$ of non-negative weights to vertices, the objective is to find a subset $X\subseteq V(G)$ such that $G[X]$ does not contain a cycle and subject to that, $w(X)\coloneqq\sum_{v\in X}w(v)$ is maximized. * • Maximum $(F,r)$-Matching, for a fixed connected graph $F$ and $r\in\mathbb{Z}^{+}$: Given a graph $G$, the objective is to find a largest subset $X\subseteq V(G)$ such that $G[X]$ can be partitioned into vertex- disjoint copies of $F$ such that distance in $G$ between any two vertices belonging to different copies is at least $r$. To be precise, to fall into the scope of our work, the problem must satisfy the following conditions: * • It must be a maximization problem on certain subsets of vertices of an input graph, possibly with non-negative weights. That is, the problem specifies which subsets of vertices of the input graph are _admissible_ , and the goal is to find an admissible subset of largest size or weight. * • The problem must be defined in terms of distances between the vertices, up to some fixed bound. That is, there exists a parameter $r\in\mathbb{Z}^{+}$ such that for any graphs $G$ and $G^{\prime}$, sets $X\subseteq V(G)$ and $X^{\prime}\subseteq V(G^{\prime})$, and a bijection $f:X\to X^{\prime}$, if $\min(r,d_{G}(u,v))=\min(r,d_{G^{\prime}}(f(u),f(v)))$ holds for all $u,v\in X$, then $X$ is admissible in $G$ if and only if $X^{\prime}$ is admissible in $G^{\prime}$. * • The problem must be monotone (i.e., all subsets of an admissible set must be admissible), or at least near-monotone (as happens for example for Maximum $(F,r)$-Matching) in the following sense: There exists a parameter $c\in\mathbb{Z}^{+}$ such that for any admissible set $A$ in a graph $G$, there exists a system $\\{R_{v}\subseteq A:v\in A\\}$ of subsets of $A$ such that every vertex belongs to $R_{v}$ for at most $c$ vertices $v\in A$, $v\in R_{v}$ for each $v\in A$, and for any $Z\subseteq A$, the subset $X\setminus\bigcup_{v\in Z}R_{v}$ is admissible in $G$. * • The problem must be tractable in graphs of bounded treewidth, that is, there must exist a function $g$ and a polynomial $p$ such that given any graph $G$, its tree decomposition of width $t$, an assignment $w$ of non-negative weights to the vertices of $G$, and a set $X_{0}\subseteq X$, it is possible to find a maximum-weight admissible subset of $X_{0}$ in time $g(t)p(|V(G)|)$. Let us call such problems _$(\leq\\!r)$ -distance determined $c$-near-monotone $(g,p)$-tw-tractable_. Note that a convenient way to verify these assumptions is to show that the problem is expressible in _solution-restricted Monadic Second-Order Logic_ ($\operatorname{\mathsf{MSOL}}$) _with bounded-distance predicates_ , i.e., by a $\operatorname{\mathsf{MSOL}}$ formula with one free variable $X$ such that the quantification is restricted to subsets and elements of $X$, and using binary predicates $d_{1}$, …, $d_{r}$, where $d_{i}(u,v)$ is interpreted as testing whether the distance between $u$ and $v$ in the whole graph is at most $i$. This ensures that the problem is $(\leq\\!r)$-distance determined, and $(g,O(n))$-tw-tractable for some function $g$ by Courcelle’s metaalgorithmic result [5]. Of course, the problems satisfying the assumptions outlined above are typically hard to solve optimally, even in rather restrictive circumstances. For example, Maximum Independent Set is $\operatorname{\mathsf{NP}}$-hard even in planar graphs of maximum degree at most $3$ and arbitrarily large (fixed) girth [1]. Moreover, it is hard to approximate it within factor of $0.995$ in graphs of maximum degree at most three [4]. Hence, to obtain polynomial-time approximation schemes ($\operatorname{\mathsf{PTAS}}$), i.e., polynomial-time algorithms for approximating within any fixed precision, further restrictions on the considered graphs are needed. A natural restriction that has been considered in this context is the requirement that the graphs have sublinear separators (a set $S$ of vertices of a graph $G$ is a _balanced separator_ if every component of $G\setminus S$ has at most $|V(G)|/2$ vertices, and a hereditary class $\mathcal{G}$ of graphs has _sublinear separators_ if for some $c<1$, every graph $G\in\mathcal{G}$ has a balanced separator of size $O(|V(G)|^{c})$). This restriction still lets us speak about many interesting graph classes (planar graphs [18] and more generally proper minor-closed classes [2], many geometric graph classes [20], …). Moreover, the problems discussed above admit $\operatorname{\mathsf{PTAS}}$ in all classes with sublinear separators or at least in substantial subclasses of these graphs: * • Maximum Independent Set has been shown to admit $\operatorname{\mathsf{PTAS}}$ in graphs with sublinear separators already in the foundational paper of Lipton and Tarjan [19]. * • For any positive integer, Maximum $r$-Independent Set and several other problems are known to admit $\operatorname{\mathsf{PTAS}}$ in graphs with sublinear separators by a straightforward local search algorithm [16]. * • All of the problems mentioned above (an more) are known to admit $\operatorname{\mathsf{PTAS}}$ in planar graphs by a layering argument of Baker [3]; this approach can be extended to some related graph classes, including all proper minor-closed classes [6, 12]. * • The problems also admit $\operatorname{\mathsf{PTAS}}$ in graph classes that admit thin systems of overlays [11], a technical property satisfied by all proper minor-closed classes and by all hereditary classes with sublinear separators and bounded maximum degree. * • Bidimensionality arguments [7] apply to a wide range of problems in proper minor-closed graph classes. However, each of the outlined approaches has drawbacks. On one side, the local search approach only applies to specific problems and does not work at all in the weighted setting. On the other side of the spectrum, Baker’s approach is quite general as far as the problems go, but there are many hereditary graph classes with sublinear separators to which it does not seem to apply. The approach through thin systems of overlays tries to balance these concerns, but it is rather technical and establishing this property is difficult. Another option that has been explored is via _fractional treewidth-fragility_. For a function $f\colon\mathbb{Z}^{+}\times\mathbb{Z}^{+}\to\mathbb{Z}^{+}$ and a polynomial $p$, a class of graphs $\mathcal{G}$ is _$p$ -efficiently fractionally treewidth-$f$-fragile_ if there exists an algorithm that for every $k\in\mathbb{Z}^{+}$ and a graph $G\in\mathcal{G}$ returns in time $p(|V(G)|)$ a collection of subsets $X_{1},X_{2},\dots X_{m}\subseteq V(G)$ such that each vertex of $G$ belongs to at most $m/k$ of the subsets, and moreover, for $i=1,\ldots,m$, the algorithm also returns a tree decomposition of $G\setminus X_{i}$ of width at most $f(k,|V(G)|)$. We say a class is _$p$ -efficiently fractionally treewidth-fragile_ if $f$ does not depend on its second argument (the number of vertices of $G$). This property turns out to hold for basically all known natural graph classes with sublinear separators. In particular, a hereditary class $\mathcal{G}$ of graphs is efficiently fractionally treewidth-fragile if * • $\mathcal{G}$ has sublinear separator and bounded maximum degree [9], * • $\mathcal{G}$ is proper minor-closed [8, 12], or * • $\mathcal{G}$ consists of intersection graphs of convex objects with bounded aspect ratio in a finite-dimensional Euclidean space and the graphs have bounded clique number, as can be seen by a modification of the argument of Erlebach et al. [15]. This includes all graph classes with polynomial growth [17]. In fact, Dvořák conjectured that every hereditary class with sublinear separators is fractionally treewidth-fragile, and gave the following result towards this conjecture. ###### Theorem 1 (Dvořák [10]). There exists a polynomial $p$ so that the following claim holds. For every hereditary class $\mathcal{G}$ of graphs with sublinear separators, there exists a polynomial $q$ such that $\mathcal{G}$ is $p$-efficiently fractionally treewidth-$f$-fragile for the function $f(k,n)=q(k\log n)$. Moreover, Dvořák [9] observed that weighted Maximum Independent Set admits a $\operatorname{\mathsf{PTAS}}$ in any efficiently fractionally treewidth- fragile class of graphs. Indeed, the algorithm is quite simple, based on the observation that for the sets $X_{1}$, …, $X_{m}$ from the definition of fractional treewidth-fragility, at least one of the graphs $G\setminus X_{1}$, …, $G\setminus X_{m}$ (of bounded treewidth) contains an independent set whose weight is within the factor of $1-1/k$ from the optimal solution. A problem with this approach is that it does not generalize to more general problems; even for the Maximum $2$-Independent Set problem, the approach fails, since a $2$-independent set in $G\setminus X_{i}$ is not necessarily $2$-independent in $G$. Indeed, this observation served as one of the motivations behind more restrictive (and more technical) concepts employed in [11, 12]. As our main result, we show that this intuition is in fact false: There is a simple way how to extend the approach outlined in the previous paragraph to all bounded distance determined near-monotone tw-tractable problems. ###### Theorem 2. For every class $\mathcal{G}$ of graphs with bounded expansion, there exists a function $h:\mathbb{Z}^{+}\times\mathbb{Z}^{+}\to\mathbb{Z}^{+}$ such that the following claim holds. Let $c$ and $r$ be positive integers, $g:\mathbb{Z}^{+}\to\mathbb{Z}^{+}$ and $f:\mathbb{Z}^{+}\times\mathbb{Z}^{+}\to\mathbb{Z}^{+}$ functions and $p$ and $q$ polynomials. If $\mathcal{G}$ is $q$-efficiently fractionally treewidth-$f$-fragile, then for every $(\leq\\!r)$-distance determined $c$-near-monotone $(g,p)$-tw-tractable problem, there exists an algorithm that given a graph $G\in\mathcal{G}$, an assignment of non-negative weights to vertices, and a positive integer $k$, returns in time $h(r,c)|V(G)|+q(|V(G)|)\cdot p(|V(G)|)\cdot g(f(h(r,c)k,|V(G)|))$ an admissible subset of $V(G)$ whose weight is within the factor of $1-1/k$ from the optimal one. Note that the assumption that $\mathcal{G}$ has bounded expansion is of little consequence—it is true for any hereditary class with sublinear separators [14] as well as for any fractionally treewidth-fragile class [9]; see Section 2 for more details. The time complexity of the algorithm from Theorem 2 is polynomial if $f$ does not depend on its second argument, and quasipolynomial (exponential in a polylogaritmic function) if $f$ is logarithmic in the second argument and $g$ is single-exponential (i.e., if $\log\log g(n)=O(\log n)$). Hence, we obtain the following corollaries. ###### Corollary 3. Let $c$ and $r$ be positive integers, $g:\mathbb{Z}^{+}\to\mathbb{Z}^{+}$ a function and $p$ a polynomial. Every $(\leq\\!r)$-distance determined $c$-near-monotone $(g,p)$-tw-tractable problem admits a $\operatorname{\mathsf{PTAS}}$ in any efficiently fractionally treewidth- fragile class of graphs. We say a problem admits a quasipolynomial-time approximation schemes ($\mathsf{QPTAS}$) if there exist quasipolynomial-time algorithms for approximating the problem within any fixed precision. Combining Theorems 1 and 2, we obtain the following result. ###### Corollary 4. Let $c$ and $r$ be positive integers, $g:\mathbb{Z}^{+}\to\mathbb{Z}^{+}$ a single-exponential function, and $p$ a polynomial. Every $(\leq\\!r)$-distance determined $c$-near-monotone $(g,p)$-tw-tractable problem admits a $\mathsf{QPTAS}$ in any hereditary class of graphs with sublinear separators. The idea of the algorithm from Theorem 2 is quite simple: We consider the sets $X_{1},\ldots,X_{m}$ from the definition of fractional treewidth-$f$-fragility, extend them to suitable supersets $Y_{1}$, …, $Y_{m}$, and argue that for $i=1,\ldots,m$, any admissible set in $G\setminus X_{i}$ disjoint from $Y_{i}$ is also admissible in $G$, and that for some $i$, the weight of the heaviest admissible set in $G\setminus X_{i}$ disjoint from $Y_{i}$ is within the factor of $1-1/k$ from the optimal one. The construction of the sets $Y_{1}$, …, $Y_{m}$ is based on the existence of orientations with bounded outdegrees that represent all short paths, a result of independent interest that we present in Section 2. Let us remark one can develop the idea of this paper in further directions. Dvořák proved in [13](via a substantially more involved argument) that every monotone maximization problem expressible in first-order logic admits a $\operatorname{\mathsf{PTAS}}$ in any efficiently fractionally treewidth- fragile class of graphs. Note that this class of problems is incomparable with the one considered in this paper (e.g., Maximum Induced Forest is not expressible in the first-order logic, while Maximum Independent Set consisting of vertices belonging to triangles is expressible in the first-order logic but does not fall into the scope of the current paper). Finally, it is worth mentioning that our results only apply to maximization problems. We were able to extend the previous uses of fractional treewidth- fragility by giving a way to handle dependencies over any bounded distance. However, for the minimization problems, we do not know whether fractional treewidth-fragility is sufficient even for the distance-$1$ problems. For a simple example, consider the Minimum Vertex Cover problem in fractionally treewidth-fragile graphs, or more generally in hereditary classes with sublinear separators. While the unweighted version can be dealt with by the local search method [16], we do not know whether there exists a $\operatorname{\mathsf{PTAS}}$ for the weighted version of this problem. ## 2 Paths and orientations in graphs with bounded expansion For $r\in\mathbb{Z}^{+}_{0}$, a graph $H$ is an _$r$ -shallow minor_ of a graph $G$ if $H$ can be obtained from a subgraph of $G$ by contracting pairwise vertex-disjoint connected subgraphs, each of radius at most $r$. For a function $f\colon\mathbb{Z}^{+}\to\mathbb{Z}^{+}$, a class $\mathcal{G}$ of graphs has _expansion bounded_ by $f$ if for all non-negative integers $r$, all $r$-shallow minors of graphs from $\mathcal{G}$ have average degree at most $f(r)$. A class has bounded expansion if its expansion is bounded by some function $f$. The theory of graph classes with bounded expansion has been developed in the last 15 years, and the concept has found many algorithmic and structural applications; see [22] for an overview. Crucially for us, this theory includes a number of tools for dealing with short paths. Moreover, as we have pointed out before, all hereditary graph classes with sublinear separators [14] as well as all fractionally treewidth-fragile classes [9] have bounded expansion. Let $\vec{G}$ be an orientation of a graph $G$, i.e, $uv$ is an edge of $G$ if and only if the directed graph $\vec{G}$ contains at least one of the directed edges $(u,v)$ and $(v,u)$; note that we allow $\vec{G}$ to contain both of them at the same time, and thus for the edge $uv$ to be oriented in both directions. We say that a directed graph $\vec{H}$ with the same vertex set is a _$1$ -step fraternal augmentation of $\vec{G}$_ if $\vec{G}\subseteq\vec{H}$, for all distinct edges $(x,y),(x,z)\in E(\vec{G})$, either $(y,z)$ or $(z,y)$ is an edge of $\vec{H}$, and for each edge $(y,z)\in E(\vec{H})\setminus E(\vec{G})$, there exists a vertex $x\in V(\vec{G})\setminus\\{y,z\\}$ such that $(x,y),(x,z)\in E(\vec{G})$. That is, to obtain $\vec{H}$ from $\vec{G}$, for each pair of edges $(x,y),(x,z)\in E(\vec{G})$ we add an edge between $y$ and $z$ in one of the two possible directions (we do not specify the direction, but in practice we would choose directions of the added edges that minimize the maximum outdegree of the resulting directed graph). For an integer $a\geq 0$, we say $\vec{F}$ is an _$a$ -step fraternal augmentation of $\vec{G}$_ if there exists a sequence $\vec{G}=\vec{G}_{0},\vec{G}_{1},\ldots,\vec{G}_{a}=\vec{F}$ where for $i=1,\ldots,a$, $\vec{G}_{i}$ is a $1$-step fraternal augmentation of $\vec{G}_{i-1}$. We say $\vec{F}$ is an $a$-step fraternal augmentation of an undirected graph $G$ if $\vec{F}$ is an $a$-step fraternal augmentation of some orientation of $G$. A key property of graph classes with bounded expansion is the existence of fraternal augmentations with bounded outdegrees. Let us remark that whenever we speak about an algorithm returning an $a$-step fraternal augmentation $\vec{H}$ or taking one as an input, this implicitly includes outputing or taking as an input the whole sequence of $1$-step fraternal augmentations ending in $\vec{H}$. ###### Lemma 5 (Nešetřil and Ossona de Mendez [21]). For every class $\mathcal{G}$ with bounded expansion, there exists a function $d:\mathbb{Z}^{+}_{0}\to\mathbb{Z}^{+}$ such that for each $G\in\mathcal{G}$ and each non-negative integer $a$, the graph $G$ has an $a$-step fraternal augmentation of maximum outdegree at most $d(a)$. Moreover, such an augmentation can be found in time $O(d(a)|V(G)|)$. As shown already in [21], fraternal augmentations can be used to succintly represent distances between vertices of the graph. For the purposes of this paper, we need a more explicit representation by an orientation of the original graph (without the additional augmentation edges). By a _walk_ in a directed graph $\vec{G}$, we mean a sequence $W=v_{0}v_{1}v_{2}\ldots v_{b}$ such that for $i=1,\ldots,b$, $(v_{i-1},v_{i})\in E(\vec{G})$ or $(v_{i},v_{i-1})\in E(\vec{G})$; that is, the walk does not have to respect the orientation of the edges. The walk $W$ is _inward directed_ if for some $c\in\\{0,\ldots,b\\}$, we have $(v_{i},v_{i+1})\in E(\vec{G})$ for $i=0,\ldots,c-1$ and $(v_{i},v_{i-1})\in E(\vec{G})$ for $i=c+1,\ldots,b$. For a positive integer $r$, an orientation $\vec{G}$ of a graph $G$ _represents $(\leq\\!r)$-distances_ if for each $u,v\in V(G)$ and each $b\in\\{0,\ldots,r\\}$, the distance between $u$ and $v$ in $G$ is at most $b$ if and only if $\vec{G}$ contains an inward-directed walk of length at most $b$ between $u$ and $v$. Note that given such an orientation with bounded maximum outdegree for a fixed $r$, we can determine the distance between $u$ and $v$ (up to distance $r$) by enumerating all (constantly many) walks of length at most $r$ directed away from $u$ and away from $v$ and inspecting their intersections. Our goal now is to show that graphs from classes with bounded expansion admit orientations with bounded maximum outdegree that represent $(\leq\\!r)$-distances. Let us define a more general notion used in the proof of this claim, adding to the fraternal augmentations the information about the lengths of the walks in the original graph represented by the added edges. A _directed graph with $(\leq\\!r)$-length sets_ is a pair $(\vec{H},\ell)$, where $\vec{H}$ is a directed graph and $\ell$ is a function assigning a subset of $\\{1,\ldots,r\\}$ to each _unordered_ pair $\\{u,v\\}$ of vertices of $\vec{H}$, such that if neither $(u,v)$ nor $(v,u)$ is an edge of $\vec{H}$, then $\ell(\\{u,v\\})=\emptyset$. We say that $(\vec{H},\ell)$ is an _orientation_ of a graph $G$ if $G$ is the underlying undirected graph of $\vec{H}$ and $\ell(\\{u,v\\})=\\{1\\}$ for each $uv\in E(G)$. We say that $(\vec{H},\ell)$ is an _$(\leq\\!r)$ -augmentation_ of $G$ if $V(\vec{H})=V(G)$, for each $uv\in E(G)$ we have $1\in\ell(\\{u,v\\})$, and for each $u,v\in V(G)$ and $b\in\ell(\\{u,v\\})$ there exists a walk of length $b$ from $u$ to $v$ in $G$. Let $(\vec{H}_{1},\ell_{1})$ be another directed graph with $(\leq\\!r)$-length sets. We say $(\vec{H}_{1},\ell_{1})$ is a _1-step fraternal augmentation_ of $(\vec{H},\ell)$ if $\vec{H}_{1}$ is a $1$-step fraternal augmentation of $\vec{H}$ and for all distinct $u,v\in V(\vec{H})$ and $b\in\\{1,\ldots,r\\}$, we have $b\in\ell_{1}(\\{u,v\\})$ if and only if $b\in\ell(\\{u,v\\})$ or there exist $x\in V(\vec{H})\setminus\\{u,v\\}$, $b_{1}\in\ell(\\{x,u\\})$, and $b_{2}\in\ell(\\{x,v\\})$ such that $(x,u),(x,v)\in E(\vec{H})$ and $b=b_{1}+b_{2}$. Note that a $1$-step fraternal augmentation of an $(\leq\\!r)$-augmentation of a graph $G$ is again an $(\leq\\!r)$-augmentation of $G$. The notion of an $a$-step fraternal augmentation of a graph $G$ is then defined in the natural way, by starting with an orientation of $G$ and peforming the $1$-step fraternal augmentation operation $a$-times. Let us now restate Lemma 5 in these terms (we just need to maintain the edge length sets, which can be done with $O(a^{2})$ overhead per operation). ###### Lemma 6. Let $\mathcal{G}$ be a class of graphs with bounded expansion, and let $d:\mathbb{Z}^{+}_{0}\to\mathbb{Z}^{+}$ be the function from Lemma 5. For each $G\in\mathcal{G}$ and each non-negative integer $a$, we can in time $O(a^{2}d(a)|V(G)|)$ construct a directed graph with $(\leq\\!a+1)$-length sets $(\vec{H},\ell)$ of maximum outdegree at most $d(a)$ such that $(\vec{H},\ell)$ is an $a$-step fraternal augmentation of $G$. Let $(\vec{H},\ell)$ be an $(\leq\\!r)$-augmentation $(\vec{H},\ell)$ of a graph $G$. For $b\leq r$, a _length $b$ walk_ in $(\vec{H},\ell)$ is a tuple $(v_{0}v_{1}\ldots v_{t},b_{1},\ldots,b_{t})$, where $v_{0}v_{1}\ldots v_{t}$ is a walk in $\vec{H}$, $b_{i}\in\ell(\\{v_{i-1},v_{i}\\}$ for $i=1,\ldots,t$, and $b=b_{1}+\ldots+b_{t}$. Note that if there exists a length $b$ walk from $u$ to $v$ in $(\vec{H},\ell)$, then there also exists a walk of length $b$ from $u$ to $v$ in $G$. We say that $(\vec{H},\ell)$ _represents $(\leq\\!r)$-distances_ in $G$ if for all vertices $u,v\in V(G)$ at distance $b\leq r$ from one another, $(\vec{H},\ell)$ contains an inward-directed length $b$ walk between $u$ and $v$. Next, we show that this property always holds for sufficient fraternal augmentations. ###### Lemma 7. Let $G$ be a graph and $r$ a positive integer and let $(\vec{H},\ell)$ be a directed graph with $(\leq\\!r)$-length sets. If $(\vec{H},\ell)$ is obtained as an $(r-1)$-step fraternal augmentation of $G$, then it represents $(\leq\\!r)$-distances in $G$. ###### Proof. For $b\leq r$, consider any length $b$ walk $W=(v_{0}v_{1}\ldots v_{t},b_{1},\ldots,b_{t})$ in an $(\leq\\!r)$-augmentation $(\vec{H}_{1},\ell_{1})$ of $G$, and let $(\vec{H}_{2},\ell_{2})$ be a $1$-step augmentation of $(\vec{H}_{1},\ell_{1})$. Note that $W$ is also a length $b$ walk between $v_{0}$ and $v_{t}$ in $(\vec{H}_{2},\ell_{2})$. Suppose that $W$ is not inward-directed in $(\vec{H}_{1},\ell_{1})$, and thus there exists $i\in\\{1,\ldots,t-1\\}$ such that $(v_{i},v_{i-1}),(v_{i},v_{i+1})\in E(\vec{H}_{1})$. By the definition of $1$-step fraternal augmentation, this implies $b_{i}+b_{i+1}\in\ell_{2}(v_{i-1},v_{i+1})$, and thus $(v_{0}\ldots v_{i-1}v_{i+1}\ldots v_{t},b_{1},\ldots,b_{i}+b_{i+1},\ldots b_{t})$ is a length $b$ walk from $v_{0}$ to $v_{t}$ in $(\vec{H}_{2},\ell_{2})$. Let $(\vec{G}_{0},\ell_{0})$, …, $(\vec{G}_{r-1},\ell_{r-1})$ be a sequence of $(\leq\\!r)$-augmentations of $G$, where $(\vec{G},\ell_{0})$ is an orientation of $G$, $(\vec{G}_{r-1},\ell_{r-1})=(\vec{H},\ell)$, and for $i=1,\ldots,r-1$, $(\vec{G}_{i},\ell_{i})$ is a $1$-step fraternal augmentation of $(\vec{G}_{i-1},\ell_{i-1})$. Let $u$ and $v$ be any vertices at distance $b\leq r$ in $G$, and let $P$ be a shortest path between them. Then $P$ naturally corresponds to a length $b$ walk $P_{0}$ in $(\vec{G}_{0},\ell_{0})$. For $i=1,\ldots,r-1$, if $P_{i-1}$ is inward- directed, then let $P_{i}=P_{i-1}$, otherwise let $P_{i}$ be a length $b$ walk in $(\vec{G}_{i},\ell_{i})$ obtained from $P_{i-1}$ as described in the previous paragaph. Since each application of the operation decreases the number of vertices of the walk, we conclude that $P_{r-1}$ is an inward- directed length $b$ walk between $u$ and $v$ in $(\vec{H},\ell)$. Hence, $(\vec{H},\ell)$ represents $(\leq\\!r)$-distances in $G$. ∎ Next, let us propagate this property back through the fraternal augmentations by orienting some of the edges in both directions. We say that $(\vec{H},\ell)$ is an _$a$ -step fraternal superaugmentation_ of a graph $G$ if there exists an $a$-step fraternal augmentation $(\vec{F},\ell)$ of $G$ such that $V(\vec{F})=V(\vec{H})$, $E(\vec{F})\subseteq E(\vec{H})$ and for each $(u,v)\in E(\vec{H})\setminus E(\vec{F})$, we have $(v,u)\in E(\vec{F})$. We say that $(\vec{F},\ell)$ is a _support_ of $(\vec{H},\ell)$. ###### Lemma 8. Let $G$ be a graph and $r$ a positive integer and let $(\vec{H},\ell)$ be an $(\leq\\!r)$-augmentation of $G$ of maximum outdegree $\Delta$ representing $(\leq\\!r)$-distances. For $a\geq 1$, suppose that $(\vec{H},\ell)$ is an $a$-step fraternal superaugmentation of $G$. Then we can in time $O(r^{2}\Delta|V(G)|)$ obtain an $(a-1)$-step fraternal superaugmentation of $G$ representing $(\leq\\!r)$-distances, of maximum outdegree at most $(r+1)\Delta$. ###### Proof. Let $(\vec{F},\ell)$ be an $a$-step fraternal augmentation of $G$ forming a support of $(\vec{H},\ell)$, obtained as a $1$-step fraternal augmentation of an $(a-1)$-step fraternal augmentation $(\vec{F}_{1},\ell_{1})$ of $G$. Let $(\vec{H}_{1},\ell_{1})$ be the $(a-1)$-step fraternal superaugmentation of $G$ obtained from $(\vec{F}_{1},\ell_{1})$ as follows: * • For all distinct vertices $y,z\in V(G)$ such that $(y,z),(z,y)\in E(\vec{H})$, $(y,z)\in E(\vec{F}_{1})$, and $(z,y)\not\in E(\vec{F}_{1})$, we add the edge $(z,y)$. * • For each edge $(y,z)\in E(\vec{H})$ and integer $b\in\ell(\\{y,z\\})\setminus\ell_{1}(\\{y,z\\})$, we choose a vertex $x\in V(G)\setminus\\{y,z\\}$ such that $(x,y),(x,z)\in E(\vec{F}_{1})$ and $b=b_{1}+b_{2}$ for some $b_{1}\in\ell_{1}(\\{x,y\\})$ and $b_{2}\in\ell_{1}(\\{x,z\\})$, and add the edge $(y,x)$. Note that such a vertex $x$ and integers $b_{1}$ and $b_{2}$ exist, since $b$ was added to $\ell(\\{y,z\\})$ when $(\vec{F},\ell)$ was obtained from $(\vec{F}_{1},\ell_{1})$ as a $1$-step fraternal augmentation. Each edge $(y,x)\in E(\vec{H}_{1})\setminus E(\vec{H})$ arises from an edge $(y,z)\in E(\vec{H})$ leaving $y$ and an element $b\in\ell(\\{y,z\\})\setminus\ell_{1}(\\{y,z\\})$, and each such pair contributes at most one edge leaving $y$. Hence, the maximum outdegree of $\vec{H}_{1}$ is at most $(r+1)\Delta$. Consider a length $b$ inwards-directed walk $(v_{0}v_{1}\ldots v_{t},b_{1},\ldots,b_{t})$ in $\vec{H}$, for any $b\leq r$. Then $\vec{H}$ contains a length $b$ inwards-directed walk from $v_{0}$ to $v_{t}$ obtained by natural edge replacements: For any edge $(y,z)\in E(\vec{H})$ of this walk and $b^{\prime}\in\ell_{i}(\\{y,z\\})$, the construction described above ensures that if $(y,z)\not\in E(\vec{H}_{1})$ or $b^{\prime}\not\in\ell_{1}(\\{y,z\\})$, then there exists $x\in V(G)\setminus\\{y,z\\}$ such that $(y,x),(x,z)\in E(\vec{H}_{1})$ and $b^{\prime}=b^{\prime\prime}+b^{\prime\prime\prime}$ for some $b^{\prime\prime}\in\ell_{1}(\\{x,y\\})$ and $b^{\prime\prime\prime}\in\ell_{1}(\\{x,z\\})$, and we can replace the edge $(y,z)$ in the walk by the edges $(y,x)$ and $(x,z)$ of $E(\vec{H}_{1})$. Since $\vec{H}$ represents $(\leq\\!r)$-distances in $G$, this transformation shows that so does $\vec{H}_{1}$. ∎ We are now ready to prove the main result of this section. ###### Lemma 9. For any class $\mathcal{G}$ with bounded expansion, there exists a function $d^{\prime}:\mathbb{Z}^{+}\to\mathbb{Z}^{+}$ such that for each $G\in\mathcal{G}$ and each positive integer $r$, the graph $G$ has an orientation with maximum outdegree at most $d^{\prime}(r)$ that represents $(\leq\\!r)$-distances in $G$. Moreover, such an orientation can be found in time $O(r^{2}d^{\prime}(r)|V(G)|)$. ###### Proof. Let $d$ be the function from Lemma 5, and let $d^{\prime}(r)=(r+1)^{r-1}d(r-1)$. By Lemma 6, we obtain an $(r-1)$-step fraternal augmentation $(\vec{H},\ell)$ of $G$ of maximum outdegree at most $d(r-1)$. By Lemma 7, $(\vec{H},\ell)$ represents $(\leq\\!r)$-distances in $G$. Repeatedly applying Lemma 8, we obtain a $0$-step fraternal superaugmentation $(\vec{G},\ell_{0})$ of $G$ of maximum outdegree at most $d^{\prime}(r)$ representing $(\leq\\!r)$-distances. Clearly, $\vec{G}$ is an orientation of $G$ of maximum outdegree at most $d^{\prime}(r)$ representing $(\leq\\!r)$-distances. ∎ ## 3 Approximation schemes Let us now prove Theorem 2. To this end, let us start with a lemma to be applied to the sets arising from fractional treewidth-fragility. ###### Lemma 10. Let $\vec{G}$ be an orientation of a graph $G$ with maximum outdegree $\Delta$. Let $A$ be a set of vertices of $G$ and for a positive integer $c$, let $\\{R_{v}:v\in A\\}$ be a system of subsets of $A$ such that each vertex belongs to at most $c$ of the subsets. For $X\subseteq V(G)$ and a positive integer $r$, let $D_{\vec{G},r}(X)$ be the union of the sets $R_{v}$ for all vertices $v\in V(G)$ such that $\vec{G}$ contains a walk from $v$ to $X$ of length at most $r$ directed away from $v$. For a positive integer $k$, let $X_{1}$, …, $X_{m}$ be a system of subsets of $V(G)$ such that each vertex belongs to at most $\frac{m}{c(\Delta+1)^{r}k}$ of the subsets. For any assignment $w$ of non-negative weights to vertices of $G$, there exists $i\in\\{1,\ldots,m\\}$ such that $w(A\setminus D_{\vec{G},r}(X_{i}))\geq(1-1/k)w(A)$. ###### Proof. For a vertex $z\in A$, let $B(z)$ be the set of vertices reachable in $\vec{G}$ from vertices $v\in A$ such that $z\in R_{v}$ by walks of length at most $r$ directed away from $v$. Note that $|B(z)|\leq c(\Delta+1)^{r}$ and that for each $X\subseteq V(G)$, we have $z\in D_{\vec{G},r}(X)$ if and only if $B(z)\cap X\neq\emptyset$. Suppose for a contradiction that for each $i$ we have $w(A\setminus D_{\vec{G},r}(X_{i}))<(1-1/k)w(A)$, and thus $w(D_{\vec{G},r}(X_{i}))>w(A)/k$. Then $\displaystyle\frac{m}{k}w(A)$ $\displaystyle<\sum_{i=1}^{m}w(D_{\vec{G},r}(X_{i}))=\sum_{i=1}^{m}\sum_{z\in D_{\vec{G},r}(X_{i})}w(z)=\sum_{i=1}^{m}\sum_{z\in A:B(z)\cap X_{i}\neq\emptyset}w(z)$ $\displaystyle\leq\sum_{i=1}^{m}\sum_{z\in A}w(z)|B(z)\cap X_{i}|=\sum_{z\in A}w(z)\sum_{i=1}^{m}|B(z)\cap X_{i}|$ $\displaystyle=\sum_{z\in A}w(z)\sum_{x\in B(z)}|\\{i\in\\{1,\ldots,m\\}:x\in X_{i}\\}|\leq\sum_{z\in A}w(z)\sum_{x\in B(z)}\frac{m}{c(\Delta+1)^{r}k}$ $\displaystyle=\sum_{z\in A}w(z)|B(z)|\frac{m}{c(\Delta+1)^{r}k}\leq\sum_{z\in A}w(z)\frac{m}{k}=\frac{m}{k}w(A),$ which is a contradiction. ∎ Next, let us derive a lemma on admissibility for $(\leq\\!r)$-distance determined problems. ###### Lemma 11. For a positive integer $r$, let $\vec{G}$ be an orientation of a graph $G$ representing $(\leq\\!r)$-distances. For a set $X\subseteq V(G)$, let $Y_{\vec{G},r}(X)$ be the set of vertices $y$ such that $\vec{G}$ contains a walk from $y$ to $X$ of length at most $r$ directed away from $y$. For any $(\leq\\!r)$-distance determined problem, a set $B\subseteq V(G)\setminus Y_{\vec{G},r}(X)$ is admissible in $G$ if and only if it is admissible in $G-X$. ###### Proof. Since the problem is $(\leq\\!r)$-distance determined, it suffices to show that $\min(r,d_{G}(u,v))=\min(r,d_{G-X}(u,v))$ holds for all $u,v\in B$. Clearly, $d_{G}(u,v)\leq d_{G-X}(u,v)$, and thus it suffices to show that if the distance between $u$ and $v$ is $G$ is $b\leq r$, then $G-X$ contains a walk of length $b$ between $u$ and $v$. Since $\vec{G}$ represents $(\leq\\!r)$-distances, there exists an inward-directed walk $P$ of length $b$ between $u$ and $v$ in $\vec{G}$. Since $u,v\not\in Y_{\vec{G},r}(X)$, we have $V(P)\cap X=\emptyset$, and thus $P$ is also a walk of length $b$ between $u$ and $v$ in $G-X$. ∎ We are now ready to prove the main result. ###### Proof of Theorem 2. Let $d^{\prime}$ be the function from Lemma 9 for the class $\mathcal{G}$. Let us define $h(r,c)=c(d^{\prime}(r)+1)^{r}$. The algorithm is as follows. Since $\mathcal{G}$ is $q$-efficiently fractionally treewidth-$f$-fragile, in time $q(|V(G)|)$ we can find sets $X_{1},\ldots,X_{m}\subseteq V(G)$ such that each vertex belongs to at most $\frac{m}{h(r,c)k}$ of them, and for each $i$, a tree decomposition of $G-X_{i}$ of width at most $f(h(r,c)k,|V(G)|)$. Clearly, $m\leq q(|V(G)|)$. Next, using Lemma 9, we find an orientation $\vec{G}$ of $G$ that represents $(\leq\\!r)$-distances. Let $Y_{\vec{G},r}$ be defined as in the statement of Lemma 11. Since the problem is $(g,p)$-tw-tractable problem, for each $i$ we can in time $p(|V(G)|)\cdot g(f(h(r,c)k,|V(G)|))$ find a subset $A_{i}$ of $V(G)\setminus Y_{\vec{G},r}(X_{i})$ admissible in $G-X_{i}$ of largest weight. By Lemma 11, each of these sets is admissible in $G$; the algorithm return the heaviest of the sets $A_{1}$, …, $A_{m}$. As the returned set is admissible in $G$, it suffices to argue about its weight. Let $A$ be a heaviest admissible set in $G$. Let $\\{R_{v}\subseteq A:v\in A\\}$ be the system of subsets from the definition of $c$-near- monotonicity, and let $D_{\vec{G},r}$ be defined as in the statement of Lemma 10. By the definition of $c$-near-monotonicity, for each $i$ the set $A\setminus D_{\vec{G},r}(X_{i})$ is admissible in $G$. Since $v\in R_{v}$ for each $v\in A$, we have $Y_{\vec{G},r}(X_{i})\subseteq D_{\vec{G},r}(X_{i})$, and thus by Lemma 11, $A\setminus D_{\vec{G},r}(X_{i})$ is also admissible in $G-X_{i}$, and by the choice of $A_{i}$, we have $w(A_{i})\geq w(A\setminus D_{\vec{G},r}(X_{i}))$. By Lemma 10, we conclude that for at least one $i$, we have $w(A_{i})\geq(1-1/k)w(A)$, as required. ∎ ## References * [1] Vladimir E. Alekseev, Vadim V. Lozin, Dmitriy S. Malyshev, and Martin Milanic. The maximum independent set problem in planar graphs. In Mathematical Foundations of Computer Science 2008, 33rd International Symposium, MFCS 2008, Torun, Poland, August 25-29, 2008, Proceedings, volume 5162 of Lecture Notes in Computer Science, pages 96–107. Springer, 2008. https://doi.org/10.1007/978-3-540-85238-4_7. * [2] Noga Alon, Paul D. Seymour, and Robin Thomas. A separator theorem for graphs with an excluded minor and its applications. In Proceedings of the 22nd Annual ACM Symposium on Theory of Computing, May 13-17, 1990, Baltimore, Maryland, USA, pages 293–299. ACM, 1990. https://doi.org/10.1145/100216.100254. * [3] Brenda S. Baker. Approximation algorithms for np-complete problems on planar graphs. J. ACM, 41(1):153–180, 1994. https://doi.org/10.1145/174644.174650. * [4] Piotr Berman and Marek Karpinski. On some tighter inapproximability results (extended abstract). In Automata, Languages and Programming, 26th International Colloquium, ICALP’99, Prague, Czech Republic, July 11-15, 1999, Proceedings, volume 1644 of Lecture Notes in Computer Science, pages 200–209. Springer, 1999. https://doi.org/10.1007/3-540-48523-6_17. * [5] Bruno Courcelle. The monadic second-order logic of graphs. i. recognizable sets of finite graphs. Inf. Comput., 85(1):12–75, 1990. https://doi.org/10.1016/0890-5401(90)90043-H. * [6] Anuj Dawar, Martin Grohe, Stephan Kreutzer, and Nicole Schweikardt. Approximation schemes for first-order definable optimisation problems. In 21th IEEE Symposium on Logic in Computer Science (LICS 2006), 12-15 August 2006, Seattle, WA, USA, Proceedings, pages 411–420. IEEE Computer Society, 2006. https://doi.org/10.1109/LICS.2006.13. * [7] Erik D. Demaine and Mohammad Taghi Hajiaghayi. Bidimensionality: new connections between FPT algorithms and ptass. In Proceedings of the Sixteenth Annual ACM-SIAM Symposium on Discrete Algorithms, SODA 2005, Vancouver, British Columbia, Canada, January 23-25, 2005, pages 590–601. SIAM, 2005. http://dl.acm.org/citation.cfm?id=1070432.1070514. * [8] Matt DeVos, Guoli Ding, Bogdan Oporowski, Daniel P. Sanders, Bruce A. Reed, Paul D. Seymour, and Dirk Vertigan. Excluding any graph as a minor allows a low tree-width 2-coloring. J. Comb. Theory, Ser. B, 91(1):25–41, 2004. https://doi.org/10.1016/j.jctb.2003.09.001. * [9] Zdenek Dvorák. Sublinear separators, fragility and subexponential expansion. Eur. J. Comb., 52:103–119, 2016. https://doi.org/10.1016/j.ejc.2015.09.001. * [10] Zdenek Dvorák. On classes of graphs with strongly sublinear separators. Eur. J. Comb., 71:1–11, 2018. https://doi.org/10.1016/j.ejc.2018.02.032. * [11] Zdenek Dvorák. Thin graph classes and polynomial-time approximation schemes. In Proceedings of the Twenty-Ninth Annual ACM-SIAM Symposium on Discrete Algorithms, SODA 2018, New Orleans, LA, USA, January 7-10, 2018, pages 1685–1701. SIAM, 2018. https://doi.org/10.1137/1.9781611975031.110. * [12] Zdenek Dvorák. Baker game and polynomial-time approximation schemes. In Proceedings of the 2020 ACM-SIAM Symposium on Discrete Algorithms, SODA 2020, Salt Lake City, UT, USA, January 5-8, 2020, pages 2227–2240. SIAM, 2020. https://doi.org/10.1137/1.9781611975994.137. * [13] Zdenek Dvorák. Approximation metatheorem for fractionally treewidth-fragile graphs. CoRR, abs/2103.08698, 2021. https://arxiv.org/abs/2103.08698. * [14] Zdenek Dvorák and Sergey Norin. Strongly sublinear separators and polynomial expansion. SIAM J. Discret. Math., 30(2):1095–1101, 2016. https://doi.org/10.1137/15M1017569. * [15] Thomas Erlebach, Klaus Jansen, and Eike Seidel. Polynomial-time approximation schemes for geometric intersection graphs. SIAM J. Comput., 34(6):1302–1323, 2005. https://doi.org/10.1137/S0097539702402676. * [16] Sariel Har-Peled and Kent Quanrud. Approximation algorithms for polynomial-expansion and low-density graphs. SIAM J. Comput., 46(6):1712–1744, 2017. https://doi.org/10.1137/16M1079336. * [17] Robert Krauthgamer and James R. Lee. The intrinsic dimensionality of graphs. Comb., 27(5):551–585, 2007. https://doi.org/10.1007/s00493-007-2183-y. * [18] Richard J. Lipton and Robert Endre Tarjan. A separator theorem for planar graphs. SIAM J. Appl. Math., 36(2):177–189, 1979. https://doi.org/10.1137/0136016. * [19] Richard J. Lipton and Robert Endre Tarjan. Applications of a planar separator theorem. SIAM J. Comput., 9(3):615–627, 1980. https://doi.org/10.1137/0209046. * [20] Gary L. Miller, Shang-Hua Teng, William P. Thurston, and Stephen A. Vavasis. Separators for sphere-packings and nearest neighbor graphs. J. ACM, 44(1):1–29, 1997. https://doi.org/10.1145/256292.256294. * [21] Jaroslav Nesetril and Patrice Ossona de Mendez. Grad and classes with bounded expansion II. algorithmic aspects. Eur. J. Comb., 29(3):777–791, 2008. https://doi.org/10.1016/j.ejc.2006.07.014. * [22] Jaroslav Nesetril and Patrice Ossona de Mendez. Sparsity - Graphs, Structures, and Algorithms, volume 28 of Algorithms and combinatorics. Springer, 2012. https://doi.org/10.1007/978-3-642-27875-4.
# Dual CFT on Nariai limit for Kerr-Sen-dS black holes Muhammad Fitrah Alfian Rangga Sakti<EMAIL_ADDRESS>Piyabut Burikham <EMAIL_ADDRESS>High Energy Physics Theory Group, Department of Physics, Faculty of Science, Chulalongkorn University, Bangkok 10330, Thailand ###### Abstract In this work, we study the Kerr-Sen-de Sitter black hole (BH) in the Nariai limit where the event and cosmological horizon coincide. We show that the near-horizon Kerr-Sen-de Sitter black hole in Nariai limit is a fiber over AdS2 with an appropriate coordinate transformation, instead of fiber over dS2. Hence, we can compute the associated central charge and CFT temperature by using the Kerr/CFT method. It is remarkably exhibited that through Cardy’s growth of states, the Bekenstein-Hawking entropy on cosmological horizon is reproduced. Moreover, we show that the radial equation of the quantum scalar field in $J$\- and $Q$-pictures on this charged rotating background in Nariai limit can be portrayed in quadratic Casimir operator form with $SL(2,R)\times SL(2,R)$ isometry. We also compute the corresponding thermodynamic quantities from CFT to find the absorption cross-section and real-time correlator in $J$-picture. In $Q$-picture, we do not find a well-defined CFT description. We then extend the study of quantum scalar field in Nariai limit for Kerr-Newman- dS black hole solution and show that the hidden conformal symmetry on this black hole’s background in in $J$\- and $Q$-pictures is well-defined. ††preprint: APS/123-QED ## I Introduction Establishing the holographic correspondence between de Sitter space (dS) and quantum theory is a significant challenge in theoretical physics. The holographic duality, most notably captured by the Anti-de Sitter/Conformal Field Theory (AdS/CFT) correspondence, has provided profound insights into the quantum theory of gravity. However, extending this framework to dS space has proven to be more problematic. One of the main difficulties in the holographic description of dS space arises from the absence of a global timelike Killing vector and spatial infinity. In the AdS/CFT correspondence, the asymptotic behavior of AdS space allows for a well-defined boundary at infinity, which plays a crucial role in the holographic interpretation. This boundary serves as a holographic screen, encoding the dynamics of the gravitational theory in terms of a dual quantum field theory living on the boundary. In contrast, dS space does not possess a similar asymptotic structure or boundary at infinity. As a result, the standard techniques used in the AdS/CFT correspondence cannot be straightforwardly applied to dS space. It is an intriguing and challenging problem that requires a deeper understanding of the quantum nature of gravity, as well as the dynamics of dS space itself. Inspite of the challenges coming from the asymptotically de Sitter spacetime, this solution may provide a correct interpretation to explain our physical world in the context of cosmological model. The standard model of cosmology, whose details are explained extensively e.g. in [1], is surprisingly consistent with this de Sitter solution. The observed accelerated expansion of our universe, supported by astronomical observations such as the cosmic microwave background radiation and the distribution of galaxies, suggests that our universe could be described by a dS spacetime [2, 3]. In particular, the connection between the cosmological constant and the behavior of smaller objects, such as black holes, is also of great interest. The presence of a non-zero cosmological constant affects the configurations and properties of black holes. In fact, the cosmological constant can influence the formation, growth, and evaporation of black holes. Moreover, one can study the quantum gravity from black hole’s properties through the gravitational waves observations [4, 5, 6]. Hence, in the future, we expect that the gravitational wave observations may give us more precise information about quantum structure of black holes including when the cosmological constant is present. Regarding the black hole solutions, one of the central challenges in quantum gravity is to understand the microscopic origin of black hole entropy, which is given by the famous Bekenstein-Hawking formula stating that the entropy is one quarter of the black hole’s event horizon area. Remarkably, the AdS/CFT correspondence provides a potential framework for addressing this problem. It is assumed that the partition function of conformal field theory to be identical to black hole’s partition function. This relation was firstly proven by computing the entropy the extremal Kerr black hole [7]. That study is inspired from the success of the investigation of the asymptotic symmetries of Banados-Teitelboim-Zanelli (BTZ) black hole which contains two-dimensional (2D) local conformal algebra. CFT techniques can then be used to calculate the state degeneracy [8] leading to the agreement of CFT entropy with the Bekenstein-Hawking entropy. In the following, we consider a class of charged rotating black hole solution in Einstein-Maxwell-Dilaton-Axion (EMDA) supergravity theory and Einstein- Maxwell theory, namely dyonic Kerr-Sen-dS (KSdS) and dyonic Kerr-Newman-dS (KNdS) black holes, respectively. Both black holes possess more than two horizons because of the existence of the positive cosmological constant. The largest horizon is the cosmological horizon for which the first thermodynamics law on this horizon resembles the first thermodynamic law on black hole’s horizon. In the cosmological horizon, one can also study the thermal spectrum with non-zero entropy that is also proportional to the horizon area, likewise in the event horizon. The existence of three horizons leads to different limits. When the Cauchy and event horizons coincide, one can obtain an extremal black hole. When the cosmological and event horizon coincide, the resulting metric will be Nariai solution. Furthermore, when all three horizons coincide, the ultracold solution will be produced. The first holographic description of the cosmological entropy of the Nariai solution obtained from Kerr-dS and KNdS solutions has been investigated in Ref. [9]. The Nariai solution possesses $SL(2,R)\times U(1)$ isometry where the spacetime metric is a fiber over dS2. We want to show that in Nariai limit, we can write the solution with similar isometry, although with the spacetime metric as a fiber over AdS2 by choosing different sign on the radial near-horizon coordinate. This metric form on Nariai limit denotes the similar form with near-horizon extremal metric [10, 11, 12, 13]. Furthermore, we want to study the scalar wave equation on the geometry when we take the Nariai limit. As has been studied for generic Kerr solution [14], we exhibit that in Nariai limit, the scalar wave equation possesses $SL(2,R)\times SL(2,R)$ isometry. This isometry happens to appear in scalar wave equation of some black hole solutions [15, 16, 17, 18] and also for higher-spin field [19, 20, 21]. Remarkably, one can also compute the absorption cross-section and real- time correlator from 2D CFT of the solution on Nariai limit. We want to prove that for both KNdS and KSdS black holes in Nariai limit, the conformal symmetry can be shown on the spacetime metric directly and on the scalar wave equation. Moreover, CFT dual will be constructed trough the matching between the entropy, absorption cross-section, and the real-time correlator. We will consider two different pictures, $J$\- and $Q$ pictures. $J$\- picture denotes the CFT description when the scalar probe is neutral while $Q$-picture denotes the CFT description when the scalar probe is electrically charged. We organize the paper as follows. In Sec. II, we carry out the KSdS solution and its thermodynamic quantities on the event and cosmological horizons. In Sec. III, we show the ”fin” diagrams which portray the horizons and the extremal limits. In the next section, we study the Nariai limit on KSdS solution and find the entropy using the Kerr/CFT correspondece. In Sec. V, we study the conformal symmetry on the scalar wave equation in $J$\- and $Q$-pictures in the background of KSdS solution in Nariai limit. In Sec. VI, we further extend the study of the scalar field on KNdS black hole in Nariai limit. In Sec. VII, we summarize our findings for the whole paper. ## II Dyonic Kerr-Sen-de Sitter solution and its thermodynamics The dyonic KSdS black hole is the exact solution to EMDA supergravity theory with positive cosmological constant. The line element of KSdS spacetime reads as [46] $ds^{2}=-\frac{\Delta}{\varrho^{2}}\hat{X}^{2}+\frac{\varrho^{2}}{\Delta}d\hat{r}^{2}+\frac{\varrho^{2}}{\Delta_{\theta}}d\theta^{2}+\frac{\Delta_{\theta}\sin^{2}\theta}{\varrho^{2}}\hat{Y}^{2},\ $ (1) where $\hat{X}=d\hat{t}-a\sin^{2}\theta\frac{d\hat{\phi}}{\Xi},~{}\hat{Y}=ad\hat{t}-(\hat{r}^{2}-d^{2}-k^{2}+a^{2})\frac{d\hat{\phi}}{\Xi},\ $ (2) $\Delta=(\hat{r}^{2}-d^{2}-k^{2}+a^{2})\left(1-\frac{\hat{r}^{2}-d^{2}-k^{2}}{l^{2}}\right)-2m\hat{r}+p^{2}+q^{2},\ \ $ (3) $\Delta_{\theta}=1+\frac{a^{2}}{l^{2}}\cos^{2}\theta,~{}\Xi=1+\frac{a^{2}}{l^{2}},~{}\varrho^{2}=\hat{r}^{2}-d^{2}-k^{2}+a^{2}\cos^{2}\theta.\ $ (4) Note that $m,a,d,k,p,q,l$ are the parameters of mass, spin, dilaton charge, axion charge, magnetic charge, electric charge, and dS length, respectively. The KSdS solution above is obtained from the gauged dyonic Kerr-Sen black hole solution [46, 45] via analytic continuation [28]. There is an important relation between charges of KSdS solution given by $d=\frac{p^{2}-q^{2}}{2m},~{}~{}~{}~{}~{}k=\frac{pq}{m}.\ $ (5) Therefore, when the magnetic charge vanishes, the axion charge also vanishes while when $p=q$, the dilaton charge will vanish. One also can write $d^{2}+k^{2}=\left(\frac{p^{2}+q^{2}}{2m}\right)^{2}.\ $ (6) KSdS black hole possesses three positive horizons. The three positive horizons are inner ($r_{-}$), outer/event ($r_{+}$) and cosmological ($r_{c}$) horizons where $r_{c}\geq r_{+}\geq r_{-}\geq 0$. We can study the thermodynamics on its horizons. Within this paper, we emphasize the study on thermodynamics on the event horizon and cosmological horizon. On the event horizon, the KSdS black hole satisfies the thermodynamic relation $dM=T_{H}dS_{BH}+\Omega_{H}dJ+\Phi_{H}dQ+\Psi_{H}dP+Vd\mathcal{P},$ (7) with the following quantities $M=\frac{m}{\Xi},~{}~{}~{}J=\frac{ma}{\Xi},~{}~{}~{}Q=\frac{q}{\Xi},~{}~{}~{}P=\frac{p}{\Xi},\ $ (8) $T_{H}=\frac{r_{+}(l^{2}-2r_{+}^{2}+2d^{2}+2k^{2}-a^{2})-ml^{2}}{2\pi(r_{+}^{2}-d^{2}-k^{2}+a^{2})l^{2}},$ (9) $S_{BH}=\frac{\pi}{\Xi}(r_{+}^{2}-d^{2}-k^{2}+a^{2}),~{}~{}~{}\Omega_{H}=\frac{a\Xi}{r_{+}^{2}-d^{2}-k^{2}+a^{2}},$ (10) $\Phi_{H}=\frac{q(r_{+}+d-p^{2}/m)}{r_{+}^{2}-d^{2}-k^{2}+a^{2}},~{}\Psi_{H}=\frac{p(r_{+}+d-p^{2}/m)}{r_{+}^{2}-d^{2}-k^{2}+a^{2}},$ (11) $V=\frac{4}{3}r_{+}S_{BH},~{}~{}~{}\mathcal{P}=-\frac{3}{8\pi l^{2}}.$ (12) where those are physical mass, angular momentum, electric charge, magnetic charge, Hawking temperature, Bekenstein-Hawking entropy, angular velocity, electric potential, magnetic potential, volume and pressure. We can also consider the thermodynamic quantities on the cosmological horizon $r_{c}$. Those thermodynamic quantities on $r_{c}$ are given as follows $M_{c}=-\frac{m}{\Xi},~{}~{}~{}J_{c}=-\frac{ma}{\Xi},~{}~{}~{}Q_{c}=-\frac{q}{\Xi},~{}~{}~{}P_{c}=-\frac{p}{\Xi},\ $ (13) $T_{c}=\frac{r_{c}(2r_{c}^{2}-2d^{2}-2k^{2}+a^{2}-l^{2})+ml^{2}}{2\pi(r_{c}^{2}-d^{2}-k^{2}+a^{2})l^{2}},$ (14) $S_{c}=\frac{\pi}{\Xi}(r_{c}^{2}-d^{2}-k^{2}+a^{2}),~{}~{}~{}\Omega_{c}=\frac{a\Xi}{r_{c}^{2}-d^{2}-k^{2}+a^{2}},$ (15) $\Phi_{c}=\frac{q(r_{c}+d-p^{2}/m)}{r_{c}^{2}-d^{2}-k^{2}+a^{2}},~{}\Psi_{c}=\frac{p(r_{c}+d-p^{2}/m)}{r_{c}^{2}-d^{2}-k^{2}+a^{2}},$ (16) $V_{c}=\frac{4}{3}r_{c}S_{c},~{}~{}~{}\mathcal{P}_{c}=\frac{3}{8\pi l^{2}}.$ (17) These thermodynamic quantities are obtained by considering the black hole’s event horizon as the boundary [26]. This is in contrast with the previous thermodynamic quantities on the black hole’s event horizon where the cosmological horizon is considered as the boundary. For this black hole, we consider the cosmological constant as dynamical variable as a consequence of considering mass of the black hole as enthalpy of the spacetime [27, 28]. ## III Structure of Dyonic Kerr-Sen-de Sitter Spacetime In order to gain insight into the structure of dyonic KSdS spacetime, we explore the horizon solutions by plotting the “fin” diagram by slicing through the parameter space with 4 planes; (a) $p=0$, (b) $q=0$, (c) $p=1$, (d) $q=1$, for $a=0.5,l=10$ as shown in Fig. 1. The zeroes of metric function $\Delta/\varrho^{2}$ give four roots for the horizons, $\hat{r}=r_{i},~{}(i=1-4)$. We observe the continuity of the real part of horizon ${\rm Re}[r_{i}]$ connecting solutions of merging horizons. In addition, there is a ring singularity at $\varrho=0$ or $\hat{r}=\sqrt{d^{2}+k^{2}}=\displaystyle{\frac{p^{2}+q^{2}}{2m}}\equiv r_{s}$. For zero-charge spinning KSdS spacetime at fixed $a,l$, there is an extremal horizon where both inner (Cauchy) horizon $r_{-}$ and outer (black hole) horizon $r_{+}$ emerge at the critical $m$, above which $r_{-}$ and $r_{+}$ move away from one another. Finally at certain mass, $r_{+}$ will merge with cosmological horizon $r_{c}$ and the outer physical spacetime region disappears leaving the naked singularity at $\hat{r}=0$. When either electric ($q$) or magnetic ($p$) charge is turned on, the spacetime structure becomes more complicated as depicted in Fig. 1. Region I contains an outer horizon, naked singularity and cosmological horizon where $r_{+}<r_{s}<r_{c}$. Region II, IV, and VII are spacetime with naked singularity and cosmological horizon and no black holes. Region III and VI are spacetime with black holes and cosmological horizon, there exists singulairty hidden behind the Cauchy horizon. Region V intriguingly has black hole region with inner and outer horizon behind the naked singularity and cosmological horizon at the furthest. The boundary lines between each Region where extremal horizons occur are categorized as follows. 1. (i) I and II: $r_{-}=r_{+}<r_{s}$, 2. (ii) II and III+VI: $r_{-}=r_{+}>r_{s}$, 3. (iii) III+VI and IV: $r_{+}=r_{c}>r_{s}$, Nariai BH, 4. (iv) II and V: $r_{-}=r_{+}<r_{s}$, 5. (v) I and V: $r_{-}=0,r_{+}<r_{s}$, 6. (vi) IV and VII: $r_{1}=r_{4}\equiv r_{c}>r_{s}$, 7. (vii) II and IV, VII: $r_{1}=r_{2}\equiv r_{c}>r_{s}$. The Nariai limit of all cases occurs in the high mass region where $r_{+}=r_{c}$, the boundary line between region III+VI and IV. On the other hand, the boundary line between region II and III+VI represents extremal solutions where $r_{-}=r_{+}$. In comparison to the “shark fin” figure in Ref. [29], the dependence of $d,k$ on $m$ and $a$ distort the fin shape in the small mass and charge region in KSdS case. Interestingly, similar kinds of “flag” diagram where extremal structures are explored for generalized spherically symmetric metric are also presented in Ref. [30, 31]. | ---|--- (a) $p=0$ KSdS | (b) $q=0$ KSdS | (c) $p=1$ KSdS | (d) $q=1$ KSdS Figure 1: Horizon structure of KSdS spacetime for $a=0.5,l=10$, color region represents parameter space with the horizon $r_{i}\geq 0$ for $i=1-4$. Region I consists of an outer horizon behind a naked singularity and cosmological horizon. Region II, IV, and VII are spacetime with naked singularity and cosmological horizon. Region III and VI are black hole spacetime with singularity behind the Cauchy horizon, event horizon, and cosmological horizon. Region V contains naked singularity and cosmological horizon, and inner and outer horizon of black hole locating behind a naked singularity. ## IV Charged Rotating Nariai/CFT Correspondence The rotating Nariai solutions from the Kerr-dS and KNdS solutions haves been obtained in Ref. [9]. In this section, we will exhibit that this rotating solution in Nariai limit can also be expressed in the metric form with a fiber over AdS2 as the common near-horizon metric form in the Kerr/CFT correspondence [10, 11, 12, 13]. Another purpose of this section is to give a more detailed derivation of the central charge, temperature and entropy for solution in Nariai limit obtained from KSdS black hole. Furthermore, we will also study the massless scalar wave equation on this rotating background in Nariai limit which may possesses $SL(2,R)\times SL(2,R)$ isometry. ### IV.1 Geometry of Nariai Limit To find the rotating geometry in Nariai limit, we will assume two different limits which are Nariai limit and near-horizon limit. The parameter that parameterizes the Nariai limit is defined by $\varepsilon=\frac{r_{c}-r_{+}}{\lambda r_{0}}.$ (18) $r_{0}$ is a scaling constant that we define as $r_{0}^{2}=r_{+}^{2}-d^{2}-k^{2}+a^{2}$ for KSdS solution. When $\varepsilon\rightarrow 0$, we can obtain $r_{c}=r_{+}$. Furthermore, the near-horizon coordinate transformations are given by [22] $\hat{r}=r_{+}+\lambda r_{0}r,~{}~{}\hat{t}=\frac{r_{0}}{\lambda}t,~{}~{}\hat{\phi}=\phi+\frac{\Omega_{c}r_{0}}{\lambda}t.\ $ (19) The constant $\lambda$ is to parameterize the near-horizon limit. $\lambda\rightarrow 0$ denotes the near-horizon limit. Using those coordinate transformations, one can find the charged rotating solution in Nariai limit as follows $\displaystyle ds^{2}$ $\displaystyle=$ $\displaystyle\Gamma(\theta)\left(-r(r-\epsilon)dt^{2}+\frac{dr^{2}}{r(r-\epsilon)}+\alpha(\theta)d\theta^{2}\right)$ (20) $\displaystyle+\gamma(\theta)\left(d\phi+erdt\right)^{2},\ $ where the metric functions are given by $\Gamma(\theta)=\frac{\varrho_{+}^{2}}{\upsilon},~{}~{}\alpha(\theta)=\frac{\upsilon}{\Delta_{\theta}},~{}~{}\gamma(\theta)=\frac{r_{0}^{4}\Delta_{\theta}\sin^{2}\theta}{\varrho_{+}^{2}\Xi^{2}},\ $ (21) $\varrho_{+}^{2}=r_{+}^{2}-d^{2}-k^{2}+a^{2}\cos^{2}\theta,~{}~{}e=\frac{2ar_{+}\Xi}{r_{0}^{2}\upsilon}.$ (22) Note that we already approximate the function $\Delta$ as $\Delta\simeq\upsilon(r-r_{c})(r-r_{+}),\ $ (23) where $\upsilon=1-(6r_{+}^{2}-d^{2}-k^{2}+a^{2})/l^{2}$. The rotating solution in Nariai limit (20) is a fiber over AdS2. This fact can obviously be seen from the factor on time-like and radial coordinates of the metric in Nariai limit. One can also write the spacetime metric (20) in Poincaré coordinates by applying the following coordinate transformations [41] $\displaystyle t$ $\displaystyle=$ $\displaystyle-\frac{2}{\varepsilon}\text{log}\frac{y}{\sqrt{\tau^{2}y^{2}-1}},$ (24) $\displaystyle r$ $\displaystyle=$ $\displaystyle\frac{\varepsilon}{2}(1+\tau y),$ (25) $\displaystyle\phi$ $\displaystyle=$ $\displaystyle\varphi+\frac{1}{2}\text{log}\frac{\tau y+1}{\tau y-1}.\ $ (26) From those transformations, one can find $\displaystyle ds^{2}$ $\displaystyle=$ $\displaystyle\Gamma(\theta)\left(-y^{2}d\tau^{2}+\frac{dy^{2}}{y^{2}}+\alpha(\theta)d\theta^{2}\right)$ (27) $\displaystyle+\gamma(\theta)\left(d\varphi+eyd\tau\right)^{2},\ $ This metric form is obviously similar with the common near-horizon metric that is used in the Kerr/CFT correspondence. This metric possesses $SL(2,R)\times U(1)$ isometry generated by the following vector fields $\zeta_{0}=\partial_{\varphi},\ $ (28) which denote the rotational $U(1)$ isometry and $X_{1}=\partial_{\tau},~{}~{}~{}X_{2}=\tau\partial_{\tau}-y\partial_{y},\ $ (29) $X_{3}=\left(\frac{1}{2y^{2}}+\frac{\tau^{2}}{2}\right)\partial_{\tau}-\tau y\partial_{y}-\frac{e}{y}\partial_{\varphi},$ (30) denoting $SL(2,R)$ isometry. As we have mentioned in the beginning of this section, in Ref. [9] they manage to show another form of charged rotating Nariai geometry from dyonic KNdS solution that is a fiber over dS2 in Eq. (B.2) in their appendix, which is given by $\displaystyle ds^{2}$ $\displaystyle=$ $\displaystyle\Gamma(\theta)\left(-(1-r^{2})dt^{2}+\frac{dr^{2}}{1-r^{2}}+\alpha(\theta)d\theta^{2}\right)$ (31) $\displaystyle+\gamma(\theta)\left(d\phi+krdt\right)^{2}.\ $ For the detail forms of $\Gamma(\theta),\alpha(\theta),\gamma(\theta),k$, one can see in the reference that we have mentioned. This metric can be found by using the near-horizon coordinate transformations (2.13) in Ref. [9] then using the coordinate changes (2.18). In that paper, they use $\hat{r}=r_{c}-\lambda r_{c}r$ in Eq. (2.13), instead of using positive new radial coordinate, that will result in metric with a fiber over dS2. This is different with our result which is a fiber over AdS2. Hence, the charged rotating geometry in Nariai limit can be portrayed in both fibers over dS2 and AdS2, depending on the sign of the near-horizon radial coordinate transformation we choose. In the other words, we can also obtain the charged rotating solution of KSdS black hole in Nariai limit with a fiber over dS2 when we use $\hat{r}=r_{c}-\lambda r_{c}r$. ### IV.2 CFT Duals For rotating Nariai geometry in Ref. [9], they basically apply the similar Kerr/CFT method to study the relation between black hole’s thermodynamics macroscopically and microscopically using 2D CFT. They propose the similar fall-off conditions for the metric deviations although the spacetime structure contains dS2 slice, unlike in the common dual CFT which contains AdS2 slice. #### IV.2.1 Central charge In the Kerr/CFT correspondence, the entropy is assumed to originate from partition function of 2D CFT resulting in Cardy entropy formula. The main ingredients for this formula are the central charges and temperatures. The basic procedure to compute the central charges is to employ the approach of Brown and Henneaux [32] where the asymptotic symmetry group (ASG) needs to satisfy some certain boundary conditions. In this work, we consider the similar boundary conditions for the metric deviations as given in Ref. [7], $\displaystyle h_{\mu\nu}\sim\left(\begin{array}[]{cccc}\mathcal{O}(r^{2})&\mathcal{O}\left(\frac{1}{r^{2}}\right)&\mathcal{O}\left(\frac{1}{r}\right)&\mathcal{O}(1)\\\ &\mathcal{O}\left(\frac{1}{r^{3}}\right)&\mathcal{O}\left(\frac{1}{r^{2}}\right)&\mathcal{O}\left(\frac{1}{r}\right)\\\ &&\mathcal{O}\left(\frac{1}{r}\right)&\mathcal{O}\left(\frac{1}{r}\right)\\\ &&&\mathcal{O}(1)\end{array}\right).$ (36) in the basis $(t,r,\theta,\phi)$ for metric (20). These metric deviations are subleading with respect to the background metric. These boundary conditions are basically chosen to eliminate excitations above extremality. The asymptotic symmetry of the general black hole family includes diffeomorphisms $\xi$ that satisfy $\delta_{\xi}g_{\mu\nu}=\mathcal{L}_{\xi}g_{\mu\nu}=\xi^{\sigma}(\partial_{\sigma}g_{\mu\nu})+g_{\mu\sigma}(\partial_{\nu}\xi^{\sigma})+g_{\sigma\nu}(\partial_{\mu}\xi^{\sigma}),\ $ (37) where the metric deviation is denoted by $\delta_{\xi}g_{\mu\nu}=h_{\mu\nu}$. The most general diffeomorphism symmetry that preserves such boundary conditions (36) in the asymptotic infinity is generated by the following Killing vector field $\displaystyle\zeta$ $\displaystyle=$ $\displaystyle\left[c_{t}+\mathcal{O}\left(r^{-3}\right)\right]\partial_{t}+\left[-r\epsilon^{\prime}(\phi)+\mathcal{O}(1)\right]\partial_{r}$ (38) $\displaystyle+\mathcal{O}\left(r^{-1}\right)\partial_{\theta}+\left[\epsilon(\phi)+\mathcal{O}\left(r^{-2}\right)\right]\partial_{\phi},\ $ where $c_{t}$ is an arbitrary constant and the prime $(^{\prime})$ denotes the derivative with respect to $\phi$. This ASG contains one copy of the conformal group of the circle which is generated by $\displaystyle\zeta_{\epsilon}=\epsilon(\phi)\partial_{\phi}-r\epsilon^{\prime}(\phi)\partial_{r},$ (39) that will be the part of the near-horizon metric in Nariai limit. We know that the azimuthal coordinate is periodic under the rotation $\phi\sim\phi+2\pi$. Hence, we may define $\epsilon_{\hat{n}}=-e^{-i\hat{n}\phi}$ and $\zeta_{\epsilon}=\zeta_{\epsilon}(\epsilon_{\hat{n}})$. By the Lie bracket, the symmetry generator (39) satisfies the Witt algebra, $\displaystyle i[\zeta_{\hat{m}},\zeta_{\hat{n}}]_{LB}=({\hat{m}}-{\hat{n}})\zeta_{{\hat{m}}+{\hat{n}}}.\ $ (40) ${\hat{m}},{\hat{n}}$ are just integers. The zero mode is the azimuthal translation or $U(1)$ isometry, $\zeta_{0}=-\partial_{\phi}$. The associated conserved charge is [33] $\displaystyle Q_{\xi}=\frac{1}{8\pi}\int_{\partial\Sigma}k^{g}_{\zeta}[h;g].$ (41) This given integral is over the boundary of a spatial slice. The contribution of the metric tensor on the central charge is given explicitly by $\displaystyle k^{g}_{\zeta}[h;g]$ $\displaystyle=$ $\displaystyle-\frac{1}{4}\epsilon_{\rho\sigma\mu\nu}\bigg{\\{}\zeta^{\nu}D^{\mu}h-\zeta^{\nu}D_{\lambda}h^{\mu\lambda}$ (42) $\displaystyle+\frac{h}{2}D^{\nu}\zeta^{\mu}-h^{\nu\lambda}D_{\lambda}\zeta^{\mu}+\zeta_{\lambda}D^{\nu}h^{\mu\lambda}$ $\displaystyle\left.+\frac{h^{\lambda\nu}}{2}\left(D^{\mu}\zeta_{\lambda}+D_{\lambda}\zeta^{\mu}\right)\right\\}dx^{\rho}\wedge dx^{\sigma}.\ $ We should note that the last two terms in Eq. (42) vanish for an exact Killing vector and an exact symmetry, respectively. The charge $Q_{\zeta}$ generates symmetry through the Dirac brackets. The ASG possesses algebra which is given by the Dirac bracket algebra of the following charges [33] $\displaystyle\\{Q_{\zeta},Q_{\bar{\zeta}}\\}_{DB}$ $\displaystyle=$ $\displaystyle\frac{1}{8\pi}\int k^{g}_{\zeta}\left[\mathcal{L}_{\bar{\zeta}}g;g\right]$ (43) $\displaystyle=$ $\displaystyle Q_{[\zeta,\bar{\zeta}]}+\frac{1}{8\pi}\int k^{g}_{\zeta}\left[\mathcal{L}_{\bar{\zeta}}\bar{g};\bar{g}\right].\ \ $ By using (39) and upon quantization, we can transform the Dirac bracket algebra into a commutation relation that allows us to interpret the classical central charge as a quantum central charge of the dual CFT. For the quantization, we replace the classical charges $Q_{\zeta}$ by their quantum counterpart $L_{{\hat{n}}}$, $\displaystyle Q_{\zeta}\equiv L_{{\hat{n}}}-x\delta_{{\hat{n}},0},$ (44) so that we obtain the conserved charges algebra in quantum form, such that $\displaystyle\left[L_{\hat{m}},L_{\hat{n}}\right]=({\hat{m}}-{\hat{n}})L_{{\hat{m}}+{\hat{n}}}+\frac{c_{L}}{12}\hat{m}({\hat{m}}^{2}-1)\delta_{{\hat{m}}+{\hat{n}},0}.$ (45) $x$ is a free parameter to scale the central charge on the term $\sim\hat{m}$ (see Appendix B). The relation (40) and (44) have been used to find the Virasoro algebra above. For the charged rotating solution in Nariai limit obtained from KSdS solution, the corresponding central charge is then $\displaystyle c_{L}$ $\displaystyle=$ $\displaystyle\frac{3e}{\Xi}\int^{\pi}_{0}d\theta\sqrt{\Gamma(\theta)\alpha(\theta)\gamma(\theta)}$ (46) $\displaystyle=$ $\displaystyle\frac{12ar_{+}}{1-\frac{6r_{+}^{2}-d^{2}-k^{2}+a^{2}}{l^{2}}}.\ $ The main difference between this central charge and the Nariai solution from KNdS black hole [9] is the presence of dilaton and axion charges. #### IV.2.2 Temperature In the Nariai limit where $r_{c}\rightarrow r_{+}$, the variation of the entropy can be expressed completely in terms of a variation of the angular momentum, electric charge, magnetic charge, and the pressure. It takes the following form $\displaystyle dS_{BH}=\frac{dJ}{T_{L}}+\frac{dQ}{T_{q}}+\frac{dP}{T_{p}}+\frac{d\mathcal{P}}{T_{\mathcal{P}}},\ $ (47) where $T_{L},T_{q},T_{p},T_{\mathcal{P}}$ are the left-moving temperature, conjugate temperature of electric charge, conjugate temperature of magnetic charge, and conjugate temperature of pressure, respectively. Hence, we can derive that $\displaystyle T_{L}=-\frac{\partial T_{H}/\partial r_{+}}{\partial\Omega_{H}/\partial r_{+}}\bigg{|}_{r_{+}=r_{c}}\frac{\upsilon(r_{+}^{2}-d^{2}-k^{2}+a^{2})}{4\pi ar_{+}\Xi}.\ $ (48) Since we already take $r_{c}\rightarrow r_{+}$, so the right-moving temperature is equal to zero which is proportional to the Hawking temperature on the cosmological horizon. However, in the next section, we will see that $T_{R}\sim\varepsilon$ generally which is fairly small. #### IV.2.3 Cardy Entropy We have calculated the central extension and the corresponding temperature which are the main ingredients in Cardy entropy. So, the main upshot of this section is to provide the derivation of Cardy entropy [7] for the charged rotating solution in Nariai limit. Firstly, Cardy entropy formula is given as follows $S_{CFT}=\frac{\pi^{2}}{3}\left(c_{L}T_{L}+c_{R}T_{R}\right).$ (49) The right-moving part is exactly zero, so only left-moving part contributes. On the cosmological horizon, by using above formula, we find the following entropy for charged rotating solution in Nariai limit $S_{CFT}=\frac{\pi}{\Xi}(r_{+}^{2}-d^{2}-k^{2}+a^{2})=S_{BH}=S_{c}.\ $ (50) This result exhibits that the Kerr/CFT correspondence can be applied for charged rotating solution in Nariai limit obtained from KSdS black hole which precisely possesses $SL(2,R)\times U(1)$ isometry like the near-horizon extremal black hole. The entropy on the event horizon is now similar with the entropy on the cosmological horizon since we take $r_{c}\rightarrow r_{+}$. So, for an observer in the visible region, the total entropy in Nariai limit is [34] $S_{tot}=S_{c}+S_{BH}=2S_{c}.\ $ (51) Note that this is the total entropy for charged rotating solution in Nariai limit with non-vanishing dyonic charge. So, when $p=0$, it is just the total entropy for electrically charged one. Furthermore, when all electromagnetic charges vanish, it will be the total entropy of rotating Nariai solution obtained from Kerr-dS solution [9]. For the entropy on the event horizon for extremal KSdS black hole, we calculate in Appendix A where we also prove the result in Ref. [45] on KSdS solution that is mentioned therein. In case of asymptotically AdS spacetime, it has been carried out in Ref. [45]. For vanishing cosmological constant and $p$, this result recovers the result in Ref. [47]. ## V Scattering of Scalar Field and Hidden Conformal Symmetry In this section, we will further study the scattering of massless scalar field on the background of charged rotating Nariai solution in $J$ (angular momentum)- and $Q$ (electric charge)-pictures. We will also compute the absorption cross-section as well as the real-time correlator on the corresponding background. Yet, first we will show that there are hidden conformal symmetries on this solution in Nariai limit. Lastly, we also consider $Q$-picture where the scalar probe is assumed to be electrically charged. ### V.1 Scalar wave equation To explore the hidden conformal symmetries in $J$-picture, we assume a massless neutral scalar field in the background of charged rotating solution in Nariai limit. The massless scalar wave equation for the scalar probe is given by $\nabla_{\alpha}\nabla^{\alpha}\Phi=0.$ (52) We know that the charged rotating solution in Nariai limit is conserved under time-like and azimuthal translations. Hence, we can separate the coordinates in the scalar wave equation as $\Phi(\hat{t},\hat{r},\theta,\hat{\phi})=\mathrm{e}^{-i\omega\hat{t}+in\hat{\phi}}R(\hat{r})S(\theta),$ (53) where $\omega$ and $n$ are the asymptotic energy and angular momentum of the scalar field. Plugging Eq. (53) into Eq. (52), leads to two differential equations i.e., the angular $S(\theta)$ and radial $R(\hat{r})$ wave functions, $\displaystyle\left[\frac{1}{\sin\theta}\partial_{\theta}(\sin\theta\partial_{\theta})-\frac{n^{2}\Xi^{2}}{\Delta_{\theta}\sin^{2}\theta}\right]S(\theta)$ $\displaystyle+\left[\frac{2an\omega\Xi-a^{2}\omega^{2}\sin^{2}\theta}{\Delta_{\theta}}\right]S(\theta)=-K_{l}S(\theta),\ $ (54) $\bigg{[}\partial_{\hat{r}}(\Delta\partial_{\hat{r}})+\frac{\left[(\hat{r}^{2}-d^{2}-k^{2}+a^{2})\omega- an\Xi\right]^{2}}{\Delta}-K_{l}\bigg{]}R(\hat{r})=0,\ $ (55) where $K_{l}$ is the separation constant. Note that we still use the generic KSdS metric (1). These equations are identical with the wave equations in the Kerr-dS solutions. We do not consider the backreaction of the scalar field in this case. Firstly, we will study the radial equation (55) at the near region, which is defined by $\omega\hat{r}\ll 1$. Furthermore, we assume that the frequency of the scalar field to be very small $\omega M\ll 1$. Consequently, we also impose $\omega a\ll 1,\,\omega d\ll 1,$ and $\omega k\ll 1$. On the other hand, we also have $\omega q\ll 1,$ and $\omega p\ll 1$. Since we have approximated $\Delta$ as (23), we may set up the radial equation in a suitable form for exploring its hidden conformal symmetry. The radial equation (55) reduces to $\displaystyle\partial_{\hat{r}}\left[(\hat{r}-r_{c})(\hat{r}-r_{+})\partial_{\hat{r}}\right]R(\hat{r})+\frac{r_{c}-r_{+}}{\hat{r}-r_{c}}AR(\hat{r})=0,$ $\displaystyle+\left[\frac{r_{c}-r_{+}}{\hat{r}-r_{+}}B+C\right]R(\hat{r})=0,$ (56) where $A=\frac{\left[(r_{c}^{2}-d^{2}-k^{2}+a^{2})\omega- an\Xi\right]^{2}}{\upsilon^{2}(r_{c}-r_{+})^{2}},\ $ (57) $B=-\frac{\left[(r_{+}^{2}-d^{2}-k^{2}+a^{2})\omega- an\Xi\right]^{2}}{\upsilon^{2}(r_{c}-r_{+})^{2}},~{}~{}~{}C=\frac{-K_{l}}{\upsilon}.\ $ (58) It is worth noting that the angular wave equation does not have $SL(2,R)\times SL(2,R)$ isometry, yet $SU(2)\times SU(2)$ isometry [39]. Hence, we will not consider to investigate further the angular part. In order to reveal the hidden symmetries of Eq. (56), we need to perform the following conformal coordinate transformations [14, 41] $\displaystyle\omega^{c}=\sqrt{\frac{\hat{r}-r_{c}}{\hat{r}-r_{+}}}e^{2\pi T_{R}\hat{\phi}+2n_{R}\hat{t}},$ $\displaystyle\omega^{+}=\sqrt{\frac{\hat{r}-r_{c}}{\hat{r}-r_{+}}}e^{2\pi T_{L}\hat{\phi}+2n_{L}\hat{t}},$ $\displaystyle\hat{y}=\sqrt{\frac{r_{c}-r_{+}}{\hat{r}-r_{+}}}e^{\pi(T_{L}+T_{R})\hat{\phi}+(n_{L}+n_{R})\hat{t}}.\ $ (59) Then we may define three locally conformal operators in terms of the new conformal coordinates $\omega^{c},\,\omega^{+}$ and $\hat{y}$ as $\displaystyle H_{1}=i\partial_{c},$ $\displaystyle H_{-1}=i\left(\omega^{c2}\partial_{c}+\omega^{c}\hat{y}\partial_{\hat{y}}-\hat{y}^{2}\partial_{+}\right),$ $\displaystyle H_{0}=i\left(\omega^{c}\partial_{c}+\frac{1}{2}\hat{y}\partial_{\hat{y}}\right),$ (60) as well as $\displaystyle\bar{H}_{1}=i\partial_{+},$ $\displaystyle\bar{H}_{-1}=i\left(\omega^{+2}\partial_{+}+\omega^{+}\hat{y}\partial_{\hat{y}}-\hat{y}^{2}\partial_{c}\right),$ $\displaystyle\bar{H}_{0}=i\left(\omega^{+}\partial_{+}+\frac{1}{2}\hat{y}\partial_{\hat{y}}\right).$ (61) Note that we have used $\partial_{c}=\partial/\partial\omega^{c},\partial_{+}=\partial/\partial\omega^{+}$. The set of operators (V.1) satisfies the $SL(2,R)$ Lie algebra $\displaystyle\left[H_{0},H_{\pm 1}\right]=\mp iH_{\pm 1},~{}~{}~{}\left[H_{-1},H_{1}\right]=-2iH_{0},$ (62) while a similar $SL(2,R)$ algebra exists for the set of operators (V.1). From every set of operators, we can construct the quadratic Casimir operator as given by $\displaystyle\mathcal{H}^{2}$ $\displaystyle=$ $\displaystyle\bar{\mathcal{H}}^{2}=-H_{0}^{2}+\frac{1}{2}(H_{1}H_{-1}+H_{-1}H_{1})$ (63) $\displaystyle=$ $\displaystyle\frac{1}{4}(\hat{y}^{2}\partial_{\hat{y}}^{2}-y\partial_{\hat{y}})+\hat{y}^{2}\partial_{c}\partial_{+}.$ We have found that the radial equation (V.1) could be re-written in terms of the $SL(2,R)$ quadratic Casimir operator as $\mathcal{H}^{2}R(r)=\bar{\mathcal{H}}^{2}R(r)=-CR(r)$ where, in $J$-picture, we should identify the constants as $n_{L}=-\frac{\upsilon}{2(r_{c}+r_{+})},~{}~{}~{}n_{R}=0,$ (64) $T_{L}=\frac{\upsilon(r_{c}^{2}+r_{+}^{2}-2d^{2}-2k^{2}+2a^{2})}{4\pi a(r_{c}+r_{+})\Xi},~{}~{}~{}T_{R}=\frac{\upsilon(r_{c}-r_{+})}{4\pi a\Xi}.$ (65) $T_{R,L}$ are identified as the CFT temperatures that emerge as a result of the spontaneously symmetry breaking of the partition function on $SL(2,R)\times SL(2,R)$ theory to the partition function of $U(1)\times U(1)$ CFT. One can see that the periodic identification of the azimuthal coordinate $\hat{\phi}\sim\hat{\phi}+2\pi$ causes the $SL(2,R)\times SL(2,R)$ symmetry to spontaneously break down to $U(1)\times U(1)$ symmetry by temperatures $T_{R},T_{L}$, $\omega^{c}\sim e^{4\pi^{2}T_{R}\omega^{c}},~{}\omega^{+}\sim e^{4\pi^{2}T_{L}\omega^{+}},~{}\hat{y}\sim e^{2\pi^{2}(T_{L}+T_{L})\hat{y}}.$ (66) This identification is generated by the $SL(2,R)\times SL(2,R)$ group element, $e^{-i4\pi^{2}T_{R}H_{0}-i4\pi^{2}T_{L}\bar{H}_{0}}$. After finding the temperatures (65), we can also compute conjugate charges $E_{L},E_{R}$ for this charged rotating solution. These conjugate charges can be obtained from the entropy via [41] $\delta S_{CFT}=\frac{\delta E_{L}}{T_{L}}+\frac{\delta E_{R}}{T_{R}}.$ (67) In order to compute the conjugate charges, we consider the first law of thermodynamics for the charged rotating solution in Nariai limit (7) with the quantities (13)-(17). Since we consider the neutral scalar field, we have $\delta Q=\delta P=\delta\mathcal{P}=0$. Now we can find the conjugate charges via $\delta S_{BH}=\delta S_{CFT}$. Using the identification $\delta M$ as $\omega$ and $\delta J$ as $n$ yields to the identification of $\delta E_{L,R}$ as $\omega_{L,R}$. Hence, in $J$-picture, we obtain that $\displaystyle\omega_{L}=\frac{r_{c}^{2}+r_{+}^{2}-2d^{2}-2k^{2}+2a^{2}}{2a\Xi}\omega,~{}\omega_{R}=\omega_{L}-n.\ $ (68) So, we find the left and right frequencies of 2D CFT for the charged rotating solution in Nariai limit for KSdS black hole in $J$-picture. For the study of the scalar field in KSAdS black hole, one can see in Ref. [18]. ### V.2 Absorption cross-section in $J$-picture In the previous section, we have derived one remarkable result of the dual CFT, i.e. the entropy. Another realization of this duality is the equivalence of the absorption cross-section of the scalar probe. In investigating the absorption cross-section, we need to consider the asymptotic region. Since we have approximated $\Delta$ in near-horizon region, this approximation in the asymptotic region will break down, except we consider the Nariai limit. This is similar in the near-extremal case of the scalar wave equation [15, 16, 20, 38, 42]. Beside the Nariai limit on (18), we consider the near-horizon coordinate transformations (19). We also consider the scalar probe with frequencies around the superradiant bound $\omega=n\Omega_{H}+\hat{\omega}\frac{\lambda}{r_{0}},$ (69) where $\Omega_{H}$ are given by (10). We can re-write the radial equation (V.1) by $\left(\partial_{r}\left[r\left(r-\varepsilon\right)\right]\partial_{r}+\frac{A_{s}}{r-\varepsilon}+\frac{B_{s}}{r}+C_{s}\right)R(r)=0,$ (70) where $\displaystyle A_{s}=\frac{(\hat{\omega}+2nr_{+}\varepsilon\Omega_{H})^{2}}{\upsilon^{2}\varepsilon},~{}~{}~{}B_{s}=-\frac{\hat{\omega}^{2}}{\upsilon^{2}\varepsilon},~{}~{}~{}C_{s}=C_{s}(\hat{\omega}),\ $ and $C_{s}$ is the new separation constant that is dependent on the frequency $\hat{\omega}$ that can be obtained by solving the angular wave equation. We then apply the coordinate transformation $z=(r-\varepsilon)/r$. In this new radial coordinate, the radial equation (70) becomes $\left[z(1-z)\partial_{z}^{2}+(1-z)\partial_{z}+\frac{\hat{A_{s}}}{z}+\hat{B_{s}}+\frac{C_{s}}{1-z}\right]R(z)=0,$ (71) where $\displaystyle\hat{A_{s}}=\frac{(\hat{\omega}+2nr_{+}\varepsilon\Omega_{H})^{2}}{\upsilon^{2}\varepsilon^{2}},~{}~{}~{}\hat{B_{s}}=-\frac{\hat{\omega}^{2}}{\upsilon^{2}\varepsilon^{2}}.\ $ The ingoing solution to differential equation (71) is also given by hypergeometric function $R(z)=z^{-i\sqrt{\hat{A_{s}}}}(1-z)^{1+h}~{}_{2}F_{1}(a_{s},b_{s};c_{s};z),\ $ (72) with the parameters $a_{s}=1+h-i(\sqrt{\hat{A}_{s}}+\sqrt{-\hat{B}_{s}})$, $b_{s}=1+h-i(\sqrt{\hat{A}_{s}}-\sqrt{-\hat{B}_{S}})$, and $c_{s}=1-2i\sqrt{\hat{A}_{s}}$. For this superradiant case, the relation between $h$ and $C_{s}$ is given by $h=\frac{1}{2}\left(-1+\sqrt{1-4C_{s}}\right).$ (73) For the asymptotic region of the radial coordinate $\hat{r}$ (or equivalently $r\gg\varepsilon$), where $z\sim 1$, the solution (72) reduces to $\displaystyle R(y)\sim D_{0}r^{h}+D_{1}r^{-1-h},\ $ (74) where $D_{0}=\frac{\Gamma(c_{s})\Gamma(1+2h)}{\Gamma(a_{s})\Gamma(b_{s})},~{}~{}~{}D_{1}=\frac{\Gamma(c_{s})\Gamma(-1-2h)}{\Gamma(c_{s}-a_{s})\Gamma(c_{s}-b_{s})}.\ $ (75) The conformal weight of the scalar field is equal to $h+1$. For the coefficient (75), we find the absorption cross-section of the scalar fields as $P_{abs}\sim\left|D_{0}\right|^{-2}\sim\sinh\left({2\pi\hat{A}_{s}^{1/2}}\right){\left|{\Gamma\left(a_{s}\right)}{\Gamma\left(b_{s}\right)}\right|^{2}}.$ (76) To be more supporting the correspondence between the charged rotating solution in Nariai limit and 2D CFT, we show that the absorption cross-section for the scalar fields (76) can be obtained from the absorption cross-section in a 2D CFT [14] $\displaystyle P_{abs}$ $\displaystyle\sim$ $\displaystyle{T_{L}}^{2h_{L}-1}{T_{R}}^{2h_{R}-1}\sinh\left({\frac{{\omega_{L}}}{{2{T_{L}}}}+\frac{{\omega_{R}}}{{2{T_{R}}}}}\right)$ (77) $\displaystyle\times$ $\displaystyle\left|{\Gamma\left({h_{L}+i\frac{{\omega_{L}}}{{2\pi{T_{L}}}}}\right)}{\Gamma\left({h_{R}+i\frac{{\omega_{R}}}{{2\pi{T_{R}}}}}\right)}\right|^{2}.\ $ The agreement between (76) and (77) can be shown when we choose proper left and right frequencies $\omega_{L},\omega_{R}$. Since in the previous subsection we have found the CFT temperatures and frequencies, we can directly calculate the similar quantities for the frequency of the scalar field in the near-superradiant bound by neglecting the second-order and higher corrections from $\lambda$. Hence, in terms of Nariai limit parameter and in near- superradiant bound, we obtain $T_{L}=\frac{\upsilon}{4\pi\Omega_{H}r_{+}},~{}~{}~{}T_{R}=\frac{\upsilon\lambda r_{0}}{4\pi a\Xi}\varepsilon.$ (78) When we take the Nariai limit $\varepsilon\rightarrow 0$ the right-moving temperature will vanish. Then we recover the result for $T_{R}$ and $T_{L}$ in the previous section. The frequencies are now $\omega_{L}=n,~{}~{}~{}\omega_{R}=\frac{r_{0}}{a\Xi}\left(\hat{\omega}+nr_{+}\varepsilon\Omega_{c}\right),\ $ (79) The conformal weights in CFT are given by $h_{L}=h_{R}=h+1.\ $ (80) This gives rise to another nontrivial evidence to support the Kerr/CFT correspondence for black holes in Nariai limit. ### V.3 Real-time correlator in $J$-picture Furthermore, one can also compute the real-time correlator. The asymptotic behaviors of the scalar field with ingoing boundary condition on the background of charged rotating solution in Nariai limit (74) indicate that two coefficients possess different roles where $D_{1}$ indicates the response and $D_{0}$ indicates the source. Hence, the two-point retarded correlator is simply [43, 44] $G_{R}\sim\frac{D_{1}}{D_{0}}=\frac{\Gamma(-1-2h)}{\Gamma(1+2h)}\frac{\Gamma(a_{s})\Gamma(b_{s})}{\Gamma(c_{s}-a_{s})\Gamma(c_{s}-b_{s})}.\ $ (81) From Eq. (81), it is easy to check that $G_{R}\sim\frac{\Gamma(h_{L}-i\frac{\omega_{L}}{2T_{L}})\Gamma(h_{R}-i\frac{\omega_{R}}{2T_{R}})}{\Gamma(1-h_{L}-i\frac{\omega_{L}}{2T_{L}})\Gamma(1-h_{R}-i\frac{\omega_{R}}{2T_{R}})}.$ (82) Then by using the relation $\Gamma(z)\Gamma(1-z)=\pi/\sin(\pi z)$, we can write above two-point retarded correlator into $\displaystyle G_{R}$ $\displaystyle\sim$ $\displaystyle\sin\left(\pi h_{L}+i\frac{\omega_{L}}{2T_{L}}\right)\sin\left(\pi h_{R}+i\frac{\omega_{R}}{2T_{R}}\right)$ (83) $\displaystyle\bigg{|}\Gamma\left(h_{L}+i\frac{\omega_{L}}{2T_{L}}\right)\Gamma\left(h_{R}+i\frac{\omega_{R}}{2T_{R}}\right)\bigg{|}^{2}.$ Since $h_{L},h_{R}$ are integers, so we have $\displaystyle\sin\left(\pi h_{L}+i\frac{\omega_{L}}{2T_{L}}\right)\sin\left(\pi h_{R}+i\frac{\omega_{R}}{2T_{R}}\right)$ $\displaystyle=(-1)^{h_{L}+h_{R}}\sin\left(i\frac{\omega_{L}}{2T_{L}}\right)\sin\left(i\frac{\omega_{R}}{2T_{R}}\right).\ $ (84) In CFT, the Euclidean correlator is given by $\displaystyle G_{E}(\omega_{EL},\omega_{ER})\sim T_{L}^{2h_{L}-1}T_{R}^{2h_{R}-1}e^{i\omega_{EL}/2T_{L}}e^{i\omega_{ER}/2T_{R}}$ $\displaystyle\times\bigg{|}\Gamma\left(h_{L}+\frac{\omega_{EL}}{2T_{L}}\right)\Gamma\left(h_{R}+\frac{\omega_{EL}}{2T_{R}}\right)\bigg{|}^{2}.$ (85) Where we can define the Euclidean frequencies $\omega_{EL}=i\omega_{L}$, and $\omega_{ER}=i\omega_{R}$. It is important noting that $G_{E}$ corresponds to the values of retarded correlator $G_{R}$. The retarded Green function $G_{R}$ is analytic on the upper half complex $\omega_{L,R}$-plane. The value of $G_{R}$ along the positive imaginary $\omega_{L,R}$-axis gives the following correlator $G_{E}(\omega_{EL},\omega_{ER})=G_{R}(i\omega_{L},i\omega_{R}),~{}~{}~{}\omega_{EL,ER}>0.$ (86) At finite temperature, $\omega_{EL,ER}$ should take discrete values of the Matsubara frequencies, given by $\omega_{EL}=2\pi m_{L}T_{L},~{}~{}~{}\omega_{ER}=2\pi m_{R}T_{R},$ (87) where $m_{L}$ and $m_{R}$ are half integers for fermionic modes and integers for bosonic modes. At these certain frequencies, the gravity computation for correlator (83) matches precisely with CFT result trough Eq. (86) up to a numerical normalization factor. ### V.4 Hidden conformal symmetry in $Q$-Picture In the previous subsections, we have shown the existence of hidden conformal symmetry on the $J$-picture with the background charged rotating black hole in Nariai limit. In this subsection, we will further study the scattering of massless scalar field on the background of charged rotating black hole solution in Nariai limit in $Q$-picture. On the other hand, we will consider charged scalar field on the background solution. For some black hole backgrounds, $Q$-picture has been explored well to exhibit the existence of the conformal symmetries [15, 16, 37, 38, 35, 36]. Nonetheless, the $Q$-picture is not well-defined for non-extremal Kerr-Sen black holes [35]. In this section, we will investigate the $Q$-picture when the cosmological constant exists. Since we will only consider $Q$-picture, for simplicity, we may assume that $p=0$, so that $k=0$. The hidden conformal symmetries can be explored by assuming a massless charged scalar probe in the background of charged rotating black hole solution in Nariai limit. It is given by $(\nabla_{\alpha}-igA_{\alpha})(\nabla^{\alpha}-igA^{\alpha})\Phi=0,$ (88) where $g$ is the electric charge of the scalar probe. In $Q$-picture, in addition to the modes of asymptotic energy and angular momentum, we need to add the charge $g$ as the eigenvalue of the operator $\partial_{\chi}$ that denotes the additional internal direction $\chi$ to four dimensions. Actually, $g$ has a natural geometrical interpretation as the radius of extra circle when the the black hole solution is considered to be embedded into 5D. In fact, this coordinate possesses similar $U(1)$ symmetry like the azimuthal coordinate. However, so far, we do not find that 4D Kerr-Sen(-dS) black hole can be uplifted to 5D solution. It is different with Kerr-Newman(-dS) black hole that can be uplifted to 5D [22]. Here, we will just first assume the fifth coordinate to reveal the conformal symmetries of the charged probe in $Q$-picture. The ansatz in $Q$-picture is given by $\Phi(\hat{t},\hat{r},\theta,\hat{\phi},\chi)=\mathrm{e}^{-i\omega\hat{t}+in\hat{\phi}+ig\chi}R(\hat{r})S(\theta).$ (89) Beside the low-frequency assumption, we need to assume also that the probe’s charge to be very small ($gq\ll 1$). By plugging Eq. (89) into Eq. (88) and assuming small charge and low frequency, in the near-horizon region, we can find the radial equation $\displaystyle\partial_{\hat{r}}\left[(\hat{r}-r_{c})(\hat{r}-r_{+})\partial_{\hat{r}}\right]R(\hat{r})+\frac{r_{c}-r_{+}}{\hat{r}-r_{c}}AR(\hat{r})=0,$ $\displaystyle+\left[\frac{r_{c}-r_{+}}{\hat{r}-r_{+}}B+C\right]R(\hat{r})=0,$ (90) where $B=-\frac{\left[(r_{+}^{2}-d^{2}+a^{2})\omega-an\Xi- gq(r_{+}+d)\right]^{2}}{\upsilon^{2}(r_{c}-r_{+})^{2}},\ $ (91) $A=\frac{\left[(r_{c}^{2}-d^{2}+a^{2})\omega-an\Xi- gq(r_{c}+d)\right]^{2}}{\upsilon^{2}(r_{c}-r_{+})^{2}},~{}~{}~{}C=\frac{-K_{l}}{\upsilon}.\ $ (92) In $Q$-picture, we take $n=0$. As the previous computation for $J$-picture, we can identify that $n_{L}=-\frac{\upsilon(r_{c}+r_{+})}{4\left[(r_{c}+d)(r_{+}+d)-a^{2}\right]},\ $ (93) $n_{R}=-\frac{\upsilon(r_{c}-r_{+})}{4\left[(r_{c}+d)(r_{+}+d)-a^{2}\right]},$ (94) $T_{L}=\frac{\upsilon(r_{c}^{2}+r_{+}^{2}-2d^{2}+2a^{2})}{4\pi g\left[(r_{c}+d)(r_{+}+d)-a^{2}\right]},\ $ (95) $T_{R}=\frac{\upsilon(r^{2}_{c}-r^{2}_{+})}{4\pi g\left[(r_{c}+d)(r_{+}+d)-a^{2}\right]}.$ (96) The temperatures (96) are irregular. We cannot check directly that $(r_{c}+d)(r_{+}+d)-a^{2}=0$ because we do not have the explicit analytical form of $r_{c},r_{+}$. However, we can easily check that the form of CFT temperatures is similar with the case when the cosmological constant vanishes. So, when $1/l^{2}=0$, one can obtain $(r_{+}+d)(r_{-}+d)-a^{2}=0,\ $ (97) where $r_{\pm}=M\pm\sqrt{M^{2}+d^{2}-a^{2}-q^{2}}.$ (98) This is the reason that $Q$\- picture is not well-defined for Kerr-Sen black hole. With the similar form of temperatures, we conclude also that $Q$-picture for Kerr-Sen-dS black hole in Nariai limit is not well-defined. ## VI Hidden Conformal Symmetry on Nariai limit for Kerr-Newman-de Sitter solution In the previous sections, we have carried out the calculation of the dual CFT for charged rotating black hole in Nariai limit obtained from KSdS solution and also the hidden conformal symmetry of the scalar probe in the black hole’s background. In this section, we will study the scattering of scalar field on Nariai limit for KNdS solution both in $J$\- and $Q$-pictures. Yet, we will revisit the computation of the entropy from CFT for KNdS solution in Nariai limit. The KNdS black hole solution is given by the following metric [22, 23] $\displaystyle ds^{2}=-\frac{\Delta}{\varrho^{2}}X^{2}+\frac{\varrho^{2}}{\Delta}d\hat{r}^{2}+\frac{\varrho^{2}}{\Delta_{\theta}}d\theta^{2}+\frac{\Delta_{\theta}\sin^{2}\theta}{\varrho^{2}}Y^{2},\ $ (99) where $X=d\hat{t}-\frac{a\sin^{2}\theta}{\Xi}d\hat{\phi},~{}~{}~{}Y=ad\hat{t}-\frac{(\hat{r}^{2}+a^{2})}{\Xi}d\hat{\phi},\ $ (100) $\Delta_{\theta}=1+\frac{a^{2}}{l^{2}}\cos^{2}\theta,~{}~{}\Xi=1+\frac{a^{2}}{l^{2}},~{}~{}~{}\varrho^{2}=\hat{r}^{2}+a^{2}\cos^{2}\theta,\\\ $ (101) $\Delta=(\hat{r}^{2}+a^{2})\left(1-\frac{\hat{r}^{2}}{l^{2}}\right)-2m\hat{r}+p^{2}+q^{2}.\ $ (102) The parameters $a$, $m$, $p$, $q$, and $l$ are spin, mass, magnetic charge, electric charge, and de Sitter radius, respectively. The electromagnetic potential and its dual are given by $\displaystyle\textbf{A}=-\frac{q\hat{r}}{\varrho^{2}}X-\frac{p\cos\theta}{\varrho^{2}}Y,~{}~{}~{}\textbf{B}=-\frac{p\hat{r}}{\varrho^{2}}X+\frac{q\cos\theta}{\varrho^{2}}Y,$ (103) At the event horizon, the thermodynamic quantities are given by [22, 23] $M=\frac{m}{\Xi},~{}~{}~{}J=\frac{ma}{\Xi},~{}~{}~{}Q=\frac{q}{\Xi},~{}~{}~{}P=\frac{p}{\Xi},$ (104) $T_{H}=\frac{r_{+}(l^{2}-2r_{+}^{2}-a^{2})-ml^{2}}{2\pi(r_{+}^{2}+a^{2})l^{2}},$ (105) $S_{BH}=\frac{\pi}{\Xi}(r_{+}^{2}+a^{2}),$ (106) $\Omega_{H}=\frac{a\Xi}{r_{+}^{2}+a^{2}},~{}~{}\Phi_{H}=\frac{qr_{+}}{r_{+}^{2}+a^{2}},~{}~{}\Psi_{H}=\frac{pr_{+}}{r_{+}^{2}+a^{2}},$ (107) $V=\frac{4}{3}r_{+}S_{BH},~{}~{}~{}\mathcal{P}=-\frac{3}{8\pi l^{2}}.$ (108) where those are physical mass, angular momentum, electric charge, magnetic charge, Hawking temperature, Bekenstein-Hawking entropy, angular velocity, electric potential, magnetic potential, thermodynamic volume and pressure. These thermodynamic quantities also satisfy the relation (7). Furthermore, one can find the similar relation for thermodynamics on the cosmological horizon with the following quantities [24, 25] $M_{c}=-\frac{m}{\Xi},~{}~{}~{}J_{c}=-\frac{ma}{\Xi},~{}~{}~{}Q_{c}=-\frac{q}{\Xi},~{}~{}~{}P_{c}=-\frac{p}{\Xi},$ (109) $T_{c}=\frac{r_{c}(2r_{c}^{2}+a^{2}-l^{2})+ml^{2}}{2\pi(r_{c}^{2}+a^{2})l^{2}},$ (110) $S_{c}=\frac{\pi}{\Xi}(r_{c}^{2}+a^{2}),$ (111) $\Omega_{c}=\frac{a\Xi}{r_{c}^{2}+a^{2}},~{}~{}\Phi_{c}=\frac{qr_{c}}{r_{c}^{2}+a^{2}},~{}~{}\Psi_{c}=\frac{pr_{c}}{r_{c}^{2}+a^{2}},$ (112) $V_{c}=\frac{4}{3}r_{c}S_{c},~{}~{}~{}\mathcal{P}_{c}=\frac{3}{8\pi l^{2}}.$ (113) These thermodynamic quantities are obtained also by considering the event horizon of the black hole as the boundary. Likewise the KSdS solution, the cosmological constant can be assumed as a dynamical quantity. ### VI.1 Nariai limit on Kerr-Newman-de Sitter black hole revisited On the Nariai limit, likewise the KSdS solution, we can find the geometry which is fiber over AdS2. Using the similar transformations (18) and (19) on the spacetime metric (1), we can obtain $\displaystyle ds^{2}$ $\displaystyle=$ $\displaystyle\Gamma(\theta)\left(-r(r-\epsilon)dt^{2}+\frac{dr^{2}}{r(r-\epsilon)}+\alpha(\theta)d\theta^{2}\right)$ (114) $\displaystyle+\gamma(\theta)\left(d\phi+erdt\right)^{2},\ $ where the metric functions are given by $\Gamma(\theta)=\frac{\varrho_{+}^{2}}{\upsilon},~{}~{}\alpha(\theta)=\frac{\upsilon}{\Delta_{\theta}},~{}~{}\gamma(\theta)=\frac{r_{0}^{4}\Delta_{\theta}\sin^{2}\theta}{\varrho_{+}^{2}\Xi^{2}},\ $ (115) $\varrho_{+}^{2}=r_{+}^{2}+a^{2}\cos^{2}\theta,~{}~{}e=\frac{2ar_{+}\Xi}{r_{0}^{2}\upsilon}.$ (116) and now we have $r_{0}^{2}=r_{+}^{2}+a^{2}$ and $\upsilon=1-(6r_{+}^{2}+a^{2})/l^{2}$. When the spin vanishes, the spacetime becomes AdS${}_{2}\times S^{2}$. Even we have different slice in the asymptotic region, by using similar analysis of ASG, we can obtain the exact central charge as found in [9]. It is precisely given by $c_{L}=\frac{12ar_{+}}{1-\frac{6r_{+}^{2}+a^{2}}{l^{2}}}.\ $ (117) The main different of this central charge with that of the solution in Nariai limit obtained from KSdS solution is the presence of the dilaton and axion charges. The temperature can be calculated directly as the previous result for Nariai limit on KNdS solution. We can find that $\displaystyle T_{L}=\frac{\upsilon(r_{+}^{2}+a^{2})}{4\pi ar_{+}\Xi},~{}~{}T_{R}=0.\ $ (118) We have obtained the corresponding central charge and the CFT temperature for the solution in Nariai limit obtained from KNdS black hole. On the cosmological horizon, by using Cardy formula, we find the following entropy for the charged rotating black hole from KNdS black hole in Nariai limit, $S_{CFT}=\frac{\pi}{\Xi}(r_{+}^{2}+a^{2})=S_{BH}=S_{c}.\ $ (119) So, in this case, the total entropy will be $2S_{c}$. ### VI.2 Hidden conformal symmetry in $J$-picture For the charged rotating solution in Nariai limit obtained from KNdS black hole, we also assume the massless neutral scalar probe. We also consider the low-frequency limit for the scalar field to exhibit the conformal symmetry. For the small frequency, we have $\omega M\ll 1$, $\omega a\ll 1,\,\omega q\ll 1,$ and $\omega p\ll 1$. The radial equation is then given by $\displaystyle\partial_{\hat{r}}\left[(\hat{r}-r_{c})(\hat{r}-r_{+})\partial_{\hat{r}}\right]R(\hat{r})+\frac{r_{c}-r_{+}}{\hat{r}-r_{c}}AR(\hat{r})=0,$ $\displaystyle+\left[\frac{r_{c}-r_{+}}{\hat{r}-r_{+}}B+C\right]R(\hat{r})=0,$ (120) where $A=\frac{\left[(r_{c}^{2}+a^{2})\omega- an\Xi\right]^{2}}{\upsilon^{2}(r_{c}-r_{+})^{2}},\ $ (121) $B=-\frac{\left[(r_{+}^{2}+a^{2})\omega- an\Xi\right]^{2}}{\upsilon^{2}(r_{c}-r_{+})^{2}},~{}~{}~{}C=\frac{-K_{l}}{\upsilon}.\ $ (122) The investigation of hidden conformal symmetry for KNdS solution is also done in Ref. [40], however, the author does not consider Nariai limit. The conformal symmetry can be revealed using the conformal coordinates transformations (59) which similarly result in $SL(2,R)\times SL(2,R)$ isometry for the set of operators (V.1) and (V.1). Using the similar computation, we can calculate the temperatures. In this case, we obtain $n_{L}=-\frac{\upsilon}{2(r_{c}+r_{+})},~{}~{}~{}n_{R}=0,$ (123) $T_{L}=\frac{\upsilon(r_{c}^{2}+r_{+}^{2}+2a^{2})}{4\pi a(r_{c}+r_{+})\Xi},~{}~{}~{}T_{R}=\frac{\upsilon(r_{c}-r_{+})}{4\pi a\Xi}.$ (124) Note that those quantities are different to those of the Nariai solution obtained from KSdS solution since there exist dilaton and axion charges. When we take $\varepsilon\rightarrow 0$, the temperatures (124) reduce to (118). After computing the CFT temperatures (124), we will compute the conjugate charges $E_{L},E_{R}$ for this solution. These conjugate charges can be obtained from the entropy via Eq. (67) by considering the first law of thermodynamics for the charged rotating black hole solution (7) with the quantities (109)-(113). For the neutral scalar field, we have $\delta Q=\delta P=\delta\mathcal{P}=0$. Using again the identification $\delta M$ as $\omega$ and $\delta J$ as $n$ yields to the identification of $\delta E_{L,R}$ as $\omega_{L,R}$. We obtain the following left and right frequencies, $\omega_{L}=\frac{r_{c}^{2}+r_{+}^{2}+2a^{2}}{2a\Xi}\omega,~{}~{}~{}\omega_{R}=\omega_{L}-n.\ $ (125) These are the left and right frequencies of 2D CFT for the charged rotating solution for KNdS solution in Nariai limit. Regarding the absorption cross-section and real-time correlator, we need to employ the radial equation (120). Similarly with KNdS case, the approximation of $\Delta$ in the asymptotic region will break down, except we consider the Nariai limit which is identical to near-extremal limit. We also consider the scalar probe with frequencies around the superradiant bound (69). With the similar lengthy computation, we find the absorption cross-section of the scalar fields as given in Eq. (76) which is precisely similar to (77) for the charged rotating solution in Nariai limit obtained from KNdS solution. Once again, the agreement between (76) and (77) for this solution can be shown when we choose proper left and right frequencies $\omega_{L},\omega_{R}$ as given in Eq. (125). In the superradiant bound, we also find the form of temperatures (78) and the frequencies (79) with the conformal weights (80). Yet, we have to note again that in the solution coming from Einstein-Maxwell theory, there is no contribution from $d,k$ on $r_{+}$. We can also compute the real-time correlator. This is given by Eq. (82). This further exhibits that the Kerr/CFT correspondence for KNdS black hole in Nariai limit is valid likewise the solution from KSdS black hole. ### VI.3 Hidden conformal symmetry in $Q$-picture In exploring the hidden conformal symmetries on the wave equation, again, we assume a massless charged scalar probe in the background (99) in Nariai limit as given by Eq. (88). By plugging Eq. (89) into Eq. (88) and assuming small charge and low frequency, in the near-horizon region of (99) in Nariai limit, we can find the radial equation $\displaystyle\partial_{\hat{r}}\left[(\hat{r}-r_{c})(\hat{r}-r_{+})\partial_{\hat{r}}\right]R(\hat{r})+\frac{r_{c}-r_{+}}{\hat{r}-r_{c}}AR(\hat{r})=0,$ $\displaystyle+\left[\frac{r_{c}-r_{+}}{\hat{r}-r_{+}}B+C\right]R(\hat{r})=0,$ (126) where $B=-\frac{\left[(r_{+}^{2}+a^{2})\omega-an\Xi- gqr_{+}\right]^{2}}{\upsilon^{2}(r_{c}-r_{+})^{2}},\ $ (127) $A=\frac{\left[(r_{c}^{2}+a^{2})\omega-an\Xi- gqr_{c}\right]^{2}}{\upsilon^{2}(r_{c}-r_{+})^{2}},~{}~{}~{}C=\frac{-K_{l}}{\upsilon}.\ $ (128) For this charged rotating solution obtained from KNdS black hole in Nariai limit, in $Q$-picture, we can identify that $n_{L}=-\frac{\upsilon(r_{c}+r_{+})}{4\left(r_{c}r_{+}-a^{2}\right)},~{}n_{R}=-\frac{\upsilon(r_{c}-r_{+})}{4\left(r_{c}r_{+}-a^{2}\right)},$ (129) $T_{L}=\frac{\upsilon(r_{c}^{2}+r_{+}^{2}+2a^{2})}{4\pi g\left(r_{c}r_{+}-a^{2}\right)},~{}T_{R}=\frac{\upsilon(r^{2}_{c}-r^{2}_{+})}{4\pi g\left(r_{c}r_{+}-a^{2}\right)}.$ (130) These CFT temperatures are regular, unlike to that of KSdS solution in Nariai limit. Since there exists electric charge of the scalar probe, in order to satisfy the entropy relation (67), we need the conjugate of the electric charge, namely the chemical potential ($\mu_{L,R}$), so that $E_{L,R}=\hat{\omega}_{L,R}=\omega_{L,R}-\mu_{L,R}q_{L,R}.\ $ (131) So, in $Q$-picture, the charged scalar probe on the charged rotating solution obtained from KNdS black hole in Nariai limit is related to the following CFT frequencies, $\displaystyle\omega_{L}$ $\displaystyle=$ $\displaystyle\frac{(r_{c}+r_{+})(r_{c}^{2}+r_{+}^{2}+2a^{2})}{2g(r_{c}r_{+}-a^{2})}\omega,$ $\displaystyle\mu_{L}$ $\displaystyle=$ $\displaystyle\frac{r_{c}^{2}+r_{+}^{2}+2a^{2}}{2(r_{c}r_{+}-a^{2})},~{}~{}~{}q_{L}=q,$ $\displaystyle\omega_{R}$ $\displaystyle=$ $\displaystyle\omega_{L}-\frac{2a(r_{c}+r_{+})}{2g(r_{c}r_{+}-a^{2})}n,$ $\displaystyle\mu_{R}$ $\displaystyle=$ $\displaystyle\frac{(r_{c}+r_{+})^{2}}{2(r_{c}r_{+}-a^{2})},~{}~{}~{}q_{R}=q.\ $ (132) One can apply the frequencies (131), temperatures (130), and conformal weights (80) in order to compute the absorption cross-section and real-time correlator. Note that to study the superradiant bound in $Q$-picture, we have assume the frequency near the superradiant bound $\omega=n\Omega_{H}+g\Phi_{H}+\hat{\omega}\frac{\lambda}{r_{0}}$. ## VII Summary In this work, we have shown the horizon solutions and extremal limits in Fig. 1. The Nariai limit could be achieved in some values of parameters. We then have carried out the calculation of the entropy for KSdS solutions in Nariai limit parameterized by a constant $\varepsilon$. When $\varepsilon\rightarrow 0$, the cosmological horizon and event horizon coincide. This solution in Nariai limit can be written in the metric form with AdS2 structure. Hence, we have shown that, in the Nariai limit, we could also portray the solution as a fiber over AdS2, instead of a fiber over dS2 as exhibited in Ref. [9]. We have computed the corresponding central charge and CFT temperature where the right- moving temperature is proportional to $\varepsilon$ denoting that it will vanish when $\varepsilon\rightarrow 0$. It is found that by employing Cardy entropy formula, the Bekenstein-Hawking entropy on the cosmological constant is reproduced. It denotes that the charged rotating solutions from KSdS black hole in Nariai limit is holographically dual with 2D CFT. To further support the CFT dual on KSdS black hole in Nariai limit, we have investigated the neutral ($J$-picture) and charged ($Q$-picture) massless scalar probes on that background. Similarly with generic rotating black hole, we could exhibit the conformal symmetries on the radial wave equation. With the appropriate locally conformal coordinate transformations, it has been shown that the radial equation possesses $SL(2,R)\times SL(2,R)$ isometry. This is similar with AdS3 space. The periodic identification of azimuthal coordinate portrays the spontaneous symmetry breaking from $SL(2,R)\times SL(2,R)$ to $U(1)\times U(1)$ by the left- and right-moving temperatures. In $J$-picture, the temperatures produced on this symmetry breaking are precisely similar with the temperatures that are obtained when the conformal symmetry appears directly on the spacetime metric. Hence, if we employ the central charges with the given temperatures, we can again reproduce the Bekenstein- Hawking entropy on the cosmological horizon. Moreover, we also have computed the absorption cross-section and the real-time correlator that correspond with the KSdS solution in Nariai limit. However, in $Q$-picture, we coud not find a well-defined CFT description because the temperatures are irregular. We have also extended the calculation to the KNdS black hole in Nariai limit both in $J$\- and $Q$-pictures. The CFT description in both pictures are well- defined. The results from gravity calculation are also exactly in agreement with the result from 2D CFT. So, this calculation is another proof that the black hole solution in Nariai limit is holographically dual to 2D CFT. ## Appendix A Entropy on the Event Horizon for Extremal KSdS solution We will briefly derive the entropy of extremal KSdS solution in this section. The entropy of the gauged dyonic Kerr-Sen black hole solution or KSAdS solution has been calculated previously in Ref. [45]. For the asymptotically de Sitter solution, we just need to follow the computation given in Ref. [45] where therein, we have argued on the central charge for KSdS solution by taking $l^{2}\rightarrow-l^{2}$. It has been noted in Ref. [45] that in the extremal case of KSAdS solution, the mass can have more than two branches as it should be similar to KSdS solution. Let start directly from the near- horizon extremal form of KSdS solution, $\displaystyle ds^{2}$ $\displaystyle=$ $\displaystyle\Gamma(\theta)\left(-r^{2}dt^{2}+\frac{dr^{2}}{r^{2}}+\alpha(\theta)d\theta^{2}\right)$ (133) $\displaystyle+\gamma(\theta)\left(d\phi+erdt\right)^{2},\ $ where the metric functions are given by $\Gamma(\theta)=\frac{\varrho_{+}^{2}}{\upsilon},~{}~{}~{}\alpha(\theta)=\frac{\upsilon}{\Delta_{\theta}},~{}~{}~{}\gamma(\theta)=\frac{r_{0}^{4}\Delta_{\theta}\sin^{2}\theta}{\varrho_{+}^{2}\Xi^{2}},\ $ (134) $\varrho_{+}^{2}=r_{+}^{2}-d^{2}-k^{2}+a^{2}\cos^{2}\theta,~{}~{}~{}e=\frac{2ar_{+}\Xi}{r_{0}^{2}\upsilon}.\ $ (135) $\upsilon=1-\frac{6r_{+}^{2}-2d^{2}-2k^{2}+a^{2}}{l^{2}}.$ (136) In this case, we can precisely derive the corresponding central charge as $\displaystyle c_{L}$ $\displaystyle=$ $\displaystyle\frac{3e}{\Xi}\int^{\pi}_{0}d\theta\sqrt{\Gamma(\theta)\alpha(\theta)\gamma(\theta)}$ (137) $\displaystyle=$ $\displaystyle\frac{12ar_{+}}{1-\frac{6r_{+}^{2}-d^{2}-k^{2}+a^{2}}{l^{2}}}.\ $ Note that although this central charge looks identical with that of in Nariai limit, both central charges are differnt because we are considering different limit on the horizons. When the magnetic charge vanishes, one can obtain the central charge that corresponds to non-dyonic KSAdS solution. Furthermore, when the cosmological constant vanishes, the central charge recovers the central charge of Kerr-Sen black hole [47, 35]. For the CFT temperature, we just need to follow Ref. [45]. For KSdS solution, we can find $\displaystyle T_{L}=-\frac{\partial T_{H}/\partial r_{+}}{\partial\Omega_{H}/\partial r_{+}}\bigg{|}_{ex}=\frac{\upsilon(r_{+}^{2}-d^{2}-k^{2}+a^{2})}{4\pi ar_{+}\Xi}.\ $ (138) By implementing Cardy entropy formula and inserting (137) and (138) into it, we successfully derive $S_{CFT}=\frac{\pi}{\Xi}(r_{+}^{2}-d^{2}-k^{2}+a^{2}).\ $ (139) This is precisely the Bekenstein-Hawking entropy of the KSdS black hole on the event horizon. This result shows that we can change $l^{2}$ on those of KSAdS solution into $-l^{2}$ by assuming analytical continuation in order to find the result on those of KSdS solution. This is identical with the KNAdS(dS) case [22]. ## Appendix B Virasoro algebra The Dirac bracket of two classical conserved charges are given by $\displaystyle\left\\{Q_{\zeta_{m}},Q_{\zeta_{n}}\right\\}=Q_{[\zeta_{m},\zeta_{n}]}+K[\zeta_{m},\zeta_{n}],$ (140) where $K[\xi,\zeta]$ is the central term. Using the quantum version of charge $Q_{\xi}$ (44), from Eq. (140) we can find $\displaystyle[L_{m},L_{n}]$ $\displaystyle=$ $\displaystyle i\\{Q_{\zeta_{m}},Q_{\zeta_{n}}\\}$ $\displaystyle=$ $\displaystyle i\left(Q_{[\zeta_{m},\zeta_{n}]}+K[\zeta_{m},\zeta_{n}]\right)$ $\displaystyle=$ $\displaystyle(m-n)L_{m+n}-2mx\delta_{m+n}+iK[\zeta_{m},\zeta_{n}].\ $ In order to find the Virasoro algebra, we need to find the form of term $K[\zeta_{m},\zeta_{n}]$. By comparing above equation with Virasoro algebra, $\left[L_{m},L_{n}\right]=({m}-{n})L_{m+n}+\frac{c_{L}}{12}m(m^{2}-1)\delta_{m+n,0},$ (142) we obtain $K[\zeta_{m},\zeta_{n}]=-i\frac{c_{L}}{12}m\left(m^{2}-1+\frac{24x}{c_{L}}\right)\delta_{m+n,0}.$ (143) So, the central charge is determined by the coefficient $m^{3}$ in term $K[\zeta_{m},\zeta_{n}]$. The term linear in $m$ is not important because $x$ is a free parameter to scale the last term in the bracket of Eq. (143). Acknowledgments The authors thank the reviewer for the fruitful suggestions on this manuscript. This work is supported by the Second Century Fund (C2F), Chulalongkorn University, Thailand. P. B. is supported in part by National Research Council of Thailand (NRCT) and Chulalongkorn University under Grant N42A660500. This research has received funding support from the NSRF via the Program Management Unit for Human Resources & Institutional Development, Research and Innovation [grant number B39G660025]. ## References * [1] V. Mukhanov. Physical Foundations of Cosmology. Cambridge University Press 2006. * [2] P. J. E. Peebles and B. Ratra, The cosmological constant and dark energy, Rev. Mod. Phys. 75 (2003) 559–606. * [3] T. Padmanabhan, Cosmological constant-the weight of the vacuum, Phys. Rep. 380 (2003) 235–320. * [4] S. Alexander, L. S. Finn, and N. Yunes, Gravitational-wave probe of effective quantum gravity, Phys. Rev. D 78 (2008) 066005. * [5] J. Abedi, H. Dykaar, and N. Afshordi, Echoes from the abyss: Tentative evidence for Planck-scale structure at black hole horizons, Phys. Rev. D 96 (2017) 082004. * [6] N. Oshita, D. Tsuna, and N. Afshordi, Quantum black hole seismology. I. Echoes, ergospheres, and spectra, Phys. Rev. D 102 (2020) 024045. * [7] M. Guica, T. Hartman, W. Song and A. Strominger, The Kerr/CFT Correspondence, Phys. Rev. D 80 (2009) 124008. * [8] A. Strominger, Black hole entropy from near-horizon microstates, JHEP 02 (1998) 009. * [9] D. Anninos and T. Hartman, Holography at an extremal de Sitter horizon, JHEP 03 (2010) 096. * [10] M. F. A. R. Sakti, A. Suroso and F. P. Zen, CFT duals on extremal rotating NUT black holes, Int. J. Mod. Phys. D 27 (2018) 1850109. * [11] M. F. A. R. Sakti, A. Suroso and F. P. Zen, Kerr/CFT correspondence on Kerr-Newman-NUT-Quintessence black hole, Eur. Phys. J. Plus 134 (2019) 580. * [12] M. F. A. R. Sakti, A. Suroso and F. P. Zen, Kerr-Newman-NUT-Kiselev black holes in Rastall theory of gravity and Kerr/CFT correspondence, Ann. Phys. 413 (2020) 168062. * [13] M. F. A. R. Sakti and F. P. Zen, CFT duals on rotating charged black holes surrounded by quintessence, Phys. Dark Universe 31 (2021) 100778. * [14] A. Castro, A. Maloney and A. Strominger, Hidden conformal symmetry of the Kerr black hole, Phys. Rev. D 82 (2010) 024008. * [15] M. F. A. R. Sakti, A. M. Ghezelbash, A. Suroso and F. P. Zen, Deformed conformal symmetry of Kerr-Newman-NUT-AdS black holes, Gen. Relativ. Grav. 51 (2019) 151. * [16] M. F. A. R. Sakti, A. M. Ghezelbash, A. Suroso and F. P. Zen, Hidden conformal symmetry for Kerr-Newman-NUT-AdS black holes, Nuc. Phys. B 953 (2020) 114970. * [17] M. F. A. R. Sakti, A. Suroso, A. Sulaksono and F. P. Zen, Rotating black holes and exotic compact objects in the Kerr/CFT correspondence within Rastall gravity, Phys. Dark Universe 35 (2022) 100974. * [18] M. F. A. R. Sakti, Hidden conformal symmetry for dyonic Kerr-Sen black hole andits gauged family, Eur. Phys. J. C 83 (2023) 255. * [19] B. Chen, J. Long and J.-J. Zhang, Hidden conformal symmetry of extremal black holes, Phys. Rev. D 82 (2010) 104017. * [20] B. Chen and J. Long, On holographic description of the Kerr-Newman-AdS-dS black holes, JHEP 08 (2010) 065. * [21] M. Z. Djogama, M. F. A. R. Sakti, F. P. Zen, and M. Satriawan, Near-extremal Kerr-like ECO in the Kerr/CFT Correspondence in Higher Spin Perturbations, arXiv:2308.08205 [hep-th]. * [22] T. Hartman, K. Murata, T. Nishioka and A. Strominger, CFT duals for extreme black holes, JHEP 04 (2009) 019. * [23] M. Sinamuli and R. B. Mann, Super-entropic black holes and the Kerr-CFT correspondence, JHEP 08 (2016) 148. * [24] M. H. Dehghani, Quasilocal thermodynamics of Kerr de Sitter spacetimes and the dS/CFTcorrespondence, Phys. Rev. D 65 (2002) 104030. * [25] A. M. Ghezelbash and R. B. Mann, Entropy and mass bounds of Kerr-de Sitter spacetimes, Phys. Rev. D 72 (2005) 064024. * [26] A. Gomberoff and C. Teitelboim, de Sitter black holes with either of the two horizons as a boundary, Phys. Rev. D 67 (2003) 104024. * [27] M. Cvetič, G. W. Gibbons, D. Kubizňák, and C. N. Pope, Black hole enthalpy and an entropy inequality for the thermodynamic volume, Phys. Rev. D 84 (2011) 024037. * [28] Brian P. Dolan, David Kastor, David Kubizňák, Robert B. Mann, and Jennie Traschen, Thermodynamic volumes and isoperimetric inequalities for de Sitter black holes, Phys. Rev. D 87 (2013) 104017. * [29] A. Castro, F. Mariani and C. Toldo, JHEP 07 (2023), 131 doi:10.1007/JHEP07(2023)131 [arXiv:2212.14356 [hep-th]]. * [30] P. Burikham, S. Ponglertsakul and T. Wuthicharn, Eur. Phys. J. C 80 (2020) no.10, 954 doi:10.1140/epjc/s10052-020-08528-0 [arXiv:2010.05879 [gr-qc]]. * [31] T. Wuthicharn, S. Ponglertsakul and P. Burikham, Int. J. Mod. Phys. D 31 (2022) no.01, 2150127 doi:10.1142/S0218271821501273 [arXiv:1911.11448 [gr-qc]]. * [32] J. D. Brown and M. Henneaux, Central charges in the canonical realization of asymptotic symmetries: an example from three-dimensional gravity, Commun. Math. Phys. 104 (1986) 207. * [33] G. Barnich and F. Brandt, Covariant theory of asymptotic symmetries, conservation laws and central charges, Nuc. Phys. B 633 (2002) 3. * [34] S. Bhattacharya, A note on entropy of de Sitter black holes, Eur. Phys. J. C 76 (2016) 112. * [35] A. M. Ghezelbash and H. M. Siahaan, Hidden and generalized conformal symmetry of Kerr–Sen spacetimes, Class. Quantum Grav. 30 (2013) 135005. * [36] A. M. Ghezelbash and H. M. Siahaan, Deformed hidden conformal symmetry for rotating black holes, Gen. Relativ. Grav. 46 (2014) 1783. * [37] C.-M. Chen, Y.-M. Huang, J.-R. Sun, M.-F. Wu and S.-J. Zou, Twofold hidden conformal symmetries of the Kerr-Newman black hole, Phys. Rev. D 82 (2010) 066004. * [38] B. Chen, C.-M. Chen and B. Ning, Holographic Q-picture of Kerr-Newman-AdS-dS black hole, Nuc. Phys. B 853 (2011) 196-209. * [39] D. A. Lowe, Antun Skanata, and I. Messamah, Hidden Kerr/CFT correspondence at finite frequencies, Phys. Rev. D 89 (2014) 064005. * [40] H. M. Siahaan, Hidden conformal symmetries in Kerr–de Sitter spacetime revisited, Eur. Phys. J. C 82 (2022) 842. * [41] G. Compére, The Kerr/CFT correspondence and its extensions, Living Rev. Rel. 15 (2012) 11 [Addendum ibid. 20 (2017) 1]. * [42] B. Chen, A. M. Ghezelbash, V. Kamali and M. R. Setare, Holographic description of Kerr-Bolt-AdS-dS spacetimes, Nuc. Phys. B 848 (2011) 108-120. * [43] B. Chen and C.-S. Chu, Real-time correlators in Kerr/CFT correspondence, JHEP 05 (2010) 004. * [44] B. Chen and J. Long, Real-time correlators and hidden conformal symmetry in the Kerr/CFT correspondence, JHEP 06 (2010) 018. * [45] M. F. A. R. Sakti and P. Burikham, Dual CFT on Dyonic Kerr-Sen black hole and its gauged and ultraspinning counterparts, Phys. Rev. D 106 (2022) 106006. * [46] D. Wu, S.-Q. Wu, P. Wu, and H. Yu, Aspects of the dyonic Kerr-sen-AdS4 black hole and its ultraspinning version, Phys. Rev. D 103 (2021) 044014. * [47] A. M. Ghezelbash, Kerr/CFT correspondence in the low energy limit of heterotic string theory, JHEP 08 (2009) 045.
11institutetext: INAF, Osservatorio Astrofisico di Catania, Via S. Sofia 78, 95123 Catania, Italy 22institutetext: Observatoire Astronomique de l’Université de Genève, Chemin Pegasi 51, CH-1290 Versoix, Switzerland 33institutetext: ETH Zurich, Department of Physics, Wolfgang-Pauli-Strasse 2, CH-8093 Zurich, Switzerland 44institutetext: Cavendish Laboratory, JJ Thomson Avenue, Cambridge CB3 0HE, UK 55institutetext: Physikalisches Institut, University of Bern, Gesellschaftsstrasse 6, 3012 Bern, Switzerland 66institutetext: Instituto de Astrofisica e Ciencias do Espaco, Universidade do Porto, CAUP, Rua das Estrelas, 4150-762 Porto, Portugal 77institutetext: Department of Astronomy, Stockholm University, AlbaNova University Center, 10691 Stockholm, Sweden 88institutetext: Centre for Exoplanet Science, SUPA School of Physics and Astronomy, University of St Andrews, North Haugh, St Andrews KY16 9SS, UK 99institutetext: Aix Marseille Univ, CNRS, CNES, LAM, 38 rue Frédéric Joliot-Curie, 13388 Marseille, France 1010institutetext: Space sciences, Technologies and Astrophysics Research (STAR) Institute, Université de Liège, Allée du 6 Août 19C, 4000 Liège, Belgium 1111institutetext: Center for Space and Habitability, University of Bern, Gesellschaftsstrasse 6, 3012 Bern, Switzerland 1212institutetext: Instituto de Astrofisica de Canarias, Via Lactea s/n, 38200 La Laguna, Tenerife, Spain 1313institutetext: Departamento de Astrofisica, Universidad de La Laguna, Astrofísico Francisco Sanchez s/n, 38206 La Laguna, Tenerife, Spain 1414institutetext: Institut de Ciencies de l’Espai (ICE, CSIC), Campus UAB, Can Magrans s/n, 08193 Bellaterra, Spain 1515institutetext: Institut d’Estudis Espacials de Catalunya (IEEC), Gran Capità 2-4, 08034 Barcelona, Spain 1616institutetext: Admatis, 5. Kandó Kálmán Street, 3534 Miskolc, Hungary 1717institutetext: Depto. de Astrofisica, Centro de Astrobiologia (CSIC-INTA), ESAC campus, 28692 Villanueva de la Cañada (Madrid), Spain 1818institutetext: Departamento de Fisica e Astronomia, Faculdade de Ciencias, Universidade do Porto, Rua do Campo Alegre, 4169-007 Porto, Portugal 1919institutetext: Space Research Institute, Austrian Academy of Sciences, Schmiedlstrasse 6, A-8042 Graz, Austria 2020institutetext: Université Grenoble Alpes, CNRS, IPAG, 38000 Grenoble, France 2121institutetext: INAF, Osservatorio Astronomico di Padova, Vicolo dell’Osservatorio 5, 35122 Padova, Italy 2222institutetext: Université de Paris Cité, Institut de physique du globe de Paris, CNRS, 1 Rue Jussieu, F-75005 Paris, France 2323institutetext: ESTEC, European Space Agency, Keplerlaan 1, 2201AZ, Noordwijk, NL 2424institutetext: Institute of Planetary Research, German Aerospace Center (DLR), Rutherfordstrasse 2, 12489 Berlin, Germany 2525institutetext: INAF, Osservatorio Astrofisico di Torino, Via Osservatorio, 20, I-10025 Pino Torinese To, Italy 2626institutetext: Centre for Mathematical Sciences, Lund University, Box 118, 221 00 Lund, Sweden 2727institutetext: Astrobiology Research Unit, Université de Liège, Allée du 6 Août 19C, B-4000 Liège, Belgium 2828institutetext: Centre Vie dans l’Univers, Faculté des sciences, Université de Genève, Quai Ernest-Ansermet 30, CH-1211 Genève 4, Switzerland 2929institutetext: Leiden Observatory, University of Leiden, PO Box 9513, 2300 RA Leiden, The Netherlands 3030institutetext: Department of Space, Earth and Environment, Chalmers University of Technology, Onsala Space Observatory, 439 92 Onsala, Sweden 3131institutetext: Dipartimento di Fisica, Universita degli Studi di Torino, via Pietro Giuria 1, I-10125, Torino, Italy 3232institutetext: Department of Astrophysics, University of Vienna, Türkenschanzstrasse 17, 1180 Vienna, Austria 3333institutetext: Science and Operations Department - Science Division (SCI-SC), Directorate of Science, European Space Agency (ESA), European Space Research and Technology Centre (ESTEC), Keplerlaan 1, 2201-AZ Noordwijk, The Netherlands 3434institutetext: Konkoly Observatory, Research Centre for Astronomy and Earth Sciences, 1121 Budapest, Konkoly Thege Miklós út 15-17, Hungary 3535institutetext: ELTE Eötvös Loránd University, Institute of Physics, Pázmány Péter sétány 1/A, 1117 Budapest, Hungary 3636institutetext: German Aerospace Center (DLR), Institute of Optical Sensor Systems, Rutherfordstraße 2, 12489 Berlin 3737institutetext: IMCCE, UMR8028 CNRS, Observatoire de Paris, PSL Univ., Sorbonne Univ., 77 av. Denfert- Rochereau, 75014 Paris, France 3838institutetext: Institut d’astrophysique de Paris, UMR7095 CNRS, Université Pierre & Marie Curie, 98bis blvd. Arago, 75014 Paris, France 3939institutetext: Astrophysics Group, Lennard Jones Building, Keele University, Staffordshire, ST5 5BG, United Kingdom 4040institutetext: Institute of Optical Sensor Systems, German Aerospace Center (DLR), Rutherfordstrasse 2, 12489 Berlin, Germany 4141institutetext: Dipartimento di Fisica e Astronomia ”Galileo Galilei”, Universita degli Studi di Padova, Vicolo dell’Osservatorio 3, 35122 Padova, Italy 4242institutetext: Department of Physics, University of Warwick, Gibbet Hill Road, Coventry CV4 7AL, United Kingdom 4343institutetext: Zentrum für Astronomie und Astrophysik, Technische Universität Berlin, Hardenbergstr. 36, D-10623 Berlin, Germany 4444institutetext: Institut fuer Geologische Wissenschaften, Freie Universitaet Berlin, Maltheserstrasse 74-100,12249 Berlin, Germany 4545institutetext: Department of Astrophysics, University of Vienna, Tuerkenschanzstrasse 17, 1180 Vienna, Austria 4646institutetext: Physikalisches Institut, University of Bern, Sidlerstrasse 5, 3012 Bern, Switzerland 4747institutetext: Université de Liège, Allée du 6 Août 19C, 4000 Liège, Belgium 4848institutetext: ELTE Eötvös Loránd University, Gothard Astrophysical Observatory, 9700 Szombathely, Szent Imre h. u. 112, Hungary 4949institutetext: MTA-ELTE Exoplanet Research Group, 9700 Szombathely, Szent Imre h. u. 112, Hungary 5050institutetext: Institute of Astronomy, University of Cambridge, Madingley Road, Cambridge, CB3 0HA, United Kingdom 5151institutetext: Institute for Theoretical Physics and Computational Physics, Graz University of Technology, Petersgasse 16, 8010 Graz, Austria # Constraining the reflective properties of WASP-178 b using CHEOPS photometry.††thanks: The CHEOPS program ID is CH_PR100016.††thanks: The CHEOPS photometric data used in this work are only available in electronic form at the CDS via anonymous ftp to cdsarc.cds.unistra.fr (130.79.128.5) or via https://cdsarc.cds.unistra.fr/cgi-bin/qcat?J/A+A/ I. Pagano ${}^{\href{https://orcid.org/0000-0001-9573-4928}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 11 G. Scandariato ${}^{\href{https://orcid.org/0000-0003-2029-0626}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 11 V. Singh ${}^{\href{https://orcid.org/0000-0002-7485-6309}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 11 M. Lendl ${}^{\href{https://orcid.org/0000-0001-9699-1459}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 22 D. Queloz ${}^{\href{https://orcid.org/0000-0002-3012-0316}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 3344 A. E. Simon ${}^{\href{https://orcid.org/0000-0001-9773-2600}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 55 S. G. Sousa ${}^{\href{https://orcid.org/0000-0001-9047-2965}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 66 A. Brandeker ${}^{\href{https://orcid.org/0000-0002-7201-7536}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 77 A. Collier Cameron ${}^{\href{https://orcid.org/0000-0002-8863-7828}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 88 S. Sulis ${}^{\href{https://orcid.org/0000-0001-8783-526X}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 99 V. Van Grootel ${}^{\href{https://orcid.org/0000-0003-2144-4316}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 1010 T. G. Wilson ${}^{\href{https://orcid.org/0000-0001-8749-1962}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 88 Y. Alibert ${}^{\href{https://orcid.org/0000-0002-4644-8818}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 111155 R. Alonso ${}^{\href{https://orcid.org/0000-0001-8462-8126}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 12121313 G. Anglada ${}^{\href{https://orcid.org/0000-0002-3645-5977}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 14141515 T. Bárczy ${}^{\href{https://orcid.org/0000-0002-7822-4413}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 1616 D. Barrado Navascues ${}^{\href{https://orcid.org/0000-0002-5971-9242}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 1717 S. C. C. Barros ${}^{\href{https://orcid.org/0000-0003-2434-3625}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 661818 W. Baumjohann ${}^{\href{https://orcid.org/0000-0001-6271-0110}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 1919 M. Beck ${}^{\href{https://orcid.org/0000-0003-3926-0275}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 22 T. Beck 55 W. Benz ${}^{\href{https://orcid.org/0000-0001-7896-6479}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 551111 N. Billot ${}^{\href{https://orcid.org/0000-0003-3429-3836}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 22 X. Bonfils ${}^{\href{https://orcid.org/0000-0001-9003-8894}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 2020 L. Borsato ${}^{\href{https://orcid.org/0000-0003-0066-9268}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 2121 C. Broeg ${}^{\href{https://orcid.org/0000-0001-5132-2614}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 551111 G. Bruno ${}^{\href{https://orcid.org/0000-0002-3288-0802}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 11 L. Carone 1919 S. Charnoz ${}^{\href{https://orcid.org/0000-0002-7442-491X}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 2222 C. Corral van Damme 2323 Sz. Csizmadia ${}^{\href{https://orcid.org/0000-0001-6803-9698}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 2424 P. E. Cubillos 25251919 M. B. Davies ${}^{\href{https://orcid.org/0000-0001-6080-1190}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 2626 M. Deleuil ${}^{\href{https://orcid.org/0000-0001-6036-0225}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 99 A. Deline 22 L. Delrez ${}^{\href{https://orcid.org/0000-0001-6108-4808}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 27271010 O. D. S. Demangeon ${}^{\href{https://orcid.org/0000-0001-7918-0355}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 661818 B.-O. Demory ${}^{\href{https://orcid.org/0000-0002-9355-5165}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 111155 D. Ehrenreich ${}^{\href{https://orcid.org/0000-0001-9704-5405}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 222828 A. Erikson 2424 A. Fortier ${}^{\href{https://orcid.org/0000-0001-8450-3374}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 551111 L. Fossati ${}^{\href{https://orcid.org/0000-0003-4426-9530}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 1919 M. Fridlund ${}^{\href{https://orcid.org/0000-0002-0855-8426}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 29293030 D. Gandolfi ${}^{\href{https://orcid.org/0000-0001-8627-9628}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 3131 M. Gillon ${}^{\href{https://orcid.org/0000-0003-1462-7739}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 2727 M. Güdel 3232 M. N. Günther 2323 Ch. Helling 19195151 S. Hoyer ${}^{\href{https://orcid.org/0000-0003-3477-2466}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 99 K. G. Isaak ${}^{\href{https://orcid.org/0000-0001-8585-1717}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 3333 L. L. Kiss 34343535 E. Kopp 3636 K. W. F. Lam ${}^{\href{https://orcid.org/0000-0002-9910-6088}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 2424 J. Laskar ${}^{\href{https://orcid.org/0000-0003-2634-789X}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 3737 A. Lecavelier des Etangs ${}^{\href{https://orcid.org/0000-0002-5637-5253}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 3838 D. Magrin ${}^{\href{https://orcid.org/0000-0003-0312-313X}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 2121 P. F. L. Maxted ${}^{\href{https://orcid.org/0000-0003-3794-1317}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 3939 C. Mordasini 551111 M. Munari ${}^{\href{https://orcid.org/0000-0003-0990-050X}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 11 V. Nascimbeni ${}^{\href{https://orcid.org/0000-0001-9770-1214}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 2121 G. Olofsson ${}^{\href{https://orcid.org/0000-0003-3747-7120}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 77 R. Ottensamer 3232 E. Pallé ${}^{\href{https://orcid.org/0000-0003-0987-1593}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 12121313 G. Peter ${}^{\href{https://orcid.org/0000-0001-6101-2513}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 4040 G. Piotto ${}^{\href{https://orcid.org/0000-0002-9937-6387}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 21214141 D. Pollacco 4242 R. Ragazzoni ${}^{\href{https://orcid.org/0000-0002-7697-5555}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 21214141 N. Rando 2323 H. Rauer ${}^{\href{https://orcid.org/0000-0002-6510-1828}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 242443434444 C. Reimers 4545 I. Ribas ${}^{\href{https://orcid.org/0000-0002-6689-0312}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 14141515 M. Rieder 46461111 N. C. Santos ${}^{\href{https://orcid.org/0000-0003-4422-2919}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 661818 D. Ségransan ${}^{\href{https://orcid.org/0000-0003-2355-8034}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 22 A. M. S. Smith ${}^{\href{https://orcid.org/0000-0002-2386-4341}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 2424 M. Stalport 4747 M. Steller ${}^{\href{https://orcid.org/0000-0003-2459-6155}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 1919 Gy. M. Szabó 48484949 N. Thomas 55 S. Udry ${}^{\href{https://orcid.org/0000-0001-7576-6236}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 22 J. Venturini ${}^{\href{https://orcid.org/0000-0001-9527-2903}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 22 N. A. Walton ${}^{\href{https://orcid.org/0000-0003-3983-8778}{\includegraphics[scale={0.5}]{figures/orcid.jpg}}}$ 5050 ###### Abstract Context. Multiwavelength photometry of the secondary eclipses of extrasolar planets is able to disentangle the reflected and thermally emitted light radiated from the planetary dayside. This leads to the measurement of the planetary geometric albedo $A_{g}$, which is an indicator of the presence of clouds in the atmosphere, and the recirculation efficiency $\epsilon$, which quantifies the energy transport within the atmosphere. Aims. In this work we aim to measure $A_{g}$ and $\epsilon$ for the planet WASP-178 b, a highly irradiated giant planet with an estimated equilibrium temperature of 2450 K. Methods. We analyzed archival spectra and the light curves collected by CHEOPS and TESS to characterize the host WASP-178, refine the ephemeris of the system and measure the eclipse depth in the passbands of the two respective telescopes. Results. We measured a marginally significant eclipse depth of 70$\pm$40 ppm in the TESS passband and statistically significant depth of 70$\pm$20 ppm in the CHEOPS passband. Conclusions. Combining the eclipse depth measurement in the CHEOPS ($\lambda_{\rm eff}=6300\leavevmode\nobreak\ \AA$) and TESS ($\lambda_{\rm eff}=8000\leavevmode\nobreak\ \AA$) passbands we constrained the dayside brightness temperature of WASP-178 b in the 2250-2800 K interval. The geometric albedo 0.1¡$\rm A_{g}$¡0.35 is in general agreement with the picture of poorly reflective giant planets, while the recirculation efficiency $\epsilon>$0.7 makes WASP-178 b an interesting laboratory to test the current heat recirculation models. ###### Key Words.: techniques: photometric – planets and satellites: atmospheres – planets and satellites: detection – planets and satellites: gaseous planets – planets and satellites: individual: WASP-178 b ## 1 Introduction In the last few years we have gained access to a detailed characterization of exoplanets. Ground-based and space-born instrumentation have progressed such to allow the analysis of the atmosphere of exoplanets, in terms of thermodynamic state and chemical composition (Sing et al. 2016; Giacobbe et al. 2021). In particular, current photometric facilities allow to observe the secondary eclipse of giant exoplanets in close orbits (Stevenson et al. 2017; Lendl et al. 2020; Wong et al. 2020; Singh et al. 2022). In this research area, CHEOPS (Benz et al. 2021) is bringing a valuable contribution given its ultra-high photometric accuracy capabilities (Lendl et al. 2020; Deline et al. 2022; Hooton et al. 2022; Brandeker et al. 2022; Parviainen et al. 2022; Scandariato et al. 2022; Demory et al. 2023). The depth of the eclipse quantifies the brightness of the planetary dayside with respect to its parent star. Depending on the temperature of the planet and the photometric band used for the observations, the eclipse depth provides insight into the reflectivity and energy redistribution of the atmosphere. WASP-178 b (HD 134004 b) is a hot Jupiter discovered by Hellier et al. (2019) and independently announced by Rodríguez Martínez et al. (2020) as KELT-26 b. It orbits an A1 IV-V dwarf star (Table 1) at a distance of 7 stellar radii: these features place WASP-178 b among the planets that receive the highest energy budget from their respective host stars. It is thus an interesting laboratory to test atmospheric models in the presence of extreme irradiation. In this paper we analyze the eclipse depths measured using the data collected by the CHEOPS and TESS space telescopes. It is organized as follows. Sect. 2 describes the data acquisition and reduction, while in Sect. 3 we describe how we derive the stellar radius, mass and age. In Sect. 4 we update the orbital solution of WASP-178 b, put an upper limit on the eclipse depth using TESS data and get a significant detection using CHEOPS photometry. Finally, in Sect. 5 we discuss the implication of the extracted eclipse signal in terms of geometric albedo and atmospheric recirculation efficiency. Table 1: Stellar and system parameters. Parameter | Symbol | Units | Value | Ref. ---|---|---|---|--- V mag | | | 9.95 | Hellier et al. (2019) Spectral Type | | | A1 IV-V | Hellier et al. (2019) Effective temperature | $\rm T_{eff}$ | K | $9350\pm 150$ | Hellier et al. (2019) Surface gravity | $\log g$ | — | $4.35\pm 0.15$ | Hellier et al. (2019) Metallicity | [Fe/H] | — | $0.21\pm 0.16$ | Hellier et al. (2019) Projected rotational velocity | $v\sin{i}$ | km/s | $8.2\pm 0.6$ | Hellier et al. (2019) Stellar radius | $\rm R_{\star}$ | $\rm R_{\sun}$ | $1.722\pm 0.020$ | this work Stellar mass | $\rm M_{\star}$ | $\rm M_{\sun}$ | $2.169_{-0.089}^{+0.083}$ | this work Stellar age | $t_{\star}$ | Gyr | $0.05_{-0.05}^{+0.06}$ | this work Radial velocity semi-amplitude | $\rm K_{RV}$ | m/s | $139\pm 9$ | Hellier et al. (2019) ## 2 Observations and data reduction ### 2.1 TESS observations TESS (Ricker et al. 2014) observed the WASP-178 system in sector 11 (from 2019 April 23 to 2019 May 20) with a 30 min cadence and in sector 38 (from 2020 April 29 to 2020 May 26) with a 2 min cadence. Rodríguez Martínez et al. (2020) claimed a modulation with period of 0.369526 days in the photometry of sector 11, interpreting it as a $\delta$ Scuti pulsation mode. Later, Lothringer et al. (2022) found out that the TESS photometry is heavily contaminated by the background eclipsing binary ASASSN-V J150908.07-424253.6, located at a projected distance of 50.4″ from the target, whose orbital period matches the periodicity in the photometry of WASP-178. In order to optimize the photometric extraction and avoid the background variable contamination, we extracted the Light Curve (LC) corresponding to each pixel in the aperture mask defined by the TESS pipeline, and we excluded the pixels for which the periodogram shows a peak at the same orbital period of the binary system. We thus re-extracted the photometry by retrieving the calibrated Full Frame Imagess and Target Pixel Files for sector 11 and 38 respectively. We used a custom extraction pipeline combined with the default quality bitmask. The extracted LCs were then background-corrected after determining the sky level using custom background masks on the FFIs and TPFs. A principal component analysis was then conducted on the pixels in these background masks across all frames in order to measure the flux contribution of scattered light in the TESS cameras. We detrended the data using these principal components as a linear model. Lastly, we further corrected for any photometric trends due to spacecraft pointing jitter by retrieving the Co-trending Basis Vectorss and two-second cadence engineering quaternion measurements for the specific cameras WASP-178 was observed in. For each sector we computed the mean of the quaternions over the length of a science observation, i.e. 30 min for the FFIs and 2 min for the TPFs. We subsequently used these vectors along with the CBVs to detrend the TESS photometry in a similar manner as was done in Delrez et al. (2021). We clip out the photometry taken before BJD 2459334.7 and in the windows 2458610–2458614.5 and 2459346–2459348: these data present artificial trends due to the momentum dumps of the telescope. We also visually identified and excluded a short bump in the LC between BJD 2459337.6 and 2459337.9, most likely an instrumental artifact or some short term photometric variability feature. The final LCs cover 7 and 8 transits for sector 11 and 38 respectively. ### 2.2 CHEOPS observations CHEOPS (Benz et al. 2021) observed the WASP-178 system during six secondary eclipses of the planet WASP-178 b with a cadence of 60 s. The aim is the measurement of the eclipse depth and derivation of the brightness of the planet in the CHEOPS passband (3500–11000 Å). Each visit is $\sim$12 hr long, scheduled in order to bracket the eclipse and equally long pre- and post- eclipse photometry. The logbook of the observations, which are part of the CHEOPS Guaranteed Time Observation (GTO) program, is summarized in Table 2. Table 2: Logbook of the CHEOPS observations of WASP-178. The filekey is the unique identifier associated with each dataset processed by the CHEOPS DRP. Filekey | Start time ]UT] | Visit duration [hr] | Exposure time [s] | N. frames | Efficiency [%] ---|---|---|---|---|--- PR100016_TG014201_V0200 | 2021-04-05 12:49:30 | 11.54 | 60.0 | 427 | 61.7 PR100016_TG014202_V0200 | 2021-04-15 13:54:09 | 12.14 | 60.0 | 473 | 64.9 PR100016_TG014203_V0200 | 2021-05-02 06:56:09 | 11.54 | 60.0 | 465 | 67.2 PR100016_TG014204_V0200 | 2021-05-22 08:52:09 | 11.54 | 60.0 | 431 | 62.3 PR100016_TG014205_V0200 | 2021-05-28 23:49:09 | 11.44 | 60.0 | 415 | 60.5 PR100016_TG014206_V0200 | 2022-05-21 22:33:49 | 11.54 | 60.0 | 445 | 64.3 The data were reduced using version 13 of the CHEOPS Data Reduction Pipeline (DRP) (Hoyer et al. 2020). This pipeline performs the standard calibration steps (bias, gain, non-linearity, dark current and flat fielding) and corrects for environmental effects (cosmic rays, smearing trails from nearby stars, and background) before the photometric extraction. As for the case of TESS LCs, the aperture photometry is significantly contaminated by ASASSN-V J150908.07-424253. To decontaminate the LC of WASP-178 we performed the photometric extraction using a modified version of the package PIPE111https://github.com/alphapsa/PIPE (Brandeker et al. 2022; Morris et al. 2021; Szabó et al. 2021), upgraded in order to compute the simultaneous Point Spread Function (PSF) photometry of the target and the background contaminant. The extracted LCs present gaps due to Earth occultations which cover $\sim$40% of the visits. The exposures close to the gaps are characterized by high value of the background flux, due to stray-light from Earth. The corresponding flux measurements are thus affected by a larger photometric scatter. To avoid these low quality data, we applied a 5$\sigma$ clipping to the background measurements: this selection criterion removes less than 20% of the data with the highest background counts. Finally, for a better outlier rejection, we smoothed the data with a Savitzky-Golay filter, computed the residuals with respect to the filtered LC and 5$\sigma$-clipped the outliers. This last rejection criterion excludes a handful of data points in each LC. Finally, we normalize the unsmoothed LCs by the median value of the photometry. These normalized LCs are publicly available at CDS PUT LINK. ## 3 Stellar radius, mass and age We use a Infra-Red Flux Method (IRFM) in a Markov Chain Monte Carlo (MCMC) approach to determine the stellar radius of WASP-178 (Blackwell & Shallis 1977; Schanche et al. 2020). We downloaded the broadband fluxes and uncertainties from the most recent data releases for the following bandpasses: Gaia G, GBP, and GRP, 2MASS J, H, and K, and WISE W1 and W2 (Skrutskie et al. 2006; Wright et al. 2010; Gaia Collaboration et al. 2021). Then we matched the observed photometry with synthetic photometry computed in the same bandpasses by using the theoretical stellar Spectral Energy Distributions corresponding to stellar atmospheric parameters (Table 1). The fit is performed in a Bayesian framework and, to account for uncertainties in stellar atmospheric modeling, we averaged the atlas (Kurucz 1993; Castelli & Kurucz 2003) and phoenix (Allard 2014) catalogs to produce weighted averaged posterior distributions. This process yields a $R_{\star}=1.722\pm 0.020\,R_{\odot}$. Assuming the $T_{\mathrm{eff}}$, [Fe/H], and $R_{\star}$ listed in Table 1 as input parameters, we also computed the stellar mass $M_{\star}$ and age $t_{\star}$ by using two different sets of stellar evolutionary models. In detail, we employed the isochrone placement algorithm (Bonfanti et al. 2015, 2016) and its capability of interpolating within pre-computed grids of PARSEC222PAdova and TRieste Stellar Evolutionary Code: http://stev.oapd.inaf.it/cgi-bin/cmd v1.2S (Marigo et al. 2017) for retrieving a first pair of mass and age estimates. A second pair of mass and age values, instead, was computed by CLES (Code Liègeois d’Évolution Stellaire; Scuflaire et al. 2008), which generates the best-fit evolutionary track of the star by entering the input parameters into the Levenberg-Marquadt minimisation scheme as described in Salmon et al. (2021). After carefully checking the mutual consistency of the two respective pairs of estimates through the $\chi^{2}$-based criterion broadly discussed in Bonfanti et al. (2021), we finally merged the outcome distributions and we obtained $M_{\star}=2.169_{-0.089}^{+0.083}\,M_{\odot}$ and $t_{\star}=50_{-50}^{+60}$ Myr. ## 4 Light curve analysis ### 4.1 TESS photometry To compare in a homogeneous way the LCs of sectors 11 and 38, we rebinned the photometry of sector 38 to 30 min. The standard deviation of the TESS LCs, after transits and secondary eclipses are clipped, is of $\sim$300 ppm for both sectors, and in both cases the photometric uncertainty can account for only $\sim$70% of the variance. This indicates that there is some noise in the LCs due to astrophysical signals and/or instrumental leftovers. To investigate if the unexplained variance is related to periodic signals, for both sectors we compute the Generalized Lomb-Scargle (GLS) periodogram (Zechmeister & Kürster 2009) of the out-of-transit and out-of-eclipse photometry (Fig. 1). For sector 11 we did not find any significant periodic signal, while for sector 38 there is a strong peak at frequency $\rm\nu=0.304\pm 0.003\leavevmode\nobreak\ d^{-1}$ with False Alarm Probability (FAP) lower than 0.1%. The amplitude of the corresponding sinusoidal signal is $90\pm 10$ ppm. Figure 1: Out-of-transit and out-of-eclipse 30 min cadence TESS photometry of WASP-178 during sector 11 (left panel) and sector 38 (central panel). In each panel, the dashed vertical lines mark the planetary transit, while the solid blue line is a smoothing of the data points to emphasize the correlated noise. The right panel shows the GLS periodogram of sector 11 and 38 (top and bottom box respectively). In each box, we report the bootstrap-computed 0.1% and 1% FAP levels (horizontal dashes) and the planetary orbital period (vertical red dotted line). The periodicity detected in sector 38 is consistent within uncertainties with the planetary orbital period. Nonetheless, we exclude that this signal is of planetary origin because it remained undetected in sector 11 and is not coherent from orbit to orbit in sector 38. This signal might be due to stellar rotation, and is consistent with the typical amplitude reported by Balona (2011) for A-type stars. Its periodicity would correspond to the $v\sin i_{\star}$ in Table 1 if the inclination of the stellar rotation axis is $\sim 15^{\circ}$. This speculation supports the hypothesis of Rodríguez Martínez et al. (2020) that the star is seen nearly pole-on. A significant misalignment between the stellar rotation axis and the planetary orbit axis is not surprising, both because of its young age (no time for realignment to occur), and the stellar temperature. As a matter of fact, hot stars (Teff¿6200 K) are observed to often have large misalignments, which is thought to be due to their lack of a convective zone (Winn et al. 2010), needed to tidally align the planetary orbit with the stellar spin axis. Furthermore, Albrecht et al. (2021) found that misaligned orbits are most often polar, or close to polar, as seems to be the case of WASP-178 b. We expand the discussion on the photometric variability and the orientation of the stellar rotation axis in Appendix B. The fact that the period of the correlated noise is similar to the orbital period of WASP-178 b makes it difficult to extract the complete planetary Phase Curve (PC) signal. We thus first tried a simpler and more robust approach to analyze the planetary transits and eclipses (Sect. 4.1.1), then we attempted a more complex analysis framework aimed at retrieving the full PC of WASP-178 b (Sect. 4.1.2). #### 4.1.1 Fit of transits and eclipses We computed the ephemeris of the planet by trimming segments of the LCs centered on the transit events (7 transits in sector 11 and 8 transits in sector 8) and as wide as 3 times the transit duration. To further constrain the ephemeris of WASP-178 b, we included in our analysis the WASP-South photometry (2006 May – 2014 Aug) and EulerCAM I-band photometry (2018 Mar 26) presented in Hellier et al. (2019). Also for the case of the WASP-South photometry, we trimmed the LC around the transits and kept the 24 intervals containing more than 20 data points. We fit the data using the same Bayesian approach described in Scandariato et al. (2022). In summary, it consists in the fit of a model - in a likelihood- maximization framework - which includes the transits, a linear term for each transit to detrend against stellar/instrumental systematics and a jitter term to fit the white noise not included in the photometric uncertainties. The transit profile is formalized using the quadratic limb darkening (LD) law indicated by Mandel & Agol (2002) with the reparametrization of the LD coefficients suggested by Kipping (2013). For TESS sector 11 the model is rebinned to the same 30 min cadence of the data. The likelihood maximization is performed with MCMC using the python emcee package version 3.1.3 (Foreman- Mackey et al. 2013), using a number of samplings long enough to ensure convergence. We used flat priors for all the fitting parameters but the stellar density, for which we used the Gaussian prior $N(0.43,0.02)$ given by the stellar mass and radius in Table 1. We also used the Gaussian priors for the Limb Darkening (LD) coefficients given by the LDTk package 333https://github.com/hpparvi/ldtk. For simplicity, we use the same LD coefficients for the three datasets. This is motivated by the fact that the WASP-South photometry is not accurate enough to constrain the LD profile, and that the TESS passband is basically centered on the standard I band, thus the same LD profile is expected for the TESS and EulerCAM LCs. Since the model fitting is computationally demanding, we ran the code in the HOTCAT computing infrastructure (Bertocco et al. 2020; Taffoni et al. 2020). The result of the model fitting is listed in Table 3. The orbital solution we derived is consistent with previous studies (Hellier et al. 2019; Rodríguez Martínez et al. 2020) within 5$\sigma$. The best fit model, corresponding to the Maximum A Posteriori (MAP) parameters, is over-plotted to the phase-folded data in Fig 2, where we rebinned the model and the LC of sector 38 to the same 30 min cadence of sector 11 (we do not show the WASP-South and EulerCAM photometry not to clutter the plot). In Fig. 9 we show the corner plot of the system’s parameters from the fit of the transit LCs. Table 3: Model parameters for the fit of the TESS WASP-South and EulerCAM data. Jump parameters | Symbol | Units | MAP | C.I.aa$a$Uncertainties expressed in parentheses refer to the last digit(s). | Prior ---|---|---|---|---|--- Time of transit | $T_{0}$ | BJDTDB-2400000 | 56612.6581 | 56612.6581(3) | $U$(56612.6, 56612.7) Orbital frequency | $\nu_{\rm orb}$ | days-1 | 0.29896856 | 0.29896855(3) | $U$(0.2989,0.2990) Stellar density | $\rho_{\star}$ | $\rho_{\sun}$ | 0.45 | 0.44(1) | $N$(0.43,0.02) Radii ratiobb$b$Fitting WASP-South, EulerCAM and TESS data altogether. | $R_{p}/R_{\star}$ | — | 0.1124 | 0.1125(2) | $U$(0.05,0.12) Radii ratiocc$c$Fitting TESS sector 11 only. | $R_{p}/R_{\star}$ | — | 0.1109 | 0.1108(4) | $U$(0.05,0.12) Radii ratiodd$d$Fitting TESS sector 38 only. | $R_{p}/R_{\star}$ | — | 0.1141 | 0.1141(4) | $U$(0.05,0.12) Impact parameter | b | — | 0.52 | 0.51(1) | $U$(0.01,0.9) First LD coef.bb$b$Fitting WASP-South, EulerCAM and TESS data altogether. | q1 | — | 0.147 | 0.147(8) | $N$(0.133,0.014) Second LD coef.bb$b$Fitting WASP-South, EulerCAM and TESS data altogether. | q2 | — | 0.344 | 0.34(2) | $N$(0.333,0.023) Secondary eclipse depth | $\delta_{\rm ecl}$ | ppm | 70 | 70(40) | $U$(0,400) Derived parameters | Symbol | Units | MAP | C.I. | Planetary radiusbb$b$Fitting WASP-South, EulerCAM and TESS data altogether. | Rp | RJ | 1.88 | 1.88(2) | including the stellar radius uncertainty Planetary radiuscc$c$Fitting TESS sector 11 only. | Rp | RJ | 1.86 | 1.85(2) | including the stellar radius uncertainty Planetary radiusdd$d$Fitting TESS sector 38 only. | Rp | RJ | 1.91 | 1.91(2) | including the stellar radius uncertainty Orbital period | Porb | day | 3.3448332 | 3.3448333(4) | Transit duration | T14 | hr | 3.488 | 3.488(8) | Scaled semi-major axis | $a/R_{\star}$ | — | 7.20 | 7.19(6) | Orbital inclination | $i$ | degrees | 85.8 | 85.8(1) | 444 Since the TESS LCs of sector 11 and 38 show different levels of variability, we investigated any seasonal dependence of the apparent planet-to-star radius ratio. We used the same Bayesian framework as above, where we fix the ephemeris of WASP-178. The planet-to-star radius ratio we derived is 0.1108$\pm$0.0004 for sector 11 and 0.1141$\pm$0.0004 for sector 38. The difference is thus 0.0033$\pm$0.0006, that is we found different transit depths with a 5.5$\sigma$ significance. In particular, we remark that the planet looks larger in sector 38, where the rotation signal is stronger. We thus speculate that WASP-178 was in a low variability state during sector 11, while one year later ($\sim$80 stellar rotations) during sector 38 the stellar surface hosted dark spots in co-rotation with the star. This hypothesis is consistent with Hümmerich et al. (2018), according to which magnetic chemically peculiar stars may show complex photometric patterns due to surface inhomogeneities. We used a similar framework to extract the secondary eclipse signal of WASP-178: we simultaneously fit segments of the TESS LCs centered on the planetary eclipses keeping fixed the ephemeris of WASP-178 to the values listed in Table 3. We did not attempt the disjoint analysis of the two sectors as the expected eclipse depth is of the order of $\sim$100 ppm (see below), and the photometric precision of the LCs is not good enough to appreciate differences in the eclipse depth with enough statistical evidence. The only free planetary parameter of the model is thus the eclipse depth, which turned out to be $\delta_{\rm ecl}=70\pm 40$ ppm (MAP: 70 ppm). The detrended phase- folded eclipses are shown in Fig 2 together with the MAP model. Figure 2: Best fit of the transits and eclipses observed by TESS. Left - Detrended and phase-folded planetary transits (top panel). The photometry of sector 11 and 38 are shown with different colors to emphasize the difference in transit depth (the photometry of sectors 11 and 38 is systematically offset respectively upwards and downwards with respect to the bestfit transit profile). The solid blue line is the best fit model, while the cyan lines represent 100 models corresponding to random samples of the MCMC fit. For the sake of comparison, the photometry of sector 38 and the theoretical models are rebinned to the same 30 min cadence of sector 11. The black dots represent the rebinned photometry. The corresponding O-C diagram is shown in the bottom panel. Right - Same as in the left panel, but centered on the eclipse. For clarity, we do not mark the two sectors with different colors. #### 4.1.2 Fit of the phase curve As a more advanced data analysis, we jointly fit the two TESS sectors to extract the full planetary PC (for a homogeneous analysis, we rebinned sector 38 to the same 30 min cadence of sector 11). The fitting model now includes the transits, secondary eclipses, the planetary PC, a Gaussian Process (GP) to fit the correlated noise in the data and, for each TESS sector, a long-term linear trend and a jitter term. Given the system parameters listed in Table 1, the expected amplitude of ellipsoidal variations (Morris 1985) and Doppler boosting (Barclay et al. 2012) in the planetary PC are 2.5 ppm and 1 ppm respectively, out of reach for the TESS photometry. Hence, for the sake of simplicity, we do not include them in the model fitting. We jointly fit the two TESS sectors, starting with the simplest model where the out of transit PC is flat (that is, we assumed that the planetary PC is not detectable) and the putative stellar rotation is modeled as GP with the SHO kernel555Provided by the celerite2 python package version 0.2.1 (Foreman- Mackey et al. 2017; Foreman-Mackey 2018), indicated for quasi-periodic signals. Checking the residuals of the fit we notice that some aperiodic correlated noise is left. We thus increased the complexity of the GP model by adding a Matérn3/2 kernel to capture the remaining correlated noise adding only two free parameters to the model. This composite GP model turned out to ensure better convergence to the fit of the model. Finally, we also included the planetary PC. Assuming that the planet is basically composed of two homogeneous dayside and nightside, the PC is in principle the combination of three components: the PC due to reflected light (with amplitude $A_{\rm refl}$), the PC of the planetary night side (with amplitude $A_{\rm n}$) and the PC of the planetary day side, whose amplitude $A_{\rm d}$ is parameterized such to be larger than $A_{\rm n}$ by an increment $\delta_{\rm d}$ (this parametrization avoids the nonphysical case of a planetary night side brighter than the dayside). The three terms are modeled assuming Lambert’s cosine law. For the reflected PC, the amplitude can be expressed as: $A_{\rm refl}=A_{\rm g}\left(\frac{R_{\rm p}}{a}\right)^{2},$ (1) where $A_{\rm g}$ is the planetary geometric albedo (Seager 2010). Assuming a Lambertian reflective planetary surface, the geometric albedo is fixed by the Bond albedo $A_{\rm B}$ by $A_{\rm g}$=$\frac{2}{3}A_{\rm B}$. In the optimistic case of a perfectly reflecting body ($A_{\rm B}$=1), the maximum amplitude for the reflected PC is thus $A_{\rm refl}=167\pm 5$ ppm, where we have used the system parameters in Table 3. With a slight re-adaptation of the formalism in Seager (2010), the amplitude thermal component $A_{\rm therm}$ can be expressed as: $A_{\rm therm}=\left(\frac{R_{\rm p}}{R_{\star}}\right)^{2}\frac{\int\eta(\lambda)B(\lambda,T)d\lambda}{\int\eta(\lambda)I_{\star}(\lambda,T_{\rm eff})d\lambda},$ (2) where $\eta(\lambda)$ is the optical throughput of the telescope and $I_{\star}(\lambda,T_{\rm eff})$ is the expected stellar intensity computed with the NextGen model (Hauschildt et al. 1999) corresponding to the effective temperature $T_{\rm eff}$. For the planetary thermal emission we lack the infrared information which can constrain the emission spectrum. We thus assume a black body spectrum $B(\lambda,T)$ at temperature $T$. The amplitude of the thermal emission from the dayside and the nightside depends on the respective temperatures, which we estimated following Cowan & Agol (2011). The expected temperature of the substellar point of WASP-178 b is: $T_{\rm 0}=T_{\rm eff}\sqrt{\frac{R_{\star}}{a}}=3530\pm 60\leavevmode\nobreak\ {\rm K},$ (3) where we used the stellar effective temperature in Table 1 and the scaled semi-major axis in Table 3. The nightside temperature $T_{\rm n}$ and the dayside temperature $T_{\rm d}$ depend on the Bond albedo $A_{B}$ and heat recirculation coefficient $\epsilon$: $\displaystyle T_{\rm n}$ $\displaystyle=T_{\rm 0}\left(1-A_{\rm B}\right)^{1/4}\left(\frac{\epsilon}{4}\right)^{1/4}$ (4) $\displaystyle T_{\rm d}$ $\displaystyle=T_{\rm 0}\left(1-A_{\rm B}\right)^{1/4}\left(\frac{2}{3}-\frac{5}{12}\epsilon\right)^{1/4}$ (5) The highest dayside temperature, corresponding to $A_{\rm B}$=0 and no heat recirculation ($\epsilon=0$), is: $T_{\rm d}=T_{\rm 0}\left(\frac{2}{3}\right)^{1/4}=3190\pm 60\leavevmode\nobreak\ {\rm K},$ (6) while the maximum nightside temperature, obtained in the case of full energy recirculation ($\epsilon=1$), is: $T_{\rm n}=T_{\rm 0}\left(\frac{1}{4}\right)^{1/4}=2500\pm 40\leavevmode\nobreak\ {\rm K}.$ (7) These maximum temperatures lead to the upper limit on the dayside ($A_{\rm d}$) and nightside ($A_{\rm n}$) thermal emission amplitude of: $A_{\rm d}=225\pm 27\leavevmode\nobreak\ {\rm ppm};A_{\rm n}=48\pm 7\leavevmode\nobreak\ {\rm ppm}.$ (8) In Fig. 3 we plot the upper limits for the three PCs discussed so far. The figure shows that the three components may reach amplitudes of comparable orders of magnitude, meaning that none of them can in principle be neglected in the extraction of the planetary PC. We note that the two components belonging to the dayside (reflection and thermal emission) have the same shape, thus it is not possible to disentangle them. To simplify the model and avoid the degeneracy between dayside reflection and emission, we thus artificially set $A_{\rm refl}=0$ and let $A_{\rm d}$ absorb the whole signal belonging to the planetary dayside. Figure 3: Upper limits on the reflected PC, dayside thermal PC and nightside thermal PC for WASP-178 b. The Akaike Information Criterion (AIC, Burnham & Anderson 2002) favors the model that includes the planetary PCs. The simpler model with flat out-of- transit PC has a relative likelihood of 55% and cannot be rejected. The posterior distributions of the common parameters (system ephemeris and planet- to-star radii ratio) between the two models do not differ significantly. For both models, the SHO GP has a frequency $\nu\sim$0.30 d-1, an amplitude $\sim$70 ppm and a timescale $\sim$11 d. Frequency $\nu$ and the amplitude of the quasi-periodic GP are consistent with the frequency and amplitude of the periodic signal found in the periodogram (Sect. 4.1). While the correspondence between the orbital period of WASP-178 b and the periodicity of the GP suggests that the red noise in the data is of planetary origin, we do not have other strong evidence that support this hypothesis. On the contrary, the planetary origin seems odd for the reasons discussed in Sect. 4.1. The aperiodic red noise fitted as a Matérn3/2 GP has an amplitude $\sim$150 ppm and a timescale of $\sim$50 min. Despite the aperiodic red noise evolves on timescales shorter than the transit duration ($T_{\rm 14}$=3.53 hr), the retrieved posterior distributions of the orbital parameters are similar to the ones obtained in Sect. 4.1.1. Regarding the retrieval of the planetary PCs, we derived an upper limit of $A_{\rm n}<$174 ppm at the 99.9% confidence level, while for the dayside we derived a 1$\sigma$ confidence band of 110$\pm$40 ppm which includes both the reflection and thermal emission components. We remark that the $A_{\rm d}$ parameter corresponds in our model to the eclipse depth $\delta_{\rm ecl}$ and is consistent within uncertainties with the estimate reported in Table 3. The fit of the full TESS LC thus confirms the results previously obtained in Sect. 4.1.1. Given the similarities between the results obtained with the two approaches, we preferred the former one as it is less prone to interference between the data detrending and the extraction of the orbital parameters and the eclipse depth. ### 4.2 CHEOPS photometry To analyze the CHEOPS LCs we use the same approach as for the fit of the eclipses observed by TESS (Sect. 4.1.1), with an additional module in the fitting model that takes into account the systematics in the CHEOPS data. As a matter of fact, CHEOPS photometry is affected by variable contamination from background stars in the field of view (e.g. Lendl et al. 2020; Deline et al. 2022; Hooton et al. 2022; Wilson et al. 2022; Scandariato et al. 2022). This variability is due to the interplay of the asymmetric PSF with the rotation of the field of view (Benz et al. 2021). PIPE by design uses nominal magnitudes and coordinates of the stars in the field to fit the stellar PSFs, but residual correlated noise is present in the LCs due to inaccuracies in the assumptions. This signal is phased with the roll angle of the CHEOPS satellite and, in the case of WASP-178, is clearly visible in the GLS periodogram of the data together with its harmonics. To remove this signal, we included in our algorithm a module which fits independently for each visit the harmonic expansions of the telecope’s orbital period and its harmonics. A posteriori, we found that the roll angle phased modulation is adequately suppressed (that is, there is no peak in the periodograms at the frequencies in the harmonic series) if we include in the model the fundamental harmonic and its first two harmonics. We also noticed that the photometry is significantly correlated with the coordinates of the centroid of the PSF on the detector. To decorrelate against this instrumental jitter we thus included in the model a bi-linear function of the centroid coordinates. Finally, to take into account a white noise not included in the formal photometric uncertainties, we add to our model a diagonal GP kernel of the form: $k(t_{i},t_{j})=j_{v}^{2}\delta_{i,j},$ (9) with an independent jitter term $j_{v}$ for each CHEOPS visit. As for the fit of the TESS data (Sect. 4.1.1), we search the bestfit parameters through likelihood-maximization in a MCMC framework, and we find $\rm\delta_{ecl}=70\pm 20$ ppm. The ephemeris of the planet were locked to the MAP values listed in Table 3, which guarantee an uncertainty on the eclipse time less than 23 s for each CHEOPS visit. The phase-folded data detrended against stellar and instrumental correlated noise are shown in Fig. 4. Figure 4: Same as the right panel in Fig. 2 for the eclipses of WASP-178 observed by CHEOPS. ## 5 Discussion In the previous sections we have analyzed the TESS and CHEOPS PCs of WASP-178 b in order to extract its planetary eclipse depth $\delta_{\rm ecl}$, obtaining respectively $70\pm 40$ ppm and $70\pm 20$ ppm. These two measurements take into account both the reflection from the planetary dayside and its thermal emission. With the assumptions discussed below, the combined information on the eclipse depth from two different instruments allows to disentangle these two contributions. For a given telescope with passband $\eta(\lambda)$ we can combine Eq. 1 and Eq. 2 obtaining: $\delta_{\rm ecl}=A_{\rm g}\left(\frac{R_{\rm p}}{a}\right)^{2}+\left(\frac{R_{\rm p}}{R_{\star}}\right)^{2}\frac{\int\eta(\lambda)B(\lambda,T_{\rm d})d\lambda}{\int\eta(\lambda)I_{\star}(\lambda,T_{\rm eff})d\lambda},$ (10) where $T_{\rm d}$ is the planetary dayside temperature. Solving for $A_{\rm g}$, and indicating with the subscripts $C$ (for CHEOPS) and $T$ (for TESS) the wavelength-dependent quantities, we obtain: $\begin{cases}A_{\rm g}^{C}=\left(\frac{a}{R_{\star}}\right)^{2}\left[\frac{\delta_{\rm ecl}^{C}}{\left(\frac{R_{\rm p}}{R_{\star}}\right)^{2}}-\frac{\int\eta^{C}(\lambda)B(\lambda,T_{\rm d})d\lambda}{\int\eta^{C}(\lambda)I_{\star}(\lambda,T_{\rm eff})d\lambda}\right],\\\ A_{\rm g}^{T}=\left(\frac{a}{R_{\star}}\right)^{2}\left[\frac{\delta_{\rm ecl}^{T}}{\left(\frac{R_{\rm p}}{R_{\star}}\right)^{2}}-\frac{\int\eta^{T}(\lambda)B(\lambda,T_{\rm d})d\lambda}{\int\eta^{T}(\lambda)I_{\star}(\lambda,T_{\rm eff})d\lambda}\right].\end{cases}$ (11) In the most general case, $A_{g}^{C}$ and $A_{g}^{T}$ differ according to the planetary reflection spectrum. Since we do not have any spectroscopic analysis of the dayside of WASP-178 b, we define the proportionality coefficient $\alpha=A_{g}^{T}/A_{g}^{C}$ to account for differences between the two passbands. Solving Eq. 11 for $T_{\rm d}$ we thus obtain the implicit function: $\frac{\alpha\delta_{\rm ecl}^{C}-\delta_{\rm ecl}^{T}}{\left(\frac{R_{\rm p}}{R_{\star}}\right)^{2}}+\frac{\int\eta^{T}(\lambda)B(\lambda,T_{\rm d})d\lambda}{\int\eta^{T}(\lambda)I_{\star}(\lambda)d\lambda}-\alpha\frac{\int\eta^{C}(\lambda)B(\lambda,T_{\rm d})d\lambda}{\int\eta^{C}(\lambda)I_{\star}(\lambda)d\lambda}=0.$ (12) We initially assumed the simplest scenario of a gray albedo spectrum ($\alpha$=1) in the spectral range covered by CHEOPS and TESS (the respective bandpasses are shown in Fig. 6). In this scenario, we expect that the contribution of reflection to the eclipse depth is the same in the CHEOPS and TESS passbands. We also expect that the thermal emission from a $T\simeq$3000 K black body has a larger contribution in the TESS passband, because it favors redder wavelengths compared with CHEOPS (see Fig. 6). To estimate the dayside temperature $T_{\rm d}$ that best explains our best observations together with its corresponding uncertainty, we randomly extracted 10 000 steps from the Monte Carlo chains obtained in the fit of the transit and eclipse LCs (Sect. 4.1.1 and 4.2). We also generated 10 000 samples of the stellar effective temperature using a normal distribution with mean and standard deviations as reported in Table 1. For each of the 10 000 samples we thus numerically solved Eq. 12 using the fsolve method of the scipy Python package, thus obtaining 10 000 samples of $T_{\rm d}$. Then, we plugged all the samples back into Eq. 11 to derive $A_{g}^{C}$(=$A_{g}^{T}$). The root finding algorithm failed for $\sim$50% of the samples. These failures correspond to the combination of parameters that prevent Eq. 12 from having a $T_{\rm d}$ root in the domain of real numbers. This happens in particular for the samples where $\delta_{\rm ecl}^{T}<\delta_{\rm ecl}^{C}$ that, as discussed above, are not consistent with the gray albedo hypothesis. We thus reject these samples and show in Fig. 7 the $A_{g}^{C}$–$T_{\rm d}$ density map of the remaining ones. Our results indicate $A_{g}^{C}=0.2\pm 0.1$ and $T_{\rm d}=2400\pm 300$ K, consistent with the temperature-pressure profile derived by (Lothringer et al. 2022) at pressure higher that 1$\rm\mu$bar by means of ultraviolet transmission spectroscopy. Our result also follows the general trend of $A_{\rm g}$ increasing with $T_{\rm d}$ indicated by Wong et al. (2021) (Fig. 5). Figure 5: Adaptation of Fig. 10 in Wong et al. (2021) including our analysis of WASP-178 b (in blue). The green squares indicate the systems from the first and second year of the TESS primary mission. The black circles indicate the Kepler-/CoRoT-band geometric albedos for the targets that were observed by those missions. By definition, the geometric albedo $A_{\rm g}$ refers to the incident light reflected back to the star at a given wavelength (or bandpass). Integrating at all angles, the spherical albedo $A_{\rm S}$ is related to $A_{\rm g}$ by the phase integral $q$: $A_{\rm S}$=$q$$A_{\rm g}$ (see for example Seager 2010). Depending on the scattering law, exoplanetary atmospheres have 1¡$q$¡1.5 (Pollack et al. 1986; Burrows & Orton 2010). Unfortunately, for the reasons explained in Sect. 4.1.2, we could not extract a robust phase curve for WASP-178 b, hence we could not place any better constraint on $q$. In the following we thus consider the two limiting scenarios $A_{S}^{\rm min}=A_{g}$ and $A_{S}^{\rm max}=1.5A_{g}$. The Bond albedo $A_{\rm B}$ is computed as the average of $A_{\rm S}$ weighted over the incident stellar spectrum: $A_{\rm B}=\frac{\int_{0}^{\infty}A_{\rm S}(\lambda)I_{\star}(\lambda)d\lambda}{\int_{0}^{\infty}I_{\star}(\lambda)d\lambda}.$ (13) The conversion into $A_{\rm B}$ thus relies on the measurement of $A_{\rm S}$ across the stellar spectrum. In this study we only covered the optical and near-infrared part of the spectrum, as shown in Fig. 6, and we lack the necessary information in the UV, mid, and far infrared domains. Following the approach of Schwartz & Cowan (2015), we explore the scenario of minimum Bond albedo $A_{\rm B}^{\rm min}$ obtained through Eq. 13 assuming $A_{\rm S}$=$A_{\rm S}^{\rm min}$ in the spectral range covered by CHEOPS and TESS and $A_{\rm S}$=0 otherwise. The opposite limiting case assumes $A_{\rm S}$$(\lambda)=A_{\rm S}^{\rm max}$ at all wavelength, which leads to $A_{\rm B}^{\rm max}=A_{\rm S}^{\rm max}$. To compute the integrals in Eq. 13, we used the synthetic spectrum in the BT-Settl library corresponding to the parameters of WASP-178 (Allard et al. 2012). Figure 6: BT-Settl synthetic stellar spectrum for WASP-178 together with the CHEOPS and TESS passbands. Under the hypothesis that $\alpha=1$, we obtained samples of $A_{\rm g}^{C}=A_{\rm g}^{T}=$$A_{\rm g}$, which now translate into samples of $A_{\rm B}^{\rm min}$ and $A_{\rm B}^{\rm max}$ with the assumptions discussed above. Moreover, inverting Eq. 5 yields to: $\epsilon=\frac{1}{5}\left[8-\left(\frac{T_{\rm d}}{T_{\rm eff}}\right)^{4}\left(\frac{a}{R_{\star}}\right)^{2}\frac{12}{1-A_{\rm B}}\right].$ (14) Equation 14 indicates that $\epsilon$ is a decreasing function of $A_{\rm B}$. Plugging in the samples of $T_{\rm d}$, Teff, $a/R_{\star}$ and alternatively $A_{\rm B}^{\rm min}$ and $A_{\rm B}^{\rm max}$, we obtained the corresponding samples $\epsilon^{\rm max}$ and $\epsilon^{\rm min}$. The albedo–recirculation density maps of the two scenarios are shown in the right panel of Fig. 7. Unsurprisingly, we find that the upper limit on $A_{\rm B}$ is larger ($\sim$0.6) in the scenario where all the factors concur to push up the reflectivity of the atmosphere (high $q$ and maximum $A_{\rm S}$), and it decreases to $\sim$0.3 in the opposite scenario of minimum reflectivity. The recirculation coefficient $\epsilon$ tends towards large values, eventually exceeding 1 due to measurement uncertainties, in both scenarios. While $\epsilon>1$ is physically meaningless, still the posterior distribution indicates a high level of atmospheric energy recirculation in both cases of minimum and maximum albedo ($\epsilon^{\rm max}=1.0\pm 0.3$ and $\epsilon^{\rm min}=0.8\pm 0.3$ respectively). Following different atmospheric models (e.g., Perez-Becker & Showman 2013; Komacek & Showman 2016; Schwartz et al. 2017; Parmentier & Crossfield 2018), WASP-178 b is in a temperature regime where heat recirculation of zonal winds is suppressed and recirculation efficiency is expected to be low. Nonetheless, Zhang et al. (2018) collected observational evidence that Hot Jupiters with irradiation temperatures similar to the one of WASP-178 b are characterized by efficient day-to-night recirculation. Using the global circulation models of Kataria et al. (2016), they ascribed this efficiency to the presence of zonal winds, that also explain the eastward offset of the PC of the same planets. While the theoretical predictions on the offset (and the recirculation efficiency correspondingly) likely overestimate the truth (see Fig. 15 in Zhang et al. 2018), still there is an indication that zonal winds might indeed explain the energy transfer from the dayside to the nightside of WASP-178 b. Unfortunately, as discussed in Sect.4.1.2, it is difficult to disentangle the planetary phase curve and the stellar variability. By consequence, any measurement of the phase curve offset is precluded. Another possibility is that ionized winds flow from the dayside, where temperatures higher that 2500 K leads to complete dissociation of H2, to the nightside, where lower temperatures allow the molecular recombination and the consequent energy release. This scenario is supported by several recent studies (Bell & Cowan 2018; Tan & Komacek 2019; Mansfield et al. 2020; Helling et al. 2021, 2023). Figure 7: Left panel - $A_{g}^{C}$–$T_{\rm d}$ density maps in the case of $\alpha=1$. Right panel - Albedo–recirculation density maps assuming maximum $A_{\rm B}$ (red contours) and minimum $A_{\rm B}$ (blue contours). According to the synthetic models computed by Sudarsky et al. (2000), the albedo spectrum of highly irradiated giant planets is expected to show molecular absorption bands by H2O at around 1 $\mu m$ and longer wavelength. Since the passband of TESS is more sensitive in the near-infrared than CHEOPS, it is plausible to assume that the geometric albedo in the TESS passband is lower than for CHEOPS. In order to assess how the assumption on $\alpha$ affects our results, we re-run our analysis assuming an extreme $\alpha=0.5$. The results are shown in Fig. 8: we find almost the same posterior distribution for dayside temperature, albedo and recirculation, indicating that the most important source of uncertainties are not the assumptions we make but the measurement uncertainties on the eclipse depth extracted from the CHEOPS and TESS LCs. Figure 8: Same as in Fig. 7 in the case $\alpha=0.5$. ## 6 Summary and conclusion In this work we have analyzed the space-borne photometry of the WASP-178 system obtained with TESS and CHEOPS. For both telescopes we have tailored the photometric extraction in order to avoid the strong contamination by a background eclipsing binary. We have found evidence that the stellar host is rotating with a period of $\sim$3 days, consistent with the orbital period of its HJ, and that it is seen in a pole-on geometry. The similarity between the stellar rotation period and the orbital period of the planet, coupled with the quality of the TESS data, does not allow a robust analysis of the full PC of the planet, nor we could put strong constraints on the planetary nightside emission. Nonetheless, focusing on the transit and eclipse events, we could update the ephemeris of the planet and measure an eclipse depth of 70$\pm$40 ppm and 70$\pm$20 ppm in the TESS and CHEOPS passband respectively. The joint analysis of the eclipse depth measured in the two passbands allowed us to constrain the temperature in the 2250–2750 K range, consistent with the temperature-pressure profile derived by (Lothringer et al. 2022) at pressure higher that 1$\rm\mu$bar. Moreover, we could also constrain the optical geometric albedo $A_{\rm g}$¡0.4, which fits the general increasing albedo with stellar irradiation indicated by Wong et al. (2021). Finally, we found indication of an efficient atmospheric heat recirculation of WASP-178 b. If confirmed, this evidence challenges the models that predict a decreasing heat transport by zonal winds as the equilibrium temperature increases (e.g., Perez-Becker & Showman 2013; Komacek & Showman 2016; Schwartz et al. 2017; Parmentier & Crossfield 2018). Conversely, WASP-178 b is an interesting target to test atmospheric models where heat transport is granted by ionized winds from the dayside to the nightside, where the recombination of H2 takes place (e.g. Mansfield et al. 2020; Helling et al. 2021, 2023). Additional observations of WASP-178 b are needed to better constrain its atmospheric recirculation and rank current competing models. The WASP-178 system is going to be observed again by TESS in May 2023, but this is not expected to add much to the two sectors analyzed in this work. Conversely, dedicated observations from larger space observatories might provide helpful insights of the physical and chemical equilibrium of the atmosphere of WASP-178 b. ###### Acknowledgements. We thank the referee, N. B. Cowan, for his valuable comments and suggestions. CHEOPS is an ESA mission in partnership with Switzerland with important contributions to the payload and the ground segment from Austria, Belgium, France, Germany, Hungary, Italy, Portugal, Spain, Sweden, and the United Kingdom. The CHEOPS Consortium would like to gratefully acknowledge the support received by all the agencies, offices, universities, and industries involved. Their flexibility and willingness to explore new approaches were essential to the success of this mission. IPa, GSc, VSi, LBo, GBr, VNa, GPi, and RRa acknowledge support from CHEOPS ASI-INAF agreement n. 2019-29-HH.0. ML acknowledges support of the Swiss National Science Foundation under grant number PCEFP2_194576. This work was also partially supported by a grant from the Simons Foundation (PI Queloz, grant number 327127). S.G.S. acknowledge support from FCT through FCT contract nr. CEECIND/00826/2018 and POPH/FSE (EC). ABr was supported by the SNSA. ACCa and TWi acknowledge support from STFC consolidated grant numbers ST/R000824/1 and ST/V000861/1, and UKSA grant number ST/R003203/1. V.V.G. is an F.R.S-FNRS Research Associate. YAl acknowledges support from the Swiss National Science Foundation (SNSF) under grant 200020_192038. We acknowledge support from the Spanish Ministry of Science and Innovation and the European Regional Development Fund through grants ESP2016-80435-C2-1-R, ESP2016-80435-C2-2-R, PGC2018-098153-B-C33, PGC2018-098153-B-C31, ESP2017-87676-C5-1-R, MDM-2017-0737 Unidad de Excelencia Maria de Maeztu-Centro de Astrobiologí­a (INTA-CSIC), as well as the support of the Generalitat de Catalunya/CERCA programme. The MOC activities have been supported by the ESA contract No. 4000124370. S.C.C.B. acknowledges support from FCT through FCT contracts nr. IF/01312/2014/CP1215/CT0004. XB, SC, DG, MF and JL acknowledge their role as ESA-appointed CHEOPS science team members. This project was supported by the CNES. The Belgian participation to CHEOPS has been supported by the Belgian Federal Science Policy Office (BELSPO) in the framework of the PRODEX Program, and by the University of Liège through an ARC grant for Concerted Research Actions financed by the Wallonia-Brussels Federation. L.D. is an F.R.S.-FNRS Postdoctoral Researcher. This work was supported by FCT - Fundação para a Ciência e a Tecnologia through national funds and by FEDER through COMPETE2020 - Programa Operacional Competitividade e Internacionalizacão by these grants: UID/FIS/04434/2019, UIDB/04434/2020, UIDP/04434/2020, PTDC/FIS-AST/32113/2017 & POCI-01-0145-FEDER- 032113, PTDC/FIS-AST/28953/2017 & POCI-01-0145-FEDER-028953, PTDC/FIS-AST/28987/2017 & POCI-01-0145-FEDER-028987, O.D.S.D. is supported in the form of work contract (DL 57/2016/CP1364/CT0004) funded by national funds through FCT. B.-O. D. acknowledges support from the Swiss State Secretariat for Education, Research and Innovation (SERI) under contract number MB22.00046. This project has received funding from the European Research Council (ERC) under the European Union’s Horizon 2020 research grant agreement No 724427). It has also been carried out in the frame of the National Centre for Competence in Research PlanetS supported by the Swiss National Science Foundation (SNSF). DE acknowledges financial support from the Swiss National Science Foundation for project 200021_200726. and innovation programme (project Four Aces. MF and CMP gratefully acknowledge the support of the Swedish National Space Agency (DNR 65/19, 174/18). DG gratefully acknowledges financial support from the CRT foundation under Grant No. 2018.2323 “Gaseousor rocky? Unveiling the nature of small worlds”. M.G. is an F.R.S.-FNRS Senior Research Associate. MNG is the ESA CHEOPS Project Scientist and Mission Representative, and as such also responsible for the Guest Observers (GO) Programme. MNG does not relay proprietary information between the GO and Guaranteed Time Observation (GTO) Programmes, and does not decide on the definition and target selection of the GTO Programme. SH gratefully acknowledges CNES funding through the grant 837319. KGI is the ESA CHEOPS Project Scientist and is responsible for the ESA CHEOPS Guest Observers Programme. She does not participate in, or contribute to, the definition of the Guaranteed Time Programme of the CHEOPS mission through which observations described in this paper have been taken, nor to any aspect of target selection for the programme. This work was granted access to the HPC resources of MesoPSL financed by the Region Ile de France and the project Equip@Meso (reference ANR-10-EQPX-29-01) of the programme Investissements d’Avenir supervised by the Agence Nationale pour la Recherche. PM acknowledges support from STFC research grant number ST/M001040/1. IRI acknowledges support from the Spanish Ministry of Science and Innovation and the European Regional Development Fund through grant PGC2018-098153-B- C33, as well as the support of the Generalitat de Catalunya/CERCA programme. GyMSz acknowledges the support of the Hungarian National Research, Development and Innovation Office (NKFIH) grant K-125015, a a PRODEX Experiment Agreement No. 4000137122, the Lendület LP2018-7/2021 grant of the Hungarian Academy of Science and the support of the city of Szombathely. NAW acknowledges UKSA grant ST/R004838/1. NCS acknowledges funding by the European Union (ERC, FIERCE, 101052347). Views and opinions expressed are however those of the author(s) only and do not necessarily reflect those of the European Union or the European Research Council. Neither the European Union nor the granting authority can be held responsible for them. KWFL was supported by Deutsche Forschungsgemeinschaft grants RA714/14-1 within the DFG Schwerpunkt SPP 1992, Exploring the Diversity of Extrasolar Planets. In this work we use the python package PyDE available at https://github.com/hpparvi/PyDE. This research has made use of the SVO Filter Profile Service (http://svo2.cab.inta- csic.es/theory/fps/) supported from the Spanish MINECO through grant AYA2017-84089. ## References * Albrecht et al. (2021) Albrecht, S. H., Marcussen, M. L., Winn, J. N., Dawson, R. I., & Knudstrup, E. 2021, ApJ, 916, L1 * Allard (2014) Allard, F. 2014, in Exploring the Formation and Evolution of Planetary Systems, ed. M. Booth, B. C. Matthews, & J. R. Graham, Vol. 299, 271–272 * Allard et al. (2012) Allard, F., Homeier, D., & Freytag, B. 2012, Philosophical Transactions of the Royal Society of London Series A, 370, 2765 * Balona (2011) Balona, L. A. 2011, MNRAS, 415, 1691 * Barclay et al. (2012) Barclay, T., Huber, D., Rowe, J. F., et al. 2012, ApJ, 761, 53 * Béky et al. (2014) Béky, B., Holman, M. J., Kipping, D. M., & Noyes, R. W. 2014, ApJ, 788, 1 * Bell & Cowan (2018) Bell, T. J. & Cowan, N. B. 2018, The Astrophysical Journal, 857, L20 * Benz et al. (2021) Benz, W., Broeg, C., Fortier, A., et al. 2021, Experimental Astronomy, 51, 109 * Bertocco et al. (2020) Bertocco, S., Goz, D., Tornatore, L., et al. 2020, in Astronomical Society of the Pacific Conference Series, Vol. 527, Astronomical Data Analysis Software and Systems XXIX, ed. R. Pizzo, E. R. Deul, J. D. Mol, J. de Plaa, & H. Verkouter, 303 * Blackwell & Shallis (1977) Blackwell, D. E. & Shallis, M. J. 1977, MNRAS, 180, 177 * Böhm et al. (2015) Böhm, T., Holschneider, M., Lignières, F., et al. 2015, A&A, 577, A64 * Bonfanti et al. (2021) Bonfanti, A., Delrez, L., Hooton, M. J., et al. 2021, A&A, 646, A157 * Bonfanti et al. (2016) Bonfanti, A., Ortolani, S., & Nascimbeni, V. 2016, A&A, 585, A5 * Bonfanti et al. (2015) Bonfanti, A., Ortolani, S., Piotto, G., & Nascimbeni, V. 2015, A&A, 575, A18 * Brandeker et al. (2022) Brandeker, A., Heng, K., Lendl, M., et al. 2022, A&A, 659, L4 * Burnham & Anderson (2002) Burnham, K. P. & Anderson, D. R., eds. 2002, Basic Use of the Information-Theoretic Approach (New York, NY: Springer New York), 98–148 * Burrows & Orton (2010) Burrows, A. & Orton, G. 2010, in Exoplanets, ed. S. Seager, 419–440 * Castelli & Kurucz (2003) Castelli, F. & Kurucz, R. L. 2003, in IAU Symposium, Vol. 210, Modelling of Stellar Atmospheres, ed. N. Piskunov, W. W. Weiss, & D. F. Gray, A20 * Cowan & Agol (2011) Cowan, N. B. & Agol, E. 2011, ApJ, 729, 54 * Deline et al. (2022) Deline, A., Hooton, M. J., Lendl, M., et al. 2022, A&A, 659, A74 * Delrez et al. (2021) Delrez, L., Ehrenreich, D., Alibert, Y., et al. 2021, Nature Astronomy, 5, 775 * Demory et al. (2023) Demory, B. O., Sulis, S., Meier Valdés, E., et al. 2023, A&A, 669, A64 * Foreman-Mackey (2018) Foreman-Mackey, D. 2018, Research Notes of the American Astronomical Society, 2, 31 * Foreman-Mackey et al. (2017) Foreman-Mackey, D., Agol, E., Ambikasaran, S., & Angus, R. 2017, AJ, 154, 220 * Foreman-Mackey et al. (2013) Foreman-Mackey, D., Hogg, D. W., Lang, D., & Goodman, J. 2013, PASP, 125, 306 * Gaia Collaboration et al. (2021) Gaia Collaboration, Brown, A. G. A., Vallenari, A., et al. 2021, A&A, 649, A1 * Giacobbe et al. (2021) Giacobbe, P., Brogi, M., Gandhi, S., et al. 2021, Nature, 592, 205 * Hauschildt et al. (1999) Hauschildt, P. H., Allard, F., & Baron, E. 1999, ApJ, 512, 377 * Hellier et al. (2019) Hellier, C., Anderson, D. R., Barkaoui, K., et al. 2019, MNRAS, 490, 1479 * Helling et al. (2023) Helling, C., Samra, D., Lewis, D., et al. 2023, A&A, 671, A122 * Helling et al. (2021) Helling, C., Worters, M., Samra, D., Molaverdikhani, K., & Iro, N. 2021, A&A, 648, A80 * Hooton et al. (2022) Hooton, M. J., Hoyer, S., Kitzmann, D., et al. 2022, A&A, 658, A75 * Hoyer et al. (2020) Hoyer, S., Guterman, P., Demangeon, O., et al. 2020, A&A, 635, A24 * Hümmerich et al. (2018) Hümmerich, S., Mikulášek, Z., Paunzen, E., et al. 2018, A&A, 619, A98 * Kataria et al. (2016) Kataria, T., Sing, D. K., Lewis, N. K., et al. 2016, ApJ, 821, 9 * Kipping (2013) Kipping, D. M. 2013, MNRAS, 435, 2152 * Komacek & Showman (2016) Komacek, T. D. & Showman, A. P. 2016, ApJ, 821, 16 * Kurucz (1993) Kurucz, R. L. 1993, SYNTHE spectrum synthesis programs and line data (Astrophysics Source Code Library) * Lendl et al. (2020) Lendl, M., Csizmadia, S., Deline, A., et al. 2020, A&A, 643, A94 * Lothringer et al. (2022) Lothringer, J. D., Sing, D. K., Rustamkulov, Z., et al. 2022, Nature, 604, 49 * Mandel & Agol (2002) Mandel, K. & Agol, E. 2002, ApJ, 580, L171 * Mansfield et al. (2020) Mansfield, M., Bean, J. L., Stevenson, K. B., et al. 2020, ApJ, 888, L15 * Marigo et al. (2017) Marigo, P., Girardi, L., Bressan, A., et al. 2017, ApJ, 835, 77 * Morris et al. (2021) Morris, B. M., Delrez, L., Brandeker, A., et al. 2021, A&A, 653, A173 * Morris (1985) Morris, S. L. 1985, ApJ, 295, 143 * Parmentier & Crossfield (2018) Parmentier, V. & Crossfield, I. J. M. 2018, in Handbook of Exoplanets, ed. H. J. Deeg & J. A. Belmonte, 116 * Parviainen et al. (2022) Parviainen, H., Wilson, T. G., Lendl, M., et al. 2022, A&A, 668, A93 * Perez-Becker & Showman (2013) Perez-Becker, D. & Showman, A. P. 2013, The Astrophysical Journal, 776, 134 * Pollack et al. (1986) Pollack, J. B., Podolak, M., Bodenheimer, P., & Christofferson, B. 1986, Icarus, 67, 409 * Ricker et al. (2014) Ricker, G. R., Winn, J. N., Vanderspek, R., et al. 2014, in Society of Photo-Optical Instrumentation Engineers (SPIE) Conference Series, Vol. 9143, Space Telescopes and Instrumentation 2014: Optical, Infrared, and Millimeter Wave, ed. J. Oschmann, Jacobus M., M. Clampin, G. G. Fazio, & H. A. MacEwen, 914320 * Rodríguez Martínez et al. (2020) Rodríguez Martínez, R., Gaudi, B. S., Rodriguez, J. E., et al. 2020, AJ, 160, 111 * Salmon et al. (2021) Salmon, S. J. A. J., Van Grootel, V., Buldgen, G., Dupret, M. A., & Eggenberger, P. 2021, A&A, 646, A7 * Scandariato et al. (2017) Scandariato, G., Nascimbeni, V., Lanza, A. F., et al. 2017, A&A, 606, A134 * Scandariato et al. (2022) Scandariato, G., Singh, V., Kitzmann, D., et al. 2022, A&A, 668, A17 * Schanche et al. (2020) Schanche, N., Hébrard, G., Collier Cameron, A., et al. 2020, MNRAS, 499, 428 * Schwartz & Cowan (2015) Schwartz, J. C. & Cowan, N. B. 2015, MNRAS, 449, 4192 * Schwartz et al. (2017) Schwartz, J. C., Kashner, Z., Jovmir, D., & Cowan, N. B. 2017, ApJ, 850, 154 * Scuflaire et al. (2008) Scuflaire, R., Théado, S., Montalbán, J., et al. 2008, Ap&SS, 316, 83 * Seager (2010) Seager, S. 2010, Exoplanet Atmospheres: Physical Processes * Sikora et al. (2020) Sikora, J., Wade, G. A., & Rowe, J. 2020, MNRAS, 498, 2456 * Sing et al. (2016) Sing, D. K., Fortney, J. J., Nikolov, N., et al. 2016, Nature, 529, 59 * Singh et al. (2022) Singh, V., Bonomo, A. S., Scandariato, G., et al. 2022, A&A, 658, A132 * Skrutskie et al. (2006) Skrutskie, M. F., Cutri, R. M., Stiening, R., et al. 2006, AJ, 131, 1163 * Stevenson et al. (2017) Stevenson, K. B., Line, M. R., Bean, J. L., et al. 2017, AJ, 153, 68 * Sudarsky et al. (2000) Sudarsky, D., Burrows, A., & Pinto, P. 2000, ApJ, 538, 885 * Szabó et al. (2021) Szabó, G. M., Gandolfi, D., Brandeker, A., et al. 2021, A&A, 654, A159 * Taffoni et al. (2020) Taffoni, G., Becciani, U., Garilli, B., et al. 2020, in Astronomical Society of the Pacific Conference Series, Vol. 527, Astronomical Data Analysis Software and Systems XXIX, ed. R. Pizzo, E. R. Deul, J. D. Mol, J. de Plaa, & H. Verkouter, 307 * Tan & Komacek (2019) Tan, X. & Komacek, T. D. 2019, ApJ, 886, 26 * Wilson et al. (2022) Wilson, T. G., Goffo, E., Alibert, Y., et al. 2022, MNRAS, 511, 1043 * Winn et al. (2010) Winn, J. N., Fabrycky, D., Albrecht, S., & Johnson, J. A. 2010, ApJ, 718, L145 * Wong et al. (2020) Wong, I., Benneke, B., Shporer, A., et al. 2020, The Astronomical Journal, 159, 104 * Wong et al. (2021) Wong, I., Kitzmann, D., Shporer, A., et al. 2021, AJ, 162, 127 * Wright et al. (2010) Wright, E. L., Eisenhardt, P. R. M., Mainzer, A. K., et al. 2010, AJ, 140, 1868 * Zechmeister & Kürster (2009) Zechmeister, M. & Kürster, M. 2009, A&A, 496, 577 * Zhang et al. (2018) Zhang, M., Knutson, H. A., Kataria, T., et al. 2018, AJ, 155, 83 ## Appendix A Posterior distributions of the model parameters from the fit of the TESS transit light curves Figure 9: Corner plot of the MCMC chains of planetary parameters from the fit of the TESS transits (see Sect. 4.1.1). In each plot, the solid blue lines mark the MAP values. ## Appendix B Analysis of the TESS transits in high cadence In Sect. 4.1 we analyzed the TESS photometry at low cadence (30 min), and we found a significant difference between the transit depth between sectors 11 and 38. We speculated that the difference in transit depth is due to photospheric variability, which manifests also as a quasi-periodic signal in the LC of sector 38. This hypothesis is supported by the growing evidence that A-type stars show rotational signals due to photospheric inhomogeneities (e.g., Balona 2011; Böhm et al. 2015; Sikora et al. 2020). This scenario is further supported by the fact that the WASP-178 is classified as an Am star by Hellier et al. (2019). To test if the visible hemisphere of the star hosts dark spots, we searched for transit anomalies in the TESS LCs, that are localized bumps in the residuals of the transit fits. If present, they indicate that the planet’s projection on the stellar surface crosses a darker area compared to the quiescent photosphere (see for example Béky et al. 2014; Scandariato et al. 2017). To this purpose, the low cadence photometry analyzed in Sect. 4.1 is of little help, as a finer time sampling is needed. We thus compared the high cadence (2 min) TESS photometry of sector 38 with the bestfit model discussed in Sect. 4.1.2, computed using $R_{\rm p}/R*$=0.1141 (see Sect. 4.1.1). This comparison does not show any bump inside the transits (Figs. 10–13), and we concluded that in the eight transits observed in sector 38 there is no spot- crossing event. We also remark that, in contrast with Rodríguez Martínez et al. (2020), we did not detect any systematic asymmetry in the transit profile. The stellar flux distribution along the transit path is symmetric with respect to the transit center. This either indicates that the star does not show any gravity darkening, which is unlikely if the stellar rotation period of $\sim$3.2 days is confirmed, or supports the hypothesis that the star is in a pole-on geometry, which leads to a radially symmetric flux distribution over the visible stellar hemisphere. Figure 10: Detrended short cadence LC of the first (left panel) and second (right panel) transit observed in TESS sector 38. In each panel, the top plot shows the model computed with the MAP parameters in Table 3 as a blue solid line. The bottom plots show the residuals of the short cadence photometry with respect to the planetary model shown in the top panel. As a guide, we plot with a blue line the smoothing of the residuals obtained with a Savitzky-Golay filter. In all plots, the vertical dashed lines mark the first and fourth contacts. Figure 11: Detrended short cadence LC of the third (left panel) and fourth (right panel) transit observed in TESS sector 38. Details are the same as in Fig. 10. Figure 12: Detrended short cadence LC of the fifth (left panel) and sixth (right panel) transit observed in TESS sector 38. Details are the same as in Fig. 10. Figure 13: Detrended short cadence LC of the seventh (left panel) and eighth (right panel) transit observed in TESS sector 38. Details are the same as in Fig. 10. *[LC]: Light Curve *[LCs]: Light Curve *[FFIs]: Full Frame Images *[TPFs]: Target Pixel File *[CBVs]: Co-trending Basis Vectors *[DRP]: Data Reduction Pipeline *[PSF]: Point Spread Function *[IRFM]: Infra-Red Flux Method *[MCMC]: Markov Chain Monte Carlo *[GLS]: Generalized Lomb-Scargle *[FAP]: False Alarm Probability *[PC]: Phase Curve *[LD]: Limb Darkening *[MAP]: Maximum A Posteriori *[GP]: Gaussian Process *[PCs]: Phase Curve *[PSFs]: Point Spread Function *[HJ]: Hot Jupiter
# Representations of Materials for Machine Learning111Accepted for publication in Annual Review of Materials Research Volume 53, https://www.annualreviews.org/. James Damewood Department of Materials Science and Engineering, Massachusetts Institute of Technology, 77 Massachusetts Avenue, Cambridge, USA, 02129 Jessica Karaguesian Department of Materials Science and Engineering, Massachusetts Institute of Technology, 77 Massachusetts Avenue, Cambridge, USA, 02129 Center for Computational Science and Engineering, Massachusetts Institute of Technology, 77 Massachusetts Avenue, Cambridge, MA, USA, 02139 Jaclyn R. Lunger Department of Materials Science and Engineering, Massachusetts Institute of Technology, 77 Massachusetts Avenue, Cambridge, USA, 02129 Aik Rui Tan Department of Materials Science and Engineering, Massachusetts Institute of Technology, 77 Massachusetts Avenue, Cambridge, USA, 02129 Mingrou Xie Department of Materials Science and Engineering, Massachusetts Institute of Technology, 77 Massachusetts Avenue, Cambridge, USA, 02129 Department of Chemical Engineering, Massachusetts Institute of Technology, 77 Massachusetts Avenue, Cambridge, MA, USA, 02139 Jiayu Peng Department of Materials Science and Engineering, Massachusetts Institute of Technology, 77 Massachusetts Avenue, Cambridge, USA, 02129 Rafael Gómez- Bombarelli<EMAIL_ADDRESS>Department of Materials Science and Engineering, Massachusetts Institute of Technology, 77 Massachusetts Avenue, Cambridge, USA, 02129 ###### Abstract High-throughput data generation methods and machine learning (ML) algorithms have given rise to a new era of computational materials science by learning relationships among composition, structure, and properties and by exploiting such relations for design. However, to build these connections, materials data must be translated into a numerical form, called a representation, that can be processed by a machine learning model. Datasets in materials science vary in format (ranging from images to spectra), size, and fidelity. Predictive models vary in scope and property of interests. Here, we review context-dependent strategies for constructing representations that enable the use of materials as inputs or outputs of machine learning models. Furthermore, we discuss how modern ML techniques can learn representations from data and transfer chemical and physical information between tasks. Finally, we outline high-impact questions that have not been fully resolved and thus, require further investigation. ###### Contents 1. 1 INTRODUCTION 2. 2 STRUCTURAL FEATURES FOR ATOMISTIC GEOMETRIES 1. 2.1 Local Descriptors 2. 2.2 Global Descriptors 3. 2.3 Topological Descriptors 3. 3 LEARNING ON PERIODIC CRYSTAL GRAPHS 4. 4 CONSTRUCTING REPRESENTATIONS FROM STOICHIOMETRY 5. 5 DEFECTS, SURFACES, AND GRAIN BOUNDARIES 6. 6 TRANSFERABLE INFORMATION BETWEEN REPRESENTATIONS 7. 7 GENERATIVE MODELS FOR INVERSE DESIGN 8. 8 DISCUSSION 1. 8.1 Trade-offs of Local and Global Structural Descriptors 2. 8.2 Prediction from Unrelaxed Crystal Prototypes 3. 8.3 Applicability of Compositional Descriptors 4. 8.4 Extensions of Generative Models ## 1 INTRODUCTION Energy and sustainability applications demand the rapid development of scalable new materials technologies. Big data and machine learning (ML) have been proposed as strategies to rapidly identify “needle-in-the-haystack” materials that have the potential for revolutionary impact. High-throughput experimentation platforms based on robotized laboratories can increase the efficiency and speed of synthesis and characterization. However, in many practical open problems, the number of possible design parameters is too large to be analyzed exhaustively. Virtual screening somewhat mitigates this challenge by using physics-based simulations to suggest the most promising candidates, reducing the cost but also the fidelity of the screens[1, 2, 3]. Over the past decade, hardware improvements, new algorithms, and the development of large-scale repositories of materials data [4, 5, 6, 7, 8, 9] have enabled a new era of ML methods. In principle, predictive ML models can identify and exploit nontrivial trends in high-dimensional data to achieve accuracy comparable with or superior to first-principles calculations, but with orders of magnitude reduction in cost. In practice, while a judicious model choice is helpful in moving towards this ideal, such ML methods are also highly dependent on the numerical inputs used to describe systems of interest—the so-called representations. Only when the representation is composed of a set of features and descriptors from which the desired physics and chemistry are emergent can the promise of ML be achieved. Thus, the problem that materials informatics researchers must answer is: how can we best construct this representation? Previous works have provided practical advice for constructing materials representations [10, 11, 12, 13, 14], namely that: (1) the similarity/difference between two data points should match the similarity/difference between representations of those two data points, (2) the representation should be applicable to the entire materials domain of interest, (3) the representation should be easier to calculate than the target property. Representations should reflect the degree of similarity between data points such that similar data have similar representations and as data points become more different their representations diverge. Indeed, the definition of similarity will depend on the application. Consider, as an example, a hypothetical model predicting the electronegativity of an element, excluding Nobel gases. One could attempt to train the model using atomic number as input, but this representation violates the above principle, as atoms with a similar atomic number can have significantly different electronegativities (e.g. fluorine and sodium), forcing the model to learn a sharply varying function whose changes appear at irregular intervals. Alternatively, a representation using period and group numbers would closely group elements with similar atomic radii and electron configurations. Over this new domain, the optimal prediction will result in a smoother function that is easier to learn. The approach used to extract representation features from raw inputs should be feasible over the entire domain of interest—all data points used in training and deployment. If data required to construct the representation is not available for a particular material, ML screening predictions cannot be made. Finally, for the ML approach to remain a worthwhile investment, the computational cost of obtaining representation features and descriptors for new data should be smaller than that of obtaining the property itself through traditional means, either experimentally or with first-principles calculations. If, for instance, accurately predicting a property calculated by density functional theory (DFT) with ML requires input descriptors obtained from DFT on the same structure and at the same level of theory, the machine learning model does not offer any benefit. A practicing materials scientist will notice a number of key barriers to forming property-informative representations that satisfy these criteria. First, describing behavior often involves quantifying structure-to-property relationships across length scales. The diversity of possible atomistic structure types considered can vary over space groups, supercell size, and disorder parameters. This challenge motivates researchers to develop flexible representations capable of capturing local and global information based on atomic positions. Beyond this idealized picture, predicting material performance relies upon understanding the presence of defects, the characteristics of the microstructure, and reactions at interfaces. Addressing these concerns requires extending previous notions of structural similarity or developing new specialized tools. Furthermore, atomistic structural information is not available without experimental validation or extensive computational effort [15, 16]. Therefore, when predictions are required for previously unexplored materials, models must rely on more readily available descriptors such as those based on elemental composition and stoichiometry. Lastly, due to experimental constraints, datasets in materials science can often be scarce, sparse, and restricted to relatively few and self-similar examples. The difficulty in constructing a robust representation in these scenarios has inspired strategies to leverage information from high-quality representations built for closely related tasks through transfer learning. In this review, we will analyze how representations of solid-state materials (Figure 1) can be developed given constraints on the format, quantity, and quality of available data. We will discuss the justification, benefits, and trade-offs of different approaches. This discussion is meant to highlight methods of particular interest rather than provide exhaustive coverage of the literature. We will discuss current limitations and open problems whose solutions would have high impact. In summary, we intend to provide readers with an introduction to current state of the field and exciting directions for future research. Figure 1: Summary of representations for perovskite SrTiO3. Top Left. 2D cross section of Voronoi decomposition. Predictive features can be constructed from neighbors and geometric shape of cells [17]. Middle Left. Crystal graph of SrTiO3 constructed assuming periodic boundary conditions and used as input to graph neural networks [18]. Bottom Left. Compositional data including concentrations and easily accessible atomic features including electronegativities and atomic radii [19]. Data taken from Reference [20]. Top Right. Deviations on a pristine bulk structure induced by an oxygen vacancy to predict formation energy [21]. Middle Right. Representations can be learned from large repositories using deep neural networks. The latent physical and chemical information can be leveraged in related but data-scare tasks. Bottom Right. Training of generative models capable of proposing new crystal structures by placing atoms in discretized volume elements [22, 23, 24, 25]. ## 2 STRUCTURAL FEATURES FOR ATOMISTIC GEOMETRIES Simple observations in material systems (e.g. higher ductility of face- centered cubic metals compared to body-centered cubic metals) have made it evident that material properties are highly dependent on crystal structure—from coordination and atomic ordering to broken symmetries and porosity. For a computational material scientist, this presents the question of how to algorithmically encode information from a set of atoms types (${a_{1},a_{2},a_{3},...}$), positions (${x_{1},x_{2},x_{3},...}$), and primitive cell parameters into a feature set that can be effectively utilized in machine learning. For machine learning methods to be effective, it is necessary that the machine-readable representation of a material’s structure fulfills the criteria as outlined in the introduction [10, 11, 12, 13, 14]. Notably, scalar properties (such as heat capacity or reactivity) do not change when translations, rotations, or permutations of atom indexing are applied to the atomic coordinates. Therefore, to ensure representations reflect the similarities between atomic structures, the representations should also be invariant to those symmetry operations. ### 2.1 Local Descriptors One strategy to form a representation of a crystal structure is to characterize the local environment of each atom and consider the full structure as a combination local representations. This concept was applied by Behler and Parinello[26], who proposed the atom-centered symmetry functions (ACSF). ACSF descriptors (Figure 2a) can be constructed using radial, $G_{i}^{1}$, and angular, $G_{i}^{2}$, symmetry functions centered on atom i, $G_{i}^{1}=\sum^{\text{neighbors}}_{j\neq i}e^{-\eta(R_{ij}-R_{s})^{2}}f_{c}(R_{ij})$ (1) $G_{i}^{2}=2^{1-\zeta}\sum^{\text{neighbors}}_{j,k\neq i}(1+\lambda\cos\theta_{ijk})^{\zeta}e^{-\eta(R_{ij}^{2}+R_{ik}^{2}+R_{jk}^{2})}f_{c}(R_{ij})f_{c}(R_{ik})f_{c}(R_{jk})$ (2) with the tunable parameters $\lambda$, $R_{s}$, $\eta$, and $\zeta$. $R_{ij}$ is the distance between the central atom $i$ and atom $j$, and $\theta_{ijk}$ corresponds to the angle between the vector from the central atom to atom $j$ and the vector from the central atom to atom $k$. The cutoff function $f_{c}$ screens out atomic interactions beyond a specified cutoff radius and ensures locality of the atomic interactions. Because symmetry functions rely on relative distances and angles, they are rotationally and translationally invariant. Local representations can be constructed from many symmetry functions of the type $G_{i}^{1}$ and $G_{i}^{2}$ with multiple settings of tunable parameters to probe the environment at varying distances and angular regions. With the set of localized symmetry functions, neural networks can then predict local contributions to a particular property and approximate global properties as the sum of local contributions. The flexibility of this approach allows for modification of the $G_{i}^{1}$ and $G_{i}^{2}$ functions [27, 28] or higher capacity neural networks for element-wise prediction [28]. In search of a representation with fewer hand-tuned parameters and a more rigorous definition of similarity, Bartok et al. [12] proposed a rotationally invariant kernel for comparing environments based on local atomic density. Given a central atom, the Smooth Overlap of Atomic Positions (SOAP) defines the atomic density function $\rho(\mathbf{r})$ as a sum of Gaussian functions centered at each neighboring atom within a cutoff radius (Figure 2b). The choice of Gaussian function is motivated by the intuition that representations should be continuous such that small changes in atomic positions should result in correspondingly small changes in the metric between two configurations. With a basis of radial functions $g_{n}(\mathbf{r})$ and spherical harmonics $Y_{lm}(\theta,\phi)$, $\rho(\mathbf{r})$ for central atom $i$ can be expressed as: $\rho_{i}(\mathbf{r})=\sum_{j}\exp{-\frac{|\mathbf{r}-\mathbf{r}_{ij}|^{2}}{2\sigma^{2}}}=\sum_{nlm}c_{nlm}g_{n}(\mathbf{r})Y_{lm}(\mathbf{\hat{r}})$ (3) and the kernel can be computed[12, 29]: $K(\rho,\rho^{\prime})=\mathbf{p}(\mathbf{r})\cdot\mathbf{p}^{\prime}(\mathbf{r})$ (4) $\mathbf{p}(\mathbf{r})\equiv\sum_{m}c_{nlm}(c_{n^{\prime}lm})^{*}$ (5) where $c_{nlm}$ are the expansion coefficients in Equation 3. In practice, $\mathbf{p}(\mathbf{r})$ can be used as a vector descriptor of the local environment and is also referred to as a power spectrum [12]. SOAP has demonstrated extraordinary versatility for materials applications both as a tool for measuring similarity [30] and as a descriptor for machine learning algorithms [31]. Furthermore, the SOAP kernel can be used to compare densities of different elements by adding an additional factor that provides a definition for similarity between atoms, where for instance, atoms in the same group could have higher similarity [29]. The mathematical connections between different local atomic density representations including ACSFs and SOAP are elucidated by a generalized formalism introduced by Willatt et al. [32], offering a methodology through which the definition of new variants can be clarified. Instead of relying on the density of nearby atoms, local representations can be derived from a Voronoi tessellation of a crystal structure. The Voronoi tessellation segments space into cells such that each cell contains one atom and all points in space such that atom A is the closest atom are contained in the same cell as atom A (Figure 2c). From these cells, Ward et al. [17] identified a set of descriptive features including an effective coordination number computed using the area of the faces, the lengths and volumes of nearby cells, ordering of the cells based on elements, and atomic properties of nearest neighbors weighted by the area of the intersecting face. When combined with compositional features [19], their representation results in better performance on predictions of formation enthalpy for ICSD than partial radial distribution functions [33] (Figure 1 in Reference [17]). In subsequent work, these descriptors have facilitated the prediction of experimental heat capacities in MOFs [34]. Similarly, Isayev et al. [35] replaced faces of the Voronoi tessellation with virtual bonds and separated the resulting framework into sets of linear (up to four atoms) and shell-based (up to nearest neighbors) fragments. Additional features related to the atomic properties of constituent elements were associated with each fragment, and the resulting vectors were concatenated with attributes of the supercell. In addition to demonstrating accurate predictive capabilities, models could be interpreted through the properties of the various fragments. For instance, predictions of band gap could be correlated with the difference in ionization potential in two-atom linear fragments, a trend that could be exploited to design material’s properties through tuning of composition[35]. Figure 2: (a) Examples of radial, $G^{1}_{i}$, and angular, $G^{2}_{i}$, symmetry functions from the local atom-centered symmetry function descriptor proposed by Behler and Parinello[26]. (b) In the Smooth Overlap of Atomic Positions (SOAP) descriptor construction, the atomic neighborhood density of a central atom is defined by a sum of Gaussian functions around each neighboring atom. A kernel function can then be built to compare the different environments by computing the density overlap of the atomic neighborhood functions. Figure is reprinted from reference [36]. (c) Voronoi tessellation in two and three dimensions. Yellow circles and spheres show particles while the blue lines divide equidistantly the space between two neighboring particles. Polygonal spaces encompassed by the blue lines are the Voronoi cells. Figure is reprinted from reference [37]. (d) Illustration of a Coulomb matrix where each element in the matrix shows Coulombic interaction between the labeled particles in the system on the left. Diagonal elements show self- interactions. (e) The births and deaths of topological holes in a point cloud (left) are recorded on a persistence diagram (right). Persistent features lie far from the parity line and indicate more significant topological features. Feature is reprinted from reference [38]. ### 2.2 Global Descriptors Alternatively, to more explicitly account for interactions beyond a fixed cutoff, atom types and positions can be encoded into a global representation that reflects geometric and physical insight. Inspired by the importance of electrostatic interactions in chemical stability, Rupp et al. [39] proposed the Coulomb matrix (Figure 2d), which models the potential between electron clouds: $M_{i,j}=\begin{cases}Z_{i}^{2.4}&\text{for }i=i\\\ \frac{Z_{i}Z_{j}}{|r_{i}-r_{j}|}&\text{for }i\neq j\end{cases}$ (6) Due to the fact that off-diagonal elements are only dependent on relative distances, Coulomb matrices are rotation and translation invariant. However, the representation is not permutation invariant since changing the labels of the atoms will rearrange the elements of the matrix. While originally developed for molecules, the periodicity of crystal structures can be added to the representation by considering images of atoms in adjacent cells, replacing the $\frac{1}{|r_{i}-r_{j}|}$ dependence with another function with the same small distance limit and periodicity that matches the parent lattice, or using an Ewald sum to account for long range interactions [11]. BIGDML [40] further improved results by restricting predictions from the representation to be invariant to all symmetry operations within the space group of the parent lattice and demonstrated effective implementations on tasks ranging from H interstitial diffusion in Pt to phonon density of states. While this approach has been able to effectively model long-range physics, these representations rely on a fixed supercell and may not be able to achieve the same chemical generality as local environments [40]. Global representations have also been implemented with higher-order tensors. Partial radial distribution functions (PRDF) are 3D non-permutation invariant matrices $g_{\alpha\beta r}$ whose elements correspond to the density of element $\beta$ in the environments of element $\alpha$ at radius $r$ [33]. The many-body tensor representation (MBTR) provides a more general framework [10] that can quantify k-body interactions and account for chemical similarity between elements. The MBTR is translationally, rotationally, and permutation invariant and can be applied to crystal structures by only summing over atoms in the primitive cell. While MBTR exhibited better performance than SOAP or Coulomb matrices for small molecules, its accuracy may not extend to larger systems [10]. Another well-established method for representing crystal structures in materials science is the cluster expansion. Given a parent lattice and a decoration $\mathbf{\sigma}$ defining the element that occupies each site, Sanchez et al. sought to map this atomic ordering to material properties and proposed evaluating the correlations between sites through a set of cluster functions. Each cluster function $\Phi$ is constructed from a product of basis functions ${\phi}$, over a subset of sites [41]. To ensure the representation is appropriately invariant, symmetrically equivalent clusters are grouped into classes denoted by $\alpha$. The characteristics of the atomic ordering can be quantified by averaging cluster functions over the decoration $<\Phi_{\alpha}>_{\mathbf{\sigma}}$, and properties $q$ of the configuration can be predicted as: $q(\mathbf{\sigma})=\sum_{\alpha}J_{\alpha}m_{\alpha}<\Phi_{\alpha}>_{\sigma}$ (7) where $m_{\alpha}$ is a multiplicity factor that accounts for the rate of appearance of different cluster types, and $J_{\alpha}$ are parameters referred to as effective cluster interactions that must be determined from fits to data [42]. While cluster expansions have been constructed for decades and provided useful models for configurational disorder and alloy thermodynamics [42], cluster expansions assume the structure of the parent lattice and models cannot generally be applied across different crystal structures [43, 44]. Furthermore, due to the increasing complexity of selecting cluster functions, implementations are restricted to binary and ternary systems without special development [45]. Additional research has extended the formalism to continuous environments (Atomic Cluster Expansion) by treating $\mathbf{\sigma}$ as pairwise distances instead of site- occupancies and constructing ${\phi}$ from radial functions and spherical harmonics [46]. The Atomic Cluster Expansion framework has provided a basis for more sophisticated deep learning approaches [47]. ### 2.3 Topological Descriptors Topological data analysis (TDA) has found favor over the past decade in characterizing structure in complex, high-dimensional datasets. When applied to the positions of atoms in amorphous or crystalline structures, topological methods reveal underlying geometric features that inform behavior in downstream predictions such as phase changes, reactivity, and separations. In particular, persistent homology (PH) is able to identify significant structural descriptors that are both machine readable and physically interpretable. The data can be probed at different length scales (formally called filtrations) by computing a series of complexes that each include all sets of points where all pairwise distances are less than the corresponding length [48]. Analysis of complexes by homology in different dimensions reveals holes or voids in the data manifold, which can be described by the range of length scales they are observed at (persistences), as well as when they are produced (births) and disappear (deaths). Emergent features with significant persistence values are less likely to be caused by noise in the data or as an artifact of the chosen length scales. In practice, multiple persistences, births, and deaths produced from a single material can be represented together by persistent diagrams (Figure 2e) or undergo additional feature engineering to generate a variety of descriptors as machine learning inputs [49]. While persistent homology has been applied to crystal structures in the Open Quantum Material Database [50], the method is particularly useful in the analysis of porous materials. The identified features (births, deaths, persistences) hold direct physical relevance to traditional structural features used to describe the pore geometries. For instance, persistent 2D deaths represent the largest sphere that can be included inside the pores of the materials. Krishnapriyan et al. has showed that these topological descriptors outperform traditional structural descriptors when predicting carbon dioxide adsorption under varying conditions for metal-organic frameworks [38], as did Lee et al. for zeolites for methane storage capacities [51]. Representative cycles can trace the topological features back to the atoms that are responsible for the hole or the void, creating a direct relationship between structure and predicted performance (Figure 4 in Reference [38]). Similarity methods for comparing barcodes can then be used to identify promising novel materials with similar pore geometries for targeted applications. A caveat is that PH does not inherently account for system size and is thus size-dependent. The radius cutoff, or the supercell size, needs to be carefully considered to encompass all significant topological features and allow comparison across systems of interest. In the worst case scenario, the computation cost per filtration for a structure is $O(N^{3})$, where $N$ is the number of sets of points defining a complex. Although the cost is alleviated by the sparsity of the boundary matrix [52], the scaling is poor for structures whose geometric features exceed unit cell lengths. The benefit of using PH features to capture more complex structural information has to be carefully balanced with the cost of generating these features. ## 3 LEARNING ON PERIODIC CRYSTAL GRAPHS In the previous section, we described many physically-inspired descriptors that characterize materials and can be used to efficiently predict properties. The use of differentiable graph-based representations in convolutional neural networks, however, mitigates the need for manual engineering of descriptors [53, 54]. Indeed, advances in deep learning and the construction of large- scale materials databases [4, 5, 6, 7, 8, 9] have made it possible to learn representations directly from structural data. From a set of atoms ${a_{1},a_{2},...}$ located at positions ${x_{1},x_{2},x_{3},...}$, materials can be converted to a graph $G(V,E)$ defined as the set of atomic nodes $V$ and the set of edges $E$ connecting neighboring atoms. Many graph-based neural network architectures were originally developed for molecular systems, with edges representing bonds. By considering periodic boundary conditions and defining edges as connections between neighbors within a cutoff radius, graphical representations can be leveraged for crystalline systems. The connectivity of the crystal graph thus naturally encodes local atomic environments [18]. When used as input to machine learning algorithms, the graph nodes and edges are initialized with an associated set of features. Nodal features can be as simple as a one-hot vector of the atomic number or can explicitly include other properties of the atomic species (e.g. electronegativity, group, period). Edge features are typically constructed from the distance between the corresponding atoms. Subsequently, a series of convolutions parameterized by neural networks modify node and/or edge features based on the current state of their neighborhood (Figure 3a). As the number of convolutions increases, interactions from further away in the structure can propagate, and graph features become tuned to reflect the local chemical environment. Finally, node and edge features can be pooled to form a single vector representation for the material [53, 55]. Crystal Graph Convolution Neural Networks (CGCNN) [18] and Materials Graph Networks (MEGNet) [56] have become benchmark algorithms capable of predicting properties across solid-state materials domains including bulk, surfaces, disordered systems, and 2D materials [57, 58]. Atomistic Line Graph Neural Network (ALIGNN) extended these approaches by including triplet three-body features in addition to nodes and edges and exhibited superior performance to CGCNN over a broad range of regression tasks including formation energy, bandgap, and shear modulus [59]. Other variants have used information from Voronoi polyhedra to construct graphical neighborhoods and augment edge features [60] or initialized node features based on the geometry and electron configuration of nearest neighbor atoms [61]. Figure 3: (a) General architecture of graph convolutional neural networks for property prediction in crystalline systems. Three-dimensional crystal structure is represented as a graph with nodes representing atoms and edges representing connections between nearby atoms. Features (e.g. nodal, edge, angular) within local neighborhoods are convolved, pooled into a crystal-wide vector, then mapped to the target property. Figure adapted from [62]. (b) Information loss in graphs built from pristine structures. Geometric distortions of ground-state crystal structures are captured as differing edge features in graphical representations. This information is lost in graphs constructed from corresponding unrelaxed structures. (c) Graph-based models can struggle to capture periodicity-dependent properties, such as cell lattice parameters. $R^{2}$ scores presented here were reported by Gong et al. for lattice parameter, a, predictions in short and long 1D single carbon chain toy structures. Figure adapted from [63]. (d) Ability of graphical representations to distinguish toy structures. Assuming a sufficiently small cutoff radius, the invariant representation—using edge lengths and/or angles—cannot distinguish the two toy arrangements, while the equivariant representation with directional features can. Figure adapted from [64]. While these methods have become widespread for property prediction, graph convolution updates based only on the local neighborhood may limit the sharing of information related to long-range interactions or extensive properties. Gong et al. demonstrated that these models can struggle to learn materials properties reliant on periodicity, including characteristics as simple as primitive cell lattice parameters (Figure 3c)[63]. As a result, while graph- based learning is a high-capacity approach, performance can vary substantially by the target use case. In some scenarios, methods developed primarily for molecules can be effectively implemented “out-of-the-box” with the addition of periodic boundary conditions, but especially in the case of long-range physical phenomena, optimal results can require specialized modeling. Various strategies to account for this limitation have been proposed. Gong et al. found that if the pooled representation after convolutions was concatenated with human-tuned descriptors, errors could be reduced by $90\%$ for related predictions, including phonon internal energy and heat capacity [63]. Algorithms have attempted to more explicitly account for long-range interactions by modulating convolutions with a mask defined by a local basis of Gaussians and a periodic basis of plane waves [65], employing a unique global pooling scheme that could include additional context such as stoichiometry [66], or constructing additional features from the reciprocal representation of the crystal [67]. Other strategies have leveraged assumptions about the relationships among predicted variables, such as representing phonon spectra using a Gaussian mixture model [68]. Given the promise and flexibility of graphical models, improving the data- efficiency, accuracy, generalizability, and scalability of these representations are active areas of research. While our previous discussion of structure-based material representations relied on the invariance of scalar properties to translation and rotation, this characteristic does not continue to hold for higher-order tensors. Consider a material with a net magnetic moment. If the material is rotated $180^{\circ}$ around an axis perpendicular to the magnetization, the net moment then points in the opposite direction. The moment was not invariant to the rotation but instead, transformed alongside the operation in an equivariant manner [69]. For a set of transformations described by group $G$, equivariant functions $f$ satisfy $g*f(x)=f(g*x)$ for every input $x$ and every group element $g$ [69, 70]. Recent efforts have shown that by introducing higher-order tensors to node and edge features (Figure 3d) and restricting the update functions such that intermediate representations are equivariant to the group E3 (encompassing translations, rotations, and reflections in $R^{3}$), models can achieve state-of-the-art accuracy on benchmark datasets and even exhibit comparable performance to structural descriptors in low-data ($\sim$100 datapoints) regimes [71, 69, 64]. Further accuracy improvements can be made by explicitly considering many-body interactions beyond edges [72, 73, 47]. Such models, developed for molecular systems, have since been extended to solid-state materials and shown exceptional performance. Indeed, Chen et al. trained an equivariant model to predict phonon density of states and was able to screen for high heat capacity targets [74], tasks identified to be particularly challenging for baseline CGCNN and MEGNet models [63]. Therefore, equivariant representations may offer a more general alternative to the specialized architectures described above. A major restriction of these graph-based approaches is the requirement for the positions of atomic species to be known. In general, ground-state crystal structures exhibit distortions that allow atoms to break symmetries, which are computationally modeled with expensive DFT calculations. Graphs generated from pristine structures lack representation of relaxed atomic coordinates (Figure 3b) and resulting model accuracy can degrade substantially [75, 76]. These graph-based models are therefore often most effective at predicting properties of systems for which significant computational resources have already been invested, thus breaking advice (3) from Section 1. As a result, their practical usage often remains limited when searching broad regions of crystal space for an optimal material satisfying a particular design challenge. Strategies have therefore been developed to bypass the need for expensive quantum calculations and use unrelaxed crystal prototypes as inputs. Gibson et al. trained CGCNN models on datasets composed of both relaxed structures and a set of perturbed structures that map to the same property value as the fully relaxed structure. The data-augmentation incentivizes the CGCNN model to predict similar properties within some basin of the fully relaxed structure and was demonstrated to improve prediction accuracy on an unrelaxed test set [76]. Alternatively, graph-based energy models can be used to modify unrelaxed prototypes by searching through a fixed set of possibilities [77] or using Bayesian optimization [78] to find structures with lower energy. Lastly, structures can be relaxed using cheap surrogate model (e.g. a force field) before a final prediction is made. The accuracy and efficiency of such a procedure will fundamentally rely on the validity and compositional generalizability of the surrogate relaxation approach [75]. ## 4 CONSTRUCTING REPRESENTATIONS FROM STOICHIOMETRY The phase, crystal system, or atomic positions of materials are not always available when modeling materials systems, rendering structural and graphical representations impossible to construct. In the absence of this data, material representations can also be built purely from stoichiometry (the concentration of the constituent elements) and without knowledge of the geometry of the local atomistic environments. Despite their lack of structural information and apparent simplicity, these methods provide unique benefits for materials science researchers. First, descriptors used to form compositional representations such as common atomic properties (e.g. atomic radii, electronegativity) do not require computational overhead and can be readily found in existing databases [19]. In addition, effective models can often be built using standard algorithms for feature selection and prediction that are implemented in freely available libraries [79], increasing accessibility to non-experts when compared with structural models. Lastly, when used as tools for high-throughput screening, compositional models identify a set of promising elemental concentrations. Compared with the suggestion of particular atomistic geometries, stoichiometric approaches may be more robust, as they make weaker assumptions about the outcomes of attempted syntheses. Compositional-based rules have long contributed to efficient materials design. Hume-Rothery and Linus Pauling designed rules for determining the formation of solid solutions and crystal structures that include predictions based on atomic radii and electronic valence states [80, 81]. However, many exceptions to their predictions can be found [82]. Machine learning techniques offer the ability to discover and model relationships between properties and physical descriptors through statistical means. Meredig et al. demonstrated that a decision tree ensemble trained using a feature set of atomic masses, positions in the periodic table, atomic numbers, atomic radii, electronegativities, and valence electrons could outperform a conventional heuristic on predicting whether ternary compositions would have formation energies $<100$ meV/atom [83]. Ward et al. significantly expanded this set to 145 input properties, including features related to the distribution and compatibility of the oxidation states of constituent atoms [19]. Their released implementation, MagPie, can be a useful benchmark or staring point for the development of further research methods [79, 84, 85]. Furthermore, if a fixed structural prototype (e.g. elapsolite) is assumed, these stoichiometric models can be used to analyze compositionally-driven variation in properties [86, 87]. Even more subtle yet extremely expressive low-dimensional descriptors can be obtained by initializing a set with standard atomic properties and computing successive algebraic combinations of features, with each calculation being added to the set and used to compute higher order combinations in the next round. While the resulting set will grow exponentially, compressive sensing can then be used to identify the most promising descriptors from sets that can exceed $10^{9}$ possibilities [88, 89]. Ghiringhelli et al. found descriptors that could accurately predict whether a binary compound would form in a zincblende or rocksalt structure [90], and Bartel et al. identified an improved tolerance factor $\tau$ for the formation of perovskite systems [91] (Table 1). While these approaches do not derive their results from a known mechanism, they do provide enough interpretability to enable the extraction of physical insights for the screening and design of materials. Table 1: Example Descriptors Determined through Compressive Sensing Descriptor | Prediction | Variables ---|---|--- | | IP-Ionization Potential $\frac{IP(B)-EA(B)}{r_{p}(A)^{2}}$ | Ordering in AB Compound | EA-Electron Afinity | | $r_{p}$-Radius of Maximum Density of p-Orbital $\frac{r_{X}}{r_{B}}-n_{A}(n_{A}-\frac{r_{A}/r_{B}}{ln[r_{A}/r_{B}]})$ | Stability of ABX3 Perovskite | $n_{Y}$-Oxidation State of Y | | $r_{y}$-Ionic Radius of Y When large datasets are available, deep neural networks tend to outperform traditional approaches, and that is also the case for compositional representations. The size of modern materials science databases have enabled the development of information-rich embeddings that map elements or compositions to vectors as well as the testing and validation of deep learning models. Chemically meaningful embeddings can be constructed by counting all compositions in which that element appeared in the Materials Project [92] or learned through the application of natural language processing to previously reported results in the scientific literature [93]. These data-hungry methods were able to demonstrate that their representations could be clustered based on atomic group [92] and could be used to suggest new promising compositions based on similarity with the best known materials [93]. The advantages of training deep learning algorithms with large datasets are exemplified by ElemNet, which only uses a vector of fractional stoichiometry as input. Despite its apparent simplicity, when $>3,000$ training points where available, ElemNet performed better than a MagPie-based model at predicting formation enthalpies [94]. While the applicability of ElemNet is limited to problem domains with $O(10^{3})$ datapoints, more recent methods have significantly reduced this threshold. ROOST [95] represented each composition as a fully-connected graph with nodes as elements, and properties were predicted using a message-passing scheme with an attention mechanism that relied on the stoichiometric fraction of each element. ROOST substantially improved on ElemNet, achieving better performance than MagPie in cases with only hundreds of training examples. Meanwhile, CrabNet [96] forms element-derived matrices as a sum of embeddings of each element’s identity and stoichiometric fraction. This approach achieves similar performance to ROOST by updating the representation using self- attention blocks. The fractional embedding can take log-scale data as input such that even dopants in small concentrations can have a significant effect on predictions. Despite the inherent challenges of predicting properties purely from composition, these recent and significant modeling improvements suggest that continued algorithmic development could be an attractive and impactful direction for future research projects. Compositional models have the advantage that they can suggest new systems to experimentalists without requiring a specific atomic geometry and, likewise, can learn from experimental data without necessitating an exact crystal structure [97]. Owing to their ability to incorporate experimental findings into ML pipelines and provide suggestions with fewer experimental requirements (e.g. synthesis of a particular phase), compositional models have become attractive methods for materials design. Zhang el al. trained a compositional model using atomic descriptors on previous experimental data to predict Vicker’s harness and validated their model by synthesizing and testing eight metal disilicides [97]. Oliynik et al. identified new Heusler compounds, while also verifying their approach on negative cases where they predicted a synthesis would fail [87]. Another application of their approach enabled the prediction of the crystal structure prototype of ternary compounds with greater than 96% accuracy. By training their model to predict the probability associated with each structure, they were able to experimentally engineer a system (TiFeP) with multiple competing phases [98]. While researchers have effectively implemented compositional models as methods for materials design, their limitations should be considered when selecting a representation for ML studies. Fundamentally, compositional models will only provide a single prediction for each stoichiometry regardless of the number of synthesizeable polymorphs. While training models to only predict properties of the lowest-energy structure is physically justifiable [99], extrapolation to technologically relevant meta-stable systems may still be limited. Additionally, graph-based structural models such as CGCNN [18] or MEGNet [56] generally outperform compositional models [84]. Therefore, composition models are most practically applicable when atomistic resolution of materials is unavailable, and thus structural representations cannot be effectively constructed. ## 5 DEFECTS, SURFACES, AND GRAIN BOUNDARIES Mapping the structure of small molecules and unit cells to materials properties has been a reasonable starting point for many applications of materials science modeling. However, materials design often requires understanding of larger length scales beyond the small unit cell, such as in defect and grain boundary engineering, and in surface science [100]. In catalysis, for example, surface activity is highly facet dependent and cannot be modeled using the bulk unit cell alone. It has been shown that the (100) facet of RuO2, a state-of-the-art catalyst for the oxygen evolution reaction (OER), has an order of magnitude higher current for OER than the active site on the thermodynamically stable (110) facet [101]. Similarly, small unit cells are not sufficient for modeling transport properties, where size, orientation, and characteristics of grain boundaries play a large role. In order to apply machine learning to practical materials design, it is therefore imperative to construct representations that can characterize environments at the relevant length scales. Figure 4: (a) Point defect properties are learned from a representation of the pristine bulk structure and additional relevant information on conduction and valence band levels. (b) Surface properties are learned from a combination of pristine bulk structure representation, miller index, and density of states information. (c) Local environments of atoms near a grain boundary versus atoms in the pristine bulk are compared to learn grain boundary properties. Figure 4a adapted from [102]. Defect engineering offers a common and significant degree of freedom through which materials can be tuned. Data science can contribute to the design of these systems as fundamental mechanisms are often not completely understood even in long-standing cases such as carbon in steels [103]. Dragoni et al. [31] developed a Gaussian Approximate Potential (GAP) [104] using SOAP descriptors for face-centered cubic iron that could probe vacancies, interstitials, and dislocations, but their model was confined to a single phase of one element and required DFT calculations incorporating $O(10^{6})$ unique environments to build the interpolation. Considering that even a small number of possible defects significantly increases combinatorial complexity, a general approach for predicting properties of defects from pristine bulk structure representations could accelerate computation by orders of magnitude (Figure 4a). For example, Varley et al. observed simple and effective linear relationships between vacancy formation energy and descriptors derived from the band structure of the bulk solid [102]. While their model only considered one type of defect, their implementation limits computational expense by demonstrating that only DFT calculations on the pristine bulk were required [102]. Structure- and composition-aware descriptors of the pristine bulk have additionally been shown to be predictive of vacancy formation in metal oxides [105, 106] and site/antisite defects in AB intermetallics [107]. To develop an approach that can be used over a broad range of chemistries and defect types, Frey et al. formed a representation by considering relative differences in characteristics (atomic radii, electronegativity, etc.) of the defect structure compared to the pristine parent[21]. Furthermore, because reference bulk properties could be estimated using surrogate ML models, no DFT calculations were required for prediction of either formation energy or changes in electronic structure [21]. We also note that in some cases it may be judicious to design a model that does not change significantly in the presence of defects. For these cases, representations based on simulated diffraction patterns are resilient to site- based vacancies or displacements [108]. Like in defect engineering, machine learning for practical design of catalyst materials requires representations beyond the single unit cell. Design of catalysts with high activity crucially depends on interactions of reaction intermediates with materials surfaces based on the Sabatier principle, which argues that activity is greatest when intermediates are bound neither too weakly nor too strongly [109]. From a computational perspective, determining absorption energies involves searches over possible adsorption active sites, surface facets, and surface rearrangements, leading to a combinatorial space that can be infeasible to exhaustively cover with DFT. Single dimension descriptors based on electronic structure have been established that can predict binding strengths and provide insight on tuning catalyst compositions, such as metal d-band center for metals [110] and oxygen 2p-band center for metal oxides [111]. Additional geometric approaches include describing the coordination of the active site (generalized coordination number in metals, adjusted generalized coordination number in metal oxides) [112]. Based on the success of these simple descriptors, machine learning models have been developed to learn binding energy using the density of states and geometric descriptors of the pristine bulk structure as features (Figure 4b) [113]. However, these structural and electronic descriptors are often not generalizable across chemistries [110, 114], limiting the systems over which they can be applied and motivating the development of more sophisticated machine learning techniques. To reduce the burden on high-throughput DFT calculations, active learning with surrogate models using information from pure metals and active-site coordination has been used to identify alloy and absorbate pairs that have the highest likelihood of producing near-optimal binding energies [115]. Furthermore, when sufficient data ($>10,000$ examples) is available, modifications of graph-convolutional models have also predicted binding energies with high accuracy even in datasets with up to 37 elements, enabling discovery without detailed mechanistic knowledge [114]. To generalize these results, the release of Open Catalyst 2020 and its related competitions [6, 9] has provided both over one million DFT energies for training new models and a benchmark through which new approaches can be evaluated [75]. While significant advancements have been made, state-of-the-art models still exhibit high-errors for particular absorbates and non-metallic surface elements, constraining chemistry over which effective screening can be conducted [75]. Furthermore, the complexity of the design space relevant for ML models grows considerably when accounting for interactions between absorbates and different surface facets [116]. Beyond atomistic interactions, the mechanical and thermal behavior of materials can be significantly modulated by processing conditions and the resulting microstructure. Greater knowledge of local distortions introduced at varying grain boundary incident angles would give computational materials scientists a more complete understanding of how experimentally chosen chemistries and synthesis parameters will translate into device performance. Strategies to quantify characteristics of grain boundary geometry have included reducing computational requirements by identifying the most promising configurations with virtual screening [117], estimating grain boundary free volume as a function of temperature and bulk composition [118], treating the microstructure as a graph of nodes connected across grain boundaries [119, 54], and predicting the energetics, and hence feasiblity, of solute segregation [120]. While the previous approaches did not include features based on the constituent atoms and were only benchmarked on systems with up to three elements, recent work has demonstrated that the excess energy of the grain boundary relative to the bulk can be approximated across compositions with five variables defining its orientation and the bond lengths within the grain boundary (Figure 4c)[121]. Further research has tried to map local grain boundary structure to function. Algorithmic approaches to grain boundary structure classification have been developed (see for example VoroTop [122]), but such approaches typically rely on expert users and do not provide a continuous representation that can smoothly interpolate between structures [123]. To eliminate these challenges, Rosenbrock et al. proposed computing SOAP descriptors for all atoms in the grain boundary, clustering vectors into classes, and identifying grain boundaries through its local environment classes. The representation was not only predictive of grain boundary energy, temperature-dependent mobility, and shear coupling but also provides interpretable effects of particular structures within the grain boundary [124]. A related approach computed SOAP vectors relative to the bulk structure when analyzing thermal conductivity [125]. Representations based on radial and angular structure functions can also quantify the mobility of atoms within a grain boundary [126]. When combined, advancing models for grain boundary stability as well as structure to property relationships opens the door for functional design of grain boundaries. ## 6 TRANSFERABLE INFORMATION BETWEEN REPRESENTATIONS Applications of machine learning to materials science are limited by the scope of compositions and structures over which algorithms can maintain sufficient accuracy. Thus, building large-scale, diverse datasets is the most robust strategy to ensure trained models can capture the relevant phenomena. However, in most contexts, materials scientists are confronted with sparsely distributed examples. Ideally, models can be trained to be generalizable and exhibit strong performance across chemistries and configurations even with few to no data points in a given domain. In order to achieve this, representations and architectures must be chosen such that models can learn to extrapolate beyond the space observed in the training set. Effective choices often rely on inherent natural laws or chemical features that are shared between the training set and extrapolated domain such as physics constraints [127, 128], the geometric [129, 130] and electronic [131, 132] structure of local environments, and positions of elements in the periodic table [133, 134]. For example, Li et al. were able to predict absorption energies on high entropy alloy surfaces after training on transition metal data by using the coordination number and electronic properties of neighbors at the active site [129]. While significant advancements have been made in the field, extrapolation of machine learning models across materials spaces typically requires specialized research methods and is not always feasible. Likewise, it is not always practical for a materials scientist to improve model generality by just collecting more data. In computational settings, some properties can only be reliably estimated with more expensive, higher levels of theory, and for experimentalists, synthetic and characterization challenges can restrict throughput. The deep learning approaches that have demonstrated exceptional performance over a wide range of tests cases discussed in this review can require at least $~{}10^{3}$ training points, putting them seemingly out for the realm of possibility for many research projects. Instead, predictive modeling may fall back on identifying relationships between a set of human-engineered descriptors and target properties. Alternatively, the hidden, intermediate layers of deep neural networks can be conceptualized as a learned vector representation of the input data. While this representation is not directly interpretable, it must still contain physical and chemical information related to the prediction task, which downstream layers for the network utilize to generate model outputs. Transfer learning leverages these learned representations from task A and uses them in the modeling of task B. Critically, task A can be chosen to be one for which a large number of data points are accessible (e.g. prediction all DFT formation energies in the Materials Project), and task B can be of limited size (e.g. predicting experimental heats of formation of a narrow class of materials). In principle, if task A and task B share an underlying physical basis (the stability of the material), the features learned when modeling task A may be more informationally-rich than a human-designed representation [135]. With this more effective starting point, subsequent models for task B can reach high accuracy with relatively few new examples. The most straightfoward methods to implement transfer learning in the materials science community follow a common procedure: (1) train a neural network model to predict a related property (task A) for which $>O(1,000)$ data points are available (pretraining), (2) fix parameters of the network up a chosen depth $d$ (freezing), and (3) given the new dataset for task B, either retrain the remaining layers, where parameters can be initialized randomly or from the task A model (finetuning), or treat the output of the model at depth d as in input representation to another ML algorithm (feature extraction) [136, 137]. The robustness of this approach has been demonstrated across model classes including those using composition only (ElemNet [135, 137], ROOST [95]), crystal graphs (CGCNN) [138], and equivariant convolutions (GemNet) [139]. Furthermore, applications of task B, range from experimental data [135, 95] to DFT-calculated surface absorption energies [139]. The sizes of the datasets for task A and task B will determine the effectiveness of a transfer learning approach in two ways. First, the quality and robustness of the representation learned for task A will increase as the number of observed examples (the size of dataset A) increases. Secondly, as the size of dataset B decreases, data becomes too sparse for a ML model to learn a reliable representation alone and prior information from the solution to task A can provide an increasingly useful method to interpolate between the few known points. Therefore, transfer learning typically exhibits the greatest boosts in performance when task A has orders of magnitude more data than task B [135, 138]. In addition, the quality of information sharing through transfer learning depends on the physical relationship between task A and task B. Intuitively, the representation from task A provides a better guide for task B if the tasks are closely related. For example, Kolluru et al. demonstrated that transfer learning from models trained on the Open Catalyst Dataset [6] exhibited significantly better performance when applied to absorption of new species than energies of less-related small molecules [139]. While it is difficult to choose the optimal task A for a given task B a priori, shotgun transfer learning [136] has demonstrated that the best pairing can be chosen experimentally by empirically validating a large pool of possible candidates and selecting top performers. The depth $d$ from which features should be extracted from task A to form a representation can also be task dependent. Kolluru et al. provided evidence that to achieve optimal performance more layers of the network should be allowed to be retrained in step (3) as the connection between task A and task B becomes more distant [139]. Gupta et al. arrived at a similar conclusion that the early layers of deep neural networks learned more general representations and performed better in cross-property transfer learning [137]. Inspired by this observation that representations at different neural network layers contain information with varying specificity to a particular prediction task, representations for transfer learning that combine activations from multiple depths have been proposed [139, 140]. When tasks are sufficiently different, freezing neural network weights may not be the optimal strategy and instead representations for task B can include predictions for task A as descriptors. For instance, Cubuk et al. observed that structural information was critical to predict Li conductivity but was only available for a small set of compositions for which crystal structures were determined. By training a separate surrogate model to predict structural descriptors from composition and using those approximations in subsequent Li conductivity models, the feasible screening domain was expanded by orders of magnitude [141]. Similarly, Greenman et al. [142] used $O(10,000)$ TD-DFT calculations to train a graph neural network whose estimates could be used as an additional descriptor for a model predicting experimental peaks in absorption spectra. Representations have also been sourced from the output of generative models. Kong et al. trained a Generative Adversial Network (GAN) to sample electronic density of states (DOS) given a particular material composition. Predictions of absorption spectra of a particular composition were improved by concatenating stoichiometric data with the average DOS sampled from the generative model [143]. ## 7 GENERATIVE MODELS FOR INVERSE DESIGN While, in principle, machine learning methods can significantly reduce the time required to compute materials properties, and material scientists can employ these models to screen for a set of target systems by rapidly estimating the stability and performance, the space of feasible materials precludes a naive global optimization strategy in most cases. Generative models including Variational Autoencoders (VAE) [144, 1], Generative Adversarial Networks (GAN) [145, 146], and diffusion models [147, 148] can be trained to sample from a target distribution and have proved to be capable strategies for optimization in high-dimensional molecular spaces [1, 149]. While some lessons can be drawn from the efforts of researchers in the computational chemistry community, generative models face unique challenges for proposing crystals [150, 151]. First, the diversity of atomic species increases substantially when compared with small organic molecules. In addition, given a composition, a properly defined crystal structure requires both the positions of the atoms within the unit cell as well as the lattice vectors and angles that determine the systems periodicity. This definition is not unique, and the same material can be described after rotations or translations of atomic coordinates as well as integer scaling of the original unit cell. Lastly, many state-of-the-art materials for catalysis (e.g. zeolites, metal organic frameworks) can have unit cells including $>100$ of atoms, increasing the dimensionality of the optimization problem [150, 151]. Figure 5: Approaches for crystal structure generative models. (Left) Initial models based on voxel representations defined positions of atoms by discretizing space into finite volume elements but were not applied generally over the space of crystal structures [23, 22, 24, 25]. (Center) Restricting the generation process to be invariant to permutation, translation, and rotations, through an appropriately constrained periodic decoder (PGNNDec) results in sampling structures exhibiting more diversity and stability. (Right) When features of the material can be assumed, such as a finite number of possible topologies connecting substructures, the dimensionality of the problem can be substantially reduced and samples over larger unit cell materials can be generated. Figures on left, center, and right are adapted from [23], [152], and [153], respectively. One attempt to partially address the challenges of generative modeling for solid materials design is a voxel representation [150], in which unit cells are divided into volume elements and models are built using techniques from computer vision. Hoffman represented unit cells using a density field that could be further segmented into atomic species and was able to generate crystals with realistic atomic spacings. However, atoms could be mistakenly decoded into other species with nearby atomic number and most of the generated structures could not be stably optimized with a DFT calculation [22]. Alternate approaches could obtain more convincing results, but over a confined region of material space [154]. iMatgen (Figure 5a) invertibly mapped all unit cells into a cube with Gaussian-smeared atomic density and trained a VAE coupled with a surrogate energy prediction. The model was able to rediscover stable structures but was constrained over the space of Vanadium oxides [23]. A similar approach constructed a separate voxel representation for each element and employed a GAN trained alongside an energy constraint to explore the phases of Bi-Se [155]. In order to resolve some of Hoffman el al’s limitations, Court et al. [24] reduced segmentation errors by augmenting the representation with a matrix describing the occupation (0,1) of each voxel and a matrix recording the atomic number of occupied voxels. Their model was able to propose new materials that exhibited chemical diversity and could be further optimized with DFT but restricted analysis to cubic systems. Likewise, compositions of halide perovskites with optimized band gaps could be proposed using a voxelized representation of a fixed perovskite prototype [25]. Voxel representations can be relaxed to continuous coordinates in order to develop methods that are more comprehensively applicable over crystal space. Kim et al. represented materials using a record of the unit cell as well as a point cloud of fractional coordinates of each element. The approach proposed lower energy structures than iMatgen for V-O binaries and was also applicable over more diverse chemical spaces (Mg-Mn-O ternaries) [156]. Another representation including atomic positions along with elemental properties could be leveraged for inverse design over spaces that vary in both composition and lattice structure. In a test case, the model successfully generated new materials with negative formation energy and promising thermoelectric power factor [154]. While these models have demonstrated improvements in performance, they lack the translational, rotational, and scale invariances of real materials and are restricted to sampling particular materials classes [156, 152]. Recently, alternatives that account for these symmetries have been proposed. Fung et al. proposed a generative model for rotationally and translationally invariant atom-centered symmetry functions (ACSF) from which target structures could be reconstructed [157]. Crystal Diffusion VAEs (Figure 5b) leveraged periodic graphs and SE(3) equivariant message-passing layers to encode and decode their representation in a translationally and rotationally invariant way [152]. They also proposed a two step generation process during which they first predicted the crystal lattice from a latent vector and subsequently sampled the composition and atomic positions through Langevin dynamics. Furthermore, they established well-defined benchmark tasks and demonstrated that for inverse design their method was more flexible than voxel models with respect to crystal system and more accurate than point cloud representations at identifying crystals with low formation energy. Scaling solid-state generative modeling techniques to unit cells with $(10^{4})$ atoms would enable inverse design of porous materials that are impossible to explore exhaustively but demonstrate exceptional technological relevance. Currently, due to the high number of degrees of freedom, sampling from these spaces requires imposing physical constraints in the modeling process. Such restrictions can be implemented as post-processing steps or integrated into the model representation. ZeoGAN [158] generated positions of oxygen and silicon atoms in a 32x32x32 grid to propose new Zeolites. While some of the atomic positions proposed directly from their model violated conventional geometric rules, they could obtain feasible structures by filtering out divergent compositions and repairing bond connectivity through the insertion or deletion of atoms. Alternatively, Yao et al. designed geometric constraints directly into the generative model by representing Metal Organic Frameworks (MOFs) by their edges, metal/organic vertexes, and distinct topologies (RFcodes) (Figure 5c) [153]. Because this representation is invertible, all RFcodes correspond to a structurally possible MOFs. By training a VAE to encode and decode this RFcode representation, they demonstrated the ability to interpolate between structures and optimize properties. In general, future research should balance more stable structure generation against the possible discovery of new motifs and topologies. ## 8 DISCUSSION In this review, we have introduced strategies for designing representations for machine learning in the context of challenges encountered by materials scientists. We discussed local and global structural features as well as representations learned from atomic-scale data in large repositories. We noted additional research that extends beyond idealized crystals to include the effects of defects, surfaces, and microstructure. Furthermore, we acknowledged that in practice the availability of data both in quality and quantity can be limited. We described methods to mitigate this including developing models based on compositional descriptors alone or leveraging information from representations built for related tasks through transfer learning. Finally, we analyzed how generative models have improved by incorporating symmetries and domain knowledge. As data-based methods have become increasingly essential for materials design, optimal machine learning techniques will play a crucial role in the success of research programs. The previous sections demonstrate that the choice of representation will be among these pivotal factors and that novel approaches can open the door to new modes of discovery. Motivated by these observations, we conclude by summarizing open problems with the potential to have high impact on the field of materials design. ### 8.1 Trade-offs of Local and Global Structural Descriptors Local structural descriptors including SOAP [12] have become reliable metrics to compare environments with a specific cutoff radius, and when properties can be defined through short-range interactions, have demonstrated strong predictive performance. Characterizing systems based off local environments allows models to extrapolate to cases where global representations may vary substantially (e.g. an extended supercell of a crystal structure)[14] and enables highly-scalable methods of computation that can extend the practical limit of simulations to much larger systems [159]. However, Unke et al. notes that the required complexity of the representation can grow quickly when modeling systems with many distinct elements and the quality of ML predictions will be sensitive to the selected hyperparameters, such as the characteristics distances and angles in atom-centered symmetry functions[160]. Furthermore, it is unclear if these high quality results extend to materials characteristics that depend strongly on long-range physics or periodicity of the crystal. On the other hand, recent global descriptors [40] can more explicitly model these phenomena, but have not exhibited the same generality across space groups and system sizes. Strategies exploring appropriate combinations of local and long- range features [161] have the potential to break through these trade-offs to provide more universal models for material property prediction. ### 8.2 Prediction from Unrelaxed Crystal Prototypes If relaxed structures are required to form representations, the space over which candidates can be screened is limited to those materials for which optimized geometries are known. Impressively, recent work [162, 163] has shown that ML force-fields, even simple models with relatively high errors, can be used optimize structures and obtain converged results that are lower in energy than those obtained using VASP [164]. Their benchmarking on the OC20 [6] dataset and lower accuracy requirements suggest that the approach could be generalizable across a wide-class of material systems and thus significantly expand the availability of structural descriptors. Similarly, Chen et al. demonstrated that a variant of MEGNET could perform high fidelity relaxations of unseen materials with diverse chemistries and that leveraging the resulting structures could improve downstream ML predictions of energy when compared with unrelaxed inputs [165]. The strong performance of these approaches and their potential to significantly increase the scale and effectiveness of computational screening motivates high-value research questions concerning the scale of data sets required for training, the generalizabiltiy over material classes, and the applicability to prediction tasks beyond stability. ### 8.3 Applicability of Compositional Descriptors Compositional descriptors are typically readily available as tabulated values, but even state-of-the-art models do not perform as well as the best structural approaches. However, there is some evidence that the scale of improvement when including structural information is property dependent. System energies can be conceptualized as a sum of site energies that are highly dependent on the local environment, and graph neural networks provide significantly more robust predictions of materials stability [84]. On the other hand, for properties dependent on global features such as phonons (vibrations) or electronic band structure (band gap) the relative improvement may not be as large [99, 166, 167]. Identifying common trends connecting tasks for which this difference is the least significant would provide more intuition on which scenarios compositional models are most appropriate. Furthermore, in some modeling situations, structural information is available but only over a small fraction of the dataset. To maximize the value of this data, more general strategies involving transfer learning [141] or combining separate composition and structural models [85] should be developed. ### 8.4 Extensions of Generative Models Additional symmetry considerations and the implementation of diffusion-based architectures led to generative models that improved significantly over previous voxel approaches. While this strategy is a promising direction for small unit cells, efforts pertaining to other parameters critical to material performance including microstructure [168], dimensionality [169] and surfaces [170] should also be pursued. In addition, research groups have side-stepped some of the challenges of materials generation by designing approaches that only sample material stoichiometry [171]. While this strategy limits the full characterization of new materials through a purely computational pipeline, there may be cases where they are sufficient to propose promising regions for experimental analysis. ## DISCLOSURE STATEMENT The authors are not aware of any affiliations, memberships, funding, or financial holdings that might be perceived as affecting the objectivity of this review. ## ACKNOWLEDGMENTS JD was involved in the writing of all sections. AT and MX collaborated on the writing and designed the figure for Atomistic Structure section, JK collaborated on the writing and designed the figure for the Periodic Graph section, JL collaborated on the writing and designed the figure for the Defects, Surfaces, and Grain Boundaries section. JP provided valuable insights for the organization and content of the article. RGB selected the topic and focus of the review, contributed to the central themes and context, and supervised the project. All authors participated in discussions and the reviewing of the final article. The authors would like to thank Anna Bloom for editorial contributions. The authors acknowledge financial support from the Advanced Research Projects Agency–Energy (ARPA-E), US Department of Energy under award number DE- AR0001220. JD, MX and ART thank the National Defense Science and Engineering Graduate Fellowship, the National Science Scholarship from Agency for Science, Technology and Research, and Asahi Glass Company, respectively, for financial support. RGB thanks the Jeffrey Cheah Chair in Engineering. ## References * [1] Gómez-Bombarelli R, Wei JN, Duvenaud D, Hernández-Lobato JM, Sánchez-Lengeling B, et al. 2018. Automatic chemical design using a data-driven continuous representation of molecules. ACS Central Science 4(2):268–276 * [2] Peng J, Schwalbe-Koda D, Akkiraju K, Xie T, Giordano L, et al. 2022. Human- and machine-centred designs of molecules and materials for sustainability and decarbonization. Nature Reviews Materials 7:991–1009 * [3] Pyzer-Knapp EO, Suh C, Gómez-Bombarelli R, Aguilera-Iparraguirre J, Aspuru-Guzik A. 2015. What is high-throughput virtual screening? a perspective from organic materials discovery. Annual Review of Materials Research 45(1):195–216 * [4] Jain A, Ong SP, Hautier G, Chen W, Richards WD, et al. 2013. Commentary: The materials project: A materials genome approach to accelerating materials innovation. APL Materials 1(1):011002 * [5] Kirklin S, Saal JE, Meredig B, Thompson A, Doak JW, et al. 2015. The open quantum materials database (oqmd): assessing the accuracy of dft formation energies. npj Computational Materials 1(1):15010 * [6] Chanussot L, Das A, Goyal S, Lavril T, Shuaibi M, et al. 2021. Open catalyst 2020 (oc20) dataset and community challenges. ACS Catalysis 11(10):6059–6072 * [7] Curtarolo S, Setyawan W, Wang S, Xue J, Yang K, et al. 2012. Aflowlib.org: A distributed materials properties repository from high-throughput ab initio calculations. Computational Materials Science 58:227–235 * [8] Ward L, Dunn A, Faghaninia A, Zimmermann NE, Bajaj S, et al. 2018. Matminer: An open source toolkit for materials data mining. Computational Materials Science 152:60–69 * [9] Tran R, Lan J, Shuaibi M, Goyal S, Wood BM, et al. 2022. The open catalyst 2022 (oc22) dataset and challenges for oxide electrocatalysis. ArXiv: 2206.08917 * [10] Huo H, Rupp M. 2017. Unified representation of molecules and crystals for machine learning. ArXiv: 1704.06439 * [11] Faber F, Lindmaa A, Lilienfeld OAV, Armiento R. 2015. Crystal structure representations for machine learning models of formation energies. International Journal of Quantum Chemistry 115(16):1094–1101 * [12] Bartók AP, Kondor R, Csányi G. 2013. On representing chemical environments. Physical Review B 87(18):184115 * [13] Lilienfeld OAV, Ramakrishnan R, Rupp M, Knoll A. 2015. Fourier series of atomic radial distribution functions: A molecular fingerprint for machine learning models of quantum chemical properties. International Journal of Quantum Chemistry 115(16):1084–1093 * [14] Musil F, Grisafi A, Bartók AP, Ortner C, Csányi G, Ceriotti M. 2021. Physics-inspired structural representations for molecules and materials. Chemical Reviews 121(16):9759–9815 * [15] Glass CW, Oganov AR, Hansen N. 2006. Uspex—evolutionary crystal structure prediction. Computer Physics Communications 175(11-12):713–720 * [16] Wang Y, Lv J, Zhu L, Ma Y. 2010. Crystal structure prediction via particle-swarm optimization. Physical Review B 82(9):094116 * [17] Ward L, Liu R, Krishna A, Hegde VI, Agrawal A, et al. 2017. Including crystal structure attributes in machine learning models of formation energies via voronoi tessellations. Physical Review B 96(2):024104 * [18] Xie T, Grossman JC. 2018. Crystal graph convolutional neural networks for an accurate and interpretable prediction of material properties. Physical Review Letters 120(14):145301 * [19] Ward L, Agrawal A, Choudhary A, Wolverton C. 2016. A general-purpose machine learning framework for predicting properties of inorganic materials. npj Computational Materials 2(1):16028 * [20] Haynes WM. 2016. CRC Handbook of Chemistry and Physics. CRC Press * [21] Frey NC, Akinwande D, Jariwala D, Shenoy VB. 2020. Machine learning-enabled design of point defects in 2d materials for quantum and neuromorphic information processing. ACS Nano 14(10):13406–13417 * [22] Hoffmann J, Maestrati L, Sawada Y, Tang J, Sellier JM, Bengio Y. 2019. Data-driven approach to encoding and decoding 3-d crystal structures. ArXiv: 1909.00949 * [23] Noh J, Kim J, Stein HS, Sanchez-Lengeling B, Gregoire JM, et al. 2019. Inverse design of solid-state materials via a continuous representation. Matter 1(5):1370–1384 * [24] Court CJ, Yildirim B, Jain A, Cole JM. 2020. 3-d inorganic crystal structure generation and property prediction via representation learning. Journal of Chemical Information and Modeling 60(10):4518–4535 * [25] Choubisa H, Askerka M, Ryczko K, Voznyy O, Mills K, et al. 2020. Crystal site feature embedding enables exploration of large chemical spaces. Matter 3(2):433–448 * [26] Behler J, Parrinello M. 2007. Generalized neural-network representation of high-dimensional potential-energy surfaces. Physical Review Letters 98(14):146401 * [27] Behler J. 2011. Atom-centered symmetry functions for constructing high-dimensional neural network potentials. The Journal of Chemical Physics 134(7):074106 * [28] Smith JS, Isayev O, Roitberg AE. 2017. Ani-1: an extensible neural network potential with dft accuracy at force field computational cost. Chemical Science 8(4):3192–3203 * [29] De S, Bartók AP, Csányi G, Ceriotti M. 2016. Comparing molecules and solids across structural and alchemical space. Physical Chemistry Chemical Physics 18(20):13754–13769 * [30] Schwalbe-Koda D, Jensen Z, Olivetti E, Gómez-Bombarelli R. 2019. Graph similarity drives zeolite diffusionless transformations and intergrowth. Nature Materials 18(11):1177–1181 * [31] Dragoni D, Daff TD, Csányi G, Marzari N. 2018. Achieving dft accuracy with a machine-learning interatomic potential: Thermomechanics and defects in bcc ferromagnetic iron. Physical Review Materials 2(1):013808 * [32] Willatt MJ, Musil F, Ceriotti M. 2019. Atom-density representations for machine learning. The Journal of Chemical Physics 150(15):154110 * [33] Schütt KT, Glawe H, Brockherde F, Sanna A, Müller KR, Gross EK. 2014. How to represent crystal structures for machine learning: Towards fast prediction of electronic properties. Physical Review B 89(20):205118 * [34] Moosavi SM, Álmos Novotny B, Ongari D, Moubarak E, Asgari M, et al. 2022. A data-science approach to predict the heat capacity of nanoporous materials. Nature Materials 21(12):1419–1425 * [35] Isayev O, Oses C, Toher C, Gossett E, Curtarolo S, Tropsha A. 2017. Universal fragment descriptors for predicting properties of inorganic crystals. Nature Communications 8(1):15679 * [36] Bartók AP, De S, Poelking C, Bernstein N, Kermode JR, et al. 2017. Machine learning unifies the modeling of materials and molecules. Science Advances 3(12):e1701816 * [37] Lazar EA, Lu J, Rycroft CH. 2022. Voronoi cell analysis: The shapes of particle systems. American Journal of Physics 90(6):469 * [38] Krishnapriyan AS, Montoya J, Haranczyk M, Hummelshøj J, Morozov D. 2021. Machine learning with persistent homology and chemical word embeddings improves prediction accuracy and interpretability in metal-organic frameworks. Scientific Reports 11(1):1–11 * [39] Rupp M, Tkatchenko A, Müller KR, Lilienfeld OAV. 2012. Fast and accurate modeling of molecular atomization energies with machine learning. Physical Review Letters 108(5):058301 * [40] Sauceda HE, Gálvez-González LE, Chmiela S, Paz-Borbón LO, Müller KR, Tkatchenko A. 2022. Bigdml—towards accurate quantum machine learning force fields for materials. Nature Communications 13(1):3733 * [41] Sanchez JM, Ducastelle F, Gratias D. 1984. Generalized cluster description of multicomponent systems. Physica A: Statistical Mechanics and its Applications 128(1-2):334–350 * [42] Chang JH, Kleiven D, Melander M, Akola J, Garcia-Lastra JM, Vegge T. 2019. Clease: a versatile and user-friendly implementation of cluster expansion method. Journal of Physics: Condensed Matter 31(32):325901 * [43] Hart GL, Mueller T, Toher C, Curtarolo S. 2021. Machine learning for alloys. Nature Reviews Materials 6(8):730–755 * [44] Nyshadham C, Rupp M, Bekker B, Shapeev AV, Mueller T, et al. 2019. Machine-learned multi-system surrogate models for materials prediction. npj Computational Materials 5(1):1–6 * [45] Yang JH, Chen T, Barroso-Luque L, Jadidi Z, Ceder G. 2022. Approaches for handling high-dimensional cluster expansions of ionic systems. npj Computational Materials 8(1):1–11 * [46] Drautz R. 2019. Atomic cluster expansion for accurate and transferable interatomic potentials. Physical Review B 99(1):014104 * [47] Batatia I, Kovács DP, Simm GNC, Ortner C, Csányi G. 2022a. Mace: Higher order equivariant message passing neural networks for fast and accurate force fields. ArXiv: 2206.07697 * [48] Carlsson G. 2020. Topological methods for data modelling. Nature Reviews Physics 2(12):697–708 * [49] Pun CS, Lee SX, Xia K. 2022. Persistent-homology-based machine learning: a survey and a comparative study. Artificial Intelligence Review 55(7):5169–5213 * [50] Jiang Y, Chen D, Chen X, Li T, Wei GW, Pan F. 2021. Topological representations of crystalline compounds for the machine-learning prediction of materials properties. npj Computational Materials 7(1):1–8 * [51] Lee Y, Barthel SD, Dłotko P, Moosavi SM, Hess K, Smit B. 2018. High-throughput screening approach for nanoporous materials genome using topological data analysis: Application to zeolites. Journal of Chemical Theory and Computation 14(8):4427–4437 * [52] Buchet M, Hiraoka Y, Obayashi I. 2018. Persistent homology and materials informatics. In: Tanaka, I. (eds) Nanoinformatics. Springer, Singapore * [53] Duvenaud D, Maclaurin D, Aguilera-Iparraguirre J, Gómez-Bombarelli R, Hirzel T, et al. 2015. Convolutional networks on graphs for learning molecular fingerprints. 28th Conference on Advances in Neural Information Processing Systems * [54] Reiser P, Neubert M, Eberhard A, Torresi L, Zhou C, et al. 2022. Graph neural networks for materials science and chemistry. Communications Materials 3(1):1–18 * [55] Schütt KT, Sauceda HE, Kindermans PJ, Tkatchenko A, Müller KR. 2018. Schnet – a deep learning architecture for molecules and materials. The Journal of Chemical Physics 148(24):241722 * [56] Chen C, Ye W, Zuo Y, Zheng C, Ong SP. 2019. Graph networks as a universal machine learning framework for molecules and crystals. Chemistry of Materials 31(9):3564–3572 * [57] Fung V, Zhang J, Juarez E, Sumpter BG. 2021a. Benchmarking graph neural networks for materials chemistry. npj Computational Materials 7(1):84 * [58] Chen C, Zuo Y, Ye W, Li X, Ong SP. 2021a. Learning properties of ordered and disordered materials from multi-fidelity data. Nature Computational Science 1(1):46–53 * [59] Choudhary K, DeCost B. 2021. Atomistic line graph neural network for improved materials property predictions. npj Computational Materials 7(1):185 * [60] Park CW, Wolverton C. 2020. Developing an improved crystal graph convolutional neural network framework for accelerated materials discovery. Physical Review Materials 4(6):063801 * [61] Karamad M, Magar R, Shi Y, Siahrostami S, Gates ID, Farimani AB. 2020. Orbital graph convolutional neural network for material property prediction. Physical Review Materials 4(9):093801 * [62] Karaguesian* J, Lunger* JR, Shao-Horn Y, Gomez-Bombarelli R. 2021. Crystal graph convolutional neural networks for per-site property prediction. Workshop on Machine Learning and the Physical Sciences at the 35th Conference on Neural Information Processing Systems * [63] Gong S, Xie T, Shao-Horn Y, Gomez-Bombarelli R, Grossman JC. 2022. Examining graph neural networks for crystal structures: limitations and opportunities for capturing periodicity. ArXiv: 2208.05039 * [64] Schütt KT, Unke OT, Gastegger M. 2021. Equivariant message passing for the prediction of tensorial properties and molecular spectra. 38th International Conference on Machine Learning * [65] Cheng J, Zhang C, Dong L. 2021. A geometric-information-enhanced crystal graph network for predicting properties of materials. Communications Materials 2(1):92 * [66] Louis SY, Zhao Y, Nasiri A, Wang X, Song Y, et al. 2020. Graph convolutional neural networks with global attention for improved materials property prediction. Physical Chemistry Chemical Physics 22(32):18141–18148 * [67] Yu H, Hong L, Chen S, Gong X, Xiang H. 2023. Capturing long-range interaction with reciprocal space neural network. ArXiv 2211.16684 * [68] Kong S, Ricci F, Guevarra D, Neaton JB, Gomes CP, Gregoire JM. 2022. Density of states prediction for materials discovery via contrastive learning from probabilistic embeddings. Nature Communications 13(1):949 * [69] Batzner S, Musaelian A, Sun L, Geiger M, Mailoa JP, et al. 2021. E(3)-equivariant graph neural networks for data-efficient and accurate interatomic potentials. Nature Communications 13(1):2453 * [70] Geiger M, Smidt T. 2022. e3nn: Euclidean neural networks. ArXiv 2207.09453 * [71] Thomas N, Smidt T, Kearnes S, Yang L, Li L, et al. 2018. Tensor field networks: Rotation- and translation-equivariant neural networks for 3d point clouds. ArXiv: 1802.08219 * [72] Gasteiger J, Becker F, Günnemann S. 2021. Gemnet: Universal directional graph neural networks for molecules. 35th Conference on Neural Information Processing Systems * [73] Gasteiger J, Shuaibi M, Sriram A, Günnemann S, Ulissi Z, et al. 2022. How do graph networks generalize to large and diverse molecular systems? ArXiv: 2204.02782 * [74] Chen Z, Andrejevic N, Smidt T, Ding Z, Xu Q, et al. 2021b. Direct prediction of phonon density of states with euclidean neural networks. Advanced Science 8(12):2004214 * [75] Kolluru A, Shuaibi M, Palizhati A, Shoghi N, Das A, et al. 2022a. Open challenges in developing generalizable large-scale machine-learning models for catalyst discovery. ACS Catalysis 12(14):8572–8581 * [76] Gibson J, Hire A, Hennig RG. 2022. Data-augmentation for graph neural network learning of the relaxed energies of unrelaxed structures. npj Computational Materials 8(1):211 * [77] Schmidt J, Pettersson L, Verdozzi C, Botti S, Marques MA. 2021. Crystal graph attention networks for the prediction of stable materials. Science Advances 7(49):7948 * [78] Zuo Y, Qin M, Chen C, Ye W, Li X, et al. 2021. Accelerating materials discovery with bayesian optimization and graph deep learning. Materials Today 51:126–135 * [79] Dunn A, Wang Q, Ganose A, Dopp D, Jain A. 2020. Benchmarking materials property prediction methods: the matbench test set and automatminer reference algorithm. npj Computational Materials 6(1):138 * [80] Callister WD, Rethwisch DG. 2018. Materials science and engineering: an introduction, 10th edition. Wiley Global Education * [81] Pauling L. 1929. The principles determining the structure of complex ionic crystals. Journal of the American Chemical Society 51(4):1010–1026 * [82] George J, Waroquiers D, Stefano DD, Petretto G, Rignanese GM, Hautier G. 2020. The limited predictive power of the pauling rules. Angewandte Chemie 132(19):7639–7645 * [83] Meredig B, Agrawal A, Kirklin S, Saal JE, Doak JW, et al. 2014. Combinatorial screening for new materials in unconstrained composition space with machine learning. Physical Review B 89(9):094104 * [84] Bartel CJ, Trewartha A, Wang Q, Dunn A, Jain A, Ceder G. 2020. A critical examination of compound stability predictions from machine-learned formation energies. npj Computational Materials 6(1):97 * [85] Stanev V, Oses C, Kusne AG, Rodriguez E, Paglione J, et al. 2018. Machine learning modeling of superconducting critical temperature. npj Computational Materials 4(1):29 * [86] Faber FA, Lindmaa A, Lilienfeld OAV, Armiento R. 2016. Machine learning energies of 2 million elpasolite (abc2d6) crystals. Physical Review Letters 117(13):135502 * [87] Oliynyk AO, Antono E, Sparks TD, Ghadbeigi L, Gaultois MW, et al. 2016. High-throughput machine-learning-driven synthesis of full-heusler compounds. Chemistry of Materials 28(20):7324–7331 * [88] Ouyang R, Curtarolo S, Ahmetcik E, Scheffler M, Ghiringhelli LM. 2018. Sisso: A compressed-sensing method for identifying the best low-dimensional descriptor in an immensity of offered candidates. Physical Review Materials 2(8):083802 * [89] Ghiringhelli LM, Vybiral J, Ahmetcik E, Ouyang R, Levchenko SV, et al. 2017. Learning physical descriptors for materials science by compressed sensing. New Journal of Physics 19(2):023017 * [90] Ghiringhelli LM, Vybiral J, Levchenko SV, Draxl C, Scheffler M. 2015. Big data of materials science: Critical role of the descriptor. Physical Review Letters 114(10):105503 * [91] Bartel CJ, Sutton C, Goldsmith BR, Ouyang R, Musgrave CB, et al. 2019. New tolerance factor to predict the stability of perovskite oxides and halides. Science Advances 5(2):eaav069 * [92] Zhou Q, Tang P, Liu S, Pan J, Yan Q, Zhang SC. 2018. Learning atoms for materials discovery. Proceedings of the National Academy of Sciences of the United States of America 115(28):E6411–E6417 * [93] Tshitoyan V, Dagdelen J, Weston L, Dunn A, Rong Z, et al. 2019. Unsupervised word embeddings capture latent knowledge from materials science literature. Nature 571(7763):95–98 * [94] Jha D, Ward L, Paul A, keng Liao W, Choudhary A, et al. 2018. Elemnet: Deep learning the chemistry of materials from only elemental composition. Scientific Reports 8(1):17593 * [95] Goodall RE, Lee AA. 2020. Predicting materials properties without crystal structure: deep representation learning from stoichiometry. Nature Communications 11(1):6280 * [96] Wang AYT, Kauwe SK, Murdock RJ, Sparks TD. 2021. Compositionally restricted attention-based network for materials property predictions. npj Computational Materials 7(1):77 * [97] Zhang Z, Tehrani AM, Oliynyk AO, Day B, Brgoch J. 2021. Finding the next superhard material through ensemble learning. Advanced Materials 33(5):2005112 * [98] Oliynyk AO, Adutwum LA, Rudyk BW, Pisavadia H, Lotfi S, et al. 2017. Disentangling structural confusion through machine learning: Structure prediction and polymorphism of equiatomic ternary phases abc. Journal of the American Chemical Society 139(49):17870–17881 * [99] Tian SIP, Walsh A, Ren Z, Li Q, Buonassisi T. 2022. What information is necessary and sufficient to predict materials properties using machine learning? ArXiv: 2206.04968 * [100] Artrith N. 2019. Machine learning for the modeling of interfaces in energy storage and conversion materials. Journal of Physics: Energy 1(3):032002 * [101] Rao RR, Kolb MJ, Giordano L, Pedersen AF, Katayama Y, et al. 2020. Operando identification of site-dependent water oxidation activity on ruthenium dioxide single-crystal surfaces. Nature Catalysis 3:516–525 * [102] Varley JB, Samanta A, Lordi V. 2017. Descriptor-based approach for the prediction of cation vacancy formation energies and transition levels. Journal of Physical Chemistry Letters 8(20):5059–5063 * [103] Zhang X, Wang H, Hickel T, Rogal J, Li Y, Neugebauer J. 2020. Mechanism of collective interstitial ordering in fe–c alloys. Nature Materials 19(8):849–854 * [104] Deringer VL, Bartók AP, Bernstein N, Wilkins DM, Ceriotti M, Csányi G. 2021. Gaussian process regression for materials and molecules. Chemical Reviews 121(16):10073–10141 * [105] Wan Z, Wang QD, Liu D, Liang J. 2021. Data-driven machine learning model for the prediction of oxygen vacancy formation energy of metal oxide materials. Physical Chemistry Chemical Physics 23(29):15675–15684 * [106] Witman M, Goyal A, Ogitsu T, McDaniel A, Lany. S. 2022. Graph neural network modeling of vacancy formation enthalpy for materials discovery and its application in solar thermochemical water splitting. ChemRxiv 10.26434 * [107] Medasani B, Gamst A, Ding H, Chen W, Persson KA, et al. 2016. Predicting defect behavior in b2 intermetallics by merging ab initio modeling and machine learning. npj Computational Materials 2(1):1 * [108] Ziletti A, Kumar D, Scheffler M, Ghiringhelli LM. 2018. Insightful classification of crystal structures using deep learning. Nature Communications 9(1):2775 * [109] Medford AJ, Vojvodic A, Hummelshøj JS, Voss J, Abild-Pedersen F, et al. 2015. From the sabatier principle to a predictive theory of transition-metal heterogeneous catalysis. Journal of Catalysis 328:36–42 * [110] Zhao ZJ, Liu S, Zha S, Cheng D, Studt F, et al. 2019. Theory-guided design of catalytic materials using scaling relationships and reactivity descriptors. Nature Reviews Materials 4(12):792–804 * [111] Hwang J, Rao RR, Giordano L, Katayama Y, Yu Y, Shao-Horn Y. 2017. Perovskites in catalysis and electrocatalysis. Science 358(6364):751–756 * [112] Calle-Vallejo F, Tymoczko J, Colic V, Vu QH, Pohl MD, et al. 2015. Finding optimal surface sites on heterogeneous catalysts by counting nearest neighbors. Science 350(6257):185–189 * [113] Fung V, Hu G, Ganesh P, Sumpter BG. 2021b. Machine learned features from density of states for accurate adsorption energy prediction. Nature Communications 12(1):88 * [114] Back S, Yoon J, Tian N, Zhong W, Tran K, Ulissi ZW. 2019. Convolutional neural network of atomic surface structures to predict binding energies for high-throughput screening of catalysts. Journal of Physical Chemistry Letters 10(15):4401–4408 * [115] Tran K, Ulissi ZW. 2018. Active learning across intermetallics to guide discovery of electrocatalysts for co2 reduction and h2 evolution. Nature Catalysis 1(9):696–703 * [116] Ghanekar PG, Deshpande S, Greeley J. 2022. Adsorbate chemical environment-based machine learning framework for heterogeneous catalysis. Nature Communications 13(1):5788 * [117] Kiyohara S, Oda H, Miyata T, Mizoguchi T. 2016. Prediction of interface structures and energies via virtual screening. Science Advances 2(11):1600746 * [118] Hu C, Zuo Y, Chen C, Ong SP, Luo J. 2020. Genetic algorithm-guided deep learning of grain boundary diagrams: Addressing the challenge of five degrees of freedom. Materials Today 38:49–57 * [119] Dai M, Demirel MF, Liang Y, Hu JM. 2021. Graph neural networks for an accurate and interpretable prediction of the properties of polycrystalline materials. npj Computational Materials 7:103 * [120] Huber L, Hadian R, Grabowski B, Neugebauer J. 2018. A machine learning approach to model solute grain boundary segregation. npj Computational Materials 4(1):64 * [121] Ye W, Zheng H, Chen C, Ong SP. 2022. A universal machine learning model for elemental grain boundary energies. Scripta Materialia 218:114803 * [122] Lazar EA. 2017. Vorotop: Voronoi cell topology visualization and analysis toolkit. Modelling and Simulation in Materials Science and Engineering 26(1):015011 * [123] Priedeman JL, Rosenbrock CW, Johnson OK, Homer ER. 2018. Quantifying and connecting atomic and crystallographic grain boundary structure using local environment representation and dimensionality reduction techniques. Acta Materialia 161:431–443 * [124] Rosenbrock CW, Homer ER, Csányi G, Hart GL. 2017. Discovering the building blocks of atomic systems using machine learning: application to grain boundaries. npj Computational Materials 3(1):29 * [125] Fujii S, Yokoi T, Fisher CA, Moriwake H, Yoshiya M. 2020. Quantitative prediction of grain boundary thermal conductivities from local atomic environments. Nature Communications 11(1):1854 * [126] Sharp TA, Thomas SL, Cubuk ED, Schoenholz SS, Srolovitz DJ, Liu AJ. 2018. Machine learning determination of atomic dynamics at grain boundaries. Proceedings of the National Academy of Sciences of the United States of America 115(43):10943–10947 * [127] Batatia I, Batzner S, Kovács DP, Musaelian A, Simm GNC, et al. 2022b. The design space of e(3)-equivariant atom-centered interatomic potentials. ArXiv: 2205.06643 * [128] Axelrod S, Shakhnovich E, Gómez-Bombarelli R. 2022. Excited state non-adiabatic dynamics of large photoswitchable molecules using a chemically transferable machine learning potential. Nature Communications 13(1):3440 * [129] Li X, Li B, Yang Z, Chen Z, Gao W, Jiang Q. 2022. A transferable machine-learning scheme from pure metals to alloys for predicting adsorption energies. Journal of Materials Chemistry A 10(2):872–880 * [130] Harper DR, Nandy A, Arunachalam N, Duan C, Janet JP, Kulik HJ. 2022. Representations and strategies for transferable machine learning improve model performance in chemical discovery. The Journal of Chemical Physics 156(7):074101 * [131] Unke OT, Chmiela S, Gastegger M, Schütt KT, Sauceda HE, Müller KR. 2021a. Spookynet: Learning force fields with electronic degrees of freedom and nonlocal effects. Nature Communications 12(1):7273 * [132] Husch T, Sun J, Cheng L, Lee SJ, Miller TF. 2021. Improved accuracy and transferability of molecular-orbital-based machine learning: Organics, transition-metal complexes, non-covalent interactions, and transition states. The Journal of Chemical Physics 154(6):064108 * [133] Zheng X, Zheng P, Zhang RZ. 2018. Machine learning material properties from the periodic table using convolutional neural networks. Chemical Science 9(44):8426–8432 * [134] Feng S, Fu H, Zhou H, Wu Y, Lu Z, Dong H. 2021. A general and transferable deep learning framework for predicting phase formation in materials. npj Computational Materials 7(1):10 * [135] Jha D, Choudhary K, Tavazza F, keng Liao W, Choudhary A, et al. 2019. Enhancing materials property prediction by leveraging computational and experimental data using deep transfer learning. Nature Communications 10(1):5316 * [136] Yamada H, Liu C, Wu S, Koyama Y, Ju S, et al. 2019. Predicting materials properties with little data using shotgun transfer learning. ACS Central Science 5(10):1717–1730 * [137] Gupta V, Choudhary K, Tavazza F, Campbell C, keng Liao W, et al. 2021. Cross-property deep transfer learning framework for enhanced predictive analytics on small materials data. Nature Communications 12(1):6595 * [138] Lee J, Asahi R. 2021. Transfer learning for materials informatics using crystal graph convolutional neural network. Computational Materials Science 190:110314 * [139] Kolluru A, Shoghi N, Shuaibi M, Goyal S, Das A, et al. 2022b. Transfer learning using attentions across atomic systems with graph neural networks (taag). The Journal of Chemical Physics 156(18):184702 * [140] Chen C, Ong SP. 2021. Atomsets as a hierarchical transfer learning framework for small and large materials datasets. npj Computational Materials 7(1):173 * [141] Cubuk ED, Sendek AD, Reed EJ. 2019. Screening billions of candidates for solid lithium-ion conductors: A transfer learning approach for small data. The Journal of Chemical Physics 150(21):214701 * [142] Greenman KP, Green WH, Gómez-Bombarelli R. 2022. Multi-fidelity prediction of molecular optical peaks with deep learning. Chemical Science 13(4):1152–1162 * [143] Kong S, Guevarra D, Gomes CP, Gregoire JM. 2021. Materials representation and transfer learning for multi-property prediction. Applied Physics Reviews 8(2):021409 * [144] Kingma DP, Welling M. 2013. Auto-encoding variational bayes. 2nd International Conference on Learning Representations * [145] Goodfellow IJ, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, et al. 2014. Generative adversarial networks. 27th Conference on Advances in Neural Information Processing Systems * [146] Nouira A, Sokolovska N, Crivello JC. 2018. Crystalgan: Learning to discover crystallographic structures with generative adversarial networks. 2019 Spring Symposium on Combining Machine Learning with Knowledge Engineering * [147] Shi C, Luo S, Xu M, Tang J. 2021. Learning gradient fields for molecular conformation generation. 38th International Conference on Machine Learning * [148] Ho J, Jain A, Abbeel P. 2020. Denoising diffusion probabilistic models. 34th Conference on Neural Information Processing Systems * [149] Schwalbe-Koda D, Gómez-Bombarelli R. 2020. Generative models for automatic chemical design. Lecture Notes in Physics 968:445–467 * [150] Noh J, Gu GH, Kim S, Jung Y. 2020. Machine-enabled inverse design of inorganic solid materials: promises and challenges. Chemical Science 11(19):4871–4881 * [151] Fuhr AS, Sumpter BG. 2022. Deep generative models for materials discovery and machine learning-accelerated innovation. Frontiers in Materials 9:182 * [152] Xie T, Fu X, Ganea OE, Barzilay R, Jaakkola T. 2021. Crystal diffusion variational autoencoder for periodic material generation. 10th International Conference on Learning Representations * [153] Yao Z, Sánchez-Lengeling B, Bobbitt NS, Bucior BJ, Kumar SGH, et al. 2021. Inverse design of nanoporous crystalline reticular materials with deep generative models. Nature Machine Intelligence 3(1):76–86 * [154] Ren Z, Tian SIP, Noh J, Oviedo F, Xing G, et al. 2022. An invertible crystallographic representation for general inverse design of inorganic crystals with targeted properties. Matter 5(1):314–335 * [155] Long T, Fortunato NM, Opahle I, Zhang Y, Samathrakis I, et al. 2021. Constrained crystals deep convolutional generative adversarial network for the inverse design of crystal structures. npj Computational Materials 7(1):66 * [156] Kim S, Noh J, Gu GH, Aspuru-Guzik A, Jung Y. 2020. Generative adversarial networks for crystal structure prediction. ACS Central Science 6(8):1412–1420 * [157] Fung V, Jia S, Zhang J, Bi S, Yin J, Ganesh P. 2022. Atomic structure generation from reconstructing structural fingerprints. Machine Learning: Science and Technology 3(4):045018 * [158] Kim B, Lee S, Kim J. 2020. Inverse design of porous materials using artificial neural networks. Science Advances 6(1):aax9324 * [159] Musaelian A, Batzner S, Johansson A, Sun L, Owen CJ, et al. 2022. Learning local equivariant representations for large-scale atomistic dynamics. ArXiv: 2204.05249 * [160] Unke OT, Chmiela S, Sauceda HE, Gastegger M, Poltavsky I, et al. 2021b. Machine learning force fields. Chemical Reviews 121(16):10142–10186 * [161] Grisafi A, Nigam J, Ceriotti M. 2021. Multi-scale approach for the prediction of atomic scale properties. Chemical Science 12:2078 * [162] Schaarschmidt M, Riviere M, Ganose AM, Spencer JS, Gaunt AL, et al. 2022. Learned force fields are ready for ground state catalyst discovery. ArXiv: 2209.12466 * [163] Lan J, Palizhati A, Shuabi M, Wood BM, Wander B, et al. 2023. Adsorbml: Accelerating adsorption energy calculations with machine learning. ArXiv 2211.16486 * [164] Kresse G, Hafner J. 1993. Ab initio molecular dynamics for liquid metals. Phys. Rev. B 47(1):558–561 * [165] Chen C, Ong SP. 2022. A universal graph deep learning interatomic potential for the periodic table. Nature Computational Science 2(11):718–728 * [166] Legrain F, Carrete J, Roekeghem AV, Curtarolo S, Mingo N. 2017. How chemical composition alone can predict vibrational free energies and entropies of solids. Chemistry of Materials 29(15):6220–6227 * [167] Zhuo Y, Tehrani AM, Brgoch J. 2018. Predicting the band gaps of inorganic solids by machine learning. Journal of Physical Chemistry Letters 9(7):1668–1673 * [168] Hsu T, Epting WK, Kim H, Abernathy HW, Hackett GA, et al. 2020. Microstructure generation via generative adversarial network for heterogeneous, topologically complex 3d materials. JOM 73(1):90–102 * [169] Song Y, Siriwardane EM, Zhao Y, Hu J. 2020. Computational discovery of new 2d materials using deep learning generative models. ACS Applied Materials and Interfaces 13(45) * [170] Liu Z, Zhu D, Rodrigues SP, Lee KT, Cai W. 2018. Generative model for the inverse design of metasurfaces. Nano Letters 18(10):6570–6576 * [171] Dan Y, Zhao Y, Li X, Li S, Hu M, Hu J. 2020. Generative adversarial networks (gan) based efficient sampling of chemical composition space for inverse design of inorganic materials. npj Computational Materials 6(1):84
[a]Wolfgang Unger # Hamiltonian Lattice QCD from Strong Coupling Expansion Pratitee Pattanaik ###### Abstract We present generalizations of Hamiltonian Lattice QCD as derived from the continuous time limit of strong coupling lattice QCD: we discuss the flavor dependence and the effect of gauge corrections. This formalism can be applied at finite temperature and baryon density as well as isospin density and allows both for analytic and numeric investigations that are sign problem-free. ## 1 Introduction The Hamiltonian formulation of lattice QCD has been discussed in detail in [1] in the strong coupling limit for ${N_{f}}=1$. In contrast to Hamiltonian formulations in the early days of lattice QCD [2] this formulation is based on the dual representation that is obtained when integrating out the gauge links first, and then the Grassmann-valued fermions [3]. The resulting dual degrees of freedom are color singlets, such as mesons and baryons. In this representation, the finite density sign problem is much milder, as it is re- expressed in terms of the geometry of baryonic world-lines. The dual representation of lattice QCD with staggered fermions has been studied in the strong coupling limit both via mean-field theory [4] and Monte Carlo [5, 6] and has been extended in both approaches to include gauge corrections [7, 8]. The Hamiltonian formulation is based on the continuous time limit of the dual representation: at fixed bare temperature $aT=\frac{a}{a_{t}N_{t}}$, the limits $a_{t}\rightarrow 0$ and $N_{t}\rightarrow\infty$ are taken simultaneously [9]. This limit requires to determine the non-perturbative relation between the bare anisotropy $\gamma$ and the physical anisotropy $\frac{a}{a_{t}}\equiv\xi(\gamma)$, based on pion current fluctuations. A conservation law for the pion current can be identified: if a quantum number $\mathfrak{m}(x)$ is raised/lowered by a spatial dimer, then at the site connected by the spatial dimer the quantum number is lowered/raised. This is a direct consequence of the even/odd decomposition for staggered fermions. With interactions derived from a high temperature series, the resulting partition sum can be expressed in terms of a Hamiltonian that is composed of mesonic annihilation and creation operators $\hat{J}^{\pm}$: $\displaystyle Z_{\rm CT}(\mathcal{T},\mu_{\mathcal{B}})={\rm Tr}_{\mathfrak{h}}\left[e^{(\hat{\mathcal{H}}+\hat{\mathcal{N}}\mu_{\mathcal{B}})/\mathcal{T}}\right],\qquad\hat{\mathcal{H}}=\frac{1}{2}\sum_{\langle\vec{x},\vec{y}\rangle}\left(\hat{J}^{+}_{\vec{x}}\hat{J}^{-}_{\vec{y}}+\hat{J}^{-}_{\vec{x}}\hat{J}^{+}_{\vec{y}}\right),\qquad\hat{\mathcal{N}}=\sum_{\vec{x}}\hat{\omega}_{x},$ $\displaystyle\hat{J}^{+}={\footnotesize\left(\begin{array}[]{cccc|cc}0&0&0&0&&\\\ \hat{v}_{\hskip 0.56003pt\text{\lx@text@tweaked{yoffset=-0.5ex}{\lx@mvs@LSteel}}\hskip 0.56003pt}&0&0&0&&\\\ 0&\hat{v}_{\hskip 0.56003pt\text{\lx@text@tweaked{yoffset=0.5ex}{\lx@mvs@TSteel}}\hskip 0.56003pt}&0&0&&\\\ 0&0&\hat{v}_{\hskip 0.56003pt\text{\lx@text@tweaked{yoffset=-0.5ex}{\lx@mvs@LSteel}}\hskip 0.56003pt}&0&&\\\ \hline\cr&&&&0&0\\\ &&&&0&0\\\ \end{array}\right)},\qquad\hat{J}^{-}=(\hat{J}^{+})^{T},\qquad\hat{\omega}={\footnotesize\left(\begin{array}[]{cccc|cc}0&0&0&0&&\\\ 0&0&0&0&&\\\ 0&0&0&0&&\\\ 0&0&0&0&&\\\ \hline\cr&&&&1&0\\\ &&&&0&-1\\\ \end{array}\right)}$ (13) with the local Hilbert space given by $|\mathfrak{h}\rangle=|\mathfrak{m};\mathfrak{b}\rangle=|0,\pi,2\pi,3\pi;B^{+},B^{-}\rangle.$ The matrix elements $\hat{v}_{\hskip 0.70004pt\text{\lx@text@tweaked{yoffset=-0.5ex}{\lx@mvs@LSteel}}\hskip 0.70004pt}=1$ and $\hat{v}_{\hskip 0.70004pt\text{\lx@text@tweaked{yoffset=0.5ex}{\lx@mvs@TSteel}}\hskip 0.70004pt}=\frac{2}{\sqrt{3}}$ are computed from the local weights of the corresponding meson vertices. Since Pauli saturation holds on the level of the quarks and mesons have a fermionic substructure, the meson occupation numbers $|\mathfrak{m}\rangle$ are also bounded from above. This results in a particle-hole symmetry, leading to an SU(2) algebra which is $d={N_{c}}+1$-dimensional: $\displaystyle\hat{J}_{1}$ $\displaystyle=\frac{\sqrt{{N_{c}}}}{2}\left(\hat{J}^{+}+\hat{J}^{-}\right),\quad\hat{J}_{2}=\frac{\sqrt{{N_{c}}}}{2i}\left(\hat{J}^{+}-\hat{J}^{-}\right),$ $\displaystyle\hat{J}_{3}$ $\displaystyle=i[J_{1},J_{2}]=\frac{{N_{c}}}{2}[\hat{J}^{+},\hat{J}^{-}],\quad\hat{J}^{2}=\frac{{N_{c}}({N_{c}}+2)}{4},$ $\displaystyle\hat{J}_{3}\left|\frac{{N_{c}}}{2},\mathfrak{s}\right\rangle$ $\displaystyle=\mathfrak{s}\left|\frac{{N_{c}}}{2},\mathfrak{s}\right\rangle,$ $\displaystyle\hat{J}^{2}\left|\frac{{N_{c}}}{2}\mathfrak{s}\right\rangle$ $\displaystyle=\frac{{N_{c}}\left({N_{c}}+2\right)}{4}\left|\frac{{N_{c}}}{2},\mathfrak{s}\right\rangle,$ $\displaystyle[\hat{J}^{2},\hat{J}_{3}]$ $\displaystyle=0$ (14) with $\mathfrak{m}\mapsto\mathfrak{s}=\mathfrak{m}-\frac{{N_{c}}}{2}$. The ${N_{f}}=1$ formulation at strong coupling has many similarities with full QCD, and via Quantum Monte Carlo the grand-canonical and canonical phase diagram could be determined [1]. The nuclear interactions are of entropic nature: the nucleons, which are point-like in the strong coupling limit, attract each other due to the modification of the pion bath that surrounds them. However, pion exchange cannot be realized in the strong coupling limit for ${N_{f}}=1$ since the Grassmann constraint does not allow for pions and nucleons to overlap. This is clearly a lattice artifact of the strong coupling limit, but this is overcome by formulations with ${N_{f}}>1$ and/or by including gauge corrections. The aim of the generalizations discussed in this proceedings is to allow for Quantum Monte Carlo simulations (QMC) at both non-zero baryon and isospin chemical potential, and possibly also strangeness chemical potential. We will argue that in the strong coupling limit, the ${N_{f}}=2$ formulation remains sign problem-free. ## 2 Hamiltonian formulation in the strong coupling limit for any $N_{f}$ Strong coupling lattice QCD with staggered fermions for ${N_{f}}>1$ admits more than one baryon per site and the Grassmann constraint allows for pion exchange between them, modifying nuclear interactions substantially. It has so far only been studied via mean-field theory [10]. A Hamiltonian formulation for ${N_{f}}=2$ allows for QMC and provides a more realistic scenario for nuclear interactions and the phase diagram. It also compares better to the strong coupling regime with Wilson fermions in a world-line formulation, as discussed in the context of the 3-dim. Polyakov effective theory (for a review, see [11]). Also for ${N_{f}}>1$ the suppression of spatial bonds $\gamma^{-k}$, $k>2$ applies, hence the continuous time limit is well defined. The first step to derive the Hamiltonian is to determine the local Hilbert space $\mathbbm{H}_{\mathfrak{h}}$ via canonical sectors $B\in\\{-{N_{f}},\ldots,{N_{f}}\\}$, i.e. we need to consider all possible single-site quantum states $\mathfrak{h}$ in a non-interacting theory to establish the basis of quantum states that generalize the ${N_{f}}=1$ states $|\mathfrak{m}\rangle$ and $|\mathfrak{b}\rangle$. The static partition sum for any ${N_{c}}$, ${N_{f}}$ is: $\displaystyle Z_{\rm stat}({N_{c}},{N_{f}})\sum_{B=-{N_{f}}}^{{N_{f}}}\prod_{a=0}^{N_{c}}\frac{a!(2{N_{f}}+a)!}{({N_{f}}+a+B)!(Nf+a-B)!}e^{B\mu_{B}/T}$ (15) where the coefficient encode the number of hadronic states, i.e. the dimension of $\mathfrak{h}$. For ${N_{c}}=3$ (separated in canonical sectors) * • for ${N_{f}}=1$: $d=[1,4,1]=6$, * • for ${N_{f}}=2$: $d=[1,20,50,20,1]=92$, * • for ${N_{f}}=3$: $d=[1,56,490,980,490,56,1]=2074$. Unfortunately there is no general formula known including the isospin sectors, required for $\mu_{I}\neq 0$. The individual states $\mathfrak{h}$ are also required for constructing the flavored $\hat{J}^{\pm}$. To arrive at this basis, we consider the ${\rm SU}(3)$ one-link integrals in terms of the fermion matrix $(\mathcal{M})_{ij}=\bar{\chi}^{\alpha}_{i}(x)\chi^{\alpha}_{i}(y)$ and $(\mathcal{M}^{\dagger})_{kl}=\chi^{\beta}_{k}(y)\bar{\chi}^{\beta}_{l}(x)$, valid also for ${N_{f}}\geq 1$: $\displaystyle\mathcal{J}(\mathcal{M},\mathcal{M}^{\dagger})$ $\displaystyle=\int\limits_{{\rm SU}(3)}dUe^{{\rm tr}[U\mathcal{M}^{\dagger}+U^{\dagger}\mathcal{M}]}=\sum_{B=-{N_{f}}}^{N_{f}}\sum_{n_{1},n_{2},n_{3}}C_{B,n_{1},n_{2},n_{3}}\frac{E^{B}}{|B|!}\prod_{i=1}^{3}\frac{X_{i}^{n_{i}}}{n_{i}!},\quad E=\begin{cases}\det{\mathcal{M}}&B>0\\\ 1&B=0\\\ \det{\mathcal{M}^{\dagger}}&B<0\\\ \end{cases}$ $\displaystyle C_{B,n_{1},n_{2},n_{3}}$ $\displaystyle=2\binom{n_{1}+2n_{2}+4n_{3}+2|B|+2}{n_{3}+|B|}\frac{|B|!}{(n_{1}+2n_{2}+3n_{3}+2|B|+2)!(n_{2}+2n_{3}+|B|+1)!}$ (16) which in contrasts to [12] is now expressed in a more suitable basis involving the baryon sectors $B$. The sum over $B$ and $n_{i}$ $(i=1,\ldots 3)$ terminates due to the Grassmann integration, depending on ${N_{f}}$. The corresponding invariants $D$, $X_{i}$ can be evaluated as follows: $\displaystyle X_{1}$ $\displaystyle={\rm tr}[\mathcal{M}\mathcal{M}^{\dagger}]={\rm Tr}[M_{x}M_{y}]=k_{U}+k_{D}+\ldots+k_{\pi^{+}}+k_{\pi^{-}}+\ldots$ $\displaystyle X_{2}$ $\displaystyle=\frac{1}{2}\left({\rm tr}[\mathcal{M}\mathcal{M}^{\dagger}]-{\rm tr}[(\mathcal{M}\mathcal{M}^{\dagger})^{2}]\right)=\frac{1}{2}\left({\rm Tr}[M_{x}M_{y}]^{2}+{\rm Tr}[(M_{x}M_{y})^{2}]\right)=X_{1}^{2}-D_{2}$ $\displaystyle X_{3}$ $\displaystyle=\det[\mathcal{M}\mathcal{M}^{\dagger}]=\frac{1}{6}\left({\rm Tr}[M_{x}M_{y}]^{3}+3{\rm Tr}[M_{x}M_{y}]{\rm Tr}[(M_{x}M_{y})^{2}]+2{\rm Tr}[(M_{x}M_{y})^{3}]\right)=X_{1}^{3}-2X_{1}D_{2}+D_{3}$ $\displaystyle D_{2}$ $\displaystyle=\frac{1}{2}\left({\rm Tr}[M_{x}M_{y}]^{2}-{\rm Tr}[(M_{x}M_{y})^{2}]\right)=k_{U}k_{D}+k_{\pi^{+}}k_{\pi^{-}}+\ldots-(k^{(2)}_{\pi^{+}\pi^{-},UD}+k^{(2)}_{UD,\pi^{+}\pi^{-}}+\ldots)$ $\displaystyle D_{3}$ $\displaystyle=\frac{1}{6}\left({\rm Tr}[M_{x}M_{y}]^{3}-3{\rm Tr}[M_{x}M_{y}]{\rm Tr}[(M_{x}M_{y})^{2}]+2{\rm Tr}[(M_{x}M_{y})^{3}]\right)=\frac{1}{6}(3X_{1}D_{2}-X_{1}^{3})$ $\displaystyle E$ $\displaystyle=\det[\mathcal{M}]=\sum_{f\leq g\leq h}B_{fgh}\quad(B>0)\quad\text{or}\quad E=\det[\mathcal{M}^{\dagger}]=\sum_{f\leq g\leq h}\bar{B}_{fgh},\quad(B<0)$ (17) where the invariants are expressed in terms of nearest neighbors $(M_{x}M_{y})^{n}=(-1)^{n+1}(\mathcal{M}\mathcal{M}^{\dagger})^{n}$. To be explicit, for ${N_{f}}=2$, the (anti-) baryons and flavored dimers in terms of quarks are: $\displaystyle B_{uud}$ $\displaystyle=\bar{u}\bar{u}\bar{d}_{x}uud_{y},$ $\displaystyle B_{udd}$ $\displaystyle=\bar{u}\bar{d}\bar{d}_{x}udd_{y},$ $\displaystyle B_{uuu}$ $\displaystyle=\bar{u}\bar{u}\bar{u}_{x}uuu_{y},$ $\displaystyle B_{ddd}$ $\displaystyle=\bar{d}\bar{d}\bar{d}_{x}ddd_{y},$ $\displaystyle\bar{B}_{uud}$ $\displaystyle=uud_{x}\bar{u}\bar{u}\bar{d}_{y},$ $\displaystyle\bar{B}_{udd}$ $\displaystyle=udd_{x}\bar{u}\bar{d}\bar{d}_{y},$ $\displaystyle\bar{B}_{uuu}$ $\displaystyle=uuu_{x}\bar{u}\bar{u}\bar{u}_{y},$ $\displaystyle\bar{B}_{ddd}$ $\displaystyle=ddd_{x}\bar{d}\bar{d}\bar{d}_{y},$ $\displaystyle k_{U}$ $\displaystyle=\bar{u}u(x)\bar{u}u(y),$ $\displaystyle k_{D}$ $\displaystyle=\bar{d}d(x)\bar{d}d(y),$ $\displaystyle k_{\pi^{+}}$ $\displaystyle=\bar{u}d(x)\bar{d}u(y),$ $\displaystyle k_{\pi^{-}}$ $\displaystyle=\bar{d}u(x)\bar{u}d(y)$ $\displaystyle k^{(2)}_{\pi^{+}\pi^{-},UD}$ $\displaystyle=\bar{u}d(x)\bar{d}u(x)\bar{u}u(y)\bar{d}d(y),$ $\displaystyle k^{(2)}_{UD,\pi^{+}\pi^{-}}$ $\displaystyle=\bar{u}u(x)\bar{d}d(x)\bar{u}d(y)\bar{d}u(y).$ (18) Every state of the local Hilbert space can be described by a set of ${N_{f}}^{2}$ charges: baryon number $B\in\\{-{N_{f}},\ldots{N_{f}}\\}$, isospin number $I\in\\{-{N_{c}},\ldots{N_{c}}\\}$, diagonal flavor numbers $U,D,S,\ldots\in\\{0,\ldots{N_{c}}\\}$ and for ${N_{f}}>2$ additional generalizations of isospin $K_{1},K_{2},\ldots\in\\{-{N_{c}},\ldots{N_{c}}\\}$. The one-link integral $J_{k}^{(B)}$ from Eq. (16) of order $k=n_{1}+2n_{2}+3n_{3}$ can be evaluated for various ${N_{f}}$. We show here the general result for SU(3) in terms of $X=X_{1}$, $D_{2}$, $D_{3}$ and $E$: in black the ${N_{f}}=1$ contributions, in blue the additional ${N_{f}}=2$ contributions and in red the additional ${N_{f}}=3$ contributions: $\displaystyle J_{0}^{(0)}$ $\displaystyle=1,\qquad J_{1}^{(0)}=\frac{1}{3}X,\qquad J_{2}^{(0)}=\frac{1}{12}X^{2}-{\color[rgb]{0,0,1}{\frac{1}{24}{D_{2}}}},\qquad J_{3}^{(0)}=\frac{1}{36}X^{3}-{\color[rgb]{0,0,1}{\frac{1}{24}X{D_{2}}}}+{\color[rgb]{1,0,0}{\frac{1}{60}{D_{3}}}},$ $\displaystyle J_{4}^{(0)}$ $\displaystyle=\frac{1}{720}\left({\color[rgb]{0,0,1}{\frac{37}{12}X^{4}-\frac{11}{2}X^{2}{D_{2}}+\frac{1}{6}{D_{2}}^{2}}}+{\color[rgb]{1,0,0}{\frac{7}{3}X{D_{3}}}}\right),$ $\displaystyle J_{5}^{(0)}$ $\displaystyle=\frac{1}{840}\left({\color[rgb]{0,0,1}{\frac{47}{120}X^{5}-\frac{31}{36}X^{3}{D_{2}}+\frac{1}{4}X{D_{2}}^{2}}}+{\color[rgb]{1,0,0}{\frac{1}{3}X^{2}{D_{3}}-\frac{1}{9}\,D_{2}D_{3}}}\right),$ $\displaystyle J_{6}^{(0)}$ $\displaystyle=\frac{1}{448}\left({\color[rgb]{0,0,1}{\frac{53}{2700}X^{6}-\frac{1}{18}X^{4}{D_{2}}+\frac{5}{144}X^{2}{D_{2}}^{2}+\frac{1}{45}{D_{3}}X^{3}-\frac{1}{6480}D_{2}^{3}}}-{\color[rgb]{1,0,0}{\frac{1}{40}XD_{2}{D_{3}}+\frac{1}{240}{D_{3}}^{2}}}\right),$ $\displaystyle J_{7}^{(0)}$ $\displaystyle=\frac{1}{25920}\left({\color[rgb]{1,0,0}{\frac{241}{2940}X^{7}-\frac{19}{70}D_{2}X^{5}+\frac{227}{1008}{D_{2}}^{2}X^{3}+\frac{1}{9}\,{D_{3}}X^{4}-\frac{13}{1008}D_{2}^{3}X-\frac{29}{168}D_{2}D_{3}X^{2}+\frac{1}{168}{D_{2}}^{2}D_{3}}}+{\color[rgb]{1,0,0}{\frac{11}{336}\,D_{3}^{2}X}}\right),$ $\displaystyle J_{8}^{(0)}$ $\displaystyle=\frac{1}{21772800}\left({\color[rgb]{1,0,0}{\frac{13259}{3360}\,X^{8}-\frac{149}{10}D_{2}X^{6}+\frac{631}{40}{D_{2}}^{2}X^{4}+\frac{121}{20}\,D_{3}X^{5}-\frac{403}{120}\,D_{2}^{3}X^{2}-\frac{143}{12}D_{2}D_{3}X^{3}+\frac{1}{240}\,D_{2}^{4}}}\right.$ $\displaystyle\left.\hskip 45.5244pt{\color[rgb]{1,0,0}{+\frac{11}{4}X{D_{2}}^{2}D_{3}+\frac{11}{5}X^{2}D_{3}^{2}-\frac{11}{20}D_{2}D_{3}^{2}}}\right),$ $\displaystyle J_{9}^{(0)}$ $\displaystyle=\frac{1}{17107200}\left({\color[rgb]{1,0,0}{\frac{63163}{423360}X^{9}-\frac{377}{588}X^{7}D_{2}+\frac{1429}{1680}X^{5}{D_{2}}^{2}+\frac{73}{280}X^{6}D_{3}-\frac{1663}{5040}X^{3}{D_{2}}^{3}-\frac{551}{840}X^{4}{D_{2}}{D_{3}}}}\right.$ $\displaystyle\hskip 45.5244pt\left.{\color[rgb]{1,0,0}{+\frac{17}{3360}X{D_{2}}^{4}+\frac{71}{210}X^{2}{D_{2}}^{2}D_{3}+\frac{13}{105}X^{3}{D_{3}}^{2}-\frac{1}{420}{D_{2}}^{3}{D_{3}}-\frac{13}{120}X{D_{2}}{D_{3}}^{2}+\frac{13}{1260}{D_{3}}^{3}}}\right),$ $\displaystyle J_{0}^{(1)}$ $\displaystyle=\frac{E}{6},\qquad J_{1}^{(1)}=\frac{E}{6}\left({\color[rgb]{0,0,1}{\frac{1}{4}X}}\right),\qquad J_{2}^{(1)}=\frac{E}{6}\left({\color[rgb]{0,0,1}{\frac{1}{24}X^{2}-\frac{1}{60}D_{2}}}\right),\qquad J_{3}^{(1)}=\frac{E}{6}\left({\color[rgb]{0,0,1}{\frac{1}{144}X^{3}-\frac{1}{120}XD_{2}}}+{\color[rgb]{1,0,0}{\frac{1}{360}{D_{3}}}}\right),$ $\displaystyle J_{4}^{(1)}$ $\displaystyle=\frac{E}{6}{\color[rgb]{1,0,0}{\left(\frac{1}{1344}X^{4}-\frac{1}{840}X^{2}D_{2}+\frac{1}{20160}{D_{2}}^{2}+\frac{1}{2240}X{D_{3}}\right)}},$ $\displaystyle J_{5}^{(1)}$ $\displaystyle=\frac{E}{6}{\color[rgb]{1,0,0}{\left(\frac{143}{2419200}X^{5}-\frac{29}{241920}\,D_{2}X^{3}+\frac{1}{32256}D_{2}^{2}X+\frac{1}{23040}\,D_{3}X^{2}-\frac{1}{80640}D_{2}D_{3}\right)}},$ $\displaystyle J_{6}^{(1)}$ $\displaystyle=\frac{E}{6}{\color[rgb]{1,0,0}{\left(\frac{19}{4838400}X^{6}-\frac{221}{21772800}X^{4}{D_{2}}+\frac{241}{43545600}X^{2}{D_{2}}^{2}+\frac{11}{2903040}X^{3}D_{3}-\frac{1}{21772800}{D_{2}}^{3}-\frac{11}{3110400}X{D_{2}}{D_{3}}+\frac{11}{21772800}{D_{3}}^{2}\right)}},$ $\displaystyle J_{0}^{(2)}$ $\displaystyle={\color[rgb]{0,0,1}{\frac{E^{2}}{144}}},\qquad J_{1}^{(2)}={\color[rgb]{0,0,1}{\frac{E^{2}}{144}}}\left({\color[rgb]{1,0,0}{\frac{1}{5}X}}\right),\qquad J_{2}^{(2)}={\color[rgb]{0,0,1}{\frac{E^{2}}{144}}}\left({\color[rgb]{1,0,0}{\frac{1}{40}X^{2}-\frac{1}{120}D_{2}}}\right),\qquad J_{3}^{(2)}={\color[rgb]{0,0,1}{\frac{E^{2}}{144}}}\left({\color[rgb]{1,0,0}{\frac{1}{360}X^{3}-\frac{1}{360}XD_{2}+\frac{1}{1260}D_{3}}}\right),$ $\displaystyle J_{0}^{(3)}$ $\displaystyle={\color[rgb]{1,0,0}{\frac{1}{8640}E^{3}}}$ (19) The number of hadronic states in each conserved charge sector $(B,I,S,\ldots)$ can be obtained from the one-link integral by combining them in alternating chains $J_{k}^{(B)}J_{(N_{f}-B){N_{c}}-k}^{(B)}$, with $k$ denoting the mesonic occupation number $\mathfrak{m}_{k}$. To obtain the Hamiltonian, we still have to integrate out the Grassmann variables. Here we consider the chiral limit only. The Grassmann constraint then dictates that all quarks $u,d,s,\ldots$ and anti-quarks $\bar{u},\bar{d},\bar{s},\ldots$ within mesons or baryons appear exactly ${N_{c}}$ times. The Grassmann integral in the chiral limit for U(3) (i.e. SU(3) with $B=0$) for a given site $x$ is $\displaystyle I_{G}$ $\displaystyle=\int\prod_{\alpha}\prod_{f}[{\rm d}\bar{f}_{\alpha}{\rm d}f_{\alpha}]\prod_{f,g}(\bar{f}g)^{k_{\bar{f}g}}(-1)^{\frac{1}{2}\sum\limits_{f\neq g}k_{\bar{f}g}}F_{N_{f}}({N_{c}},\\{k_{\bar{f}g}\\})$ (20) with $k_{\bar{f}g}$ the sum of flavored dimers attached to site $x$, and with $F_{N_{f}}({N_{c}})$ some ${N_{c}}$-dependent function, given the power $k_{fg}$ of charged fluxes ($f\neq g$). The result for non-zero baryon number $B\neq 0$ and details on $F_{N_{f}}({N_{c}})$ will be presented in a forthcoming publication. On a given configuration, the Grassmann integration simplifies due to flux conservation: for the non-diagonal pseudo-scalar mesons $k_{\bar{f}g}=k_{\bar{g}f}$ such that all minus signs from the Grassmann integration cancel. Only the non-trivial contribution of type $k^{(2)}_{UD,\pi^{+}\pi^{-}}$ and its generalizations for ${N_{f}}>2$ induce minus signs. Those link states can be resummed and diagonalized, e.g. the following states have an eigenvalue 1: $\displaystyle\pi_{0}^{2}$ $\displaystyle=\left(k_{U}k_{D}+\frac{1}{\sqrt{{N_{c}}}}k^{(2)}_{\pi^{+}\pi^{-},UD}\right),$ $\displaystyle\bar{\pi}_{0}^{2}$ $\displaystyle=\left(k_{\pi^{+}}k_{\pi^{-}}+\frac{1}{\sqrt{{N_{c}}}}k^{(2)}_{UD,\pi^{+}\pi^{-}}\right)$ (21) Likewise all dimer-based states are resummed, as states are only distinguishable on the quark level. This reduces the state space drastically to the physical Hilbert state, e.g. for ${N_{f}}=2$: $B=0:340\mapsto 50$, $B=\pm 1:152\mapsto 20$, $B=\pm 2:4\mapsto 1$, resulting in the local Hilbert space $\mathbbm{H}_{\mathfrak{h}}$ as given in Tab. 1, classified by baryon number $B$, isospin number $I$ and meson occupation number $\mathfrak{m}$. The particle-hole symmetry generalizes to $\mathfrak{m}\mapsto\mathfrak{s}=\mathfrak{m}-\frac{{N_{c}}}{2}({N_{f}}-|B|)$ as the meson raising and lowering operators fulfill an SU(2) algebra for each meson charge $Q_{i}$. All states $\mathfrak{h}\in\mathbbm{H}_{\mathfrak{h}}$ for ${N_{f}}=2,3$ contribute with a weight 1, although some dimer-based link weights are negative. $B$ | $I$ | $\mathfrak{s}=\mathfrak{m}-\frac{3}{2}(2-|B|)$ | $\Sigma$ ---|---|---|--- | | $-3$ | $-\frac{5}{2}$ | $-2$ | $-\frac{3}{2}$ | $-1$ | $-\frac{1}{2}$ | $\,0\,$ | $+\frac{1}{2}$ | $+1$ | $+\frac{3}{2}$ | $+2$ | $+\frac{5}{2}$ | $+3$ | -2 | 0 | | | | | | | 1 | | | | | | | 1 -1 | $-\frac{3}{2}$ | | | | 1 | | 1 | | 1 | | 1 | | | | 4 -1 | $-\frac{1}{2}$ | | | | 1 | | 2 | | 2 | | 1 | | | | 6 -1 | $+\frac{1}{2}$ | | | | 1 | | 2 | | 2 | | 1 | | | | 6 -1 | $+\frac{3}{2}$ | | | | 1 | | 1 | | 1 | | 1 | | | | 4 0 | -3 | | | | | | | 1 | | | | | | | 1 0 | -2 | | | | | 1 | | 2 | | 1 | | | | | 4 0 | -1 | | | 1 | | 2 | | 4 | | 2 | | 1 | | | 10 0 | 0 | 1 | | 2 | | 4 | | 6 | | 4 | | 2 | | 1 | 20 0 | -1 | | | 1 | | 2 | | 4 | | 2 | | 1 | | | 10 0 | -2 | | | | | 1 | | 2 | | 1 | | | | | 4 0 | -3 | | | | | | | 1 | | | | | | | 1 1 | $-\frac{3}{2}$ | | | | 1 | | 1 | | 1 | | 1 | | | | 4 1 | $-\frac{1}{2}$ | | | | 1 | | 2 | | 2 | | 1 | | | | 6 1 | $+\frac{1}{2}$ | | | | 1 | | 2 | | 2 | | 1 | | | | 6 1 | $+\frac{3}{2}$ | | | | 1 | | 1 | | 1 | | 1 | | | | 4 2 | 0 | | | | | | | 1 | | | | | | | 1 $\Sigma$ | | 1 | 0 | 4 | 8 | 10 | 12 | 22 | 12 | 10 | 8 | 4 | 0 | 1 | 92 Table 1: All 92 possible quantum states for the ${N_{f}}=2$ Hamiltonian formulation with ${\rm SU}(3)$ gauge group. The number of states are given for the sectors specified baryon number $B$ and isospin number $I$, and symmetrized meson occupation number $\mathfrak{s}=\mathfrak{m}-\frac{{N_{c}}}{2}({N_{f}}-|B|)$. Note the mesonic particle-hole symmetry $\mathfrak{s}\leftrightarrow-\mathfrak{s}$ which corresponds to the shift symmetry by ${a_{\tau}}$. Only single meson exchange is possible (multiple meson exchange becomes resolved into single mesons in the continuous time limit), the resulting interaction Hamiltonian has ${N_{f}}^{2}$ terms: $\displaystyle\hat{\mathcal{H}}$ $\displaystyle=\frac{1}{2}\sum_{\langle\vec{x},\vec{y}\rangle}\sum_{Q_{i}}\left({\hat{J}_{Q_{i},\vec{x}}}^{+}{\hat{J}_{Q_{i},\vec{y}}}^{-}+{\hat{J}_{Q_{i},\vec{x}}}^{-}{\hat{J}_{Q_{i},\vec{y}}}^{+}\right),$ (22) e.g. for ${N_{f}}=2$: $Q_{i}\in\\{\pi^{+},\pi^{-},\pi_{U},\pi_{D}\\}$ and for ${N_{f}}=3$: $Q_{i}\in\\{\pi^{+},\pi^{-},K^{+},K^{-},K_{0},\bar{K_{0}},\pi_{U},\pi_{D},\pi_{S}\\}$. For the transition $\mathfrak{h}_{1}\mapsto\mathfrak{h}_{2}$, the matrix elements ${\langle\mathfrak{h}_{1}|Q_{i}|\mathfrak{h}_{2}\rangle}$ of $\hat{J}_{Q_{i}}^{\pm}$ are determined from Grassmann integration and a square root per participating link weight (in, hopping, out); only those matrix elements are non-zero which are consistent with current conservation for all $Q_{i}$. Some matrix elements are negative, but in the combination of a closed charged meson loop the total weight remains positive. Some examples for ${N_{f}}=2$ are: $\displaystyle{\langle\pi^{+},\pi^{-}|\pi^{+}|\pi_{U},\pi_{D}\pi^{-}\rangle}$ $\displaystyle=\frac{\sqrt{3}}{2},$ $\displaystyle{\langle\pi_{U},\pi^{-}|\pi^{+}|\pi_{U},2\pi^{-}\rangle}$ $\displaystyle=\frac{\sqrt{6}}{3},$ $\displaystyle{\langle B_{uud}|\pi^{+}|B_{uud},\pi_{U},\pi_{D}\rangle}$ $\displaystyle=\frac{2}{\sqrt{5}},$ $\displaystyle{\langle B_{uud}\pi^{+}|\pi^{+}|B_{uud},\pi_{D}\rangle}$ $\displaystyle=-\frac{\sqrt{3}}{6}$ (23) There are 130 non-zero matrix elements for $B=0$ and 40 matrix elements for $B=\pm 1$. The $B=\pm 2$ states do not allow for pion exchange. Only after contraction of the matrix elements the resummation is carried out for the internal hadronic states, resulting in positive weights. An important application of the ${N_{f}}=2$ partition function is to determine the QCD phase diagram with both finite baryon and isospin chemical potential [13, 14]. Our formulation is still sign-problem free in the continuous time limit. As we have not yet performed dynamical simulations, we can only obtain analytic results in a high temperature expansion of the partition sum $Z$, which is an expansion around the static limit $\displaystyle Z\left(\frac{\mu_{B}}{T},\frac{\mu_{I}}{T}\right)=$ $\displaystyle 2\cosh\frac{3\mu_{I}}{T}+8\cosh\frac{2\mu_{I}}{T}+20\cosh\frac{\mu_{I}}{T}+20$ $\displaystyle+2\cosh\frac{\mu_{B}}{T}\left(8\cosh{\frac{\frac{3}{2}\mu_{I}}{T}}+12\cosh\frac{\frac{1}{2}\mu_{I}}{T}\right)+2\cosh\frac{2\mu_{B}}{T}$ (24) in the number of spatial mesons. This will be presented in a forthcoming publication. ## 3 Leading order gauge corrections to the Hamiltonian formulation for $N_{f}=1$ The QCD Partition can be expanded via the strong coupling expansion in $\beta$: $\displaystyle Z_{QCD}$ $\displaystyle=\int d\psi d\bar{\psi}dUe^{S_{G}+S_{F}}=\int d\psi d\bar{\psi}Z_{F}{\langle e^{S_{G}}\rangle}_{Z_{F}},$ $\displaystyle{\langle e^{S_{G}}\rangle}_{Z_{F}}$ $\displaystyle\simeq 1+{\langle S_{G}\rangle}_{Z_{F}}+\mathcal{O}(\beta^{2})=1+\frac{\beta}{2{N_{c}}}\sum_{P}{\langle{\rm tr}[U_{P}+U_{P}^{\dagger}]\rangle}_{Z_{F}}+\mathcal{O}(\beta^{2})$ (25) Additional color singlet link states are due to plaquette excitations. For discrete time lattices, the dual formulation has been extended beyond $\mathcal{O}(\beta)$ in terms of a tensor network [8]. Here we want to consider the leading order gauge corrections in a Hamiltonian formulation. On anisotropic lattices, the anisotropy $\xi=\frac{a_{s}}{a_{t}}$ is a function of two bare anisotropies $\gamma_{F}$ and $\gamma_{G}=\sqrt{\frac{\beta_{t}}{\beta_{s}}}$, i.e $\xi=\xi(\gamma_{F},\gamma_{G},\beta)$. However, in the continuous time limit $a_{t}\rightarrow 0$ ($\xi\rightarrow\infty$) and for small $\beta$, spatial plaquettes are suppressed over temporal plaquettes by $(\gamma_{G}\gamma_{F})^{-2}$, hence only temporal plaquettes need to be considered. They are of the same order as meson exchange, hence the $\mathcal{O}(\beta)$ weights will contribute to the corresponding operators $\hat{J}^{\pm}$. Note that $\hat{J}^{\pm}$ still has block-diagonal structure, but they will also allows to couple to baryons. Fig. 1 shows the relevant weights as computed in [8]. Figure 1: Tensors and their weights of $\mathcal{O}(\beta)$ that can be incorporated into the Hamiltonian formulation. Blue: dimers, red: baryon flux, green: plaquette flux due to gauge corrections. ## 4 Summary and Outlook We have presented two extensions to the Hamiltonian formulation of lattice QCD: (1) generalization from ${N_{f}}=1$ to ${N_{f}}>1$ in the strong coupling limit, (2) $\mathcal{O}(\beta)$ gauge corrections for ${N_{f}}=1$. In a forthcoming publication, we also address gauge corrections to the ${N_{f}}=2$ Hamiltonian. It turns out that - although not sign problem-free - the sign problem is much milder than in the corresponding formulation at finite ${N_{\tau}}$ as for ${N_{\tau}}\rightarrow\infty$ only temporal plaquettes contribute. Meson exchange between baryons can thus be either due to the gauge corrections or the flavor content. All results presented here are valid in the chiral limit. We have argued in [1] that also a Hamiltonian formulation at finite quark mass $am_{q}$ is well-defined, which extends this approach to staggered lattice QCD further. We are preparing Quantum Monte Carlo simulations that will help to obtain the QCD phase diagram on the lattice in the parameter space $(aT,a\mu_{B},a\mu_{I},am_{q},\beta)$ in lattice units or in units of the baryon mass $am_{B}$. This work was supported by the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation) – project number 315477589 – TRR 211.“ ## References * [1] Marc Klegrewe and Wolfgang Unger. Strong Coupling Lattice QCD in the Continuous Time Limit. Phys. Rev. D, 102(3):034505, 2020. * [2] John B. Kogut and Leonard Susskind. Hamiltonian Formulation of Wilson’s Lattice Gauge Theories. Phys. Rev. D, 11:395–408, 1975. * [3] Pietro Rossi and Ulli Wolff. Lattice QCD With Fermions at Strong Coupling: A Dimer System. Nucl. Phys., B248:105–122, 1984. * [4] N. Kawamoto and J. Smit. Effective Lagrangian and Dynamical Symmetry Breaking in Strongly Coupled Lattice QCD. Nucl. Phys., B192:100, 1981. [,556(1981)]. * [5] F. Karsch and K. H. Mutter. Strong Coupling QCD at Finite Baryon Number Density. Nucl. Phys., B313:541–559, 1989. * [6] Philippe de Forcrand and Michael Fromm. Nuclear Physics from lattice QCD at strong coupling. Phys. Rev. Lett., 104:112005, 2010. * [7] Philippe de Forcrand, Jens Langelage, Owe Philipsen, and Wolfgang Unger. Lattice QCD Phase Diagram In and Away from the Strong Coupling Limit. Phys. Rev. Lett., 113(15):152002, 2014. * [8] Giuseppe Gagliardi and Wolfgang Unger. New dual representation for staggered lattice QCD. Phys. Rev. D, 101(3):034509, 2020. * [9] Philippe de Forcrand, Wolfgang Unger, and Helvio Vairinhos. Strong-Coupling Lattice QCD on Anisotropic Lattices. Phys. Rev., D97(3):034512, 2018. * [10] Neven Bilic, Frithjof Karsch, and Krzysztof Redlich. Flavor dependence of the chiral phase transition in strong coupling QCD. Phys. Rev., D45:3228–3236, 1992. * [11] Owe Philipsen. Strong coupling methods in QCD thermodynamics. Indian J. Phys. 95, 1599–1611 (2021) * [12] K. E. Eriksson, N. Svartholm, and B. S. Skagerstam. On Invariant Group Integrals in Lattice QCD. J. Math. Phys., 22:2276, 1981. * [13] Yusuke Nishida. Phase structures of strong coupling lattice QCD with finite baryon and isospin density. Phys. Rev., D69:094501, 2004. * [14] B. B. Brandt, G. Endrodi, and S. Schmalzbauer. QCD phase diagram for nonzero isospin-asymmetry. Phys. Rev. D, 97(5):054514, 2018.
# ZipLM: Inference-Aware Structured Pruning of Language Models Eldar Kurtic IST Austria <EMAIL_ADDRESS> &Elias Frantar IST Austria <EMAIL_ADDRESS> &Dan Alistarh IST Austria & Neural Magic <EMAIL_ADDRESS> ###### Abstract The breakthrough performance of large language models (LLMs) comes with major computational footprints and high deployment costs. In this paper, we progress towards resolving this problem by proposing a novel structured compression approach for LLMs, called ZipLM. ZipLM achieves state-of-the-art accuracy-vs- speedup, while matching a set of desired target runtime speedups in any given inference environment. Specifically, given a model, a dataset, an inference environment, as well as a set of speedup targets, ZipLM iteratively identifies and removes components with the worst loss-runtime trade-off. Unlike prior methods that specialize in either the _post-training/one-shot_ or the _gradual compression_ setting, and only for specific families of models such as BERT (_encoder_) or GPT (_decoder_), ZipLM produces state-of-the-art compressed models across all these settings. Furthermore, ZipLM achieves superior results for a fraction of the computational cost relative to prior distillation and pruning techniques, making it a cost-effective approach for generating an entire family of smaller, faster, and highly accurate models, guaranteed to meet the desired inference specifications. In particular, ZipLM outperforms all prior $\textrm{BERT}_{\small{\textrm{base}}}\,$distillation and pruning techniques, such as CoFi, MiniLM, and TinyBERT. Moreover, it matches the performance of the heavily optimized MobileBERT model, obtained via extensive architecture search, by simply pruning the baseline $\textrm{BERT}_{\small{\textrm{large}}}\,$model. When compressing GPT2, ZipLM outperforms DistilGPT2 while being 60% smaller and 30% faster. Our code is available at: https://github.com/IST-DASLab/ZipLM. ## 1 Introduction The high accuracy of modern language models from the Transformer family [1] comes at the price of massive computational cost, which hinders their practical adoption in resource-constrained settings. This has motivated the development of _model compression_ techniques, which can be categorized into _pruning_ [2], _quantization_ [3], and _distillation_ [4]. In this paper, we focus on _structural compression_ , whose goal is to reduce model size by removing entire sub-components, such as rows or columns from the model’s weight matrices. The key advantage of structured pruning, relative to unstructured pruning of individual weights, is that the model can be reshaped to new dimensions, and the resulting computational savings can be leveraged on any hardware, without specialized computational support. At the same time, structured pruning introduces significant challenges. First, models are usually highly-sensitive to structured compression, and most methods require _gradual compression_ , including retraining cycles designed to allow the model to recover accuracy. In addition, structural compression significantly complicates the use of knowledge distillation [5], which is usually done via manual or dynamic layer mapping [6, 7]. On the practical side, another challenge is that most existing techniques do not provide _runtime speedup_ guarantees: the model is pruned to a fixed sparsity or FLOPS target, and then must be evaluated in the target inference environment. If the pruned model fails to meet the target inference specifications, the whole process must be repeated from scratch. #### Overview. In this paper, we resolve these issues and provide a novel structured pruning approach called ZipLM, which achieves state-of-the-art performance, both in the _post-training/one-shot_ setting, where retraining is not desirable, as well as in the popular _gradual compression_ setting, where retraining is possible. We accomplish this via an inference-aware algorithm, which successfully balances the loss-runtime trade-off at each pruning step. By taking runtime into account, we avoid removing components that do not bring significant speedup gains. Additionally, our algorithm provides speedup guarantees for compressed models, a highly-desirable property in practical applications. We summarize our contributions as follows: * • We introduce a novel structured pruning approach, which unifies the saliency criteria investigated by prior work–weight magnitude, activation impact, and removal of linearly-redundant structures, while considering local (layer-wise) and global correlations. We augment it to be _inference-aware_ , ensuring desired latency or throughput in any given configuration. * • We complement the algorithm with a novel _layer-wise token-level distillation_ , which consistently boosts accuracy on small datasets and does not require manual layer matching, circumventing a limitation of prior structured pruning techniques. * • ZipLM is the first structured pruning approach that achieves state-of-the-art results for both, _post-training/one-shot_ compression and _gradual pruning_ settings, while being applicable to both, BERT (_encoder_) and GPT (_decoder_) language models, without any modifications. * • ZipLM is practical and efficient. For a set of desired speedups (e.g. 2x, 5x, 10x) in the target inference environment (e.g. batch-size=128, sequence- length=384, device=V100), in a single run and under the same set of hyper- parameters, it produces the entire family of compressed models, one for each speedup target. Consequently, it leads to state-of-the-art results in _GPU- based_ inference environments. Moreover, it is compatible with unstructured pruning and quantization, leading to state-of-the-art results even for _CPU- based_ environments. ## 2 Related Work #### Distillation-based compression methods focus on training a smaller student model to mimic the representations of a larger teacher model. The “distance” between the representations of student and teacher is often architecture-specific. MiniLM [8] uses a deep self- attention mechanism to replicate the attention mechanism of the teacher, and TinyBERT [6] employs a bespoke distillation mechanism, for a manually-picked subset of layers. Both methods offer a very strong baseline, generally outperforming other approaches, except for MobileBERT. MobileBERT [9] involves first training a custom large BERT teacher model from scratch, and then deviates from the standard architecture [10] by introducing heavily-optimized components with reduced latency, whose combinations are decided in neural architecture search (NAS)-like fashion. It achieves strong results in terms of accuracy-per-parameter, at the cost of significant computational costs in the search process. DistilBERT and DistilGPT2 [11] involve training a fixed student obtained by removing every other layer from the teacher, while BERT- PKD [12] employs incremental knowledge extraction through the distillation of intermediate layers. Well-Read-Students [13] reduces the size of the standard BERT architecture through principled downscaling of internal dimensions. DynaBERT [14], on the other hand, distills knowledge to a student model that is both depth- and width-adaptive. #### Structural pruning methods usually start from a large pre-trained model, and iteratively reduce the dimensions of weight matrices. Block Movement Pruning [15] identifies and removes redundant rectangular blocks of weights while following the movement pruning intuition [16] that weights moving towards zero during fine-tuning should be removed. FLOP [17] and Low-Rank [18] use matrix decomposition techniques to progressively remove rank-1 components from factorized weight matrices during training. BERT-of-Theseus [19] employs a similar approach, but replaces entire submodules with smaller counterparts. Methods like LayerDrop [20] and Poor Man’s BERT [21] address structured compression through various layer-dropping techniques. LayerDrop uses structured layer-dropout regularization to train a model resilient to sub-network selection during inference, while Poor Man’s BERT explores a wide range of layer-dropping strategies. The recent CoFi method [7] employs masks of different granularities to jointly prune coarse and fine-grained submodules during fine- tuning, combined with an optional customized distillation technique. CoFi is the state-of-the-art _structural pruning_ method; relative to distillation methods, CoFi outperforms MiniLM and TinyBERT, but not MobileBERT, in terms of accuracy-vs-speedup. #### Other compression methods such as the ones that exploit dynamic forms of sparsity which appear at runtime [22], or the ones that utilize lower bit-width representation of weights and/or activations [23, 24] are complementary to our approach. We demonstrate this in Section 5 where we apply quantization to obtain even higher compression ratios for edge deployment environments like commodity CPUs. ## 3 Method Removing large structures like entire matrix columns or attention heads from a language model quickly leads to severe accuracy degradation, from which it is often difficult to recover even with extensive finetuning. This is why current state-of-the-art approaches like Block Movement Pruning [15] or CoFi [7] opt for integrating pruning directly into training (via sampling or differentiable approximations), rather than performing it in the standard gradual pruning fashion of discrete steps with finetuning in between. However, as we will show, by designing a new highly accurate pruning algorithm which is able to account for both local correlations of structures within single layers as well as global correlations across layers, we can actually apply the gradual pruning paradigm, with all its advantages, to improve significantly over the current state-of-the-art. ### 3.1 The ZipLM Structured Pruning Algorithm (Local Correlations) Most existing structured pruning criteria [25, 26] are based on one or two of the following assumptions about saliency: structures with lower (average) weight magnitude are easier to prune [27, 28], structures with small input activations can be removed at little loss [29], and structures that are close to a linear combination of other structures are the most redundant [30, 31]. We will now show how all these aspects can be jointly considered in a principled manner via our new ZipLM technique. #### Problem formulation. Our approach starts from the idea of applying structured compression layer- wise, in a way that allows the layer to preserve most of its output characteristics. This setup is popular in the post-training quantization and unstructured pruning literature [32, 33, 34], and can be implemented as follows. We are given a small amount of calibration data, which we run through the network, to obtain “reference” inputs and outputs for each layer. Then, for each layer, given the calibration inputs $\mathbf{X}$ and the original layer weights $\mathbf{{W}}$, we aim to find compressed weights $\mathbf{\widehat{W}}$ respecting the compression constraint $\mathcal{C}$, which best approximate the original output, measured via the squared error metric. If we assume that the input and weight matrices have an appropriate rectangular form, the problem can be formalized as: $\text{argmin}_{\mathbf{\widehat{W}}}\,||\mathbf{\widehat{W}}\mathbf{X}-\mathbf{W}\mathbf{X}||_{2}^{2}\quad\text{subject to}\quad\mathbf{\widehat{W}}\in\mathcal{C}.$ (1) This objective can be decomposed across the rows of $\mathbf{W}$, leading to a set of sparse linear regression problems, one per row. These row-wise problems are independent, which forms the basis of related work [34]; yet, since we do structured pruning, they become dependent, as we would like to prune the same weight indices _across all rows_ , i.e. prune entire columns. Thus, finding the optimal weights $\mathbf{\widehat{W}}\in\mathcal{C}$ is equivalent to finding: 1) the optimal structure $\mathbf{S}$ of the desired shape to be removed, which we assume to be applied across all rows, with corresponding pruning mask $\mathbf{M_{S}}$, where pruned indices have value $1$ in the mask, and others are $0$; and 2) the corresponding update $\boldsymbol{\delta_{S}}$ to all of the remaining weights, optimally compensating for the error caused by the removal of weights in $\mathbf{S}$. #### Saliency scores and weight update. Let $\mathbf{H}=\mathbf{X}\mathbf{X}^{\top}$ be the Hessian matrix for the $\ell_{2}$-minimization problem in Equation 1, which is independent of the weights. Define $\mathbf{W}_{i,\mathbf{M_{S}}}$ to be the subset of weights under the mask $\mathbf{M_{S}}$ in row $i$, and by $(\mathbf{H}^{-1})_{\mathbf{M_{S}},\mathbf{M_{S}}}$ the submatrix of the inverse Hessian corresponding to the entries under the mask $\mathbf{M_{S}}$. Then, we can obtain the optimal mask and weight update as follows: $\text{argmin}_{\mathbf{S}}\,\sum_{i=0}^{d_{\text{row}}}\mathbf{W}_{i,\mathbf{M_{S}}}\cdot\left(\left(\mathbf{H}^{-1}\right)_{\mathbf{M_{S}},\mathbf{M_{S}}}\right)^{-1}\cdot\mathbf{W}^{\top}_{i,\mathbf{M_{S}}}$ (2) $\boldsymbol{\delta_{S}}=-\mathbf{W}_{:,\mathbf{M_{S}}}\cdot\left(\left(\mathbf{H}^{-1}\right)_{\mathbf{M_{S}},\mathbf{M_{S}}}\right)^{-1}\cdot\left(\mathbf{H}^{-1}\right)_{\mathbf{M_{S}},:}$ (3) We obtain this by extending the Optimal Brain Surgeon [35, 36] formulas for solving Equation 1 to cover all $d_{\text{row}}$ weight matrix rows simultaneously. Importantly, the subselection of the inverse Hessian $((\mathbf{H}^{-1})_{\mathbf{M_{S}},\mathbf{M_{S}}})^{-1}$ is shared between all rows. Further, since we generally consider only non-overlapping sets $\mathbf{S}$ of the same size, we pay just $O(d_{\text{col}}\cdot|\mathbf{M_{S}}|^{2})$ total cost for all extra inversions. Since the number of structures in the mask $|\mathbf{M_{S}}|$ is usually small, e.g. attention heads usually consist of 64 columns, the overall cost of these inversions is low. Simply selecting the structures to prune according to the criterion in Equation 2 unifies the weight magnitude and activation influence criteria (via the Hessian), but still ignores any correlations between structures. We address this by pruning structures _one-at-a-time_ , while always applying update $\boldsymbol{\delta_{S}}$ and fully recomputing $\mathbf{H}^{-1}$ relative to the remaining structures. For example, if there exist two redundant structures $S_{1}$ and $S_{2}$, we will first drop $S_{1}$ and update $S_{2}$ to compensate for this removal, at which point $S_{2}$ is no longer easy to prune. Without this one-at-a-time removal, both structures would have been incorrectly removed as they each individually seem easy to prune according to Equation 2. Executing this strategy naively will require a full $O(d_{\text{col}}^{3})$ recomputation of the inverse Hessian relative to the remaining structures at each step, which would be very slow. However, this can be avoided by removing the rows and columns corresponding to $\mathbf{M_{S}}$ directly in the inverse with one step of Gaussian elimination [34], applied block-wise to cover larger structures, as follows: $\mathbf{H}^{-1}-\mathbf{H}^{-1}_{:,\mathbf{M_{S}}}\cdot\left(\left(\mathbf{H}^{-1}\right)_{\mathbf{M_{S}},\mathbf{M_{S}}}\right)^{-1}\cdot\mathbf{H}^{-1}_{\mathbf{M_{S}},:},$ (4) which takes only $O(|\mathbf{M_{S}}|\cdot d_{\text{col}}^{2})$ time. We provide complete pseudocode in Algorithm 1. Algorithm 1 The ZipLM pruning algorithm. Given inverse Hessian $\mathbf{H}^{-1}=(2\mathbf{X}\mathbf{X}^{\top}+\lambda\mathbf{I})^{-1}$, we remove exactly $k$ structures from the corresponding weight matrix $\mathbf{W}$. $\mathbf{R}\leftarrow$ set of all possible structures for $k$ times do $\mathbf{S}\leftarrow\text{argmin}_{\mathbf{S}}\,\sum_{i=0}^{d_{\text{row}}}\mathbf{W}_{i,\mathbf{M_{S}}}\cdot((\mathbf{H}^{-1})_{\mathbf{M_{S}},\mathbf{M_{S}}})^{-1}\cdot\mathbf{W}^{\top}_{i,\mathbf{M_{S}}}$ $\boldsymbol{\delta_{S}}\leftarrow-\mathbf{W}_{:,\mathbf{M_{S}}}\cdot((\mathbf{H}^{-1})_{\mathbf{M_{S}},\mathbf{M_{S}}})^{-1}\cdot(\mathbf{H}^{-1})_{\mathbf{M_{S}},:}$ $\mathbf{W}\leftarrow\mathbf{W}+\boldsymbol{\delta_{S}}$ $\mathbf{H}^{-1}\leftarrow\mathbf{H}^{-1}-\mathbf{H}^{-1}_{:,\mathbf{M_{S}}}\cdot((\mathbf{H}^{-1})_{\mathbf{M_{S}},\mathbf{M_{S}}})^{-1}\cdot\mathbf{H}^{-1}_{\mathbf{M_{S}},:}$ $\mathbf{R}\leftarrow\mathbf{R}-\\{\mathbf{S}\\}$ end for $\mathbf{W}\leftarrow\mathbf{W}\odot\mathbf{M_{R}}$ We utilize the fact that the values corresponding to pruned weights in $\mathbf{W}$ and in the inverse Hessian $\mathbf{H}^{-1}$ do not affect any subsequent calculations and can therefore be ignored even if they are not exactly zero. However, in the end we have to prune them explicitly again by multiplying with the overall mask to ensure that they are exactly zero. In a practical implementation, $((\mathbf{H}^{-1})_{\mathbf{M_{S}},\mathbf{M_{S}}})^{-1}$ should only be computed once and reused when computing the corresponding sum across all rows. #### Pruned structures. Focusing on Transformers, we consider three types of structural removal: dropping attention heads, shrinking the expanded intermediate dimension of the fully-connected network (FC) layers, and removing entire residual parts, i.e. attention or FC-modules. We implement this by dropping $d_{\text{head}}$ consecutive columns in the out-matrix of the attention block and individual columns in the second linear layer of the feed-forward network. Once these column-structures are zeroed out, corresponding rows in previous layers can be safely removed without any output change. Crucially, by pruning e.g. columns in the FC2 layer rather than equivalent rows in FC1, we can utilize the input correlations via Hessian-information using the ZipLM pruner. #### Novelty relative to existing Optimal Brain Surgeon (OBS) approaches. The original framework [35], as well as modern efficient versions [37, 38, 34], have been explicitly developed for _unstructured pruning_ , i.e. removing individual weights. It is nontrivial to extend them to structured pruning, as this involves considering additional correlations, both within as well as across multiple blocks (such blocks are usually employed for computational tractability). For example, the state-of-the-art layer-wise approach of [34], performs unstructured pruning by handling weight matrix rows separately, and then greedily merging results. In contrast, we perform structured pruning jointly across multiple rows, which is not only necessary for correctness but additionally enables us to design an algorithm with a computational complexity that is lower by a full factor of the hidden dimension size. Additionally, structured pruning requires explicitly matching matrix shapes for consecutive layers and a dedicated strategy for utilizing weight updates even when entire blocks/rows are pruned. ### 3.2 Inference-Aware Structured Pruning (Global Correlations) We now describe how to augment the algorithm to be _inference-aware_ , in the sense that it accepts inference specifications, such as batch-size, sequence- length, and speedup on the target hardware, as additional inputs to optimize for. #### Motivation. The main benefit of inference-aware structured pruning is the fact that pruning decisions are not guided purely by saliency scores, but instead by loss-vs-speedup trade-offs associated with the removal of each component in the model. Prior methods, e.g. [36, 15, 7] focus solely on pruning until a specific sparsity threshold is reached, without taking into account the real- world speedups corresponding to the compression threshold, which can vary significantly between settings. For example, a 95% sparse BERT produced by CoFi [7] has 12x speedup on a V100 GPU, but only 5x on an A100 GPU. With existing methods, if real-world timings fail to meet the inference requirements, the entire process has to be repeated with different sparsity values until the target speedup is achieved, which is both time-consuming and error-prone. An additional advantage of inference-awareness, which we showcase in our GPT experiments in Section 4, is that it enables optimizing for different real-world metrics, such as latency or throughput. #### Runtime awareness. We integrate runtime constraints via a latency table [39] for our target inference environment, where we record the time to run an attention block, including all overheads, with $0,\dots,N_{\text{heads}}-1$ heads pruned and similarly for the fully-connected block with the intermediate dimension shrunk by a factor of $0.9^{i}$, for $i=0,\dots,42$; in relative steps of $10\%$ up until $\approx 99\%$ sparsity, following [40]. This allows rapid runtime estimation for different per-layer sparsity configurations. We provide an example of our latency table in Appendix E. #### Finding the optimal sparsity configuration. Ultimately, our goal is to find a per-layer-sparsity configuration that satisfies a certain speedup-constraint while maximizing accuracy. A popular paradigm of doing this [27, 41] is to produce a large number of pruned models with different sparsity distributions across layers and then select the one, satisfying a target constraint, with the highest accuracy. To make this computationally feasible, it is crucial that pruning is cheap, yet accurate. ZipLM treats each layer independently, which makes it possible to precompute a database of several pruned versions with different sparsities for each layer. The entire database can be produced in a single run, utilizing the algorithm’s one-at-a-time nature. While our algorithm is compatible with various search methods for finding layer-wise profiles [27, 42], we adapt the recent SPDY approach [40]. #### Structured SPDY search. The SPDY approach is designed for unstructured pruning and assigns a quadratic prior to per-layer sensitivity of different sparsity levels. This is not valid in our structured pruning scenario, since for instance it would suggest that dropping a full layer is only slightly more difficult than pruning it to 99% sparsity. Thus, using standard SPDY would lead the algorithm to explore a large number of sub-optimal configurations, significantly wasting computational resources. To alleviate this problem, for a structured sparsity $s$, we introduce a better prior $p_{s}$ as the relative layer-wise squared error incurred by pruning, defined as $p_{s}=||\mathbf{\widehat{W}_{s}}\mathbf{X}-\mathbf{W}\mathbf{X}||_{2}/||\mathbf{W}\mathbf{X}||_{2}$, which simply has a value of 1 for a fully dropped layer. Furthermore, the original SPDY approach uses shrinking neighborhood search, which has high variance in both runtime and solution quality for structured compression. Therefore, we perform a fixed number of $1000$ steps, randomly mutating in expectation 10% of the layer-wise sensitivity coefficients. Finally, we note that any candidate evaluated by this procedure actually achieves the target speedup, leading to significantly decreased search time. We validate our approach in Appendix F, where we demonstrate that our speedup estimations are indeed very accurate in practice. Specifically, real-world on-device measurements deviate at most by 5.28% from their expected values. ### 3.3 Layer-wise Token Distillation For structured pruning, it is common to apply _layer-wise distillation_ objectives to transfer intermediate representations. However, structured pruning creates compatibility issues relative to the fixed teacher architecture, leading most methods to develop customized distillation strategies. A popular approach, introduced in [6] and improved by [7], solves the problem via static [6] or dynamic [7] mapping of a subset of teacher layers to a subset of student layers. Their main limitation is manual layer selection, where making the “optimal” choice would require evaluating all possible combinations, which can be very expensive. Another limitation is shape-matching between intermediate layers, which is solved by introducing a learnable linear transformation matrix attached to student outputs. #### Our approach. We address these challenges differently, by leveraging the fact that ZipLM preserves the hidden dimension size, and propose to use distillation of intermediate token representations across the entire model. The resulting minimization objective consists of three components: $\displaystyle\mathcal{L}(\theta^{\mathrm{s}},\theta^{\mathrm{t}}|x)=\lambda_{\mathrm{1}}\mathcal{L}_{\mathrm{task}}(\theta^{\mathrm{s}}|x)\,+\,\lambda_{\mathrm{2}}\mathcal{L}_{\mathrm{logit}}(\theta^{\mathrm{s}},\theta^{\mathrm{t}}|x)+\lambda_{3}\mathcal{L}_{\mathrm{token}}(\theta^{\mathrm{s}},\theta^{\mathrm{t}}|x),$ (5) where $\theta^{\mathrm{s}}$ and $\theta^{\mathrm{t}}$ represent student and teacher models respectively, $x$ are the inputs, $\mathcal{L}_{\mathrm{task}}$ is the loss associated with the task (e.g. cross-entropy for text- classification), $\mathcal{L}_{\mathrm{logit}}$ is the KL-divergence between output logits as described in [5], and $\mathcal{L}_{\mathrm{token}}$ is our token-level distillation loss. Hidden tensors passed between consecutive transformer layers are of constant shape $\mathbf{H}\in\mathbb{R}^{B\times seq\times H}$, where $B$ stands for the batch-size, $seq$ for the sequence length, and $H$ for the hidden size defined by the model architecture. This tensor can be interpreted as a collection of $B\times seq$ vectors $\mathbf{h}\in\mathbb{R}^{H}$, each carrying intermediate model representations of input tokens $x$. We define the loss $\mathcal{L}_{\mathrm{token}}$ as an Euclidean distance $\Delta$ between vectors $\mathbf{h}$ corresponding to each non-padded token in the input sequence, averaged over all unpruned layers. Formally, for a layer $k$, it is defined as $\mathcal{L}_{\mathrm{token}}^{k}=\frac{1}{\sum_{j=1}^{B\times seq}\mathbbm{1}[j\notin\mathbf{P}]}\sum_{j=1}^{B\times seq}\mathbbm{1}[j\notin\mathbf{P}]\cdot\Delta(\mathbf{h}^{\theta_{\mathrm{s}}},\mathbf{h}^{\theta_{\mathrm{t}}}),$ (6) where $\mathbf{P}$ stands for the set of padding tokens. This formulation encourages the student model to generate vector representations for each token that are similar to those produced by the teacher model. In Appendix B, we present ablation studies and comparisons for ZipLM and CoFi, with and without their respective distillation objectives. ## 4 Experiments #### Setup. Given a pre-trained model, a dataset, and a set of desired speedups in a target inference environment, we iteratively fine-tune and prune the model in a structured way such that in the end we obtain a set of accurate compressed models, one for each speedup target. We consider pruning of the standard $\textrm{BERT}_{\small{\textrm{base}}}\,$and $\textrm{BERT}_{\small{\textrm{large}}}\,$architectures, evaluating on dev- sets of established benchmarks: SQuADv1.1 [43], and a subset of GLUE [44] tasks: SST-2 [45], QNLI [44], MNLI [46], and QQP [47], selected to match publicly-available checkpoints from prior work. For a precise comparison to prior work [7], our inference environment is a single NVIDIA V100 16GB GPU, batch size of 128, and sequence lengths of 384 and 128 for SQuAD and GLUE tasks, respectively. In addition to encoder-based BERT models, we also consider pruning of the decoder-based GPT2 model on the OpenWebTextCorpus [48], for which we consider two inference environments: pruning for throughput (batch-size=16, sequence-length=1024), and pruning for latency (batch-size=1, a set of prompts with varying lengths). For illustration, our pipeline is depicted in Figure 1. In Appendix H and I, we report exact values for all results, as well as hyper-parameters for reproducibility. Figure 1: Illustration of the ZipLM pipeline: 1) inference specifications, 2) runtime benchmarking of candidates for pruning, 3) gradual structured pruning until all speedup targets are met. #### Baselines. In the _gradual pruning_ setting, we explore the performance of ZipLM pruning of BERT- and GPT2-family models, across a wide range of inference speedup targets, ranging from 2x to 15x, in unit increments. This allows us to compare the effectiveness of our approach against a diverse set of structured pruning and distillation-based techniques, including state-of-the-art CoFi pruning, competitive Block Movement Pruning, and distillation approaches including TinyBERT, DistilBERT, DistilGPT2, MobileBERT, MiniLM, and DynaBERT. Additionally, we include comparisons with other relevant methods. For fairness, we follow [7] and report TinyBERT and DynaBERT results without data augmentations. In the _post-training/one-shot_ setting, which does not allow retraining, we demonstrate that ZipLM outperforms the prior state-of-the-art approach of [49]. We evaluate inference speedups of all models in the same environment, unless the models are not publicly available, in which case we report speedups from their respective papers. We refer to ZipLM compressed BERT models as ZipBERT, and to ZipLM compressed GPT2 models as ZipGPT2. ### 4.1 Gradual Structured Pruning #### $\textrm{BERT}_{\small{\textrm{base}}}\,$results. In Figure 2 we compare structured compression methods on the SQuADv1.1 task. ZipLM outperforms both CoFi and TinyBERT, prior state-of-the-art techniques, by 3 points in the F1 score at the same speedup factor, while at the same F1 score it is able to improve inference speedups by at least 60%. In Figure 3, we extend this comparison to a subset of GLUE tasks and provide an exhaustive overview of various structured compression techniques. Results on the other four remaining GLUE tasks are provided in Appendix Figure 7. As can be observed, distillation-based methods usually provide either one or a few structurally-compressed models, due to the massive costs associated with training from scratch for each new model. Relative to the most competitive approaches, such as TinyBERT, CoFi, and MiniLM, ZipLM provides consistent improvements in terms of both, accuracy and speedup, while providing guarantees for each compressed model in terms of the expected speedup in the target inference environment. Interestingly, on tasks like QQP and SST-2, ZipLM is able to compress the $\textrm{BERT}_{\small{\textrm{base}}}\,$model up to 6x and 10x speedups, respectively, while maintaining the accuracy of the _dense_ model. In Appendix D, we provide additional comparisons against CoFi on test-set results from the official GLUE evaluation server. Figure 2: Structured compression of $\textrm{BERT}_{\small{\textrm{base}}}\,$(left) and $\textrm{BERT}_{\small{\textrm{large}}}\,$(right) on the SQuADv1.1 task. Dashed horizontal lines represent full and 99% accuracy recovery of the uncompressed model. Figure 3: Structured compression of $\textrm{BERT}_{\small{\textrm{base}}}\,$on QNLI, MNLI, SST-2, and QQP tasks. Dashed horizontal lines represent full and 99% accuracy recovery of the uncompressed model. #### $\textrm{BERT}_{\small{\textrm{large}}}\,$results. To verify that our approach does not pertain only to the $\textrm{BERT}_{\small{\textrm{base}}}\,$model, we apply ZipLM structured pruning to the 3x larger $\textrm{BERT}_{\small{\textrm{large}}}\,$model on the SQuADv1 task. In this setup, we compare against the only two approaches that attempted to structurally compress this larger model, Block Movement Pruning and distillation-based MobileBERT. As can be seen in Figure 2, ZipLM is able to compress $\textrm{BERT}_{\small{\textrm{large}}}\,$up to 4x faster inference while maintaining the F1 score of the uncompressed model. At the same F1 score as the fastest Block Movement Pruning model (3x), ZipLM doubles the inference speedup (6x). A result worth emphasizing is that ZipLM is even able to match the performance of the highly optimized MobileBERT model by simply compressing the baseline BERT architecture, without the many additional optimizations and custom components used by MobileBERT. Specifically, some of the module- and operator-level optimizations used by MobileBERT include: bottleneck structures and carefully-balanced self-attention and feed-forward modules, embedding layer factorization, a bespoke closed-source teacher model, replacement of LayerNorm layers with lower-latency NoNorm layers, and replacement of GELU activation functions with ReLU activations. #### 99% recovery. The MLPerf Benchmark [50] targets recovery of $>$99% of the baseline accuracy. At this industry-defined threshold, ZipLM models set new state-of-the-art performance across all of the considered datasets with the following $\textrm{BERT}_{\small{\textrm{base}}}\,$inference speedups: 5x on the SQuADv1 task, 6x on QNLI and MNLI, and, surprisingly, 13x and 15x on SST-2 and QQP, respectively. When compressing $\textrm{BERT}_{\small{\textrm{large}}}\,$on the SQuADv1 task, ZipLM produces a 6x faster model at 99% recovery. #### GPT2 results. To validate that our approach does not only apply to encoder-based models, we apply ZipLM structured pruning to the decoder-based GPT2 model. In addition to this, to further demonstrate the inference-awareness property of our approach and its importance for real-world applications, we consider two different regimes: pruning for throughput and pruning for latency. An example application for the former regime is a server-side deployment where the model processes many queries at the same time, while an application for the latter regime is a text-generation scenario where the model is used in an online fashion to auto-complete user’s text. For a fair comparison, we follow the DistilGPT2 setup [11] and prune the 124M parameters GPT2 variant on the OpenWebTextCorpus dataset, followed by zero- shot evaluations, without any fine-tuning, on the test-split of the WikiText [51] dataset. Because of the enormous vocabulary size, the maximum achievable speedup in the throughput regime for this model is roughly 3.5x. Thus, we run ZipLM pruning to 1.5x, 2x, 2.5x, and 3x speedup targets. For the latency regime, we report the median time to process sequences of various lengths when generating text with Top-K sampling [52]. In Table 1, we present zero-shot evaluations of the uncompressed GPT2 model which serves as a baseline relative to the competing DistilGPT2 approach, and four variants of our ZipLM pruned GPT2. In the pruning for throughput scenario, at similar speedup and decoder size (1.6x-vs-1.5x and 42.5M-vs-47.3M), ZipGPT2 achieves significantly lower perplexities relative to DistilGPT2. Further, at slightly better (lower) perplexities, ZipGPT2 reduces the decoder size from 42.5M to only 26.5M parameters (60% reduction) and improves speedup from 1.6x to 2.1x (30% faster). In the pruning for latency scenario, at a similar speedup of 1.9x-vs-2.0x, ZipGPT2 reduces the decoder size by 3M params while providing almost 2 points improvement in the zero-shot perplexity. Table 1: Zero-Shot perplexity (PPL) of compressed GPT2 in two regimes: pruning for throughput and pruning for latency. ∗GPT2 was trained by OpenAI [53] on a much larger closed-source dataset and for significantly longer. The only direct comparison is between DistilGPT2 and ZipGPT2. Model | Pruning for throughput | Pruning for latency ---|---|--- Speedup | Decoder size | Wiki Text-103 PPL $\downarrow$ | Speedup | Decoder size | Wiki Text-103 PPL $\downarrow$ GPT2∗ | 1.0x | 85.0M | 28.5 | 1.0x | 85.0M | 28.5 DistilGPT2 | 1.6x | 42.5M | 43.0 | 1.9x | 42.5M | 43.0 ZipGPT2 (ours) | 1.5x | 47.3M | 35.4 | 1.6x | 48.7M | 37.8 2.1x | 26.5M | 41.5 | 2.0x | 39.2M | 41.2 2.7x | 14.0M | 50.4 | 2.2x | 26.6M | 49.0 3.3x | 15.7M | 72.1 | 2.5x | 20.7M | 55.0 Table 2: One-shot (post-training) structured pruning of $\textrm{BERT}_{\small{\textrm{base}}}\,$on three downstream datasets and two speedup targets. Speedup | Kwon et al. [49] | ZipBERT base ---|---|--- SQuAD, F1 | | 1.5x 2.0x | 86.2 76.5 | 87.1 84.1 QQP, acc. | | 1.5x 2.0x | 89.5 83.9 | 89.7 84.8 MNLI, acc. | | 1.5x 2.0x | 82.8 78.1 | 83.0 78.2 ### 4.2 On the Importance of Inference-Awareness #### Depth vs. width pruning. A particularly interesting illustration of the importance of inference- awareness in the pruning algorithm is given by our GPT2 models running directly in the PyTorch-HuggingFace framework, which can be used in two different modes: batch-prediction (throughput-constrained) and text-generation (latency-constrained). For the former, inputs are typically large, and shrinking weight matrices is an effective way to achieve speedups. However, for the latter, the inputs are much smaller, and the size of weight matrices is no longer the primary bottleneck. In this scenario, the only way to achieve substantial speedups is to completely drop some modules, which prior methods cannot account for as they solely optimize for overall model sparsity. However, with ZipLM, runtime measurements from the target inference environment guide pruning decisions, allowing it to learn the best way to compress the model for an optimal speedup-accuracy trade-off. Our GPT2 compression results in Table 1 clearly illustrate and support these statements. Even though pruned for the same speedup target, the final architectures of ZipGPT2 models are drastically different. For the throughput- constrained scenario, the model’s depth was preserved but the matrix dimensions were significantly reduced (roughly by a factor of 10) making the corresponding multiplications with large input tensors much faster. In contrast, for the latency-constrained scenario, the model’s width (shapes of weight matrices) was mostly preserved but the depth was shrunk almost by a factor of 4, making the forward pass with small inputs faster by reducing the effective number of modules. #### Inference device capabilities. Incorporating capabilities of the inference device is another important aspect for effective structured pruning which prior methods do not account for as they solely optimize for higher sparsities. As noted in Section 3.2, this reflects in larges discrepancies between speedups obtained on different devices, e.g. a compressed model with 12x speedup on a V100 is only 5x faster on an A100 GPU. This arises because the A100 GPU is significantly more powerful and thus faster on the dense model; at the same time, it is highly underutilized for small matrices, which significantly limits the speedups for very high sparsity. To illustrate this, we have measured the speedup from reducing the MLP size for both GPU types (see Table 3). As can be seen, pruning to $\approx$90% sparsity (3072 $\rightarrow$ 302) gives $\approx$7x speedup on a V100 but only $\approx$3x speedup on an A100. Such differences are automatically captured by ZipLM, where pruning for sparsity is replaced by pruning for speedup. #### Pruning for speedup vs. pruning for sparsity. In Figure 4 we compare results with ZipLM pruning when the target for pruning is sparsity (like prior approaches) and when the target for pruning is speedup (the ZipLM approach). Pruning for speedup brings significant improvements, up to 10 points, especially at higher speedups where inference-awareness is very important as the algorithm does not remove components that do not bring any further speed and therefore helps preserving accuracy. Figure 4: Ablation study for the impact of the pruning target: pruning for sparsity (like prior approaches) versus pruning for speedup (the ZipLM approach). Table 3: Speedups from shrinking the intermediate size of MLPs in the FFN section of a Transformer layer, on different GPUs. | Speedup ---|--- MLP size | V100 | A100 3072 | 11.0x | 1.0x 1814 | 11.6x | 1.1x 1322 | 12.0x | 1.4x 1302 | 16.9x | 3.1x 1130 | 11.8x | 4.4x 1176 | 13.1x | 4.4x 1133 | 14.8x | 4.4x ### 4.3 Post-training/One-shot Structured Pruning We now study the performance of ZipLM when applied purely in one-shot, without any retraining. In this setting, we compare against the state-of-the-art method of Kwon et al. [49] which combines several heuristics: Fisher-based mask search, mask rearrangement, and mask tuning. Instead of heuristics, our pruning framework utilizes direct end-to-end loss information to find the optimal sparsity configuration. During the warm-start phase, [49] utilizes a diagonal Fisher matrix to estimate the significance of heads and filters, which discards correlations caused by off-diagonal elements. Although the approach attempts to address this limitation by approximating correlations within a single layer, it will not capture global dependencies. Furthermore, the weights are adapted for layer-wise reconstruction at the very end of the compression step, whereas our method does it continuously during the pruning (please see Section 4 for the significance of doing this). For a fair comparison, we apply the authors’ own implementation in latency-constrained mode on the exact same model weights. Table 2 presents results on several datasets and speedups, showing that ZipLM is even more accurate than the approach designed and optimized specifically for the post-training/one-shot pruning. #### Sensitivity to calibration data. Additionally, we have found that ZipLM is very robust to the amount of calibration data. In Table 4 we present a sensitivity analysis with respect to the number of calibration samples. We one-shot prune $\textrm{BERT}_{\small{\textrm{base}}}\,$on the SQuADv1.1 task for two speedup targets: 1.5x and 2.0x. In this setup, we compare results against Kwon et al. [49], which uses 2048 samples by default. As can be seen from the table, ZipLM outperforms prior state-of-the-art starting at only 32 samples. As we increase the number of samples, the results improve, up to 2 points in F1 score. ## 5 Discussion and Extensions #### CPU as an LLM-inference environment. In Section 4 we have focused on various GPU-based inference environments as it enabled us to conduct fair comparisons against prior structural compression techniques. However, CPUs present another compelling inference environment focused on edge deployment of LLMs. Therefore, we target the recently proposed compound compression pipeline of [36], which involves three steps: structured pruning, unstructured pruning, and quantization. We replace their structured pruning approach based on layer dropping with ZipLM. As a result, at full accuracy recovery, we are able to improve speedup from 3x to 13x, and at the largest compression ratio from 30x to 50x. Due to space constraints, we provide full results in Appendix A. #### Computational efficiency. Relative to distillation-based methods, structured pruning is an order of magnitude more efficient in terms of GPU hours due to the massive costs associated with pretraining from scratch for each compressed model [7, 9, 6]. For efficiency comparisons to CoFi, we consider the task of producing a full family of compressed $\textrm{BERT}_{\small{\textrm{base}}}\,$models with speedup targets ranging from 2x to 15x. In this setup, ZipLM requires only 115 epochs in total, whereas CoFi would require 560 epochs. Therefore, ZipLM is _4.87 times more efficient_ than CoFi. In terms of end-to-end runtime, ZipLM produces the entire family of compressed $\textrm{BERT}_{\small{\textrm{base}}}\,$models on a single RTX A6000 GPU in $\sim$35 hours on larger datasets (e.g. MNLI) and only $\sim$10 hours on smaller ones (e.g. SST2). Finally, it is worth emphasizing that we have not taken into account the cost of hyper-parameter tuning in the above comparisons, but that this is very favorable to ZipLM: it uses a single set of hyper-parameters to produce an entire family of compressed models while other methods require hyper-parameter tuning for each model independently. Figure 5: Scaling laws of structured pruning vs. distillation on the standard BERT architecture. Table 4: Sensitivity to the number of calibration samples. | | F1 score at ---|---|--- Method | Num samples | 1.5x | 2.0x ZipLM | 4 | 82.3 | 48.4 32 | 86.8 | 82.6 128 | 86.8 | 83.6 512 | 86.8 | 84.1 2048 | 87.1 | 84.1 4096 | 87.6 | 84.7 Kwon et al. | 2048 | 86.2 | 76.5 #### Scaling laws for structured pruning. To further understand the accuracy-speedup trade-offs, we run ZipLM on larger speedup ratios, up to 55x for $\textrm{BERT}_{\small{\textrm{large}}}\,$ and 75x for $\textrm{BERT}_{\small{\textrm{base}}}\,$. To the best of our knowledge, this is the first result in literature demonstrating that such extreme compression ratios are achievable with structured pruning without model collapse. In Figure 5, we compare these results against distillation- based downscaling of the BERT architecture [13]. The results clearly demonstrate that each of the pruned models, based either on $\textrm{BERT}_{\small{\textrm{large}}}\,$or $\textrm{BERT}_{\small{\textrm{base}}}\,$, significantly outperforms comparable pre-trained variants. An emergent behavior that can be observed is that structurally pruned models tend to follow a linear scaling law, meaning that the accuracy decreases linearly with the increase of the speedup ratio, at a slope given by the original model. Fitting linearly via least squares produces the following expressions for the accuracy-speedup relationship: $\texttt{F1}_{\texttt{large}}\approx 92.1-0.3\times\texttt{speedup}_{\texttt{large}}$, and $\texttt{F1}_{\texttt{base}}\approx 90.3-0.6\times\texttt{speedup}_{\texttt{base}}$. Thus, the rate of decrease in accuracy for $\textrm{BERT}_{\small{\textrm{base}}}\,$is twice as large as that of $\textrm{BERT}_{\small{\textrm{large}}}\,$, which can be attributed to the presence of more redundant representations in the larger model, making it more resilient to pruning. In Appendix G we provide additional analysis of the structure of pruned models. ## References * [1] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. Attention is all you need. Advances in neural information processing systems, 30, 2017. * [2] Torsten Hoefler, Dan Alistarh, Tal Ben-Nun, Nikoli Dryden, and Alexandra Peste. Sparsity in deep learning: Pruning and growth for efficient inference and training in neural networks. Journal of Machine Learning Research, 22(241):1–124, 2021. * [3] Amir Gholami, Sehoon Kim, Zhen Dong, Zhewei Yao, Michael W Mahoney, and Kurt Keutzer. A survey of quantization methods for efficient neural network inference. arXiv preprint arXiv:2103.13630, 2021. * [4] Jianping Gou, Baosheng Yu, Stephen J Maybank, and Dacheng Tao. Knowledge distillation: A survey. International Journal of Computer Vision, 129(6):1789–1819, 2021\. * [5] Geoffrey Hinton, Oriol Vinyals, Jeff Dean, et al. Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531, 2(7), 2015. * [6] Xiaoqi Jiao, Yichun Yin, Lifeng Shang, Xin Jiang, Xiao Chen, Linlin Li, Fang Wang, and Qun Liu. Tinybert: Distilling bert for natural language understanding. In Findings of the Association for Computational Linguistics: EMNLP 2020, pages 4163–4174, 2020. * [7] Mengzhou Xia, Zexuan Zhong, and Danqi Chen. Structured pruning learns compact and accurate models. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 1513–1528, 2022. * [8] Wenhui Wang, Furu Wei, Li Dong, Hangbo Bao, Nan Yang, and Ming Zhou. Minilm: Deep self-attention distillation for task-agnostic compression of pre-trained transformers. Advances in Neural Information Processing Systems, 33:5776–5788, 2020. * [9] Zhiqing Sun, Hongkun Yu, Xiaodan Song, Renjie Liu, Yiming Yang, and Denny Zhou. Mobilebert: a compact task-agnostic bert for resource-limited devices. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 2158–2170, 2020. * [10] Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 4171–4186, Minneapolis, Minnesota, June 2019. Association for Computational Linguistics. * [11] Victor Sanh, Lysandre Debut, Julien Chaumond, and Thomas Wolf. Distilbert, a distilled version of bert: smaller, faster, cheaper and lighter. arXiv preprint arXiv:1910.01108, 2019. * [12] Siqi Sun, Yu Cheng, Zhe Gan, and Jingjing Liu. Patient knowledge distillation for bert model compression. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 4323–4332, 2019. * [13] Iulia Turc, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. Well-read students learn better: The impact of student initialization on knowledge distillation. ArXiv, abs/1908.08962, 2019. * [14] Lu Hou, Zhiqi Huang, Lifeng Shang, Xin Jiang, Xiao Chen, and Qun Liu. Dynabert: Dynamic bert with adaptive width and depth. Advances in Neural Information Processing Systems, 33:9782–9793, 2020. * [15] François Lagunas, Ella Charlaix, Victor Sanh, and Alexander Rush. Block pruning for faster transformers. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 10619–10629, Online and Punta Cana, Dominican Republic, November 2021. Association for Computational Linguistics. * [16] Victor Sanh, Thomas Wolf, and Alexander Rush. Movement pruning: Adaptive sparsity by fine-tuning. Advances in Neural Information Processing Systems, 33:20378–20389, 2020. * [17] Ziheng Wang, Jeremy Wohlwend, and Tao Lei. Structured pruning of large language models. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 6151–6162, 2020. * [18] Matan Ben Noach and Yoav Goldberg. Compressing pre-trained language models by matrix decomposition. In Proceedings of the 1st Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics and the 10th International Joint Conference on Natural Language Processing, pages 884–889, 2020. * [19] Canwen Xu, Wangchunshu Zhou, Tao Ge, Furu Wei, and Ming Zhou. Bert-of-theseus: Compressing bert by progressive module replacing. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 7859–7869, 2020. * [20] Angela Fan, Edouard Grave, and Armand Joulin. Reducing transformer depth on demand with structured dropout. In International Conference on Learning Representations, 2019. * [21] Hassan Sajjad, Fahim Dalvi, Nadir Durrani, and Preslav Nakov. Poor man’s bert: Smaller and faster transformer models. arXiv preprint arXiv:2004.03844, 2020. * [22] Zichang Liu, Jue Wang, Tri Dao, Tianyi Zhou, Binhang Yuan, Zhao Song, Anshumali Shrivastava, Ce Zhang, Yuandong Tian, Christopher Re, et al. Deja vu: Contextual sparsity for efficient llms at inference time. In International Conference on Machine Learning, pages 22137–22176. PMLR, 2023. * [23] Elias Frantar, Saleh Ashkboos, Torsten Hoefler, and Dan Alistarh. Gptq: Accurate post-training quantization for generative pre-trained transformers. arXiv preprint arXiv:2210.17323, 2022. * [24] Ji Lin, Jiaming Tang, Haotian Tang, Shang Yang, Xingyu Dang, and Song Han. Awq: Activation-aware weight quantization for llm compression and acceleration. arXiv preprint arXiv:2306.00978, 2023. * [25] Liyang Liu, Shilong Zhang, Zhanghui Kuang, Aojun Zhou, Jing-Hao Xue, Xinjiang Wang, Yimin Chen, Wenming Yang, Qingmin Liao, and Wayne Zhang. Group fisher pruning for practical network compression. In International Conference on Machine Learning, pages 7021–7032. PMLR, 2021. * [26] Lucas Liebenwein, Cenk Baykal, Harry Lang, Dan Feldman, and Daniela Rus. Provable filter pruning for efficient neural networks. arXiv preprint arXiv:1911.07412, 2019. * [27] Yihui He, Ji Lin, Zhijian Liu, Hanrui Wang, Li-Jia Li, and Song Han. Amc: Automl for model compression and acceleration on mobile devices. In Proceedings of the European conference on computer vision (ECCV), pages 784–800, 2018. * [28] Hao Li, Asim Kadav, Igor Durdanovic, Hanan Samet, and Hans Peter Graf. Pruning filters for efficient convnets. arXiv preprint arXiv:1608.08710, 2016. * [29] Zhuang Liu, Jianguo Li, Zhiqiang Shen, Gao Huang, Shoumeng Yan, and Changshui Zhang. Learning efficient convolutional networks through network slimming. In Proceedings of the IEEE international conference on computer vision, pages 2736–2744, 2017. * [30] Yang He, Ping Liu, Ziwei Wang, Zhilan Hu, and Yi Yang. Filter pruning via geometric median for deep convolutional neural networks acceleration. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 4340–4349, 2019. * [31] Yang Sui, Miao Yin, Yi Xie, Huy Phan, Saman Aliari Zonouz, and Bo Yuan. Chip: Channel independence-based pruning for compact neural networks. Advances in Neural Information Processing Systems, 34:24604–24616, 2021. * [32] Markus Nagel, Rana Ali Amjad, Mart Van Baalen, Christos Louizos, and Tijmen Blankevoort. Up or down? adaptive rounding for post-training quantization. In International Conference on Machine Learning, pages 7197–7206. PMLR, 2020. * [33] Itay Hubara, Brian Chmiel, Moshe Island, Ron Banner, Joseph Naor, and Daniel Soudry. Accelerated sparse neural training: A provable and efficient method to find n: m transposable masks. Advances in Neural Information Processing Systems, 34:21099–21111, 2021. * [34] Elias Frantar, Sidak Pal Singh, and Dan Alistarh. Optimal Brain Compression: A framework for accurate post-training quantization and pruning. arXiv preprint arXiv:2208.11580, 2022. Accepted to NeurIPS 2022, to appear. * [35] Babak Hassibi and David Stork. Second order derivatives for network pruning: Optimal brain surgeon. Advances in neural information processing systems, 5, 1992. * [36] Eldar Kurtic, Daniel Campos, Tuan Nguyen, Elias Frantar, Mark Kurtz, Benjamin Fineran, Michael Goin, and Dan Alistarh. The optimal bert surgeon: Scalable and accurate second-order pruning for large language models. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 4163––4181, 2022. * [37] Sidak Pal Singh and Dan Alistarh. Woodfisher: Efficient second-order approximation for neural network compression. Advances in Neural Information Processing Systems, 33, 2020. * [38] Elias Frantar, Eldar Kurtic, and Dan Alistarh. M-fac: Efficient matrix-free approximations of second-order information. Advances in Neural Information Processing Systems, 34, 2021. * [39] Han Cai, Chuang Gan, Tianzhe Wang, Zhekai Zhang, and Song Han. Once-for-all: Train one network and specialize it for efficient deployment. arXiv preprint arXiv:1908.09791, 2019. * [40] Elias Frantar and Dan Alistarh. SPDY: Accurate pruning with speedup guarantees. arXiv preprint arXiv:2201.13096, 2022. * [41] Yawei Li, Kamil Adamczewski, Wen Li, Shuhang Gu, Radu Timofte, and Luc Van Gool. Revisiting random channel pruning for neural network compression. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 191–201, 2022. * [42] Zichao Guo, Xiangyu Zhang, Haoyuan Mu, Wen Heng, Zechun Liu, Yichen Wei, and Jian Sun. Single path one-shot neural architecture search with uniform sampling. In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XVI 16, pages 544–560. Springer, 2020. * [43] Pranav Rajpurkar, Jian Zhang, Konstantin Lopyrev, and Percy Liang. Squad: 100,000+ questions for machine comprehension of text. In EMNLP, 2016. * [44] Alex Wang, Amanpreet Singh, Julian Michael, Felix Hill, Omer Levy, and Samuel R. Bowman. Glue: A multi-task benchmark and analysis platform for natural language understanding. ArXiv, abs/1804.07461, 2018. * [45] Richard Socher, Alex Perelygin, Jean Wu, Jason Chuang, Christopher D Manning, Andrew Y Ng, and Christopher Potts. Recursive deep models for semantic compositionality over a sentiment treebank. In Proceedings of the 2013 conference on empirical methods in natural language processing, pages 1631–1642, 2013. * [46] Adina Williams, Nikita Nangia, and Samuel Bowman. A broad-coverage challenge corpus for sentence understanding through inference. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), pages 1112–1122. Association for Computational Linguistics, 2018. * [47] S. Shankar. Identifying quora question pairs having the same intent. 2017\. * [48] Aaron Gokaslan and Vanya Cohen. Openwebtext corpus, 2019. * [49] Woosuk Kwon, Sehoon Kim, Michael W Mahoney, Joseph Hassoun, Kurt Keutzer, and Amir Gholami. A fast post-training pruning framework for transformers. arXiv preprint arXiv:2204.09656, 2022. * [50] Peter Mattson, Vijay Janapa Reddi, Christine Cheng, Cody Coleman, Greg Diamos, David Kanter, Paulius Micikevicius, David Patterson, Guenther Schmuelling, Hanlin Tang, et al. Mlperf: An industry standard benchmark suite for machine learning performance. IEEE Micro, 40(2):8–16, 2020. * [51] Stephen Merity, Caiming Xiong, James Bradbury, and Richard Socher. Pointer sentinel mixture models, 2016. * [52] Angela Fan, Mike Lewis, and Yann Dauphin. Hierarchical neural story generation. arXiv preprint arXiv:1805.04833, 2018. * [53] Alec Radford, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, Ilya Sutskever, et al. Language models are unsupervised multitask learners. OpenAI blog, 1(8):9, 2019. * [54] Benoit Jacob, Skirmantas Kligys, Bo Chen, Menglong Zhu, Matthew Tang, Andrew Howard, Hartwig Adam, and Dmitry Kalenichenko. Quantization and training of neural networks for efficient integer-arithmetic-only inference. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 2704–2713, 2018. * [55] NeuralMagic. Deep sparse: A fast cpu inference engine, 2021. * [56] Eldar Kurtic and Dan Alistarh. Gmp*: Well-tuned global magnitude pruning can outperform most bert-pruning methods. arXiv preprint arXiv:2210.06384, 2022. * [57] Thomas Wolf, Lysandre Debut, Victor Sanh, Julien Chaumond, Clement Delangue, Anthony Moi, Pierric Cistac, Tim Rault, Rémi Louf, Morgan Funtowicz, Joe Davison, Sam Shleifer, Patrick von Platen, Clara Ma, Yacine Jernite, Julien Plu, Canwen Xu, Teven Le Scao, Sylvain Gugger, Mariama Drame, Quentin Lhoest, and Alexander M. Rush. Transformers: State-of-the-art natural language processing. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, pages 38–45, Online, October 2020. Association for Computational Linguistics. * [58] Quentin Lhoest, Albert Villanova del Moral, Yacine Jernite, Abhishek Thakur, Patrick von Platen, Suraj Patil, Julien Chaumond, Mariama Drame, Julien Plu, Lewis Tunstall, Joe Davison, Mario Šaško, Gunjan Chhablani, Bhavitvya Malik, Simon Brandeis, Teven Le Scao, Victor Sanh, Canwen Xu, Nicolas Patry, Angelina McMillan-Major, Philipp Schmid, Sylvain Gugger, Clément Delangue, Théo Matussière, Lysandre Debut, Stas Bekman, Pierric Cistac, Thibault Goehringer, Victor Mustar, François Lagunas, Alexander Rush, and Thomas Wolf. Datasets: A community library for natural language processing. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, pages 175–184. Association for Computational Linguistics, November 2021. * [59] Mark Kurtz, Justin Kopinsky, Rati Gelashvili, Alexander Matveev, John Carr, Michael Goin, William Leiserson, Sage Moore, Bill Nell, Nir Shavit, and Dan Alistarh. Inducing and exploiting activation sparsity for fast inference on deep neural networks. In Hal Daumé III and Aarti Singh, editors, Proceedings of the 37th International Conference on Machine Learning, volume 119 of Proceedings of Machine Learning Research, pages 5533–5543, Virtual, 13–18 Jul 2020. PMLR. * [60] Aran Komatsuzaki. One epoch is all you need. arXiv preprint arXiv:1906.06669, 2019. ## Appendix A Compound Compression for Edge Deployment Deploying large language models in edge environments requires running inference on low-power devices such as CPUs. Therefore, we follow the compound compression approach from [36] which bundles together structured, unstructured pruning, and quantization for efficient inference on CPUs. We start with ZipLM structurally pruned models, and apply on top the state-of-the-art oBERT unstructured pruning method [36] to 80% sparsity. After structured and unstructured pruning, we apply quantization-aware-training (QAT) [54] to quantize FP32 weights into INT8 representations. We benchmark these compound compressed models by running inference in the DeepSparse [55] engine, on a single-core of Intel Cascade Lake CPU. In this setting, we compare our results against the compound compression pipeline of [36] which applies layer dropping as a form of structured pruning. As can be seen from Figure 6, when we substitute layer dropping with a principled structured pruning via ZipLM, the resulting compound compressed models achieve very competitive latency-vs- accuracy performance in the edge-inference regime. At full accuracy recovery, ZipLM improves the speedup from 3x to 13x, while at the largest compression ratio ZipLM improves the speedup from 30x to 50x. Figure 6: Improvements in CPU-inference speedups for compound compressed $\textrm{BERT}_{\small{\textrm{base}}}\,$models on the SQuADv1.1 task when ZipLM is used for structured pruning. End-to-end latency indicated by the dashed line. ## Appendix B Ablation Studies In Table 5, we present ablation results for ZipLM and CoFi, with and without their respective layer-wise distillation techniques. ZipLM outperforms CoFi in all tasks when both methods use distillation, and in three out of four when distillation is not used. For example, ZipLM outperforms CoFi with a significant 3 point increase in F1 score on the SQuAD task in both setups. Furthermore, when comparing ZipLM results with and without layer-wise distillation, it can be observed that benefits are pronounced for low data tasks, where accuracy improvements reach up to 2 points. Table 5: Comparison of ZipLM and CoFi dev-set results, with and without layer-wise distillation. | SST-2 acc. | QNLI acc. | MNLI m-acc. | SQuAD F1 ---|---|---|---|--- CoFi | 90.4 | 86.1 | 80.6 | 82.6 $\textrm{ZipBERT}_{\small{\textrm{base}}}$ | 91.7 | 88.6 | 81.7 | 85.7 CoFi w/o $\mathcal{L}_{\mathrm{layer}}$ | 91.1 | 85.1 | 79.7 | 82.5 $\textrm{ZipBERT}_{\small{\textrm{base}}}$ w/o $\mathcal{L}_{\mathrm{token}}$ | 89.2 | 86.5 | 81.2 | 85.7 ## Appendix C Additional GLUE results Due to space constraints, in Figure 3 we present results only on four GLUE tasks. Therefore, for completeness, in Figure 7 we present results on the remaining four GLUE tasks, namely: CoLA, MRPC, STS-B, and RTE. Results show the same trends as the other four GLUE tasks, with large improvements for ZipLM, especially at higher compression rates. Figure 7: Structured compression of $\textrm{BERT}_{\small{\textrm{base}}}\,$on CoLA, MRPC, STS-B, and RTE tasks. Dashed horizontal lines represent full accuracy recovery of the uncompressed model. ## Appendix D Additional Validation Evaluating and comparing compressed models on the development set (dev-set) is standard practice, as it enables comparisons with off-the-shelf results from the literature. However, an implicit assumption behind such comparisons is that all methods tune their hyper-parameters only on a subset of the dev-set before evaluating and reporting results on all samples, which is not always the case. Moreover, specifically-tuned hyper-parameters can lead to large performance differences, especially when compressing LLMs [56]. To ensure that there is no such “overfitting” on the dev-set, in Table 6 we compare ZipLM against the prior state-of-the-art CoFi approach on _unseen test-set_ , obtained by submitting predictions to the official GLUE evaluation server. The results show consistent improvements over CoFi, on both dev- and test-sets. Table 6: Dev- and test-set comparison of $\textrm{ZipBERT}_{\small{\textrm{base}}}$ and CoFi models with comparable speedups. | dev-set | test-set ---|---|--- | CoFi | $\textrm{ZipBERT}_{\small{\textrm{base}}}$ | CoFi | $\textrm{ZipBERT}_{\small{\textrm{base}}}$ QNLI, acc. | 86.1 | 88.6 | 85.8 | 88.4 SST-2, acc. | 90.4 | 91.7 | 88.2 | 91.8 MNLI, m-acc. | 80.6 | 81.7 | 80.7 | 81.9 MNLI, mm-acc. | 80.7 | 82.0 | 79.9 | 80.6 SQuAD, F1 | 82.6 | 85.7 | N/A | N/A ## Appendix E Latency table used for ZipLM pruning As described in Section 3.2, we record the time to run an attention block, including all overheads, with $0,\dots,N_{\text{heads}}-1$ heads pruned (pruning everything with runtime 0 is also considered) and similarly for the fully-connected block with the intermediate dimension shrunk by a factor of $0.9^{i}$, for $i=0,\dots,42$; in relative steps of $10\%$ up until $\approx 99\%$ sparsity, following [40]. In Table 7 we present an example of such a latency table used in ZipLM pruning approach. Table 7: An example of the latency table used by the ZipLM pruning approach. Intermediate size | Latency (ms) | Number of heads | Latency (ms) ---|---|---|--- 3072 | 11.9 | 12 | 7.9 1814 | 7.4 | 10 | 6.7 1322 | 5.8 | 8 | 5.8 302 | 1.6 | 6 | 4.4 130 | 1.0 | 4 | 3.2 76 | 0.9 | 2 | 1.9 33 | 0.7 | 0 | 0 ## Appendix F Speedup Evaluations As shown in Figure 1, ZipLM is based on measuring runtimes of higher-level modules, such as attention heads and fully connected matrices, rather than low-level operators. This makes our approach independent of underlying optimizations in different inference engines and frameworks, which usually perform further optimizations such as operator-fusion. Our runtime lookup table contains information about the runtime of a Transformer layer with different numbers of attention heads, and various dimensions of the fully connected matrices. This implies that we measure runtimes of a layer with 12 heads, 11 heads, 10 heads, and so on, as well as the runtimes of fully connected matrices with hidden sizes ranging from 3072 to 0. We utilize this information to guide pruning decisions. To fully validate the ability of ZipLM to compress the model while satisfying desired speedup constraints via the described approach, we provide the timing results in Table 8, comparing the desired (target) speedup and the achieved (measured) speedup for different models. Table 8: Comparison of target (desired) inference speedups with achieved (on-device measured) speedups obtained with our ZipLM pruning approach. $\textrm{BERT}_{\small{\textrm{base}}}\,$on SQuADv1.1 | $\textrm{BERT}_{\small{\textrm{large}}}\,$on SQuADv1.1 ---|--- Target speedup | Achieved speedup | Deviation | Target speedup | Achieved speedup | Deviation 2 | 1.98 | -1.00% | 2 | 2.01 | +0.50% 4 | 4.05 | +1.25% | 4 | 4.05 | +1.25% 6 | 6.16 | +2.67% | 6 | 6.09 | +1.50% 8 | 8.25 | +3.12% | 8 | 8.27 | +3.37% 10 | 10.36 | +3.60% | 10 | 10.33 | +3.30% 12 | 12.31 | +2.58% | 12 | 12.46 | +3.83% 14 | 14.33 | +2.35% | 14 | 14.74 | +5.28% As can be seen from the Table 8, the deviation between the desired (target) and the achieved (measured) speedup is at most 5.28%. This confirms that our approach indeed provides reliable runtime information to guide the pruning decisions. ## Appendix G Structure of Pruned Models Through a comprehensive examination of ZipLM pruned BERT models across all datasets considered in Section 4, we aim to identify trends in the pruning of key components of the Transformer layer, namely attention heads and intermediate size, needed to achieve a specific speedup target. As illustrated in Figure 8, we observe that the intermediate size is pruned at a higher rate relative to attention heads, which aligns with the fact that the intermediate size dictates the dimensions of the two large linear layers in the feed- forward part of the Transformer block. For instance, to attain a 2x speedup, roughly 60% of the intermediate size and 40% of the attention heads need to be removed. Additionally, in Figure 9, we visualize the entire encoder size needed to reach a specific speedup target. Interestingly, we find that 15x faster models retain on average only 2% of intermediate size and 6% of attention heads which amounts to only 2.9M parameters overall, while at the same time recovering more than 95% of the uncompressed model’s accuracy (see Figure 3). Figure 8: Percentage of pruned attention heads and intermediate size to reach a specific speedup target with ZipLM. Figure 9: Encoder size vs. speedup factor of ZipLM pruned $\textrm{BERT}_{\small{\textrm{base}}}\,$models, averaged over all considered datasets in Section 4. Additionally, in Figures 10, 11, 12, 13 we visualize the number of remaining heads and intermediate size across all Transformer layers and various speedup targets on a subset of GLUE datasets. Figure 10: Remaining number of attention heads and intermediate size across all layers of the ZipLM compressed $\textrm{BERT}_{\small{\textrm{base}}}\,$model at various speedups and MNLI dataset. Figure 11: Remaining number of attention heads and intermediate size across all layers of the ZipLM compressed $\textrm{BERT}_{\small{\textrm{base}}}\,$model at various speedups and QNLI dataset. Figure 12: Remaining number of attention heads and intermediate size across all layers of the ZipLM compressed $\textrm{BERT}_{\small{\textrm{base}}}\,$model at various speedups and QQP dataset. Figure 13: Remaining number of attention heads and intermediate size across all layers of the ZipLM compressed $\textrm{BERT}_{\small{\textrm{base}}}\,$model at various speedups and SST-2 dataset. ## Appendix H Experiments - Additional Results In Table 9 we report accuracy and model size of ZipLM pruned models visualized in Section 4, in Figures 2 and 3. Table 9: Accuracy and model size for ZipLM pruned models in Section 4. | $\textrm{BERT}_{\small{\textrm{base}}}\,$ | $\textrm{BERT}_{\small{\textrm{large}}}\,$ ---|---|--- | QNLI | MNLI | SST2 | QQP | SQuADv1 | SQuADv1 Speedup | Acc. | Encoder size (M) | Acc. | Encoder size (M) | Acc. | Encoder size (M) | Acc. | Encoder size (M) | F1 | Encoder size (M) | F1 | Encoder size (M) 32x | 91.4 | 38.0 | 84.8 | 38.5 | 93.4 | 38.7 | 91.3 | 37.8 | 89.1 | 37.3 | 91.6 | 141.1 33x | 91.1 | 23.8 | 84.8 | 23.5 | 93.4 | 24.1 | 91.3 | 23.8 | 88.6 | 23.4 | 91.4 | 188.3 34x | 90.9 | 16.9 | 84.0 | 17.1 | 93.0 | 17.2 | 91.3 | 16.8 | 88.0 | 16.8 | 91.1 | 163.1 35x | 90.8 | 12.5 | 84.0 | 13.5 | 93.0 | 13.5 | 91.1 | 13.0 | 87.5 | 13.0 | 90.8 | 148.5 36x | 90.4 | 39.5 | 83.5 | 10.5 | 93.0 | 11.0 | 91.1 | 10.2 | 86.7 | 10.4 | 90.2 | 139.1 37x | 89.8 | 38.0 | 83.2 | 38.8 | 93.0 | 39.0 | 90.9 | 38.1 | 86.1 | 38.7 | 89.9 | 132.7 38x | 89.2 | 36.4 | 83.1 | 37.5 | 93.0 | 37.6 | 90.9 | 36.8 | 85.7 | 37.5 | 89.7 | 127.5 39x | 89.1 | 35.7 | 82.8 | 36.3 | 93.0 | 36.7 | 90.8 | 35.8 | 85.3 | 36.2 | 89.3 | 123.8 10x | 88.6 | 34.9 | 82.7 | 35.4 | 93.0 | 35.7 | 90.8 | 34.9 | 84.2 | 35.3 | 89.1 | 120.9 11x | 88.6 | 34.0 | 82.5 | 34.7 | 92.7 | 34.9 | 90.7 | 34.3 | 83.8 | 34.7 | 88.8 | 118.4 12x | 87.8 | 33.6 | 81.7 | 34.1 | 91.7 | 34.2 | 90.6 | 34.1 | 83.2 | 34.0 | 88.4 | 116.4 13x | 87.6 | 33.2 | 81.3 | 33.5 | 91.7 | 33.8 | 90.6 | 33.7 | 82.5 | 33.4 | 87.9 | 114.9 14x | 87.4 | 32.8 | 81.2 | 33.3 | 91.7 | 33.6 | 90.3 | 33.3 | 81.7 | 33.2 | 87.7 | 113.7 15x | 87.2 | 32.6 | 80.8 | 32.9 | 90.7 | 33.2 | 90.3 | 32.9 | 81.4 | 32.9 | 87.6 | 112.5 ## Appendix I Hyper-parameters for Reproducibility To facilitate reproducibility, we conduct experiments in the open-source Transformers library [57], and use publicly available datasets [58]. We plan to open-source our entire framework which supports one-shot and gradual structured pruning via SparseML [59], making it very easy to experiment with other models and datasets. In addition to our code, we plan to open-source all of our compressed models via the popular HuggingFace Hub. In Table 10 we report hyper-parameters used to produce our ZipLM pruned models in Section 4. Because of the excessive memory overhead, we don’t make use of any kind of knowledge distillation when pruning the GPT2 model. Following insights from DistilGTP2, we hypothesize that this can further improve our results. We follow [60] and disable dropout regularization while pre-training ZipGPT2 models at OpenWebTextCorpus dataset. Table 10: Hyper-parameters used for gradual ZipLM runs in Section 4. | $\textrm{BERT}_{\small{\textrm{base}}}\,$ | $\textrm{BERT}_{\small{\textrm{large}}}\,$ | GPT2 ---|---|---|--- batch-size | 16 SQuADv1 32 GLUE | 128 max-seq-length | 384 SQuADv1 128 GLUE | 1024 finetune before pruning | 3 epochs | 50k steps finetune in-between pruning steps | 8 epochs | 10 epochs | 2 epochs LR schedule in-between pruning steps | linear decay | linear decay initial LR | 8e-5 | 5e-5 | 1e-3 #calibration samples | 2048 | 512 speedup-targets | {2, 3, 4, 5, …, 15}x | {1.5, 2, 2.5, 3}x knowledge distillation $\lambda_{1}$ | 0 | 1.0 knowledge distillation $\lambda_{2}$ | 1.0 SQuADv1 0.5 GLUE | 0 knowledge distillation $\lambda_{3}$ | 0.0 SQuADv1 0.5 GLUE | 0 weight-decay | 0.03 | 0.05 | 0 ## Appendix J Broader Impact and Limitations Our results contribute to the line of work on efficient language models. Thus, it should help reduce the energy and monetary cost of inference over such models, and allow them to be used without access to powerful hardware. While this is a mainly positive outcome, it also reduces the cost of employing these models for detrimental purposes, such as spam generation. Thus, this significant cost reduction for inference should also be seen as further motivation for methods to ensure safe usage of these models, such as watermarking or alignment. As any academic study, our work is not without its limitations. All of our benchmarks are focused on English-language datasets and therefore our results do not provide insights into compression effects for low-data languages. Unfortunately, this limitation is inherent to all of the existing works on compression due to the lack of standardized benchmarks. Given that our structured pruning approach relies on a small sample of calibration data to perform pruning decisions, we hypothesize that our approach should be able to provide satisfying results in the low-data setup as well. At the moment we do not have data to support these claims, but we see it as an opportunity for future work.
# Cross-Validated Decision Trees with Targeted Maximum Likelihood Estimation for Nonparametric causal mixtures analysis David B. McCoy Department of Environmental Health Sciences University of California, Berkeley Berkeley, CA 94704 <EMAIL_ADDRESS> & Alan E. Hubbard Department of Biostatistics University of California, Berkeley Berkeley, CA 94704 <EMAIL_ADDRESS> & Alejandro Schuler Department of Biostatistics University of California, Berkeley Berkeley, CA 94704 <EMAIL_ADDRESS> & Mark J. van der Laan Department of Biostatistics University of California, Berkeley Berkeley, CA 94704 <EMAIL_ADDRESS> ###### Abstract Exposure to a mixture of chemicals such as drugs, pollutants and nutrients occur in most realistic exposure or treatment situations. We can imagine that, within this exposure space, there are arbitrary regions wherein we can measure the covariate adjusted outcome within the region compared to the complimentary exposure space. Ideally, it is most useful to estimate a causal estimand that maximizes this mean difference. A statistical estimator that identifies regions which maximize this difference and delivers the relevant effect unbiasedly would be valuable to public health efforts which aim to understand what levels of pollutants or drugs have the strongest effect. This estimator would take in as input a vector of exposures $A$ which can be a variety of data types (binary, multinomial, continuous), a vector of baseline covariates $W$ and outcome $Y$ and outputs 1. a region of the exposure space that attempts to optimize this maximum mean difference and 2. an unbiased estimate of the causal effect comparing the average outcomes if every unit were forced to self-select exposure within that region compared to the equivalent space outside of that region (the average regional-exposure effect (ARE)). Rather than the region of interest being of arbitrary shape, searching for rectangular regions is most helpful for policy implications by helping policymakers decide on what combination of thresholds to set on allowable combinations of exposures. This is because rectangular regions can be expressed as a series of thresholds, for example, $A_{1}\geq a_{1}$ while $A_{2}\leq a_{2}$ where $a_{1},a_{2}$ are specific doses of exposures $A_{1},A_{2}$ respectively. Non-parametric methods such as decision trees are a useful tool to evaluate combined exposures by finding partitions in the joint-exposure (mixture) space that best explain the variance in an outcome. We present a methodology for evaluating the causal effects of a data- adaptively determined mixture region using decision trees. This approach uses K-fold cross-validation and partitions the full data in each fold into a parameter-generating sample and an estimation sample. In the parameter- generating sample, decision trees are applied to a mixed exposure to obtain the region and estimators for our statistical target parameter. The region indicator and estimators are then applied to the estimation sample where the average regional-exposure effect is estimated. Targeted learning is used to update our initial estimates of the ARE in the estimation sample to optimize bias and variance towards our target parameter. This results in a plug-in estimator for our data-adaptive decision tree parameter that has an asymptotically normal distribution with minimum variance from which we can derive confidence intervals. Likewise, our approach uses the full data with no loss of power due to sample splitting. The open source software, CVtreeMLE, a package in R, implements our proposed methodology. Our approach makes possible the non-parametric estimation of the causal effects of a mixed exposure producing results that are both interpretable and asymptotically efficient. Thus, CVtreeMLE allows researchers to discover important mixtures of exposure and also provides robust statistical inference for the impact of these mixtures. ## 1 Introduction In most environmental epidemiology studies, researchers are interested in how a joint exposure affects an outcome. This is because, in most real world exposure settings, an individual is exposed to a multitude of chemicals concurrently or, a mixed exposure. Individuals are exposed to a range of multi-pollutant chemical exposures from the environment including air pollution, endocrine disrupting chemicals, pesticides, and heavy metals. Because many of these chemicals may affect the same underlying biological pathway which lead to a disease state, the toxicity of these chemicals can be modified by simultaneous or sequential exposure to multiple agents. In these mixed exposure settings, the joint impact of the mixture on an outcome may not be equal to the additive effects of each individual agent. Mixed exposures may have impacts that are greater than expected given the sum of individual exposures or effects may be less than additive expectations if certain exposures antagonize the affects of others. Likewise, the effects of a mixed exposure may be different for subpopulations of individuals based on environmental stressors, genetic, and psychosocial factors that may modify the impact of a mixed exposure. Safe (1993); Kortenkamp (2007) Causal inference of mixed exposures has been limited by reliance on parametric models and, in most cases, by researchers considering only one exposure at a time, usually estimated as a coefficient in a generalized linear regression model (GLM). This independent assessment of exposures poorly estimates the joint impact of a collection of the same exposures in a realistic exposure setting. Given that most researchers simply add individual effects to estimate the joint impact of an exposure, it is almost certain that the current evidence of the total impact environmental toxicants have on chronic disease is incorrectly estimated. The impact of using linear modeling is not limited to just potential bias: in the case where linearity does not hold, it’s not even clear what is being estimated. The limitation in effective estimation of the joint effects of mixed exposure is (in-part) due to the lack of robust statistical methods. There has been some method development for estimation of joint effects of mixed exposures, such as Weighted Quantile Sum Regression Keil et al. (2019), Bayesian Mixture Modeling De Vocht et al. (2012), and Bayesian Kernel Machine Regression Bobb et al. (2014). However, these mixture methods have strong assumptions built into them, including directional homogeneity (e.g. all mixtures having a positive effect), linear/additive assumptions and/or require information priors. Many methods suffer from human bias due to choice of priors or poor model fit. More flexible models remain more or less a black-box and describe the mixture through a series of plots rather than with an interpretable summary statistic Bobb et al. (2014). Given that the National Institute for Environmental Health Sciences (NIEHS) has included the study of mixtures as a key goal in its 2018-2023 strategic plan National Institute of Environmental Health Sciences (2018) (NIEHS), it is imperative to develop new statistical methods for mixtures that are less biased, rely less on human input, use machine learning (ML) to model complex interactions, and are designed to return an interpretable parameter of interest. Decision trees are a useful tool for outcome prediction based on exposures because they are fast, nonparametric (i.e. can discover and model interaction effects), and interpretable Leo Breiman (1984). However, it is not immediately clear how to adapt outcome prediction methods to inference about the effect of some kind of hypothetical intervention on the mixture of exposures- especially because in these settings we don’t have a particular intervention in mind. Rather than leveraging decision trees for a simple prediction model, we introduce a target parameter on top of the prediction model, which is the average outcome within a fixed region of the exposure space. When an ensemble of decision trees is applied to an exposure mixture, this coincides with a leaf in the best fitting decision tree. By cross-estimating the average outcome given exposure to this region which maximizes the outcome difference we are able to build an estimator that is asymptotically unbiased with the smallest variance for our causal parameter of interest. Previous work, in the most naive approach, confidence intervals (CI) and hypothesis testing of decision trees is done by constructing a $(1-\alpha)\times 100$% confidence interval for a node mean $\bar{y}_{t}$ as $\bar{y}_{t}\pm±z_{1-\alpha/2}(\frac{s_{t}}{\sqrt{n_{t}}})$ where $\bar{y}_{t}$ is the node mean and $s_{t}$ is the standard deviation estimates in the node. Of course, these CI intervals tend to be overly optimistic because 1. decision trees are adaptive and greedy algorithms, meaning that they have a tendency to overfit and 2. the target parameter, in this case the node average, is estimated on the same data by which the node was created. Because of this the estimated CIs are too narrow. The best approach is to use an independent test set to derive inference for the expected outcome in each leaf. However, this approach is costly if additional data is gathered or power is greatly reduced if sample-splitting is done. Sampling splitting is done in previous work for causal inference of decision trees using so-called "honest estimation" for estimation of heterogeneous causal effects of a binary treatment. This approach Athey and Imbens (2016) uses one part of the data for constructing the partition nodes and and and another for estimating effects within leaves of the partition. Our proposed approach follows a similar sample-splitting technique where one part of the data is used to determine the partition nodes and the other is used to estimate the parameter of interest; however, we extend this technique to K-fold cross-validation where we rotate through the full data. Additionally, rather than estimating heterogenous treatment effects, we are interested in mapping a set of exposures that are of a variety of data types (continuous, binary, multinomial) into a set of partitioning rules using the best fitting decision tree from which we can estimate the average regional-exposure effect, or the expected outcome difference if all individuals were exposed to an exposure region compared to if no individuals were exposed to this region. In most research scenarios, the analyst is interested in causal inference for an a priori specified treatment or exposure. However, in the evaluation of a mixed exposure it is not known what mixture components, levels of these components and combinations of these component levels contribute to the most to a change in the outcome. In the ideal scenario, the analyst has knowledge of the full, multidimensional dose-response curve $E[Y(A_{1},A_{2},...A_{k})]$ where $A$ are the exposures and $Y$ is the outcome. However, even in this case, it is difficult to estimate and/or interpret this curve. Estimation is hard because 1. we need unrealistic assumptions to get identifiability for the full curve and 2. the curve isn’t pathwise differentiable which means there aren’t any robust methods to build confidence intervals. Therefore, a sensible approach is to instead categorize the joint exposure and compare averages between categories as one would for a binary exposure. This approach is helpful because we can define interpretable categories like $(A_{1}>a_{1})\&(A_{2}<a_{2})$ where $a_{i}$ are specific values in $A$ (vs complement of this space) which are of clear interest to policymakers. Identification assumptions are also more transparent in this setting. However, we don’t know a priori what the right categorization of the exposure space are given some objective function. We have to use the data to tell us what regions are determined given a predefined objective function. In our case, we want a categorization that shows a maximal mean difference in outcomes. Regression trees are a nice way to do this while respecting the fact that we want interpretable rules like the above. The idea is to fit a kind of decision tree to figure out what thresholds in the exposure space produce a maximal exposure effect. As discussed, the result can be biased if we use the same data to define the thresholds and to estimate the effects in each leaf. We solve that problem by splitting the data, doing threshold estimation in one part and regional-exposure effect estimation (given the fixed thresholds) in the other. We can even redo the splits in a round-robin fashion (K-fold cross-validation) to efficiently use all of the data. Lastly, once we have thresholds, we want to get the best possible inference for the effect. We could always do a difference in outcome means between the samples in each category/region, but that estimate would be 1. biased by confounding and 2. have a large confidence interval because we haven’t used covariates to soak up residual variance. Our approach is thus to use a doubly-robust efficient estimator (TMLE) that simultaneously addresses both these problems. Building on prior work related to data-adaptive parameters Hubbard et al. (2016) and cross-validated targeted minimum loss-based estimation (CV-TMLE) Zheng and van der Laan (2010), our method, called CVtreeMLE, is a novel approach for estimating the joint impact of a mixed exposure by using CV-TMLE which guarantees consistency, efficiency, and multiple robustness despite using highly flexible learners (ensemble machine learning) to estimate a data- adaptive parameter. CVtreeMLE summarizes the effect of a joint exposure on the outcome of interest by first doing an iterative backfitting procedure, similar to general additive models Hastie and Tibshirani (1990), to fit $f(A)$, a Super Learner van der Laan Mark et al. (2007) of decision trees, and $h(W)$, an unrestricted Super Learner, in a semi-parametric model; $E(Y|A,W)=f(A)+h(W)$, where $A$ is a vector of exposures and $W$ is a vector of covariates. In many public health settings, the analyst is first interested in a parsimonious set of thresholds focusing on the exposure space that best explains some outcome across the whole population rather partitions that also include baseline covariates. This additive model approach allows us to identify partitioning nodes in the exposure space while flexibly adjusting for covariates. In this way, we can data-adaptively find the best fitting decision tree model which has the lowest cross-validated model error while flexibly adjusting for covariates. This procedure is done to find partitions in the mixture space which allows for an interpretable mixture contrast parameter, "What is the expected difference in outcomes if all individuals were exposed to this region of the mixed exposure vs. if no individuals were exposed?". This approach easily extends to marginal case (partitions on individual exposures) as well. Our approach for integrating decision trees as a data- adaptive parameter with cross-validated targeted minimum loss-based estimation (CV-TMLE) allows for flexible machine learning estimators to be used to estimate nuisance parameter functionals while preserving desirable asymptotic properties of our target parameter. We provide implementations of this methodology in our free and open source software CVtreeMLE package, for the R language and environment for statistical computing [R Core Team, 2022]. This manuscript is organized as follows, in Section 2.1 we give a background of semi-parametric methodology, in section 2.2 we discuss the ARE target parameter for a fixed exposure region and in 2.3 the assumptions necessary for our statistical estimate to have a causal interpretation. In section 3 we discuss estimation and inference of the ARE for a fixed region. In section 4 we discuss data-adaptively determining the region which maximizes the W-controlled mean outcome difference. In section 5 we show how this requires cross-estimation which builds from 2.2 for a fixed region ARE. In section 5 we expand this to cross-estimation to k-fold CV and discuss methods for pooling estimates across the folds. Lastly, in section 5.3, because we may have different data-adaptively identified regions across the CV folds, we discuss the union rule which pairs with the pooled estimates. In section 6 we discuss simulations with two and three exposures and show our estimator is asymptotically unbiased with a normally sampling distribution. In section 7 we apply CVtreeMLE to the NIEHS mixtures workshop data and identify interactions built into the synthetic data. In section 7.1 we compare CVtreeMLE to the popular quantile sum g-computation method. In section 7.2 we apply CVtreeMLE to NHANES data to determine if there is association between mixed metals and leukocyte telomere length. Section 8 describes our CVtreeMLE software. We end with a brief discussion of the CVtreeMLE method in Section 9. ## 2 The Estimation Problem ### 2.1 Setup and Notation Our setting is an observational study with baseline covariates ($W\in\mathbb{R}^{p}$), multiple exposures ($A\in\mathbb{R}^{m}$), and a single-timepoint outcome ($Y$). Let $O=(W,A,Y)$ denote the observable data. We presume that there exists a potential outcome function $Y(a)$ (i.e. $Y(a)$ is a random variable for each value of $a$) that generates the outcome that would have obtained for each observation had exposure been forced to the value $A=a$. These potential outcomes are unobserved but the observed outcome $Y$ corresponds to the potential outcome for the observed value $A$ of the exposure, i.e. $Y=Y(A)$. Let $E[Y(a)|W=w]=\mu(a,w)$ denote the causal dose- response curve for observations with covariates $w$ so that $E[\mu(a,W)]$ represents the average outcome we would observe if we forced treatment to $a$ for all observations. We use $P_{0}$ to denote the data-generating distribution. That is, each sample from $P_{0}$ results in a different realization of the data and if sampled many times we would eventually learn the true $P_{0}$ distribution. We assume our $O_{1},O_{2},...,O_{n}$ are iid draws of $O=(W,A,Y)\sim P_{0}$. We decompose the joint density as $p_{Y,A,W}(y,a,w)=p_{Y|A,W}(y,a,w)p_{A|W}(a,w)p_{W}(w)$ and make no assumptions about the forms of these densities. Compare this to to many methods which assume a parametric model which is one where each probability distribution $P\in\mathcal{M}$ can be uniquely described with a finite-dimensional set of parameters. Many methods assume $O$ to be identically distributed normal random variables, which means that the model can be described by the mean and standard deviation. Models like GLMs assume normal-linear relationships and assume $Y=X\beta+\mathcal{N}(\mu,\sigma^{2})$. Thus, methods for mixtures that use this approach have three parameters: the slope $\beta$, mean $\mu$ and standard deviation $\sigma$ of the normally-distributed random noise. This model assumes that the true relationship between $X$ and conditional mean of $Y$ is additive and linear and that the conditional distribution of $Y$ given $X$ is normal with a standard deviation that is fixed and doesn’t depend on $X$. Of course, these are very strict assumptions especially in the case of mixtures where exposures from a common source may be highly correlated, may interact on the outcome in a non-additive way, and may have non-normal distributions. As such, simply adding coefficients attached to variables in a mixture to estimate the overall joint effect may be biased. Our statistical target parameter, $\Psi(P_{0})$, is defined as a mapping from the statistical model, $\mathcal{M}$, to the parameter space (i.e., a real number) $\mathbb{R}$. That is, $\Psi$: $\mathcal{M}\rightarrow\mathbb{R}$. We can think of this as, if $\Psi$ were given the true distribution $P_{0}$ it would provide us with our true estimand of interest. We can think of our observed data $(O_{1}\dots O_{n})$ as a (random) probability distribution $P_{n}$ that places probability mass $1/n$ at each observation $O_{i}$. Our goal is to obtain a good approximation of the estimand $\Psi$, thus we need an estimator, which is an a-priori specified algorithm that is defined as a mapping from the set of possible empirical distributions, $P_{n}$ to the parameter space. More concretely, the estimator is a function that takes as input the observed data, a realization of $P_{n}$, and gives as output a value in the parameter space, which is the estimate, $\hat{\Psi}(P_{n})$. Since the estimator $\hat{\Psi}$ is a function of the empirical distribution $P_{n}$, the estimator itself is a random variable with a sampling distribution. So, if we repeat the experiment of drawing $n$ observations we would every time end up with a different realization of our estimate. We would like an estimator that is provably unbiased relative to the true (unknown) target parameter and which has the smallest possible sampling variance so that our estimation error is as small as it can be on average. ### 2.2 Defining the Differential Effect Given Regional Exposure In problems with binary treatment $A\in\\{0,1\\}$ the standard counterfactual model defines potential outcomes $Y(0)$ and $Y(1)$ describing what would happen to each individual had they been forced onto either treatment. The estimand of interest is most often the average treatment effect $E[Y(1)]-E[Y(0)]$. In our setting $A\in\mathbb{R}^{m}$ is continuous and we must thus define a potential outcome $Y(a)$ for each of the infinite possible values of the exposure. There is no singular, obvious “effect of treatment”. In this work, we propose asking about the differential policy effect of allowing self-selection within a region of the exposure space. For example, we may be interested in the effect of a law that limits arsenic soil levels to a certain level while also limiting cadmium to a different level. In this setting, communities are still left to “self-select” exposures, but they must conform to the legal limits. In order to estimate such an effect, we have to specify how each individual would modify their “preference” for each exposure level when presented with a more limited set of choices. In this work we presume that relative self-selection preferences are preserved. For example, if a person were twice as likely to choose level $a_{1}$ as they were to choose $a_{2}$, that relative preference would continue to hold whether or not $a_{3}$ were a legal option.111 Another way of looking at this is to define a random variable $T$ that represents the activation of the policy (e.g. law) that nominally forces the exposures into the desired region $\mathcal{A}$ and ask about the ATE of $T$ on $Y$. In this view, we continue by presuming that $T$ only affects $Y$ through the ”potential exposures” $A(t)$ and that the relationship between counterfactual exposure levels $A(0)$ and $A(1)$ is: $p_{A(1)|W}(a,w)=\frac{1_{\mathcal{A}}(a)}{\pi_{\mathcal{A}}(w)}p_{A(0)|W}(a,w)$ e.g., we assume that by enacting the “law” $T$ we are changing the distribution of exposures a way that does not violate said law and that maintains relative self-selection preferences, conditional on covariates. Then, to estimate the ATE of enacting this policy it suffices to estimate the ARE for $\mathcal{A}$. This is not always reasonable a reasonable assumption to make! For example, with a single continuous exposure with $\mathcal{A}=[\alpha,\infty]$, it could be more reasonable to assume that the modified treatment takes the value $\tilde{A}=\text{max}(A,\alpha)$. This has important and general implications for estimating the effect of “binary” policies that are really thresholds on continuous exposures. This is a special kind of modified treatment policy Hejazi et al. (2021, 2022); Iván Díaz Muñoz and Mark van der Laan* (2012). Formally, let $\mathcal{A}\subset\mathbb{R}^{m}$ denote a subset of the exposure space. For example, presume that $\mathcal{A}$ represents dosages of drugs that have been deemed safe for combination. Let $\mathcal{A}$ also denote the binary random variable $1_{\mathcal{A}}(A)$ indicating whether an observation conformed to the policy. Let $\pi_{\mathcal{A}}(w)=P(A\in\mathcal{A}|W=w)$ be the probability that an observation with covariates $w$ naturally self-selects treatment in the requested region $\mathcal{A}$. We define the modified treatment variable $\tilde{A}$ to represent the distribution of exposure once all exposures are forced to be self-selected within the region $\mathcal{A}$. The modified treatment has density: $p_{\tilde{A}|W}(a,w)=\frac{1_{\mathcal{A}}(a)}{\pi_{\mathcal{A}}(w)}p_{A|W}(a,w)$ which preserves the relative self-selection preferences for each available exposure and sets preferences for “outlawed” exposures to zero. Now we can define the population expected outcome ($\mu$) if we were to impose this policy: $\displaystyle E[\mu(\tilde{A},W)]$ $\displaystyle=\int\mu(a,w)\,dP_{\tilde{A},W}$ $\displaystyle=\int\mu(a,w)\frac{1_{\mathcal{A}}(a)}{\pi_{\mathcal{A}}(w)}\,p_{A|W}(a,w)p_{W}(w)\,da\,dw$ $\displaystyle=\int_{w}\underbrace{\left[\int_{a\in\mathcal{A}}\mu(a,w)\frac{p_{A|W}(a,w)}{\pi_{\mathcal{A}}(w)}\ da\right]}_{Q(\mathcal{A}=1,w)}\ p_{W}(w)dw$ $\displaystyle=E[Q(\mathcal{A}=1,W)]$ What we have shown is that this parameter is a population average of some function $m$ when $\mathcal{A}$ is forced to 1. For any value of $w$, $m(1,w)$ is a particular convexly weighted average of the causal dose-response curve across the different exposure levels, thus collapsing it down to a single number for each $w$. In a similar fashion we define the population average outcome under the complementary policy $\mathcal{A}^{c}$, which is $E[Q(\mathcal{A}=0,W)]$. With this, we are ready to introduce our parameter of interest, which we call the average regional effect (ARE): $\psi^{*}=E[Q(\mathcal{A}=1,W)]-E[Q(\mathcal{A}=0,W)]$ representing the difference in average outcomes if we forced exposure to self- selection within $\mathcal{A}$ vs. to self-selection within $\mathcal{A}^{c}$. In most applications, it is not known a-priori what region $\mathcal{A}$ should be set. For example, we may not know how various chemicals or drugs interact and how to set safe limits for all of them. Therefore $\mathcal{A}$ itself is in practice something that should be estimated in order to maximize some objective. That said, for the purposes of establishing our theory we should first imagine $\mathcal{A}$ as known and fixed but depends on the original distribution of $A$. We will then show how we can choose what policy to enact (i.e. choose $\mathcal{A}$) while also unbiasedly estimating its effect. ### 2.3 Identification and Causal Assumptions Our target parameter is defined on the causal data-generating process, so it remains to show that we can define it only in reference to observable quantities under certain assumptions. Standard conditioning arguments show that $\psi=E[E[Y|\mathcal{A}=1,W]-E[Y|\mathcal{A}=0,W]]$ identifies the causal effect as long as the following assumptions hold: 1. 1. Conditional Randomization: $A\perp Y(a)\mid W$ for all $a$ 2. 2. Positivity: $P(\mathcal{A}=1|W)>0$ for all $w$ Our identification result shows that we can get at the causal ARE by estimating an observable “ATE” under certain conditions. Our goal is now to show how to efficiently estimate the observable ATE without imposing any additional assumptions (e.g. linearity, normality, etc.). While our identification assumptions may not always hold in all applications, we can at least eliminate all model misspecification bias and minimize random variation. Once we’ve established how to estimate the ARE for a fixed region, we’ll turn our attention to the problem of finding a good region $\mathcal{A}$ and lastly how to do that without incurring selection bias in estimating the ARE for that region. ## 3 Estimating ARE with TMLE In the previous sections we established that the causal ARE is equivalent to the observable ATE $E[E[Y|\mathcal{A}=1,W]-E[Y|\mathcal{A}=0,W]]$ under standard identifying assumptions. Therefore to estimate it all we need to do is 1) create a new binary random variable $\mathcal{A}_{i}=1_{\mathcal{A}}(A_{i})$ and 2) proceed as if we were estimating the observable ATE from the observational data structure $(Y,\mathcal{A},W)$. There is an extensive literature on estimating the ATEs from observational data Nichols (2007); Winship and Morgan (1999); Rubin (2006). Using split- sample machine learning we can construct estimators that are provably unbiased (modulo bias from any violations of identifying assumptions), have the minimum possible sampling variance, and which are “doubly robust” Zheng and van der Laan (2010); Zivich and Breskin (2021). Augmented inverse propensity-weighting (AIPW) and targeted maximum likelihood (TMLE) are two established estimation approaches that accomplish these goals. Although they are usually very similar in practice, TMLE is often better for smaller samples Luque-Fernandez et al. (2018); Smith et al. (2022); van der Laan and Rose (2011, 2018) and should generally be preferred. In what follows we use the TMLE estimator of the ATE, which we briefly describe here. The TMLE estimator is inspired by the fact that if we knew the true conditional mean $Q(\mathcal{A},W)=E[Y|\mathcal{A},W]$ we could estimate the ATE with the empirical average $\frac{1}{n}\sum_{i}Q(1,W_{i})-Q(0,W_{i})$. Of course, we do not know $Q$, but we can estimate it by regressing the outcome $Y$ onto the exposure $\mathcal{A}$ and covariates $W$. However a detailed mathematical analysis shows that we incur bias if we use our estimate $\hat{Q}$ instead of the truth. This bias might decrease as sample size increases, but it dominates relative to random variability, making it impossible to establish p-values or confidence intervals. TMLE solves this problem by computing a correction to the regression model $\hat{Q}$ that removes the bias. In other words, it “targets” the estimate $\hat{Q}$ to the parameter of interest (here the ATE). The process is as follows: 1. 1. Use cross-validated ensembles of machine learning algorithms (a “super learner”) to generate estimates of the conditional means of treatment: $\hat{g}(\mathcal{A}=a,W)\approx P(\mathcal{A}=a|W)$ (i.e. propensity score) and outcome: $\hat{Q}(\mathcal{A},W)\approx E[Y|\mathcal{A},W]$ 2. 2. Regress $Y$ (scaled to $[0,1]$) onto the “clever covariate” $H_{i}=\frac{1_{1}(\mathcal{A}_{i})}{\hat{g}(1,W_{i})}-\frac{1_{0}(\mathcal{A}_{i})}{\hat{g}(0,W_{i})}$ using a logistic regression with a fixed offset term $\text{logit}(\hat{Q}(\mathcal{A},X))$. The (rescaled) output of this is our targeted regression model $\hat{Q}^{*}$ 3. 3. Compute the plug-in estimate using the targeted model: $\hat{\psi}=\frac{1}{n}\sum_{i}\hat{Q}^{*}(1,W_{i})-\hat{Q}^{*}(0,W_{i})$ An estimated standard error for $\hat{\psi}$ is given by $\hat{\sigma}^{2}=\frac{1}{n^{2}}\sum_{i}\left[\left(\frac{1_{1}(\mathcal{A}_{i})}{\hat{g}(1,W_{i})}-\frac{1_{0}(\mathcal{A}_{i})}{\hat{g}(0,W_{i})}\right)\bigg{(}Y_{i}-\hat{Q}^{*}(\mathcal{A}_{i},W_{i})\bigg{)}+\bigg{(}\hat{Q}^{*}(1,W_{i})-\hat{Q}^{*}(0,W_{i})\bigg{)}-\hat{\psi}\right]^{2}$ with corresponding 95% confidence interval $\hat{\psi}\pm 1.96\hat{\sigma}$. Explaining why the targeting step takes the form of a logistic regression and how the estimated standard error is derived are beyond the scope of this work. van der Laan and Rubin (2006); van der Laan and Rose (2011, 2018) offer explanations targeted to audiences with varying levels of mathematical sophistication. To obtain these estimates we need only to specify the ensemble of machine learning algorithms used to estimate the propensity and initial outcome regressions $\hat{g}$ and $\hat{Q}$. The theoretical guarantees hold as long as a sufficiently rich library is chosen. For estimating the ARE, we must also specify the region $\mathcal{A}$ so that we can compute our binary “exposure” variable. The issue of course is that we have been treating $\mathcal{A}$ as a known region, whereas in many applications the important question is figuring out what guidelines to impose in the first place. This is the focus of the next section. ## 4 Defining the Target Region Thus far we have not focused on how we define the target region $\mathcal{A}$. First, let’s think of $\mathcal{A}$ as nonparametrically defined as the maximizer of some criterion, independent of an estimator. We can think of this as any region on the exposure gradient that maximizes the outcome (can take any shape). However, such a region isn’t interpretable. Therefore, it is easier to constrain the optimization so $\mathcal{A}$ is a rectangle in the exposure space. This is because these sections can be easily described using $\geq$ and $\leq$ rules. This is also important from a public health standpoint where these rules effectively are thresholds of exposures found to have the most severe (or least severe) affects. For this purpose, regression trees are an ideal estimator to get at such a region. Each decision tree algorithm uses some objective function to split a node into two or more sub- nodes. Of course, it is generally impossible to know a priori which learner will perform best for a given prediction problem and data set. Decision trees have many hyper-parameters such as the maximum depth, minimum samples in a leaf, and criteria for splitting amongst others. As such, we need to find the decision tree estimator that best fits the data given a set of nodes. We do this by creating a library of decision tree estimators to be applied to the exposure data and use cross-validation to select nodes based on the best fitting decision tree. This CV selection of the best fitting decision tree algorithm defines our exposure Super Learner $f(A)$ in our additive semi- parametric model $E(Y|A,W)=f(A)+h(W)$. This additive model is needed because we are interested in finding regions that maximize an outcome within only an exposure space not including the baseline covariates. ### 4.1 Discovering Regions in Multiple Exposures using Ensemble Trees To discover regions in multiple exposures and therefore discover interactions in the exposure space, we use predictive learning via rule ensembles Friedman and Popescu (2008). Thus, as part of the data-adaptive procedure the $f(A)$ is a regression model constructed as a linear combinations of simple rules derived from the exposure data. Each rule consists of a conjunction of a small number of simple statements concerning the values of individual input exposures. Machine learning using rule ensembles have not only been shown to have predictive accuracy comparable to the best methods but also result in a linear combination of interpretable rules. Prediction rules used in the ensemble are logical if [conditions] then [prediction] statements, which in our case the conditions are regions in the exposure space that are predictive of the outcome. Learning ensembles have the structure: $F(x)=a_{0}+\sum_{m=1}^{M}a_{m}f_{m}(x)$ where $M$ is the size of the ensemble (total number of trees) and each ensemble member $f_{m}(x)$ is a different function of the input exposures $A$ derived from the training data in the cross-estimation procedure (discussed later). Ensemble predictions from $F(x)$ are derived from a linear combination of the predictions of each ensemble member with ${a_{m}}_{0}^{M}$ being the parameters specifying the linear combination. Given a set of base learners, trees constructed using the exposures, ${{f_{m}(x)}_{1}^{M}}$ the parameters for the linear combination are obtained by a regularized linear regression using the training data. Ideally, each tree in the ensemble is limited to including only 2-3 exposures at a time which enhances interpretability. For instance, given the noise and small sample size in most public health studies, it is unlikely that signal is strong enough to detect interactions with 4 or more variables. Not only that but trees with partitions across many variables become less interpretable. Therefore, in our case we are interested in using an ensemble algorithm that creates a linear combination of smaller trees but also shows optimal prediction performance. To accomplish this goal we use the PRE package Fokkema (2020a) which is similar to the original RuleFit algorithm Friedman and Popescu (2008) with some enhancements including 1. unbiased recursive partitioning algorithms, 2. complete implementation in R, 3. capacity to handle many outcome types and 4. includes a random forest approach to generating prediction rules in addition to bagging and boosting methods. Here, the package PRE is fit to the exposure data in the training sample. Mechanically the procedure is 1. generate an ensemble of trees using exposure data, 2. fit a lasso regression using these trees to predict the outcome, 3. extract the tree basis with nonzero coefficients, 4. store these basis as rules which when evaluated on the exposure space demarcate an exposure region $\mathcal{A}$. There may then be many exposure regions such as $\mathcal{A}_{X_{1},X_{2}}$ which is a region including exposures $X_{1},X_{2}$ or another region in the exposure space which uses variables $X_{4},X_{5}$, $\mathcal{A}_{X_{4},X_{5}}$ etc. The ARE is then calculated for each of these regions which are based off of trees found to be predictive in the ensemble. In the case that multiple trees are included in the ensemble which are composed of the same set of exposures, we select the tree with the largest coefficient. This procedure is done in each fold of the cross- validation procedure. ### 4.2 Discovering Regions in Single Exposures using Decision Trees In addition to finding interactions in the exposure space, the analyst may also be interested in identifying what exposures have a marginal impact and at what levels the outcome changes the most in these exposures. To answer this question we include a marginal tree fitting procedure which is very similar to the method described in 4.1. Here, $f(A)$ is a Super Learner of decision trees fit onto one exposure at a time. We then extract the rules determined from the best fitting tree. Each terminal leaf demarcates a region in the exposure and thus similarly we may have several $\mathcal{A}$ for 1 exposure which are the regions found when creating the partitions which best explains the outcome. Here, rather than calculating an ARE for each region we calculate an ARE comparing each region to the reference region. The reference region is defined as the region that captures the lowest values of $A$. For example, consider our resulting decision tree when fit to variable $A_{1}$ resulted in terminal leaves $A_{1}<0.6$, $A_{1}>0.6$ & $A_{1}<0.9$, $A_{1}>0.9$, in this case the reference region would be $A_{1}<0.6$. We would then have two ARE estimates for the two regions above the reference region. Mechanically, in the training sample, we find the best fitting decision tree which finds partitions in one exposure that best explains an outcome, these rules are evaluated as if statements on the exposure to create $\mathcal{A}_{i}=1_{\mathcal{A}}(A_{i})$ for the respective exposure. We then subset the reference level out and row bind it with each region above the reference region and pass that data to our estimators of the ARE. This approach was chosen to give users a dose-response type estimate for data-adaptively determined thresholds in the univariate exposure space. ### 4.3 Iterative Backfitting We need an algorithm that will allow us to fit $f(A)$ while controlling for $W$ but not including $W$ in the partitions (trees). As such, we iteratively backfit two Super Learners $f(A)$ a Super Learner of decision trees and $h(W)$ an unrestricted Super Learner applied to the covariates. However, both algorithms need to use the same convergence criteria (here maximum likelihood estimation). Thus, $f(A)$ uses an ensemble of regression trees and $h(W)$ uses an ensemble of flexible MLE based algorithms (MARS, elastic net, Highly Adaptive Lasso amongst others). The algorithm first initializes by getting predictions from $f(A)$ and $h(W)$, that is, simply fitting a Super Learner to the exposures and covariates separately and then getting predictions. Then we begin fitting each algorithm offset by the predictions of the other. So at iteration 1 we fit $f(A,\text{offset}=h(W)_{\text{iter 0}})$ and likewise $h(W,\text{offset}=f(A)_{\text{iter 0}})$; where the offsets are predictions of the models fit individually (without offset at iteration 0). The predictions of these models without an offset then gives us $f(A)_{\text{iter 1}}$ and $h(W)_{\text{iter 1}}$. These predictions are then used as offsets at iteration 2. This process continues until convergence where convergence is is defined as the absolute mean difference between the two models being less than some very small number $\delta$ where $\delta$ by default is 0.001. In this way, for both where $A$ in $f(A)$ is a vector of exposures (resulting rules include combinations of different exposure levels) and when $A$ is a single exposure, we are able to identify cut-points in the exposure space while controlling for $W$ in the additive model that converges in maximum likelihood. We evaluate the best fitting decision tree onto the exposure space which results in an indicator of the exposure region and calculate our k-fold specific and pooled target parameters give this region. ### 4.4 Convergence to the True Region The cross-validation selector is simply the recursive partitioning algorithm which performed best in terms of cross-validated risk within the parameter- generating sample. Assuming that a partition $\\{\mathcal{D}_{d}\\}d=1,...,D$ of the space $A=A_{1},...,A_{Z}$ exists with $D$ cells (or segments) such that the $\mathcal{D}_{d}$ cells are used to define the $\mathcal{A}$ boundaries, then the oracle selector as defined in Theorem 2 of van der Laan et al. (2008) is the estimator, among the decision tree learners considered, which minimizes risk under the true data-generating distribution $P_{0}$. That is, the oracle selector is the best possible estimator given the set of candidate decision tree learners considered; however, it of course depends on both the observed data and $P_{0}$, and thus is unknown. Theorem 2 in van der Laan et al. (2008) shows that the Super Learner performs as well (in terms of expected risk difference) as the oracle selector, up to a typically second order term. Therefore, as long as the number of candidate decision tree learners considered is polynomial in sample size ($n^{x}$), the Super Learner is the optimal learner. Given that the Super Learner performs asymptotically as well (in the risk difference sense) as the oracle selector, which chooses the best of the decision tree candidate learners, and given that, this class of learners are restricted to algorithms which partition the mixture space into $\mathcal{D}_{d}$ nodes of finite depth, it follows then that the selection of $\mathcal{D}_{d}$ nodes used in the decision tree estimator selected by the Super Learner converges asymptotically to the $\mathcal{D}_{d}$ nodes used by the oracle selector under depth constraints. Here the finite tree depth is needed because without tree depth limits, as sample size increases the best fitting estimator will always be the decision tree with increased depth and therefore there is no convergence to some true set of $\mathcal{D}_{d}$ nodes. Likewise, for interpretability, we are interested in some small set of nodes, or thresholds, per variable that is informative for public policy. That is to say, we may limit the ensemble of decision trees to a depth of 3 for interpretability. For example, we may want a more concise rule such as, "if arsenic is greater than X and lead is greater than Y", or without monotonic assumptions, "if arsenic is between X1 and X2 and lead is between Y1 and Y2". Overall, under functional averaging theory developed and additive model assumptions we can perhaps say there is some convergence to a true region which best differentiates the values of the outcome as sample size goes to infinity. That being said, we show simulations in the next section to measure how good CVtreeMLE is at identifying the true ARE and true exposure region built into a data-generating process as sample size increases. ## 5 K-fold Cross-Estimation Of course, the mixture region used to estimate the ARE is not defined a priori. If we were to use the same data to both identify the region and make the ARE our estimates will be biased. Thus, for desirable asymptotic properties to hold without additional assumptions, we need our conditional means to be cross-estimated from the observed data. We split the data into $P_{n_{-k}}$ (parameter-generating) and $P_{n_{k}}$ (estimation) samples. These splits or folds are part of a k-fold cross-validation framework. K-fold cross-validation involves: (i) ${1,...,n}$, observations, is divided into $K$ equal size subgroups, (ii) for each $k$, an estimation-sample, notationally $P_{k}$, is defined by the k-th subgroup of size $n/K$, while the parameter- generating sample, $P_{n_{-k}}$, is its complement. In this round robin manner we rotate through our data and thus, in the case of $K=10$ get 10 difference target parameter mappings $\mathcal{A}_{n}$, outcome estimators $Q_{n}$ and propensity estimators $g_{n}$. We want one summary measure of the target parameter found across the folds, such as the average. With $P_{n_{-k}}$ we find thresholds in our exposure space (using the results of a decision tree) which designates exposure region. Then given this exposure region using the same $P_{n_{-k}}$ we train our $g_{n}$ and $Q_{n}$ estimators which are needed for our TMLE update step to debias our initial estimates of the ARE and give us an asymptotically unbiased estimator. We then plug-in our $P_{n_{k}}$ to this unbiased estimator to get our ARE estimate in this estimation sample. Let $\bar{Q}_{n}$ denote a substitution estimator that plugs in the empirical distribution with weight 1/n for each observation which approximates the true conditional mean $\bar{Q}_{0}$ in $P_{0}$, this estimator, in our case is a Super Learner, or ensemble machine learning algorithm, our substitution estimator looks like: $\Psi(Q_{P_{n_{k}}})=\frac{1}{V}\sum_{v=1}^{V}{\bar{Q}_{n_{-k}}(\mathcal{A}_{n_{-k}}=1,W_{v})-\bar{Q}_{n_{-k}}(\mathcal{A}_{n_{-k}}=0,W_{v})}$ Let’s focus first on the $k$ subscripts, we split data into $k\in 1...K$ non- overlapping folds and fit $K$ different models. Thus, $\bar{Q}_{n_{-k}}$ denotes our outcome regression function fit when excluding the data for fold $k$. $P_{n_{k}}$ denotes our estimation-data and $P_{n_{-k}}$ is the parameter-generating sample, that is, our parameter-generating sample is used to train our estimators and then we pass our estimation-data in to get estimates. $\bar{Q}_{n_{-k}}$ then, in our case, is a Super Learner fit using the parameter-generating data. Likewise, $\mathcal{A}_{n_{-k}}$ is a decision tree fit using the parameter-generating data. $\Psi(Q_{P_{n_{k}}})$ then indicates that we pass the estimation-sample data into our estimators trained with the parameter-generating data; so here we first fit a decision tree to the exposure space of the parameter-generating data, then apply the rules found to the estimation-sample data to create an exposure region indicator. Then using this exposure and the estimation-sample covariates, we feed this into the outcome regression model trained on the parameter-sample data. We then get predicted outcomes under different counterfactuals for a data- adaptively determined exposure using our estimation-sample data. Our cross- estimated TMLE estimator for this data-adaptively defined exposure produces an unbiased, efficient substitution estimator of target parameters of a data- generating distribution we are interested in. This estimator looks like: $\Psi(Q_{P_{n_{k}}}^{\star})=\frac{1}{V}\sum_{v=1}^{V}\\{\bar{Q}_{n_{-k}}^{\star}(\mathcal{A}_{n_{-k}}=1,W_{v})-\bar{Q}_{n_{-k}}^{\star}(\mathcal{A}_{n_{-k}}=0,W_{v})\\}$ Here we can see the only change to our above equation is $\bar{Q}^{\star}$ which is the TMLE augmented estimate. This new function, $f(\bar{Q}_{n_{-k}}^{\star}(A,W))=f(\bar{Q}_{n_{-k}}(A,W))+\epsilon_{n_{-k}}\cdot h_{n_{-k}}(A,W)$, where $f(\cdot)$ is the appropriate link function (e.g., logit), $\epsilon_{n}$ is an estimated coefficient and $h_{n}(A,W)$ is a "clever covariate" which is now cross-estimated. Here what we mean is that, the initial estimates for the estimation-sample using models trained using the parameter-generating data are updated through this so-called, least-favorable submodel. The cross-estimated clever covariate looks like: $h_{n_{-k}}(\mathcal{A},W)=\frac{\mathbbm{I}(\mathcal{A}_{n_{-k}}=1)}{g_{n_{-k}}(\mathcal{A}_{n_{-k}}=1|W)}-\frac{\mathbbm{I}({\mathcal{A}_{n_{-k}}=0)}}{g_{n_{-k}}(\mathcal{A}_{n_{-k}}=0|W)}$ Here, $g_{n_{-k}}(W)=\mathbb{P}(\mathcal{A}_{n_{-k}}=1\mid W)$, the propensity score of the data-adaptively determined exposure region, is being estimated using a Super Learner with the parameter-generating data. That is, in our parameter generating sample we get the exposure region, and an estimator $g_{n}$ we apply this exposure region to the estimation sample and then get predictions for the probability of that exposure region indicator using the estimation sample, we then plug these estimates into the above cross-estimated clever covariate used in the TMLE update. We can see that by using v-fold cross-validation, we can do better than traditional sample splitting as v-fold allows us to make use of the full data which results in tighter confidence intervals because our variance is estimated over the full data. Similarly, our estimate is an average of the v-fold specific estimates: $\Psi_{n}(P)=Ave\\{\Psi_{P_{n_{-k}}}(P)\\}\equiv\frac{1}{V}\sum_{v=1}^{V}\Psi_{P_{n,-k}}(P)$ We do this in a pooled TMLE update manner where we stack the estimation-sample estimates for each nuisance parameter and then do a pooled TMLE update across all the initial estimates using clever covariates across all the folds to get our estimate $\epsilon$ we then update our counterfactuals across all the folds and take the average. More concretely, in each fold we have our initial estimates from that fold from $Q_{n_{-k}}(Y|\mathcal{A},W)$ and the fold specific clever covariate $h_{n_{-k}}(\mathcal{A}|W)$ of length $k$ for a fold specific exposure found using $\mathcal{A}_{n_{-k}}$. We stack all the $Q_{n_{-k}}$’s and $h_{n_{-k}}(\mathcal{A}|W)$’s together along with the outcomes in each validation fold and do our fluctuation step: $f(\bar{Q}_{n}^{\star}(\mathcal{A},W))=f(\bar{Q}_{n}(\mathcal{A},W))+\epsilon_{n}\cdot h_{n}(\mathcal{A},W)$ Notice here the $k$ subscripts are removed, this is because we are using our cross-estimates for all of $n$. Using the $\epsilon$ from this model, we then update the counterfactuals across all the folds and take the difference for our final ARE. In a similar fashion, we use the updated conditional means, counterfactuals, and clever covariates to solve the IC across the whole sample. By pooling the cross-estimates across the folds and then calculating the SE for this pooled IC we are able to derive more narrow confidence intervals compared to if we were to average the IC estimated in each of the folds (because the IC is scaled by $n$ and not $n/K$). This pooled estimate still provides us with proper intervals because all estimates in its construction were cross-estimated. An alternative to this pooled approach is to simply report the k-fold specific estimates of the ARE and fold specific variance estimates for this ARE using the fold specific IC. We do this as well. We do this because, if the exposure region $\mathcal{A}$ identified in each region is highly variable, that is, if the region that that maximizes the difference for sets of exposure variables are very different across the folds, then interpreting the pooled ARE is difficult. By calculating and providing both k-fold specific and pooled results users can investigate how variable a pooled result is across the folds. ### 5.1 Inverse-Variance Method for Combining K-fold Results In addition to the pooled TMLE approach to aggregate k-fold specific data- adaptive target estimates, we also calculate the inverse-variance method (IVM) commonly used in meta-analyses. We call this method the k-fold harmonic mean. Here each fold is given a weight defined as: $w_{k_{i}}=\frac{1}{SE(\hat{\theta}_{k_{i}})^{2}}$ Which is simply an inverse of the standard error such that estimates with smaller SE are given a higher weight. The inverse-variance pooled ARE across the folds is given as: $\hat{\theta}_{IVM}=\frac{\sum w_{k_{i}}\hat{\theta}_{k_{i}}}{\sum w_{k_{i}}}$ And lastly, the pooled SE is calculated as: $SE(\hat{\theta}_{IVM})\frac{1}{\sqrt{\sum w_{k_{i}}}}$ For which confidence intervals and p-values are derived for the pooled IVM estimate. This pooled estimate is given because, in the event of high inconsistency of the k-fold estimates in lower sample size, the confidence intervals from pooled influence curve may not cover the true ARE if the pooled ARE was applied to $P_{0}$. This is because the union rule attached to the pooled ARE is a conservative rule which covers all observations across the folds (discussed later). The IVM derived CIs are wider and provide better coverage in the event of high inconsistency (which we show in simulations). We explain rule stability metrics and establishing a union rule across the folds in the next section. ### 5.2 Defining the Union Region The pooled TMLE ARE is matched with a pooled region that encompasses all the observation indicated by each fold specific regions. We group the trees across the folds according to what variable sets the trees are composed of. That is, a linear combination of tree ensembles is fit to each training sample specific to the fold. There may be variability in where the partition is set for trees with the same variable sets across the folds, or certain ensembles don’t use certain variable sets at all in some folds but used in others. We need a method of creating a pooled region and give stability metrics for how consistently trees with a respective variable set are found in the cross- validation procedure. For this we create a union region. There are other possible ways of pooling the regions, such as averaging the partitions per exposure variable across the folds. Here we choose a conservative approach. This is the union region of the k-fold regions in the sense that, we create a new region that is the OR combination of each k-fold specific tree. For three folds and therefore three partitions say, $X_{1}<2$ & $X_{2}>5$, $X_{1}<2.3$ & $X_{2}>5.2$ and $X_{1}<1.9$ & $X_{2}>5.3$, the union rule is $X_{1}<2$ & $X_{2}>5$ OR $X_{1}<2.3$ & $X_{2}>5.2$ OR $X_{1}<1.9$ & $X_{2}>5.3$ forms the rule: $X_{1}<2.3$ & $X_{2}>5$ because this region covers all the observations indicated in the fold specific regions. For variables where the logic is $>$ we take the minimum value across the folds and likewise for $<$ we take the maximum. This union region is conservative and sensitive to outlier partition points found across the folds and therefore higher $K$ folds will lead to more stable partitions if there is signal in the data. Additionally, the analyst should investigate the fold specific regions to determine the interpretability of the pooled region. If there is high variability or outliers, there may be bias in the TMLE pooled estimate when compared to the expected difference in outcomes if the respective pooled region was applied to the true population $P_{0}$. ### 5.3 Stability Metrics Given a pooled region, we simply give the proportion of folds trees with a respective variable set are found across the folds. For example, consider a study of mixed metals that uses CVtreeMLE and the results across three folds are: 1. lead > 2.2 & arsenic > 1.3, 2. lead > 2.1 & arsenic > 1.2, 2. lead > 2.0 & arsenic > 1.1. Our pooled region is lead > 2.0 & arsenic > 1.1 because this region contains all the fold specific regions. The stability metric here is 100% because a tree with lead and arsenic was found in all three folds. If however, this tree was only found in 2 of three folds, the stability metric is 67%. ## 6 Simulations In this section, we demonstrate using simulations that our approach identifies the correct exposure region which maximizes the difference in conditional means and estimates the correct difference built into a DGP for this region. ### 6.1 Data-Generating Processes Because a two dimensional exposure space is easier to visualize and describe compared to higher dimensional spaces, we start by investigating a squared dose-response relationship between two exposure variables where an interaction occurs between the exposures when each meets a particular threshold value. We extend simulations to the three dimensional case. In both 2-D and 3-D exposure simulations there are specific outcome values generated for each subspace of the mixture based on split points $\mathcal{D}_{d}$ but there exists one region with the maximum outcome (the truth that we want). In both scenarios, the goal is to determine if our data adaptive target parameter is targeting the region that maximizes the conditional mean outcome for the given sample and evaluate how CVtreeMLE approaches this desired oracle parameter as sample size increases. To meet this goal, we construct a data-generating process (DGP) where $Y$ is generated from a tree-structured covariate-adjusted relationship of a mixture consisting of components, $A_{1},A_{2},A_{3},A_{n}$. That is, generally in each simulation we generate exposure regions, where the density of the region is driven by covariates and there is one region that has the maximum difference compared to outside the respective region. More details for each simulation are given below. #### 6.1.1 Two-Dimensional Exposure Simulations This DGP has the following characteristics, $O=(W,A,Y)$. $W$ are three baseline covariates $W_{1}\sim\mathcal{N}(\mu=37,\sigma=3),W_{2}\sim\mathcal{N}(\mu=20,\sigma=1),W_{3}\sim\mathcal{B}(\mu=0.5)$ Where $\mathcal{B}$ is a Bernoulli distribution and $\mathcal{N}$ is normal. These distributions and values were chosen to represent a study with covariates for age, BMI and sex. Our generated exposures were likewise created to represent a chemical exposure quantized into 5 discrete levels. The values and range of the outcome were chosen to represent common environmental health outcomes such as telomere length or epigenetic expression. We are interested in sampling observations into a 2-dimensional exposure grid. Here a $5\times 5$ grid is based on combinations of two discrete exposure levels with values 1-5. We want the number of observations in each of these cells to be affected by covariates. To do this we define a conditional categorical distribution $P\\{(A_{1},A_{2})=(a_{1},a_{2})|W=w\\}$ and sample from it. $P\\{(A_{1},A_{2})=(a_{k},a_{l})|W\\}=\frac{e^{W^{\top}\beta_{k,l}}}{1+\sum_{k,l}e^{W^{\top}\beta_{k,l}}}$ Here the $\beta$’s attached to each covariate were drawn from a normal distribution with means 0.3, 0.4, 0.5 and 0.5 respectively all with a standard deviation of 2. This then gives us 25 unique exposure regions with densities dependent on the covariates. We then want to assign an outcome in each of these regions based on main effects and interactions between the exposures. We use the relationship $Y=0.2A_{1}^{2}+0.5A_{1}A_{2}+0.5A_{2}^{2}+0.2*\text{age}+0.4*\text{sex}+\epsilon(0,0.1)$ Which indicates there is a slightly weaker squared effect for $A_{2}$ relative to $A_{1}$ and a strong interaction between the exposures and confounding due to age and sex. The resulting data distribution and generating process is shown in Figure 1. Figure 1: 2D Exposure Simulation Of course, it is also possible to explore other dose-response relationships (such as logarithmic) by changing the coefficient matrix. ##### Computing Ground Truth The fact that our exposures are discrete in this simulation lets us easily compute the ground-truth ARE for any region $\mathcal{A}$ because we can explicitly compute the conditional mean function $m$ $\displaystyle m(\mathcal{A}=1,w)$ $\displaystyle=\int_{a\in\mathcal{A}}\mu(a,w)\frac{p_{A|W}(a,w)}{\pi_{\mathcal{A}}(w)}\ da$ $\displaystyle=\frac{\sum_{a\in\mathcal{A}}\mu(a,w)p_{A|W}(a,w)}{\sum_{a\in\mathcal{A}}p_{A|W}(a)}$ Therefore to approximate the ARE to arbitrary precision we can 1. 1. Sample a large number of times (e.g. $b=100,000$) from the covariate distribution to obtain $W_{\\{1,\dots i,\dots b\\}}$. 2. 2. Compute the values $m(\mathcal{A}=1,W_{i})$ using the above formula. This is possible because the functions $\mu$ and $p{A|W}$ are known for the data- generating process 222If the exposure space were not discrete, this step would require numerical approximation of an integral for each different value of $w$ which would be generally impractical.. In a similar fashion compute $m(\mathcal{A}=0,w)$. 3. 3. Compute $\text{ARE}(\mathcal{A})=\frac{1}{b}\sum_{i}^{b}m(1,W_{i})-m(0,W_{i})$. #### 6.1.2 Three-Dimensional Exposure Simulations This DGS has the same general structure, $O=W,A,Y$. $W$ and baseline covariates $W_{1}=\mathcal{N}(\mu=37,\sigma=3),W_{2}=\mathcal{N}(\mu=20,\sigma=1),W_{3}=\mathcal{B}(\mu=0.5)$ In this 3D simulation we are interested in keeping the exposures continuous as this is more realistic compared to the 2D simulation. Here $A$ are three continuous mixtures from a multivariate normal distribution: $\displaystyle\begin{pmatrix}A_{1}\\\ A_{2}\\\ A_{3}\end{pmatrix}$ $\displaystyle\sim N\begin{bmatrix}\begin{pmatrix}0\\\ 0\\\ 0\\\ \end{pmatrix}\\!\\!,&\begin{pmatrix}1.0&0.5&0.8\\\ 0.5&1.0&0.7\\\ 0.8&0.7&1.0\end{pmatrix}\end{bmatrix}$ We assign one partition point value to each exposure which creates 8 possible regions in the $2x2x2$ 3D grid for which we want to assign outcomes. Just as the first simulation we want the number of observations in each of the cells in the "mixture cube" to be affected by covariates. To do this we define a conditional categorical distribution $P\\{(A_{1},A_{2},A_{3})=(a_{1},a_{2},a_{3})|W=w\\}$ and sample from it. $P\\{(A_{1},A_{2},A_{3})=(a_{k},a_{l},a_{j})|W_{i}\\}=\frac{e^{W_{i}^{\top}\beta_{k,l,j}}}{1+\sum_{k,l,j}e^{W_{i}^{\top}\beta_{k,l,j}}}$ For each of these categories which defines a region in the exposure space we need to assign exposure values while also preserving the local correlation structure within that region. To do this, we convert the cumulative distribution function of the exposures to a uniform distribution then back transform this uniform distribution to the original exposure distribution with bounds for each exposure region. So for instance, in the region where each exposure is less than each threshold value, we back transform the uniform distribution with the minimum value set as the minimum for each exposure and max as the partition value for each exposure. These values then are attached to the categorical variables generated which represent the mixture region. This then generates continuous exposure values with a correlation structure in each region. The outcome $Y$ is then generated via a linear regression of the form: $Y=\beta_{0}+\beta_{1}\mathbbm{1}(A=1)+\beta_{2}\mathbbm{1}(A=2)+...+\beta_{7}\mathbbm{1}(A=7)+\beta_{W_{1}}W_{1}+\beta_{W_{2}}W_{2}+\epsilon,\epsilon~{}N(0,\sigma)$ Where the $\beta_{j}$ are chosen so some mixture groups have a high mean, some have a low mean and $\mathbbm{1}$ represent indicators of each of the possible 8 regions. Thus, the outcome in each region of the mixture cube is determined by the $\beta$ assigned to that region. Given this formulation of a DGP it is possible to then generate $Y$ by shifting the drivers or "hot spots" around the mixture space, thereby simulating possible agonist and antagonistic relationships. We could assign something like $\beta_{2}=2$ with all other regions having a $\beta_{\neq 2}=0$. This then would mean the ARE in the true DGP is 2. Likewise we could assign $\beta$’s in each region in which case the truth by our definition is the region with the max ARE. The process for this DGP is shown in Figure 2. Figure 2: 3D Exposure Simulation Overall, our 3D example is very similar to the 2D exposure simulation but we aim to test CVtreeMLE in identifying thresholds used to generate an outcome in a space of three continuous exposures. Also, because we keep the space of possible outcomes relatively simple here, we simply generate individual outcomes for each mixture subspace. This allows us to create situations where only one region drives the outcome while the complementary space is 0 or there is an outcome in each region and we are interested in identifying the region with the maximum outcome. In each simulation we are interested in the bias/variance of our estimates compared to the truth, the bias of our rule compared to the true rule and the bias of our data-adaptive rule compared to the expected ARE if that rule was applied to the true population. We discuss this next. ##### Computing Ground Truth Previously, in the discrete exposure case, we could directly estimate ground truth by inverse weighting given the summed probability in the exposure region multiplied by the outcome. This is not possible in the continuous case. To make things simpler, we z-score standardize the covariates so the mean of each covariate is 0. Therefore we can directly compute the mean in the region indicated by the ground-truth rule and the mean outcome in the complementary space and take the difference. This is the same as the max coefficient minus the mean of the other coefficients in the linear model, this is the true ARE. ### 6.2 Evaluating Performance The following steps breakdown how each simulation was tested to determine 1. asymptotic convergence to the true mixture region used in the DGP, 2. convergence to the true ARE based on this true region and 3. convergence to the true data-adaptive ARE, that is CVtreeMLE’s ability to correctly estimate the ARE if the data-determined rule was applied to the population. We do this by: 1. 1. To approximate $P_{0}$, we draw a very large sample (500,000) from the above described DGP. 2. 2. We then generate a random sample from this DGP of size $n$ which is broken into $K$ equal size estimation samples of size $n_{k}=n/K$ with corresponding parameter generating samples of size $n-n/K$. 3. 3. At each iteration the parameter generating fold defines the region and is used to create the necessary estimators. The estimation fold is used to get our TMLE updated causal parameter estimate, we then do this for all folds. 4. 4. For an iteration, we output the ARE estimates given pooled TMLE, k-fold specific TMLE and the harmonic mean. The region identified in the fold is applied to the large sample $P_{0}$ to estimate the data-adaptive bias. Likewise, each estimate is compared to the ground-truth ARE and region. For each iteration we calculate metrics for bias, variance, MSE, CI coverage, and confusion table metrics for the true maximal region compared to the estimated region. For each type of estimate (pooled TMLE, k-fold specific TMLE estimates, and harmonic mean) we have bias when comparing our estimate to 1. the ARE based on the true region in the DGP that maximizes the mean difference and 2. the ARE when the data-adaptively determined region is applied to the population. Therefore, when comparing to the true "oracle" region ARE we have: 1. 1. $\psi^{0}_{\text{pooled tmle bias}}$: This is the bias of the pooled TMLE ARE compared to the ground-truth ARE for the true region built into the DGP which maximizes the mean difference in adjusted outcomes. 2. 2. $\psi^{0}_{\text{mean v-fold tmle bias}}$: This is the bias of the mean k-fold specific AREs compared to the ground-truth ARE for the true region built into the DGP. 3. 3. $\psi^{0}_{\text{harmonic mean v-fold tmle bias}}$ This is the bias of the harmonic mean of k-fold specific AREs compared to the ground-truth ARE for the true region built into the DGP. The above bias metrics are each compared to the true ARE for the oracle region in the DGP. We are also interested in the ARE if the data-adaptively determined region, the region estimated to maximizes the difference in outcomes in the sample data, were applied to $P_{0}$ the true population. Therefore, there are also bias estimates for: 1. 1. $\psi^{DA}_{\text{pooled tmle bias}}$: This is the bias of the pooled TMLE ARE compared to the ARE of the union region across the folds applied to $P_{0}$. 2. 2. $\psi^{DA}_{\text{mean v-fold tmle bias}}$: This is the bias of the mean k-fold specific AREs compared to the mean ARE when all the k-fold specific rules are applied to $P_{0}$. 3. 3. $\psi^{DA}_{\text{harmonic mean k-fold tmle bias}}$ This is the bias of the harmonic mean of k-fold specific AREs compared to the ARE of the union region across the folds applied to $P_{0}$. We multiple each bias estimate by $\sqrt{n}$ to ensure the rate of convergence is at or faster than $\sqrt{n}$. For each ARE estimate we calculate the variance and subsequently the mean-square error as: $\text{MSE}=\text{bias}^{2}+\text{variance}$. MSE estimates were also multiplied by $n$. For each ARE estimate we calculate the confidence interval coverage of the true ARE parameter givent the oracle region and the ARE given the data-adaptively determined region applied to $P_{0}$. For the TMLE pooled estimates these are lower and upper confidence intervals based on the pooled influence curve. For the k-fold specific coverage, we take the mean lower and upper bounds. For the harmonic pooled coverage, we calculate confidence intervals from the pooled standard error. In each case, we check to see if the ground-truth rule ATE and data-adaptive rule ATE are within the interval. Lastly, we compare the data-adaptively identified region to the ground-truth region using the confusion table metrics for true positive, true negative, false positive and false negative to determine whether, as sample size increases, we converge to the true region. These performance metrics were calculated at each iteration, where 50 iterations were done for each sample size $n=$ (200, 350, 500, 750, 1000, 1500, 2000, 3000, 5000). It was ensured that, for each data sample, at least one observation existed in the ground-truth region to ensure confusion table estimates could be calculated. CVtreeMLE was run with 5 fold CV (to speed up calculations in the simulations) with default learner stacks for each nuisance parameter and data-adaptive parameter. Our data-adaptive parameter for interactions was the tree with the max ARE (positive coefficient) for each variable set in the ensemble. ### 6.3 Default Estimators As discussed, CVtreeMLE needs estimators for $\bar{Q}=E(Y|A,W)$ and $g_{n}=P(A|W)$. CVtreeMLE has built in default algorithms to be used in a Super Learner van der Laan Mark et al. (2007) that are fast and flexible. These include random forest, general linear models, elastic net, and xgboost. These are used to create Super Learners for both $\bar{Q}$ and $g_{n}$. CVtreeMLE also comes with a default tree ensemble which is fit to the exposures during the iterative backfitting procedure. These trees are built from the partykit package Hothorn and Zeileis (2015) in R. By default we include 7 trees in the tree Super Learner that have various levels for the hyper-parameters alpha (p-value to partition on), max-depth (maximum depth of the tree), bonferroni correction (whether to adjust alpha by bonferroni) and min-size (minimum number of observations in terminal leaves). These trees are used during the iterative backfitting in estimating partitions for each individual exposure. For the rule ensemble, the predictive rule ensemble package (pre) Fokkema (2020b) is used with default settings and 10-fold cross- validation. Users can pass in their own libraries for these nuisance and data- adaptive parameters. For these simulations, we use these default estimators in each Super Learner. ### 6.4 Results #### 6.4.1 CVtreeMLE Algorithm Identifies the True Region with Maximum ARE First we describe results for identifying the true region built into the DGP. It is obviously necessary for this to converge to the truth as sample size increases in order for the $\psi^{0}$ estimates to be asymptotically unbiased. Overall we find the tree algorithm identifies the true region in the DGP and therefore provides results which have high-value for treatment policies. Figure 3 shows metrics comparing observations covered by the estimated pooled region to those indicated by the true region in the DGP for two discrete exposures. From this figure it can be seen that, at around 1500 observations, the pooled region is the true region. Figure 4 shows the confusion table metrics comparing the data-adaptive pooled region to the oracle region in the three continuous exposure scenario. As sample size increases, the false positives approach 0 which is what we would desire in this continuous case. From this, we see that in both instances of discrete and continuous exposures, CVtreeMLE is able to identify the correct region in the exposure data which has the maximum ARE. There is some small disparity in the discovered region compared to the truth in the continuous case, this is because for false- positives to perfectly match the true region, the tree search algorithm must identify the exact set of continuous digits that delineate the region which is very difficult. In our case, this region is is $M_{3}$ <= 2.5 & $M_{1}$ >= 0.99 & $M_{2}$ >= 2.0. In this three exposure case there is antagonism of $M_{3}$. Given the exposures are continuous, it is likely that the tree search algorithm gets very close but not absolutely exact to these boundaries. In the two exposure case where the exposures were discretized finding the boundaries is easier. As such, our future evaluation is focused more on the data-adaptive estimates (comparing estimates to the ARE given applying the data-adaptive rule to $P_{0}$). Ultimately, the data-adaptive target parameter theory only holds for the data-adaptive parameter and not the parameter given an oracle rule; however, we include both again to investigate how CVtreeMLE approaches the oracle rule as sample size increases. Figure 3: 2D Exposure Confusion Table Metrics of Rule Coverage Figure 4: Three Exposure Confusion Table Metrics of Rule Coverage #### 6.4.2 CVtreeMLE Unbiasedly Estimates the Data-Adaptive Parameter Looking at the bias for the ARE estimate given two discrete exposures compared to the data-adaptively discovered region applied to $P_{0}$ TMLE unbiasedly estimates the data-adaptive parameter at root $n$ rates with good coverage. Below, Figure 5 A shows the data-adaptive rule ARE bias ($\psi^{DA}_{\text{pooled tmle bias}}$, $\psi^{DA}_{\text{mean k-fold tmle bias}}$, $\psi^{DA}_{\text{harmonic mean k-fold tmle bias}}$) and MSE (B). Figure 5: 2D Exposure Bias and MSE In Figure 5 A the data-adaptive rule ARE bias is larger for the pooled estimates (pooled TMLE ARE and harmonic mean ARE compared to ARE if the pooled rule was applied to $P_{0}$) compared to the average folds bias (mean k-fold ARE compared to the mean of each k-fold rule applied to $P_{0}$). This is because inconsistent rule estimates in lower sample sizes can bias the pooled ARE compared to the pooled region. Consider a 3-fold situation where for variables $X_{1}$ and $X_{2}$ the region was designated by $X_{1}>4\text{ \& }X_{2}>4$, $X_{1}>4\text{ \& }X_{2}>4$, and $X_{1}>2\text{ \& }X_{2}>4$; because $X_{1}>2\text{ \& }X_{2}>4$ is found in one of the folds, this is the pooled region (as it covers observations for $X_{1}>4$) and thus (if the true ARE for $X_{1}>4\text{ \& }X_{2}>4$ applied to $P_{0}$) is higher, our pooled results would be biased to this higher ARE because two of three of our folds have an ARE for this region. This bias converges to average k-fold bias at a sample size of 1500. Effectively, once the trees across the folds stabilizes there is less bias in the pooled estimate compared to the pooled region ATE. This similar pattern is reflected in the pooled estimates MSE (given higher bias in smaller samples). For the user, this indicates that, in smaller sample sizes ($n$ < 1000) the analyst should look at fold specific results to ensure the trees are close in the cut-off values in order to interpret the pooled result. If not, k-fold specific results should be reported as these show very low bias/MSE even in smaller sample sizes. The bias and MSE for all estimates compared to the ground-truth rule ATE show an $1/sqrt(n)$ reduction as sample size increases. In sum, as sample size increases the bias for all estimates converge to 0 which which is necessary for our estimator to have valid confidence intervals. Figure 6: 3D Exposure Bias and MSE Figure 6 A and C likewise show the asymptotic bias in the three continuous exposure case. All estimates show bias decreasing when evaluated against the ARE when the data-adaptively determined region is applied to $P_{0}$; however, these estimates do not go to 0 exactly (at max sample size equal to 5000) as the data-adaptive rule is still not exactly the true rule, which is expected. Figure 7 A shows the confidence interval coverage for each estimate compared to the data-adaptive region applied to $P_{0}$. Figure 7: 2D Exposure Confidence Interval Coverage For coverage of the ARE of the pooled rule applied to $P_{0}$, the CIs calculated from the pooled k-fold standard errors showed coverage between 95% - 100%. The pooled TMLE CIs showed poorer coverage at lower sample sizes, this is likely due to the bias of the pooled ARE estimate compared to the pooled region applied to $P_{0}$ paired with the more narrow confidence intervals calculated across the full sample. The harmonic pooled k-fold CIs were wider and thus covered the truth in this pooled setting. Coverage for the k-fold specific CIs were almost always at or above 90% and converge to 95% at higher sample sizes. Figure 8: 3D Exposure CI Coverage Figure 8 A shows the CI coverage of the data-adaptive rule in the three exposures simulation. As expected, the average k-fold CI converges to 95%. The pooled estimates are lower given the conservative pooled rule. #### 6.4.3 CVtreeMLE Unbiasedly Estimates the Oracle Target Parameter Now we look at comparing estimates to the true ARE given the oracle region in the DGP. Figure 5 B and D show the bias and MSE for this comparison in the two discrete exposures. As can be seen, both decrease at root $n$ rate for all estimates. Figure 6 B and D likewise show this same rate of convergence for the three continuous exposure case. Based on these simulations, CVtreeMLE unbiasedly estimates the oracle target parameter at root $n$ rates. We next look at the coverage. Figure 7 B shows coverage of the true ARE given the true region. The CIs calculated from the harmonic pooled k-fold standard errors had consistent 95% coverage, the k-fold specific CIs converged to 95% when sample sizes reached 1500 and the pooled TMLE CIs converged to 75% coverage. The same is shown for the three continuous exposures in Figure 8 B the inverse variance CI converges to 95% for coverage of the true ATE with the mean k-fold slightly lower around 82%. Table 2 gives the bias, SD, MSE, and coverage for sample sizes 200, 1000, and 5000, comparing estimates to the data-adaptive truth. #### 6.4.4 CVtreeMLE has a Normal Sampling Distribution for Valid Inference For our estimator to have valid inference, we must ensure that the estimator has a normal sampling distribution centered at 0 that gets more narrow as sample size increases. To confirm this, we next examine the empirical distribution of the standardized differences, ($\psi_{n}$ \- $\psi^{0}$)/SE($\psi_{n}$), this is the ARE estimate bias compared to the true ARE given the true region divided by the standard error of the estimates over the iterations and $\psi_{n}$ \- $\psi^{DA}$)/SE($\psi_{n}$) which is the same standardized difference but compared to the resulting ARE when the data- adaptive region is applied to $P_{0}$. Figure 9 shows the sampling distribution for each sample size with 50 iterations per sample size to estimate the probability density distribution of the standardized bias compared to the data-adaptive ARE. We see convergence to a mean 0 normal sampling distribution as sample size increases for all estimates. Figure 9 A shows the sampling distribution of the standardized bias of the mean k-fold AREs compared to the ground-truth ARE. We can see that this sampling distribution is quite tight around 0. Figures 9 B and C show the sampling distribution for the harmonic mean and the pooled TMLE estimates which are mirror reflections of each other. For both estimates, lower sample sizes (such as in purple $n$ = 200) there is a wider spread of bias (estimates vary more widely) with z-scores out to 2 or 4 but this distribution gets tighter as sample size increases. Figure 9: Bias Standardized by Standard Error Compared to ATE of Data-Adaptive Rule Likewise, Figure 10 A-C show the standardized bias of each estimate compared to the ground-truth region ARE. All estimates generally follow the same distribution and converge to a 0 mean normal distribution as sample size increases. N | Absolute Bias | SD | MSE | Coverage ---|---|---|---|--- 200 | 0.574 | 2.058 | 4.565 | 1 1000 | 0.379 | 1.458 | 2.268 | 0.97 5000 | 0.140 | 1.058 | 1.138 | 0.97 Table 1: Simulation results for Estimating the Data-Adaptive ARE using the Average k-fold Estimates Table 1 shows the results of the simulations based on comparing the mean fold estimated ARE to the mean ARE of data-adaptive rules applied to $P_{0}$. It can be seen that the estimation is unbiased, and the coverage of confidence intervals based IC-based estimates of the standard errors is slightly high. Figure 11 shows the sampling distribution for each sample size for each type of estimate in the three continuous exposures. We see each estimate converge to a mean 0 normal sampling distribution as sample size increases with the average k-fold estimate having a tighter distribution. Figure 10: Bias Standardized by Standard Error Compared to ATE of True Rule N | Absolute Bias | SD | MSE | Coverage ---|---|---|---|--- 200 | 0.608 | 2.10 | 4.797 | 0.95 1000 | 0.382 | 1.437 | 2.210 | 0.95 5000 | 0.178 | 0.894 | 0.831 | 0.96 Table 2: Simulation results for Estimating the Data-Adaptive ARE using the Average k-fold Estimates in Three Exposure Simulations Figure 11: Bias Standardized by Standard Error Compared to ARE of Data-Adaptive Rule for Three Exposures ## 7 Applications ### 7.1 NIEHS Synthetic Mixtures The NIEHS synthetic mixtures data (found here on github) is a commonly used data set to evaluated the performance of statistical methods for mixtures. This synthetic data can be considered the results of a prospective cohort study. The outcome cannot cause the exposures (as might occur in a cross- sectional study). Correlations between exposure variables can be thought of as caused by common sources or modes of exposure. The nuisance variable Z can be assumed to be a potential confounder and not a collider. There are 7 exposures ($X_{1}-X_{7}$) which have a complicated dependency structure with a biologically-based dose response function based on endocrine disruption. For details the github page synthetic data key for data set 1 (used here) gives a description as to how the data was generated. Largely, there are two exposure clusters ($X_{1},X_{2},X_{3}$ and $X_{5},X_{6}$). And therefore, correlations within these clusters are high. $X_{1},X_{2},X_{7}$ contribute positively to the outcome; $X_{4},X_{5}$ contribute negatively; $X_{3}$ and $X_{6}$ do not have an impact on the outcome which makes rejecting these variables difficult given their correlations with cluster group members. This correlation and effects structure is biologically plausible as different congeners of a group of compounds (e.g., PCBs) may be highly correlated, but have different biological effects. There are various agonistic and antagonistic interactions that exist in the exposures. Table 3 gives a breakdown of the variable sets and their relationships. Variables | Interaction Type ---|--- X1 and X2 | Toxic equivalency factor, a special case of concentration addition (both increase Y) X1 and X4 | Competitive antagonism (similarly for X2 and X4) X1 and X5 | Competitive antagonism (similarly for X2 and X4) X1 and X7 | Supra-additive (“synergy”) (similarly for X2 and X7) X4 and X5 | Toxic equivalency factor, a type of concentration addition (both decrease y) X4 and X7 | Antagonism (unusual kind) (similarly for X5 and X7) Table 3: NIEHS Synthetic Data Interactions Given these toxicological interactions we can expect certain statistical interactions determined as cut-points for sets of variables from CVtreeMLE. For example, we might expect a positive ARE attached to a rule for $X_{1}>=x_{1}\text{ \& }X_{2}>=x_{2}$ where $x_{1},x_{2}$ are certain values for the respective exposures because these two exposures both have a positive impact on Y. Likewise, in the case for antagonistic relationships such as in the case of $X_{2},X_{4}$, we would expect a positive ARE attached to a rule $X_{2}>=x_{2}\text{ \& }X_{4}<=x_{4}$. This is because we might expect the outcome to be highest in a region where $X_{2}$ is high and $X_{4}$ is low given the antagonistic interaction. The NIEHS data set has 500 observations and 9 variables. Z is a binary confounder. Of course, in this data there is no ground-truth, like in the above simulations, but we can gauge CVtreeMLE’s performance by determining if the correct variable sets are used in the interactions and if the correct variables are rejected. Because many machine learning algorithms will fail when fit with one predictor (in our case this happens for g(Z)), we simulate additional covariates that have no effects on the exposures or outcome but prevent these algorithms from breaking. We apply CVtreeMLE to this NIEHS synthetic data using 10-fold CV and the default stacks of estimators used in the Super Learner for all parameters. We select for trees with positive coefficients in the ensemble during the data- adaptive estimation and therefore report results as positive AREs. We parallelize over the cross-validation to test computational run-time on a newer personal machine an analyst might be using. Mixture ATE | Standard Error | Lower CI | Upper CI | P-value | P-value Adj | Vars | Union_Rule ---|---|---|---|---|---|---|--- 8.24 | 0.56 | 7.14 | 9.34 | 0.00 | 0.00 | X1-X5 | X1 $>$= 0.267 & X5 $<$= 3.189 8.16 | 0.56 | 7.07 | 9.26 | 0.00 | 0.00 | X1-X7 | X1 $>$= 0.326 & X7 $>$= 0.22 6.68 | 0.62 | 5.46 | 7.89 | 0.00 | 0.00 | X2-X5 | X2 $>$= 0.602 & X5 $<$= 3.189 6.82 | 0.58 | 5.68 | 7.95 | 0.00 | 0.00 | X2-X7 | X2 $>$= 0.619 & X7 $>$= 1.171 7.29 | 0.51 | 6.29 | 8.29 | 0.00 | 0.00 | X5-X7 | X5 $<$= 3.269 & X7 $>$= 0.138 Table 4: NIEHS Synthetic Data Consistent Interaction Results Table 4 shows the results from CVtreeMLE when applied to this NIEHS synthetic data set using the aforementioned settings. We filter results to only interactions that were found in all 10-folds, that is, trees with variable sets found across all the folds and therefore have consistent "signal" in the data. Let’s focus on the second row with variables $X_{1}$ and $X_{7}$. Table 3 shows that these two variables have a supra-additive or synergistic non- additive relationship. The union rule for trees including these two variables was $X1>=0.326\text{ \& }X7>=0.22$ meaning this rule covers all observations indicated by the fold-specific rules. The mixture ARE is then interpreted as, if all individuals were exposed to $X_{1}$ at levels at or greater than 0.326 and exposed to levels of $X_{7}$ at or greater than 0.22 the outcome would be 8.16 units greater compared to if all individuals were exposed to levels less than these respective levels. The subsequent standard errors derived from the pooled influence curve (column 2) are used to derive the confidence intervals and p-values for hypothesis testing. Overall, comparing these statistical interactions to the toxicological interactions listed CVtreeMLE identifies 5 of 9 interactions. The other interactions in the above table are interpreted in the same way as the $X_{1}$ and $X_{7}$ interaction. We next can investigate how consistent the results are across the folds by looking at the k-fold specific results, this gives us a sense of how reliable our ARE estimates are for the pooled rule. Let’s dig deeper into this $X_{1}$ and $X_{7}$ interaction. Table 5 shows the k-fold specific results for the interactions found for the variables $X_{1}$ and $X_{7}$. Each row is the results for each fold and the final row is the inverse variance weighted pooled result, pooling estimates across the folds. Estimates show stability across the folds with only one fold, fold 8, deviating from the trend. Cut- points at $X_{1}$ were either at 0.991 or 0.998 with fold 8 having a lower cut-point of 0.319. Likewise, $X_{7}$ was partitioned at 0.48 in most folds. Each fold-specific result has valid inference however it is also necessary to evaluate how consistent results were across the folds and thus determine if partitions are stable. Here we see the $X1>=0.99\text{ \& }X7>=0.48$ partition for these two variables is stable and found in 8 of the folds. The ARE estimate for these rules ranges from 8-9 all with a significant effect. CVtreeMLE also provides plots of k-fold estimates to more easily assess for trends, Figure 12 gives an example of this plot for the interaction $X_{1}$ and $X_{7}$. Mixture ATE | SE | Lower CI | Upper CI | P-Value | P-Value Adj | Mix Rule | Fold ---|---|---|---|---|---|---|--- 8.09 | 1.03 | 6.08 | 10.10 | 0.00 | 0.00 | X1 $>$ 0.998 & X7 $>$ 0.48 | 1 7.12 | 1.63 | 3.92 | 10.32 | 0.00 | 0.00 | X1 $>$ 0.998 & X7 $>$ 2.108 | 2 8.38 | 1.51 | 5.41 | 11.34 | 0.00 | 0.00 | X1 $>$ 0.991 & X7 $>$ 0.48 | 3 9.42 | 2.23 | 5.05 | 13.80 | 0.00 | 0.00 | X1 $>$ 0.991 & X7 $>$ 0.441 | 4 8.81 | 1.76 | 5.36 | 12.26 | 0.00 | 0.00 | X1 $>$ 0.998 & X7 $>$ 0.439 | 5 8.43 | 1.53 | 5.42 | 11.43 | 0.00 | 0.00 | X1 $>$ 1.049 & X7 $>$ 0.482 | 6 8.84 | 1.84 | 5.23 | 12.44 | 0.00 | 0.00 | X1 $>$ 0.991 & X7 $>$ 0.413 | 7 5.20 | 2.69 | -0.07 | 10.48 | 0.05 | 0.53 | X7 $>$ 0.217 & X1 $>$ 0.319 | 8 8.41 | 1.04 | 6.36 | 10.46 | 0.00 | 0.00 | X1 $>$ 0.991 & X7 $>$ 0.482 | 9 8.94 | 1.39 | 6.22 | 11.67 | 0.00 | 0.00 | X1 $>$ 0.991 & X7 $>$ 0.482 | 10 8.30 | 5.49 | -2.45 | 19.05 | 0.13 | 0.13 | X1 $>$= 0.326 & X1 $<$= 4.687 & X7 $>$= 0.22 & X7 $<$= 4.886 | Pooled Table 5: $X_{1}$ and $X_{7}$ k-fold Interaction Results Figure 12: K-fold specific results for the interaction $X_{1}$ and $X_{7}$ Overall, CVtreeMLE is able to determine subspaces in the respective variables that have the most impact on the endocrine disrupting outcome. Of note is the fact that no interactions include the variables $X_{3}$ and $X_{6}$ both of which have no impact on the outcome. #### 7.1.1 Comparison to Existing Methods Currently, quantile g-computation is a popular method for mixture analysis in environmental epidemiology. The method yields estimates of the effect of increasing all exposures by one quantile, simultaneously under linear model assumptions. Quantile g-computation looks like: $Y_{i}=\beta_{0}+\sum_{j=1}^{d}\beta_{j}X_{ji}^{q}+\beta Z_{i}+\epsilon_{i}$ Where $X^{q}$ are the quantized mixture components and $Z$ are the covariates. Which works by first transforming mixture components into quantiles. Then the negative and positive coefficients from a linear model for the mixture components are summed to give a mixture ($\Psi$) summary measure which characterizes the joint impact. There are many assumptions that should be poignant after our discussion of mixtures. Firstly, quantiles may not characterize the exposure-response relationship (could be non-monotonic) which occurs in endocrine disrupting compounds. For interpretable weights and mixture estimate $\Psi$, assumes additive relationship of quantiles ($\Psi$ is just sum of $\beta$’s in front of mixture components). After our discussion, in mixtures our main goal is model possible interactions in the data because we expect exposures to have non-additive, possible non-monotonic, antagonistic and agonistic relationships. Therefore, we should expect interactions in our mixture data. In quantile g-computation, with the inclusion of interactions, the proportional contribution of an exposure to the overall effect then varies according to levels of other variables and therefore weights cannot be estimated. Because we can never assume no interactions, quantile g-computation then boils down to getting conditional expectations when setting mixtures to quantiles through a linear model with interaction terms specified by the analyst. After our discussion of mixtures this should feel incorrect. As we argue, the important variables, relationships, and thresholds in a mixture are all unknown to the analyst which makes this a data-adaptive target parameter problem. Even testing quantile g-computation on the NIEHS data is difficult because we don’t know what interactions to include a priori. The best we can do is run it out of the box and with two-way interactions and compare results to the ground-truth measures. Lastly, quantile g-computation does not flexibly control for covariates. We run quantile g-computation on the NIEHS data using 4 quantiles with no interactions to investigate results using this model. The scaled effect size (positive direction, sum of positive coefficients) was 6.28 and included $X_{1},X_{2},X_{3},X_{7}$ and the scaled effect size (negative direction, sum of negative coefficients) was -3.68 and included $X_{4},X_{5},X_{6}$. Compared to the NIEHS ground-truth, $X_{3},X_{6}$ are incorrectly included in these estimates. However the positive and negative associations for the other variables are correct. Next, because we expect interactions to exist in the mixture data, we would like to assess for them but the question is which interaction terms to include? Our best guess is to include interaction terms for all the exposures. We do this and show results in Table 6. | Estimate | Std. Error | Lower CI | Upper CI | Pr($>$$|$t$|$) ---|---|---|---|---|--- (Intercept) | 21.29 | 1.58 | 18.19 | 24.39 | 0.00 psi1 | 0.02 | 1.62 | -3.16 | 3.20 | 0.99 psi2 | 0.59 | 0.67 | -0.71 | 1.90 | 0.37 Table 6: Quantile G-Computation Interaction Results from NIEHS Synthetic Data In Table 6 $\Psi_{1}$ is the summary measure for main effects and $\Psi_{2}$ for interactions. As can be seen, when including all interactions neither of the estimates are significant. Of course this is to be expected given the number of parameters in the model and sample size $n=500$. However, moving forward with interaction assessment is difficult, if we were to assess for all 2-way interaction of 7 exposures the number of sets is 21 and with 3-way interactions is 35. We’d have to run this many models and then correct for multiple testing. Hopefully this example shows why mixtures are inherently a data-adaptive problem and why popular methods such as this, although succinct and interpretable, fall short even in a simple synthetic data set. ### 7.2 NHANES Data Environmental chemical and metal exposure can affect telomere length, which is a mediating pathway for adverse health outcomes including cancer. Studies on the association of metals with leucocyte telomere length (LTL) are mainly limited to single-metal effect Xia et al. (2022); Zota et al. (2015); Clarity et al. (2021). Some research has investigated the overall joint associations metal mixture with LTL using parameteric models like multiple linear regression or quantile-sum g-computation Keil et al. (2019); Lai et al. (2022). Since environmental stressors are known to disrupt telomere length homeostasis which plays a crucial role in cellular aging and disease we investigate the association of mixed metal exposure on LTL. Our desires are two fold, 1. to show CVtreeMLE results applied to real-world data and 2. to provide such data and data processing code with the open-source CVtreeMLE package. As such, we develop a pipeline to download and clean National Health and Nutrition Examination Survey (NHANES) dataset. We download and format the relevant NHANES 1999–2002 dataset containing demographic data, disease history, nine urine metals, and LTL. The demographic data used as possible confounders ($W$) include age, gender, race, education level, marital status, alcohol, smoking (cotinine) , body mass index (BMI), family poverty ratio (PIR), fasting glucose, systolic and diastolic blood pressure, exercise and birth country. Urine metal contained barium (Ba), cadmium (Cd), cobalt (Co), cesium (Cs), molybdenum (Mo), lead (Pb), antimony (Sb), thallium (Tl) and tungsen (W). The outcome is LTL. The number of observations in this test data is 2510. The coding pipeline and data are available in the CVtreeMLE package. We apply CVtreeMLE using the default learners in each stack. We use 10-fold CV and set the max number of iterations in the iterative backfitting to 10 as well. Because previous research has shown the exposure to metals shortens LTL, we set the ATE direction to negative to select trees in the data-adaptive procedure which have the minimum (negative) impact and thus return negative ATEs for each fold. Mixture ATE | Standard Error | Lower CI | Upper CI | P-value | P-value Adj | Vars | Union_Rule | %Fold ---|---|---|---|---|---|---|---|--- 0.06 | 0.04 | -0.02 | 0.13 | 0.17 | 1.00 | cadmium-molybdenum | cadmium $>$= 0 & cadmium $<$= 0.715 & molybdenum $>$= 19.8 & molybdenum $<$= 436.8 | 0.80 -0.03 | 0.01 | -0.06 | -0.01 | 0.02 | 0.37 | cadmium-thallium | cadmium $>$= 0.027 & cadmium $<$= 36.777 & thallium $>$= 0.01 & thallium $<$= 0.38 | 1.00 Table 7: Consistent Pooled TMLE Results NHANES Metal Mixture-LTL Table 7 shows the pooled TMLE ARE results for rules found in more than 75% of the folds. Here we see rules including cadmium and thallium were found in all the folds and rules including cadmium and molybdenum were found in 80% of the folds. The cadmium-thallium interaction had a significant ARE of -0.03 and the cadmium-molybdenum was borderline significant with an ARE of 0.06. These results show that, exposure to high levels of cadmium >= 0.027 and low levels of thallium <= 0.38 is associated with a reduced telomere length of 0.03 compared to exposure levels of cadmium levels lower than 0.027 and thallium levels greater than 0.38. This result implies an antagonistic relationship between cadmium and thallium. Likewise, telomere length was longer (0.06) for those exposed to low levels of cadmium <= 0.715 and high levels of molybdenum >= 19.8 compared to those exposed to the inverse exposure region for these two metals. Like the NIEHS synthetic data results, we can investigate the k-fold specific results for these pooled results. Let’s look at the cadmium and thallium interaction in each fold to see how stable the partition points were for each metal. ATE | SE | Lower CI | Upper CI | P-Value | P-Value Adj | Mix Rule ---|---|---|---|---|---|--- -0.03 | 0.06 | -0.15 | 0.09 | 0.67 | 1.00 | thallium $<$= 0.21 & cadmium $>$ 0.243 -0.04 | 0.03 | -0.10 | 0.03 | 0.24 | 1.00 | cadmium $>$ 0.295 & thallium $<$= 0.31 -0.03 | 0.03 | -0.09 | 0.03 | 0.32 | 1.00 | thallium $<$= 0.21 & cadmium $>$ 0.101 -0.05 | 0.04 | -0.13 | 0.02 | 0.14 | 1.00 | cadmium $>$ 0.097 & thallium $<$= 0.38 -0.03 | 0.03 | -0.09 | 0.03 | 0.37 | 1.00 | cadmium $>$ 0.295 & thallium $<$= 0.21 -0.03 | 0.05 | -0.12 | 0.06 | 0.54 | 1.00 | thallium $<$= 0.21 & cadmium $>$ 0.143 -0.04 | 0.08 | -0.20 | 0.12 | 0.62 | 1.00 | cadmium $>$ 0.29 & thallium $<$= 0.21 -0.04 | 0.05 | -0.14 | 0.06 | 0.44 | 1.00 | thallium $<$= 0.36 & cadmium $>$ 0.092 -0.02 | 0.06 | -0.13 | 0.10 | 0.76 | 1.00 | cadmium $>$ 0.027 & thallium $<$= 0.14 -0.03 | 0.04 | -0.11 | 0.05 | 0.52 | 1.00 | thallium $<$= 0.22 & cadmium $>$ 0.254 -0.03 | 0.16 | -0.34 | 0.27 | 0.83 | 0.83 | cadmium $>$= 0.027 & thallium $<$= 0.38 Table 8: K-fold specific results for cadmium-thallium interactions associated with LTL Table 8 shows the k-fold specific results for cadmium and thallium interaction. This interaction was found in all the folds with an ARE ranging from -0.02 to -0.05. None of the fold specific results were significant due to the variance estimates being calculated on the 251 observations in each validation fold, making standard errors high. However, we see consistent partitioning of thallium between 0.14 and 0.38 and partitioning of cadmium between 0.027 and 0.29. Overall, we see consistent cut-points across the folds which indicates this interaction is stable. The last row in this table is the inverse weighted pooled results. Here we can see that we gain much power by using the pooled influence curve in the pooled TMLE procedure which is able to borrow variance information across the folds because all estimates are cross- estimated. Here, we can see the pooled estimated has much higher variance and wider confidence intervals. ATE | SE | Lower CI | Upper CI | P-Value | P-Value Adj | Mix Rule | fold ---|---|---|---|---|---|---|--- 0.05 | 0.04 | -0.04 | 0.14 | 0.30 | 1.00 | molybdenum $>$ 55.2 & cadmium $<$= 0.384 | 2 0.04 | 0.10 | -0.15 | 0.23 | 0.68 | 1.00 | molybdenum $>$ 52.9 & cadmium $<$= 0.368 | 3 0.01 | 0.04 | -0.07 | 0.09 | 0.77 | 1.00 | cadmium $<$= 0.715 & molybdenum $>$ 19.7 | 4 0.11 | 0.14 | -0.16 | 0.37 | 0.42 | 1.00 | cadmium $<$= 0.35 & molybdenum $>$ 102.5 | 5 0.05 | 0.05 | -0.04 | 0.14 | 0.27 | 1.00 | cadmium $<$= 0.292 & molybdenum $>$ 57.2 | 6 0.02 | 0.22 | -0.41 | 0.46 | 0.92 | 1.00 | cadmium $<$= 0.124 & molybdenum $>$ 21.7 | 7 0.02 | 0.06 | -0.09 | 0.14 | 0.71 | 1.00 | cadmium $<$= 0.429 & molybdenum $>$ 44.5 | 8 0.14 | 0.28 | -0.42 | 0.69 | 0.63 | 1.00 | cadmium $<$= 0.131 & molybdenum $>$ 55.6 | 9 0.04 | 0.41 | -0.77 | 0.84 | 0.93 | 0.93 | cadmium $>$= 0 & cadmium $<$= 0.715 & molybdenum $>$= 19.8 & molybdenum $<$= 436.8 | Pooled Table 9: K-fold specific results for cadmium-molybdenum interactions associated with LTL Lastly, we look at the cadmium-molybdenum interactions in Table 9 . As we can see here, interactions are not found in every fold and the partition points have a larger range although they all point in the same direction (low cadmium and high molybdenum) and all fold specific results are positive. This makes sense given that molybdenum processes proteins and genetic material like DNA and helps break down drugs and toxic substances that enter the body. Therefore, we would expect low cadmium and high molybdenum to be associated with higher telomere length. Overall, in this NHANES example, we show that in real world data, CVtreeMLE can answer questions regarding expected outcomes under different exposure levels of a mixture which are otherwise occult given the limitation of existing methods. ## 8 Software The development of asymptotically linear estimators for data-adaptive parameters are critical for the field of mixed exposure statistics. However, the development of open-source software which translates semi-parametric statistical theory into well-documented functional software is a formidable challenge. Such implementation requires understanding of causal inference, semi-parametric statistical theory, machine learning, and the intersection of these disciplines. The CVtreeMLE R package provides researchers with an open- source tool for evaluating the causal effects of a mixed exposure using the methodology described here. The CVtreeMLE package is well documented and includes a vignette detailing semi-parametric theory for data-adaptive parameters, examples of output, results with interpretations under various real-life mixture scenarios, and comparison to existing methods. The NIEHS synthetic data and the NHANES mixed metal exposure data are provided. The NIEHS synethetic data application is used in the vignette of the package which makes these results reproducible to any researcher and likewise the NHANES data and code are provided for reproducibility. CVtreeMLE can run sequentially or parallelized across folds using the furrr package Vaughan and Dancho (2022). New statistical software using machine learning often presume the availability of significant computational resources in order to run in a timely manner. Here, our applications of NIEHS and NHANES were all run on a personal macbook machine in under 30 minutes by utilizing parallelization and using flexible yet efficient estimators. Of course, for the simulations high performance computing was used to parallelize iteration over clusters. To-date in scientific publication, the release of reproducible software is the exception rather than the rule. In an effort to make robust statistical software adopted in the future, rather than reliance of simple parameteric models, we make CVtreeMLE available with clear, easily accessible, highly detailed documentation of the coding methods. We also make all functions user- accessible, and develop numerous tests and examples. Coding notebooks show simulations of mixed exposure data and CVtreeMLE output with detailed summaries of interpretation. Lastly, the CVtreeMLE package is well maintained to ensure accessibility with ongoing improvements tested at each iteration. The CVtreeMLE package has been made publicly available via GitHub. The pseduo- code that describes the code which executes the described method is shown in Figure 13. Figure 13: CVtreeMLE Pseudo-Code ## 9 Discussion In this paper we introduce a new method for estimating the effects of a mixed exposure. Our approach treats ensemble decision trees as a data-adaptive target parameter for which we estimate the average effects of exposure for regions identified in the best fitting decision trees. This is done within a cross-validated framework paired with targeted learning of our target parameter which provides estimates that are asymptotically unbiased and have the lowest variance for studies which satisfy the unconfoundedness and positivity assumptions. Our proposed method provides valid confidence intervals without restrictions on the number of exposure, covariates, or the complexity of the data-generating process. Our method first partitions the exposure space into subspaces or regions that best explains the outcome. The output of our method is the exposure effect and respective confidence intervals if all individual were exposed to the exposure region compared to if all individual were not exposed to this region. Our approach has potentially many important applications including identifying what combinations of drugs lead to the most beneficial patient outcomes as well as finding what combinations of pollution chemicals have the most deleterious outcomes on public health. Our approach allows for "dredging with dignity" wherein exposure regions can be discovered in the data which are not known a priori and still provide unbiased estimates for the target parameter with valid confidence intervals. This approach of course comes with some cost as construction of a pooled region across the folds is rather ad hoc. This is the main limitation in the proposed method and other alternatives may exist such as using the average partitioning values of each exposure variable rather than our union approach which is conservative. Our simulations with ground-truth, NIEHS synethetic data and real-world data application show the robustness and interpretability of our approach. In an effort to make adoption of semi- parametric methods such as this more seamless we provide the CVtreeMLE R package on github which is well documented for analysts to apply to their respective data. ## 10 Appendix ## References * Safe [1993] S. Safe. Toxicology, structure-function relationship, and human and environmental health impacts of polychlorinated biphenyls: Progress and problems. _Environmental Health Perspectives_ , 100:259–268, 1993\. ISSN 00916765. doi:10.1289/ehp.93100259. * Kortenkamp [2007] Andreas Kortenkamp. Ten years of mixing cocktails: A review of combination effects of endocrine-disrupting chemicals. _Environmental Health Perspectives_ , 115(SUPPL1):98–105, 2007. ISSN 00916765. doi:10.1289/ehp.9357. * Keil et al. [2019] Alexander P. Keil, Jessie P. Buckley, Katie M. O’Brien, Kelly K. Ferguson, Shanshan Zhao, and Alexandra J. White. A quantile-based g-computation approach to addressing the effects of exposure mixtures. _arXiv_ , 128(April):1–10, 2019. ISSN 23318422. doi:10.1097/01.ee9.0000606120.58494.9d. * De Vocht et al. [2012] Frank De Vocht, Nicola Cherry, and Jon Wakefield. A Bayesian mixture modeling approach for assessing the effects of correlated exposures in case-control studies. _Journal of Exposure Science and Environmental Epidemiology_ , 22(4):352–360, 2012. ISSN 15590631. doi:10.1038/jes.2012.22. * Bobb et al. [2014] Jennifer F. Bobb, Linda Valeri, Birgit Claus Henn, David C. Christiani, Robert O. Wright, Maitreyi Mazumdar, John J. Godleski, and Brent A. Coull. Bayesian kernel machine regression for estimating the health effects of multi-pollutant mixtures. _Biostatistics_ , 16(3):493–508, 2014. ISSN 14684357. doi:10.1093/biostatistics/kxu058. * National Institute of Environmental Health Sciences (2018) [NIEHS] National Institute of Environmental Health Sciences (NIEHS). 2018-2023 Strategic Plan. _National Institutes of Health_ , 2018. URL https://www.niehs.nih.gov/about/strategicplan/strategicplan20182023_508.pdf. * Leo Breiman [1984] Charles J. Stone R.A. Olshen Leo Breiman, Jerome Friedman. _Classification and Regression Trees_. Chapman and Hall/CRC, 1984. * Athey and Imbens [2016] Susan Athey and Guido Imbens. Recursive partitioning for heterogeneous causal effects. _Proceedings of the National Academy of Sciences of the United States of America_ , 113(27):7353–7360, 2016. ISSN 10916490. doi:10.1073/pnas.1510489113. * Hubbard et al. [2016] Alan E. Hubbard, Sara Kherad-Pajouh, and Mark J. Van Der Laan. Statistical Inference for Data Adaptive Target Parameters. _International Journal of Biostatistics_ , 12(1):3–19, 2016. ISSN 15574679. doi:10.1515/ijb-2015-0013. * Zheng and van der Laan [2010] Wenjing Zheng and MJ van der Laan. Asymptotic theory for cross-validated targeted maximum likelihood estimation. _U.C. Berkeley Division of Biostatistics Working Paper Series_ , (273), 2010. URL http://biostats.bepress.com/ucbbiostat/paper273/. * Hastie and Tibshirani [1990] Trevor Hastie and Robert Tibshirani. _Generalized additive models_. Wiley Online Library, 1990. * van der Laan Mark et al. [2007] J. van der Laan Mark, Polley Eric C, and Hubbard Alan E. Super learner. _Statistical Applications in Genetics and Molecular Biology_ , 6(1):1–23, 2007. URL https://EconPapers.repec.org/RePEc:bpj:sagmbi:v:6:y:2007:i:1:n:25. * Hejazi et al. [2021] Nima S. Hejazi, Mark J. van der Laan, Holly E. Janes, Peter B. Gilbert, and David C. Benkeser. Efficient nonparametric inference on the effects of stochastic interventions under two-phase sampling, with applications to vaccine efficacy trials. _Biometrics_ , 77(4):1241–1253, 2021. ISSN 15410420. doi:10.1111/biom.13375. * Hejazi et al. [2022] Nima S Hejazi, Kara E Rudolph, Mark J Van Der Laan, and Iván Díaz. Nonparametric causal mediation analysis for stochastic interventional (in)direct effects. _Biostatistics_ , pages 1–22, 2022. ISSN 1465-4644. doi:10.1093/biostatistics/kxac002. * Iván Díaz Muñoz and Mark van der Laan* [2012] Iván Díaz Muñoz and Mark van der Laan*. Population Intervention Causal Effects Based on Stochastic Interventions. _Biometrics._ , 68(2):541–549, 2012. ISSN 15378276. doi:10.1111/j.1541-0420.2011.01685.x.Population. URL https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3624763/pdf/nihms412728.pdf. * Nichols [2007] Austin Nichols. Causal inference with observational data. _The Stata Journal_ , 7(4):507–541, 2007. doi:10.1177/1536867X0800700403. * Winship and Morgan [1999] Christopher Winship and Stephen L. Morgan. The estimation of causal effects from observational data. _Annual Review of Sociology_ , 25(May):659–707, 1999. ISSN 03600572. doi:10.1146/annurev.soc.25.1.659. * Rubin [2006] Donald B. Rubin. Estimating causal effects from large data sets using propensity scores. _Matched Sampling for Causal Effects_ , pages 443–453, 2006. ISSN 0003-4819. doi:10.1017/CBO9780511810725.035. * Zivich and Breskin [2021] Paul N. Zivich and Alexander Breskin. Machine Learning for Causal Inference: On the Use of Cross-fit Estimators. _Epidemiology_ , 32(3):393–401, 2021. ISSN 15315487. doi:10.1097/EDE.0000000000001332. * Luque-Fernandez et al. [2018] Miguel Angel Luque-Fernandez, Michael Schomaker, Bernard Rachet, and Mireille E. Schnitzer. Targeted maximum likelihood estimation for a binary treatment: A tutorial. _Statistics in Medicine_ , 37(16):2530–2546, 2018\. ISSN 10970258. doi:10.1002/sim.7628. * Smith et al. [2022] Matthew J. Smith, Mohammad A. Mansournia, Camille Maringe, Paul N. Zivich, Stephen R. Cole, Clémence Leyrat, Aurélien Belot, Bernard Rachet, and Miguel A. Luque-Fernandez. Introduction to computational causal inference using reproducible Stata, R, and Python code: A tutorial. _Statistics in Medicine_ , 41(2):407–432, 2022\. ISSN 10970258. doi:10.1002/sim.9234. * van der Laan and Rose [2011] M.J. van der Laan and S. Rose. _Targeted Learning: Causal Inference for Observational and Experimental Data_. Springer Series in Statistics. Springer New York, 2011. ISBN 9781441997821. URL https://books.google.com/books?id=RGnSX5aCAgQC. * van der Laan and Rose [2018] M.J. van der Laan and S. Rose. _Targeted Learning in Data Science: Causal Inference for Complex Longitudinal Studies_. Springer Series in Statistics. Springer International Publishing, 2018\. ISBN 9783319653044. URL https://books.google.com/books?id=vKFTDwAAQBAJ. * van der Laan and Rubin [2006] Mark J. van der Laan and Daniel Rubin. Targeted maximum likelihood learning. _The International Journal of Biostatistics_ , 2(1), 2006\. doi:doi:10.2202/1557-4679.1043. URL https://doi.org/10.2202/1557-4679.1043. * Friedman and Popescu [2008] Jerome H. Friedman and Bogdan E. Popescu. Predictive learning via rule ensembles. _Annals of Applied Statistics_ , 2(3):916–954, 2008. ISSN 19326157. doi:10.1214/07-AOAS148. * Fokkema [2020a] Marjolein Fokkema. Fitting prediction rule ensembles with R package pre. _Journal of Statistical Software_ , 92(12), 2020a. ISSN 15487660. doi:10.18637/jss.v092.i12. * van der Laan et al. [2008] Mark van der Laan, Eric Polley, and Alan Hubbard. Super Leaner. _UC Berklely Division of Biostatistics Working Paper Series_ , 2008\. * Hothorn and Zeileis [2015] Torsten Hothorn and Achim Zeileis. partykit: A modular toolkit for recursive partytioning in R. _Journal of Machine Learning Research_ , 16:3905–3909, 2015\. URL https://jmlr.org/papers/v16/hothorn15a.html. * Fokkema [2020b] Marjolein Fokkema. Fitting prediction rule ensembles with R package pre. _Journal of Statistical Software_ , 92(12):1–30, 2020b. doi:10.18637/jss.v092.i12. * Xia et al. [2022] Fang Xia, Qingwen Li, Xin Luo, and Jinyi Wu. Association between urinary metals and leukocyte telomere length involving an artificial neural network prediction: Findings based on NHANES 1999–2002. _Frontiers in Public Health_ , 10, 2022. ISSN 22962565. doi:10.3389/fpubh.2022.963138. * Zota et al. [2015] Ami R. Zota, Belinda L. Needham, Elizabeth H. Blackburn, Jue Lin, Sung Kyun Park, David H. Rehkopf, and Elissa S. Epel. Associations of cadmium and lead exposure with leukocyte telomere length: Findings from National Health And Nutrition Examination Survey, 1999-2002. _American Journal of Epidemiology_ , 181(2):127–136, 2015. ISSN 14766256. doi:10.1093/aje/kwu293. * Clarity et al. [2021] Cassidy Clarity, Jessica Trowbridge, Roy Gerona, Katherine Ona, Michael McMaster, Vincent Bessonneau, Ruthann Rudel, Heather Buren, and Rachel Morello-Frosch. Associations between polyfluoroalkyl substance and organophosphate flame retardant exposures and telomere length in a cohort of women firefighters and office workers in San Francisco. _Environmental Health: A Global Access Science Source_ , 20(1):1–14, 2021. ISSN 1476069X. doi:10.1186/s12940-021-00778-z. * Lai et al. [2022] Xuefeng Lai, Yu Yuan, Miao Liu, Yang Xiao, Lin Ma, Wenting Guo, Qin Fang, Huihua Yang, Jian Hou, Liangle Yang, Handong Yang, Mei an He, Huan Guo, and Xiaomin Zhang. Individual and joint associations of co-exposure to multiple plasma metals with telomere length among middle-aged and older Chinese in the Dongfeng-Tongji cohort. _Environmental Research_ , 214(P3):114031, 2022\. ISSN 10960953. doi:10.1016/j.envres.2022.114031. URL https://doi.org/10.1016/j.envres.2022.114031. * Vaughan and Dancho [2022] Davis Vaughan and Matt Dancho. _furrr: Apply Mapping Functions in Parallel using Futures_ , 2022\. https://github.com/DavisVaughan/furrr, https://furrr.futureverse.org/.
11institutetext: Paper ID 555611institutetext: King Abdullah University of Science and Tehchnology (KAUST), Saudi Arabia 22institutetext: University of Oxford, United Kingdom 22email<EMAIL_ADDRESS> # On the Robustness of Quality Measures for GANs Anonymous ECCV submission Motasem Alfarra 11 Juan C. Pérez 11 Anna Frühstück 11 Philip H. S. Torr 22 Peter Wonka 11 Bernard Ghanem 11 ###### Abstract This work evaluates the robustness of quality measures of generative models such as Inception Score (IS) and Fréchet Inception Distance (FID). Analogous to the vulnerability of deep models against a variety of adversarial attacks, we show that such metrics can also be manipulated by additive pixel perturbations. Our experiments indicate that one can generate a distribution of images with very high scores but low perceptual quality. Conversely, one can optimize for small imperceptible perturbations that, when added to real world images, deteriorate their scores. We further extend our evaluation to generative models themselves, including the state of the art network StyleGANv2. We show the vulnerability of both the generative model and the FID against additive perturbations in the latent space. Finally, we show that the FID can be robustified by simply replacing the standard Inception with a robust Inception. We validate the effectiveness of the robustified metric through extensive experiments, showing it is more robust against manipulation.111Code: https://github.com/R-FID-Robustness-of-Quality-Measures- for-GANs ###### Keywords: Generative Adversarial Networks, Perceptual Quality, Adversarial Attacks, Network Robustness ## 1 Introduction Figure 1: Does the Fréchet Inception Distance (FID) accurately measure the distances between image distributions? We generate datasets that demonstrate the unreliability of FID in judging perceptual (dis)similarities between image distributions. The top left box shows a sample of a dataset constructed by introducing imperceptible noise to each ImageNet image. Despite the remarkable visual similarity between this dataset and ImageNet (bottom box), an extremely large FID (almost 8000) between these two datasets showcases FID’s failure to capture perceptual similarities. On the other hand, a remarkably low FID (almost 1.0) between a dataset of random noise images (samples shown in the top right box) and ImageNet illustrates FID’s failure to capture perceptual dissimilarities. Deep Neural Networks (DNNs) are vulnerable to small imperceptible perturbations known as adversarial attacks. For example, while two inputs $x$ and $(x+\delta)$ can be visually indistinguishable to humans, a classifier $f$ can output two different predictions. To address this deficiency in DNNs, adversarial attacks [11, 7] and defenses [20, 27] have prominently emerged as active areas of research. Starting from image classification [28], researchers also assessed the robustness of DNNs for other tasks, such as segmentation [1], object detection [30], and point cloud classification [18]. While this lack of robustness questions the reliability of DNNs and hinders their deployment in the real world, DNNs are still widely used to evaluate performance in other computer vision tasks, such as that of generation. Metrics in use for assessing generative models in general, and Generative Adversarial Networks (GANs) [10] in particular, are of utmost importance in the literature. This is because such metrics are widely used to establish the superiority of a generative model over others, hence guiding which GAN should be deployed in real world. Consequently, such metrics are expected to be not only useful in providing informative statistics about the distribution of generated images, but also reliable and robust. In this work, we investigate the robustness of metrics used to assess GANs. We first identify two interesting observations that are unique to this context. First, current GAN metrics are built on pretrained classification DNNs that are nominally trained (i.e. trained on clean images only). A popular DNN of choice is the Inception model [25], on which the Inception Score (IS) [22] and Fréchet Inception Distance (FID) [12] rely. Since nominally trained DNNs are generally vulnerable to adversarial attacks [7], it is expected that DNN-based metrics for GANs also inherit these vulnerabilities. Second, current adversarial attacks proposed in the literature are mainly designed at the instance level (e.g. fooling a DNN into misclassifying a particular instance), while GAN metrics are distribution-based. Therefore, attacking these distribution-based metrics requires extending attack formulations from the paradigm of instances to that of distributions. In this paper, we analyze the robustness of GAN metrics and recommend solutions to improve their robustness. We first attempt to assess the robustness of the quality measures used to evaluate GANs. We check whether such metrics are actually measuring the quality of image distributions by testing their vulnerability against additive pixel perturbations. While these metrics aim at measuring perceptual quality, we find that they are extremely brittle against imperceptible but carefully-crafted perturbations. We then assess the judgment of such metrics on the image distributions generated by StyleGANv2 [15] when its input is subjected to perturbations. While the output of GANs is generally well behaved, we still observe that such metrics provide inconsistent judgments where, for example, FID favors an image distribution with significant artifacts over more naturally-looking distributions. At last, we endeavor to reduce these metrics’ vulnerability by incorporating robustly- trained models. We summarize our contributions as follows: * • We are the first to provide an extensive experimental evaluation of the robustness of the Inception Score (IS) and the Fréchet Inception Distance (FID) against additive pixel perturbations. We propose two instance-based adversarial attacks that generate distributions of images that fool both IS and FID. For example, we show that perturbations $\delta$ with a small budget (i.e. $\|\delta\|_{\infty}\leq 0.01$) are sufficient to increase the FID between ImageNet [8] and a perturbed version of ImageNet to $\sim 7900$, while also being able to generate a distribution of random noise images whose FID to ImageNet is $1.05$. We illustrate both cases in Figure 1. * • We extend our evaluation to study the sensitivity of FID against perturbations in the latent space of state-of-the-art generative models. In this setup, we show the vulnerability of both StyleGANv2 and FID against perturbations in both its $z$\- and $w$\- spaces. We found that FID provides inconsistent evaluation of the distribution of generated images compared to their visual quality. Moreover, our attack in the latent space causes StyleGANv2 to generate images with significant artifacts, showcasing the vulnerability of StyleGANv2 to additive perturbations in the latent space. * • We propose to improve the reliability of FID by using adversarially-trained models in its computation. Specifically, we replace the traditional Inception model with its adversarially-trained counterpart to generate the embeddings on which the FID is computed. We show that our robust metric, dubbed R-FID, is more resistant against pixel perturbations than the regular FID. * • Finally, we study the properties of R-FID when evaluating different GANs. We show that R-FID is better than FID at distinguishing generated fake distributions from real ones. Moreover, R-FID provides more consistent evaluation under perturbations in the latent space of StyleGANv2. ## 2 Related Work GANs and Automated Assessment. GANs [10] have shown remarkable generative capabilities, specially in the domain of images [14, 15, 4]. Since the advent of GANs, evaluating their generative capabilities has been challenging [10]. This challenge spurred research efforts into developing automated quantitative measures for GAN outputs. Metrics of particular importance for this purpose are the Inception Score (IS), introduced in [22], and the Fréchet Inception Distance (FID), introduced in [12]. Both metrics leverage the ImageNet- pretrained Inception architecture [25] as a rough proxy for human perception. The IS evaluates the generated images by computing conditional class distributions with Inception and measuring (1) each distribution’s entropy—related to Inception’s certainty of the image content, and (2) the marginal’s entropy—related to diversity across generated images. Noting the IS does not compare the generated distribution to the (real world) target distribution, Heusel et al. [12] proposed the FID. The FID compares the generated and target distributions by (1) assuming the Inception features follow a Gaussian distribution and (2) using each distribution’s first two moments to compute the Fréchet distance. Further, the FID was shown to be more consistent with human judgement [24]. Both the original works and later research criticized these quantitative assessments. On one hand, IS is sensitive to weight values, noisy estimation when splitting data, distribution shift from ImageNet, susceptibility to adversarial examples, image resolution, difficulty in discriminating GAN performance, and vulnerability to overfitting [2, 22, 3, 29]. On the other hand, FID has been criticized for its over-simplistic assumptions (“Gaussianity” and its associated two-moment description), difficulty in discriminating GAN performance, and its inability to detect overfitting [3, 19, 29]. Moreover, both IS and FID were shown to be biased to both the number of samples used and the model to be evaluated [6]. In this work, we provide extensive empirical evidence showing that both IS and FID are not robust against perturbations that modify image quality. Furthermore, we also propose a new robust FID metric that enjoys superior robustness. Adversarial Robustness. While DNNs became the de facto standard for image recognition, researchers found that such DNNs respond unexpectedly to small changes in their input [26, 11]. In particular, various works [5, 20] observed a widespread vulnerability of DNN models against input perturbations that did not modify image semantics. This observation spurred a line of research on adversarial attacks, aiming to develop procedures for finding input perturbations that fool DNNs [7]. This line of work found that these vulnerabilities are pervasive, casting doubt on the nature of the impressive performances of DNNs. Further research showed that training DNNs to be robust against these attacks [20] facilitated the learning of perceptually-correlated features [13, 9]. Interestingly, a later work [23] even showed that such learnt features could be harnessed for image synthesis tasks. In this work, we show (1) that DNN-based scores for GANs are vulnerable against adversarial attacks, and (2) how these scores can be “robustified” by replacing nominally trained DNNs with robustly trained ones. ## 3 Robustness of IS and FID To compare the output of generative models, two popular metrics are used: the _Inception Score_ (IS) and the _Fréchet Inception Distance_ (FID). These metrics depend only on the statistics of the distribution of generated images in an ImageNet-pretrained Inception’s embedding space, raising the question: _What do quality measures for generative models, such as IS and FID, tell us about image quality?_ We investigate this question from the robustness perspective. In particular, we analyze the sensitivity of these metrics to carefully crafted perturbations. We start with preliminary background about both metrics. ### 3.1 Preliminaries We consider the standard image generation setup where a generator $G:\mathbb{R}^{d_{z}}\rightarrow\mathbb{R}^{d_{x}}$ receives a latent code $z\in\mathbb{R}^{d_{z}}$ and outputs an image $x\in\mathbb{R}^{d_{x}}$. Upon training, $G$ is evaluated based on the quality of the generated distribution of images $\mathcal{D}_{G}$ by computing either the IS [22] or the FID [12]. Both metrics leverage an ImageNet-pretrained [8] InceptionV3 [25]. Salimans et al. [22] proposed measuring the perceptual quality of the generated distribution $\mathcal{D}_{G}$ by computing the IS as: $\text{IS}(\mathcal{D}_{G})=\exp\left(\>\mathbb{E}_{x\sim\mathcal{D}_{G}}\left(\text{KL}\left(p(y|x)\>||\>p(y)\right)\right)\>\right),$ (1) where $p(y|x)$ is the output probability distribution of the pretrained Inception model. While several works have argued about the effectiveness of the IS and its widely-used implementation [2], its main drawback is that it disregards the relation between the generated distribution, $\mathcal{D}_{G}$, and the real one, $\mathcal{D}_{R}$, used for training $G$ [12]. Consequently, Heusel et al. proposed the popular FID, which involves the statistics of the real distribution. In particular, FID assumes that the Inception features of an image distribution $\mathcal{D}$ follow a Gaussian distribution with mean $\mu_{\mathcal{D}}$ and covariance $\Sigma_{\mathcal{D}}$, and it measures the squared Wasserstein distance between the two Gaussian distributions of real and generated images. Hence, $\text{FID}(\mathcal{D}_{R},\mathcal{D}_{G})$, or FID for short, can be calculated as: $\text{FID}=\|\mu_{R}-\mu_{G}\|^{2}+\text{Tr}\left(\Sigma_{R}+\Sigma_{G}-2(\Sigma_{R}\Sigma_{G})^{\nicefrac{{1}}{{2}}}\right),$ (2) where $._{R},._{G}$ are the statistics of the real and generated image distributions, respectively, and $\text{Tr}(\cdot)$ is the trace operator. Note that the statistics of both distributions are empirically estimated from their corresponding image samples. In principle, FID measures how close (realistic) the generated distribution $\mathcal{D}_{G}$ is to $\mathcal{D}_{R}$. We remark that the FID is the _de facto_ metric for evaluating image generation-related tasks. Therefore, our study focuses mostly on FID. We note here that both the IS and the FID are oblivious to $G$’s training process and can be computed to compare two arbitrary sets of images $\mathcal{D}_{R}$ and $\mathcal{D}_{G}$. In generative modeling, this is typically a set of real images (photographs) and a set of generated images. However, it is also possible to compare two sets of photographs, two sets of generated images, manipulated photographs with real photographs, _etc_. This flexibility allows us to study these metrics in a broader context next, where no generative model is involved. ### 3.2 Robustness under Pixel Perturbations We first address the question presented earlier in Section 3 by analyzing the sensitivity of IS and FID to additive pixel perturbations. In particular, we assume $\mathcal{D}_{R}$ to be either CIFAR10 [17] or ImageNet [8] and ask: (i) can we generate a distribution of imperceptible additive perturbations $\delta$ that deteriorates the scores for $\mathcal{D}_{G}=\mathcal{D}_{R}+\delta$? Or, alternatively, (ii) can we generate a distribution of low visual quality images, i.e. noise images, that attain good quality scores? If the answer is yes to both questions, then FID and IS have limited capacity for providing information about image quality in the worst case. #### 3.2.1 Good Images - Bad Scores Figure 2: Sensitivity of Inception Score (IS) against pixel perturbations. First row: real-looking images (sampled from $\mathcal{D}_{G}=\mathcal{D}_{R}+\delta$) with a low IS (below 3). Second row: random noise images with a high IS (over 135). We aim at constructing a distribution of real-looking images with _bad_ quality measures, i.e. low IS or high FID. While both metrics are distribution-based, we design instance-wise proxy optimization problems to achieve our goal. Minimizing IS. Based on Eq. (1), one could minimize the IS by having both the posterior $p(y|x)$ and the prior $p(y)$ be the same distribution. Assuming that $p(y)$ is a uniform distribution, we minimize the IS by maximizing the entropy of $p(y|x)$. Therefore, we can optimize a perturbation $\delta^{*}$ for each real image $x_{r}\sim\mathcal{D}_{R}$ by solving the following problem: $\displaystyle\delta^{*}=\operatorname*{arg\,max}_{\|\delta\|_{\infty}\leq\epsilon}~{}\mathcal{L}_{\text{ce}}\left(p(y|x_{r}+\delta),\hat{y}\right),$ (3) $\displaystyle\text{s.t. }\hat{y}=\operatorname*{arg\,max}_{i}~{}p^{i}(y|x_{r}+\delta),$ where $\mathcal{L}_{\text{ce}}$ is the Cross Entropy loss. We solve the problem in Eq. (3) with 100 steps of Projected Gradient Descent (PGD) and zero initialization. We then compile the distribution $\mathcal{D}_{G}$, where each image $x_{g}=x_{r}+\delta^{*}$ is a perturbed version of the real dataset $\mathcal{D}_{R}$. Note that our objective aims to minimize the network’s confidence in predicting all labels for each $x_{g}$. In doing so, both $p(y|x_{g})$ and $p(y)$ tend to converge to a uniform distribution, thus, minimizing the KL divergence between them and effectively lowering the IS. Note how $\epsilon$ controls the allowed perturbation amount for each image $x_{r}$. Therefore, for small $\epsilon$ values, samples from $\mathcal{D}_{G}$ and $\mathcal{D}_{R}$ are perceptually indistinguishable. Table 1: Robustness of IS and FID against pixel perturbations. We assess the robustness of IS and FID against perturbations with a limited budget $\epsilon$ on CIFAR10 and ImageNet. In the last row, we report the IS and FID of images with carefully-designed random noise having a resolution similar to CIFAR10 and ImageNet. $\epsilon$ | CIFAR10 | ImageNet ---|---|--- | IS | FID | IS | FID 0.00 | 11.54 | 0.00 | 250.74 | 0.00 $5\times 10^{-3}$ | 2.62 | 142.45 | 3.08 | 3013.33 $0.01$ | 2.50 | 473.19 | 2.88 | 7929.01 random noise | 94.87 | 9.94 | 136.82 | 1.05 Maximizing FID. Next, we extend our attack setup to the more challenging FID. Given an image $x$, we define $f(x):\mathbb{R}^{d_{x}}\rightarrow\mathbb{R}^{d_{e}}$ to be the output embedding of an Inception model. We aim to maximize the FID by generating a perturbation $\delta$ that pushes the embedding of a real image away from its original position. In particular, for each $x_{r}\sim\mathcal{D}_{R}$, we aim to construct $x_{g}=x_{r}+\delta^{*}$ where: $\delta^{*}=\operatorname*{arg\,max}_{\|\delta\|_{\infty}\leq\epsilon}~{}\left\|f(x_{r})-f(x_{r}+\delta)\right\|_{2}.$ (4) In our experiments, we solve the optimization problem in Eq. (4) with 100 PGD steps and a randomly initialized $\delta$ [20]. Maximizing this objective indirectly maximizes FID’s first term (Eq. (2)), while resulting in a distribution of images $\mathcal{D}_{G}$ that is visually indistinguishable from the real $\mathcal{D}_{R}$ for small $\epsilon$ values. Experiments. We report our results in Table 1. Our simple yet effective procedure illustrates how both metrics are very susceptible to attacks. In particular, solving the problem in Eq. (3) yields a distribution of noise that significantly decreases the IS from 11.5 to 2.5 in CIFAR10 and from 250.7 to 2.9 in ImageNet. We show a sample from $\mathcal{D}_{G}$ in Figure 2, first row. Similarly, our optimization problem in Eq. (4) can create imperceptible perturbations that maximize the FID to $\approx$7900 between ImageNet and its perturbed version (examples shown in Figure 1). #### 3.2.2 Bad Images - Good Scores While the previous experiments illustrate the vulnerability of both the IS and FID against small perturbations (i.e. good images with bad scores), here we evaluate if the converse is also possible, i.e. bad images with good scores. In particular, we aim to construct a distribution of noise images (e.g. second row of Figure 2) that enjoys good scores (high IS or low FID). Maximizing IS. The IS has two terms: Inception’s confidence on classifying a generated image, i.e. $p(y|x_{g})$, and the diversity of the generated distribution of predicted labels, i.e. $p(y)$. One can maximize the IS by generating a distribution $\mathcal{D}_{G}$ such that: (i) each $x_{g}\sim\mathcal{D}_{G}$ is predicted with high confidence, and (ii) the distribution of predicted labels is uniform across Inception’s output $\mathcal{Y}$. To that end, we propose the following procedure for constructing such $\mathcal{D}_{G}$. For each $x_{g}$, we sample a label $\hat{y}\sim\mathcal{Y}$ uniformly at random and solve the problem: $x_{g}=\operatorname*{arg\,min}_{x}~{}\mathcal{L}_{ce}(p(y|x),\hat{y}).$ (5) In our experiments, we solve the problem in Eq. (5) with 100 gradient descent steps and random initialization for $x$. Minimizing FID. Here, we analyze the robustness of FID against such a threat model. We follow a similar strategy to the objective in Eq. (4). For each image $x_{r}\sim\mathcal{D}_{R}$, we intend to construct $x_{g}$ such that: $x_{g}=\operatorname*{arg\,min}_{x}~{}\left\|f(x)-f(x_{r})\right\|_{2}$ (6) with a randomly initialized $x$. In our experiments, we solve Eq. (6) with 100 gradient descent steps. As such, each $x_{g}$ will have a similar Inception representation to a real-world image, i.e. $f(x_{g})\approx f(x_{r})$, while being random noise. ##### Experiments. We report our results in the last row of Table 1. Both the objectives in Eqs. (5) and (6) are able to fool the IS and FID, respectively. In particular, we are able to generate distributions of noise images with resolutions $32\times 32$ and $224\times 224$ (i.e. CIFAR10 and ImageNet resolutions) but with IS of 94 and 136, respectively. We show a few qualitative samples in the second row of Figure 2. Furthermore, we generate noise images that have embedding representations very similar to those of CIFAR10 and ImageNet images. This lowers the FID of both datasets to 9.94 and 1.05, respectively (examples are shown in Figure 1). ### 3.3 Robustness under Latent Perturbations In the previous section, we established the vulnerability of both the IS and FID against pixel perturbations. Next, we investigate the vulnerability against perturbations in a GAN’s latent space. Designing such an attack is more challenging in this case, since images can only be manipulated indirectly, and so there are fewer degrees of freedom for manipulating an image. To that end, we choose $G$ to be the state of the art generator StyleGANv2 [14] trained on the standard FFHQ dataset [14]. We limit the investigation to the FID metric, as IS is not commonly used in the context of unconditional generators, such as StyleGAN. Note that we always generate $70$k samples from $G$ to compute the FID. Recall that our generator $G$ accepts a random latent vector $z\sim\mathcal{N}(0,\text{I})$222The appendix presents results showing that sampling $z$ from different distributions still yields good looking StyleGANv2-generated images. and maps it to the more expressive latent space $w$, which is then fed to the remaining layers of $G$. It is worthwhile to mention that “truncating” the latent $w$ with a pre-computed $\bar{w}$333$\bar{w}$ is referred to as the mean of the $w$-space. It is computed by sampling several latents $z$ and averaging their representations in the $w$-space. and constant $\alpha\in\mathbb{R}$ (i.e. replacing $w$ with $\alpha w+(1-\alpha)\bar{w}$) controls both the quality and diversity of the generated images [14]. Figure 3: Effect of attacking truncated StyleGANv2’s latent space on the Fréchet Inception Distance (FID). We conduct attacks on the latent space of StyleGANv2 and record the effect on the FID. We display the resulting samples of these attacks for two truncation values, $\alpha=0.7$ (top row) and $\alpha=1.0$ (bottom row). Despite the stark differences in realism between the images in the top and bottom rows—i.e. the top row’s remarkable quality and the bottom row’s artifacts—the FID to FFHQ reverses this ranking, wherein the bottom row is judged as farther away from FFHQ than the top row. ##### Effect of Truncation on FID. We first assess the effect of the truncation level $\alpha$ on both image quality and FID. We set $\alpha\in[0.7,1.0,1.3]$ and find FIDs to be $[21.81,2.65,9.31]$, respectively. Based on our results, we assert the following observation: while the visual quality of generated images at higher truncation levels, e.g. $\alpha=0.7$, is better and has fewer artifacts than the other $\alpha$ values, the FID does not reflect this fact, showing lower (better) values for $\alpha\in\\{1.0,1.3\\}$. We elaborate on this observation with qualitative experiments in the appendix. ##### FID-Guided Sampling. Next, we extend the optimization problem in Eq. (4) from image to latent perturbations. In particular, we aim at constructing a perturbation $\delta^{*}_{z}$ for each sampled latent $z$ by solving: $\displaystyle\delta^{*}_{z}$ $\displaystyle=\operatorname*{arg\,max}_{\delta}~{}\left\|f(G(z+\delta))-f(x_{r})\right\|_{2}.$ (7) Thus, $\delta^{*}_{z}$ perturbs $z$ such that $G$ produces an image whose embedding differs from that of real image $x_{r}$. We solve the problem in Eq. (7) for $\alpha\in\\{0.7,1.0\\}$. ##### Experiments. We visualize our results in Figure 3 accompanied with their corresponding FID values (first and second rows correspond to $\alpha=0.7\text{ and }1.0$, respectively). While our attack in the latent space is indeed able to significantly increase the FID (from 2.65 to 31.68 for $\alpha=1.0$ and 21.33 to 34.10 for $\alpha=0.7$), we inspect the results and draw the following conclusions. (i) FID provides inconsistent evaluation of the generated distribution of images. For example, while both rows in Figure 3 have comparable FID values, the visual quality is significantly different. This provides practical evidence of this metric’s unreliability in measuring the performance of generative models. (ii) Adding crafted perturbations to the input of a state of the art GAN deteriorates the visual quality of its output space (second row in Figure 3). This means that GANs are also vulnerable to adversarial attacks. This is confirmed in the literature for other generative models such as GLOW [16, 21]. Moreover, we can formulate a problem similar to Eq. (7) but with the goal of perturbing the $w$-space instead of the $z$-space. We leave results of solving this formulation for different $\alpha$ values to the appendix. ##### Section Summary. In this section, we presented an extensive experimental evaluation investigating if the quality measures (IS and FID) of generative models actually measure the perceptual quality of the output distributions. We found that such metrics are extremely vulnerable to pixel perturbations. We were able to construct images with very good scores but no visual content (Section 3.2.2), as well as images with realistic visual content but very bad scores (Section 3.2.1). We further studied the sensitivity of FID against perturbations in the latent space of StyleGANv2 (Section 3.3), allowing us to establish the inconsistency of FID under this setup as well. Therefore, we argue that such metrics, while measuring useful properties of the generated distribution, lead to questionable assessments of the visual quality of the generated images. ## 4 R-FID: Robustifying the FID After establishing the vulnerability of IS and FID to perturbations, we analyze the cause of such behavior and propose a solution. We note that, while different metrics have different formulations, they rely on a pretrained Inception model that could potentially be a leading cause of such vulnerability. This observation suggests the following question: _Can we robustify the FID by replacing its Inception component with a robustly trained counterpart?_ We first give a brief overview of adversarial training. ### 4.1 Leveraging Adversarially Trained Models Adversarial training is arguably the _de facto_ procedure for training robust models against adversarial attacks. Given input-label pairs $(x,y)$ sampled from a training set $\mathcal{D}_{tr}$, $\ell_{2}$-adversarial training solves the following min-max problem: $\min_{\theta}~{}\mathbb{E}_{(x,y)\sim\mathcal{D}_{tr}}\left[\max_{\|\delta\|_{2}\leq\kappa}\mathcal{L}\left(x+\delta,y;\theta\right)\right]$ (8) for a given loss function $\mathcal{L}$ to train a robust network with parameters $\theta$. We note that $\kappa$ controls the robustness-accuracy trade-off: models trained with larger $\kappa$ tend to have higher robust accuracy (accuracy under adversarial attacks) and lower clean accuracy (accuracy on clean images). Since robust models are expected to resist pixel perturbations, we expect such models to inherit robustness characteristics against the attacks constructed in Section 3.2. Moreover, earlier works showed that robustly-trained models tend to learn more semantically-aligned and invertible features [13]. Therefore, we hypothesize that replacing the pretrained Inception model with its robustly trained counterpart could increase FID’s sensitivity to the visual quality of the generated distribution (i.e. robust against attacks in Section 3.3). To that end, we propose the following modification to the FID computation. We replace the pretrained Inception model with a robustly trained version on ImageNet following Eq. (8) with $\kappa\in\\{64,128\\}$. The training details are left to the appendix. We refer to this alternative as R-FID, and analyze its robustness against perturbations next. Figure 4: Attacking R-FID with pixel perturbations. We attack two variants of R-FID ($\kappa=64$ and $\kappa=128$) and visualize samples from the resulting datasets. Attempting to fool these R-FIDs at the pixel level yields perturbations that correlate with semantic patterns, in contrast to those obtained when attempting to fool the standard FID (as shown in Figure 1). Table 2: R-FID against attacks in the pixel space. We study the robustness of R-FID against the adversarial attacks in Eq. 4. $\epsilon$ | CIFAR10 | ImageNet ---|---|--- | $\kappa=64$ | $\kappa=128$ | $\kappa=64$ | $\kappa=128$ $0.01$ | 1.5 | 0.3 | 21.0 | 4.5 $0.02$ | 20.7 | 7.8 | 293.8 | 92.1 $0.03$ | 46.4 | 19.7 | 657.9 | 264.6 ### 4.2 R-FID against Pixel Perturbations We first test the sensitivity of R-FID against additive pixel perturbations. For that purpose, we replace the Inception with a robust Inception, and repeat the experiments from Section 3.2.1 to construct real images with bad scores. We conduct experiments on CIFAR10 and ImageNet with $\epsilon\in\\{0.01,0.02,0.03\\}$ for the optimization problem in Eq. (4), and we report the results in Table 2. We observe that the use of a robustly- trained Inception significantly improves robustness against pixel perturbations. Our robustness improvement for the same value of $\epsilon=0.01$ is of 3 orders of magnitude (an FID of 4 for $\kappa=128$ compared to 7900 reported in Table 1). While both models consistently provide a notable increase in robustness against pixel perturbations, we find that the model most robust to adversarial attacks (i.e. $\kappa=128$) is also the most robust to FID attacks. It is worthwhile to mention that this kind of robustness is expected since our models are trained not to alter their prediction under additive input perturbations. Hence, their feature space should enjoy robustness properties, as measured by our experiments. In Figure 4 we visualize a sample from the adversarial distribution $\mathcal{D}_{G}$ (with $\epsilon=0.08$) when $\mathcal{D}_{R}$ is ImageNet. We observe that our adversaries while aiming only at pushing the feature representation of samples of $\mathcal{D}_{G}$ away from those of $\mathcal{D}_{R}$, are also more correlated with human perception. This finding aligns with previous observations in the literature, which find robustly-trained models have a more interpretable (more semantically meaningful) feature space [13, 9]. We leave the evaluation under larger values of $\epsilon$, along with experiments on unbounded perturbations, to the appendix. Table 3: Truncation’s effect on R-FID. We study how truncation affects R-FID against FFHQ (first two rows), and across different truncation levels (last two rows). $(\mathcal{D}_{G}(\alpha)$, $\mathcal{D}_{R})$ | 0.7 | 0.9 | 1.0 ---|---|---|--- $\kappa=64$ | 98.3 | 90.0 | 88.1 $\kappa=128$ | 119.9 | 113.7 | 113.8 $(\mathcal{D}_{G}(\alpha_{i}),\mathcal{D}_{G}(\alpha_{j}))$ | (0.7, 1.0) | (0.7, 0.9) | (0.9, 1.0) $\kappa=64$ | 10.5 | 4.9 | 0.48 $\kappa=128$ | 9.9 | 4.6 | 0.46 ### 4.3 R-FID under Latent Perturbations In Section 4.2, we tested R-FID’s robustness against pixel-level perturbations. Next, we study R-FID for evaluating generative models. For this, we follow the setup in Section 3.3 using an FFHQ-trained StyleGANv2 as generator $G$. Figure 5: Robustness of R-FID against perturbations in StyleGANv2 latent space. We conduct attacks on two variants of R-FID ($\kappa=64$ on the left, and $\kappa=128$ on the right) and two truncation values ($\alpha=0.7$ on the top, and $\alpha=1.0$ on the bottom) by perturbing the latent space. We also visualize samples from the generated distributions. For the pairs $(\kappa,\alpha)\in\\{(64,0.7),(64,1.0),(128,0.7),(128,1.0)\\}$, we find corresponding R-FID values of {128.1, 157.8, 126.6, 162,8}. In contrast to the minimal changes required to fool the standard FID (Fig. 3), fooling the R-FID leads to a dramatic degradation in visual quality of the generated images. Effect of Truncation on R-FID. Here, we analyze the R-FID when the generator is using different truncation levels. In particular, we choose $\alpha\in\\{0.7,0.9,1.0\\}$ and report results in Table 3. We observe that the robust Inception model clearly distinguishes the distribution generated by StyleGANv2 from the FFHQ dataset, regardless of the truncation $\alpha$. In this case, we obtain an R-FID of 113.8, substantially larger than the 2.6 obtained when the nominally-trained Inception model is used. This result demonstrates that, while the visual quality of StyleGANv2’s output is impressive, the generated image distribution is far from the FFHQ distribution. We further evaluate if the R-FID is generally large between any two distributions by measuring the R-FID between two distributions of images generated at two truncation levels $(\alpha_{i},\alpha_{j})$. Table 3 reports these results. We observe that (i) the R-FID between a distribution and itself is $\approx 0$, e.g. R-FID = $10^{-3}$ at (1.0, 1.0). Please refer to the appendix for details. (ii) The R-FID gradually increases as the image distributions differ, e.g. R-FID at (0.9, 1.0) $<$ (0.7, 1.0). This observation validates that the large R-FID values found between FFHQ and various truncation levels are a result of the large separation in the embedding space that robust models induce between real and generated images. R-FID Guided Sampling. Next, we assess the robustness of the R-FID against perturbations in the latent space of the generator $G$. For this purpose, we conduct the attack proposed in Eq. (7) with $f$ now being the robustly-trained Inception. We report results and visualize few samples in Figure 5. We make the following observations. (i) While the R-FID indeed increases after the attack, the relative increment is far less than that of the non-robust FID. For example, R-FID increases by 44% at $\kappa=64$ and $\alpha=0.7$ compared to an FID increase of 1000% under the same setup. (ii) The increase in R-FID is associated with a significantly larger amount of artifacts introduced by the GAN in the generated images. This result further evidences the vulnerability of the generative model. However, it also highlights the changes in the image distribution that are required to increase the R-FID. We leave the $w$\- space formulation for the attack on the R-FID, along with its experiments, to the appendix. Section Summary. In this section, we robustified the popular FID by replacing the pretrained Inception model with a robustly-trained version. We found this replacement results in a more robust metric (R-FID) against perturbations in both the pixel (Section 4.2) and latent (Section 4.3) spaces. Moreover, we found that pixel-based attacks yield much more perceptually-correlated perturbations when compared to the attacks that used the standard FID (Figure 2). Finally, we observed that changing R-FID values requires a more significant and notable distribution shift in the generated images (Figure 5). ### 4.4 R-FID against Quality Degradation Table 4: Sensitivity of R-FID against noise and blurring. We measure R-FID $(\kappa=128)$ between ImageNet and a transformed version of it under Gaussian noise and blurring. As $\sigma$ increases, the image quality decreases and R-FID increases. $\nicefrac{{\sigma_{N}}}{{\sigma_{B}}}$ | $\nicefrac{{0.1}}{{1.0}}$ | $\nicefrac{{0.2}}{{2.0}}$ | $\nicefrac{{0.3}}{{3.0}}$ | $\nicefrac{{0.4}}{{4.0}}$ ---|---|---|---|--- Gaussian (N)oise | 16.65 | 61.33 | 128.8 | 198.3 Gaussian (B)lur | 15.54 | 54.07 | 78.67 | 89.11 At last, we analyze the effect of transformations that degrade image quality on R-FID. In particular, we apply Gaussian noise and Gaussian blurring on ImageNet and report the R-FID $(\kappa=128)$ between ImageNet and the degraded version in Table 4. Results show that as the quality of the images degrades (i.e. as $\sigma$ increases), the R-FID steadily increases. Thus, we find that R-FID is able to distinguish a distribution of images from its degraded version. ## 5 Discussion, Limitations, and Conclusions In this work, we demonstrate several failure modes of popular GAN metrics, specifically IS and FID. We also propose a robust counterpart of FID (R-FID), which mitigates some of the robustness problems and yields significantly more robust behavior under the same threat models. Measuring the visual quality for image distributions has two components: (1) the statistical measurement (e.g. Wasserstein distance) and (2) feature extraction using a pretrained model (e.g. InceptionV3). A limitation of our work is that we only focus on the second part (the pretrained model). As an interesting avenue for future work, we suggest a similar effort to assess the reliability of the statistical measurement as well, i.e. analyzing and finding better and more robust alternatives to the Wasserstein distance. Current metrics mainly focus on comparing the distribution of features. In these cases, visual quality is only hoped to be a side effect and not directly optimized for nor tested by these metrics. Developing a metric that directly assesses visual quality remains an open problem that is not tackled by our work but is recommended for future work. Acknowledgments. This work was supported by the King Abdullah University of Science and Technology (KAUST) Office of Sponsored Research (OSR) under Award No. OSR-CRG2019-4033. ## References * [1] Arnab, A., Miksik, O., Torr, P.H.: On the robustness of semantic segmentation models to adversarial attacks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. pp. 888–897 (2018) * [2] Barratt, S., Sharma, R.: A note on the inception score (2018) * [3] Borji, A.: Pros and cons of gan evaluation measures. Computer Vision and Image Understanding 179, 41–65 (2019) * [4] Brock, A., Donahue, J., Simonyan, K.: Large scale GAN training for high fidelity natural image synthesis. In: International Conference on Learning Representations (ICLR) (2019) * [5] Carlini, N., Wagner, D.: Towards evaluating the robustness of neural networks. In: 2017 IEEE Symposium on Security and Privacy (SP) (2017) * [6] Chong, M.J., Forsyth, D.: Effectively unbiased fid and inception score and where to find them. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. pp. 6070–6079 (2020) * [7] Croce, F., Hein, M.: Reliable evaluation of adversarial robustness with an ensemble of diverse parameter-free attacks. In: International Conference on Machine Learning (ICML) (2020) * [8] Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., Li, F.F.: Imagenet: A large-scale hierarchical image database. In: The IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2009) * [9] Engstrom, L., Ilyas, A., Santurkar, S., Tsipras, D., Tran, B., Madry, A.: Adversarial robustness as a prior for learned representations (2020), https://openreview.net/forum?id=rygvFyrKwH * [10] Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., Bengio, Y.: Generative adversarial nets. In: Advances in Neural Information Processing Systems (2014) * [11] Goodfellow, I., Shlens, J., Szegedy, C.: Explaining and harnessing adversarial examples. In: International Conference on Learning Representations (ICLR) (2015) * [12] Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., Hochreiter, S.: Gans trained by a two time-scale update rule converge to a local nash equilibrium. In: Advances in Neural Information Processing Systems (NeurIPS) (2017) * [13] Ilyas, A., Santurkar, S., Tsipras, D., Engstrom, L., Tran, B., Madry, A.: Adversarial examples are not bugs, they are features. In: Advances in Neural Information Processing Systems (NeurIPS) (2019) * [14] Karras, T., Laine, S., Aila, T.: A style-based generator architecture for generative adversarial networks. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (June 2019) * [15] Karras, T., Laine, S., Aittala, M., Hellsten, J., Lehtinen, J., Aila, T.: Analyzing and improving the image quality of stylegan. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (June 2020) * [16] Kingma, D.P., Dhariwal, P.: Glow: Generative flow with invertible 1x1 convolutions. In: Bengio, S., Wallach, H., Larochelle, H., Grauman, K., Cesa-Bianchi, N., Garnett, R. (eds.) Advances in Neural Information Processing Systems. vol. 31. Curran Associates, Inc. (2018), https://proceedings.neurips.cc/paper/2018/file/d139db6a236200b21cc7f752979132d0-Paper.pdf * [17] Krizhevsky, A., Hinton, G.: Learning multiple layers of features from tiny images. In: University of Toronto, Canada (2009) * [18] Liu, H., Jia, J., Gong, N.Z.: Pointguard: Provably robust 3d point cloud classification. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 6186–6195 (2021) * [19] Lucic, M., Kurach, K., Michalski, M., Gelly, S., Bousquet, O.: Are gans created equal? a large-scale study. Advances in Neural Information Processing Systems (NeurIPS) 31 (2018) * [20] Madry, A., Makelov, A., Schmidt, L., Tsipras, D., Vladu, A.: Towards deep learning models resistant to adversarial attacks. In: International Conference on Learning Representations (ICLR) (2018) * [21] Pope, P., Balaji, Y., Feizi, S.: Adversarial robustness of flow-based generative models. In: Chiappa, S., Calandra, R. (eds.) Proceedings of the Twenty Third International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 108, pp. 3795–3805. PMLR (26–28 Aug 2020) * [22] Salimans, T., Goodfellow, I., Zaremba, W., Cheung, V., Radford, A., Chen, X., Chen, X.: Improved techniques for training gans. In: Advances in Neural Information Processing Systems (NeurIPS) (2016) * [23] Santurkar, S., Ilyas, A., Tsipras, D., Engstrom, L., Tran, B., Madry, A.: Image synthesis with a single (robust) classifier. In: Advances in Neural Information Processing Systems (NeurIPS) (2019) * [24] Shmelkov, K., Schmid, C., Alahari, K.: How good is my gan? In: Proceedings of the European Conference on Computer Vision (ECCV). pp. 213–229 (2018) * [25] Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J., Wojna, Z.: Rethinking the inception architecture for computer vision. In: Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR) (2016) * [26] Szegedy, C., Zaremba, W., Sutskever, I., Bruna, J., Erhan, D., Goodfellow, I., Fergus, R.: Intriguing properties of neural networks. In: International Conference on Learning Representations (ICLR) (2014) * [27] Wang, Y., Zou, D., Yi, J., Bailey, J., Ma, X., Gu, Q.: Improving adversarial robustness requires revisiting misclassified examples. In: International Conference on Learning Representations (ICLR) (2019) * [28] Wu, D., Xia, S.T., Wang, Y.: Adversarial weight perturbation helps robust generalization. In: Advances in Neural Information Processing Systems (NeurIPS) (2020) * [29] Xu, Q., Huang, G., Yuan, Y., Guo, C., Sun, Y., Wu, F., Weinberger, K.: An empirical study on evaluation metrics of generative adversarial networks. arXiv preprint arXiv:1806.07755 (2018) * [30] Zhao, Y., Zhu, H., Liang, R., Shen, Q., Zhang, S., Chen, K.: Seeing isn’t believing: Towards more robust adversarial attack against real world object detectors. In: Proceedings of the 2019 ACM SIGSAC Conference on Computer and Communications Security. pp. 1989–2004 (2019) ## Appendix 0.A Sampling $z$ Outside Standard Gaussian In this section, we check the effect of sampling the latent $z$ from distributions other than the one used in training. In particular, instead of sampling $z$ from a standard Gaussian distribution, we try the following setups: * • $z\sim\mathcal{N}(\mu,I)$ where $\mu\in\\{0.1,0.2,0.7,0.8,0.9,1.0,2.0,6.0,7.0\\}$. * • $z\sim\mathcal{N}(\mu,I)+\mathcal{U}[0,1]$ where $\mathcal{U}$ is a uniform distribution. * • $z\sim\mathcal{U}[0,1]$ We report the results in Figures 6 and 7, setting the truncation to $\alpha=0.5$. We observe that the effect of the distribution from which $z$ is sampled has a minor effect on the quality of the generated output image from StyleGAN. Therefore, we run our latent attack as an _unconstrained_ optimization. Figure 6: Effect of shifting the mean of the Gaussian distribution on the output visual quality. We notice that, for a truncation level $\alpha=0.5$, shifting the mean of the Gaussian distribution from which we sample the latent $z$ has a _very minor_ effect on the visual quality of the generated images. Figure 7: Sampling from other distributions than standard Gaussian. We analyze the effect of adding a random uniform vector to the sampled $z$ from a standard Gaussian in the first row. In the second row, we sample $z$ from a uniform distribution as opposed to the standard Gaussian. In both cases, and for truncation level of $\alpha=0.5$, we note that StyleGANv2 is capable of producing output images with good visual quality. ## Appendix 0.B Visualizing the Output of StyleGANv2 at Different Truncation Levels In Section 3.3, we argued that FID favours a distribution of images with more artifacts. That is, FID values for a distribution of images generated with truncation of $\alpha=0.7$ are worse than the ones for $\alpha\in\\{1.0,1.3\\}$, while the latter suffer from significantly more artifacts. We visualize some examples in Figure 8 for completeness. Figure 8: Visualizing the output of StyleGANv2 at different truncation levels. We observe that while outputs with $\alpha=0.7$ are more stable in terms of visual quality, the FID for $\alpha\in\\{1.0,1.3\\}$ is better. ## Appendix 0.C Maximizing FID in the $w-$ Space In Section 3.3, we showed the vulnerability of both the FID and StyleGANv2 against perturbations in the latent space $z$. One natural question that could arise is whether this vulnerability propagated to the $w-$ space as well. To that end, we replicate the setup in Section 3.3 with the following procedure: for each $z_{i}\in\mathcal{N}(0,I)$, we map it to the $w-$ space and construct the perturbation $\delta^{*}_{w}$ by solving the following optimization problem: $\displaystyle\delta^{*}_{w}$ $\displaystyle=\operatorname*{arg\,max}_{\delta}~{}\left\|f(\hat{G}(w+\delta))-f(x_{r})\right\|_{2}.$ (9) We note here that $\hat{G}$ is a StyleGANv2 model excluding the mapping layers from the $z-$ space to the $w-$ space. We solve the optimization problem in (9) with 20 iterations of SGD and learning rate of 0.3. We note that the number of iterations is set to a relatively small value compared to the attacks conducted in the $z-$space for computational purposes. We visualize the results in Figure 9. For a truncation value of $\alpha=1.0$, the FID increases from 2.65 to 6.42. We note here that, similar to earlier observations, the FID is providing inconsistent judgement by favouring a distribution with larger artifacts (comparing Figure 9 with the first row of Figure 8). Moreover, even with the small learning rate and number of iterations, we observe the StyleGANv2 is vulnerable against manipulations in the $w-$space. Figure 9: Robustness of FID against perturbations in $w-$space. We analyze the sensitivity of StyleGANv2 and FID against perturbations in the $w-$space. We report an FID value of 6.4, as opposed to 2.65 without perturbations $(\alpha=1.0)$. ## Appendix 0.D Training Details and Code We conducted $\ell_{2}$ PGD adversarial training by solving the problem in Equation (8). At each iteration, we compute the adversary using 2 steps of PGD attack and random initialization with Gaussian noise. We train the network for 90 epochs with SGD optimizer and a learning rate of $0.1$. We drop the learning rate by a factor of $10$ after each 30 epochs. We train on ImageNet’s training set from scratch. We release our implementation and pre-trained models at https://github.com/R-FID-Robustness-of-Quality-Measures-for-GANs. ## Appendix 0.E Attacking R-FID with Larger $\epsilon$ In Section 4.2, we tested the sensitivity of R-FID against pixel perturbations that are limited by an $\epsilon$ budget. In the main paper, we reported the results after attacking R-FID with a budget of $\epsilon\in\\{0.01,0.02,0.03\\}$. For completeness, we conduct experiments with $\epsilon\in\\{0.04,0.05,0.06,0.07,0.08\\}$ for the robust Inception model trained with $\kappa=128$. We find R-FID values of $\\{503.6,663.2,817.1,891,960.7\\}$, respectively. We note that, even under the largest $\epsilon$ value we considered ($\epsilon=0.08$), the R-FID is still one order of magnitude smaller than of FID when being attacked with $\epsilon=0.01$. This provides further evidence to the effectiveness of R-FID in defending against pixel perturbations. ##### Unbounded Perturbations. Here we test the robustness of R-FID against noisy images. In Section 3.2.2, we showed the sensitivity of FID in assigning good scores to noisy images. We replicate our setup from Table 1 for ImageNet and conducted the attack on R-FID. For the optimized noise images (noise images in this case should be assigned low R-FID), we found the R-FID to be 340, significantly higher than when attacking FID (Table 1 reports an FID of 1.05 for random noise images). We note that while better metrics could be proposed in the future, we believe that R-FID is a step towards a more reliable metri—more robust to both pixel and latent perturbations). ## Appendix 0.F Effect of Truncation on R-FID In Section 4.3, and specifically Table 3, we analyzed whether R-FID outputs large values for any pair of distributions. We provided R-FID values for distributions generated from StyleGANv2 with pairs of truncation values $(\alpha_{i},\alpha_{j})$. For completeness, we report the results for the rest of the pairs, including the R-FID between two splits of FFHQ dataset in Table 5. We observe that the R-FID is very small for identical distributions (e.g. two splits of FFHQ, or at the same truncation level (1.0, 1.0)). Moreover, R-FID increases gradually as the distributions differ. This fact confirms our earlier observation that R-FID better discriminates the generated distribution from the real one. Figure 10: Robustness of R-FID against perturbations in the $w-$space. We report R-FID (at $\alpha=1.0$) of 114.3, as opposed to 113.8 without perturbations. Table 5: R-FID between two distributions. We analyze the R-FID between distributions of images generated at different truncation levels. The last column is the R-FID between two non-overlapping splits of the FFHQ dataset. $(\mathcal{D}_{G}(\alpha_{i}),\mathcal{D}_{G}(\alpha_{j}))$ | (0.7, 1.0) | (0.7, 0.9) | (0.9, 1.0) | (1.0, 1.0) | $(\hat{\mathcal{D}}_{R},\hat{\mathcal{D}}_{R})$ ---|---|---|---|---|--- $\kappa=64$ | 10.5 | 4.9 | 0.48 | 0.007 | 0.004 $\kappa=128$ | 9.9 | 4.6 | 0.46 | 0.008 | 0.006 ## Appendix 0.G Maximizing R-FID in the $w-$ Space We replicate our setup in Appendix 0.C to analyze the sensitivity of R-FID against perturbations in the $w-$ space. To that end, we leverage our attack in Equation (9) but replace the pretrained Inception with a robustly trained version with $\kappa=128$. We visualize the results accompanied by the R-FID value in Figure 10. We draw the following observations: (i): The increase in the R-FID under the same threat model is much smaller than the increase of FID (113.8 $\rightarrow$ 114.3 compared to 2.54 $\rightarrow 6.4$). That is, R-FID is more robust than FID against latent perturbations in the $w-$space. (ii): Changes in the R-FID are accompanied by significant changes in the visual quality of the generated image from StyleGANv2. This is similar to the earlier observation noted in Section 4.3. This constitutes further evidence about the effectiveness of R-FID for providing a robust metric against manipulation. ## Appendix 0.H How Large is $\delta^{*}$ In Section 3.3, we constructed $\delta^{*}$ to perturb the latent code in an unbounded fashion. While the random latent $z$ belongs to a standard normal distribution, there are no bounds on how each latent $z$ should look like. Nevertheless, we analyze the latent perturbation $\delta^{*}$ to better assess the robustness under latent perturbations. To that end, we measure the Wasserstein distance between the unperturbed and perturbed latent codes. We found this value to be small ($\sim$0.07 on average across experiments). We attribute such a small value to using a small step size and a moderate number of iterations for solving the optimization problem ## Appendix 0.I Additional Comments on the Motivation This work aims at characterizing the reliability of the metrics used to judge generative models. Such metrics play a sensitive role in determining whether a generative model is doing a better job than the other. Throughout our assessment, we found that both IS and FID can be easily manipulated by perturbing either the pixel or the latent space. That is, GAN designers could potentially improve the scores of their generative model by simply adding small imperceptible perturbations to the generated distribution of images or latents. This makes the IS and FID less trustworthy, urging for more reliable metrics. In this work, we also proposed one possible fix to increase the reliability of FID, by replacing the pretrained InceptionV3 with a robustly trained version. We note, at last, that while better metrics could appear in the future, we conjecture that R-FID will be part of future solutions to this problem. Table 6: Robust Inception Score against pixel perturbations on CIFAR10. $\epsilon$ | 0.0 | $5\times 10^{-3}$ | $0.01$ | random noise ---|---|---|---|--- R-IS | 9.94 | 5.49 | 3.91 | 1.01 ## Appendix 0.J Robust Inception Score (R-IS) Finally and for completeness, we explore the robustness enhancements that the robust model provide to the Inception Score (IS). Thus, we replicate the setup from Table 1 and conduct pixel perturbations on CIFAR10 dataset. We report the results for $\kappa=128$ in Table 6. We observe that the variation of R-IS against pixel perturbations is much more stable than regular IS. For instance, R-IS drops from 9.94 to 5.49 at $\epsilon=5\times 10^{-3}$ compared to IS which drops from 11.54 to 2.62 for the same value of $\epsilon$. Moreover, running the same optimization for constructing noise images with good IS does not yield a good R-IS. This demonstrates an additional advantage of deploying robust models in GANs quality measures. ## Appendix 0.K Additional Visualizations In the main paper, and due to space constraints, we provided only six samples from the analyzed distributions. For completeness and fairer qualitative comparison, we show additional samples from each considered distribution. In particular, we visualize the output of StyleGANv2 after attacking the latent space by: (i): Maximizing FID with truncation $\alpha=0.7$ (Figure 11) (ii): Maximizing FID with truncation $\alpha=1.0$ (Figure 12) (iii): Maximizing R-FID $(\kappa=128)$ with truncation $\alpha=0.7$ (Figure 13) (iv): Maximizing R-FID $(\kappa=128)$ with truncation $\alpha=1.0$ (Figure 14) (v): Maximizing R-FID $(\kappa=64)$ with truncation $\alpha=0.7$ (Figure 15) (vi): Maximizing R-FID $(\kappa=64)$ with truncation $\alpha=1.0$ (Figure 16). Figure 11: Visualizing samples after attacking the latent space $z$ for StyleGANv2 to maximize FID with truncation $\alpha=0.7$. Figure 12: Visualizing samples after attacking the latent space $z$ for StyleGANv2 to maximize FID with truncation $\alpha=1.0$. Figure 13: Visualizing samples after attacking the latent space $z$ for StyleGANv2 to maximize R-FID with $\kappa=128$ and truncation $\alpha=0.7$. Figure 14: Visualizing samples after attacking the latent space $z$ for StyleGANv2 to maximize R-FID with $\kappa=128$ and truncation $\alpha=1.0$. Figure 15: Visualizing samples after attacking the latent space $z$ for StyleGANv2 to maximize R-FID with $\kappa=64$ and truncation $\alpha=0.7$. Figure 16: Visualizing samples after attacking the latent space $z$ for StyleGANv2 to maximize R-FID with $\kappa=64$ and truncation $\alpha=1.0$.
# Pandemic Culture Wars: Partisan Asymmetries in the Moral Language of COVID-19 Discussions Ashwin Rao 1, 2, Siyi Guo 1, 2, Sze-Yuh Nina Wang 3, Fred Morstatter 2, Kristina Lerman 2 ###### Abstract Effective response to the COVID-19 pandemic required coordinated adoption of mitigation measures, like masking and quarantines, to curb virus’s spread. However, political divisions that emerged early in the pandemic hindered consensus on the appropriate response. To better understand these divisions, our study examines a vast collection of COVID-19-related tweets. We focus on five contentious issues: coronavirus origins, lockdowns, masking, education, and vaccines. We describe a weakly supervised method to identify issue- relevant tweets and employ state-of-the-art computational methods to analyze moral language and infer political ideology. We explore how ideological divisions and moral language shape conversations about these issues. Our findings reveal ideological differences in issue salience and the dynamics of moral language. We find that conservatives use more negatively-valenced moral language than liberals, but moral language use by conservatives is less persistent and appears to be driven by dynamics of the news cycle. Furthermore, we find that political elites use moral rhetoric to a greater extent than non-elites across most issues. Examining the evolution and moralization on divisive issues can provide valuable insights into the dynamics of COVID-19 discussions and assist policymakers in better understanding the emergence of ideological divisions. ## Introduction The COVID-19 pandemic presented a major challenge to US society, testing its resilience and capacity to respond to a crisis. Slowing the rapidly spreading illness required coordinated adoption of non-pharmaceutical interventions recommended by public health experts, such as wearing a mask, staying home, social distancing, and getting vaccines once they became available. However, the response to the pandemic became politically polarized, with Republicans and Democrats disagreeing both on the gravity of the crisis and the appropriate measures to address it (Funk and Tyson 2020). Some Republican leaders downplayed the severity of the virus and expressed skepticism about the effectiveness of measures like mask-wearing and social distancing, while some Democratic leaders advocated for more stringent measures like closing schools, businesses, and even parks and beaches. Policy differences in pandemic mitigation emerged at the state level as well. Republican-led states lifted restrictions on businesses and public gatherings earlier than Democratic-led states. When the COVID-19 vaccine was approved, many Democratic-led states introduced vaccine and mask mandates for returning to work and school, while some Republican-led states passed bills to ban such mandates. These political divisions hampered effective pandemic response, leading to more than a million deaths in US, one of the highest mortality rates among advanced economies (Mordecai and Connaughton 2020). To understand how the pandemic response became so polarized, we study a large corpus of pandemic-related tweets (Chen et al. 2020), comprising over 270M tweets from 2.1M users around United States. We focus on five highly politicized and contentious wedge issues: coronavirus origins, lockdowns and business closures, masking, online education, and vaccines. These issues have been previously found to be salient during the pandemic (Schaeffer 2020; Connaughton 2021; Rojas 2020; Luttrell and Trentadue 2023; Chan 2021; Pierri et al. 2022; Rathje et al. 2022). We classify user ideology based on the partisanship of information sources users share and also describe a method for identifying issues raised in a tweet. These models enable us to label millions of tweets and users, thereby allowing us to study polarized discussions at scale. We study online conversations about the pandemic from the lens of Moral Foundations Theory (MFT) (Haidt and Graham 2007). This theory provides a framework for understanding how moral values shape people’s political attitudes and behaviors. MFT proposes that individuals’ values and judgments can be described by five moral foundations: care/harm, fairness/cheating, loyalty/betrayal, authority/subversion, and sanctity/degradation. Individuals and groups vary in how much emphasis they place on each foundation. While both liberals and conservatives consider the care and fairness foundations to be important, conservatives are thought to endorse the latter three foundations to a greater extent than liberals (Graham, Haidt, and Nosek 2009). Studying online conversations from the perspective of MFT can provide insights into the ways in which partisans use moral appeals to articulate and defend their political positions. It can also help us understand how to frame messaging so as to better appeal to the values of each group. We organize our investigations around the following research questions: 1. 1. What COVID-19 issues did liberals and conservatives discuss online? 2. 2. How do liberals and conservatives differ in their use of moral language on different issues? 3. 3. Does moral language persist over time or quickly die out? What moral foundations resonate with partisans on different issues? 4. 4. How does moral language use differ between political elites and non-elites? We propose and evaluate a weakly-supervised method to identify issue-relevant tweets. This novel method extracts relevant phrases from publicly available Wikipedia pages and then uses them to detect issues in the text of tweets. In addition, we use state-of-the-art computational methods to detect moral foundations in tweets (Guo, Mokhberian, and Lerman 2023) and to infer the political ideology of users (Rao et al. 2021). These models enable us to track partisan attention to polarized issues at scale from the start of the COVID-19 pandemic to November 2021. We uncover important differences in the issues that liberals and conservatives discuss over time, as well as their use of moral language over time. We also unveil differences in the moral appeals of political elites and non-elites across the ideological spectrum. Understanding the moralization on divisive issues can provide valuable insights into the emergence of polarization and help public health experts better tailor messaging to the values and concerns of each group. ## Related Work Table 1: Issue Detection: Wikipedia Articles Issue | Wikipedia Articles ---|--- Origins | Plandemic, Investigations into the origin of COVID-19, COVID-19 lab leak theory Lockdowns | COVID-19 lockdowns, COVID-19 protests in the United States, U.S. federal government response to the COVID-19 pandemic, Protests against responses to the COVID-19 pandemic, Social impact of the COVID-19 pandemic in the United States, Stay-at-home order, Social distancing measures related to the COVID-19 pandemic, Quarantine Masking | Face masks during the COVID-19 pandemic, Face masks during the COVID-19 pandemic in the United States, Maskne, Anti-mask sentiment Education | Impact of the COVID-19 pandemic on education, Homeschooling during the COVID-19 pandemic, Impact of the COVID-19 pandemic on education in the United States Vaccines | COVID-19 vaccination in the United States, COVID-19 vaccine, COVID-19 vaccine hesitancy in the United States, COVID-19 vaccine misinformation and hesitancy, Deaths of anti-vaccine advocates from COVID-19, Herman Cain Award Baseline | COVID-19 pandemic, COVID-19 pandemic in the United States, Politics of the United States #### Polarization American society has been divided by “culture wars” (Hunter 1992), with liberals and conservatives disagreeing on controversial issues (also known as wedge issues) like abortion, women’s and LGBTQ+ rights (DiMaggio, Evans, and Bryson 1996; Evans 2003; Abramowitz and Saunders 2008). This polarization was further exacerbated by the COVID-19 pandemic. Partisans disagreed on the virus’s existence and severity, its origins, mitigation strategies like social distancing and masking, and vaccine mandates. Political ideology was found to explain differences in compliance with health guidelines and mandates (Gollwitzer et al. 2020; Grossman et al. 2020). A study of elite messaging on Twitter at the onset of the pandemic (Green et al. 2020) found that Democratic elites discussed the pandemic more frequently and emphasized its threats to public health, while Republican elites focused on China and the pandemic’s impact on businesses. These divisions extend beyond political elites: nearly 3 out of 10 Americans believed that COVID-19 was created in a lab, with 4 in 10 Republicans endorsing this claim (Schaeffer 2020). Government efforts to restrict the spread of COVID-19 through lockdowns were met with resistance in the form of protests and demonstrations, which were largely driven by conservatives: nearly $52\%$ of conservatives in the United States preferred to have fewer COVID-19 restrictions as opposed to only $7\%$ of liberals (Connaughton 2021). These divisions can have important behavioural consequences: following and interacting with right-leaning political elites, influencers and hyper- partisan media sources was found to be associated with vaccine hesitancy in the US (Pierri et al. 2022; Rathje et al. 2022). #### Moral Foundations Differences in moral values may underlie these partisan divisions. We analyze these through the framework of moral foundations theory (MFT), which proposes five moral foundations of care, fairness, in-group loyalty, respect for authority, and purity/sanctity (Haidt and Graham 2007). These foundations can be further categorized into virtues (care, authority, fairness, loyalty & purity) and vices (harm, subversion, cheating, betrayal & degradation). A number of language analysis techniques have been used to quantify the moral foundations in text data (e.g. (Graham and Haidt 2012; Garten et al. 2016; Guo, Mokhberian, and Lerman 2023)), with past work suggesting that moral appeals may increase message diffusion (Brady et al. 2017; Wang and Inbar 2022). To comprehend and mitigate issue-based political divisions, it is crucial to understand how individuals attach moral significance to their viewpoints (Koleva et al. 2012). Some past work has done this in the context of COVID-19. For example, (Díaz and Cova 2022) found that valuing the care foundation was correlated with compliance with COVID-19 public health recommendations. Similarly, (Chan 2021) found that endorsing the fairness and care moral foundations predicted compliance with calls for staying-at-home, wearing masks, and social distancing, while purity predicted compliance with face masks and social distancing. Others have found that framing masking-related health messaging using ideology-matched moral arguments was effective for liberals, but not conservatives (Luttrell and Trentadue 2023). Moral values can also be used to predict county-level vaccination rates: counties with residents who prioritize moral concerns about purity had lower vaccination rates while counties whose residents prioritized fairness and loyalty had higher vaccination rates (Reimer et al. 2022). Some initial work has assessed COVID-19 vaccine discourse on Twitter. (Borghouts et al. 2023) found that liberal tweets about the COVID-19 vaccine expressed care, fairness, liberty111Note that some researchers include a sixth foundation, liberty/oppression. and authority moral foundations more than conservative tweets, while oppression and harm were referenced more by conservatives. Similarly, (Pacheco et al. 2022) found that care/harm was associated with pro-vaccine sentiment whereas, liberty/oppression was correlated with anti-vaccine attitudes. While vaccinations are a critical polarizing issue in the discussion of COVID-19, no work as of yet has explored differences in moral appeals across a broader range of contentious COVID-19 issues. ## Methods ### Data We use a publicly available dataset (Chen et al. 2020) consisting of 1.4B tweets about COVID-19 posted between January 21, 2020 and November 4, 2021. These tweets contained one or more COVID-19-related keywords, such as coronavirus, pandemic, and Wuhan, among others. We utilized Carmen (Dredze et al. 2013), a geo-location identification technique for Twitter data to assign tweets to locations within US. This method leverages tweet metadata, including “place” and “coordinates” objects that encode location information, such as country, city, and geographic coordinates. Additionally, the technique uses mentions of locations in a user’s bio on Twitter to infer their location. A manual review confirmed that this approach was effective in identifying a user’s home state. As a result, we were left with 270 million tweets generated by 2.1 million geo-located users in the United States. (a) Origins (b) Lockdowns (c) Masking (d) Education (e) Vaccines Figure 1: Wordclouds showing phrases identified from Wikipedia articles using SAGE for each issue. We show the top-50 most frequently occurring phrases in tweets. Bigrams and trigrams are connected by __ for better visualization. ### Issue Detection Table 2: Sample tweets highlighting wedge issues in COVID-19 online discussions. Issue | Tweets ---|--- Origins | No matter what the Chinese Communist Party says, given the mounting evidence, the most likely origins for the China virus are the Wuhan labs studying bats and coronavirus. Lockdowns | This is a GREAT idea. We’re all in this together. Take care of each other. #StayHome #TakeItSeriously #FlattenTheCurve #COVID19 Masking | We’re in the middle of a pandemic and y’all are still coughing and sneezing without covering your mouths? Come on now. Education | More glimmers of hope as we “safely” move forward and open up Texas A&M University while containing #COVID19. Vaccines | You are joking right? Zero sympathy for anti-vaxxers who quit their jobs rather than get vaccinated. They put us all at risk and make the pandemic prolonged for the world. As the pandemic progressed, a set of polarizing issues such as origins of the coronavirus, lockdowns and masking mandates, school closures and online schooling and vaccines, emerged and grew to dominate conversations about COVID-19 (Schaeffer 2020; Connaughton 2021; Rojas 2020; Pierri et al. 2022; Rathje et al. 2022). Previous studies (Garimella et al. 2018; Brady et al. 2017) relied on a small set of manually-selected keywords and phrases to harvest issue-relevant tweets. However, there is a lack of data on the reliability of these methods, and a systematic framework to identify issue- relevant conversations is largely missing. While issue detection has analogues to topic-detection (Blei, Ng, and Jordan 2003; Grootendorst 2022), one cannot control for the stochasticity of topics identified in traditional topic- detection approaches. Although (Card et al. 2015) presented a corpus of several thousand news articles annotated with frames on traditional policy issues such as smoking, same-sex marriages, and immigration, there has not been an attempt to propose a framework for identifying COVID-19 issues in social media discourse. The lack of expertly annotated data on the issues we explore in this work prevents us from employing supervised learning techniques. We define the origins issue to encompass discussions surrounding the origins of the pandemic, including topics such as pangolins, gain of function research, wet-markets, and bats. The lockdown issue comprises content pertaining to early state and federal mitigation efforts, such as quarantines, stay-at-home orders, business closures, reopening, and calls for social distancing. The masking issue is defined by discussions on the use of face coverings, mask mandates, mask shortages, and anti-mask sentiment. Education- related content involves tweets about school closures, reopening of educational institutions, homeschooling, and online learning during the pandemic. The vaccines issue pertains to discussions about COVID-19 vaccines, vaccine mandates, anti-vaccine sentiment, and vaccine hesitancy in the US. We describe a weakly-supervised method to harvest relevant keywords from Wikipedia pages discussing these issues (see Table 1). We use SAGE (Eisenstein, Ahmed, and Xing 2011) to identify distinctive keywords and phrases that are relevant to a specific issue. SAGE calculates the deviation in log-frequencies of words from a baseline lexical distribution. As a baseline, we use Wikipedia articles discussing general aspects of the pandemic and politics in the US (see 1). We concatenate these pages for each issue. Depending on how we tokenize the corpus, we can use SAGE to identify issue- relevant n-grams. In this study, we restrict analysis to unigrams, bigrams and trigrams. Using SAGE we then compare tokens from issue-relevant articles to baseline tokens to identify keywords and phrases that uniquely define each issue. We manually verify the keywords identified by SAGE to ensure precision and relevance. Word clouds in Fig. 1 show the top 50 keywords/phrases extracted from Wikipedia articles for each issue. We assess the relevance of a tweet to a issue by the presence of these keywords and phrases. For example, terms such as ‘Wuhan labs’ or ‘wet markets’ would indicate a tweet is about the origins of the pandemic, while ‘cover your mouth’, ‘N-95 masks’ would indicate a masking-related tweet. Examples of tweets discussing these issues are shown in Table 2. Table 3: Inter-rater Agreement and Evaluation of Issue Detection. Issue | Pairwise | Multi- | F1- | Support ---|---|---|---|--- | Cohen’s $\kappa$ | annotator | Score | | | Fleiss’ $\kappa$ | | Origins | $0.73\pm 0.06$ | 0.45 | 0.51 | 25 Lockdowns | $0.79\pm 0.04$ | 0.47 | 0.90 | 108 Masking | $0.92\pm 0.04$ | 0.55 | 0.83 | 101 Education | $0.71\pm 0.07$ | 0.50 | 0.74 | 54 Vaccines | $0.87\pm 0.05$ | 0.53 | 0.92 | 84 ##### Validation We evaluate issue detection on a subset of tweets, which we draw at random with the constraint that there are at least five tweets for each combination of issues and moral foundation. This gave us a test set of 784 tweets, which were then labeled by five trained annotators. Table 3 shows that the mean Cohen’s Kappa between each pair of annotators for different issues were all above 0.7. The Multi-annotator Fleiss’ Kappa values are lower because it relies on agreement among all five annotators rather than pairs of annotators. Finally, we evaluate performance of issue detection method on this test set and show that F1 scores for most issues are above 0.7 (Table 3), indicating good model performance. ### Morality Detection We capture the moral sentiments in tweets based on the Moral Foundations Theory. We train our morality detection model on top of the transformer-based pretrained language model BERT (Devlin et al. 2018). We train BERT with three Twitter datasets, including a manually annotated COVID dataset (Rojecki, Zheleva, and Levine 2021), the Moral Foundation Twitter Corpus dataset with six different topics (Hoover et al. 2020), and a dataset of political tweets published by US congress members (Johnson and Goldwasser 2018). We test our model on a test subset from the COVID dataset and show the F1 scores from 10 random runs in Table 4. The method has good performance on the major categories such as care and harm, although the performance inevitably varies with the amount of support for different moral categories, which is also observed in previous studies (Hoover et al. 2020; Trager et al. 2022). However, by fusing an in-domain training set that is also about COVID-19, along with other datasets consisting of various topics, we are improving the model generalizability when applied to the target data (Guo, Mokhberian, and Lerman 2023). We compare our method with a widely used dictionary-based method Distributed Dictionary Representations (DDR) (Garten et al. 2018) and find our method outperforms DDR for most moral categories (Refer Table 4). Table 4: Evaluation of Morality Detection Method on Test Data. We show the mean and standard deviation of F1 scores over 10 random runs, and the median of support of 10 runs. Moral Foundation | DDR | Ours | Support ---|---|---|--- Care | $0.71\pm 0.02$ | $\mathbf{0.82\pm 0.02}$ | 239 Harm | $0.21\pm 0.04$ | $\mathbf{0.75\pm 0.03}$ | 111 Fairness | $0.09\pm 0.08$ | $\mathbf{0.24\pm 0.17}$ | 5 Cheating | $0.10\pm 0.12$ | $\mathbf{0.24\pm 0.18}$ | 7 Loyalty | $0.33\pm 0.08$ | $\mathbf{0.59\pm 0.06}$ | 56 Betrayal | $0.03\pm 0.02$ | $\mathbf{0.16\pm 0.07}$ | 11 Authority | $0.00\pm 0.00$ | $\mathbf{0.68\pm 0.09}$ | 42 Subversion | $0.20\pm 0.15$ | $\mathbf{0.35\pm 0.15}$ | 11 Purity | $\mathbf{0.47\pm 0.26}$ | $0.33\pm 0.11$ | 8 Degradation | $0.05\pm 0.04$ | $\mathbf{0.20\pm 0.35}$ | 2 ### Ideology Detection A number of methods identify the political ideology of social media users, with recent works (Le et al. 2019; Nikolov, Flammini, and Menczer 2020; Cinelli et al. 2021; Rao et al. 2021) leveraging URLs that users share in tweets. We use the method discussed in (Rao et al. 2021) to estimate the ideology of individual users. This method relies on Media Bias-Fact Check (MBFC) (Check 2023) to compute ideology scores for individuals. MBFC provides ideological leaning for over 6K Pay-Level Domains (PLDs) with leaning categorized as Left/Hardline Liberal ($0$), Left-Center ($0.25$), Least- Biased/Center ($0.5$), Right-Center ($0.75$), Right/Hardline Conservative ($1$). Each individual’s ideological leaning is estimated to be the weighted average of the leanings of the URLs they share. These scores are then binarized with a threshold of $\leq 0.4$ for liberals ($0$) and $\geq 0.6$ for conservatives ($+1$). The model described in (Rao et al. 2021) then leverages a fastText embedding model pre-trained on Twitter data (Pagliardini, Gupta, and Jaggi 2018), to generate embeddings for user’s tweets over time. These embeddings serve as features to train a Logistic Regression classifier to predict the binarized ideological leanings for all users, including users who have not shared URLs from domains in the MBFC. We were able to identify ideology for 2.1M US users comprising of 1.6M liberal users and 500K conservatives. ##### Validation We assess the level of agreement between the ideology scores estimated by the methods proposed in (Rao et al. 2021) and (Barberá 2015), which uses the follower graph to estimate user ideology. We discretize the continuous ideal point estimates in (Barberá 2015) using a threshold of $<0$ for liberals and $>0$ for conservatives, which we respectively label as $0$ and $+1$. We identify approximately 35K users who appear in both datasets and calculate the F1-score between the two models. The resulting F1-score of $0.75$ and Jaccard overlap of $0.87$ indicates a high level of agreement between the two methods. Figure 2: Dynamics of attention to issues. Daily fraction of (a) original tweets, (b) retweets and (c) replies related to each issue. We use 7-day rolling average to reduce noise. Major events are marked with vertical lines of the same color as the issue and placed along the timeline. Figure 3: Difference in the daily share of original tweets by liberals and conservatives on issues. Positive (negative) values indicate that the issue is $x\%$ more significant for liberals (resp. conservatives). Seven-day rolling average is leveraged to reduce noise. Vertical lines mark significant events that brought shifts in issue-related discussions. (a) Origins (b) Lockdowns (c) Masking (d) Education (e) Vaccines Figure 4: Partisan differences in framing each issue. Values show the log odds ratio of phrases used by liberals (blue) and conservatives (red) in issue- relevant tweets across the time period. ## Results Figure 2 shows the daily share of tweets, retweets and replies discussing each issue, smoothed using a 7-day rolling window, with vertical lines marking major issue-related events. The five issues we monitored accounted for small share of all messages at the beginning of the pandemic, but grew to dominate all discussions. Lockdowns were among the first mitigation efforts carried out by state governments, with California the first state in the mainland US to issue a stay-at-home order on March 19, 2020. This date marks the start of an increase in lockdown related discussions. Although the peak in the masking discussions did not occur until the Black Lives Matter protests in June 2020, Fig. 2 indicates an early upward trend on April 3rd, 2020, when the CDC recommended face coverings to prevent infections. On July 8, 2020 President Trump promised to cut funding to schools that did not reopen for in-person instruction, which is associated with a spike in education related discussions. Similarly, Pfizer-BioNTech’s announcement (Pfizer 2020) on November 9, 2020 that their COVID-19 vaccine was highly effective in Phase 3 clinical trials created a spike in vaccine related tweets. A spike in the discussions of COVID-19 origins is seen on May 23, 2021 when the Wall Street Journal published an article citing an US Intelligence Report that three researchers from the Wuhan Institute of Virology had sought hospital care in November 2019 shortly before the outbreak (Gordon, Strobel, and Hinshaw 2021). The story marked the first major report on the origins of the pandemic by a major news organization. These results provide a qualitative check of robustness of our issue detection approach, as well as insights into importance of issues at different times. Fig. 2 also shows differences in the engagement on different issues. For example, tweets about masking seemed to garner more replies, suggesting that they may have generated more discussion or controversy. (a) Time series: Education (b) Auto-correlation: Education Figure 5: Partisan asymmetries in the persistence of moral language. (a) Time series of moral language used by liberal and conservative accounts in the discussions of education, separated by moral foundation. We show only the three more widely present foundations, and for benefits of visualization, reflect the time series of conservative tweets about the x-axis. White line shows the difference between the two groups, indicating which side mentions the issue more. (b) Auto-correlation function (ACF), showing decay of the correlation of the time series with its lagged version for different lags (in days). (a) ACF: Pre-Vaccine (b) ACF: Post-Vaccine Figure 6: Persistence of moral language. Each cell of the heatmap shows the minimum lag time in day required for the auto-correlation function to drop below the upper confidence level when time series are split into the pre- vaccine period (before 12/11/2020) and post-vaccine period. (a) Origins (b) Lockdowns (c) Masking (d) Education (e) Vaccines Figure 7: Partisan differences in moral language use along each issue. Values indicate percentage of tweets shared by liberals (blue) and conservatives (red) containing the specific moral foundation. Values for Care and Harm are downscaled by a factor of two for better visualization. Figure 8: Moral language differences between elites and non-elites. Boxplot shows the differences in the share of daily tweets made by each group on the issues: (a) origins, (b) lockdowns,(c) masking, (d) education and (e) vaccines. ### Issue Salience by Ideology Next, we examine the relative prevalence of issues in tweets by conservatives and liberals. Since the Twitter user population in our sample is ideologically unbalanced, with three times as many liberals and conservatives, we cannot directly compare the volume of messages posted by each side. Instead, we compare the share of all messages posted by liberals and conservatives that are devoted to each issue. Specifically, we calculate the difference $\Delta$ along an issue $i$ at time (day) $d$ as follows: $\Delta(i,d)=(\frac{\tau_{l}(i,d)}{\tau_{l}(d)}-\frac{\tau_{c}(i,d)}{\tau_{c}(d)})$ where $\tau_{l}$ and $\tau_{c}$ represent the number of all original tweets made by liberals and conservatives on that day respectively, and $\tau_{l}(i,d)$ and $\tau_{c}(i,d)$ represent the number of original tweets about the issue posted by each group. Fig. 3 illustrates the difference in the share of issue-relevant original tweets posted by liberals and conservatives. Positive (negative) values on the y-axis indicate higher prevalence for liberals (conservatives). Vertical lines mark events where significant shifts in conversations occurred. While liberals discussed education more than conservatives, conservatives discussed lockdowns and COVID-19 origins more than liberals. Early in the pandemic, masking was an issue marginally more important to liberals, but after George Floyd’s killing, it was more discussed among conservatives. Ideological divisions about masking remained small, however, until the summer of 2021. We also see interesting changes in vaccine-related discussions. While there were minimal differences between liberals and conservatives prior to the vaccine release, we see that liberals were discussing the vaccine more than conservatives after it was made available in the US on December 11, 2020. However, following vaccine mandates for federal workers on September 9, 2021 (Biden 2021), the issue became a greater focus for conservatives, likely due to their outrage about the mandates. After the Wall Street Journal’s report (Gordon, Strobel, and Hinshaw 2021) in May 2021 claiming that COVID-19 originated at the Wuhan Institute of Virology, there was an increase in conservative’s share of COVID-19 origins tweets. While we limit our discussion here to original tweets due to space constraints, we see similar trends in our analysis of retweets and replies. To characterize differences in liberal and conservative tweets about issues, we identify adjectives that were used to modify issue-relevant phrases (anchors) extracted earlier. We leverage SpaCy’s Dependency Parser to extract dependency relations of the form: XX-$amod\rightarrow ANC$ and XX–$amod\rightarrow YY\leftarrow amod$–$ANC$, where $XX$ refers to the extracted adjective and $ANC$ refers to an issue-relevant phrase. We then pair the extracted adjective with its corresponding anchor as a phrase of the form $XX$_$ANC$ and compute the log-odds ratio of these phrases being expressed in liberal and conservative tweets. Wordclouds in Fig. 4 depict the top 10 most- likely used phrases in liberal (blue) and conservative (red) issue-discourse. Conservatives employ conspiratorial phrases such as “batshit conspiracy”, “fake plandemic”, and “military bioweapon” to discuss the origins issue. In contrast, liberals use phrases like “likely bats” and “medical virology”, suggesting uncertainty. When it comes to lockdowns, conservatives express negative sentiment, framing them as “deadly”, “unconstitutional”, and “stringent”. Liberals, however, approach the topic cautiously, using phrases such as “premature reopening”, “safe reopening”, and “soft lockdowns”. Similar patterns emerge in the discourse on masking, with conservatives deeming mask mandates as “unconstitutional” and “useless”, while liberals focus on “reusable masks” and “mandatory masking”. Concerning education, conservatives express distrust in mitigation strategies employed by educational institutions, using terms like “fascist education” and “unmasked students”, while liberals highlight “hybrid learning”, “vulnerable students”, and “early education.” On the issue of vaccines, conservatives exhibit skepticism, referring to “experimental mRNA” and “dangerous vaccines”. In contrast, liberals emphasize the benefits, discussing “free vaccinations” and “mobile vaccines”. ### Persistence of Moral Language The fluctuations in the daily volume of tweets, even when disaggregated by issue (Fig. 2), obscure the dynamics and complexity of moral expression. To better understand how partisans use moral appeals, we split original tweets on a particular issue by moral foundation label and user ideology. Fig. 5 shows the share of tweets posted by liberals and conservatives that express each moral foundation in the discussions of education. To ease visualization, the time series for conservatives tweets is reflected around the x-axis. In this analysis, both virtues and vices of a moral foundation contribute to its presence in a tweet. As with other issues, the care/harm foundation is the most widely expressed dimension of morality, followed by authority/subversion and fairness/cheating. The two remaining moral foundations are far less common. The time series of moral expression displays complex dynamics, with strong symmetry between liberals and conservatives suggesting that both groups are driven by external events. However, there are also notable differences. The care/harm language used by liberals shows characteristic weekly patterns of activity, with dips on weekends, that are less visible in conservative tweets. In contrast, conservative language shows more short-lived spikes that dissipate as quickly as they appear. To quantify differences in the dynamics of moral language, we compute the auto-correlation function (ACF) of each time series. This function measures the correlation between a time series and its own past values at different lags. Fig. 5 shows this function at different lags from 0 to 60 days. By design, ACF starts at one and decays at larger lags, becoming insignificant when it drops below the confidence bound (horizontal line). For liberals, the ACF values for the care/harm and fairness/cheating foundations have peaks at lag 7, 14, 21, etc. days, a signature of weekly pattern in the time series. The ACF for fairness/cheating decays quickly for conservatives, becoming insignificant for lags larger than 10 days. The fast decaying ACF is associated with short-lived spikes in the time series of moral language that occur at irregular time intervals. This pattern suggests that conservatives react to external events but their attention quickly fades. In contrast, a slowly-decaying ACF for liberals suggests that the moral language resonates within a community and continues to be used long after the precipitating event. We measure how long it takes for ACF to decay to insignificant values. Fig. 6 shows these values for different issues discussed by liberals and conservatives. We split the time series into two periods: before the first COVID-19 vaccine became available in the US (December 11, 2020) and after. The interesting differences in the two heatmaps suggest changes in the moral language. In general, the care/harm and authority/subversion foundations resonate with both communities, taking a long time to decay. Moral language tends to persist longer in the discussions of liberals compared to conservatives. The exceptions to this observations are seen in the discussions about vaccines, lockdowns and COVID-19 origins in the post-vaccine period. Given that pandemic origins is a partisan issue of greater interest to conservatives, we expect their discussions on this topic to persist longer, consistently with the trends we observe in the heatmap. In the discussions of lockdowns, care and fairness resonate more for conservatives, while loyalty resonates more for liberals. Purity language is no longer resonant for either community; it is largely driven by the 3–4 day news cycles. Another surprising reversal is for conservative discussions of vaccines, where purity language becomes far more resonant in the post-vaccine period. This likely reflects the worries of conservatives to inject themselves with an “experimental vaccine”. This is even more surprising as purity language becomes less resonant among both liberals and conservatives in the post-vaccine period, making its persistence among conservative discussions of vaccines that much more remarkable. The findings suggest that informational campaigns aimed at different groups should appeal to moral foundations that resonate within each group. Similar analysis of the moral language of retweets shows that ACF quickly decays, with correlations losing significance after three to five days. This suggests that moral language in retweets is driven largely by the news cycle. ### Partisan Differences in Moral Language To study which moral values are important to liberals and conservatives, we measure the share of issue-related tweets posted by each group that expresses a certain moral foundation. We examine moral valence by separately calculating the share for vices and virtues. The proportion $\rho(m,i)$ of original tweets related to issue $i$ that express moral foundation vice or virtue $m$ is: $\rho(m,i)=\frac{\tau(m,i)}{\tau(i)},$ where, $\tau(i)$ returns the number of original tweets on issue $i$. Figure 7 shows partisan differences in the proportions $\rho$. Across issues, we find that care and harm the most prevalent moral virtues and vices, followed by authority and subversion, with very few tweets expressing purity and degradation. Conservatives express more vices than virtues. There are other partisan differences in the use of moral language. For example, while both appealed to subversion, betrayal, and cheating when discussing the origins of COVID-19, liberals focus more on harm (see Fig. 7(a)). This could reflect greater concern among liberals about the potential harms of these speculations, as they led to a rise in anti-Asian hate crimes across the nation (Gover, Harper, and Langton 2020). Across lockdowns, masking, education, and vaccine issues (Figs. 7(b)-(e)), conservatives expressed greater levels of harm and subversion than liberals, while liberals expressed more care. This may indicate greater support for mandates (e.g., school shutdowns and vaccine mandates) among liberals, and opposition by conservatives, consistent with partisan divides on shutdowns, social distancing, and masks (Deane, Parker, and Gramlich 2021). Overall, we find similarities and interesting differences in the moral values that liberals and conservatives appeal to when discussing COVID-19. Conservatives appear to make more negatively valenced moral appeals when it comes to lockdowns, masking, education, and vaccines. ### Differences in Moral Language Use by Elites and Non-Elites Past research has shown that moral language used by political elites (Clifford and Jerit 2013; Wang and Inbar 2021) in the US can influence the opinions and moral reasoning of non-elites (Clayton et al. 2021). In this work, we compare the moral language used by elites and non-elites. Political elites in US include prominent members of Congress, politicians, journalists and media pundits. Previous studies (Wojcieszak et al. 2022; Shugars et al. 2021) have curated extensive lists of Twitter handles for over 4K US political elites. To ensure a fair comparison with non-elites, we randomly sample non-elite original tweets such that the number matches the number of tweets from elites on a given day. The random sampling also ensures that tweets from non-elites with higher overall activity are selected. We then calculate the daily proportion of issue-related tweets that express a certain moral foundation. Specifically, we calculate the proportion $\rho$ of original tweets relevant to issue $i$ that express moral foundation $m$ on day $d$: $\rho(m,i,d)=\frac{\tau(m,i,d)}{\tau(i,d)}$ where, $\tau$ returns the number of original tweets. We calculate this separately for elites and non-elites. To account for variation, we bootstrap non-elite tweets 100 times. Figure 8((a)-(e)) compares moral language use by elites and non-elites across issues. Boxplots show the distribution of the daily share of tweets in each category. We do not show loyalty/betrayal and purity/degradation foundations, as they were not substantially expressed by elites. The lower variance for non-elites is an attribute of bootstrapping. We rely on the non-parametric Mann-Whitney U Test to test the significance of our comparisons. All differences are assessed for significance at $p<0.001$ unless otherwise specified. On the issue of origins (Fig. 8(a)), non-elites express more care, subversion and cheating than elites. Conservative non-elites use more harm language than conservative elites, a difference that is not significant for liberals. Lockdowns, masking, and education tweets all show a pattern of elites using more moral language than non-elites across most moral foundations, with liberal elites and non-elites tending to refer more to care (Fig. 8(b-d)). Discussions of vaccines deviate somewhat from these trends (Fig. 8(e)), with conservative non-elites expressing more harm than both liberals and conservative elites. Conservatives (both elites and non-elites) also express more subversion than liberals. ## Conclusions The COVID-19 pandemic presented a significant challenge to society. The ensuing response to the pandemic quickly became polarized, with liberals and conservatives disagreeing on the severity of the pandemic and appropriate measures to address it. Sharp differences in issue positions were observed in discussions relating to the origins of the pandemic, lockdowns and business closures, masking mandates, disruptions to education, and vaccines. Previous studies (Schaeffer 2020; Connaughton 2021; Rojas 2020; Luttrell and Trentadue 2023; Chan 2021; Pierri et al. 2022; Rathje et al. 2022) attest to the salience of these issues during the pandemic. We use a massive corpus of over 200M tweets (Chen et al. 2020) to assess these ideological divisions. We find that importance of these issues varies by political ideology. While liberals dominated the discussion on education, tweets about COVID-19 origins and lockdowns were much more prevalent among conservative. Tweets about the vaccine were predominantly made by liberals once the vaccine became available, but switched to being more important to conservatives after federal vaccine mandate were announced. We also found strong differences in moral language among liberals and conservatives. We found that tweets expressing the care/harm and authority/subversion foundations were most persistent, suggesting that these values resonated with both groups. Our time series analysis of moral expression suggests that conservatives were more reactive, driven by the dynamics of the news cycle. Our analysis revealed further differences in the moral language of partisans. While conservatives relied on negatively-valenced moral language, like harm and subversion, liberals primarily leveraged care. We also found that political elites in the US tended to use more moral language compared to non-elites, regardless of ideology. In sum, our findings suggest that differences in the expressions of moral values drove ideological divisions on different issues, fueling polarization. However, it is worth noting that our methods have several limitations. We infer geo-location of tweets based on tweet metadata and user bios’ which may not be accurate. Expression of ideological preferences and moral appeals can be both subtle and highly subjective. Methods leveraged to identify ideology and moral foundations, while state-of-the-art, may suffer from inconsistencies and biases given the task’s inherent complexity. While our method to identify issue relevant tweets is highly scalable, less explicit references to issues in tweets may be missed by this method. Future work could test a supervised classifier that may more accurately identify all relevant tweets. While we made an effort to combine available lists of US political elites on Twitter, this list is by no means exhaustive. ### Broader perspective, ethics and competing interests All data used in this study is publicly available (Chen et al. 2020; Check 2023). The study was deemed exempt from review by the Institutional Review Board (IRB) at the University of Southern California, as it relied solely on publicly available data. Our study also adheres to Twitter’s terms of service (Twitter 2022). Tweet objects contain user information and this brings user anonymity into question. Users can restrict their tweets by switching to a private account or by deleting their tweets. Additionally, we preserve user anonymity during analysis by relying on user IDs instead of screen names. However, we acknowledge that many Twitter users may not be aware their data is used for research purposes (Fiesler and Proferes 2018).In this article, we present aggregated statistics to address further mitigate this concern. The authors declare that there are no competing interests. ## References * Abramowitz and Saunders (2008) Abramowitz, A. I.; and Saunders, K. L. 2008. Is polarization a myth? _The Journal of Politics_ , 70(2): 542–555. * Barberá (2015) Barberá, P. 2015. Birds of the same feather tweet together: Bayesian ideal point estimation using Twitter data. _Political analysis_ , 23(1): 76–91. * Biden (2021) Biden, P. J. 2021. Executive Order on Requiring Coronavirus Disease 2019 Vaccination for Federal Employees. https://www.whitehouse.gov. Accessed: 2023-05-06. * Blei, Ng, and Jordan (2003) Blei, D. M.; Ng, A. Y.; and Jordan, M. I. 2003. Latent dirichlet allocation. _JMLR_ , 3(Jan): 993–1022. * Borghouts et al. (2023) Borghouts, J.; Huang, Y.; Gibbs, S.; Hopfer, S.; Li, C.; and Mark, G. 2023. Understanding underlying moral values and language use of COVID-19 vaccine attitudes on twitter. _PNAS nexus_ , 2(3): pgad013. * Brady et al. (2017) Brady, W. J.; Wills, J. A.; Jost, J. T.; Tucker, J. A.; and Van Bavel, J. J. 2017\. Emotion shapes the diffusion of moralized content in social networks. _PNAS_ , 114(28): 7313–7318. * Card et al. (2015) Card, D.; Boydstun, A.; Gross, J. H.; Resnik, P.; and Smith, N. A. 2015. The media frames corpus: Annotations of frames across issues. In _ACL-IJCNLP 2015_ , 438–444. * Chan (2021) Chan, E. Y. 2021. Moral foundations underlying behavioral compliance during the COVID-19 pandemic. _Personality and individual differences_ , 171: 110463. * Check (2023) Check, M. B.-F. 2023. The media bias chart. https://mediabiasfactcheck.com. Accessed: 2023-05-06. * Chen et al. (2020) Chen, E.; Lerman, K.; Ferrara, E.; et al. 2020. Tracking social media discourse about the covid-19 pandemic: Development of a public coronavirus twitter data set. _JMIR public health and surveillance_ , 6(2): e19273. * Cinelli et al. (2021) Cinelli, M.; De Francisci Morales, G.; Galeazzi, A.; Quattrociocchi, W.; and Starnini, M. 2021. The echo chamber effect on social media. _PNAS_ , 118(9): e2023301118. * Clayton et al. (2021) Clayton, K.; Davis, N. T.; Nyhan, B.; Porter, E.; Ryan, T. J.; and Wood, T. J. 2021\. Elite rhetoric can undermine democratic norms. _PNAS_ , 118(23): e2024125118. * Clifford and Jerit (2013) Clifford, S.; and Jerit, J. 2013. How words do the work of politics: Moral foundations theory and the debate over stem cell research. _The Journal of Politics_ , 75(3): 659–671. * Connaughton (2021) Connaughton, A. 2021. Those on ideological right favor fewer COVID-19 restrictions in most advanced economies. Pew Research Center. Accessed: 2023-04-29. * Deane, Parker, and Gramlich (2021) Deane, C.; Parker, K.; and Gramlich, J. 2021. U.S. Public Opinion on the Coronavirus Pandemic. Pew Research Center. Accessed: 2023-05-11. * Devlin et al. (2018) Devlin, J.; Chang, M.-W.; Lee, K.; and Toutanova, K. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. _arXiv preprint arXiv:1810.04805_. * Díaz and Cova (2022) Díaz, R.; and Cova, F. 2022. Reactance, morality, and disgust: the relationship between affective dispositions and compliance with official health recommendations during the COVID-19 pandemic. _Cognition and Emotion_ , 36(1): 120–136. * DiMaggio, Evans, and Bryson (1996) DiMaggio, P.; Evans, J.; and Bryson, B. 1996. Have American’s social attitudes become more polarized? _American journal of Sociology_ , 102(3): 690–755. * Dredze et al. (2013) Dredze, M.; Paul, M. J.; Bergsma, S.; and Tran, H. 2013. Carmen: A twitter geolocation system with applications to public health. In _AAAI workshop on HIAI_ , volume 23, 45. Citeseer. * Eisenstein, Ahmed, and Xing (2011) Eisenstein, J.; Ahmed, A.; and Xing, E. P. 2011. Sparse additive generative models of text. In _ICML-11_ , 1041–1048. * Evans (2003) Evans, J. H. 2003. Have Americans’ attitudes become more polarized?—An update. _Social Science Quarterly_ , 84(1): 71–90. * Fiesler and Proferes (2018) Fiesler, C.; and Proferes, N. 2018. ”Participant” Perceptions of Twitter Research Ethics. _Social Media + Society_ , 1–14. * Funk and Tyson (2020) Funk, C.; and Tyson, A. 2020. Partisan Differences Over the Pandemic Response Are Growing. Pew Research Center. Accessed: 2023-04-29. * Garimella et al. (2018) Garimella, K.; De Francisci Morales, G.; Gionis, A.; and Mathioudakis, M. 2018. Political discourse on social media: Echo chambers, gatekeepers, and the price of bipartisanship. In _WWW-18_ , 913–922. * Garten et al. (2016) Garten, J.; Boghrati, R.; Hoover, J.; Johnson, K. M.; and Dehghani, M. 2016. Morality between the lines: Detecting moral sentiment in text. In _IJCAI 2016 workshop on Computational Modeling of Attitudes_. * Garten et al. (2018) Garten, J.; Hoover, J.; Johnson, K.; Boghrati, R.; Iskiwitch, C.; and Dehghani, M. 2018. Dictionaries and distributions: Combining expert knowledge and large scale textual data content analysis: Distributed dictionary representation. _Behavior Research Methods, Instruments, and Computers_ , 50(1): 344–361. * Gollwitzer et al. (2020) Gollwitzer, A.; Martel, C.; Brady, W. J.; Pärnamets, P.; Freedman, I. G.; Knowles, E. D.; and Van Bavel, J. J. 2020. Partisan differences in physical distancing are linked to health outcomes during the COVID-19 pandemic. _Nature Human Behaviour_ , 1–12. * Gordon, Strobel, and Hinshaw (2021) Gordon, M.; Strobel, W.; and Hinshaw, D. 2021. Intelligence on Sick Staff at Wuhan Lab Fuels Debate on Covid-19 Origin. Wall Street Journal. Accessed: 2023-05-06. * Gover, Harper, and Langton (2020) Gover, A. R.; Harper, S. B.; and Langton, L. 2020. Anti-Asian hate crime during the COVID-19 pandemic: Exploring the reproduction of inequality. _American journal of criminal justice_ , 45: 647–667. * Graham and Haidt (2012) Graham, J.; and Haidt, J. 2012. Moral Foundation Dictionary. https://moralfoundations.org. Accessed: 2023-04-29. * Graham, Haidt, and Nosek (2009) Graham, J.; Haidt, J.; and Nosek, B. A. 2009. Liberals and conservatives rely on different sets of moral foundations. _Journal of personality and social psychology_ , 96(5): 1029. * Green et al. (2020) Green, J.; Edgerton, J.; Naftel, D.; Shoub, K.; and Cranmer, S. J. 2020. Elusive consensus: Polarization in elite communication on the COVID-19 pandemic. _Science advances_ , 6(28): eabc2717. * Grootendorst (2022) Grootendorst, M. 2022. BERTopic: Neural topic modeling with a class-based TF-IDF procedure. _arXiv preprint arXiv:2203.05794_. * Grossman et al. (2020) Grossman, G.; Kim, S.; Rexer, J. M.; and Thirumurthy, H. 2020. Political partisanship influences behavioral responses to governors’ recommendations for COVID-19 prevention in the United States. _PNAS_ , 117(39): 24144–24153. * Guo, Mokhberian, and Lerman (2023) Guo, S.; Mokhberian, N.; and Lerman, K. 2023. A Data Fusion Framework for Multi-Domain Morality Learning. _arXiv preprint arXiv:2304.02144_. * Haidt and Graham (2007) Haidt, J.; and Graham, J. 2007. When morality opposes justice: Conservatives have moral intuitions that liberals may not recognize. _Social justice research_ , 20(1): 98–116. * Hoover et al. (2020) Hoover, J.; Portillo-Wightman, G.; Yeh, L.; Havaldar, S.; Davani, A. M.; Lin, Y.; Kennedy, B.; Atari, M.; Kamel, Z.; Mendlen, M.; Moreno, G.; Park, C.; Chang, T. E.; Chin, J.; Leong, C.; Leung, J. Y.; Mirinjian, A.; and Dehghani, M. 2020. Moral Foundations Twitter Corpus: A Collection of 35k Tweets Annotated for Moral Sentiment. _Social Psychological and Personality Science_ , 11(8): 1057–1071. * Hunter (1992) Hunter, J. D. 1992. _Culture wars: The struggle to control the family, art, education, law, and politics in America_. Avalon Publishing. * Johnson and Goldwasser (2018) Johnson, K.; and Goldwasser, D. 2018. Classification of Moral Foundations in Microblog Political Discourse. In _ACL_ , 720–730. Melbourne, Australia: Association for Computational Linguistics. * Koleva et al. (2012) Koleva, S. P.; Graham, J.; Iyer, R.; Ditto, P. H.; and Haidt, J. 2012. Tracing the threads: How five moral concerns (especially Purity) help explain culture war attitudes. _Journal of research in personality_ , 46(2): 184–194. * Le et al. (2019) Le, H.; Maragh, R.; Ekdale, B.; High, A.; Havens, T.; and Shafiq, Z. 2019. Measuring political personalization of Google news search. In _WWW-19_ , 2957–2963. * Luttrell and Trentadue (2023) Luttrell, A.; and Trentadue, J. T. 2023. Advocating for Mask-Wearing Across the Aisle: Applying Moral Reframing in Health Communication. _Health Communication_ , 1–13. * Mordecai and Connaughton (2020) Mordecai, M.; and Connaughton, A. 2020. Public opinion about coronavirus is more politically divided in U.S. than in other advanced economies. Pew Research Center. Accessed: 2023-04-29. * Nikolov, Flammini, and Menczer (2020) Nikolov, D.; Flammini, A.; and Menczer, F. 2020. Right and left, partisanship predicts (asymmetric) vulnerability to misinformation. _arXiv preprint arXiv:2010.01462_. * Pacheco et al. (2022) Pacheco, M. L.; Islam, T.; Mahajan, M.; Shor, A.; Yin, M.; Ungar, L.; and Goldwasser, D. 2022. A holistic framework for analyzing the covid-19 vaccine debate. _arXiv preprint arXiv:2205.01817_. * Pagliardini, Gupta, and Jaggi (2018) Pagliardini, M.; Gupta, P.; and Jaggi, M. 2018. Unsupervised Learning of Sentence Embeddings using Compositional n-Gram Features. In _NAACL 2018_. * Pfizer (2020) Pfizer. 2020. Pfizer and BioNTech Announce Vaccine Candidate Against COVID-19. https://www.pfizer.com. Accessed: 2023-05-06. * Pierri et al. (2022) Pierri, F.; Perry, B. L.; DeVerna, M. R.; Yang, K.-C.; Flammini, A.; Menczer, F.; and Bryden, J. 2022. Online misinformation is linked to early COVID-19 vaccination hesitancy and refusal. _Scientific reports_ , 12(1): 1–7. * Rao et al. (2021) Rao, A.; Morstatter, F.; Hu, M.; Chen, E.; Burghardt, K.; Ferrara, E.; and Lerman, K. 2021. Political partisanship and antiscience attitudes in online discussions about COVID-19: Twitter content analysis. _Journal of medical Internet research_ , 23(6): e26692. * Rathje et al. (2022) Rathje, S.; He, J. K.; Roozenbeek, J.; Van Bavel, J. J.; and van der Linden, S. 2022\. Social media behavior is associated with vaccine hesitancy. _PNAS Nexus_ , 1(4): pgac207. * Reimer et al. (2022) Reimer, N. K.; Atari, M.; Karimi-Malekabadi, F.; Trager, J.; Kennedy, B.; Graham, J.; and Dehghani, M. 2022. Moral values predict county-level COVID-19 vaccination rates in the United States. _American Psychologist_ , 77(6): 743. * Rojas (2020) Rojas, R. 2020. Masks Become a Flash Point in the Virus Culture Wars. New York Times. Accessed: 2023-04-29. * Rojecki, Zheleva, and Levine (2021) Rojecki, A.; Zheleva, E.; and Levine, L. 2021. The Moral Imperatives of Self-Quarantining. _Annual meeting of the American Political Science Association_. * Schaeffer (2020) Schaeffer, K. 2020. Nearly three-in-ten Americans believe COVID-19 was made in a lab. Pew Research Center. Accessed: 2023-04-29. * Shugars et al. (2021) Shugars, S.; Gitomer, A.; McCabe, S.; Gallagher, R. J.; Joseph, K.; Grinberg, N.; Doroshenko, L.; Welles, B. F.; and Lazer, D. 2021. Pandemics, protests, and publics: Demographic activity and engagement on Twitter in 2020. _Journal of Quantitative Description: Digital Media_ , 1. * Trager et al. (2022) Trager, J.; Ziabari, A. S.; Davani, A. M.; Golazazian, P.; Karimi-Malekabadi, F.; Omrani, A.; Li, Z.; Kennedy, B.; Reimer, N. K.; Reyes, M.; et al. 2022. The Moral Foundations Reddit Corpus. _arXiv preprint arXiv:2208.05545_. * Twitter (2022) Twitter. 2022. Developer Agreement and Policy. https://developer.twitter.com. Accessed: 2023-05-06. * Wang and Inbar (2021) Wang, S.-Y. N.; and Inbar, Y. 2021. Moral-language use by US political elites. _Psychological Science_ , 32(1): 14–26. * Wang and Inbar (2022) Wang, S. Y. N.; and Inbar, Y. 2022. Re-examining the spread of moralized rhetoric from political elites: Effects of valence and ideology. _Journal of Experimental Psychology: General_ , 151: 3292–3303. * Wojcieszak et al. (2022) Wojcieszak, M.; Casas, A.; Yu, X.; Nagler, J.; and Tucker, J. A. 2022. Most users do not follow political elites on Twitter; those who do show overwhelming preferences for ideological congruity. _Science advances_ , 8(39): eabn9418.
# Hadron yield correlation and constituent quark degree of freedom in heavy ion collisions Rui-qin Wang Department of Physics, Qufu Normal University, Shandong 273165, People’s Republic of China Feng-lan Shao<EMAIL_ADDRESS>Department of Physics, Qufu Normal University, Shandong 273165, People’s Republic of China Jun Song Department of Physics, Shandong University, Shandong 250100, People’s Republic of China Qu-bing Xie Department of Physics, Shandong University, Shandong 250100, People’s Republic of China ###### Abstract Based on the assumption of the production of deconfined quark matter, we use a quark combination model to systematically investigate hadron yields in heavy ion collisions from RHIC $\sqrt{s_{NN}}=200,130,62.4$ GeV to SPS $E_{beam}=158,80,40,30,20$ AGeV. We find that as the collision energy is greater than or equal to 30 AGeV the yields of various hadrons, their correlations, in particular, the observables $A=\frac{\overline{\Lambda}~{}k^{-}~{}p}{\Lambda~{}k^{+}~{}\overline{p}}$ and $B=\frac{\Lambda~{}k^{-}~{}\overline{\Xi}^{+}}{\overline{\Lambda}~{}k^{+}~{}\Xi^{-}}$ , are all reproduced; however, as the collision energy drops to 20 AGeV quark combination fails. This indicates that the constituent quark degrees of freedom represent a decisive factor in thermal hadron production above 30 AGeV and seem to be invalid at 20 AGeV. In addition, hadron yields as well as particle ratios at midrapidity in the most central Pb+Pb collisions at $\sqrt{s_{NN}}=5.5$ TeV are predicted. ###### pacs: 25.75.Dw, 25.75.Gz, 25.75.Nq, 25.75.-q ## I INTRODUCTION The question at which collision energy in heavy ion collisions the deconfinement is first reached has attracted more and more attentions in recent years Andronic2009PLB ; YGMa2009nuclex ; Hove1982PLB ; Gazdzicki ; Gazdz ; GAGo99 ; Bugaev2003PLB ; Akkelin . The Beam Energy Scan programme of NA49 experiment at the CERN-SPS has suggested a preliminary answer — around 30 AGeV20A30A2008PRC . The ongoing Beam Energy Scan programme of STAR Collaboration at Brookhaven National Lab provides an opportunity to study it in more detailstarESC ; besRHIC . Once the deconfined hot and dense quark matter is produced in heavy ion collisions, the observables of various thermal hadrons after hadronization, e.g. yields and momentum spectra etc, have some correlations originated from early quark degrees of freedom. One of the most typical examples is the elliptic flow (v2) of hadrons measured at RHIC energies. As both v2 and transverse momentum($p_{T}$) are divided by the constituent quark number of hadron, the rescaled v2 of various baryons and mesons, which are just that of constituent quarks, almost coincide with each other in the intermediate $p_{T}$ range AAdare2007PRL . If the hot and dense quark matter is hadronized by quark (re-)combination/coalescenceFries2003PRL ; Greco2003PRL ; FLShao2005PRC , as is commonly accepted, these correlations of hadrons can be beautifully explained. In quark (re-)combination/coalescence scenario, quarks and antiquarks are available in unbound state before hadronization and they can coalesce freely into various hadrons, and thereby these correlations from early quark degrees of freedom among different hadron species are naturally formed. On the other hand, if the deconfined quark matter is not produced at all in collisions, there is no free quarks and antiquarks (much less their subsequent combination) and these so-called “quark-level” correlations of hadrons maybe disappear or contort. Therefore, we can study whether the deconfinement is achieved by investigating these correlations among various hadrons produced in heavy ion collisions. Hadron yield is one of the most significant observables from which one can obtain a lot of important information on the hot nuclear matter produced at the early stage of relativistic heavy ion collisions. At the high collision energies RHIC and top SPS, the yields of different hadron species have shown an explicit “quark-level” correlation in quark (re-)combination/coalescence scenario A.Bialas1998PLB ; J.Zimanyi2000PLB ; FLShao2005PRC . In this paper, we make an energy scan from RHIC energies $\sqrt{s_{NN}}=200,130,62.4$ GeV to SPS energies $E_{beam}=158,80,40,30,20$ AGeV to study at which collision energy this “quark-level” correlation of hadron yields first breaks. In particular, we define two correlation quantities $A=\frac{\overline{\Lambda}~{}k^{-}~{}p}{\Lambda~{}k^{+}~{}\overline{p}}$ and $B=\frac{\Lambda~{}k^{-}~{}\overline{\Xi}^{+}}{\overline{\Lambda}~{}k^{+}~{}\Xi^{-}}$, sensitive to quark degrees of freedom. The values of A and B are equal to one in the framework of quark (re-)combination/coalesce, independent of models. The deviation of A and B from one or not can be regarded as a possible signal of deconfinement in heavy ion collisions. We apply a quark combination model, which can exclusively describe hadron production and well reproduce the yields and momentum spectra of final-state hadrons in relativistic heavy ion collisionsFLShao2007PRC ; CEShao2009PRC ; JSong2009MPA ; DMWei2008MPA ; YFWang2008CPC ; WHan2009PRC , to carry out the concrete calculations. The paper is organized as follows. In Sec.II, we present the relations among yields of various hadrons in quark (re-)comibination/coalescence scenario. In Sec.III, we use the quark combination model to calculate the yields of various hadrons, their yield ratios and correlation quantities $A=\frac{\overline{\Lambda}~{}k^{-}~{}p}{\Lambda~{}k^{+}~{}\overline{p}}$ and $B=\frac{\Lambda~{}k^{-}~{}\overline{\Xi}^{+}}{\overline{\Lambda}~{}k^{+}~{}\Xi^{-}}$ at midrapidity in the most central A+A collisions at different energies. Sec.IV summaries our work. ## II hadron yields in quark (re-)combination/coalescence scenario Let us start from the general inclusive formula of hadron production in quark (re-)combination/coalescence scenario $\displaystyle N_{M}$ $\displaystyle=$ $\displaystyle\int dp_{1}dp_{2}F_{q\bar{q}}(p_{1},p_{2})\,\mathcal{R}_{M}(p_{1},p_{2})$ (1) $\displaystyle N_{B}$ $\displaystyle=$ $\displaystyle\int dp_{1}dp_{2}dp_{3}F_{qqq}(p_{1},p_{2},p_{3})\,\mathcal{R}_{B}(p_{1},p_{2},p_{3}).$ (2) Here $F_{q\bar{q}}$ ($F_{qqq}$) is the joint quark-antiquark (three quark) distribution. $\mathcal{R}_{M}$ ($\mathcal{R}_{B}$) is the combination function which stands for the formation probability of quark antiquark (three quarks) into a meson (baryon), dominated by chromodynamics. In sudden approximation, it is equal to the overlap between two (three) quark wave functions and the wave function of meson (baryon). Neglecting exotic (multi- quark) states, mesons and baryons exhaust all fate of quarks and antiquarks. One reaches the following relations : $\sum N_{M}+3\sum N_{B}=\sum N_{q}$ and $\sum N_{M}+3\sum N_{\bar{B}}=\sum N_{\bar{q}}$, where $N_{q}$ ($N_{\bar{q}}$) is the quark (antiquark) number of flavor $q$($\bar{q}$). Extracting $N_{q}$ and $N_{\bar{q}}$ from the joint two (three) quark distribution $F_{q\bar{q}}$ ($F_{qqq}$) and putting them out of the integral, one has the following schematic relations between hadron yields and quark numbers after integrating over quark momenta $N_{M(q\bar{q})}\propto C_{M}N_{q}N_{\bar{q}},\hskip 28.45274ptN_{B({qqq})}\propto C_{B}N_{q}N_{q}N_{q},$ (3) and quark number conservation will fix the proportionality coefficient. The effects of combination function on hadron yields are characterized with the factors $C_{M}$ and $C_{B}$, and $C_{M}=C_{\overline{M}}$ and $C_{B}=C_{\overline{B}}$ are assumed. One realized method of quark number conservation during combination is adding a factor $b_{q}$ for each quark flavor in above equations, as did in ALCOR model alcor95 , $\displaystyle N_{M(q\bar{q})}$ $\displaystyle=$ $\displaystyle C_{M}(b_{q}N_{q})(b_{\bar{q}}N_{\bar{q}}),$ (4) $\displaystyle N_{B({qqq})}$ $\displaystyle=$ $\displaystyle C_{B}(b_{q}N_{q})(b_{q}N_{q})(b_{q}N_{q}).$ (5) Then $b_{q}$ can be uniquely determined by quark number conservation. According to Eqs.(4) and (5), we have the following relations between hadrons and the corresponding antihadrons $\displaystyle\frac{\overline{p}}{p}$ $\displaystyle=$ $\displaystyle(\frac{b_{\overline{q}}\overline{q}}{b_{q}q})^{3},\hskip 45.52458pt\frac{k^{+}}{k^{-}}=(\frac{b_{q}q}{b_{\overline{q}}\overline{q}})(\frac{b_{\overline{s}}\overline{s}}{b_{s}s}),$ $\displaystyle\frac{\overline{\Lambda}}{\Lambda}$ $\displaystyle=$ $\displaystyle(\frac{b_{\overline{q}}\overline{q}}{b_{q}q})^{2}\frac{b_{\overline{s}}\overline{s}}{b_{s}s},\hskip 25.6073pt\frac{\Xi^{-}}{\overline{\Xi}^{+}}=(\frac{b_{s}s}{b_{\overline{s}}\overline{s}})^{2}\frac{b_{q}q}{b_{\overline{q}}\overline{q}}.$ Here, we use particle symbols stand for their numbers for short, $q$ for light quark number and $s$ for strange quark number. Hiding the quark content in hadron yield, we obtain the following interesting relations among different hadron species $\frac{\overline{\Lambda}~{}k^{-}}{\Lambda~{}k^{+}}=\frac{\overline{p}}{p},\hskip 28.45274pt\frac{\Lambda~{}k^{-}}{\overline{\Lambda}~{}k^{+}}=\frac{\Xi^{-}}{\overline{\Xi}^{+}}.$ We define correlation quantities $A$ and $B$ as follows $A=\frac{\overline{\Lambda}~{}k^{-}~{}p}{\Lambda~{}k^{+}~{}\overline{p}}\hskip 28.45274ptB=\frac{\Lambda~{}k^{-}~{}\overline{\Xi}^{+}}{\overline{\Lambda}~{}k^{+}~{}\Xi^{-}}.$ (6) If the quark matter exists and hadronizes via sudden (re-)combination/coalesce in heavy ion collisions, A and B should be equal to one for directly produced hadrons. It is a general result under the constraint of quark number conservation, which is independent of specific models. Different from the well-known recombination and coalescence models Fries2003PRL ; Greco2003PRL , the quark combination model FLShao2005PRC ; QBXie1988PRD is unique for its combination rule. The main idea of the combination rule is to line up the (anti)quarks in a one-dimensional order in phase space, e.g., in rapidity, and then let them combine into initial hadrons one by one according to this orderFLShao2005PRC . Three (anti)quarks or a quark-antiquark pair in the neighborhood form a (anti)baryon or a meson, respectively. At last all quarks and antiquarks are combined into hadrons. The relations between hadron yields and the corresponding quark numbers are easily obtained CEShao2009PRC . With this rule, the model can give the yields and momentum distributions of all hadrons (included in the model) in an event, possessing some exclusive nature. The decay of the short-life resonances is systematically taken into account in the model. The model has been realized in Monte Carlo program and has described many properties of hadron production in relativistic heavy ion collisions FLShao2007PRC ; CEShao2009PRC ; JSong2009MPA ; DMWei2008MPA ; YFWang2008CPC ; WHan2009PRC . ## III RESULTS AND DISCUSSIONS In this section, we use the quark combination model to calculate the hadron yields and their ratios as well as correlation quantities $A=\frac{\overline{\Lambda}~{}k^{-}~{}p}{\Lambda~{}k^{+}~{}\overline{p}}$ and $B=\frac{\Lambda~{}k^{-}~{}\overline{\Xi}^{+}}{\overline{\Lambda}~{}k^{+}~{}\Xi^{-}}$ at midrapidity in the most central A+A collisions at $\sqrt{s_{NN}}=200,~{}130,~{}62.4$ GeV and $E_{beam}=158,~{}80,~{}40,~{}30,~{}20$ AGeV. The predictions at LHC are also presented. The necessary input of the model, i.e., quark distribution just before hadronization, can be obtained by applying the hydrodynamics to describe the evolution of hot and dense quark matter just before hadronization, see Appendix A for details. ### III.1 Hadron yields, their ratios and correlation quantities A and B Table 1: The calculated hadron yields $dN/dy$ at midrapidity in the most central A+A collisions at different energies. The experimental data are taken from Refs20A30A2008PRC ; 20A30A40A80A158A2006PRC ; 20A30A40A80A158A2008PRC ; 40A80A158A2004PRL ; 40A158A2005PRL ; 158A2002PRC ; 622009nuclex ; 1302002PRL ; 1302004PRL ; 1302004PRC ; 2002004PRC ; 2002007PRL . | Au+Au 200 GeV | Au+Au 130 GeV | Au+Au 62.4 GeV | Pb+Pb 158 AGeV ---|---|---|---|--- | data | model | data | model | data | model | data | model $\pi^{+}$ | $286.4\pm 24.2$ | 281.0 | $276\pm 3\pm 35.9$ | 267.2 | $233\pm 17$ | 227.4 | $170.1\pm 0.7\pm 9$ | 165.1 $\pi^{-}$ | $281.8\pm 22.8$ | 281.8 | $270\pm 3.5\pm 35.1$ | 270.5 | $237\pm 17$ | 233.5 | $175.4\pm 0.7\pm 9$ | 176.5 $k^{+}$ | $48.9\pm 6.3$ | 48.6 | $46.7\pm 1.5\pm 7.0$ | 45.2 | $37.6\pm 2.7$ | 38.3 | $29.6\pm 0.3\pm 1.5$ | 27.2 $k^{-}$ | $45.7\pm 5.2$ | 46.1 | $40.5\pm 2.3\pm 6.1$ | 42.4 | $32.4\pm 2.3$ | 32.2 | $16.8\pm 0.2\pm 0.8$ | 16.2 $p$ | $18.4\pm 2.6$ | 17.0 | $28.7\pm 0.9\pm 4.0$ | 25.7 | $29.0\pm 3.8$ | 29.1 | $29.6\pm 0.9\pm 2.9$ | 30.1 $\overline{p}$ | $13.5\pm 1.8$ | 12.5 | $20.1\pm 1.0\pm 2.8$ | 18.2 | $13.6\pm 1.7$ | 13.5 | $1.66\pm 0.17\pm 0.16$ | 1.91 $\Lambda$ | $16.7\pm 0.2\pm 1.1$ | 15.3 | $17.3\pm 1.8\pm 2.8$ | 14.5 | $14.9\pm 0.2\pm 1.49$ | 13.7 | $10.9\pm 1.0\pm 1.3$ | 13.3 $\overline{\Lambda}$ | $12.7\pm 0.2\pm 0.9$ | 12.1 | $12.7\pm 1.8\pm 2.0$ | 10.9 | $8.02\pm 0.11\pm 0.8$ | 7.28 | $1.62\pm 0.16\pm 0.2$ | 1.66 $\Xi^{-}$ | $2.17\pm 0.06\pm 0.19$ | 2.05 | $2.04\pm 0.14\pm 0.2$ | 1.93 | $1.64\pm 0.03\pm 0.014$ | 1.67 | $1.44\pm 0.10\pm 0.15$ | 1.18 $\overline{\Xi}^{+}$ | $1.83\pm 0.05\pm 0.20$ | 1.69 | $1.74\pm 0.12\pm 0.17$ | 1.53 | $0.989\pm 0.057\pm 0.057$ | 1.02 | $0.31\pm 0.03\pm 0.03$ | 0.23 $\Omega^{-}$ | | | | | | | $0.14\pm 0.03\pm 0.01$ | 0.11 $\overline{\Omega}^{+}$ | $0.53\pm 0.04\pm 0.04$ | 0.56 | $0.56\pm 0.11\pm 0.06$ | 0.52 | $0.356\pm 0.046\pm 0.014$ | 0.379 | $0.07\pm 0.02\pm 0.01$ | 0.04 $\chi^{2}/ndf$ | $2.8/8$ | $2.0/8$ | $2.1/8$ | $8.5/9$ | Pb+Pb 80 AGeV | Pb+Pb 40 AGeV | Pb+Pb 30 AGeV | Pb+Pb 20 AGeV ---|---|---|---|--- | data | model | data | model | data | model | data | model $\pi^{+}$ | $132.0\pm 0.5\pm 7$ | 129.9 | $96.6\pm 0.4\pm 6$ | 97.8 | $83.0\pm 0.4\pm 4.2$ | 84.8 | $72.9\pm 0.3\pm 3.6$ | 73.6 $\pi^{-}$ | $140.4\pm 0.5\pm 7$ | 141.8 | $106.1\pm 0.4\pm 6$ | 110.1 | $96.5\pm 0.5\pm 4.8$ | 99.3 | $84.8\pm 0.4\pm 4.2$ | 85.0 $k^{+}$ | $24.6\pm 0.2\pm 1.2$ | 23.8 | $20.1\pm 0.3\pm 1.0$ | 19.4 | $21.2\pm 0.8^{+1.5}_{-0.9}$ | 21.6 | $16.4\pm 0.6\pm 0.4$ | 16.6 $k^{-}$ | $11.7\pm 0.1\pm 0.6$ | 11.9 | $7.58\pm 0.12\pm 0.4$ | 7.27 | $7.8\pm 0.1\pm 0.2$ | 7.3 | $5.58\pm 0.07\pm 0.11$ | 5.01 $p$ | $30.1\pm 1.0\pm 3.0$ | 31.5 | $41.3\pm 1.1\pm 4.1$ | 38.6 | $42.1\pm 2.0\pm 4.2$ | 37.7 | $46.1\pm 2.1\pm 4.6$ | 37.0 $\overline{p}$ | $0.87\pm 0.07\pm 0.09$ | 0.83 | $0.32\pm 0.03\pm 0.03$ | 0.36 | $0.16\pm 0.02\pm 0.02$ | 0.17 | $0.06\pm 0.01\pm 0.006$ | 0.07 $\Lambda$ | $13.5\pm 0.7\pm 1.0$ | 14.5 | $15.3\pm 0.6\pm 1.0$ | 14.7 | $14.7\pm 0.2\pm 1.2$ | 14.2 | $13.4\pm 0.1\pm 1.1$ | 11.5 $\overline{\Lambda}$ | $1.06\pm 0.08\pm 0.1$ | 0.95 | $0.42\pm 0.04\pm 0.04$ | 0.49 | $0.21\pm 0.02\pm 0.02$ | 0.20 | $0.10\pm 0.02\pm 0.01$ | 0.08 $\Xi^{-}$ | $1.22\pm 0.14\pm 0.13$ | 1.23 | $1.15\pm 0.11\pm 0.13$ | 1.10 | $1.17\pm 0.13\pm 0.13$ | 1.38 | $0.93\pm 0.13\pm 0.10$ | 0.92 $\overline{\Xi}^{+}$ | $0.21\pm 0.03\pm 0.02$ | 0.15 | $0.07\pm 0.01\pm 0.01$ | 0.08 | $0.05\pm 0.01\pm 0.01$ | 0.07 | $------$ | 0.02 $\Omega^{-}+\overline{\Omega}^{+}$ | $------$ | 0.13 | $0.10\pm 0.02\pm 0.02$ | 0.09 | $------$ | 0.14 | $------$ | 0.07 $\chi^{2}/ndf$ | $2.9/7$ | $3.0/8$ | $5.6/7$ | $15.3/6$ Table 1 shows the hadron density dN/dy at midrapidity in the most central A+A collisions at RHIC energies 200, 130, 62.4 GeV and SPS energies 158, 80, 40, 30, 20 AGeV. The experimental data are taken from Refs20A30A2008PRC ; 20A30A40A80A158A2006PRC ; 20A30A40A80A158A2008PRC ; 40A80A158A2004PRL ; 40A158A2005PRL ; 158A2002PRC ; 622009nuclex ; 1302002PRL ; 1302004PRL ; 1302004PRC ; 2002004PRC ; 2002007PRL . The agreement between the calculation results and the experimental data is good except at 20 AGeV where $\chi^{2}/ndf$ is far greater than one. We note that ALCOR model alcor95 can also describe well the hadron yields at top SPS energy, which gives a cross- verification of quark combination. On the contrary, HIJING or HIJING/B model which is in compliance with the hadronic scenario for the early evolution of heavy ion collisions via a fragmentation hadronization, can not self- consistently explain the data of multi-strange hadrons topor94 ; Csizmadia99 . Fig. 1 shows the ratios of antihadrons to hadrons at midrapidity as the function of collision energy. The filled triangles are the calculated results, and the experimental data are taken from Refs20A30A2008PRC ; 20A30A40A80A158A2006PRC ; 20A30A40A80A158A2008PRC ; 40A80A158A2004PRL ; 158A2002PRC ; 622009nuclex ; 1302002PRL ; 1302004PRC ; 2002004PRC ; 2002007PRL . These ratios are mainly influenced by the no-zero baryon number density. As the collision energy increases, the nuclear transparency power becomes strong and baryon number density at midrapidity becomes small. This results in a rapid increase for ratios of $K^{-}/K^{+}$, $\overline{p}/p$, $\overline{\Lambda}/\Lambda$ with the increasing collision energy. $\pi^{-}/\pi^{+}$ follows a different pattern. At low collision energies $\pi^{-}/\pi^{+}$ is slightly higher than one, while it is close to one at high energies. This is caused by the asymmetry of decay contribution from hyperons and anti-hyperons (e.g. $\Lambda\rightarrow p\,\pi^{-}$). As the collision energy increases, the yields of hyperons are close to that of anti- hyperons and their decay contributions to pion yields are almost the same and therefore the ratio of $\pi^{-}/\pi^{+}$ is close to one. Figure 1: (Color online)The yield ratios of antihadrons to hadrons at midrapidity as the function of the collision energy. The filled symbols are the calculated results, and the experimental data, open symbols with error bar, are from Refs20A30A2008PRC ; 20A30A40A80A158A2006PRC ; 20A30A40A80A158A2008PRC ; 40A80A158A2004PRL ; 158A2002PRC ; 622009nuclex ; 1302002PRL ; 1302004PRC ; 2002004PRC ; 2002007PRL . Fig. 2 shows the ratios of baryons to mesons $p/\pi$ and the strangeness ratios $K/\pi$ at midrapidity as the function of collision energy. The filled triangles (up and down) are the computed results, and the experimental data are taken from Refs20A30A2008PRC ; 20A30A40A80A158A2006PRC ; 158A2002PRC ; 622009nuclex ; 1302002PRL ; 1302004PRC ; 2002004PRC . The big splits between $p/\pi^{+}$ and $\overline{p}/\pi^{-}$ and between $K^{+}/\pi^{+}$ and $K^{-}/\pi^{-}$ at low collision energies is due to the high baryon number density. At 20 AGeV, the result of $p/\pi^{+}$ deviates seriously from the data. This is probably because the participant nucleons are not broken completely in collisions. These nucleon fragments deposited in midrapidity region lead to the extra contribution of proton production besides those from quark combination and lead to the excessively high ratio of $p/\pi^{+}$. Figure 2: (Color online)The values of the relative production yields $p/\pi$ and $k/\pi$ at different energies. The triangles down are the numerical results for $p/\pi^{+}$ and $k^{+}/\pi^{+}$, and the triangles up are the numerical results for $\overline{p}/\pi^{-}$ and $k^{-}/\pi^{-}$. The experimental data, open symbols with error bar, are from Refs20A30A2008PRC ; 20A30A40A80A158A2006PRC ; 158A2002PRC ; 622009nuclex ; 1302002PRL ; 1302004PRC ; 2002004PRC . Following the experimental data in Table 1, A and B at different collision energies are evaluated and the results are presented with squares in Fig.3. We find that the data of A and B at RHIC energies are almost equal to one while at SPS energies they deviate from one; particularly, at 20 AGeV the value of A amounts to two, seriously deviating from one (the data for B are unavailable). This is probably because the decay of resonances will blur A and B to a certain degree. In order to explore the decay effect, we use the quark combination model to compute the values of A and B for the directly produced hadrons and the final-state hadrons, respectively. The dashed lines are the results of directly produced hadrons. Just as analyzed above, the dashed lines keep an invariant value of one for both A and B, independent of collision energy. The very small fluctuations of dashed lines are due to the fact that the rapidity distributions of the formed hadrons are slightly different from those of quarks, which leads to a small amount of hadrons formed by midrapidity quarks escape from the midrapidity region FLShao2007PRC . The filled triangles down are model results of final-state hadrons. We see that as collision energy is greater than 30 AGeV these triangles agree well with the experimental data within statistical uncertainties. Removing the part of resonance decay, the data of A and B for directly produced hadrons should be equal to one, respectively. This suggests the existence of the “quark level” correlation of directly produced hadrons at these collision energies. However, at 20 AGeV the value of A calculated via quark combination for final state hadrons seriously deviates the data. This in itself indicates the quark degrees of freedom do not represent a decisive factor in hadron production at 20 AGeV. Figure 3: (Color online)The correlation quantities A and B as the function of collision energy. The experimental data, open symbols with error bar, are from Refs20A30A2008PRC ; 20A30A40A80A158A2006PRC ; 20A30A40A80A158A2008PRC ; 40A80A158A2004PRL ; 158A2002PRC ; 622009nuclex ; 1302002PRL ; 1302004PRL ; 1302004PRC ; 2002004PRC ; 2002007PRL . From the above analysis of hadron yields, hadron ratios and correlation quantities A and B, we find that as the collision energy is greater than or equal to 30 AGeV the quark combination can describe reasonably well all these observables which indicates the existence of constituent quark degrees of freedom in the energy region. When the collision energy drops to 20 AGeV, however, the quark combination mechanism can not self-consistently describe these quantities, in particular the correlation quantity A. Note that in order to best fit the data of hadron yields, we have adjusted the strangeness parameter $\lambda_{s}$ (see Table 4) to the quite high values at 20, 30 AGeV which are far greater than the saturated values at higher collision energies and are also greater than those at lower energies thermalreview . But even so the quark combination can not self-consistently explain the data at 20 AGeV. In addition, the high strangeness entangled with baryon density leads to the Sawtooth-like shape for calculated final-state A and B at low SPS energies. In fact this peak behavior of strangeness (i.e. $K^{+}/\pi^{+}$ ratio or $\lambda_{s}$ determined mainly by the former ) at low SPS energies has been interpreted in Ref Gazdzicki ; GAGo99 as a result of onset of deconfinement, i.e. the result of strangeness carrier changing from strange hadrons to strange quarks at the onset of deconfinement. The failure of quark combination energy indicates the (partial) disappearance of constituent quark degrees of freedom and closely relates to the onset of deconfinement around 30 AGeV observed by NA49 Collaboration 20A30A2008PRC . ### III.2 Predictions at LHC With the increasing collision energy, the strangeness $\lambda_{s}$ of the hot and dense quark matter in Table 4 tends to be saturated and the squared sound velocity $c_{s}^{2}$ approaches an ideal value $1/3$ and the baryon number density $n_{0}$ decreases regularly. So we take $\lambda_{s}$=0.43, $c_{s}^{2}=1/3$ and $n_{0}=0.0fm^{-3}$ at $\sqrt{s_{NN}}=5.5$ TeV. The initial entropy density of the central point at the beginning of hydrodynamic evolution is taken to be 256.5 /fm3 according to Eq. (13). In Tables 2 and 3, we present our predictions of hadron yields and hadron ratios at LHC energy ($\sqrt{s_{NN}}=5.5$ TeV). Our computed particle ratios are consistent with the results of the thermal model in RefAndronic2009PLB . Table 2: The predicted $dN/dy$ of identified hadrons at midrapidity in the most central ($0-5\%$) Pb+Pb collisions at $\sqrt{s_{NN}}=5.5$ TeV. $\pi^{+}$ | $\pi^{-}$ | $k^{+}$ | $k^{-}$ | $p$ | $\overline{p}$ | $\Lambda$ | $\overline{\Lambda}$ | $\Xi^{-}$ | $\overline{\Xi}^{+}$ | $\Omega^{-}$ | $\overline{\Omega}^{+}$ ---|---|---|---|---|---|---|---|---|---|---|--- 435.5 | 435.5 | 71.6 | 71.6 | 34.6 | 34.6 | 20.5 | 20.5 | 2.9 | 2.9 | 0.4 | 0.4 Table 3: The predicted hadron ratios at midrapidity in the most central Pb+Pb collisions at $\sqrt{s_{NN}}=5.5$ TeV. $p/\pi^{+}$ | $\overline{p}/\pi^{-}$ | $k^{+}/\pi^{+}$ | $k^{-}/\pi^{-}$ | $\Lambda/\pi^{-}$ | $\Xi^{-}/\pi^{-}$ | $\Omega^{-}/\pi^{-}$ ---|---|---|---|---|---|--- 0.079 | 0.079 | 0.165 | 0.165 | 0.047 | 0.007 | 0.001 ## IV Summary In this paper, we used the quark combination model to make a systematical study of the hadron yields in heavy ion collision in a broad collision energy region. At the collision energies where the deconfined quark matter has been created, the yields of various hadrons after hadronization have some correlations inherited from the early quark degrees of freedom. We investigate at which collision energy this “quark-level” correlation of hadron yields first breaks. We apply the hydrodynamics to describe the evolution of deconfined quark matter and to obtain the quark distribution just before hadronization, then utilize the quark combination model to describe hadronization. We find that as the collision energy is greater than or equal to 30 AGeV, the quark combination well reproduce the yields of various hadrons, their ratios and correlation quantities A and B; however, as the collision energy drops to 20 AGeV, the mechanism can not self-consistently describe these quantities. This indicates that the constituent quark degrees of freedom represent a decisive factor in thermal hadron production above 30 AGeV and seem to be invalid at 20 AGeV. It is related to the onset of deconfinement observed at collision energy 20-30 AGeV. Finally, we predict the yields of various hadrons and their ratios in the most central Pb+Pb collisions at $\sqrt{s_{NN}}=5.5$ TeV. ### ACKNOWLEDGMENTS The authors thank Z. T. Liang, Q. Wang and G. Li for helpful discussions. R. Q. Wang would like to thank L. G. Pang and J. Deng for fruitful discussions. The work is supported in part by the National Natural Science Foundation of China under the grant 10775089, 10947007 and 10975092. ## Appendix A Quark distribution just before hadronization Table 4: The values of the initial baryon density $n_{0}$, strangeness factor $\lambda_{s}$ and squared sound velocity $c_{s}^{2}$ in central A+A collisions at different energies. Energy | 200GeV | 130GeV | 62.4GeV | 158AGeV | 80AGeV | 40AGeV | 30AGeV | 20AGeV ---|---|---|---|---|---|---|---|--- $n_{0}$ ($fm^{-3}$) | 0.30 | 0.34 | 0.74 | 1.31 | 1.46 | 1.64 | 1.74 | 1.56 $\lambda_{s}$ | 0.43 | 0.43 | 0.43 | 0.44 | 0.50 | 0.57 | 0.80 | 0.70 $c_{s}^{2}$ | $1/3.1$ | $1/3.4$ | $1/4.0$ | $1/6.0$ | $1/6.0$ | $1/6.0$ | $1/6.0$ | $1/6.0$ The quark distribution just before hadronization is needed for the quark combination to describe the hadron production in relativistic heavy ion collisions. Here, we use relativistic hydrodynamics Landau53 ; DirkH1995NPA ; Kolb0305084nuth to describe the time-space evolution of the hot and dense quark matter before hadronization. The evolution equation of hydrodynamics follows from the local conservation laws for energy, momentum, and other conserved charges, e.g. baryon number, $\displaystyle\partial_{\mu}T^{\mu\nu}(x)=0,~{}~{}(\nu=0,1,2,3)$ (7) $\displaystyle\partial_{\mu}j^{\mu}(x)=0,$ (8) by inserting the ideal fluid decomposition $\displaystyle T^{\mu\nu}(x)=\bigg{(}e(x)+p(x)\bigg{)}u^{\mu}(x)u^{\nu}(x)-g^{\mu\nu}p(x),$ (9) $\displaystyle j^{\mu}(x)=n(x)u^{\mu}(x).$ (10) Here, $u^{\mu}(x)=\gamma(1,v_{x},v_{y},v_{z})$ with $\gamma=1/\sqrt{1-v_{x}^{2}-v_{y}^{2}-v_{z}^{2}}$ is the local four velocity of a thermalized fluid cell; $e(x)$ is the energy density, $p(x)$ the pressure, and $n(x)$ the conserved number density. As the energy density of the fluid cell drops to 1.0 GeV/fm3 (a common criteria of phase transition from Lattice QCDKarsch2002NPA ), we stop the hydrodynamic evolution and let the constituent quarks and antiquarks freeze out according to Cooper-Frye formalism Coopfrye74 $E\frac{dN_{i}}{d^{3}p}=\frac{dN_{i}}{dyp_{T}dp_{T}d\varphi}=\frac{g_{i}}{(2\pi)^{3}}\int_{\Sigma}f_{i}(p\cdot u(x),x)p\cdot d^{3}\sigma(x),$ (11) where $d^{3}\sigma(x)$ is the outward normal vector on the freeze-out surface $\Sigma(x)$, $g_{i}$ the degeneracy factor of quarks ($g_{i}=6$). The phase- space distribution $f$ in the formula is taken to be a local equilibrium distribution, $f_{i}(E,x)=\frac{1}{\exp[(E-\mu_{i}(x))/T(x)]+1}.$ (12) We consider only the freeze-out of light and strange quarks and antiquarks. The three chemical potentials $\mu_{q}$, $\mu_{\bar{q}}$ and $\mu_{s}=\mu_{\bar{s}}$ at freeze-out can be determined uniquely by the global conservation of energy and baryon number plus an ancillary constraint of strangeness. We study in this paper the hadron production in midrapidity region only. The code of KolbKolb2000PRC ; Kolb2001NPA ; Kolb2003PRC for 2+1- dimensional hydrodynamics with the longitudinal boost invariance is used to simulate the evolution of quark system before hadronization. These frozen-out quarks and antiquarks with momentum distributions in Eq. (11) are hadronized by the quark combination modelFLShao2005PRC ; QBXie1988PRD . In this method of hydrodynamics + quark combination, the quark combination model is default setting while some inputs for hydrodynamics need fixing. The first is initial entropy density $s_{0}$ of the central point at the beginning of hydrodynamic evolution. The distribution of transverse entropy density is determined by optical Glauber modelGlauber1959 . We apply the following empirical formula of $s_{0}$entropy2008nuth as the function of collision energy $\sqrt{s_{NN}}$ $s_{0}(\sqrt{s_{NN}})=6.99\times 10^{-3}\Big{(}312.5\log_{10}{\sqrt{s_{NN}}}-64.8\Big{)}^{3/2}.$ (13) The second is the initial baryon number density $n_{0}$ at central point fixed by the data of the net-proton rapidity density. The distribution of transverse baryon number density is also determined by optical Glauber model. The third is the equation of state $p=p(e)$ for the hot and dense quark matter which is taken to be the simplest pattern $p=c_{s}^{2}e$. The squared sound velocity $c_{s}^{2}$ is obtained by fitting the data of transverse momentum spectra of protons. The fourth is the strangeness of hot and dense quark matter denoted by the factor $\lambda_{s}=2\langle s\bar{s}\rangle/\langle u\bar{u}+d\bar{d}\rangle$. In order to make a better description of strange hadrons, we regard it as a free parameter in the present paper. The values of $n_{0}$, $\lambda_{s}$ and $c_{s}^{2}$ in central A+A collisions at different energies are shown in Table 4. We note that the extracted $\lambda_{s}$ around 30 AGeV are quite high which are consistent with the analytic results of thermal model thermalreview . ## References * (1) L. Van Hove, Phys. Lett. B 118, 138 (1982). * (2) M. Gaździcki and D. Röhrich, Z. Phys. C 65, 215 (1995); 71, 55 (1996), and references therein. * (3) M. Gaździcki, Z. Phys. C 66, 659 (1995). * (4) M. Gaździcki and M. I. Gorenstein, Acta Phys. Polon. B 30, 2705 (1999). * (5) M. I. Gorenstein, M. Gaździcki, and K. A. Bugaev, Phys. Lett. B 567, 175 (2003). * (6) S. V. Akkelin and Yu. M. Sinyukov, Phys. Rev. C 73, 034908 (2006). * (7) A. Andronic, P. Braun-Munzinger and J. Stachel, Phys. Lett. B 673, 142(2009). * (8) S. Zhang, J. H. Chen, H. Crawford, D. Keane, Y. G. Ma and Z. B. Xu, arXiv:nucl-ex/0908.3357. * (9) C. Alt et al. (NA49 Collaboration), Phys. Rev. C 77 024903 (2008). * (10) L. Kumar, J. Phys. G 36 064066 (2009), and references therein. * (11) B. I. Abelev et al. STAR Note SN0493: http://drupal.star.bnl.gov/STAR/starnotes/public/sn0493. * (12) A. Adare, et al. (PHENIX Collaboration), Phys. Rev. Lett. 98, 162301 (2007). * (13) R. J. Fries, B. Müller, C. Nonaka, and S. A. Bass, Phys. Rev. Lett. 90, 202303 (2003). * (14) V. Greco, C. M. Ko, and P. Lévai, Phys. Rev. Lett. 90, 202302 (2003). * (15) F. L. Shao, Q. B. Xie and Q. Wang, Phys. Rev. C 71, 044903 (2005). * (16) A. Bialas, Phys. Lett. B 442, 449(1998). * (17) J. Zimányi, T. S. Biró, T. Csörgö and P. Lévai, Phys. Lett. B 472 243 (2000). * (18) F. L. Shao, T. Yao and Q. B. Xie, Phys. Rev. C 75, 034904 (2007). * (19) C. E. Shao, J. Song, F. L. Shao and Q. B. Xie, Phys. Rev. C 80, 014909 (2009). * (20) J. Song, F. L. Shao and Q. B. Xie, Int. J. Mod. Phys. A 24, 1161(2009). * (21) D. M. Wei, F. L. Shao, J. Song and Y. F. Wang, Int. J. Mod. Phys. A 23, 5217(2008). * (22) Y. F. Wang, F. L. Shao, J. Song, D. M. Wei and Q. B. Xie, Chin. Phys. C 32, 976(2008). * (23) W. Han, S. Y. Li, Y. H. Shang, F. L. Shao and T. Yao, Phys. Rev. C 80, 035202 (2009). * (24) T. S. Biró, P. Lévai and J. Zimányi, Phys. Lett. B 347 6 (1995). * (25) Q. B. Xie and X. M. Liu , Phys. Rev. D 38, 2169 (1988). * (26) R. J. Glauber, in“Lectures on T Physics” 1 (W. E. Brittin, L. G. Dunham eds.,Interscience, NY, 1959). * (27) Gregory Kestin and Ulrich Heinz, arXiv:nucl-th/0806.4539. * (28) C. Alt et al. (NA49 Collaboration), Phys. Rev. C 73 044910 (2006). * (29) C. Alt et al. (NA49 Collaboration), Phys. Rev. C 78 034918 (2008). * (30) T. Anticic et al. (NA49 Collaboration), Phys. Rev. Lett. 93 022302 (2004). * (31) C. Alt et al. (NA49 Collaboration), Phys. Rev. Lett. 94 192301 (2005). * (32) S. V. Afanasiev et al. (NA49 Collaboration), Phys. Rev. C 66 054902 (2002). * (33) B. I. Abelev et al. (STAR Collaboration), Phys. Rev. C 79 034909 (2009). * (34) K. Adcox et al. (PHENIX Collaboration), Phys. Rev. Lett. 89 092302 (2002). * (35) J. Adams et al. (STAR Collaboration), Phys. Rev. Lett. 92 182301 (2004). * (36) K. Adcox et al. (PHENIX Collaboration), Phys. Rev. C 69 024904 (2004). * (37) S. S. Adler et al. (PHENIX Collaboration), Phys. Rev. C 69 034909 (2004). * (38) J. Adams et al. (STAR Collaboration), Phys. Rev. Lett. 98 062301 (2007). * (39) V. Topor Pop, A. Andrighetto, M. Morando, F. Pellegrini, R. A. Ricci, G. Segato, arXiv:hep-ph/9407262 (1994). * (40) P. Csizmadia, P. Lévai, S. E. Vance, T. S. Biró, M. Gyulassy and J. Zimányi, J. Phys. G: Nucl. Part. Phys. 25 321 (1999). * (41) P. Braun-Munzinger, K. Redlich and J. Stachel, arXiv:nucl-th/0304013 and references therein. Invited review in: R.C. Hwa, X.N. Wang (Eds.), Quark Gluon Plasma 3, World Scientific, Singapore, 2004. * (42) A. Andronic, P. Braun-Munzinger, J. Stachel, Nucl. Phys. A 772 167 (2006). * (43) L. D. Landau, Izv. Akad. Nauk SSSR Ser. Fiz. 17 51 (1953). * (44) Dirk H. Rischke, Stefan Bernard and Joachim A. Maruhn, Nucl. Phys. A 595 346 (1995). * (45) Peter F. Kolb and Ulrich Heinz, arXiv:nucl-th/0305084. Invited review in: R.C. Hwa, X.N. Wang (Eds.), Quark Gluon Plasma 3, World Scientific, Singapore, 2004. * (46) F. Karsch, Nucl. Phys. A 698, 199 (2002). * (47) F. Cooper and G. Frye, Phys. Rev. D 10 186 (1974). * (48) Peter F. Kolb, Josef Sollfrank and Ulrich Heinz, Phys. Rev. C 62, 054909 (2000). * (49) Peter F. Kolb, U. Heinz, P. Huovinen, K.J. Eskola, K. Tuominen, Nucl. Phys. A 696 197 (2001). * (50) Peter F. Kolb, and Ralf Rapp, Phys. Rev. C 67, 044903 (2003).
# Learning Relationships between Text, Audio, and Video via Deep Canonical Correlation for Multimodal Language Analysis Zhongkai Sun,1 Prathusha K Sarma,2 William A Sethares,1 Yingyu Liang1 1University of Wisconsin-Madison, 2Curai <EMAIL_ADDRESS><EMAIL_ADDRESS><EMAIL_ADDRESS><EMAIL_ADDRESS>Work done while at UW-Madison ###### Abstract Multimodal language analysis often considers relationships between features based on text and those based on acoustical and visual properties. Text features typically outperform non-text features in sentiment analysis or emotion recognition tasks in part because the text features are derived from advanced language models or word embeddings trained on massive data sources while audio and video features are human-engineered and comparatively underdeveloped. Given that the text, audio, and video are describing the same utterance in different ways, we hypothesize that the multimodal sentiment analysis and emotion recognition can be improved by learning (hidden) correlations between features extracted from the outer product of text and audio (we call this text-based audio) and analogous text-based video. This paper proposes a novel model, the Interaction Canonical Correlation Network (ICCN), to learn such multimodal embeddings. ICCN learns correlations between all three modes via deep canonical correlation analysis (DCCA) and the proposed embeddings are then tested on several benchmark datasets and against other state-of-the-art multimodal embedding algorithms. Empirical results and ablation studies confirm the effectiveness of ICCN in capturing useful information from all three views. ## 1 Introduction Human language communication occurs in several modalities: via words that are spoken, by tone of voice, and by facial and bodily expressions. Understanding the content of a message thus requires understanding all three modes. With the explosive growth in availability of data, several machine learning algorithms have been successfully applied towards multimodal tasks such as sentiment analysis (?; ?), emotion recognition (?), image-text retrieval (?), and aiding medical diagnose (?; ?) etc. Among multimodal language sentiment or emotion experiments involving unimodal features (?; ?; ?; ?), it is commonly observed that text based features perform better than visual or auditory modes. This is plausible for at least three reasons: 1) Text itself contains considerable sentiment-related information. 2) Visual or acoustic information may sometimes confuse the sentiment or emotion analysis task. For instance: “angry” and “excited” may have similar acoustic performances (high volume and high pitch) even though they belong to opposite sentiments. Similarly, “sad” and “disgusted” may have different visual features though they both belong to the negative sentiment. 3) Algorithms for text analysis have a richer history and are well studied. Based on this observation, learning the hidden relationship between verbal information and non-verbal information is a key point in multi-modal language analysis. This can be approached by looking at different ways of combining multi-modal features. The simplest way to combine text (T), audio (A) and video (V) for feature extraction and classification is to concatenate the A, V, and T vectors. An alternative is to use the outer product (?; ?) which can represent the interaction between pairs of features, resulting in 2D or 3D arrays that can be processed using advanced methods such as convolutional neural networks (CNNs) (?). Other approaches (?; ?; ?; ?) study multi-modal interactions and intra-actions by using either graph or temporal memory networks with a sequential neural network LSTM (?). While all these have contributed towards learning multi-modal features, they typically ignore the hidden correlation between text-based audio and text-based video. Individual modalities are either combined via neural networks or passed directly to the final classifier stage. However, it is obvious that attaching both audio and video features to the same textual information may enable non-text information to be better understood, and in turn the non-text information may impart greater meaning to the text. Thus, it is reasonable to study the deeper correlations between text-based audio and text-based video features. This paper proposes a novel model which uses the outer-product of feature pairs along with Deep Canonical Correlation Analysis (DCCA) (?) to study useful multi-modal embedding features. The effectiveness of using an outer- product to extract cross-modal information has been explored in (?; ?). Thus, features from each mode are first extracted independently at the sentence (or utterance) level and two outer-product matrices ($T\otimes V$ and $T\otimes A$) are built for representing the interactions between text-video and between text-audio. Each outer-product matrix is connected to a convolutional neural network (CNN) for feature extraction. Outputs of these two CNNs can be considered as feature vectors for text-based audio and text-based video and should be correlated. In order to better correlate the above text-based audio and text-based video, we use Canonical Correlation Analysis (CCA) (?), which is a well-known method for finding a linear subspace where two inputs are maximally correlated. Unlike cosine similarity or Euclidean distance, CCA is able to learn the direction of maximum correlation over all possible linear transformations and is not limited by the original coordinate systems. However, one limitation of CCA is that it can only learn linear transformations. An extension to CCA named Deep CCA (DCCA) (?) uses a deep neural network to allow non-linear relationships in the CCA transformation. Recently several authors (?; ?) have shown the advantage of using CCA-based methods for studying correlations between different inputs. Inspired by these, we use DCCA to correlate text- based audio and text-based video. Text-based audio and text-based video features derived from the two CNNs are input into a CCA layer which consists of two projections and a CCA Loss calculator. The two CNNs and the CCA layer then form a DCCA, the weights of the two CNNs and the projections are updated by minimizing the CCA Loss. In this way, the two CNNs are able to extract useful features from the outer-product matrices constrained by the CCA loss. After optimizing the whole network, outputs of the two CNNs are concatenated with the original text sentence embedding as the final multi-modal embedding, which can be used for the classification. We evaluate our approach on three benchmark multi-modal sentiment analysis and emotion recognition datasets: CMU-MOSI (?), CMU-MOSEI (?), and IEMOCAP(?). Additional experiments are presented to illustrate the performance of the ICCN algorithm. The rest of the paper is organized as follows: Section 2 presents related work, Section 3 introduces our proposed model and Section 4 describes our experimental setup. Section 5 presents a discussion on the empirical observations, Section 6 concludes this work. ## 2 Related Work The central themes of this paper are related to learning (i) multi-modal fusion embeddings and (ii) cross-modal relationships via canonical correlation analysis (CCA). Multi-modal fusion embedding: Early work (?) concatenates the audio, video and text embeddings to learn a larger multi-modal embedding. But this may lead to a potential loss of information between different modalities. Recent studies on learning multi-modal fusion embeddings train specific neural network architectures to combine all three modalities. In their work (?) propose improvements to multi-modal embeddings using reinforcement learning to align the multi-modal embedding at the word level by removing noises. A multi-modal tensor fusion network is built in (?) by calculating the outer-product of text, audio and video features to represent comprehensive features. However this method is limited by the need of a large computational resources to perform calculations of the outer dot product. In their work (?) developed an efficient low rank method for building tensor networks which reduce computational complexity and are able to achieve competitive results. A Memory Fusion Network (MFN) is proposed by (?) which memorizes temporal and long-term interactions and intra-actions between cross-modals, this memory can be stored and updated in a LSTM. (?) learned multistage fusion at each LSTM step so that the multi-modal fusion can be decomposed into several sub-problems and then solved in a specialized and effective way. A multimodal transformer is proposed by (?) that uses attention based cross-modal transformers to learn interactions between modalities. Cross-modal relationship learning via CCA: Canonical Correlation Analysis(CCA) (?) learns the maximum correlation between two variables by mapping them into a new subspace. Deep CCA (DCCA) (?) improves the performance of CCA by using feedforward neural networks in place of the linear transformation in CCA. A survey of recent literature sees applications of CCA-based methods in analyzing the potential relationship between different variables. For example, a CCA based model to combine domain knowledge and universal word embeddings is proposed by (?). Models proposed by (?) use Deep Partial Canonical Correlation Analysis (DPCCA), a variant of DCCA, for studying the relationship between two languages based on the same image they are describing. Work by (?) investigates the application of DCCA to simple concatenations of multimodal- features, while (?) applied CCA methods to learn joint-representation for detecting sarcasm. Both approaches show the effectiveness of CCA methods towards learning potential correlation between two input variables. ## 3 Methodology This section first introduces CCA and DCCA. Next, the interaction canonical correlation network (ICCN), which extracts the interaction features of a CNN in a DCCA-based network, is introduced. Finally, the whole pipe-line of using this method in a multimodal language analysis task is described. ### CCA and DCCA Given two sets of vectors $X\in\mathbb{R}^{n_{1}\times m}$ and $Y\in{\mathbb{R}}^{n_{2}\times m}$, where $m$ denotes the number of vectors, CCA learns two linear transformations $A\in\mathbb{R}^{n_{1}\times r}$ and $B\in\mathbb{R}^{n_{2}\times r}$ such that the correlation between $A^{T}X$ and $B^{T}Y$ is maximized. Denote the covariances of $X$ and $Y$ as $S_{11},S_{22}$, and the cross-covariance of $X,Y$ as $S_{12}$. The CCA objective is $\begin{split}A^{*},B^{*}&=\mathop{\arg\max}_{A,B}\textrm{corr}(A^{T}X,B^{T}Y)\\\ &=\mathop{\arg\max}_{A,B}\frac{A^{T}S_{12}B}{\sqrt{A^{T}S_{11}A}\sqrt{B^{T}S_{22}B}}.\end{split}$ (1) The solution of the above equation is fixed and can be solved in multiple ways (?; ?). One method suggested by (?) lets $U,S,V^{T}$ be the Singular Value Decomposition ($SVD$) of the matrix $Z$ = $S_{11}^{-\frac{1}{2}}S_{12}S_{22}^{-\frac{1}{2}}$. Then $A^{*},B^{*}$ and the total maximum canonical correlation are $\begin{split}A^{*}&=S_{11}^{-\frac{1}{2}}U\\\ B^{*}&=S_{22}^{-\frac{1}{2}}V\\\ \textrm{corr}(A^{*T}X,B^{*T}Y)&=\textrm{trace}(Z^{T}Z)^{\frac{1}{2}}.\end{split}$ (2) One limitation of CCA is that it only considers linear transformations. DCCA (?) learns non-linear transformations using a pair of neural networks. Let $f,g$ denote two independent neural networks, the objective of DCCA is to optimize parameters $\theta_{f},\theta_{g}$ of $f,g$ so that the canonical correlation between the output of $f$ and $g$, denoted as $F_{X}=$ $f(X;\theta_{1})$ and $F_{Y}=$ $g(Y;\theta_{2})$, can be maximized by finding two linear transformations $C^{*},D^{*}$. The objective of DCCA is $\begin{split}\theta_{f}^{*},\theta_{g}^{*}&=\mathop{\arg\max}_{\theta_{f},\theta_{g}}\textrm{CCA}(F_{X},F_{Y})\\\ &=\mathop{\arg\max}_{\theta_{f},\theta_{g}}\textrm{corr}(C^{*T}F_{X},D^{*T}F_{Y}).\end{split}$ (3) In order to update the parameters of $f,g$, a loss for measuring the canonical correlation must be calculated and back-propagated. Let $R_{11},R_{22}$ be covariances of $F_{X},F_{Y}$, the cross-covariance of $F_{X},F_{Y}$ as $R_{12}$. Let $E=R_{11}^{-\frac{1}{2}}R_{12}R_{22}^{-\frac{1}{2}}$. According to (2), the canonical correlation loss for updating $F_{x},F_{Y}$ can be defined by $\textrm{CCA Loss}=-\textrm{trace}(E^{T}E)^{\frac{1}{2}}.$ (4) Networks $f(X;\theta_{f}),g(Y;\theta_{g})$’s parameters can be updated by minimizing the CCA Loss (4) (i.e. maximize the total canonical correlation). ### Text Based Audio Video Interaction Canonical Correlation Previous work of (?; ?) on multi-modal feature fusion has shown that the outer-product is able to learn interactions between different features effectively. Thus, we use the outer-product to represent text-video and text- audio features. Given that outer-product and DCCA are applied at the utterance (sentence)-level, we extract utterance level features for each uni-modal independently in order to test the effectiveness of ICCN more directly. Let $H_{t}\in\mathbb{R}^{d_{t}}$ be the utterance-level text feature embedding, and $H_{v}\in\mathbb{R}^{d_{v}\times l_{v}},H_{a}\in\mathbb{R}^{d_{a}\times l_{a}}$ be the video and audio input sequences. A 1D temporary convolutional layer is used to extract local structure of the audio and video sequences, and the outputs of the 1D-CNN are denoted as $H_{a1}\in\mathbb{R}^{d_{a1}\times l_{a}},H_{v1}\in\mathbb{R}^{d_{v1}\times l_{v}}$. Next, two LSTMs process the audio and video sequences. The final hidden state of each LSTM is used as the utterance-level audio or video feature, denoted as $H_{a2}\in\mathbb{R}^{d_{a2}},H_{v2}\in\mathbb{R}^{d_{v2}}$. Once each utterance-level feature has been obtained, the text-based audio feature matrix and text-based video feature matrix can then be learned using the outer- product on $H_{t},H_{v2},H_{a2}$: $\begin{split}H_{ta}&=H_{t}\otimes H_{a2},H_{ta}\in\mathbb{R}^{d_{t}\times d_{a2}}\\\ H_{tv}&=H_{t}\otimes H_{v2},H_{tv}\in\mathbb{R}^{d_{t}\times d_{v2}}.\end{split}$ (5) In order to extract useful features from the outer-product matrices $H_{ta},H_{tv}$, a Convolutional Neural Network is used as the basic feature extractor. $H_{ta}$ and $H_{tv}$ are connected by multiple 2D-CNN layers with max pooling. Outputs of the two 2D-CNNs are reshaped to 1D vector and then be used as inputs to the CCA Loss calculation. 1D-CNN, LSTM, and 2D-CNN’s weights are again updated using the back-propagation of the CCA Loss (4). Thus the two 2D-CNNs learn to extract features from $H_{tv}$ and $H_{ta}$ so that their canonical correlation is maximized. Figure 1: ICCN method for aligning text-based audio features and text-based video features. Sentence level uni-modal features are extracted independently. Outer-product matrices of text-audio and text-video are used as input to the Deep CCA network. After learning the CNN’s weights using the CCA Loss, outputs of the two CNNs are concatenated with the original text to form the multi- modal embedding. This can be used as input to independent downstream tasks. Algorithm 1 provides the pseudo-code for the whole Interaction Canonical Correlation Network (ICCN). 0: Data $H_{t}\in\mathbb{R}^{d_{t}\times N}$, $H_{a}\in\mathbb{R}^{d_{a}\times l_{a}\times N}$, $H_{v}\in\mathbb{R}^{d_{v}\times l_{v}\times N}$, $epoch$, $\eta$ Initialize $W_{a}$ of ($\textrm{CNN1D}_{a},\textrm{LSTM}_{a}$, and $\textrm{CNN2D}_{ta}$) Initialize $W_{v}$ of ($\textrm{CNN1D}_{v},\textrm{LSTM}_{v}$, and $\textrm{CNN2D}_{tv}$) while $epoch>0$ do $H_{a1}=\textrm{CNN1D}_{a}(H_{a})$ $H_{v1}=\textrm{CNN1D}_{v}(H_{v})$ $H_{a2}=\textrm{LSTM}{a}(H_{a1})$ $H_{v2}=\textrm{LSTM}{v}(H_{v1})$ $H_{tv}=H_{t}\otimes H_{v2}$, $H_{ta}=H_{t}\otimes H_{a2}$ $K_{tv}=\textrm{CNN2D}_{tv}(H_{tv})$ $K_{ta}=\textrm{CNN2D}_{ta}(H_{ta})$ Compute gradients $\nabla W_{v}$, $\nabla W_{a}$ of: $\min\limits_{W_{v},W_{a}}\textrm{CCALoss}(K_{tv},K_{ta})$ Update: $W_{v}\leftarrow W_{v}-\eta\nabla W_{v}$ $W_{a}\leftarrow W_{a}-\eta\nabla W_{a}$ $epoch\leftarrow epoch-1$ end while $K_{tv},K_{ta}$ Algorithm 1 Interaction Canonical Correlation Network ### Pipe-line for Downstream Tasks The ICCN method acts as a feature extractor. In order to test its performance, an additional downstream classifier is also required. Uni-modal features can be extracted using a variety of simple extraction schemes as well as by learning features using more complex neural network based models such as a sequential LSTM. Once uni-modal features for text, video, and audio have been obtained, the ICCN can be used to learn text-based audio feature $K_{ta}$ and text-based video feature $K_{tv}$. The final multi-modal embedding can be formed as the concatenation of the text-based audio, the original text, and the text-based video features, which are denoted as $[K_{ta};H_{t};K_{tv}]$. This $[K_{ta};H_{t};K_{tv}]$ can then be used as an input to downstream classifiers such as logistic regression or multilayer perceptron. Figure 1 shows the pipe-line using the ICCN for downstream tasks in our work. ## 4 Experiment Settings This section describes the experimental datasets and baseline algorithms against which ICCN is compared. ### Datasets The proposed algorithm is tested using three public benchmark multi-modal sentiment analysis and emotion recognition datasets: CMU-MOSI (?), CMU-MOSEI (?), and IEMOCAP(?). Both CMU-MOSI and CMU-MOSEI’s raw features, and most of the corresponding extracted features can be acquired from the CMU- MultimodalSDK (?). * • CMU-MOSI: This dataset is a multi-modal dataset built on 93 Youtube movie reviews. Videos are segmented to 2198 utterance clips, and each utterance example is annotated on a scale of [-3, 3] to reflect sentiment intensity. -3 means an extremely negative and 3 means an extremely positive sentiment. This data set is divided into three parts, training (1283 samples), validation (229 samples) and test (686 samples). * • CMU-MOSEI: This dataset is similar to the CMU-MOSI, but is larger in size. It consists of 22856 annotated utterances extracted from Youtube videos. Each utterance can be treated as an individual multi-modal example. Train, validation, and test sets contain 16326, 1871, and 4659 samples respectively. * • IEMOCAP: This dataset contains 302 videos in which speakers performed 9 different emotions (angry, excited, fear, sad, surprised, frustrated, happy, disappointed and neutral). Those videos are divided into short segments with emotion annotations. Due to the imbalance of some emotion labels, we follow experiments in previous papers (?; ?) where only four emotions (angry, sad, happy, and neutral) are used to test the performance of the algorithm. Train, validation, and test partitions contain 2717, 789, and 938 data samples respectively. ### Multi-modal Features The following uni-modal features are used prior to combinations, * • Text Features: For MOSI and MOSEI, we use a pre-trained transformer model BERT (?) to extract utterance level text features, (many other approaches use Glove word-level embeddings followed by a LSTM). The motivation behinds using BERT is 1) BERT is the state-of-the-art in sentence encoding algorithms and has demonstrated tremendous success in several downstream text applications such as sentiment analysis, question-answering, semantic similarity tasks etc, 2) using BERT simplifies the training pipe-line, with a large focus now towards improving the performance of ICCN on a particular downstream task. We input the raw text to the pre-trained uncased BERT-Base model (without fine-tuning). Sentence encodings output from BERT are used as the text features. Each individual text feature is of size 768. For IEMOCAP, we used InferSent(?) to encode utterance level text. Since data is provided in the form of word indices for GLOVE embeddings rather than raw text, we use InferSent; a BiLSTM layer followed by a max-pooling layer to learn sentence embeddings. * • Audio Features: The audio feature is extracted by using COVAREP (?), which is a public software used for extracting acoustic features such as pitch, volume, and frequency. The CMU-MultimodalSDK provides COVAREP feature sequence for every multi-modal example, the dimension of each frame’s audio feature is 74. * • Video Features: Facet(?) has been used for extracting facial expression features such as action units and face pose. Similarly, every multimodal example’s video feature sequence is also obtained from the CMU-MultimodalSDK. The size of each frame’s video feature is 35. ### Baseline Methods We consider a variety of baseline methods for multi-modal embedding comparison. In order to focus on the multi-modal embedding itself, we input each multi-modal embedding to the same downstream task classifier (or regressor). Experimental comparisons are reported in two parts 1) we report the effectiveness of DCCA over the simpler CCA based methods when used as inputs to the ICCN and 2) we compare ICCN against newer utterance level embeddings algorithms that learn features for a down stream task in an end-to- end fashion. The following baselines are used in our comparisons, * • Uni-modal and Concatenation: This is the simplest baseline in which uni-modal features are concatenated to obtain a multi-modal embedding. * • Linear CCA: CCA (?) considers linear transformations for different inputs. We use the CCA to learn a new common space for audio and video modes, and combine the learned audio and video features with the original text embedding. This is because, (?) showed that using a CCA-based method to correlate audio and video is more effective that correlating audio-text or video-text. * • Kernel-CCA: Kernel-CCA (?) introduces a nonlinearity via kernel maps. Kernel- CCA can be used exactly like CCA. * • GCCA: Generalized CCA (?) learns a common subspace across multiple views. We use GCCA in two ways: 1) use the GCCA output embedding directly and 2) combine the GCCA output embedding with the original text embedding. * • DCCA: A Deep CCA based algorithm proposed by (?). Audio and video features are simply concatenated and then be correlated with text features using DCCA. Outputs of the DCCA are again concatenated with raw text, audio, and video features to formulate the multimodal embedding. In the proposed ICCN algorithm, text features are encoded by a pre-trained BERT transformer or by InferSent. This is unlike most of the state-of-the-art algorithms that obtain sentence level encodings by passing word embeddings through variants of RNNs. However, since the idea is to compare modal features, we also choose the following three state-of-the-art utterance-level (i.e. sentence-level) fusion models (whose core algorithm is agnostic to the text encoding architecture) as additional baselines. To make the comparison fair, these methods use the same features as ours. * • TFN: Tensor Fusion Network (TFN) (?) combines individual modal’s embeddings via calculating three different outer-product sub-tensors: unimodal, bimodal, and trimodal. All tensors will then be flattened and used as a multi-modal embedding vector. * • LMF: Low-Rank Multimodal Fusion (LMF) (?) learns the multimodal embedding based the similar tensor processing of TFN, but with an additional low-rank factor for reducing computation memory. * • MFM: Multimodal Factorization Model (MFM) (?) is consists of a discriminative model for prediction and a generative model for reconstructing input data. A comprehensive multimodal embedding is learned via optimizing the generative- discriminative objective simultaneously. Data View | CMU-MOSI | CMU-MOSEI ---|---|--- | Acc-2 | F-score | MAE | Acc-7 | Corr | Acc-2 | F-score | MAE | Acc-7 | Corr Audio | 45.15 | 45.83 | 1.430 | 16.21 | 0.248 | 58.75 | 59.23 | 0.785 | 38.59 | 0.298 Video | 48.10 | 49.06 | 1.456 | 15.51 | 0.339 | 59.25 | 59.90 | 0.770 | 36.09 | 0.288 Text | 80.80 | 80.17 | 0.897 | 35.92 | 0.688 | 82.83 | 83.02 | 0.582 | 48.76 | 0.681 Text+Video | 81.00 | 80.91 | 0.920 | 35.11 | 0.676 | 82.86 | 83.01 | 0.581 | 47.92 | 0.674 Text+Audio | 80.59 | 80.56 | 0.909 | 35.08 | 0.672 | 82.80 | 82.96 | 0.582 | 49.02 | 0.689 Audio+Video+Text | 80.94 | 81.00 | 0.895 | 36.41 | 0.689 | 82.72 | 82.87 | 0.583 | 50.11 | 0.692 CCA | 79.45 | 79.35 | 0.893 | 34.15 | 0.690 | 82.94 | 83.06 | 0.573 | 50.23 | 0.690 KCCA | 79.82 | 79.91 | 0.889 | 34.76 | 0.689 | 83.05 | 83.14 | 0.574 | 50.09 | 0.692 GCCA | 62.50 | 62.15 | 1.403 | 17.29 | 0.533 | 75.12 | 75.46 | 0.653 | 45.33 | 0.602 GCCA+Text | 77.80 | 77.87 | 1.107 | 25.94 | 0.658 | 82.75 | 82.90 | 0.613 | 46.06 | 0.644 DCCA | 80.60∗ | 80.57∗ | 0.874 | 35.51 | 0.703 | 83.62∗ | 83.75∗ | 0.579 | 50.12 | 0.707 TFN | 80.82 | 80.77 | 0.901 | 34.94 | 0.698 | 82.57 | 82.09 | 0.593 | 50.21 | 0.700 LMF | 82.53 | 82.47 | 0.917 | 33.23 | 0.695 | 82.03 | 82.18 | 0.623 | 48.02 | 0.677 MFM | 81.72 | 81.64 | 0.877 | 35.42 | 0.706 | 84.40 | 84.36 | 0.568 | 51.37 | 0.717 ICCN | 83.07 | 83.02 | 0.862 | 39.01 | 0.714 | 84.18 | 84.15 | 0.565 | 51.58 | 0.713 Table 1: Results for experiments on CMU-MOSI and CMU-MOSEI. Best numbers are in bold. For accuracy, F-score, and Correlation, higher is better. For mean absolute error, lower is better. Results marked with $*$ are reported in original papers. For TFN, LMF, and MFM, we re-did experiments with using our features for a fair comparison. ### Ablation studies of ICCN In order to analyze the usefulness of different components of the ICCN, we consider the following two questions: Question 1: Is using canonical correlation better than using other methods like Cosine-Similarity? Question 2: Is learning the interactions between text and video (or audio) useful? We design several experiments to address these two questions, First, we replace the CCA Loss with Cosine-Similarity Loss while leaving other parts of the ICCN unchanged. Second, instead of using the outer-product of audio (or video) and text as input to the CCA Loss, we use audio and video directly as the input to DCCA network. We compare different ICCN variants’ performance to prove the usefulness of each component of the network. ### Evaluation Methods To evaluate ICCN against previous baselines, the following performance metrics as described in (?; ?; ?) are reported, * • On the CMU-MOSI and CMU-MOSEI we report four performance metrics, i) binary accuracy, ii) F1-score iii) mean absolute error and iv) 7-class sentiment level / Correlation with human labeling. * • On the IEMOCAP we used i) binary accuracy and ii) F1-score for evaluation. Evaluations Details on CMU-MOSI, CMU-MOSEI: The original MOSI and MOSEI datasets are labeled in the range [-3,3]. The author of the datasets suggests a criterion for building binary labels: examples with label in [-3, 0) are considered to have negative sentiment while examples with label in (0, 3] are considered to have positive sentiment. 7-class sentiment level is also calculated based on the label distribution in [-3,3]. The correlation of predicted results with human labeling is also used as a criteria. ### Hyperparameter Tuning A basic Grid-Search is used to tune hyperparameters, and the best hyperparameter settings for the ICCN are chosen according to its performance on the validation dataset. For ICCN, hyperparameters and tuning ranges are: learning rate ($1e-5$–$1e-3$), mini-batch size ($128$–$512$), the number of epoch ($10$–$100$), hidden dimensions of MLP ($64$–$512$), and output dimension of the CCA projection ($30$–$100$). ReLU is used as the activation function, RMSProp is used as the optimizer. Whenever the training of the ICCN with a specific hypyerparameter setting has finished, features learned from the ICCN are used as input to the same downstream task models (a simple MLP is used in this work). Test results are reported by using the best hyperparameter setting learned above. Data View | IEMOCAP ---|--- Emotions | Happy | Angry | Sad | Neutral | Acc-2 | F-score | Acc-2 | F-score | Acc-2 | F-score | Acc-2 | F-score Audio | 84.03 | 81.09 | 85.49 | 84.03 | 82.75 | 80.26 | 63.08 | 59.24 Video | 83.14 | 80.36 | 85.91 | 83.27 | 81.19 | 80.35 | 62.30 | 58.19 Text | 84.80 | 81.17 | 85.12 | 85.21 | 84.18 | 83.63 | 66.02 | 63.52 Text+Video | 85.32 | 82.01 | 85.22 | 85.10 | 83.33 | 82.96 | 65.82 | 65.81 Text+Audio | 85.10 | 83.47 | 86.09 | 84.99 | 83.90 | 83.91 | 66.96 | 65.02 Audio+Video+Text | 86.00 | 83.37 | 86.37 | 85.88 | 84.02 | 83.71 | 66.87 | 65.93 CCA | 85.91 | 83.32 | 86.17 | 84.39 | 84.19 | 83.71 | 67.22 | 64.84 KCCA | 86.54 | 84.08 | 86.72 | 86.32 | 85.03 | 84.91 | 68.29 | 65.93 GCCA | 81.15 | 80.33 | 82.47 | 78.06 | 83.22 | 81.75 | 65.34 | 59.99 GCCA+Text | 87.02 | 83.44 | 88.01 | 88.00 | 84.79 | 83.26 | 68.26 | 67.61 DCCA | 86.99 | 84.32 | 87.94 | 87.85 | 86.03 | 84.36 | 68.87 | 65.93 TFN | 86.66 | 84.03 | 87.11 | 87.03 | 85.64 | 85.75 | 68.90 | 68.03 LMF | 86.14 | 83.92 | 86.24 | 86.41 | 84.33 | 84.40 | 69.62 | 68.75 MFM | 86.67 | 84.66 | 86.99 | 86.72 | 85.67 | 85.66 | 70.26 | 69.98 ICCN | 87.41 | 84.72 | 88.62 | 88.02 | 86.26 | 85.93 | 69.73 | 68.47 Table 2: Results for experiments on IEMOCAP. Best numbers are in bold. For accuracy and F-score, higher is better. For TFN, LMF, and MFM, we re-did experiments with using our features for a fair comparison. Data View | CMU-MOSI | CMU-MOSEI ---|---|--- | Acc-2 | F-score | MAE | Acc-7 | Corr | Acc-2 | F-score | MAE | Acc-7 | Corr ICCN | 83.07 | 83.02 | 0.862 | 39.01 | 0.714 | 84.18 | 84.15 | 0.565 | 51.58 | 0.713 ICCN1(no text) | 82.13 | 82.05 | 0.874 | 35.51 | 0.703 | 83.01 | 83.10 | 0.575 | 50.12 | 0.707 ICCN2(cos) | 82.32 | 82.27 | 0.876 | 36.01 | 0.702 | 82.98 | 82.90 | 0.575 | 50.63 | 0.700 ICCN3(no text + cos) | 81.49 | 81.58 | 0.889 | 35.77 | 0.692 | 82.59 | 82.73 | 0.578 | 50.21 | 0.696 Table 3: Ablation studies of ICCN on CMU-MOSI and CMU-MOSEI. ICCN1-3 denote different variants, “no text” means applying DCCA to audio and video directly instead of applying to the outer-product with text; “cos” means replacing CCA Loss with Cosine-Similarity Loss;“no text + cca” means removing outer-product with text and use Cosine-Similarity Loss. ## 5 Discussion of Empirical Results This section presents and discusses results on the CMU-MOSI, CMU-MOSEI, and IEMOCAP datasets. ### Performance on Benchmark Datasets Tables 1 and 2 present results of experiments on the CMU-MOSI, CMU-MOSEI, and IEMOCAP datasets. * • First, when compared with results of using uni-modal and simple concatenations, ICCN outperforms all of them in all of the criteria. Note that the performance of the text feature is always better than the performance of the audio and video, and that a simple concatenation of text, video, and audio does not work well. This shows the advance of highly-developed pre-trained text features capable of improving the overall multimodal task performance. However, it also shows the challenge of how to effectively combine such a highly developed text feature with audio and video features. * • Second, ICCN also outperforms other CCA-based methods. The results of using other CCA-based methods show that they cannot improve the multimodal embedding’s performance. We argue this occurs because 1) CCA / KCCA / GCCA do not exploit the power of neural network architectures so that their learning capacities are limited. 2) Using DCCA without learning the interactions between text-based audio and text-based video may sacrifice useful information. * • Third, the ICCN still achieves better or similar results when compared with other neural network based state-of-the-art methods (TFN, LMF, and MFM). These results demonstrate the ICCN’s competitive performance. ### Results of Ablation Studies Table 3 shows results of using variants of ICCN on CMU-MOSI and CMU-MOSEI datasets. First, using the CCA Loss performs better than using Cosine-Similarity Loss with or without the outer-product. This is reasonable as the DCCA is able to learn the hidden relationships (with the help of non-linear transformations) but cosine-similarity is restrained by the original coordinates. To further verify this, we also record changes of canonical correlation and cosine similarity between text-based audio and text-based video (i.e., between the two outputs of the CNNs in the ICCN) by using CCA Loss or Cosine Similarity Loss for the ICCN with using the CMU-MOSI dataset. Curves in Figures 2 and 3 summarize the results of the experiments. Results show that maximizing the canonical correlation by using the CCA Loss does not necessarily increase the cosine similarity, and vice versa. This demonstrates that the canonical correlation is a genuinely different objective function than cosine similarity, and explains the different behaviors in the downstream applications. CCA is capable of learning hidden relationships between inputs that the cosine similarity does not see. Second, learning the interactions between non-text and text performs better than using audio and video directly. This also makes sense because audio and video are more correlated when they are based on the same text, thus learning text-based audio and text-based video performs better. In summary, Table 3 shows the usefulness of using a text-based outer-product together with DCCA. Figure 2: Changes of mean canonical correlation and mean cosine similarity between text-based audio and text-based video when training with CCA Loss: The network learns to maximize the canonical correlation but the cosine similarity isn’t affected. Figure 3: Changes of mean canonical correlation and cosine similarity between text-based audio and text-based video when training with Cosine Similarity Loss: convergence of cosine similarity doesn’t affect canonical correlation. ## 6 Conclusion and Future Work This paper has proposed the ICCN method, which uses canonical correlation to analyze hidden relationships between text, audio, and video. Testing on a multi-modal sentiment analysis and emotion recognition task shows that the multi-modal features learned from the ICCN model can achieve state-of-the-art performance, and shows the effectiveness of the model. Ablation studies confirm the usefulness of different part of the network. There is, of course, considerable room for improvement. Possible directions include learning dynamic intra-actions in each model together with inter- actions between different modes; learning the trade-off between maximum canonical correlation and best downstream task performance; and developing an interpretable end-to-end multi-modal canonical correlation model. In the future, we hope to move forward in the development of multi-modal machine learning. ## References * [Akaho 2006] Akaho, S. 2006\. A kernel method for canonical correlation analysis. arXiv preprint cs/0609071. * [Andrew et al. 2013] Andrew, G.; Arora, R.; Bilmes, J.; and Livescu, K. 2013\. Deep canonical correlation analysis. In ICML, 1247–1255. * [Busso et al. 2008] Busso, C.; Bulut, M.; Lee, C.-C.; Kazemzadeh, A.; Mower, E.; Kim, S.; Chang, J. N.; Lee, S.; and Narayanan, S. S. 2008\. Iemocap: Interactive emotional dyadic motion capture database. Language resources and evaluation 42(4):335. * [Chen et al. 2017] Chen, M.; Wang, S.; Liang, P. P.; Baltrušaitis, T.; Zadeh, A.; and Morency, L.-P. 2017\. Multimodal sentiment analysis with word-level fusion and reinforcement learning. In ICMI, 163–171. ACM. * [Conneau et al. 2017] Conneau, A.; Kiela, D.; Schwenk, H.; Barrault, L.; and Bordes, A. 2017\. Supervised learning of universal sentence representations from natural language inference data. arXiv preprint arXiv:1705.02364. * [Degottex et al. 2014] Degottex, G.; Kane, J.; Drugman, T.; Raitio, T.; and Scherer, S. 2014\. Covarep—a collaborative voice analysis repository for speech technologies. In ICASSP, 960–964. IEEE. * [Devlin et al. 2018] Devlin, J.; Chang, M.-W.; Lee, K.; and Toutanova, K. 2018\. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805. * [Gers, Schmidhuber, and Cummins 1999] Gers, F. A.; Schmidhuber, J.; and Cummins, F. 1999\. Learning to forget: Continual prediction with lstm. ICANN. * [Haq and Jackson 2011] Haq, S., and Jackson, P. J. 2011\. Multimodal emotion recognition. In MAPAS. IGI Global. 398–423. * [Hazarika et al. 2018] Hazarika, D.; Poria, S.; Gorantla, S.; Cambria, E.; Zimmermann, R.; and Mihalcea, R. 2018\. Cascade: Contextual sarcasm detection in online discussion forums. arXiv preprint arXiv:1805.06413. * [Hotelling 1936] Hotelling, H. 1936\. Relations between two set of variables. Biometrika 28(3-4):321–377. * [iMotions 2017] iMotions. 2017\. Facial expression analysis. imotions.com. * [Lawrence et al. 1997] Lawrence, S.; Giles, C. L.; Tsoi, A. C.; and Back, A. D. 1997\. Face recognition: A convolutional neural-network approach. IEEE transactions on neural networks 8(1):98–113. * [Lee et al. 2014] Lee, G.; Singanamalli, A.; Wang, H.; Feldman, M. D.; Master, S. R.; Shih, N. N.; Spangler, E.; Rebbeck, T.; Tomaszewski, J. E.; and Madabhushi, A. 2014\. Supervised multi-view canonical correlation analysis (smvcca): Integrating histologic and proteomic features for predicting recurrent prostate cancer. IEEE transactions on medical imaging 34(1):284–297. * [Liang et al. 2018] Liang, P. P.; Liu, Z.; Zadeh, A.; and Morency, L.-P. 2018\. Multimodal language analysis with recurrent multistage fusion. arXiv preprint arXiv:1808.03920. * [Liu et al. 2018] Liu, Z.; Shen, Y.; Lakshminarasimhan, V. B.; Liang, P. P.; Zadeh, A.; and Morency, L.-P. 2018\. Efficient low-rank multimodal fusion with modality-specific factors. ACL. * [Liu et al. 2019] Liu, F.; Guan, B.; Zhou, Z.; Samsonov, A.; G Rosas, H.; Lian, K.; Sharma, R.; Kanarek, A.; Kim, J.; Guermazi, A.; and Kijowski, R. 2019\. Fully-automated diagnosis of anterior cruciate ligament tears on knee mr images using deep learning. Radiology 1\. * [Martin and Maes 1979] Martin, N., and Maes, H. 1979\. Multivariate analysis. Academic press London. * [Morency, Mihalcea, and Doshi 2011] Morency, L.-P.; Mihalcea, R.; and Doshi, P. 2011\. Towards Multimodal Sentiment Analysis: Harvesting Opinions from The Web. In ICMI 2011. * [Poria et al. 2016] Poria, S.; Chaturvedi, I.; Cambria, E.; and Hussain, A. 2016\. Convolutional mkl based multimodal emotion recognition and sentiment analysis. In ICDM, 439–448. IEEE. * [Rotman, Vulić, and Reichart 2018] Rotman, G.; Vulić, I.; and Reichart, R. 2018\. Bridging languages through images with deep partial canonical correlation analysis. In ACL(Volume 1: Long Papers), 910–921. * [Sarma, Liang, and Sethares 2018] Sarma, P. K.; Liang, Y.; and Sethares, W. A. 2018\. Domain adapted word embeddings for improved sentiment classification. arXiv preprint arXiv:1805.04576. * [Soleymani et al. 2017] Soleymani, M.; Garcia, D.; Jou, B.; Schuller, B.; Chang, S.-F.; and Pantic, M. 2017\. A survey of multimodal sentiment analysis. Image and Vision Computing 65:3–14. * [Sun et al. 2019] Sun, Z.; Sarma, P. K.; Sethares, W.; and Bucy, E. P. 2019\. Multi-modal sentiment analysis using deep canonical correlation analysis. arXiv preprint arXiv:1907.08696. * [Tenenhaus and Tenenhaus 2011] Tenenhaus, A., and Tenenhaus, M. 2011\. Regularized generalized canonical correlation analysis. Psychometrika 76(2):257. * [Tsai et al. 2018] Tsai, Y.-H. H.; Liang, P. P.; Zadeh, A.; Morency, L.-P.; and Salakhutdinov, R. 2018\. Learning factorized multimodal representations. arXiv preprint arXiv:1806.06176. * [Tsai et al. 2019] Tsai, Y.-H. H.; Bai, S.; Liang, P. P.; Kolter, J. Z.; Morency, L.-P.; and Salakhutdinov, R. 2019\. Multimodal transformer for unaligned multimodal language sequences. arXiv preprint arXiv:1906.00295. * [Wang et al. 2018] Wang, Y.; Shen, Y.; Liu, Z.; Liang, P. P.; Zadeh, A.; and Morency, L.-P. 2018\. Words can shift: Dynamically adjusting word representations using nonverbal behaviors. arXiv preprint arXiv:1811.09362. * [Wang, Li, and Lazebnik 2016] Wang, L.; Li, Y.; and Lazebnik, S. 2016\. Learning deep structure-preserving image-text embeddings. In CVPR, 5005–5013. * [Zadeh et al. 2016] Zadeh, A.; Zellers, R.; Pincus, E.; and Morency, L.-P. 2016\. Mosi: multimodal corpus of sentiment intensity and subjectivity analysis in online opinion videos. arXiv preprint arXiv:1606.06259. * [Zadeh et al. 2017] Zadeh, A.; Chen, M.; Poria, S.; Cambria, E.; and Morency, L.-P. 2017\. Tensor fusion network for multimodal sentiment analysis. arXiv preprint arXiv:1707.07250. * [Zadeh et al. 2018a] Zadeh, A.; Liang, P. P.; Mazumder, N.; Poria, S.; Cambria, E.; and Morency, L.-P. 2018a. Memory fusion network for multi-view sequential learning. In AAAI. * [Zadeh et al. 2018b] Zadeh, A.; Liang, P. P.; Poria, S.; Vij, P.; Cambria, E.; and Morency, L.-P. 2018b. Multi-attention recurrent network for human communication comprehension. In AAAI. * [Zadeh et al. 2018c] Zadeh, A. B.; Liang, P. P.; Poria, S.; Cambria, E.; and Morency, L.-P. 2018c. Multimodal language analysis in the wild: Cmu-mosei dataset and interpretable dynamic fusion graph. In ACL, 2236–2246.
# Quantum Gravity in Flat Spacetime Jarmo Mäkelä111Vaasa University of Applied Sciences, Wolffintie 30, 65200 Vaasa, Finland, email<EMAIL_ADDRESS> ###### Abstract Inspired by Einstein’s Strong Principle of Equivalence we consider the effects of quantum mechanics to the gravity-like phenomena experienced by an observer in a uniformly accelerating motion in flat spacetime. Among other things, our model of quantum gravity, derived from the first principles, predicts the Unruh effect, and a discrete area spectrum for spacelike two-surfaces. ## 1 Introduction When Albert Einstein began to formulate his general theory of relativity, he used as a starting point his Strong Principle of Equivalence. In broad terms, this principle states that it is impossible to decide, by means of local measurements, whether one is in a gravitational field or in an accelerating frame of reference. In this sense the effects of gravity are equivalent with those of acceleration. [1, 2] For instance, if an observer finds all bodies in his surroundings to fall with the same acceleration, the observer has no way to tell, whether he is in a uniform gravitational field, caused by the curvature of spacetime, or in a uniformly accelerating motion in flat spacetime. One of the greatest challenges of modern physics is to create the quantum theory of gravity, which is supposed to bring together general relativity and quantum mechanics. It appears to the author that in the attempts to quantize gravity developed so far the Equivalence Principle has received insufficient attention. It is possible that we should take the Equivalence Principle as the starting point of quantum gravity in the same way as Einstein took it as the starting point of his general relativity. Inspired by these thoughts we shall, in this paper, quantize the effects of gravity in flat spacetime from the point of view of an observer in a uniformly accelerating motion. According to the Equivalence Principle quantization of gravity in curved spacetime should, at least locally, be equivalent with the quantization of gravity in flat spacetime from the point of view of an acccelerating observer. We shall see that the resulting ”quantum gravity in flat spacetime” may indeed be developed systematically, beginning from the first principles. In our approach we consider the gravity-like effects on a spacelike two-plane at rest with respect to our accelerating observer, and it plays a role somewhat similar to that of the event horizon of a black hole. The standard rules of quantum mechanics applied to our model imply that the accelerating plane has a discrete structure in the sense that it consists of a finite number of separate constituents, all of them having an equally spaced area spectrum. It also predicts the Unruh effect, according to which an observer in an accelerating motion detects thermal radiation with a characteristic temperature, which is proportional to the proper acceleration of the observer. Unless otherwise stated, we shall always use the natural units, where $\hbar=G=c=k_{B}=1$. ## 2 Action Whenever we attempt to quantize gravity, we may use the classical Einstein- Hilbert action $S_{EH}:=\frac{1}{16\pi}\int_{M}R\sqrt{-g}\,d^{4}x$ (2.1) as the starting point. In this equation $R$ is the Riemann curvature scalar, and $g$ is the determinant of the metric tensor of spacetime. We have integrated over the whole spacetime $M$. However, if the spacetime region under consideration possesses a boundary, we must supplement the Einstein- Hilbert action with the Gibbons-Hawking boundary term [3] $S_{GH}:=\frac{1}{8\pi}\int_{\partial M}K\,dV,$ (2.2) where $K$ is the trace of the exterior curvature tensor induced on the boundary $\partial M$ of spacetime. $dV$ is the volume element on the boundary, and we have integrated over the whole boundary. So the whole action takes the form: $S=S_{EH}+S_{GH}=\frac{1}{16\pi}\int_{M}R\sqrt{-g}\,d^{4}x+\frac{1}{8\pi}\int_{\partial M}K\,dV.$ (2.3) In flat spacetime the Riemann curvature scalar $R\equiv 0$, and we are left with the Gibbons-Hawking boundary term only. Hence the action becomes to: $S=\frac{1}{8\pi}\int_{\partial M}K\,dV.$ (2.4) Even if spacetime were flat, the action may still be non-zero, if the boundary of spacetime has been chosen appropriately. Because of the presence of the potentially non-zero boundary term, the quantum-mechanical properties of the gravity-like effects caused by a choice of a non-inertial frame of reference are far from trivial. In this paper we first pick up from the four-dimensional flat spacetime an inertial frame of reference, where each point of spacetime is identified by means of four coordinates $(T,X,Y,Z)$. The coordinate $T$ is timelike, and the coordinates $X$, $Y$ and $Z$ spacelike. When written in terms of these coordinates the line element of spacetime takes the flat Minkowski form: $ds^{2}=-dT^{2}+dX^{2}+dY^{2}+dZ^{2}.$ (2.5) We choose the boundary of spacetime to consist of five timelike hyperplanes, where $X=X_{1}$, $Y=Y_{1}$, $Y=Y_{2}$, $Z=Z_{1}$ and $Z=Z_{2}$ such that $Y_{2}>Y_{1}$ and $Z_{2}>Z_{1}$. In other words, the boundary of spacetime is a three-dimensional, rectangular box, which proceeds in time, and where the faces stay still in our frame of reference. One of the faces of the box, however, is not assumed to stay still, but it is assumed to accelerate with constant proper acceleration $a$ to the direction of the positive $X$-axis. As a consequence, the $X$-coordinates of the ponts of the face satisfy an equation: $X^{2}-T^{2}=\frac{1}{a^{2}},$ (2.6) which implies: $X=\sqrt{T^{2}+\frac{1}{a^{2}}}.$ (2.7) So the boundary of spacetime consists, as a whole, of five timelike hyperplanes, and of a one timelike hypersurface, which is created, when a spacelike two-plane accelerates along the $X$-axis. The exterior curvature of all of the timelike hyperplanes is zero, and so they do not contribute to the action. The exterior curvature of the hypersurface created by the accelerating plane, however, is non-zero. To calculate the trace of the exterior curvature tensor on that hypersurface we introduce the Rindler coordinates to our spacetime. As it is well known, between the Rindler coordinates $(t,x)$ and the coordinates $(T,X)$ there is the relationship: [4] $\displaystyle t$ $\displaystyle=$ $\displaystyle\tanh^{-1}\left(\frac{T}{X}\right),$ (2.8a) $\displaystyle x$ $\displaystyle=$ $\displaystyle\sqrt{X^{2}-T^{2}}.$ (2.8b) When written in terms of the Rindler coordinates, together with the coordinates $Y$ and $Z$, the line element of spacetime takes the form: $ds^{2}=-x^{2}\,dt^{2}+dx^{2}+dY^{2}+dZ^{2}.$ (2.9) An observer with constant Rindler coordinate $x$, together with constant coordinates $Y$ and $Z$, has constant proper acceleration $a$, which is related to $x$ such that $x=\frac{1}{a},$ (2.10) which follows from Eqs. (2.6) and (2.8b). Eq. (2.8 a), in turn, implies that the Rindler time coordinate $t$ gives the boost angle $\phi$ of an observer with constant proper acceleration. On the timelike hypersurface created, when a plane is in an accelerating motion with constant proper acceleration $a$ the Rindler coordinate $x$ is a constant. The only non-zero component of the exterior curvature tensor on that hypersurface is: $K_{tt}=\Gamma_{tt}^{x}=x,$ (2.11) and therefore its trace is: $K=g^{tt}K_{tt}=-\frac{1}{x}.$ (2.12) The volume element on our hypersurface is: $dV=x\,dY\,dZ\,dt,$ (2.13) and so we find that the action in Eq. (2.7) takes, if the Rindler time coordinate $t$ lies within an interval $[t_{i},t_{f}]$, the form: $S=-\frac{1}{8\pi}A(t_{f}-t_{i}),$ (2.14) where $A$ is the area of the accelerating plane. [5] ## 3 Hamiltonian The bridge from the action to the Hamiltonian $H(q_{j},p_{j};t)$ of any system is formed by the Hamilton-Jacobi equation: $H(q_{j},\frac{\partial S}{\partial q_{j}};t)+\frac{\partial S}{\partial t}=0,$ (3.1) where $S$ is the principal function of the system. The quantities $q_{j}$ are the coordinates of the configuration space, and the quantities $p_{j}=\frac{\partial S}{\partial q_{j}}$ the corresponding coordinates of the momentum space. The principal function of the system has been denoted by the same symbol as its action for a very good reason: As it is well known, the principal function agrees with the action integral calculated along the classical path. Hence it follows that if we know the action $S$ calculated along the classical path, the Hamiltonian of the system will be: [6] $H=-\frac{\partial S}{\partial t}.$ (3.2) Our derivation of the classical Hamiltonian of the gravitational field is based on Eq. (3.2). Before proceeding to the use of Eq. (3.2), however, we must decide, which time coordinate to use. The very idea of this paper is to consider the effects of gravity from the point of view of a uniformly accelerating observer. It is therefore sensible to take the proper time $\tau$ of the observer as the time coordinate. Eq. (2.9) implies that between the proper time $\tau$ and the Rindler time $t$ there is the relationship: $d\tau=x\,dt.$ (3.3) Writing $H=-\frac{\partial S}{\partial\tau},$ (3.4) and using Eqs. (2.10) and (2.14) we therefore find: $H=\frac{a}{8\pi}A.$ (3.5) Eq. (3.5) gives the classical Hamiltonian of the gravitational field from the point of view of an observer moving along a plane with area $A$ and proper acceleration $a$. Remarkably, the Hamiltonian in Eq. (3.5) is identical to the one obtained in several papers in spacetimes with horizon, from the point of view of an observer with constant proper acceleration $a$, just outside of a horizon with area $A$. [7-14] When the proper acceleration $a$ of the observer tends to infinity, Eq. (2.10) implies that the Rindler coordinate $x$ tends to zero, which means that in this limit our plane tends to a Rindler horizon of flat spacetime. We may always define a diffeoforphism $f$ from the unit square $I^{2}:=[0,1]\times[0,1]$ to the accelerating plane, which we shall henceforth denote by $B$. In general, this map $f:I^{2}\longrightarrow B$ is of the form: $f(\chi^{1},\chi^{2})=(Y(\chi^{1},\chi^{2}),Z(\chi^{1},\chi^{2}))$ (3.6) for all $\chi^{1},\chi^{2}\in[0,1]$, where the coordinates $Y(\chi^{1},\chi^{2})$ and $Z(\chi^{1},\chi^{2})$ of the point $(Y,Z)$ on the plane $B$ are functions of the coordinates $\chi^{1}$ and $\chi^{2}$ of the point $\chi=(\chi^{1},\chi^{2})$ in $I^{2}$. Defining the function $p(\chi):=p(\chi^{1},\chi^{2}):=\frac{1}{8\pi}\left|\begin{matrix}\frac{\partial Y(\chi^{1},\chi^{2})}{\partial\chi^{1}}&\frac{\partial Y(\chi^{1},\chi^{2})}{\partial\chi^{2}}\\\ \frac{\partial Z(\chi^{1},\chi^{2})}{\partial\chi^{1}}&\frac{\partial Z(\chi^{1},\chi^{2})}{\partial\chi^{2}}\end{matrix}\right|$ (3.7) we may write the area of the plane $B$ as: $A=8\pi\int_{I^{2}}p(\chi)\,d^{2}\chi,$ (3.8) where we have integrated over the unit square. As a consequence, the Hamiltonian of the gravitational field takes, from the point of view of our accelerating observer, the form: $H=a\int_{I^{2}}p(\chi)\,d^{2}\chi.$ (3.9) It must be emphasized that the proper acceleration $a$ is not a dynamical variable of the system, but just a parameter, which determines the observer. We shall now take the quantities $p(\chi)$ associated with the points $\chi$ as the coordinates of the momentum space. The corresponding coordinates of the configuration space we shall denote by $q(\chi)$. They obey the Hamiltonian equations of motion: $\dot{q}(\chi)=\frac{\delta H}{\delta p(\chi)}=a,$ (3.10) where the dot means the derivative with respect to the proper time $\tau$ of the accelerating observer. The general solution to Eq. (3.10) is: $q(\chi)=a\tau+C(\chi),$ (3.11) where $C(\chi)$ is an arbitrary constant function of the proper time $\tau$, and depends on the coordinates $(\chi^{1},\chi^{2})$ of the point $\chi$ only. Eqs. (2.10) and (3.3) imply that the quantity $a\tau$ may be identified as the boost angle $\phi$ of the observer. Hence we may write Eq. (3.11) as: $q(\chi)=\phi+C(\chi).$ (3.12) Since our Hamiltonian in Eq. (3.10) does not involve $q(\chi)$, the Hamiltonian equations of motion written for the variables $p(\chi)$ read as: $\dot{p}(\chi)=-\frac{\delta H}{\delta q(\chi)}=0,$ (3.13) which means that the quantities $p(\chi)$ are constants of motion of the system. ## 4 Quantization When we go over from the classical to the quantum-mechanical consideration of the gravitational field in flat spacetime we must replace the classical quantities $q(\chi)$ and $p(\chi)$ by the corresponding operators $\hat{q}(\chi)$ and $\hat{p}(\chi)$ obeying the canonical commutation relations: $\displaystyle[\hat{q}(\chi),\hat{q}(\chi^{\prime})]=[\hat{p}(\chi),\hat{p}(\chi^{\prime})]$ $\displaystyle=$ $\displaystyle 0,$ (4.1a) $\displaystyle[\hat{q}(\chi),\hat{p}(\chi^{\prime})]$ $\displaystyle=$ $\displaystyle i\delta^{2}(\chi,\chi^{\prime})$ (4.1b) at all points $\chi,\chi^{\prime}\in I^{2}$. In Eq. (4.1b) $\delta^{2}(\chi,\chi^{\prime})$ is the two-dimensional delta function in $I^{2}$. When written in terms of these operators the classical Hamltonian $H$ in Eq. (3.9) becomes replaced by the corresponding Hamiltonian operator $\hat{H}=a\int_{I^{2}}\hat{p}(\chi)\,d^{2}\chi,$ (4.2) and therefore the time-independent Schrödinger equation $\hat{H}|\psi\rangle=E|\psi\rangle$ (4.3) written for the energy eigenstates $|\psi\rangle$ of the gravitational field takes the form: $a\int_{I^{2}}\hat{p}(\chi)|\psi\rangle=E|\psi\rangle.$ (4.4) Identifying the energy $E$ of the gravitational field with the classical Hamiltonian $H$ in Eq. (3.9) we observe that this equation reduces to: $\int_{I^{2}}\hat{p}(\chi)|\psi\rangle=\int_{I^{2}}p(\chi)|\psi\rangle.$ (4.5) Eq. (4.5) has an obvious solution, where $|\psi\rangle$ is the common eigenstate of all of the operators $\hat{p}(\chi)$, which means that: $\hat{p}(\chi)|\psi\rangle=p(\chi)|\psi\rangle$ (4.6) for all $\chi\in I^{2}$. If the eigenstate $|\psi\rangle$ is normed to unity, we must have: $\langle\psi|\psi\rangle=1.$ (4.7) We shall take Eqs. (4.6) and (4.7) as the postulates of our model, and see, where this will take us. In the position representation the energy eigenstates $|\psi\rangle$ are expressed as functionals $\psi[q(\chi)]$ of the functions $q(\chi)$, and they may be regarded as the wave functions of the gravitational field. The inner product between the wave functions $\psi_{1}[q(\chi)]$ and $\psi_{2}[q(\chi)]$ is defined as: $\langle\psi_{1}|\psi_{2}\rangle:=\int\psi_{1}^{*}[q(\chi)]\psi_{2}[q(\chi)]\,\mathcal{D}[q(\chi)],$ (4.8) where $\mathcal{D}[q(\chi)]$ is an appropriate integration measure in the space of the functions $q(\chi)$. The momentum operators $\hat{p}(\chi)$ are expressed as functional differential operators: $\hat{p}(\chi):=-i\frac{\delta}{\delta q(\chi)},$ (4.9) and so Eq. (4.6) takes the form: $-i\frac{\delta\psi[q(\chi)]}{\delta q(\chi)}=p(\chi)\psi[q(\chi)].$ (4.10) Fortunately, this functional deifferential equation may be solved explicitly. Its general solution is: $\psi[q(\chi)]=\mathcal{N}\exp\left[i\int_{I^{2}}p(\chi)q(\chi)\,d^{2}\chi\right],$ (4.11) where $\mathcal{N}$ is an arbitrary complex number. As one may observe, the functional $\psi[q(\chi)]$ in Eq. (4.11) is uniquely determined, up to the constant coefficient $\mathcal{N}$, by the function $p(\chi)$, which gives, at each point $\chi$, the eigenvalue of the operator $\hat{p}(\chi)$. If we denote by $p^{\prime}(\chi)$ another eigenvalue of $\hat{p}(\chi)$, the corresponding eigenfunction is: $\psi^{\prime}[q(\chi)]=\mathcal{N}^{\prime}\exp\left[i\int_{I^{2}}p^{\prime}(\chi)q(\chi)\,d^{2}\chi\right],$ (4.12) and since eigenstates associated with different eigenvalues must be orthogonal, we must have: $\int\psi^{*}[q(\chi)]\psi^{\prime}[q(\chi)]\,\mathcal{D}[q(\chi)]=1,$ (4.13) whenever $p^{\prime}(\chi)\equiv p(\chi),$ (4.14) and $\int\psi^{*}[q(\chi)]\psi^{\prime}[q(\chi)]\,\mathcal{D}[q(\chi)]=0,$ (4.15) if $p^{\prime}(\chi)$ differs from $p(\chi)$ at any point $\chi$. In Eq. (4.12) $\mathcal{N}^{\prime}$ is an arbitrary complex number such that Eq. (4.13) is satisfied. We now divide the unit square $I^{2}$ in $N$ non-intersecting subsets $D_{j}$ $(j=1,2,\dots,N)$ such that $D_{j}\cap D_{k}=\emptyset$ for all $j\neq k$, where $j,k\in\\{1,2,\dots,N\\}$, and $I^{2}=D_{1}\cup D_{2}\cup\cdots\cup D_{N}.$ (4.16) We shall assume that each subset $D_{j}$ has the same area $\Delta\chi:=\frac{1}{N}.$ (4.17) We shall also assume that in the limit, where $N$ tends to infinity, the supremum of the diameters of the subsets $D_{j}$ tends to zero. Picking up from each of the subsets $D_{j}$ a point $\chi_{j}$ we may write the integral in the exponential in Eq. (4.11) as a Riemann integral: $\int_{I^{2}}p(\chi)q(\chi)\,d^{2}\chi=\lim_{N\rightarrow\infty}\left(\sum_{j=1}^{N}p_{j}q_{j}\Delta\chi\right),$ (4.18) where we have denoted: $\displaystyle p_{j}$ $\displaystyle:=$ $\displaystyle p(\chi_{j}),$ (4.19a) $\displaystyle q_{j}$ $\displaystyle:=$ $\displaystyle q(\chi_{j})$ (4.19b) for all $j=1,2,\dots,N$. Hence we have: $\psi[q(\chi)]=\mathcal{N}\lim_{N\rightarrow\infty}\left[\prod_{j=1}^{N}\exp(ip_{j}q_{j}\Delta\chi)\right].$ (4.20) Eq. (4.20) enables us to define the integral on the left hand side of Eq. (4.13): $\int\psi^{*}[q(\chi)]\psi^{\prime}[q(\chi)]\,\mathcal{D}[q(\chi)]:=\mathcal{N}^{*}\mathcal{N}^{\prime}\lim_{N\rightarrow\infty}\left\\{\prod_{j=1}^{N}\int_{-L/2}^{L/2}\exp[i(p^{\prime}_{j}-p_{j})q_{j}\,\Delta\chi]\,dq_{j}\right\\},$ (4.21) where we have denoted ${p^{\prime}}_{j}:=p^{\prime}(\chi_{j})$. As one may see, we have introduced a new, positive number $L$. We might, of course, have taken $L$ to infinity, and thereby integrate $q_{j}$ from the negative to the positive infinity. Such a choice, however, would make the integral in Eq. (4.21) infinite. Because of that we shall henceforth take $L$ to be a fixed number. If we want Eqs. (4.13) and (4.15) to be satisfied, we must have, for every $j=1,2,\dots,N$: $p_{j}\Delta\chi=n_{j}\frac{2\pi}{L},$ (4.22) where $n_{j}=0,\pm 1,\pm 2\,\dots$. The diffeomorphism $f:I^{2}\longrightarrow B$ defined in Eq. (3.6) maps every subset $D_{j}$ of $I^{2}$ to a certain piece, or constituent $U_{j}=f(D_{j})$ (4.23) of the accelerating plane $B$. These constituents have the properties: $U_{j}\cap U_{k}=\emptyset$ (4.24) for all $j\neq k$, and $B=U_{1}\cup U_{2}\cup\cdots\cup U_{N}.$ (4.25) In the limit, where the number $N$ of the constituents becomes very large, we may view the quantity $A_{j}:=8\pi p_{j}\Delta\chi$ (4.26) as the area of the constituent $j$. According to Eq. (4.22) the possible area eigenvalues of the constituent are: $A_{j}=n_{j}\frac{16\pi^{2}}{L}.$ (4.27) Obviously, negative area eigenvalues do not make sense, and therefore we shall ignore the negative values of the quantum numbers $n_{j}$, and keep only the non-negative ones. In other words, $n_{j}=0,1,2,\dots$ in Eq. (4.27). As one may observe from Eq. (4.27), the areas of the constituents of the accelerating plane $B$ have a discrete spectrum with an equal spacing. Unless the plane has an infinite number of constituents with zero area, we are therefore forced to conclude that the number of the constituents of the plane is finite. This number we shall denote by $N$. The total area of the plane is the sum of the areas of its constituents: $A=A_{1}+A_{2}+\cdots+A_{N}.$ (4.28) So we see that the postulates (4.6) and (4.7) of our model imply that our plane necessarily has a discrete structure in the sense that it consists of a finite number of separate constituents. The possible area eigenvalues of the plane are of the form: $A=\frac{16\pi^{2}}{L}(n_{1}+n_{2}+\cdots+n_{N}).$ (4.29) The eigenfunctions associated with these area eigenvalues are of the form: $\psi(q_{1},q_{2},\dots,q_{N})=\mathcal{N}\exp\left[\frac{i}{8\pi}(A_{1}q_{1}+A_{2}q_{2}+\cdots+A_{N}q_{N})\right].$ (4.30) Hence the wave functions are no more functionals, but ordinary functions of a finite number of variables $q_{1},q_{2},\dots,q_{N}$. The inner product between the wave functions is defined as: $\langle\psi_{1}|\psi_{2}\rangle:=\int_{-L/2}^{L/2}dq_{1}\int_{-L/2}^{L/2}dq_{2}\cdots\int_{-L/2}^{L/2}dq_{N}\,\psi^{*}_{1}(q_{1},q_{2},\dots,q_{N})\psi_{2}(q_{1},q_{2},\dots,q_{N}).$ (4.31) and if we want the wave function $\psi(q_{1},q_{2},\dots,q_{N})$ to be normed to unity, we may take: $\mathcal{N}=\frac{1}{\sqrt{L^{N}}}.$ (4.32) So we find: $\psi(q_{1},q_{2},\dots,q_{N})=\frac{1}{\sqrt{L^{N}}}\exp\left[\frac{i}{8\pi}(A_{1}q_{1}+A_{2}q_{2}+\cdots+A_{N}q_{N})\right].$ (4.33) Eqs. (3.5) and (4.29) imply that the energy eigenvalues of the gravitational field from the point of view of our accelerating observer are: $E=\frac{\alpha a}{8\pi}(n_{1}+n_{2}+\cdots n_{N}),$ (4.34) where $\alpha:=\frac{16\pi^{2}}{L}.$ (4.35) As the reader may have noticed, quantization of the gravitational field in flat spacetime from the point of view of an accelerating observer is, in our model, very similar to the quantization of the system of free particles in non-relativistic quantum mechanics. In such systems the coordinates of the momentum space are constants of motion, and the coordinates of the configuration space are, possibly up to additive constants, linear functions of time, whereas in our system the momentum variables $p(\chi)$ are constants , and the corresponding coordinates $q(\chi)$ of the configuration space depend, up to additive constants, linearly on the proper time $\tau$ of the observer. When quantizing the system of free particles in non-relativistic quantum mechanics one must use box normalization for the wave function of the system. When using the box normalization one assumes that the particles of the system lie in a rectangular box. Finally, at the end of the calculations, the edge kength of the box is taken to infinity. In our model we also applied a sort of box normalization, where the coordinates of the confuguration space of the system were assumed to lie within the interval $[-L/2,L/2]$. However, we did not take the ”edge length” $L$ of the box to infinity, but we took $L$ to be a fixed number. As a consequence, $L$ appears as a parameter of the model, which must be fixed such that the predictions of the model agree with experiments. ## 5 Unruh Effect In Section 4 we completed, in our model, the quantization of gravity in flat spacetime. The most important outcome of our model was that the area of an accelerating plane has a discrete spectrum with equal spacing. We are now prepared to go to the thermodynamical properties of gravity. As it is well known, the thermodynamical properties of any system may be deduced from its partition function $Z(\beta):=\sum_{n}e^{-\beta E_{n}},$ (5.1) where $\beta$ is the temperature parameter, and we have summed over the energy eigenstates $n$ of the system. In what follows, we shall base our calculation of the partition function on Eq. (4.34). We say that constituent $j$ is in vacuum, if $n_{j}=0$; otherwise the constituent is in an excited state. When obtaining an expression for the partition function of our system we shall assume that at least one of the constituents of the accelerated plane is in an excited state, and when the quantum states of the constituents are interchanged, the quantum state of the system will also change. With these assumptions the partition function takes the form: $\begin{split}Z(\beta)=&\,\,\,\,\sum_{n_{1}=1}^{\infty}\exp\left(-\frac{\alpha\beta a}{8\pi}n_{1}\right)\\\ &+\sum_{n_{1}=1}^{\infty}\exp\left(-\frac{\alpha\beta a}{8\pi}n_{1}\right)\sum_{n_{2}=1}^{\infty}\exp\left(-\frac{\alpha\beta a}{8\pi}n_{2}\right)\\\ &+\cdots\\\ &+\sum_{n_{1}=1}^{\infty}\exp\left(-\frac{\alpha\beta a}{8\pi}n_{1}\right)\cdots\sum_{n_{N}=1}^{\infty}\exp\left(-\frac{\alpha\beta a}{8\pi}n_{N}\right).\end{split}$ (5.2) In the first term on the right hand side of Eq. (5.2) just one of the constituents is in an excited state, and we have summed over all of those states. In the second term two of the constituents are in excited states. Finally, in the last term all of the $N$ constituents are in excited states. Defining the characteristic temperature $T_{C}:=\frac{\alpha a}{8\pi\ln(2)}$ (5.3) we may write Eq. (5.2) as: $\begin{split}Z(\beta)&=\frac{1}{2^{\beta T_{C}}-1}+\left(\frac{1}{2^{\beta T_{C}}-1}\right)^{2}+\cdots+\left(\frac{1}{2^{\beta T_{C}}-1}\right)^{N}\\\ &=\frac{1}{2^{\beta T_{C}}-2}\left[1-\left(\frac{1}{2^{\beta T_{C}}-1}\right)^{N}\right],\end{split}$ (5.4) whenever $\beta T_{C}\neq 1$. When $\beta T_{C}=1$, we have: $Z(\beta)=N.$ (5.5) The energy of the gravitational field may be calculated from its partition function $Z(\beta)$ as: $E=-\frac{\partial}{\partial\beta}\ln[Z(\beta)],$ (5.6) and if we define the energy per constituent as $\bar{E}:=\frac{E}{N},$ (5.7) Eqs. (5.4) and (5.6) imply: $\bar{E}=\bar{E}_{1}+\bar{E}_{2},$ (5.8) where we have denoted: $\displaystyle\bar{E}_{1}$ $\displaystyle:=$ $\displaystyle\frac{1}{N}\frac{2^{\beta T_{C}}}{2^{\beta T_{C}}-2}T_{C}\ln(2),$ (5.9a) $\displaystyle\bar{E}_{2}$ $\displaystyle:=$ $\displaystyle\frac{2^{\beta T_{C}}}{2^{\beta T_{C}}-1-(2^{\beta T_{C}}-1)^{N+1}}T_{C}\ln(2),$ (5.9b) whenever $\beta T_{C}\neq 1$. When $\beta T_{C}=1$, a calculation similar to the one performed in Ref. [7] gives, in the large $N$ limit, the result: $\bar{E}=T_{C}\ln(2).$ (5.10) Eqs. (4.34), (5.3), (5.8) and (5.9) imply that the average $\bar{n}:=\frac{n_{1}+n_{2}+\cdots n_{N}}{N}$ (5.11) of the quantum numbers $n_{1},n_{2},\dots,n_{N}$ depends of the temperature parameter $\beta$ as: $\bar{n}=\frac{1}{N}\frac{2^{\beta T_{C}}}{2^{\beta T_{C}}-2}+\frac{2^{\beta T_{C}}}{2^{\beta T_{C}}-1-(2^{\beta T_{C}}-1)^{N+1}}.$ (5.12) So far everything has been quite exact, and no approximations have been made. At this point we proceed to consider, what happens in the limit, where the number $N$ of the constituents of the accelerating plane tends to infinity. In this limit the first term on the right hand side of Eq. (5.12) will vanish, and we have, in effect: $\bar{n}=\frac{2^{\beta T_{C}}}{2^{\beta T_{C}}-1-(2^{\beta T_{C}}-1)^{N+1}}.$ (5.13) When $\beta T_{C}>1$, which means that the temperature $T$ experienced by the observer moving along the plane is less than the characteristic temperature $T_{C}$ we have: $\lim_{N\rightarrow\infty}(2^{\beta T_{C}}-1)^{N+1}=\infty,$ (5.14) and so the constituents of the plane are, in effect, in vacuum. However, when $\beta T_{C}<1$, which means that $T>T_{C}$, we find: $\lim_{N\rightarrow\infty}(2^{\beta T_{C}}-1)^{N+1}=0,$ (5.15) and so we may write, as an excellent approximation: $\bar{n}=\frac{2^{\beta T_{C}}}{2^{\beta T_{C}}-1}.$ (5.16) An interesting property of this expression is that, when $T$ tends to $T_{C}$ from the right hand side, which means that $\beta T_{C}$ tends to $1$ from the left hand side, we have: $\bar{n}=2.$ (5.17) So we find that at the characteristic temperature $T_{C}$ the accelerating plane performs a phase transition, where the constituents of the plane jump, in average, from the vacuum to the second excited states. The latent heat per constituent associated with this phase transition is $\bar{L}=2T_{C}\ln(2).$ (5.18) Since the constituents of the plane are effectively in the vacuum, when $T<T_{C}$, and jump to the second excited states at the characteristic temperature $T_{C}$, we may view the characteristic temperature $T_{C}$, defined in Eq. (5.3), as the lowest possible temperature experienced by the accelerating observer. Interestingly, we find that if we choose: $\alpha=4\ln(2),$ (5.19) then: $T_{C}=\frac{a}{2\pi},$ (5.20) which exactly agrees with the Unruh temperature [15] $T_{U}:=\frac{a}{2\pi}$ (5.21) experienced by the observer. In this sense we have been able to derive the Unruh effect from our model of quantum gravity in flat spacetime: An accelerating observer detects thermal radiation with a temperature, which is proportional to the proper acceleration $a$ of the observer. That radiation is produced, when the constituents of the plane perform quantum jump from the excited states to the vacuum. ## 6 Concluding Remarks In this paper we have constructed, step by step, a model of quantum gravity in flat spacetime from the point of view of a unformly accelerating observer. An observer of this kind detects phenomena identical to those in a uniform gravitational field in his surroundings, and we considered those effects quantum-mechanically. Among other things, our nodel of quantum gravity predicted that spacetime has a discrete structure in the sense that the plane at rest with respect to the accelerating observer consists of separate constituents, each of which has an area, which is an integer times a certain fundamental area. As a consequence, the area eigenvalues of the accelerating plane are of the form: $A=\alpha(n_{1}+n_{2}+\cdots n_{N}),$ (6.1) where the quantum numbers $n_{j}$ are non-negative integers for all $j=1,2,\dots,N$, and $N$ is the number of the constituents of the plane. We found that with the choice $\alpha=4\ln(2)$ (6.2) for the parameter $\alpha$ of our model an observer with constant proper acceleration $a$ detects thermal radiation with certain minimum temperature, which agrees with the Unruh temperature $T_{U}=\frac{a}{2\pi}$ (6.3) of the observer. In this sense our model of quantum gravity in flat spacetime implies the Unruh effect. Basically, the results of the paper were outcomes of our decision to supplement the standard Einstein-Hilbert action with the standard Gibbons- Hawking boundary term. We first picked up a rectangular box on a spacelike hypersurface of the flat Minkowski spacetime spacetime, where the time coordinate is a constant, and then put one of its faces into a uniformly accelerating motion, while the other faces were kept at rest in the flat Minkowski system of coordinates. When the faces of the box proceeded in the flat spacetime, a three-dimensional boundary was created for spacetime. It turned out that the trace of the exterior curvature tensor on the three- dimensional timelike hypersurface created by the accelerating plane was non- zero, while the exterior curvature tensor on the timelike hypersurfaces created by the other faces of the box vanished identically. As a consequence, the Gibbons-Hawking boundary term was non-zero, even though the Einstein- Hilbert action vanished. Thus, the presence of the Gibbons-Hawking boundary term produced non-zero gravitational action, despite of the fact that spacetime was flat. From the action we obtained the classical Hamiltonian with respect to the accelerating observer. The discrete structure of the accelerating plane followed from the straightforward replacement of the classical Hamiltonian by the corresponding Hamiltonian operator according to the standard rules of quantum mechanics. The coordinates $p_{j}$ of the momentum space agree, up to the coefficient $\frac{1}{8\pi}$, with the areas of the constituents of the plane, whereas the corresponding coordinates $q_{j}$ of the configuration space agree, up to additive constants, with the boost angle of the observer. In the process we were compelled to introduce a box normalization for the wave function of the gravitational field, where the coordinates $q_{j}$ of the configuration space are confined to live inside of a box with edge length $L$. 222Not to be confused with the edge length of the accelerating box in flat spacetime! Defining a parameter $\alpha$ in terms of the edge length $L$ as: $\alpha=\frac{16\pi^{2}}{L}$ (6.4) we obtained the area spectrum in Eq. (6.1). As we have seen, quantization of gravity in flat spacetime from the point of view of an accelerating observer is really very simple and straightforward, and may be carried out explicitly. Nevertheless, there are still some shortcomings in our procedure. For instance, our model of quantum gravity predicted for the areas of the constituents of the accelerating plane both positive and negative eigenvalues, and we excluded the negative eigenvalues ”by hand”. An even more serious problem in our approach is the presence of an undetermined parameter, the edge length $L$ of the box used in the normalization of the wave function. Eqs. (6.2) and (6.4) imply that if we choose $L=\frac{4\pi^{2}}{\ln(2)}\approx 57.0,$ (6.5) our model predicts the Unruh effect. The close relationship between the variables $q_{j}$ and the boost angle $\phi$ of the observer suggests that the edge length $L$ sets the bound for $\phi$ such that, for symmetry reasons, $-\frac{L}{2}\leq\phi\leq\frac{L}{2}$. Whether such interpetation is valid, is still an open question, as well as are its possible consequences. Taken as a whole, our paper highlights an aspect of the potential quantum theory of gravity, which so far has received very little attention: The quantum states of the gravitational field are not absolute, but relative, and they depend on the state of motion of the observer. For instance, an accelerating observer in flat spacetime detects gravity-like effects, which may be quantized by means of the standard rules of quantum mechanics, whereas inertial observers detect no effects of gravity whatsoever. It is to be hoped that the results obtained this paper would pave the way for a proper quantum theory of gravitation in curved spacetime. ## References * [1] A. Einstein, Jahrb. Radioaktivität Elektronik 4 (1907) 414. * [2] A. Einstein, Ann. Phys. 35 (1911) 898. * [3] G. W. Gibbons, S. W. Hawking, Phys. Rev. D15 (1977) 2752. * [4] See, for example, R. M. Wald, General Relativity (The University of Chicago Press, Chicago 1984). * [5] There are still conributions to the boundary term emerging from the edges of our box. Those contributions, however, may be cancelled, if necessary, by means of an addition of an appropriate, constant boundary term. * [6] See, for example, H. Goldstein, Classical Mechanics (Second Edition) (Addison-Wesley, 1980). * [7] J. Mäkelä, Entropy 13 (2011) 1324. * [8] J. Mäkelä, Int. J. Mod. Phys. D23 (2014) 1450001. * [9] J. Mäkelä, Phys. Rev. D87 (2013) 104040. * [10] J. Mäkelä, Phys. Rev. D91 (2015) 124050. * [11] J. Mäkelä, Phys. Rev. D93 (2016) 084002. * [12] J. Mäkelä, Int. J. Mod. Phys. D28 (2019) 1950129. * [13] J. Mäkelä, Phys. Rev. D103 (2021) 027002. * [14] J. Mäkelä, Int. J. Mod. Phys. D31 (2022) 2250082. * [15] W. G. Unruh, Phys. Rev. D14 (1976) 5670.
# Activation of a supercooled liquid confined in a nanopore. Felix Mercier Laboratoire de Photonique d’Angers EA 4464, Université d’Angers, Physics Department, 2 Bd Lavoisier, 49045 Angers, France Gaetan Delhaye Laboratoire de Photonique d’Angers EA 4464, Université d’Angers, Physics Department, 2 Bd Lavoisier, 49045 Angers, France Victor Teboul <EMAIL_ADDRESS>Laboratoire de Photonique d’Angers EA 4464, Université d’Angers, Physics Department, 2 Bd Lavoisier, 49045 Angers, France ###### Abstract It is well stablished that confinement of supercooled liquids in nano-pores induces various effects as a strong modification of the dynamics and a layering of the local structure. In this work we raise the issue as how these confinement effects are modified when the liquid is out of equilibrium. To answer that question, we use molecular dynamics simulations to investigate the effect of confinement on a supercooled liquid activated by the periodic folding of a molecular motor. We find that the motor’s opening angle controls the activation of the medium and use that result to study the effect of different activations on the confined supercooled liquid. We observe an increase of the activation effect on the dynamics when the medium is confined. We find that the confinement slowing down dependence with the pore radius depends on the activation of the liquid. We argue that these findings result from a modification of dynamic correlation lengths with activation. In this picture the activation permits to control the liquid correlation length and dynamics inside the pore. Studying the local structure we observe a modification of the layering organization induced by the activation. We also find that the mobility inside the pore depends on the layers, being larger where the local density is small. dynamic heterogeneity,glass-transition ###### pacs: 64.70.pj, 61.20.Lc, 66.30.hh ## I Introduction Liquids and solids constituted or comprising molecules like molecular motors that are able to move by themselves are called activeactive1 ; active5 ; active-confi . An example is cells of living biological organisms that are constituted of active soft materials. Active matter opens a window to the statistical physics of non equilibrium systems and to biological physics making it a fascinating new domain of research that develops rapidlyactive1 ; active2 ; active3 ; active4 ; active5 ; active6 ; active7 ; active8 ; active9 ; active10 ; active11 ; active12 ; active13 ; active14 ; active15 ; active16 ; active17 ; active18 ; active-confi . The different statistical physics associated with non equilibrium systems, induces a number of new phenomena, as for example non-equilibrium phase transitionsactive4 . When liquids are cooled fast enough to remain liquid below their melting temperature $T_{m}$, they are subject to a large increase of their viscosity. The viscosity below $T_{m}$ then increases at least exponentially when the temperature drops, leading eventually to a glass at the glass-transition temperature $T_{g}$. That dramatic increase of the viscosity and the associated decrease of most dynamical properties like the diffusion coefficient, strangely appears without significant structural modifications of the liquid. The nature of that transition from the liquid to the glass, known as the long standing glass-transition problem, is still puzzling scientistsanderson ; gt0 ; gt1 ; gt2 ; dh0 ; c3 ; c4 ; md16 . As phase transitions are usually associated with the divergence of correlation lengths and the apparition of cooperative mechanisms, increasing correlation lengths have been searched extensively in supercooled liquids. Confinement of the liquid inside nano-pores has been seen as a way to cutoff the correlation lengths and obtain indirect informations on them by studying the associated dynamical modifications in the liquid. For that fondamental reason as well as multiple practical applications, confinement has been the subject of a number of studies and is still a subject of active researchconff1 ; conff2 ; conff3 ; conff4 ; conff5 ; conff6 ; conf-1 ; conf0 ; conf1 ; conf2 ; conf3 ; conf4 ; conf5 ; conf6 ; confine ; conf7 ; active-confi . Indeed, in supercooled liquids, confinement induces various effects. As suggested by the theoretical perspective described above, confinement in most cases induces a slowing down of the dynamics reminiscent of the slowing down induced by a temperature decrease. It also induces a structural organization called layering due to the fixation of the local structure in front of the immobile walls of the pore. In this work we raise the question to how the confinement effects are modified for active matter. In order to answer that question, we use a simple butterfly-like flat molecular motormotoro1 ; motoro2 ; motoro3 ; motoro4 ; motoro5 ; motoro6 ; motoro7 ; motoro8 ; motoro9 ; motoro10 ; motoro11 (Figure 1), to activate with its periodic folding the supercooled liquid surrounding it, confined in nano-pores of various radii (Figures 2 and 3). Experimentally our motor can be seen as a simplification of the azobenzene molecule or derivatives, which undergoes a photo-isomerization processaz1 ; az2 ; az3 ; az4 ; az5 ; az6 ; az7 ; az8 ; az9 ; cage ; md16 ; az10 when subject to a light stimulus. We study the effect of different activations induced by different opening angles of our motor (Figure 1). We observe an increase of the activation effect on the dynamics when the medium is confined. Studying the local structure we observe a modification of the layering organization induced by the activation. We also find that the motility is dependent on the layers, being larger where the local density is small. Figure 1: (color online) Picture of the folded and unfolded motor molecule. Through this work we use the parameter $d_{z}$ to tune and quantify the activation of the medium by the motor’s folding. We validate that choice at posteriori in the excitation density section. Figure 2: (color online) Snapshots of simulations for pores of radii (a) $R=5$Å, (b) $R=8$Å, (c) $R=10$Å, and (d) $R=12.5$Å. The colors are arbitrary. $T=500K$. The simulations use $8$ pores of radii varying from $5$ to $12.5$Å separated by steps of $1$ Å ($1.5$ Å for the last step). Figure 3: (color online) Snapshot of a simulation observed from the side to show the motor. The pore radius is $R=10$Å. $T=500K$. ## II Calculation In addition to theoretical and experimental methods, simulations in particular molecular dynamics and Monte Carlo simulationsmd1 ; md2 ; md2b ; md4 together with model systemsms1 ; ms2 ; ms3 ; ms4 ; ms5 are now widely used to unravel unsolved problems in condensed matter and complex systems physicskeys ; md3 ; md4b ; md6 ; md7 ; md8 ; md9 ; ee2 ; md11 ; md12 ; md13 ; md14 ; md15 ; ee1 ; finite1 ; u1 . The reader will find details on our simulation procedure in previous paperspccp ; prefold ; ariane , however for convenience we will resume it here. Our simulations use one motor molecule (see Figure 1 for its description) imbedded inside a medium constituted of $1760$ linear molecules, in a parallelepiped box $31.1$ Å wide and $62.2$ Å high. After aging the simulation box during $10ns$ at the temperature of studies, we create the pore by suddenly freezing molecules located at a distance larger than $R$ from the center axis of our simulation box. This procedure is intended to minimize the structure modification of the liquid by the confinement. After their sudden freezing, the glassy molecules of the pores are no more allowed to move, leading to pure elastic interactions with the confined liquid. Consequently the liquid temperature is not modified by the interactions with the pore walls. After that procedure, as the dynamics changes with the confinement, the liquid is again aged during $10ns$ before the beginning of any study. We integrate the equations of motion using the Gear algorithm with a quaternion decompositionmd1 and a time step $\Delta t=10^{-15}s$. When the motor is active our simulations are out of equilibrium, because the motor’s folding periodically release energy into the system. However our system is in a steady state and does not age as explained below. For that purpose we use the Berendsen thermostatberendsen to evacuate from the system the energy created by the motor’s folding. Notice that the thermostat applies to the medium’s molecules but not to the motor. Thus the motor’s cooling results only from the molecular interactions with the medium’s molecules. We use the NVT canonic thermodynamic ensemble as approximated by Berendsen thermostat (see ref.finite2 for an evaluation of the effect of the thermostat on our calculations), and periodic boundary conditions. The molecules of the medium (host)ariane are constituted of two rigidly bonded atoms ($i=1,2$) at the fixed interatomic distance $l_{h}$$=1.73$Å. These atoms interact with atoms of other molecules with the following Lennard-Jones potentials: $V_{ij}=4\epsilon_{ij}((\sigma_{ij}/r)^{12}-(\sigma_{ij}/r)^{6})$ (1) with the parametersariane : $\epsilon_{11}=\epsilon_{12}=0.5KJ/mol$, $\epsilon_{22}=0.4KJ/mol$, $\sigma_{11}=\sigma_{12}=3.45$Å, $\sigma_{22}=3.28$Å. The mass of the motor is $M=540g/mole$ (constituted of $18$ atoms, each one of mass $30g/mole$) and the mass of the host molecule is $m=80g/mole$ ($2$ atoms with a mass of $40g/mole$ each). We model the motor with $18$ atoms in a rectangular shape constituted of two rows of $9$ rigidly bonded atoms. The width of the swimmer is $L_{s}=4.4$Å and its length $l_{s}=15.4$Å. The length of the host molecule is $l_{h}=5.09$Å and its width $L_{h}=3.37$Å. The motor’s atoms interact with the medium’s atoms using mixing rules and a Lennard-Jones interatomic potential on each atom of the motor, defined by the parameters: $\epsilon_{33}=1.88KJ/mol$, $\sigma_{33}=3.405$Å. We use the following mixing rules mix1 ; mix2 : $\epsilon_{ij}=(\epsilon_{ii}.\epsilon_{jj})^{0.5};\sigma_{ij}=(\sigma_{ii}.\sigma_{jj})^{0.5}$ (2) for the interactions between the motor and the host atoms. Our medium is a fragile liquidfragile1 ; fragile2 that falls out of equilibrium in bulk simulations below $T=340K$, i.e. $T=340$ K is the smallest temperature for which we can equilibrate the bulk liquid when the motor is not active. Consequently the bulk medium above that temperature behaves as a viscous supercooled liquid in our simulations while below it behaves as a solid (as $t_{simulation}<\tau_{\alpha}$). Notice however that when active, the motor’s motions induce a fluidization of the medium around itmd16 ; flu1 ; flu2 ; flu3 ; flu4 ; carry ; flu5 ; rate , an effect that has been experimentally demonstratedflu1 ; flu2 ; flu3 ; flu4 ; flu5 with azobenzene photo- isomerizing molecules embedded in soft matter. We evaluate the glass transition temperature $T_{g}$ in the bulk to be slightly smaller $T_{g}\approx 250K$. Notice that as they are modeled with Lennard-Jones atoms, the host and motor potentials are quite versatile. Due to that property, a shift in the parameters $\epsilon$ will shift all the temperatures by the same amount, including the glass-transition temperature and the melting temperature of the material. Each folding is modeled as continuous, using a constant quaternion variation, with a folding time $\tau_{f}=0.3ps$. The total cycle period is constant and fixed at $\tau_{p}=400ps$. The folded and unfolded part of the cycle have the same duration equal to $\tau_{p}/2=200ps$ including the $0.3ps$ folding or unfolding. Our system, while out of equilibrium, is in a steady state and is not aging. That behavior is obtained because the energy released by the motor into the medium is small enough and the time lapse between two stimuli large enough for the system to relax before a new stimuli appears. In other words we are in the linear response regimepccp . Through this work we use the mean square displacement to obtain informations on the molecules mobility and diffusion inside the pore. The mean square displacement is defined asmd1 : $\displaystyle{<r^{2}(t)>={1\over N.N_{t_{0}}}\sum_{i,t_{0}}\mid{{\bf r}_{i}(t+t_{0})-{\bf r}_{i}(t_{0})}}\mid^{2}$ (3) From the time evolution of the mean square displacement we then calculate the diffusion coefficient $D$ for diffusive displacements using the Stokes- Einstein equation: $\displaystyle{\lim_{t\to\infty}<r^{2}(t)>=6Dt}$ (4) Finally we will use the Non Gaussian parameter $\alpha_{2}(t)$ to measure the extent of cooperative motions inside the medium. $\displaystyle{\alpha_{2}(t)=\frac{3<r^{4}(t)>}{5<r^{2}(t)>^{2}}-1}$ (5) Unless otherwise specified the results presented in this work correspond to a temperature $T=500K$ at which our model liquid is supercooled. Notice however that our system is a model system with Lennard-Jones interactions only. Therefore, due to the properties of the Lennard-Jones potential, one can shift the temperatures by any factor $\alpha$ ($T^{\prime}=\alpha T$) creating a new material. That new material will have all the $\epsilon$ parameters of the LJ potentials (including for the motor) modified in the same way $\epsilon^{\prime}=\alpha\epsilon$. The results corresponding to that new material will have to be translated from the results of our work by a shift of time $t^{\prime}=t/\alpha^{0.5}$. Similarly, the size of the atoms can be scaled by a factor $\gamma$ ($\sigma^{\prime}=\gamma\sigma$) provided that all the distances of the system are scaled by the same factor. Our results will in that case have to be translated by a shift of distance $\gamma$ ($r^{\prime}=\gamma r$) and a shift of time $\gamma$ ($t^{\prime}=\gamma t$). In that way our simulations are valid for a large number of materials although approximately, and experimentalists can adjust our data to their system of concern. ## III Results and discussion ### III.1 Diffusion inside the pore Figure 4: (color online) Mean square displacement of the medium molecules located at a distance $r<10$ Å from the motor at time $t=0$, for different pore radii $R$. The activation parameter $d_{z}$ is different for each Figure. $Z$ is the direction of the pore axis and $z$ the direction perpendicular to the motor’s plane in the motor’s molecular frame. In Figure (a) $d_{z}=0$ meaning that there is no activation and the motor acts as a passive probe. Figures 4a 4b 4c and 4d show the mean square displacement of the confined medium for different pore radii ranging from the Bulk to a pore radius $R=4$Å, each Figure corresponding to a different activation quantified by the motion of the arms $d_{z}$ (Figure 1). For a purely passive motor molecule (Figure 4a), the pore effect is important. As the pore size decreases, the diffusion inside the medium decreases (i.e. the viscosity increases) leading for pores of radii smaller than 11Å to a solid non diffusive medium. This confinement effect is due to the decrease of the probability of molecules near the wall to escape the cage created by the molecules surrounding them, as part of their cage is created by the pore wall and thus never opens. Notice that, this well established confinement effect conf-1 ; conf0 ; conf1 ; conf2 ; conf3 ; conf4 ; conf5 ; conf6 ; confine ; conf7 , displays a number of similarities with the effect of a decrease in temperature in supercooled liquids. As expected, when the motor activates the medium in Figures 4b to 4d, the diffusion inside the pores increases. However, surprisingly it increases more in the small pores (see for example the rapid increase for $R=5$Å), than in the larger ones, leading to a decrease of the difference in diffusions. Eventually, for the larger activation investigated in Figure 4d, the diffusion inside the different pores is rather similar. Thus we observe in Figure 4 that the activation progressively suppresses the effect of confinement. Figure 5: (color online) Diffusion coefficient of the medium molecules for different pore radii $R$ and activation parameters $d_{z}$. The evolution of the diffusion coefficient in Figure 5 resumes that behavior. The upper circles on the Figure show the effect of the larger activation by the motor on the medium’s diffusion for various pore radii. The diffusion coefficients appear roughly constant for that large activation, even increasing slightly when the pore radius decreases. As discussed before, the activation of the medium inside the pore washes out the slowing down induced by the confinement. For an activation slightly weaker the second rank of circles from the top show a constant diffusion for the different pores. When the activation decreases the pore slowing down effect increases but is still much smaller in the Figure than for the non-activated medium ($d_{z}=0$). ### III.2 Activation effect Figure 5 shows that the diffusion coefficients follow exponentially decreasing laws with the inverse of the pore radius $1/R$. $\displaystyle{D=D_{0}exp(-{\zeta}/R)}$ (6) Where $\zeta$ is a correlation length that depends on the activation parameter $d_{z}$. As a result, one can explain the decrease of the pore effect with the activation, by the decrease of that correlation length of the medium when it is activated. For our largest activation the correlation length $\zeta$ tends to zero and the pore effect disappears. Figure 6: (color online) Mean square displacement of the medium molecules for various activation parameters $d_{z}$; (a) in the bulk ($R=\infty$); (b) for a pore of radius $R=5$Å. The activation dependence appears much larger in the pore. Figures 6(a) and (b) compare the effect of the activation in the bulk and in a pore of radius $R=5$Å. Figure 6 shows a much larger activation effect on the medium’s displacements inside the small pore than in the bulk. For a time lapse of $10ns$ the mean square displacement in the pore varies indeed from $10^{-1}$Å2 to $4.10^{1}$Å2, that is by a factor $400$ depending on the activation, while in the bulk it varies by a mere factor $3$. This result agrees well with the findings of previous section that the diffusion for large activation doesn’t depend on the pore size. As the pore radius decreases, the slowing down by the pore wall increases, but the activation effect also increases compensating that slowing down. Notice that as the pore radius decreases, the relative number of motor to free molecules increases due to the geometry. While results we show take only into account molecules a distance $r<10$Å apart from the motor, reducing that geometrical effect, it doesn’t eliminate it. Geometry appears consequently as a possible cause for the increase of the activation impact on the diffusion for small pores. In order to conclude on that possibility we calculated the mean square displacements at a smaller distance $r<5$Å from the motor. The results (not shown) were very similar to the results of Figure 6, showing that the increase of the activation effect on diffusion for small pores, is not arising from the geometry (i.e. from the mean distance to the motor). It is interesting to point out however that the increase of the effect of activation when the pore size decreases doesn’t increase the diffusion to values larger than bulk values but tends instead to reach the bulk values from below. That point comes in support of the decrease of the correlation lengths explanation. ### III.3 Cooperative motions Figure 7: (color online) Non Gaussian parameter $\alpha_{2}(t)=\frac{3<r^{4}(t)>}{5<r^{2}(t)>^{2}}-1$ around the motor ($r<10$Å) for various activation parameters $d_{z}$ and for a pore of radius $R=5$Å, at $T=500K$. Spontaneous (thermal)dh0 and activation-induced cooperative motionsmd16 ; c3 ; c4 are present in supercooled liquids and have been well documented. Because cooperativity is directly related to the physical correlation length of the liquid, its increase or decrease can give us information on that correlation length. Therefore, to shed some light on the correlation length explanation, we now study the extent of cooperative motions in our system. For that purpose we display in Figure 7 the Non Gaussian parameter $\alpha_{2}(t)$ evolution with the activation. Cooperative motions in supercooled liquids induce a tail in the Van Hove self correlation function, leading to its departure from the Gaussian shape. For that reason the departure from the Gaussian shape of the Van Hove, measured by the Non Gaussian coefficient, has been used extensively as a measure of the cooperative motions. In Figure 7 on the right $\alpha_{2}(t)$ shows peaks at times corresponding to multiple of the motor’s folding period, that are the result of the excitations induced by the motor’s folding on the medium. The activation-induced cooperativity as measured by $\alpha_{2}(t)$ decreases when we increase the activation parameter $d_{z}$ from $3.3$Å to $4.9$Å. That result comes again in support of the decrease of the medium’s correlation length with the activation. Notice however that as expected when the activation parameter is null the induced cooperativity disappears as there is no more activation. This result suggests an increase of the cooperativity for small activations and an optimum. ### III.4 Excitation density To better understand previous results, we now turn our attention to the density of excitations inside the pores. Following Keys et al. keys we define excitations as elementary diffusion processes. In our work, excitations are molecules that move in a time lapse $\Delta t=10ps$ a distance larger than $\Delta r=1$ Å from their previous position. Figure 8 shows the excitation density for various pore radii and excitation processes. We find in Figure 8 that the density of excitation increases with the activation whatever the pore radius. That result validates at posteriori our choice of the activation parameter as the displacement of the motor’s arm $d_{z}$. Interestingly the maximum excitation density is not always located at the pore center. This effect could be due to layering inside the pores. Layering induces oscillations of the density inside the pores and low densities promote excitations as cages are then loose, letting molecules escape easier. We will test that interpretation in Figure 9. Figure 8 shows that when the pore size decreases the excitation density decreases much less rapidly for large activations. Therefore the excitation density is larger than expected for small pores when the activation is on, in agreement with the activation effect described in a previous section. That result suggests that the larger than expected diffusion observed for small pores (the activation effect) is due to a larger density of excitations that is a larger number of mobile molecules as opposed to larger motions of some molecules. Figure 8: (color online) Excitation density $\rho_{excitation}$ as a function of the distance $r$ from the pore central axis, for different pore radii $R$ and activation parameters $d_{z}$. Excitations are here defined as medium’s molecules motion larger than $\Delta r=1$ Å occuring during a time lapse $\Delta t=10ps$. ### III.5 Structural organization We will now investigate the local structure inside the pores. In Figure 9 we display the local density inside the pores as a function of the distance $r$ from the pore center. That is, $r$ is the radius of cylindrical coordinates, $z$ being the pore axis. The Figure shows a clear layering effect, that we highlight with a line when the activation is off. The local density is in most cases as expected maximum at the center of the pore ($r=0$) and minimum around the wall. However this is not always the case. For instance for a pore of $R=5$Å radius, the maximum density is not located in the center of the pore, and around the wall for large activations the density is not null. For a pore of $R=8$Å radius, there is a maximum density at the pore center but it corresponds to a thin peak, and we observe a wider maximum apart from the center. As expected, the layering explains the excitation density maxima shifted from the center of the pore, as the local density is actually smaller at the location of excitation density maxima. Moreover an unexpected effect appears, the layering changes with the activation. Not only it decreases slightly but the oscillations locations are also modified. Work is in progress to better understand that interesting effect. Figure 9: (color online) Density $\rho$ of medium molecules inside the pore as a function of the distance $r$ from the pore central axis. The Figures show that result for different pore radii $R$ and activation parameters $d_{z}$. The values are normalized by the average density inside the pore $\rho_{0}$. The dashed line displays the layering when the motor is not active. The figures show that the layering is strongly modified with the activation. ## IV Conclusion In supercooled liquids, nano-confinement modifies the dynamical properties of the liquid usually inducing an important slowing down of the dynamics. Confinement also induces a structural organization called layering. In this work we investigated how these confinement effects are modified when the liquid is out of equilibrium. Results show that the dynamical slowing down induced by the confinement is strongly modified by the activation, leading to a diffusion that may not depend anymore on the pore diameter. In agreement with that finding our results show that the activation effect is much larger upon confinement than in the bulk, increasing rapidly when the pore size decreases. We interpret these effects as arising from an activation induced decrease of the dynamical correlation lengths of the liquid. In agreement with that picture we observe a decrease of the cooperative motions upon activation. While the structure of the liquid is unchanged, we observe a decrease of the layering inside the pore but also a qualitative modification of the layering with the activation. Eventually we observe a relation between the layering and the elementary diffusion processes called excitations, a result strongly suggesting that the layering triggers an increase of the mobility in the less dense parts of the pore. Conflict of interest There are no conflict of interest to declare. Data availability The data that support the findings of this study are available from the corresponding author upon reasonable request. ## References * (1) O. Dauchot, H. Lowen, J. Chem. Phys. 151, 114901 (2019). * (2) C. Bechinger, R. Di Leonardo, H. Lowen, C. Reichhardt, G. Volpe, G. Volpe, Rev. Mod. Phys. 88, 045006 (2016). * (3) P. Liu, H. Zhu, Y. Zeng, G. Du, L. Ning, D. Wang, K. Chen, Y. Lu, N. Zheng, F. Ye, M. Yang, P.N.A.S. 117, 11901 (2020). * (4) A.P. Solon, J. Stenhammar, R. Wittkowski, M. Kardar, Y. Kafri, M.E. Cates, J. Tailleur, Phys. Rev. Lett. 114, 198301 (2015). * (5) A.P. Solon, J. Stenhammar, M.E. Cates, Y. Kafri, J. Tailleur, Phys. Rev. E 97, 020602(R) (2018). * (6) T. Vicsek, A. Czirok, E. Ben-Jacob, I. Cohen, O. Shochet, Phys. Rev. Lett. 75, 1226 (1995). * (7) J. Elgeti, R.G. Winkler, G. Gompper, Rep. Prog. Phys. 78, 056601 (2015). * (8) E.W. Burkholder, J.F. Brady, J. Chem. Phys. 150, 184901 (2019). * (9) G. Junot, G. Briand, R. Ledesma-Alonso, O. Dauchot, Phys. Rev. Lett. 119, 028002 (2017). * (10) J.U. Klamser, S. Kapfer, W. Krauth, J. Chem. Phys. 150, 144113 (2019). * (11) N. Nikola, A.P. Solon, Y. Kafri, M. Kardar, J. Tailleur, R. Voituriez, Phys. Rev. Lett. 117, 098001 (2016). * (12) H. Wang, T. Qian, X. Xu, Soft Matter 17, 3634 (2021). * (13) S. Dal Cengio, D. Levis, I. Paganobarraga, J. Stat. Mech. 043201 (2021). * (14) H.G. Wood, J.A. Hanna, Soft Matter 17, 3137 (2021). * (15) C. Reichhardt, C.J.O. Reichhardt, Phys. Rev. E 103, 022602 (2021). * (16) J. Denk, E. Frey, P.N.A.S. 117, 31623 (2020). * (17) V. Teboul, G. Rajonson, J. Chem. Phys. 150, 214505 (2019). * (18) P. Herrera, M. Sandoval, Phys. Rev. E 103, 012601 (2021). * (19) D. Dattler, G. Fuks, J. Heiser, E. Moulin, A. Perrot, X. Yao, N. Giuseppone, Chem. Rev. 120, 310 (2020). * (20) P.W. Anderson, Science 267, 1610 (1995) * (21) K. Binder, W. Kob, Glassy materials and disordered solids World Scientific (Singapore) 2011 * (22) P.G. Debenedetti, Metastable liquids Princeton University Press (Princeton) 1996 * (23) P.G. Wolynes, V. Lubchenko, Structural glasses and supercooled liquids Wiley (Hoboken) 2012 * (24) L. Berthier, G. Biroli, J.P. Bouchaud, L. Cipelletti, W. Van Saarloos, Dynamical heterogeneities in glasses, colloids and granular media, Oxford Science Publications, Oxford 2011. * (25) V. Teboul, J. B. Accary, Phys. Rev. E 89, 012303 (2014). * (26) V. Teboul, J. B. Accary, M. Chrysos, Phys. Rev. E 87, 032309 (2013). * (27) V. Teboul, M. Saiddine, J.M. Nunzi, Phys. Rev. Lett. 103, 265701 (2009). * (28) A. Jani, M. Busch, J.B. Mietner, et al., J. Chem. Phys. 154, 094505 (2021). * (29) R. Mhanna, P. Catrou, S. Dutta, et al., J. Phys. Chem. B 124, 3152 (2020). * (30) B. Malfait, A. Pouessel, A. Jani, D. Morineau, J. Phys. Chem. Lett. 11, 5763 (2020). * (31) S. Fei, W.L. Hsu, J.J. Delaunay, et al., J. Chem. Phys. 154, 144503 (2021). * (32) P. Bampoulis, J. Chem. Phys. 154, 134705 (2021). * (33) P. Sepulveda-Medina, M. Tyagi, C. Wang, et al., J. Chem. Phys. 154, 154903 (2021). * (34) L. Stefanutti, L.E. Bove, G. Lelong, M.A. Ricci, A.K. Soper, F. Bruni, Phys. Chem. Chem. Phys. 21, 4931 (2019). * (35) J. Swenson, Phys. Chem. Chem. Phys. 20, 30095 (2018). * (36) A.A. Milischuk, B.M. Ladanyi, J. Chem. Phys. 135, 174709 (2011). * (37) N. Kuon, A.A. Milischuk, B. M. Ladanyi, E. Flenner, J. Chem. Phys. 146, 214501 (2017). * (38) J. Swenson, H. Jansson, R. Bergman, Phys. Rev. Lett. 96, 247802 (2006). * (39) S.M. Saparov, J.R. Pfeifer, L. Al-Momani, G. Portella, B.L. deGroot, U. Koert, P. Pohl, Phys. Rev. Lett. 96, 148101 (2006). * (40) V. Teboul, Int. J. Nanotechnol. 5, 851 (2008). * (41) D.J. Mann, M.D. Halls, Phys. Rev. Lett. 90, 195503 (2003). * (42) M.C. Bellissent-Funel, J. Phys.: Condens. Matter 13, 9165-9177 (2001). * (43) R. Bergman, J. Swenson, Nature 403, 283-286 (2000). * (44) I. Aprahamian, ACS Central Science 6, 347 (2020). * (45) J. Wang, Nanomachines: Fundamentals and Applications, Wiley, Weinheim 2013. * (46) A.P. Davis, Nature 401, 120-121 (1999). * (47) J.P. Sauvage, _Molecular machines and motors_ , Springer, Berlin, 2001. * (48) J.R. Howse, R.A.L. Jones, A.J. Ryan, T. Gough, R. Vafabakhsh, R. Golestanian, Phys. Rev. Lett. 99, 048102 (2007). * (49) H. Karani, G.E. Pradillo, P.M. Vlahovska, Phys. Rev. Lett. 123, 208002 (2019). * (50) F. Novotny, M. Pumera, Scien. Rep. 9, 13222 (2019). * (51) X. Arque, A. Romero-Rivera, F. Feixas, T. Patino, S. Osuna, S. Sanchez, Nature Comm. 10, 2826 (2019). * (52) A.M. Brooks, M. Tasinkevych, S. Sabrina, D. Velegol, A. Sen, K.J.M. Bishop, Nature Comm. 10, 495 (2019). * (53) P. Pietzonka, E. Fodor, C. Lohrmann, M.E. Cates, U. Seifert, Phys. Rev. X 9, 041032 (2019). * (54) C. Calero, J. Garcia-Torres, A. Ortiz-Ambriz, F. Sagues, I. Pagonabarraga, P. Tierno, Nanoscale 11, 18723 (2019). * (55) A. Natansohn, P. Rochon, Chem. Rev. 102, 4139-4175 (2002). * (56) J.A. Delaire, K. Nakatani, Chem. Rev. 100, 1817 (2000). * (57) G.S. Kumar, D.C. Neckers, Chem. Rev. 89, 1915 (1989). * (58) K.G. Yager, C.J. Barrett, Curr. Opin. Solid State Mater. Sci. 5, 487 (2001). * (59) T.G. Pedersen, P.M. Johansen, Phys. Rev. Lett. 79, 2470-2473 (1997). * (60) T.G. Pedersen, P.M. Johansen, N.C.R. Holme, P.S. Ramanujam, S. Hvilsted, Phys. Rev. Lett. 80, 89-92 (1998). * (61) J. Kumar, L. Li, X.L. Jiang, D.Y. Kim, T.S. Lee, S. Tripathy, Appl. Phys. Lett. 72, 2096-2098 (1998). * (62) C.J. Barrett, P.L. Rochon, A.L. Natansohn, J. Chem. Phys. 109, 1505-1516 (1998). * (63) C.J. Barrett, A.L. Natansohn, P.L. Rochon, J. Phys. Chem. 100, 8836-8842 (1996). * (64) V. Teboul, M. Saiddine, J.M. Nunzi, J.B. Accary, J. Chem. Phys. 134, 114517 (2011). * (65) P. Lefin, C. Fiorini, J.M. Nunzi, Pure Appl. Opt. 7, 71-82 (1998). * (66) M.P. Allen, D.J. Tildesley, _Computer Simulation of Liquids_ , Oxford University Press, New York 1990. * (67) M. Griebel, S. Knapek, G. Zumbusch, _Numerical Simulation in Molecular Dynamics_ , Springer-Verlag, Berlin 2007. * (68) D. Frenkel, B. Smit, _Understanding Molecular Simulation_ , Academic Press, San Diego 1996. * (69) S. Chaussedent, V. Teboul, A. Monteil, Curr. Opin. Solid State Mater. Science 7, 111-116 (2003). * (70) F. Ritort, P. Sollich, Adv. Phys. 52, 219 (2003). * (71) G. H. Fredrickson, H. C. Andersen, Phys. Rev. Lett. 53, 1244 (1984). * (72) J. Jackle, S. Eisinger, Z. Phys. B 84, 115 (1991). * (73) W. Kob and H. C. Andersen, Phys. Rev. E 48, 4364 (1993). * (74) V. Teboul, J. Chem. Phys. 141, 194501 (2014). * (75) C.F.E. Schroer, A. Heuer, Phys. Rev. Lett. 110, 067801 (2013). * (76) J.B. Accary, V. Teboul, J. Chem. Phys. 136, 0194502 (2012). * (77) A. Furukawa, K. Kim, S. Saito, H. Tanaka, Phys. Rev. Lett. 102, 016001 (2009). * (78) T. Iwashita, T. Egami, Phys. Rev. Lett. 108, 196001 (2012). * (79) T. Gleim, W. Kob, K. Binder, Phys. Rev. Lett. 81, 4404 (1998). * (80) E.J. Saltzman, K.S. Schweitzer, J. Chem. Phys. 125, 044509 (2006). * (81) Y. Le Duff, V. Teboul, Phys. Lett. A 157, 44 (1991). * (82) V. Teboul, Y. Le Duff, T. Bancewicz, J. Chem. Phys. 103, 1384 (1995). * (83) B. Uralcan, I.A. Aksay, P.G. Debenedetti, D. Limmer, J. Phys. Chem. Lett. 7, 2733 (2016). * (84) J. Zhang, F. Muller-Plathe, F. Leroy, Langmuir 31, 7544 (2015). * (85) E. Flenner, G. Szamel, J. Phys. Chem. B 119, 9188 (2015). * (86) E. Flenner, G. Szamel, Nature Comm. 6, 7392 (2015). * (87) L.C. Fai, V. Teboul, A. Monteil, S. Maabou, I. Nsangou, Cond. Matt. Phys. 8, 639 (2005). * (88) Y.S. Elmatad, A.S. Keys, Phys. Rev. E 85, 061502 (2012). * (89) V. Teboul, Eur. Phys. J. B 51, 111 (2006). * (90) A. S. Keys, L. O. Hedges, J.P. Garrahan, S.C. Glotzer, D. Chandler, Phys. Rev. X 1, 021013 (2011). * (91) A.P. Kerasidou, Y. Mauboussin, V. Teboul, Chem. Phys. 450, 91 (2015). * (92) S. Ciobotarescu, S. Bechelli, G. Rajonson, S. Migirditch, B. Hester, N. Hurduc, V. Teboul, Phys. Rev. E 96, 062614 (2017). * (93) S. Ciobotarescu, N. Hurduc, V. Teboul, Phys. Chem. Chem. Phys. 18, 14654 (2016). * (94) H.J.C. Berendsen, J.P.M. Postma, W. Van Gunsteren, A. DiNola, J.R. Haak, J. Chem. Phys. 81, 3684-3690 (1984). * (95) S. Taamalli, J. Hinds, S. Migirditch, V. Teboul, Phys. Rev. E 94, 052604 (2016). * (96) R.J. Good, C.J. Hope, J. Chem. Phys. 53, 540 (1970). * (97) J. Delhommelle, P. Millie, Mol. Phys. 99, 619 (2001). * (98) C.A. Angell, Science 267, 1924-1935 (1995). * (99) N.A. Mauro, M. Blodgett, M.L. Johnson, A.J. Vogt, K.F. Kelton, Nature Comm. 5, 4616 (2014). * (100) M. Saiddine, V. Teboul, J.M. Nunzi, J. Chem. Phys. 133, 044902 (2010). * (101) V. Teboul, R. Barille, P. Tajalli, S. Ahmadi-Kandjani, H. Tajalli, S. Zielinska, E. Ortyl, Soft Matt. 11, 6444 (2015). * (102) J.B. Accary, V. Teboul, J. Chem. Phys. 139, 034501 (2013). * (103) P. Karageorgiev, D. Neher, B. Schulz, B. Stiller, U. Pietsch, M. Giersig, L. Brehmer, Nature Mater. 4, 699-703 (2005). * (104) G.J. Fang, J.E. Maclennan, Y. Yi, M.A. Glaser, M. Farrow, E. Korblova, D.M. Walba, T.E. Furtak, N.A. Clark, Nature Comm. 4, 1521 (2013). * (105) N. Hurduc, B.C. Donose, A. Macovei, C. Paius, C. Ibanescu, D. Scutaru, M. Hamel, N. Branza-Nichita, L. Rocha, Soft Mat. 10, 4640-4647 (2014). * (106) J. Vapaavuori, A. Laventure, C.G. Bazuin, O. Lebel, C. Pellerin, J. Amer. Chem. Soc. 137, 13510 (2015).
Original Article Journal Section Cuneyt Gurcan Akcora, Departments of Computer Science and Statistics, University of Manitoba, Winnipeg, MB R3T-2N2, Canada <EMAIL_ADDRESS>NSF DMS 1925346, NSF ECCS 2039716, RGPIN-2020-05665 # Blockchain Networks: Data Structures of Bitcoin, Monero, Zcash, Ethereum, Ripple and Iota Cuneyt Gurcan Akcora Departments of Statistics and Computer Science, University of Manitoba, Winnipeg, MB R3T-2N2, Canada Yulia R. Gel Department of Mathematical Sciences, University of Texas at Dallas, Richardson, TX75080, USA Murat Kantarcioglu Department of Computer Science, University of Texas at Dallas, Richardson, TX75080, USA ###### Abstract Blockchain is an emerging technology that has enabled many applications, from cryptocurrencies to digital asset management and supply chains. Due to this surge of popularity, analyzing the data stored on blockchains poses a new critical challenge in data science. To assist data scientists in various analytic tasks on a blockchain, in this tutorial, we provide a systematic and comprehensive overview of the fundamental elements of blockchain network models. We discuss how we can abstract blockchain data as various types of networks and further use such associated network abstractions to reap important insights on blockchains’ structure, organization, and functionality. ###### keywords: _blockchain_ , _bitcoin and litecoin_ , _ethereum_ , _monero_ , _zcash_ , _ripple_ , _iota_ ## 1 Introduction On October 31, 2008, an unknown person called Satoshi Nakamoto posted a white paper titled “Bitcoin: A Peer-to-Peer Electronic Cash System” to the Cyberpunks mailing list. In eight pages [1], Satoshi explained the network, transactions, incentives, and other building blocks of a digital currency that he called Bitcoin. Bitcoin solves the problem of sending and receiving digital currency on the world wide web. However, the idea of a digital currency is as old as the Internet. For similar purposes, traditional banks and Internet companies have created online payment services, such as Paypal, Visa, and Master. However, a trusted entity intermediates currency flows in these solutions and updates user balances as transactions are processed. Blockchain removes the trusted entity and provides a framework to process transactions and maintain user balances correctly and securely. Bitcoin and blockchain have been used interchangeably in the past, but Bitcoin is just one financial application among many other use cases of blockchain technology. Blockchain stores a limited number of transactions (i.e., coin transfers) in a data structure called a block, which is in turn stored in a public ledger. We may consider the ledger as a notebook that has information to calculate user balances written in it. Blockchain represents a user with a blockchain address as a fixed-length string of characters (e.g., 1aw345….). A transaction can be as simple as sending bitcoins from one address to another, along with the required digital signatures. Transaction size increases if more addresses are involved. Blocks contain executed transactions. Block sizes (which determine the number of transactions in a block) are usually small (e.g., 1MB in Bitcoin that allows approximately 3000 transactions). Blockchain uses an underlying peer-to-peer network to transmit blocks and proposed transactions between blockchain users worldwide. We will elaborate on “users” in later sections; however, users of the first blockchain, Bitcoin, are ordinary web users who can join the network by downloading and installing an application called a wallet. A copy of the ledger is stored locally at every participant (i.e., node) of the peer-to-peer Bitcoin network. Every user is supposed to check its blockchain copy to learn about user balances. Bitcoin ledger is extended by appending a new block to the end of the chain every 10 minutes through a process that is called mining. The mining is where transaction approval/verification and coin creation occur, and blockchain removes financial institutions from this process and allows ordinary users to serve in their role. Mining is achieved by solving a cryptographic puzzle by trial and error. A valid solution is a proof that the miner has completed some work and effort to find the answer to the puzzle. The answer itself is an integer called nonce that once appended to the end of block content, the hash of the content satisfies a predetermined difficulty goal. The nonce and the resulting hash are known as proof-of-work, which has its roots in a research article [2]. Mining of each block is an open competition worldwide. Any user may compete to solve the puzzle to earn a handsome sum called the block reward, creating new bitcoins. This process is similar to fiat money printing; however, coins are created in predefined quantities (50 BTC in 2009, the amount halves every four years) at each block and given to the miner. Mining is, by purpose, designed to be difficult so that miners will not litter the blockchain with blocks. However, the difficulty is adaptive and depends on the number of miners and their computing power. Furthermore, Bitcoin uses an adaptive difficulty level to reduce the frequency of block mining. The blockchain peer-to-peer network participants will receive the latest block in time and be aware of its transactions. Awareness prevents malicious users from spending the same coin multiple times. For example, Bitcoin updates the mining difficulty to create a ten-minute gap between two blocks on average, and Ethereum aims at 12-15 seconds per block. Figure 1: Block structure. We show transactions with squares within blocks. The canonical blockchain is the main chain that contains blocks and continues to grow in time. New blocks are expected to be mined on the top (rightmost part) of this blockchain. Non-transactional block data, such as block hash or nonce, are not shown in this figure, but edges indicate parents. Blocks that we depict in gray are called stale blocks because they do not appear in the canonical blockchain. At the right end of the blockchain, two blocks are competing to be the $19$th block. Both blocks are valid, but eventually, the canonical blockchain will include only one of them, and the other will be a stale block. The long transactions at the top of each block are coinbase transactions (rewards of miners) which contain the sum of newly minted coins (in red) and transaction fees (in green). Block 15 has no transaction other than the coinbase one. As nodes compete worldwide, miners create candidate blocks and append them to the chain. The competition is resolved by adopting the candidate that appears on the longest blockchain (see Figure 1). As miners append more blocks to the longest chain, the earlier blocks are deemed less likely to change, i.e., data in earlier blocks are considered final. In Figure 1 a blockchain has 19 blocks (the latest block, $b_{19}$ has two competing candidates). The longest chain is called the canonical path, and blocks on the chain are deemed valid (i.e., $b_{11}$ to $b_{19}$). Gray boxes show stale blocks that have failed to be a part of the blockchain. An arrow shows a parent-child relationship between blocks. For example, $b_{14}$ is the parent of $b_{15}$ ($b_{15}$ is mined later than $b_{14}$). Blockchains proliferate, and the ecosystem is in evolution. A diverse set of blockchains have proposed novel mechanisms in storing data and representing users, coin transfers, smart-contract operations, and more. As a result, mining blockchain data requires domain expertise which may seem daunting to data scientists. In this survey, we take a holistic view and define salient characteristics of six significant blockchains in terms of their data structures: Bitcoin, ZCash, Monero, Ethereum, Ripple, and IOTA. As Bitcoin is the first blockchain, we first teach the UTXO blockchains and design choices of Nakamoto that have significantly affected how new blockchains function and store data. The Ethereum section covers the basics of smart contracts, tokens, and other decentralized finance constructs. In the Ripple and IOTA sections, we teach their use cases beyond cryptocurrency and platform aspects. We conclude the survey with an overview of graph types found in these blockchains.pears on the longest blockchain ## 2 A Taxonomy of Blockchains and Transaction Networks After Bitcoin’s popularity, various improvements to blockchain have been suggested and implemented in other digital currencies. Bitcoin, so far, has been very conservative in changes to its core technology. A few common points stand out between all blockchain implementations in mining, coding capabilities, and data scalability solutions. We can broadly classify blockchains as i) private or public blockchains and ii) currency or platform blockchains. A third discussion point is about first vs. second layer technologies that determine how much data blockchains should store on-chain. Private, Consortium vs. Public. Most blockchains, such as Bitcoin and Ethereum, are permissionless. They store data publicly, and any user can mine a block (Ethereum 2.0, when implemented, will change the mining algorithm). Anyone can join the Blockchain network and download blocks to view the transactions in them. When used by companies, a public blockchain would reveal sensitive corporate data to outsiders. Instead, private blockchains, such as Hyperledger fabric by IBM [3], have been developed to restrict block mining rights or data access to one participant (i.e., private blockchain) or a few verified participants (i.e., consortium blockchain). Currency vs Platform. Bitcoin and other cryptocurrencies store coin transactions in blocks as data. However, blockchain is oblivious to the type of stored data, which can be multimedia files, weblogs, digitized books, and any other data that we can fingerprint by using a hash function. A hash function takes an unlimited length of data and creates a fixed length (e.g., 256 bits in SHA-256) representation using mathematical operations. In 1997, Nick Szabo, an American computer scientist, had envisioned embedding what he called smart contracts as “contractual clauses in the hardware and software to make a breach of contract expensive” [4]. Almost 20 years later, Blockchain enthusiasts saw a path to implement this vision. Blockchain platforms Neo (2014), Nem (2015), Ethereum (2015), and Waves (2016) have stored and run software code, called smart contracts, on a blockchain. Smart contracts are written in Turing-complete language. However, the block gas limit restricts what transactions can include in the code. For example, on Ethereum, a smart contract cannot be arbitrarily long to encode an algorithm. As a result, blockchain opponents argue that the Ethereum gas limit prevents Turing-complete smart contracts. The benefits of storing code on a blockchain are multi-fold. Every blockchain network node can read smart contracts, and transactions can execute contracts by sending call messages to them, and all blockchain nodes will run the contract with the message. These aspects mean that smart contracts allow unstoppable, unmodifiable, and publicly verifiable code execution as transactions between blockchain participants. Any blockchain classification can be flexed to allow public-private functionalities. For example, Hyperledger Fabric uses smart contracts in permissioned settings. The Ripple credit network is a permissioned blockchain that restricts block mining but stores data publicly. First vs. Second Layer Technologies. Over time, blockchains started to run into scalability issues due to limited block sizes and increased user participation. Blockchains developed initial solutions, such as SegregatedWitness,111https://en.bitcoin.it/wiki/Segregated_Witness to leave some of the encryption signatures and other non-transactional data out of blocks to make space for more transactions. Scalability efforts have culminated in second layer solutions, such as the Lightning Network [5], where the network executes most of the transactions off-the-blockchain. The first layer (i.e., the blockchain itself) only stores a summary of transactions that occur on the second layer. Second layer networks, such as the Lightning Network, are a hot research area in network analysis that can be the topic of an extensive review article on its own. Due to space limitations, this manuscript will only teach blockchain networks that we can extract from the first layer, i.e., the blockchain itself. Blockchains have come to contain various data such as sensor messages on IOTA, software code on Ethereum, and international banking transactions on Ripple. We give an outline by considering two major blockchain types: cryptocurrencies and platforms. Cryptocurrencies are designed to store financial transactions between addresses. A Blockchain platform stores financial transactions as well but furthermore stores software code in smart contracts. Blockchains can further be categorized into two broad categories in terms of transaction type: account-based (e.g., Ethereum) and unspent transaction output (UTXO)-based (e.g., Bitcoin, Litecoin) blockchains. Traditionally cryptocurrencies are UTXO-based, whereas platforms are account-based. The difference between UTXO and account-based models has a profound impact on blockchain networks. In the first type of account-based blockchains, an account (i.e., address) can spend a fraction of its coins and keep the remaining balance. An analogy to account-based blockchains is a bank account that makes payments and keeps the remaining balance in the account. A transaction has exactly one input and one output address (an address is a unique identifier on the blockchain transaction network). We may use an address to receive and send coins multiple times. The resulting network is similar to traditional social networks, which implies that we can apply social network analysis tools directly to account networks. In turn, the second type (i.e., UTXO-based blockchains), such as Bitcoin, are the earliest and most valuable blockchains. Bitcoin constitutes around 45-60% of total cryptocurrency market capitalization. Litecoin has approximately 2% capitalization. On UTXO based blockchains, nodes (i.e., addresses) are mostly one-time-use only, complicating network analysis. See Section 8 for issues that complicate network analyses. ###### Remark 2.1. A blockchain platform uses smart contracts to implement complex transactions, resulting in a diverse set of networks. Traditionally, blockchain platforms have followed an account-based blockchain model. For this reason, we use the terms account blockchain and blockchain platform interchangeably. However, in theory, a platform can also employ a UTXO transaction model. Similarly, a cryptocurrency could use an account-based transaction model. The reader must discern the distinction between account and UTXO based blockchain networks. Table 1: Blockchain types. The private/public columns indicate block mining permissions. IOTA and Ripple are maintained by consortiums that own the mining rights. Many projects (such as Ripple and IOTA) developed smart contract functionality many years after their launch. Although IOTA uses a directed acyclic graph instead of a canonical blockchain, the IOTA transaction model is UTXO based. Privacy coins Monero and Zcash use UTXO models with cryptographic security. | UTXO | Account | DAG | Platform | Cryptocurrency | Private | Public ---|---|---|---|---|---|---|--- Bitcoin | | | | | | | ZCash | | | | | | | Monero | | | | | | | Ripple | | | | | | | Ethereum | | | | | | | IOTA | | | | | | | Table 1 shows the focus of this manuscript and lists the blockchains that we will cover. Specifically, we teach two projects that differ from cryptocurrencies and platforms in critical aspects: Ripple and IOTA. Ripple is a credit network that predates blockchain fame. Ripple keeps a ledger of transactions where participants can trade user-issued currencies along with the native cryptocurrency of the Ripple, which is called XRP. The Ripple network moves currencies across the world, and financial institutions mainly use it. Recently, Ripple has also implemented smart contracts (called Hooks). We include Ripple in this manuscript since Ripple uses key blockchain technologies, such as hash-based identities and smart contracts, although not a blockchain by definition. IOTA Tangle is a directed acyclic graph of transactions. However, the underlying transaction model is UTXO based. IOTA relegates mining to each transaction’s creator and does not incur a transaction fee. Although IOTA has its associated cryptocurrency, mainly Internet-of-Things devices use IOTA to share and store data in industrial applications. ## 3 Bitcoin, Monero and ZCash: UTXO Networks A cryptocurrency transaction is a construct that consumes one or more outputs of previous transactions and creates one or more outputs. As Bitcoin is a good representative of cryptocurrencies, we will use its data to teach this section. Note that Litecoin data is in the same format as Bitcoin, and in the past, we could have parsed them by using the same software library (e.g., Bitcoin4J). We will conclude the section with Monero and Zcash networks. Figure 2: A coinbase transaction $t_{1}$ with two outputs. Note that $t_{1}$ has no input address. Compare this transaction with the ordinary spending transaction of Figure 3. Bitcoin stores sequential block data in blk*.dat files on the disk. For example, block1 is serialized and appended into the blk00000.dat file. Next, Bitcoin appends a magic byte to separate block 1 from the upcoming block 2 in the dat file. For example, the Bitcoin mainnet uses 0xD9B4BEF9 as the magic byte. As the number and size of transactions in a block determine the block size, each dat file may contain a different number of blocks. Each block contains one coinbase transaction and zero or more spending transactions. A coinbase transaction is the first transaction of the block and contains the block reward plus a sum of fees from the block’s transactions, if any. Note that a transaction may leave nothing as a transaction fee but still get mined in the block. The Bitcoin protocol sets the mining reward. If a miner records the block reward higher than the set amount, the network will reject the block. However, if the miner sets the block reward less than the set amount, the block will be accepted, and the miner will have lost bitcoins. In reality, these bitcoins are lost to the protocol, and the total bitcoin supply decreases from 21 million bitcoins. Figure 3: A spending transaction that consumes a previous output and creates two new outputs. A coinbase transaction has no input but one or more outputs (Figure 2). In a sense, in a coinbase transaction, the miner can write itself a check in the amount of mining reward plus transaction fees. A coinbase transaction increases the total bitcoin supply by the mining reward amount. This is where new coins are created in Bitcoin. For this reason, a coinbase transaction lists no input. All other transactions must list one or more inputs that are being spent. Usually, a coinbase transaction has a single output address, and the address belongs to the miner. Some mining pools pay their miners in coinbase transactions. Such transactions may have hundreds of output addresses where each address belongs to an individual pool member. In Figure 2 the coinbase transaction has two outputs that total 1.270B satoshis, out of which 1.25B (12.5 bitcoins) are the mining reward, whereas 20M are the sum of transaction fees. Satoshi is the subunit of bitcoin (as Cent is the subunit of US dollar), and one bitcoin contains 100 million satoshis. As Figure 2 shows, an output has two useful information: amount and address. An address can next use the received output of $o_{1}$ in a spending transaction. Figure 3 shows the associated spending transaction that consumes $o_{1}$ and creates two new outputs. A look at Figure 3 reveals an interesting point about how Bitcoin records transactions. Transaction $t_{2}$ gives the consumed output with its previous transaction id (the hash of $t_{1}$, which created this output) and the output index. From this information, we can deduce that we are spending the zeroth output of $t_{1}$, but this does not tell us the amount or address ($a_{1}$). By only linking the previous transaction $t_{1}$, we can learn the corresponding amount and address. From Figure 2 we see that the zeroth output of $t_{1}$ holds 1B satoshis. We will denote the satoshi amount in an output with $A()$, i.e., $A(o_{1})=1B$. Next, in Figure 3 500M and 495M satoshis are sent to $a_{3}$ and $a_{4}$. The difference between $A(o_{1})-[A(o_{3})+A(o_{4})]=5M$ satoshis are implicitly left (by the transaction creator) as the transaction fee. The miner’s job is to validate that output amounts are greater than or equal to the input amounts. Furthermore, the miner must observe all previous blocks and ensure that another transaction has not already spent $o_{1}$. If the miner makes a mistake and includes an invalid (e.g., already spent output, too great an output amount) transaction, the network will reject the miner’s block. In practice, starting from an output, it is possible to go back in time and trace the lineage of a coin to reach a coinbase transaction. For example, the output linked graph in Figure 4 shows the third transaction $t_{3}$, which has two inputs. The output $o_{6}$ which has two lineages: $o_{1}\rightarrow o_{4}\rightarrow o_{6}$ and $o_{2}\rightarrow o_{6}$. However, the lineage can be obscured by mixing schemes [6]. Figure 4: Bitcoin transaction network where we link transaction inputs to previous transaction outputs. As well as multiple outputs, a Bitcoin transaction allows multiple inputs. The owner of each input must sign its portion of the transaction to prove an authorized coin transfer. A many-to-many transaction is an interesting data type that we do not encounter in banking and finance; multiple addresses merge their coins and send them to multiple receiving addresses. For the output, a Bitcoin address can be created for free and shared with the sender. Output addresses of a transaction need not belong to the same real-life entity: $a_{5}$ and $a_{6}$ may belong to people who have never met in real life. The same reasoning is less likely to apply to input addresses. There can be multiple scenarios considering the inputs. * • The user behind $a_{2}$ signs the input of $t_{3}$ and sends the partially signed transaction to the user behind $a_{4}$, who signs its input and finally sends the transaction to the mempool. This scenario implies that the users are different people but know and communicate with each other. * • The user behind $a_{2}$ signs the input of $t_{3}$ and sends the partially signed transaction to a public forum where other users can add their inputs, and the transaction is finally sent to the mempool by one of them. This scenario implies that the users are different people, do not know but communicate with each other. * • A single user owns both $a_{2}$ and $a_{4}$. The user signs both inputs and sends the transaction to the mempool. The tree cases mean that we can never be sure about the joint ownership of inputs. However, in practice, most of the time, all input addresses belong to the same user. We can use this information to link addresses across transactions. ###### Remark 3.1. When a transaction sends coins to an address, the address appears in the transaction, but the public key of the receiver address is unknown. The receiver discloses their public key only when spending the received coins, and any user in the network can hash the key to verify that the hash equals the address. The matching hash is additional proof that the assets belong to the receiver. In a P2PSH (pay to script hash) address (which starts with ’3’), the owner must show the script whose hash equals the address. ### 3.1 Graph Rules for UTXO Blockchains Before modeling the Bitcoin network, we emphasize three graph rules that restrict how transactions can transfer coins. These rules are due to Bitcoin design choices made by Satoshi Nakamoto. For clarity, we will replace outputs with the associated addresses and work with the toy graph shown in Figure 5. Figure 5: A blockchain network with six transactions and 12 addresses. Addresses $a_{1},a_{2},a_{3},a_{4}$ receive coins from earlier transactions that are not shown here. Transaction $t_{5}$ uses a previously used address $a_{1}$ in edge $e_{1}$. Edge $e_{2}$ denotes an address reuse where the change is sent back to the sender address $a_{10}$. Dashed edge $e_{3}$ cannot be created with the given transactions, because $a_{5}$ receives only a single output in $t_{1}$. In this network, outputs of $t_{5}$ (to addresses $a_{1}$ and $a_{11}$) and $t_{6}$ (to addresses $a_{12}$ and $a_{10}$) remain unspent. 1. 1. Balance Rule: Bitcoin transactions consume and create outputs. Outputs are indivisible units; we can spend each output in one transaction only. Consequently, we must spend coins received from one output in a single transaction. Any amount that we do not send to an output address is considered a transaction fee, and the miner collects the change as the transaction fee. To spend a portion of the coins in output and keep the change, we must create a new address and send the remaining balance to this new address. Another option is to use our existing address as one of the output addresses and re- direct the balance (as done in edge $e_{2}$ in Figure 5). As a community practice, reuse of the spender’s address (i.e., address reuse) is discouraged. As a result, most address nodes appear in the graph two times: first when they receive coins and second when they spend them. 2. 2. Source Rule: We can merge input coins from multiple transactions and spend them in a single transaction (e.g., the address $a_{10}$ receives coins from $t_{3}$ and $t_{4}$ to spent in $t_{5}$ and $t_{6}$ in Fig. 5). However, $a_{10}$ could have spent all its coins in a single transaction as well. 3. 3. Mapping Rule: In a transaction the input-output address mappings are not explicitly recorded. For instance, consider the transaction $t_{2}$ in Fig. 5. The output to address $a_{6}$ may come from either $a_{3}$ or $a_{4}$. We can make an analogy with lakes where in-flowing rivers (inputs) bring water (coins) to the lake (transaction), and outgoing emissaries (outputs) take the water (coin) out. Besides the three rules, in our research, we have encountered the following Bitcoin spending practices: * • A transaction may list the same address in multiple outputs. We have encountered transactions where all outputs have the same address. As the address must be recorded more than once, this (spam) practice unnecessarily increases the transaction size. * • In time, an address may receive outputs from multiple transactions. Typically, an address collects all outputs and spends them at once. In rare cases, an address can receive outputs again after it has spent its coins. For example, edge $e_{1}$ in Figure 5 brings an output to $a_{1}$ after $a_{1}$ has spent its coins in $t_{1}$. * • Address reuse, as shown in $e_{2}$, is widespread. Although the community frowns upon address reuse, users are still reluctant to create a new address for change. Hierarchically deterministic wallets facilitate automated address creation (https://en.bitcoin.it/wiki/Deterministic_wallet). However, address reuse did not abate. * • Community practice is to wait for six confirmations (blocks) after receiving a transaction output before spending it. The practice protects bitcoin buyers against history reversion attacks. Thereby, we must spend coins of an output in a minimum of six blocks after they are received ($6\times 10=60$ minutes later). A day sees 144 blocks; a coin should move at most 24 times (in 24 blocks). In reality, however, once a transaction reaches the mempool, it can be considered final, and absent a double-spending, its coins can be used in another transaction immediately, even in the same block. * • It is possible to have $n$ ordered transactions in a single block, where $1<x<y<n$ and $t_{y}$ consumes the outputs generated by $t_{x}$. As a result, some coins move in the Blockchain network more than 144 times a day. ###### Remark 3.2. In theory, Monero, Dash, and ZCash privacy coins are UTXO blockchains; their Blockchain networks are similar to Bitcoin’s. In practice, privacy coins employ cryptographic techniques to hide node and edge attributes in the blockchain network. For example, ZCash hides information in its shielded pool, whereas Monero adds decoy UTXOs to the input UTXO set. This section explains how to model the blockchain network when the node and edge attributes are public (as in Bitcoin). Data scientists model Bitcoin transaction networks as address or transaction graphs. Address graph omits transactions, and the transaction graph omits addresses. Both approaches are influenced by traditional social network analysis, which employs graphs with one node type only. Additionally, one can extract chainlet substructures from the blockchain network. In the following sections, we will cover these graph models. ### 3.2 UTXO Transaction graph Transaction graphs omit address nodes from the transaction network and create edges among transactions only. Figure 6 shows the transaction graph of the network shown in Figure 5. The most important aspect of the transaction graph is that a node can appear only once. There will be no future edges that reuse a transaction node. The transaction graph contains far fewer nodes than the network it models. We can immediately observe a few drawbacks from Figure 6. By omitting addresses, we lose the information that $t_{5}$ and $t_{1}$ are connected by $a_{1}$. The address reuse of $a_{10}$ is hidden in the transaction graph as well. Additionally, unspent transaction outputs are not visible; we cannot know how many outputs are there in $t_{5}$ and $t_{6}$. Similarly, if $t_{3}$ had an unspent output, we would not learn this information from the graph. In Bitcoin, many outputs stay unspent for years; the transaction graph will ignore all of them. Figure 6: Transaction graph representation of the Blockchain network in Figure 5. The advantages of the transaction graph are multiple. First, we may be more interested in analyzing transactions than addresses. For example, anti-money laundering tools aim at detecting mixing transactions, and once they are found, we can analyze the involved addresses next. Many chain analysis companies focus their efforts on identifying e-crime transactions. Second, the graph order (node count) and size (edge count) are smaller, which is better for large-scale network analysis. The reduction is because, on UTXO networks, transaction nodes are typically less than half the number of address nodes. For example, Bitcoin contains 400K-800K unique daily addresses but 200K-400K transactions only. As we will explain in the next section, the address graph contains many more edges than the transaction graph. ### 3.3 UTXO Address Graph The address graph is the most commonly used graph model for UTXO networks. The address graph omits transactions and creates edges between addresses only. Address nodes may appear multiple times, which implies that addresses may create new transactions or receive coins from new transactions in the future. Address graphs are larger than transaction graphs in node and edge counts. As the mapping rule states, a UTXO transaction does not explicitly create an edge between input and output addresses in the blockchain transaction network. When omitting the intermediate transaction node, we cannot know how to connect input-output address pairs. As a result, data scientists create an edge between every pair. For example, in Figure 7 we create six edges between inputs ($a_{3},~{}a_{4}$) and outputs ($a_{6},~{}a_{7},~{}a_{8}$). If there are few addresses in the transaction, this may not be a big problem. However, large transactions can easily end up creating millions of edges. For example, the highest number of inputs in a Bitcoin transaction was 20000 (821 on Litecoin), whereas the highest number of outputs in Bitcoin was 13107 (5094 on Litecoin). The address graph approach will have to create one million edges for a transaction with one thousand inputs and one thousand outputs. Graph size is not the only problem. The address graph loses the association of input or output addresses. For example, the address graph in Figure 7 loses the information that edges $a_{3}$ and $a_{4}$ were used in a single transaction; address graph edges would be identical if the addresses had used two separate transactions to transfer coins to $a_{6}~{}a_{7}$ and $a_{8}$. We can solve this issue by adding an attribute (e.g., transaction id) to the edge; however, this requires additional edge features. Address graph edge weighting is done as follows. Consider a transaction $t$ with its input addresses $I$ and output addresses $O$. We will denote the amount an address $a_{x}$ sends to or receives from $t$ as $w_{x}$. Amount of coins in an address is denoted as $A(a_{x})$. The edge between an input address $a_{i}\in I$ and an output address $a_{j}\in O$ is assigned an edge weight as $w_{i\rightarrow j}=A(a_{i})\times\dfrac{A(a_{j})}{\sum_{a_{x}\in O}{A(a_{x})}}.$ ###### Example 3.3. Consider a transaction $t$ with input addresses $I=\\{a_{1},a_{2}\\}$ and amounts as $A(a_{1})=1$, $A(a_{2})=3$ bitcoin. The transaction has two output addresses $O=\\{a_{3},a_{4}\\}$ and their associated amounts as $A(a_{3})=0.9$, $A(a_{4})=2$ bitcoin. The transaction fee is left as the difference between inputs and outputs: $1+3-(0.9+2)$ bitcoin. When creating the address graph, $w_{2\rightarrow 3}$ is weighted as $3\times(0.9/2.9)$, whereas $w_{2\rightarrow 4}=3\times(2/2.9)$. Compared to the transaction graph, the address graph loses less information from the Blockchain network. For example, the address graph does not fail to record past address reuse (edge from $a_{9}$ to $a_{1}$) and change address reuse (as self-loop at $a_{10}$). Furthermore, unspent output addresses remain visible in the graph. ###### Remark 3.4. Address reuse: A blockchain address can be used in multiple transactions as a coin sender or receiver. However, address reuse is discouraged. Ideally, a Bitcoin user should create a new address each time it receives bitcoins in a transaction. Address clustering methods aim at linking multiple addresses to identify real-life entities behind the addresses. Figure 7: Address graph representation of the Blockchain network in Figure 5. Once we create the address graph, we may be inclined to run traditional network science tools and algorithms on it. However, some of these methods are ineffective in blockchain networks. We can count address clustering, motif analysis, and core decomposition among these methods. UTXO networks are sparse and devoid of closed triangles. Furthermore, address reuse is discouraged on UTXO blockchains as a community practice; most addresses appear in two transactions (i.e., receiving and spending coins). As a result, off-the-shelf Data Science algorithms and software libraries are relatively inefficient on address graphs. For example, network motif analysis [7] aims at finding repeating subgraphs of specific orders (usually three nodes). Searching for motifs in Figure 7 will try to find shapes without considering how specific addresses appear together as inputs or outputs. Addresses $a_{3}~{}a_{6}$, and $a_{4}$ will most likely never form a triangle. Also, a triangle between $a_{6},~{}a_{7}$, and $a_{8}$ is very unlikely. As a second example, motifs do not consider which addresses can be active. For instance, once $a_{3}$ spends its output in a transaction without receiving a new output, it will never create another edge in the future. Keeping such nodes in memory and searching for their future edges are unnecessary. Past address reuse is discouraged, but motifs do not use this information and search for (non-existent) edges to past addresses in a large address graph. As a third example, consider that many coins are moved in the network more than 100 times a day, and each time they are received in newly-created addresses. The resulting graph will be huge, but there will be almost no closed edge triangles. Due to these issues, we argue that we should model UTXO networks as a forward branching tree rather than networks. More importantly, Data Scientists should develop tools to incorporate domain practices such as aversion to address- reuse. ### 3.4 Monero and Zcash Transaction Networks Monero and Zcash are two UTXO based cryptocurrencies that hide transaction details from the public. For this reason, Monero and Zcash are called privacy coins. #### 3.4.1 Monero Networks Figure 8: Monero transaction network. We show the actual transaction addresses with red nodes and decoy UTXO addresses with blue nodes. Each ring surrounds nodes in light blue. Dashes indicate the remaining UTXOs that are linked to previous transaction outputs. Note that each input ring has exactly ten decoys and one actual input. However, outputs are listed individually and not in a ring. Address $a_{1}$ is truly spent in $t_{1}$, but listed as a ring member in $t_{5}$ as well. By only looking at transaction inputs, one cannot learn which transaction spends the coins at $a_{1}$. Since the RingCT update, Monero has hidden the UTXO amount and address as well. The reader should note how difficult, if not impossible, it would be to link addresses to transactions in such a graph. Monero (created in 2014) uses ring signatures to hide inputs of a transaction. In a ring signature scheme [8], any one of the $m$ members of the ring can sign a document. Any third party can publicly verify the signature but cannot deduce which ring participant signed it. Monero treats each transaction input as a document that ring members will sign. The process to create a transaction is carried out as follows: 1. 1. Choose one or more of your UTXOs that you want to spend in a transaction. 2. 2. For each UTXO that you plan to spend, choose 10 foreign UTXOs (i.e., ring member UTXOs that do not necessarily belong to you) that you will use as decoys. 3. 3. Prepare the ring signature by using the private key of your address. 4. 4. After signing each input, forward the transaction to the Peer-to-Peer network. The reader should note that decoy UTXOs can belong to any other user, and the transaction creator does not need the consent of these users to include their UTXOs as decoys in a ring. Choosing the best decoy UTXOs to hide the identity of the actual input UTXO is an active research area. Monero initially allowed using 0-decoys, i.e., not adding any decoys to the ring. In this case, a Monero transaction was equivalent to a Bitcoin transaction where all information is public. Monero also allowed a user to set the ring size, which could be very big. Privacy research has shown the drawbacks of this liberal approach [9, 10]. First, 0-decoy transactions jeopardize the privacy of other blockchain users [10] since a 0-decoy transaction input reveals that no other transaction could have spent it. As a result, we can remove the input from the rings of future transactions. Second, we can use a meta-analysis on similar ring sizes and link transactions of the same user. Monero banned 0-decoy transactions in 2016, incremented the mandatory ring size from 5 to 7 and eventually to 11. With the RingCT update, Monero has also hidden UTXO addresses and amounts. As a result, very few pieces of transactions are visible to the public, making it an ideal privacy coin. Figure 8 shows a Monero address-transaction graph. Note that although inputs list 11 member rings, outputs are listed individually. However, the receiving address and received amount are hidden in the UTXO. Monero uses commitments and range-proofs to make sure that amounts are spent correctly (See Chapter 4 in [11]). #### 3.4.2 Zcash Networks Zcash uses zero-knowledge proofs to hide information about some transactions. In cryptography, a zero-knowledge proof is a method by “which one party can prove to another party that they know a value $x$, without conveying any information apart from the fact that they know the value $x$”. 222https://z.cash/technology/ We can broadly categorize Zcash transactions into shielded and public transactions. Public transactions are identical to Bitcoin transactions where amounts, addresses, and all other pieces of information are public. Shielded transactions hide all transaction information from the public by using zero- knowledge proofs. Addresses that are used in public transactions are called t-addresses and start with the letter “t” such as t1K2UQ5VzGHGC1ZPqGJXXSocxtjo5s6peSJ. Shielded transactions use shielded (also called private) addresses that start with the letter “z”. Figure 9: A Zcash address-transaction graph that contains six transactions. Amounts on edges are coins (we assume zero transaction fees). Although we show the shielded pool nodes and edges here, in reality, the pool is not visible to the public at all. Transactions which involve $z$-addresses are computationally costly to create and validate. As a result, only around 10% of Zcash transactions are shielded [12]. With $t$ and $z$ addresses, there are five Zcash transactions types. We will show these types by using Figure 9, and explain them as follows: 1. 1. $t$-to-$t$ (public) transactions: all transactions details are public ($t_{1}$ in Figure 9). 2. 2. $t$-to-$z$ (shielding) transactions: $t$ address and its sending amount are public, $z$ address is encrypted ($t_{2}$ in Figure 9). 3. 3. $z$-to-$t$ (de-shielding) transactions: $t$ address and amount it receives are public, $z$ address is encrypted ($t_{6}$ in Figure 9). 4. 4. $z$-to-$z$ (private) transactions: the addresses, transaction amount, and the memo field are all encrypted and not publicly visible, i.e., $t$ (without an index) in the shielded pool of Figure 9. 5. 5. $tz$-to-$tz$ (mixed) transactions: $z$-addresses are involved, but there are public inputs or outputs in the transaction ($t_{3}$ and $t_{5}$ in Figure 9). A few design choices have made an impact on Zcash transactions. First, the Zcash protocol includes a consensus rule that coinbase rewards must be sent to a shielded address, and typically these coins are forwarded to $t$-addresses very soon. Second, Zcash initially used a zero-knowledge scheme that supported at most two hidden inputs and two hidden outputs. As a result, including more than two input or output addresses required more cryptographic work, which was costly. A newer scheme called sapling does not have this limitation. Both Monero and Zcash are much more costly than Bitcoin in terms of computational costs and resource usage. Monero transactions store ten decoys for each actual input, which considerably increases transaction size on disk, and Zcash transaction validation takes too much memory and time. As a result, online exchanges avoid storing balances of $z$-addresses. However, Monero and Zcash both use UTXO models that we can analyze by graph analysis tools developed for Bitcoin. ### 3.5 Chainlets for UTXO Transaction Networks Subgraph encoding on the Blockchain network is an alternative to address and transaction graph approaches. In traditional graphs, we consider nodes and edges the building blocks because nodes are created in time and may establish new edges at various time points. However, UTXo transactions create multiple edges at once. That is, we can think of a transaction itself as a building block of the network. With its nodes and edges, a transaction represents an immutable decision that is encoded as a substructure on the UTXO network. Rather than using individual edges or nodes, we can use this substructure as the building block in network analysis. If we consider multiple transactions and their connections through addresses, we may extend the substructure idea to blockchain subgraphs by considering multiple transactions. We use the term chainlet to refer to such subgraphs [13]. Consider a UTXO network with transaction and address nodes. The node set $V=\\{\text{address},\text{transaction}\\}$. An edge $e\in E$ connects an address $a_{x}$ to a transaction $t_{y}$ (i.e., $a_{x}\rightarrow t_{y}$) or a transaction $t_{y}$ to an address $a_{z}$ (i.e., $t_{y}\rightarrow a_{z}$). This implies that there are no edges between two nodes of the same type. A UTXO chainlet $\mathcal{G^{\prime}}=(V^{\prime},E^{\prime},B)$ is a subgraph of $\mathcal{G}$, if $V^{\prime}\subseteq V$ and $E^{\prime}\subseteq E$. If $\mathcal{G^{\prime}}=(V^{\prime},E^{\prime},B)$ is a subgraph of $\mathcal{G}$ and $E^{\prime}$ contains all edges $e_{u,v}\in E$ such that $(u,v)\in V^{\prime}$, then $G^{\prime}$ is called an induced subgraph of $G$. Figure 10: Examples of the first order Merge ($\mathbb{C}_{2\rightarrow 1}$), Transition ($\mathbb{C}_{2\rightarrow 2}$) and Split ($\mathbb{C}_{2\rightarrow 3}$) $k=1$ chainlets from the UTXO network of Figure 5. An address can appear in multiple chainlets, each time spending or receiving a different UTXO. Transaction nodes can appear in a single chainlet only. Let $k$-chainlet $\mathcal{G}_{k}=(V_{k},E_{k},B)$ be a subgraph of $\mathcal{G}$ with $k$ nodes of type “transaction”. If there exists an isomorphism between $\mathcal{G}_{k}$ and $\mathcal{G}^{\prime}$, $\mathcal{G}^{\prime}\in\mathcal{G}$, we say that there exists an occurrence of $\mathcal{G}_{k}$ in $\mathcal{G}$. A $\mathcal{G}_{k}$ is called a blockchain $k$-chainlet. As a starting point, we can focus on the first order chainlet ($k=1$), which consists of a single transaction node and the address nodes. The first order chainlet is, by definition, a substructure of the network. We denote a chainlet of $x$ inputs and $y$ outputs with $\mathbb{C}_{x\rightarrow y}$. A natural classification of first-order chainlets can be made regarding the number of inputs $x$ and outputs $y$ since there is only one transaction involved. We can quickly identify three main types of first-order chainlets. If the transaction merges input UTXOs, it will have a higher number of inputs than outputs. We call these merge chainlets, i.e., $\mathbb{C}_{x\rightarrow y}$ such that $x>y$, which show an aggregation of coins into fewer addresses. Two other classes of chainlets are transition and split chainlets with $x=y$ and $x<y$, respectively, as shown in Figure 10. We refer to these three chainlet types as the aggregate chainlets. Figure 11 visualizes the percentage of aggregate Bitcoin chainlets in time. For example, the transition chainlets are those $\mathbb{C}_{x\rightarrow x}$ for $x\geq 1$. Figure 11 shows that the Bitcoin network, starting as an unknown project, stabilized only after summer 2011. From 2014 and on-wards, the split chainlets continued to rise steadily, compared to merge and transition chainlets. Spam attacks on the Bitcoin blockchain, which created too many transactions on the mempool to slow down transaction processing and force a block size increase, were visible around late 2015. Figure 11: Percentage of daily aggregate chainlets in Bitcoin. Splits constitute around 75% of all transactions. Extending this discussion, higher-order chainlets, as shown in Figure 12, can be classified in terms of their shapes. Figure 12: Two second order ($k=2$) chainlets from the Blockchain network of Figure 5. ### 3.6 Occurrence and Amount Information in Chainlets Chainlets provide two lenses to look at the transaction network; we can consider amounts or counts of chainlets. For a given time granularity, such as one day, we can take snapshots of the blockchain network and extract chainlets from it. From the Blockchain network snapshot for a given granularity (e.g., daily), we mine two information for each chainlet type: amount to store volume of coin transfers by using the chainlet and occurrence to store instances (i.e., counts) of the chainlet. Figure 13: A UTXO transaction network with amounts. Transaction fees range from 0.02 to 0.1 coins. For example, Figure 13 contains six transactions and 12 addresses. However, if we look at the first order chainlets, we find five unique chainlets: Transaction $t_{1}$ creates a split chainlet $\mathbb{C}_{2\rightarrow 1}$. Transaction $t_{2}$ creates split chainlet $\mathbb{C}_{2\rightarrow 3}$. Transactions $t_{3}$ and $t_{6}$ create split chainlets $\mathbb{C}_{1\rightarrow 2}$. Transaction $t_{4}$ creates merge chainlet $\mathbb{C}_{3\rightarrow 1}$. Transaction $t_{5}$ creates transition chainlet $\mathbb{C}_{2\rightarrow 2}$. From these chainlets, we can create two $3\times 3$ matrices to hold occurrence and amount information as $O=\begin{bmatrix}0&2&0\\\ 1&1&1\\\ 1&0&0\end{bmatrix},~{}A=\begin{bmatrix}0&3.58&0\\\ 1.9&1.75&3\\\ 2.8&0&0\end{bmatrix}.$ where $\mathcal{O}[i,j]$ and $\mathcal{A}[i,j]$ store the number and transferred amount of chainlets of the type $\mathbb{C}_{i\rightarrow j}$, respectively, where $i\geq 1$ and $j\geq 1$. For example, transactions $t_{3}$ and $t_{6}$ are stored in $\mathcal{O}[{1,2}]=2$. The amounts output in these transactions are $\mathcal{A}[1,2]=1+0.8+1+1.78=3.58$. ###### Remark 3.5. A coinbase transaction has no input address but $\geq 1$ output addresses. If we plan to use the occurrence matrix $\mathcal{O}$ to store coinbase transactions, we need to extend the matrix with $i=0$ as well. Coinbase transactions would then be stored in $\mathcal{O}[{0,j}]$. For the transaction network in Figure 13, $i=j=3$ suffices to store the matrices since transactions have at most three input addresses and three output addresses. However, in a real blockchain matrix, dimensions can easily reach thousands. UTXO blockchains restrict the number of input and output addresses in a transaction by limiting the block size (1MB in Bitcoin), but the number of inputs and outputs can still reach thousands. As a result, we can have large chainlets (e.g., $\mathbb{C}_{1000\rightarrow 200}$, $\mathbb{C}_{901\rightarrow 200}$ or $\mathbb{C}_{100\rightarrow 951}$). Consider the case where we choose to create $1000\times 1000$ to store the occurrence and amount matrices with 1 million cells. Having so many cells is neither useful nor practical since most transaction types will not exist in the network. In turn, corresponding matrix cells in occurrence and amount would hold many zero values, implying that the matrices would be very sparse. As an alternative, we can choose a suitable value for matrix dimensions. For the optimal matrix dimension $N$, we have analyzed the history of Bitcoin and Litecoin. We have found that % 91.38 of Bitcoin and % 91.27 of Litecoin chainlets have $N$ of 5 (i.e., $\mathbb{C}_{x\rightarrow y}$ s.t., $x<5$ and $y<5$) in average for daily snapshots. This value reaches % 98.10 and % 96.14 for $N$ of 20, for the respective coins. We have selected $N$ of 20 since it can distinguish a sufficiently large number (i.e., 400) of chainlets and still offers a dense matrix. The choice of $N$ requires a strategy to deal with transactions whose dimensions are bigger than $N$. We record these chainlets in the last columns/rows of the matrices, which are defined as follows: ###### Definition 3.6. Amount Matrix. We denote the total amount of coins transferred by a chainlet $\mathbb{C}_{x\rightarrow y}$ in a graph snapshot as $\mathbb{A}(\mathbb{C}_{x\rightarrow y})$. Amount of coins transferred by chainlets in the graph snapshot are stored as an $N\times N$-matrix $\mathcal{A}$ such that for $i\leq N,j\leq N$ $\mathcal{A}[i,j]=\left\\{\begin{array}[]{l l}\mathbb{A}(\mathbb{C}_{i\rightarrow j})&\quad\text{if $i<N$ and $j<N$, }\\\ \sum\limits_{z=N}^{\infty}\mathbb{A}(\mathbb{C}_{i\rightarrow z})&\quad\text{if $i<N$ and $j=N$,}\\\ \sum\limits_{y=N}^{\infty}\mathbb{A}(\mathbb{C}_{y\rightarrow j})&\quad\text{if $i=N$ and $j<N$,}\\\ \sum\limits_{y=N}^{\infty}\sum\limits_{z=n}^{\infty}\mathbb{A}(\mathbb{C}_{y\rightarrow z})&\quad\text{if $i=N$ and $j=N$.}\\\ \end{array}\right.$ ###### Definition 3.7. Occurrence Matrix. We denote the total number of chainlet $\mathbb{C}_{x\rightarrow y}$ in a graph snapshot as $\mathbb{O}(\mathbb{C}_{x\rightarrow y})$. Chainlet counts obtained from the graph snapshot are stored as an $N\times N$-matrix $\mathcal{O}$ such that for $i\leq N,j\leq N$ $\mathcal{O}[i,j]=\left\\{\begin{array}[]{l l}\mathbb{O}(\mathbb{C}_{i\rightarrow j})&\quad\text{if $i<N$ and $j<N$, }\\\ \sum\limits_{z=N}^{\infty}\mathbb{O}(\mathbb{C}_{i\rightarrow z})&\quad\text{if $i<N$ and $j=N$,}\\\ \sum\limits_{y=N}^{\infty}\mathbb{O}(\mathbb{C}_{y\rightarrow j})&\quad\text{if $i=N$ and $j<N$,}\\\ \sum\limits_{y=N}^{\infty}\sum\limits_{z=N}^{\infty}\mathbb{O}(\mathbb{C}_{y\rightarrow z})&\quad\text{if $i=N$ and $j=N$.}\\\ \end{array}\right.$ ###### Example 3.8. Consider the following example matrices $O=\begin{bmatrix}0&2&1\\\ 1&1&1\\\ 1&0&3\end{bmatrix},~{}A=\begin{bmatrix}0&3.58&0.5\\\ 1.9&1.75&3\\\ 2.8&0&4\end{bmatrix}.$ If we define $N=2$ for this example, $\mathcal{O}[1,3]$,$\mathcal{O}[2,3]$, $\mathcal{O}[3,1]$, $\mathcal{O}[3,3]$ and $\mathcal{A}[1,3]$,$\mathcal{A}[2,3]$, $\mathcal{A}[3,1]$, $\mathcal{A}[3,3]$ values must be stored inside the $2\times 2$ matrices. The updated matrices are $O=\begin{bmatrix}0&2+1\\\ 1+1&1+1+3\\\ \end{bmatrix},~{}A=\begin{bmatrix}0&3.58+0.5\\\ 1.9+2.8&1.75+3+4\\\ \end{bmatrix}.$ Extreme Chainlets: In occurrence and amount matrices, choosing an $N$ value, such as $N=20$, means that a chainlet with more than 20 inputs/outputs (i.e., $\mathbb{C}_{x\rightarrow y}$ s.t., $x\geq 20$ or $y\geq 20$) is recorded in the $N$-th row or column. That is, we aggregate chainlets with large dimensions that would otherwise fall outside matrix dimensions. We use the term extreme chainlets to refer to these aggregated chainlets on the $N$-th row and column. Extreme chainlets are large transactions that may involve more than 10 thousand addresses and are also large in coin amounts. We see two behavior in extreme chainlets. In the first case, chainlets are of the type $\mathbb{C}_{i<20\rightarrow j>20}$; less than 20 input addresses (usually one or two) sell coins to more than 20 addresses. An example is the Bitcoin transaction a79b970c17d97557357ec0661a2b9de44724440e1c635e1b603381c53ece725d in 2018. These extreme chainlets are split-chainlets which may indicate selling behavior. In the second case, chainlets are of the type $\mathbb{C}_{i>20\rightarrow j<20}$. This chainlet type is useful in finding large coin buys. Usually, an online exchange collects coins of many individual sellers and creates an extreme chainlet that has few (usually one or two) output addresses. Figure 14: Percentage of Bitcoin chainlets for $N=6$. Dark colors indicate higher percentages. The first row shows the percentages of one input chainlets. For example, 8.45% of all Bitcoin transactions are 1-to-1 chainlets, whereas 57.04% are 1-to-2 chainlets. Extreme chainlets capture large coin movements. In our research, we have found that information from extreme chainlets shows utility in volatility, risk, and price prediction in cryptocurrencies [14, 15]. ## 4 Ethereum: Account Networks Account blockchains, such as Ethereum, do not use the UTXO data structure of Bitcoin. Unlike UTXO coin transactions involving as few as two or as many as thousands of addresses, coin transactions on account blockchains involve only two addresses: sender and receiver. Many design choices behind account blockchains originate from the most popular account-based blockchain Ethereum. In this section, we will teach account networks by using Ethereum as our data source. However, the reader will find similar, if not the same, network and graph models in other account blockchains. A major difference between UTXO and account blockchains is the type and variety of networks. UTXO networks are transaction and lightning networks, whereas, on account blockchains, we can observe the following networks: 1. 1. Coin transaction network. Similar to the UTXO transaction network, this network is created from the coin (ether) transfers between addresses. Network edges only carry the native currency (coin) of the blockchain. 2. 2. Token transaction networks. Asset trading networks that are created by internal smart contract transactions. 3. 3. Trace network. Interactions between all address types. The name trace implies that a transaction triggers a cascade of calls to smart contracts or externally owned addresses. In this section, we discuss the three networks and their building blocks separately. We begin by first outlining network node and edge types. It is useful to re-consider the Ethereum address types. In account networks, we classify addresses into two node types and one special address. 1. 1. Externally owned address (EOA) has a private key. An EOA is managed by a real- life entity such as an investor or a centralized exchange. An entity may create and use multiple EOAs at the same time. Typically, an EOA is used over a long time in many transactions. 2. 2. Smart contract address does not have a private key. We can only distinguish Smart contract addresses by searching for smart contract code at the address. 3. 3. Null address 0x000.. node that has two use cases. First, a smart contract is created by using the NULL address as the receiver. Second, the NULL address is used to dispose of crypto assets. Any coin or asset sent to the address cannot be reclaimed and considered burned. The address may have a private key, but finding it by trial is considered impossible. Figure 15 visualizes the three-node symbols that we will use to teach account networks. Figure 15: Three address types and corresponding node shapes in account networks. ### 4.1 Account Transaction Network The account transaction network contains coin transfers between the three node types. The network edges always start from an EOA because a contract address or a NULL address cannot initiate a transaction. However, a smart contract can send coins to an EOA when a trace initiates the transfer (we will cover traces shortly). Specifically, transaction network edges are from i) EOA to EOA, ii) EOA to contract, iii) EOA to NULL address. iv) contract to EOA. Network edges may have i) coin amount, ii) account nonce, iii) gas price, and iv) timestamp features. The coin amount in an edge is in subunits (Wei). Account nonce is a number that orders transactions initiated by an EOA. Miners must mine transactions of an address in nonce order. For example, in Table 2 $a_{1}$ creates four transactions with nonce values 0 to 3. A future transaction of $a_{1}$ with nonce 5 has to wait because the transaction of nonce 4 has not been mined. Nonce order ensures that the network cannot have out-of-order or missing edges. However, miners can mine multiple transactions from an EOA in the same block according to the nonce order. Figure 16: A toy network of five nodes and edges with ether amounts (in wei) from four blocks. Edge features are given in Table 2. The address $a_{1}$ owns 300 weis from an earlier transaction (not shown). Transaction fee is one wei. Although we mention transaction time as an edge feature, transactions do not have timestamps themselves. The timestamp comes from the timestamp of the block that contains the transaction. Transactions of a block are ordered in the block by the miner, and online explorers record the order as the block index. However, even the miner cannot truly know when a transaction was created. Another edge feature could be the used gas field of the transaction. However, the feature is not useful for the transaction network because the gas cost equals the base fee (currently 21000 gas) for coin transfers. The input data field of the transaction is empty for coin transfer transactions. Figure 16 shows an example account transaction network. Table 2 gives the edge table of this network. We easily model the account transaction network as a directed and weighted multigraph (i.e., multiple edges exist between node pairs). Although address creation is cheap, address reuse is not discouraged on Ethereum, and a node may appear in multiple blocks and send and receive coins for an extended period. Furthermore, smart contracts have permanent addresses. These factors help us track node behavior and network dynamics in time. block height | from | to | amount (wei) | nonce | block index | timestamp ---|---|---|---|---|---|--- 10646423 | $a_{1}$ | $a_{2}$ | 100 | 0 | 1 | Aug-12-2020 05:11:17 PM +UTC 10646423 | $a_{1}$ | $a_{2}$ | 200 | 1 | 2 | Aug-12-2020 05:11:17 PM +UTC 10646424 | $a_{2}$ | $a_{3}$ | 249 | 0 | 1 | Aug-12-2020 05:11:18 PM +UTC 10646424 | $a_{2}$ | $a_{1}$ | 49 | 1 | 2 | Aug-12-2020 05:11:18 PM +UTC 10646425 | $a_{1}$ | NULL | 24 | 2 | 1 | Aug-12-2020 05:11:26 PM +UTC 10646426 | $a_{1}$ | $a_{4}$ | 24 | 3 | 1 | Aug-12-2020 05:11:43 PM +UTC Table 2: Network edges and their features of some transaction from blocks 10646423 to 10646426. ### 4.2 Token Transaction Networks A token is created by deploying a smart contract where the token’s features and business logic are defined. Any blockchain participant can create a token and facilitate its trade. The token’s contract defines meta attributes about the token, such as its symbol, total token supply, or decimals. Only the address of a token is unique in the blockchain; multiple tokens can have the same symbol, which creates confusion in trades. Currently, there are more than a hundred thousand tokens on Ethereum [16]. We consider each token to have its network and set of traders. A token’s current supply is the number of token instances created by the smart contract. If the instances are given unique ids, the token is considered non- fungible (as in the ERC 721 standard); each instance has its characteristics, owner, and price. Otherwise, one token instance will be equal to any other; the token is considered to be fungible (as in the ERC 20 standard). In this section, we will consider networks of fungible tokens, but we can easily extend our analysis to the non-fungible case. A token transaction network has EOA, NULL, and smart contract addresses as nodes. We outline the following three types of transactions that a Data Scientist must know to analyze token networks. * • The creation transaction that assigns an address for the token initializes its smart contract and state variables. * • A trade transaction that moves some tokens between addresses. * • A management transaction that can only be initiated by the smart contract creator (or any address that the owner specifies). The transaction may delete the contract or forward its balance (in ether or token) to another address. We will study management transactions in Section 4.3. token | block height | from | to | amount (Gwei) | tx fee (Gwei) | input data ---|---|---|---|---|---|--- Binance: BNB Token | 3978343 | 0x00c5…454 | 0xb8c7…d52 | 0 | 32643080 | […..] ChainLink: LINK Token | 4281611 | 0xf550…780 | 0x5149…6ca | 0 | 21895728 | […..] Tether USD | 4634748 | 0x3692…d57 | 0xdac1…ec7 | 0 | 12683176 | […..] Table 3: Contract creation transactions for three popular tokens. The input data field contains the contract code. The from address is the owner of the smart contract, which resides at the to address. The Ethereum protocol deterministically computes the address from i) the owner’s address and ii) the account nonce when the transaction is created (by the owner). Table 3 lists the creation transactions of three popular tokens on Ethereum. The from field lists the contract creator; this address is the owner of the token. Contract creation is expensive, and the transaction fee increases with more elaborate contracts. The contract code appears in the input data field of the transaction. ###### Remark 4.1. Usually, the transaction that creates a smart contract carries no ether (as shown with 0 amounts in Table 3). The actual payload is the smart contract code written in the input data field of the transaction. Token networks evolve with changing user balances that we denote as edges between addresses. However, token transfers are internal transactions which are not broadcast to the network in the form of an ordinary Ethereum transaction. In other words, what we call a token trade is, in reality, an update of balances in smart contract variables of the token. It is as simple as changing the values of two keys in a hashmap. (a) Two transactions are mined in blocks. (b) Account transaction network records an edge for transferring 100 wei (c) Token transaction network records an edge for transferring two tokens Figure 17: A trade between addresses $a_{1}$ and $a_{2}$ for the tokens issued by $a_{3}$. Transaction and token networks capture partial views from the transactions given in (17(a)). The zero amount transaction is omitted in the token transaction network, whereas the transfer of two tokens are not captured in an Ethereum transaction network. We will explain the observed transactions and states with the help of Figure 17. A token trade involves the following steps. 1. 1. Address $a_{2}$ owns tokens that are issued by $a_{3}$. $a_{2}$ could have received tokens through various channels. For example, $a_{2}$ may have bought them in an earlier transaction, or the tokens might have been given to $a_{2}$ by the owner of $a_{3}$. 2. 2. A price per token is agreed upon between $a_{1}$ or $a_{2}$ by any off-the- blockchain channel or through a clause in the contract (e.g., the contract sets token price as 50 Wei). 3. 3. $a_{1}$ pays $a_{2}$ an ether amount (100 Wei in Figure 17(a)) by creating a transaction, where $a_{2}$ is the receiver and the input data field of the transaction is blank. Here $a_{1}$ sends the transaction to the network to be mined. 4. 4. $a_{2}$ downloads the latest blocks and sees that 100 Weis have been sent to its address from $a_{1}$. In turn, $a_{2}$ uses the conversion rate and decides to send 2 tokens to $a_{1}$. 5. 5. $a_{2}$ creates a transaction where the receiver is $a_{3}$, and the transferred ether amount is 0. However, the input data lists $a_{1}$ as the recipient of two tokens. Typically the input data is a call to the transfer function of the smart contract at $a_{3}$ with two parameters: to_address:$a_{1}$ and amount:$2$. $a_{2}$ sends this transaction to the network to be mined. 6. 6. At every node of the Ethereum network, the Ethereum Virtual Machine executes the transaction of step 5. The smart contract at $a_{3}$ decreases the balance of $a_{2}$ by two tokens and increases the balance of $a_{1}$ by two tokens. This balance update, which is an internal transaction, is recorded as an edge from $a_{2}$ to $a_{3}$ in the token network of Figure 17(c), but not sent to the network as a transaction. This step fails if $a_{2}$ does not own two tokens. 7. 7. $a_{1}$ downloads and observes the transaction from $a_{2}$ to $a_{3}$. The node at $a_{1}$ can run the transaction in its local Ethereum Virtual Machine to ensure that tokens are transferred without any error. As Figure 17 shows, transaction and token networks have different views from the two mined transactions. Without running a virtual machine and executing the transactions, we cannot observe the internal transactions nor create the token transaction network. Token networks attract trades and traders every day. As Ethereum addresses (both EOA and contract) are reused in multiple days, a token network may see the same node trading in multiple days. However, daily token networks are sparse and consist of disconnected components, and very few traders appear in a token’s network every day. Some traders appear in networks of multiple tokens, making token networks an ideal setting for studying multi-layer networks. Each token network constitutes a layer with edges and nodes, and nodes overlap between layers. ### 4.3 Trace Network Ethereum stores an ecosystem of addresses, smart contracts, and decentralized organizations. In transaction and token networks, we studied financial relationships between addresses. This section now shifts our focus to relationships, call-dependencies, inheritances, and other interactions between Ethereum addresses. A trace network stores these relationships where nodes are EOA and smart contract addresses, and edges are interactions between EOA- contract, contract-EOA, and contract-contract pairs. An EOA creates a transaction that is directed to the address of a smart contract. The smart contract can execute a function and terminate or call functions of other smart contracts. The contract can also move coins to EOAs. The trace can be extended until the transaction gas limit is exhausted. All the addresses that are involved in this call/interaction chain create a trace. In graph terms, a trace is a hyper-edge (i.e., an edge that connects more than two nodes). A trace can involve every operation that a smart contract can execute. For example, a trace can create a smart contract, call a smart contract function, or delete a smart contract. As such, we can label parts of the trace with the executed operation. Figure 18: Transactions and message calls. The account 0x0..ow initiates the first transaction (1st step) whose amount is forfeited as it is less than 1 ether. The forfeited amount is deposited to the contract at 0x0..25 for the owner (0x0..1e) of the smart contract to withdraw it in the third step. Next, the account creates another transaction (4th step) that successfully converts 2 ethers to 6 tokens. ###### Example 4.2. We construct a trace network from a toy scenario given in Figure 18. We show addresses with their first and last characters (e.g., 0x0..57). Figure 18 contains three smart contracts (0x0..57, 0x0..a4 and 0x0..25) and two externally owned accounts (0x0..ow and 0x0..1e). Edges denote transactions or message calls. Transactions are created by externally owned accounts explicitly and mined in blocks, whereas message calls are not. In this example, a contract acts as a salesman for a cryptoasset: 0x0..a4 receives ethers from addresses, and stores their amount of token in its storage. The contract has a rule that the minimum deposit amount is 1 ether. 0x0..ow is not aware of this rule at first and creates a transaction that sends 0.2 ethers to 0x0..a4, which calls a deposit function at 0x0..25 that seizes the amount and stores it in the address. 0x0..1e can observe 1 in the blockchain as a transaction, but 2 is a message call that requires 0x0..1e to execute the contract call to discover. Usually contracts have withdraw functions that allow the contract creator to remove the deposited ethers from the contract. In 3 0x0..1e makes a contract call transaction to withdraw these ethers. 0x0..ow learns about the minimum amount rule, and sees that its 0.2 ethers are forfeited. 0x0..ow makes yet another attempt to buy tokens by sending 2 ethers again in transaction 4. This time, 0x0..a4 accepts the amount, and DELEGATECALLS a conversion function from 0x0..57. which directly stores a state variable that records account 0x0..57 has 6 tokens (three for each ether). So far, three transactions ( 1, 4 and 3) have been initiated by EOAs and mined in blocks. We observe the following three traces from the three transactions: 1. 1. Trace 1: 0x0..ow $\xrightarrow{\leavevmode\hbox to14.18pt{\vbox to14.18pt{\pgfpicture\makeatletter\hbox{\hskip 7.09111pt\lower-7.09111pt\hbox to0.0pt{\pgfsys@beginscope\pgfsys@invoke{ }\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\pgfsys@setlinewidth{0.4pt}\pgfsys@invoke{ }\nullfont\hbox to0.0pt{\pgfsys@beginscope\pgfsys@invoke{ }{ {{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }\definecolor[named]{pgffillcolor}{rgb}{1,.5,0}\pgfsys@color@rgb@fill{1}{.5}{0}\pgfsys@invoke{ }{{}{{{}}}{{}}{}{}{}{}{}{}{}{}{}{\pgfsys@beginscope\pgfsys@invoke{ }\definecolor[named]{pgffillcolor}{rgb}{1,.5,0}\pgfsys@color@rgb@fill{1}{.5}{0}\pgfsys@invoke{ }{}\pgfsys@moveto{6.89111pt}{0.0pt}\pgfsys@curveto{6.89111pt}{3.8059pt}{3.8059pt}{6.89111pt}{0.0pt}{6.89111pt}\pgfsys@curveto{-3.8059pt}{6.89111pt}{-6.89111pt}{3.8059pt}{-6.89111pt}{0.0pt}\pgfsys@curveto{-6.89111pt}{-3.8059pt}{-3.8059pt}{-6.89111pt}{0.0pt}{-6.89111pt}\pgfsys@curveto{3.8059pt}{-6.89111pt}{6.89111pt}{-3.8059pt}{6.89111pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{0.0pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{ } \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{-2.5pt}{-3.22221pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{{1}} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} } \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope{{{}}}{}{}\hss}\pgfsys@discardpath\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope\hss}}\lxSVG@closescope\endpgfpicture}}}$ 0x0..a4$\xrightarrow{\leavevmode\hbox to14.18pt{\vbox to14.18pt{\pgfpicture\makeatletter\hbox{\hskip 7.09111pt\lower-7.09111pt\hbox to0.0pt{\pgfsys@beginscope\pgfsys@invoke{ }\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\pgfsys@setlinewidth{0.4pt}\pgfsys@invoke{ }\nullfont\hbox to0.0pt{\pgfsys@beginscope\pgfsys@invoke{ }{ {{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }\definecolor[named]{pgffillcolor}{rgb}{1,.5,0}\pgfsys@color@rgb@fill{1}{.5}{0}\pgfsys@invoke{ }{{}{{{}}}{{}}{}{}{}{}{}{}{}{}{}{\pgfsys@beginscope\pgfsys@invoke{ }\definecolor[named]{pgffillcolor}{rgb}{1,.5,0}\pgfsys@color@rgb@fill{1}{.5}{0}\pgfsys@invoke{ }{}\pgfsys@moveto{6.89111pt}{0.0pt}\pgfsys@curveto{6.89111pt}{3.8059pt}{3.8059pt}{6.89111pt}{0.0pt}{6.89111pt}\pgfsys@curveto{-3.8059pt}{6.89111pt}{-6.89111pt}{3.8059pt}{-6.89111pt}{0.0pt}\pgfsys@curveto{-6.89111pt}{-3.8059pt}{-3.8059pt}{-6.89111pt}{0.0pt}{-6.89111pt}\pgfsys@curveto{3.8059pt}{-6.89111pt}{6.89111pt}{-3.8059pt}{6.89111pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{0.0pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{ } \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{-2.5pt}{-3.22221pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{{2}} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} } \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope{{{}}}{}{}\hss}\pgfsys@discardpath\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope\hss}}\lxSVG@closescope\endpgfpicture}}}$ 0x0..25 2. 2. Trace 2: 0x0..1e $\xrightarrow{\leavevmode\hbox to14.18pt{\vbox to14.18pt{\pgfpicture\makeatletter\hbox{\hskip 7.09111pt\lower-7.09111pt\hbox to0.0pt{\pgfsys@beginscope\pgfsys@invoke{ }\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\pgfsys@setlinewidth{0.4pt}\pgfsys@invoke{ }\nullfont\hbox to0.0pt{\pgfsys@beginscope\pgfsys@invoke{ }{ {{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }\definecolor[named]{pgffillcolor}{rgb}{1,.5,0}\pgfsys@color@rgb@fill{1}{.5}{0}\pgfsys@invoke{ }{{}{{{}}}{{}}{}{}{}{}{}{}{}{}{}{\pgfsys@beginscope\pgfsys@invoke{ }\definecolor[named]{pgffillcolor}{rgb}{1,.5,0}\pgfsys@color@rgb@fill{1}{.5}{0}\pgfsys@invoke{ }{}\pgfsys@moveto{6.89111pt}{0.0pt}\pgfsys@curveto{6.89111pt}{3.8059pt}{3.8059pt}{6.89111pt}{0.0pt}{6.89111pt}\pgfsys@curveto{-3.8059pt}{6.89111pt}{-6.89111pt}{3.8059pt}{-6.89111pt}{0.0pt}\pgfsys@curveto{-6.89111pt}{-3.8059pt}{-3.8059pt}{-6.89111pt}{0.0pt}{-6.89111pt}\pgfsys@curveto{3.8059pt}{-6.89111pt}{6.89111pt}{-3.8059pt}{6.89111pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{0.0pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{ } \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{-2.5pt}{-3.22221pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{{3}} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} } \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope{{{}}}{}{}\hss}\pgfsys@discardpath\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope\hss}}\lxSVG@closescope\endpgfpicture}}}$ 0x0..25 3. 3. Trace 3: 0x0..ow $\xrightarrow{\leavevmode\hbox to14.18pt{\vbox to14.18pt{\pgfpicture\makeatletter\hbox{\hskip 7.09111pt\lower-7.09111pt\hbox to0.0pt{\pgfsys@beginscope\pgfsys@invoke{ }\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\pgfsys@setlinewidth{0.4pt}\pgfsys@invoke{ }\nullfont\hbox to0.0pt{\pgfsys@beginscope\pgfsys@invoke{ }{ {{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }\definecolor[named]{pgffillcolor}{rgb}{1,.5,0}\pgfsys@color@rgb@fill{1}{.5}{0}\pgfsys@invoke{ }{{}{{{}}}{{}}{}{}{}{}{}{}{}{}{}{\pgfsys@beginscope\pgfsys@invoke{ }\definecolor[named]{pgffillcolor}{rgb}{1,.5,0}\pgfsys@color@rgb@fill{1}{.5}{0}\pgfsys@invoke{ }{}\pgfsys@moveto{6.89111pt}{0.0pt}\pgfsys@curveto{6.89111pt}{3.8059pt}{3.8059pt}{6.89111pt}{0.0pt}{6.89111pt}\pgfsys@curveto{-3.8059pt}{6.89111pt}{-6.89111pt}{3.8059pt}{-6.89111pt}{0.0pt}\pgfsys@curveto{-6.89111pt}{-3.8059pt}{-3.8059pt}{-6.89111pt}{0.0pt}{-6.89111pt}\pgfsys@curveto{3.8059pt}{-6.89111pt}{6.89111pt}{-3.8059pt}{6.89111pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{0.0pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{ } \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{-2.5pt}{-3.22221pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{{4}} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} } \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope{{{}}}{}{}\hss}\pgfsys@discardpath\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope\hss}}\lxSVG@closescope\endpgfpicture}}}$ 0x0..25$\xrightarrow{\leavevmode\hbox to14.18pt{\vbox to14.18pt{\pgfpicture\makeatletter\hbox{\hskip 7.09111pt\lower-7.09111pt\hbox to0.0pt{\pgfsys@beginscope\pgfsys@invoke{ }\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\pgfsys@setlinewidth{0.4pt}\pgfsys@invoke{ }\nullfont\hbox to0.0pt{\pgfsys@beginscope\pgfsys@invoke{ }{ {{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }\definecolor[named]{pgffillcolor}{rgb}{1,.5,0}\pgfsys@color@rgb@fill{1}{.5}{0}\pgfsys@invoke{ }{{}{{{}}}{{}}{}{}{}{}{}{}{}{}{}{\pgfsys@beginscope\pgfsys@invoke{ }\definecolor[named]{pgffillcolor}{rgb}{1,.5,0}\pgfsys@color@rgb@fill{1}{.5}{0}\pgfsys@invoke{ }{}\pgfsys@moveto{6.89111pt}{0.0pt}\pgfsys@curveto{6.89111pt}{3.8059pt}{3.8059pt}{6.89111pt}{0.0pt}{6.89111pt}\pgfsys@curveto{-3.8059pt}{6.89111pt}{-6.89111pt}{3.8059pt}{-6.89111pt}{0.0pt}\pgfsys@curveto{-6.89111pt}{-3.8059pt}{-3.8059pt}{-6.89111pt}{0.0pt}{-6.89111pt}\pgfsys@curveto{3.8059pt}{-6.89111pt}{6.89111pt}{-3.8059pt}{6.89111pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{0.0pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{ } \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}{{{{}}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@transformcm{1.0}{0.0}{0.0}{1.0}{-2.5pt}{-3.22221pt}\pgfsys@invoke{ }\hbox{{\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{ }\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{ }\hbox{{5}} }}\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}} } \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope{{{}}}{}{}\hss}\pgfsys@discardpath\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope\hss}}\lxSVG@closescope\endpgfpicture}}}$ 0x0..57 Note that 6 creates a state change (an internal transaction) but it is not recorded in the trace network. Figure 19: A hyper-graph from Ethereum traces. Three hyper-edges connect five nodes. Nodes that appear on a trace are all considered neighbors. A finer grained hyper graph can add edge features based on interaction types between node pairs. Placing all these traces (hyper-edges) together, we obtain the hypergraph shown in Figure 19. ###### Remark 4.3. The Ethereum virtual machine executes every transaction in order and executes calls accordingly. There is no concurrency in Ethereum transaction execution; EVM does not make two calls simultaneously. When a contract calls two contracts sequentially, EVM will make the second call only after the first contract call (and any further contract calls it makes) ends. In theory, traces can reach large sizes in the shape of trees with many branches. However, the transaction fee grows with additional calls and prevents the creation of huge traces. Note that even when a transaction sets a high gas limit to pay a big transaction fee, the transaction will not be mined if the gas used exceeds the Ethereum block gas limit (currently 12.5K). If a trace creates a call loop, this may deplete smart contract balances before the transaction gas limit is reached. We can understand smart contract behavior by trace network analysis. For example, we can study the network to discover unexpected call branches in smart contracts. ## 5 Ripple: Credit Networks Ripple (www.ripple.com) and Stellar (www.stellar.org) are two credit networks that closely resemble the ancient Hawala system (in Arabic, hawala means to transfer or trust). The main idea of Hawala is to allow a money sender to use connections of people who trust each other to make a payment in a distant geographical location. Merchants have historically used these types of networks to transfer money between countries [17]. We model credit networks as directed, weighted graphs that we build from trust lines between address pairs. A trust line is a directed edge between two addresses, which implies that the source address trusts the target address. We can add weight (i.e., money amount) to the edge to show the limits of that trust. In a credit network built from trust lines, a directed edge is a promise by the source node that it will let the target node use a loan amount in a future transaction. Trust lines can be deleted or updated for amounts in time. It is helpful to explain a few confusion points in Ripple to a blockchain researcher. Ripple uses the term ledger instead of a block. Transactions have multiple types and may involve financial constructs (e.g., checks), user- issued currencies (e.g., USD), and path-based settlements. The reader should remember that Ripple contains elaborate business logic in its building blocks (e.g., transactions). Before we delve into Ripple, we will use a few scenarios to explain how we can use a network of trust lines to make a payment. ###### Remark 5.1. Academic articles have conflicting views on how to represent a trust line as a directed graph edge. Here, we follow the notation used in the official Ripple documentation, i.e., the edge is from the lender (source) to the borrower (target). The lender trusts the borrower. Figure 20: Two Ripple trust lines between Charlie, Bob, and Alice. Charlie has created a trust line for Alice for 500USD. Alice has already used 5USD from the trust line. The amount below both edges, 500 USD, is the maximum limit (set by Charlie and Bob) that Alice can use. Figure 20 shows two trust lines between Alice, Charlie, and Bob. Each edge has two edge features. First, a value below the edge (500 USD for both edges) is the maximum credit amount (known as the limit) that Charlie and Bob trust with Alice that Alice can use in future transactions. Second, a value above the edge (5 and 15 USD) is the balance amount that Alice has already used with Charlie and Bob. For example, Alice can use $500-15=$485 in a future transaction by using the trust line from Bob. Users may create the Ripple trust line due to two reasons. First, Alice may have paid 500 USD to Charlie and Bob each in an off-the-Ripple transaction. Second, Charlie and Bob may have trusted Alice based on her reputation and created trust lines for her to use. In practice, the amounts are almost always due to the first reason, i.e., offline deposits. By default, Ripple limits are lower-bounded by 0 and upper-bounded by $\infty$. A high limit creates credit risks for the lender. We make a Ripple trust graph from such trust lines. We can delete an existing trust line by setting its trust limit to zero. We can use trust lines to make payments to third addresses over a path that traverses individual trust lines in the same currency, such as USD. Traveling trust lines is known as rippling. Figure 21 will help the reader learn basic rippling across trust lines. (a) Trust lines. Before Sarah sends any currency, the Ripple trust graph includes five nodes and five trust lines. (b) Payment. After the rippling, owed amounts are updated and shown in red rectangles. Figure 21: An example of the Ripple trust graph before (left) and after (right) Sarah uses rippling to send 50 USD to Bob. On the (left) network, Sarah has two paths that connect her to Bob: 1) Bob $\rightarrow$ Alice $\rightarrow$ Sarah and 2) Bob $\rightarrow$ John $\rightarrow$ Tim $\rightarrow$ Sarah. However, in the 1st path that traverses two trust lines, the trust line between Sarah and Alice has a max limit of 20 USD (2 USD of which is already used), which is less than the 50 USD that Sarah wants to send. As a result, Sarah cannot use the first path to send all 50 USD. The second path is longer than the first because it traverses three trust lines; however, the limits allow the 50 USD transfer. Sarah uses the second path for rippling. Owed amounts (values above edges) are increased by the transferred amount: 50 USD. Sarah needs to consider updated (right) values in future transactions; she can no longer send 50 USD to Bob because the trust line John $\rightarrow$ Tim has a current capacity of USD 100 - USD 75 = USD 25 only. The updated (increased) amounts imply that after the payment, Sarah owes 50 USD to Tim, who owes 75 USD to John, who owes 65 USD to Bob. ### 5.1 Ripple Networks On cryptocurrencies, a single type of coin, such as bitcoin, is traded between addresses. Similarly, Ripple has its currency called XRP with a drop sub-unit (1 XRP = 1 million drops). However, Ripple also allows users to issue currencies and create trust lines in the issued currencies. Ripple officially recognizes two main currency types: i) Ripple (abbreviated as XRP) and ii) user-issued currencies. Academic articles further divide user- issued currencies into country currencies (e.g., US dollar, European Euro, Japanese Yen), cryptocurrencies (e.g., Bitcoin), and fictional currencies such as tokens. For example, in Figure 20, USD amount is a user-issued currency. User-issued digital tokens are fictional currencies with no outside backing. Traders must be aware that such tokens have no inherent value. XRP is issued natively and has no issuer. All other currencies are represented as “currency.issuer” in transactions. We represent a currency with i) three characters such as USD, EUR, or ii) less frequently by a 40 character string. USD issued by address1 is USD.address1 on the Ripple ledger (also called the XRP Ledger). Multiple issuers can issue a currency: USD.address1 and USD.address2 are considered two different currencies, even though they are both USD currencies. However, in rippling user issued currencies of the same denomination (e.g., USD) are considered to be the same currency. Trust in the issuer determines the fungibility of a user-issued currency. If Alice does not trust address1 but trusts address2, 1 USD.address1 will not be worth 1 USD.address2 for Alice. Accordingly, Alice may refuse to participate in rippling transactions that bring her USD.addres2 amounts. Each Ripple address needs to store a non-trivial XRP amount as a reserve that it cannot spend; otherwise, Ripple considers the address deleted. The reserve requirement (currently 20 XRP) discourages multiple address creation and usage. If a security breach occurs, a user may change its address; however, best practice reuses the same address for multiple transactions. ###### Info 1 A currency issuer may set a transfer fee for every transaction in that currency between other addresses. The fee is similar to earning a commission for every transaction in the currency. The issuer of a popular currency thus accumulates transfer fees, which reduces the amount of offline liability it must hold to exist as an entity in the Ripple network. The issuer can also freeze all transactions of its currency (XRP cannot be frozen by anyone). However, holders of the currency can still redeem the currency by returning it to the issuer and getting paid outside of the XRP Ledger. However, in real life, the issuer may have already gone bankrupt or may ignore redeeming requests. Every address on the ledger must make its own trust decisions by considering these issues and create/allow trust lines accordingly. Figure 22: Ripple address types. We outline Ripple Ledger address types in Figure 22 as a gateway, cold wallet (issuing address), hot wallet (operational addresses), market maker, and wallets (i.e., ordinary traders). Gateways are the currency issuers of the Ripple network and link the XRP Ledger to the rest of the world. All network nodes can buy Ripple or any other issued currency by paying money (e.g., US dollars) to the gateway (outside the Ripple ledger). The following steps, taken verbatim from the Ripple documentation, explain the six necessary steps. 1. 1. A customer sends money to a gateway’s offline accounts. This could be fiat money, Bitcoin, or any other asset not native to the XRP Ledger. 2. 2. The gateway takes custody of the money and records it. 3. 3. The gateway issues a balance in the XRP Ledger, denominated in the same currency, to an address belonging to the customer. This is done by creating a trust line to the customer’s Ripple address. 4. 4. The customer uses the issued currency in the XRP Ledger, such as by sending cross-currency payments or trading in the decentralized exchange. 5. 5. A customer (not necessarily the one who deposited the money initially) sends the issued currency to the gateway’s XRP Ledger address. 6. 6. The gateway confirms the customer’s identity who sent the balance in the XRP Ledger funds and gives the corresponding amount of money outside the XRP Ledger to that customer. An institution may use both cold and hot wallets on the XRP ledger. Connected to the Internet, a hot wallet signs ordinary transactions, and the cold wallet stores currencies offline and is used less frequently. Figure 23: A Ripple trust graph. The market has trust lines to two nodes. Note that it is the gateway and market that have credit lines to the users. If the balance (edge weight above the line) is non-zero, the node has already used the credit to make a payment to another node. Assume that market and gateway are low accounts. Below lines, we show low and high limits, respectively. A market maker is an address that converts a currency to another and takes a conversion fee through its offer, which is broadcast to the ledger in an offer create transaction. The offer may immediately consume earlier offers and gain the market maker what currency it desires. If the offer is not fully consumed, it is stored in the ledger as an offer object that can be used in future payments or consumed by future offers (of other addresses). Offers must be updated in real-time as currency conversion rates change (for country currencies, the conversion rate must be updated according to international exchange rates). A standing exchange offer is canceled in an offer cancel transaction. We will teach Ripple networks in two subsections: Ripple trust graph and Ripple payment graph. The trust graph stores trust lines and balances. The payment graph utilizes the trust graph to process transactions. ### 5.2 Ripple trust graph We create the Ripple trust graph from trust lines issued by addresses. In our basic examples of Figures 20 and 21, trust lines were not reciprocal. However, in Ripple, both addresses can create a trust line to each other in a currency. Furthermore, two addresses can have trust lines in multiple currencies. Ripple creates a single RippleState object to connect two accounts per currency, sorts the account addresses numerically, and labels the numerically lower address as low account and the other as the high account. Assume that Alice and Bob have trust lines in USD and EUR amounts. Ripple creates two RippleState objects for Alice and Bob; one for EUR and one for USD. The object has high and low limit fields to record the limit from high $\rightarrow$ low and low $\rightarrow$ high trust lines. However, the net balance is a single value shared by the addresses, and Ripple stores the net balance of the trust line from the low account’s perspective. ###### Remark 5.2. So far, we have used trust lines between Ripple users (e.g., John $\rightarrow$ Sarah) in our payment scenarios. However, ordinary users do not issue currencies on Ripple, which means that actual trust lines connect currency issuer-user pairs (gateway $\rightarrow$ Sarah). We can check the currency issuer in a RippleState with balance. If the balance is positive, the high account is the issuer. If the balance is negative, the low account is the issuer. Often, the issuer has its limit set to 0, and the other account has a positive limit, but this is not reliable because limits can be increased or decreased without affecting an existing balance. The Ripple trust graph is a directed, weighted multi-graph where an edge has four features: currency name (i.e., type of the edge), balance, low limit, and high limit. Multiple currency edges can connect node pairs. For example, in Figure 23 the gateway and rf1…Jpn have trust lines in USD and EUR currencies. We give the graph data of this figure in Table 4 where we have not excluded 0 high limits. low | high | currency | balance | low limit | high limit ---|---|---|---|---|--- gateway | rSA…Adw | USD | 250 | 500 | 0 gateway | rf1…Jpn | USD | 0 | 300 | 0 gateway | rf1…Jpn | EUR | 0 | 30 | 0 market | rf1…Jpn | EUR | 50 | 500 | 0 market | r4Y…b3n | USD | 0 | 370 | 0 Table 4: Ripple trust graph data. Gateway and market are assumed to be low accounts, whereas user addresses are high accounts. High limit values are 0 because high accounts are all user addresses. A trust line deactivates in three cases. First, the lender may remove the trust line by updating trust parameters in a TrustSet transaction. Second, the lender may freeze the trust line temporarily. Third, the currency issuer can freeze all transactions of the currency (except for the redeeming transaction). In two cases, a node may hold a balance on a trust line greater than the limit, that is, when the node acquires more of that currency through trading and when the node decreases the limit on the trust line. The average node degree is less than four in the trust graph, and the clustering coefficient is around 0.10. ### 5.3 Ripple payment graph The Ripple trust graph shares edge features with the Ripple payment graph. Additionally, the payment graph stores a node feature called reserve (i.e., XRP balance of the node). We cannot use the reserve amount in transactions. In addition to the base reserve, the reserve amount increases for every object (such as a trust line) that we create. The reserve increase applies only to the node extending trust, not to the node receiving it. We create the Ripple payment graph from 1) direct payments, 2) path-based settlements, and 3) a few other financial constructs. We will teach these payment types before giving a graph model for the payment graph. #### 5.3.1 Direct payments A direct payment is an XRP transfer between two Ripple addresses and does not require a trust line. Regardless of what trust lines exist in the trust graph, two ripple addresses can send to and receive XRP from each other. The XRP balance of an address must not fall below the total required reserve. Otherwise, the address is not allowed to create some types of transactions. When an address’s XRP balance falls below the reserve, the address may issue OfferCreate transactions to buy more XRP or other currencies on its existing trust lines. However, these transactions cannot create credit that we can use to buy XRP to satisfy the reserve requirement (i.e., when the account balance$<$ the required reserve). For example, the address may pay a transaction fee and create an offer for XRP on the ledger. Once a buyer consumes the offer, the address will receive its XRP payment and satisfy the reserve requirement. When an address is below the reserve requirement, it can send new OfferCreate transactions to acquire more XRP or other currencies on its existing trust lines. These transactions cannot create new trust lines, nor Offer objects in the ledger, so they can only execute trades that consume existing Offers. #### 5.3.2 Path-based settlements A path-based settlement transaction connects a source and a destination node pair via a network path through rippling. For example, in Figure 23 the node rsA…AdW can make a payment to rf1…Jpn through the gateway. We can make a payment by consuming exchange offers along the path. For example, if the sender wants to make a payment in USD, the path can use trust lines that follow USD$\rightarrow$EUR$\rightarrow$USD, where the arrow indicates a currency exchange. Traversing exchanges is called bridging. If the transaction offers a better cost, currency$\rightarrow$XRP and XRP$\rightarrow$currency conversions can automatically be executed on the path as well. This is called auto-bridging. A path-based settlement can explicitly write one or more paths (called a pathset) in a payment transaction. All paths in a pathset must start and end with the same currency, and the sender can also leave path selection to Ripple. Pathfinding is difficult because user XRP balances and trust lines change every few seconds as new ledgers (i.e., blocks of transactions) are validated. As a result, Ripple is not designed to compute the absolute best path. The simplest possible path to connect the steps of the transaction is called the default path. In addition to the paths written in the transaction, Ripple can choose to use the default path. The sender must set the tfNoDirectRipple flag to avoid the default path and force Ripple to use the paths of the pathset. In a path-based settlement, we can set a few fields of the transaction to restrict path selection. These are the Account (sender), Destination (receiver), Amount (currency and amount to be delivered), and SendMax (currency and amount to be sent, if specified) variables of the transaction. The variables shape the path in these ways: * • The Account (sender) is set to be the first step of the path. * • If the SendMax variable is set to be a currency issuer, the issuer must be the second step of the path. * • If the Amount variable is set to be a currency issuer, the issuer must be the second-to-last step of the path. * • The path ends at the Destination address. The actual cost of a path-based payment can change between submission and execution of a transaction based on updated information (e.g., the limit of a trust line) from transactions. A path-based payment can fail for multiple reasons; first, the receiving address may not have the reserve XRP amount. Ripple checks whether the payment would deliver enough XRP to meet the reserve requirement in such a case. If not, the payment fails. Next, the receiving address may have limitations on receiving payments, such as Deposit Authorization or RequireDest (We will cover these constructs shortly). 333Currency issuers set such rules. Third, the paths that the sender specifies may have dried up (i.e., trust line limits are exceeded). A rippling transaction can redistribute credit from a more trusted to a less trusted issuer without the specific consent of the involved address’ owner [18]. Gateway and market addresses should allow rippling; however, ordinary addresses may benefit from disallowing rippling on their trust lines. Since 2013, Ripple has allowed setting a no_ripple flag to disable rippling transactions involving the trust line. Since March 2015, the flag has been set to true by default; users must opt-in to allow rippling. Furthermore, the defaultRipple flag enables rippling among all the wallet’s trust lines. Researchers in 2016 had found that most paths are short [18]: 15.1% of transactions are direct payments. 52.7% of transactions use a single intermediate address and represent the interactions of gateways with their users following the hot-cold wallet mechanism. 32% of transactions use two or more intermediate addresses. In the network, the average path length was 6.83, whereas the diameter was 13. #### 5.3.3 Financial Concepts Ripple adopts and uses the following concepts from traditional finance: offers, checks, and escrows. Offer. Offers are orders to exchange two currencies (XRP or issued currencies in any combination). Furthermore, we can trade the same currencies issued by different issuers in offers. An offer is created in an OfferCreate transaction by the sender, and Ripple tries to match the offer by using existing offers fully. If a matching offer is available, the currency balances of both offer senders are updated accordingly. If the offer is not fully consumed (for the requested amount), the offer is written as an offer object to the ledger for the remaining amount. Future transactions can consume the remaining part of the offer. Offers are also used on path-based settlements to convert one currency to another. This auto-bridging of currencies happens automatically on any OfferCreate transaction. ###### Example 5.3. An offer $o_{2}$ that proposes exchanging 10 USD for 7 EUR ($o_{2}$ sender has 10 USD) will consume an existing offer $o_{1}$ that proposes exchanging 7 EUR for 10 USD ($o_{1}$ sender has 7 EUR). Balances of both senders will be updated; $o_{2}$ sender will have 7 EUR whereas $o_{1}$ sender will have 10 USD. If the exchange rate favors the sender of $o_{2}$, it will sell less currency. For example, if $o_{1}$ offers 7 EUR for 9 USD, $o_{2}$ will take the order and save 1 USD; the sender of $o_{2}$ buys 7 EUR with 9 USD and keeps its 1 USD. Offers are typically used to create cross-currency payments which convert one currency to another by consuming Offers. A cross-currency payment involves two or more currencies; at least one of the currencies must be non-XRP. Cross- currency payments are fully atomic; either the payment fully executes or it fails. Note that an address can buy user-issued currency through an offer. In that case, the user is deemed to hold an amount of the issued currency, which creates a trust line to the currency issuer. This new trust line is written to the ledger in an object, which increases the reserve XRP requirement of the offer’s owner. For this reason, if the owner does not have enough XRP as a reserve, the offer will be considered unfunded and fail. Offers are explicit instructions to acquire certain issued currencies, so they can go beyond limits set in a trust line. Check. A check is a deferred payment that its intended recipient can cash out. Checks record sender, receiver, amount information, and expiration date. The receiver can partially cash a check, and only at the cashing time must the sender have the required amount of the currency in its account. Otherwise, the check fails. Differing from bank checks, both the sender and receiver can cancel a Ripple check before it is cashed out. Checks can send either XRP or issued currencies. Checks are helpful for financial institutions to avoid receiving unwanted payments and comply with financial regulations. If a user wants to deposit an amount, it can write a check for a Ripple address of the financial institution and expect the institution to cash it. The receiving address can fully or partially cash or refuse the check. Ripple allows rejecting all incoming payments by default, which we can achieve by enabling Deposit Authorization on an address by setting the asDepositAuth flag in an AccountSet transaction. Our address can receive payments from i) pre-authorized addresses or 2) Escrow, Payment Channels, or Checks. Escrow. Escrows are conditional payments that set aside XRP and deliver it when certain conditions are met. The conditions can be time-based unlocks and crypto-conditions. Escrows can expire if not finished in time. Escrows lock up XRP to be used by a specific receiving address. The escrowed XRP cannot be spent elsewhere nor used or destroyed until the escrow has been successfully finished or canceled. After the expiration time, the unused XRP will return to the sender. Payment Channel. Similar to Bitcoin, Ripple uses Payment Channels to set aside XRP that can be used in high-volume microtransactions without recording all channel transactions in ledgers. Payment Channels and Escrow can use XRP only, and an address can send XRP to itself through Escrow. However, an address cannot use Payment Channels to send XRP to itself. Checks are more constringent; an address cannot use checks to send XRP nor issued currency to itself. Partial Payment. The Amount field of a Payment transaction specifies the exact amount to deliver after charging exchange rates and transfer fees. However, the trust graph may not allow the exact amount to be delivered due to trust line limits, exchange rates, and transfer fees. Rather than failing the transaction altogether, the sender can set the Partial Payment flag (tfPartialPayment) to automatically reduce the starting amount to a deliverable amount. Partial payments are notorious because they have been used to exploit naive integrations with the XRP Ledger to steal money from exchanges and gateways. Assume that an address sends 100 USD to a gateway with the partial payment flag set. The delivered amount (e.g., 30 USD) may be much less than the intended amount. The gateway should not use the amount field to redeem the money. Otherwise, it will pay the address 100 USD and lose 70 USD. Partial Payments have the following limitations: a partial payment cannot provide XRP to satisfy the reserve requirement of an address. Also, direct XRP-to-XRP payments cannot be partial payments. With all the financial concepts and path- based settlements, 79% of transactions involve user-issued currencies, whereas 21% transfer the native cryptocurrency XRP. ## 6 IOTA: Tangle Networks The IOTA blockchain uses a Proof-of-Work algorithm for mining a transaction, and the concept of a block does not exist. The approach is markedly different from other blockchains, where miners mine blocks of transactions rather than individual transactions. We will cover IOTA networks in two sections: the IOTA Tangle graph and the IOTA transaction graph. As a starting point for both graph types, we will discuss the seed, subseed, private key, difficulty, and hashing concepts that IOTA uses in address and transaction. IOTA uses two types of addresses; one-time addresses and Merkle root addresses. One-time addresses send and receive payments. In a sense, one-time addresses are similar to ordinary Bitcoin addresses. Merkle root addresses are used in special applications such as Masked Message Authentication, allowing data channels with publishers and subscribers. ###### Remark 6.1. This section will teach one-time addresses because IOTA is phasing out some of the applications that use Merkle root addresses. For example, Masked Message Authentication is being replaced with IOTA Streams. We do not yet know how Merkle root addresses will be used in the new technologies. IOTA uses a concept of seed which is a string of 81 characters. Each seed character is generated from a tryte, which in turn is three characters (trits) that can take -1, 0, or 1 values. The use of trits, rather than bits that can take 0 or 1, is quite unusual in the blockchain. IOTA defends the practice by arguing that “ternary computing is considered to be more efficient as it can represent data in three states rather than just two”. For example, $\\{0,0,0\\}$, $\\{0,-1,1\\}$ and $\\{1,0,1\\}$ are three example trytes. As each trit can take three values, a tryte of three trits can take $3^{3}$ possible values. IOTA uses the number 9 and the uppercase letters A-Z (27 characters in total) to map a tryte to an easy-to-read character. Table 5 shows three example trytes and their character mappings. Decimal number | Tryte = 3 Trits | Tryte-encoded character ---|---|--- 0 | 0, 0, 0 | 9 1 | 1, 0, 0 | A 2 | -1, 1, 0 | B Table 5: IOTA uses trytes and encodes each tryte in a character from a 27 character alphabet for ease of use. A second concept is called an index, which can take values between 0 and 9,007,199,254,740,991. We use the index to create multiple addresses from the same seed. The hash value of seed+index is called the subseed. IOTA uses the KERL hash function, which is the Ternary version of the Keccak-384 hashing algorithm. The hash produces a 243 trit long (81 trytes) subseed. ###### Info 2 In cryptography, a sponge function or sponge construction is a class of algorithms with a finite internal state that takes an input bitstream of any length and produces an output bitstream of any desired length. Wikipedia. IOTA uses three security levels to determine the length of the private key: 1, 2, and 3. The private key is computed by absorbing and squeezing the subseed in a sponge function 27 times for each security level. The default security level is 2, which produces a private key of 4374 trytes. Longer private keys bring better security; however, they increase the signature length of a transaction (see Figure 24). Figure 24: IOTA address creation by using index and security parameters. An address is created by hashing the computed private key. Note that we compute a different address for each difficulty. A seed can create $9^{57}$ addresses. Signature length causes efficiency problems in transactions. IOTA uses empty transactions that carry neither currency nor a message to store the signature of an address that has security level 2 or 3. Assume that for a security level 2 address $a_{i}$, we create a (regular) transaction $t_{0}$ and prepare its signature, which will be 4374 trytes. However, signatureMessageFragment field of a single transaction such as $t_{0}$ can contain only 2187 trytes. Consequently, we need to create another empty transaction $t_{1}$ to store the second fragment of the signature. A data structure called bundle links transactions $t_{0}$ and $t_{1}$ to each other, and the transactions are submitted to the Tangle (i.e., the directed acyclic graph) together. The bundle hash is created from individual transactions of the bundle. As such, any change in any transaction invalidates the bundle hash. For security level 3, we would need to create two empty transactions. Considering that each transaction requires its POW computations and takes up disk storage, the bundling scheme reduces the efficiency of IOTA. We will cover bundle usages in both the tangle graph and transaction graph. The private key is split into 81-tryte segments where each segment is hashed 26 times. Afterward, segment hashes are hashed together to create an 81 tryte long IOTA address. Sometimes, a checksum of 9 trytes is appended to the address resulting in 90 trytes. An example IOTA address is OGMMQJUDMNNYSOAXMJWAMNAJPHWMGVAY9UWBXRGTXXVEDIEWSNYRNDQY99NDJQB9QQBPCRRNFAIUPGPLZ. ###### Info 3 One-time signatures are public-key signature schemes that have the property that we can use the signatures to sign one single message only. On the other hand, such schemes are usually highly efficient and easy to implement in very constrained devices since they only require the implementation of hash functions and not any advanced arithmetic operations [19]. IOTA uses the Winternitz one-time signature scheme to generate digital signatures [19]. Winternitz has the advantage of being robust, easy to compute, and resistant to attacks from quantum computers. On the downside, Winternitz reveals a part of the private key whenever the key signs a transaction. For example, if the address signs two transactions, the signing process reveals more than 50% of its private key. As a result, an address should sign a single transaction only to minimize risk. The address can receive coins from multiple transactions, but it should spend all of them at one attempt. If the address leaves a balance or receives future payments after it has spent its coins, malicious users can attempt to predict the address private key by using brute-force attacks. As a result, address reuse in IOTA is possible but not recommended. ### 6.1 IOTA tangle graph An IOTA transaction validates and lists two past transactions, which are called tips. Afterward, the transaction is mined with a POW that is considerably easier to solve than the Bitcoin POW. This way, IOTA distributes the mining task to individual users who do not need powerful POW mining devices, which makes IOTA an ideal blockchain for low-powered Internet-of- Thing devices. The resulting transaction dependency network is called the Tangle, as shown in Figure 26. Figure 25: An IOTA transaction with its two linked past transactions. Edges to tips have dashed and solid fills (as used in official IOTA documentation) depending on their type. Tip types are used to mark information in certain transactions such as in bundles. Each new transaction must select two existing transactions and list them as trunk and branch. This process is called tip selection, which is at the heart of IOTA security. Although theoretically, a new transaction could be spending the coins received from one of its tip transactions, the new transaction does not have to have any relation with tips. We can select any previous transaction as a tip. Multiple new transactions can select an existing transaction as a tip. A transaction $t$ validates its trunk and branch tip directly. Future transactions that use $t$ as a tip confirm the trunk and branch tips of $t$ indirectly. Unfortunately, confirmations by individual transactions do not prevent double- spending attacks on IOTA. In Figure 26, assume that $t_{1}$ and $t_{2}$ spend the same IOTA coins in an address $a_{1}$. Each double-spending transaction can reach a different set of users, who select either $t_{1}$ or $t_{2}$ as one of their tips. Note that users would not select both $t_{1}$ and $t_{2}$ because they spend the same coins. In traditional blockchains, the block miner would choose to include either $t_{1}$ or $t_{2}$, and the network would be safe from the double-spending problem. Through an entity known as the Coordinator, the IOTA foundation creates and uses milestone transactions, which protect the Tangle against double-spending attacks. Milestones arrive every minute and act as anchor points where the confirmed subtangle is considered to be valid and in consensus. The coordinator address is hard-coded in the IOTA software so that all tangle participants can locate and trust milestone transactions. In this aspect, IOTA is a mix of private and public blockchains because although the Tangle is open to the public, the consensus on the Tangle state is guaranteed by the IOTA foundation. In the future, IOTA is aiming to transition to a coordinator-free POW scheme known as Coordicide (i.e., killing the coordinator). Figure 26: IOTA tangle graph with milestone transactions $m_{0}$ and $m_{1}$. Assume that transactions $t_{1}$ and $t_{2}$ double spend the same IOTA coins. As $m_{1}$ indirectly confirms $t_{1}$, all transactions (marked with a cross sign) that confirm $t_{2}$ are building on an invalid subgraph. As a result, these transactions will not be used as tips in future transactions. If a transaction such as Self1 is not confirmed in time, we can create a new zero value transaction Self2 that confirms Self1. New transactions will use Self2 as a tip so that Self1 will be indirectly confirmed by them. Consequently, we hope that Self1 will be indirectly confirmed by a future milestone transaction. In Figure 26, we show two milestones $m_{0}$ and $m_{1}$ which indirectly confirm $t_{1}$ and invalidate $t_{2}$. The reason is due to the fact that tangle users will see $t_{1}$ being confirmed by $m_{1}$, and the users can mark $t_{2}$ as double-spending the coins used in $t_{1}$. tx hash | epoch value | value | bundle | tag | address | branch tip | trunk tip ---|---|---|---|---|---|---|--- FPX..999 | 1603985801 | 0 | XTR..QZX | COR..VU | COR..CJZ | EJR..999 | RBL..999 EJR..999 | 1603985790 | 0 | DKR..NF9 | COR..UM | COR..CJZ | MNF..999 | NWX..999 RBL..999 | 1603985788 | 0 | PZY..SP9 | COR..UL | COR..CJZ | DPW..999 | YBD..999 Table 6: Graph edge list for the Tangle for three zero-value transactions with the same address (we shorten addresses and tx hashes for clarity). A value 0 implies zero-value (message) transactions where the address is the transaction creator. In value transactions, the address may belong to the transaction sender or receiver. The first transaction FPXNJVTEVMSAQCUMGRTLGLLNZQOLOUHWXNNYZAXBBEIDXOTTXT9PIRKKBGNTKHPMATIQUHSMNELV99999 has the next two transactions as its tips. Note that these transactions do not belong to the same bundle. The transaction tag (attachmentTag field) is used to label transactions. The epoch value is an integer that counts the number of milliseconds since Jan 1, 1970. ### 6.2 IOTA transaction graph The tangle graph is extended by appending new transactions that confirm tip transactions. We classify IOTA transactions into value and zero-value transactions. Zero-value transactions: IOTA creates a zero-value transaction (see Table 27) for two purposes. First, the transaction may store signature fragments for addresses that use security levels 2 or 3. Second, the transaction may store any data such as sensor data, voting results, or encrypted messages. In both types, data is stored in signatureMessageFragment or attachmentTag fields. All other fields of the transaction are trivial. Figure 27: Bundle structure. Two Bitcoin UTXO transactions (top) and their corresponding IOTA bundles (bottom). Dashed edges denote trunk tips, whereas solid edges show branch edges. Tips of bundles ($t_{0},\ldots,t_{3}$) are past transactions that are unrelated to these IOTA bundles. We show trunk tips with dashed edges. The figure assumes that security level=2 for addresses, which requires creating one zero-value transaction (in gray color) to store the second fragment of each input address signature. For the 3rd level, there would be two zero-value transactions for each input address. We give bundle indices (e.g., $0/4$) next to shapes. The left bundle has five transactions, whereas the right bundle has four. Value transactions: IOTA adopts the UTXO transaction model with a few modifications. UTXO model assumes that a transaction has one or more inputs and one or more outputs. For each input, IOTA creates an input transaction which lists an address and the withdrawn token value. For each output, IOTA creates an output transaction with an address and a value. The value signs can distinguish input and output transactions: negative for input and positive for output transactions. For the input transaction of an address with security level 1, the address signature is stored in the signatureMessageFragment field of the input transaction. If the address security level is 2 or 3, additional zero-value transactions are created to store second and third key fragments. Output transactions do not require any address signatures. transaction | address | time | type | value ---|---|---|---|--- AXI..99 | OOC..H9X | 1603951001 | input | -142 998 000 MVP..99 | EM9..SYW | 1603951006 | output | 1 000 EKG..99 | NBN..GJA | 1603950941 | output | 142 997 000 Table 7: One input and two outputs transactions from bundle SGEAO9WIHW HEEWZRQIVCNIS HBFBE ZCTEYTKTSRUBLXKLD ETKG9ISSUGBOQGFFTDLD GGFBNEIYPYUBEK9C. IOTA tokens are transferred from the first row address to the next two row addresses. As the input transaction uses security level 2, the bundle contains an additional zero-value transaction (omitted here) to store the second fragment of the signature of OOC..H9X. In Table 7 we show a UTXO transaction and its input and output transactions on IOTA. We chose to show the “type” field to help the reader, but we can as well omit the field because the value can distinguish transaction type. The table omits zero-value transactions that are required to store signatures. In Figure 27, we show two UTXO transactions with all required IOTA bundle transactions. In the figure, note that zero-value transactions are indexed just after their input transactions; $a_{1}$ signature is stored in index $1/4$. However, the order of the input-output transactions is not important in a bundle; output transaction of address $a_{5}$ could also be given at index $0/4$. As Figure 27 clearly shows, an IOTA bundle uses far more pointers and stores many more data pieces than a UTXO transaction. This inefficiency, coupled with low POW difficulty, results in a fast-flowing and large Tangle, which is too costly to store. IOTA uses snapshots to limit tangle size periodically. A snapshot resets all transaction history but saves a list of address balances. Only IOTA permanodes store all IOTA history. Snapshots occur every few months, and IOTA users are expected to observe the snapshot and manage their IOTA usage accordingly. ### 6.3 IOTA Streams IOTA Streams is a framework for building cryptographic messaging protocols on IOTA in the Rust programming language. An older version of Streams is known as the Masked Authenticated Messaging, which allowed publishers to attach periodic data (e.g., home temperature readings every hour) as zero value transactions to the Tangle. Streams has a native protocol for attaching messages to the Tangle, but users can extend Streams to send messages from other mediums, such as HTTP URLs. IOTA Streams allows attaching data in public, private, and subscriber-only modes. In the private case, we can only read Streams messages if the encryption key is known. In the subscriber case, we use a subscription key in addition to the private key. Streams messages contain a reference for the next (future) message in the Tangle. If the message is encrypted or subscriber- only, we cannot locate the future message in the Tangle. Even if we locate the message, we cannot decrypt its content without the private and subscriber keys. Streams is still in development while its prior-form, the Masked Authentication Messaging, is widely used in the Tangle. We can model the Streams graph as a hypergraph where nodes are transactions, and directed edges between nodes indicate the shared stream. Note that Streams messages may use previous Streams messages as tips, but this is unnecessary for a well-formed stream. Streams transactions may be value or zero-value transactions. In the zero-value case, Streams transactions do not have to be confirmed by any other IOTA transaction. ## 7 Tools for Blockchain Data Analytics We can access blockchain data easily by downloading wallet software and connecting to the P2P network. Recently, Google has implemented ETL libraries (https://github.com/blockchain-etl) to help parse Bitcoin, Monero, ZCash, Ethereum, and a few other blockchains. However, blockchain sizes have become prohibitively large to run data analytics in a single machine. For example, the BlockSci tool [20] requires around 60GB of memory to load the Bitcoin transaction graph. Many blockchain analytics companies offer REST APIs for blockchain data. Examples are etherscan.io, blockcypher.com, and infura.io. However, APIs are either costly or allow limited access to the API. We recommend running your wallet and using the Blockchain-etl libraries of Google to extract blockchain data. Most tools and algorithms for blockchain data are related to e-crime or financial (e.g., price, investor) analytics. For example, Barjasic et al [21] “analyze the time-series of minute price returns on the Bitcoin market through the statistical models of the generalized autoregressive conditional heteroskedasticity family”. From ransomware payment detection [22] to sextortion discovery [23], transaction network analysis has proven useful to study blockchain address importance and to cluster blockchain addresses. Network flow algorithms [24], random walks [25] and Petri nets [26] are the main unsupervised methods in this line of research. For example, starting from Bitcoin addresses of potential interest, Egret [23] analyzes neighborhood subgraphs in terms of path length and confluence to detect suspicious Bitcoin flow and other wallet addresses controlled by malicious actors. A rather curious case of clustering research involves identifying heuristics that can link multiple addresses used by a real-life entity. Although heuristics are error-prone, they are widely used for blockchain data analytics (e.g., see [27, 28, 29])). Another emerging promising approach for ransomware payment detection is topological data analysis (TDA) [30, 31]. TDA systematically infers qualitative and quantitative geometric and topological structures of blockchain transaction graphs at multiple resolutions [32, 33]. As a result, TDA allows us to capture subtler patterns in the transaction graphs, including changes in chainlet dynamics, which are often associated with illicit or malicious activity and which are inaccessible with more conventional methods based on various forms of information aggregation [22, 34]. ## 8 Limitations of Data Science on Blockchains Having covered blockchain data structures, we now turn our attention to the difficulties faced by data scientists when mining blockchain data. First of all, many exchanges create off-channel transactions and do not store them on the blockchain, and this means that many transactions are executed without being recorded on the blockchain. All of this off-chain transaction information is not directly accessible to data scientists. A further concern is to integrate, model, and query data arriving from hundreds of blockchains. The Bitcoin blockchain itself produces approximately 300K transactions per day, arriving from 400K-800K user addresses (see blockchain.com/en/charts), in a temporal resolution scale of 10 minutes. Some
# Increasing Fault Tolerance and Throughput with Adaptive Control Plane in Smart Factories Cao Vien Phung and Admela Jukan Technische Universität Braunschweig, Germany Email: {c.phung<EMAIL_ADDRESS> ###### Abstract Future smart factories are expected to deploy an emerging dynamic Virtual Reality (VR) applications with high bandwidth wireless connections in the THz communication bands, where a factory worker can follow activities through 360°video streams with high quality resolution. THz communications, while promising as a high bandwidth wireless communication technology, are however known for low fault tolerance, and are sensible to external factors. Since THz channel states are in general hard to estimate, what is needed is a system that can adaptively react to transceiver configurations in terms of coding and modulation. To this end, we propose an adaptive control plane that can help us configure the THz communication system. The control plane implements a workflow algorithm designed to adaptively choose between various coding and modulation schemes depending on THz channel states. The results show that an adaptive control plane can improve throughput and signal resolution quality, with theoretically zeroed bit error probability and a maximum achievable throughput in the scenarios analayzed. ## I Introduction Future smart factory scenarios are envisioned with the downlink applications using high bandwidth wireless links able to provide high throughput and fault tolerance. A typical scenario, as illustrated in Fig. 1a, includes several robots, a factory worker is using an augmented or virtual reality (VR) and high quality resolution device (as a receiver), and a Base Station (BS) (as a transmitter). The high data rate of communications in Gb/s is today achievable with Terahertz (THz) frequency band ($0.3-10$ THz). A system that can combine THz communications, and virtual- and augmented reality will enable several advanced features in the manufacturing process, such as the real-time placement of images onto the real working environment to see the proper instructions on how to assemble a particular product or a component, hands- free [1]. Such applications in general are expected to improve the manufacturing efficiency, decrease the training time of factory works, and enhance the safety. To guarantee high throughput and fault tolerance, the system needs to be able to adaptively respond to changed transmission and applications conditions. First, this is necessary since VR users move, leading to the outdated channel states, making static system configuration obsolete. Second, and more critically, THz band is sensitive to molecular absorption [2], leading to an easily deteriorated Signal-to-Noise Ratio (SNR), i.e., a high Bit Error Rate (BER). Thus, to provide the desired robustness, the adoption and development of the adequate coding schemes are subject of ongoing research [3]. In regard to coding, a few FEC codes with redundant data are under discussion for THz communication to protect original data against losses, such as Low Density Parity Check (LDPC), RS or Polar code [4, 5]. To increase the transmission performance, data can also be transferred with either a higher modulation formats, or over longer paths, which also is a major challenge in THz spectrum. Previous work analyzed time-varying channel modeling and tracking for THz indoor communications [6]. Papers [7] and [8] analyzed to this end different mobility scenarios in THz indoor communications. In this paper, we conceptualize the need to provide an adaptive system configuration in response to mobility and channel state changes, with the aim of improving throughput and fault tolerance. To this end we propose data and control plane architectures, akin to what is known in high speed networks, e.g., [9]. In the control plane, we design a workflow feedback loop algorithm which can adaptively update the channel states and generate adaptive coding and modulation THz system configurations. In the data plane, we consider four modulation schemes: BPSK, QPSK, 8PSK and 16QAM. For coding, the system can choose between and Multidimensional Parity Check (MDPC) [10] and Reed-Solomon (RS) codes [11]. The reasons for the above selections is because these modulations and coding schemes are well known and can be analyzed, though our approach is applicable to other schemes as they emerge. The results show an adaptive control plane can choose the best coding and modulation schemes with the lowest transmission overhead, and the highest code rate and throughput, therefore improving the system performance with theoretically zeroed bit error probability and a maximum achievable throughput in the scenarios analayzed. (a) (b) Figure 1: (a) An augmented reality smart factory application with wireless communications in THz band; (b) A THz control plane architecture with feedback loop workflow protocol. ## II System Design ### II-A THz data plane The THz data plane includes the transmitter (Tx), also referred to as THz base station, and the receiver (Rx) module, here corresponding to the VR device. In our model we assume that Rx (VR) can move, whereas Tx (BS) is static. At the Transmitter (Tx) (THz base station), the data source with a long bit stream is split into substreams. Each substream is referred to as a coding generation of $K$ bits, whereby the coding generation includes a set of bits coded and decoded together. Based on channel condition, $K$ can be adaptively chosen and all Ks are temporarily stored at the coding cache for the coding process. A coding generation of $K$ bits generates $R$ redundant bits from the coding process, where $R$ bits are used to support for correcting error bits from $K$ bits. We consider two coding schemes: RS code, and MDPC($n$D/$m$L) code with $n$ dimensions (D) and the same length (L) of $m$ bits for each dimension, whereby the original bits stay unchanged after coding, meaning that only redundant coded bits are generated by coding. The control plane can choose one of these two schemes adaptively, based on an algorithm. If MDPC($n$D/$m$L) code is chosen, the encoding process for $R=(m+1)^{n}-m^{n}$ parity bits [10] generated from $K=m^{n}$ bits can be performed as follows: A special case of MDPC($n$D/$m$L) code with n = 2 has column and row parity bits in the last row and column of the two-dimensional matrix, respectively. The original data bits are put at the remaining columns and rows of the matrix. Data bits in the column of the matrix and in the row of the matrix are secured by one column parity bit in the last row and by one row parity bit in the last column, respectively. The parity check bit on check bits for securing column and row parity bits is at the bottom right corner of the matrix. The construction of MDPC($n$D/$m$L) code with $n\geq 3$ is a multidimensional hypercube combined from basic three-dimensional parity check cubes. A parity check cube of three dimensions is a set of the two-dimensional parity matrices layered into a third dimension and the parity bits across the layers of data bits belong to the last layer. If RS code is chosen, the related coding mechanism as well known from [11] can be performed as follows: The original data $M(X)$ with the size of $K$ bits is operated on the coding process with the modulo-$g(X)$ function to create the redundant data $CK(X)$, i.e., $CK(X)=X^{\frac{R}{s}}\cdot M(X)\;mod\;g(X)$, whereby $X^{\frac{R}{s}}$ represents the displacement shift, $s$ is the coding symbol size in bits, $R$ is the size of redundant coded information bits and $g(X)$ represents the generator. Therefore, the RS code word $C(X)=X^{\frac{R}{s}}\cdot M(X)+CK(X)$ includes the redundant data $CK(X)$ appended systematically onto the original data $M(X)$, i.e., $K$ original bits are increased to $K+R$ bits during the coding process. $R+K$ bits are divided into symbols of $s$ bits each and all coding operations are performed via finite field $\mathbb{F}_{2^{s}}$. The maximum codeword length of RS code is $\frac{K+R}{s}=2^{s}-1$. If $R$ is fixed and $\frac{K+R}{s}<2^{s}-1$, then $Z$ zero padding symbols can be added into $\frac{K}{s}$ original symbols by the encoder so that the codeword length $\frac{K+R}{s}+Z=2^{s}-1$ can be achieved, and $\frac{R}{s}$ redundant coded symbols are generated. These zero padding symbols will not be sent out of transmission channels, but the decoder needs to use them for the decoding process. After the coding process, $K+R$ data bits are modulated to generate $X$ modulation symbols. The control plane can choose one of the four modulation schemes: BPSK, QPSK, 8PSK and 16QAM, depending on the THz channel state. After the modulation process, $X$ modulation symbols are sent over the THz channel. At the Receiver (Rx) side (VR device), we collect $X$ symbols to demodulate. $X$ modulation symbols are demodulated to re-create $K$ original data bits and $R$ redundant bits, for decoding. $K+R$ bits will be then delivered to the decoding block. As large amount of data arrives at the same time, a receiving buffer is necessary to temporally store all arriving data for the decoding process, whereby the decoded generations are serialized. If MDPC($n$D/$m$L) code is used, the iterative decoding method can be applied [10] to correct the data bits with the highest error probability in each decoding iteration. All parity bits and their related data bits are checked in each decoding iteration, and then we will count a failed dimension marker (FDM). The FDM of the arbitrary bit is set to $k$, if that bit discovers a parity check error in the $k$ dimension. If the FDM reaches the highest value less than $2$, then the decoding process stops. Otherwise, all bits are inverted with the highest FDM value. In the next step, the iteration number limit is checked to decide whether the decoding process stops or all steps are repeated in the next iteration, based on the predefined maximum iteration. Finally, if $FDM_{max}=0$ after the decoding process, the frame is considered error-free. Otherwise, we face an uncorrectable error. If RS code [11] is chosen, the data of RS code is decoded as follows. The syndrome components are calculated from the received word at the first stage. The second stage is to calculate the error-locator word from the syndrome components. At the third stage, the error locations are calculated from the error-locator numbers which are from the error-locator word. The fourth stage is for calculating the error values from the syndrome components and the error-locator numbers. Finally, we calculate the decoded code word from the received word, the error locations, and the error values. Figure 2: BW 8.64 GHz at center frequency 300.24 GHz. ### II-B THz Control Plane The control plane module includes the two main components referred to as Transmitter Controller and Adaptive Configuration Module. In between, a messaging protocol (feedback messaging protocol) is used for communications over a wireless channel. The latter can be either in THz or out of THz band, and is not critical in the overall design due to its low bandwidth required. Adaptive configuration module generates new system configurations when the channel state changes. This module collects the necessary measurements (e.g., BER) used for estimating the receiver performance as well as channel state information. Such measurements are not foreseen at the transmitter module, here THz BS, but can be easily added to the system. For instance, transmitter can measure the noise levels and calibrate the transmitting power as the noise increases (e.g., due to equipment aging or heat). The arrow labelled with the message "Retrieve BER value $BER_{m}$" illustrated that the measured BER value is periodically updated from the receiver module to the adaptive configuration module. The messaging arrows labelled as "Set demodulation scheme" and "Set decoding scheme: RS($K,s$)/MDPC($K$)" illustrate the actuation messaging to inform the data plane when the best demodulation and decoding schemes are chosen, respectively. This messages are sent out after the new system configurations are generated by the adaptive configuration module. It is important to note that a buffer needs to be defined sufficiently large for the system to work and store the messages about the system states. Let us define a buffer $B$ with an enough size for $N$ messages. This buffer is used to save measured BER values $BER_{m}$ (before decoding) sent by the receiver. $BER_{m}$ after buffering is denoted to be $BER_{f}$. The $BER_{f}s$ stored in the buffer are used to anticipate the status of the receiver (e.g., moving or standing idly at a certain position), or the transmission distance. For example, changed $BER_{f}s$ mean that it is likely that the receiver is moving. Consider the BER values of bandwidth $8.64$ GHz at center frequency $300.24$ GHz over different transmission distances and modulation schemes conducted using the Simulator for Mobile Networks (SiMoNe) in Fig. 2 from [12], e.g., if we have $BER_{m}$ or $BER_{f}$ is approximately equal to $0.0579$, then the transmission distance is estimated to be $d=20$ m with using BPSK, whereby $BER_{m}$ and $BER_{f}$ can be deduced for the current transmission distance and previous transmission distance, respectively. At the beginning, we store the value of $0$ in the buffer. The values $BER_{m}$ are periodically updated from the receiver module at a constant time interval $t$. We assume that a small moving speed of a pedestrian is at $1$ m/s, i.e., a footstep of worker with $0.5$ m will take $0.5$ s. So, to estimate the moving of receiver, we recommend setting the constant time interval $t=0.5$ s for periodically updating $BER_{m}$. In Fig. 2, on the other hand, the fluctuations of channel state are continuous when the receiver is moving to different transmission distances. Hence, the adaptive configuration module only generates the new configurations, when the receiver is predicted to stop at a certain position. This can be explained as follows: The channel state changes so fast, when the receiver is moving. So, the new configurations have not been able to actuate yet, and are already obsolete. Also, too frequent channel state update can cause an extremely large overhead in each transmitted frame. At the beginning, let us assume that we set up the system using RS code $(224,240)$ and the main THz channel using 16QAM, and they are known by transmitter, receiver and adaptive configuration module. ### II-C Adaptive Algorithm To generate new states, an Algorithm 1 is used. For line 1, the Selection function includes $BER_{m}$ periodically updated by the receiver module. For line 2-line 3, we compare $BER_{m}$ with the other previous BER values $BER_{f}$s stored in the buffer. If there exists at least a difference of $BER_{m}$ and any $BER_{f}$ stored in the buffer $B$, where that difference is more than or equal to $\varepsilon$, then it is likely that the receiver is moving, whereby $\varepsilon>0$. For example, based on BER values over different distances and modulation schemes in Fig. 2, if the difference between $BER_{m}$ and $BER_{f}$ is at least $\varepsilon\approx 7.646\cdot 10^{-8}$ for BPSK, $\varepsilon\approx 6.649\cdot 10^{-9}$ for QPSK, $\varepsilon\approx 9.375\cdot 10^{-7}$ for 8PSK and $\varepsilon\approx 2.992\cdot 10^{-7}$ for 16QAM for bandwidth $8.64$ GHz at center frequency $300.24$ GHz, then the receiver is predicted to be moving. When the receiver is moving, $BER_{m}$s will be continuously changed. Therefore, $BER_{f}$s previously stored in the buffer will be removed, and then the new $BER_{m}$ is added into the buffer. The reason for the deletion is that when the receiver is moving, $BER_{f}$s become outdated. If $\left|BER_{m}-BER_{f}\right|<\varepsilon$, $\forall BER_{f}\in B$ (line 4), then it is likely that the receiver stops at a certain position. However, no new configuration is generated because we are not totally sure that whether the receiver keeps walking or not, i.e., we need more $BER_{m}$ updates to give a decision of generating new configurations. If the buffer $B$ is not full (line 5), then we add the new $BER_{m}$ into it (line 6). We keep adding new $BER_{m}$s, which satisfy the condition of line 4, until the buffer is full (line 7). For line 8-10, the adaptive configuration module starts generating new adaptive configurations because highly likely the receiver stops moving. Based on $BER_{m}$ received with the current modulation scheme used by the transmitter and the BER database over different transmission distances and modulation schemes stored in this module, we can approximately predict the current transmission distance between the transmitter and receiver (line 8). Based on the current distance of THz transmission channel and the BER database over different transmission distances and modulation schemes stored in this module, we can approximately predict the bit error probability $p_{e}$, if the transmitter uses modulation scheme $e\in\\{$ BPSK, QPSK, 8PSK,16QAM $\\}$ at that transmission distance of main THz channel (line 9). Next, we discuss line 10. First of all, we consider the number of error bits $t_{MDPC}$ corrected by MDPC code and the number of error symbols $t_{RS}$ corrected by RS code: $\text{ }\left\\{\begin{matrix}t_{MDPC}=2^{n-1}-1,&\\\ t_{RS}=\frac{R}{2s}&,\end{matrix}\right.$ (1) whereby $t_{V}$ ($V\in$ {MDPC code, RS code}) is assumed to be fixed. On the other hand, we consider two parameters: Code rate $R_{F}$ and coding system throughput $TH$ defined to be the number of useful data bits transferred by the THz transmission system to a certain destination per unit of time. These two ones with formulas are expressed by: $\text{ }\left\\{\begin{matrix}R_{F}=\frac{K}{K+R},&\\\ TH=R_{F}\cdot D\cdot(1-P_{re}^{V}),&\end{matrix}\right.$ (2) where $D$ is the data transmission rate and $P_{re}^{V}$ is the residual bit error probability after decoding. With fixed value $t_{MDPC}$ (Eq. (1)) and $p_{e}$ from four modulation schemes considered from line 9, we can choose $4$ optimized values of $K=m^{n}$ original bits with $R=(m+1)^{n}-m^{n}$ redundant bits from $4$ modulation schemes for MDPC code, where $m,n\in\mathbb{N}$, so that they satisfy the eqution as: $(m+1)^{n}\cdot p_{e}\leq t_{MDPC}.$ (3) At the same time, similarly, with fixed value $t_{RS}$ (Eq. (1)) and $p_{e}$ from four modulation schemes considered from line 9, we can choose $4$ optimized values of $K$ original bits with $R$ redundant bits from $4$ modulation schemes for RS code, so that they satisfy the eqution as: $\text{ }\left\\{\begin{matrix}\frac{K+R}{s}\cdot P_{s}\leq t_{RS}\;\;;\;\;K\geq 0&\\\ 2^{s-1}\leq\frac{K}{s}+\frac{R}{s}\leq 2^{s}-1&,\end{matrix}\right.$ (4) where $P_{s}=1-(1-p_{e})^{s}$ is the symbol error probability on the main THz channel and the maximum codeword length of RS code is $\frac{K+R}{s}=2^{s}-1$. Note that the values of $K$, $R$ or $s$ chosen from Eq. (3) for MDPC code and Eq. (4) for RS code over different modulation schemes, the total number of error bits/symbols occurred is lower than or equal to the ability of correcting bit/symbol errors by MDPC($n$D/$m$L) and RS code, when $K+R$ bits or $\frac{K+R}{s}$ symbols are sent over the main THz channel with the bit error probability $p_{e}$ or symbol error probability $P_{s}$, with the aim of optimizing the fault tolerance system. Theoretically, with this approach, the residual bit error probability $P_{re}^{V}$ after decoding is approximately equal to $0$. From eight optimized values of $K$ original bits/symbols with $R$ redundant bits/symbols of MDPC and RS code over four modulation schemes, we can choose the best coding and modulation scheme with the highest code rate $R_{F}$ and throughput $TH$, based on Eq.(2). As the adaptive configurations are chosen, the THz data transmission can achieve high throughput and code rate with an optimized system fault tolerance. If $BER_{m}$ arrives when the buffer of adaptive configuration module is full and $\left|BER_{m}-BER_{f}\right|<\varepsilon$, $\forall BER_{f}\in B$ (line 12), then we only remove the oldest $BER_{f}$ stored in the buffer, and then add $BER_{m}$ into it. In this case, we do not generate any new system configurations because the receiver does not still change its position, i.e., the current channel state is similar to the previous one with the adaptive configurations already generated from line 7 to line 10. Note that if MDPC code is chosen, this module generates the configuration: MDPC($K$), whereby $K$ is the size of original input data. Else if, RS code is chosen, this module generates the configuration: RS($K,s$), whereby $s$ is the size of coding symbol in bits. #### II-C1 Analysis of algorithm complexity In this subsection, we calulate the overhead for generating a new adaptive configuration, when a $BER_{m}$ is updated at the adaptive configuration module. As analyzed from line 5-line 10 in Algorithm 1, this module only generates new configurations, if the new $BER_{m}$ arrives at the time when the buffer $B$ stores $N-1$ $BER_{f}$ messages. Assume that a computation for the process of generating adaptive configurations is counted as one unit and we ignore the simple processes of removing or storing from the bufffer. We need $N-1$ units to compare $BER_{m}$ and $BER_{f}$ at line 2. At line 8, we need one unit for predicting the transmission distance of VR user. At line 9, we need $M=4$ units for predicting the bit error probability for four modulation schemes. Finally, for each coding scheme, we need $M\cdot 3=12$ units at line 10 because we have $M=4$ values of bit error probability $p_{e}$ applied for $3$ equations (Eqs. (2), (3), (4)) to find the best coding and throughput with the highest code rate and throughput. As a result, in general, we need $N+M(1+3C)$ units for generating a new system configuration, where $M$ is the number of modulation schemes considered and $C$ is the number of coding schemes considered. Algorithm 1 Adaptive Workflow Configuration Algorithm 1:function selection($BER_{m}$) 2: if $\exists$ $\left|BER_{m}-BER_{f}\right|\geq\varepsilon$, $\forall BER_{f}\in B$ then 3: Remove all $BER_{f}$s, and buffering $BER_{m}$; 4: else 5: if $B$ is not full then 6: Add $BER_{m}$ into the buffer; 7: if $B$ is full then 8: Based on $BER_{m}$ $\rightarrow$ transmission distance. 9: Based on transmission distance $\rightarrow$ bit error probabilities $p_{e}$, if using BPSK, QPSK, 8PSK or 16QAM. 10: Using $p_{e}$ from four modulation schemes, $t_{V}$ fixed and Eqs. (2), (3), (4) $\rightarrow$ chosen coding and modulation scheme with the highest $R_{F}$ and $TH$. 11: end if 12: else 13: Remove the oldest $BER_{f}$, and buffer $BER_{m}$; 14: end if 15: end if 16:end function (a) Transmission distance. (b) Code rate. (c) Transmission overhead. Figure 3: BER vs. distance without coding, transmission distance randomly chosen, code rate and transmission overhead. (a) Chosen coding scheme. (b) Chosen modulation scheme. (c) Throughput. Figure 4: Theoretical results. For chosen coding scheme, $1$ and $2$ denote chosen RS code and MDPC code, respectively. For chosen modulation scheme, $4,5,6$ and $7$ represent chosen BPSK, QPSK, 8PSK and 16QAM, respectively.. ## III Performance Evaluation In this section, we show the performance results in the smart factory applications shown, considering THz data and control planes proposed. All data including over the THz channel, which has bandwidth $8.64$ GHz at center frequency $300.24$ GHz. For the BER values, we make use of the data shown in Fig. (2). These BER values are used to estimate the transmission distance between VR user and THz BS and bit error probability $p_{e}$ over different modulations under the Algorithm 1. To estimate the moving of VR user, we assume setting the constant time interval $t=0.5$ s for periodically updating $BER_{m}$ from VR user to the adaptive configuration module. Note that at the beginning, we store temporally the value $0$ in the buffer and use RS code $(224,240)$ with the modulation scheme of 16QAM. The adaptive configuration module only generates the new configurations, when VR user stops at a certain position. As discussed in Algorithm 1, based on the database of BER values from Fig. 2, we define $\varepsilon\approx 7.646\cdot 10^{-8}$ for BPSK, $\varepsilon\approx 6.649\cdot 10^{-9}$ for QPSK, $\varepsilon\approx 9.375\cdot 10^{-7}$ for 8PSK and $\varepsilon\approx 2.992\cdot 10^{-7}$ for 16QAM to predict whether the VR user is moving or stops at a certain place. The data transmission rate $D$ for BPSK, QPSK, 8PSK and 16QAM is $7.04$ Gbps, $14.08$ Gbps, $21.12$ Gbps and $28.16$ Gbps, respectively. Table I: List of main values applied for evaluation. Values | Meaning ---|--- $t_{MDPC}=1$ | Number of bits can be corrected by MDPC code. $t_{RS}=1$ | Number of symbols can be corrected by RS code. $N=4$ | Maximum number of messages stored in buffer $B$. $t=0.5$ | Time for periodically updating $BER_{m}$ in second. $\varepsilon\approx 7.646\cdot 10^{-8}$ | Likely VR user is moving if using BPSK. $\varepsilon\approx 6.649\cdot 10^{-9}$ | Likely VR user is moving if using QPSK. $\varepsilon\approx 9.375\cdot 10^{-7}$ | Likely VR user is moving if using 8PSK. $\varepsilon\approx 2.992\cdot 10^{-7}$ | Likely VR user is moving if using 16QAM. $D=7.04$ | Data transmission rate for BPSK in Gbps. $D=14.08$ | Data transmission rate for QPSK in Gbps. $D=21.12$ | Data transmission rate for 8PSK in Gbps. $D=28.16$ | Data transmission rate for 16QAM in Gbps. Fig. 3a shows the different standing positions (transmission distances) of VR user around $60600$ s, where they are randomly chosen between $0.5$ m and $20.0$ m in steps of $0.5$ m from the transmitter. When VR user moves between two positions, the channel state changes. Assume that based on $BER_{m}$ periodically updated from the VR user (line 8 in Algorithm 1). Based on the channel states in Fig. 3a, our adaptive algorithm can choose the best coding scheme (Fig. 4a) and modulation scheme (Fig. 4b) with the lowest transmission overhead (Fig. 3c), the highest code rate (Fig. 3b) and throughput (Fig. 4c). Assume that the time when the VR user is idly standing at any position is randomly chosen in $[3,4,5,6,7]$ minutes. We assume $R$ redundant bits are generated from $K$ original bits so that the number of bits can be corrected to be $t_{MDPC}=1$ bit for MDPC code and to be $t_{RS}=1$ symbol for RS code. The reason for this assumption is because with a low redundant coding data, the coding and decoding time is optimized with a low latency, while it can still satisfy the optimized fault tolerance with high code rate and throughput, as discussed in section II-C. Assume that the processing time for the feedback controller and adaptive configuration module is small and ignored. Fig. 3b, Fig. 3c, Fig. 4a, Fig. 4b and Fig. 4c theoretically show the code rate, transmission overhead, chosen coding and modulation scheme and throughput, respectively, where the throughput is observed in the time interval, when VR user idly stands at a specific position. $1$ and $2$ denote the chosen coding scheme, respectively. $4$, $5$, $6$ and $7$ denote the chosen modulation scheme of BPSK, QPSK, 8PSK and 16QAM, respectively. Based on the standing positions of VR user predicted in Fig. 3a (line 8 in Algorithm 1), we can approximately estimate four bit error probabilities $p_{e}$ from the database of BER values over different transmission distances and modulation schemes in Fig. 2, stored at the adaptive configuration module, if THz BS uses BPSK, QPSK, 8PSK or 16QAM (line 9 in Algorithm 1). With four bit error probabilities from four modulation schemes if used by THz BS (line 9 in Algorithm 1), $t_{MDPC}=1$ bit and $t_{RS}=1$ symbol, we will find four values of $K$ for MDPC code satisfying Eq. (3) and four values of $K$ and $s$ for RS code satisfying Eq. (4). Theoretically, if we can choose a reasonable value of input data $K$ (section II-C), then the residual bit error probability $P_{re}^{V}\approx 0$ after decoding. Therefore, the throughput is equivalent to the information bit rate, i.e., based in Eq. (2), $TH=R_{F}\cdot D$. With $R$ redundant bits deduced from $t_{MDPC}=1$ bit and $t_{RS}=1$ symbol by Eq. (1), and eight values of $K$ are previouly found, all of them are applied from Eq. (2). As a result, the adaptive configuration module will find the best coding scheme (Fig. 4a) and modulation scheme (Fig. 4b) with the highest code rate (Fig. 3b) and throughput (Fig. (4c)). The transmission overhead in Fig. (3c) is given by: $\theta=1-R_{F}$, where $R_{F}$ is code rate. For code rate in Fig. 3b, $0.25\leq R_{F}\leq 1$, i.e., $0\leq\theta\leq 0.75$ in Fig. 3c; the longer the transmission distance, the lower the code rate For the chosen coding scheme in Fig. 4a, RS code performs better than MDPC. For the chosen modulation scheme in Fig. 4b, 16QAM is the best, while BPSK is the lowest selection because its data transmission rate is the lowest. We see that no modulation scheme 8PSK is chosen in our scenario. This can be explained since BER of 8PSK and 16QAM are quite similar, while the data transmission rate of 16QAM is higher than that of 8PSK. Therefore Algorithm 1 always chooses the coding and modulation scheme with the code rate and throughput as the highest. As discussed above, with the suitable code rates or transmission overheads chosen from the best coding and modulation schemes in Fig. 3b and Fig. 3c over different channel states, we can theoretically improve the system fault tolerance with the residual bit error probability $P_{re}^{V}\approx 0$. Also, since $P_{re}^{V}\approx 0$, the throughput can achieve the information bit rate $TH=R_{F}\cdot D$, and due to the flexibility in choosing the best coding and modulation schemes with the highest code rate, sometime we can theoretically get a maximum throughput of $28.16$ Gbps in Fig. 4c, equivalent to data transmission rate of 16QAM. ## IV Conclusion We proposed a THz data and control plane system architecture for VR applications in smart factories. The results shows that an adaptive design system can improve throughput and fault tolerance, whereby the residual bit error probability is $P_{re}^{V}\approx 0$ after decoding and we get can theoretically get a maximum throughput of $28.16$ Gbps, equivalent to data transmission rate of 16QAM. Future work needs to address algorithm computation offloading and its response time. ## Acknowledgment The authors acknowledge the financial support by the Federal Ministry of Education and Research, Germany, program "Souveran. Digital. Vernetzt." project 6G-RIC, 16KISK031 and partial support by the DFG Project Nr. JU2757/12-1. ## References * [1] Boeing, “Boeing tests augmented reality in the factory.” [Online]. Available: https://www.boeing.com/features/2018/01/augmented-reality-01-18.page * [2] F. Moshir and S. Singh, “Rate adaptation for terahertz communications,” in _2016 13th IEEE Annual Consumer Communications Networking Conference (CCNC)_ , Jan 2016, pp. 816–819. * [3] N. Wehn and O. Sahin, “Next-Generation Channel Coding Towards Terabit/S Wireless Communications,” Aug. 2018. * [4] _Next-Generation Channel Coding Towards Terabit/s Wireless Communications_. Zenodo, Aug. 2018. [Online]. Available: https://doi.org/10.5281/zenodo.1346686 * [5] N. Wehn, “Channel coding for tb/s communications,” Aug. 2019. [Online]. Available: https://doi.org/10.5281/zenodo.3360520 * [6] S. Nie and I. F. Akyildiz, “Three-dimensional dynamic channel modeling and tracking for terahertz band indoor communications,” in _2017 IEEE 28th Annual International Symposium on Personal, Indoor, and Mobile Radio Communications (PIMRC)_ , 2017, pp. 1–5. * [7] R. Singh and D. Sicker, “Parameter modeling for small-scale mobility in indoor thz communication,” in _2019 IEEE Global Communications Conference (GLOBECOM)_ , 2019, pp. 1–6. * [8] B. Peng, K. Guan, and T. Kürner, “Cooperative dynamic angle of arrival estimation considering space-time correlations for terahertz communications,” _IEEE Trans. Wirel. Commun._ , vol. 17, no. 9, pp. 6029–6041, 2018. * [9] M. Chamania and A. Jukan, _Dynamic Control of Optical Networks_. Cham: Springer International Publishing, 2020, pp. 535–552. [Online]. Available: https://doi.org/10.1007/978-3-030-16250-4_15 * [10] I. Vertat and L. Dudacek, “Multidimensional cross parity check codes as a promising solution to cubesat low data rate downlinks,” in _2019 International Conference on Applied Electronics (AE)_ , 2019, pp. 1–5. * [11] I. S. Reed and X. Chen, _Reed-Solomon Codes_. Boston, MA: Springer US, 1999, pp. 233–284. * [12] B. K. Jung, C. Herold, J. M. Eckhardt, and T. Kürner, “Link-level and system-level simulation of 300 ghz wireless backhaul links,” in _2020 International Symposium on Antennas and Propagation (ISAP)_ , 2021, pp. 619–620.
# M Subdwarf Research<EMAIL_ADDRESS>Atmospheric Parameters and Kinematics Shuo Zhang CAS Key Laboratory of Optical Astronomy, National Astronomical Observatories, Beijing 100101, China University of Chinese Academy of Sciences, Beijing 100049, China Department of Astronomy, School of Physics, Peking University, Beijing 100871, P. R. China Kavli institute of Astronomy and Astrophysics, Peking University, Beijing 100871, P. R. China A-Li Luo CAS Key Laboratory of Optical Astronomy, National Astronomical Observatories, Beijing 100101, China University of Chinese Academy of Sciences, Beijing 100049, China School of Information Management & Institute for Astronomical Science, Dezhou University, Dezhou 253023, China Department of Physics and Astronomy, University of Delaware, Newark, DE 19716, USA Georges Comte Aix- Marseille Univ, CNRS, CNES, LAM, Laboratoire d’Astrophysique de Marseille, Marseille, France Rui Wang CAS Key Laboratory of Optical Astronomy, National Astronomical Observatories, Beijing 100101, China University of Chinese Academy of Sciences, Beijing 100049, China Yinbi Li CAS Key Laboratory of Optical Astronomy, National Astronomical Observatories, Beijing 100101, China Bing Du CAS Key Laboratory of Optical Astronomy, National Astronomical Observatories, Beijing 100101, China Wen Hou CAS Key Laboratory of Optical Astronomy, National Astronomical Observatories, Beijing 100101, China Li Qin CAS Key Laboratory of Optical Astronomy, National Astronomical Observatories, Beijing 100101, China School of Information Management & Institute for Astronomical Science, Dezhou University, Dezhou 253023, China John Gizis Department of Physics and Astronomy, University of Delaware, Newark, DE 19716, USA Jian-Jun Chen CAS Key Laboratory of Optical Astronomy, National Astronomical Observatories, Beijing 100101, China Xiang-Lei Chen CAS Key Laboratory of Optical Astronomy, National Astronomical Observatories, Beijing 100101, China University of Chinese Academy of Sciences, Beijing 100049, China Yan Lu CAS Key Laboratory of Optical Astronomy, National Astronomical Observatories, Beijing 100101, China School of Information Management & Institute for Astronomical Science, Dezhou University, Dezhou 253023, China Yi-Han Song CAS Key Laboratory of Optical Astronomy, National Astronomical Observatories, Beijing 100101, China Hua-Wei Zhang Department of Astronomy, School of Physics, Peking University, Beijing 100871, P. R. China Kavli institute of Astronomy and Astrophysics, Peking University, Beijing 100871, P. R. China Fang Zuo CAS Key Laboratory of Optical Astronomy, National Astronomical Observatories, Beijing 100101, China University of Chinese Academy of Sciences, Beijing 100049, China ###### Abstract Applying the revised M subdwarf classification criteria discussed in Paper i@ to LAMOST DR7, combining the M subdwarf sample from Savcheva et al, a new M subdwarf sample was constructed for further study. The atmospheric parameters for each object were derived fitting with the PHOENIX grid, combining with Gaia DR2, the relationship between the gravity and metallicity were explored according to the locus both in the color-absolute magnitude diagram and the reduced proper motion diagram. Objects that have both the largest gravity and the lowest metallicity are located away from the main-sequence cloud and may be considered as the intrinsic M subdwarfs, which can be classified as luminosity class<EMAIL_ADDRESS>Another group of objects whose spectra show typical M subdwarf characters have lower gravity and relatively moderate metal deficiency and occupy part of the ordinary M dwarf region in both diagrams. The Galactic $U$, $V$, $W$ space velocity components and their dispersion show that the local Galactic halo population sampled in the solar neighborhood is represented by objects of high gravity and an inconspicuous bimodal metallicity distribution, with a fraction of prograde orbits. The other M subdwarfs seem to partly belong to the thick disk component with a significant fraction of thin disk moderately metal-poor objects intricately mixed with them. However, the selection effects, especially the favored anti-center direction of investigation in the LAMOST sub-sample, but also contamination by multiplicity and parameter coupling could play important roles and need to be further investigated. Stellar astronomy — Stellar types — Late-type stars— M stars, Stellar kinematics, Atmospheric parameters ††journal: ApJ††software: astropy (Astropy Collaboration et al., 2013), matplotlib (Barrett et al., 2005), pandas(McKinney et al., 2010), numpy (Oliphant, 2007), scipy (Virtanen et al., 2019) ## 1 Introduction M subdwarfs are usually recognized as the oldest members of the low-mass stellar populations and metal-poor counterparts of the late-type M dwarfs. With their extremely long nuclear-burning lifetimes, subdwarfs are supposed to represent early generations of star formation and are thus potential tracers of Galactic dynamic and chemical evolution history (Savcheva et al., 2014). Despite their scarcity in the solar neighborhood, M subdwarfs are supposed to comprise the majority of stars in the Milky Way stellar halo (Bochanski et al., 2013), while a number of them also populate the thick disk, whose population is supposed to be older than that of the thin disk. Additionally, the cool atmospheres of M subdwarfs provide conditions for studying molecule and dust formation as well as radiative transfer in metal-poor environments (Allard et al., 1997). In many early investigations, M subdwarfs are considered as metal-poor main- sequence dwarfs (e.g. Lépine et al. 2007; Rajpurohit et al. 2014) solely with some noteworthy kinematic and spectral properties: red subdwarfs, exhibiting large space velocities relative to the Sun, kinematically associated with the Galactic halo (Gizis, 1997) and thick disk, display shallower TiO absorption bands in the optical spectra than the ordinary red dwarfs dominating the Galactic thin disk. Historically, they were usually searched based on their low luminosity combined with large proper motion (Sandage, 1969; Jones, 1972; Hartwick et al., 1984; Sandage & Kowal, 1986; Gizis, 1997; Lépine et al., 2003, 2007; Smith et al., 2009; Jao et al., 2011) and characteristic spectral features (Mould, 1976; Mould & McElroy, 1978; Ake & Greenstein, 1980; Lépine & Scholz, 2008; Savcheva et al., 2014; Bai et al., 2016; Zhang et al., 2019; Hejazi et al., 2020). The research of metal-poor subdwarfs also extends to the late M-type and L-type in recent years (e.g. Zhang et al. 2017b, a, 2018; Zhang 2019). On the other hand, some studies have suggested that subdwarfs should have a luminosity classification independent of dwarfs—vi@ because of 1$\sim$2 magnitude less luminous than the main sequence dwarfs in the H-R diagram, forming an independent sequence, and more or less showing larger surface gravity than the dwarfs (Jao et al., 2008). Results from Kesseli et al. (2019) show that ultra subdwarfs can be as much as five times smaller than their solar-metallicity counterparts for a given temperature. However, subdwarf identification in most of the researches depends much more on the metallicity rather than the gravity, and whether gravity constrain for subdwarfs is necessary still remains an open question. Characterizing M subdwarfs in terms of atmospheric parameters has long been a challenge. Like M dwarfs, the opacity sources are dominated by the molecular absorption bands such as TiO, VO, MgH, FeH and CaH in the optical, as well as H${}_{\text{2}}$O and CO in the infrared, each of which has millions of spectral lines. These complex band structures leave no window to access the true continuum and only allow the strongest atomic lines such as Cai@, Nai@, Ki@ to show out at low spectral resolution. In addition, dust clouds form with decreasing temperature, the stellar atmosphere becomes more complicated, and the spectra show increasingly diatomic and triatomic molecules (Rajpurohit et al., 2016). Thus, the traditional technique of estimating the effective temperature ($T_{\text{eff}}$) of a star based on black body approximation and broadband photometry is not applicable to cool M subdwarfs in which the true continuum is covered by a wide variety of molecular absorbers(Allard et al., 1997). It is also difficult to disentangle the effect of reduced metal abundance from that of the increasing surface gravity or the temperature decrease because these three parameters affect the pressure structure of the photosphere in a similar way (Jao et al., 2008). Since accurate metallicities for most M subdwarfs are difficult to measure, proxies for metallicity have been introduced. The most widely used metallicity-tracer parameter $\zeta$, (a parameter derived from CaH2, CaH3, and TiO5 molecular bandhead indices originally used by Gizis 1997), was introduced by Lépine et al. (2007) and revised by Dhital et al. (2012) and Zhang et al. (2019). With the parameter $\zeta$ or revised definition, two large samples of M subdwarfs are spectroscopically identified by Savcheva et al. (2014) and Zhang et al. (2019) using Sloan Digital Sky Survey (SDSS; York et al. 2000) and the Large Sky Area Multi-Object Fiber Spectroscopic Telescope (LAMOST; Cui et al. 2012) respectively. According to the decreasing order of $\zeta$, the M subdwarfs can be divided into three metallicity subclasses: M subdwarfs (sdM), extreme metal-poor M subdwarfs (esdM), and ultra metal-poor M subdwarfs (usdM). The first model grid based on the PHOENIX radiative transfer code (Allard & Hauschildt, 1995) has been used by several authors to compare synthetic spectra to observed ones and to the broad-band colors: Dahn et al. (1995); Leggett et al. (1996); Jones et al. (1996). Gizis (1997) was thus able to constrain the metallicity scale of subdwarfs from $-$2 dex for esdM to $-$1 to $-$1.5 for sdM, while the stars not classified as subdwarfs were spanning the range $-$1 dex to 0. However, this pioneering model grid was far from being optimized because of the lack of theoretical molecular opacities, especially for the very important species CaH. The NextGen model grid (Hauschildt et al., 1999) brought a number of improvements and was in turn used by e.g. Woolf & Wallerstein (2005) who demonstrated the $\alpha$ enrichment of halo metal-poor M stars. Several papers using NextGen models to interpret spectra of cool M dwarfs and subdwarfs were subsequently published (e.g.Leggett et al. 2000; Lépine et al. 2004) while Burgasser & Kirkpatrick (2006) compared the performances of NextGen models and AmesCond models on the spectrum of an esdM star. Progress in modeling led the PHOENIX-based computations to the BT-Settl series of grids, the last generation of which proved highly successful in reproducing the features of the observed spectra in detail. Comparisons of low mass stars between a few high-resolution spectroscopic spectra and theoretical synthetic ones show that the latest PHOENIX models have already reached a milestone.(Rajpurohit et al., 2014, 2018a, 2018b). Gaia Data Release 2 (Gaia Collaboration et al., 2018b) has provided astrometry and photometry on more than one billion stars. As a result, a large number of subdwarfs that were out of the reach of parallax studies now have accurate available trigonometric parallaxes permitting distance determination and hence their placement in the Hertzsprung-Russell diagram, while their 3D kinematics can be studied thanks to high accuracy proper motions from Gaia and radial velocity from spectroscopy. M subdwarfs have been found both at high Galactic latitudes, as expected for halo population members, and at much lower latitudes in fields located towards the Galactic anti-center, suggesting that one could use them to study the mix of various population components in the Solar neighborhood. It becomes thus especially interesting to explore the links between their atmospheric parameters, especially gravity and metallicity, and their kinematic properties. For this, we need a reasonably accurate estimation of the atmospheric parameters of a large sample of subdwarfs for which we have accurate photometry and astrometry and the same data on a control sample of ordinary M dwarfs. The present work, based only on low-resolution spectroscopy, has its limitations and must be considered as a preliminary attempt towards this goal. The paper is organized as follows. In Section 2, we briefly summarize the observational material that will be used in Sections 4 and 5. In Section 3, we analyze the properties of the PHOENIX-based BT-Settl CIFIST model grid to evaluate its performance by studying spectral indices of BT-Settl synthetic spectra compared with those measured on real-world template spectra of M dwarfs and subdwarfs. Then, we fit the template spectra with the synthetic grid to get the atmospheric parameter ranges for each template spectra. In section 4, we derive the atmospheric parameters for spectra of a sample of M dwarf and subdwarf from the best-fit PHOENIX BT-Settl synthetic spectra. This leads to the construction of a cleaned genuine subdwarf sample, a normal dwarf sample and a third population whose properties point towards non-genuine subdwarfs but rather metal-deficient dwarfs. In section 5, we explore the links between the distribution of the atmospheric parameters of the stars belonging to these three samples and their photometric and kinematic properties as provided by the Gaia Data Release 2, using color-absolute magnitude diagram, reduced proper motion diagram and 3D Galactic motion distributions. Finally, conclusions and summary are provided in Section 6. ## 2 Observational data The low-resolution spectra of M dwarfs and subdwarfs are from LAMOST DR7 and the SDSS DR7, and the astrometric data is from Gaia DR2. LAMOST is a reflecting Schmidt telescope located in Xinglong Station of National Astronomical Observatory, China (40∘N, 105∘E) with a mean aperture of 4.3 meters and a field of view of 5∘. The observable sky area extends from $-$10∘ to +90∘ declination. 4000 optical fibers positioned on the focal plane yield a high spectrum acquisition rate per night. Until March 2020, DR7 published more than 10 million low-resolution spectra (R$\sim$1800) covering 3800$-$9000$\rm\AA$ (Luo et al., 2015) and 2.17 million medium resolution spectra (R$\sim$7500) covering 3700$-$5900$\rm\AA$ and 5700$-$9000$\rm\AA$ (Liu et al., 2020). The SDSS began regular survey operations in 2000, and has progressed through several phases until now: SDSS-i@ (2000$-$2005), SDSS-ii@ (2005$-$2008), SDSS- iii@ (2008$-$2014), and SDSS-iv@ (2014$-$2020). The SDSS DR7 contains over 1.6 million low-resolution spectra in total, including 930,000 galaxies, 120,000 quasars, and 460,000 stars. It was collected by a dedicated wide-field 2.5 m telescope at Apache Point Observatory (Gunn et al., 2006). The telescope employed a drift-scan technique, imaging the sky in $u$,$g$,$r$,$i$,$z$ wide bands along five camera columns (Gunn et al., 1998). The fiber-fed spectrographs acquired 640 spectra (R$\sim$2000) simultaneously, and the spectral range is 3800$-$9200 $\rm\AA$. The proper motions, trigonometric parallaxes, and photometry used in this work come from Gaia Data Release 2 made available in April 2018 by European Space Agency. Gaia DR2 contains the position and brightness information of more than 1.69 billion stars, the parallax and proper-motion measurements for 1.33 billion stars (Gaia Collaboration et al., 2018a; Lindegren et al., 2018; Luri et al., 2018), the color information for 1.38 billion stars (Gaia Collaboration et al., 2018a), the radial velocities for more than 7 million stars (Katz et al., 2019), etc. The mean parallax error is up to 0.7 mas for sources having G=20 mag, and the proper motion uncertainty is up to 0.2 and 1.2 mas yr-1 for G = 17 mag and G = 20 mag respectively. ## 3 PHOENIX BT-Settl Model Analysis ### 3.1 PHOENIX BT-Settl Model Grid Classical theoretical model calculation codes include ATLAS code from Kurucz (1973) and Castelli & Kurucz (2004), MARCS code from Gustafsson et al. (1975, 2008), PHOENIX code from Allard & Hauschildt (1995), and so on. In the atmosphere of very low mass stars, various molecular absorption (each with hundreds of thousands to millions of spectral lines) and the presence of numerous condensates make accurate modeling extremely complicated; the convection zone extends to the most outer photosphere layer, which means that the evolution of the model strictly depends on the precise treatment of surface boundary (Allard et al., 2013). It is worth celebrating that in recent years, the PHOENIX code differs from previous methods by computing the opacities during the model execution (or on-the-fly), considering non-local thermodynamic equilibrium effects, etc., and combining with the latest solar abundance (Asplund et al., 2009; Caffau et al., 2011). The atmospheric model, especially for modeling very low mass stars such as M-type stars, brown dwarfs, and even Jupiter-like planets, has made breakthrough progress. The latest version PHOENIX BT-Settl models represent decisive progress compared to the previous model grids (Allard et al., 2001) in different aspects, including updated molecular line lists, better solar abundance estimates and treatment of convection, though residual incompleteness of opacities and a limitation of the mixing length theory formalism still exist. Further studies on a complete, comprehensive and uniform grid of models are addressing these shortcomings and will better reproduce observational constraints (Allard, 2016). The synthetic spectra generated by BT-Settl model atmospheres (Allard et al., 2012, 2013; Baraffe et al., 2015) have been compared with observed spectra and their consistency verified in a number of recent M dwarf/subdwarf studies (e.g., Rajpurohit et al. 2014, 2016, 2018a, 2018b). In this work, we use the version BT-Settl CIFIST2011 model atmosphere which is valid across the entire parameter range (Allard et al., 2011, 2012). A part of pre-calculated grid as list in 1 is used in the analysis and the following parameter measurement process. The models are calculated with effective temperatures from $T_{\text{eff}}$ = 400 to 8000 K in 100 K steps, surface grativity ranging from log $g$ = 0.0 to 6.0 in steps of 0.5 dex, and metallicity from [M/H]=$-$2.5 to +0.5 in steps of 0.5 dex (M stands for all elements heavier than H and He; assuming [X/H] = [Fe/H] for most elements and so [Fe/H] denotes the overall metallicity). Alpha element (O, Ne, Mg, Si, S, Ar, Ca, and Ti) enhancement is taken into account as follows: for [M/H] = 0.0, no enhancement; for [M/H]=$-$0.5, a value of [$\alpha$/M]=+0.2 has been used; for [M/H] $\leq-$1.0, [$\alpha$/M]=+0.4. Note that in PHOENIX the [$\alpha$/M] is similar to [$\alpha$/Fe] since it is computed as the change in dex of the abundance of alpha elements relative to the non-alpha average, and the heaviest is Fe. Table 1: Parameter Space of the Grid Used in this Work Variable | Range | Step size ---|---|--- T${}_{\text{eff}}$ | 2700 $-$ 4500 | 100 K log g | 0 $-$ 6.0 | 0.5 dex $[\text{M/H}]$ | $-$2.5 $-$ +0.5 | 0.5 dex $[\alpha/\text{M}]$ | 0 $-$ +0.4 | 0.2 dex ### 3.2 Comparison of Models with Real-world Template Spectra in the “canonical” diagram In this section, we firstly compare M-type template spectra from Zhong et al. (2015) with BT-Settl models. The templates, assembled from SDSS low-resolution (R$\sim$2000) spectra and classified using Lépine et al. (2007, 2013) calibration, are spanning the spectral subtypes K7.0 to M8.5 and covering every half-subtype across the range. The relevant pre-calculated synthetic spectra111https://phoenix.ens- lyon.fr/Grids/BT-Settl/CIFIST2011/SPECTRA/ generated by PHOENIX code have been convolved down to the spectral resolution of SDSS, using a series of Gaussian kernels with parameters sigma and corresponding FWHM listed in Table 2. To determine the sigma value of a convolution kernel, we fit gaussians profiles to night sky lines (assuming they have a natural width much smaller than the spectral psf). In the red range, clean lines of OH molecule are numerous, in the blue the 5577$\rm\AA$ is also good, as well as some mercury lines, except 5461$\rm\AA$ which has fine structure. Na D lines of city lights are avoided because they have a very large width (high pressure lamps). Table 2: The gaussian kernels with the following sigmas and corresponding fwhm used for convolving synthetic spectra. Central $\rm\lambda(\rm\AA)$ | $\rm\sigma$ | Pixels of 0.5 $\rm\AA$ | FWHM ---|---|---|--- 4200 | 1.05 | 2.10 | 4.935 5100 | 1.08 | 2.16 | 5.075 6000 | 1.25 | 2.50 | 5.875 6900 | 1.43 | 2.86 | 6.720 7800 | 1.40 | 2.80 | 6.580 8700 | 1.39 | 2.78 | 6.535 Figure 1: The “canonical” diagram classically used to identify subdwarfs from dwarfs: spectral composite index CaH2+CaH3 versus TiO5. Models are plotted with the following codes : continuous lines: log $g$ = 4.5 models, dashed lines: log $g$ = 5.0 models, and dotted lines: log $g$ = 5.5 models. Black : solar metallicity, blue: $-$1.0 dex, and magenta: $-$2.0 dex. The small 2-digit labels indicate $T_{\text{eff}}$ points (e.g. “32” means 3200 K). Observed template spectra: black dots represent normal main-sequence dwarfs, subdwarf templates colored in blue are for “sdM” subclass, green for “esdM” and magenta for “usdM” subclass. The M0v@ to M7v@ “standard dwarf” spectral type sequence is labeled on all graphs. Figure 1 is the “canonical” [TiO5, CaH2+CaH3] diagram classically used by many authors to identify and spectroscopically separate the subdwarfs from the normal dwarfs. The spectral indices were defined initially by Reid et al. (1995). The boundaries of the d/sd/esd/usd metallicity classification scheme are also plotted. This diagram may represent more or less a metallicity- gravity diagram, since TiO5 is a metallicity (and alpha) indicator while CaH2+CaH3 is rather a gravity indicator, and both are naturally strongly $T_{\text{eff}}$ dependent. The subdwarfs are supposed to order along temperature sequences, while the fan-shape of their distribution reflects their atmospheric metal deficiencies. The normal dwarf template sequence is well-ordered and fits quite well with the model family {log $g$ = 4.5, [Fe/H] = 0.0} track, although with the coolest types (M6$\sim$M7) being a bit offset, as it would be expected from an insufficient value of the model gravity.This is an important point and needs to be further discussed. All late-type M dwarfs are expected to have higher surface gravity than early M dwarfs due to the near 1-to-1 mass radius relation for low-mass stars. Note that between metallicities of 0.0 and $-$1.0, the displacement of the model track is quite small even at the lowest $T_{\text{eff}}$, making the estimates of moderate metal deficiencies rather difficult on this diagram. The subdwarf templates also draw quite well-defined and well separated sequences with increasing slopes from sd to usd, the latter one being almost vertical at a constant TiO5 index of almost null absorption. The coolest types of dwarfs and sd are somehow confused, this questions the validity of the templates themselves which are built from spectra either in insufficient number or too noisy to be really representative of the genuine subtype definition. The metallicity of the sd subclass does not comply with a $-$1.0 value, it must be more deficient, or affected by a higher gravity, or both. The model family {log $g$ = 5.0, [Fe/H] = $-$2.0} could provide a reasonable fit for usdM0$\sim$3, but the track moves towards significant values of the TiO5 opacity for later subtypes, thus the real stars must be more deficient. The model family {log $g$ = 5.5, [Fe/H] = $-$2.0} track yields more or less an outer envelope for the subdwarfs, implying that most of the right-hand distribution of their template representative points is driven more by large metal deficiencies than by high gravity. ### 3.3 Fitting the dM and sdM Templates with the Model Using template spectra rather than hundreds of individual spectra to compare with the models enables to smooth accidental differences due to noise and uncertainties in individual classification, and in principle gives access to high signal-to-noise data. However, the final quality of a template depends on the number of individual spectra co-added to build it. We shall hereafter use the COMPLETE collection of templates of Zhong et al. (2015), i.e. the interpolated 221 spectra in total spanning 12 subclasses (from dMr to usdMp) and 18 subtypes (from M0 to M8.5). The parameter grid of the synthetic spectra is shown in Table 1. Note that the grid is not uniform: the value log $g$ = 6.0 only pertains for $T_{\text{eff}}\geq$ 3000 K, and the [$\alpha$/M] ratio is dependent on the metallicity. The spectral fitting region extends from 6000$\rm\AA$ to 8800$\rm\AA$ with a step size of 1$\rm\AA$ which covers most prominent spectral features used for classification. Before the fitting process, the synthetic spectra were smoothed at the resolution of the observed spectra. In the fitting process, we set the spectral flux as a function of $T_{\text{eff}}$, log $g$, and [M/H], and assume: $F_{\text{o},\lambda}=F_{\text{m},\lambda}\times P_{\text{n},\lambda}+\epsilon_{\lambda},$ (1) where $F_{\text{o},\lambda}$ is the flux of the observed spectrum at wavelength $\lambda$, $F_{\text{m},\lambda}$ is the flux of the synthetic spectrum generated by PHOENIX model wavelength $\lambda$, P${}_{\text{n},\lambda}$ is an n-order polynomial factor used to correct the inaccuracies of the continuum of the observed spectrum (here we set n = 4 based on the optimal experimental results), $\epsilon_{\lambda}$ is the difference between the synthetic spectrum. The equation above can be solved by minimizing the loss function L=$\sum_{\lambda=1}^{\text{N}}(F_{\text{m},\lambda}\times$ P${}_{\text{n},\lambda}-F_{\text{o},\lambda})^{2}$ using the least square method to obtain the coefficient matrix solution of P${}_{\text{n},\lambda}$. Then we obtain the best-fit synthetic spectrum by minimizing the $\chi^{2}$ distance, $\chi^{2}\ =\ \frac{\sum_{i\text{=1}}^{N}[(F_{\text{o}_{i}}-F^{\prime}_{\text{m}_{i}})^{2}/\epsilon_{i}^{2}]}{N}$ (2) where $F^{\prime}_{\text{m}_{i}}$ is the flux of the $i^{\text{th}}$ pixel of the synthetic spectrum corrected by polynomial, $\epsilon_{i}$ is the flux difference of the $i^{\text{th}}$ pixel, and $N$ is the number of pixels involved in the calculation. For each given template spectrum, we calculate the $\chi^{2}$ values when matching all the synthetic spectra in the grid and then interpolate between the parameters by synthesizing linear combinations of top-five best-fit synthetic spectra, as the same way in Yee et al. (2017). Figure 2: Template spectra of subtype dM0, sdM2, esdM4 and usdM6 with their corresponding synthetic spectra and fitting residuals. Figure 3: The residuals between template spectra of different spectral subtypes and their best-fit synthetic spectra. Black, blue, purple, and red represent subclass dM, sdM, esdM, and usdM respectively. The error bar is the standard deviation of the difference between the flux of the template spectrum and the flux of the synthetic spectrum. The number beside each node represents the number of spectra used to construct the template spectrum. Figure 4: The distributions of estimated atmospheric parameters of the templates spectra of different subclasses. The colors represent the same as in Figure 3. The fitting results of four template spectra of different spectral subtypes and metallicity subclasses (dM0, sdM2, esdM4 and usdM6) with their corresponding estimated parameters are shown in Figure 2 as examples. The shape of the residual spectra indicate that the errors mainly come from the region of 6000$-$7000Å, where there remain some incompleteness in line lists and/or missing/incomplete opacities on some molecules used in the modeling (CaOH, CN, CaH, etc.). Besides, the far red region of the observed spectra could be contaminated by telluric absorption. Figure 3 shows the residual levels of the fitting results of all the templates. The residual level of the fitting results gradually increases from $\sim$5% for early type M stars to 10% for the late types: both spectral quality and the number of objects for template building decrease dramatically for the latest types. Finally, the distributions of the estimated results of log $g$ and [Fe/H] of the templates above are shown in Figure 4. The dispersions are large for both the two parameters, with the basic trend being a clear decrease of the average metallicity from sd to usd accompanied by a more confuse increase of gravity. Note that when using the complete set of templates (221 in total), we can meet spectra of slightly atypical characteristics, as for instance illustrated in the top panel of Figure 2 in which a dM0 is metal-poor and 100 K cooler than a standard dM0 of solar metallicity. ## 4 Sample Construction and Atmospheric Parameter Estimates To estimate the atmospheric parameters $T_{\text{eff}}$, log $g$ and [M/H] of M subdwarfs from low-resolution spectra and investigate the trends of these parameters with the kinematic properties of the stars, we need a sufficiently large sample of subdwarfs spanning the broadest possible range of parameters. In Paper i@, we had explored LAMOST DR4 in which 2,791 M subdwarf were identified. Extending the search to LAMOST DR7, a total of 4,767 visually selected candidates are retrieved including the previously obtained DR4 ones. Combining this sample with SDSS objects published by Savcheva et al. (2014), we submitted these spectra to a least-squares based fitting process with PHOENIX synthetic spectra. The determination of atmospheric parameters is not always possible with a reasonable accuracy because the error budget of the fitting process depends on the signal-to-noise (S/N) of the spectra. Thus, the final combined LAMOST + SDSS sample of stars with known atmospheric parameters will be reduced with respect to the initial numbers. ### 4.1 Sample Construction and Individual Spectra Fitting Because of magnitude limitation, most M dwarfs of LAMOST range in spectral type before M3, and only a few cooler and fainter spectra of low S/N are available. The data of SDSS DR7 covers a more extended parameter space, with cooler and lower metallicity objects. The strategies of two surveys point to different fields, i.e. SDSS at high Galactic latitude favoring halo and thick disk objects detection, while LAMOST survey has in priority been targeted at regions of the Galactic anti-center. Therefore, in this study we combine the subdwarfs from SDSS DR7 (SEGUE-I/II) and LAMOST DR7 to construct a consistent sample of M subdwarfs, more representative of the total population of these stars. First, it is useful to discuss three points on which the processing of the LAMOST and SDSS spectra sets will differ: (1) selecting criteria, (2) spectroscopic resolution, and (3) spectral data reduction. (1) Selecting criteria. Several works have been conducted to identify M subdwarfs from SDSS and LAMOST low resolution spectral data sets (West et al., 2004; Lépine & Scholz, 2008; Bochanski et al., 2013; Savcheva et al., 2014; Zhang et al., 2016; Bai et al., 2016), Paper i@, and the two largest samples published were from Savcheva et al. (2014) and Paper i@ respectively. The selecting conditions differ for these two latter samples: the principal separator used is the standard parameter $\zeta$ initially introduced by Lépine et al. (2007) based on a combination of CaH2, CaH3 and TiO5 spectral indices: $\zeta_{\text{CaH}/\text{TiO}}=\frac{1-\text{TiO5}}{1-[\text{TiO5}]_{Z_{\odot}}}$ (3) in which the [TiO5]${}_{Z_{\odot}}$ is a third order polynomial fit of the TiO5 spectral index as a function of the CaH2+CaH3 index as: $\begin{split}[\text{TiO5}]_{Z_{\odot}}\ &=\ a\\\ +&b\times(\text{CaH2}\ +\ \text{CaH3})\\\ +&c\times(\text{CaH2}\ +\ \text{CaH3})^{2}\\\ +&d\times(\text{CaH2}\ +\ \text{CaH3})^{3}\end{split}$ (4) so that $\zeta$ = 1 for solar metallicity and $\zeta$ = 0 for the most metal- poor objects where TiO5 absorption becomes undetectable. The coefficients $a$, $b$, $c$, $d$ derived for SDSS objects by Lépine et al. (2007) led to the calibration $\zeta_{\text{L07}}$ which was later revised by Lépine et al. (2013) (hereafter $\zeta_{\text{L13}}$). The condition for a star to belong to the subdwarf population was set as $\zeta_{\text{L07}}<$0.825 and $\zeta_{\text{L13}}<$0.825 Since $\zeta$ was found to be dependent on the M dwarfs reference sample, instrumental setup and various factors influential in the reduction process (Lépine et al., 2013), we proposed a new calibration adapted to LAMOST spectra in Paper i@ (hereafter $\zeta_{\text{Z19}}$), and we suggested the condition for LAMOST subdwarfs being set as $\zeta_{\text{Z19}}<$0.75 with an additional screening condition based on the CaH1 index. (2) Spectroscopic resolution. Along the wavelength of a spectrum from SDSS or LAMOST, the physical resolution (depending on the grating$\times$focal length of camera$\times$pixel size of the detector) varies in a complicated way. For LAMOST, when spectra are extracted and rebinned in constant log-wavelength steps by the 1D pipeline, this variation is +/- “dissolved” into another variation. Additionally, though in theory, all spectra obtained from the same instrument are supposed to have an identical resolving power, in practice for a large-scale survey project, the resolution changes slightly at different observation dates, due to imperfect instrument stability. In the actual calculation, assuming a gaussian instrumental profile, the resolution R of a spectrum at wavelength $\lambda$ can be derived from $\lambda$/$\Delta\lambda$, where $\Delta\lambda$ is the full width at half maximum (FWHM) of the spectral line broadening at that wavelength. The Gaussian kernel $\sigma$ = FWHM/2.355 used to convolve the synthetic spectrum is derived from $\sqrt{{\sigma_{\text{o}}}^{2}-{\sigma_{\text{m}}}^{2}}$, where $\sigma_{\text{m}}$ corresponds to the synthetic spectrum, and $\sigma_{\text{o}}$ corresponds to the observed spectrum. Since the resolution of the synthetic spectra is extremely high, we set $\sigma_{\text{m}}$ as 0 here. Prior to the fitting process, the observed spectra are shifted to rest-frame based on the measured radial velocities. The fitting range was set to 6000$-$8800Å, in which the pixels marked by “bad” flags and the spectral intervals heavily contaminated by telluric absorption (7210$-$7350Å, 7560$-$7720Å, and 8105$-$8240Å) were masked. We divided the region 6000$-$8800$\rm\AA$ into 7 bands evenly and used a dedicated convolution kernel for each 400$\rm\AA$ band. For SDSS and LAMOST spectra, we adopted different convolution kernels calculated as follows. For an SDSS spectrum, we computed the mean value of the $\sigma$ in each band by the “dispersion” value of each pixel stored in the fits file. For a LAMOST spectrum, we calculated the mean value of the FWHM of the arc lamp spectrum in each band, obtained on the same day for wavelength calibration purpose. (3) Data reduction. In the fitting process described above, we first correct the continuum of an observed spectrum using a fourth-order polynomial and then choose the best-fit synthetic spectrum based on the chi-square minimum principle. We find that the continuum slopes of some observed spectra of LAMOST set differ largely from their corresponding corrected spectra. This deformation is mainly caused by flux calibration problems (Du et al., 2016) and would affect the accuracy of the spectral indices. To control this unwanted bias, we measure the spectral indices of each polynomial-corrected observed spectrum, and calculate the difference between them and the original indices measured on the uncorrected spectrum. We then construct the distribution of each index differences, as shown in Figure 5. Although the offset of a single index is generally small, the error on the $\zeta$ parameter composed of multiple indices may be considerably enlarged. Hence, targets with at least one spectral index difference beyond 3$\sigma$ of the corresponding difference distribution are removed from the final subdwarf sample. Figure 5: Each subdiagram shows a spectral index error distribution for the LAMOST subdwarfs. The errors are the differences between the values measured from the original spectra minus the values measured from the continuum- corrected spectra. They reflect the deformation level of the local pseudocontinuum, due to imperfect flux calibration. The bottom right panel shows the distribution of the derived parameter $\zeta$ values with a scatter 0.133 although the scatters of each of the indices (CaH2, CaH3 and TiO5) used to calculate $\zeta$ are quite smaller. We summarize the specific solution to build the combined LAMOST + SDSS subdwarf sample in seven steps as follows: (1) Select 4,767 visually inspected subdwarfs from LAMOST DR7. (2) Recalculate the indices of the SDSS subdwarfs from Savcheva et al. (2014) with the radial velocities recorded in the catalog (see Paper i@ for more comments). (3) Calculate the resolution of each spectrum using the profile of corresponding arc-lamp spectrum (LAMOST) or the “dispersion” value stored in the fits file (SDSS) and get the gaussian convolution kernels. (4) Smooth the synthetic spectra to fit every observed spectrum and estimate the atmospheric parameters. Figure 6 shows the fitting result of a LAMOST extreme subdwarf in the upper panel and an SDSS subdwarf spectrum showing H$\alpha$ emission in the lower panel. The spectral bands contaminated by telluric absorptions are masked. (5) For each index, measure its value on each observed spectrum and on its corresponding continuum-corrected spectrum, and get their difference. (6) Exclude the objects with one or more spectral index differences beyond 3$\sigma$ of the corresponding difference distribution. (7) Combine the two samples above, and apply the selecting cuts by removing the objects of $\zeta_{\text{L07}}\geq$0.825, $\zeta_{\text{L13}}\geq$0.825, and $\zeta_{\text{Z19}}\geq$0.75. Finally, a sample of 3,131 subdwarfs is assembled, 1,722 of which have reliable kinematic properties. To provide better understanding, we summarize in Table 3 the various subsamples used along this work. Table 3: The various stellar samples used in the present work. (1) | LAMOST M dwarfs (2) | LAMOST M subdwarfs initial sample | SDSS M subdwarfs initial sample (3) | LAMOST+SDSS combined subdwarf sample with atmospheric param. before CaOH/CaH1 cut | LAMOST+SDSS green-marked “subdwarfs” | LAMOST+SDSS “genuine” subdwarfs after CaOH/CaH1 cut (4) | LAMOST M giants (5) ---|---|---|---|---|---|---|--- N | $\sim$70,000 | 4,767 | 2,780 | 3,131 (LAMOST: 1,852) (SDSS: 1,279) | 855 | 2,276 | 7,200 N${}_{\text{Gaia}}$ | 35,382 | - | - | 1,722 (LAMOST: 1,160) (SDSS: 562) | 526 | 1,196 | - * • * • Notes to Table 3: * • (1) The numbers corresponding to N${}_{\text{Gaia}}$ means the objects (N) have available data from Gaia DR2 and can be used for kinematic analysis. The flag is set as 1 if all of the following conditions are true: a. phot_bp_mean_flux_over_error $>$ 10 b. phot_rp_mean_flux_over_error $>$ 10 c. phot_g_mean_flux_over_error $>$ 10 d. ResFlag = 1 $\&$ ModFlag = 1 (from Bailer-Jones et al. 2018) e. astrometric_excess_noise_sig $\leq$ 2 f. ruwe $\leq$ 1.4 * • (2) A visually identified sample from Zhang et al. (2019) for comparison. * • (3) Savcheva et al. (2014), the spectral indices are recalculated using the radial velocities published in the catalog. * • (4) i.e. using the Equation 5 in the present paper. * • (5) A visually identified sample from Zhang et al. (2019) for comparison. Figure 6: Upper panel: a LAMOST extreme subdwarf spectrum and its best-fit synthetic spectrum. The spectral bands contaminated by telluric absorptions are masked. Lower panel: an SDSS subdwarf spectrum showing H$\alpha$ emission and its best-fit synthetic spectrum. Figure 7: The M subdwarf sample constructed in this work, combining the low-resolution spectra from SDSS (blue dots) and LAMOST (red dots) data sets. The M dwarfs used for comparison (black dots) are visually confirmed from LAMOST data (Zhang et al., 2019). The bottom right internal subdiagram shows the spectral type distributions of the two sub-samples. We must underline that we haven’t used the CaH1 index as a filter until now. For comparison, we also estimate the atmospheric parameters of a M dwarf sample which contains around 70,000 visually inspected spectra from LAMOST (Zhang et al., 2019). Figure 7 shows the combined LAMOST+SDSS sample of subdwarfs together with the comparison sample of LAMOST M dwarfs. The LAMOST sample contains a large number of early-type and medium metal-poor stars, while the SDSS sample covers a broader range of spectral types and metal abundances. ### 4.2 Results of Parameter Estimation #### 4.2.1 Signal-to-Noise impact on quality Statistically, the fit residuals between the observed spectra and the synthetic spectra vary with signal-to noise ratio (S/N). Figure 8 shows the distribution of fitting residuals for subdwarfs from the two surveys. The left panel shows the LAMOST subdwarfs mainly covering M0$-$M3 while the SDSS subdwarfs on the right panel cover M0 to M8. The residual levels stabilize below 10% after the S/N becomes larger than $\sim$30\. The later-type subdwarfs from both samples show higher residual levels because of an average lower spectrum quality. Figure 8: The spectral fitting residual of subdwarfs as a function of the signal-to-noise in $i$ band. Left panel: LAMOST subdwarfs, right panel: SDSS subdwarfs. #### 4.2.2 Comparison with literature results In order to estimate the accuracy of the derived parameters, we cross-match the entire LAMOST DR7 M-type stars with objects having known parameters from literature. As a result, we find a total of 269 observations for 167 stars studied with high or medium spectral resolution in 9 papers, and the comparison is shown in Figure 9. Although a systematic bias in the effective temperature appears from around 3700K to the cool end, it generally follows a 1-to-1 trend with the literature values. The metallicity also seems to follow a 1-to-1 trend, while the gravity values show a larger range than the literature ones. Note that we have masked several telluric absorption bands where gravity-sensative K and Na lines are located in the fitting process, the accuracy of our estimated gravity values is therefore affected somehow. The distributions of the differences (measurements from this work $-$ literature values) for the three parameters are shown in Figure 9. The bias and scatter are 165 K and 170 K for $T_{\text{eff}}$, 0.03 dex and 0.32 dex for log $g$, and $-$0.17 dex and 0.41 dex for [Fe/H], respectively. Figure 9: The upper panels show the comparison of atmospheric parameters of 167 M stars estimated from 269 LAMOST spectra (including multi-observations) with the ones compiled from literature. Parameter values from different literature are shown in different colors as listed in legend. Error estimates provided are shown as corresponding error bars. The uncertainties for the values from LAMOST spectra are the upper values of internal errors determined from signal-to-ratio and $\sqrt{\chi^{2}}$. The bottom panels show the distribution of differences along with the corresponding mean value and standard deviation. ## 5 Sample Properties Analysis In this section, we analyze the properties of the subdwarf sample by comparing them with the dwarf sample on spectral index diagrams and Gaia DR2 H-R diagram, and provide a preliminary account of their kinematic properties from the reduced proper motion diagram and in the 3D galactic motion system. In Paper i@, we suggested a screening condition associated with the CaH1 index be combined with the classical $\zeta$ condition to identify a subdwarf via its low-resolution spectrum. To show the importance of gravity and gravity- dependent index—CaH1 for identifying a “genuine” subdwarf, we divide the subdwarf sample selected above into two groups by the separator curve (equation 9 in Paper i@) on the [CaOH, CaH1] index diagram: CaH1 $\displaystyle<0.4562\times\text{CaOH}^{3}-0.1977\times\text{CaOH}^{2}$ (5) $\displaystyle-0.01899\times\text{CaOH}-0.7631$ The objects that do not meet Equation (5) are marked in green in the following figures. For further analysis, we retrieve proper motions, and photometric magnitudes in $G$, $BP$, and $RP$ bands of these stars from Gaia Data Release 2, and estimated distance from Bailer-Jones et al. (2018). To ensure reliability, we set a series of filter conditions as listed in the notes to Table 3, and finally 1,722 objects among the 3,131 selected in Section 4 meet the conditions, 526 of which are green-marked “subdwarfs”. Figure 10: The gravity and metallicity distributions of the green-marked “subdwarfs” vs. the other sdMs. The two upper panels show the distribution of spectral indices of four groups of subdwarfs: green-marked “subdwarfs” (the subdwarfs that don’t meet Equation 5), sdMs, esdMs, and usdMs (the remaining subdwarfs classified into to categories according to the $\zeta$ values). The sdMs, esdMs and usdMs are colored in blue, purple and magenta respectively. The bottom left and middle panels show the comparison of green-marked “subdwarfs” with the sdMs. The bottom right panel compares the total proper- motion distribution of the green-marked “subdwarfs” with that of the other sdMs. ### 5.1 Spectral Index Diagram As shown in Figure 10, the upper left panel shows the green marked “subdwarfs”. According to the $\zeta$ values, we also divide the remaining subdwarfs into sdM/esdM/usdM subclasses on the [TiO5, CaH2+CaH3] diagram, as can be seen in the upper right panel, and the objects in the three subclasses are plotted as blue, purple and magenta points respectively. Most of the green marked “subdwarfs” appear located in the “sdM” region, which is colored in blue. Therefore, we compare the gravity and metallicity of the two groups and display those distributions in the bottom left and middle panels. As shown in the bottom middle panel, the green objects share similar metal abundance distribution with the blue labeled ones, however, their gravities are systematically lower than the latter ones by some 0.2 dex. With the available proper-motions of these targets from Gaia DR2, we also compare the total proper motions of the two groups in bottom right panel of Figure 10. The proper-motions of green marked “subdwarfs” are overall smaller than those of the sdMs. Generally speaking, large proper motion is a classical property of subdwarfs, though some objects may have relatively small proper motion due to the 3D motion direction. Proper motions of esdMs and usdMs are larger than sdMs (not shown in the figure), which means that the green ones do behave more like dwarfs than subdwarfs. Figure 11: M-type stars with different gravity shown on the spectral index diagram CaH1 vs. CaOH. Color coding is the same as in the upper left panel of Figure 10. To illustrate the sensitivity to gravity of this diagram, an M giant sample (7,200 objects visually identified from Zhang et al. 2019 and colored in yellow) and the M dwarf sample are added. The gravity differences can be seen more clearly in Figure 11, in which we add the LAMOST M dwarfs and a new sample of M giants onto the [CaOH, CaH1] index diagram. The giants were also spectroscopically identified by visual inspection (Zhang et al., 2019). Giants, and dwarfs subdwarfs are classified as luminosity class iii@, v@ respectively and Jao et al. (2008) have proposed that subdwarfs be classified as vi@, this sequence tracing the gravity variation trend. Therefore, Figure 11 clearly illustrates the gravity segregation between the luminosity classes of giants/dwarfs/subdwarfs thanks to the sensitivity of CaH1 index to log $g$ especially when compared to CaOH index. ### 5.2 H-R Diagram Using estimated distance by Bailer-Jones et al. (2018), we derive the absolute magnitudes M${}_{\text{G}}$, M${}_{\text{BP}}$ and M${}_{\text{RP}}$ of the dwarfs and subdwarfs. Since most of the M-type stars we analyze here are located within 1 kpc, we choose to ignore the effect of extinction in the calculation. Figure 12 shows the stars on Gaia color-absolute magnitude diagram [BP$-$RP, M${}_{\text{RP}}$]. On this diagram, which is an observational Hertzsprung-Russel diagram, the gray dots represent the dwarf sample, the green dots represent the “subdwarfs” that do not meet Equation (5), and the red dots represent the remaining subdwarfs (all subclasses together). The green-marked “subdwarfs” fall almost completely in the dwarf region, and thus once again verify the conclusion that they behave more like dwarfs than subdwarfs. Figure 12: All the subdwarfs and dwarfs on the Gaia color-magnitude diagram. The absolute magnitudes in RP band are derived from the Gaia DR2 apparent magnitudes and estimated distances from Bailer-Jones et al. (2018). Most of the green-marked “subdwarfs” fall into the main sequence region populated by the M dwarfs plotted as black dots, and the remaining subdwarfs plotted as red dots lie in part below the main sequence with a large dispersion, but some are also distributed across the main sequence region. Figure 13: After the removal of the green-marked “subdwarfs”, the remaining samples are shown on the Gaia color-absolute magnitude diagram again. On the left panel, subdwarf are color-coded for different metallicities, while the same ones on the right panel are color-coded for different gravities. Note that the subdwarfs with log $g>$ 5.5 or log $g<$ 4.5 (38 objects in total) are not plotted on the right panel to make the color bar more tight. Besides, we find that a number of Equation (5) selected subdwarfs (red dots) also overlap with dwarfs on this diagram. This point will be developed below. Figure 13 shows the metal abundance and gravity variation of these stars on the Gaia color-magnitude diagram. The subdwarfs falling in the dwarf region are the ones with the highest metal abundances or the smallest surface gravities. In other words, stars that are away from the main sequence have both large gravity and low metal abundance, and they should be the “genuine” subdwarfs. ### 5.3 Kinematics #### 5.3.1 Reduced Proper Motion diagram Aside from the H-R diagram, another useful tool for effectively separating different stellar populations is the reduced proper motion (RPM; Jones 1972) diagram (e.g. Salim & Gould 2002; Subasavage et al. 2005; Lépine et al. 2007; Faherty et al. 2009), which uses photometric information (color) and proper motions of the stars. Even without knowledge of the distance to the stars, the RPM diagram can be used to constrain kinematic properties of large stellar samples, because a given stellar population whose members are orbiting in the Galactic potential has a characteristic space velocity distribution reflected in a combination of luminosity and transverse motion. The definition of the reduced proper motion $H$ is $H=m+5log\ \mu+5=M+5log\ T$ (6) where $m$ is the apparent magnitude of a star in a given photometric band, $\mu$ is the annual total proper motion (in arcsec yr-1), $M$ is the absolute magnitude, and $T$ is the tangential velocity (in a.u. yr-1). Since the dwarfs belong to the thin disk population and the subdwarfs are claimed to belong to the thick disk and halo populations, the comparison between blue to red or optical to near-infrared colors and the reduced proper motions can in principle relatively clearly distinguish the Population i@ dwarfs and Population ii@ subdwarfs : (1) given an absolute magnitude, subdwarfs are bluer in color; (2) subdwarfs being members of a population spanning a spatial volume more extended out of the Galactic plane with more eccentric orbits, they have higher average tangential velocities (Salim & Gould, 2002; Lépine et al., 2007). The RPM diagram combines these two features so that the subdwarfs and main sequence stars fall into two separate regions. This method has long been used to separate high-velocity subdwarfs from dwarfs (Jones, 1972; Yong & Lambert, 2003a, b). Figure 14: After the removal of the green-marked “subdwarfs”, the remaining samples are shown on the reduced proper motion diagram. The color coding is the same as in Figure 12. The fiducial separation line between Cloud A and Cloud B objects is drawn (see text). Figure 14 shows the RPM diagram for our stars, using Gaia DR2 color and proper motions. A clear similarity with the information contained in the H-R diagram of Figure 14 can be noted. Although the selection biases of the combined LAMOST+SDSS subdwarf sample are poorly known (mix of pointing directions, different limiting depth of surveys, velocity ellipsoid sampling, density distribution sampling, spatial inhomogeneities, etc.), basic trends can be outlined. First, the subdwarfs falling below the region occupied by the main- sequence dwarfs have low metal abundances and higher gravities. Second, remaining subdwarfs appear clearly divided in two clouds (noted A and B in the following) with a quite clean separation: Cloud A is not distinct from the lower part of the main-sequence dwarf cloud, while Cloud B is centered around a half magnitude below the center of Cloud A. The fact that Cloud A and B are especially distinct in the RPM diagram imply that these two clouds sample two distinct populations whose kinematic behaviors are different. #### 5.3.2 Galactic velocity components Using the usual textbook transformation formula, we compute the three- dimensional space velocity components $U$, $V$, $W$ in the right-handed Galactic frame: the $U$ and $V$ are contained in the Galactic plane, $U$ points toward the Galactic center, while $V$ is the velocity component in the direction of the solar motion. The $W$ component is perpendicular to the Galactic plane and points towards the North Galactic pole. $U$, $V$, $W$ for our targets are determined using the radial velocities, coordinates, proper motions and distances derived from the parallaxes. The components are reduced to the Local Standard of Rest (LSR; Fuchs et al. 2009) by correcting from the solar motion assumed to be ($U_{\odot}$, $V_{\odot}$, $W_{\odot}$)=[11.1, 12.24, 7.25] km s-1 (Schönrich et al., 2010). Figure 15: Velocity component distribution of dwarfs and subdwarfs with different gravities in the Galactic 3D motion system ($U$, $V$, $W$). From left to right, the abscissa represents the velocity component of $U$, $V$, and $W$ respectively. The ordinate represents the distance from Galactic plane. Except for the green-marked “subdwarfs”, the remaining subdwarfs are divided into groups with different gravities and distinguished by different colors. Each dwarf bin contains 7,000 stars and each subdwarf bin contains 100 stars. The node and the error bar show the mean value and the standard deviation of each bin respectively. Figure 16: Velocity component distribution of dwarfs and subdwarfs with different metallicities in the Galactic 3D motion system ($U$, $V$, $W$). Except for the green-marked “subdwarfs”, the remaining subdwarfs are divided into groups with different metallicities and distinguished by different colors. The meaning of the axes and the number of stars contained in each bin are the same as in Figure 15. In Figures 15 and 16, we divide our remaining subdwarfs into different gravity and metallicity groups and compare them with the green-marked “subdwarfs” and dwarfs. Each dwarf bin contains 7,000 objects and each subdwarf bin contains 100 objects. As expected, the green-marked “subdwarfs” behave mostly like field dwarfs with only a slightly larger dispersion in all the velocity components. In addition, the yellow bins in Figure 15 representing the subdwarf group with the lowest gravity and the light blue bins in Figure 16 representing the subdwarf group with the highest metallicity have nearly the same behaviors in $V$ as the dwarfs and the green-marked “subdwarfs”. To further explore the kinematic behavior of the full subsamples, we draw several histograms in Figure 17, in which the dwarfs, green-marked “subdwarfs”, Cloud A and Cloud B subdwarfs are plotted separately. The mean values and standard deviations are also shown on each panel, except for $|$z$|$, the absolute value of the distance to the Galactic plane, for which the quartiles are computed. The separation between Cloud A and Cloud B objects is defined by the fiducial line drawn on Figure 14. Several points emerge: Figure 17: Histograms of some characteristics for the various sub-populations of dwarfs and subdwarfs for which Gaia astrometry is available. From left to right, log $g$, [Fe/H], $U$, $V$, $W$ space velocity components and $|$Z$|$. From top to bottom: (a) all spectroscopically identified M dwarfs, (b) green- marked “subdwarfs”, (c) Cloud A subdwarfs (see Figure 14 for definition), (d) Cloud B subdwarfs, (e) Extended Cloud A (see text for definition), (f) Disk outliers (spectroscopically identified M dwarfs whose total space velocity w.r.t. the LSR is larger than that of 99 % of the population, see Figure 19 for their location on Toomre diagram). The $\mu$ and $\sigma$ on each panel are indicated for the mean value and standard deviation respectively. For the $|$Z$|$ distributions, the quartiles are indicated. (a) The gravities increase systematically from dwarfs (mean close to 5.0, in agreement with the results from Section 3) to Cloud B subdwarfs (mean around 5.2). (b) Metallicity decreases systematically from dwarfs (mean around $-$0.4 dex) to Cloud B subdwarfs. The metallicity of the dwarfs, quite surprisingly, appears systematically poorer than solar, with a distribution tail down to $-$1\. dex. Although the error assessment and control is difficult in the spectra fitting process, we recall that the comparison in Subsection 4.2 with the sample of 167 stars having a metallicity measurement in the literature did not reveal any major systematic deviation. Figure 18: The bimodal distribution of metallicity of the halo candidates, which are selected from Cloud B subdwarfs with $v_{tot}\geq$220 km s-1. The mean values and standard deviations of each Gaussian fit are also shown on the diagram. (c) As shown in Figure 18, the halo candidates in Cloud B subdwarfs exhibit an inconspicuous bimodal distribution of metallicity, with two sub-populations, one around [Fe/H] = $-$1 dex and the other around $-1.5$ dex. (d) The mean values and dispersions of the $U$, $V$, and $W$ components agree quite well with the dwarf sample belonging to the thin disk, with a velocity lag of $V$ = $-$10 km s-1. The green-marked and Cloud A objects have more velocity dispersion suggesting possible membership of the thick disk, but their mean $V$ value significantly deviates from the usually adopted $-$60 km s-1. Cloud B objects are obviously dominated by halo members, with a mean $V$ largely retrograde and considerable dispersion. (e) The quartiles of the $|$z$|$ distributions show that the dwarf sample is the one which is the most concentrated around the Galactic plane, while the Cloud B contains the most diffuse population out of this plane. These histograms also show that the LAMOST+SDSS surveys sample the M dwarf population inside scale height of the thin disk above the Galactic plane (600 pc) (Jurić et al., 2008), while the subdwarf one is sampled only up to less than 1 kpc (one scale height of the thick disk or less) (Gilmore & Reid, 1983; Veltz et al., 2008; Carollo et al., 2010). (f) The striking similarity of all the indicators (except for a marginal difference in log $g$ ) between the green-marked objects and the Cloud A ones imply that they are members of the same stellar population as defined kinematically, only differing by the CaH1 and CaOH indices. In the following, these two groups will be considered as a unique subsample which will be designated as Extended Cloud A. #### 5.3.3 The Toomre diagram Trying to push forward the analysis, we have plotted in Figure 19 the Toomre diagram for all our stars. The dwarf cloud appears slightly asymetric, with an extension towards negative $V$ velocity component. This is not unexpected since 2 or 3 - component disk (thin, thick and intermediate) model populations (Schönrich & Binney 2009 and references herein) predict such an asymmetry. The Extended Cloud A group seems to be centered more or less coincident with the dwarf cloud, while Cloud B objects are scattered across a large area of the diagram. 1. 1. Halo objects. To select halo objects, a stringent condition on their total velocity (Bonaca et al., 2017) which would avoid pollution by possibly numerous thick disk stragglers (Smith et al., 2009), is: $|V-V_{\text{LSR}}|\geq 220\ \text{km s}^{-1}$ (7) which translates, which the adopted velocity components of the LSR, ($U$ = 0, $V$ = 220, $W$ = 0 km s-1) into: $v_{\text{tot}}=\sqrt{U^{2}+V^{2}+W^{2}}\geq 220\ \text{km s}^{-1}$ (8) where $v_{\text{tot}}$ is the total space velocity with respect to the LSR. Figure 19: Left panel: Toomre diagram for all stars in the present study. Black points: all spectroscopically identified M dwarfs, green circles: the green-marked “subdwarfs”, blue circles: Cloud A subdwarfs, red circles: Cloud B subdwarfs. The black dashed curve encloses 99 % of the M dwarf population: all black points outside this curve are considered as representative of “outliers”. The dashed red circle is our adopted limit $v_{\text{tot}}$ = 220 km s-1 for halo population assignment. The dashed magenta vertical line separates retrograde orbits on its left from prograde orbits on its right. Right panel: the M dwarfs with thin disk kinematics have been removed. The green-marked “subdwarfs” and Cloud A subdwarfs are merged into a single group: Extended Cloud A. The M dwarfs outliers have been separated based on their total space velocity $v_{\text{tot}}$ w.r.t. the LSR. The Cloud B are color- coded in four categories based on their metallicity and $v_{\text{tot}}$. This selection provides a kinematic sample of the local halo haunting the Solar Neighbourhood, which contains essentially stars from Cloud B, (220 objects), with a handful from Extended Cloud A, (9 objects) accompanied by targets considered as outliers from the dwarf cloud (71 objects)(see below). Its kinematic characteristics are in fair agreement with those of already identified halo population stars selected across widely more extended volumes and using more luminous targets (Smith et al., 2009; Bond et al., 2010; Bochanski et al., 2013): $<$$U$$>$ = 5 km s-1 and $\sigma_{U}$ = 173 km s-1, $<$$V$$>$ = $-$248 km s-1 and $\sigma_{V}$ = 77 km s-1, and $<$$W$$>$ = 6 km s-1 and $\sigma_{W}$ = 89 km s-1. With the adopted value of the LSR velocity, the fiducial line at $V$ = $-$220 km s-1 separates the stars which have retrograde orbits from those who have prograde ones. A striking fact is the quite high fraction of prograde objects (258 stars or 40%) observed in the present local halo sample, thus the strongly negative value of $<$$V$$>$ quoted above does not account completely for a complex orbital distribution. The metallicity of this halo sample, in spite of the kinematical cut applied, remains bimodal. Previous studies, mostly based on F-G-K stars in which metallicity is derived from atomic lines have already found a substantial metal-rich component in the halo population, (and the metallicity distribution of globular clusters is one historic and well-documented example (see e.g. Zinn 1985; Harris 1996) but the metal-rich peak is generally found around [Fe/H] = $-$0.6 dex: Bonaca et al. (2017). The prograde objects are also reputedly more metal-rich than the retrograde ones, implying that they could belong to different subpopulations, the metal-rich one containing inner halo objects while the metal-poor one is made of outer halo objects, these two subpopulations having different formation histories and different space distributions. In the present local halo sample, we do not find such a segregation, the number of prograde metal-poor objects remaining important while many higher metallicity stars are on obviously retrograde orbits, as shown in the right panel of Figure 19. 2. 2. Thin disk dwarfs: Since some spectroscopically identified dwarfs are found to kinematically belong to the local halo, we could expect some continuity in the dwarf high velocity distribution tail, and therefore a number of them would also be kinematically mixed with the Extended Cloud A population and need to be disentangled. To select thin disk dwarfs, a condition of type $v_{\text{tot}}$ = constant is not completely satisfactory because of the asymmetry of the dwarf cloud in the Toomre diagram. A correct way to proceed would be to compute for each star its probability of membership to one of the three populations - halo, thick disk and thin disk as proposed e.g. by Bensby et al. (2003) but we prefer to postpone such a calculation because the selection effects in the present sample are not correctly assessed. We choose a simplified and empirical procedure as follows: we assume that the distribution of the dwarf stars in the Toomre diagram is consistent with an incomplete ellipse, and that the isodensity contours of this ellipse trace constant probability level of a star to belong to its parent population. We select the isodensity contour containing 99% of the dwarf cloud and fit an ellipse to this contour. All stars found outside are considered as “outliers” from the thin disk dwarf population. The kinematically “purified” dwarf sample shows typical velocity component values of the thin disk population (Edvardsson et al., 1993; Reddy et al., 2003): $<$$U$$>$ = 0 km s-1 and $\sigma_{U}$ = 44 km s-1, $<$$V$$>$ = $-$13 km s-1 and $\sigma_{V}$ = 30 km s-1, and $<$$W$$>$ = 0 km s-1 and $\sigma_{W}$ = 26 km s-1, and its stars have exclusively prograde orbits as expected. The metallicity, as outlined above, is in average slightly deficient, the histogram showing three peaks, one solar or slightly super-solar, the principal one around $-$0.4 dex and the third one around $-$1 dex. Since the selection effects are not under control, we doubt that this three-component would be significant. 3. 3. Thick disk population and its elusive limits The Extended Cloud A occupies a large area, somewhat fan-shaped vertically, in the Toomre diagram, broadly covering the dwarf cloud and merging beyond it until the halo region, with practically all orbits prograde. The shape of the $|$Z$|$ histogram, with its peak displaced from that of the thin disk dwarfs and its quartiles systematically larger shows that this population is clearly thicker in $z$. The average metal content is poorer than that of the thin disk dwarfs at [Fe/H] = $-$0.74 dex with a [Fe/H] distribution having a long metal- poor tail extending to $-$2.5 dex. We suspect that the Extended Cloud A population is an intricate mix of objects really belonging to a thick disk component with pollution by an important contribution of thin disk “mild” subdwarfs very difficult to disentangle, with unknown selection effects playing in background. A possible way to separate these contributions is the detailed computation of individual orbits and study of the $z_{\text{max}}$ and eccentricities, but such a work is far beyond the scope of the present study. In the Toomre diagram, the Extended Cloud A population is mixed with a fraction of the Cloud B population, corresponding to stars that preferentially belong to the high metallicity peak of the bimodal [Fe/H] distribution of Cloud B in Figure 17. These stars likely belong to the halo for a part, because our halo selection kinematic cut is very stringent, and to the thick disk for another part, but their respective assessment to the two components would probably also need a detailed orbital calculation. The [Fe/H] distribution of the outliers from the dwarf cloud show that they span the same broad metallicity range as the principal dwarf population, peaking at $-$0.6 dex, only slightly more metal-poor than the ordinary disk dwarfs. Their distribution in $|$Z$|$ is clearly more extended, making them probable members of the thick disk. Their velocity dispersions $\sigma_{U}$ = 119 km s-1, $\sigma_{V}$ = 90 km s-1, $\sigma_{W}$ = 75 km s-1 suggest that they contain a majority of objects belonging to the thick disk with some pollution by an inner halo fraction. To conclude, the combined subdwarf sample extracted from LAMOST DR7 and SDSS DR7 contains a mix of objects belonging to various components: clear halo members with an implicit bimodal metallicity and considerable orbital mixing between retrograde and prograde cases, and an important thick disk fraction merging both into halo and thin disk. The spectroscopically selected M dwarfs themselves, whose quasi-totality belong to the standard thin disk, contain a small fraction (less than 1% of the total) of outliers belonging both to the thick disk and the halo. It is quite unclear if the present situation is dependent or not on the selection effects, especially the favoured directions of investigation towards the Galactic anticenter in the LAMOST survey. ## 6 Summary and Conclusion In the most widely used classification system, M-type main sequence stars are classified into dM/sdM/esdM/usdM subclasses corresponding to different metal abundances according to the CaH2, CaH3 and TiO5 molecular bands in the low- resolution spectrum. The metal-poor subclasses are named as subdwarfs, which have typical kinematic characteristics of thick disk and halo population. However, in the present work, we suggest the “genuine” subdwarfs belonging to population ii@ be classified as luminosity class vi@ on the H-R diagram. For these stellar objects, having larger surface gravity is a necessary condition as important as “being metal deficient”, and the three spectral indices in classical classification system can not be fully competent for selecting these subdwarfs. This conclusion is drawn by a series of experiments including the comparison of synthetic spectra with the templates, measuring atmospheric parameters for a subdwarf sample and combine with kinematics analysis, etc). According to the spectral index CaOH-CaH1 relationship and the empirical relationship on RPM diagram, we divide a spectroscopically identified subdwarfs into multiple groups and analyze the multi-population properties in this sample from their multi-dimensional parameters. The sample including subdwarfs from both LAMOST and SDSS combined with a control group of dwarfs selected from LAMOST, and estimate their atmospheric parameters fitting with the PHOENIX grid. The technical details of the whole fitting procedure of individual spectra is developed, including the use of different convolution kernels to smooth the synthetic spectra and the pseudocontinuum flux correction specifically implemented for the LAMOST data. The residuals of the fitting procedure can be used to determine the precision which depend on the S/N of the observed spectra. The accuracy of the process is evaluated from 167 M-type stars with measured atmospheric parameters from literature which have LAMOST spectra: and the bias and scatter of the comparison of parameters are 165$\pm$170 K for $T_{\text{eff}}$, 0.03$\pm$0.32 dex for gravity, and -0.17$\pm$0.41 dex for metallicity. A systematic bias in effective temperature exists for $T_{\text{eff}}$ lower than 3700 K while the gravity and metallicity both appear in fair agreement. The effective temperature scale produced by these detailed fits is reasonably consistent with those published by Rajpurohit et al. (2014) for the subdwarfs and Rajpurohit et al. (2018a) for the dwarfs. The following analysis are mainly dependent on the latter two parameters. For the sample of subdwarfs and dwarfs, apparent magnitudes, broad-band colors, and proper-motions are obtained from cross-matching these two sets with Gaia DR2, and the distances are from Bailer-Jones et al. (2018). The objects are selected so as to have acceptable quality flags in photometry, distance and accuracy. The spectroscopic “subdwarfs” do not entirely lie below the main-sequence but in an extended region which overlaps that occupied by the M dwarfs in the color-absolute magnitude diagram, while the objects that lie away (and below) from the main sequence are the ones with both larger gravity and lower metal abundance. They are the “genuine” subdwarfs, for which the luminosity class vi@ ((Jao et al., 2008)) is justified. Alternatively, some “subdwarfs” with gravity similar as that of the dwarfs or only slightly higher also behave like dwarfs in many aspects although they are medium metal- deficient through the definition using the CaH and TiO spectral indices, we have been able to sort out a significant subsample of objects which, in spite of average low metallicity, appear intricately mixed with the ordinary main- sequence dwarfs and also sharing their location in a reduced proper motion diagram. These “subdwarfs” should perhaps better be classified as “metal-poor dwarfs”. A provisional kinematic analysis of these samples show that the sdM population lying below the main-sequence cloud is dominated by halo stars, in which an implicit bimodal metallicity distribution appears, with a peak around $-$1.0 dex and a second peak around $-$1.5 dex. This group, which samples the galactic halo only in a very local volume (typically less than 1 kpc radius), contains a mix of objects of prograde and retrograde orbits, as revealed by their location in the Toomre diagram, without clear trend with the metallicity. The other subdwarfs, including those which we prefer to consider as “metal-poor dwarfs” appear as a mixed population of thick disk and thin disk members. If “lying below the main sequence” and “associated with halo kinematics” (large negative $V$ component and large dispersions in $U$, $V$ and $W$) define intrinsic features of the genuine subdwarfs belonging to Population ii@ stars, then both large gravity (average log $g$ = 5.2 to 5.4) $and$ low metal abundance should be essential attributes of them. The fact that the two properties must exist simultaneously needs to be underlined: we have only found a handful of halo objects among the other subdwarfs, even those which have a metallicity lower than $-$1.0 dex. The ordinary dwarfs populate in their vast majority the Galactic thin disk, as expected, with an average metallicity slightly lower than solar. A small fraction (less than 1%) form a set of kinematic outliers, halo and thick disk probable members. Finally, CaH1 index is a good indicator of gravity experimentally, which behaves better in separating subdwarfs from dwarfs than the composite index CaH2+CaH3. Though the contamination by low ratio of multiplicity (12.5$\pm$1.9%: Ziegler et al. 2015) and the parameter coupling effect would influence the accuracy of parameters measurement, a preliminary indicative conclusion can still be drawn from the analysis results above. Besides, accurate discrimination of subdwarfs has to be revisited with more information such as high-resolution spectra and multi-band data, and the estimates of parameters also need a more complete model grid (an independent alpha enrichment such as the analysis in Hejazi et al. 2020, more complete opacity and line lists, etc). The sub-population membership assessment of the subdwarfs found in numbers in the large area spectroscopic surveys, in turn needs careful selection effects control and detailed orbital calculation. ZS thanks Dr. Yan. H.L. for the thoughtful discussion about smoothing the synthetic spectra, and Dr. Allard, F. for her helpful private communication about PHOENIX models. This work is supported by National Key R&D Program of China(No. 2019YFA0405502), National Science Foundation of China (No. U1931209). Guoshoujing Telescope (the Large Sky Area Multi-Object Fiber Spectroscopic Telescope, LAMOST) is a National Major Scientific Project built by the Chinese Academy of Sciences. Funding for the project has been provided by the National Development and Reform Commission. LAMOST is operated and managed by the National Astronomical Observatories, Chinese Academy of Sciences. ## References * Ake & Greenstein (1980) Ake, T. B., & Greenstein, J. L. 1980, ApJ, 240, 859 * Allard (2016) Allard, F. 2016, in SF2A-2016: Proceedings of the Annual meeting of the French Society of Astronomy and Astrophysics, ed. C. Reylé, J. Richard, L. Cambrésy, M. Deleuil, E. Pécontal, L. Tresse, & I. Vauglin, 223–227 * Allard & Hauschildt (1995) Allard, F., & Hauschildt, P. H. 1995, ApJ, 445, 433 * Allard et al. (1997) Allard, F., Hauschildt, P. H., Alexander, D. R., & Starrfield, S. 1997, Annual Review of Astronomy and Astrophysics, 35, 137 * Allard et al. (2001) Allard, F., Hauschildt, P. H., Alexander, D. R., Tamanai, A., & Schweitzer, A. 2001, ApJ, 556, 357 * Allard et al. (2011) Allard, F., Homeier, D., & Freytag, B. 2011, Astronomical Society of the Pacific Conference Series, Vol. 448, Model Atmospheres From Very Low Mass Stars to Brown Dwarfs, ed. C. Johns-Krull, M. K. Browning, & A. A. West, 91 * Allard et al. (2012) —. 2012, Philosophical Transactions of the Royal Society of London Series A, 370, 2765 * Allard et al. (2013) —. 2013, Mem. Soc. Astron. Italiana, 84, 1053 * Asplund et al. (2009) Asplund, M., Grevesse, N., Sauval, A. J., & Scott, P. 2009, ARA&A, 47, 481 * Astropy Collaboration et al. (2013) Astropy Collaboration, Robitaille, T. P., Tollerud, E. J., et al. 2013, A&A, 558, A33 * Bai et al. (2016) Bai, Y., Luo, A. L., Comte, G., et al. 2016, Research in Astronomy and Astrophysics, 16, 107 * Bailer-Jones et al. (2018) Bailer-Jones, C. A. L., Rybizki, J., Fouesneau, M., Mantelet, G., & Andrae, R. 2018, AJ, 156, 58 * Baraffe et al. (2015) Baraffe, I., Homeier, D., Allard, F., & Chabrier, G. 2015, A&A, 577, A42 * Barrett et al. (2005) Barrett, P., Hunter, J., Miller, J. T., Hsu, J. C., & Greenfield, P. 2005, Astronomical Society of the Pacific Conference Series, Vol. 347, matplotlib – A Portable Python Plotting Package, ed. P. Shopbell, M. Britton, & R. Ebert, 91 * Bensby et al. (2003) Bensby, T., Feltzing, S., & Lundström, I. 2003, A&A, 410, 527 * Bochanski et al. (2013) Bochanski, J. J., Savcheva, A., West, A. A., & Hawley, S. L. 2013, AJ, 145, 40 * Bonaca et al. (2017) Bonaca, A., Conroy, C., Wetzel, A., Hopkins, P. F., & Kereš, D. 2017, ApJ, 845, 101 * Bond et al. (2010) Bond, N. A., Ivezić, Ž., Sesar, B., et al. 2010, ApJ, 716, 1 * Burgasser & Kirkpatrick (2006) Burgasser, A. J., & Kirkpatrick, J. D. 2006, ApJ, 645, 1485 * Caffau et al. (2011) Caffau, E., Ludwig, H. G., Steffen, M., Freytag, B., & Bonifacio, P. 2011, Solar Physics, 268, 255 * Carollo et al. (2010) Carollo, D., Beers, T. C., Chiba, M., et al. 2010, ApJ, 712, 692 * Castelli & Kurucz (2004) Castelli, F., & Kurucz, R. L. 2004, A&A, 419, 725 * Cui et al. (2012) Cui, X.-Q., Zhao, Y.-H., Chu, Y.-Q., et al. 2012, Research in Astronomy and Astrophysics, 12, 1197 * Dahn et al. (1995) Dahn, C. C., Liebert, J., Harris, H. C., & Guetter, H. H. 1995, in The Bottom of the Main Sequence - and Beyond, ed. C. G. Tinney, 239 * Dhital et al. (2012) Dhital, S., West, A. A., Stassun, K. G., et al. 2012, AJ, 143, 67 * Du et al. (2016) Du, B., Luo, A. L., Kong, X., et al. 2016, ApJS, 227, 27 * Edvardsson et al. (1993) Edvardsson, B., Andersen, J., Gustafsson, B., et al. 1993, A&A, 500, 391 * Faherty et al. (2009) Faherty, J. K., Burgasser, A. J., Cruz, K. L., et al. 2009, AJ, 137, 1 * Fuchs et al. (2009) Fuchs, B., Dettbarn, C., Rix, H.-W., et al. 2009, AJ, 137, 4149 * Gaia Collaboration et al. (2018a) Gaia Collaboration, Babusiaux, C., van Leeuwen, F., et al. 2018a, A&A, 616, A10 * Gaia Collaboration et al. (2018b) Gaia Collaboration, Brown, A. G. A., Vallenari, A., et al. 2018b, A&A, 616, A1 * Gilmore & Reid (1983) Gilmore, G., & Reid, N. 1983, MNRAS, 202, 1025 * Gizis (1997) Gizis, J. E. 1997, AJ, 113, 806 * Gunn et al. (1998) Gunn, J. E., Carr, M., Rockosi, C., et al. 1998, AJ, 116, 3040 * Gunn et al. (2006) Gunn, J. E., Siegmund, W. A., Mannery, E. J., et al. 2006, AJ, 131, 2332 * Gustafsson et al. (1975) Gustafsson, B., Bell, R. A., Eriksson, K., & Nordlund, A. 1975, A&A, 500, 67 * Gustafsson et al. (2008) Gustafsson, B., Edvardsson, B., Eriksson, K., et al. 2008, A&A, 486, 951 * Harris (1996) Harris, W. E. 1996, AJ, 112, 1487 * Hartwick et al. (1984) Hartwick, F. D. A., Cowley, A. P., & Mould, J. R. 1984, ApJ, 286, 269 * Hauschildt et al. (1999) Hauschildt, P. H., Allard, F., & Baron, E. 1999, ApJ, 512, 377 * Hejazi et al. (2020) Hejazi, N., Lépine, S., Homeier, D., Rich, R. M., & Shara, M. M. 2020, AJ, 159, 30 * Jao et al. (2008) Jao, W.-C., Henry, T. J., Beaulieu, T. D., & Subasavage, J. P. 2008, AJ, 136, 840 * Jao et al. (2011) Jao, W.-C., Henry, T. J., Subasavage, J. P., et al. 2011, AJ, 141, 117 * Jones (1972) Jones, E. M. 1972, ApJ, 173, 671 * Jones et al. (1996) Jones, H. R. A., Longmore, A. J., Allard, F., & Hauschildt, P. H. 1996, MNRAS, 280, 77 * Jurić et al. (2008) Jurić, M., Ivezić, Ž., Brooks, A., et al. 2008, ApJ, 673, 864 * Katz et al. (2019) Katz, D., Sartoretti, P., Cropper, M., et al. 2019, A&A, 622, A205 * Kesseli et al. (2019) Kesseli, A. Y., Kirkpatrick, J. D., Fajardo-Acosta, S. B., et al. 2019, AJ, 157, 63 * Kurucz (1973) Kurucz, R. L. 1973, PhD thesis, HARVARD UNIVERSITY. * Leggett et al. (1996) Leggett, S. K., Allard, F., Berriman, G., Dahn, C. C., & Hauschildt, P. H. 1996, ApJS, 104, 117 * Leggett et al. (2000) Leggett, S. K., Allard, F., Dahn, C., et al. 2000, ApJ, 535, 965 * Lépine et al. (2013) Lépine, S., Hilton, E. J., Mann, A. W., et al. 2013, AJ, 145, 102 * Lépine et al. (2003) Lépine, S., Rich, R. M., & Shara, M. M. 2003, AJ, 125, 1598 * Lépine et al. (2007) —. 2007, ApJ, 669, 1235 * Lépine & Scholz (2008) Lépine, S., & Scholz, R.-D. 2008, ApJ, 681, L33 * Lépine et al. (2004) Lépine, S., Shara, M. M., & Rich, R. M. 2004, ApJ, 602, L125 * Lindegren et al. (2018) Lindegren, L., Hernández, J., Bombrun, A., et al. 2018, A&A, 616, A2 * Liu et al. (2020) Liu, C., Fu, J., Shi, J., et al. 2020, arXiv e-prints, arXiv:2005.07210 * Luo et al. (2015) Luo, A.-L., Zhao, Y.-H., Zhao, G., et al. 2015, Research in Astronomy and Astrophysics, 15, 1095 * Luri et al. (2018) Luri, X., Brown, A. G. A., Sarro, L. M., et al. 2018, A&A, 616, A9 * McKinney et al. (2010) McKinney, W., et al. 2010, in Proceedings of the 9th Python in Science Conference, Vol. 445, Austin, TX, 51–56 * Mould (1976) Mould, J. R. 1976, ApJ, 210, 402 * Mould & McElroy (1978) Mould, J. R., & McElroy, D. B. 1978, ApJ, 220, 935 * Oliphant (2007) Oliphant, T. E. 2007, Computing in Science Engineering, 9, 10 * Rajpurohit et al. (2018a) Rajpurohit, A. S., Allard, F., Rajpurohit, S., et al. 2018a, A&A, 620, A180 * Rajpurohit et al. (2018b) Rajpurohit, A. S., Allard, F., Teixeira, G. D. C., et al. 2018b, A&A, 610, A19 * Rajpurohit et al. (2016) Rajpurohit, A. S., Reylé, C., Allard, F., et al. 2016, A&A, 596, A33 * Rajpurohit et al. (2014) —. 2014, A&A, 564, A90 * Reddy et al. (2003) Reddy, B. E., Tomkin, J., Lambert, D. L., & Allende Prieto, C. 2003, MNRAS, 340, 304 * Reid et al. (1995) Reid, I. N., Hawley, S. L., & Gizis, J. E. 1995, AJ, 110, 1838 * Salim & Gould (2002) Salim, S., & Gould, A. 2002, ApJ, 575, L83 * Sandage (1969) Sandage, A. 1969, ApJ, 158, 1115 * Sandage & Kowal (1986) Sandage, A., & Kowal, C. 1986, AJ, 91, 1140 * Savcheva et al. (2014) Savcheva, A. S., West, A. A., & Bochanski, J. J. 2014, ApJ, 794, 145 * Schönrich & Binney (2009) Schönrich, R., & Binney, J. 2009, MNRAS, 399, 1145 * Schönrich et al. (2010) Schönrich, R., Binney, J., & Dehnen, W. 2010, MNRAS, 403, 1829 * Smith et al. (2009) Smith, M. C., Evans, N. W., Belokurov, V., et al. 2009, MNRAS, 399, 1223 * Subasavage et al. (2005) Subasavage, J. P., Henry, T. J., Hambly, N. C., et al. 2005, AJ, 130, 1658 * Veltz et al. (2008) Veltz, L., Bienaymé, O., Freeman, K. C., et al. 2008, A&A, 480, 753 * Virtanen et al. (2019) Virtanen, P., Gommers, R., Oliphant, T. E., et al. 2019, arXiv e-prints, arXiv:1907.10121 * West et al. (2004) West, A. A., Hawley, S. L., Walkowicz, L. M., et al. 2004, AJ, 128, 426 * Woolf & Wallerstein (2005) Woolf, V. M., & Wallerstein, G. 2005, MNRAS, 356, 963 * Yee et al. (2017) Yee, S. W., Petigura, E. A., & von Braun, K. 2017, ApJ, 836, 77 * Yong & Lambert (2003a) Yong, D., & Lambert, D. L. 2003a, PASP, 115, 22 * Yong & Lambert (2003b) —. 2003b, PASP, 115, 796 * York et al. (2000) York, D. G., Adelman, J., Anderson, John E., J., et al. 2000, AJ, 120, 1579 * Zhang et al. (2016) Zhang, L., Pi, Q., Han, X. L., et al. 2016, New A, 44, 66 * Zhang et al. (2019) Zhang, S., Luo, A. L., Comte, G., et al. 2019, ApJS, 240, 31 * Zhang (2019) Zhang, Z. 2019, MNRAS, 489, 1423 * Zhang et al. (2017a) Zhang, Z. H., Homeier, D., Pinfield, D. J., et al. 2017a, MNRAS, 468, 261 * Zhang et al. (2017b) Zhang, Z. H., Pinfield, D. J., Gálvez-Ortiz, M. C., et al. 2017b, MNRAS, 464, 3040 * Zhang et al. (2018) Zhang, Z. H., Galvez-Ortiz, M. C., Pinfield, D. J., et al. 2018, MNRAS, 480, 5447 * Zhong et al. (2015) Zhong, J., Lépine, S., Hou, J., et al. 2015, AJ, 150, 42 * Ziegler et al. (2015) Ziegler, C., Law, N. M., Baranec, C., Riddle, R. L., & Fuchs, J. T. 2015, ApJ, 804, 30 * Zinn (1985) Zinn, R. 1985, ApJ, 293, 424 Table 4: Data description of the subdwarf catalog. Column Name | Unit | Description ---|---|--- ID | - | SpecObjID for SDSS objects and obsid for LAMOST objects specname | - | Name of spectrum ra | degrees | Right Ascension (J2000) dec | degrees | Declination (J2000) l | degrees | Galactic longitude b | degrees | Galactic latitude rv | km s-1 | Heliocentric radial velocity Teff | K | Estimated effective temperature logg | dex | Estimated surface gravity FeH | dex | Estimated overall metallicity snri | - | Average spectrum signal to noise ratio in i band spt | - | Spectral subtype CaOH | - | Spectral index of CaOH band CaH1 | - | Spectral index of CaH1 band CaH2 | - | Spectral index of CaH2 band CaH3 | - | Spectral index of CaH3 band TiO5 | - | Spectral index of TiO5 band ZetaL07 | - | The value of $\zeta$ parameter calibrated by Lépine et al. (2007) ZetaD12 | - | The value of $\zeta$ parameter calibrated by Dhital et al. (2012) ZetaL13 | - | The value of $\zeta$ parameter calibrated by Lépine et al. (2013) ZetaZ19 | - | The value of $\zeta$ parameter calibrated by Zhang et al. (2019) source | - | The source of spectrum (lamost/sdss) parallax | mas | Gaia DR2 parallax parallax_error | mas | Standard error of Gaia DR2 parallax pmra | mas yr-1 | Gaia DR2 proper motion in right ascension direction pmra_error | mas yr-1 | Standard error of Gaia DR2 proper motion in right ascension direction pmdec | mas yr-1 | Gaia DR2 proper motion in declination direction pmdec_error | mas yr-1 | Standard error of Gaia DR2 proper motion in declination direction astrometric_gof_al | - | Goodness of fit statistic of model wrt along-scan observations astrometric_excess_noise_sig | - | Significance of excess noise phot_g_mean_flux_over_error | - | G-band mean flux divided by its error phot_g_mean_mag | mag | Gaia DR2 G-band mean magnitude phot_bp_mean_flux_over_error | - | Integrated BP mean flux divided by its error phot_bp_mean_mag | mag | Gaia DR2 integrated BP mean magnitude phot_rp_mean_flux_over_error | - | Integrated RP mean flux divided by its error phot_rp_mean_mag | mag | Gaia DR2 integrated RP mean magnitude bp_rp | mag | Gaia DR2 BP-RP colour. ruwe | - | Renormalized Unit Weight Error associated to each Gaia source rest | pc | Estimated distance from Bailer-Jones et al. (2018), based on Gaia DR2 ResFlag | - | The Result Flag from Bailer-Jones et al. (2018) ModFlag | - | The Modality Flag from Bailer-Jones et al. (2018) flag | - | Set as 1 for the final sample used in kinematic part, see table 3 for details $V_{total}$ | km s-1 | Total spacial velocity $Z$ | pc | Vertical distance away from Galactic plane $U$ | km s-1 | Velocity component points towards the galactic center $V$ | km s-1 | The tangential velocity component rotating around galactic center $W$ | km s-1 | Velocity component points towards the north galactic pole Note. — The complete catalog can be accessed online222http://paperdata.china- vo.org/work/subdwarf-catalog.csv.
* Farsalinos et al., (2021) Farsalinos, K., Poulas, K., Kouretas, D., Vantarakis, A., Leotsinidis, M., Kouvelas, D., Docea, A. O., Kostoff, R., Gerotziafas, G. T., & Antoniou, M. N. (2021). Improved strategies to counter the COVID-19 pandemic: Lockdowns vs. primary and community healthcare. Toxicology Reports, 8, 1–9. * Federico & Ferrari, (2020) Federico, S. & Ferrari, G. (2020). Taming the spread of an epidemic by lockdown policies. Journal of Mathematical Economics, 93, 102453. * Fernández-Villaverde & Jones, (2020) Fernández-Villaverde, J. & Jones, C. M. (2020). Estimating and Simulating a SIRD Model of COVID-19 for Many Countries, States, and Cities. Working Paper 27128, National Bureau of Economic Research. * Forneron & Ng, (2018) Forneron, J.-J. & Ng, S. (2018). The ABC of simulation estimation with auxiliary statistics. Journal of Econometrics, 205(1), 112–139. * Frankel & Kotti, (2021) Frankel, J. A. & Kotti, R. (2021). The Virus, Vaccination, and Voting. Technical report, National Bureau of Economic Research. * Freed et al., (2020) Freed, M., Cubanski, J., Neuman, T., Kates, J., & Michaud, J. (2020). What share of people who have died of COVID-19 are 65 and older – and how does it vary by state? https://www.kff.org/coronavirus-covid-19/issue-brief/what-share-of-people-who-have-died-of-covid-19-are-65-and-older-and-how-does-it-vary-by-state/. Accessed: 2021-10-12. * Fruchterman & Reingold, (1991) Fruchterman, T. M. & Reingold, E. M. (1991). Graph drawing by force-directed placement. Software: Practice and Experience, 21(11), 1129–1164. * Galeotti et al., (2020) Galeotti, A., Golub, B., & Goyal, S. (2020). Targeting interventions in networks. Econometrica, 88(6), 2445–2471. * Ganesh et al., (2005) Ganesh, A., Massoulié, L., & Towsley, D. (2005). The effect of network topology on the spread of epidemics. In Proceedings IEEE 24th Annual Joint Conference of the IEEE Computer and Communications Societies., volume 2 (pp. 1455–1466). * Garetto et al., (2003) Garetto, M., Gong, W., & Towsley, D. (2003). Modeling malware spreading dynamics. In IEEE INFOCOM 2003. Twenty-second Annual Joint Conference of the IEEE Computer and Communications Societies (IEEE Cat. No. 03CH37428), volume 3 (pp. 1869–1879). * Gertler & Waldman, (1992) Gertler, P. J. & Waldman, D. M. (1992). Quality-adjusted cost functions and policy evaluation in the nursing home industry. Journal of Political Economy, 100(6), 1232–1256. * (53) Gollier, C. (2020a). Cost–benefit analysis of age-specific deconfinement strategies. Journal of Public Economic Theory, 22(6), 1746–1771. * (54) Gollier, C. (2020b). Pandemic economics: optimal dynamic confinement under uncertainty and learning. The Geneva Risk and Insurance Review, 45(2), 80–93. * Green et al., (2006) Green, D. M., Kiss, I. Z., & Kao, R. R. (2006). Parameterization of individual-based models: comparisons with deterministic mean-field models. Journal of Theoretical Biology, 239(3), 289–297. * Harris, (2020) Harris, J. E. (2020). The Subways Seeded the Massive Coronavirus Epidemic in New York City. Working Paper 27021, National Bureau of Economic Research. * Hashem Pesaran & Yang, (2021) Hashem Pesaran, M. & Yang, C. F. (2021). Matching theory and evidence on Covid-19 using a Stochastic Network SIR Model. CESifo Working Paper, (8695). * Heap et al., (2020) Heap, S. P. H., Koop, C., Matakos, K., Unan, A., & Weber, N. (2020). Valuating health vs wealth: The effect of information and how this matters for COVID-19 policymaking. Article, VoxEU.org. * Hethcote, (2000) Hethcote, H. W. (2000). The mathematics of infectious diseases. SIAM review, 42(4), 599–653. * Hethcote & Yorke, (2014) Hethcote, H. W. & Yorke, J. A. (2014). Gonorrhea transmission dynamics and control, volume 56. Springer. * Hong et al., (2021) Hong, B., Bonczak, B. J., Gupta, A., Thorpe, L. E., & Kontokosta, C. E. (2021). Exposure density and neighborhood disparities in COVID-19 infection risk. Proceedings of the National Academy of Sciences, 118(13). * IMF, (2020) IMF (2020). CHAPTER 1 THE GREAT LOCKDOWN. World Economic Outlook. * Ip, (2020) Ip, G. (2020). New thinking on Covid lockdowns: They’re overly blunt and costly. Wall Street Journal. Accessed: 2021-09-30. * Jeong et al., (2000) Jeong, H., Tombor, B., Albert, R., Oltvai, Z. N., & Barabási, A.-L. (2000). The large-scale organization of metabolic networks. Nature, 407(6804), 651–654. * John, (2020) John, M. (2020). How is the world dealing with lockdown? This poll has captured the public’s mood. Opinion, World Economic Forum. * Karaivanov, (2020) Karaivanov, A. (2020). A social network model of covid-19. Plos One, 15(10), e0240878. * Karanikolos & McKee, (2020) Karanikolos, M. & McKee, M. (2020). How comparable is COVID-19 mortality across countries? Eurohealth, 26(2), 45–50. * Keeling & Eames, (2005) Keeling, M. J. & Eames, K. T. (2005). Networks and epidemic models. Journal of the Royal Society Interface, 2(4), 295–307. * Kephart & White, (1992) Kephart, J. O. & White, S. R. (1992). Directed-graph epidemiological models of computer viruses. In Computation: the micro and the macro view (pp. 71–102). World Scientific. * Kermack & McKendrick, (1927) Kermack, W. O. & McKendrick, A. G. (1927). A contribution to the mathematical theory of epidemics. Proceedings of the Royal Society of London. Series A, containing papers of a mathematical and physical character, 115(772), 700–721. * Kermack & McKendrick, (1932) Kermack, W. O. & McKendrick, A. G. (1932). Contributions to the mathematical theory of epidemics. ii.—the problem of endemicity. Proceedings of the Royal Society of London. Series A, containing papers of a mathematical and physical character, 138(834), 55–83. * Knodel, (2020) Knodel, J. (2020). Texas Lt. Gov. Dan Patrick suggests he, other seniors willing to die to get economy going again. https://www.nbcnews.com/news/us-news/texas-lt-gov-dan-patrick-suggests-he-other-seniors-willing-n1167341. Accessed: 2020-10-14. * Kochańczyk & Lipniacki, (2021) Kochańczyk, M. & Lipniacki, T. (2021). Pareto-based evaluation of national responses to covid-19 pandemic shows that saving lives and protecting economy are non-trade-off objectives. Scientific reports, 11(1), 1–9. * Kuchler et al., (2020) Kuchler, T., Russel, D., & Stroebel, J. (2020). The Geographic Spread of COVID-19 Correlates with Structure of Social Networks as Measured by Facebook. Working Paper 26990, National Bureau of Economic Research. * Liljeros et al., (2001) Liljeros, F., Edling, C. R., Amaral, L. A. N., Stanley, H. E., & Åberg, Y. (2001). The web of human sexual contacts. Nature, 411(6840), 907–908. * Lloyd et al., (2006) Lloyd, A. L., Valeika, S., & Cintrón-Arias, A. (2006). Infection dynamics on small-world networks. Contemporary Mathematics, 410, 209–234. * Ma et al., (2021) Ma, L., Shapira, G., De Walque, D., Do, Q.-T., Friedman, J., & Levchenko, A. A. (2021). The Intergenerational Mortality Tradeoff of COVID-19 Lockdown Policies. Working Paper 28925, National Bureau of Economic Research. * National Center for Health Statistics, (2020) National Center for Health Statistics (2020). Provisional COVID-19 deaths by sex and age. https://data.cdc.gov/NCHS/Provisional-COVID-19-Deaths-by-Sex-and-Age/9bhg-hcku. Accessed: 2021-10-12. * Neelon et al., (2021) Neelon, B., Mutiso, F., Mueller, N. T., Pearce, J. L., & Benjamin-Neelon, S. E. (2021). Associations between governor political affiliation and COVID-19 cases, deaths, and testing in the US. American Journal of Preventive Medicine, 61(1), 115–119. * Newman, (2002) Newman, M. E. (2002). Spread of epidemic disease on networks. Physical Review E, 66(1), 016128. * Pastor-Satorras et al., (2015) Pastor-Satorras, R., Castellano, C., Van Mieghem, P., & Vespignani, A. (2015). Epidemic processes in complex networks. Reviews of Modern Physics, 87(3), 925. * Pastor-Satorras & Vespignani, (2001) Pastor-Satorras, R. & Vespignani, A. (2001). Epidemic spreading in scale-free networks. Physical Review Letters, 86(14), 3200. * Peng et al., (2020) Peng, L., Yang, W., Zhang, D., Zhuge, C., & Hong, L. (2020). Epidemic analysis of covid-19 in china by dynamical modeling. arXiv preprint arXiv:2002.06563. * Pindyck, (2020) Pindyck, R. S. (2020). COVID-19 and the welfare effects of reducing contagion. Working Paper 27121, National Bureau of Economic Research. * Pongou, (2010) Pongou, R. (2010). The economics of fidelity in network formation. PhD thesis, Brown University. * Pongou & Serrano, (2013) Pongou, R. & Serrano, R. (2013). Fidelity networks and long-run trends in HIV/AIDS gender gaps. American Economic Review, 103(3), 298–302. * Pongou & Serrano, (2016) Pongou, R. & Serrano, R. (2016). Volume of trade and dynamic network formation in two-sided economies. Journal of Mathematical Economics, 63, 147–163. * Pongou & Tondji, (2018) Pongou, R. & Tondji, J.-B. (2018). Valuing inputs under supply uncertainty: The bayesian Shapley value. Games and Economic Behavior, 108, 206–224. * Prem et al., (2020) Prem, K., Liu, Y., Russell, T. W., Kucharski, A. J., Eggo, R. M., Davies, N., Flasche, S., Clifford, S., Pearson, C. A., Munday, J. D., et al. (2020). The effect of control strategies to reduce social mixing on outcomes of the covid-19 epidemic in wuhan, china: a modelling study. The Lancet Public Health, 5(5), e261–e270. * Reyes-Santías et al., (2020) Reyes-Santías, F., Cordova-Arevalo, O., & Rivo-Lopez, E. (2020). Using flexible regression models for calculating hospital’s production functions. BMC Health Services Research, 20(1), 1–11. * Rice et al., (2021) Rice, B. L., Annapragada, A., Baker, R. E., Bruijning, M., Dotse-Gborgbortsi, W., Mensah, K., Miller, I. F., Motaze, N. V., Raherinandrasana, A., Rajeev, M., et al. (2021). Variation in SARS-CoV-2 outbreaks across sub-Saharan Africa. Nature Medicine, 27(3), 447–453. * Rodrigues, (2019) Rodrigues, F. A. (2019). Network centrality: an introduction. In A mathematical modeling approach from nonlinear dynamics to complex systems (pp. 177–196). Springer. * Schellekens & Sourrouille, (2020) Schellekens, P. & Sourrouille, D. (2020). The unreal dichotomy in COVID-19 mortality between high-income and developing countries. The Brookings Institution. * Seierstad & Sydsaeter, (1986) Seierstad, A. & Sydsaeter, K. (1986). Optimal control theory with economic applications. Elsevier North-Holland, Inc. * Smith Jr, (1993) Smith Jr, A. A. (1993). Estimating nonlinear time-series models using simulated vector autoregressions. Journal of Applied Econometrics, 8(S1), S63–S84. * Sorci et al., (2020) Sorci, G., Faivre, B., & Morand, S. (2020). Explaining among-country variation in COVID-19 case fatality rate. Scientific reports, 10(1), 1–11. * Stiglitz, (2020) Stiglitz, J. (2020). Joseph Stiglitz on priorities for the post-COVID economy. Opinion, World Economic Forum. * Van den Driessche & Watmough, (2002) Van den Driessche, P. & Watmough, J. (2002). Reproduction numbers and sub-threshold endemic equilibria for compartmental models of disease transmission. Mathematical Biosciences, 180(1-2), 29–48. * Van Mieghem et al., (2008) Van Mieghem, P., Omic, J., & Kooij, R. (2008). Virus spread in networks. IEEE/ACM Transactions On Networking, 17(1), 1–14. * Wadhera et al., (2020) Wadhera, R. K., Wadhera, P., Gaba, P., Figueroa, J. F., Maddox, K. E. J., Yeh, R. W., & Shen, C. (2020). Variation in COVID-19 hospitalizations and deaths across New York City boroughs. JAMA, 323(21), 2192–2195. * Wallinga et al., (2006) Wallinga, J., Teunis, P., & Kretzschmar, M. (2006). Using data on social contacts to estimate age-specific transmission parameters for respiratory-spread infectious agents. American Journal of Epidemiology, 164(10), 936–944. * Wang et al., (2003) Wang, Y., Chakrabarti, D., Wang, C., & Faloutsos, C. (2003). Epidemic spreading in real networks: An eigenvalue viewpoint. In 22nd IEEE International Symposium on Reliable Distributed Systems, 2003. Proceedings. (pp. 25–34). * Watts & Strogatz, (1998) Watts, D. J. & Strogatz, S. H. (1998). Collective dynamics of ‘small-world’ networks. Nature, 393(6684), 440–442. * Wichmann & Wichmann, (2020) Wichmann, B. & Wichmann, R. (2020). Nonparametric estimation of a primary care production function in urban brazil. Health Economics Review, 10(1), 1–10. * Yang, (2021) Yang, J. (2021). Impact of the COVID-19 pandemic on older adults in the U.S.- statistics & facts. https://www.statista.com/topics/8276/impact-of-the-covid-19-pandemic-on-older-adults-in-the-us/. Accessed: 2020-10-14. * Young, (2009) Young, H. P. (2009). Innovation diffusion in heterogeneous populations: Contagion, social influence, and social learning. American Economic Review, 99(5), 1899–1924. * Young, (2011) Young, H. P. (2011). The dynamics of social innovation. Proceedings of the National Academy of Sciences, 108(Supplement 4), 21285–21291.
# The rational hull of modules gangyong lee Gangyong Lee, Department of Mathematics Education, Chungnam National University Daejeon 34134, Republic of Korea e-mail: <EMAIL_ADDRESS> ###### Abstract. In this paper, we provide several new characterizations of the maximal right ring of quotients of a ring by using the relatively dense property. As a ring is embedded in its maximal right ring of quotients, we show that the endomorphism ring of a module is embedded into that of the rational hull of the module. In particular, we obtain new characterizations of rationally complete modules. The equivalent condition for the rational hull of the direct sum of modules to be the direct sum of the rational hulls of those modules under certain assumption is presented. For a right $H$-module $M$ where $H$ is a right ring of quotients of a ring $R$, we provide a sufficient condition to be $\text{End}_{R}(M)=\text{End}_{H}(M)$. Also, we give a condition for the maximal right ring of quotients of the endomorphism ring of a module to be the endomorphism ring of the rational hull of a module. ###### 2020 Mathematics Subject Classification: Primary 16D70; 16S50, Secondary 16D50 Key Words: rational hull, injective hull, maximal right ring of quotients ## 1\. Introduction The theory of rings of quotients has its origin in the work of Ø. Ore [11] and K. Asano [2] on the construction of the total ring of fractions, in the 1930’s and 40’s. But the subject did not really develop until the end of the 1950’s, when a number of important papers appeared (by R.E. Johnson [6], Y. Utumi [15], A.W. Goldie [5], J. Lambek [8] et al). In particular, Johnson(1951), Utumi(1956), and Findlay $\&$ Lambek(1958) have studied the maximal right ring of quotients of a ring which is an extended ring of the base ring. For the well-known example, the maximal right ring of quotients of integers is rational numbers. It is the same as the injective hull of integers. For a commutative ring, the classical right ring of quotients of a ring is to its total quotient ring as the maximal right ring of quotients of a ring is to its complete ring of quotients. As we know, the study of the rational hull of a module is the same as that of the maximal right ring of quotients in a different way. Also, like every module has the injective hull, it is known that every module has the rational hull in [4, Theorem 2.6]. Now, we introduce the definition of the rational hull of a module and present its well-known results, briefly. Let $M$ be a right $R$-module and $T=\text{End}_{R}(E(M))$. Put $\widetilde{E}(M)=\\{x\in E(M)|\,\vartheta(M)=0~{}\text{with}~{}\vartheta\in T~{}\Rightarrow\ \vartheta(x)=0\\}=\displaystyle\bigcap_{\begin{subarray}{c}M\subseteq\text{Ker}\vartheta\\\ \vartheta\in T\end{subarray}}\text{Ker}\vartheta=\mathbf{r}_{E(M)}\left(\mathbf{l}_{T}(M)\right).$ Then $\widetilde{E}(M)$ is the unique maximal rational extension of $M$. We call it the _rational hull_ of $M$. Also, it is known that $\mathbf{r}_{E(M)}(J(T))\leq\mathbf{r}_{E(M)}\left(\mathbf{l}_{T}(M)\right)=\widetilde{E}(M)$ because $\mathbf{l}_{T}(M)\subseteq J(T)$ where $J(T)=\\{\alpha\in T\,|\,\text{Ker}\alpha\leq^{\text{ess}}E(M)\\}$ is a Jacobson radical of a ring $T$. Note that the maximal right ring of quotients of $R$ is $Q(R)=\mathbf{r}_{E(R)}(\mathbf{l}_{H}(R))$ where $H=\text{End}_{R}(E(R))$ (see [8, Proposition 2]). After the necessary background history, results, and notations in this section, we provide several characterizations of the rational hull of a module in Section 2 (see Theorem 2.3 and Corollary 2.10). In addition, characterizations of rationally complete modules are presented. As a corollary, we obtain several new characterizations of the maximal right ring of quotients of a ring. In particular, we show that the endomorphism ring of a module is embedded into that of the rational hull of the module as the inherited property of its maximal right ring of quotients (see Theorem 2.15). Our focus, in Section 3, is on the question of when is the rational hull of the direct sum of modules the direct sum of the rational hulls of those modules. For $M=\bigoplus_{k\in\Lambda}M_{k}$, we prove that $\widetilde{E}(M)=\bigoplus_{k\in\Lambda}\widetilde{E}(M_{k})$ if and only if $M_{i}$ is $M_{j}$-dense in $\widetilde{E}(M_{i})$ for all $i,j\in\Lambda$ when either $R$ is right noetherian or $|\Lambda|$ is finite (see Theorem 3.6). In the last section, we obtain a condition to be $\text{End}_{R}(M)=\text{End}_{H}(M)$ where $H$ is a right ring of quotients of a ring $R$ (Theorem 4.1). This condition is called the _relatively dense property_ to a module. Also, we provide a sufficient condition for the maximal right ring of quotients of the endomorphism ring of a module to be the endomorphism ring of the rational hull of a module (see Theorem 4.5). Throughout this paper, $R$ is a ring with unity and $M$ is a unital right $R$-module. For a right $R$-module $M$, $S=\text{End}_{R}(M)$ denotes the endomorphism ring of $M$; thus $M$ can be viewed as a left $S$\- right $R$-bimodule. For $\varphi\in S$, $\text{Ker}\varphi$ and $\text{Im}\varphi$ stand for the kernel and the image of $\varphi$, respectively. The notations $N\leq M$, $N\leq^{\text{ess}}M$, $N\leq^{\text{den}}M$ or $N\leq^{\oplus}M$ mean that $N$ is a submodule, an essential submodule, a dense submodule, or a direct summand of $M$, respectively. By $E(M)$, $\widehat{M}$, and $\widetilde{E}(M)$ we denote the injective hull, the quasi-injective hull, and the rational hull of $M$, respectively, and $T=\text{End}_{R}(E(M))$. $Q(R)$ denotes the maximal right ring of quotients of $R$. The direct sum of $\Lambda$ copies of $M$ is denoted by $M^{(\Lambda)}$ where $\Lambda$ is an arbitrary index set. $\mathsf{CFM}_{\mathbb{N}}(F)$ denotes the $\mathbb{N}\times\mathbb{N}$ column finite matrix ring over a field $F$. By $\mathbb{Q}$, $\mathbb{Z}$, and $\mathbb{N}$ we denote the set of rational, integer, and natural numbers, respectively. $\mathbb{Z}_{n}$ denotes the $\mathbb{Z}$-module $\mathbb{Z}/n\mathbb{Z}$. For $x\in M$, $x^{-1}K=\\{r\in R\,|\,xr\in K\\}\leq R_{R}$ with a right $R$-submodule $K$ of $M$. We also denote $\mathbf{r}_{M}(I)=\\{m\in M\,|\,Im=0\\}$ for $I\leq S$ and $\mathbf{l}_{S}(N)=\\{\varphi\in S\,|\,\varphi N=0\\}$ for $N\leq M$. We give some properties of dense submodules. Recall that a submodule $N$ of $M$ is said to be _dense_ in $M$ if for any $x,0\neq y\in M,$ there exists $r\in R$ such that $xr\in N$ and $0\neq yr$. ###### Proposition 1.1 ([3, Proposition 1.3.6]). Let $N\leq M$ be right $R$-modules. Then the following conditions are equivalent: 1. (a) $N$ is dense in $M$; 2. (b) $\emph{Hom}_{R}(M/N,E(M))=0$; 3. (c) for any submodule $P$ such that $N\leq P\leq M$, $\emph{Hom}_{R}(P/N,M)=0$. ###### Proposition 1.2 ([7, Proposition 8.7]). Let $L,N$ be submodules of a right $R$-module $M$: 1. (i) If $L\leq^{\emph{den}}M$ and $N\leq^{\emph{den}}M$ then $L\cap N\leq^{\emph{den}}M$. 2. (ii) Let $L\leq V\leq M$. Then $L\leq^{\emph{den}}M$ if and only if $L\leq^{\emph{den}}V$ and $V\leq^{\emph{den}}M$. ###### Proposition 1.3 ([3, Proposition 1.3.7]). Let $M$ be a right $R$-module and $M\leq V\leq{E}(M)$. Then $M\leq^{\emph{den}}V$ if and only if $V\leq\widetilde{E}(M)$. We remind of some important characterizations of the rational hull of a module. ###### Proposition 1.4. The following hold true for a right $R$-module $M$ and $T=\emph{End}_{R}(E(M))$: 1. (i) _([9, Exercises 5])_ $\widetilde{E}(M)=\\{x\in E(M)|\,~{}\vartheta|_{M}=1_{M}~{}\emph{with}~{}\vartheta\in T\ ~{}\Rightarrow\ \vartheta(x)=x\\}$. 2. (ii) _([7, Proposition 8.16])_ $\widetilde{E}(M)=\\{x\in E(M)\,|\,~{}\forall y\in E(M)\\!\setminus\\!\\{0\\},~{}y\cdot x^{-1}M\neq 0\\}$. ## 2\. The rational hull of a module As the injective hull of a module $M$ is the minimal injective module including $M$, the next result shows that the rational hull of a module $M$ is the minimal rationally complete module including $M$. Recall that a right $R$-module $M$ is said to be _rationally complete_ if it has no proper rational (or dense) extensions, or equivalently $\widetilde{E}(M)=M$. Thus, the rational hull $\widetilde{E}(M)$ of a module $M$ is rationally complete. ###### Theorem 2.1. The following conditions are equivalent for right $R$-modules $M$ and $F$: 1. (a) $F$ is maximal dense over $M$; 2. (b) $F$ is rationally complete, and is dense over $M$; 3. (c) $F$ is minimal rationally complete, and is essential over $M$. Note that a right $R$-module $F$ is exactly the rational hull of a module $M$ if $F$ satisfies any one of the above equivalent conditions. ###### Proof. (a)$\Rightarrow$(b) From Proposition 1.3, it is easy to see that $F$ has no proper dense extension. So, $F$ is a rationally complete module. (b)$\Rightarrow$(c) Let $F^{\prime}$ be rationally complete such that $M\leq F^{\prime}\leq F$. Since $M\leq^{\text{den}}F$, from Proposition 1.2(ii) $M\leq^{\text{den}}F^{\prime}\leq^{\text{den}}F$. Thus, from Proposition 1.3 $F\leq^{\text{den}}\widetilde{E}(F^{\prime})=F^{\prime}$ because $F^{\prime}$ is rationally complete. Therefore $F=F^{\prime}$. (c)$\Rightarrow$(a) Let $F$ be minimal rationally complete over $M$. Since $F$ is essential over $M$, $M\leq F\leq E(M)$. Since $M\leq^{\text{den}}\widetilde{E}(M)$, $\text{Hom}_{R}(\widetilde{E}(M)/M,E(M))=0$. Also, since $E(F)=E(M)$, $\text{Hom}_{R}(\widetilde{E}(M)/M,E(F))=0$. From [7, Theorem 8.24], an inclusion map $\iota:M\rightarrow F$ extends to $\rho:\widetilde{E}(M)\rightarrow F$ as $F$ is rationally complete (see also Proposition 2.13). Note that $\rho$ is a monomorphism. Since $\widetilde{E}(M)$ is rationally complete and $F$ is minimal, $\widetilde{E}(M)=F$. ∎ The next example shows that the condition “essential over $M$” in Theorem 2.1(c) is not superfluous. ###### Example 2.2. Let $M=\mathbb{Z}$ and $F=\mathbb{Z}_{(p)}\oplus\mathbb{Z}_{p}$ be right $\mathbb{Z}$-modules where $\mathbb{Z}_{(p)}$ is the localization of $\mathbb{Z}$ at the prime ideal $(p)$. It is easy to see that $M$ is not essential in $F$, so $F$ is not a rational hull of $M$. In fact, $F$ is minimal rationally complete over $M$: From [7, Example 8.21], $F$ is rationally complete because $F$ is the rational hull of $L=\mathbb{Z}\oplus\mathbb{Z}_{p}$. It is enough to show that $F$ is minimal over $M$: Let $K$ be a rationally complete module such that $M\leq K\leq F$. Hence $1=\text{u.dim}(M)\leq\text{u.dim}(K)\leq\text{u.dim}(F)=2$. Assume that $\text{u.dim}(K)=1.$ Then $M\leq^{\text{ess}}K$, and hence $K$ is nonsingular since $M$ is nonsingular. Thus $M\leq^{\text{den}}K$, which implies that $K\cong\mathbb{Q}$ since $K$ is rationally complete and $\widetilde{E}(M)=\mathbb{Q}$. It follows that $\mathbb{Q}$ can be embedded into $F=\mathbb{Z}_{(p)}\oplus\mathbb{Z}_{p}$, a contradiction. Therefore, $\text{u.dim}(K)=2.$ Then $K\leq^{\text{ess}}F$, and hence $K\cap\mathbb{Z}_{p}\neq 0$. Thus $\mathbb{Z}_{p}\leq K$, which implies that $L=\mathbb{Z}\oplus\mathbb{Z}_{p}\leq K$. Note that $L\leq^{\text{den}}F$ since $F=\widetilde{E}(L)$. Hence $K\leq^{\text{den}}F$, so that $K=F$ due to the fact that $K$ is rationally complete. We provide another characterization for the rational hull of a module using the relatively dense property. A right ideal $I$ of a ring $R$ is called _relatively dense to a right $R$-module_ $M$ (or $M$-_dense_) in $R$ if for any $r\in R$ and $0\neq m\in M$, $m\cdot r^{-1}I\neq 0$. It is denoted by $I\leq^{\text{den}}_{M}R$. ###### Theorem 2.3. Let $M$ be a right $R$-module. Then $\widetilde{E}(M)=\\{x\in E(M)\,|\,x^{-1}M\leq^{\emph{den}}_{M}R\\}$. ###### Proof. Let $x\in\widetilde{E}(M)$ be arbitrary. Consider a right ideal $x^{-1}M\leq R$. Let $0\neq m\in M$ and $r\in R$. Since $M\leq^{\text{den}}\widetilde{E}(M)$, there exists $s\in R$ such that $ms\neq 0$ and $(xr)s=x(rs)\in M$, that is, $rs\in x^{-1}M$. Hence $x^{-1}M\leq^{\text{den}}_{M}R$. For the reverse inclusion, let $x\in E(M)$ such that $x^{-1}M\leq^{\text{den}}_{M}R$. For an arbitrary nonzero element $0\neq y\in E(M)$, it suffices to show that $y\cdot x^{-1}M\neq 0$. As $M\leq^{\text{ess}}E(M)$, $0\neq yr\in M$ for some $r\in R$. Since $x^{-1}M\leq^{\text{den}}_{M}R$, there exists $s\in R$ such that $yrs\neq 0$ and $rs\in x^{-1}M$. Hence $0\neq yrs\in y\cdot x^{-1}M$. Therefore $x\in\widetilde{E}(M)$. ∎ The next definition was shown in [4, pp79] as $N\leq M(K)$, so we call a submodule $N$ relatively dense to a module $K$ in a module $M$. (For details, see [17].) ###### Definition 2.4. A submodule $N$ of a right $R$-module $M$ is said to be _relatively dense to a right $R$-module_ $K$ (or $K$-_dense_) in $M$ if for any $m\in M$ and $0\neq x\in K$, $x\cdot m^{-1}N\neq 0$, denoted by $N\leq^{\text{den}}_{K}M$. Note that $N$ is $M$-dense in $M$ if and only if $N$ is dense in $M$. We provide some characterizations of the relative density property. One can compare the following characterizations to Proposition 1.1. The equivalence (a)$\Leftrightarrow$(c) in the following proposition is provided by [4, pp79]. ###### Proposition 2.5. The following are equivalent for right $R$-modules $M,K$ and $N\leq M$: 1. (a) $N$ is $K$-dense in $M$; 2. (b) $\emph{Hom}_{R}(M/N,E(K))=0$; 3. (c) for any submodule $P$ such that $N\leq P\leq M$, $\emph{Hom}_{R}(P/N,K)=0$. ###### Proof. (a)$\Rightarrow$(b) Assume that there exists $0\neq\alpha\in\text{Hom}_{R}(M,E(K))$ with $\alpha N=0$. Since $\alpha M\cap K\neq 0$, there exist $x\in M$ and $0\neq y\in K$ such that $\alpha(x)=y$. Since $N$ is $K$-dense in $M$, there exists $r\in R$ such that $xr\in N$ and $0\neq yr$. However, $0=\alpha(xr)=\alpha(x)r=yr\neq 0$, a contradiction. Hence $\text{Hom}_{R}(M/N,E(K))=0$. (b)$\Rightarrow$(c) Assume that for any submodule $P$ such that $N\leq P\leq M$, there exists $0\neq\eta\in\text{Hom}_{R}(P/N,K)$. Then by the injectivity of $E(K)$, we can extend $\eta$ to a nonzero homomorphism from $M/N$ to $E(K)$, a contradiction. Hence $\text{Hom}_{R}(P/N,K)=0$. (c)$\Rightarrow$(a) Assume that $y\cdot x^{-1}N=0$ for some $x\in M$ and $0\neq y\in K$. We define $\gamma:N+xR\rightarrow K$ given by $\gamma(n+xr)=yr$ for $n\in N$ and $r\in R$. It is easy to see that $\gamma$ is a well-defined $R$-homomorphism vanishing on $N$. Since $N\leq N+xR\leq M$, by hypothesis $0=\gamma(x)=y\neq 0$, a contradiction. Thus $N$ is $K$-dense in $M$. ∎ We obtain another characterization of the relative density property related to homomorphisms. ###### Proposition 2.6. Let $M,K$ be right $R$-modules. Then a submodule $N$ is $K$-dense in $M$ if and only if $\mathbf{l}_{H}(N)=0$ where $H=\emph{Hom}_{R}(M,E(K))$. ###### Proof. Suppose $N$ is $K$-dense in $M$. Assume that $0\neq\varphi\in H$ such that $\varphi N=0$. Then there exists $m\in M\setminus N$ such that $\varphi(m)\neq 0$. Since $\varphi(m)\in E(K)$, $0\neq\varphi(m)r\in K$ for some $r\in R$. Hence there exists $s\in R$ such that $mrs\in N$ and $\varphi(m)rs\neq 0$ because $N\leq^{\text{den}}_{K}M$. That yields a contradiction that $0\neq\varphi(m)rs=\varphi(mrs)\in\varphi N=0$. Therefore $\mathbf{l}_{H}(N)=0$. Conversely, assume that $x\cdot m^{-1}N=0$ for some $0\neq x\in K$ and $m\in M$. We define $\gamma:N+mR\rightarrow E(K)$ by $\gamma(n+mt)=xt$ for $n\in N$ and $t\in R$. Clearly, $\gamma$ is a nonzero $R$-homomorphism vanishing on $N$. Also, there exists $\overline{\gamma}:M\rightarrow E(K)$ such that $\overline{\gamma}|_{N+mR}=\gamma$. Since $0=\overline{\gamma}N$, $0\neq\overline{\gamma}\in\mathbf{l}_{H}(N)$, a contradiction. Therefore $x\cdot m^{-1}N\neq 0$. ∎ If $M=R$, the following result is directly provided. ###### Corollary 2.7 ([14, Proposition 1.1]). Let $K$ be a right $R$-module and $I$ be a right ideal of a ring $R$. Then $I$ is $K$-dense in $R$ if and only if $\mathbf{l}_{E(K)}(I)=0$. ###### Proposition 2.8. Let $K$ be a right $R$-module and $I$ be an ideal of a ring $R$. Then $\mathbf{l}_{K}(I)=0$ if and only if $\mathbf{l}_{E(K)}(I)=0$. ###### Proof. Since one direction is trivial, we need to show the other direction. Suppose $\mathbf{l}_{K}(I)=0$. Assume that $\mathbf{l}_{E(K)}(I)\neq 0$. Then there exists $0\neq x\in E(K)$ such that $xI=0$. Also, $0\neq xr\in K$ for some $r\in R$ because $K\leq^{\text{ess}}E(K)$. Since $xrI\subseteq xI=0$, $0\neq xr\in\mathbf{l}_{K}(I)$, a contradiction. Therefore $\mathbf{l}_{E(K)}(I)=0$. ∎ ###### Corollary 2.9 ([3, Proposition 1.3.11(iv)]). Let $I$ be an ideal of a ring $R$. Then $I\leq^{\emph{den}}R_{R}$ if and only if $\mathbf{l}_{R}(I)=0$. ###### Proof. The proof also follows from Corollary 2.7 and Proposition 2.8. ∎ Using Theorem 2.3 and Corollary 2.7, we obtain another characterization for the rational hull of a module. Also, using the characterization of the relatively dense property, new characterization for the rational hull of a module is provided. ###### Corollary 2.10. Let $M$ be a right $R$-module. Then the following statements hold true: 1. (i) _([14, Proposition 1.4(b)])_ $\widetilde{E}(M)=\\{x\in E(M)\,|\,\mathbf{l}_{E(M)}(x^{-1}M)=0\\}$. 2. (ii) $\widetilde{E}(M)=\\{x\in E(M)\,|\,\emph{Hom}_{R}(R/x^{-1}M,E(M))=0\\}$. ###### Proof. It directly follows from Theorem 2.3, Corollary 2.7, and Proposition 2.5. ∎ Several new characterizations for the maximal right ring of quotients of a ring are provided as the following. ###### Theorem 2.11. Let $R$ be a ring. Then the following statements hold true: 1. (i) A right ideal $I$ is dense in $R$ if and only if $\mathbf{l}_{E(R)}(I)=0$. 2. (ii) $Q(R)=\\{x\in E(R)\,|\,x^{-1}R\leq^{\emph{den}}R\\}$. 3. (iii) $Q(R)=\\{x\in E(R)\,|\,\mathbf{l}_{E(R)}(x^{-1}R)=0\\}$. 4. (iv) $Q(R)=\\{x\in E(R)\,|\,\emph{Hom}_{R}(R/x^{-1}R,E(R))=0\\}$. We give characterizations for a rationally complete module. ###### Theorem 2.12. The following conditions are equivalent for a right $R$-module $M$: 1. (a) $M$ is a rationally complete module; 2. (b) $\\{\overline{x}\in E(M)/M\,|\,\mathbf{l}_{\text{E}(\text{M})}(\mathbf{r}_{\text{R}}(\overline{x}))=0\\}=\overline{0}$; 3. (c) For any $I\leq^{\emph{den}}_{M}R$, $\varphi\in\emph{Hom}_{R}(I,M)$ can be uniquely extended to $\widetilde{\varphi}\in\emph{Hom}_{R}(R,M)$. ###### Proof. Take $A:=\\{\overline{x}\in E(M)/M\,|\,\mathbf{l}_{\emph{E}(\emph{M})}(\mathbf{r}_{\text{R}}(\overline{x}))=0\\}$. (a)$\Rightarrow$(b) Assume that $x\in E(M)\setminus M$ such that $\overline{x}\in A$. From Corollary 2.7, $\mathbf{r}_{\text{R}}(\overline{x})\leq^{\text{den}}_{M}R$. Since $\mathbf{r}_{\text{R}}(\overline{x})=x^{-1}M$, $x^{-1}M\leq^{\text{den}}_{M}R$. Hence from Theorem 2.3 $x\in\widetilde{E}(M)=M$ because $M$ is rationally complete, a contradiction. Therefore $A=\overline{0}$. (b)$\Rightarrow$(c) Assume to the contrary of the condition (c). For $I\leq^{\text{den}}_{M}R$, since $M\subseteq E(M)$, there exists $\varphi\in\text{Hom}_{R}(I,M)$ such that $\widetilde{\varphi}\in\text{Hom}_{R}(R,E(M))$, $\widetilde{\varphi}|_{I}=\varphi$, and $\widetilde{\varphi}(1)\notin M$. Since $\overline{0}\neq\widetilde{\varphi}(1)+M\in E(M)/M$ and $I\subseteq\mathbf{r}_{\text{R}}(\widetilde{\varphi}(1)+M)\leq^{\text{den}}_{M}R$, $\mathbf{l}_{\emph{E}(\emph{M})}(\mathbf{r}_{\text{R}}(\widetilde{\varphi}(1)+M))=0$ from Corollary 2.7, a contradiction that $A=\overline{0}$. Therefore $\varphi\in\text{Hom}_{R}(I,M)$ is extended to $\widetilde{\varphi}\in\text{Hom}_{R}(R,M)$. For the uniqueness, the proof is similar to that of Proposition 2.13. (c)$\Rightarrow$(a) Assume that $M$ is not rationally complete. Then there exists $x\in\widetilde{E}(M)\setminus M$ such that $x^{-1}M\leq^{\text{den}}_{M}R$ from Theorem 2.3. Define $\varphi:x^{-1}M\rightarrow M$ given by $\varphi(r)=xr$. By hypothesis, $\widetilde{\varphi}(1)=x1=x\in M$, a contradiction. Therefore $M$ is rationally complete. ∎ Next, as a ring is embedding into its maximal right ring of quotients, we provide the relationship between the endomorphism rings of a module and its rational hull. ###### Proposition 2.13. Let $M$ and $K$ be right $R$-modules. For any $N\leq^{\emph{den}}_{K}M$, $\varphi\in\emph{Hom}_{R}(N,K)$ is uniquely extended to $\widetilde{\varphi}\in\emph{Hom}_{R}(M,\widetilde{E}(K))$ and $\widetilde{\varphi}|_{N}=\varphi$. In addition, $\varphi N\leq^{\emph{den}}\widetilde{\varphi}M$. ###### Proof. (Existence) Let $\varphi\in\text{Hom}_{R}(N,K)$ be arbitrary. Then there exists $\widetilde{\varphi}\in\text{Hom}_{R}(M,E(K))$ such that $\widetilde{\varphi}|_{N}=\varphi$. Since $\widetilde{\varphi}$ induces a surjection from $M/N$ to $(\widetilde{\varphi}M+\widetilde{E}(K))/\widetilde{E}(K)$ and $\text{Hom}_{R}(M/N,E(K))=0$ (see Proposition 2.5), $\text{Hom}_{R}\left(\frac{\widetilde{\varphi}M+\widetilde{E}(K)}{\widetilde{E}(K)},E(K)\right)=0$. Hence $\widetilde{E}(K)\leq^{\text{den}}\widetilde{\varphi}M+\widetilde{E}(K)$ by Proposition 1.1. As $\widetilde{E}(K)$ is rationally complete, $\widetilde{\varphi}M\subseteq\widetilde{E}(K)$. (Uniqueness) Suppose $\widetilde{\varphi}$ and $\widetilde{\psi}$ are in $\text{Hom}_{R}(M,\widetilde{E}(K))$ such that $\widetilde{\varphi}|_{N}=\widetilde{\psi}|_{N}$. It is enough to show that $\widetilde{\varphi}=\widetilde{\psi}$. Assume that $\widetilde{\varphi}(x)\neq\widetilde{\psi}(x)$ for some $x\in M$. Take $0\neq y=(\widetilde{\varphi}-\widetilde{\psi})(x)\in\widetilde{E}(K)$. Thus, there exists $r\in R$ such that $0\neq yr\in K$. Since $N\leq^{\text{den}}_{K}M$, there exists $s\in R$ such that $xrs\in N$ and $yrs\neq 0$. This yields a contradiction that $0\neq yrs=(\widetilde{\varphi}-\widetilde{\psi})(xrs)=(\widetilde{\varphi}|_{N}-\widetilde{\psi}|_{N})(xrs)=0$. Therefore $\widetilde{\varphi}=\widetilde{\psi}$. In addition, let $x_{1}\in\widetilde{\varphi}M$ and $0\neq x_{2}\in\widetilde{\varphi}M$. Then $\widetilde{\varphi}(m_{1})=x_{1},\widetilde{\varphi}(m_{2})=x_{2}$ for some $m_{1},m_{2}\in M$. As $\widetilde{\varphi}M\subseteq\widetilde{E}(K)$, $0\neq x_{2}r\in K$ for some $r\in R$. Since $N\leq^{\text{den}}_{K}M$ and $m_{1}r\in M$, there exists $s\in R$ such that $m_{1}rs\in N$ and $0\neq x_{2}rs$. Thus $x_{1}rs=\widetilde{\varphi}(m_{1}rs)\in\varphi N$ and $0\neq x_{2}rs$. Therefore $\varphi N\leq^{\text{den}}\widetilde{\varphi}M$. ∎ Note that the dense property implies the essential property, however the relatively dense property does not imply the essential property in general: See $\mathbb{Z}_{p}\leq^{\text{den}}_{\mathbb{Z}}\mathbb{Z}_{p}\oplus\mathbb{Z}_{p}$ but $\mathbb{Z}_{p}\nleq^{\text{ess}}\mathbb{Z}_{p}\oplus\mathbb{Z}_{p}$ as a $\mathbb{Z}$-module. However, Proposition 2.13 shows that $\varphi N\leq^{\text{den}}\widetilde{\varphi}M$ when $N\leq^{\text{den}}_{K}M$ for any $\varphi\in\text{Hom}_{R}(N,K)$. As a corollary, we have a generalized result of Theorem 2.12((a)$\Rightarrow$(b)). ###### Corollary 2.14. Let $M$ be a right $R$-module. If $K$ is rationally complete, then for any $N\leq^{\emph{den}}_{K}M$, $\varphi\in\emph{Hom}_{R}(N,K)$ is uniquely extended to $\widetilde{\varphi}\in\emph{Hom}_{R}(M,K)$ and $\widetilde{\varphi}|_{N}=\varphi$. ###### Theorem 2.15. Let $M$ be a right $R$-module. Then $\emph{End}_{R}(M)$ is considered as a subring of $\emph{End}_{R}(\widetilde{E}(M))$. ###### Proof. Since $M\leq^{\text{den}}\widetilde{E}(M)$, from Proposition 2.13 $\varphi\in\text{End}_{R}(M)$ can be uniquely extended to $\widetilde{\varphi}\in\text{End}_{R}(\widetilde{E}(M))$ because $\text{End}_{R}(M)\subseteq\text{Hom}_{R}(M,\widetilde{E}(M))$. Thus we have a one-to-one correspondence between $\text{End}_{R}(M)$ and $\\{\widetilde{\varphi}\in\text{End}_{R}(\widetilde{E}(M))\,|\,\widetilde{\varphi}|_{M}={\varphi}\in\text{End}_{R}(M)\\}$ given by $\Omega(\varphi)=\widetilde{\varphi}$. We need to check that $\Omega$ is a ring homomorphism. (i) Since $\Omega(\varphi+\psi)|_{M}=(\widetilde{\varphi+\psi})|_{M}=\varphi+\psi=\Omega(\varphi)|_{M}+\Omega(\psi)|_{M}=(\Omega(\varphi)+\Omega(\psi))|_{M}$, from the uniqueness of Proposition 2.13 we have $\Omega(\varphi+\psi)=\Omega(\varphi)+\Omega(\psi)$. (ii) Since $\Omega{(\varphi\circ\psi)}|_{M}=(\widetilde{\varphi\circ\psi})|_{M}=\varphi\circ\psi=\Omega(\varphi)|_{M}\circ\Omega(\psi)|_{M}=(\Omega(\varphi)\circ\Omega(\psi))|_{M}$ because $\Omega(\varphi)|_{M}\leq M$, from the uniqueness of Proposition 2.13 we have $\Omega(\varphi\circ\psi)=\Omega(\varphi)\circ\Omega(\psi)$. Thus $\text{End}_{R}(M)$ is isomorphic to a subring of $\text{End}_{R}(\widetilde{E}(M))$. Therefore we consider $\text{End}_{R}(M)$ as a subring of $\text{End}_{R}(\widetilde{E}(M))$. ∎ We conclude this section with results for the rational hulls of quasi- continuous modules and quasi-injective modules. ###### Theorem 2.16. The following statements hold true for a module $M$: 1. (i) If $M$ is a quasi-continuous module then $\widetilde{E}(M)$ is a quasi- continuous module. 2. (ii) If $M$ is a quasi-injective module then $\widetilde{E}(M)$ is a quasi- injective module. ###### Proof. (i) Let $T=\text{End}_{R}(E(\widetilde{E}(M)))=\text{End}_{R}(E(M))$. From [10, Theorem 2.8], we need to show that $f\widetilde{E}(M)\leq\widetilde{E}(M)$ for all idempotents $f^{2}=f\in T$: Assume that $f\widetilde{E}(M)\nleq\widetilde{E}(M)$ for some idempotent $f^{2}=f\in T$. Then there exists $x\in\widetilde{E}(M)$ such that $f(x)\notin\widetilde{E}(M)$. Thus, there exists $g\in T$ such that $gM=0$ and $gf(x)\neq 0$. Since $gf(x)\in E(M)$, there exists $r\in R$ such that $0\neq gf(xr)\in\widetilde{E}(M)$. Thus, as $M\leq^{\text{den}}\widetilde{E}(M)$ and $xr\in\widetilde{E}(M)$, there exists $s\in R$ such that $0\neq gf(xrs)$ and $xrs\in M$. Note that $fM\leq M$ for all idempotents $f^{2}=f\in T$ because $M$ is quasi-continuous. However, $0\neq gf(xrs)\in gfM\leq gM=0$, a contradiction. Therefore $\widetilde{E}(M)$ is a quasi-continuous module. (ii) The proof is similar to that of part (i) by using [10, Corollary 1.14]. ∎ ###### Remark 2.17 ([1, Theorem 5.3]). The rational hull of every extending module is an extending module. Note that if $M$ is an injective module then $M=\widetilde{E}(M)$ (see [7, Examples 8.18(1)]). The next examples exhibit that the converses of Theorem 2.16 and Remark 2.17 do not hold true. ###### Example 2.18. (i) Consider $\mathbb{Z}$ as a $\mathbb{Z}$-module. Then $\widetilde{E}(\mathbb{Z})=\mathbb{Q}$ is (quasi-)injective, while $\mathbb{Z}$ is not quasi-injective. (ii)([10, Example 2.9]) Consider a ring $R=\left(\begin{smallmatrix}F&F\\\ 0&F\end{smallmatrix}\right)$ where $F$ is a field. Then $\widetilde{E}(R_{R})=\left(\begin{smallmatrix}F&F\\\ F&F\end{smallmatrix}\right)$ is injective (hence, quasi-continuous), while $R_{R}$ is not quasi-continuous. (iii) Consider $\mathbb{Z}^{(\mathbb{N})}$ as a $\mathbb{Z}$-module. Then $\widetilde{E}(\mathbb{Z}^{(\mathbb{N})})=\mathbb{Q}^{(\mathbb{N})}$ is injective (hence, extending), while $\mathbb{Z}^{(\mathbb{N})}$ is not extending. ###### Corollary 2.19. The maximal right ring of quotients of a quasi-continuous ring is also a quasi-continuous ring. ###### Remark 2.20 ([7, Exercises 13.8]). The maximal right ring of quotients of a simple (resp., prime, semiprime) ring is also a simple (resp., prime, semiprime) ring. ###### Open Question 1. Is the rational hull of a continuous module always a continuous module? ## 3\. Direct sum of rational hulls of modules As we know, the injective hull of the direct sum of two modules is the direct sum of the injective hulls of each module without any condition. However, the rational hull case is different from the injective hull case. In this section, we discuss the condition for the rational hull of the direct sum of two modules to be the direct sum of the rational hulls of those modules. The next example shows that the rational hull of the direct sum of two modules is not the direct sum of the rational hulls of each module, in general. ###### Example 3.1. Consider $M=\mathbb{Z}\oplus\mathbb{Z}_{p}$ as a $\mathbb{Z}$-module where $p$ is prime. Then $\widetilde{E}(\mathbb{Z})=\mathbb{Q}$ and $\widetilde{E}(\mathbb{Z}_{p})=\mathbb{Z}_{p}$. However, by [7, Example 8.21] $\widetilde{E}(M)=\mathbb{Z}_{(p)}\oplus\mathbb{Z}_{p}\neq\mathbb{Q}\oplus\mathbb{Z}_{p}$ where $\mathbb{Z}_{(p)}=\\{\frac{m}{n}\in\mathbb{Q}\,|\,{m,n}\in\mathbb{Z},(n,p)=1\\}$. Hence $M$ is not a dense submodule of $\mathbb{Q}\oplus\mathbb{Z}_{p}$: For $(\frac{1}{p},\overline{0})$ and $0\neq(0,\overline{1})\in\mathbb{Q}\oplus\mathbb{Z}_{p}$, there is no $n\in\mathbb{Z}$ such that $n(\frac{1}{p},\overline{0})\in\mathbb{Z}\oplus\mathbb{Z}_{p}$ and $n(0,\overline{1})\neq 0$. ###### Proposition 3.2. Let $M=\bigoplus_{k\in\Lambda}M_{k}$ where $M_{k}$ be a right $R$-module and $\Lambda$ is any index set. If either $R$ is right noetherian or $|\Lambda|$ is finite, then $\widetilde{E}(M)\leq\bigoplus_{k\in\Lambda}\widetilde{E}(M_{k})$. ###### Proof. Suppose $0\neq m\in\widetilde{E}(M)$. Since $\widetilde{E}(M)\subseteq E(M)=\oplus_{k\in\Lambda}E(M_{k})$ because $R$ is right noetherian or $|\Lambda|$ is finite, there exists $\ell\in\mathbb{N}$ such that $m\in\oplus_{i=1}^{\ell}E(M_{i})$. Thus, $m=(m_{1},\dots,m_{\ell})$ where $m_{i}\in E(M_{i})$. Since $(0,\dots,0,y_{i},0,\dots,0)\cdot m^{-1}M\neq 0$ for all $0\neq y_{i}\in E(M_{i})$ and $m^{-1}M=m_{1}^{-1}M_{1}\cap\cdots\cap m_{\ell}^{-1}M_{\ell}$, $y_{i}\cdot m_{i}^{-1}M_{i}\neq 0$ for all $0\neq y_{i}\in E(M_{i})$. Thus, $m_{i}\in\widetilde{E}(M_{i})$ for all $1\leq i\leq\ell$ from Proposition 1.4. So, $m=(m_{1},\dots,m_{\ell})\in\oplus_{i=1}^{\ell}\widetilde{E}(M_{i})\subseteq\oplus_{k\in\Lambda}\widetilde{E}(M_{k})$. Therefore $\widetilde{E}(M)\leq\oplus_{k\in\Lambda}\widetilde{E}(M_{k})$. ∎ ###### Remark 3.3. Example 3.1 illustrates Proposition 3.2 because $R=\mathbb{Z}$ is a noetherian ring, that is, $\widetilde{E}(\mathbb{Z}\oplus\mathbb{Z}_{p})=\mathbb{Z}_{(p)}\oplus\mathbb{Z}_{p}\lneq\mathbb{Q}\oplus\mathbb{Z}_{p}=\widetilde{E}(\mathbb{Z})\oplus\widetilde{E}(\mathbb{Z}_{p})$. However, Example 3.7 shows that the condition “either $R$ is right noetherian or $|\Lambda|$ is finite” is not superfluous because $\widetilde{E}(\oplus_{k\in\Lambda}\mathbb{Z}_{2})=\prod_{k\in\Lambda}\mathbb{Z}_{2}\gneq\oplus_{k\in\Lambda}\mathbb{Z}_{2}=\oplus_{k\in\Lambda}\widetilde{E}(\mathbb{Z}_{2})$ with a non-noetherian ring $R=\langle\oplus_{k\in\Lambda}\mathbb{Z}_{2},1\rangle$. To get the reverse inclusion of Proposition 3.2, first we provide the properties of the relatively dense property. ###### Lemma 3.4. Let $N\leq M$ and $K_{i}$ be right $R$-modules for all $i\in\Lambda$. Then $N$ is $K_{i}$-dense in $M$ for all $i\in\Lambda$ if and only if $N$ is $\bigoplus_{i\in\Lambda}K_{i}$-dense in $M$ if and only if $N$ is $\bigoplus_{i\in\Lambda}\widetilde{E}(K_{i})$-dense in $M$. ###### Proof. Let $P$ be any submodule such that $N\leq P\leq M$. Since $N$ is $K_{i}$-dense in $M$, $\text{Hom}_{R}(P/N,K_{i})=0$ for all $i\in\Lambda$ from Proposition 2.5. Consider the sequence $0\rightarrow\oplus_{i\in\Lambda}K_{i}\rightarrow\prod_{i\in\Lambda}K_{i}$. Then we have $0\rightarrow\text{Hom}_{R}(P/N,\oplus_{i\in\Lambda}K_{i})\rightarrow\text{Hom}_{R}(P/N,\prod_{i\in\Lambda}K_{i})\cong\prod_{i\in\Lambda}(\text{Hom}_{R}(P/N,K_{i})=0$. Thus $\text{Hom}_{R}(P/N,\oplus_{i\in\Lambda}K_{i})=0$. Therefore $N$ is $\oplus_{i\in\Lambda}K_{i}$-dense in $M$ from Proposition 2.5. Conversely, since $\text{Hom}_{R}(P/N,\oplus_{i\in\Lambda}K_{i})=0$, $\text{Hom}_{R}(P/N,K_{i})=0$ for each $i\in\Lambda$. Hence $N$ is $K_{i}$-dense in $M$ for all $i\in\Lambda$. For the second equivalence, since $E(\widetilde{E}(K_{i}))=E(K_{i})$, from Proposition 2.5 it is easy to see that $N$ is $K_{i}$-dense in $M$ if and only if $N$ is $\widetilde{E}(K_{i})$-dense in $M$, for all $i\in\Lambda$. After using the first equivalence, we have the second equivalence. ∎ Using Lemma 3.4, we obtain a characterization for $\oplus_{k\in\Lambda}N_{k}$ to be a dense submodule of $\oplus_{k\in\Lambda}M_{k}$ where $N_{i}$ is a submodule of $M_{i}$ for each $i\in\Lambda$. ###### Proposition 3.5. Let $N_{i}\leq M_{i}$ be right $R$-modules for all $i\in\Lambda$ where $\Lambda$ is any index set. Let $N=\bigoplus_{k\in\Lambda}N_{k}$ and $M=\bigoplus_{k\in\Lambda}M_{k}$. Then $N\leq^{\emph{den}}M$ if and only if $N_{i}$ is $M_{j}$-dense in $M_{i}$ for all $i,j\in\Lambda$. ###### Proof. Suppose $N\leq^{\text{den}}M$. Then $N$ is $M$-dense in $M$ by the definition. From Lemma 3.4 $N$ is $M_{j}$-dense in $M$ for all $j\in\Lambda$. Let $x_{i}\in M_{i}$ and $0\neq y_{j}\in M_{j}$ be arbitrary for each $i,j\in\Lambda$. Since $(0,\dots,0,x_{i},0,\dots)\in M$ and $0\neq y_{j}\in M_{j}$, there exists $r\in R$ such that $(0,\dots,0,x_{i},0,\dots)r\in N$ and $y_{j}r\neq 0$. Since $x_{i}r\in N_{i}$ and $y_{j}r\neq 0$, $N_{i}$ is $M_{j}$-dense in $M_{i}$ for all $i,j\in\Lambda$. Conversely, suppose $N_{i}$ is $M_{j}$-dense in $M_{i}$ for all $i,j\in\Lambda$. From Lemma 3.4, $N_{i}$ is $\oplus_{k\in\Lambda}M_{k}$-dense in $M_{i}$ for all $i\in\Lambda$. Let $x\in M$ and $0\neq y\in M$ be arbitrary. Then there exists $\ell\in\mathbb{N}$ such that $x=(x_{1},\dots,x_{\ell})\in\oplus_{k=1}^{\ell}M_{k}\leq M$. Since $N_{1}$ is $M$-dense in $M_{1}$, there exists $r_{1}\in R$ such that $x_{1}r_{1}\in N_{1}$ and $0\neq yr_{1}\in M$. Also, since $N_{2}$ is $M$-dense in $M_{2}$, there exists $r_{2}\in R$ such that $x_{2}r_{1}r_{2}\in N_{2}$ and $0\neq yr_{1}r_{2}\in M$. By the similar processing, we have $r=r_{1}r_{2}\cdots r_{\ell}\in R$ such that $xr\in\oplus_{k=1}^{\ell}N_{k}\leq N$ and $yr\neq 0$. Therefore $N\leq^{\text{den}}M$. ∎ From Propositions 3.2 and 3.5, we have a characterization for the rational hull of the direct sum of modules to be the direct sum of the rational hulls of each module. ###### Theorem 3.6. Let $M=\bigoplus_{k\in\Lambda}M_{k}$ where $M_{k}$ is a right $R$-module and $\Lambda$ is any index set. If either $R$ is right noetherian or $|\Lambda|$ is finite, then $\widetilde{E}(M)=\bigoplus_{k\in\Lambda}\widetilde{E}(M_{k})$ if and only if $M_{i}$ is $M_{j}$-dense in $\widetilde{E}(M_{i})$ for all $i,j\in\Lambda$. ###### Proof. Suppose $\widetilde{E}(M)=\oplus_{k\in\Lambda}\widetilde{E}(M_{k})$. Since $M\leq^{\text{den}}\oplus_{k\in\Lambda}\widetilde{E}(M_{k})$, from Proposition 3.5 $M_{i}$ is $\widetilde{E}(M_{j})$-dense in $\widetilde{E}(M_{i})$ for all $i,j\in\Lambda$. Thus, $M_{i}$ is $M_{j}$-dense in $\widetilde{E}(M_{i})$ for all $i,j\in\Lambda$ from Lemma 3.4. Conversely, suppose $M_{i}$ is $M_{j}$-dense in $\widetilde{E}(M_{i})$ for all $i,j\in\Lambda$. Then $M_{i}$ is $\widetilde{E}(M_{j})$-dense in $\widetilde{E}(M_{i})$ for all $i,j\in\Lambda$ from Lemma 3.4. Thus, from Proposition 3.5 $M\leq^{\text{den}}\oplus_{k\in\Lambda}\widetilde{E}(M_{k})$. Hence $\oplus_{k\in\Lambda}\widetilde{E}(M_{k})\leq\widetilde{E}(M)$ from Proposition 1.3. Also, from Proposition 3.2 $\widetilde{E}(M)\leq\oplus_{k\in\Lambda}\widetilde{E}(M_{k})$. Therefore $\widetilde{E}(M)=\oplus_{k\in\Lambda}\widetilde{E}(M_{k})$. ∎ The next examples show that the condition “$R$ is right noetherian or $|\Lambda|$ is finite” in Theorem 3.6 is not superfluous. ###### Example 3.7. (i) Let $R=\langle\oplus_{k\in\Lambda}\mathbb{Z}_{2},1\rangle$ and $M=\oplus_{k\in\Lambda}M_{k}$ where $M_{k}=\mathbb{Z}_{2}$. Note that $R$ is not noetherian. Since $\mathbb{Z}_{2}$ is an injective $R$-module, $\widetilde{E}(\mathbb{Z}_{2})=\mathbb{Z}_{2}$. Thus $M_{i}$ is $M_{j}$-dense in $\widetilde{E}(M_{i})$ for all $i,j\in\Lambda$. However, $\widetilde{E}(\oplus_{k\in\Lambda}\mathbb{Z}_{2})=\prod_{k\in\Lambda}\mathbb{Z}_{2}\gneq\oplus_{k\in\Lambda}\mathbb{Z}_{2}=\oplus_{k\in\Lambda}\widetilde{E}(\mathbb{Z}_{2})$. (ii) Let $R=\\{(a_{k})\in\prod_{k\in\Lambda}\mathbb{Z}\,|\,a_{k}~{}\text{is eventually constant}\\}$ and $M=\oplus_{k\in\Lambda}\mathbb{Z}$. Note that $R$ is not noetherian. Then $\widetilde{E}(\mathbb{Z})=\mathbb{Q}$ and $\mathbb{Z}$ is $\mathbb{Z}$-dense in $\widetilde{E}(\mathbb{Z})$. However, $\widetilde{E}(\oplus_{k\in\Lambda}\mathbb{Z})=\prod_{k\in\Lambda}\mathbb{Q}\gneq\oplus_{k\in\Lambda}\mathbb{Q}=\oplus_{k\in\Lambda}\widetilde{E}(\mathbb{Z})$. The next example illustrates Theorem 3.6. ###### Example 3.8. Consider $M=\mathbb{Z}\oplus\mathbb{Z}_{p}$ as a $\mathbb{Z}$-module where $p$ is prime. Then $\mathbb{Z}_{p}$ is $\mathbb{Z}$-dense in $\widetilde{E}(\mathbb{Z}_{p})=\mathbb{Z}_{p}$, but $\mathbb{Z}$ is not $\mathbb{Z}_{p}$-dense in $\mathbb{Q}$ because for $\frac{1}{p}\in\mathbb{Q},\overline{1}\in\mathbb{Z}_{p}$, there is no element $t\in\mathbb{Z}$ such that $t\frac{1}{p}\in\mathbb{Z}$ and $t\overline{1}\neq 0$. Thus, from Theorem 3.6 $\widetilde{E}(M)=\mathbb{Z}_{(p)}\oplus\mathbb{Z}_{p}\lneq\mathbb{Q}\oplus\mathbb{Z}_{p}=\widetilde{E}(\mathbb{Z})\oplus\widetilde{E}(\mathbb{Z}_{p})$. (See Example 3.1 for details.) ###### Corollary 3.9. Let $M$ be a right $R$-module. If either $R$ is right noetherian or $\Lambda$ is a finite index set, then $\widetilde{E}(M^{(\Lambda)})=(\widetilde{E}(M))^{(\Lambda)}$. ###### Corollary 3.10. Let $\\{M_{k}\\}_{k\in\Lambda}$ be a class of rationally complete right $R$-module for any index set $\Lambda$. If either $R$ is right noetherian or $|\Lambda|$ is finite, then $M=\bigoplus_{k\in\Lambda}M_{k}$ is rationally complete. ###### Proof. Since $\widetilde{E}(M_{i})=M_{i}$, $M_{i}$ is $M_{j}$-dense in $\widetilde{E}(M_{i})$ for all $i,j\in\Lambda$. From Theorem 3.6, $\widetilde{E}(M)=\oplus_{k\in\Lambda}\widetilde{E}(M_{k})=\oplus_{k\in\Lambda}M_{k}=M$. Therefore $\oplus_{k\in\Lambda}M_{k}$ is rationally complete. ∎ ###### Proposition 3.11 ([14, Proposition 1.9]). Let $\\{S_{i}\\}_{i\in\Lambda}$ be a set of nonisomorphic simple modules, representing all singular simple modules. Then every module containing the module $P=\bigoplus_{i\in\Lambda}S_{i}$ is rationally complete. ## 4\. The endomorphism ring of a module over a right ring of quotients of a ring In this section, we obtain some condition to be $\text{End}_{R}(M)=\text{End}_{H}(M)$ where $H$ is a right ring of quotients of a ring $R$. Recall that an extension ring $H$ of a ring $R$ is called a _right ring of quotients_ of $R$ if for any two elements $x\neq 0$ and $y$ of $H$, there exists an element $r\in R$ such that $xr\neq 0$ and $yr\in R$. ###### Theorem 4.1. Let $M$ be a right $H$-module where $H$ is a right ring of quotients of a ring $R$. If $R$ is $M_{R}$-dense in $H_{R}$ then $\emph{End}_{R}(M)=\emph{End}_{H}(M)$. ###### Proof. Since $\text{End}_{H}(M)\subseteq\text{End}_{R}(M)$, it suffices to show that $\text{End}_{R}(M)\subseteq\text{End}_{H}(M):$ Let $\varphi\in\text{End}_{R}(M)$ be arbitrary. Assume that $\varphi\notin\text{End}_{H}(M)$. Then there exist $m\in M,t\in H$ such that $\varphi(mt)-\varphi(m)t\neq 0.$ Since $R$ is $M_{R}$-dense in $H_{R}$, there exists $r\in R$ such that $(\varphi(mt)-\varphi(m)t)r\neq 0$ and $tr\in R$. Hence $0\neq(\varphi(mt)-\varphi(m)t)r=\varphi(mt)r-\varphi(m)(tr)=\varphi(mtr)-\varphi(mtr)=0$, a contradiction. Therefore $\text{End}_{R}(M)=\text{End}_{H}(M)$. ∎ ###### Remark 4.2. (i) $R$ is always $E(R)$-dense in $H_{R}$ where $H$ is a right ring of quotients of $R$. For, let $x\in H_{R}$ and $0\neq y\in E(R)$. Since $H\leq^{\text{ess}}E(R)_{R}$, there exists $s\in R$ such that $0\neq ys\in H$. Also, $xs\in H$. Since $R\leq^{\text{den}}H_{R}$, there exists $t\in R$ such that $xst\in R$ and $0\neq yst$. Therefore $R$ is $E(R)$-dense in $H_{R}$. (ii) If $M$ is a nonsingular $R$-module, then $R$ is $M_{R}$-dense in $H_{R}$: For, let $0\neq m\in M$ and $t\in H$ be arbitrary. Take $t^{-1}R=\\{r\in R\,|\,tr\in R\\}$ a right ideal of $R$. Note that $t^{-1}R\leq^{\textrm{ess}}R_{R}$. Since $t^{-1}R\nleq^{\text{ess}}\mathbf{r}_{R}(m)$, there exists $r\in t^{-1}R$ and $r\notin\mathbf{r}_{R}(m)$. Thus, $tr\in R$ and $mr\neq 0$. Therefore $R$ is $M_{R}$-dense in $H_{R}$. (iii) If $M$ is a submodule of a projective right $H$-module, then $R$ is $M_{R}$-dense in $H_{R}$. For, let $P$ be a projective right $R$-module including $M$, that is, $M\leq P$ where $P\leq^{\oplus}H^{(\Lambda)}$ with some index set $\Lambda$. Then there is a right $R$-module $K\leq E(P)$ such that $E(P)=E(M)\oplus K.$ Since $R\leq^{\text{den}}H_{R}$, we get that $R$ is $H^{(\Lambda)}$-dense in $H_{R}$ from Lemma 3.4. Hence $R$ is $P$-dense in $H_{R}$. Thus $\text{Hom}_{R}(H/R,E(P))=0$ from Proposition 2.5. Since $\text{Hom}_{R}(H/R,E(P))\cong\text{Hom}_{R}(H/R,E(M))\oplus\text{Hom}_{R}(H/R,K)$, we obtain $\text{Hom}_{R}(H/R,E(M))=0$. It follows that $R$ is $M_{R}$-dense in $H_{R}.$ ###### Corollary 4.3. Let $M$ be a projective right $H$-module where $H$ is a right ring of quotients of $R$. Then $\emph{End}_{R}(M)=\emph{End}_{H}(M)$. The next example illustrates Corollary 4.3. ###### Example 4.4. Let $Q=\prod_{n=1}^{\infty}\mathbb{Z}_{2}$ and $R=\\{(a_{n})\in Q~{}|~{}a_{n}$ is eventually constant$\\}$. Then $Q$ is a maximal right ring of quotients of $R$. Hence from Theorem 4.1, End${}_{R}(Q^{(\Lambda)})$= End${}_{Q}(Q^{(\Lambda)})=\mathsf{CFM}_{\Lambda}(Q)$. ###### Theorem 4.5. Let $M$ be a finitely generated free $R$-module with $S=\emph{End}_{R}(M)$. If either $R$ is right noetherian or $\Lambda$ is any finite index set, then $\emph{End}_{R}\left(\widetilde{E}(M^{(\Lambda)})\right)=\mathsf{CFM}_{\Lambda}\left(Q(S)\right)$. ###### Proof. Let $M=R^{(n)}$ for some $n\in\mathbb{N}$. From Corollary 3.9, $\widetilde{E}(R^{(n)})=\widetilde{E}(R)^{(n)}=Q(R)^{(n)}$ as $\widetilde{E}(R)=Q(R)$. Hence $\text{End}_{R}\left(\widetilde{E}(M^{(\Lambda)})\right)=\text{End}_{R}\left(\widetilde{E}(M)^{(\Lambda)}\right)=\text{End}_{R}\left((Q(R)^{(n)})^{(\Lambda)}\right)=\text{End}_{Q(R)}\left((Q(R)^{(n)})^{(\Lambda)}\right)=\mathsf{CFM}_{\Lambda}\left(\text{End}_{Q(R)}(Q(R)^{(n)})\right)=\mathsf{CFM}_{\Lambda}\left(\mathsf{Mat}_{n}(Q(R))\right)$ from Theorem 4.1. Therefore $\text{End}_{R}\left(\widetilde{E}(M^{(\Lambda)})\right)=\mathsf{CFM}_{\Lambda}\left(Q(\text{End}_{R}(M))\right)$ because $\mathsf{Mat}_{n}(Q(R))=Q(\mathsf{Mat}_{n}(R))$ by [15, 2.3] and $\text{End}_{R}(M)=\mathsf{Mat}_{n}(R)$. ∎ The next result is generalized from [15, 2.3]. ###### Corollary 4.6. Let $M$ be a finitely generated free $R$-module. Then $Q(\emph{End}_{R}(M))=\emph{End}_{R}(\widetilde{E}(M))$. The following example shows that the above result can not be extended to flat modules. This example also shows that $R$ is not $M_{R}$-dense in $Q_{R}$ where $Q$ is a right ring of quotients of $R$. ###### Example 4.7. Let $Q=\prod_{n=1}^{\infty}\mathbb{Z}_{2}$, $R=\\{(a_{n})\in Q~{}|~{}a_{n}$ is eventually constant$\\}$, and $I=\\{(a_{n})\in Q~{}|~{}a_{n}=0~{}\textrm{eventually}\\}$. Note that $Q=Q(R)$. Let $M=Q/I$, which is a flat $Q$-module but not projective. We claim that End${}_{Q}(M)\subsetneq$ End${}_{R}(M)$. Indeed, define $f:M\to M$ via $f[(a_{1},a_{2},\dots,a_{n},a_{n+1},\dots)+I]=(a_{1},0,a_{2},0,\dots,a_{n},0,a_{n+1},0,\dots)+I,$ for any $\overline{a}=a+I=(a_{1},a_{2},\dots,a_{n},a_{n+1},\dots)+I\in M$. It is easy to see that $f(\overline{a}+\overline{b})$ = $f(\overline{a})$ \+ $f(\overline{b})$ for any $\overline{a}$, $\overline{b}\in M$. Meanwhile, for any $r=(r_{1},r_{2},\dots,r_{n},r_{n+1},\dots)\in R$, we have $(a+I)r=ar+I=\left\\{\begin{array}[]{ll}(0,\dots,0,a_{n},a_{n+1},\dots)+I,&\mbox{if }r_{n}=r_{n+1}=\dots=1;\\\ (0,\dots,0,0,0,\dots)+I,&\mbox{if }r_{n}=r_{n+1}=\dots=0.\end{array}\right.$ Note that $a+I=(0,0,\dots,0,a_{n},a_{n+1},\dots)+I$ for some $n\in\mathbb{N}$. One can easily see that $f[(a+I)r]$ = $[f(a+I)]r$ for all $a\in Q$, $r\in R$. This shows $f\in$ End${}_{R}(M)$. However, for $q=(0,1,0,1,\dots)=q^{2}\in Q$, we have $[f(q+I)]q=0+I$ while $f[(q+I)q]=f(q+I)\neq 0+I$. This means $f\not\in$ End${}_{Q}(M)$. Thus, End${}_{Q}(M)\subsetneq$ End${}_{R}(M)$. Note that $R$ is not $M_{R}$-dense in $Q$. For, let $q\in Q\setminus R$ and $m=1+I\in M$. Since $(1+I)r=0+I$ for all $r\in R\setminus{1}$, it has to be $r=1$ to get $mr\neq 0+I$. However, $qr\notin R$. Recall that a module $M$ is said to be _polyform_ if every essential submodule of $M$ is a dense submodule. ###### Lemma 4.8. A module $M$ is polyform iff $\widetilde{E}(M)$ is a polyform quasi-injective module. ###### Proof. Let $X$ be essential in $\widetilde{E}(M)$. Then $X\cap M\leq^{\text{ess}}M$. Hence $X\cap M$ is a dense submodule of $M$ because $M$ is polyform. Since $X\cap M\leq^{\text{den}}M\leq^{\text{den}}\widetilde{E}(M)$, $X\cap M\leq^{\text{den}}\widetilde{E}(M)$. Thus $X$ is a dense submodule of $\widetilde{E}(M)$ from Proposition 1.2(ii). Therefore $\widetilde{E}(M)$ is a polyform module. In addition, $\widehat{M}$ is also a polyform module from [16, 11.1]. Since $M\leq^{\text{ess}}\widehat{M}$, $M\leq^{\text{den}}\widehat{M}$. Thus $\widetilde{E}(M)=\widetilde{E}(\widehat{M})$. Since the rational hull of a quasi-injective module is also quasi-injective from Theorem 2.16, $\widetilde{E}(M)$ is a quasi-injective module. Therefore $\widetilde{E}(M)$ is a polyform quasi-injective module. Conversely, let $N$ be any essential submodule of $M$. Then $N$ is also essential in $\widetilde{E}(M)$. Hence $N$ is a dense submodule of $\widetilde{E}(M)$ as $\widetilde{E}(M)$ is polyform. So $N$ is a dense submodule of $M$. Therefore $M$ is polyform. ∎ We show from Theorem 2.15 that there is a canonical embedding of the ring $\text{End}_{R}(M)$ into the ring $\text{End}_{R}(\widetilde{E}(M))$. Next, we obtain a condition when $\text{End}_{R}(M)$ and $\text{End}_{R}(\widetilde{E}(M))$ are isomorphic. It is a generalization of [7, Exercises 7.32]. ###### Proposition 4.9. If $M$ is a quasi-injective module then $\emph{End}_{R}(M)\stackrel{{\scriptstyle\Omega}}{{\cong}}\emph{End}_{R}(\widetilde{E}(M))$. In particular, if $M$ is a polyform module, then the converse holds true. ###### Proof. In the proof of Theorem 2.15, we only need to show that $\Omega:\text{End}_{R}(M)\rightarrow\text{End}_{R}(\widetilde{E}(M))$ given by $\Omega(\varphi)=\widetilde{\varphi}$, is surjective: Let $\psi\in\text{End}_{R}(\widetilde{E}(M))$ be arbitrary. Then there exists $\widehat{\psi}\in\text{End}_{R}(E(M))$ such that $\widehat{\psi}|_{\widetilde{E}(M)}=\psi$. Since $\widehat{\psi}M\leq M$ as $M$ is quasi-injective, $\widehat{\psi}|_{M}=\psi|_{M}\in\text{End}_{R}(M)$. Thus, $\Omega(\psi|_{M})=\psi$, which shows that $\Omega$ is surjective. In addition, suppose that $M$ is a polyform module. Then from Lemma 4.8, $\widetilde{E}(M)$ is quasi-injective. Thus, for any $\vartheta\in\text{End}_{R}(E(M))$, $\vartheta\widetilde{E}(M)\leq\widetilde{E}(M)$. Since $\vartheta|_{\widetilde{E}(M)}\in\text{End}_{R}(\widetilde{E}(M))$ and $\text{End}_{R}(M)\stackrel{{\scriptstyle\Omega}}{{\cong}}\text{End}_{R}(\widetilde{E}(M))$, there exists $\varphi\in\text{End}_{R}(M)$ such that $\Omega(\varphi)=\vartheta|_{\widetilde{E}(M)}$. Also by Theorem 2.15, $\vartheta|_{M}=\varphi$. Thus, $\vartheta M=\varphi M\leq M$. Therefore $M$ is a quasi-injective module. ∎ ###### Corollary 4.10. If $M$ is a quasi-injective module, then $Q(\emph{End}_{R}(M))\cong\emph{End}_{R}(\widetilde{E}(M))$. ###### Proof. Since $M$ is a quasi-injective module $\text{End}_{R}(M)$ is a right self- injective ring. So, $Q(\text{End}_{R}(M))=\text{End}_{R}(M)$. Thus, $Q(\text{End}_{R}(M))\cong\text{End}_{R}(\widetilde{E}(M))$ by Proposition 4.9. ∎ Remark that if $M$ is a quasi-injective module then $\widetilde{E}(M)$ is a quasi-injective module from Theorem 2.16 and $\text{End}_{R}(M)\cong\text{End}_{R}(\widetilde{E}(M))$ from Proposition 4.9. However, the next example shows that there exists a quasi-injective module $M$ such that $M\neq\widetilde{E}(M)$. ###### Example 4.11. Let $R=\left(\begin{smallmatrix}F&F\\\ 0&F\end{smallmatrix}\right)$ and $M=\left(\begin{smallmatrix}0&0\\\ 0&F\end{smallmatrix}\right)$ where $F$ is a field. Then $M$ is a quasi-injective $R$-module. However, $\widetilde{E}(M)=E(M)=\left(\begin{smallmatrix}0&0\\\ F&F\end{smallmatrix}\right)$ because $M$ is nonsingular. Thus $M$ is a quasi- injective $R$-module such that $M\lneq\widetilde{E}(M)$ and $\text{End}_{R}(M)\cong\left(\begin{smallmatrix}0&0\\\ 0&F\end{smallmatrix}\right)\cong\text{End}_{R}(\widetilde{E}(M))$. Because $\widetilde{E}(M)=E(M)$ for a right nonsingular module $M$, we have the following well-known results as a consequence of Proposition 4.9. ###### Corollary 4.12 ([7, Exercises 7.32]). For any nonsingular module $M$, the following statements hold true: 1. (i) there is a canonical embedding $\Omega$ of the ring $\emph{End}_{R}(M)$ into the ring $\emph{End}_{R}(E(M))$. 2. (ii) $M$ is a quasi-injective $R$-module if and only if $\Omega$ is an isomorphism. ###### Corollary 4.13. Let $M$ be a right $H$-module where $H$ is a right ring of quotients of a ring $R$. The following statements hold true: 1. (i) If $M$ is a nonsingular $R$-module then $\emph{End}_{R}(M)=\emph{End}_{H}(M)$. 2. (ii) If $M$ is a submodule of a projective $H$-module, then $\emph{End}_{R}(M)=\emph{End}_{H}(M)$. 3. (iii) If $M$ is a nonsingular quasi-injective $R$-module then $\emph{End}_{R}(M)\cong\emph{End}_{R}(E(M))$ and $\emph{End}_{H}(M)\cong\emph{End}_{H}(E(M))$. 4. (iv) If $M$ is a quasi-injective $R$-module and is a submodule of a projective $H$-module then $\emph{End}_{R}(M)\cong\emph{End}_{R}(\widetilde{E}(M))$ and $\emph{End}_{H}(M)\cong\emph{End}_{H}(\widetilde{E}(M))$. ## References * [1] E. Akalan; G.F. Birkenmeier; A. Tercan, Goldie extending modules, Comm. Algebra, 2009 _37(2)_ , 663–683 * [2] K. Asano, Über die Quotientenbildung von Schiefringen, J. Math. Soc. Japan, 1949 _1(2)_ , 73–78 * [3] G.F. Birkenmeier; J.K. Park; S.T. Rizvi, _Extensions of Rings and Modules_ , Birkhäuser/Springer, New York-Heidelberg-Dordrecht-London (2013) * [4] G.D. Findlay; J. Lambek, A generalized ring of quotients I, Canad. Math. Bull., 1958 _1_ , 77–85 * [5] A.W. Goldie, Localization in non-commutative noetherian rings, J. Algebra, 1967 _5_ , 89–105 * [6] R.E. Johnson, The extended centralizer of a ring over a module, Proc. Amer. Math. Soc., 1951 _2_ , 891–895 * [7] T.Y. Lam, _Lectures on Modules and Rings_ , GTM 189, Springer Verlag, Berlin-Heidelberg-New York (1999) * [8] J. Lambek, On Utumi’s ring of quotients, Canadian J. Math., 1963 _15_ , 363–370 * [9] J. Lambek, _Lectures on Rings and Modules_ , Blaisdell Publishing Co. [Ginn and Co.], Waltham, Mass.-Toronto, Ont.-London (1966) * [10] S.H. Mohamed; B.J. M$\ddot{\textrm{u}}$ller, _Continuous and Discrete Modules_ , Lecture Note Ser. No. 147, London Math. Soc., Cambridge Univ. Press (1990) * [11] Ø. Ore, Linear equations in non-commutative fields, Ann. of Math. Second Series, 1931 _32(3)_ , 463–477 * [12] S.T. Rizvi; C.S. Roman, On $\mathcal{K}$-nonsingular modules and applications, Comm. Algebra, 2007 _35_ , 2960–2982 * [13] B. Stenström, _Rings of Quotients_ , Band 217, Springer Verlag, Berlin-Heidelberg-New York (1975) * [14] H. Storrer, Rational extensions of modules, Pacific Journal of Mathematics, 1971 _38(3)_ , 785–794 * [15] Y. Utumi, On quotient rings, Osaka Math. J., 1956 _8_ , 1–18 * [16] R. Wisbauer, _Modules and Algebras: bimodule structure and group actions on algebras_ , Pitman Monographs and Surveys in Pure and Applied Mathematics 81, Longman, Harlow (1996) * [17] X. Zhang; G. Lee; N.K. Tung, Relatively polyform modules, Journal of Algebra Appl. 2025 2550046 (17 pages) DOI: 10.1142/S021949882550046X
# A Complexity Dichotomy in Spatial Reasoning via Ramsey Theory Manuel Bodirsky and Bertalan Bodor Institut für Algebra, Fakultät für Mathematik, TU Dresden and Department of Algebra, Faculty of Mathematics and Physics, Charles University, Prague ###### Abstract. Constraint satisfaction problems (CSPs) for first-order reducts of finitely bounded homogeneous structures form a large class of computational problems that might exhibit a complexity dichotomy, P versus NP-complete. A powerful method to obtain polynomial-time tractability results for such CSPs is a certain reduction to polynomial-time tractable finite-domain CSPs defined over $k$-types, for a sufficiently large $k$. We give sufficient conditions when this method can be applied and apply these conditions to obtain a new complexity dichotomy for CSPs of first-order expansions of the basic relations of the well-studied spatial reasoning formalism RCC5. We also classify which of these CSPs can be expressed in Datalog. Our method relies on Ramsey theory; we prove that RCC5 has a Ramsey order expansion. The authors have received funding from the European Research Council (Grant Agreement no. 681988, CSP-Infinity). ###### Contents 1. 1 Introduction 2. 2 Countably Categorical Structures 3. 3 Ramsey Theory 4. 4 Oligomorphic Clones 5. 5 Canonical Functions 1. 5.1 Behaviours 2. 5.2 Canonicity 3. 5.3 Canonisation 4. 5.4 Complexity 6. 6 Results 1. 6.1 The Unique Interpolation Property 2. 6.2 Other Characterisations of the UIP 7. 7 Proof of the Extension Lemma 1. 7.1 Diagonal Interpolation 2. 7.2 Rich subsets 3. 7.3 Interpolation invariance 4. 7.4 Proofs of main results 8. 8 Verification of the UIP 9. 9 First-order expansions of the basic relations of RCC5 1. 9.1 A Ramsey order expansion of $\mathfrak{R}$ 2. 9.2 Polymorphisms of $\mathfrak{R}$ that are canonical with respect to $(\mathfrak{R},\prec)$ 3. 9.3 Independent substructures 4. 9.4 Uniformly continuous minor-preserving maps 1. 9.4.1 Cyclic polymorphisms 2. 9.4.2 Factoring $\prec$ 3. 9.4.3 Restricting to $\smash{{}^{\bot}_{\prec}}$ 4. 9.4.4 Canonical pseudo-cyclic polymorphisms 5. 9.5 Verifying the UIP 6. 9.6 Proof of the complexity classification 10. 10 Expressibility in Datalog and Extensions 11. 11 Future Work ## 1\. Introduction The _Constraint satisfaction problem (CSP)_ of a relational structure $\mathfrak{B}$ is the computational problem of deciding whether a given finite structure $\mathfrak{A}$ has a homomorphism to $\mathfrak{B}$. If the structure $\mathfrak{B}$ is finite, then $\operatorname{CSP}(\mathfrak{B})$ is in P or NP-complete; this was conjectured by Feder and Vardi [FV99] and proved by Bulatov [Bul17] and, independently, by Zhuk [Zhu17]. Both approaches use concepts and methods from universal algebra; in particular, they use that the computational complexity of $\operatorname{CSP}(\mathfrak{B})$ is fully determined by the set $\operatorname{Pol}(\mathfrak{B})$ of _polymorphisms_ of $\mathfrak{B}$, which is the set of homomorphisms from $\mathfrak{B}^{n}$ to $\mathfrak{B}$ for $n\in{\mathbb{N}}$. In fact, $\operatorname{CSP}(\mathfrak{B})$ is NP-complete if $\operatorname{Pol}(\mathfrak{B})$ has a minor-preserving map to the clone $\operatorname{Proj}$ of projections on a two-element set, and is in P otherwise. Some of the results about finite structures can be lifted to countably infinite structures whose automorphism group satisfies a certain finiteness condition, called _oligomorphicity_ : the requirement is that for every $n\in{\mathbb{N}}$ the componentwise action of the automorphism group on $n$-tuples has only finitely many orbits. Examples of such structures arise systematically in model theory: every homogeneous structure $\mathfrak{B}$ with a finite relational signature is of this type; such structures will be called _finitely homogeneous_. An additional finiteness condition is to require that the class of finite substructures of $\mathfrak{B}$ is described by finitely many forbidden substructures, in which case $\mathfrak{B}$ is called _finitely bounded_. The class of reducts of finitely bounded homogeneous structures is a huge generalisation of the class of all finite structures. There is also a generalisation of the Feder-Vardi dichotomy conjecture, which is still open, to reducts of finitely bounded homogeneous structures. In fact, there is a known NP-hardness condition for the CSPs of such structures which is conjectured to be at the border between polynomial-time tractable and NP- complete CSPs [BPP21, BOP18, BKO+17]. The _infinite-domain tractability conjecture_ states that the CSP for every structure that does not satisfy the mentioned hardness condition is in P (details can be found in Section 4). There are many classes of infinite-domain structures where the infinite-domain tractability conjecture has been verified. Often, these classes consist of the _first-order reducts_ of some fixed underlying structure $\mathfrak{B}$. By a _first-order reduct of $\mathfrak{B}$_ we mean a reduct of the expansion of $\mathfrak{B}$ by all relations that are first-order definable in $\mathfrak{B}$. For example, the infinite-domain tractability conjecture has been verified for 1. (1) all CSPs of structures preserved by all permutations [BK08], which is precisely the class of first-order reducts of pure sets (structures with no relations); 2. (2) all CSPs of structures with a highly set-transitive automorphism group [BK09b] (i.e., for all finite subsets $X,Y$ with $|X|=|Y|$ there exists an automorphism which maps $X$ to $Y$), which is precisely the class of first- order reducts of unbounded dense linear orders; 3. (3) all CSPs of first-order reducts of the homogeneous universal poset [KP18]; 4. (4) all CSPs of first-order reducts of the binary branching C-relation [BJP17]; 5. (5) all CSPs of first-order reducts of homogeneous graphs [BMPP19]; 6. (6) all CSPs of first-order reducts of _unary structures_ , i.e., structures with a signature that consists of finitely many unary relation symbols [BM18]; 7. (7) all CSPs expressible in MMSNP [BMM18]; MMSNP is a fragment of existential second-order logic introduced by Feder and Vardi [FV99]. In some, but not in all cases above the polynomial-time tractability results for $\operatorname{CSP}(\mathfrak{B})$ can be obtained by reducing $\operatorname{CSP}(\mathfrak{B})$ to $\operatorname{CSP}(\mathfrak{T})$ where $\mathfrak{T}$ is a certain finite structure whose domain is the set of complete $k$-types of $\mathfrak{B}$, for a sufficiently large $k$; this method works if the _clone of canonical polymorphisms of $\mathfrak{B}$_ does not have a minor-preserving map to the projections [BM18]. In this case, $\operatorname{Pol}(\mathfrak{T})$ does not have a minor-preserving map to $\operatorname{Proj}$, and hence satisfies the condition for the algorithms of Bulatov and of Zhuk. The polynomial-time tractability of $\operatorname{CSP}(\mathfrak{T})$ then implies the polynomial-time tractability of $\operatorname{CSP}(\mathfrak{B})$, by a reduction from [BM18]. The publications cited above show that the canonical polymorphisms determine the complexity of the CSPs in (1), (3), (5), (6), and (7), but not for the CSPs in (2) and (4). We introduce a property, which is implicit in [BM18, BMM18], and which we call the _unique interpolation property (UIP)_. We present equivalent characterisations of the unique interpolation property; in particular, we show that it suffices to verify the unique interpolation property for binary polymorphisms. For our characterisation we need an additional assumption, which holds in all the mentioned examples that appear in (1)-(7) above, and which makes the connection between CSPs and universal algebra particularly strong: we require that $\mathfrak{B}$ is a reduct of a finitely homogeneous _Ramsey structure_ $\mathfrak{A}$. While the assumption to be Ramsey is quite strong, the assumption that a finitely homogeneous structure is a reduct of a finitely homogeneous Ramsey structure is weak: in fact, it might be that _every_ finitely bounded homogeneous structure has a finitely bounded homogeneous Ramsey expansion (see [Bod15, BPT13]). We use our general results to prove a new complexity dichotomy for qualitative spatial reasoning. A fundamental formalism in spatial reasoning is RCC5 [Ben94], which involves five binary relations defined on some general set of _regions_ ; a formal definition can be found in Section 9. Renz and Nebel [RN01] showed that the CSP for RCC5 is NP-complete, and Nebel [Neb95] showed that the CSP for the five basic relations of RCC5 is in P (via a reduction to 2SAT111The proof of Nebel was formulated for RCC8 instead of RCC5, but the result for RCC5 can be shown analogously.). Renz and Nebel [RN01] have extended the polynomial-time tractability result to a superclass of the basic relations, and they showed that their expansion is _maximal_ in the sense that every larger subset of the RCC5 relations has an NP-hard CSP. Drakengren and Jonsson [JD97] classified the computational complexity of the CSP for all subsets of the RCC5 relations. We classify the complexity of the CSP for expansions of the basic relations of RCC5 by first-order definable relations of arbitrary finite arity, solving the essential part of Problem 4 in [BJ17]. In particular, we verify the tractability conjecture for these CSPs, and show that the tractable cases can be solved by reducing them to polynomial-time tractable finite-domain CSPs using the reduction from [BM18]. For our method, we need a homogeneous Ramsey expansion of RCC5; to prove the Ramsey property of the expansion that we construct we apply a recent Ramsey transfer result of Mottet and Pinsker [MP21], using the well-known fact that a certain ordered version of the atomless Boolean algebra has the Ramsey property [KPT05]. The results mentioned so far have been announced in the proceedings of LICS 2021 [BB21a], but most proofs were omitted because of space restrictions. We additionally show that if the condition of the infinite-domain tractability conjecture applies to a first-order expansion $\mathfrak{B}$ of the basic relations of RCC5, then $\operatorname{CSP}(\mathfrak{B})$ can be solved by Datalog [FV99] (in polynomial time); otherwise, it cannot be expressed in the much stronger _fixed-point logic with counting (FPC)_ [ABD09]. ### Related Work There exists another powerful approach to CSP complexity classification under similar conditions, the theory of _smooth approximations_ [MP22, MNPW21]. The theory of smooth approximations also gives a sufficient condition for the UIP in certain cases (see the discussion at the end of Section 3.2 in [MP22]); but the machinery and the general statements there do not seem to directly relate to the statements that we prove here. The smooth approximations approach proved to be very efficient in re-deriving existing classifications and proving new classifications where a surprisingly large proportion of the work follows from the general results of smooth approximations, and only relatively few arguments are needed to adapt to the classes under consideration. However, we do not know how to apply their approach to derive our spatial reasoning classification. We believe that also the general results presented here will be useful in other settings; one example of a different setting where the present approach has been applied is the class of monadically stable $\omega$-categorical structures, for which the second author confirmed the infinite-domain tractability conjecture in his PhD thesis [Bod22], using results of this article. In contrast to the present article, the smooth approximations approach also provides a systematic approach to Datalog expressibility [MNPW21]; but also here we do not know how to obtain the particular results about Datalog expressibility in spatial reasoning from their general theorems. ### Outline In Section 2, Section 3, and Section 4 we introduce some fundamental notions and results from model theory, Ramsey theory, and universal algebra, respectively, that we need to state our results. In Section 3 we also need a consequence of the Ramsey property of an $\omega$-categorical structure $\mathfrak{A}$ for the existence of independent elementary substructures of $\mathfrak{A}$ (Theorem 3.2); this result might be of independent interest in model theory. The formal definition of the unique interpolation property and our first general main result about the UIP with an application to complexity classification is presented in Section 6.1. We then present two results that facilitate the verification of the UIP in concrete settings (Section 6.2). The applications of the general results to the spatial reasoning formalism RCC5 are in Section 9, and the classification of the respective CSPs in Datalog in Section 10. ## 2\. Countably Categorical Structures All structures considered in this text are assumed to be relational unless stated otherwise. If $\mathfrak{A}$ is a structure, then $A$ denotes the domain of $\mathfrak{A}$. A structure $\mathfrak{A}$ is called _$\omega$ -categorical_ if the set of all first-order sentences that hold in $\mathfrak{A}$ has exactly one countable model up to isomorphism. This concept has an equivalent characterisation based on the automorphism group $\operatorname{Aut}(\mathfrak{A})$ of $\mathfrak{A}$. For $l\in{\mathbb{N}}$, the _$l$ -orbit_ of $a=(a_{1},\dots,a_{l})\in A^{l}$ in $\operatorname{Aut}(\mathfrak{A})$ is the set $\\{\alpha(a)\colon\alpha\in\operatorname{Aut}(\mathfrak{A})\\}\text{ where }\alpha(a)\coloneqq(\alpha a_{1},\dots,\alpha a_{l}).$ We write ${\mathcal{O}}_{l}(\mathfrak{A})$ for the set of $l$-orbits of $\operatorname{Aut}(\mathfrak{A})$. Engeler, Svenonius, and Ryll-Nardzewski (see, e.g., Theorem 6.3.1 in [Hod97]) proved that a countable structure is $\omega$-categorical if and only if $\operatorname{Aut}(\mathfrak{A})$ is _oligomorphic_ , i.e., if ${\mathcal{O}}_{l}(\mathfrak{A})$ is finite for every $l\in{\mathbb{N}}$. In fact, $a,b\in A^{l}$ have the same $l$-orbit if and only if they satisfy the same first-order formulas over $\mathfrak{A}$; we write $\operatorname{typ}^{\mathfrak{A}}(a)$ for the set of all first-order formulas satisfied by $a$ over $\mathfrak{A}$, called the _type of $a$ over $\mathfrak{A}$_. For $B\subseteq A$, we write $\operatorname{typ}^{\mathfrak{A}}(a/B)$ for the set of all first-order formulas with parameters from $B$ (i.e., formulas over an expanded signature that also contains constant symbols denoting elements from $B$) satisfied by $a$ over $\mathfrak{A}$. Homogeneous structures. Many examples of $\omega$-categorical structures arise from structures $\mathfrak{A}$ that are _homogeneous_. A relational structure is _homogeneous_ if every isomorphism between finite substructures of $\mathfrak{A}$ can be extended to an automorphism of $\mathfrak{A}$. It follows from the above that if $\mathfrak{A}$ is finitely homogeneous, then $\mathfrak{A}$ is $\omega$-categorical. If $a\in A^{n}$, we write $\operatorname{qf-typ}^{\mathfrak{A}}(a)$ for the set of all quantifier-free formulas that hold on $a$ over $\mathfrak{A}$. Clearly, in homogeneous structures $\mathfrak{A}$ the quantifier-free type of $a$ determines the type of $a$. Also note that every $\omega$-categorical structure can be turned into a homogeneous $\omega$-categorical structure by expanding it by all first- order definable relations. Every homogeneous structure $\mathfrak{A}$ is uniquely given (up to isomorphism) by its _age_ , i.e., by the class $\operatorname{Age}(\mathfrak{A})$ of all finite structures that embed into $\mathfrak{A}$ (see, e.g., Lemma 6.1.4 in [Hod97]). Conversely, every class $\mathcal{C}$ of structures with finite relational signature which is an _amalgamation class_ , i.e., is closed under isomorphism, substructures, and which has the _amalgamation property_ , is the age of a homogeneous structure, which we call the _Fraïssé-limit of ${\mathcal{C}}$_ (see, e.g., Theorem 6.1.2 in [Hod97]). We say that $\mathfrak{B}$ is _finitely bounded_ if there exists a finite set of finite structures $\mathcal{F}$ such that the age of $\mathfrak{B}$ equals the class of all finite structures $\mathfrak{A}$ such that no structure from $\mathcal{F}$ embeds into $\mathfrak{A}$. Model-complete cores. An $\omega$-categorical structure $\mathfrak{C}$ is _model-complete_ if every self-embedding $e\colon\mathfrak{C}\hookrightarrow\mathfrak{C}$ preserves all first-order formulas. An $\omega$-categorical structure $\mathfrak{C}$ is called a _core_ if every endomorphism of $\mathfrak{C}$ is a self-embedding of $\mathfrak{C}$. If there is a homomorphism from a structure $\mathfrak{A}$ to a structure $\mathfrak{B}$ and vice versa, then $\mathfrak{A}$ and $\mathfrak{B}$ are called _homomorphically equivalent_. Every $\omega$-categorical structure $\mathfrak{A}$ is homomorphically equivalent to a model-complete core, which is unique up to isomorphism, and again $\omega$-categorical, and which is called _the_ model-complete core of $\mathfrak{A}$ [Bod07, BHM12]. ## 3\. Ramsey Theory For $\tau$-structures $\mathfrak{A},\mathfrak{B}$ we write ${\mathfrak{B}\choose\mathfrak{A}}$ for the set of all embeddings of $\mathfrak{A}$ into $\mathfrak{B}$. If $\mathfrak{A},\mathfrak{B},\mathfrak{C}$ are $\tau$-structures and $r\in{\mathbb{N}}$ then we write $\mathfrak{C}\to(\mathfrak{B})^{\mathfrak{A}}_{r}$ if for every function $\chi\colon{\mathfrak{C}\choose\mathfrak{A}}\to\\{1,\dots,r\\}$ (also referred to as a _colouring_ , where $\\{1,\dots,r\\}$ are the different colours) there exists $e\in{\mathfrak{C}\choose\mathfrak{B}}$ such that $\chi$ is constant on $\left\\{e\circ f\colon f\in{\mathfrak{B}\choose\mathfrak{A}}\right\\}\subseteq{\mathfrak{C}\choose\mathfrak{A}}.$ ###### Definition 1. A class of finite $\tau$-structures has the _Ramsey property_ if for all $\mathfrak{A},\mathfrak{B}$ and $r\in{\mathbb{N}}$ there exists $\mathfrak{C}$ such that $\mathfrak{C}\to(\mathfrak{B})^{\mathfrak{A}}_{r}$. A homogeneous structure has the Ramsey property if its age has the Ramsey property. Let $r\in{\mathbb{N}}$. We write $\mathfrak{C}\to(\mathfrak{B})_{r}$ if for every $\chi\colon C^{|B|}\to\\{1,\dots,r\\}$ there exists $e\colon\mathfrak{B}\hookrightarrow\mathfrak{C}$ such that if $\operatorname{qf-typ}^{\mathfrak{B}}(s)=\operatorname{qf- typ}^{\mathfrak{B}}(t)$ for $s,t\in B^{|B|}$, then $\chi(e(s))=\chi(e(t))$. The Ramsey property has the following well-known consequence (see, e.g., Proposition 2.21 in [Bod15]). ###### Lemma 3.1. Let $\mathcal{C}$ be a Ramsey class. Then for every $r\in{\mathbb{N}}$ and $\mathfrak{B}\in\mathcal{C}$ there exists $\mathfrak{C}\in\mathcal{C}$ such that $\mathfrak{C}\to(\mathfrak{B})_{r}$. Another important consequence of the Ramsey property is presented in Section 5.3. We also need the following consequence of the Ramsey property (in the proof of Proposition 7.3), which appears to be new. Let $\mathfrak{D}$ be a structure and let $A,B\subseteq D$. We say that $A$ is _independent from $B$ in $\mathfrak{D}$_ if for all $\bar{a},\bar{b}\in A^{n}$, if $\operatorname{typ}^{\mathfrak{D}}(\bar{a})=\operatorname{typ}^{\mathfrak{D}}(\bar{b})$, then $\operatorname{typ}^{\mathfrak{D}}(\bar{a}/B)=\operatorname{typ}^{\mathfrak{D}}(\bar{b}/B)$. Two substructures $\mathfrak{A},\mathfrak{B}$ of $\mathfrak{D}$ are called _independent_ if $A$ is independent from $B$ in $\mathfrak{D}$ and $B$ is independent from $A$ in $\mathfrak{D}$. A substructure $\mathfrak{B}$ of a $\tau$-structure $\mathfrak{A}$ is called _elementary_ if the identity mapping from $\mathfrak{B}$ to $\mathfrak{A}$ preserves all first-order $\tau$-formulas. ###### Theorem 3.2. Let $\mathfrak{A}$ be a countable homogeneous $\omega$-categorical Ramsey structure. Then $\mathfrak{A}$ contains two independent elementary substructures. In the proof of Theorem 3.2 we use the following consequence of the compactness theorem of first-order logic. ###### Proposition 3.3. Let $\mathfrak{A}$ be a countable $\omega$-categorical homogeneous structure. Then the following are equivalent. 1. (1) $\mathfrak{A}$ contains two independent elementary substructures. 2. (2) For all $\mathfrak{B}_{1},\mathfrak{B}_{2}\in\operatorname{Age}(\mathfrak{A})$ there exists embeddings $e_{i}\colon\mathfrak{B}_{i}\hookrightarrow\mathfrak{A}$, for $i\in\\{1,2\\}$, such that $e_{1}(B_{1})$ and $e_{2}(B_{2})$ are independent in $\mathfrak{A}$. ###### Proof. The forward implication is immediate from the definitions. For the converse implication, let $a_{1},a_{2},\dots$ be an enumeration of $A$ and let $B=\\{b_{1},b_{2},\dots\\}$ and $C=\\{c_{1},c_{2},\dots\\}$ be sets of new constant symbols. Let $\Phi$ be the set of all sentences that express that for all $n\in{\mathbb{N}}$, $\bar{b},\bar{b}^{\prime}\in B^{n}$, and $\bar{c},\bar{c}^{\prime}\in C^{n}$ * • if $\operatorname{typ}^{\mathfrak{A}}(\bar{b})=\operatorname{typ}^{\mathfrak{A}}(\bar{b}^{\prime})$ then $\operatorname{typ}^{\mathfrak{A}}(\bar{b},\bar{c})=\operatorname{typ}^{\mathfrak{A}}(\bar{b}^{\prime},\bar{c})$. * • if $\operatorname{typ}^{\mathfrak{A}}(\bar{c})=\operatorname{typ}^{\mathfrak{A}}(\bar{c}^{\prime})$ then $\operatorname{typ}^{\mathfrak{A}}(\bar{c},\bar{b})=\operatorname{typ}^{\mathfrak{A}}(\bar{c}^{\prime},\bar{b})$. For $n\in{\mathbb{N}}$ let $\phi_{n}(x_{1},\dots,x_{n})$ be a formula that expresses that $\operatorname{typ}^{\mathfrak{A}}(x_{1},\dots,x_{n})=\operatorname{typ}^{\mathfrak{B}}(a_{1},\dots,a_{n})$. By assumption, all finite subsets of $T\coloneqq\operatorname{Th}(\mathfrak{A})\cup\Phi\cup\\{\phi_{n}(b_{1},\dots,b_{n})\wedge\phi_{n}(c_{1},\dots,c_{n}):n\in{\mathbb{N}}\\}$ are satisfiable. By compactness of first-order logic, it follows that $T$ has a model $\mathfrak{A}^{\prime}$. By the downward Löwenheim-Skolem theorem, we may assume that $\mathfrak{A}^{\prime}$ is countably infinite. The reduct of $\mathfrak{A}^{\prime}$ with the same signature as $\mathfrak{A}$ satisfies $\operatorname{Th}(\mathfrak{A})$, and since $\mathfrak{A}$ is $\omega$-categorical this reduct is isomorphic to $\mathfrak{A}$, so that we may assume that $\mathfrak{A}^{\prime}$ is an expansion of $\mathfrak{A}$. Let $\mathfrak{B}$ be the substructure of $\mathfrak{A}$ induced by the constants from $B$. Since $\mathfrak{A}^{\prime}\models\\{\phi_{n}(b_{1},\dots,b_{n}):n\in{\mathbb{N}}\\}$ and by the homogeneity of $\mathfrak{A}$ we have that $\mathfrak{B}$ is an elementary substructure of $\mathfrak{A}$. Likewise, the substructure $\mathfrak{C}$ of $\mathfrak{A}$ induced by the constants from $C$ is an elementary substructure of $\mathfrak{A}$. Since $\mathfrak{A}^{\prime}\models\Phi$ the two substructures $\mathfrak{B}$ and $\mathfrak{C}$ are independent. ∎ To find independent sets, we use the Ramsey property via the following lemma. ###### Lemma 3.4. Let $\mathfrak{A}$ be a countable homogeneous $\omega$-categorical Ramsey structure, $C\subseteq A$ finite, and $\mathfrak{B}\in\operatorname{Age}(\mathfrak{A})$. Then there exists $\mathfrak{D}\in\operatorname{Age}(\mathfrak{A})$ such that for every $e\colon\mathfrak{D}\hookrightarrow\mathfrak{A}$ there exists $f\colon\mathfrak{B}\hookrightarrow\mathfrak{A}$ such that $f(B)\subseteq e(D)$ and for all $\bar{a},\bar{b}\in B^{|B|}$, if $\operatorname{typ}^{\mathfrak{A}}(f(\bar{a}))=\operatorname{typ}^{\mathfrak{A}}(f(\bar{b}))$ then $\operatorname{typ}^{\mathfrak{A}}(f(\bar{a})/C)=\operatorname{typ}^{\mathfrak{A}}(f(\bar{b})/C)$. ###### Proof. Let $C=\\{c_{1},\dots,c_{n}\\}$. Let $r$ be the number of types of $|B|+n$-tuples in $\mathfrak{A}$. Since $\mathfrak{A}$ is homogeneous Ramsey there exists $\mathfrak{D}\in\operatorname{Age}(\mathfrak{A})$ such that $\mathfrak{D}\to(\mathfrak{B})_{r}$. We claim that $\mathfrak{D}$ satisfies the statement of the lemma. Let $e\colon\mathfrak{D}\hookrightarrow\mathfrak{A}$ be an embedding. We color $e(D)^{|B|}$ as follows: the colour of $t\in e(D)^{|B|}$ is the orbit of $(t_{1},\dots,t_{|B|},c_{1},\dots,c_{n})$. There are at most $r$ such orbits. Since $\mathfrak{D}\to(\mathfrak{B})_{r}$ we get that there exists an $f\colon\mathfrak{B}\hookrightarrow\mathfrak{A}$ such that $f(B)\subseteq e(D)$ and if $\bar{a},\bar{b}\in B^{|B|}$ are such that $\operatorname{typ}^{\mathfrak{A}}(f(\bar{a}))=\operatorname{typ}^{\mathfrak{A}}(f(\bar{b}))$, then $\operatorname{typ}^{\mathfrak{A}}(f(\bar{a})/C)=\operatorname{typ}^{\mathfrak{A}}(f(\bar{b})/C)$. ∎ ###### Proof of Theorem 3.2. We use Proposition 3.3. Let $\mathfrak{B}_{1},\mathfrak{B}_{2}\in\operatorname{Age}(\mathfrak{A})$; we may assume that $\mathfrak{B}_{1}$ and $\mathfrak{B}_{2}$ are substructures of $\mathfrak{A}$. Then Lemma 3.4 implies that there exists $\mathfrak{D}_{1}\in\operatorname{Age}(\mathfrak{A})$ such that for every embedding $e_{1}\colon\mathfrak{D}_{1}\hookrightarrow\mathfrak{A}$ there exists an embedding $f_{1}\colon\mathfrak{B}_{1}\hookrightarrow\mathfrak{A}$ such that $f_{1}(B_{1})\subseteq e_{1}(D_{1})$ and for all $n\in{\mathbb{N}}$, $b,b^{\prime}\in(B_{1})^{n}$ (1) $\displaystyle\text{ if }\operatorname{typ}^{\mathfrak{A}}(f_{1}(b))=\operatorname{typ}^{\mathfrak{A}}(f_{1}(b^{\prime}))$ $\displaystyle\text{ then }\operatorname{typ}^{\mathfrak{A}}(f_{1}(b)/B_{2})=\operatorname{typ}^{\mathfrak{A}}(f_{1}(b^{\prime})/B_{2}).$ We may assume that $\mathfrak{D}_{1}$ is a substructure of $\mathfrak{A}$. Another application of Lemma 3.4 gives us an embedding $f_{2}\colon\mathfrak{B}_{2}\hookrightarrow\mathfrak{A}$ such that for all $n\in{\mathbb{N}}$, $b,b^{\prime}\in(B_{2})^{n}$, if $\operatorname{typ}^{\mathfrak{A}}(f_{2}(b))=\operatorname{typ}^{\mathfrak{A}}(f_{2}(b^{\prime}))$ then $\operatorname{typ}^{\mathfrak{A}}(f_{2}(b)/D_{1})=\operatorname{typ}^{\mathfrak{A}}(f_{2}(b^{\prime})/D_{1})$. By the homogeneity of $\mathfrak{A}$, there exists $\alpha\in\operatorname{Aut}(\mathfrak{A})$ extending $f_{2}$. The property of $\mathfrak{D}_{1}$ implies that there exists an embedding $f_{1}\colon\mathfrak{B}_{1}\hookrightarrow\mathfrak{A}$ such that $f_{1}(B_{1})\subseteq\alpha^{-1}(D_{1})$ such that (1) holds for all $b,b^{\prime}\in(B_{1})^{n}$. We claim that $f_{1}(B_{1})$ and $B_{2}$ are independent. Clearly, $f_{1}(B_{1})$ is independent from $B_{2}$. To show that $B_{2}$ is independent from $f_{1}(B_{1})$, let $n\in{\mathbb{N}}$ and $b,b^{\prime}\in(B_{2})^{n}$ be such that $\operatorname{typ}^{\mathfrak{A}}(b)=\operatorname{typ}^{\mathfrak{A}}(b^{\prime})$. Then $\operatorname{typ}^{\mathfrak{A}}(\alpha(b))=\operatorname{typ}^{\mathfrak{A}}(\alpha(b^{\prime}))$, and hence that $\operatorname{typ}^{\mathfrak{A}}(\alpha(b)/D_{1})=\operatorname{typ}^{\mathfrak{A}}(\alpha(b^{\prime})/D_{1})$ by the choice of $f_{2}$. This in turn implies that $\operatorname{typ}^{\mathfrak{A}}(b/\alpha^{-1}(D_{1}))=\operatorname{typ}^{\mathfrak{A}}(b^{\prime}/\alpha^{-1}(D_{1}))$ which implies the statement since $f_{1}(B_{1})\subseteq\alpha^{-1}(D_{1})$. ∎ ## 4\. Oligomorphic Clones In this section we introduce fundamental concepts and results about polymorphism clones of finite and countably infinite $\omega$-categorical structures that are needed for the statement of our results. The _$k$ -th direct power_ of a structure $\mathfrak{A}$ (also called the _categorical power_), denoted by $\mathfrak{A}^{k}$. It is the $\tau$-structure with domain $A^{k}$ such that $((a_{1,1},\dots,a_{1,k}),\dots,(a_{n,1},\dots,a_{n,k}))\in R^{\mathfrak{A}^{k}}$ if and only if $(a_{1,i},\dots,a_{n,i})\in R^{\mathfrak{A}}$ for every $i\in\\{1,\dots,k\\}$. A _polymorphism_ of $\mathfrak{A}$ is a homomorphism from a finite direct power $\mathfrak{A}^{k}$ to $\mathfrak{A}$. The set of all polymorphisms of $\mathfrak{A}$, denoted by $\operatorname{Pol}(\mathfrak{A})$, forms a _clone (over $A$)_, i.e., it is closed under composition and contains for every $k$ and $i\leq k$ the projection $\pi^{k}_{i}$ of arity $k$ that always returns the $i$-th argument. Moreover, $\operatorname{Pol}(\mathfrak{A})$ is closed with respect to the _topology of pointwise convergence_ , i.e., if $f\colon A^{k}\to A$ is such that for every finite $F\subseteq A$ there exists $g\in\operatorname{Pol}(\mathfrak{A})$ such that $g(a_{1},\dots,a_{k})=f(a_{1},\dots,a_{k})$ for all $a_{1},\dots,a_{k}\in F$, then $f\in\operatorname{Pol}(\mathfrak{A})$. If ${\mathscr{S}}$ is a set of operations over $A$, then $\overline{\mathscr{S}}$ denotes the smallest closed set that contains ${\mathscr{S}}$. If ${\mathscr{S}}$ is a set of operations, we write ${\mathscr{S}}^{(k)}$ for the set of operation in ${\mathscr{S}}$ of arity $k$. The clone of all projections on $\\{0,1\\}$ is denoted by $\operatorname{Proj}$. ###### Definition 2. A function $\zeta\colon{\mathscr{C}}_{1}\to{\mathscr{C}}_{2}$ between two clones over a countably infinite set $A$ is called * • _minor-preserving_ if $\zeta\big{(}f(\pi^{k}_{i_{1}},\dots,\pi^{k}_{i_{n}})\big{)}=\zeta(f)(\pi^{k}_{i_{1}},\dots,\pi^{k}_{i_{n}})$ for all $f\in{\mathscr{C}}^{(n)}_{1}$ and $i_{1},\dots,i_{n}\in\\{1,\dots,k\\}$. * • _uniformly continuous_ if and only if for any finite $G\subseteq A$ there is a finite $F\subseteq A$ such that for all $n\in{\mathbb{N}}$ and $f,g\in{\mathscr{C}}_{1}^{(n)}$, if $f|_{F^{n}}=g|_{F^{n}}$ then $\zeta(f)|_{G^{n}}=\zeta(g)|_{G^{n}}$. The set ${\mathscr{O}}_{A}\coloneqq\bigcup_{k\in{\mathbb{N}}}(A^{k}\to A)$ of all operations on $A$ can be equipped with a complete metric such that the closed clones with respect to this metric are precisely the polymorphism clones of relational structures with domain $A$, and such that the uniformly continuous maps are precisely the maps that are uniformly continuous with respect to this metric in the standard sense, justifying our terminology; see, e.g., [Bod21]. Now suppose that $\mathfrak{B}$ is a finite structure with a finite relational signature; the following results imply that the existence of a minor-preserving map from $\operatorname{Pol}(\mathfrak{B})$ to $\operatorname{Proj}$ describes the border between containment in P and NP- completeness. ###### Theorem 4.1 (follows from [BKJ05, BOP18]). Let $\mathfrak{B}$ be a structure with a finite domain and a finite relational signature. If there is a minor-preserving map $\operatorname{Pol}(\mathfrak{B})\to\operatorname{Proj}$ then $\operatorname{CSP}(\mathfrak{B})$ is NP-hard. ###### Theorem 4.2 ([MM08]). Let $\mathfrak{B}$ be a structure with a finite domain. If there is no minor- preserving map $\operatorname{Pol}(\mathfrak{B})\to\operatorname{Proj}$ then $\operatorname{Pol}(\mathfrak{B})$ contains a _weak near unanimity (WNU) operation_ , i.e., an operation $f$ of some arity $k\geq 2$ that satisfies for all $a,b\in B$ that $f(a,\dots,a,b)=f(a,\dots,b,a)=\dots=f(b,a,\dots,a).$ ###### Theorem 4.3 (Bulatov [Bul17], Zhuk [Zhu17]). Let $\mathfrak{B}$ be a structure with a finite domain and a finite relational signature. If $\mathfrak{B}$ has a WNU polymorphism, then $\operatorname{CSP}(\mathfrak{B})$ is in P. The hardness result in Theorem 4.1 can be generalised to a large class of countable structures $\mathfrak{B}$. ###### Theorem 4.4 ([BOP18]). Let $\mathfrak{B}$ be a countable $\omega$-categorical structure with finite relational signature such that $\operatorname{Pol}(\mathfrak{B})$ has a uniformly continuous minor-preserving map to $\operatorname{Proj}$. Then $\operatorname{CSP}(\mathfrak{B})$ is NP-hard. The present paper is motivated by the following conjecture from [BPP21], in a reformulation from [BKO+17], which generalises the Bulatov-Zhuk dichotomy (Theorem 4.3). ###### Conjecture 4.5. Let $\mathfrak{B}$ be a reduct of a countable finitely bounded homogeneous structure such that there is no uniformly continuous minor-preserving map from $\operatorname{Pol}(\mathfrak{B})$ to $\operatorname{Proj}$. Then $\operatorname{CSP}(\mathfrak{B})$ is in P. ## 5\. Canonical Functions Some of the strongest partial results towards Conjecture 4.5 is based on the concept of _canonical_ polymorphisms. Let $\mathfrak{A}$ be an $\omega$-categorical structure. An operation $f\colon A^{k}\to A$ is called _canonical over ${\mathfrak{A}}$_ if for all $n\in{\mathbb{N}}$ and $a_{1},\dots,a_{k},b_{1},\dots,b_{k}\in A^{n}$, if $\operatorname{typ}^{\mathfrak{A}}(a_{i})=\operatorname{typ}^{\mathfrak{A}}(b_{i})$ for all $i\in\\{1,\dots,k\\}$, then $\operatorname{typ}^{\mathfrak{A}}(f(a_{1},\dots,a_{k}))=\operatorname{typ}^{\mathfrak{A}}(f(b_{1},\dots,b_{k}))$ (the function is applied to the tuples componentwise). Note that if $\mathfrak{A}$ is a finitely homogeneous structure, and $m$ is the maximal arity of (the relations of) $\mathfrak{A}$, then $f$ is canonical with respect to $\mathfrak{A}$ if and only if it is satisfies the condition above for $n=m$. The set of all canonical operations over $\mathfrak{A}$ is denoted by $\operatorname{Can}(\mathfrak{A})$; note that $\operatorname{Can}(\mathfrak{A})$ is a clone and closed. ### 5.1. Behaviours To conveniently work with canonical functions and with functions that locally resemble canonical functions, we need the concept of a _behaviour_ of functions $f\colon A^{k}\to A$ over a structure $\mathfrak{A}$, following [BP11, BP21]. In order to have some flexibility when using the terminology, we work in the setting of _partial functions_ from $A^{k}\to A$, i.e., functions that are only defined on a subset of $A^{k}$. ###### Definition 3 (Behaviours). Let $n,k\in{\mathbb{N}}$. Then an _$n$ -behaviour over $\mathfrak{A}$ (of a $k$-ary partial map from $A^{k}$ to $A$)_ is a partial function from ${\mathcal{O}_{n}(\mathfrak{A})}^{k}$ to ${\mathcal{O}}_{n}(\mathfrak{A})$. A _behaviour over $\mathfrak{A}$_ is a partial function $\mathcal{B}$ from $\bigcup_{i=1}^{\infty}{\mathcal{O}_{i}(\mathfrak{A})}^{k}$ to $\bigcup_{i=1}^{\infty}{\mathcal{O}}_{i}(\mathfrak{A})$ so that for every $i$ we have $\mathcal{B}({\mathcal{O}}_{i}(\mathfrak{A})^{k})\subseteq{\mathcal{O}}_{i}(\mathfrak{A})$. An $n$-behaviour (behaviour) is called _complete_ if it is defined on all of ${\mathcal{O}}_{n}(\mathfrak{A})^{k}$ (on all of $\bigcup_{i=1}^{\infty}{\mathcal{O}_{i}(\mathfrak{A})^{k}}$). Note that every $n$-behaviour is in particular a behaviour. ###### Definition 4 (Behaviours of functions). Let $\mathfrak{A}$ be a structure and $k\in{\mathbb{N}}$. A partial function $f\colon A^{k}\to A$ _realises_ a behaviour $\mathcal{B}$ over $\mathfrak{A}$ if for all $a^{1}_{1},\dots,a^{k}_{n}\in A$, if $O_{i}$ is the orbit of $(a^{i}_{1},\dots,a^{i}_{n})$ for $i\in\\{1,\dots,k\\}$, then ${\mathcal{B}}(O_{1},\dots,O_{k})$ is the orbit of $(f(a^{1}_{1},\dots,a^{k}_{1}),\dots,f(a^{1}_{n},\dots,a^{k}_{n}))$. If $S\subseteq A^{k}$ and $\mathcal{B}$ is a behaviour over $\mathfrak{A}$, then we say that $f\colon A^{k}\to A$ _realises $\mathcal{B}$ on $S$_ if the restriction $f|_{S}$ realises $\mathcal{B}$. The following proposition can be shown by a standard compactness argument. ###### Proposition 5.1. Let $\mathfrak{A}$ be an $\omega$-categorical structure and $k,n\in{\mathbb{N}}$. Then an $n$-behaviour over $\mathfrak{A}$ is realised by some $f\in\operatorname{Pol}(\mathfrak{A})^{(k)}$ if and only if it is realised by some $f\in\operatorname{Pol}(\mathfrak{A})^{(k)}$ on $F^{k}$ for every finite subset $F\subseteq A$. ### 5.2. Canonicity Let $n,k\in{\mathbb{N}}$. A partial function $f\colon A^{k}\to A$ is called _$n$ -canonical over $\mathfrak{A}$_ if it realises a complete $n$-behaviour over $\mathfrak{A}$, and _canonical over $\mathfrak{A}$_ if it realises some complete behaviour over $\mathfrak{A}$; note that this is consistent with the definition given in Section 4. If $S\subseteq A$, then we say that $f$ is _( $n$-) canonical on $S$_ if the restriction $f|_{S}$ is ($n$-) canonical. Suppose that $\mathfrak{B}$ is a homogeneous relational structure with maximal arity $m$. If $f\colon B\to C$ realises some complete $m^{\prime}$-behaviour over $(\mathfrak{B},\mathfrak{C})$ for $m^{\prime}=\max(m,2)$, then $f$ is canonical over $(\mathfrak{B},\mathfrak{C})$ (in other words, in this situation a complete $m$-behaviour uniquely determines a complete behaviour; the requirement $m^{\prime}\geq 2$ is necessary so that the $m^{\prime}$-behaviour allows, for instance, to distinguish constant functions from injective functions). Hence, if $\mathfrak{B}$ is finitely homogeneous, then there are only finitely many complete behaviours for functions of arity $k$. Note that the binary relation $\\{(u,\alpha(u))\colon u\in A^{n},\alpha\in\operatorname{Aut}(\mathfrak{A})\\}$ is a congruence of $\operatorname{Can}(\mathfrak{A})$, and so there is a uniformly continuous clone homomorphism $\xi^{\mathfrak{A}}_{n}$ from $\operatorname{Can}(\mathfrak{A})$ to a clone whose domain ${\mathcal{O}}_{n}(\mathfrak{A})$ is the (finite) set of congruence classes, and which we denote by $\xi^{\mathfrak{A}}_{n}(\operatorname{Can}(\mathfrak{A}))$. If ${\mathscr{C}}\subseteq\operatorname{Can}(\mathfrak{A})$, then we write $\xi^{\mathfrak{A}}_{n}({\mathscr{C}})$ for the subclone of $\xi^{\mathfrak{A}}_{n}(\operatorname{Can}(\mathfrak{A}))$ induced by the image of $\mathscr{C}$ under $\xi^{\mathfrak{A}}_{n}$. Note that if $f\in\operatorname{Can}(\mathfrak{A})$, then $g\colon A^{k}\to A$ has the same behaviour as $f$ over $\mathfrak{A}$ if and only if $f$ interpolates $g$ and $g$ interpolates $f$ over $\mathfrak{A}$. ###### Lemma 5.2 (Proposition 6.6 in [BPP21]). Let $\mathfrak{A}$ be a finitely homogeneous structure. Let $m$ be the maximal arity of $\mathfrak{A}$ and suppose that $f,g\in\operatorname{Can}(\mathfrak{A})$ are such that $\xi^{\mathfrak{A}}_{m}(f)=\xi^{\mathfrak{A}}_{m}(g)$. Then there are $e_{1},e_{2}\in\overline{\operatorname{Aut}(\mathfrak{A})}$ such that $e_{1}\circ f=e_{2}\circ g$. The following proposition can be shown by a compactness argument. ###### Lemma 5.3. Let $\mathfrak{B}$ be a first-order reduct of an $\omega$-categorical structure $\mathfrak{A}$. Let $k,n,m\in{\mathbb{N}}$ and let ${\mathcal{B}}_{1},\dots,{\mathcal{B}}_{m}$ be $n$-behaviours over $(\mathfrak{A}^{(k)},\mathfrak{A})$. Suppose that for every finite $F\subseteq A$ and $i\in\\{1,\dots,m\\}$ there exist $\alpha_{i,1},\dots,\alpha_{i,k}$ and $f\in\operatorname{Pol}(\mathfrak{B})$ such that $f(\alpha_{i,1},\dots,\alpha_{i,k})$ realises ${\mathcal{B}}_{i}$ on $F^{k}$. Then there exist $e_{1,1},\dots,e_{m,k}\in\overline{\operatorname{Aut}(\mathfrak{A})}$ and $f\in\operatorname{Pol}(\mathfrak{B})$ such that for every $i\in\\{1,\dots,m\\}$ the operation $f(e_{i,1},\dots,e_{i,k})$ realises ${\mathcal{B}}_{i}$ on all of $A^{k}$. Moreover, if $I\subseteq\\{1,\dots,k\\}$ and $\alpha_{i,j}=\operatorname{id}$ for all $j\in I$ and every finite $X\subseteq A$ then we can guarantee $e_{i,j}=\operatorname{id}$ for all $j\in I$. ### 5.3. Canonisation The following is from [BPT13]; also see [BP21]. ###### Lemma 5.4 (Canonisation lemma). Let $\mathfrak{A}$ be a countable $\omega$-categorical homogeneous Ramsey structure and let $f\colon A^{k}\to A$. Then there is a canonical function over $\mathfrak{A}$ which is interpolated by $f$ over $\operatorname{Aut}(\mathfrak{A})$. In some situations it is practical to use a finite version of the canonisation lemma (e.g. in the proof of Lemma 7.5 or Lemma 8.1), which can be derived easily from Lemma 5.4 by a compactness argument, or directly from Lemma 3.1. ###### Lemma 5.5. Let $\mathfrak{A}$ be a homogeneous $\omega$-categorical Ramsey structure and $B\subseteq A$ finite. Then there exists a finite $C\subseteq A$ such that for every $f\colon A^{k}\to A$ there are $\alpha_{1},\dots,\alpha_{k}\in\operatorname{Aut}(\mathfrak{A})$ such that $\alpha_{1}(B)\subseteq C,\dots,\alpha_{k}(B)\subseteq C$ and $f(\alpha_{1},\dots,\alpha_{k})$ is canonical on $B^{k}$. ### 5.4. Complexity The following theorem is essentially from [BM18] and uses [Bul17, Zhu17]; since we are not aware of a formulation of the result in the form below, which is the form that is most useful for the purposes of this article, we added a guide to the literature in the proof. ###### Theorem 5.6 ([BM18]). Let $\mathfrak{B}$ be a reduct of a countable finitely bounded homogeneous structure $\mathfrak{A}$ such that there is no uniformly continuous minor- preserving map from $\operatorname{Pol}(\mathfrak{B})\cap\operatorname{Can}(\mathfrak{A})$ to $\operatorname{Proj}$. Then $\operatorname{CSP}(\mathfrak{B})$ is in P. ###### Proof. Let $m$ be the maximal arity of the relations of $\mathfrak{B}$. If $\xi_{m}^{\mathfrak{A}}$ has a (uniformly continuous) minor-preserving map to $\operatorname{Proj}$ then so has $\operatorname{Pol}(\mathfrak{B})$ by composing the maps, contrary to our assumptions. Therefore, $\xi_{m}^{\mathfrak{A}}$ contains a _Siggers operation_ [Sig10]. Lemma 5.2 implies that $\operatorname{Pol}(\mathfrak{B})\cap\operatorname{Can}(\mathfrak{A})$ has a _Siggers operation modulo $\overline{\operatorname{Aut}(\mathfrak{A})}$_ [BP20]. The polynomial-time tractability of $\operatorname{CSP}(\mathfrak{B})$ then follows from Corollary 4.13 in [BM18], which uses [Bul17, Zhu17]. ∎ For many classes of reducts of finitely bounded homogeneous structures, the conditions given in Theorem 4.4 and in Theorem 5.6 cover all the cases. The next section presents a condition from [BM18] that turned out to be highly relevant to understand this phenomenon. ## 6\. Results In this section we state the main results of the paper that we then prove in Section 7 and in Section 8. The applications to RCC5 in spatial reasoning can be found in Section 9. ### 6.1. The Unique Interpolation Property Let $\mathscr{M}$ be a set of function from $A$ to $A$ and let $f,g\colon A^{k}\to A$ be operations. We say that _$f$ interpolates $g$ over ${\mathscr{M}}$_ if for every finite $F\subseteq A$ there are $u,v_{1},\dots,v_{k}\in{\mathscr{M}}$ such that $g(a_{1},\dots,a_{k})=u(f(v_{1}(a_{1}),\dots,v_{k}(a_{k})))$ for all $a_{1},\dots,a_{k}\in F$. ###### Definition 5. Let $\mathfrak{A}$ be a structure and let $\mathscr{C}$ be an operation clone that contains $\operatorname{Aut}(\mathfrak{A})$. A function $\zeta$ defined on $\mathscr{D}\subseteq\mathscr{C}$ has the _unique interpolation property (UIP) with respect to ${\mathscr{C}}$ over $\mathfrak{A}$_ if $\zeta(g)=\zeta(h)$ for all $g,h\in\mathscr{D}$ that are interpolated by the same operation $f\in{\mathscr{C}}$ over $\operatorname{Aut}(\mathfrak{A})$. The main result of this section, Lemma 6.1, gives a sufficient condition for the existence of an extension of a minor-preserving map defined on the canonical polymorphisms of a structure $\mathfrak{B}$ to a uniformly continuous minor-preserving map defined on all of $\operatorname{Pol}(\mathfrak{B})$. The statement is inspired by Theorem 5.5 in [BM18], and similar statements played an important role in [BMM18]. However, the original proof in [BM18] contains a mistake (in the proof the mashup theorem, Theorem 5.5 in [BM18], more specifically in the claim that $\phi$ is a minor-preserving map) and fixing this with a general approach is one of the contributions of the present article. ###### Lemma 6.1 (The extension lemma). Let $\mathfrak{B}$ be a reduct of a countable finitely homogeneous Ramsey structure $\mathfrak{A}$ and let $\mathscr{D}$ be an operation clone. Suppose that $\zeta\colon\operatorname{Pol}(\mathfrak{B})\cap\operatorname{Can}(\mathfrak{A})\to\mathscr{D}$ is minor-preserving and has the UIP with respect to $\operatorname{Pol}(\mathfrak{B})$ over $\mathfrak{A}$. Then $\zeta$ can be extended to a uniformly continuous minor-preserving map $\tilde{\zeta}\colon\operatorname{Pol}(\mathfrak{B})\to\mathscr{D}$. The extension lemma motivates the study of the UIP in the context of complexity classification of CSPs, and we obtain the following in Section 7.4. ###### Corollary 6.2. Let $\mathfrak{C}$ be a reduct of a finitely bounded homogeneous structure $\mathfrak{B}$, which is itself a reduct of a finitely homogeneous Ramsey structure $\mathfrak{A}$. Suppose that whenever there is a uniformly continuous minor-preserving map from $\operatorname{Pol}(\mathfrak{C})\cap\operatorname{Can}(\mathfrak{B})$ to $\operatorname{Proj}$ then there is also a minor-preserving map from $\operatorname{Pol}(\mathfrak{C})\cap\operatorname{Can}(\mathfrak{A})$ to $\operatorname{Proj}$ which has the UIP with respect to $\operatorname{Pol}(\mathfrak{C})$ over $\mathfrak{A}$. Then $\operatorname{CSP}(\mathfrak{C})$ is in P or NP-complete. ### 6.2. Other Characterisations of the UIP To verify the UIP, the following characterisation of the UIP is helpful, since it allows us to focus on binary polymorphisms. Let ${\mathscr{C}}$ be an operation clone and let ${\mathscr{M}}\subseteq{\mathscr{C}}^{(1)}$. If $v_{1},\dots,v_{k}\in{\mathscr{M}}$ we write $f(v_{1},\dots,v_{k})$ for the operation that maps $(a_{1},\dots,a_{k})$ to $f(v_{1}(a_{1}),\dots,v_{k}(a_{k}))$. We say that a function $\zeta$ defined on $\mathscr{C}$ is _$\mathscr{M}$ -invariant_ if for all $f\in{\mathscr{C}}^{(k)}$ and $u,v_{1},\dots,v_{k}\in\mathscr{M}$ we have (2) $\displaystyle\zeta(f)=\zeta\big{(}u(f(v_{1},\dots,v_{k}))\big{)}.$ It follows from known results (see Proposition 7.7) that if $\mathfrak{A}$ is a finitely homogeneous structure and $\mathscr{C}\subseteq\operatorname{Can}(\mathfrak{A})$ is a closed clone, then there exists a uniformly continuous minor-preserving map from $\mathscr{C}$ to $\operatorname{Proj}$ if and only if there exists a minor-preserving map $\zeta\colon{\mathscr{C}}\to\operatorname{Proj}$ which is $\overline{\operatorname{Aut}(\mathfrak{A})}$-invariant. ###### Theorem 6.3 (Binary UIP verification). Let $\mathfrak{A}$ be a countable finitely homogeneous Ramsey structure and let $\mathfrak{C}$ be a reduct of $\mathfrak{A}$. Let $\mathscr{C}$ be the clone of all polymorphisms of $\mathfrak{C}$ that are canonical over $\mathfrak{A}$ and let $\zeta\colon{\mathscr{C}}\to\operatorname{Proj}$ be an $\overline{\operatorname{Aut}(\mathfrak{A})}$-invariant minor-preserving map. Then the following are equivalent. 1. (1) $\zeta$ has the UIP with respect to $\operatorname{Pol}(\mathfrak{C})$ over $\mathfrak{A}$. 2. (2) for all $f\in\operatorname{Pol}(\mathfrak{C})^{(2)}$ and $u_{1},u_{2}\in{\overline{\operatorname{Aut}(\mathfrak{A})}}$, if $f(\operatorname{id},u_{1})$ and $f(\operatorname{id},u_{2})$ are canonical over $\mathfrak{A}$ then $\zeta(f(\operatorname{id},u_{1}))=\zeta(f(\operatorname{id},u_{2})).$ Theorem 6.3 can be strengthened further. We have already seen that every homogeneous $\omega$-categorical Ramsey structure contains two independent elementary substructures. The idea of the following theorem is that it suffices to verify the UIP on such a pair of substructures. ###### Theorem 6.4. Let $\mathfrak{A}$ be a countable finitely homogeneous Ramsey structure and let $\mathfrak{C}$ be a reduct of $\mathfrak{A}$. Let ${\mathscr{C}}$ be the clone of all polymorphisms of $\mathfrak{C}$ that are canonical over $\mathfrak{A}$ and let $\zeta\colon{\mathscr{C}}\to\operatorname{Proj}$ be an $\overline{\operatorname{Aut}(\mathfrak{A})}$-invariant minor-preserving map. Then the following are equivalent. 1. (1) $\zeta$ has the UIP with respect to $\operatorname{Pol}(\mathfrak{C})$ over $\mathfrak{A}$. 2. (2) for every pair $(\mathfrak{A}_{1},\mathfrak{A}_{2})$ of independent elementary substructures of $\mathfrak{A}$ such that for all $f\in\operatorname{Pol}(\mathfrak{C})^{(2)}$, $u_{1},u_{2}\in{\overline{\operatorname{Aut}(\mathfrak{A})}}$ with $u_{1}(A)\subseteq A_{1},u_{2}(A)\subseteq A_{2}$, if $f(\operatorname{id},u_{1})$ and $f(\operatorname{id},u_{2})$ are canonical over $\mathfrak{A}$ then $\zeta(f(\operatorname{id},u_{1}))=\zeta(f(\operatorname{id},u_{2})).$ ## 7\. Proof of the Extension Lemma This section contains the proof of Lemma 6.1; Corollary 6.2 about the complexity of CSPs is then an easy consequence. ### 7.1. Diagonal Interpolation In our proofs we need the concept of _diagonal interpolation_ , which is a more restricted form of interpolation, and was already used in [BMM18]. ###### Definition 6. Let $\mathscr{M}$ be a set of functions from $A$ to $A$. We say that $f\colon A^{k}\to A$ _diagonally interpolates $g\colon A^{k}\to A$ over ${\mathscr{M}}$_ if for every finite $F\subseteq A$ there are $u,v\in{\mathscr{M}}$ such that $g(a_{1},\dots,a_{k})=u(f(v(a_{1}),\dots,v(a_{k})))$ for all $a_{1},\dots,a_{k}\in F$. Diagonal interpolation is well-behaved with respect to minor identities. ###### Lemma 7.1. Let $\mathscr{M}$ be a set of functions from $A$ to $A$, let $f\colon A^{k}\to A$ be an operation, and let $\pi=(\pi^{n}_{i_{1}},\dots,\pi^{n}_{i_{k}})$ be a vector of $k$ projections, each of arity $n$. Suppose that $f$ diagonally interpolates $g$ over $\mathscr{M}$. Then $f\circ\pi$ diagonally interpolates $g\circ\pi$ over ${\mathscr{M}}$. ###### Proof. By assumption $g\in\overline{\\{u(f(v,\dots,v))\colon u,v\in\mathscr{M}\\}}$. Since composition with $\pi$ is continuous, $\displaystyle g\circ\pi$ $\displaystyle\in\overline{\\{u(f(v,\dots,v))\circ\pi\colon u,v\in\mathscr{M}\\}}$ $\displaystyle=\overline{\\{u(f\circ\pi)(v,\dots,v)\colon u,v\in\mathscr{M}\\}}$ and thus $f\circ\pi$ diagonally interpolates $g\circ\pi$ over ${\mathscr{M}}$. ∎ We want to stress that the innocent-looking Lemma 7.1 fails for interpolation instead of diagonal interpolation: the following example shows that Lemma 7.1 fails for interpolation instead of diagonal interpolation. ###### Example 7.2. Let $\mathfrak{A}$ be a structure with countable domain and two disjoint infinite unary relations $U_{1}$ and $U_{2}$. Let $f\colon A^{3}\to A$ be an injective function such that for every $i\in\\{1,2\\}$ * • $f(a,b,c)\in U_{i}$ if $a\in U_{i}$ and $b\neq c$, * • $f(a,b,c)\in U_{i}$ if $b\in U_{i}$ and $b=c$. Let $e_{1},e_{2}$ be two self-embeddings of $\mathfrak{A}$ with disjoint images and let $g\colon A^{3}\to A$ be given by $g(x,y,z)\coloneqq f(x,e_{1}(y),e_{2}(z)).$ Then $g$ is injective and for each $i\in\\{1,2\\}$ we have $g(U_{i},A,A)\subseteq U_{i}$. Hence, $g$ is canonical and $\xi_{1}(g)=\pi^{3}_{1}$. We claim that $f^{\prime}\coloneqq f\circ(\pi^{2}_{1},\pi^{2}_{2},\pi^{2}_{2})$ does not interpolate $g^{\prime}\coloneqq g\circ(\pi^{2}_{1},\pi^{2}_{2},\pi^{2}_{2})$. For each $i\in\\{1,2\\}$ we have $f^{\prime}(A,U_{i})\subseteq U_{i}$, so $f^{\prime}$ is canonical and $\xi_{1}(f^{\prime})=\pi^{2}_{2}$. On the other hand $\xi_{1}(g^{\prime})=\xi_{1}(g)(\pi^{2}_{1},\pi^{2}_{2},\pi^{2}_{2})=\pi^{3}_{1}(\pi^{2}_{1},\pi^{2}_{2},\pi^{2}_{2})=\pi^{2}_{1}$ which proves the claim. The next lemma plays an important role in the proof of Lemma 6.1. ###### Lemma 7.3. Let $\mathfrak{B}$ be a reduct of a countable finitely homogeneous Ramsey structure $\mathfrak{A}$. Let $f\in\operatorname{Pol}(\mathfrak{B})$. Then there exists $g\in\operatorname{Pol}(\mathfrak{B})$ that diagonally interpolates both $f$ and an operation in $\operatorname{Can}(\mathfrak{A})$ over $\operatorname{Aut}(\mathfrak{A})$. ###### Proof. Let $k$ be the arity of $f$. We first show a local version of the statement, and then derive the statement by a compactness argument. The local version is that for every finite $X\subseteq A$ there exists $g\in\operatorname{Pol}(\mathfrak{B})$ and $e\in\overline{\operatorname{Aut}(\mathfrak{A})}$ such that $g(e,\dots,e)$ is canonical on $X^{k}$ over $\mathfrak{B}$ and $g$ agrees with $f$ on $X^{k}$. Let $c=(c_{1},\dots,c_{n})$ be such that $\\{c_{1},\dots,c_{n}\\}=X$. By Theorem 3.2 there are two independent elementary substructures $\mathfrak{C}_{1}$ and $\mathfrak{C}_{2}$ of $\mathfrak{A}$. Since $\mathfrak{A}$ is $\omega$-categorical, for $i\in\\{1,2\\}$ there exists an isomorphism $e_{i}$ from $\mathfrak{A}$ to $\mathfrak{C}_{i}$; by the homogeneity of $\mathfrak{A}$ we have $e_{i}\in\overline{\operatorname{Aut}(\mathfrak{A})}$. The restriction of $e_{1}$ to $X$ can be extended to an automorphism $\delta\in\operatorname{Aut}(\mathfrak{A})$. Then $\delta^{-1}(C_{1})$ and $\delta^{-1}(C_{2})$ induce independent elementary substructures of $\mathfrak{A}$, and $X\subseteq\delta^{-1}(C_{1})$. So we may assume without loss of generality that $X\subseteq C_{1}$. Since $\mathfrak{A}$ is Ramsey, $f(e,\dots,e)$ interpolates a canonical function over $\operatorname{Aut}(\mathfrak{A})$; so there are $\delta_{1},\dots,\delta_{k}\in\operatorname{Aut}(\mathfrak{A})$ such that $f(e\circ\delta_{1},\dots,e\circ\delta_{k})$ is canonical on $X^{k}$. In other words, for $Y_{i}\coloneqq e\circ\delta_{1}(X)$, the map $f$ is canonical on $Y_{1}\times\cdots\times Y_{k}$ over $\mathfrak{A}$. Since $X\subseteq C_{1}$ and $Y_{i}\subseteq C_{2}$, the tuples $e\circ\delta_{1}(c),\dots,e\circ\delta_{k}(c)$ all have the same type over $c$. Hence, there exist $\alpha_{i}\in\operatorname{Aut}(\mathfrak{A})$ such that $\alpha_{i}(Y_{1})=Y_{i}$ and $\alpha(c)=c$. Define $g\coloneqq f(\alpha_{1},\dots,\alpha_{k})$. Then $f$ and $g$ agree on $X^{k}$. We claim that $h\coloneqq g(e\delta_{1},\dots,e\delta_{1})$ is canonical on $X^{k}$ over $\mathfrak{A}$. This follows from $f$ being canonical on $\displaystyle Y_{1}\times\cdots\times Y_{k}$ $\displaystyle=\alpha_{1}(Y_{1})\times\cdots\times\alpha_{k}(Y_{1})$ $\displaystyle=\alpha_{1}e\delta_{1}(X)\times\cdots\alpha_{k}e\delta_{k}(X)$ since $h$ equals $f(\alpha_{1}e\delta_{1},\dots,\alpha_{k}e\delta_{k})$ on $X^{k}$. To show how the local version implies the statement of the lemma, let $\sigma$ be the signature of $\mathfrak{B}$ and let $\mathfrak{C}$ be an expansion of $\mathfrak{B}$ by countably many constants such that every element of $\mathfrak{C}$ is named by a constant symbol. Let $\rho$ be the signature of $\mathfrak{C}$ together with a new a $k$-ary function symbol $g$. Consider the $\rho$-theory $T$ consisting of the union of the following first-order sentences: 1. (1) $\operatorname{Th}(\mathfrak{C})$; 2. (2) A first-order sentence which asserts that $g$ preserves all relations from $\sigma$; 3. (3) For all $l\in{\mathbb{N}}$ and constant symbols $c_{0},c_{1},\dots,c_{k}\in\rho$ such that $f(c_{1}^{\mathfrak{C}},\dots,c_{k}^{\mathfrak{C}})=c_{0}^{\mathfrak{C}}$ the sentence $g(c_{1},\dots,c_{k})=c_{0}$. 4. (4) for all constant symbols $c_{1},\dots,c_{l}\in\sigma$ the sentence $\psi_{c_{1},\dots,c_{l}}$ which expresses that there exist $y_{1},\dots,y_{l}$ such that $\operatorname{typ}^{\mathfrak{A}}(y_{1},\dots,y_{l})=\operatorname{typ}^{\mathfrak{A}}(c_{1},\dots,c_{l})$ and $g$ is canonical on $\\{y_{1},\dots,y_{l}\\}^{k}$. It follows from the basic facts about $\omega$-categorical structures from Section 2 that such sentences exist. Claim. Every finite subset $S$ of $T$ has a model. Let $X$ be the (finite) set of all constants $c$ such that the respective constant symbol appears in a sentence of $S$ or in $\\{c_{1},\dots,c_{l}\\}$ for some $\psi_{c_{1},\dots,c_{l}}\in S$. Let $d_{1},\dots,d_{n}$ be some enumeration of $X$. Then $\psi_{d_{1},\dots,d_{n}}$ implies every sentence in $S^{\prime}$ from item (4). By the local version of the statement there exists $g\in\operatorname{Pol}(\mathfrak{B})$ and $e\in\overline{\operatorname{Aut}(\mathfrak{A})}$ such that * • $g(e,\dots,e)$ is canonical on $X^{k}$ over $\mathfrak{A}$, and * • $g$ agrees with $f$ on $X^{k}$. Let $\mathfrak{D}$ be the $\rho$-expansion of $\mathfrak{C}$ where $g^{\mathfrak{D}}\coloneqq g$. Then $\mathfrak{D}$ satisfies $S$. This is clear for the sentences from item (1), (2), and (3). Finally, since $g(e,\dots,e)$ is canonical on $X^{k}$ over $\mathfrak{A}$ the sentences of $S^{\prime}$ from item (4) can be satisfied by setting $y_{i}\coloneqq e(c_{i})$. By the compactness theorem of first-order logic, $T$ has a model $\mathfrak{D}$. By the downward Löwenheim-Skolem theorem we may also assume that $\mathfrak{D}$ is countable. Then the $\sigma$-reduct of $\mathfrak{D}$ satisfies $\operatorname{Th}(\mathfrak{B})$ and hence is isomorphic to $\mathfrak{B}$ by the $\omega$-categoricity of $\mathfrak{B}$. We may therefore identify the element of $\mathfrak{D}$ with the elements of $\mathfrak{B}$ along this isomorphism and henceforth assume that $\mathfrak{D}$ is an expansion of $\mathfrak{A}$. Because of the sentences under (2) we have that $g^{\mathfrak{D}}$ is a polymorphism of $\mathfrak{B}$. The map $e\colon B\to B$ given by $c^{\mathfrak{C}}\mapsto c^{\mathfrak{D}}$ is an elementary self-embedding. The sentences under (3) imply that $g(e,\dots,e)$ equals $e(f)$, so $g$ diagonally interpolates $f$. Finally, the sentences under (4) imply that $g$ diagonally interpolates some function from $\operatorname{Can}(\mathfrak{B})$. ∎ ### 7.2. Rich subsets To prove the existence of canonical functions it is useful to introduce a notion of _rich_ substructures of a structure; this will be needed in the full version of the proofs for Lemma 6.1, but also explicitly in Section 9 (e.g., in Lemma 9.18) when we apply the general results to the spatial reasoning formalism RCC5. ###### Definition 7. Let $\mathfrak{A}$ be a countable finitely homogeneous structure and let $m$ be the maximal arity of $\mathfrak{A}$. Let ${\mathscr{C}}\subseteq{\mathscr{O}}_{A}$ be a closed clone containing $\operatorname{Aut}(\mathfrak{A})$ and let $k\in{\mathbb{N}}$. Then we say that $X\subseteq A$ is _$k$ -rich with respect to ${\mathscr{C}}$_ if * • every $m$-orbit of $\mathfrak{A}^{(k)}$ contains a tuple from $(X^{k})^{m}$; * • every behaviour over $\mathfrak{A}$ which is realised on $X^{k}$ by some operation in $\mathscr{C}$ is also realised on $A^{k}$ by some operation in $\mathscr{C}$. Note that if $X$ is $k$-rich then every behaviour which is realisable on $X^{k}$ is complete. It is clear from the definition that if $X\subseteq A$ is $k$-rich, and $X\subseteq Y$, then $Y$ is also $k$-rich. ###### Lemma 7.4. Let $\mathfrak{A}$ be a countable finitely homogeneous structure. Let $\mathscr{C}$ be a closed clone containing $\operatorname{Aut}(\mathfrak{A})$ and let $k\in{\mathbb{N}}$. Then there exists a finite $X\subseteq A$ which is $k$-rich with respect to ${\mathscr{C}}$. ###### Proof. Let $m$ be the maximal arity of $\mathfrak{A}$. The structure $\mathfrak{A}^{(k)}$ is $\omega$-categorical and hence has finitely many $m$-orbits. Therefore, we can choose a finite subset $X_{0}$ of $A$ such that every orbit of $m$-tuples in $\operatorname{Aut}(\mathfrak{A}^{(k)})$ intersects $(X_{0}^{k})^{m}$. Let $X_{0}\subset X_{1}\subset\cdots$ be finite subsets of $A$ such that $\bigcup_{i=1}^{\infty}X_{i}=A$. Let $\mathcal{B}_{i}$ denote the (finite) set of all $m$-behaviours which are realised on $X^{k}_{i}$ by some operation of ${\mathscr{C}}$. Then ${\mathcal{B}}_{0}\supseteq{\mathcal{B}}_{1}\supseteq\cdots$ and hence there exists an $N$ such that ${\mathcal{B}}_{N}={\mathcal{B}}_{N+1}=\cdots$. Then Proposition 5.1 implies that every behaviour in ${\mathcal{B}}_{N}$ is realised on $A^{k}$ by some operation in ${\mathscr{C}}$. This proves that $X_{N}$ is $k$-rich: the first item is satisfied because $X_{0}\subseteq X_{N}$, and the second item because every behaviour that is realised in $X^{k}_{N}$ is also realised on $A^{k}$. ∎ ###### Definition 8. Let $\mathfrak{A}$ and $\mathscr{C}$ be as in Definition 7, let $\eta$ be a function defined on ${\mathscr{C}}\cap\operatorname{Can}(\mathfrak{A})$, let $F$ be $k$-rich with respect to ${\mathscr{C}}$ for some $k\geq 1$, and let $f\in\mathscr{C}^{(k)}$. Suppose that $f$ has on $F^{k}$ the same complete behaviour as $g\in\operatorname{Can}(\mathfrak{A})$. Then we will also write $\eta(f|_{F^{k}})$ for $\eta(g)$. ### 7.3. Interpolation invariance It has been shown in [BP15] (see Theorem 6.4 in [BOP18]) that if $\mathscr{C}\subseteq{\mathscr{O}_{B}}$ is a closed clone and $\mathscr{G}$ an oligomorphic permutation group over the same base set $B$, then every $\mathscr{G}$-invariant continuous map $\zeta$ defined on $\mathscr{C}$ is uniformly continuous. In Lemma 7.5 we present a sufficient condition for uniform continuity which does not require that the map $\zeta$ we start from is continuous, for the special case that $\mathscr{G}=\operatorname{Aut}(\mathfrak{C})$ for a finitely homogeneous Ramsey structure $\mathfrak{C}$. ###### Definition 9. Let $\mathscr{C}$ be a closed operation clone on a set $B$ and let $\mathscr{M}$ be a set of functions from $B$ to $B$. A function $\zeta$ defined on ${\mathscr{C}}$ is called _interpolation invariant over $\mathscr{M}$_ if $\zeta(f)=\zeta(g)$ whenever $f\in{\mathscr{C}}$ interpolates $g\in{\mathscr{C}}$ over $\mathscr{M}$. Clearly, interpolation invariance over $\mathscr{M}$ implies ${\mathscr{M}}$-invariance; see 2. ###### Lemma 7.5. Let $\mathfrak{B}$ be a first-order reduct of a countable finitely homogeneous Ramsey structure $\mathfrak{A}$. Let $\zeta$ be a function defined on $\operatorname{Pol}(\mathfrak{B})$ which is interpolation invariant over $\operatorname{Aut}(\mathfrak{A})$. Then $\zeta$ is uniformly continuous. ###### Proof. Let $m$ be the maximal arity of $\mathfrak{A}$. Suppose for contradiction that $\zeta$ is not uniformly continuous. Then there exists a $k\in{\mathbb{N}}$ so that for every finite $X\subseteq A$ there exist $f_{X},g_{X}\in\operatorname{Pol}(\mathfrak{B})$ such that $(f_{X})|_{X^{k}}=(g_{X})|_{X^{k}}$, but $\zeta(f_{X})\neq\zeta(g_{X})$. By Lemma 7.4, there exists a finite $A_{0}\subseteq A$ which is $k$-rich with respect to $\operatorname{Pol}(\mathfrak{B})$. Claim. For every finite $F\subseteq A$ that contains $A_{0}$ there exists an $h\in\operatorname{Pol}(\mathfrak{B})^{(k)}$ and $h^{\prime}\in\operatorname{Pol}(\mathfrak{B})^{(k)}\cap\operatorname{Can}(\mathfrak{A})$ such that $h$ and $h^{\prime}$ have the same complete behaviour on $F^{k}$ and $\zeta(h)\neq\zeta(h^{\prime})$. By Lemma 5.5 there exists a finite $C\subseteq A$ such that for every $h\colon A^{k}\to A$ there are $\alpha_{1},\dots,\alpha_{k}\in\operatorname{Aut}(\mathfrak{A})$ such that $\alpha_{1}(F)\subseteq C,\dots,\alpha_{k}(F)\subseteq C$ and $h(\alpha_{1},\dots,\alpha_{k})$ is canonical on $F^{k}$. In particular, this holds for the operations $f_{C},g_{C}$ introduced above, so that $f^{\prime}\coloneqq f_{C}(\alpha_{1},\dots,\alpha_{k})$ and $g^{\prime}\coloneqq g_{C}(\alpha_{1},\dots,\alpha_{k})$ are canonical on $F^{k}$. We have $(f_{C})|_{C^{k}}=(g_{C})|_{C^{k}}$ and $\zeta(f_{C})\neq\zeta(g_{C})$, and $\zeta(f_{C})=\zeta(f^{\prime})$ and $\zeta(g_{C})=\zeta(g^{\prime})$ because $f_{C}$ interpolates $f^{\prime}$ and $g_{C}$ interpolates $g^{\prime}$ over $\operatorname{Aut}(\mathfrak{A})$. Since $A_{0}\subseteq F$ is $k$-rich there exist $h^{\prime}\in\operatorname{Pol}^{(k)}(\mathfrak{B})\cap\operatorname{Can}(\mathfrak{A})$ such that $h^{\prime}$ has the same complete behaviour as $f^{\prime}$ on $F^{k}$, and therefore also the same complete behaviour as $g^{\prime}$ on $F^{k}$. Hence, we must have $\zeta(h^{\prime})\neq\zeta(f^{\prime})$ or $\zeta(h^{\prime})\neq\zeta(g^{\prime})$, and hence either $(f^{\prime},h^{\prime})$ or $(g^{\prime},h^{\prime})$ provide us witnesses for the claim. Let $A_{0}\subseteq A_{1}\subseteq\cdots$ be finite such that $\bigcup_{l\in{\mathbb{N}}}A_{l}=A$. By the claim above, for each $l\in{\mathbb{N}}$ we can find a function $h\in\operatorname{Pol}(\mathfrak{B})$ and $h^{\prime}\in\operatorname{Pol}(\mathfrak{B})^{(k)}\cap\operatorname{Can}(\mathfrak{A})$ such that $h$ and $h^{\prime}$ have the same $m$-behaviour ${\mathcal{H}}_{l}$ on $F^{k}$ and $\zeta(h)\neq\zeta(h^{\prime})$. By Lemma 5.4 the operation $h$ interpolates an operation $g$ with a complete behaviour ${\mathcal{B}}_{l}$ over $\mathfrak{A}$, and $\zeta(g)=\zeta(h)\neq\zeta(h^{\prime})$. Since there are finitely many $m$-behaviours over $\mathfrak{A}$, there exists $({\mathcal{H}},{\mathcal{B}})$ such that $({\mathcal{H}}_{l},{\mathcal{B}}_{l})=({\mathcal{H}},{\mathcal{B}})$ for infinitely many $l\in{\mathbb{N}}$. Lemma 5.3 shows that there exists $f\in\operatorname{Pol}(\mathfrak{B})$ and $e_{1},e^{\prime}_{1},\dots,e_{k},e_{k}^{\prime}\in\overline{\operatorname{Aut}(\mathfrak{A})}$ such that $f\circ(e_{i},\dots,e_{k})$ has behaviour ${\mathcal{H}}$ and $f\circ(e^{\prime}_{i},\dots,e^{\prime}_{k})$ has behaviour ${\mathcal{B}}$. Hence, $f$ interpolates two functions that $\zeta$ maps to different values, contradicting interpolation invariance of $\zeta$. ∎ ###### Lemma 7.6. Let $\mathfrak{A}$ be a finitely related homogeneous structure and let $\mathscr{C}$ be a subclone of $\operatorname{Can}(\mathfrak{A})$. Let $\zeta\colon{\mathscr{C}}\to\operatorname{Proj}$ be $\overline{\operatorname{Aut}(\mathfrak{A})}$-invariant. Then $\zeta$ is uniformly continuous. ###### Proof. Let $m$ be the maximal arity of $\mathfrak{A}$, and let $X\subseteq A$ be finite so that every orbit of $m$-tuples contains a witness in $X^{m}$. Let $k\in{\mathbb{N}}$ and $f,g\in{\mathscr{C}}^{(k)}$ be such that $f|_{X^{k}}=g|_{X^{k}}$. Then $\zeta(f)=\zeta(g)$ by the choice of $X$. By Lemma 5.2, there are $e_{1},e_{2}\in\overline{\operatorname{Aut}(\mathfrak{A})}$ such that $e_{1}\circ f=e_{2}\circ g$. The $\overline{\operatorname{Aut}(\mathfrak{A})}$-invariance of $\zeta$ implies that $\zeta(f)=\zeta(e_{1}\circ f)=\zeta(e_{2}\circ g)=\zeta(g)$, proving the uniform continuity of $\zeta$. ∎ The following can be obtained by combining known results in the literature. ###### Proposition 7.7. Let $\mathfrak{B}$ be finitely related homogeneous with maximal arity $m\in{\mathbb{N}}$. Let $\mathfrak{C}$ be a first-order reduct of $\mathfrak{B}$ which is a model-complete core. Suppose that ${\mathscr{C}}\coloneqq\operatorname{Pol}(\mathfrak{C})\subseteq\operatorname{Can}(\mathfrak{B})$. Then the following are equivalent. 1. (1) There is no uniformly continuous minor-preserving map from ${\mathscr{C}}\to\operatorname{Proj}$. 2. (2) There is no $\overline{\operatorname{Aut}(\mathfrak{B})}$-invariant minor- preserving map ${\mathscr{C}}\to\operatorname{Proj}$. 3. (3) $\xi^{\mathfrak{B}}_{m}({\mathscr{C}})$ has no minor-preserving map to $\operatorname{Proj}$. 4. (4) $\xi^{\mathfrak{B}}_{m}({\mathscr{C}})$ has a cyclic operation. ###### Proof. $(1)\Rightarrow(2)$: An immediate consequence of Lemma 7.6. $(2)\Rightarrow(3)$: If $\xi^{{\mathfrak{C}}}_{m}({\mathscr{C}})$ has a minor- preserving map to $\operatorname{Proj}$, then we can compose it with $\xi^{{\mathfrak{C}}}_{m}$ and obtain a $\overline{\operatorname{Aut}(\mathfrak{B})}$-invariant minor-preserving map from ${\mathscr{C}}$ to $\operatorname{Proj}$, and $(2)$ does not hold. $(3)\Rightarrow(4)$ has been shown in [BOP18, BK12]. $(4)\Rightarrow(1)$. Since ${\mathfrak{C}}$ is finitely homogeneous, it has less than doubly exponential growth. Therefore, $\mathfrak{B}$ has less than doubly exponential growth as well, and the statement follows from results in [BKO+19]. ∎ ### 7.4. Proofs of main results ###### Proof of Lemma 6.1. Let $f\in\operatorname{Pol}(\mathfrak{B})^{(k)}$. Since $\mathfrak{A}$ is a finitely homogeneous Ramsey structure, we can apply the canonisation lemma (Lemma 5.4), which implies that $f$ interpolates a function $g\in\operatorname{Can}(\mathfrak{A})$ over $\operatorname{Aut}(\mathfrak{A})$. Define $\bar{\zeta}(f)\coloneqq\zeta(g)$; this is well-defined because $\zeta$ has the UIP with respect to $\operatorname{Pol}(\mathfrak{B})$. Note that if $f\in\operatorname{Can}(\mathfrak{A})$, then $\bar{\zeta}(f)=\zeta(f)$, so $\bar{\zeta}$ extends $\zeta$. Claim 1. $\bar{\zeta}$ is $\overline{\operatorname{Aut}(\mathfrak{A})}$-invariant. Suppose that $f,g\in\operatorname{Pol}(\mathfrak{B})$ are such that $f$ interpolates $g$ over $\operatorname{Aut}(\mathfrak{A})$. Lemma 5.4 implies that $g$ interpolates over $\operatorname{Aut}(\mathfrak{A})$ an operation $h\in\operatorname{Can}(\mathfrak{A})$. Then $f$ interpolates $h$ over $\operatorname{Aut}(\mathfrak{A})$, too, and we have $\bar{\zeta}(f)=\zeta(h)=\bar{\zeta}(g)$. Claim 2. $\bar{\zeta}$ is uniformly continuous. This follows from Claim 1 by Lemma 7.5. Claim 3. $\bar{\zeta}$ is minor-preserving. Arbitrarily choose $f\in\operatorname{Pol}(\mathfrak{B})^{(k)}$ and a vector $\pi=(\pi^{n}_{i_{1}},\dots,\pi_{i_{k}}^{n})$ of projections of arity $n\in{\mathbb{N}}$. We have to show that $\bar{\zeta}(f\circ\pi)=\bar{\zeta}(f)\circ\pi$. By Lemma 7.3, there exists $g\in\operatorname{Pol}(\mathfrak{B})^{(k)}$ that diagonally interpolates $f$ and diagonally interpolates $h\in\operatorname{Pol}(\mathfrak{B})^{(k)}\cap\operatorname{Can}(\mathfrak{A})$. We obtain $\displaystyle\bar{\zeta}(f\circ\pi)$ $\displaystyle=\bar{\zeta}(g\circ\pi)\quad\quad\quad\quad\text{(Lem.~{}\ref{lem:diag- minor} and interpol.-inv.\ of $\bar{\zeta}$)}$ $\displaystyle=\bar{\zeta}(h\circ\pi)\quad\quad\quad\quad\text{(Lem.~{}\ref{lem:diag- minor} and interpol.-inv.\ of $\bar{\zeta}$)}$ $\displaystyle=\zeta(h\circ\pi)\quad\quad\quad\quad\text{($h\circ\pi\in\operatorname{Pol}(\mathfrak{B})\cap\operatorname{Can}(\mathfrak{A})$)}$ $\displaystyle=\zeta(h)\circ\pi\quad\quad\quad\quad\text{($\zeta$ is minor preserving)}$ $\displaystyle=\bar{\zeta}(h)\circ\pi=\bar{\zeta}(g)\circ\pi=\bar{\zeta}(f)\circ\pi.$ Claim 2 and 3 imply the statement of the theorem. ∎ ###### Proof of Corollary 6.2. Let $\mathfrak{C}\in{\mathcal{C}}$. First suppose that $\operatorname{Pol}(\mathfrak{C})\cap\operatorname{Can}(\mathfrak{B})$ does not have a uniformly continuous minor-preserving map to the projections. Theorem 5.6 implies that $\operatorname{CSP}(\mathfrak{C})$ is in P. Now suppose that $\operatorname{Pol}(\mathfrak{C})\cap\operatorname{Can}(\mathfrak{B})$ has a uniformly continuous minor-preserving map to $\operatorname{Proj}$. Then the assumptions imply that there is also a uniformly continuous minor-preserving map $\zeta\colon\operatorname{Pol}(\mathfrak{C})\cap\operatorname{Can}(\mathfrak{A})\to\operatorname{Proj}$ that has the UIP with respect to $\operatorname{Pol}(\mathfrak{C})$ over $\mathfrak{A}$. Lemma 6.1 shows that $\zeta$ can be extended to a uniformly continuous minor-preserving map $\bar{\zeta}\colon\operatorname{Pol}(\mathfrak{C})\to\operatorname{Proj}$. Then $\operatorname{CSP}(\mathfrak{C})$ is NP-hard by Theorem 4.4. Since CSPs of reducts of finitely bounded structures are in NP, this concludes the proof. ∎ Note that the proof of Corollary 6.2 shows that Conjecture 4.5 holds for all structures $\mathfrak{C}\in{\mathcal{C}}$. ## 8\. Verification of the UIP In this section we show that if a minor-preserving map to the clone of projections does not have the UIP, then this is witnessed by binary operations of a very special form, proving Theorem 6.3 and its strengthening Theorem 6.4. We need the following ‘higher-dimensional checker board’ canonisation lemma. ###### Lemma 8.1. Let $\mathfrak{A}$ be a countable finitely homogeneous Ramsey structure and $f\colon A^{k}\to A$. Suppose that $f$ interpolates over $\mathfrak{A}$ the operations $h_{1},\dots,h_{m}\in\operatorname{Can}(\mathfrak{A})$. Then for every finite $X\subseteq A$ there exist $\alpha_{1,1},\dots,\alpha_{m,k}\in\operatorname{Aut}(\mathfrak{A})$ such that * • for every $i\in\\{1,\dots,m\\}$ the operation $f(\alpha_{i,1},\dots,\alpha_{i,k})$ has the same behaviour as $h_{i}$ on $X^{k}$, and * • for all $u_{1},\dots,u_{k}\in\\{1,\dots,m\\}$ the operation $f(\alpha_{u_{1},1},\dots,\alpha_{u_{k},k})$ is canonical on $X^{k}$. ###### Proof. Let $u^{1},\dots,u^{p}$ be an enumeration of $\\{1,\dots,m\\}^{k}$. We show by induction on $q\in\\{1,\dots,p\\}$ that for every finite $X\subseteq A$ there exist $\alpha_{1,1},\dots,\alpha_{m,k}\in\operatorname{Aut}(\mathfrak{A})$ such that for every $i\in\\{1,\dots,m\\}$ the operation $f(\alpha_{i,1},\dots,\alpha_{i,k})$ has the same behaviour as $h_{i}$ on $X^{k}$, and for all $l\in\\{1,\dots,q\\}$ the operation $f(\alpha_{u^{l}_{1},1},\dots,\alpha_{u^{l}_{k},k})$ is canonical on $X^{k}$. For $q=p$ we obtain the statement of the lemma. If $q=0$ the statement of the claim follows from the assumptions of the lemma. For the inductive step, we assume that the statement holds for $q-1$ and prove the statement for $q$. Let $X\subseteq A$ be finite. Lemma 5.5 asserts the existence of a finite $C\subseteq A$ such that for every $f\colon A^{k}\to A$ there are $\beta_{1},\dots,\beta_{k}\in\operatorname{Aut}(\mathfrak{A})$ such that $\beta_{1}(X)\subseteq C,\dots,\beta_{k}(X)\subseteq C$ and $f(\beta_{1},\dots,\beta_{k})$ is canonical on $X^{k}$. We apply the induction hypothesis to $C$, and obtain $\gamma_{1,1},\dots,\gamma_{m,k}$ such that for every $i\in\\{1,\dots,m\\}$ the operation $f(\gamma_{i,1},\dots,\gamma_{i,k})$ has the same behaviour as $h_{i}$ on $C^{k}$ and $f(\gamma_{u^{l}_{1},1},\dots,\gamma_{u^{l}_{k},k})$ is canonical on $C^{k}$ for every $l\in\\{1,\dots,q-1\\}$. Let $f^{\prime}\coloneqq f(\gamma_{u^{q}_{1},1},\dots,\gamma_{u^{q}_{k},k})$. The property of $C$ implies that there are $\beta_{1},\dots,\beta_{k}\in\operatorname{Aut}(\mathfrak{A})$ such that $\beta_{1}(X)\subseteq C,\dots,\beta_{k}(X)\subseteq C$ and $f^{\prime}(\beta_{1},\dots,\beta_{k})$ is canonical on $X^{k}$. For $i\in\\{1,\dots,m\\}$ and $j\in\\{1,\dots,k\\}$ define $\alpha_{i,j}\coloneqq\gamma_{i,j}\circ\beta_{j}$. Observe $\alpha_{i,j}(X)\subseteq\gamma_{i,j}(C)$ and hence * • $f(\alpha_{u^{l}_{1},1},\dots,\alpha_{u^{l}_{k},k})$ is canonical on $X^{k}$ for $l\in\\{1,\dots,q-1\\}$: this follows from the inductive assumption that $f(\gamma_{u^{l}_{1},1},\dots,\gamma_{u^{l}_{k},k})$ is canonical on $C^{k}$; * • $f(\alpha_{u^{q}_{1},1},\dots,\alpha_{u^{q}_{k},k})$ is canonical on $X^{k}$: this follows from the property of $\beta_{1},\dots,\beta_{k}$ that $f^{\prime}(\beta_{1},\dots,\beta_{k})=f(\alpha_{u^{q}_{1},1},\dots,\alpha_{u^{q}_{k},k})$ is canonical on $X^{k}$; * • for every $i\in\\{1,\dots,m\\}$, the operation $f(\alpha_{i,1},\dots,\alpha_{i,k})$ has the same behaviour as $h_{i}$ on $X^{k}$ because for every $j\in\\{1,\dots,k\\}$ and $f(\gamma_{i,1},\dots,\gamma_{i,k})$ has the same behaviour as $h_{i}$ on $C^{k}$ by the inductive assumption. This concludes the proof that $\alpha_{1,1},\dots,\alpha_{m,k}$ satisfy the inductive statement. ∎ We introduce useful notation for the proofs of Theorem 6.3 and Theorem 6.4. If $f\colon B^{k}\to B$, $u\colon B\to B$, and $\ell\in\\{1,\dots,k\\}$, we write $f^{u}_{\ell}$ for the $k$-ary operation defined by $(x_{1},\dots,x_{k})\mapsto f(x_{1},\dots,x_{\ell-1},u(x_{\ell}),x_{\ell+1},\dots,x_{k}).$ ###### Proof of Theorem 6.3. The forward implication is trivial. For the converse implication, suppose that $\zeta$ does not have the UIP with respect to $\operatorname{Pol}(\mathfrak{C})$ over $\mathfrak{A}$. That is, there are operations $g,g_{1},g_{2}\in{\mathscr{C}}^{(k)}$ such that $g$ interpolates both $g_{1}$ and $g_{2}$ over $\operatorname{Aut}(\mathfrak{A})$ and $\zeta(g_{1})\neq\zeta(g_{2})$. By Lemma 7.4 there exists a finite $X_{0}\subseteq A$ which is $k$-rich with respect to $\operatorname{Pol}(\mathfrak{B})$. By Lemma 8.1 applied to $g$ and $X_{0}$ there exist $\alpha_{1,1},\dots,\alpha_{2,k}\in\operatorname{Aut}(\mathfrak{A})$ such that * • for all $u_{1},\dots,u_{k}\in\\{1,2\\}$ the operation $g(\alpha_{u_{1},1},\dots,\alpha_{u_{k},k})$ is canonical on $X_{0}^{k}$ over $\mathfrak{A}$, and * • for $i\in\\{1,2\\}$ the operation $g(\alpha_{i,1},\dots,\alpha_{i,k})$ has the same behaviour as $g_{i}$ on $X_{0}^{k}$. Let $i\in\\{1,\dots,k\\}$ be the smallest index such that $\displaystyle\zeta\big{(}g(\alpha_{1,1},\dots,\alpha_{1,i-1},\alpha_{1,i},\alpha_{2,{i+1}},\dots,\alpha_{2,k})\big{)}$ $\displaystyle\neq\;$ $\displaystyle\zeta\big{(}g(\alpha_{1,1},\dots,\alpha_{1,i-1},\alpha_{2,i},\alpha_{2,{i+1}},\dots,\alpha_{2,k})\big{)}.$ We know that such an index exists since $\zeta\big{(}g(\alpha_{1,1},\dots,\alpha_{1,k})\big{)}=\zeta(g_{1})\neq\zeta(g_{2})=\zeta\big{(}g(\alpha_{2,1},\dots,\alpha_{2,k})\big{)}.$ Let $h\coloneqq g(\alpha_{1,1},\dots,\alpha_{1,i-1},\operatorname{id},\alpha_{2,{i+1}},\dots,\alpha_{2,k})$. So we have shown the following. Claim. For every finite $X\subseteq A$ that contains $X_{0}$ there exist $i\in\\{1,\dots,k\\}$, $h\in\operatorname{Pol}(\mathfrak{B})^{(k)}$, $\alpha,\beta\in\operatorname{Aut}(\mathfrak{A})$, and $h,h^{\prime}\in\operatorname{Pol}(\mathfrak{B})^{(k)}\cap\operatorname{Can}(\mathfrak{A})$ such that * • $f^{\alpha}_{\ell}$ has the same behaviour as $h$ on $X^{k}$, * • $f^{\beta}_{\ell}$ has the same behaviour as $h^{\prime}$ on $X^{k}$, and * • $\zeta(h)\neq\zeta(h^{\prime})$. Let $X_{0}\subset X_{1}\subset\cdots$ be finite subsets of $A$ such that $\bigcup_{i=0}^{\infty}X_{i}=A$. Then the claim applied to $X=X_{i}$, for $i\in{\mathbb{N}}$, asserts the existence of $f_{i}$, $\alpha_{i}$, $\beta_{i}$, and $\ell_{i}$ such that $(f_{i})_{\ell_{i}}^{\alpha_{i}}$ and $(f_{i})_{\ell_{i}}^{\beta_{i}}$ are canonical on $X_{i}^{k}$ and $\zeta((f_{i})^{\alpha_{i}}_{\ell_{i}})\neq\zeta((f_{i})^{\beta_{i}}_{\ell_{i}})$. By thinning out the sequences $(f_{i})_{i\in\mathbb{N}}$, $(\alpha_{i})_{i\in{\mathbb{N}}}$, $(\beta_{i})_{i\in{\mathbb{N}}}$, and $(\ell_{i})_{i\in{\mathbb{N}}}$ we can assume that all $\ell_{i}$ are equal, say $\ell$, and that the complete behaviour ${\mathcal{B}}_{1}$ of $f_{i}^{\alpha_{i}}$ on $X_{0}$ and ${\mathcal{B}}_{2}$ of $f_{i}^{\beta_{i}}$ on $X_{0}$ does not depend on $i$. Note that we must have ${\mathcal{B}}_{1}\neq{\mathcal{B}}_{2}$. By Lemma 5.3 (and in particular the statement at the end starting with “moreover”) there exist $g\in\operatorname{Pol}(\mathfrak{B})$ and $u,v\in\overline{\operatorname{Aut}(\mathfrak{A})}$ such that $g^{u}_{\ell}$ has behaviour ${\mathcal{B}}_{1}$ on all of $A$, and $g^{v}_{\ell}$ has behaviour ${\mathcal{B}}_{2}$ on all of $A$. So they are canonical over $\mathfrak{A}$ and $\zeta(g^{u}_{\ell})\neq\zeta(g^{v}_{\ell})$. Suppose that $\zeta(g^{u}_{\ell})=\pi^{k}_{r}$ and $\zeta(g^{v}_{\ell})=\pi^{k}_{s}$. Let $f(x,y)\coloneqq f(\underbrace{y,\dots,y}_{r-1},x,y,\dots,y).$ Then note that (3) $\displaystyle\zeta\big{(}f(\pi^{2}_{1},u(\pi^{2}_{2}))\big{)}=\;$ $\displaystyle\zeta\big{(}g(\underbrace{u(\pi^{2}_{2}),\dots,u(\pi^{2}_{2})}_{r-1},\pi^{2}_{1},u(\pi^{2}_{2}),\dots,u(\pi^{2}_{2}))\big{)}$ (4) $\displaystyle=\;$ $\displaystyle\zeta\big{(}g^{u}_{\ell}(\underbrace{\pi^{2}_{2},\dots,\pi^{2}_{2}}_{r-1},\pi^{2}_{1},\pi^{2}_{2},\dots,\pi^{2}_{2})\big{)}$ (5) $\displaystyle=\;$ $\displaystyle\pi^{k}_{r}(\underbrace{\pi^{2}_{2},\dots,\pi^{2}_{2}}_{r-1},\pi^{2}_{1},\pi^{2}_{2},\dots,\pi^{2}_{2})$ (6) $\displaystyle=\;$ $\displaystyle\pi^{2}_{1}$ where equality of (3) and (4) holds since $\zeta$ is $\overline{\operatorname{Aut}(\mathfrak{A})}$-invariant, and equality of (4) and (6) holds since $\zeta$ is minor-preserving. Similarly, since $r\neq s$ $\displaystyle\zeta\big{(}f(\pi^{2}_{1},v(\pi^{2}_{2}))\big{)}$ $\displaystyle=\pi^{k}_{s}(\underbrace{\pi^{2}_{2},\dots,\pi^{2}_{2}}_{r-1},\pi^{2}_{1},\pi^{2}_{2},\dots,\pi^{2}_{2})=\pi^{2}_{2}$ Therefore, $f,u,v$ show that item $(2)$ from the statement does not hold, concluding the proof that $(2)\Rightarrow(1)$. ∎ We finally prove Theorem 6.4, which shows that it suffices to verify the UIP on independent elementary substructures. ###### Proof of Theorem 6.4. The forward implication holds trivially. We now show the converse, $(2)\Rightarrow(1)$. It suffices to verify $(2)$ in Theorem 6.3. Let $f\in\operatorname{Pol}(\mathfrak{C})^{(2)},u_{1},u_{2}\in\overline{\operatorname{Aut}(\mathfrak{A})}$ be such that $f^{u_{1}}_{2},f^{u_{2}}_{2}\in\operatorname{Can}(\mathfrak{A})$. Let $F\subseteq A$ be 2-rich with respect to $\operatorname{Pol}(\mathfrak{C})$. We have to verify that $f^{u_{1}}_{2}$ and $f^{u_{2}}_{2}$ have the same complete behaviour on $F^{2}$. Let $(\mathfrak{A}_{1},\mathfrak{A}_{2})$ be a pair of independent elementary substructures of $\mathfrak{A}$; such a pair exists by Theorem 3.2. For $j\in\\{1,2\\}$ let $e_{j}$ be an embedding of $\mathfrak{A}$ into $\mathfrak{A}_{j}$. Since $\mathfrak{A}$ is homogeneous there exists $\epsilon\in\operatorname{Aut}(\mathfrak{A})$ such that $e_{1}$ and $\epsilon$ agree on $u_{1}(F)\cup u_{2}(F)$. By Lemma 5.4 applied to $f(\operatorname{id},\epsilon^{-1}e_{2})$ there are $\beta_{1},\beta_{2}\in\operatorname{Aut}(\mathfrak{A})$ such that $f(\beta_{1},\epsilon^{-1}e_{2}\beta_{2})$ is canonical on $F^{2}$. Let $v_{i}\coloneqq\epsilon u_{i}$ and $w\coloneqq e_{2}\beta_{2}$. Note that $v_{i}(F)\subseteq A_{1}$ and $w(F)\subseteq A_{2}$. Let $h\coloneqq f(\beta_{1},\epsilon^{-1})\in\operatorname{Pol}(\mathfrak{C})$, and note that $h_{2}^{v_{i}}=f(\beta_{1},\epsilon^{-1}\epsilon u_{i})=f(\beta_{1},u_{i})$ is canonical since $f_{2}^{u_{i}}=f(\operatorname{id},u_{i})$ is canonical, and that $h_{2}^{w}=f(\beta_{1},\epsilon^{-1}e_{2}\beta_{2})$ is canonical on $F^{2}$ as we have seen above. We thus apply (2) two times to obtain that $\zeta(h_{2}^{v_{1}})=\zeta(h_{2}^{w})=\zeta(h_{2}^{v_{2}})$. Finally, note that $\zeta(f(\operatorname{id},u_{i}))=\zeta(f(\beta_{1},u_{i}))=\zeta(f(\beta_{1},\epsilon^{-1}\epsilon u_{i}))=\zeta(h_{2}^{v_{i}}).$ We conclude that $\zeta(f(\operatorname{id},u_{1}))=\zeta(f(\operatorname{id},u_{2}))$. ∎ ## 9\. First-order expansions of the basic relations of RCC5 RCC5 is a fundamental formalism for spatial reasoning, and may be viewed as an $\omega$-categorical structure $\mathfrak{R}$ (see, e.g., [BC09, BJ17]). There are many equivalent ways of formally introducing RCC5; we follow the presentation in [BC09] and refer to [BJ17] for other definitions and their equivalence. Let $\mathfrak{S}$ be the structure with domain $S\coloneqq 2^{\mathbb{N}}\setminus\\{\emptyset\\}$, i.e., the set of all non-empty subsets of the natural numbers ${\mathbb{N}}$. The signature of $\mathfrak{S}$ consists of the five binary relation symbols $\mathtt{EQ},\mathtt{PP},\mathtt{PPI},\mathtt{DR},\mathtt{PO}$ where for $x,y\subseteq\mathbb{N}$ we have (7) $\displaystyle(x,y)\in\mathtt{EQ}$ $\displaystyle\text{ iff }x=y,$ “$x$ and $y$ are equal” (8) $\displaystyle(x,y)\in\mathtt{PP}$ $\displaystyle\text{ iff }x\subset y,$ “$x$ is strictly contained in $y$” (9) $\displaystyle(x,y)\in\mathtt{PPI}$ $\displaystyle\text{ iff }x\supset y,$ “$x$ strictly contains $y$” (10) $\displaystyle(x,y)\in\mathtt{DR}$ $\displaystyle\text{ iff }x\cap y=\emptyset,$ “$x$ and $y$ are disjoint” (11) $\displaystyle(x,y)\in\mathtt{PO}$ $\displaystyle\text{ iff }x\not\subset y\wedge y\not\subset x\wedge x\cap y\neq\emptyset,$ “$x$ and $y$ properly overlap”. Note that by definition every pair $(x,y)\in S^{2}$ is contained in exactly one of the relations $\mathtt{DR}^{\mathfrak{S}},\mathtt{PO}^{\mathfrak{S}},\mathtt{PP}^{\mathfrak{S}},\mathtt{PPI}^{\mathfrak{S}},\mathtt{EQ}^{\mathfrak{S}}$. Note that the structure $\mathfrak{S}$ is not $\omega$-categorical; however, $\operatorname{Age}(\mathfrak{S})$ is an amalgamation class (Theorem 30 in [BC09]), and hence there exists a countable homogeneous structure $\mathfrak{R}$ with the same age as $\mathfrak{S}$. We refer to the relations $\mathtt{EQ}^{\mathfrak{R}},\mathtt{PP}^{\mathfrak{R}},\mathtt{PPI}^{\mathfrak{R}},\mathtt{DR}^{\mathfrak{R}},\mathtt{PO}^{\mathfrak{R}}$ as the _basic relations of RCC5_. The _composition_ of two binary relations $R_{1}$ and $R_{2}$ is the binary relation $R_{1}\circ R_{2}\coloneqq\big{\\{}(x,y):\exists z\big{(}R_{1}(x,z)\wedge R_{2}(z,y)\big{)}\big{\\}}$. The _converse_ (sometimes also called _inverse_) of a relation $R$ is the relation $\\{(y,x)\colon(x,y)\in R\\}$, and denoted by $R^{\smile}$. The converse of $\mathtt{PP}$ is $\mathtt{PPI}$, and $\mathtt{EQ}^{\mathfrak{R}}$, $\mathtt{DR}^{\mathfrak{R}}$, and $\mathtt{PO}^{\mathfrak{R}}$ are their own converse. The full binary relation containing all pairs of elements of $\mathfrak{R}$ is denoted by $\bf 1$. It is straightforward to verify that the relations $\mathtt{EQ}^{\mathfrak{R}},\mathtt{PP}^{\mathfrak{R}},\mathtt{PPI}^{\mathfrak{R}},\mathtt{DR}^{\mathfrak{R}},\mathtt{PO}^{\mathfrak{R}}$ compose as shown in Table 1. ###### Lemma 9.1. $\mathfrak{R}$ is finitely bounded. ###### Proof. A finite structure embeds into $\mathfrak{R}$ if and only if all three-element substructures of the structure embed into $\mathfrak{R}$. For details we refer to [BJ17], implication $2\Rightarrow 3$ in Proposition 15. ∎ ###### Remark 9.2. As a _relation algebra_ , RCC5 is given by the composition table and the data about the converses. Our structure $\mathfrak{R}$ introduced above is a _representation_ of RCC5. We do not introduce relation algebras formally, because they will not be needed in the following, and rather refer to [BJ17]. ∘ DR PO PP PPI EQ DR 1 PP∪DR∪PO PP∪DR∪PO DR DR PO PPI∪DR∪PO 1 PP∪PO PPI∪PO PO PP DR PP∪DR∪PO PP 1 PP PPI PPI∪DR∪PO PPI∪PO 1 ∖DR PPI PPI EQ DR PO PP PPI EQ Figure 1. The composition table for the relations of $\mathfrak{R}$. A _first-order expansion_ of a structure $\mathfrak{A}$ is an expansion of $\mathfrak{A}$ by relations that are first-order definable over $\mathfrak{A}$. ###### Theorem 9.3. For every first-order expansion $\mathfrak{C}$ of $\mathfrak{R}$ either * • $\operatorname{Pol}(\mathfrak{C})\cap\operatorname{Can}(\mathfrak{R})$ does not have a uniformly continuous minor-preserving map to $\operatorname{Proj}$, in which case $\operatorname{CSP}(\mathfrak{C})$ is in P, or * • $\operatorname{Pol}(\mathfrak{C})$ has a uniformly continuous minor-preserving map to $\operatorname{Proj}$, in which case $\operatorname{CSP}(\mathfrak{C})$ is NP-complete. We prove Theorem 9.3 using Corollary 6.2 and hence first need to introduce a Ramsey expansion $(\mathfrak{R};\prec)$ of $\mathfrak{R}$ (Section 9.1); we use a recent Ramsey transfer result of Mottet and Pinsker [MP21]. Every first- order expansion $\mathfrak{C}$ of $\mathfrak{R}$ is clearly a model-complete core. We verify the remaining assumption in Corollary 6.2 in two steps. Assuming that there exists a uniformly continuous minor-preserving map from $\operatorname{Pol}(\mathfrak{C})\cap\operatorname{Can}(\mathfrak{R})$ to $\operatorname{Proj}$, we construct Section 9.4 either * • a minor-preserving map $\eta\colon\operatorname{Pol}(\mathfrak{C})\cap\operatorname{Can}(\mathfrak{R},\prec)\to\operatorname{Proj}$ which arises from the action of the canonical polymorphisms on the two relations $\mathtt{PP}$ and $(\mathtt{DR}\cup\mathtt{PO})\,\cap\prec$, or * • a minor-preserving map $\rho\colon\operatorname{Pol}(\mathfrak{C})\cap\operatorname{Can}(\mathfrak{R},\prec)\to\operatorname{Proj}$ which arises from the action of the canonical polymorphisms on the two relations $\mathtt{DR}\,{\cap}\,{\prec}$ and $\mathtt{PO}\,{\cap}\,{\prec}$. In the second step, we prove that if such a map $\eta,\rho\colon\operatorname{Pol}(\mathfrak{C})\cap\operatorname{Can}(\mathfrak{R},\prec)\to\operatorname{Proj}$ exists, then it has the UIP with respect to $\operatorname{Pol}(\mathfrak{C})$ over $(\mathfrak{R},\prec)$ (Section 9.5); here we use Theorem 6.4. The statement then follows from Corollary 6.2. ### 9.1. A Ramsey order expansion of $\mathfrak{R}$ The structure $\mathfrak{R}$ is not Ramsey, but it has a homogeneous Ramsey expansion by a linear order. Let $\mathcal{C}$ be the class of all expansions of structures from $\operatorname{Age}(\mathfrak{S})$ with the signature $\\{\mathtt{EQ},\mathtt{DR},\mathtt{PO},\mathtt{PP},\mathtt{PPI},\prec\\}$ such that $\prec$ denotes a linear extension of $\mathtt{PP}$. ###### Proposition 9.4. The class $\mathcal{C}$ defined above is an amalgamation class. ###### Proof. It is clear from the definition that $\mathcal{C}$ is closed under isomorphisms and substructures. It is well-known that in order to prove the amalgamation property, it suffices to verify the 1-point amalgamation property (see, e.g., [Bod21]): for all structures $\mathfrak{A},\mathfrak{B}_{1},\mathfrak{B}_{2}\in{\mathcal{C}}$ such that $A=B_{1}\cap B_{2}$ and $B_{i}=A\cup\\{b_{i}\\}$, for $i\in\\{1,2\\}$ and $b_{1}\neq b_{2}$, there exists $\mathfrak{C}\in{\mathcal{C}}$ with $C=B_{1}\cup B_{2}=A\cup\\{b_{1},b_{2}\\}$ such that $\mathfrak{B}_{1}$ and $\mathfrak{B}_{2}$ are substructures of $\mathfrak{C}$. It is easy to see that such a structure $\mathfrak{C}$ can be determined by specifying which relations of $\mathfrak{C}$ contain the pair $(b_{1},b_{2})$. Bodirsky and Chen [BC09] (also see [Bod21]) proved that there exists $R\in\\{\mathtt{DR},\mathtt{PO},\mathtt{PP},\mathtt{PPI}\\}$ such that * • if we add the pair $(b_{1},b_{2})$ to $R^{\mathfrak{C}}$ and the pair $(b_{2},b_{1})$ to $(R^{\mathfrak{C}})^{\smile}$, then the $\\{\mathtt{EQ},\mathtt{DR},\mathtt{PO},\mathtt{PP},\mathtt{PPI}\\}$-reduct of the resulting structure $\mathfrak{C}$ is in $\operatorname{Age}(\mathfrak{S})$, and * • $(b_{1},b_{2})\in\mathtt{PP}^{\mathfrak{C}}$ only if there exists $a\in A$ such that $(b_{1},a)\in\mathtt{PP}^{\mathfrak{B}_{1}}$ and $(a,b_{2})\in\mathtt{PP}^{\mathfrak{B}_{2}}$. If $R$ equals $\mathtt{PP}$ we add $(b_{1},b_{2})$ to $\prec^{\mathfrak{C}}$; if $R$ equals $\mathtt{PPI}$ we add $(b_{2},b_{1})$ to $\prec^{\mathfrak{C}}$. If $R$ is from $\\{\mathtt{DR},\mathtt{PO}\\}$, then we add $(b_{1},b_{2})$ or $(b_{2},b_{1})$ to $\prec^{\mathfrak{C}}$ according to an order-amalgam of the $\\{\prec\\}$-reducts of $\mathfrak{B}_{1}$ and $\mathfrak{B}_{2}$ over $\mathfrak{A}$. In each of these cases, $\prec^{\mathfrak{C}}$ is a linear order that extends $\mathtt{PP}^{\mathfrak{C}}$, and hence $\mathfrak{C}$ is in ${\mathcal{C}}$. ∎ One can check by an easy back-and-forth argument that the $\\{\mathtt{EQ},\mathtt{DR},\mathtt{PO},\mathtt{PP},\mathtt{PPI}\\}$-reduct of the Fraïssé-limit of ${\mathcal{C}}$ is isomorphic to $\mathfrak{R}$; hence, we may denote this Fraïssé-limit by $(\mathfrak{R},\prec)$. To show that $(\mathfrak{R},\prec)$ has the Ramsey property, we use a recent Ramsey transfer result. ###### Definition 10 (Mottet and Pinsker [MP21]). Let $\mathscr{G}$ be a permutation group on a set $X$. A function $g\colon X\to X$ is called _range-rigid with respect to $\mathscr{G}$_ if for every $k\in{\mathbb{N}}$, every orbit of $k$-tuples in the range of $g$ is preserved by $g$. Mottet and Pinsker [MP21] proved that if $\mathfrak{B}$ is a homogeneous structure and $g\colon B\to B$ is range-rigid with respect to $\operatorname{Aut}(\mathfrak{B})$, then the age of the structure induced by $\mathfrak{B}$ on the image of $g$ has the amalgamation property; we denote the Fraïssé-limit of this class by $\mathfrak{B}_{g}$. ###### Theorem 9.5 (Lemma 10 in [MP21]). Let $\mathfrak{B}$ be a homogeneous structure and let $g\colon B\to B$ be range-rigid with respect to $\operatorname{Aut}(\mathfrak{B})$. If $\mathfrak{B}$ is Ramsey, then so is $\mathfrak{B}_{g}$. Let $\mathfrak{A}=(A;\cap,\cup,\overline{\cdot},0,1)$ be the countable atomless Boolean algebra [Hod97], which we view as a subalgebra of $(2^{\mathbb{N}};\cap,\cup,\overline{\cdot},0,1)$. We use the usual shortcuts: * • $x\subseteq y$ stands for $x\cap y=x$, * • $x\subset y$ stands for $x\subseteq y$ and $x\neq y$, and * • $x+y$ stands for $(x\cap\overline{y})\cup(y\cap\overline{x})$. Let $\mathfrak{A}^{\prime}$ be the ($\omega$-categorical) relational structure with the same domain as $\mathfrak{A}$ carrying all relations that are first- order definable in $\mathfrak{A}$. Kechris, Pestov, and Todorčević described a linear order $\prec$ on $A$ so that the expanded structure $(\mathfrak{A}^{\prime},\prec)$ is a homogeneous $\omega$-categorical Ramsey structure [KPT05]. The age of $(\mathfrak{A}^{\prime},\prec)$ can be described as follows. If $\mathfrak{F}$ is a finite substructure of $\mathfrak{A}$, then $\mathfrak{F}$ is a finite Boolean algebra and there exists an enumeration $a_{1},\dots,a_{n}$ of the atoms of $\mathfrak{F}$ such that for all $u,v\in F$ with $u=\bigcup_{i=1}^{n}(d_{i}\cap a_{i})$ and $v=\bigcup_{i=1}^{n}(e_{i}\cap a_{i})$ for $d,e\in\\{0,1\\}^{n}$ we have $u\prec v$ if and only if there exists some $j\in\\{1,\dots,n\\}$ such that $d_{j}<e_{j}$ and $d_{i}=e_{i}$ for all $i>j$; such an ordering is called an _antilexicographical ordering_. Let $\mathfrak{B}$ be the relational structure with domain $B\coloneqq A\setminus\\{0\\}$ and for all $k,l\geq 1$ the relation $\displaystyle R_{k,l}\coloneqq\big{\\{}(a_{1},\dots,a_{k},b_{1},\dots,b_{l})\mid\bigcap_{i=1}^{k}a_{i}\subseteq\bigcup_{j=1}^{l}b_{j}\big{\\}}$ and for all $k\geq 1$ and $e,d\in\\{0,1\\}^{k}$ the relation $\displaystyle O_{k,d,e}\coloneqq\big{\\{}(a_{1},\dots,a_{k})\mid\bigcap_{i=1}^{k}(a_{i}+d_{i})\prec\bigcap_{i=1}^{k}(a_{i}+e_{i})\big{\\}}.$ Note that $x\prec y$ has the definition $O_{2,(0,1),(1,0)}(x,y)$ in $\mathfrak{B}$, which holds precisely if $x\cap\overline{y}\prec{\overline{x}}\cap y$. ###### Lemma 9.6. The structure $\mathfrak{B}$ is homogeneous. ###### Proof. Let $i$ be an isomorphism between finite induced substructures of $\mathfrak{B}$. We claim that $i$ preserves all quantifier-free formulas of $\mathfrak{A}$. To see this, it suffices to show that for every Boolean term $f$ of arity $n$ and every Boolean term $g$ of arity $k$, the function $i$ preserves the formula $f(x_{1},\dots,x_{n})=g(y_{1},\dots,y_{k})$. Since $f(x_{1},\dots,x_{n})=g(y_{1},\dots,y_{k})$ if and only if $f(x_{1},\dots,x_{n})+g(y_{1},\dots,y_{k})=0$, it suffices to show the statement for the special case $f(x_{1},\dots,x_{n})=0$. Every Boolean term can be rewritten as a finite union of finite intersections of arguments and complements of arguments. Therefore, $f(x_{1},\dots,x_{n})=0$ is equivalent to a conjunction of formulas of the form $x_{i_{1}}\cap\cdots\cap x_{i_{r}}\subseteq x_{j_{1}}\cup\cdots\cup x_{j_{s}}$. These formulas are preserved by $i$, because $i$ preserves $R_{r,s}$. We claim that $i$ also preserves all quantifier-free formulas of $(\mathfrak{A},\prec)$. The definition of $\prec$ implies that it suffices to show that $i$ preserves $\prec$ for the atoms of the Boolean algebra generated by the domain $\\{a_{1},\dots,a_{m}\\}$ of $i$. Let $b$ and $c$ be two such atoms with $b\prec c$. Then $b=\bigcap_{i=1}^{m}(a_{i}+d_{i})$ and $c=\bigcap_{i=1}^{m}(a_{i}+e_{i})$ for some $d,e\in\\{0,1\\}^{m}$, and hence $i(b)\prec i(c)$, because $i$ preserves the relation $O_{m,d,e}$. By the homogeneity of $(\mathfrak{A},\prec)$, the map $i$ has an extension to an automorphism of $(\mathfrak{A},\prec)$. The restriction of this map to $B$ is an automorphism of $\mathfrak{B}$, concluding the proof. ∎ ###### Lemma 9.7. The structure $\mathfrak{B}$ is Ramsey. ###### Proof. The structure $(\mathfrak{A}^{\prime},\prec)$ and its homogeneous substructure $\mathfrak{A}^{\prime\prime}$ with domain $B$ have topologically isomorphic automorphism groups; since the Ramsey property only depends on the topological automorphism group (see, e.g., [KPT05]), it follows that $\mathfrak{A}^{\prime\prime}$ is Ramsey. It is easy to see that $\mathfrak{A}^{\prime\prime}$ and $\mathfrak{B}$ have the same automorphism group, even as permutation groups, which implies the statement. ∎ Consider the $\\{\mathtt{EQ},\mathtt{DR},\mathtt{PO},\mathtt{PP},\mathtt{PPI}\\}$-structure with domain $A$ whose relations are defined over $\mathfrak{A}$ by the expressions in (7), (8), (9), (10), (11); let $\mathfrak{D}$ be the substructure of this structure with domain $B$. ###### Lemma 9.8. There exists a self-embedding $e$ of $(\mathfrak{D},\prec)$ such that 1. (1) for all $k,l\in{\mathbb{N}}$ and $a_{1},\dots,a_{k},b_{1},\dots,b_{l}\in B$ we have that $(e(a_{1}),\dots,e(a_{k}),e(b_{1}),\dots,e(b_{l}))\in R_{k,l}$ if and only if $(a_{i},b_{j})\in\mathtt{PP}\cup\mathtt{EQ}$ or $(a_{i},a_{j})\in\mathtt{DR}$ for some $i,j\leq\max(k,l)$, and 2. (2) for every $k\in{\mathbb{N}}$ and $d,e\in\\{0,1\\}^{k}$ there exists a quantifier-free formula $\phi_{k,d,e}$ in the signature of $(\mathfrak{D},\prec)$ that defines $O_{k,d,e}$. The function $e$ is range-rigid with respect to $\operatorname{Aut}(\mathfrak{B})$. ###### Proof. First note that any function $e$ that satisfies the first statement of the lemma must be range-rigid. By the homogeneity of $\mathfrak{B}$, orbits of $k$-tuples in $\mathfrak{B}$ can be described by conjunctions of relations of the form $R_{k,l}$ and of the form $O_{k,d,e}$. Therefore, every orbit of $k$-tuples in the range of $e$ has by items (1) and (2) of the statement a quantifier-free definition over the relations of $(\mathfrak{D},\prec)$. Since $e$ preserves these relations, $e$ preserves all orbits of $k$-tuples in the range of $e$. To prove the first part of the statement, by a standard compactness argument it suffices to prove that for every finite substructure $\mathfrak{F}$ of $\mathfrak{B}$ there exists a finite substructure $(\mathfrak{G},\prec)$ of $(\mathfrak{A},\prec)$ and a map $f\colon F\to G\setminus\\{0\\}$ that preserves the relations of $(\mathfrak{D},\prec)$ and satisfies the property of $e$ formulated in items (1) and (2) in the statement. Let $\displaystyle I\coloneqq\big{\\{}$ $\displaystyle\\{a_{1},\dots,a_{k}\\}\mid\;k\in{\mathbb{N}},a_{1},\dots,a_{k}\in F,(a_{i},a_{j})\in\mathtt{PO}^{\mathfrak{F}}\text{ for all }1\leq i<j\leq k\\}\big{\\}}$ and let $f\colon F\to 2^{I}$ be the map given by $f(a)\coloneqq\\{X\in I\mid\exists a^{\prime}\in X.\,a^{\prime}\subseteq a\\}.$ Define the relation $\sqsubset$ on $I$ so that $X\sqsubset Y$ holds for $X,Y\in I$ if and only if the smallest element of $X+Y=(X\cup Y)\setminus(X\cap Y)$ with respect to $\prec$ is contained in $X$. Claim 1. $\sqsubset$ defines a linear order on $I$. It is clear from the definition of $\sqsubset$ that for all $X,Y\in I$ exactly one of $X=Y$, $X\sqsubset Y$, and $Y\sqsubset X$ holds. To show that $\sqsubset$ is transitive it suffices to show that $\sqsubset$ does not contain a directed 3-cycle. Let $X,Y,Z\in I$ be pairwise distinct. Let $a$ be the smallest element in $(X\cup Y\cup Z)\setminus(X\cap Y\cap Z)$ and suppose that $a\in X\setminus Y$, so that $X\sqsubset Y$. * • If $a\in Z$ then $a$ is the the smallest element in $Z+Y$, and thus $Z\sqsubset Y$. * • If $a\notin Z$ then $a$ is the smallest element in $X+Z$, and thus $X\sqsubset Z$. In either case the restriction of $\sqsubset$ to $X,Y,Z$ does not form a directed 3-cycle. By symmetry, the assumption that $x\in X\setminus Y$ can be made without loss of generality; this proves the claim. Let $\mathfrak{G}$ be the Boolean algebra on $2^{I}$, and let $\prec$ denote the linear order on $G$ such that $S\prec T$ if the largest element of the symmetric difference $S+T$ with respect to the linear order $\sqsubset$ is contained in $T$. Note that $\prec$ is an antilexicographic linear order on $G$ and hence $(\mathfrak{G};\prec)$ embeds into $(\mathfrak{A};\prec)$. Also note that the image of $f$ lies in $G\setminus\\{0\\}$. Claim 2. Every $X\in I$ equals the largest element of $I$ with respect to $\sqsubset$ which is contained in $\bigcap_{a\in X}f(a)$. By construction, $X\in f(a)$ for all $a\in X$, that is, $X\in\bigcap_{a\in X}f(a)$. Assume that there exists $Y\in\bigcap_{a\in X}f(a)$. We have to show that $Y=X$ or $Y\sqsubset X$. If $X\subseteq Y$ then we are done. Otherwise, let $a$ be the smallest element of $X\setminus Y$ with respect to $\prec$. Since $Y\in f(a)$ there exists some $a^{\prime}\in Y$ such that $a^{\prime}\subseteq a$ by the definition of $f(a)$. Since $a\notin Y$, it follows that $a^{\prime}\neq a$. So $a^{\prime}\subset a$ and thus $a$ and $a^{\prime}$ cannot both be in $X$. Since $a\in X$ we conclude that $a^{\prime}\notin X$. Hence, $a^{\prime}\in Y\setminus X$. Since $\prec$ extends $\subset$ it follows that $a^{\prime}\prec a$. Since $a$ was chosen to be minimal in $X\setminus Y$ with respect to $\prec$ this implies that $Y\sqsubset X$. To prove that $f$ preserves the relations of $\mathfrak{D}$, it suffices to show that $f$ preserves $\mathtt{PP}$, $\mathtt{DR}$, and $\mathtt{PO}$. * • Suppose that $a,b\in F$ satisfy $\mathtt{PP}(a,b)$ in $\mathfrak{F}$. Let us assume that $X\in f(a)$ for some $X\in I$. By definition we know that $a^{\prime}\in X$ for some $a^{\prime}\subseteq a$. Then we also have $a^{\prime}\subseteq b$, and thus by definition $X\in f(b)$. This concludes the proof that $f(a)\subseteq f(b)$. On the other hand $\\{b\\}\in f(b)\setminus f(a)$, thus in fact $\mathtt{PP}(f(a),f(b))$ holds in $\mathfrak{G}$. * • Suppose that $a,b\in F$ satisfy $\mathtt{DR}(a,b)$ in $\mathfrak{F}$. If $X\in f(a)$ then by definition there exists $a^{\prime}\in X$ such that $a^{\prime}\subseteq a$. Since $X\in I$ and $a^{\prime}\cap b=0$ this implies that $b^{\prime}\notin X$ for every $b^{\prime}\subseteq b$. Therefore, $X\notin f(b)$. This implies that $\mathtt{DR}(f(a)\cap f(b))$ holds in $\mathfrak{G}$. * • Suppose that $a,b\in F$ satisfy $\mathtt{PO}(a,b)$ in $\mathfrak{F}$. Then $\\{a\\}\in f(a)\setminus f(b)$, $\\{b\\}\in f(b)\setminus f(a)$, and $\\{a,b\\}\in\gamma(a)\cap\gamma(b)$. Therefore, $\mathtt{PO}(f(a),f(b))$ holds in $\mathfrak{G}$. We now show that $f$ preserves $\prec$. Let $a,b\in F$ be such that $a\prec b$. Then by definition $\\{a\\}\sqsubset\\{b\\}$. Claim 2 implies that the largest elements contained in $f(a)$ and $f(b)$ are $\\{a\\}$ and $\\{b\\}$, respectively. It follows that $f(a)\prec f(b)$. To prove that $f$ satisfies items (1) and (2) of the statement, the following notation will be convenient. For $X\subseteq F$, we write $M(X)$ for the minimal elements in $S$ with respect to $\subset$, i.e., $M(X)\coloneqq\\{a\in X\mid\neg\exists b\in X.\,b\subset a\\}.$ Note that for every $a\in X$ there exists an $a^{\prime}\in M(X)$ such that $a^{\prime}\subseteq a$, because $X\subseteq F$ is finite. Claim 3. Let $a_{1},\dots,a_{k},b_{1},\dots,b_{l}\in F$. Then the following are equivalent. 1. (1) $(f(a_{1}),\dots,f(a_{k}),f(b_{1}),\dots,f(b_{l}))\notin R_{k,l}$; 2. (2) $c\coloneqq\bigcap_{i=1}^{k}f(a_{i})\cap\bigcap_{i=1}^{l}\overline{f(b_{i})}\neq\emptyset$; 3. (3) $a_{i}\cap a_{j}\neq 0$ and $a_{i}\not\subseteq b_{j}$ for all $1\leq i<j\leq\max(k,l)$. 4. (4) The $\sqsubset$-maximal element of $c=\bigcap_{i=1}^{k}f(a_{i})\cap\bigcap_{i=1}^{k}\overline{f(b_{i})}$ is $M(\\{a_{1},\dots,a_{k}\\})$. (1) and (2) are equivalent by definition. (2) implies (3): if there are $i,j\in\\{1,\dots,k\\}$ such that $a_{i}\cap a_{j}=\emptyset$, then $f(a_{i})\cap f(a_{j})=\emptyset$ since $f$ preserves $\mathtt{DR}$, and hence $c=\emptyset$. If there are $i\leq k$ and $j\leq l$ such that $a_{i}\subseteq b_{j}$ then $a\cap\overline{b_{j}}=\emptyset$ and hence $c=\emptyset$. (3) implies (4): Let $X\coloneqq M(\\{a_{1},\dots,a_{k}\\})$. Then by assumption $X\in I$, and by Claim 2 we know that $X$ is the maximal element of $\bigcap_{i=1}^{k}f(a_{i})$ with respect to $\sqsubset$. Suppose for contradiction that $X\in f(b_{j})$ for some $j\in\\{1,\dots,l\\}$. The definition of $f$ implies that there exists $b_{j}^{\prime}\subseteq b_{j}$ such that $b_{j}^{\prime}\in X$. In other words, $b_{j}^{\prime}=a_{i}$ for some $i\in\\{1,\dots,k\\}$, and therefore we have $a_{i}\subseteq b_{j}$, a contradiction. (4) implies (2): trivial. Note that the equivalence of (1) and (3) in Claim 3 immediately implies that $f$ satisfies item (1) of the statement. To prove that $f$ satisfies item (2), let $a_{1},\dots,a_{k}\in F$ and $d,e\in\\{0,1\\}^{k}$. Let $X\coloneqq M(\\{a_{i}\mid d_{i}=0\\})$ and $Y\coloneqq M(\\{a_{i}\mid e_{i}=0\\})$. We will prove that $(f(a_{1}),\dots,f(a_{k}))\in O_{k,d,e}$ if and only if $\bigcap_{i=1}^{k}(f(a_{i})+e_{i})\neq 0$ and either 1. (1) $\bigcap_{i=1}^{k}(f(a_{i})+d_{i})=0$, or 2. (2) $\bigcap_{i=1}^{k}(f(a_{i})+d_{i})\neq 0$ and $X\sqsubset Y$. If $\bigcap_{i=1}^{k}(f(a_{i})+e_{i})=0$ then clearly (12) $\displaystyle\bigcap_{i=1}^{k}(f(a_{i})+d_{i})\prec\bigcap_{i=1}^{k}(f(a_{i})+e_{i})$ is false and hence $(f(a_{1}),\dots,f(a_{k}))\notin O_{k,d,e}$ and we are done. So suppose that $\bigcap_{i=1}^{k}(f(a_{i})+e_{i})\neq 0$. If $\bigcap_{i=1}^{k}(f(a_{i})+d_{i})=0$ then $(\ref{eq:order})$ holds. In this case $(f(a_{1}),\dots,f(a_{k}))\in O_{k,d,e}$ and we are again done. So suppose that $\bigcap_{i=1}^{k}(f(a_{i})+d_{i})\neq 0$. Then by Claim 3 we know that $X,Y\in I$ and $X$ and $Y$ are the $\sqsubset$-maximal elements in $\bigcap_{i=1}^{k}(f(a_{i})+d_{i})$ and $\bigcap_{i=1}^{k}(f(a_{i})+e_{i})$, respectively. Thus, if $X\sqsubset Y$ then $\bigcap_{i=1}^{k}(f(a_{i})+d_{i})\prec\bigcap_{i=1}^{k}(f(a_{i})+e_{i})$ and so $(f(a_{1}),\dots,f(a_{k}))\in O_{k,d,e}$ and we are done. If $Y\sqsubset X$ then $\bigcap_{i=1}^{k}(f(a_{i})+e_{i})\prec\bigcap_{i=1}^{k}(f(a_{i})+d_{i})$ and hence $(f(a_{1}),\dots,f(a_{k}))\notin O_{k,d,e}$ and we are also done. If $X=Y$, let $i\in\\{1,\dots,k\\}$. If $d_{i}=0$ then by the definition of $X$ there exists an $a_{j}\in X$ such that $a_{j}\subseteq a_{i}$. Then $a_{j}\in Y$ and thus $e_{j}=0$. Since $\bigcap_{l=1}^{k}(f(a_{l})+e_{l})\neq 0$ is follows that in particular $f(a_{j})\cap(f(a_{i})+e_{i})=(\gamma(a_{j})+e_{j})\cap(f(a_{i})+e_{i})\neq\emptyset$. Since $f$ preserves $\mathtt{DR}$ this implies that $a_{j}\cap(a_{i}\cap e_{i})\neq 0$. Since $a_{j}\subseteq a_{i}$ this is only possible if $e_{i}=0$. Similarly one can show that $e_{i}=0$ implies $d_{i}=0$. Therefore, $d=e$, and thus $O_{k,d,e}=\emptyset$. In particular, $(f(a_{1}),\dots,f(a_{k}))\notin O_{k,d,e}$. Claim 3 implies that $\bigcap_{i=1}^{k}(f(a_{j})+e_{i})\neq 0$ and $\bigcap_{i=1}^{k}(f(a_{i})+d_{i})=0$ can be defined by quantifier-free $\\{\mathtt{PP},\mathtt{EQ},\mathtt{DR}\\}$-formulas. Note that condition (2) above can be expressed (assuming for simplicity of notation that $X=\\{a_{i}\mid d_{i}=0\\}$ and $Y=\\{a_{i}\mid e_{i}=0\\}$; the general case is similar) by the quantifier-free $\\{\prec\\}$-formula $\bigvee_{i:d_{i}=0,e_{i}=1}\bigwedge_{j:d_{j}\neq e_{j}}a_{i}\prec a_{j}$ which concludes the proof. ∎ ###### Proposition 9.9. $\operatorname{Age}(\mathfrak{D},\prec)=\operatorname{Age}(\mathfrak{R},\prec)$. ###### Proof. Note that $\mathfrak{D}$ has the same age as $\mathfrak{S}$ and hence the same age as $\mathfrak{R}$. Also, $\prec$ is a linear extension of $\mathtt{PP}^{\mathfrak{D}}$, and hence every finite substructure of $(\mathfrak{D},\prec)$ is also a substructure of $(\mathfrak{R},\prec)$. Conversely, let $(\mathfrak{F},\prec)$ be a finite substructure of $(\mathfrak{R},\prec)$. Let $g$ be an embedding of $\mathfrak{F}$ into $\mathfrak{D}$. Let $u_{1},\dots,u_{k}$ be an enumeration of $F$ such that $u_{1}\prec\cdots\prec u_{k}$. Let $v_{1},\dots,v_{k}\in B$ be such that $(v_{i},g(u_{j}))\in\mathtt{DR}^{\mathfrak{D}}$ for all $i,j\in\\{1,\dots,k\\}$ and such that $(v_{i},v_{j})\in\mathtt{DR}^{\mathfrak{D}}$ for $i\neq j$. Then we define $f\colon F\to B$ by $b(u)\coloneqq g(u)\cup\bigcup_{(u_{i},u)\in\mathtt{PP}^{\mathfrak{F}}}v_{i}.$ Note that $v_{1},\dots,v_{k}$ are atoms in the Boolean algebra generated by the elements $g(F)\cup\\{v_{1},\dots,v_{n}\\}$ in $\mathfrak{A}$, and let $w_{1},\dots,w_{\ell}$ be the other atoms. We may assume that $\prec$ is defined on this Boolean algebra according the enumeration $w_{1},\dots,w_{\ell},v_{1},\dots,v_{k}$ of the atoms. We prove that $f$ is an embedding of $(\mathfrak{F},\prec)$ into $(\mathfrak{D},\prec)$. Let $u,u^{\prime}\in F$. If $(u,u^{\prime})\in\mathtt{PP}^{\mathfrak{F}}$, then $g(u)\subset g(v)$ and $\\{i\colon(u_{i},u)\in\mathtt{PP}^{\mathfrak{F}}\\}\subset\\{i\colon(u_{i},u^{\prime})\in\mathtt{PP}^{\mathfrak{F}}\\}$ by the transitivity of $\mathtt{PP}^{\mathfrak{F}}$, so $(b(u),b(u^{\prime}))\in\mathtt{PP}^{\mathfrak{D}}$. It is also clear that $b$ preserves $\mathtt{EQ}$, $\mathtt{PPI}$, $\mathtt{DR}$. To see that $b$ preserves $\mathtt{PO}$, note that if $(u,u^{\prime})\in\mathtt{PO}^{\mathfrak{F}}$, then $(g(u),g(u^{\prime}))\in\mathtt{PO}^{\mathfrak{D}}$, so $g(u)\cap g(u^{\prime})$, $g(u)\cap\overline{g(u^{\prime})}$, and $\overline{g(u)}\cap g(u^{\prime})$ are non-empty. Note that $g(u)\cap g(u^{\prime})\subseteq b(u)\cap b(u^{\prime})$, $g(u)\cap\overline{g(u^{\prime})}\subseteq b(u)\cap\overline{b(u^{\prime})}$, and $\overline{g(u)}\cap g(u^{\prime})\subseteq\overline{b(u)}\cap b(u^{\prime})$, so $(b(u),b(u^{\prime}))\in\mathtt{PO}^{\mathfrak{D}}$. To prove that $b$ preserves $\prec$, let $i,j\in\\{1,\dots,k\\}$ be such that $i<j$ and $u_{i}\prec u_{j}$. Then $\mathtt{PP}^{\mathfrak{D}}(v_{j},b(u_{j}))$, $\mathtt{DR}^{\mathfrak{D}}(v_{j},b(u_{i}))$, and for every $m\in\\{j+1,\dots,k\\}$ we have $\mathtt{DR}^{\mathfrak{D}}(v_{m},b(u_{i}))$ and $\mathtt{DR}^{\mathfrak{D}}(v_{m},b(u_{j}))$. Indeed, if $\mathtt{DR}^{\mathfrak{D}}(v_{m},b(u_{i}))$ does not hold, then $\mathtt{PP}^{\mathfrak{D}}(v_{m},b(u_{i}))$ and hence $\mathtt{PP}^{\mathfrak{F}}(u_{m},u_{i})$ by the definition of $b$. This in turn implies that $u_{m}\prec u_{i}$ and hence $m<i$, a contradiction. By the definition of $\prec$ on $A$ this implies that $b(u_{i})\prec b(u_{j})$ and finishes the proof of the claim. ∎ Every relation of $(\mathfrak{D},\prec)$ has a quantifier-free definition in the homogeneous structure $\mathfrak{B}$; let $(\mathfrak{C},\prec)$ be the $\\{\mathtt{EQ},\mathtt{DR},\mathtt{PO},\mathtt{PP},\mathtt{PPI},\prec\\}$-structure defined over $\mathfrak{B}_{e}$ via these quantifier-free definitions. ###### Proposition 9.10. $(\mathfrak{C},\prec)$ is homogeneous. ###### Proof. It suffices to show that $(\mathfrak{C},\prec)$ and $\mathfrak{B}_{e}$ are quantifier-free interdefinable. By definition, every relation of $(\mathfrak{C},\prec)$ has a quantifier-free definition in $\mathfrak{B}_{e}$. Conversely, let $R$ be a relation of $\mathfrak{B}_{e}$. If $R$ is of the form $R_{k,l}$ then item (1) of Lemma 9.8 implies that $R$ has a quantifier-free definition over $(\mathfrak{C},\prec)$. If $R$ is of the form $O_{k,d,e}$ then item (2) of Lemma 9.8 implies that $R$ has a quantifier-free definition over $(\mathfrak{C},\prec)$. ∎ ###### Corollary 9.11. $(\mathfrak{C},\prec)$ and $(\mathfrak{R},\prec)$ are isomorphic. ###### Proof. It suffices to prove that the homogeneous structures $(\mathfrak{C},\prec)$ and $(\mathfrak{R},\prec)$ have the same age. By Proposition 9.9, the age of $(\mathfrak{R},\prec)$ equals the age of $(\mathfrak{D},\prec)$. ∎ ###### Theorem 9.12. $(\mathfrak{R},\prec)$ is Ramsey. ###### Proof. We apply the Ramsey transfer (Theorem 9.5) to the homogeneous Ramsey structure $\mathfrak{B}$ from Lemma 9.6 and the map $e$ from Lemma 9.8 which is range- rigid with respect to $\operatorname{Aut}(\mathfrak{B})$. Theorem 9.5 states that the homogeneous structure $\mathfrak{B}_{e}$ is Ramsey. The structure $\mathfrak{B}_{e}$ has the same automorphism group as the homogeneous structure $(\mathfrak{R};\prec)$; it follows that $(\mathfrak{R};\prec)$ is Ramsey as well. ∎ ### 9.2. Polymorphisms of $\mathfrak{R}$ that are canonical with respect to $(\mathfrak{R},\prec)$ From now on, we identify the symbols $\mathtt{EQ},\mathtt{PP},\mathtt{PPI},\mathtt{DR},\mathtt{PO}$ with the respective relations of $\mathfrak{R}$, and we write $\succ$ for the converse of $\prec$. Note that $\mathtt{PP}\cup\mathtt{DR}\cup\mathtt{PO}$ and $\mathtt{PPI}\cup\mathtt{DR}\cup\mathtt{PO}$ are primitively positively definable in $\mathfrak{R}$, since they are entries in Table 1. Hence, their intersection $\mathtt{DR}\cup\mathtt{PO}$ is primitively positively definable in $\mathfrak{R}$, too. We also write $\bot$ instead of $\mathtt{DR}\cup\mathtt{PO}$, and $\smash{{}^{\bot}_{\prec}}$ for the relation $\bot\,\cap\prec$. The composition table for the binary relations with a first-order definition over $(\mathfrak{R};\prec)$ can be derived from the composition table of $\mathfrak{R}$ (Table 1) using the following lemma. ###### Lemma 9.13. Let $R_{1},R_{2}\in\\{\mathtt{EQ},\mathtt{PP},\mathtt{PPI},\mathtt{DR},\mathtt{PO}\\}$ and let $O_{1},O_{2}$ be two orbits of pairs of $(\mathfrak{R};\prec)$ such that $O_{i}\subseteq R_{i}$. Then $O_{1}\circ O_{2}=\begin{cases}(R_{1}\circ R_{2})\,\cap\prec&\text{ if }O_{1},O_{2}\subseteq\;\prec\\\ (R_{1}\circ R_{2})\,\cap\succ&\text{ if }O_{1},O_{2}\subseteq\;\succ\\\ R_{1}\circ R_{2}&\text{ otherwise. }\end{cases}$ ###### Proof. It is clear that $O_{1}\circ O_{2}\subseteq R_{1}\circ R_{2}$ and that if $O_{1},O_{2}\subseteq\;\prec$ then $O_{1}\circ O_{2}\subseteq\,\prec\circ\prec\,=\,\prec.$ So the $\subseteq$-containment in the statement of the lemma holds in the first case, and by similar reasoning also in the other two cases. The reverse containment in the first two cases is also clear since the homogeneity of $(\mathfrak{R},\prec)$ implies that the expression in the statement on the right describes an orbit of pairs in $\operatorname{Aut}(\mathfrak{R},\prec)$ (we have already seen that it is non-empty). To show the equality in the third case, let $(x,z)\in R_{1}\circ R_{2}$. If $R_{1}\circ R_{2}$ equals $\mathtt{PP}$ (or $\mathtt{PPI}$) then $x\prec z$ (or $z\succ x$), and again $R_{1}\circ R_{2}$ is an orbit of pairs in $\operatorname{Aut}(\mathfrak{R},\prec)$ and the equality holds. If $R_{1}\circ R_{2}$ equals $\mathtt{EQ}$ then the statement is clear, too. Otherwise, let $x^{\prime},y^{\prime},z^{\prime}\in R$ be such that $(x^{\prime},y^{\prime})\in O_{1}$ and $(y^{\prime},z^{\prime})\in O_{2}$. If $(x^{\prime},z^{\prime})$ lies in the same orbit as $(x,z)$ in $\operatorname{Aut}(\mathfrak{R},\prec)$ then $(x,z)\in O_{1}\circ O_{2}$. Otherwise, consider the structure induced by $(\mathfrak{R},\prec)$ on $\\{x^{\prime},y^{\prime},z^{\prime}\\}$; we claim that if we replace the tuple $(x^{\prime},z^{\prime})$ in $\prec$ by the tuple $(z^{\prime},x^{\prime})$, the resulting structure still embeds into $(\mathfrak{R};\prec)$. By assumption, $O_{1}\subseteq\,\prec$ and $O_{2}\subseteq\,\succ$, or $O_{1}\subseteq\,\succ$ and $O_{2}\subseteq\,\prec$, so the modified relation $\prec$ is still acyclic. Moreover, since $R_{1}\circ R_{2}\in\\{\mathtt{DR},\mathtt{PO}\\}$, the modified relation $\prec$ is still a linear extension of $\mathtt{PP}$ and hence in $\operatorname{Age}(\mathfrak{R};\prec)$. The homogeneity of $(\mathfrak{R};\prec)$ implies that $(x^{\prime},z^{\prime})$ lies in the same orbit as $(x,z)$, and hence $(x,z)\in O_{1}\circ O_{2}$. This concludes the proof that $R_{1}\circ R_{2}\subseteq O_{1}\circ O_{2}$. ∎ ###### Corollary 9.14. Let $O\subseteq\;\prec$ be an orbit of pairs in $\operatorname{Aut}(\mathfrak{R};\prec)$. Then $\mathtt{PP}\subseteq O\circ O$. ###### Proof. If $O=\mathtt{PP}$, then $O\circ O=O=\mathtt{PP}$. If $O=(\mathtt{DR}\;\cap\prec)$ then $\displaystyle O\circ O$ $\displaystyle=(\mathtt{DR}\circ\mathtt{DR})\,\cap\prec$ (Lemma 9.13) $\displaystyle={\bf 1}\,\cap\prec$ (table in Figure 1) $\displaystyle=\;\prec\,.$ Similarly we may compute that $\mathtt{PP}$ is contained in $(\mathtt{PO}\,{\cap}\,{\prec})\circ(\mathtt{PO}\,{\cap}\,{\prec})$. ∎ The following general observations are useful when working with operations that preserve binary relations over some set $B$. If $f\colon B^{k}\to B$ is an operation that preserves $R_{1},\dots,R_{k},R^{\prime}_{1},\dots,R_{k}^{\prime}\subseteq B^{2}$ then $\displaystyle f(R_{1}\circ R_{1}^{\prime},\dots,R_{k}\circ R^{\prime}_{k})$ (13) $\displaystyle\subseteq\;$ $\displaystyle f(R_{1},\dots,R_{k})\circ f(R^{\prime}_{1},\dots,R^{\prime}_{k}).$ Also note that (14) $\displaystyle f(R_{1}^{\smile},\dots,R_{k}^{\smile})=f(R_{1},\dots,R_{k})^{\smile}.$ ###### Lemma 9.15. Let $f\in\operatorname{Pol}(\mathfrak{R})\cap\operatorname{Can}(\mathfrak{R},\prec)$. Then $f$ preserves $\prec$ and $\smash{{}^{\bot}_{\prec}}$. ###### Proof. Let $a_{1},\dots,a_{k},b_{1},\dots,b_{k}$ be elements of $\mathfrak{R}$ such that $a_{i}\prec b_{i}$ for all $i\in\\{1,\dots,k\\}$ and $f(b_{1},\dots,b_{k})\preceq f(a_{1},\dots,a_{k})$. For $i\in\\{1,\dots,k\\}$, let $O_{i}$ be the orbit of $(a_{i},b_{i})$ in $\operatorname{Aut}(\mathfrak{R},\prec)$. We then have $\displaystyle f(\mathtt{PP},\dots,\mathtt{PP})$ $\displaystyle\subseteq f(O_{1}\circ O_{1},\dots,O_{k}\circ O_{k})$ (Corollary 9.14) $\displaystyle\subseteq f(O_{1},\dots,O_{k})\circ f(O_{1},\dots,O_{k})$ (13) $\displaystyle\subseteq(\succeq\circ\succeq)$ (canonicity) $\displaystyle=\;\succeq$ which is a contradiction to $f$ being a polymorphism of $\mathfrak{R}$. Since $\bot$ is primitively positively definable in $\mathfrak{R}$, it also follows that every polymorphism of $\mathfrak{R}$ which is canonical with respect to $(\mathfrak{R},\prec)$ preserves $\smash{{}^{\bot}_{\prec}}$. ∎ ###### Lemma 9.16. Let $f\in\operatorname{Pol}(\mathfrak{R})$ be canonical over $\mathfrak{R}$ or over $(\mathfrak{R},\prec)$. Then $f$ preserves $1\setminus\mathtt{EQ}$. ###### Proof. Let $k$ be the arity of $f$. First suppose that $f$ is canonical over $(\mathfrak{R},\prec)$. Let $O_{1},\dots,O^{k}$ be orbits of pairs of $\operatorname{Aut}(\mathfrak{R};\prec)$ that are distinct from $\mathtt{EQ}$. Suppose for contradiction that $f(O_{1},\dots,O_{k})=\mathtt{EQ}$. Then $\displaystyle f(O_{1}\circ O_{1},\dots,O_{k}\circ O_{k})$ $\displaystyle\subseteq f(O_{1},\dots,O_{k})\circ\cdots\circ f(O_{1},\dots,O_{k})$ $\displaystyle\subseteq\mathtt{EQ}\circ\cdots\circ\mathtt{EQ}=\mathtt{EQ}.$ Since $O_{i}\circ O_{i}$ contains $\mathtt{PP}$ or $\mathtt{PPI}$ for every $i\in\\{1,\dots,k\\}$ by Corollary 9.14, it follows that $f(P_{1},\dots,P_{k})\subseteq\mathtt{EQ}$ for some $P_{1},\dots,P_{k}\in\\{\mathtt{PP},\mathtt{PPI}\\}$. Note that $f(P_{1}^{\smile},\dots,P_{k}^{\smile})=\mathtt{EQ}^{\smile}=\mathtt{EQ}$ by 14. Since $\mathtt{PP}\subseteq\mathtt{PP}\circ\mathtt{PPI}$ and $\mathtt{PP}\subseteq\mathtt{PPI}\circ\mathtt{PP}$ we obtain that $f(\mathtt{PP},\mathtt{PP})\subseteq\mathtt{EQ}\circ\mathtt{EQ}=\mathtt{EQ}$, a contradiction to $f(\mathtt{PP},\mathtt{PP})\subseteq\mathtt{PP}$. We conclude that $f$ preserves $1\setminus\mathtt{EQ}$. If $f$ is canonical over $\mathfrak{R}$ instead of $(\mathfrak{R},\prec)$, then the statement can be shown similarly, using the table in Figure 1 instead of Corollary 9.14. ∎ ### 9.3. Independent substructures To verify the UIP property, we will use Theorem 6.4 and therefore need certain pairs $(\mathfrak{A}_{1},\mathfrak{A}_{2})$ of independent elementary substructures of $(\mathfrak{R},\prec)$. ###### Lemma 9.17. There are elementary substructures $\mathfrak{A}_{1}$ and $\mathfrak{A}_{2}$ of $(\mathfrak{R},\prec)$ such that for all $a_{1}\in A_{1}$ and $a_{2}\in A_{2}$ we have $(a_{1},a_{2})\in\mathtt{DR}$ and $a_{1}\prec a_{2}$; in particular, $\mathfrak{A}_{1}$ and $\mathfrak{A}_{2}$ are independent. ###### Proof. By the homogeneity of $(\mathfrak{R},\prec)$ it suffices to show that for every structure $(\mathfrak{B},\prec)\in\operatorname{Age}(\mathfrak{R},\prec)$ there are embeddings $e_{1},e_{2}\colon\mathfrak{B}\to(\mathfrak{R},\prec)$ such that for all $b_{1},b_{2}\in B$ we have $(e_{1}(a_{1}),e_{2}(a_{2}))\in\mathtt{DR}$ and $e_{1}(b_{1})\prec e_{2}(b_{2})$. Choose an embedding $f$ of $\mathfrak{B}$ into the structure $\mathfrak{S}$ from the definition of $\mathfrak{R}$; so $f(b)\subseteq{\mathbb{N}}$ for each $b\in B$. Then $f_{1}\colon b\mapsto\\{2n\colon n\in f(b)\\}$ and $f_{2}\colon b\mapsto\\{2n+1\colon n\in f(b)\\}$ are two embeddings of $\mathfrak{B}$ into $\mathfrak{S}$ such that $(f_{1}(b_{1}),f_{2}(b_{2}))\in\mathtt{DR}$ for all $b_{1},b_{2}\in B$. Let $\mathfrak{B}^{\prime}$ be the substructure of $\mathfrak{S}$ with domain $B^{\prime}\coloneqq f_{1}(B)\cup f_{2}(B)$. For all $b_{1},b_{2}\in B$ with $b_{1}\prec b_{2}$, define the linear order $\prec$ on $B^{\prime}$ by $f_{1}(b_{1})\prec f_{1}(b_{2})$, $f_{2}(b_{1})\prec f_{2}(b_{2})$, $f_{1}(b_{1})\prec f_{2}(b_{2})$. Then it is straightforward to check that $\prec$ extends $\mathtt{PP}$, and hence $(\mathfrak{B}^{\prime},\prec)\in\operatorname{Age}(\mathfrak{R},\prec)$, which concludes the proof. ∎ Independent substructures are used in the proof of the following lemma that plays an important role when verifying the UIP later. ###### Lemma 9.18. Let $\mathfrak{C}$ be a first-order expansion of $\mathfrak{R}$ and let $\zeta\colon\operatorname{Pol}(\mathfrak{C})\cap{\operatorname{Can}(\mathfrak{R},\prec)}\to\operatorname{Proj}$ be a clone homomorphism which does not have the UIP with respect to $\operatorname{Pol}(\mathfrak{C})$. Then for every finite 2-rich finite subset $F$ of the domain of $\mathfrak{R}$ there exist $f\in\operatorname{Pol}(\mathfrak{C})^{(2)}$ and $\alpha_{1},\alpha_{2}\in\operatorname{Aut}(\mathfrak{R};\prec)$ such that * • $f$ is canonical on $F\times\alpha_{1}(F)$ and on $F\times\alpha_{2}(F)$ with respect to $(\mathfrak{R},\prec)$, * • $\zeta(f(\operatorname{id},\alpha_{1})|_{F^{2}})\neq\zeta(f(\operatorname{id},\alpha_{2})|_{F^{2}})$ (recall Definition 8), * • for all $a,b\in F$, if $(a,b)\in\mathtt{PP}\cup\mathtt{EQ}$, then $(\alpha_{1}(a),\alpha_{2}(b))\in\mathtt{PP}$, and * • for all $a,b\in F$, if $(a,b)\in\;\bot$ then $(\alpha_{1}(a),\alpha_{2}(b))\in\;\bot$. ###### Proof. Let $\mathfrak{A}_{1},\mathfrak{A}_{2}$ be the two independent elementary substructures of $(\mathfrak{R},\prec)$ from Lemma 9.17. Since $\zeta$ does not have the UIP with respect to $\operatorname{Pol}(\mathfrak{C})$, by Theorem 6.4 there exists $f^{\prime}\in\operatorname{Pol}(\mathfrak{C})^{(2)}$ and $u_{1},u_{2}\in\overline{\operatorname{Aut}(\mathfrak{A})}$ such that for $i\in\\{1,2\\}$ the image $\Im(u_{i})$ of $u_{i}$ is contained in $A_{i}$, the operation $f^{\prime}(\operatorname{id},u_{i})$ is canonical with respect to $(\mathfrak{R},\prec)$, and $\zeta(f^{\prime}(\operatorname{id},u_{1}))\neq\zeta(f^{\prime}(\operatorname{id},u_{2}))$. By Lemma 5.5 there exists a finite subset $X$ of the domain of $\mathfrak{R}$ such that for all $g\in\operatorname{Pol}(\mathfrak{R})^{(2)}$ there exist $\beta_{1},\beta_{2}\in\operatorname{Aut}(\mathfrak{R},\prec)$ with $\beta_{1}(F),\beta_{2}(F)\subseteq X$ such that $g$ is canonical on $\beta_{1}(F)\times\beta_{2}(F)$. For $i\in\\{1,2\\}$, let $\epsilon_{i}\in\operatorname{Aut}(\mathfrak{R},\prec)$ be such that $\epsilon_{i}(X)\subseteq\Im(u_{i})$. We may view the substructure of $\mathfrak{R}$ induced by $\epsilon_{1}(X)\cup\epsilon_{2}(X)$ as a substructure of $\mathfrak{S}$. By the homogeneity of $\mathfrak{R}$ we may also assume that the substructure of $\mathfrak{S}$ on $\epsilon_{1}(X)\cup\epsilon_{2}(X)\cup\\{\epsilon_{1}(x)\cup\epsilon_{2}(x)\colon x\in X\\}$ has an embedding $e$ into $\mathfrak{R}$ such that $e(\epsilon_{i}(x))=\epsilon_{i}(x)$ for all $x\in X$ and $i\in\\{1,2\\}$. Define $\epsilon\colon X\to\mathfrak{R}\text{ by }\epsilon(x)\coloneqq e\big{(}\epsilon_{1}(x)\cup\epsilon_{2}(x)\big{)}.$ Let $a,b\in X$. Note that 1. (1) if $(a,b)\in\mathtt{PP}\cup\mathtt{EQ}$ then $\displaystyle\epsilon_{i}(a)=e(\epsilon_{i}(a))$ $\displaystyle\subset e\big{(}\epsilon_{1}(a)\cup\epsilon_{2}(a)\big{)}$ $\displaystyle\subseteq e\big{(}\epsilon_{1}(b)\cup\epsilon_{2}(b)\big{)}=\epsilon(b)$ and hence $(\epsilon_{i}(a),\epsilon(b))\in\mathtt{PP}$; 2. (2) if $(a,b)\in\;\bot$ then $(\epsilon_{i}(a),\epsilon_{i}(b))\in\;\bot$. Since $\epsilon_{i}(a)\in A_{i}$ it follows that $\epsilon_{i}(a)$ is disjoint from $\epsilon_{j}(b)$ for $j\neq i$. This implies that $\epsilon_{i}(a)\setminus\epsilon(b)=\epsilon_{i}(a)\setminus\big{(}\epsilon_{1}(b)\cup\epsilon_{2}(b)\big{)}=\epsilon_{i}(a)\setminus\epsilon_{i}(b)\neq\emptyset$ and that for $j\neq i$ we have $\epsilon_{j}(b)=\epsilon_{j}(b)\setminus\epsilon_{i}(a)\subseteq\epsilon(b)\setminus\epsilon_{i}(a).$ In particular, $\epsilon(b)\setminus\epsilon_{i}(a)\neq\emptyset$. We obtained that the sets $\epsilon_{i}(a)\setminus\epsilon(b)$ and $\epsilon(b)\setminus\epsilon_{i}(a)$ are both non-empty. Therefore $(\epsilon_{i}(a),\epsilon(b))\in\;\bot$. We define an order $\prec^{\prime}$ on $\epsilon_{1}(X)\cup\epsilon_{2}(X)\cup\epsilon(X)$ by setting $a\prec^{\prime}b$ if one of the following holds. * • $a,b\in\epsilon_{1}(X)\cup\epsilon_{2}(X)$ and $a\prec b$; * • $a\in\epsilon_{1}(X)\cup\epsilon_{2}(X)$ and $b\in\epsilon(X)$; * • $a,b\in\epsilon(X)$ and $\epsilon^{-1}(a)\prec\epsilon^{-1}(b)$. Then it is easy to see that $\prec^{\prime}$ defines a partial order that extends $\mathtt{PP}$; let $\prec^{\prime\prime}$ be a linear order that extends $\prec^{\prime}$. By the definition of $(\mathfrak{R};\prec)$ there exists an automorphism $\gamma$ of $\mathfrak{R}$ that maps $\prec^{\prime\prime}$ to $\prec$. This shows that we may assume that $\epsilon$ preserves $\prec$ (otherwise, replace $\epsilon$ by $\gamma\circ\epsilon$). By the definition of $X$ there are $\beta_{1},\beta_{2}\in\operatorname{Aut}(\mathfrak{R},\prec)$ such that $\beta_{1}(F)\subseteq X$, $\beta_{2}(F)\subseteq X$, and $f^{\prime}(\operatorname{id},\epsilon)$ is canonical on $\beta_{1}(F)\times\beta_{2}(F)$ over $(\mathfrak{R},\prec)$. Since $\epsilon_{i}\beta_{2}(F)\subseteq\epsilon_{i}(X)\subseteq\Im(u_{i})$ for $i\in\\{1,2\\}$ we have that $\zeta(f^{\prime}|_{\beta_{1}(F)\times\epsilon_{1}\beta_{2}(F)})\neq\zeta(f^{\prime}|_{\beta_{1}(F)\times\epsilon_{2}\beta_{2}(F)})$. Then for some $i\in\\{1,2\\}$ we have that $\zeta(f^{\prime}|_{\beta_{1}(F)\times\epsilon_{i}\beta_{2}(F)})\neq\zeta(f^{\prime}|_{\beta_{1}(F)\times\epsilon\beta_{2}(F)})$. We claim that $f\coloneqq f^{\prime}(\beta_{1},\operatorname{id})$, $\alpha_{1}\coloneqq\epsilon_{i}\circ\beta_{2}$ and $\alpha_{2}\coloneqq\epsilon\circ\beta_{2}$ satisfy the conclusion of the lemma: * • $f$ is canonical on $F\times\alpha_{i}(F)$, for $i\in\\{1,2\\}$, because $f^{\prime}(\operatorname{id},\epsilon)$ is canonical on $\beta_{1}(F)\times\beta_{2}(F)$. * • for all $(a,b)\in F^{2}\cap(\mathtt{PP}\cup\mathtt{EQ})$ we have that $(\beta_{2}(a),\beta_{2}(b))\in X^{2}\cap(\mathtt{PP}\cup\mathtt{EQ})$ and by (1) we obtain $(\alpha_{1}(a),\alpha_{2}(b))=(\epsilon_{i}(\beta_{2}(a)),\epsilon(\beta_{2}(b)))\in\mathtt{PP}$. * • for all $(a,b)\in F^{2}\cap\bot$ we have that $(\beta_{2}(a),\beta_{2}(b))\in X^{2}\cap\bot$ and by (2) we obtain $\displaystyle(\alpha_{1}(a),\alpha_{2}(b))$ $\displaystyle=(\epsilon_{i}(\beta_{2}(a)),\epsilon(\beta_{2}(b)))\in\;\bot.\qed$ ### 9.4. Uniformly continuous minor-preserving maps Let $\mathfrak{B}$ be a first-order expansion of $\mathfrak{R}$. In this section we show that if there exists a uniformly continuous minor-preserving map from $\operatorname{Pol}(\mathfrak{B})\cap\operatorname{Can}(\mathfrak{R})$ to $\operatorname{Proj}$, then a specific minor-preserving map $\eta$ (introduced in Section 9.4.2) or $\rho$ (introduced in Section 9.4.3) is a uniformly continuous minor-preserving map from $\operatorname{Pol}(\mathfrak{B})\cap\operatorname{Can}(\mathfrak{R},\prec)\to\operatorname{Proj}$. For this task, we need some more results from finite universal algebra that will be recalled in the following (Section 9.4.1). #### 9.4.1. Cyclic polymorphisms An operation $f\colon A^{k}\to A$ is * • _idempotent_ if $f(x,\dots,x)=x$ for every $x\in A$. * • _cyclic_ if $k\geq 2$ and $f(x_{1},\dots,x_{k})=f(x_{2},\dots,x_{k},x_{1})$
# Mean-field limits for non-linear Hawkes processes with excitation and inhibition by P. Pfaffelhuber, S. Rotter and J. Stiefel _Albert-Ludwigs University Freiburg_ ###### Abstract We study a multivariate, non-linear Hawkes process $Z^{N}$ on the complete graph with $N$ nodes. Each vertex is either excitatory (probability $p$) or inhibitory (probability $1-p$). We take the mean-field limit of $Z^{N}$, leading to a multivariate point process $\bar{Z}$. If $p\neq\tfrac{1}{2}$, we rescale the interaction intensity by $N$ and find that the limit intensity process solves a deterministic convolution equation and all components of $\bar{Z}$ are independent. In the critical case, $p=\tfrac{1}{2}$, we rescale by $N^{1/2}$ and obtain a limit intensity, which solves a stochastic convolution equation and all components of $\bar{Z}$ are conditionally independent. †† _AMS 2000 subject classification._ 60G55 (Primary) , 60F05 (Secondary).†† _Keywords and phrases._ Multivariate Hawkes process; Volterra equation; spike train ## 1 Introduction In [19, 21], Hawkes processes were introduced as self-excitatory point processes. Today, they are used in various fields of applications including seismology [33, 17], interactions in social networks [41, 31], finance [14, 20] and neuroscience [18, 27]. In the classical univariate, linear Hawkes process, the firing rate at time $t$ is a linear function of $I_{t}:=\sum_{i}\varphi(t-T_{i})$, where the $T_{i}$’s are previous jump times. In this case, since rates cannot become negative, $\varphi\geq 0$ is required, leading to a self-excitatory process. Our main motivation to study Hawkes processes comes from the neurosciences. In a graph, vertices model neurons, whereas the (directed) edges are synapses linking the neurons. A point process indexed by the vertices models action potentials or spike trains of electrical impulses. Communication via synapses leads to correlated point processes such that each spike in one neuron influences the rate by which a neighboring vertex fires. In this field, it is known that neurons cannot only excite others, but inhibition is another important factor (see e.g. [30]). The main goal of this paper is to obtain limit results for the multivariante, non-linear Hawkes process, where the firing rate at time $t$ is $h(I_{t})$ and $\varphi\leq 0$ can occur as well, which we interpret as inhibition. Nonlinear Hawkes processes have been studied to some extent in the past decades. [4, 32] focus on ergodic properties using Lipschitz conditions of the transfer function $h$. In [39, 40], central limit theorems and large deviation results for a univariate nonlinear Hawkes process are given. The case of inhibition in a multivariate setting was studied in [6] using a thinning process representation, while [11] uses renewal techniques to establish limit theorems in the univariate setting. Mean-field models have frequently been applied in the life sciences; see e.g. [36] where particular applications in neuroscience are discussed. Mean-field limits of nonlinear Hawkes processes were first studied by [12]. This has been extended to age-dependent nonlinear Hawkes processes by [8, 7], and to specific models including multiple classes [13]. In addition to the law of large numbers in the mean-field model, a central limit theorem is proved in [7, 23]. Another branch of research studies a mean-field limit of a spatially extended (geometric) Hawkes process, showing a law of large numbers [9] and central limit theorem [10]. Here, a first proof of the neural field equation could be obtained. In the present paper, our approach is a combination and extension of [8],[23] and [15], with a special focus on models including excitatory and inhibitory neurons/vertices. We assume that each vertex/neuron is either excitatory or inhibitory, i.e. excites or inhibits all of its neighbors. We will denote by $p$ the fraction of excitatory vertices/neurons and distinguish the critical case $p=1/2$ from the non-critical one. For the latter, we obtain in Theorem 1 a classical mean-field result, i.e. by rescaling the interaction intensity by $N$, a deterministic limit of the intensity and independent point processes driven by this intensity arises. We also provide a central limit result for the intensity. In Theorem 2, we are dealing with the critical case. Here, we rescale the interaction intensity by $N^{1/2}$, leading to a limiting intensity which is itself stochastic, and drives conditionally independent point processes. We will discuss connections of our findings to previous work in Remarks 3.4 and 3.6. The mean-field model we discuss here is unrealistic for applications in neuroscience for various reasons. Above all, not all neurons are connected. Sparse graphs are one possibility for a more realistic model [5], while multiple classes of differently interconnected neurons are another option [13]. For a detailed discussion on implications of our findings for neuroscience, see Section 4. ## 2 Model and main results We use the following general model for a non-linear Hawkes process: ###### Definition 2.1 (Multi-variate, non-linear Hawkes process). Let $\mathbb{G}=(\mathbb{V},\mathbb{E})$ be some finite, directed graph, and write $ji\in\mathbb{E}$ if $j\to i$ is an edge in $\mathbb{G}$. Consider a family of measurable, real-valued functions $(\varphi_{ji})_{ji\in\mathbb{E}}$ and a family of real-valued, non-negative functions $(h_{i})_{i\in\mathbb{V}}$. Then, a point process $Z=(Z^{i})_{i\in\mathbb{V}}$ (with state space $\mathbb{N}_{0}^{\mathbb{V}}$) is a multi-variate non-linear Hawkes process with interaction kernels $(\varphi_{ji})_{ji\in\mathbb{E}}$ and transfer functions $(h_{i})_{i\in\mathbb{V}}$, if $Z^{i},Z^{j}$ do not jump simultaneously for $i\neq j$, almost surely, and the compensator of $Z^{i}$ has the form $(\int_{0}^{t}\lambda_{s}^{i}ds)_{t\geq 0}$ with $\lambda_{t}^{i}:=\lambda_{t}^{i}(Z_{s<t}):=h_{i}\Big{(}\sum_{j:ji\in\mathbb{E}}\int_{0}^{t-}\varphi_{ji}(t-s)dZ_{s}^{j}\Big{)},\qquad i\in\mathbb{V}.$ We need some mimimal conditions such that the multivariate, non-linear Hawkes process is well-defined (i.e. exists). As mentioned in [12], Remark 5, the law of the non-linear Hawkes process is well-defined, provided that the following assumption holds. ###### Assumption 2.2. All interaction kernels $\varphi_{ji},ji\in\mathbb{E}$ are locally integrable, and all transfer functions $(h_{i})_{i\in\mathbb{V}}$ are Lipschitz continuous. ###### Remark 2.3 (Interpretation and initial condition). 1. 1. If $dZ_{s}^{j}=1$, we call $\varphi_{ji}(t-s)dZ_{s}^{j}$ the influence of the point at time $s$ in vertex $j$ on vertex $i$. 2. 2. Consider the case of monotonically increasing transfer functions. If $\varphi_{ji}\leq 0$, we then say that vertex $j$ inhibits $i$, since any point in $Z^{j}$ decreases the jump rate of $Z^{i}$. Otherwise, if $\varphi_{ji}\geq 0$, we say that $j$ excites $i$. 3. 3. In our formulation, we have $Z_{0}^{i}=0,i\in\mathbb{V}$, with the consequence that the $dZ_{u}^{j}$-integral in (5.5) could also be extended to $-\infty$ without any change. We note that it would also be possible to use some initial condition, i.e. some (fixed) $(Z_{t}^{i})_{i\in\mathbb{V},t\leq 0}$, and extend the integral to the negative reals. Let us now come to the mean-field model, where we now fix some basic assumptions. Note that we will show convergence for large graphs, i.e. all processes come with a scaling parameter $N$, which determines the size of the graph. ###### Assumption 2.4 (Mean-field setting). Let 1. 1. $\mathbb{G}_{N}=(\mathbb{V}_{N},\mathbb{E}_{N})$ be the complete graph on $N$ vertices, i.e. $\mathbb{V}_{N}=\\{1,...,N\\}$ and $ji,ij\in\mathbb{E}$ for all $i,j\in\mathbb{V}$; 2. 2. $h_{i}=h$ for all $i\in\mathbb{V}_{N}$ where $h\geq 0$ is bounded, $h$ and $\sqrt{h}$ are Lipschitz with constant $h_{Lip}$; 3. 3. $\varphi_{ji}=\theta_{N}U_{j}\varphi$ for all $j,i\in\mathbb{V}_{N}$, where $U_{1},U_{2},...$ are iid with $\mathbb{P}(U_{1}=1)=1-\mathbb{P}(U_{1}=-1)=p$, $\varphi\in\mathcal{C}_{b}^{1}([0,\infty))$, the set of bounded continuously differentiable functions with bounded derivative, and $\theta_{N}\in\mathbb{R}$. The form of $\varphi_{ji}$ implies that node $j$ is exciting all other nodes with probability $p$, and inhibiting all other nodes with probability $1-p$. By the law of large numbers, we have that $\displaystyle\frac{1}{N}\sum_{j=1}^{N}U_{j}\xrightarrow{N\to\infty}2p-1$ (2.1) almost surely, and by the central limit theorem, $\displaystyle\frac{1}{\sqrt{N}}\sum_{j=1}^{N}(U_{j}+1-2p)=:W_{N}\xRightarrow{N\to\infty}W\sim N(0,4p(1-p)),$ (2.2) where $\Rightarrow$ denotes weak convergence. For convergence, this suggests to use $\theta_{N}=\tfrac{1}{N}$ for $p\neq\tfrac{1}{2}$ and $\theta_{N}=\tfrac{1}{\sqrt{N}}$ for $p=\tfrac{1}{2}$. ## 3 Results on the mean-field model Our main goal is to give a limit result on the family $Z^{N}$, the multivariate, non-linear Hawkes process on the graph $\mathbb{G}_{N}$ with interaction kernels and transfer functions as given in Assumption 2.4. We distinguish the cases $p\neq\tfrac{1}{2}$ (Theorem 1) and $p=\tfrac{1}{2}$ (Theorem 2). In both cases, the limit compensator is given by $(\int_{0}^{t}h(I_{s})ds)_{t\geq 0}$, where $I$ is the weak limit of $I^{N}$ (see (3.1) and (3.9)), and we find the following limit results: For $p\neq\tfrac{1}{2}$, we have that $(I_{t})_{t\geq 0}$ follows a linear, deterministic convolution equation, and all components of the limit of $Z^{N}$ are independent. For $p=\tfrac{1}{2}$, the critical case, the process $(I_{t})_{t\geq 0}$ is the unique solution of a stochastic convolution equation and all components of $Z^{N}$ are conditionally independent given $I$. Below, we denote by $\Rightarrow$ weak convergence in $\mathcal{D}_{\mathbb{R}^{n}}([0,\infty))$, the space of cadlag paths, which is equipped with the Skorohod topology; see e.g. Chapter 3 in [16]. The proof of the following result can be found in Section 5.4. ###### Theorem 1 (Mean-field limit of multi-variate non-linear Hawkes processes, $p\neq\tfrac{1}{2}$). Let Assumption 2.4 hold with $p\neq\tfrac{1}{2}$ and $\theta_{N}=\tfrac{1}{N}$. Let $Z^{N}=(Z^{N,1},...,Z^{N,N})$ be the multivariate, non-linear Hawkes process from Definition 2.1, and $\displaystyle I^{N}(t):=\frac{1}{N}\sum_{j=1}^{N}\int_{0}^{t-}U_{j}\varphi(t-s)dZ_{s}^{j,N}.$ (3.1) 1. 1. Then, $I^{N}\xrightarrow{N\to\infty}I$, uniformly on compact time intervals in $L^{2}$, where $I=(I_{t})_{t\geq 0}$ is the unique solution of the integral equation $\displaystyle I_{t}=(2p-1)\int_{0}^{t}\varphi(t-s)h(I_{s})ds.$ (3.2) 2. 2. For all $n=1,2,...$, $(Z^{N,1},...,Z^{N,n})\xRightarrow{N\to\infty}(\bar{Z}^{1},...,\bar{Z}^{n}),$ where $\bar{Z}^{1},...,\bar{Z}^{n}$ are independent and $\bar{Z}^{i}$ is a simple point process with intensity at time $t$ given by $h(I_{t})$, $i=1,...,n$. It is possible to build $\bar{Z}^{1},...,\bar{Z}^{n}$, such that the convergence is almost surely (in Skorohod-distance). 3. 3. Assume that $h\in\mathcal{C}^{1}(\mathbb{R})$ and that $h^{\prime}$ is bounded and Lipschitz. Then $\sqrt{N}(I^{N}-I)\xRightarrow{N\to\infty}K$, where $K$ is the unique (strong) solution of $\displaystyle K_{t}$ $\displaystyle=\int_{0}^{t}\varphi(t-s)dG_{s},$ (3.3) $\displaystyle G_{t}$ $\displaystyle=\int_{0}^{t}(Wh(I_{s})+(2p-1)h^{\prime}(I_{s})K_{s})ds+\int_{0}^{t}\sqrt{h(I_{s})}dB_{s},$ where $W$ is given by (2.2) and $B$ is a Brownian motion independent from $W$. ###### Remark 3.1. The form of (3.2) tells us that $I$ follows a linear Volterra convolution equation [3]. Turning into a differential equation, we write, using Fubini, $\displaystyle\frac{dI}{dt}$ $\displaystyle=\frac{d}{dt}(2p-1)\Big{(}\int_{0}^{t}\int_{s}^{t}\varphi^{\prime}(r-s)h(I_{s})drds+\int_{0}^{t}\varphi(0)h(I_{s})ds\Big{)}$ $\displaystyle=(2p-1)\Big{(}\int_{0}^{t}\varphi^{\prime}(t-s)h(I_{s})ds+\varphi(0)h(I_{t})\Big{)}.$ In particular, the special choice of $\varphi(s)=e^{-\lambda s}$ gives $\displaystyle\frac{dI}{dt}$ $\displaystyle={\color[rgb]{0,0,0}-\lambda(2p-1)\int_{0}^{t}\varphi(t-s)h(I_{s})ds+(2p-1)h(I_{t})=-\lambda I_{t}+(2p-1)h(I_{t}),}$ i.e. $I$ follows some ordinary differential equation in this case. While Theorem 1 is concerned with convergence of the limit intensity of the multivariate, non-linear Hawkes process, we are also in the situation to study convergence of the average intensity of $Z^{N}$. The proof of the next corollary is found in Section 5.5. ###### Corollary 3.2. Let $Z^{N}=(Z^{N,1},...,Z^{N,N})$ be as in Theorem 1, and $\bar{Z}=(\bar{Z}^{1},\bar{Z}^{2},...)$ be as in Theorem 1.2. Then, $\displaystyle\frac{1}{N}\sum_{j=1}^{N}\Big{(}Z^{N,j}-\int_{0}^{.}h(I^{N}_{s})ds\Big{)}\xrightarrow{N\to\infty}0$ (3.4) and $\displaystyle\frac{1}{N}\sum_{j=1}^{N}\Big{(}Z^{N,j}-\bar{Z}^{j}\Big{)}\xrightarrow{N\to\infty}0$ (3.5) in probability, uniformly on compact intervals. Moreover, $\displaystyle\frac{1}{\sqrt{N}}\sum_{j=1}^{N}\Big{(}Z^{N,j}-\int_{0}^{.}h(I_{s}^{N})ds\Big{)}\xRightarrow{N\to\infty}\int_{0}^{.}\sqrt{h(I_{s})}d\widetilde{B}_{s}$ (3.6) for some Brownian motion $\widetilde{B}$, and (writing $h(I):=(h(I_{t}))_{t\geq 0}$) $\displaystyle\sqrt{N}\big{(}h(I^{N})-h(I)\big{)}\xRightarrow{N\to\infty}h^{\prime}(I)K.$ (3.7) Let $B,\widetilde{B}$ be correlated Brownian motions with $\mathbf{E}[B_{t}\widetilde{B}_{t}]=(2p-1)t,t\geq 0$. Use $B$ in the definition of $K$ in (3.3). Then, $\displaystyle\frac{1}{\sqrt{N}}\sum_{j=1}^{N}\Big{(}Z^{N,j}-\int_{0}^{.}h(I_{s})ds\Big{)}\xRightarrow{N\to\infty}\int_{0}^{.}h^{\prime}(I_{s})K_{s}ds+\int_{0}^{.}\sqrt{h(I_{s})}d\widetilde{B}_{s}.$ (3.8) ###### Remark 3.3 (Correlation between $B$ and $\widetilde{B}$). Let us briefly discuss the correlated Brownian motions appearing in (3.8). Clearly, the left hand sides of (3.6) and (3.7) sum to the left hand side of (3.8). The limits $\int_{0}^{.}\sqrt{h(I_{s})}d\widetilde{B}_{s}$ and $K$ appearing on the right hand sides of (3.6) and (3.7) are weak limits of sums of compensated point processes. While in (3.6), we sum over all point processes in the system, $I^{N}$ in (3.7) distinguishes between nodes with different signs $U_{j}$. Hence, the correlation is positive for the proportion $p$ of point processes with positive sign, and negative for the proportion $1-p$ of point processes with negative sign, summing to $p-(1-p)=2p-1$. For more details, see the proof in Section 5.5. ###### Remark 3.4 (Connections of Theorem 1 to previous work). Connections of Theorem 1.1 and Theorem 1.2 to [8, 23]. First note that 1. and 2. of Theorem 1 are a particular case of [8], Theorem 4.1. In their notation, we have choosen $H_{ij}=U_{j}h$ as interaction functions, which satisfy their condition (9), as stated in their remark 2.1 on the synaptic weights. We reformulate the result to introduce 3. of of Theorem 1, which does not appear in [8], and to allow for a comparison to the result and proof of our Theorem 2. Note further that Theorem 1.1 implies that the compensator $(\Lambda_{t})_{t\geq 0}$ of each of the independent limiting processes $(\bar{Z}^{i}_{t})_{t\geq 0}$ is $\Lambda_{t}=\int_{0}^{t}h(I_{s})ds=\int_{0}^{t}h\Big{(}(2p-1)\int_{0}^{s}\varphi(s-r)h(I_{r})dr\Big{)}ds=\int_{0}^{t}h\Big{(}(2p-1)\int_{0}^{s}\varphi(s-r)d\Lambda_{r}\Big{)}ds.$ See also (8) of[12] and (5), (7), (8) of [23] for the connection to (3.2), both for the special case $p=1$ and $W=0$. In contrast to our setting, recall that [23] require $\varphi(0)=0$. Connections of Theorem 1.3 to [7], [23]. Equation (5.10) from Theorem 5.6 in [7] gives an expression for the fluctuation of age dependent Hawkes processes. In the case where the transfer function becomes independent of the age, $\Psi(s,x)\equiv\Psi(x)$ in their notation, we discover our equation (3.3) in the special case $p=1$ and $W=0$. As Corollary 2.2 of [23] is a special case of this Theorem 5.6, (33) in that corollary is precisely (3.3), and the convergence of that Corollary for the special case $p=1$ (as well as $W=0$ and $\varphi(0)=0$) coincides with the statement of Theorem 1.3. Connections of Corollary 3.2 to [23]. The convergence in (3.7) generalizes Theorem 2.4 of [23] to the case $p\neq 1$, $W\neq 0$ and $\varphi(0)\neq 0$. Moreover, the convergence in (3.8) generalizes Theorem 2.1 of [23]: the right hand side of (3.8) for $p=1$, $W=0$ has $B=\widetilde{B}$ and hence equals $G$ from (3.3). Using integration by parts and $\varphi(0)=0$ we obtain $\displaystyle 0$ $\displaystyle=\varphi(0)G_{t}-\varphi(t)G_{0}=\int_{0}^{t}\varphi(t-s)dG_{s}-\int_{0}^{t}G_{s}\varphi^{\prime}(t-s)ds.$ Therefore we can reshape $G$ from (3.3), $\displaystyle G_{t}$ $\displaystyle=\int_{0}^{t}h^{\prime}(I_{u})\int_{0}^{u}\varphi(u-s)dG_{s}du+\int_{0}^{t}\sqrt{h(I_{s})}dB_{s}$ $\displaystyle=\int_{0}^{t}h^{\prime}(I_{u})\int_{0}^{u}\varphi^{\prime}(u-s)G_{s}dsdu+\int_{0}^{t}\sqrt{h(I_{s})}dB_{s}$ $\displaystyle=\int_{0}^{t}G_{s}\int_{s}^{t}\varphi^{\prime}(u-s)h^{\prime}(I_{u})duds+\int_{0}^{t}\sqrt{h(I_{s})}dB_{s},$ to obtain their (24). We now turn to the case $p=\tfrac{1}{2}$, where we can rescale the intensity by $\sqrt{N}$ rather than $N$. This is more involved since the limit of $I^{N}$ turns out to be a stochastic process. For the proof of the following Theorem, see Section 5.3. ###### Theorem 2 (Mean-field limit of multi-variate non-linear Hawkes processes, $p=\tfrac{1}{2}$). Let Assumption 2.4 hold with $p=\tfrac{1}{2}$ and $\theta_{N}=\tfrac{1}{\sqrt{N}}$. Let $Z^{N}=(Z^{N,1},...,Z^{N,N})$ be the multivariate, non-linear Hawkes process fom Definition 2.1, and $\displaystyle I^{N}(t):=\frac{1}{\sqrt{N}}\sum_{j=1}^{N}\int_{0}^{t-}U_{j}\varphi(t-s)dZ_{s}^{j,N}.$ (3.9) 1. 1. For all $w\in\mathbb{R}$, and a Brownian motion $B$, the stochastic integral equation $\displaystyle I_{t}=I_{0}+w\int_{0}^{t}\varphi(t-s)h(I_{s})ds+\int_{0}^{t}\varphi(t-s)\sqrt{h(I_{s})}dB_{s}$ (3.10) has a unique strong solution. 2. 2. Let $W\sim N(0,1)$, $B$ be an independent Brownian motion and $I$ be the solution of (3.10) with $w$ replaced by $W$. Then, for all $n=1,2,...$, $(I^{N},Z^{N,1},...,Z^{N,n})\xRightarrow{N\to\infty}(I,\bar{Z}^{1},...,\bar{Z}^{n})$, where $\bar{Z}^{1},...,\bar{Z}^{n}$ are conditionally independent given $I$ and $\bar{Z}^{i}$ is a simple point process with intensity at time $t$ given by $h(I_{t})$, $i=1,...,n$. Again, we discuss convergence of the average intensity of $Z^{N}$. The proof is in Section 5.5. ###### Corollary 3.5. Let $Z^{N}=(Z^{N,1},...,Z^{N,N})$ be as in Theorem 2, and $(I,\bar{Z}^{1},\bar{Z}^{2},...)$ be as in Theorem 2.2. Then, (3.4), (3.5), (3.6) hold. We stress that the results corresponding to (3.7) and (3.8) do not hold here since $I^{N}\Rightarrow I$, so no further upscaling is possible. ###### Remark 3.6 (Some connections to [15]). Our model and the statements of Theorem 2.2 should be compared to the findings in [15]. In their model, the size of interactions is given by a centered probability measure, so at each spike of a neuron, the potential given to other neurons in the system may change. In contrast, in our model, the potential given to other neurons is fixed for each neuron, but varies from neuron to neuron. The latter case seems more relevant in neuroscience; see the discussion in Section 4. Still, results from Theorem 2 can be compare to Theorem 1.4(ii) and Theorem 1.7. of [15], where the same type of convergence is shown, but the interaction kernel $\varphi$ is an exponential function. As we see in Remark 3.7.2 below, this special choice makes the intensity $I$ a Markov process. However, as the intensity of both models is still comparable, we find (3) from [15] in our Remark 3.7.2 below by inserting $w=0$. ###### Remark 3.7 (Some properties of $I$). 1. 1. If $\varphi^{\prime}$ exists, it is not hard to see that $I_{t}$ is a semimartingale, since $\displaystyle\int_{0}^{t}\varphi(t-u)\sqrt{h(I_{u})}dB_{u}$ $\displaystyle=\int_{0}^{t}\int_{0}^{s}\varphi^{\prime}(t-s)\sqrt{h(I_{u})}dB_{u}ds+\int_{0}^{t}\varphi(0)\sqrt{h(I_{u})}dB_{u}$ by the stochastic Fubini Theorem [3, Theorem 4.A], where the first term on the right hand side has finite variation and the second is a (local) martingale. This calculation also shows that $I$ has finite variation provided that $\varphi(0)=0$. 2. 2. Consider the special case $\varphi(s)=e^{-\lambda s}$. Then, a solution of (3.10) also solves $dI=(wh(I)-\lambda I)dt+\sqrt{h(I)}dB$ and in particular is a Markov process. In order to see this, we write $\displaystyle I_{t}-I_{0}=e^{-\lambda t}\Big{(}w\int_{0}^{t}e^{\lambda s}h(I_{s})ds+\int_{0}^{t}e^{\lambda s}\sqrt{h(I_{s})}dB_{s}\Big{)}.$ Using integration by parts (on the processes $A_{t}:=e^{-\lambda t}$, which satisfies $dA=-\lambda Adt$, and the term $C_{t}$ in brackets), we find that $\displaystyle I_{t}-I_{0}$ $\displaystyle=A_{t}C_{t}=\int_{0}^{t}A_{s}dC_{s}+\int_{0}^{t}C_{s}dA_{s}$ $\displaystyle=w\int_{0}^{t}h(I_{s})ds+\int_{0}^{t}\sqrt{h(I_{s})}dB_{s}-\lambda\int_{0}^{t}A_{s}C_{s}ds$ $\displaystyle=\int_{0}^{t}(wh(I_{s})-\lambda I_{s})ds+\int_{0}^{t}\sqrt{h(I_{s})}dB_{s}.$ ## 4 Implications for neuroscience The human brain is very large. It comprises 86 billion neurons, along with other types of cells [24]. Each individual neuron makes contact with hundreds or thousands of other neurons, forming a very complex network. It should be thought of as a highly integrated circuit, made of biological parts, which are capable of processing electro-chemical signals and transmitting them to distant neurons with high fidelity. Neurons use electrical impulses (“action potentials” or “spikes”) to communicate with other neurons in the same network. Sequences of these impulses evolve in time (“spike trains”), involving the collective activity of many interacting neurons. These multivariate signals are believed to reflect the dynamic “computations” performed by these networks. The biological “purpose” of these processes covers a wide range from immediate control of bodily functions, like activating a muscle, to abstract information processing, like making a complex decision [30]. Neurons communicate with other neurons in a network by activating specialized functional contacts between cells (“synapses”). In the most common type of synapses, diffusing chemicals are used for signaling (“neurotransmitters”). In a nutshell, each spike of the sender neuron triggers a transient change in excitability of the receiver neuron. This can either increase or decrease the tendency of the receiver neuron to fire an action potential. We speak of “excitation” and “inhibition”, respectively. A hallmark of normal brain function is a robust balance of excitation and inhibition. Brain diseases, in contrast, are often linked with a marked imbalance of these two forces [38]. A whole zoo of mathematical models exist to mimic the dynamical processes of synaptic interaction in networks of spiking neurons. Most of them are difficult to analyze, due to the discrete nature of pulse trains and an essential nonlinearity in the spike generation mechanism. From a mathematical point of view, the Hawkes process seems like a very reasonable choice as it can directly reflect several important biological features. First, as a stochastic point process it is by construction based on pulse-coded signaling. Second, the use of interaction kernels automatically reflects the causal nature of communication in a physical system with recurrent loops, allowing only spikes in the past to influence the future evolution of the network. Third, as a multivariate process with coupled components it is able to capture and consistently describe the collective activity dynamics. The linear Hawkes process [22] has been employed as a mathematical model of networks of spiking neurons, as it can directly reflect excitatory synaptic interactions between spiking neurons using positive interaction kernels [28]. An important shortcoming of the linear model is that inhibitory synaptic interaction cannot be included in a consistent way. Negative interaction kernels are not allowed, as they might yield a negative point process intensity. To enable the study of networks with excitation-inhibition balance, however, negative interaction kernels must be admitted [35]. A nonlinear link function can fix this deficit, and effective descriptions of nonlinear biological neurons combining linear synaptic interaction with a nonlinear link function are actually well-known in computational neuroscience. They have been called cascade models [25], and from the viewpoint of statistics they lead to generalized linear models [34]. For that reason, the nonlinear Hawkes process is a natural extension of Hawkes’ classical model, greatly extending its range of applications. The specific model described in this paper can be further generalized, allowing for a different interaction kernel for each pair of nodes in the network. This includes networks with arbitrary graph topology, setting kernels to zero where a synapse is absent [35]. In brain networks, the amplitude of individual interaction kernels (“synaptic weight”) is thought to change as a result of learning. The promise for the biologist is to be able to investigate networks, the structure of which is either designed for a specific purpose, or reflects previous experience and memory. The challenge for the mathematician is to manage the lack of symmetry in these networks, and a genuinely multivariate description becomes necessary. There is an ongoing debate about the significance of scaling synaptic weights for networks of infinite size [37, 2]. As the number of neurons in brain networks is typically quite high, such limit may represent a meaningful approximation to large brains. The difficulty is that large brains also have to accommodate constraints with regard to the number of possible connections, imposed by the distance of neurons in space and by the volume occupied by fibers. So it might be more fruitful to rephrase the question a bit [1]: Which aspect of the input to a neuron from within the network is more important: Is it the mean drift, or is it the amplitude of the fluctuations? If there is a perfect balance between excitation and inhibition, the drift is zero, and the fluctuations take over and drive the neuron in a stochastic fashion. For the slightest imbalance of the inputs, however, keeping the drift bounded in very large networks forces the fluctuations to be very small, and the dynamics becomes more deterministic. This is where biophysical intuition is in line with the result of mathematical analysis presented in this paper. ## 5 Proofs We start off in Subsection 5.1 with some preliminary results on convergence of Poisson processes and convolution equations. We proceed in 5.2 with a reformulation of the multivariate linear Hawkes process using a time-change equation. Then, we prove Theorem 2 in Sebsection 5.3 and Theorem 1 in Subsection 5.4. We give the proof of Theorem 2 first because it is more involved than the proof of Theorem 1, mainly since the limit process $I$ is stochastic. ### 5.1 Preliminairies We need the following (rather standard) convergence results for Poisson processes. ###### Lemma 5.1 (Convergence of Poisson processes). Let $Y$ be a unit rate Poisson process, defined on some probability space $(\Omega,\mathcal{F},\mathbb{P})$. Then, for all $t>0$, $\displaystyle\sup_{0\leq s\leq t}\frac{1}{N}(Y(Ns)-Ns)\xrightarrow{N\to\infty}0$ (5.1) almost surely (and in $L^{2}$). In addition, $\displaystyle\Big{(}\frac{1}{\sqrt{N}}(Y(Nt)-Nt)\Big{)}_{t\geq 0}\xRightarrow{N\to\infty}B$ (5.2) for some Brownian motion $B$. We can extend $(\Omega,\mathcal{F},\mathbb{P})$ by a Brownian motion $B$ such that, for all $t\geq 0$, $\displaystyle\sup_{0\leq s\leq t}\Big{|}\frac{1}{\sqrt{N}}(Y(Ns)-Ns)-B_{s}\Big{|}\xrightarrow{N\to\infty}0$ (5.3) almost surely and in $L^{2}$. ###### Proof. For (5.1), we use Doob’s martingale inequality for the martingale $(Y(s)-s)_{s\geq 0}$ and the fact that the fourth centered moment of a Poisson random variable with parameter $\lambda$ is $3\lambda^{2}+\lambda$ in order to see $\displaystyle\mathbb{E}\Big{[}\sup_{0\leq s\leq t}\Big{(}\frac{1}{N}(Y(Ns)-Ns)\Big{)}^{4}\Big{]}\leq\Big{(}\frac{4}{3}\Big{)}^{4}\mathbb{E}\Big{[}\Big{(}\frac{1}{N}(Y(Nt)-Nt)\Big{)}^{4}\Big{]}={\color[rgb]{0,0,0}\Big{(}\frac{4}{3N}\Big{)}^{4}}\big{(}3(Nt)^{2}+Nt\big{)},$ which is summable. Using Borel-Cantelli, almost-sure convergence follows. Next, (5.2) follows by an application of Donsker’s theorem ([26] Ch VII, Cor. 3.11). By Skorohod’s Theorem, we can extend our probability space such that this convergence is almost surely with respect to Skorohod distance, and by continuity of $B$ it is equivalent to uniform convergence, i.e. (5.3) holds almost surely for all $t\geq 0$. For the $L^{2}$-convergence in (5.3), it suffices to show the uniform integrability of $\sup_{0\leq s\leq t}\Big{(}\frac{1}{\sqrt{N}}(Y(Ns)-Ns)-B_{s}\Big{)}^{2}$. Using Doob’s and Minkovski’s inequalities, we compute $\displaystyle\mathbb{E}\Big{[}\sup_{0\leq s\leq t}\Big{(}\frac{1}{\sqrt{N}}(Y(Ns)-Ns)-B_{s}\Big{)}^{4}\Big{]}^{1/4}$ $\displaystyle\leq\frac{4}{3}\mathbb{E}\Big{[}\Big{(}\frac{1}{\sqrt{N}}(Y(Nt)-Nt)-B_{t}\Big{)}^{4}\Big{]}^{1/4}$ $\displaystyle\leq\frac{4}{3}\Big{(}\mathbb{E}\Big{[}\Big{(}\frac{1}{\sqrt{N}}(Y(Nt)-Nt)\Big{)}^{4}\Big{]}^{1/4}+\mathbb{[}(B_{t})^{4}]^{1/4}\Big{)}$ $\displaystyle=\frac{4}{3}((3t^{2}+t/N)^{1/4}+(3t^{2})^{1/4})$ showing the desired uniform integrability and $L^{2}$-convergence follows. ∎ In order to bound the value of a convolution equation by its integrator we need the following ###### Lemma 5.2. Let $J$ be the sum of an Itô process with bounded coefficients and a càdlàg pure-jump process, and let $\varphi\in\mathcal{C}_{b}^{1}([0,\infty))$. Then $\displaystyle\sup_{0\leq s\leq t}\Big{(}\int_{0}^{s}\varphi(s-r)dJ_{r}\Big{)}^{2}\leq C(\varphi,t)\cdot\sup_{0\leq s\leq t}J_{s}^{2}.$ (5.4) ###### Proof. Wlog, we have $J_{0}=0$. By [3, Theorem 4.A] and Fubini’s theorem for Lebesgue-integrals we can apply the Stochastic Fubini Theorem to $J$, hence $\displaystyle\sup_{0\leq s\leq t}\Big{(}\int_{0}^{s}\varphi(s-r)dJ_{r}\Big{)}^{2}$ $\displaystyle=\sup_{0\leq s\leq t}\Big{(}\int_{0}^{s}\Big{(}\varphi(0)+\int_{r}^{s}\varphi^{\prime}(s-u)du\Big{)}dJ_{r}\Big{)}^{2}$ $\displaystyle\leq 2\,||\varphi||^{2}\cdot\sup_{0\leq s\leq t}J_{\color[rgb]{0,0,0}s}^{2}+2\sup_{0\leq s\leq t}\Big{(}\int_{0}^{s}\int_{0}^{u}\varphi^{\prime}(s-u)dJ_{r}du\Big{)}^{2}$ $\displaystyle=2\,||\varphi||^{2}\cdot\sup_{0\leq s\leq t}J_{s}^{2}+2\sup_{0\leq s\leq t}\Big{(}\int_{0}^{s}\varphi^{\prime}(s-u)J_{u}du\Big{)}^{2}$ $\displaystyle\leq 2\,(||\varphi||^{2}+t||\varphi^{\prime}||^{2})\cdot\sup_{0\leq s\leq t}J_{s}^{2}.$ ∎ ### 5.2 Reformulation of Hawkes processes Alternative descriptions of non-linear Hawkes processes have been given in the literature. Above all, the construction using a Poisson random measures is widely used; see e.g. Proposition 3 in [12]. Here, we rely on the following construction using time-change equations (see e.g. Chapter 6 of [16]), which we give here without proof. ###### Lemma 5.3. Let $\mathbb{G}=(\mathbb{V},\mathbb{E})$, $(\varphi_{ji})_{ji\in\mathbb{E}}$ and $(h_{i})_{i\in\mathbb{V}}$ be as in Definition 2.1, and let Assumption 2.2 hold. A point process $Z=(Z^{i})_{i\in\mathbb{V}}$ is a multivariate, non- linear Hawkes process with interaction kernels $(\varphi_{ji})_{ji\in\mathbb{E}}$ and transfer functions $(h_{i})_{i\in\mathbb{V}}$, if and only if it is the weak solution of the time-change equations $\displaystyle Z_{t}^{i}=Y_{i}\Big{(}\int_{0}^{t}\lambda_{s}^{i}ds\Big{)}=Y_{i}\Big{(}\int_{0}^{t}h_{i}\Big{(}\sum_{j:ji\in\mathbb{E}}\int_{0}^{s-}\varphi_{ji}(s-u)dZ_{u}^{j}\Big{)}ds\Big{)},$ (5.5) where $(Y_{i})_{i\in\mathbb{V}}$ is a family of independent unit rate Poisson processes. Under Assumption 2.4, the time-change equations from this lemma read, with independent unit rate Poisson processes $Y_{1},...,Y_{N}$, $\displaystyle Z_{t}^{N,i}$ $\displaystyle=Y_{i}\Big{(}\int_{0}^{t}h\Big{(}\sum_{j=1}^{N}\int_{0}^{s-}\theta_{N}U_{j}\varphi(s-u)dZ_{u}^{N,j}\Big{)}ds\Big{)}.$ (5.6) We rewrite this as $\displaystyle Z_{t}^{N,i}$ $\displaystyle=Y_{i}\Big{(}\int_{0}^{t}h(I_{s}^{N})ds\Big{)}$ (5.7) with $\displaystyle I_{s}^{N}$ $\displaystyle=\int_{0}^{s-}\varphi(s-u)dJ_{u}^{N}\quad\text{ and }\quad J_{u}^{N}=\theta_{N}\sum_{j=1}^{N}U_{j}Z_{u}^{N,j}.$ (5.8) Now, we rewrite $J^{N}$ further. Observe that $\sum_{j:U_{j}=1}Y_{j}$ is a Poisson process with rate $\tfrac{1}{2}\sum_{j=1}^{N}(U_{j}+1)=pN+\tfrac{1}{2}\sqrt{N}W_{N}$ (recall $W_{N}$ from (2.2)) and $\sum_{j:U_{j}=-1}Y_{j}$ is a Poisson process with rate $(1-p)N-\tfrac{1}{2}\sqrt{N}W_{N}$. Hence we have that, for two independent unit rate Poisson processes $Y^{+}$ and $Y^{-}$, $\displaystyle J_{t}^{N}$ $\displaystyle=\theta_{N}\sum_{j=1}^{N}U_{j}Y_{j}\Big{(}\int_{0}^{t}h(I_{s}^{N})ds\Big{)}$ (5.9) $\displaystyle=\theta_{N}Y^{+}\Big{(}\big{(}pN+\tfrac{1}{2}\sqrt{N}W_{N}\big{)}\int_{0}^{t}h(I_{s}^{N})ds\Big{)}-\theta_{N}Y^{-}\Big{(}\big{(}(1-p)N-\tfrac{1}{2}\sqrt{N}W_{N}\big{)}\int_{0}^{t}h(I_{s}^{N})ds\Big{)}.$ We extend our probability space such that the rescaled Poisson processes $Y^{+},Y^{-}$ converge to Brownian motions $B^{+},B^{-}$ almost surely (as in (5.3)). Consider the Brownian motion $\hat{B}=(\hat{B}(t))_{t\geq 0}$ with $\hat{B}_{t}=B^{+}_{pt}-B^{-}_{(1-p)t}$. Since $\Big{(}\hat{B}\Big{(}\int_{0}^{t}h(I_{s}^{N})ds\Big{)}\Big{)}_{t\geq 0}$ is a continuous martingale, we can extend the probability space by another Brownian motion $B^{N}$, such that $\displaystyle\hat{B}\Big{(}\int_{0}^{t}h(I_{s}^{N})ds\Big{)}=\int_{0}^{t}\sqrt{h(I_{s}^{N})}dB_{s}^{N}\quad\text{for }t\geq 0,$ (5.10) by [29, Theorem 16.10]. ### 5.3 Proof of Theorem 2 First, strong existence and uniqueness for (3.10) follows from [3, Theorem 3.A]. For the convergence result, we work on a probability space where the convergence $W_{N}\to W$ from (2.2) holds almost surely. We will show in Steps 1 and 2 convergence of $J^{N}$ to $J$ with $\displaystyle J_{t}=W\int_{0}^{t}h(I_{s})ds+\int_{0}^{t}\sqrt{h(I_{s})}dB_{s}$ (5.11) and from $I^{N}$ to $I$ from (3.10) with $w$ replaced by $W$. Step 3 then gives the desired convergence of $(I^{N},Z^{N,1},...,Z^{N,n})$. We use processes $\widetilde{J}^{N}$ and $\widetilde{I}^{N}$ which are unique strong solutions of $\displaystyle\widetilde{J}^{N}_{t}$ $\displaystyle=W_{N}\int_{0}^{t}h\big{(}\widetilde{I}_{s}^{N}\big{)}ds+\int_{0}^{t}\sqrt{h(\widetilde{I}_{s}^{N})}dB_{s}^{N},$ $\displaystyle\widetilde{I}_{s}^{N}$ $\displaystyle=\int_{0}^{s}\varphi(s-r)d\widetilde{J}^{N}_{r},$ where $B^{N}$ is the Brownian motion defined by (5.10). Again, strong existence and uniqueness follows from [3, Theorem 3.A]. In Step 1, we will show that $J^{N}-\widetilde{J}^{N}\to 0$ and $I^{N}-\widetilde{I}^{N}\to 0$ uniformly on compact sets in $L^{2}$, conditional on $(U_{i})_{i\in\mathbb{N}}$. In Step 2, we show the remaining conditional convergence $\widetilde{I}^{N}-I\to 0$ and conclude $I^{N}-I\to 0$ in probability. Step 1: Convergence $J^{N}-\widetilde{J}^{N}\to 0$ and $I^{N}-\widetilde{I}^{N}\to 0$. Recall $Y^{+}$ and $Y^{-}$ from Section 5.2. For $\theta_{N}=\tfrac{1}{\sqrt{N}}$ and $p=\tfrac{1}{2}$ in (5.9) we get that, $J_{t}^{N}=\frac{1}{\sqrt{N}}Y^{+}\Big{(}\tfrac{N+\sqrt{N}W_{N}}{2}\int_{0}^{t}h(I_{s}^{N})ds\Big{)}-\frac{1}{\sqrt{N}}Y^{-}\Big{(}\tfrac{N-\sqrt{N}W_{N}}{2}\int_{0}^{t}h(I_{s}^{N})ds\Big{)}.$ We write $\displaystyle J_{t}^{N}-\widetilde{J}_{t}^{N}$ $\displaystyle=A_{t}^{N}+W_{N}C_{t}^{N}+D_{t}^{N}\quad\text{with}$ (5.12) $\displaystyle A_{t}^{N}$ $\displaystyle=\frac{1}{\sqrt{N}}Y^{+}\Big{(}\frac{N+\sqrt{N}W_{N}}{2}\int_{0}^{t}h(I_{s}^{N})ds\Big{)}-\frac{1}{\sqrt{N}}Y^{-}\Big{(}\frac{N-\sqrt{N}W_{N}}{2}\int_{0}^{t}h(I_{s}^{N})ds\Big{)}$ $\displaystyle\qquad\qquad\qquad\qquad\qquad\qquad\qquad\qquad- W_{N}\int_{0}^{t}h(I_{s}^{N})ds-\hat{B}\Big{(}\int_{0}^{t}h(I_{s}^{N})ds\Big{)},$ $\displaystyle C_{t}^{N}$ $\displaystyle=\int_{0}^{t}h(I_{s}^{N})ds-\int_{0}^{t}h(\widetilde{I}_{s}^{N})ds,$ $\displaystyle D_{t}^{N}$ $\displaystyle=\int_{0}^{t}\sqrt{h(I_{s}^{N})}dB^{N}_{s}-\int_{0}^{t}\sqrt{h(\widetilde{I}_{s}^{N})}dB^{N}_{s}.$ We make the disclaimer, that we will use $C>0$ for a variable, which only depends on $t,h,\varphi$, and which might change from line to line. We write $\mathbb{E}_{U}$ for the conditional expectation with respect to $U=(U_{i})_{i\in\mathbb{N}}$. The function $h$ is bounded, therefore, for some unit rate Poisson process $Y$, $\displaystyle\mathbb{E}{{}_{U}}[\sup_{0\leq s\leq t}(A_{s}^{N})^{2}]$ $\displaystyle\leq\mathbb{E}{{}_{U}}\Big{[}\sup_{0\leq s\leq t||h||}\Big{(}\frac{1}{\sqrt{N}}Y^{+}\Big{(}\frac{N+\sqrt{N}W_{N}}{2}s\Big{)}-\frac{1}{\sqrt{N}}Y^{-}\Big{(}\frac{N-\sqrt{N}W_{N}}{2}s\Big{)}$ (5.13) $\displaystyle\qquad\qquad\qquad\qquad\qquad\qquad\qquad\qquad\qquad\qquad\qquad- W_{N}s-\hat{B}(s)\Big{)}^{2}\Big{]}$ $\displaystyle\leq C\cdot\Big{(}\mathbb{E}{{}_{U}}\Big{[}\sup_{0\leq s\leq t||h||}\Big{(}\frac{1}{\sqrt{N}}(Y^{+}(Ns/2)-Ns/2)-B^{+}(s/2)\Big{)}^{2}\Big{]}$ $\displaystyle\qquad+\mathbb{E}{{}_{U}}\Big{[}\sup_{0\leq s\leq t||h||}\Big{(}\frac{1}{\sqrt{N}}(Y^{-}(Ns/2)-Ns/2)-B^{-}(s/2)\Big{)}^{2}\Big{]}$ $\displaystyle\qquad+\mathbb{E}{{}_{U}}\Big{[}\sup_{0\leq s\leq|W_{N}|t||h||}\Big{(}\frac{1}{\sqrt{N}}(Y(\sqrt{N}s))-s\Big{)}^{2}\Big{]}\Big{)}$ $\displaystyle{\color[rgb]{0,0,0}\leq o(1)+C\cdot\mathbb{E}_{U}\Big{[}\Big{(}\frac{1}{\sqrt{N}}(Y(\sqrt{N}|W_{N}|t||h||s))-|W_{N}|t||h||s\Big{)}^{2}\Big{]}}$ $\displaystyle{\color[rgb]{0,0,0}\leq o(1)+C\frac{1}{\sqrt{N}}W_{N}}\xrightarrow{N\to\infty}0$ almost surely, by Lemma 5.1 and Doob’s inequality. Then, using Jensen, $\displaystyle\mathbb{E}{{}_{U}}[\sup_{0\leq s\leq t}(W_{N}C_{s}^{N})^{2}]$ $\displaystyle{\color[rgb]{0,0,0}\leq\sup_{N}W_{N}^{2}}\mathbb{E}{{}_{U}}\Big{[}\sup_{0\leq s\leq t}\Big{(}\int_{0}^{s}h(I_{r}^{N})-h(\widetilde{I}_{r}^{N})dr\Big{)}^{2}\Big{]}$ $\displaystyle\leq{\color[rgb]{0,0,0}\sup_{N}W_{N}^{2}}th_{Lip}^{2}\int_{0}^{t}\mathbb{E}{{}_{U}}[(I_{s}^{N}-\widetilde{I}_{s}^{N})^{2}]ds$ Note that $\sup_{N}W_{N}^{2}<\infty$, as $W_{N}$ converges to $W$, which is finite almost surely. By Doob’s inequality and Ito’s isometry, $\displaystyle\mathbb{E}{{}_{U}}[\sup_{0\leq s\leq t}(D_{s}^{N})^{2}]$ $\displaystyle\leq C\cdot\mathbb{E}{{}_{U}}[(D_{t}^{N})^{2}]=C\cdot\mathbb{E}{{}_{U}}\Big{[}\Big{(}\int_{0}^{t}\sqrt{h(I_{s}^{N})}dB^{N}_{s}-\int_{0}^{t}\sqrt{h(\widetilde{I}_{s}^{N})}dB^{N}_{s}\Big{)}^{2}\Big{]}$ $\displaystyle=C\cdot\mathbb{E}{{}_{U}}\Big{[}\int_{0}^{t}\big{(}\sqrt{h(I_{s}^{N})}-\sqrt{h(\widetilde{I}_{s}^{N})}\big{)}^{2}ds\Big{]}$ $\displaystyle\leq C\cdot\int_{0}^{t}\mathbb{E}{{}_{U}}[(I_{s}^{N}-\widetilde{I}_{s}^{N})^{2}]ds.$ In order to bound $\sup_{0\leq s\leq t}(W_{N}C_{s}^{N})^{2}$ and $\sup_{0\leq s\leq t}(D_{s}^{N})^{2}$ by $\sup_{0\leq s\leq t}(J_{s}^{N}-\widetilde{J}_{s}^{N})^{2}$, apply Lemma 5.2 to $J_{s}^{N}-\widetilde{J}_{s}^{N}$. We obtain $\displaystyle\mathbb{E}{{}_{U}}\big{[}\sup_{s\leq t}(J_{s}^{N}-\widetilde{J}_{s}^{N})^{2}\big{]}$ $\displaystyle\leq C\cdot\mathbb{E}{{}_{U}}[\sup_{0\leq s\leq t}(A_{s}^{N})^{2}]+C\cdot\int_{0}^{t}\mathbb{E}{{}_{U}}[(I_{s}^{N}-\widetilde{I}_{s}^{N})^{2}]ds$ $\displaystyle\leq C\cdot\mathbb{E}{{}_{U}}[\sup_{0\leq s\leq t}(A_{s}^{N})^{2}]+C\cdot\int_{0}^{t}\mathbb{E}{{}_{U}}[\sup_{r\leq s}(J_{r}^{N}-\widetilde{J}_{r}^{N})^{2}]ds.$ By the Gronwall inequality and (5.13), we conclude that, for all $t\geq 0$, $\displaystyle\mathbb{E}{{}_{U}}\big{[}\sup_{s\leq t}(J_{s}^{N}-\widetilde{J}_{s}^{N})^{2}\big{]}$ $\displaystyle\xrightarrow{N\to\infty}0$ and with Lemma 5.2 also $\displaystyle\mathbb{E}{{}_{U}}\big{[}\sup_{s\leq t}(I_{s}^{N}-\widetilde{I}_{s}^{N})^{2}\big{]}$ $\displaystyle\xrightarrow{N\to\infty}0,$ almost surely. Step 2: Convergence $\widetilde{I}^{N}-I\to 0$ Since $B^{N}$ is a Brownian motion, we have that, in distribution, $\displaystyle\widetilde{I}_{t}^{N}$ $\displaystyle=W_{N}\int_{0}^{t}\varphi(t-s)h(\widetilde{I}_{s}^{N})ds+\int_{0}^{t}\varphi(t-s)\sqrt{h(\widetilde{I}_{s}^{N})}dB_{s}^{N},$ $\displaystyle I_{t}$ $\displaystyle=W\int_{0}^{t}\varphi(t-s)h(I_{s})ds+\int_{0}^{t}\varphi(t-s)\sqrt{h(\widetilde{I}_{s})}dB_{s}^{N}.$ Therefore, on this probability space, $\displaystyle\widetilde{I}_{t}^{N}-I_{t}$ $\displaystyle=E_{t}^{N}+F_{t}^{N}+G_{t}^{n}\text{ with}$ $\displaystyle E_{t}^{n}$ $\displaystyle:=(W_{N}-W)\int_{0}^{t}\varphi(t-s)h(I_{s})ds,$ $\displaystyle F_{t}^{N}$ $\displaystyle:=W_{N}\int_{0}^{t}\varphi(t-s)(h(\widetilde{I}_{s}^{N})-h(I_{s}))ds,$ $\displaystyle G_{t}^{N}$ $\displaystyle:=\int_{0}^{t}\varphi(t-s)\Big{(}\sqrt{h(\widetilde{I}_{s}^{N})}-\sqrt{h(I_{s})}\Big{)}dB_{s}^{N}.$ Since, again using some constant $C<\infty$, which can change from line to line, $\displaystyle\mathbb{E}{{}_{U}}[\sup_{0\leq s\leq t}(E_{s}^{N})^{2}]$ $\displaystyle\leq C\cdot{\color[rgb]{0,0,0}(W_{N}-W)^{2}}\xrightarrow{N\to\infty}0,$ $\displaystyle\mathbb{E}{{}_{U}}[\sup_{0\leq s\leq t}(F_{s}^{N})^{2}]$ $\displaystyle\leq C\cdot{\color[rgb]{0,0,0}\sup_{N}W_{N}^{2}}\int_{0}^{t}\mathbb{E}{{}_{U}}[\sup_{0\leq r\leq s}(\widetilde{I}_{r}^{N}-I_{r})^{2}]ds,$ $\displaystyle\mathbb{E}{{}_{U}}[\sup_{0\leq s\leq t}(G_{s}^{N})^{2}]$ $\displaystyle\leq C\cdot\int_{0}^{t}\mathbb{E}{{}_{U}}[\sup_{0\leq r\leq s}(\widetilde{I}_{r}^{N}-I_{r})^{2}]ds,$ we find that $\displaystyle\mathbb{E}{{}_{U}}[\sup_{0\leq s\leq t}(\widetilde{I}_{s}^{N}-I_{s})^{2}]$ $\displaystyle\leq C\cdot{\color[rgb]{0,0,0}(W_{N}-W)^{2}}+C\cdot\int_{0}^{t}\mathbb{E}{{}_{U}}[\sup_{0\leq r\leq s}(\widetilde{I}_{r}^{N}-I_{r})^{2}]ds,$ and the Gronwall inequality again implies that $\mathbb{E}{{}_{U}}[\sup_{0\leq s\leq t}(\widetilde{I}_{s}^{N}-I_{s})^{2}]\xrightarrow{N\to\infty}0$, almost surely. Combining Step 1 and Step 2 we obtain $\mathbb{E}_{U}[\sup_{0\leq s\leq t}(I_{s}^{N}-I_{s})^{2}]\xrightarrow{N\to\infty}0$. We deduce $\mathbb{E}_{U}[1\wedge\sup_{0\leq s\leq t}(I_{s}^{N}-I_{s})^{2}]\xrightarrow{N\to\infty}0$ and, by dominated convergence, $\mathbb{E}[1\wedge\sup_{0\leq s\leq t}(I_{s}^{N}-I_{s})^{2}]\xrightarrow{N\to\infty}0$. Hence we have shown that $\sup_{0\leq s\leq t}(I_{s}^{N}-I_{s})^{2}\to 0$ in probability. Step 3: Convergence of $(I^{N},Z^{N,1},...,Z^{N,n})$ The predictable quadratic variation of the compensated process $Z^{N,i}_{t}-\int_{0}^{t}h(I_{s}^{N})ds$ is given by $\langle Z^{N,i}-\int_{0}^{\cdot}h(I_{s}^{N})ds\rangle_{t}=\int_{0}^{t}h(I_{s}^{N})ds$. By [26][Chapter VI, Corollary 3.33, Theorem 4.13], tightness of $(Z^{N,1},...,Z^{N,n})$ follows from convergence of $\sum_{i\leq n}\langle Z^{N,i}-\int_{0}^{\cdot}h(I_{s}^{N})ds\rangle_{t}=n\int_{0}^{t}h(I_{s}^{N})ds$. We already know that $I^{N}\Rightarrow I$, hence $(I^{N},Z^{N,1},...,Z^{N,n})$ is tight. To identify the limit, let us try to extend the state space such that the Hawkes process becomes Markovian. Recall that each $Z^{N,i}$ jumps at time $t$ at rate $h(I^{N}(t))$ with $I^{N}(t)=\int_{0}^{t}\varphi(t-s)dJ^{N}_{s}$ and $J_{s}^{N}$ as given in (5.8). The process $J^{N}$ jumps at time $t$ from $x$ to $x+\theta_{N}$ at rate $h\Big{(}\int_{0}^{t}\varphi(t-s)dJ^{N}_{s}\Big{)}\cdot u^{+}$ with $u^{+}:=\sum_{j}1_{U_{j}=1}$ and to $x-\theta_{N}$ at rate $h\Big{(}\int_{0}^{t}\varphi(t-s)dJ^{N}_{s}\Big{)}\cdot u^{-}$ with $u^{-}=\sum_{j}1_{U_{j}=-1}=N-u^{+}$. Note that the process $J^{N}$ satisfies, for smooth $f$, $\displaystyle\frac{\mathbb{E}\Big{[}f(J^{N}_{t+\varepsilon})-f(J^{N}_{t})|\mathcal{F}_{t}\Big{]}}{\varepsilon}$ $\displaystyle=h(I_{t}^{N})u^{+}\Big{[}f(J^{N}_{t}+\theta_{N})-f(J^{N}_{t})\Big{]}+h(I_{t}^{N})u^{-}\Big{[}f(J^{N}_{t}-\theta_{N})-f(J^{N}_{t})\Big{]}+o(\varepsilon)$ $\displaystyle=f^{\prime}(J^{N}_{t})\cdot\Big{(}h(I_{t}^{N})\underbrace{(u^{+}_{N}-u^{-}_{N})\theta_{N}}_{=W_{N}}\Big{)}+\tfrac{1}{2}f^{\prime\prime}(J^{N}_{t})\cdot h(I_{t}^{N})\underbrace{(u^{+}_{N}+u^{-}_{N})\theta_{N}^{2}}_{=1}+o(\varepsilon)$ $\displaystyle=f^{\prime}(J^{N}_{t})\cdot h(I_{t}^{N})W_{N}+\tfrac{1}{2}f^{\prime\prime}(J^{N}_{t})\cdot h(I_{t}^{N})+o(\varepsilon).$ We obtain a similar expression if we ignore the jumps of $Z^{N,i},i=1,...,n$, i.e. if we replace $J^{N}$ by $J^{N(n)}=\theta_{N}\sum_{j>n}U_{j}Z^{N,j}$. In this case we have to replace $W_{N}$ by $W_{N(n)}=\theta_{N}\sum_{j>n}U_{j}$ and $(u^{+}_{N}+u^{-}_{N})\theta_{N}^{2}$ by $(N-n)/N$. In the limit $N\to\infty$ we obtain the same expression, as $W_{N(n)}\to W$ and $(N-n)/N\to 1$. From now on write $J_{[0,t]}$ for the path of a process $J$ on the interval $[0,t]$ and $\displaystyle G_{J}^{N(n)}f(J_{[0,t]}^{\color[rgb]{0,0,0}N})$ $\displaystyle=f^{\prime}(J_{t}^{\color[rgb]{0,0,0}N})\cdot h(I_{t}^{\color[rgb]{0,0,0}N})W_{N(n)}+\tfrac{N-n}{2N}f^{\prime\prime}(J_{t}^{\color[rgb]{0,0,0}N})\cdot h(I_{t}^{\color[rgb]{0,0,0}N})$ $\displaystyle G_{J}f(J_{[0,t]})$ $\displaystyle=f^{\prime}(J_{t})\cdot h(I_{t})W+\tfrac{1}{2}f^{\prime\prime}(J_{t})\cdot h(I_{t})$ The generator $G^{N}$ of the Markov process $\big{(}J^{N}_{[0,t]},Z^{N,1}_{t},...,Z^{N,n}_{t}\big{)}$ with domain $\mathcal{D}$ consisting of functions of the form $f_{0,...,n}(J_{[0,t]},k_{1},...,k_{n})=f_{0}(J_{t})f_{1}(k_{1})...f_{n}(k_{n})$, where we suppose the functions $f_{1},...,f_{n}$ to be bounded functions on $\mathbb{N}$, and $f_{0}$ to be a smooth function on $\mathbb{R}$ with bounded second derivative, is given by $\displaystyle G^{N}f_{0,..,n}(J_{[0,t]}^{\color[rgb]{0,0,0}N},Z^{N,1}_{t}$ $\displaystyle,...,Z^{N,n}_{t})=G^{N(n)}_{J}f_{0}(J_{[0,t]}^{\color[rgb]{0,0,0}N})\prod_{i=1}^{n}f_{i}(Z^{N,i}_{t})$ $\displaystyle+h(I_{t})\sum_{j=1}^{n}\big{(}f_{0}(J_{t}+\theta_{N}U_{j})f_{j}(Z^{N,j}_{t}+1)-f_{0}(J_{t})f_{j}(Z^{N,j}_{t})\big{)}\prod_{i\neq j}f_{i}(Z^{N,i}_{t}).$ By [16][Chapter 4, Proposition 1.7], $\displaystyle f_{0}(J^{N}_{t})\prod_{i=1}^{n}f_{i}(Z^{N,i}_{t})-\int_{0}^{t}G^{N}f_{0}...f_{n}(J_{[0,s]}^{\color[rgb]{0,0,0}N},Z^{N,1}_{s},...,Z^{N,n}_{s})ds$ is a martingale with respect to the filtration generated by $J,Z^{N,i}$. It is equivalent that $\displaystyle\mathbb{E}\Big{[}\Big{(}f_{0}(J^{N}_{t})\prod_{i=1}^{n}f_{i}(Z^{N,i}_{t})-f_{0}(J^{N}_{s})\prod_{i=1}^{n}f_{i}(Z^{N,i}_{s})-\int_{s}^{t}G^{N}f_{0}...f_{n}(J_{[0,r]}^{\color[rgb]{0,0,0}N},Z^{N,1}_{r},...,Z^{N,n}_{r})ds\Big{)}$ $\displaystyle\prod_{j=1}^{m}g_{j}(J_{[0,t_{j}]}^{\color[rgb]{0,0,0}N},Z^{N,1}_{t_{j}},...,Z^{N,n}_{t_{j}})\Big{]}=0,$ (5.14) for all $m\in\mathbb{N},0\leq t_{1}\leq...\leq t_{m}\leq s<t$ and continuous functions $g_{j}$. To identify the limit, assume $\big{(}J^{N},Z^{N,1},...,Z^{N,n}\big{)}\Rightarrow\big{(}\bar{J},\bar{Z}^{1},...,\bar{Z}^{n}\big{)}$ in distribution with respect to Skorohod distance. By continuity of projection, $J^{N}\Rightarrow\bar{J}$, hence $\bar{J}=J$. In Step 1 and 2 we have already shown the stronger convergence $J^{N}\to J$ and $I^{N}\to I$ with respect to uniform topology. As $G^{N}f$ converges to some $Gf$, $\theta_{N}\to 0$ and the evaluation at timepoints $t$ is continuous with respect to uniform topology, we can pass to the limit in (5.14) and obtain that $\displaystyle f_{0}(J_{t})\prod_{i=1}^{n}f_{i}(\bar{Z}^{i}_{t})-\int_{0}^{t}Gf_{0,...,n}(J_{s},\bar{Z}^{1}_{s},...,\bar{Z}^{n}_{s})ds$ (5.15) is a martingale for any $f_{0,...,n}\in\mathcal{D}$, where $\displaystyle Gf_{0,..,n}(J_{[0,t]},Z^{1}_{t},...,Z^{n}_{t})=$ $\displaystyle G_{J}f_{0}(J_{[0,t]})\prod_{i=1}^{n}f_{i}(Z^{i}_{t})$ $\displaystyle+h(I_{t})f_{0}(J_{t})\sum_{j=1}^{n}\big{(}f_{j}(Z^{j}_{t}+1)-f_{j}(Z^{j}_{t})\big{)}\prod_{i\neq j}f_{i}(Z^{i}_{t}).$ It remains to read off the distribution of $(J,\bar{Z}^{1},...,\bar{Z}^{n})$ from (5.15). Choose $f_{0}=1$ and $f_{i}(k)=\exp(-u_{i}k)$ for $u_{j}\geq 0$. Then $\displaystyle 0=$ $\displaystyle\mathbb{E}\Big{[}f_{0}(J_{t})\prod_{i=1}^{n}f_{i}(\bar{Z}^{i}_{t})-\int_{0}^{t}Gf_{0}...f_{n}(J_{s},\bar{Z}^{1}_{s},...,\bar{Z}^{n}_{s})ds\Big{]}$ $\displaystyle=$ $\displaystyle\mathbb{E}\big{[}\prod_{i=1}^{n}e^{-u_{i}\bar{Z}^{i}_{t}}-\int_{0}^{t}h(I_{s})\sum_{j=1}^{n}\big{(}e^{-u_{j}}-1\big{)}\prod_{i=1}^{n}e^{-u_{i}\bar{Z}^{i}_{s}}ds\big{]}$ $\displaystyle=$ $\displaystyle\mathbb{E}\Big{[}\mathbb{E}\big{[}\prod_{i=1}^{n}e^{-u_{i}\bar{Z}^{i}_{t}}|I\big{]}-\int_{0}^{t}h(I_{s})\sum_{j=1}^{n}\big{(}e^{-u_{j}}-1\big{)}\mathbb{E}\big{[}{\color[rgb]{0,0,0}\prod_{i=1}^{n}}e^{-u_{i}\bar{Z}^{i}_{s}}|I\big{]}ds\Big{]},$ where, in the last equality, we have interchanged the $ds$-integral and conditional expectation with respect to $I$, and used the $I$-measurability of $h(I)$. The function in brackets is nonnegative, hence $\displaystyle\mathbb{E}\Big{[}\prod_{i=1}^{n}e^{-u_{i}\bar{Z}^{i}_{t}}|I\Big{]}=\prod_{j=1}^{n}\exp\Big{(}(e^{-u_{j}}-1)\int_{0}^{t}h(I_{s})ds\Big{)}.$ As claimed, given $I$ the processes $\bar{Z}^{i}$ are independent poisson processes with intensity at time $t$ given by $h(I_{t})$. ### 5.4 Proof of Theorem 1 Proof of 1.: Define $\displaystyle J_{t}=(2p-1)\int_{0}^{t}h(I_{s})ds,\text{ such that }I_{t}=\int_{0}^{t}\varphi(t-s)dJ_{s}.$ (5.16) For $\theta_{N}=\tfrac{1}{N}$ in (5.9) we get that, $\displaystyle J_{t}^{N}-J_{t}$ $\displaystyle=\frac{1}{N}\Big{(}Y^{+}\Big{(}\big{(}pN+\tfrac{1}{2}\sqrt{N}W_{N}\big{)}\int_{0}^{t}h(I_{s}^{N})ds\Big{)}-Y^{+}\Big{(}pN\int_{0}^{t}h(I_{s}^{N})ds\Big{)}\Big{)}$ $\displaystyle\quad-\frac{1}{N}\Big{(}Y^{-}\Big{(}\big{(}(1-p)N-\tfrac{1}{2}\sqrt{N}W_{N}\big{)}\int_{0}^{t}h(I_{s}^{N})ds\Big{)}-Y^{-}\Big{(}(1-p)N\int_{0}^{t}h(I_{s}^{N})ds\Big{)}\Big{)}$ $\displaystyle\qquad\qquad+\frac{1}{N}Y^{+}\Big{(}pN\int_{0}^{t}h(I_{s}^{N})ds\Big{)}-p\int_{0}^{t}h(I_{s}^{N})ds$ $\displaystyle\qquad\qquad-\frac{1}{N}Y^{-}\Big{(}(1-p)N\int_{0}^{t}h(I_{s}^{N})ds\Big{)}+(1-p)\int_{0}^{t}h(I_{s}^{N})ds$ $\displaystyle\qquad\qquad\qquad\qquad\qquad\qquad\qquad+(2p-1)\int_{0}^{t}h(I_{s}^{N})-h(I_{s})ds,$ so $\displaystyle\sup_{0\leq s\leq t}(J_{s}^{N}-J_{s})^{2}$ $\displaystyle\leq\varepsilon_{N}+C\cdot\Big{(}\sup_{0\leq s\leq t||h||}\Big{(}\frac{1}{N}Y^{+}\Big{(}pNs\Big{)}-ps\Big{)}^{2}+\Big{(}\frac{1}{N}Y^{-}\Big{(}(1-p)Ns\Big{)}-(1-p)s\Big{)}^{2}\Big{)}$ $\displaystyle\qquad\qquad\qquad+(2p-1)^{2}h_{Lip}\cdot\sup_{0\leq s\leq t}\int_{0}^{s}(I_{r}^{N}-I_{r})^{2})ds$ $\displaystyle\qquad\qquad\qquad\qquad\qquad\qquad\leq\varepsilon_{N}+C\cdot\int_{0}^{t}\sup_{0\leq r\leq s}(I_{r}^{N}-I_{r})^{2}ds$ (5.17) with $\varepsilon_{N}\xrightarrow{N\to\infty}0$ (by Lemma 5.1). By the Gronwall inequality and Lemma 5.2 we can conclude $\displaystyle\sup_{0\leq s\leq t}\Big{(}J_{s}^{N}-J_{s}\Big{)}^{2}\xrightarrow{N\to\infty}0\qquad\text{and}\qquad\sup_{0\leq s\leq t}\Big{(}I_{s}^{N}-I_{s}\Big{)}^{2}\xrightarrow{N\to\infty}0.$ (5.18) Proof of 2.: Define $\bar{Z}^{i}_{t}=Y_{i}\big{(}\int_{0}^{t}h(I_{s})ds\big{)}$, where $Y_{1},...,Y_{N}$ are independent Poisson process as in (5.7). Fix $\omega\in\Omega$, such that $\sup_{0\leq s\leq t}|I_{s}^{N}(\omega)-I_{s}|\rightarrow 0$. As $h$ is Lipschitz, $\int_{0}^{t}h(I_{s}^{N}(\omega))ds\rightarrow\int_{0}^{t}h(I_{s})ds$, hence for any point of continuity of $t\mapsto\bar{Z}^{i}_{t}(\omega)$ we can conclude $Z^{i}_{t}(\omega)\rightarrow\bar{Z}^{i}_{t}(\omega)$. As the points of continuity are dense in $[0,\infty)$, convergence in Skorohod-distance follows from [26][Theorem 2.15 c)(ii)]. Proof of 3.: First, strong existence and uniqueness for (3.3) follows from [3]. The proof proceeds similarly to the proof of Theorem 2, Step 1 and 2. Wlog, we assume that $W_{N}\to W$ almost surely and in $L^{2}$. Recall the pair $(G,K)$ from (3.3) (for some Brownian motion $B$). Further, we define the auxiliary processes $\widetilde{G}^{N},\widetilde{K}^{N}$ by (recall from (5.10) the Definition of the Brownian motion $B^{N}$) $\displaystyle\widetilde{K}^{N}_{t}$ $\displaystyle=\int_{0}^{t}\varphi(t-s)d\widetilde{G}^{N}_{s}$ (5.19) $\displaystyle\widetilde{G}^{N}_{t}$ $\displaystyle=\int_{0}^{t}Wh(I_{s})+(2p-1)h^{\prime}(I_{s})\widetilde{K}^{N}(s)ds+\int_{0}^{t}\sqrt{h(I_{s})}dB^{N}_{s}.$ In our proof, we will first show convergence of $\sqrt{N}(J^{N}-J)-\widetilde{G}^{N}\rightarrow 0$ and $\sqrt{N}(I^{N}-I)-\widetilde{K}^{N}\rightarrow 0$. We then replace $B^{N}$ by $B$ and find $\widetilde{K}^{N}\Rightarrow K$, since $(\widetilde{K}^{N},\widetilde{G}^{N})\sim(K,G)$. The process $\sqrt{N}(J^{N}-J)-\widetilde{G}^{N}$ is described by the following equation: $\displaystyle\sqrt{N}(J^{N}_{t}-J_{t})-\widetilde{G}^{N}_{t}=A^{N}_{t}+C^{N}_{t}+D^{N}_{t}+(2p-1)E^{N}_{t}$ with $\displaystyle A_{t}^{N}$ $\displaystyle=\frac{1}{\sqrt{N}}Y^{+}\Big{(}\Big{(}pN+\frac{\sqrt{N}W_{N}}{2}\Big{)}\int_{0}^{t}h(I_{s}^{N})ds\Big{)}-\frac{1}{\sqrt{N}}Y^{-}\Big{(}\Big{(}(1-p)N-\frac{\sqrt{N}W_{N}}{2}\Big{)}\int_{0}^{t}h(I_{s}^{N})ds\Big{)}$ $\displaystyle\qquad\qquad\qquad\qquad-\sqrt{N}(2p-1)\int_{0}^{t}h(I_{s}^{N})ds- W_{N}\int_{0}^{t}h(I_{s}^{N})ds-\hat{B}\Big{(}\int_{0}^{t}h(I_{s}^{N})ds\Big{)},$ $\displaystyle C_{t}^{N}$ $\displaystyle=W_{N}\int_{0}^{t}h(I_{s}^{N})ds-W\int_{0}^{t}h({I}_{s})ds,$ $\displaystyle D_{t}^{N}$ $\displaystyle=\int_{0}^{t}\sqrt{h(I_{s}^{N})}dB^{N}_{s}-\int_{0}^{t}\sqrt{h(I_{s})}dB^{N}_{s},$ $\displaystyle E_{t}^{N}$ $\displaystyle=\int_{0}^{t}\sqrt{N}\big{(}h(I^{N}_{s})-h(I_{s})\big{)}-h^{\prime}(I_{s})\widetilde{K}^{N}_{s}ds.$ As in the proof of Theorem 2, Step 1, we can show that $C^{N},D^{N}\to 0$ (as we already know that $I^{N}\to I$ and $W_{N}\rightarrow W$) uniformly on compact time intervals in $L^{2}$. The uniform $L^{2}$-convergence $A^{N}\to 0$ follows from Lemma 5.1. For $E^{N}$ we obtain $\displaystyle E^{N}_{t}=\int_{0}^{t}\big{(}\sqrt{N}(I^{N}_{s}-I_{s})-\widetilde{K}^{N}_{s}\big{)}h^{\prime}(I^{N}_{s})+\sqrt{N}R_{1}h(I^{N}_{s},I_{s})ds,$ where $R_{1}h(I^{N}_{s},I_{s})$ is the first order remainder in Taylor’s formula. Using the Peano form of the remainder we obtain for some $\xi^{N}_{s}$ between $I^{N}_{s}$ and $I_{s}$ $\displaystyle\Big{(}\sqrt{N}R_{1}h(I^{N}_{s},I_{s})\Big{)}^{2}$ $\displaystyle=N\Big{(}(h^{\prime}(\xi^{N}_{s})-h^{\prime}(I_{s}))^{2}(I^{N}_{s}-I_{s})^{2}\Big{)}$ $\displaystyle\leq(h^{\prime}_{Lip})^{2}N(I^{N}_{s}-I_{s})^{4}.$ Therefore $\displaystyle\mathbb{E}\Big{[}\sup_{0\leq s\leq t}\big{(}E^{N}_{s}\big{)}^{2}\Big{]}$ $\displaystyle\leq 2\mathbb{E}\Big{[}\sup_{0\leq s\leq t}\Big{(}\int_{0}^{s}\big{(}\sqrt{N}(I^{N}_{u}-I_{u})-\widetilde{K}^{N}_{u}\big{)}h^{\prime}(I^{N}_{u})du\Big{)}^{2}\Big{]}$ $\displaystyle\qquad\qquad+2\mathbb{E}\Big{[}\sup_{0\leq s\leq t}\Big{(}\int_{0}^{t}\big{(}\sqrt{N}R_{1}h(I^{N}_{u},I_{u})\big{)}du\Big{)}^{2}\Big{]}$ $\displaystyle\leq 2t||h^{\prime}||^{2}\int_{0}^{t}\mathbb{E}\Big{[}\big{(}\sqrt{N}(I^{N}_{s}-I_{s})-\widetilde{K}^{N}_{s}\big{)}^{2}\Big{]}ds$ $\displaystyle\qquad\qquad+2t^{2}(h^{\prime}_{Lip})^{2}\mathbb{E}\Big{[}\sup_{0\leq s\leq t}N(I^{N}_{s}-I_{s})^{4}\Big{]}.$ We can show convergence of the second term similarly to the proof of 1., as $\displaystyle N\cdot\mathbb{E}\Big{[}\sup_{0\leq s\leq t}\Big{(}\frac{1}{N}(Y^{\pm}(Ns)-Ns)\Big{)}^{4}\Big{]}$ $\displaystyle\leq\Big{(}\frac{4}{3}\Big{)}^{4}\frac{1}{N^{3}}\mathbb{E}\Big{[}\Big{(}(Y^{\pm}(Nt)-Nt)\Big{)}^{4}\Big{]}$ $\displaystyle=\Big{(}\frac{4}{3}\Big{)}^{4}\Big{(}\frac{3t^{2}}{N}+\frac{t}{N^{2}}\Big{)}\xrightarrow{N\to\infty}0.$ For the first term we apply Lemma 5.2 in order to obtain $\displaystyle\mathbb{E}\Big{[}\sup_{0\leq s\leq t}\Big{(}\big{(}\sqrt{N}(I^{N}_{s}-I_{s})-\widetilde{K}^{N}_{s}\big{)}\Big{)}^{2}\Big{]}\leq C\cdot\mathbb{E}\Big{[}\sup_{0\leq s\leq t}\Big{(}\big{(}\sqrt{N}(J^{N}_{s}-J_{s})-\widetilde{G}^{N}_{s}\big{)}\Big{)}^{2}\Big{]}.$ Summing up the results on $A^{N},C^{N},D^{N}$ and $E^{N}$ we conclude $\displaystyle\mathbb{E}\Big{[}\sup_{0\leq s\leq t}\big{(}\sqrt{N}(J^{N}_{s}-J_{s})-\widetilde{G}^{N}_{s}\big{)}^{2}\Big{]}\leq\varepsilon_{N}+C\int_{0}^{t}\mathbb{E}\Big{[}\sup_{0\leq u\leq s}\big{(}\sqrt{N}(J^{N}_{u}-J_{u})-\widetilde{G}^{N}_{u}\big{)}^{2}\Big{]}ds$ for $\varepsilon_{N}\to 0$ and by the Gronwall inequality $\displaystyle\mathbb{E}\Big{[}\sup_{0\leq s\leq t}\big{(}\sqrt{N}(J^{N}_{s}-J_{s})-\widetilde{G}^{N}_{s}\big{)}^{2}\Big{]}\xrightarrow{N\to\infty}0.$ Applying Lemma 5.2 it follows that $\displaystyle\mathbb{E}\Big{[}\sup_{0\leq s\leq t}\big{(}\sqrt{N}(I^{N}_{s}-I_{s})-\widetilde{K}^{N}_{s}\big{)}^{2}\Big{]}\xrightarrow{N\to\infty}0.$ ### 5.5 Proof of Corollaries 3.2 and 3.5 ###### Proof of Corollary 3.2. Recall $Y^{+}$ and $Y^{-}$ from Section 5.2, and which are used frequently in the proof of Theorem 1. For (3.4), using (5.7) and Lemma 5.1, for the Poisson process $Y=(Y(t))_{t\geq 0}$, given by $Y(t)=Y^{+}(pt)+Y^{-}((1-p)t)$, $\displaystyle\sup_{0\leq t\leq T}$ $\displaystyle\Big{|}\frac{1}{N}\sum_{j=1}^{N}\Big{(}Z_{t}^{N,j}-\int_{0}^{t}h(I_{s}^{N})ds\Big{)}\Big{|}=\sup_{0\leq t\leq T}\Big{|}\frac{1}{N}Y\Big{(}N\int_{0}^{t}h(I_{s}^{N})ds\Big{)}-\int_{0}^{t}h(I_{s}^{N})ds\Big{|}$ (5.20) $\displaystyle\leq\sup_{0\leq t\leq T||h||}\Big{|}\frac{1}{N}(Y(Nt)-Nt)\Big{|}=0.$ For (3.5), by the law of large numbers, $\frac{1}{N}\sum_{j=1}^{N}\bar{Z}^{j}_{T}\xrightarrow{N\to\infty}\int_{0}^{T}h(I_{t})dt$, as well as $h(I^{N})\xrightarrow{N\to\infty}h(I)$ in $L^{2}$, uniformly on compact time intervals by Theorem 1.1, therefore $\displaystyle\frac{1}{N}\sum_{j=1}^{N}\Big{(}Z^{N,j}-\bar{Z}^{j}\Big{)}$ $\displaystyle=\frac{1}{N}\sum_{j=1}^{N}\Big{(}Z^{N,j}-\int_{0}^{.}h(I_{s}^{N})ds\Big{)}$ $\displaystyle\qquad\qquad-\frac{1}{N}\sum_{j=1}^{N}\Big{(}\bar{Z}^{j}-\int_{0}^{.}h(I_{s})ds\Big{)}+\int_{0}^{.}(h(I_{s}^{N})-h(I_{s}))ds\xrightarrow{N\to\infty}0.$ For (3.6), we write, with the same Poisson process as in (5.20), and with the Brownian motions $B^{\prime}$ arising as limit of the compensated Poisson process $Y$, and some $\widetilde{B}$, $\displaystyle\frac{1}{\sqrt{N}}$ $\displaystyle\sum_{j=1}^{N}\Big{(}Z^{N,j}_{t}-\int_{0}^{t}h(I_{s}^{N})ds\Big{)}=\frac{1}{\sqrt{N}}\Big{(}Y\Big{(}N\int_{0}^{t}h(I_{s}^{N})ds\Big{)}-N\int_{0}^{t}h(I_{s}^{N})ds\Big{)}$ (5.21) $\displaystyle=B^{\prime}\Big{(}\int_{0}^{t}h(I_{s})ds\Big{)}+o(1)=\int_{0}^{t}\sqrt{h(I_{s})}d\widetilde{B}_{s}+o(1).$ Next, for (3.7), with $K$ as in Theorem 1.3, $\displaystyle\sqrt{N}(h(I_{t}^{N})-h(I_{t}))\Big{)}=\sqrt{N}h^{\prime}(I_{t})(I_{t}^{N}-I_{t})+o(1)=h^{\prime}(I_{t})K_{t}+o(1).$ Last, we note that the left hand side of (3.8) is a sum of the left hand sides of (3.6) and (3.7). We obtain the result by summing the limits of these equations, once we determine the correlation structure of the martingales $\int_{0}^{.}\sqrt{h(I_{s})}dB_{s}$ in (3.3) and $\int_{0}^{.}\sqrt{h(I_{s})}d\widetilde{B}_{s}$ in (3.6) on a joint probability space. We will show that joint weak convergence of (3.6) and (3.7) holds if $\displaystyle\mathbb{E}[\int_{0}^{t}\sqrt{h(I_{s})}dB_{s}\int_{0}^{t}\sqrt{h(I_{s})}d\widetilde{B}_{s}]=(2p-1)\int_{0}^{t}h(I_{s})ds,$ (5.22) implying the result. Recall from Section 5.2 the limits $B^{+}$ and $B^{-}$ of the rescaled Poisson processes $Y^{+}$ and $Y^{-}$. By the construction of $Y$ above (5.20), we find that $B^{\prime}_{t}=B^{+}_{pt}+B^{-}_{(1-p)t}$ is the limit of the rescaled Poisson process $Y$. In addition, $\hat{B}_{t}=B^{+}_{pt}-B^{-}_{(1-p)t}$ arises above (5.10). Clearly, by independence of $B^{+}$ and $B^{-}$, we have $\mathbb{E}[\hat{B}_{t}B_{t}^{\prime}]=\mathbb{E}[(B_{pt}^{+}+B_{(1-p)t}^{-})(B_{pt}^{+}-B_{(1-p)t}^{-})]=(2p-1)t$. From the proof of Theorem 1.3, we know that the convergence in this result is even in probability if we exchange $K$ by $\widetilde{K}^{N}$ from (5.19). Hence, using $I^{N}\to I$, we find for the integral appearing in (3.3), on a joint probability space, and taking (5.10) into account, $\int_{0}^{t}\sqrt{h(I_{s})}dB_{s}=\lim_{N\to\infty}\int_{0}^{t}\sqrt{h(I^{N}_{s})}dB^{N}_{s}=\lim_{N\to\infty}\hat{B}\Big{(}\int_{0}^{t}h(I^{N}_{s})ds\Big{)}=\hat{B}\Big{(}\int_{0}^{t}h(I_{s})ds\Big{)}.$ From this and (5.21), we can thus write $\displaystyle\mathbb{E}\Big{[}\int_{0}^{t}\sqrt{h(I_{s})}dB_{s}\int_{0}^{t}\sqrt{h(I_{s})}d\widetilde{B}_{s}\Big{]}$ $\displaystyle=\mathbb{E}\Big{[}\hat{B}\Big{(}\int_{0}^{t}h(I_{s})ds\Big{)}\cdot B^{\prime}\Big{(}\int_{0}^{t}h(I_{s})ds\Big{)}\Big{]}$ $\displaystyle=(2p-1)\int_{0}^{t}h(I_{s})ds$ and we are done. ∎ ###### Proof of Corollary 3.5. Actually, the proof of (3.4), (3.5) and (3.6) is literally the same as in the proof above. ∎ ## References * Ahmadian and Miller [2021] Ahmadian, Y. and K. D. Miller (2021, Aug). What is the dynamical regime of cerebral cortex? Neuron. * Barral and Reyes [2016] Barral, J. and A. D. Reyes (2016). Synaptic scaling rule preserves excitatory–inhibitory balance and salient neuronal network dynamics. Nature Neuroscience 19(12), 1690–1696. * Berger and Mizel [1980] Berger, M. A. and V. J. Mizel (1980). Volterra equations with Itô integrals—I. The Journal of Integral Equations 2(3), 187–245. * Brémaud and Massoulié [1996] Brémaud, P. and L. Massoulié (1996, 02). Stability of Nonlinear Hawkes Processes. Ann. Appl. Probab. 24(3), 1563–1588. * Brunel [2000] Brunel, N. (2000). Dynamics of sparsely connected networks of excitatory and inhibitory spiking neurons. J Comput Neurosci 8(3), 183–208. * Chen et al. [2019] Chen, S., A. Shojaie, E. Shea-Brown, and D. Witten (2019). The Multivariate Hawkes Process in High Dimensions: Beyond Mutual Excitation. https://arxiv.org/abs/1707.04928. * Chevallier [2017a] Chevallier, J. (2017a). Fluctuations for mean-field interacting age-dependent Hawkes processes. Elec. J. Probab. 22(42), 1–49. * Chevallier [2017b] Chevallier, J. (2017b). Mean-field limit of generalized Hawkes processes. Stoch. Proc. Appl. 127, 3870–3912. * Chevallier et al. [2019] Chevallier, J., A. Duarte, E. Löcherbach, and G. Ost (2019). Mean field limits for nonlinear spatially extended Hawkes processes with exponential memory kernels. Stoch. Proc. Appl. 129, 1–27. * Chevallier and Ost [2020] Chevallier, J. and G. Ost (2020). Fluctuations for spatially extended Hawkes processes. Stoch. Proc. Appl. 130, 5510–5542. * Costa et al. [2020] Costa, M., C. Graham, L. Marsalle, and V. C. Tran (2020). Renewal in Hawkes processes with self-excitation and inhibition. Adv. Appl. Probab. 52, 879–915. * Delattre et al. [2016] Delattre, S., N. Fournier, and M. Hoffmann (2016). Hawkes processes on large networks. Ann. Appl. Probab. 26(1), 216–261. * Ditlevsen and Löcherbach [2017] Ditlevsen, S. and E. Löcherbach (2017). Multi-class oscillating systems of interacting neurons. Stoch. Proc. Appl. 127, 1840–1869. * E. Bacry and Muzy [2015] E. Bacry, I. M. and J.-F. Muzy (2015). Hawkes processes in finance. https://arxiv.org/pdf/1502.04592;Hawkes. * Erny et al. [2020] Erny, X., E. Löcherbach, and D. Loukianova (2020). Mean field limits for interacting hawkes processes in a diffusive regime. https://arxiv.org/abs/1904.06985. * Ethier and Kurtz [1986] Ethier, S. and T. Kurtz (1986). Markov Processes. Characterization and Convergence. John Wiley, New York. * Fox et al. [2016] Fox, E., F. Schoenberg, and J. Gordon (2016). Spatially inhomogeneous background rate estimators and uncertainty quantification for nonparametric Hawkes point process models of earthquake occurrences. Ann. Appl. Stat. 10, 1725–1756. * Gerhard et al. [2017] Gerhard, F., M. Deger, and W. Truccolo (2017, 02). On the stability and dynamics of stochastic spiking neuron models: Nonlinear Hawkes process and point process GLMs. PLoS Comput Biol 13(2), e1005390. * Hawkes [1971a] Hawkes, A. (1971a). Spectra of some self-exciting and mutually exciting point processes. Biometrika 58, 83–90. * Hawkes [2018] Hawkes, A. (2018). Hawkes processes and their applications to finance: a review. Quantitative Finance 18(2), 193–198. * Hawkes and Oakes [1974] Hawkes, A. and D. Oakes (1974). A Cluster Process Representation of a Self-Exciting Process. J. Appl. Probab. 11(3), 493–503. * Hawkes [1971b] Hawkes, A. G. (1971b). Spectra of some self-exciting and mutually exciting point processes. Biometrika 58(1), 83–90. * Heesen and Stannat [2021] Heesen, S. and W. Stannat (2021). Fluctuation limits for mean-field interacting nonlinear Hawkes processes. Stoch. Proc. Appl. 139, 280–297. * Herculano-Houzel [2009] Herculano-Houzel, S. (2009). The human brain in numbers: a linearly scaled-up primate brain. Frontiers in Human Neuroscience 3, 31. https://doi.org/10.3389/neuro.09.031.2009. * Herz et al. [2006] Herz, A. V. M., T. Gollisch, C. K. Machens, and D. Jaeger (2006). Modeling single-neuron dynamics and computations: a balance of detail and abstraction. Science 314(5796), 80–85. * Jacod and Shiryaev [2003] Jacod, J. and A. Shiryaev (2003). Limit theorems for stochastic processes, Volume 288. Springer Science & Business Media. * Jovanovic et al. [2015] Jovanovic, S., J. Hertz, and S. Rotter (2015, Apr). Cumulants of Hawkes point processes. Phys Rev E Stat Nonlin Soft Matter Phys 91(4), 042802\. * Jovanović et al. [2015] Jovanović, S., J. Hertz, and S. Rotter (2015). Cumulants of Hawkes point processes. Physical Review E 91(4), 042802. * Kallenberg [2002] Kallenberg, O. (2002). Foundations of Modern Probability. 2nd ed. Probability and Its Applications. New York, NY: Springer. * Kandel et al. [2012] Kandel, E. R., J. H. Schwartz, T. M. Jessell, S. A. Siegelbaum, and A. J. Hudspeth (2012). Principles of Neural Science (5th ed.). New York: McGraw-Hill. * Lukasik et al. [2016] Lukasik, M., P. K. Srijith, D. Vu, K. Bontcheva, A. Zubiaga, and T. Cohn (2016, August). Hawkes processes for continuous time sequence classification: an application to rumour stance classification in Twitter. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), Berlin, Germany, pp. 393–398. Association for Computational Linguistics. * Massoulié [1998] Massoulié, L. (1998). Stability results for a general class of interacting point processes dynamics, and applications. Stoch. Proc. Appl. 75, 1–30. * Ogata [1988] Ogata, Y. (1988). Statistical models for earthquake occurrences and residual analysis for point processes. J. Amer. Statist. Assoc. 83, 9–27. * Paninski et al. [2004] Paninski, L., J. W. Pillow, and E. P. Simoncelli (2004). Maximum likelihood estimation of a stochastic integrate-and-fire neural encoding model. Neural computation 16(12), 2533–2561. * Pernice et al. [2011] Pernice, V., B. Staude, S. Cardanobile, and S. Rotter (2011). How structure determines correlations in neuronal networks. PLoS Comput Biol 7(5), e1002059. * Pra [2017] Pra, P. D. (2017). Stochastic mean-field dynamics and applications to life sciences. In Giambattista Giacomin, Stefano Olla, Ellen Saada, Herbert Spohn, and Gabriel Stoltz, editors, Stochastic Dynamics Out of Equilibrium, pp. 3–27. Springer. * Vreeswijk and Sompolinsky [1998] Vreeswijk, C. v. and H. Sompolinsky (1998). Chaotic balanced state in a model of cortical circuits. Neural Computation 10(6), 1321–1371. * Yizhar et al. [2011] Yizhar, O., L. E. Fenno, M. Prigge, F. Schneider, T. J. Davidson, D. J. O’shea, V. S. Sohal, I. Goshen, J. Finkelstein, J. T. Paz, et al. (2011). Neocortical excitation/inhibition balance in information processing and social dysfunction. Nature 477(7363), 171–178. * Zhu [2013] Zhu, L. (2013). Central Limit Theorem for Nonlinear Hawkes Processes. J. Appl. Probab. 50, 760–771. * Zhu [2015] Zhu, L. (2015). Large deviations for Markovian nonlinear Hawkes processes. Ann. Appl. Probab. 25, 548–581. * Ziplin et al. [2016] Ziplin, J., F. Schoenberg, K. Coronges, and A. Bertozzi (2016). Point-process models of social network interactions: Parameter estimation and missing data recovery. Euro. Jnl of Applied Mathematics 27, 502–529.
# Supplement: ``A mushy source for the geysers of Enceladus'' Colin R. Meyer<EMAIL_ADDRESS>Thayer School of Engineering, Dartmouth College, Hanover, NH 03755 Jacob J. Buffo Thayer School of Engineering, Dartmouth College, Hanover, NH 03755 Francis Nimmo Department of Earth & Planetary Sciences, University of California, Santa Cruz, CA 95064 USA Andrew J. Wells Atmospheric, Oceanic & Planetary Physics, Department of Physics, University of Oxford, OX1 3PU, UK Samuel Boury Courant Institute of Mathematical Sciences, New York University, New York, NY 10012 Tara C. Tomlinson Thayer School of Engineering, Dartmouth College, Hanover, NH 03755 Jamie R. G. Parkinson Atmospheric, Oceanic & Planetary Physics, Department of Physics, University of Oxford, OX1 3PU, UK Geoffrey M. Vasil School of Mathematics, University of Edinburgh, EH9 3FD, UK ## 1 Fracture depths It is unclear how easy it is to produce a fracture spanning the entire ice shell of Enceladus. Due to the limited tidal stresses ($\sim 100$ kPa), when compared to the overburden pressure ($\sim 600$ kPa at the base of a 6 km ice shell), it is hard for near-surface fractures to extend downwards more than 1–2 km (Smith-Konter and Pappalardo, 2008; Olgin et al., 2011; Walker et al., 2021). This is less than the estimated minimum shell thickness of 6 km at the south pole (Hemingway et al., 2018). Fractures extending upwards from the base of the shell are less affected by overburden pressure because they would be water-filled (Crawford and Stevenson, 1988; Shibley and Laughlin, 2021). However, ice at the base of the shell is more ductile and thus harder to fracture than ice at the surface (Schulson and Duval, 2009). The current paradigm for the source of geyser material is that fractures extend down to the subsurface ocean (Hurford et al., 2007). To get an idea for the depth of the tiger stripe fractures, we balance the extensional stress required to open a fracture with the overburden ice pressure acting to close it (i.e. the zero-stress Nye (1955) criterion for closely spaced glacier crevasses). Estimates for the extensional stress are on the order of $\tau\sim 1$ bar (Nimmo et al., 2007, 2011). Gravity on Enceladus is $g=0.113$ m s-2 (Hansen et al., 2006, 2008) and the salty ice density is about $\rho_{\textrm{ice}}=940$ kg m-3 (Timco and Frederking, 1996). The fracture close-off depth $h_{\textrm{top}}$ is given as $h_{\textrm{top}}\sim\frac{\tau}{\rho_{\textrm{ice}}g}\sim 1~{}\mathrm{km},$ (1) which is an estimate for the likely depth of surface fractures and agrees with more sophisticated calculations (Smith, 1976; Broberg, 1999; Smith-Konter and Pappalardo, 2008; Olgin et al., 2011). The fact that the surface fractures are only likely to propagate to $\sim 1~{}\mathrm{km}$ depth is important because $\sim 1~{}\mathrm{km}$ is shallower than the minimum predicted shell thickness of $\sim 6-30~{}\mathrm{km}$ (Hemingway et al., 2018), which casts doubt on the idea that the fractures penetrate from the surface down to the ocean. Moreover, it would be a considerable feat for a fracture to extend up from the ocean (i.e. a basal crevasse) to within $\sim 1~{}\mathrm{km}$ of the surface, based on the extensional stress just mentioned. A water-filled basal crevasse does lead to a larger close-off height $h_{\textrm{bottom}}$, given by $h_{\textrm{bottom}}\sim\frac{\tau}{\left(\rho_{\textrm{water}}-\rho_{\textrm{ice}}\right)g}\sim 10\mbox{ km},$ (2) where we use a value of seawater density, $\rho_{\textrm{water}}=1030$ kg m-3. This is sufficient to fracture the shell if (i) the fracture emanates up from the bottom and (ii) if the shell is on the lower end of thickness range (i.e. $\sim 6-30~{}\mathrm{km}$ Hemingway et al., 2018). However, the lower shell is likely ductile and less prone to fracturing (Nimmo, 2004; Schulson and Duval, 2009). On the other hand, there are additional sources of stress available to open cracks which likely exceed the tidal stresses. These include the effects of shell thickening (Rudolph et al., 2022) and flexural stresses from plume loading (Hemingway et al., 2020). Thus, opening of through-going fractures appears possible but not certain. At the same time, the dike mechanism discussed in the main text could allow for the initiation of downward- propagating through-going fractures that could be maintained by turbulent water flow and shear heating. We discuss the dike mechanism further in section 5. ## 2 Analytical temperature solution Here we demonstrate that the leading order thermal structure can be understood from steady state conduction, with reactive flow causing subsequent modifications to the porosity structure, composition and shape of the inclusion. We consider a fracture of depth $h$ in a shell of thickness $H$ with a shear heat flux of $\mu\rho_{i}ghu$. So long as the latent heat sink of partial melting and heat transport by interstitial flow are relatively small, then the steady state temperature $T$ around the fracture is given by the solution to Laplace's equation, i.e. $\nabla^{2}T=0,$ (3) where we treat the thermal conductivity as temperature independent to facilitate analytical progress. Equation (3) is subject to the boundary conditions $\displaystyle-k\frac{\partial T}{\partial x}$ $\displaystyle=$ $\displaystyle 0~{}~{}~{}\mbox{on}~{}~{}~{}x=0~{}(z<h)~{}~{}~{}\mbox{and}~{}~{}~{}x=L,$ (4) $\displaystyle-k\frac{\partial T}{\partial x}$ $\displaystyle=$ $\displaystyle\mu\rho_{i}ghu~{}~{}~{}\mbox{on}~{}~{}~{}x=0~{}(z\geq h),$ (5) $\displaystyle-k\frac{\partial T}{\partial z}$ $\displaystyle=$ $\displaystyle r_{1}+r_{2}(T-T_{s})~{}~{}~{}\mbox{on}~{}~{}~{}z=H,$ (6) $\displaystyle T$ $\displaystyle=$ $\displaystyle T_{e}~{}~{}~{}\mbox{on}~{}~{}~{}z=0.$ (7) The surface condition is a linearized radiative condition, with respect to the background temperature $T_{s}$ of space near Enceladus (see table 1), rather than the full black-body radiation condition as used by Abramov and Spencer (2009). This choice for the surface boundary condition allows for an analytical solution and is the condition used in SOFTBALL. To derive the linearized condition, we start from the surface energy balance where conductive heat fluxes from the ice balance net black-body radiation as $-k\frac{\partial T}{\partial z}=\epsilon\sigma\left(T^{4}-T_{s}^{4}\right),$ (8) where $\epsilon$ is the emissivity and $\sigma$ is the Stefan-Boltzmann constant (values in table 1). Expanding the temperature as a small deviation from the space temperature, i.e. $T=T_{s}+\tilde{T}$, where $\tilde{T}\ll T_{s}$, then we can use the binomial expansion and discard terms of second order and higher, to find that $-k\frac{\partial T}{\partial z}=4\epsilon\sigma T_{s}^{3}(T-T_{s}),$ (9) which is in the form of the boundary condition above, with $r_{1}=0$ and $r_{2}=4\epsilon\sigma T_{s}^{3}$. We nondimensionalized the problem, by taking $T=T_{e}+\Delta T\theta,~{}~{}~{}x=H\tilde{x},~{}~{}~{}z=H\tilde{z},$ which yields $\tilde{\nabla}^{2}\theta=0,$ (10) subject to the boundary conditions $\displaystyle-\frac{\partial\theta}{\partial\tilde{x}}$ $\displaystyle=$ $\displaystyle 0~{}~{}~{}\mbox{on}~{}~{}~{}\tilde{x}=0~{}(\tilde{z}<\delta)~{}~{}~{}\mbox{and}~{}~{}~{}\tilde{x}=\lambda,$ (11) $\displaystyle-\frac{\partial\theta}{\partial\tilde{x}}$ $\displaystyle=$ $\displaystyle F~{}~{}~{}\mbox{on}~{}~{}~{}\tilde{x}=0~{}(\tilde{z}\geq\delta),$ (12) $\displaystyle-\frac{\partial\theta}{\partial\tilde{z}}$ $\displaystyle=$ $\displaystyle G+b\left(\theta-\theta_{s}\right)~{}~{}~{}\mbox{on}~{}~{}~{}\tilde{z}=1,$ (13) $\displaystyle\theta$ $\displaystyle=$ $\displaystyle 0~{}~{}~{}\mbox{on}~{}~{}~{}\tilde{z}=0.$ (14) Here we define the parameters as $\delta=\frac{h}{H},~{}~{}~{}\lambda=\frac{L}{H},~{}~{}~{}F=\frac{\mu\rho_{i}ghHu}{k_{o}\Delta T},~{}~{}~{}G=\frac{Hr_{1}}{k_{o}\Delta T},~{}~{}~{}b=\frac{Hr_{2}}{k_{o}},~{}~{}~{}\theta_{s}=\frac{T_{s}-T_{e}}{\Delta T}.$ Due to convergence issues in the SOFTBALL simulations, we could not use large values for $b$, and so the value quoted in table 1 is artificially lower than what we would calculate from the parameters. The main effect of this difference is on the surface temperature profile and not the size, shape, or liquid volume of the mushy zone. Asymptotically, our analytical solutions shows that large values of $b$ pin the surface temperature to the space temperature, except very near the crack, where there is a narrow boundary layer region of rapid temperature change. The boundary layer becomes thinner as $b$ increases. Additionally, since $G=0$ here, we disregard it in subsequent calculations. To simplify the problem in the vertical direction, we write the total solution $\theta$ as a sum of a particular solution $\theta_{I}$ to the inhomogeneous problem and the homogenous solution $\theta_{H}$, i.e. $\theta=\theta_{I}(\tilde{z})+\theta_{H}(\tilde{x},\tilde{z}).$ Solving first for the inhomogeneous part, we have that $\frac{\partial^{2}\theta_{I}}{\partial\tilde{x}^{2}}+\frac{\partial^{2}\theta_{I}}{\partial\tilde{z}^{2}}=0\longrightarrow\frac{\partial^{2}\theta_{I}}{\partial\tilde{z}^{2}}=0,$ so $\theta_{I}=A\tilde{z}+B.$ Using the boundary conditions, we find that $B=0\ \ \ \mbox{ and }\ \ \ A=\frac{b\theta_{s}}{1+b}.$ For large values of $b$, the slope $A$ is approximately $\theta_{s}$, which is consistent with the asymptotic description above, that the surface temperature mostly stays close to the space temperature. Now we can examine the homogeneous problem using separation of variables, i.e. $\theta_{H}(\tilde{x},\tilde{z})=f(\tilde{x})g(\tilde{z}),~{}~{}~{}\mbox{which implies}~{}~{}~{}f^{\prime\prime}g+fg^{\prime\prime}=0.$ Dividing through and choosing the separation constant $\alpha^{2}$, we find the two ODEs $f^{\prime\prime}=\alpha^{2}f\ \ \ \mbox{ and }\ \ \ g^{\prime\prime}=-\alpha^{2}g.$ Solving the vertical equation first, we have that $g=C\sin(\alpha\tilde{z})+D\cos(\alpha\tilde{z}).$ With $g=0$ at $\tilde{z}=0$, we know that $D=0$. The homogeneous surface boundary condition is $-g^{\prime}=bg\ \ \ \mbox{ {at} }\ \ \ \tilde{z}=1,$ which results in the transcendental eigenvalue condition $-\alpha\cos(\alpha)=b\sin(\alpha).$ (15) This yields a full set of distinct eigenvalues $\alpha_{n}$ for $n=1,2,3,...$, which can be computed numerically. In the horizontal direction, we have that $f=De^{\alpha\tilde{x}}+Ee^{-\alpha\tilde{x}}.$ Using the no heat flux boundary condition at $\tilde{x}=\lambda$, we have that $-f^{\prime}=0\ \ \ \mbox{ {at} }\ \ \ \tilde{x}=\lambda\ \ \ \mbox{ {implies} }\ \ \ E=De^{2\alpha\lambda}.$ Thus, the full solution with unknown coefficients is given by $\theta=\frac{b\theta_{s}}{1+b}\tilde{z}+\sum_{n=1}^{\infty}{c_{n}\left[e^{\alpha_{n}\tilde{x}}+e^{2\alpha_{n}\lambda}e^{-\alpha_{n}\tilde{x}}\right]\sin(\alpha_{n}\tilde{z})}.$ (16) We find the values for $c_{n}$ by using orthogonality and integrating over the heat flux condition at $\tilde{x}=0$, which gives $\int_{\delta}^{1}{F\sin(\alpha_{m}\tilde{z})~{}d\tilde{z}}=\sum_{n=1}^{\infty}{c_{n}\alpha_{n}\left[1-e^{2\alpha_{n}\lambda}\right]\int_{0}^{1}{\sin(\alpha_{n}\tilde{z})\sin(\alpha_{m}\tilde{z})~{}d\tilde{z}}}.$ (17) The integral on right is zero for $\alpha_{n}\neq\alpha_{m}$, using the transcendental eigenvalue condition from equation (15) implying orthogonality of the eigenfunctions $\sin(\alpha_{n}\hat{z})$. When $\alpha_{n}=\alpha_{m}$, we have that $\int_{0}^{1}{\sin(\alpha_{m}\tilde{z})\sin(\alpha_{m}\tilde{z})~{}d\tilde{z}}=\frac{1}{2}-\frac{\sin(2\alpha_{m})}{4\alpha_{m}}.$ We evaluate the integral on the left as $\int_{\delta}^{1}{F\sin(\alpha_{m}\tilde{z})~{}d\tilde{z}}=-\frac{F}{\alpha_{m}}\left[\cos(\alpha_{m})-\cos(\alpha_{m}\delta)\right].$ Putting it all together gives, $c_{n}=-\frac{4F}{\left[2\alpha_{n}^{2}-\alpha_{n}\sin(2\alpha_{n})\right]}\frac{\left[\cos(\alpha_{n})-\cos(\alpha_{n}\delta)\right]}{1-e^{2\alpha_{n}\lambda}}.$ (18) We show the analytical solution in figure 1 and find that a mushy zone will form along the crack face in regions with $T>T_{e}$. These results are similar to what we find in the main text as well as Hammond (2020), Kalousová et al. (2016), and Stevenson (1996). This analytical solution also allows us to predict the size of the mushy zone, by looking for the contour where $T=T_{e}$. Figure 1: Analytical solution for the temperature around a heated crack with $F=260$. Figure 2: Liquid volume in the mushy zone for $F=500$: (a) evolution to a steady state and average effective heat flux as well as (b) rate of liquid volume production, which can support the entire flux observed by the Cassini spacecraft from a single tiger stripe fracture. ## 3 Mushy zone physics To analyze the effect of shear heating on a fracture in an icy shell, we use SOFTBALL (SOlidification, Flow and Thermodynamics in Binary ALLoys), a mushy layer code with adaptive mesh refinement. The github repository for this open- source code is: https://github.com/jrgparkinson/mushy-layer (accessed 8 April 2023). This state-of-the-art code solves the equations of mushy layer theory, as described below, in two-dimensional geometries (Wells et al., 2019; Parkinson et al., 2020). The code uses a finite volume discretization and adaptive time steps. In our simulations we use a spatial scale of $N_{x}=512$ grid points in the horizontal direction and $N_{z}=256$ in the vertical direction. This gives a resolution of $\sim 40$ meters in both the horizontal and vertical directions. The time steps adjust and are on the order of $\sim 30$ years. The simulations take on the order of a day to run on the linux system (2-16 cores) at Dartmouth College. Conservation of mass in a mushy layer dictates that $\frac{\partial\left(\phi\rho\right)}{\partial t}=M_{f}~{}~{}~{}\mbox{and}~{}~{}~{}\frac{\partial\left[\left(1-\phi\right)\rho\right]}{\partial t}+\boldsymbol{\nabla}\cdot\boldsymbol{q}=-M_{f},$ (19) where the solid and liquid densities $\rho$ are approximated as equal (using an extended form of the Boussinesq approximation; Parkinson et al., 2020), the solid fraction is $\phi$, the fluid flux relative to the solid is $\boldsymbol{q}$, and the rate of mass phase change per volume is $M_{f}$. The sum of equations (19) yields $\boldsymbol{\nabla}\cdot\boldsymbol{q}=0$, i.e. a solenoidal fluid flow field. Here we assume that the porous solid matrix is stationary. Fluid flux is a function of pressure gradients and gravity, which can be described by Darcy's law, as $\boldsymbol{q}=-\frac{\Pi}{\eta}\left(\nabla p+\rho_{b}g\boldsymbol{z}\right),$ (20) where $\eta$ is the liquid viscosity, $\rho_{b}$ is the density of liquid brine (which is a function of the brine concentration and temperature), $\boldsymbol{z}$ is the vertical unit vector, and $\Pi$ is the solid-fraction- dependent permeability (Tait and Jaupart, 1992; Katz and Worster, 2008), given by $\Pi=\left[\frac{12}{d^{2}}+\frac{1}{K_{0}\phi^{3}}\right]^{-1}.$ (21) Following Katz and Worster (2008), SOFTBALL uses a permeability regularization for pure liquid regions (i.e. as $\phi\rightarrow 1$) such that there is a background permeability and Darcy's law is solved everywhere, rather solving Navier-Stokes in the pure fluid region (see Parkinson et al., 2020; Buffo et al., 2021). The ratio of the background permeability $d^{2}/12$ to the permeability prefactor $K_{0}$ is the reluctance $\mathtt{R}=d^{2}/(12K_{0})$. Conservation of energy in the mushy layer is given by $\overline{\rho c_{p}}\frac{\partial T}{\partial t}+\rho_{b}c_{b}\boldsymbol{q}\cdot\boldsymbol{\nabla}T=\boldsymbol{\nabla}\cdot\left(\overline{k}\boldsymbol{\nabla}T\right)+\mathscr{L}M_{f},$ (22) where $\overline{\rho c_{p}}$ and $\overline{k}$ are the solid-fraction- weighted functions of ice and brine properties for heat capacity and thermal conductivity, respectively. The latent heat is $\mathscr{L}$. We write a conservation law for the concentration of brine, i.e. $\left(1-\phi\right)\frac{\partial C}{\partial t}+\boldsymbol{q}\cdot\boldsymbol{\nabla}C=\boldsymbol{\nabla}\cdot\left(\overline{D}\boldsymbol{\nabla}C\right)+\frac{\left(C-C_{s}\right)}{\rho}M_{f},$ (23) where $\overline{D}$ is the phase-weighted salt diffusivity arising from transport through the interstitial liquid. The evolution of temperature and concentration are coupled in the mushy region, as shown in figure 1 in the main text, where the ice and liquid phases are at an equilibrium temperature given by the liquidus. This imposes a constraint on the evolution equations (22) and (23) in mushy regions, where the concentration and temperature are tied together by the liquidus relationship $T=T_{L}(C)$. In idealized form, the liquidus relationship is given as $T_{L}=T_{m}-mC,$ (24) where $T_{m}$ is the pure substance melting temperature and $m$ is the liquidus slope. This constraint no longer applies when $T<T_{e}$, where $M_{f}=0$, $\boldsymbol{q}=\boldsymbol{0}$, $\phi=1$, (22) becomes the heat equation for thermal conduction and salt conservation (23) implies constant bulk salinity. ## 4 Liquid volume and rates of production In the main text, we show steady state temperature and porosity fields. We treat the rate of shear heating as constant, given that the thermal diffusion timescale is much longer than the tidal cycle (Nimmo and Gaidos, 2002; Nimmo et al., 2007). Here, in figure 2(left), we show the approach to steady state for the liquid volume in the mushy zone. Initially, there is no mushy zone and therefore no liquid volume. Then a mushy zone forms and the liquid volume increases rapidly. Due to the temperature field reaching a steady state and the porosity-weakening of the fault friction saturates causing the heat flux $F(1-\overline{\phi})$ to saturate in figure 2(left). In all simulations, the local value of $\phi$ is used to calculate the shear heating rate. Here we show the average value over fracture as an aggregate measure of the effect of reducing the shear heating flux. From the curve of liquid volume with time, we take the derivative to find the rate of liquid production, as shown in figure 2(right). This shows the same three regimes: no initial liquid volume, growth of a mushy zone, and approach to steady state. \begin{overpic}[width=138.76157pt]{pltU_tempcontour180.pdf} \put(33.0,75.0){$F=180$}\end{overpic}\begin{overpic}[width=138.76157pt]{pltU_tempcontour240.pdf} \put(33.0,75.0){$F=240$}\end{overpic}\begin{overpic}[width=138.76157pt]{pltU_tempcontour420.pdf} \put(33.0,75.0){$F=420$}\end{overpic}\begin{overpic}[width=138.76157pt]{pltU_porosity180.pdf} \end{overpic}\begin{overpic}[width=138.76157pt]{pltU_porosity240.pdf} \end{overpic}\begin{overpic}[width=138.76157pt]{pltU_porosity420.pdf} \end{overpic} Figure 3: Temperature and porosity plots for three different values of the nondimensional heat flux: (left) $F=180$, (middle) $F=240$, and (right) $F=420$. Note that porosity is shown with different axis than the temperature plots. Similar to the main text, we show vertical velocity arrows with logarithmic magnitude with the largest arrows representing about 1 mm/year. The observed geyser ejection rate for all four tiger stripes is $\sim$0.2 m3 s-1 (Hansen et al., 2006, 2008). Following a step change in heating and for the range of heat fluxes we used in these simulations, we find that the maximum rate of liquid volume production is greater than $0.2$ m3 s-1 ($F=500$), which is sufficient to explain the geyser flux from all of the tiger stripe fractures. Figure 2 shows that the melt rate is equal to the observed geyser ejection rate initially and stays at a comparable magnitude for $\sim$200 thousand years. At later times, the melt rate reduces and melt volume saturates because a region of large porosity develops and weakens the frictional heating on the crack. However, venting of liquid from the crack will limit porosity growth, and enable sustained heating and melting rates. Figure 4: Flow velocity in the mushy zone for $F=500$: (left) vertical velocity, showing buoyant fluid rising along the crack and (right) horizontal velocity, demonstrating the buoyant plume is stable at $\mathtt{Ra_{C}}=200$ (Boury et al., 2021) with periodic structures in the outer region of the mushy zone. The fact that peak melt rates exceeds the observed geyser flux is an exciting demonstration of the viability of the shear heating model for the genesis of plumes on Enceladus and other icy satellites. While the simple configuration with steady heating doesn't account for the inherent complexities of the transient evolution, our impulse-response calculations likely elucidate the order of magnitude and timescale of heating. In figure 3, we show the steady state temperature and porosity for three increasing nondimensional flux values, $F=180$, $F=240$, and $F=420$. These figures mirror figures 2(right) and 3(right) in the main text, where we show the temperature and porosity structure for $F=500$. For $F=260$, the maximum rate of liquid volume production is on the order of $0.05$ m3 s-1, which is less than rate required to indefinitely produce plumes with the observed ejection rate, but the mushy zone liquid reservoir could sustain geysers for tens of thousands of years. This suggests that there is a parallel to the Enceladus heat signal, which may be an indication of oscillations in thermal activity and hence plume activity (Spencer and Nimmo, 2013; Nimmo et al., 2018). In other words, the liquid volume production rate by shear heating may allow for a large reservoir to develop periodically, leading to geysers that transiently exhausts the supply. Lastly, in our simulations, we reduce the shear heating rate as liquid is generated in the vicinity of the fracture. If we were to remove the liquid as geyser material, however, the shear heating could stay high, leading to additional liquid volume production. We currently do not remove the liquid from our simulation domain and leave these simulations to future work. Within the mushy zone, there is convection of the pore fluid and the fluid near the fracture rises buoyantly, as shown in the main text and in figures 4 and 5. The intensity of convection is controlled by the thermal and compositional mushy Rayleigh numbers, $\displaystyle\mathtt{Ra_{T}}=\frac{\rho_{w}g\alpha\Delta TK_{0}h}{\kappa_{o}\eta},$ (25) $\displaystyle\mathtt{Ra_{C}}=\frac{\rho_{w}g\beta\Delta CK_{0}h}{\kappa_{o}\eta},$ (26) respectively. Boury et al. (2021) show that the rising plume along the fracture is unstable, with the development of the instability here resulting in rapid variation in horizontal velocity leading to numerical convergence issues in SOFTBALL (figure 5). To find solutions that converge in a reasonable amount of time for our grid resolution, we follow Boury et al. (2021) and restrict the Rayleigh numbers to values below the instability threshold, $\mathtt{Ra_{C}}\approx 205$ for our system and use $\mathtt{Ra_{C}}=200$ for the results we present. There is a large uncertainty in the permeability prefactor $K_{0}$ and the Rayleigh numbers that we present use a value of $K_{0}$ that is on the lower end for sea ice simulations (Polashenski et al., 2017; Buffo et al., 2018; Boury et al., 2021). The low permeability limits convection and allows the simulations to converge numerically. In figure 5, we compare three increasing Rayleigh numbers $\mathtt{Ra_{C}}=200,~{}400,~{}\mbox{and}~{}800$, where the latter two simulations do not converge. The higher Rayleigh numbers show that the flow- and-phase-change instability leads to fine-scale structures that ultimately approach the grid resolution, affecting the pattern of the horizontal velocity field. The porosity field, however, is similar across the three Rayleigh numbers, suggesting that the results for total volume and peak melt rate would approximately hold as the Rayleigh number increased. \begin{overpic}[width=138.76157pt]{pltU_horizvel500_Ra200.pdf}\put(33.0,76.0){$\mathtt{Ra_{C}}=200$}\end{overpic}\begin{overpic}[width=138.76157pt]{pltU_horizvel500_Ra400.pdf}\put(33.0,76.0){$\mathtt{Ra_{C}}=400$}\put(22.0,11.0){not converged}\end{overpic}\begin{overpic}[width=138.76157pt]{pltU_horizvel500_Ra800.pdf}\put(33.0,76.0){$\mathtt{Ra_{C}}=800$}\put(22.0,11.0){not converged}\end{overpic} Figure 5: Increasing instability with stronger convection for $F=500$: (left) $\mathtt{Ra_{C}}=200$ showing a stable interior plume and an onset of instability in the region around the liquid inclusion. (middle) $\mathtt{Ra_{C}}=400$ showing an onset of instability in the interior plume. (right) $\mathtt{Ra_{C}}=800$ showing an unstable interior plume and a stable outer region that becomes unstable at the edge of the mushy zone. The three porosity plots, however, are similar. Near the fracture, shear heating melts the ice and dissolves the salt. Therefore, the pore fluid in this region has a higher salinity (figure 6) than the surrounding shell. Here we plot the bulk salinity $C$, which is given as $C=\phi C_{\ell}+(1-\phi)C_{s},$ (27) where $\phi$ is the porosity, $C_{\ell}$ is the liquid salinity, and $C_{s}$ is the solid salinity. Starting from a step change in heating, the evolution of the bulk salinity is shown in figure 7. First, shear heating warms the salty ice above the eutectic temperature generating a permeable zone. Since the brine salinity and temperature are connected through the liquidus constraint within the mushy zone, colder and saltier brine along the top then sinks and the resulting convective desalination reduces the bulk salinity when compared to the background value. Next, the sinking brine cannot escape and collects near the base of the porous inclusion along the fracture, generating a salt-rich zone with porosity near unity. Convection continues within the mushy zone. The convective desalination leaves behind regions of high bulk salinity in the outer parts of the porous inclusion, that are trapped due to the reduction of permeability in the neighbouring desalinated ice which now has low porosity. Due to the partial melting within the mushy zone, the chemical composition of the geyser ejecta sourced from the shear heating model that we propose here will be different from geyser ejecta sourced from the ocean. In our simulations, the salinity in the fluid around the crack is higher than the background shell salinity. The shell likely froze from the underlying ocean and will have a salinity structure that varies with depth (Buffo et al., 2021). Figure 6: Salinity for the mushy zone with $F=500$. (left) Bulk salinity $C$, showing that the fluid within the interstices is saltier than the surrounding shell, which froze from the underlying ocean. (right) Liquid salinity $C_{\ell}$ showing the increase in salinity with depth in the mushy zone. High-salinity channels form at an oblique angle, due to the competition between melting and gravity drainage (Buffo et al., 2023). ## 5 Dike propagation In our steady state simulations, a region of salty brine develops along the fracture with a porosity approaching a value of $\phi\approx 1$. Here we do not allow the liquid to escape as geysers, therefore the liquid volume can build up within the mushy zone. The question then arises as to whether the brine pocket can hydraulically fracture through the ice shell and into the underlying ocean, thereby generating a dike that opens a surface-to-base connection. We analyze this possibility through a scaling analysis, following Lister (1990) and Kalousová et al. (2016). For $F=500$, we take the half width of brine region to be $a\sim 200$ m, the breadth into the page as $w\sim 500$ km, and the height as $\ell\sim 4$ km, we have two options for the balance of gravity. Either, viscous stresses resist propagation or the fracture toughness $K_{I}\sim 100$ kPa m1/2 resists opening. Starting with the case of viscous stress, we have that $\phi\Delta\rho ga\ell w\sim\mu\frac{V}{a^{2}}a\ell w,$ (28) where $\phi\sim 0.25$ is a representative average porosity in the brine pocket; $\Delta\rho\sim 90$ kg m-3 is the density difference between the brine and the surrounding ice; $\mu\sim 10^{14}$ Pa$\cdot$s is the brine-saturated temperate ice viscosity; and $V$ is the rate of dike propagation. Canceling terms, we find that the rate of propagation scales as $V\sim\frac{\phi\Delta\rho ga^{2}}{\mu}.$ (29) The time $t_{\textrm{dike}}$ for a dike to propagate a distance $\ell$, then scales as $t_{\textrm{dike}}\sim\frac{\ell}{V}\sim\frac{\mu\ell}{\phi\Delta\rho ga^{2}}\sim 110~{}\mbox{kyr}.$ (30) This timescale is on the same order of magnitude as the time to steady state in our shear heating calculations, indicating that the brine pocket could form and then propagate as a dike within that time frame, if the porosity grows to order 1 without significant venting of the liquid from the crack. Figure 7: Bulk salinity evolution with $F=500$, showing the drainage and brine concentration along the fracture: (left) 541 kyr. (middle) 865 kyr. (right) 13,000 kyr. If the balance is instead between gravity and the fracture toughness, we have that $\phi\Delta\rho ga\ell w\sim\frac{K_{I}}{\ell^{1/2}}aw,$ (31) which determines whether the fracture will propagate or not. Simplifying and plugging in numbers, we have that $\phi\Delta\rho g\ell^{3/2}\sim 700~{}\mbox{kPa m${}^{1/2}$}~{}~{}>~{}~{}K_{I}\sim 100~{}\mbox{kPa m${}^{1/2}$}.$ (32) Thus, we find that $\phi\Delta\rho g\ell^{3/2}$ is greater than the ice fracture toughness $K_{I}$ and much larger than values measured in brine- saturated sea ice, e.g. $\sim 50$ kPa m1/2 (Schulson and Duval, 2009), demonstrating that the ice will not provide significant resistance to fracture. In both cases, we see that dike propagation is possible if the porosity reaches order 1 over an extended region, before the liquid is vented. The time scale for diking is on the same order of magnitude as the for our simulations to reach a steady state and the ice fracture toughness is not large enough to prevent fracture. These scaling calculations spawn a few ideas. An internal dike is interesting because there could be periodic brine pocket generation, leading to opening a connection between surface and ocean, then Kite and Rubin (2016)-style eruptions. Or the dike could drain all of the fluid internally, leading to a stronger fault with larger shear heating, and a periodic mushy zone source for the geysers. However, by not allowing for interstitial liquid to escape, our estimates for the liquid brine volume are artificially high, meaning that a region with porosity approaching 1 may never form within the mushy zone. In this case, there would be no prospect for a dike and the material generated would only emanate out as geysers. We leave additional analysis exploring these ideas to future work. parameters | | ---|---|--- specific heat of ice | $c_{i}$ | 2000 J kg-1 K-1 specific heat of ocean | $c_{o}$ | 4000 J kg-1 K-1 ice density | $\rho_{i}$ | 940 kg m-3 seawater density | $\rho_{w}$ | 1030 kg m-3 density coefficient for temperature | $\beta_{T}$ | 2.1$\times 10^{-4}$ kg m-3 K-1 density coefficient for salt | $\beta_{C}$ | 7.7$\times 10^{-4}$ kg m-3 ppt-1 thermal conductivity of ice | $k_{i}$ | 2.0 W m-1 K-1 thermal conductivity of ocean | $k_{o}$ | 0.6 W m-1 K-1 latent heat of fusion | $\mathscr{L}$ | 335000 J kg-1 water viscosity | $\eta$ | 2$\times 10^{-3}$ Pa s salt diffusivity | $\kappa_{s}$ | 2$\times 10^{-9}$ m2 s-1 thermal diffusivity | $\kappa_{o}$ | 1.5$\times 10^{-7}$ m2 s-1 Stefan-Boltzmann constant | $\sigma$ | $5.67\times 10^{-8}$ W m-2 K-4 shell salt composition | $C_{i}$ | 35 ppt eutectic composition | $C_{e}$ | 230 ppt liquidus slope | $m$ | -0.0913 K ppt-1 eutectic temperature | $T_{e}$ | 252 K surface temperature | $T_{s}$ | 75 K gravity on Enceladus | $g$ | 0.113 m s-2 fracture depth | $h$ | 5 km height scale | $H$ | 10 km permeability prefactor | $K_{0}$ | 3.2$\times 10^{-13}$ m2 Hele-Shaw cell spacing | $d$ | 5$\times 10^{-5}$ m time scale | $t_{c}$ | 2.2$\times 10^{7}$ years slip velocity | $u$ | 3.4$\times$10-6 m s-1 temperature scale | $\Delta T$ | 18 K emissivity | $\epsilon$ | 1 coefficient of friction | $\mu$ | 0.3 Stefan number | $\mathtt{St}$ | 4.7 Prandtl number | $\mathtt{Pr}$ | 12 Lewis number | $\mathtt{Le}$ | 73 mushy compositional Rayleigh number | $\mathtt{Ra_{C}}$ | 200 mushy thermal Rayleigh number | $\mathtt{Ra_{T}}$ | 2 reluctance | $\mathtt{R}$ | 9.6 shear heating | $F$ | 500 linearized radiation, nondimensional $r_{1}$ | $G$ | 0 linearized radiation, nondimensional $r_{2}$ | $b$ | 100 linearized radiation, nondimensional $T_{s}$ | $\theta_{s}$ | -10 Table 1: Parameters used in the SOFTBALL numerical simulations, grouped by constants, variables, nondimensional parameters. Definitions follow Parkinson et al. (2020). Values from Nimmo and Gaidos (2002), Nimmo et al. (2007), Spencer and Nimmo (2013) as well as Buffo et al. (2021). ## References * Abramov and Spencer (2009) O. Abramov and J. R. Spencer. Endogenic heat from Enceladus' south polar fractures: New observations, and models of conductive surface heating. _Icarus_ , 199(1):189 – 196, 2009. doi: 10.1016/j.icarus.2008.07.016. * Boury et al. (2021) S. Boury, C. R. Meyer, G. M. Vasil, and A. J. Wells. Convection in a mushy layer along a vertical heated wall. _J. Fluid Mech._ , 926:A33, 2021. doi: 10.1017/jfm.2021.742. * Broberg (1999) K. B. Broberg. _Cracks and fracture_. Academic Press, 1999. * Buffo et al. (2018) J. J. Buffo, B. E. Schmidt, and C. Huber. Multiphase reactive transport and platelet ice accretion in the sea ice of McMurdo Sound, Antarctica. _J. Geophys. Res._ , 123(1):324–345, 2018. doi: 10.1002/2017JC013345. * Buffo et al. (2021) J. J. Buffo, C. R. Meyer, and J. R. G. Parkinson. Dynamics of a solidifying icy satellite shell. _J. Geophys. Res._ , 126(5):e2020JE006741, 2021\. doi: 10.1029/2020JE006741. * Buffo et al. (2023) J. J. Buffo, C. R. Meyer, C. J. Chivers, C. C. Walker, C. Huber, and B. E. Schmidt. Geometry of freezing impacts ice composition: Implications for icy satellites. _J. Geophys. Res._ , 128(3):e2022JE007389, 2023\. doi: 10.1029/2022JE007389. * Crawford and Stevenson (1988) G. D. Crawford and D. J. Stevenson. Gas-driven water volcanism in the resurfacing of Europa. _Icarus_ , 73:66–79, 1988. doi: 10.1016/0019-1035(88)90085-1. * Hammond (2020) N. P. Hammond. Estimating the magnitude of cyclic slip on strike-slip faults on Europa. _J. Geophys. Res._ , 125(7):e2019JE006170, 2020\. doi: 10.1029/2019JE006170. * Hansen et al. (2006) C. J. Hansen, L. Esposito, A. I. F Stewart, J. Colwell, A. Hendrix, W. Pryor, D. Shemansky, and R. West. Enceladus' water vapor plume. _Science_ , 311(5766):1422–1425, 2006. doi: 10.1126/science.1121254. * Hansen et al. (2008) C. J. Hansen, L. W. Esposito, A. I. F. Stewart, B. Meinke, B. Wallis, J. E. Colwell, A. R. Hendrix, K. Larsen, W. Pryor, and F. Tian. Water vapour jets inside the plume of gas leaving enceladus. _Nature_ , 456(7221):477–479, 2008. doi: 10.1038/nature07542. * Hemingway et al. (2018) D. Hemingway, L. Iess, R. Tajeddine, and G. Tobie. The interior of Enceladus. In P. M. Schenk, R. N. Clark, C. J. A. Howett, A. J. Verbiscer, and J. H. Waite, editors, _Enceladus and the Icy Moons of Saturn_ , pages 57–78. University of Arizona Press, 2018. ISBN 9780816537075. doi: 10.2307/j.ctv65sw2b.12. * Hemingway et al. (2020) D. J. Hemingway, M. L. Rudolph, and M. Manga. Cascading parallel fractures on Enceladus. _Nat. Astron._ , 4:234–239, 2020. doi: 10.1038/s41550-019-0958-x. * Hurford et al. (2007) T. A. Hurford, P. Helfenstein, G. V. Hoppa, R. Greenberg, and B. G. Bills. Eruptions arising from tidally controlled periodic openings of rifts on Enceladus. _Nature_ , 447:292–294, 2007. doi: 10.1038/nature05821. * Kalousová et al. (2016) K. Kalousová, O. Souček, G. Tobie, G. Choblet, and O. Čadek. Water generation and transport below Europa's strike-slip faults. _J. Geophys. Res._ , 121:2444–2462, 2016. doi: 10.1002/2016JE005188. * Katz and Worster (2008) R. F. Katz and M. G. Worster. Simulation of directional solidification, thermochemical convection, and chimney formation in a Hele-Shaw cell. _J. Comput. Phys._ , 227(23):9823–9840, 2008\. doi: 10.1016/j.jcp.2008.06.039. * Kite and Rubin (2016) E. S. Kite and A. M. Rubin. Sustained eruptions on Enceladus explained by turbulent dissipation in tiger stripes. _Proc. Natl. Acad. Sci._ , 113(15):3972–3975, 2016. doi: 10.1073/pnas.1520507113. * Lister (1990) J. R. Lister. Buoyancy-driven fluid fracture: similarity solutions for the horizontal and vertical propagation of fluid-filled cracks. _J. Fluid Mech._ , 217:213–239, 1990. doi: 10.1017/S0022112090000696. * Nimmo (2004) F. Nimmo. Stresses generated in cooling viscoelastic ice shells: Application to Europa. _J. Geophys. Res._ , 109(E12), 2004. doi: 10.1029/2004JE002347. * Nimmo and Gaidos (2002) F. Nimmo and E. Gaidos. Strike-slip motion and double ridge formation on Europa. _J. Geophys. Res._ , 107(E4):5–1–5–8, 2002\. doi: 10.1029/2000JE001476. * Nimmo et al. (2007) F. Nimmo, J. R. Spencer, R. T. Pappalardo, and M. E. Mullen. Shear heating as the origin of the plumes and heat flux on Enceladus. _Nature_ , 447(7142):289, 2007. doi: 10.1038/nature05783. * Nimmo et al. (2011) F. Nimmo, B. G. Bills, and P. C. Thomas. Geophysical implications of the long-wavelength topography of the Saturnian satellites. _J. Geophys. Res._ , 116(E11), 2011. doi: 10.1029/2011JE003835. * Nimmo et al. (2018) F. Nimmo, A. C. Barr, M. Běhounková, and W. B. McKinnon. The thermal and orbital evolution of Enceladus: Observational constraints and models. In P. M. Schenk, R. N. Clark, C. J. A. Howett, A. J. Verbiscer, and J. H. Waite, editors, _Enceladus and the Icy Moons of Saturn_ , pages 79–94. University of Arizona Press, 2018. ISBN 9780816537075. doi: 10.2307/j.ctv65sw2b.13. * Nye (1955) J. F. Nye. Comments on Dr. Loewe's letter and notes on crevasses. _J. Glaciol._ , 2(17):512–514, 1955. doi: 10.3189/S0022143000032652. * Olgin et al. (2011) J. G. Olgin, B. R. Smith-Konter, and R. T. Pappalardo. Limits of Enceladus's ice shell thickness from tidally driven tiger stripe shear failure. _Geophys. Res. Lett._ , 38:L02201, 2011. doi: 10.1029/2010GL044950. * Parkinson et al. (2020) J. R. G. Parkinson, D. F. Martin, A. J. Wells, and R. F. Katz. Modelling binary alloy solidification with adaptive mesh refinement. _J. Comput. Phys._ , 5:100043, 2020. doi: 10.1016/j.jcpx.2019.100043. * Polashenski et al. (2017) C. Polashenski, K. M. Golden, D. K. Perovich, E. Skyllingstad, Alexandra Arnsten, C. Stwertka, and N. Wright. Percolation blockage: A process that enables melt pond formation on first year arctic sea ice. _J. Geophys. Res._ , 122(1):413–440, 2017. doi: 10.1002/2016JC011994. * Rudolph et al. (2022) M. L. Rudolph, M. Manga, M. Walker, and A. R. Rhoden. Cooling crusts create concomitant cryovolcanic cracks. _Geophys. Res. Lett._ , 49(5):e2021GL094421, 2022\. doi: 10.1029/2021GL094421. * Schulson and Duval (2009) E. M. Schulson and P. Duval. _Creep and fracture of ice_. Cambridge University Press, 2009. * Shibley and Laughlin (2021) N. C. Shibley and G. Laughlin. Do oceanic convection and clathrate dissociation drive Europa’s geysers? _Planet. Sci. J._ , 2(6):221, 2021. doi: 10.1029/2022GL098621. * Smith (1976) R. A. Smith. The application of fracture mechanics to the problem of crevasse penetration. _J. Glaciol._ , 17(76):223–228, 1976. doi: 10.3198/1976JoG17-76-223-228. * Smith-Konter and Pappalardo (2008) B. Smith-Konter and R. T. Pappalardo. Tidally driven stress accumulation and shear failure of Enceladus's tiger stripes. _Icarus_ , 198(2):435–451, 2008. doi: 10.1016/j.icarus.2008.07.005. * Spencer and Nimmo (2013) J. R. Spencer and F. Nimmo. Enceladus: An active ice world in the Saturn system. _Annu. Rev. Earth Planet. Sci._ , 41(1):693–717, 2013. doi: 10.1146/annurev-earth-050212-124025. * Stevenson (1996) D. J. Stevenson. Heterogeneous tidal deformation and geysers on Europa. In _Europa Ocean Conference_ , number 5 in Capistrano conference, pages 69–70, San Juan Capistrano, Calif., 1996. San Juan Capistrano Res. Inst. * Tait and Jaupart (1992) S. Tait and C. Jaupart. Compositional convection in a reactive crystalline mush and melt differentiation. _J. Geophys. Res._ , 97(B5):6735–6756, 1992. doi: 10.1029/92JB00016. * Timco and Frederking (1996) G. W. Timco and R. M. W. Frederking. A review of sea ice density. _Cold Reg. Sci. Technol._ , 24(1):1–6, 1996. ISSN 0165-232X. doi: 10.1016/0165-232X(95)00007-X. * Walker et al. (2021) C. C. Walker, J. N. Bassis, and B. E. Schmidt. Propagation of vertical fractures through planetary ice shells: The role of basal fractures at the ice–ocean interface and proximal cracks. _Planet. Sci. J._ , 2(4):135, 2021. doi: 10.3847/PSJ/ac01ee. * Wells et al. (2019) A. J. Wells, J. R. Hitchen, and J. R. G. Parkinson. Mushy-layer growth and convection, with application to sea ice. _Philos. Trans. R. Soc. London, Ser. A_ , 377(2146):20180165, 2019. doi: 10.1098/rsta.2018.0165.
# Gravitational wave fossils in nonlinear regime: halo tidal bias and intrinsic alignments from gravitational wave separate universe simulations Kazuyuki Akitsu<EMAIL_ADDRESS>School of Natural Sciences, Institute for Advanced Study, 1 Einstein Drive, Princeton, NJ 08540, USA Yin Li Center for Computational Astrophysics & Center for Computational Mathematics, Flatiron Institute, 162 5th Avenue, New York, NY 10010, USA Teppei Okumura Institute of Astronomy and Astrophysics, Academia Sinica, No. 1, Sec. 4, Roosevelt Road, Taipei 10617, Taiwan Kavli Institute for the Physics and Mathematics of the Universe (WPI), UTIAS, The University of Tokyo, 5-1-5 Kashiwanoha, Kashiwa, Chiba 277-8583, Japan ###### Abstract We investigate impacts of long-wavelength gravitational waves (GWs) on nonlinear structure formation by utilizing the tidal separate universe simulations. Based on the equivalence of a long-wavelength GW to a uniform tidal field in a local frame, we provide a way to incorporate a long- wavelength GW into the tidal separate universe simulation as an effective anisotropic expansion. This methodology enables us to study effects of GWs on large-scale structure efficiently. We measure the anisotropic imprint in the local power spectrum from the tidal separate universe simulations with GWs, which corresponds to the scalar-scalar-tensor bispectrum in squeezed limit or the so-called power spectrum response to GWs. We also detect the halo tidal bias induced by GWs from the response of the halo-matter cross-power spectrum to GWs, as well as the linear shape bias (or the linear alignment coefficient) induced by GWs from the one-point function of the halo ellipticity. In contrast to the case of the tidal field induced by scalar perturbations, we discover that the wavenumber dependence of the temporal evolution of GWs naturally causes these biases to be scale-dependent. We also find that this scale dependence is well approximated by the second-order density induced by the coupling between scalar and tensor perturbation. This highlights that the structure formation, especially the process to determine the halo shape, is nonlocal in time. Our findings lay the foundation for predicting the impact of GWs on large-scale structure. ## I Introduction Gravitational waves (GWs) serve as vital means to observe the universe. In particular, since long-wavelength GWs in the Gpc-Mpc range are unlikely to originate from astrophysical events, they are thought to have a cosmological origin (e.g., inflation Starobinsky (1979); Sato (1981); Guth (1981); Linde (1982); Maggiore (2000)), making them an important probe in cosmology. In spite of the various experiments trying to detect the primordial $B$-mode signal in the polarization of the Cosmic Microwave Background (CMB) Komatsu _et al._ (2011); Akrami _et al._ (2020); Ade _et al._ (2018); Adachi _et al._ (2022); Abazajian _et al._ (2016), which is one of the most powerful methods to hunt for long-wavelength cosmological GWs Zaldarriaga and Seljak (1997); Kamionkowski _et al._ (1997), such GWs have not been observed yet. Compared to the numerous studies for the effect of GWs on the perturbations of the CMB, that on large-scale structure (LSS) of the universe has not received as much attention. As large-scale structure is dominantly sourced by scalar perturbations, there is a long history for the study on how the scalar perturbations have shaped LSS, including the linear and nonlinear perturbation theory Dodelson (2003); Suto and Sasaki (1991); Makino _et al._ (1992); Bernardeau _et al._ (2002); Baumann _et al._ (2012) and the $N$-body simulation Springel (2005); Potter _et al._ (2016); Springel _et al._ (2021); Angulo and Hahn (2021). On the other hand, there are fewer studies of the effect of the tensor perturbations (GWs) on nonlinear structure formation, as we list below. There are two types of effects of GWs on LSS observables: the dynamical effect and the projection effect. The former refers to the effect of GWs on nonlinear structure formation itself, whereas the latter refers to the effect of GWs on the light path emitted from distant galaxies to us. In other words, the dynamical effect would be observed by a comoving observer in a local frame, while the projection effect comes from the mapping of observables from the galaxy’s local frame to our local frame (at the earth), which includes the Sachs-Wolfe effect and the gravitational lensing effect caused by GWs. Both effects have been studied by means of the perturbation theory; Refs. Kaiser and Jaffe (1997); Dodelson _et al._ (2003); Yoo _et al._ (2009); Jeong _et al._ (2012); Schmidt and Jeong (2012a); Jeong and Schmidt (2012) formulated the projection effect on the galaxy clustering and the galaxy shape (shear) by GWs, while Ref Schmidt and Jeong (2012a) pointed out that long-wavelength GWs can also contribute the intrinsic alignments of galaxy shapes and Refs. Masui and Pen (2010); Dai _et al._ (2013); Schmidt _et al._ (2014) computed the second order matter density contrast induced by the coupling between scalar perturbations and long-wavelength tensor perturbations (GWs). For the projection effect the perturbative treatment would be adequate. For the dynamical effect, however, the perturbation theory breaks down in nonlinear scales and thus the nonlinear nature of LSS requires $N$-body simulations with GWs to capture the fully nonlinear impact of GWs on structure formation. Furthermore, given that the biased tracers of LSS such as halos are themselves nonlinear objects, such simulations are necessary to understand their biases to GWs even in the linear regime. An $N$-body simulation with GWs is generally challenging because the usual $N$-body simulations are based on Newtonian gravitational dynamics in an expanding background whereas GWs are a purely general-relativistic effect. The most straightforward way to introduce GWs into $N$-body simulations is to develop a general-relativistic cosmological simulation Adamek _et al._ (2013, 2016a, 2016b), though Refs. Adamek _et al._ (2016a, b) only considered the second-order (induced) tensor perturbations by scalar perturbations.111 Technically the induced tensor perturbations computed in Refs. Adamek _et al._ (2016a, b) involve non-GWs contributions, which are tensor modes but not propagating waves. See e.g., Ref. Domènech and Sasaki (2021) for detailed discussion on this issue. Also, since GWs are much smaller than scalar perturbations, it is difficult to single out the effect of GWs on structure formation from that of the scalar perturbations in this sort of simulations. In this paper, we utilize a separate universe approach to circumvent this issue on $N$-body simulations with GWs. In the separate universe approach, the influence of a long-wavelength perturbation is absorbed into the cosmic expansion observed in the local frame, thereby the local expansion becomes different from the global one. Accordingly, nonlinear structure formation in the local region responds to this difference in the background expansions. Using this technique, the response to the long-wavelength perturbation can be accurately measured in $N$-body simulations Sirko (2005); Gnedin _et al._ (2011); Baldauf _et al._ (2011); Li _et al._ (2014); Wagner _et al._ (2015); Lazeyras _et al._ (2016); Li _et al._ (2016); Baldauf _et al._ (2016). Recently, the anisotropic extension of the separate universe simulation was developed in Refs. Stücker _et al._ (2021); Masaki _et al._ (2020); Akitsu _et al._ (2021a). They considered a long-wavelength tidal perturbation sourced by the long-wavelength scalar tides $\propto\left(\partial_{i}\partial_{j}-\frac{1}{3}\delta^{\textsc{k}}_{ij}\right)\Phi$ Dai _et al._ (2015a); Akitsu _et al._ (2017); Akitsu and Takada (2018); Akitsu _et al._ (2019); Li _et al._ (2018); Taruya and Akitsu (2021), and the perturbation was absorbed into the local background, making the local cosmic expansion anisotropic. Taking advantage of the equivalence of a long-wavelength GW to a long- wavelength tidal field in a local region, we apply this tidal separate universe simulation to measure the impact of GWs on structure formation. In contrast to the scalar case, the time evolution of GWs is scale-dependent (or wavenumber-dependent). As a result, when mimicking the long-wavelength GW as the local anisotropic expansion, the anisotropic expansion rate would be different depending on the wavenumber of GWs, and so is the response of large- scale structure, as considered for isotropic scale-dependent long-wavelength perturbation in Refs. Hu _et al._ (2016); Chiang _et al._ (2016, 2018). The purpose of this paper is to generalize the tidal separate universe simulation for GWs in order to study the scale-dependent responses for long-wavelength GWs of different wavelengths. The remainder of this paper provides a way to implement long-wavelength GWs into the tidal separate universe simulation and presents imprints of GWs on large-scale structure measured from newly developed simulations. After giving a brief review about the local coordinates in the presence of GWs and perturbative results in Sec. II, we construct the tidal separate universe with GWs in Sec. III. In Sec. IV, we measure the power spectrum responses for matter auto-, halo-matter cross-, and halo auto-power spectra, which is related to the scalar-scalar-tensor bispectrum in squeezed limit. Sec. V and Sec. VI are devoted to the measurements of the halo tidal bias and linear alignment coefficient (or the linear shape bias) from GWs, respectively. We discuss possible observables in large-scale structure to probe GWs in Sec. VII. Throughout this paper we adopt cosmological parameters consistent with Planck result: $\Omega_{\rm r0}=4.1577\times 10^{-5}$, $\Omega_{\rm m0}=0.3089$, $\Omega_{\Lambda 0}=0.6911$, $H_{0}=67.74$ Ade _et al._ (2016). ## II Local frame and the perturbative results Here we first introduce the local coordinates in the presence of long- wavelength GWs. We then briefly summarize the derivation of the second-order density perturbations induced by the coupling between scalar and tensor perturbations (GWs), following Ref. Schmidt _et al._ (2014). We employ the Lagrangian perturbation formalism, which can be straightforwardly used to construct the tidal separate universe in the next section. ### II.1 Long-wavelength gravitational waves in the conformal Fermi coordinates Figure 1: Left panels from top to bottom: Transfer function of GWs $\mathcal{T}(a;k_{L})$, growth coefficient $\alpha(a;k_{L})$ and dilation coefficient $\beta(a;k_{L})$ in Eq. (29) as a function of scale factor $a$ for various wavenumbers $k_{L}$. Right panels from top to bottom: Same as the left panels but as a function of wavenumber of GWs, $k_{L}$, at various redshifts. Note that the functional shape of $\beta(a;k_{L})$ is equivalent to that of $\mathcal{T}(a;k_{L})$ because $\beta(a;k_{L})=-\frac{1}{2}\left[\mathcal{T}(a;k_{L})-1\right]$. In the cosmological context, GWs are defined as the trace-free transverse components in the perturbed FLRW metric, $\displaystyle ds^{2}=a^{2}[-d\eta^{2}+(\delta^{\textsc{k}}_{ij}+h_{ij})dx^{i}dx^{j}],$ (1) where $a$ is the scale factor, $\eta$ is the conformal time, $\delta^{\textsc{k}}_{ij}$ is Kronecker’s delta, and $h_{ij}$ is GWs satisfying $\tensor{h}{{}^{i}_{i}}=0$ and $\partial^{i}h_{ij}=0$. We use $k_{L}$ to denote the wavenumber of GWs in what follows since we focus on the long-wavelength GWs in this paper. We introduce the transfer function of GWs, $\mathcal{T}(\eta;k_{L})$, through $\displaystyle h_{ij}(\eta;{\bf k}_{L})=\mathcal{T}(\eta;k_{L})h^{\rm ini}_{ij}({\bf k}_{L}),$ (2) where $h_{ij}^{\rm ini}(k_{L})$ denotes the primordial value of GWs, i.e., $h_{ij}^{\rm ini}(k_{L})=h_{ij}(0;k_{L})$. The transfer function $\mathcal{T}(\eta;k_{L})$ obeys $\displaystyle\mathcal{T}^{\prime\prime}(\eta;k_{L})+2\mathcal{H}\mathcal{T}^{\prime}(\eta;k_{L})+k_{L}^{2}\mathcal{T}(\eta;k_{L})=0,$ (3) with ${}^{\prime}=\differential/\differential\eta$, and $\mathcal{H}=aH$ being the conformal Hubble paramter. The top-left panel of Fig. 1 shows $\mathcal{T}(\eta;k_{L})$ as a function of the scale factor for various wavenumbers of GWs. Similarly, the top-right panel shows that as a function of the wavenumbers of GWs at various redshifts. For any $k_{L}$, $\mathcal{T}(\eta;k_{L})$ remains unity when $k_{L}\eta\ll 1$, which means that GWs are frozen before they enter the horizon. For later convenience, we expand GWs by the helicity basis: $\displaystyle h_{ij}(\eta;{\bf k}_{L})=\sum_{\lambda=\pm 2}h_{(\lambda)}(\eta;{\bf k}_{L})e^{(\lambda)}_{ij}(\hat{k}_{L}),$ (4) where $e_{ij}^{(\pm 2)}\equiv e_{i}^{(\pm)}e_{j}^{(\pm)}$ and ${\bf e}^{(\pm)}\equiv({\bf e}_{1}\mp i{\bf e}_{2})/\sqrt{2}$ with $\\{{\bf k}_{L},{\bf e}_{1},{\bf e}_{2}\\}$ being an orthonormal set. The helicity basis satisfies $e^{(\lambda)}_{ij}e^{ij}_{(-\lambda)}=1$, $e^{(\lambda)}_{ij}e^{ij}_{(\lambda)}=0$, and $h_{(\lambda)}=e^{ij}_{(-\lambda)}h_{ij}$. The power spectrum of GWs for each helicity mode is defined as $\displaystyle\langle h_{(\lambda)}({\bf k}_{L};\eta)h^{*}_{(\lambda)}({\bf k}^{\prime}_{L};\eta^{\prime})\rangle=(2\pi)^{3}\delta_{\rm D}^{(3)}({\bf k}_{L}-{\bf k}^{\prime}_{L})P_{h_{(\lambda)}}(k_{L};\eta,\eta^{\prime}),$ (5) which is related to the primordial power spectrum $P_{h_{(\lambda)}}(k_{L})$ through $\displaystyle P_{h_{(\lambda)}}(k_{L};\eta,\eta^{\prime})=\mathcal{T}(\eta;k_{L})\mathcal{T}(\eta^{\prime};k_{L})P_{h_{(\lambda)}}(k_{L}).$ (6) For unpolarized GWs, the two power spectra has equal power, i.e., $P_{h_{(+2)}}(k)=P_{h_{(-2)}}(k)\equiv P_{h}(k)/2$ with $P_{h}(k)$ being the total power spectrum of GWs. For chiral GWs, the chiral parameter $\chi(k)$ defined as $\displaystyle\chi(k)\equiv\frac{P_{(+2)}(k)-P_{(-2)}(k)}{P_{h}(k)},$ (7) measures the degree of the parity-breaking in GWs. The total power spectrum of GWs is often characterized via $\displaystyle\frac{k^{3}P_{h}(k)}{2\pi^{2}}\equiv rA_{s}\left(\frac{k}{k_{*}}\right)^{n_{T}}$ (8) with $r$ being the scalar-tensor ratio, $A_{s}$ being the amplitude for the primordial curvature perturbations at the pivot scale $k_{*}$, and $n_{T}$ being the tensor tilt. To investigate the physical effects of long-wavelength GWs on scalar perturbations at smaller scales, consider a local region centered at the timelike geodesic of a comoving observer. In this local patch we can construct the so-called conformal Fermi coordinates (CFC), which is an extension of the Fermi normal coordinates Manasse and Misner (1963), developed in Refs. Schmidt _et al._ (2014); Dai _et al._ (2015b). The metric of this coordinates, $g_{\mu\nu}^{F}$, takes the form of the FLRW metric along the central geodesic with leading-order corrections of ${\cal O}(x^{2}_{F})$. We are interested in the interaction between long-wavelength GWs and non-relativistic matter. In this case $g_{00}^{F}$ encodes all the relevant impact from the long- wavelength perturbations because the dynamics of non-relativistic matter is solely determined by the usual Newtonian potential in $g_{00}^{F}$. One can show that given the global coordinates Eq. (1), $g_{00}^{F}$ is computed as $\displaystyle g_{00}^{F}=-a^{2}\left[1+\tau_{ij}x_{F}^{i}x_{F}^{j}\right],$ (9) with $\displaystyle\tau_{ij}(\eta;k_{L})=-\frac{1}{2}\left[a^{-1}\left(ah_{ij}^{\prime}\right)^{\prime}\right]=-\frac{1}{2}\left(h_{ij}^{\prime\prime}+\mathcal{H}h_{ij}^{\prime}\right)\equiv T(\eta;k_{L})h^{\rm ini}_{ij}$ (10) representing the effective tidal field induced by the long-wavelength GWs in the local region. The derivation is summarized in App. A. It is worth noting that $\tau_{ij}$ takes effect only after GWs cross the horizon as implied by the time derivative on $h_{ij}$. Note that $x_{F}^{i}$ corresponds to the comoving distance, unlike the physical (proper) distance in the Fermi normal coordinates. In what follows, we work in this frame and drop the subscript $F$ in $x_{F}^{i}$, namely, denote $x^{i}$ as $x_{F}^{i}$. ### II.2 Second order density induced by the interaction between GWs and scalar perturbations The equation of motion of a matter particle in the local frame is 222Strictly speaking, this is justified by considering the geodesic equation in the local frame. $\displaystyle\frac{\differential^{2}r_{i}}{\differential t^{2}}=\frac{1}{a^{2}}\left[\frac{\differential^{2}}{\differential\eta^{2}}-\mathcal{H}\frac{\differential}{\differential\eta}\right]r_{i}=-\frac{\partial}{\partial r_{i}}\left(\Phi_{\rm iso}+\phi\right),$ (11) where $r_{i}=ax_{i}$ and we split the gravitational potential into a background potential $\Phi_{\rm iso}$ and a peculiar potential $\phi$. The subscript “iso” in $\Phi$ stands for the potential sourced only by the usual isotropic background, $\displaystyle\Phi_{\rm iso}\equiv\frac{2}{3}\pi G\bar{\rho}_{\rm m}r^{2}-\frac{\Lambda}{6}r^{2},$ (12) where $\bar{\rho}_{\rm m}$ is the mean density of matter. The peculiar potential $\phi$ includes the potential sourced by local inhomogeneities $\phi_{s}$ as well as the effective tidal field sourced by the long-wave gravitational wave, $\phi=\phi_{s}+\frac{1}{2}\tau_{ij}x^{i}x^{j}$. In the next sections, the effective tidal potential is absorbed into the back ground potential, $\Phi=\Phi_{\rm iso}+\frac{1}{2}\tau_{ij}x^{i}x^{j}$ (see Eq. (33) below). The potential sourced by local inhomogeneities, $\phi_{s}$, satisfies the Poisson equation $\displaystyle\nabla_{r}^{2}\phi_{s}=4\pi G\bar{\rho}_{\rm m}\delta=\frac{3}{2}\frac{\Omega_{\rm m}(\eta)\mathcal{H}^{2}}{a^{2}}\delta,$ (13) where $\delta=\rho/\bar{\rho}_{\mathrm{m}}-1$ denotes the overdensity field. We can also decompose the left-hand side of Eq. (11) into the background and peculiar parts, resulting in $\displaystyle\mathcal{H}^{\prime}x_{i}$ $\displaystyle=-\frac{\partial}{\partial x_{i}}\Phi_{\rm iso},$ (14) $\displaystyle x_{i}^{\prime\prime}+\mathcal{H}x_{i}^{\prime}$ $\displaystyle=-\frac{\partial}{\partial x_{i}}\left(\phi_{s}+\frac{1}{2}\tau_{\ell m}x^{\ell}x^{m}\right).$ (15) One can verify that Eq. (14) with Eq. (12) gives the usual Friedmann equation. Eq. (15) can be regarded as the evolution equation for the displacement field $\Psi_{i}$, which relates Lagrangian position $q_{i}$ to Eulerian position $x_{i}$ via $x_{i}=q_{i}+\Psi_{i}$. Before shell-crossing the Jacobian determinant of this mapping gives the overdensity as $\displaystyle\delta=\left|\frac{\partial x_{i}}{\partial q_{j}}\right|^{-1}-1=\left|\delta^{\textsc{k}}_{ij}+\frac{\partial\Psi_{i}}{\partial q_{j}}\right|^{-1}-1.$ (16) At linear order this reduces to $\delta^{(1)}=-\partial\Psi^{(1)}_{i}/\partial q_{i}$. We split the linear displacement into the one sourced by pure scalar contributions and that sourced by $\tau_{ij}$, each of which satisfies $\displaystyle\Psi^{(1)\prime\prime}_{s,i}+\mathcal{H}\Psi^{(1)\prime}_{s,i}$ $\displaystyle=-\frac{3}{2}\Omega_{m}(\eta)\mathcal{H}^{2}\frac{\partial_{q}^{i}}{\partial^{2}_{q}}\delta^{(1)},$ (17) $\displaystyle\Psi^{(1)\prime\prime}_{t,i}+\mathcal{H}\Psi^{(1)\prime}_{t,i}$ $\displaystyle=-\frac{1}{2}\partial_{q}^{i}\left[\tau_{kl}q^{k}q^{l}\right],$ (18) where $\Psi^{(1)}_{s,i}$ and $\Psi^{(1)}_{t,i}$ denote the linear displacement caused by the scalar and tensor perturbations, respectively, and we have used Eq. (13) and the fact that $x_{i}$ can be replaced by $q_{i}$ at linear order. The non-decaying solutions for $\Psi^{(1)}_{s,i}$ and $\Psi^{(1)}_{t,i}$ are $\displaystyle\Psi^{(1)}_{s,i}(\eta)$ $\displaystyle=-\frac{D(\eta)}{D(\eta_{0})}\frac{\partial_{q}^{i}}{\partial_{q}^{2}}\delta^{(1)}(\eta_{0}),$ (19) $\displaystyle\Psi^{(1)}_{t,i}(\eta;k_{L})$ $\displaystyle=\frac{1}{2}\left[\mathcal{T}(\eta;k_{L})-1\right]h^{\rm ini}_{ij}(k_{L})q^{j}\equiv-\beta(\eta;k_{L})h_{ij}^{\rm ini}(k_{L})q^{j},$ (20) where we have used the boundary condition $\lim_{\eta\to 0}\mathcal{T}(\eta;k_{L})=1$ and have introduced the quantity $\beta(\eta;k_{L})$ to characterize the linear displacement due to the tensor perturbations, and $D(\eta)$ represents the linear growth function, which follows $\displaystyle D^{\prime\prime}(\eta)+\mathcal{H}D^{\prime}(\eta)-\frac{3}{2}\Omega_{\rm m}(\eta)\mathcal{H}^{2}D(\eta)=0.$ (21) From this result one can confirm that GWs do not induce the linear density: $\delta^{(1)}_{t}=-\partial\Psi^{(1)}_{t,i}/\partial q_{i}\propto h_{i}^{\ i}=0$, as expected. At second order, however, GWs do affect the density field as well as the displacement field. Expanding Eq. (16) up to second order leads to $\displaystyle\delta^{(1)}({\bf x}({\bf q}))+\delta^{(2)}_{st}({\bf x}({\bf q}))=$ $\displaystyle\delta^{(1)}({\bf q})-\left.\frac{\partial\Psi^{(2)}_{st,i}}{\partial q_{i}}\right|_{{\bf q}}+\left.\frac{\partial\Psi^{(1)}_{s,i}}{\partial q_{j}}\frac{\partial\Psi^{(1)}_{t,j}}{\partial q_{i}}\right|_{{\bf q}}$ $\displaystyle=$ $\displaystyle\delta^{(1)}({\bf x})-\Psi_{t,i}^{(1)}({\bf x})\partial_{x}^{i}\delta^{(1)}({\bf x})-\left.\frac{\partial\Psi^{(2)}_{st,i}}{\partial q_{i}}\right|_{{\bf q}={\bf x}}+\left.\frac{\partial\Psi^{(1)}_{s,i}}{\partial q_{j}}\frac{\partial\Psi^{(1)}_{t,j}}{\partial q_{i}}\right|_{{\bf q}={\bf x}},$ (22) where we focus on the second-order density arising from the coupling between the scalar and tensor modes, and $\delta_{st,i}^{(2)}$ and $\Psi_{st,i}^{(2)}$ respectively denote the second-order density and displacement induced by the coupling. Namely, we omit the second-order contributions from the auto- coupling between scalars (irrelevant) or tensors (subdominant). In deriving the first equality above, we have used the fact that $\Psi^{(1)}_{t,i}$ is divergence free. Taking the divergence of Eq. (15) with respect to ${\bf q}$ we obtain $\displaystyle\psi_{st}^{(2)\prime\prime}+{\cal H}\psi_{st}^{(2)\prime}=-\nabla_{x}^{2}\phi_{s}-\frac{\partial\Psi^{(1)}_{t,j}}{\partial q_{i}}\frac{\partial^{2}\phi_{s}}{\partial x_{i}\partial x_{j}}-\frac{\partial\Psi^{(1)}_{s,j}}{\partial q_{i}}\tau_{ij},$ (23) where we have defined $\psi^{(2)}_{st}\equiv\partial\Psi_{st,i}^{(2)}/\partial q_{i}$ and used $\partial_{q}^{i}=\partial_{x}^{i}+\partial_{q}^{i}\Psi_{j}\partial_{q}^{j}\simeq\partial_{x}^{i}+\partial_{q}^{i}\Psi_{j}\partial_{x}^{j}$. Rewriting $\phi_{s}$ in terms of $\delta$ by using Eq. (13) and using Eq. (22) yield $\displaystyle\psi_{st}^{(2)\prime\prime}+\mathcal{H}\psi_{st}^{(2)\prime}=\frac{3}{2}\Omega_{\rm m}\mathcal{H}^{2}\left[-\frac{\partial\Psi^{(2)}_{st,i}}{\partial q_{i}}+\frac{\partial\Psi^{(1)}_{s,i}}{\partial q_{j}}\frac{\partial\Psi^{(1)}_{t,j}}{\partial q_{i}}\right]-\frac{3}{2}\Omega_{\rm m}\mathcal{H}^{2}\frac{\partial\Psi^{(1)}_{t,j}}{\partial q_{i}}\frac{\partial\Psi^{(1)}_{s,i}}{\partial q_{j}}-\frac{\partial\Psi^{(1)}_{s,j}}{\partial q_{i}}\tau_{ij}.$ (24) Finally the equation for $\psi_{st}^{(2)}$ is found to be $\displaystyle\psi_{st}^{(2)\prime\prime}+\mathcal{H}\psi_{st}^{(2)\prime}-\frac{3}{2}\Omega_{\rm m}\mathcal{H}^{2}\psi_{st}^{(2)}=\left(\frac{\partial^{i}_{q}\partial^{j}_{q}}{\partial_{q}^{2}}\delta^{(1)}(\eta)\right)\tau_{ij}.$ (25) We can write the solution as $\displaystyle\psi_{st}^{(2)}({\bf q},\eta;k_{L})=D_{st}^{(2)}(\eta;k_{L})\left(\frac{\partial^{i}_{q}\partial^{j}_{q}}{\partial_{q}^{2}}\delta^{(1)}({\bf q},\eta_{0})\right)h_{ij}^{\rm ini}(k_{L}),$ (26) where the time-dependent part of $\psi_{st}^{(2)}({\bf q},\eta;k_{L})$, which we write $D_{st}^{(2)}(\eta;k_{L})$, satisfies $\displaystyle D_{st}^{(2)\prime\prime}(\eta;k_{L})+\mathcal{H}D_{st}^{(2)\prime}(\eta;k_{L})-\frac{3}{2}\Omega_{\rm m}(\eta)\mathcal{H}^{2}D_{st}^{(2)}(\eta;k_{L})=\frac{D(\eta)}{D(\eta_{0})}\cdot\left[-\frac{1}{2a}\left(a\mathcal{T}^{\prime}(\eta;k_{L})\right)^{\prime}\right].$ (27) We can solve this equation numerically given the cosmological parameters but here we derive the analytic solution assuming the matter domination for the later convenience. In the matter-dominated era where we have $\Omega_{\rm m}=1$, $\mathcal{H}=2/\eta$ and $D(\eta)\propto a\propto\eta^{2}$, we can write the solution using Green’s function: $\displaystyle D_{st}^{(2)}(\eta;k_{L})=$ $\displaystyle\frac{1}{D(\eta_{0})}\int_{0}^{\eta}\differential\tilde{\eta}~{}\frac{1}{5}\left[\frac{\eta^{2}}{\tilde{\eta}}-\frac{\tilde{\eta}^{4}}{\eta^{3}}\right]D(\tilde{\eta})\left[-\frac{1}{2a}\left[a\mathcal{T}^{\prime}(\tilde{\eta};k_{L})\right]^{\prime}\right]$ $\displaystyle=$ $\displaystyle\frac{1}{5}\frac{D(\eta)}{D(\eta_{0})}\left[\beta(\eta)+4\int_{0}^{\eta}\differential\tilde{\eta}\left(\frac{\tilde{\eta}}{\eta}\right)^{5}\beta^{\prime}(\tilde{\eta};k_{L})\right].$ (28) Finally we find the second order density induced by the interaction between the scalar and tensor perturbation as $\displaystyle\delta^{(2)}_{st}({\bf x},\eta;k_{L})=$ $\displaystyle h_{ij}^{\rm ini}\left[\beta(\eta;k_{L})x^{j}\partial_{x}^{i}\delta^{(1)}({\bf x},\eta)-D_{st}^{(2)}(\eta;k_{L})\left(\frac{\partial^{i}_{x}\partial^{j}_{x}}{\partial_{x}^{2}}\delta^{(1)}({\bf x},\eta_{0})\right)+\beta(\eta;k_{L})\left(\frac{\partial^{i}_{x}\partial^{j}_{x}}{\partial_{x}^{2}}\delta^{(1)}({\bf x},\eta)\right)\right]$ $\displaystyle=$ $\displaystyle h_{ij}^{\rm ini}\left[\alpha(\eta;k_{L})\frac{\partial^{i}_{x}\partial^{j}_{x}}{\partial_{x}^{2}}+\beta(\eta;k_{L})x^{j}\partial_{x}^{i}\right]\delta^{(1)}({\bf x},\eta),$ (29) where $\beta$ is defined in Eq. (20) and $\alpha$ is defined as $\displaystyle\alpha(\eta;k_{L})\equiv$ $\displaystyle-\frac{D(\eta_{0})}{D(\eta)}D_{st}^{(2)}(\eta;k_{L})+\beta(\eta;k_{L})$ $\displaystyle=$ $\displaystyle\frac{4}{5}\left(\beta(\eta;k_{L})-\int_{0}^{\eta}\differential\tilde{\eta}\left(\frac{\tilde{\eta}}{\eta}\right)^{5}\beta^{\prime}(\tilde{\eta};k_{L})\right).$ (30) Eq. (29) holds not only during the matter domination but also the entire history of the universe as long as we solve Eq. (27) numerically, whereas the second equality of Eq. (30) holds only for the matter domination. The first term in the last line $(\propto\alpha)$ represents the changes in the short- mode amplitude, called the growth effect, by the coupling between the scalar tidal field and long-wavelength GWs. The second term in the last line $(\propto\beta)$ represents the coordinate shift induced by the long- wavelength GWs, known as the dilation effect. The power spectrum of the density field in the presence of the long-wavelength GWs then becomes $\displaystyle P_{\rm mm}({\bf k_{S}},\eta|h_{ij}(k_{L}))=P_{\rm lin}(k_{S})\left[1+\hat{k}_{S}^{i}\hat{k}_{S}^{j}h_{ij}^{\rm ini}(k_{L})\left(2\alpha(\eta;k_{L})-\beta(\eta;k_{L})\frac{\partial\ln{P_{\rm lin}(k_{S},\eta)}}{\partial\ln{k_{S}}}\right)\right].$ (31) We show $\mathcal{T}(k_{L};\eta)$, $\alpha(k_{L};\eta)$ and $\beta(k_{L};\eta)$ as a function of the scale factor for various wavenumbers of GWs (the left panels) and as a function of the wavemunber of GWs at various redshifts at the right and left panels of Fig.1, respectively. To plot these functions, we numerically integrated Eqs. (3) and (27) without assuming the matter domination. In the limit of $k_{L}\eta\to 0$, it is obvious that there is no physical effect from GWs, i.e., $\alpha(\eta;k_{L})\to 0$ and $\beta(\eta;k_{L})\to 0$, since GWs are frozen on the super-horizon scales. On the other hand, taking the limit $k_{L}\eta\gg 1$ is more interesting because in this limit $\mathcal{T}(\eta;k_{L})\to 0$ but $\alpha(\eta;k_{L})$ and $\beta(\eta;k_{L})$ do not vanish. In other words, even long after GWs have decayed away, their impact on the growth and displacement remains, which is sometimes called the “fossil” effect Masui and Pen (2010). ## III Tidal separate universe with gravitational waves In this section, we describe how to incorporate long-wavelength GWs into the simulation background with the help of the tidal separate universe simulation technique developed in Ref. Akitsu _et al._ (2021a) (see also Refs. Stücker _et al._ (2021); Masaki _et al._ (2020)). We focus on differences arising from GWs and refer the readers to Ref. Akitsu _et al._ (2021a) about the details of the implementation. In this section we do not employ Einstein’s summation convention to avoid confusions. ### III.1 Anisotropic background In the tidal separate universe simulation we introduce anisotropic scale factors by absorbing the long-wavelength tidal perturbations into the background. In general, reflecting that the tidal perturbations are expressed by the $3\times 3$ symmetric matrix $\tau_{ij}$ the anisotropic scale factors are also written by the $3\times 3$ symmetric matrix $a_{ij}$, which relates the physical coordinate $r_{i}$ to the comoving coordinate $x_{i}$ as $r_{i}=\sum_{j}a_{ij}x_{j}$. However, we can always rotate the simulation coordinates to align with the eigenvectors of $\tau_{ij}$, leaving only the diagonal components non-zero so that $\tau_{ij}=\tau_{i}\delta^{\textsc{k}}_{ij}$ and $a_{ij}=a_{i}\delta^{\textsc{k}}_{ij}$. Now we have different scale factors for each axis and characterize these differences by $\Delta_{i}$ defined via $\displaystyle a_{i}=a(1+\Delta_{i}),$ (32) with $a$ being the global scale factor. While the equation of motion remains the same as in Sec. II.2, now we want to absorb the long-wavelength effective tidal potential induced by GWs into the background potential so that $\displaystyle\Phi=\Phi_{\rm iso}+\frac{1}{2}\sum_{i}\tau_{i}x_{i}^{2}=\frac{2}{3}\pi G\bar{\rho}_{\rm m}r^{2}-\frac{\Lambda}{6}r^{2}+\frac{1}{2}\sum_{i}\tau_{i}x_{i}^{2}.$ (33) The background and peculiar equations in an anisotropic background become $\displaystyle\frac{1}{a^{2}}\left[a_{i}^{\prime\prime}-\mathcal{H}a_{i}^{\prime}\right]x_{i}=-\frac{1}{a_{i}}\frac{\partial}{\partial x_{i}}\Phi,$ (34) $\displaystyle\frac{1}{a^{2}}\left[a_{i}x_{i}^{\prime\prime}+2a_{i}^{\prime}x_{i}^{\prime}-\mathcal{H}a_{i}x_{i}\right]=-\frac{1}{a_{i}}\frac{\partial}{\partial x_{i}}\phi_{s}.$ (35) Subtracting the isotropic background, which is determined by Eq. (14), from Eq. (34) and linearizing it in $\Delta_{i}$ we find $\displaystyle\Delta^{\prime\prime}_{i}(\eta;k_{L})+\mathcal{H}\Delta_{i}^{\prime}(\eta;k_{L})=\frac{1}{2}a^{-1}(\eta)\left[a(\eta)h_{i}^{\prime}(\eta;k_{L})\right]^{\prime}.$ (36) Note that anisotropic scale factors depend on the wavenumber of GWs as a consequence of the wavenumber dependence of the transfer function of GWs, $\mathcal{T}(\eta;k_{L})$, while they do not for the case of scalar large- scale tidal field where the linear growth of scalar perturbations is independent of their wavenumbers, as characterized by $D(\eta)$. Note also that the source term for the anisotrpic scale factors on the right-hand side is non-zero only when $h_{i}^{\prime}(\eta;k_{L})$ does not vanish; in other words long-wavelength GWs induces the anisotropic scale factors only after GWs enters the horizon as expected. In fact, integrating Eq. (36) twice yields $\displaystyle\Delta_{i}(\eta;k_{L})=\frac{h^{\rm ini}_{i}(k_{L})}{2}\left[\mathcal{T}(\eta;k_{L})-1\right]=-h^{\rm ini}_{i}(k_{L})\beta(\eta;k_{L}),$ (37) which goes to zero when $\eta\to 0$ because $\lim_{\eta\to 0}\mathcal{T}(\eta;k_{L})=1$. The appearance of the function $\beta(\eta;k_{L})$ defined in Eq. (20) is expected and this result can be understood in a more intuitive way. What is specifically done in the separate universe construction is to absorb the displacement caused by the long-wavelength perturbation into the background expansion while keeping the physical distance unchanged. In other words, we introduce the local scale factor $a_{i}$ to satisfy $\displaystyle a_{i}x_{i}=a(x_{i}+\Psi_{i}^{\rm long}),$ (38) which implies $\Delta_{i}x_{i}=\Psi_{i}^{\rm long}$. Given the displacement caused by long-wavelength GWs in Eq. (20), this immediately leads to $\Delta_{i}(\eta;k_{L})=-h_{i}^{\rm ini}(k_{L})\beta(\eta;k_{L})$. We note that this matching only works for the non-relativistic matter. In other words, the effect of long-wavelength GWs can be captured by the anisotropic expansion only when we focus on non-relativistic particles that do not care about $g_{0i}$ and $g_{ij}$ components in the metric. For example, the method presented here is not useful in order to study the impact of long- wavelentgh GWs on the radiation perturbations. However, this treatment is sufficient to study the influence on dark matter particles that we are interested in and consistent with the usual Newtonian $N$-body method. ### III.2 Initial conditions The background anisotropy induces a correction to the 2LPT solution in the isotropic background as discussed in Ref.Akitsu _et al._ (2021a). The correction depends on $\Delta_{i}(\eta)$, which is different for the scalar tidal field and GWs. Here we derive this correction induced by the background anisotropy governed by long-wavelength GWs. The equation for the displacement in the anisotropic background can be obtained by combining Eq. (35) with Eq. (13) as $\displaystyle\sum_{ij}\left|\frac{\partial{\bf x}}{\partial{\bf q}}\right|\left[\delta_{ij}+\Psi_{i,j}\right]^{-1}\left[\Psi_{i,j}^{\prime\prime}+(\mathcal{H}+2\Delta_{i}^{\prime})\Psi_{i,j}^{\prime}\right]=\frac{3}{2}\Omega_{\rm m}\mathcal{H}^{2}\left(\left|\frac{\partial{\bf x}}{\partial{\bf q}}\right|-1\right),$ (39) where we adopt the notation $\Psi_{i,j}\equiv\partial_{q}^{j}\Psi_{i}$ in this subsection. Taking $\Delta_{i}=0$ results in the usual master equation for the LPT. We introduce the correction $\epsilon^{(1)}_{i}$ of order ${\cal O}(\delta^{(1)}\Delta_{i})$ with $\delta^{(1)}$ being short-wavelength modes in the simulations: $\displaystyle\Psi_{i}=\Psi_{i}^{(1)}+\Psi_{i}^{(2)}+\epsilon^{(1)}_{i},$ (40) where $\Psi_{i}^{(1)}+\Psi_{i}^{(2)}$ is the usual 2LPT solution in the isotropic background. In the following we also introduce the potentials such that $\Psi_{i}^{(1)}=-\psi^{(1)}_{,i}$ and $\epsilon_{i}^{(1)}=-\epsilon^{(1)}_{,i}$ for the convenience. The equation for $\epsilon^{(1)}$ can be found as $\displaystyle\sum_{i}\epsilon_{,ii}^{(1)\prime\prime}+\mathcal{H}\sum_{i}\epsilon_{,ii}^{(1)\prime}-\frac{3}{2}\Omega_{\rm m}\mathcal{H}^{2}\sum_{i}\epsilon_{,ii}^{(1)}=-2\sum_{i}\Delta_{i}^{\prime}\psi_{,ii}^{(1)\prime}.$ (41) Going to Fourier space and decomposing $\epsilon^{(1)}({\bf k})$ as $\epsilon^{(1)}({\bf k})=\sum_{i}\hat{k}_{i}^{2}\varepsilon_{i}^{(1)}({\bf k})$, this equation can be rewritten as $\displaystyle\varepsilon_{i}^{(1)\prime\prime}+\mathcal{H}\varepsilon_{i}^{(1)\prime}-\frac{3}{2}\Omega_{\rm m}\mathcal{H}^{2}\varepsilon_{i}^{(1)}=-2\Delta_{i}^{\prime}\psi^{(1)\prime}.$ (42) Notice that $\varepsilon_{i}^{(1)}$ is different from $\epsilon_{i}^{(1)}$. We can derive the matter-dominated solution for $\varepsilon_{i}^{(1)}$ for initial-condition generation. Since Green’s function for this equation is the same as Eq. (27) the solution is $\displaystyle\varepsilon_{i}^{(1)}(\eta)=$ $\displaystyle\int_{0}^{\eta}\differential\tilde{\eta}~{}\frac{1}{5}\left[\frac{\eta^{2}}{\tilde{\eta}}-\frac{\tilde{\eta}^{4}}{\eta^{3}}\right]\cdot\left(-2\Delta_{i}^{\prime}(\tilde{\eta})\psi^{(1)\prime}(\tilde{\eta})\right)$ $\displaystyle=$ $\displaystyle-\frac{2}{5}\frac{\psi^{(1)}(\eta)}{D(\eta)}\frac{\Delta_{i}(\eta)}{\beta(\eta)}\int_{0}^{\eta}\differential\tilde{\eta}~{}\left[\frac{\eta^{2}}{\tilde{\eta}}-\frac{\tilde{\eta}^{4}}{\eta^{3}}\right]\cdot D^{\prime}(\tilde{\eta})\beta^{\prime}(\tilde{\eta})$ $\displaystyle=$ $\displaystyle-\frac{4}{5}\psi^{(1)}(\eta)h_{i}^{\rm ini}\left[\beta(\eta)-\int_{0}^{\eta}\differential\tilde{\eta}~{}\left(\frac{\tilde{\eta}}{\eta}\right)^{5}\beta^{\prime}(\tilde{\eta})\right]$ $\displaystyle=$ $\displaystyle\psi^{(1)}(\eta)h_{i}^{\rm ini}\alpha(\eta),$ (43) where we have used $D(\eta)\propto\eta^{2}$ and $\Delta_{i}(\eta)=h_{i}^{\rm ini}\beta(\eta)$ and $\alpha(\eta)$ is introduced in Eq. (29). Thus $\displaystyle\epsilon^{(1)}(\eta)=\psi^{(1)}(\eta)\alpha(\eta)\sum_{i}h_{i}^{\rm ini}\hat{k}_{i}^{2}.$ (44) This implies that the linear growth function has a direction-dependent modulation: $\displaystyle D(\eta,{\bf k})=D(\eta)\left[1+\alpha(\eta;k_{L})\sum_{i}h_{i}^{\rm ini}\hat{k}_{i}^{2}\right],$ (45) which is consistent with Eq. (29). Using Eq. (44), the correction to the velocity can be computed as $\displaystyle\epsilon_{i}^{(1)\prime}=$ $\displaystyle\left(f_{\alpha}+f_{1}\right)\mathcal{H}\epsilon_{i}^{(1)},$ (46) with $f_{\alpha}\equiv\differential\ln\alpha/\differential\ln a$ and $f_{1}\equiv\differential\ln D/\differential\ln a$. We implement these modifications in 2LPTIC Crocce _et al._ (2006) and generate the initial conditions at $z_{i}=99$. ### III.3 Simulations Figure 2: Fractional anisotropic scale factor in the $x$-axis $\Delta_{x}(a;k_{L})=a_{x}(a;k_{L})/a-1$ for various wavenumbers of GWs $k_{L}$ as a function of the scale factor $a$ in the case of $h^{\rm ini}_{ij}={\rm diag}(0.1,-0.1,0)$. The vertical dashed line represents the starting redshift $z_{\rm ini}=99$. We perform $N$-body simulations in the tidal backgrounds with $1024^{3}$ particles in $500~{}\mathrm{Mpc}/h$ boxes. The details of modifications of the $N$-body code based on Gadget-2 Springel (2005) in the tidal background is described in Ref. Akitsu _et al._ (2021a). One important additional modification to the code was made in the drift operator, which is discussed in Appendix C. After rotating $\tau_{ij}$ to align its eigenvectors with the simulation axis, the remaining degrees of freedom can be completely characterized by two parameters, which we can parametrize as $\tau_{\rm e}\equiv-(\tau_{1}-\tau_{2})/2$ and $\tau_{\rm p}=-\tau_{3}+(\tau_{1}+\tau_{2})/2$. Taking into account the transverse condition of GWs, we cannot consider $\tau_{\rm p}$-type tides for the tidal separate universe with GWs unlike the case of the scalar tidal field in Ref. Akitsu _et al._ (2021a). Hence in this paper we only consider $\tau_{\rm e}$-type tides for the background anisotropy. In other words, we consider GWs propagating along $z$ direction with $+$ mode polarization: $\displaystyle h^{\rm ini}_{ij}=\begin{pmatrix}\pm\epsilon&0&0\\\ 0&\mp\epsilon&0\\\ 0&0&0\end{pmatrix},$ (47) where we choose $\epsilon=0.1$. Since we aim to measure the response of large- scale structure to the long-wavelength GWs, the results should not be dependent on the choice of the direction of the propagation and the polarization basis. In order to investigate effects of GWs over a wide range of wavenumbers, we run tidal separate universe simulations with various wavenumbers of GWs: $\displaystyle k_{L}=\left\\{0.0001,\,0.0002,\,0.0005,\,0.001,\,0.002,\,0.005,\,0.01,\,0.02,\,0.05,\,0.1,\,0.2\right\\}\,[h/{\rm Mpc}].$ (48) These different wavenumbers of GWs give rise to the different time evolution of the local anisotropic scale factors, through which long-wavelength GWs affect the simulated small-scale structure formation. Some examples of the time evolution of $\Delta_{x}$ are shown in Fig. 2. Given the box size of $500~{}{\rm Mpc}/h$, some wavenumbers are larger than the fundamental mode in the simulation: $k_{\rm F}=0.013~{}h/{\rm Mpc}$. For such larger wavemubers, we cannot neglect the curvature of the long-wavelength modes and treat GW as a unifiom tidal field over the whole simulation box and the approximation is violated. Still, these GWs are longer-mode than the halo formation scale so GW can be seen as a uniform tide in that local region. Therefore we can study the impact of GWs on halos, in particular, the response of the halo shape as discussed in Sec. VI, where we come back to this issue again. For references, we also run fiducial simulations with the isotropic background, and tidal separate universe simulations induced by the scalar $\tau_{\rm e}$-type tidal field. These reference simulations share the same parameters (including random seeds) as those used in the GW tidal separate universe simulations. For each type of simulations (fiducial and tidal separate universe with GWs and scalar tides), we run four realizations, amounting to 100 simulations in total. We use the AHF code Knollmann and Knebe (2009) to identify dark matter halos in the simulations by spherical overdensity (SO) regions 200 times as dense as the mean matter density. We need to identify SO halos in the _global_ coordinates on the _isotropic_ background, i.e., $a_{i}x_{i}/a$, while AHF by default uses the local simulation coordinates $x_{i}$ on the anisotropic background. To this end, we modify the AHF code to rescale the coordinate correspondingly in distance computations. ## IV Anisotropic power spectrum response: tensor fossils in nonlinear regime In this section, as a first example of imprints of GWs on large-scale structure, we present the anisotropic impact on the matter auto-, matter-halo cross-, and halo auto-power spectra by long-wavelength GWs measured from our $N$-body simulations. ### IV.1 Growth-dilation decomposition As we derived in Sec. II.2, long-wavelength GWs leave the anisotropic imprint in the matter power spectrum in a given realization of GWs through the nonlinear interaction of tidal fields. This tidal response consists of two different contributions as in Eq. (31); the term proportional to $\alpha$ that modulates the amplitude of the power spectrum and the term proportional to $\beta$ that modulates the scales. In terms of the tidal separate universe introduced in Sec. III.1, the former known as the growth effect describes the changes in the amplitude of short-mode fluctuations measured in the local anisotropic background, while the latter known as the dilation effect stems from the anisotropic expansion of the local background with respect to the global one. Specifically, the response of the power spectrum to long- wavelength GWs can be decomposed as $\displaystyle\left.\frac{\differential{\ln P_{\cal G}}}{\differential{h^{\rm ini}_{ij}}}\right|_{{\bf k}_{\cal G}}=\left.\frac{\differential{\ln P_{\cal L}}}{\differential{h^{\rm ini}_{ij}}}\right|_{{\bf k}_{\cal G}}=$ $\displaystyle\left.\frac{\partial\ln P_{\cal L}}{\partial h^{\rm ini}_{ij}}\right|_{{\bf k}_{\cal L}}+\left.\frac{\partial\ln P_{\cal L}}{\partial\ln k_{{\cal L},i^{\prime}}}\right|_{h_{ij}^{\rm ini}}\left.\frac{\differential{\ln k_{{\cal L},i^{\prime}}}}{\partial h^{\rm ini}_{ij}}\right|_{{\bf k}_{\cal G}}$ $\displaystyle\equiv$ $\displaystyle\,\hat{k}_{i}\hat{k}_{j}\left[R^{\rm GW}_{\rm growth}(k;k_{L})+R^{\rm GW}_{\rm dilation}(k;k_{L})\right],$ (49) where in the first line the power spectra defined with respect to the global isotropic background and the local anisotropic background are denoted by $P_{\cal G}$ and $P_{\cal L}$, respectively, and correspondingly the wavenumbers by ${\bf k}_{\cal G}$ and ${\bf k}_{\cal L}$. The physical scale should be unchanged in these two coordinates, $a_{i}x_{{\cal L},i}=ax_{{\cal G},i}$, which implies $k_{{\cal L},i}=k_{{\cal G},i}(1+\Delta_{i})$. In the first equality we have used that the variances must be conserved in the coordinate transformation: $P_{\cal G}\,\differential^{3}{\bf k}_{\cal G}=P_{\cal L}\,\differential^{3}{\bf k}_{\cal L}$, together with $|\differential^{3}{\bf k}_{\cal L}/\differential^{3}{\bf k}_{\cal G}|=\prod_{i=1}^{3}(1+\Delta_{i})=1$ at leading order. In the second line, We relabel both $k_{\cal G}$ and $k_{\cal L}$ as $k$, since the responses $R^{\rm GW}_{\rm growth}$ and $R^{\rm GW}_{\rm dilation}$ are already first order in $h^{\rm ini}_{ij}$ (or $\Delta_{i}$) and hence here we do not need to distinguish $k_{\cal G}$ and $k_{\cal L}$. Notice that we distinguish $k_{L}$ from $k_{\cal L}$; the former represents the wavenumber of GWs. Notice also that we have defined the response with respect to the $h_{\rm ini}$, not $h(z)={\cal T}(z)h_{\rm ini}$, which allows for simple computations of observables in terms of the primordial tensor mode amplitude. Namely the cosmology dependence other than the initial amplitude of GWs factorizes out in the response function. The power spectrum of biased tracers in the presence of long-wavelength GWs then has a generic form, $\displaystyle P_{XY}({\bf k}|h_{ij}(k_{L}))=P_{XY}(k)\left[1+\hat{k}^{i}\hat{k}^{j}h_{ij}^{\rm ini}(k_{L})\left(R^{\rm GW}_{{\rm growth};XY}(k;k_{L})+R^{\rm GW}_{{\rm dilation};XY}(k;k_{L})\right)\right],$ (50) where $X$ and $Y$ represent tracers being considered. For the matter auto- power spectrum ($X=Y={\rm m}$), this corresponds to the nonlinear extension of Eq. (31). We will also consider the matter-halo cross-power spectrum ($X={\rm m}$ and $Y={\rm h}$) and the halo auto-power spectrum ($X=Y={\rm h}$) below. Moreover, using these GW power spectrum responses, we can write down the $X$-$Y$-GWs (scalar-scalar-tensor) bispectrum in squeezed limit as $\displaystyle\lim_{k_{L}\to 0}B_{XYh_{(\lambda)}}(k,k^{\prime},k_{L})=\hat{k}^{i}\hat{k}^{j}e^{(\lambda)}_{ij}\left[R^{\rm GW}_{{\rm growth};XY}(k;k_{L})+R^{\rm GW}_{{\rm dilation};XY}(k;k_{L})\right]P_{XY}(k)P_{h_{(\lambda)}}(k_{L}),$ (51) where we have defined the $X$-$Y$-GW bispectrum via $\langle X({\bf k})Y({\bf k}^{\prime})h_{(\lambda)}({\bf k}_{L})\rangle=(2\pi)^{3}\delta^{(3)}_{\rm D}({\bf k}+{\bf k}^{\prime}+{\bf k}_{L})B_{XYh_{(\lambda)}}(k,k^{\prime},k_{L})$ and neglected the primordial contribution.333 Although here we have neglected the primordial scalar-scalar- tensor bispectrum, its contribution to the local observables appears only at the order of ${\mathcal{O}}(k_{L}/k)^{2}$ and thus is subdominant for the single-field inflation (see Ref. Pajer _et al._ (2013)). Making use of the growth-dilation decomposition, we can go a little further than the perturbative result. Because the dilation effect purely captures the coordinate transformation, using Eq. (37), we can obtain the non-perturbative result as $\displaystyle R^{\rm GW}_{{\rm dilation};XY}(k;k_{L})=\beta(k_{L})\frac{\partial\ln P_{XY}(k)}{\partial\ln k}.$ (52) Eqs. (31) and (52) are different because former perturbative result involves the linear power spectrum while the latter involves the nonlinear power spectrum and can be applied to a non-linear regime. This means that we can compute the dilation piece without running simulations even in the nonlinear regime given the slope of the nonlinear power spectrum. On the other hand, the growth piece comes from the dynamical effect where we cannot extend the perturbative result in the nonlinear regime. We thus need to rely on simulations in order to calibrate the growth response in the nonlinear regime. Hence, in this paper we focus on measuring the growth term from the tidal separate universe simulations with GWs. By comparing the measurement of the growth term of the matter auto-power spectrum in simulations with the perturbation theory prediction at quasi-nonlinear scales, we can validate our methodology using the relation: $\displaystyle\lim_{k\ll k_{\rm NL}}R^{\rm GW}_{\rm growth;mm}(k;k_{L})=2\alpha(k_{L}).$ (53) Note that since we assume GWs are long-wavelength modes compared with scalar perturbations the perturbative result is valid only for the range of $k_{L}\ll k\ll k_{\rm NL}$. ### IV.2 Growth response of the matter auto-power spectrum from simulations In the presence of GWs of $\tau_{\rm e}$-type configuration (Eq. (47)), the matter auto-power spectrum takes a form of $\displaystyle P_{\rm mm}({\bf k}|h_{\rm e}(k_{L}))=P_{\rm mm}(k)\left[1+\frac{2}{3}R^{\rm GW}_{\rm growth;mm}(k;k_{L})\left({\cal L}_{2}(\hat{k}_{1})-{\cal L}_{2}(\hat{k}_{2})\right)h^{\rm ini}_{\rm e}\right],$ (54) where $P_{\rm mm}(k)$ is the nonlinear matter power spectrum in the isotropic background, ${\cal L}_{2}(x)$ is Legendre polynomial of order two, $\hat{k}_{1}$ and $\hat{k}_{2}$ represent the $x$ and $y$ components of $\hat{k}$ respectively, and $h_{\rm e}^{\rm ini}\equiv-\left(h_{11}^{\rm ini}-h_{22}^{\rm ini}\right)/2=\pm\epsilon$. We can estimate the growth response by taking the quadrupoles of the power spectrum along both $x$ and $y$ axes: $\displaystyle P^{\ell_{\rm e}=2}_{\rm mm}(k|h_{\rm e}(k_{L}))$ $\displaystyle\equiv P^{\ell_{x}=2}_{\rm mm}(k)-P^{\ell_{y}=2}_{\rm mm}(k)$ (55) $\displaystyle=2P_{\rm mm}(k)R^{\rm GW}_{\rm growth;mm}(k;k_{L})h^{\rm ini}_{\rm e},$ (56) where $P^{\ell_{i}=2}_{\rm mm}(k)$ ($i=\\{x,y\\}$) are defined as $\displaystyle P^{\ell_{x}=2}_{XY}(k)$ $\displaystyle\equiv 5\int\frac{\differential^{2}\hat{{\bf k}}}{4\pi}P_{XY}({\bf k}){\cal L}_{2}(\hat{k}_{1}),$ (57) $\displaystyle P^{\ell_{y}=2}_{XY}(k)$ $\displaystyle\equiv 5\int\frac{\differential^{2}\hat{{\bf k}}}{4\pi}P_{XY}({\bf k}){\cal L}_{2}(\hat{k}_{2}).$ (58) This leads to the estimator for $R_{\rm growth;mm}(k;k_{L})$ as $\displaystyle R^{\rm GW}_{\rm growth;mm}(k;k_{L})=\frac{P^{\ell_{\rm e}=2}_{\rm mm}(k|h^{\rm ini}_{\rm e}=+\epsilon)-P^{\ell_{\rm e}=2}_{\rm mm}(k|h^{\rm ini}_{\rm e}=-\epsilon)}{4\epsilon P_{\rm mm}(k)}.$ (59) Figure 3: Growth response of matter auto-power spectrum to GWs, $R^{\rm GW}_{\rm growth}(k_{L};k)$, as a function of the short-wavenmer $k$ at various wavenumvers of GWs (left) and various redshifts (right), measured from the simulations. The top-left and bottom-left panels show the responses at $z=0.5$ $z=2$, respectively. The top-right and bottom-right panels show the responses at $k_{L}=0.0002~{}h/{\rm Mpc}$ and $k_{L}=0.001~{}h/{\rm Mpc}$, respectively. The dashed lines with each different color represent the predictions from the perturbation theory at each wavenumber and redshift. Figure 4: Growth response of matter auto-power spectrum to GWs, $R^{\rm GW}_{\rm growth}(k_{L};k)$, as a function of the wavenumber of GWs $k_{L}$ at $z=2,1,0.5,$ and $0$ (from the left to the right), measured from the simulations. The upper and lower rows show the results for different $k$, $k=0.376~{}h/{\rm Mpc}$ and $k=0.881~{}h/{\rm Mpc}$, respectively. The orange line depicts the “rescaled growth response” to the scalar tides, $\frac{7}{4}\alpha(k_{L})R_{\rm growth}^{\rm scalar}(k)$ with $R_{\rm growth}^{\rm scalar}(k)$ measured from the tidal separate universe simulations with the scalar tides. We measure the growth response of the matter auto-power spectrum to GWs, $R^{\rm GW}_{\rm growth;mm}(k;k_{L})$, from the simulations which share the same initial random phase to reduce the sample variance. Since $R^{\rm GW}_{\rm growth;mm}$ depends on the wavenumbers of the shote-mode, $k$, and of GWs, $k_{L}$, as well as redshift $z$, it is hard to show all the dependences in one figure. Thus, let us start by showing the measured growth response as a function of $k$ for various $k_{L}$ at $z=0.5$ and $z=2$ in the upper- and lower-left panels of Fig. 3, respectively. We then show it for $k_{L}=0.0002h/{\rm Mpc}$ and $k_{L}=0.001h/{\rm Mpc}$ at various redshifts in the upper- and lower-right panels, respectively. In each panel, the corresponding perturbation theory predictions, Eq. (53), are shown in the dashed lines. First, onc can see the excellent agreement between the measured and predicted growth responses on large scales for all $k_{L}$, which verifies that our methodology to incorporate GWs in simulations works correctly. Second, the measured response deviates from the perturbation prediction on nonlinear scales, in particular for larger $k_{L}$. Although the way it deviates depends on $k_{L}$, the overall trend is similar among different $k_{L}$ values; (i) at the smallest scales ($k\gtrsim 2\ h/{\rm Mpc}$) the growth response decreases for all $k_{L}$ and redshifts. (ii) At earlier redshift, the growth response is slightly enhanced compared to the perturbation theory while at lower redshift it is largely suppressed. Third, the redshift dependence of the growth response is clearer as shown in the right panels. The agreement between the perturbation theory and the simulation becomes worse at lower redshifts. At $z=2$ these two are in good agreement up to $k\sim 0.2\ h/{\rm Mpc}$, while at $z=0$ the simulation results start to differ from the perturbation theory around $k\sim 0.04\ h/{\rm Mpc}$. These tendencies we found for the tensor mode above are actually very similar to those for the case of the scalar tidal field which had been extensively studied (see App.B and Refs. Akitsu _et al._ (2021a); Stücker _et al._ (2021); Masaki _et al._ (2020) for details). Given the similarity of the behavior of the growth response on nonlinear scales between the scalar and tensor cases, it is natural to ask how similar are these two quantitatively. Let us finish this subsection by answering this question. For this purpose, we consider a “rescaled growth response” to the scalar tides, $\frac{7}{4}\alpha(\eta;k_{L})R^{\rm scalar}_{\rm growth;mm}(k)$, where the scalar tidal response $R^{\rm scalar}_{\rm growth;mm}(k)$ is introduced in Eq. (96) in App. B. Since $R^{\rm scalar}_{\rm growth;mm}(k)$ approaches $8/7$ at the large-scale limit (see Eq. (95)), this scalar tidal response matches the tensor tidal response in the large scale limit (Eq. 53), $\displaystyle\lim_{k\to 0}\frac{7}{4}\alpha(\eta;k_{L})R^{\rm scalar}_{\rm growth;mm}(k)=2\alpha(\eta;k_{L}).$ (60) In Fig.4 we compare this “rescaled growth response” to the scalar tides with the measured tensor tidal response on nonlinear scales as a function of $k_{L}$ for various redshifts. The upper and lower panels show the results for $k=0.376\ h/{\rm Mpc}$ and $k=0.881\ h/{\rm Mpc}$, respectively. Overall, the rescaled response to the scalar tides captures the general feature of the growth response to GWs on nonlinear scales. Such agreement can be seen for both the two wavenumbers at all the redshifts when $k_{L}\lesssim 10^{-3}~{}h/{\rm Mpc}$. On the other hand, for $k_{L}\gtrsim 10^{-3}~{}h/{\rm Mpc}$ the response to GWs has greater values than the rescaled response, except for some $k_{L}$ at $z=0$. This difference gets larger at higher redshifts and larger $k$. These can be attributed to the different time evolution of the anisotropic scale factors in the scalar tide and GWs cases. The anisotropic scalar factor induced by the scalar tidal field grows monotonically in time, following the linear growth rate $D(\eta)$ at leading order, $\Delta^{\rm scalar}_{i}(\eta)\propto D(\eta)$ (Eq. (94)), whereas that induced by GWs has the time dependence described by $\beta(\eta;k_{L})$, which is generally monotonous in time for $k_{L}\lesssim 10^{-3}~{}h/{\rm Mpc}$ while not so for $k_{L}\gtrsim 10^{-3}~{}h/{\rm Mpc}$ (see Fig. 2). Specifically, for $k_{L}\gtrsim 10^{-3}~{}h/{\rm Mpc}$ the anisotropic scale factors from GWs reach their asymptotic value at early redshift, when the anisotropic scale factors from scalar tides are still tiny, which causes the tidal response to be stronger in these $k_{L}$. Figure 5: Growth response of the halo-matter cross-power spectrum to GWs, $R^{\rm GW;hm}_{\rm growth}(k_{L};k)$, as a function of $k$ for various $k_{L}$ (left), various redshifts (center), and various halo masses (right), measured from the simulations. The blue dotted line in the right panel shows the growth response of the matter auto-power spectrum. Figure 6: Similar to Fig. 5 but the growth response of the halo auto-power spectrum to GWs, $R^{\rm GW;hh}_{\rm growth}(k_{L};k)$ as a function of $k$ for various $k_{L}$ (left), various redshifts (center), and various halo masses (right), measured from the simulations. The upper and lower sets show the results normalized by the halo auto-power spectrum with and without a shot noise, respectively. For the latter, the shot noise contribution is subtracted assuming the Poisson distribution. In the lower-right panel, we do not show the result for the mass bin, $10^{14}M_{\odot}/h<M_{\rm vir}<10^{14.5}M_{\odot}/h$ because it is too noisy. ### IV.3 Growth responses of the halo-matter cross- and halo auto-power spectra from simulations The halo-matter cross- and halo auto-power spectra in a local region are also affected by GWs. The growth responses of these power spectra to GWs can be estimated from the simulations in the same way as the matter auto-power spectrum, $\displaystyle R^{\rm GW}_{\rm growth;XY}(k;k_{L})=\frac{P^{\ell_{\rm e}=2}_{\rm XY}(k|h^{\rm ini}_{\rm e}=+\epsilon)-P^{\ell_{\rm e}=2}_{\rm XY}(k|h^{\rm ini}_{\rm e}=-\epsilon)}{4\epsilon P_{\rm XY}(k)},$ (61) with $XY\in\\{{\rm hm},{\rm hh}\\}$. In Fig. 5 we show the growth response of the halo-matter cross-power spectrum to GWs as a function of the wavenumber of the short modes. The left and middle panels focus on the fixed halo mass ($10^{12}M_{\odot}/h<M_{\rm vir}<10^{12.5}M_{\odot}/h$), and compares the growth responses for various wavenumbers of GWs at $z=0.5$ in the left panel and various redshfits at $k_{L}=0.001~{}h/{\rm Mpc}$ in the middle panel. As for the $k_{L}$-dependence, the response tends to be greater for larger $k_{L}$, similar to the case of the matter auto response. On the other hand, the redshift dependence is slightly different from the matter auto case. For the halo-matter cross-power spectrum, the growth response at late redshifts ($z=0$ and $z=0.5$) gets slightly enhanced compared to the linear regime and persists up to $k\lesssim 1~{}h/{\rm Mpc}$, whereas the matter auto response starts to decrease around $k\simeq 0.1~{}h/{\rm Mpc}$. This can be clearly seen in the right panel of Fig. 5, which directly compares the growth response of the matter auto power-spectrum to that of the halo-matter cross-power spectrum with various halo masses at $z=0.5$ and $k_{L}=0.001~{}h/{\rm Mpc}$. First, it turns out that the scale-dependence ($k$-dependence) of the growth response varies with the halo mass. The enhancement of the growth response around $k\simeq 1~{}h/{\rm Mpc}$ is greater for less massive halos. This is mainly due to the normalization. Here we define the response with respect to the halo-matter cross-power spectrum in the fiducial simulation. However if we define the response with respect to the matter auto-power spectrum, the response of more massive halos gets more amplified in the nonlinear regime. Second, the difference of the responses between the matter auto and halo- matter cross cases stems from the halo biases. In particular at largest scales it should be explained by the halo tidal bias induced by GWs, which we investigate in the next section. Next, let us focus on the growth response of the halo auto-power spectrum. To obtain it, for the denominator of Eq. (61) we use measurements of the halo auto-power spectrum with and without the shot noise contribution. For the latter, the shot noise contribution is subtracted assuming the Poisson distribution, $1/\bar{n}_{\rm h}$. The upper and lower rows of Fig. 6 show the resulting response of the halo auto-power spectrum as a function of $k$ with and without the shot noise, respectively. As with Fig. 5, the left and middle columns show the responses with the fixed mass range $10^{12}M_{\odot}/h<M_{\rm vir}<10^{12.5}M_{\odot}/h$ for various wavenumber of GWs at $z=0.5$ and for $k_{L}=0.001~{}h/{\rm Mpc}$ at various redshfits, respectively, while the right columns show the result with various halo masses for $k_{L}=0.001~{}h/{\rm Mpc}$ at $z=0.5$. Overall, the $k_{L}$, redshift, and halo mass dependencies of the responses for the halo auto spectra both with and without the shot noise are similar to that for the halo-matter cross spectrum. In the upper set of Fig. 6, the responses approach zero on small scales where the halo auto-power spectra are dominated by shot noises. On the contrary, the peaky features around $k\sim 2~{}h/{\rm Mpc}$ in the lower set reflects the non-Poissonian behavior of the halo shot noise due to, e.g., the exclusion effect Smith _et al._ (2007); Hamaus _et al._ (2010); Baldauf _et al._ (2013). Therefore, the simple Poissonian shot-noise removal leads to zero-crossing or negative, unphysical halo auto-power spectrum on small scales, making the responses peaky. Nonetheless, up to $k\sim 0.5~{}h/{\rm Mpc}$, where the shot noise contribution is still small, the responses are not suppressed unlike the matter auto case. These trends seen in the response of the halo-matter cross- and halo auto-power spectra are the same as those in the scalar tidal field case (see App. B). ## V Halo tidal bias induced by GWs Figure 7: Halo tidal bias induced by GWs as a function of halo mass for various wavenumber of GWs at $z=0.5$ (the left panel) and at various redsfhits for $k_{L}=0.001~{}h/{\rm Mpc}$ (the right panel). As discussed in the previous section, the long-wavelength GWs can affect the halo density field. In the perturbative regime, this effect should be characterized in terms of the halo bias. When only scalar perturbations are considered, which is the standard setup, the halo density field can be expanded up to the second order as $\displaystyle\delta_{\rm h}=b_{1}\delta+\frac{1}{2}b_{2}\delta^{2}+b_{s^{2}}s^{2},$ (62) where $s^{2}=s_{ij}s^{ij}$ with $s_{ij}\equiv(\partial_{i}\partial_{j}/\partial^{2}-\delta^{\textsc{k}}_{ij}/3)\delta$. Here $b_{s^{2}}$ is called the tidal bias that captures the effect of the tidal fields on the halo density field. Since long-wavelength GWs are locally indistinguishable from tidal fields induced by the scalar perturbations, it is natural to expect that there is a tidal bias induced by GWs as well. Then, at the linear order of GWs, the halo density field should acquire the following contribution from GWs, $\displaystyle\delta_{\rm h}^{(2)}(\eta)\supset b^{\rm GW}_{s^{2}}(\eta;k_{L})s^{ij}(\eta)h_{ij}^{\rm ini}(k_{L}),$ (63) where $b_{s^{2}}^{\rm GW}$ is the tidal bias coefficient induced by the coupling between tidal fields induced by GWs and scalar density perturbations. Note that we define $b_{s^{2}}^{\rm GW}$ with respect to the GWs at the _initial_ epoch instead of the _same_ time, unlike the biases defined with respect to the scalar perturbations. Because the GWs are additional degrees of freedom to the scalar adiabatic perturbations, we expect the tidal bias to depend on the wavenumber of GWs. Collecting the second order pieces in the matter density field and halo density field, we find the tree-level halo- matter-GWs bispectrum in the squeezed limit as $\displaystyle\lim_{k_{L}\to 0}B_{{\rm hm}h_{(\lambda)}}(k,k^{\prime},k_{L})=\hat{k}^{i}\hat{k}^{j}e^{(\lambda)}_{ij}\left[2b_{1}\alpha(\eta;k_{L})+2b_{s^{2}}^{\rm GW}(\eta;k_{L})\right]P_{\rm lin}(k)P_{h_{(\lambda)}}(k_{L}),$ (64) where we have omitted the dilation piece. This implies that the local halo- matter power spectrum in the presence of long-wavelength GWs is $\displaystyle P_{\rm hm}({\bf k}|h_{ij}(k_{L}))=\left[b_{1}+\left[2b_{1}\alpha(\eta;k_{L})+2b^{\rm GW}_{s^{2}}(\eta;k_{L})\right]k^{i}k^{j}h_{ij}^{\rm ini}\right]P_{\rm lin}(k).$ (65) Thus, we can estimate $b_{s^{2}}^{\rm GW}$ from the growth response of the halo-matter power spectrum to GWs, which is presented in the previous section. Although there are several ways to estimate $b_{s^{2}}^{\rm GW}$ from the growth response of the halo-matter power spectrum, we use the following way to reduce the uncertainty of $b_{1}$ and the sample variance. We first define the local linear bias estimator as a ratio of the halo-matter and matter auto- power spectra including the quadrupoles measured in the simulations with GWs, $\displaystyle\hat{b}^{\ell_{x,y}}_{1}(k;\pm\epsilon)\equiv\frac{P^{\ell=0}_{\rm hm}(k;\pm\epsilon)+P^{\ell_{x,y}=2}_{\rm hm}(k;\pm\epsilon)}{P^{\ell=0}_{\rm mm}(k;\pm\epsilon)+P^{\ell_{x,y}=2}_{\rm mm}(k;\pm\epsilon)}=\frac{b_{1}\pm(2b_{1}\alpha(\eta;k_{L})+2b_{s^{2}}^{\rm GW})\epsilon}{1\pm 2\alpha(\eta;k_{L})\epsilon}\simeq\tilde{b}_{1}(k)\pm 2\tilde{b}_{s^{2}}^{\rm GW}(k)\epsilon,$ (66) Therefore using this local linear bias, the estimator for $b_{s^{2}}^{\rm GW}$ is now $\displaystyle\tilde{b}^{\rm GW}_{s^{2}}(k)=\frac{1}{2}\left[\frac{\hat{b}_{1}^{\ell_{x}}(k;+\epsilon)-\hat{b}_{1}^{\ell_{x}}(k;-\epsilon)}{4\epsilon}+\frac{\hat{b}_{1}^{\ell_{y}}(k;+\epsilon)-\hat{b}_{1}^{\ell_{y}}(k;-\epsilon)}{4\epsilon}\right].$ (67) We estimate $b_{s^{2}}^{\rm GW}$ utilizing the $\chi^{2}$ statistic, defined as $\chi^{2}=\sum_{k=k_{\rm min}}^{k_{\rm max}}[b_{s^{2}}^{\rm GW}-\tilde{b}_{s^{2}}^{\rm GW}(k)]^{2}/\sigma^{2}_{b_{s^{2}}^{\rm GW}}(k)$, where $\sigma^{2}_{b_{s^{2}}^{\rm GW}}(k)$ is the variance of $\tilde{b}_{s^{2}}^{\rm GW}(k)$ at each $k$-bin measured from simulations. Adopting $k_{\rm max}=0.08~{}h/{\rm Mpc}$, we obtain the best-fitting value of $b_{s^{2}}^{\rm GW}$ and its uncertainty by minimizing $\chi^{2}$. We restrict the measurement of $b_{s^{2}}^{\rm GW}$ to a range of $k_{L}\leq 0.002~{}h/{\rm Mpc}$ because we rely on the perturbative results, which are valid only when $k_{L}\ll k$. Figure 8: Upper panels: Halo tidal bias induced by GWs as a function of wavenumber of GWs with halo mass of $10^{13}~{}M_{\odot}/h<M_{\rm vir}<10^{13.5}~{}M_{\odot}/h$. Lower panels: Same as upper panels but with halo mass of $10^{14}~{}M_{\odot}/h<M_{\rm vir}<10^{14.5}~{}M_{\odot}/h$. From the left to right , we show the results at $z=1$, $0.5$ and $0$. The blue points with error bars are the measurement from the simulations. The orange line represents the ansatz $b_{s^{2}}^{\rm GW}(k_{L};M)=\frac{7}{4}b_{s^{2}}^{\rm scalar}(M)\alpha(k_{L})$ with the shaded region being the error coming from $b_{s^{2}}^{\rm scalar}(M)$. Fig. 7 shows the tidal bias from GWs determined in this way as a function of halo mass. In the left panel we plot the result for various wavenumbers of GWs at $z=0.5$, and in the right panel we plot the result for various redshifts at $k_{L}=0.001~{}h/{\rm Mpc}$. Regardless of the wavenumber of GWs and redshift, more massive halos have greater absolute values of $b_{s^{2}}^{\rm GW}$, which has the same trend as the usual tidal bias induced by the scalar tidal fields. In addition to the mass dependence, the redshift dependence is also similar to the scalar tidal bias case; the absolute value of the tidal bias is larger at higher redshift. On the other hand, in contrast to the case of the scalar tidal bias, the tidal bias from GWs has a particular wavenumber dependence as shown in the left panel of Fig. 7. This wavenumber-dependence (or scale- dependence) inherits from the wavenumber-dependent transfer function of GWs while the growth function for the scalar density fluctuations is independent of the wavenumber (Eq. (21)). To look at this scale dependence in more detail, Fig. 8 displays $b_{s^{2}}^{\rm GW}$ as a function of $k_{L}$ for several redshifts and halo masses. We compare the measurements with the following ansatz: $\displaystyle b_{s^{2}}^{\rm GW}(\eta;k_{L})=\frac{7}{4}\alpha(\eta;k_{L})b^{\rm scalar}_{s^{2}}(\eta),$ (68) where $b_{s^{2}}^{\rm scalar}$ is the tidal bias induced by the scalar tides, introduced in Eq. (62) (see also App. B). This ansatz is motivated by the fact that (i) the tidal bias term stems from the coupling of tidal fields, which is given by $s^{ij}(z)s_{ij}(z)$ for the scalar tide case ($b_{s^{2}}^{\rm scalar}$) while $s^{ij}(z)h^{\rm ini}_{ij}$ for the tensor tide case ($b_{s^{2}}^{\rm GW}$), and (ii) the second-order matter density induced by tidal fields is $\frac{4}{7}s^{ij}(z)s_{ij}(z)$ for the scalar tides case while $\alpha(z;k_{L})s^{ij}(z)h^{\rm ini}_{ij}(k_{L})$ for the tensor tides case. Despite the large error bars, overall the measurements are well explained by this ansatz. This result suggests that we do not have to introduce a new free bias parameter for GWs at leading order once $b_{s^{2}}^{\rm scalar}$ is known. ## VI Intrinsic alignments induced by GWs Figure 9: Linear shape bias induced by GWs as a function of halo mass. The left and right panels show $b_{K}^{\rm GW}$ for various wavenumber of GWs at $z=0.5$ and for various redshifts at $k_{L}=0.001~{}h/{\rm Mpc}$, respectively. As we discussed so far, the tidal fields contribute to the density field only at the second order because tidal fields are a tensor while density fields are a scalar. Conversely, the tidal fields should contribute to tensor quantities at linear order. One well-known observable of tensor quantities in large-scale structure is the intrinsic alignments of halo (or galaxy) shapes Catelan _et al._ (2001); Hirata and Seljak (2004). The deviation of the halo intrinsic shape from sphere is characterized by ellipticity at the lowest order. In other words, the halo shape can be described by the trace-free rank-2 tensor in three-dimensional space, $\gamma_{ij}$. The linear alignment model of the intrinsic alignment relates the halo shape with the tidal field as follows: $\displaystyle\gamma_{ij}(\eta)$ $\displaystyle=b^{\rm scalar}_{K}(\eta)s_{ij}(\eta),$ (69) where $b^{\rm scalar}_{K}$ is the linear alignment coefficient or the linear shape bias, which captures the sensitivity or the response of halo or galaxy shape to the tidal field. Because long-wavelength GWs are locally equivalent to the tidal field, we naturally expect that the halo shapes are also aligned by GWs, namely $\displaystyle\gamma_{ij}(\eta;k_{L})=b_{K}^{\rm GW}(\eta;k_{L})h^{\rm ini}_{ij}(k_{L}),$ (70) where we have introduced the linear shape bias $b_{K}^{\rm GW}$ for GWs, in analogy with the tidal bias, $b^{\rm scalar}_{K}$. While $b^{\rm scalar}_{K}$ is defined as a response of the halo shape with respect to the scalar tidal field at the same time, $b^{\rm scalar}_{K}$ is defined as that with respect to GWs at the _initial_ time when the GWs are frozen, as in the tidal bias case. Figure 10: Linear shape bias induced by GWs as a function of the wavenumber of GWs for the halos with $10^{13.5}~{}M_{\odot}/h<M_{\rm vir}<10^{14}~{}M_{\odot}/h$. From the left to the right panels we show the results at $z=2,1,0.5,$ and $0$, respectively. The orange line depicts the ansatz introduced in Eq. (75), $b_{K}^{\rm GW}(k_{L};M,z)=\frac{7}{4}\alpha(k_{L},z)b_{K}^{\rm scalar}(M,z)$, and is not a fitting. The orange dashed region corresponds to the 1-$\sigma$ error of $b_{K}^{\rm scalar}$. The linear shape bias introduced above can be efficiently measured in the tidal separate universe simulation Akitsu _et al._ (2021a); Stücker _et al._ (2021), as the linear halo bias can be obtained precisely in the isotropic separate universe simulation Li _et al._ (2016); Baldauf _et al._ (2016); Lazeyras _et al._ (2016). In this paper, we define the halo shape by its reduced inertia tensor, $\displaystyle J_{ij}=\sum_{n=1}^{N_{\rm p}}\frac{x_{n,i}x_{n,j}}{x_{n}^{2}},$ (71) where $N_{\rm p}$ is the number of particles in the halo and $x_{n,i}$ is the $i$-th component of the particle location with respect to the halo center. Note that here $x$ represents the distance measured in the isotropic background, i.e., the physical coordinates. As $\gamma_{ij}$ is regarded as the trace-free part of $J_{ij}$, the linear alignment model states that this halo shape can be written as $\displaystyle J_{ij}=J_{0}\left[\frac{1}{3}\delta^{\textsc{k}}_{ij}+\gamma_{ij}\right]=J_{0}\left[\frac{1}{3}\delta^{\textsc{k}}_{ij}+b_{K}K_{ij}\right],$ (72) where $J_{0}$ is the normalization, for which we use the trace part of $J_{ij}$: $J_{0}={\rm Tr}[J_{ij}]$, and $K_{ij}$ is the tidal field of either $s_{ij}$ or $h_{ij}^{\rm ini}$. Therefore, we can measure $b_{K}$ as a response of the one-point function of the ellipticity to the tidal field. Specifically, introducing the following quantity $\displaystyle J_{\rm e}\equiv\frac{J_{11}-J_{22}}{2},$ (73) $b_{K}$ can be obtained as $\displaystyle b_{K}(M,z)=\frac{J_{\rm e}(M,z|K_{\rm e}^{\rm ini}=+\epsilon)-J_{\rm e}(M,z|K_{\rm e}^{\rm ini}=-\epsilon)}{2\epsilon J_{0}(M,z)},$ (74) where $K_{\rm e}^{\rm ini}=h_{\rm e}^{\rm ini}=\left(h_{11}^{\rm ini}-h_{22}^{\rm ini}\right)/2$ for the tensor tides and $K_{\rm e}^{\rm ini}=\left(s_{11}(z)-s_{22}(z)\right)/2$ for the scalar tides. Fig. 9 shows $b_{K}^{\rm GW}$ as a function of halo mass for various wavenumbers of GWs at $z=0.5$ (the left panel) and for $k_{L}=0.001~{}h/{\rm Mpc}$ at various redshifts (the right panel). First, we find that GWs influence the halo shape, implying that indeed intrinsic alignments can be induced by GWs. Compared with the measurement of $b_{s^{2}}^{\rm GW}$, the measurement of $b_{K}^{\rm GW}$ has much greater signal-to-noise ratio as we use the one-point function to measure $b_{K}^{\rm GW}$ while $b_{s^{2}}^{\rm GW}$ is measured from the large-scale limit of the power spectrum responses. Second, the halo-mass and redshift dependences of $b_{K}^{\rm GW}$ are similar to those of $b_{K}^{\rm scalar}$; that is, massive halos tend to more strongly align with GWs and the strength of the alignment at fixed halo mass decreases as the redshift gets smaller (see Fig. 16 and Ref. Akitsu _et al._ (2021a) for details).444 Note, however, that for the tensor case $b_{K}^{\rm GW}(z)$ is defined with respect to the _initial_ amplitude of GWs and it thus represents the strength of the alignment with respect to $h_{ij}^{\rm ini}$, while for the scalar tides case, $b_{K}^{\rm scalar}(z)$ represents the strength of the alignment with respect to $s_{ij}(z)$. Taking this difference into account, however, does not change the trend of the redshift dependence though, since $s_{ij}\propto D$ that increases monotonically with time. Third, the unique feature in the case of GWs is that $b_{K}^{\rm GW}$ is wavenumber-dependent (or scale-dependent), as in the cases with the power spectrum response and the halo tidal bias. Fig. 10 plots $b_{K}^{\rm GW}$ as a function of the wavenumber of GWs for the halos with $10^{13.5}~{}M_{\odot}/h<M_{\rm vir}<10^{14}~{}M_{\odot}/h$ at $z=2,1,0.5,0$ to directly demonstrate this scale dependence. The scale dependence is clearly seen because of the high $S/N$. Figure 11: Ratio of the linear shape bias induced by GWs to the one induced by scalar tidal fields as a function of the wavenumber of GWs for various halo masses. From the left to the right, we show the results at $z=2,1,0.5,0$, respectively. The orange line depicts the ansatz introduced in Eq. (75), $b_{K}^{\rm GW}(k_{L};M,z)/b_{K}^{\rm scalar}(M,z)=\frac{7}{4}\alpha(k_{L},z)$, and is not a fitting. In Fig. 10 we compare the measurements with the following ansatz for the intrinsic alignment from GWs Schmidt _et al._ (2014),555 The original ansatz considered in Ref. Schmidt _et al._ (2014) is $b_{K}^{\rm GW}(\eta;k_{L})=\frac{7}{2}\alpha(\eta;k_{L})b_{K}^{\rm scalar}(\eta)$ (see their Eq. (74)). However, the prefactor should be $7/4$ given the factor of two that comes from exchanging long-mode and short-mode in the second-order coupling ($\delta^{(2)}=\frac{2}{7}s_{ij}s_{ij}\to\delta^{(2)}=\frac{4}{7}s^{\rm long}_{ij}s^{\rm short}_{ij}$). $\displaystyle b_{K}^{\rm GW}(\eta;k_{L})=\frac{7}{4}\alpha(\eta;k_{L})b_{K}^{\rm scalar}(\eta).$ (75) It is natural to assume that the halo shape is determined by the local tidal environment around the halo, which is affected by long-wavelength tidal field via nonlinear mode-coupling. The influence of the long-wavelength tides on the small-scalar tides is captured by the response function that is discussed in Sec. IV. In particular, in Fig. 4 we examine the relation between the response function to the scalar tides and GWs and find that $R_{\rm growth;mm}^{\rm GW}(k;k_{L})=\frac{7}{4}\alpha(\eta;k_{L})R_{\rm growth;mm}^{\rm scalar}(k)$ is a good approximation even in the nonlinear regime. As these responses serve as the amplitude of the small-scale tides induced by the large-scale tides, we expect that $\gamma^{\rm scalar}_{ij}\propto R_{\rm growth;mm}^{\rm scalar}s_{ij}$ and $\gamma^{\rm GW}_{ij}\propto R_{\rm growth;mm}^{\rm GW}h^{\rm ini}_{ij}$, leading to the above ansatz. The ansatz of Eq. (75) is shown by the orange lines in Fig. 10. Remarkably, the measurements are well explained by this prediction for all the redshifts. Fig. 11 investigates if this agreement holds for all halo mass by displaying the ratio of the linear shape biases, $b_{K}^{\rm GW}/b_{K}^{\rm scalar}$, which is equal to $\frac{7}{4}\alpha$ in the ansatz regardless of halo mass. It turns out that the trend seen in Fig. 10 remains the same for all halo masses. One important point that follows from this agreement is that the process to determine the halo shape is not local in time. For, if the halo shape responds to the tidal field locally in time, meaning that the halo shape is related to the instantaneous tides: $\gamma^{\rm scalar}_{ij}(\eta)\propto s_{ij}(\eta)$ and $\gamma^{\rm GW}_{ij}(\eta)\propto\tau_{ij}(\eta)$, we expect $b_{K}^{\rm GW}(\eta)=b_{K}^{\rm scalar}(\eta)T(\eta;k_{L})/a(\eta)$ with $T$ defined in Eq. (10); however this is not the case. Let us finish this section by considering possible reasons that cause the difference between the measurements and the ansatz. First, the deviation is also observed around $k_{L}\sim 10^{-3}~{}h/{\rm Mpc}$ at $z=0$. This might be inherited from the large difference in the response functions to the scalar and tensor tides shown in the upper row of Fig. 4, which implies that relatively large-scale tides are more responsible for halo shapes than halo- scale tides. Second, the deviation of the measurements from the ansatz gets larger as $k_{L}$ increases. This trend is also consistent with the responses of the matter power spectrum in Fig. 4. At the same time, this could be partly because the approximation we employ in this paper is no longer valid for these large $k_{L}$. In other words, GWs with larger-$k_{L}$ cannot be seen as a uniform tidal field even in the halo formation region, given that the Lagrangian halo radius $R_{M}=\left(4\pi\bar{\rho}_{\rm m}/3M_{\rm vir}\right)^{1/3}$ is close to the wavelength of GWs. Specifically, ignoring the curvature of GWs can leads the corrections to $b_{K}^{\rm GW}$, which scales as ${\cal O}(k_{L}^{2}R_{M}^{2})$. In the case of $M_{\rm vir}=1.0\times 10^{13}M_{\odot}/h$ and $k_{L}=0.1~{}h/{\rm Mpc}$, for example, the correction is about $(0.1\cdot 3)^{2}\sim 0.1$, which cannot be negligible. In order to accurately study the impact of such GWs, a different technique is necessary. We also note that the results for high-$k_{L}$ GWs can be verified by analyzing the standard $N$-body simulation with staring an anisotropic initial power spectrum because the effect of high-$k_{L}$ GWs is almost encoded in the initial conditions. As shown in Fig. 2, the anisotropic scale factors induced by high-$k_{L}$ GWs ($k_{L}\gtrsim 10^{-2}~{}h/{\rm Mpc}$) reach their asymptote already at $z_{\rm ini}$ and thus their effect in the late time can be absorbed into the overall time-independent rescaling of scale factors, implying that it does not have a physical effect on structure formation. In other words, for high-$k_{L}$ GWs halos are formed in absence of a large-scale tidal field but with anisotropic small-scale modes, while for low-$k_{L}$ GWs halos are formed in an evolving tidal field, in analogy to the difference in $b_{1}$ (the response to the large-scale density field) and $b_{\phi}$ (the response to the change of $\sigma_{8}$) in the density case Baldauf _et al._ (2011); Desjacques _et al._ (2018). This study is beyond our paper and we leave it for future work. ## VII Discussion In this paper, we have quantified the impact of GWs on large-scale structure by means of the tidal separate universe simulation. To the best of our knowledge, this is the first study for the effect of GWs on nonlinear structure formation using $N$-body simulations. We found that GWs indeed influence nonlinear structure formation in both the clustering statistics and the intrinsic alignment of halo shapes. Our main finding is that the impact of GWs on large-scale structure can be well described by combining the impact of the scalar tides with the perturbation theory. Specifically, the halo tidal bias and linear shape bias induced by GWs, $b_{s^{2}}^{\rm GW}$ and $b_{K}^{\rm GW}$ respectively, can be approximated by $b_{X}^{\rm GW}(k_{L})=\frac{7}{4}\alpha(k_{L})b_{X}^{\rm scalar}$ where $X=\\{s^{2},K\\}$. Figure 12: Shape auto-power spectra from GWs at $z=1$ assuming $r=0.1$. The blue, orange, and green lines show the monopole of the $E$-mode auto-, the monopole of the $B$-mode auto-, and the dipole of the $EB$ cross-power spectra, respectively. For the $EB$ cross power spectrum we also assume a maximally parity-violating case, i.e., $\chi(k)=1$. The blue dashed line depicts the shape noise. Let us discuss possible observables that could be used to probe GWs from LSS. As GWs affect halo shapes at linear order shown in Sec. VI, the simplest probe would be the shape correlation, namely the intrinsic alignment from GWs Schmidt and Jeong (2012a); Schmidt _et al._ (2014); Biagetti and Orlando (2020). Although the shape correlation is primarily affected by the scalar perturbations, we can use the $E/B$-decomposition to distinguish the GWs contribution from the scalar contribution at linear order. Under the flat-sky approximation and assuming the line-of-sight direction, $\hat{n}$, is parallel to the $z$-axis, we can define $E$-mode and $B$-mode via $\displaystyle E({\bf k},\hat{n})\pm iB({\bf k},\hat{n})\equiv{{}_{\pm 2}\gamma}({\bf k},\hat{n})e^{\mp 2i\phi_{k}},$ (76) where $\displaystyle{{}_{\pm 2}\gamma}({\bf k},\hat{n})\equiv m_{\mp}^{i}(\hat{n})m_{\mp}^{j}(\hat{n})\gamma_{ij}({\bf k}),$ (77) with ${\bf m}_{\pm}\equiv(1,\mp i,0)/\sqrt{2}$. At linear order, the scalar perturbations only induce $E$-mode with vanishing $B$-mode Hirata and Seljak (2004). On the other hand, GWs generate both $E$-mode and $B$-mode as $\displaystyle E({\bf k}_{L},\hat{n})$ $\displaystyle=b_{K}^{\rm GW}(k_{L})\frac{1}{8}(1+\mu_{L}^{2})\sum_{\lambda}h_{(\lambda)}({\bf k}_{L}),$ (78) $\displaystyle B({\bf k}_{L},\hat{n})$ $\displaystyle=-b_{K}^{\rm GW}(k_{L})\frac{i}{2}\mu_{L}\sum_{\lambda}\frac{\lambda}{2}h_{(\lambda)}({\bf k}_{L}),$ (79) where we have used Eq. (70) and $\mu_{L}\equiv\hat{k}_{L}\cdot\hat{n}$. Note that here we project the halo shapes onto two-dimensional plane but do not project their position; we can combine photometric and spectroscopic surveys to get the projected shapes and their three-dimensional positions (e.g., Okumura and Taruya, 2020; Kurita _et al._ , 2021; Akitsu _et al._ , 2021b; Kurita and Takada, 2022). Defining the power spectra for $E$\- and $B$-modes via $\displaystyle\langle X({\bf k})Y^{*}({\bf k}^{\prime})\rangle\equiv(2\pi)^{3}\delta^{(3)}_{\rm D}({\bf k}-{\bf k}^{\prime})P_{XY}({\bf k}),$ (80) we obtain $\displaystyle P_{EE}(k_{L},\mu_{L};z)$ $\displaystyle=\frac{1}{64}(1+\mu_{L}^{2})^{2}(b_{K}^{\rm GW}(z))^{2}P_{h}(k_{L})\simeq\frac{49}{1024}(1+\mu_{L}^{2})^{2}\alpha^{2}(k_{L};z)(b_{K}^{\rm scalar}(z))^{2}P_{h}(k_{L}),$ (81) $\displaystyle P_{BB}(k_{L},\mu_{L};z)$ $\displaystyle=\frac{1}{4}\mu_{L}^{2}(b_{K}^{\rm GW}(z))^{2}P_{h}(k_{L})\simeq\frac{49}{64}\mu_{L}^{2}\alpha^{2}(k_{L};z)(b_{K}^{\rm scalar}(z))^{2}P_{h}(k_{L}),$ (82) and $P_{EB}=0$ for unpolarized GWs, where we have used Eq. (75). Considering the chiral GWs, there appears a non-vanishing $EB$ correlation as $\displaystyle P_{EB}(k_{L},\mu;z)$ $\displaystyle=\frac{i}{16}\mu_{L}(1+\mu_{L}^{2})(b_{K}^{\rm GW}(z))^{2}\chi(k_{L})P_{h}(k_{L})\simeq i\frac{49}{256}\mu_{L}(1+\mu_{L}^{2})\alpha^{2}(k_{L};z)(b_{K}^{\rm scalar}(z))^{2}\chi(k_{L})P_{h}(k_{L}),$ (83) where $\chi$ is defined in Eq. (7). Notice that $P_{h}(k_{L})$ is the primordial power spectrum of the tensor mode. In Fig. 12, we plot the lowest order moment of the multipoles of these spectra at $z=1$, i.e., the monopole of the $EE$ and $BB$ spectra and the dipole of the $EB$ spectrum, assuming $b_{K}^{\rm scalar}=0.1$, $r=0.1$, and $\chi(k)=1$ as a demonstration. On large scales, the suppression comes from the behaviour of $\alpha(k_{L})$ while on small scales the shape obeys a power-law with $P(k_{L})\propto k_{L}^{-3}$ because $\alpha(k_{L})$ does not change much. We also display the shape noise as the dashed line, assuming $\sigma^{2}_{\gamma}=0.2$ and $\bar{n}_{g}=5.0\times 10^{-4}~{}(h/{\rm Mpc})^{3}$. As is evident, for all cases the shape noise contribution is much greater than the expected signals, meaning that the detection of GWs using these spectra is very challenging. There should also be the contributions of the scalar-mode both in $EE$ (at linear order) and $BB$ (at one-loop order), though they are absent in the $EB$ spectrum Biagetti and Orlando (2020). 666Ref. Biagetti and Orlando (2020) argued that the shape noise is absent in the $EB$ power spectrum, making it a cleaner probe of the chiral GWs. Although this is true at the signal level, the covariance of the $EB$ spectrum includes the $EE$ and $BB$ spectra with the shape noise. Thus even for the $EB$ power spectrum the detectability is limited by the shape noise. Still, this sort of probes allows us to put the upper limit on the amplitude of GWs that are generated after the recombination or the reionization, which is not constrained by the CMB. Specifically, we could obtain the 1-$\sigma$ error on $r$ of order $\sigma(r)\sim 10^{3}$ at $k\sim 10^{-3}~{}h/{\rm Mpc}$, implying the total energy spectrum of GWs, $\Omega_{\rm GW}(k)$, could be constrained as $\Omega_{\rm GW}(k)\lesssim 10^{-8}$ at $k\sim 10^{-3}~{}h/{\rm Mpc}$ from the current galaxy surveys. The density-density-shape bispectrum also involves the GWs contribution. For instance, the tree-level bispectrum of $\delta_{\rm h}$-$\delta_{\rm h}$-$B$ in squeezed limit ($k_{L}\to 0$) is found to be $\displaystyle\lim_{k_{L}\to 0}iB^{\rm grav.}_{{\rm hh}B}(k,k_{L};z)=$ $\displaystyle\frac{1}{2}b_{1}(z)b_{K}^{\rm GW}(k_{L};z)\left[\left(2b_{1}(z)\alpha(k_{L};z)+2b^{\rm GW}_{s^{2}}(k_{L};z)+b_{1}(z)\beta(k_{L};z)\frac{\partial}{\partial\ln{k}}\right)P_{\rm lin}(k;z)\right]$ $\displaystyle\hskip 241.84842pt\times\mu_{L}\hat{k}^{i}\hat{k}^{j}\sum_{\lambda}\frac{\lambda}{2}e^{(\lambda)}_{ij}(\hat{k}_{L})P_{h_{(\lambda)}}(k_{L})$ (84) $\displaystyle\simeq$ $\displaystyle\frac{7}{8}b_{1}(z)b_{K}^{\rm scalar}(z)\alpha(k_{L};z)\left[\left(2(b_{1}(z)+\frac{7}{4}b^{\rm scalar}_{s^{2}}(z))\alpha(k_{L};z)+b_{1}(z)\beta(k_{L};z)\frac{\partial}{\partial\ln{k}}\right)P_{\rm lin}(k;z)\right]$ $\displaystyle\hskip 241.84842pt\times\mu_{L}\hat{k}^{i}\hat{k}^{j}\sum_{\lambda}\frac{\lambda}{2}e^{(\lambda)}_{ij}(\hat{k}_{L})P_{h_{(\lambda)}}(k_{L}),$ (85) where we neglect the contributions from the projection effect Schmidt and Jeong (2012b) and the redshift-space distortion Kaiser (1987). Notice again that here $P_{h_{(\lambda)}}(k_{L})$ is the primordial power spectrum and all the redshift dependence is encoded in the bias coefficients ($b_{K}^{\rm GW}(k_{L};z)$ and $b^{\rm GW}_{s^{2}}(k_{L};z)$), $\alpha(k_{L};z)$, and $\beta(k_{L};z)$. The superscript “grav.” emphasises that this bispectrum is induced by the gravitational interaction in the late time universe. In other words, there could be an additional contribution from the primordial universe, e.g., the scalar-scalar-tensor non-Gaussianity Endlich _et al._ (2013); Domènech _et al._ (2017). In principle we can directly observe the scalar- scalar-tensor non-Gaussianity by looking at this density-density-shape bispectrum. Another promising observable to probe GWs from LSS is the quadrupolar anisotropic imprint in the local density power spectrum discussed in Sec. IV and Sec. V. One can construct the optimal quadratic estimator for GWs by using the anisotropic imprint in the local power spectrum Masui and Pen (2010); Jeong and Kamionkowski (2012); Dimastrogiovanni _et al._ (2014); Dai _et al._ (2013); Masui _et al._ (2017). Our result can be used to increase $k_{\rm max}$ of the estimator in this method, allowing for the improved detectability. As we demonstrated for the first time how GWs affect the biased tracer, a more realistic estimator for the biased tracer can be available. Taking the cross correlation with the CMB would also be valuable to improve the detectability Dodelson (2010); Alizadeh and Hirata (2012); Chisari _et al._ (2014); Philcox and Johnson (2022). We leave the detailed investigations on these possibilities for future work. Let us conclude by mentioning one potentially significant effect of GWs on LSS observables. In this paper, we focused on the effect of GWs on the cold dark matter (CDM) perturbations and ignored the effect on the photon-baryon fluid in the early universe. However, given that the biased tracers such as galaxies and halos trace the CDM-baryon perturbation, this could leave a distinctive signature in LSS observables as well. We will investigate it in future work. ###### Acknowledgements. We thank Giovanni Cabass, William Coulton, and Matias Zaldarriaga for insightful discussions and Fabian Schmidt for valuable comments on the draft. KA is supported by JSPS Overseas Research Fellowships. TO acknowledges support from the Ministry of Science and Technology of Taiwan under Grants Nos. MOST 110-2112-M-001-045- and 111-2112-M-001-061- and the Career Development Award, Academia Sinina (AS-CDA-108-M02) for the period of 2019 to 2023. Numerical computation was carried out on the Helios and the Typhon cluster at the Institute for Advanced Study and Popeye-Simons cluster at San Diego Supercomputer Center. The Flatiron Institute is supported by the Simons Foundation. ## Appendix A From global FLRW coordinates to local CFC coordinates Here we outline the mapping from the global perturbed FLRW metric to the local CFC coordinates. More detailed discussion can be found in Refs. Pajer _et al._ (2013); Schmidt _et al._ (2014); Dai _et al._ (2015b). We start from the perturbed FLRW metric, $\displaystyle ds^{2}=a^{2}(\eta)\left[-(1+h_{00}(\eta,{\bf x}))d\eta^{2}+\left(\delta^{\textsc{k}}_{ij}+h_{ij}(\eta,{\bf x})\right)dx^{i}dx^{j}\right],$ (86) where we neglect $h_{0i}$ component. The coordinate transformation is given by $\displaystyle x^{0}$ $\displaystyle=x_{F}^{0}+\frac{1}{2}\int_{0}^{x^{0}_{F}}h_{00}(\tilde{\eta})\differential\tilde{\eta}+v_{i}x_{F}^{i}-\frac{1}{4}h^{\prime}_{ij}x_{F}^{i}x_{F}^{j},$ (87) $\displaystyle x^{i}$ $\displaystyle=x_{F}^{i}+v^{i}(x^{0}_{F}-\eta_{F})-\frac{1}{2}h^{i}_{\ j}x_{F}^{j}-\frac{1}{4}\left[h^{i}_{\ j,k}+h^{i}_{\ k,j}-h_{jk}^{\ \ ,i}\right]x_{F}^{j}x_{F}^{k},$ (88) where $v_{i}$ is the coordinate velocity of the central geodesic and $h_{\mu\nu}$ is evaluated along the central geodesics. The CFC metric can be obtained by using the transformation law of the metric, $g_{\mu\nu}^{F}(\eta_{F},{\bf x}_{F})=g_{\alpha\beta}(\eta,{\bf x})(\partial x^{\alpha}/\partial x^{\mu}_{F})(\partial x^{\beta}/\partial x^{\nu}_{F})$, yielding $\displaystyle g_{00}^{F}$ $\displaystyle=-a_{F}^{2}(\eta_{F})\left[1-\frac{1}{2}\left(h_{00,ij}+\mathcal{H}h^{\prime}_{ij}+h^{\prime\prime}_{ij}\right)x_{F}^{i}x_{F}^{j}\right],$ (89) $\displaystyle a_{F}(x^{0}_{F})$ $\displaystyle=a\left(\eta=\eta_{F}+\frac{1}{2}\int_{0}^{\eta_{F}}h_{00}(\tilde{\eta})\differential\tilde{\eta}\right).$ (90) In Eq. (89), setting $h_{00}=0$ coincides to the main text (Eqs. (9)-(10)). On the other hand, the case where $h_{00}=-2\Phi_{L}$ and $h_{ij}=0$ corresponds to the usual tidal separate universe simulation picture (see App. B). ## Appendix B The tidal responses to the scalar tides from tidal separete universe simulations Figure 13: Growth response of matter auto-power spectrum to the scalar tides, $R^{\rm scalar;mm}_{\rm growth}(k)$ for various redshifts, measured from the simulations. The blue dashed line corresponds to the perturbation theory prediction, $R^{\rm scalar}_{\rm growth;mm}=8/7$ Akitsu _et al._ (2017). In this appendix, we summarize the tidal responses to the scalar tidal field. These include not only the tidal response of the matter auto-power spectrum and the linear shape bias, which are already presented in Refs. Akitsu _et al._ (2021a); Stücker _et al._ (2021), but also the tidal responses of halo- matter and halo auto-power spectra and the halo tidal bias. Figure 14: Growth responses involving the halo density field as a function of $k$. In the left panels the responses for various resfhits are plotted while in the right panels the responses for various halo masses are plotted. The top panels show the response of the halo-matter cross-power spectrum, $R^{\rm scalar}_{\rm hm;growth}(k)$, The middle panels show the response of the halo auto-power spectrum, $R^{\rm scalar}_{\rm hh;growth}(k)$, normalized by the halo auto-power spectrum with the shot noise, and the bottom panels show the response of the halo auto-power spectrum, $R^{\rm scalar}_{\rm hh;growth}(k)$, normalized by the halo auto-power spectrum without the shot noise. First, we sketch the construction of the tidal separate universe in the scalar tides case, focusing on the difference to the GWs case, although we refer the reader to Ref. Akitsu _et al._ (2021a) for details. The difference between the GWs case and scalar-tides case in the induced tidal field in the local region, $\tau_{ij}$, results in the different initial conditions and anisotropic scale factors. In the scalar-tides case, the induced tides become $\displaystyle\tau_{ij}(\eta)=-\frac{1}{2}h_{00,ij}$ $\displaystyle=\Phi_{L,ij}$ (91) $\displaystyle=\frac{1}{2}\Omega_{\rm m}(\eta)\mathcal{H}^{2}\delta_{L}(\eta)\delta^{\textsc{k}}_{ij}+\frac{3}{2}\Omega_{\rm m}(\eta)\mathcal{H}^{2}s_{L,ij}(\eta),$ (92) where we have used Poisson equation and decomposed into the trace (large-scale overdensity, $\delta_{L}$) and the traceless (large-scale pure tidal field, $s_{L,ij}$). The construction of the tidal separate universe with the scalar tides corresponds to replacing Eq. (10) with Eq. (92) and repeating the analysis in Sec. II.2, Sec. III.1, and Sec. III.2. Focusing on the pure tidal mode ($s_{L,ij}$), the equation that governs the evolution of the anisotropic scale factors, $\Delta_{i}$, is now $\displaystyle\Delta^{\prime\prime}_{i}(\eta)+\mathcal{H}\Delta_{i}^{\prime}(\eta)=-\frac{3}{2}\Omega_{\rm m}(\eta)\mathcal{H}^{2}s_{L,i}(\eta),$ (93) whose solution is given by $\displaystyle\Delta_{i}(\eta)=-s_{L,i}(\eta_{0})\frac{D(\eta)}{D(\eta_{0})}.$ (94) The perturbative prediction of the growth response is $\displaystyle\lim_{k\to 0}R_{\rm mm;growth}^{\rm scalar}(k)$ $\displaystyle=\frac{8}{7},$ (95) where the response with respect to the scalar tides is defined through $\displaystyle\left.\frac{\differential{\ln P_{\cal G}}}{\differential{s_{L,ij}}}\right|_{{\bf k}_{\cal G}}=\left.\frac{\differential{\ln P_{\cal L}}}{\differential{s_{L,ij}}}\right|_{{\bf k}_{\cal G}}=$ $\displaystyle\left.\frac{\partial\ln P_{\cal L}}{\partial s_{L,ij}}\right|_{{\bf k}_{\cal L}}+\left.\frac{\partial\ln P_{\cal L}}{\partial\ln k_{{\cal L},\ell}}\right|_{s_{L,ij}}\left.\frac{\differential{\ln k_{{\cal L},\ell}}}{\partial s_{L,ij}}\right|_{{\bf k}_{\cal G}}$ $\displaystyle\equiv$ $\displaystyle\,\hat{k}_{i}\hat{k}_{j}\left[R^{\rm scalar}_{\rm growth}(k)+R^{\rm scalar}_{\rm dilation}(k)\right],$ (96) which is analogue to Eq. (49). One important difference of Eq. (96) from Eq. (49) is that in Eq. (96) we define the response with respect to the scalar tides at the same epoch rather than the initail epoch. As a result, the perturbative prediction remains the same for all redshifts. Also note that these results are independent of the wavenumber of long-modes. Fig. 13 shows the tidal response of the matter auto-power spectrum to the scalar tidal field as a function of $k$ for various redshifts, measured from tidal separate universe simulation with the scalar tides. This should be contrasted with Fig. 3 in the main text. In Fig. 14 we show the tidal responses of the halo-matter cross-power spectrum (the top panels), the halo auto-power spectrum with and without the short noise (the middle and the bottom panels respectively) for various redshifts (the left panels) and for various halo masses (the right panels). This is analogue to Figs. 5-6 in the main text. The differences of the responses for various redshifts and halo masses on large scales should be explained by the halo biases (the combination of $b_{1}(z;M)$ and $b^{\rm scalar}_{s^{2}}(z;M)$). The peaky feature appeared in the halo auto response normalized by the halo auto-power without the shot noise is due to the non-Poissonian feature of the shot noise term and thus not physical (see the discussion in the last paragraph in Sec. IV.3). Figure 15: Halo tidal bias measured from the halo-matter power spectrum response. The left: The tidal bias as a function of halo mass at $z=0.5$. The right: The tidal bias as a function of the linear bias $b_{1}$ from various redshifts and halo masses. The black dashed line displays the Lagrangian local-in-matter-density (LLIMD) prediction: $b_{s^{2}}^{\rm scalar}=-\frac{2}{7}(b_{1}-1)$. Figure 16: Linear shape bias induced by the scalar tides as a function of halo mass for various redshifts. We measure the halo tidal bias $b^{\rm scalar}_{s^{2}}$ in the same way as Eq. (66) and Eq. (67) in the main text. The result is presented in Fig. 15 where the left panel shows $b_{s^{2}}^{\rm scalar}$ at $z=0.5$ for various halo masses and the right panel shows $b_{s^{2}}^{\rm scalar}$ as a function of $b_{1}$. We compare the result with the Lagrangian local-in-matter-density (LLIMD) prediction Desjacques _et al._ (2018): $b_{s^{2}}^{\rm scalar}=-\frac{2}{7}(b_{1}-1)$, which is plotted as the black-dashed line. In general the LLIMD prediction fails to capture the behaivour of $b_{s^{2}}^{\rm scalar}$, in particular at high-mass end, which is consistent with Refs. Lazeyras and Schmidt (2018); Abidi and Baldauf (2018). In Fig. 16 we show the linear shape bias (or the linear alignment coefficient) induced by the scalar tides, $b_{K}^{\rm scalar}$, (introduced in Eq. (69)) as a function of halo mass for various redshifts. The estimator for $b_{K}^{\rm scalar}$ used here is the same as Eq. (74). Fig. 16 is a counterpart of Fig. 9. ## Appendix C The modification in the drift operator Figure 17: Ratio of the approximate drift integral to the exact one as a function of the scale factor. Left: the ratio for different wavenumber of GWs with the same starting redshift $z_{\rm ini}=99$. Right: the ratio for different starting redshifts with the same wavenumber of GWs $k_{L}=0.02~{}h/{\rm Mpc}$. In the tidal separate universe simulation, the drift operator changes from the usual $N$-body simulation as $\displaystyle x_{i;n+1}=x_{i;n}+\frac{P_{i}}{m}\int^{a_{n+1}}_{a_{n}}\frac{\differential a}{a_{i}^{2}\mathcal{H}}\equiv x_{i;n}+\frac{P_{i}}{m}~{}{\cal D}^{\rm exact}_{i;n},$ (97) where $P_{i}$ is the conjugate momenta of $x_{i}$, $i=x,y,z$ and $n$ represents the time step. In the original Gadget-2 code Springel (2005), instead of computing the drift integral in Eq. (97) at each time step, first it prepares the following table $\displaystyle{\cal D}_{i}[j]\equiv\int^{a[j]}_{a_{\rm ini}}\frac{\differential a}{a_{i}^{2}\mathcal{H}},$ (98) where $1\leq j\leq N_{\rm drift}$ with $N_{\rm drift}$ being the length of the table. $a[j]$ is the $j$-th scale factor that is sampled equally spaced in logarithm from $a_{\rm ini}$ to $a=1$, regardless of the actual time step. The actual drift integral at each time step is then evaluated by linearly interpolating this drift table as $\displaystyle{\cal D}^{\rm approx.}_{i;n}=$ $\displaystyle\int^{a_{n+1}}_{a_{\rm ini}}\frac{\differential a}{a_{i}^{2}\mathcal{H}}-\int^{a_{n}}_{a_{\rm ini}}\frac{\differential a}{a_{i}^{2}\mathcal{H}}$ $\displaystyle\simeq$ $\displaystyle\ {\cal D}_{i}[j+1]+\left(a_{n+1}-a[j+1]\right)\left({\cal D}_{i}[j+2]-{\cal D}_{i}[j+1]\right)$ $\displaystyle-{\cal D}_{i}[j]-\left(a_{n}-a[j]\right)\left({\cal D}_{i}[j+1]-{\cal D}_{i}[j]\right)$ (99) with $j=\lfloor n\rfloor$. This prescription works well for a monotonic integrand, which is the case for the usual cosmological simulation. We found, however, that this approximation for the drift operator breaks down for the tidal separate universe simulation with GWs where the integrand oscillates. Fig. 17 compares the drift factor evaluated by the above procedure and the direct calculation of the integral in Eq. (97). Although the default length of the drift table is 1000 we increased it to $N_{\rm drift}=200000$. Even with this large table, the approximated drift factor fails to capture the exact result. Therefore we modified the drift operator so that at each step the drift integral is directly evaluated without using the drift table or interpolation. This modification is particularly important for $k_{L}\gtrsim 0.02~{}h/{\rm Mpc}$ when $z_{\rm ini}=99$. In fact, without this modification the results suffers from the artifact. ## References * Starobinsky (1979) A. A. Starobinsky, JETP Lett. 30, 682 (1979). * Sato (1981) K. Sato, Mon. Not. Roy. Astron. Soc. 195, 467 (1981). * Guth (1981) A. H. Guth, Phys. Rev. D 23, 347 (1981). * Linde (1982) A. D. Linde, Phys. Lett. B 108, 389 (1982). * Maggiore (2000) M. Maggiore, Phys. Rept. 331, 283 (2000), arXiv:gr-qc/9909001 . * Komatsu _et al._ (2011) E. Komatsu _et al._ (WMAP), Astrophys. J. Suppl. 192, 18 (2011), arXiv:1001.4538 [astro-ph.CO] . * Akrami _et al._ (2020) Y. Akrami _et al._ (Planck), Astron. Astrophys. 641, A10 (2020), arXiv:1807.06211 [astro-ph.CO] . * Ade _et al._ (2018) P. A. R. Ade _et al._ (BICEP2, Keck Array), Phys. Rev. Lett. 121, 221301 (2018), arXiv:1810.05216 [astro-ph.CO] . * Adachi _et al._ (2022) S. Adachi _et al._ (POLARBEAR), Astrophys. J. 931, 101 (2022), arXiv:2203.02495 [astro-ph.CO] . * Abazajian _et al._ (2016) K. N. Abazajian _et al._ (CMB-S4), (2016), arXiv:1610.02743 [astro-ph.CO] . * Zaldarriaga and Seljak (1997) M. Zaldarriaga and U. Seljak, Phys. Rev. D 55, 1830 (1997), arXiv:astro-ph/9609170 . * Kamionkowski _et al._ (1997) M. Kamionkowski, A. Kosowsky, and A. Stebbins, Phys. Rev. D 55, 7368 (1997), arXiv:astro-ph/9611125 . * Dodelson (2003) S. Dodelson, _Modern Cosmology_ (Academic Press, Amsterdam, 2003). * Suto and Sasaki (1991) Y. Suto and M. Sasaki, Phys. Rev. Lett. 66, 264 (1991). * Makino _et al._ (1992) N. Makino, M. Sasaki, and Y. Suto, Phys. Rev. D 46, 585 (1992). * Bernardeau _et al._ (2002) F. Bernardeau, S. Colombi, E. Gaztanaga, and R. Scoccimarro, Phys. Rept. 367, 1 (2002), arXiv:astro-ph/0112551 . * Baumann _et al._ (2012) D. Baumann, A. Nicolis, L. Senatore, and M. Zaldarriaga, JCAP 07, 051 (2012), arXiv:1004.2488 [astro-ph.CO] . * Springel (2005) V. Springel, Mon. Not. Roy. Astron. Soc. 364, 1105 (2005), arXiv:astro-ph/0505010 . * Potter _et al._ (2016) D. Potter, J. Stadel, and R. Teyssier, (2016), arXiv:1609.08621 [astro-ph.IM] . * Springel _et al._ (2021) V. Springel, R. Pakmor, O. Zier, and M. Reinecke, Mon. Not. Roy. Astron. Soc. 506, 2871 (2021), arXiv:2010.03567 [astro-ph.IM] . * Angulo and Hahn (2021) R. E. Angulo and O. Hahn, (2021), 10.1007/s41115-021-00013-z, arXiv:2112.05165 [astro-ph.CO] . * Kaiser and Jaffe (1997) N. Kaiser and A. H. Jaffe, Astrophys. J. 484, 545 (1997), arXiv:astro-ph/9609043 . * Dodelson _et al._ (2003) S. Dodelson, E. Rozo, and A. Stebbins, Phys. Rev. Lett. 91, 021301 (2003), arXiv:astro-ph/0301177 . * Yoo _et al._ (2009) J. Yoo, A. L. Fitzpatrick, and M. Zaldarriaga, Phys. Rev. D 80, 083514 (2009), arXiv:0907.0707 [astro-ph.CO] . * Jeong _et al._ (2012) D. Jeong, F. Schmidt, and C. M. Hirata, Phys. Rev. D 85, 023504 (2012), arXiv:1107.5427 [astro-ph.CO] . * Schmidt and Jeong (2012a) F. Schmidt and D. Jeong, Phys. Rev. D 86, 083513 (2012a), arXiv:1205.1514 [astro-ph.CO] . * Jeong and Schmidt (2012) D. Jeong and F. Schmidt, Phys. Rev. D 86, 083512 (2012), arXiv:1205.1512 [astro-ph.CO] . * Masui and Pen (2010) K. W. Masui and U.-L. Pen, Phys. Rev. Lett. 105, 161302 (2010), arXiv:1006.4181 [astro-ph.CO] . * Dai _et al._ (2013) L. Dai, D. Jeong, and M. Kamionkowski, Phys. Rev. D 88, 043507 (2013), arXiv:1306.3985 [astro-ph.CO] . * Schmidt _et al._ (2014) F. Schmidt, E. Pajer, and M. Zaldarriaga, Phys. Rev. D 89, 083507 (2014), arXiv:1312.5616 [astro-ph.CO] . * Adamek _et al._ (2013) J. Adamek, D. Daverio, R. Durrer, and M. Kunz, Phys. Rev. D 88, 103527 (2013), arXiv:1308.6524 [astro-ph.CO] . * Adamek _et al._ (2016a) J. Adamek, D. Daverio, R. Durrer, and M. Kunz, Nature Phys. 12, 346 (2016a), arXiv:1509.01699 [astro-ph.CO] . * Adamek _et al._ (2016b) J. Adamek, D. Daverio, R. Durrer, and M. Kunz, JCAP 07, 053 (2016b), arXiv:1604.06065 [astro-ph.CO] . * Domènech and Sasaki (2021) G. Domènech and M. Sasaki, Phys. Rev. D 103, 063531 (2021), arXiv:2012.14016 [gr-qc] . * Sirko (2005) E. Sirko, Astrophys. J. 634, 728 (2005), arXiv:astro-ph/0503106 . * Gnedin _et al._ (2011) N. Y. Gnedin, A. V. Kravtsov, and D. H. Rudd, The Astrophysical Journal Supplement Series 194, 46 (2011). * Baldauf _et al._ (2011) T. Baldauf, U. Seljak, L. Senatore, and M. Zaldarriaga, JCAP 10, 031 (2011), arXiv:1106.5507 [astro-ph.CO] . * Li _et al._ (2014) Y. Li, W. Hu, and M. Takada, Phys. Rev. D 89, 083519 (2014), arXiv:1401.0385 [astro-ph.CO] . * Wagner _et al._ (2015) C. Wagner, F. Schmidt, C.-T. Chiang, and E. Komatsu, Mon. Not. Roy. Astron. Soc. 448, L11 (2015), arXiv:1409.6294 [astro-ph.CO] . * Lazeyras _et al._ (2016) T. Lazeyras, C. Wagner, T. Baldauf, and F. Schmidt, JCAP 02, 018 (2016), arXiv:1511.01096 [astro-ph.CO] . * Li _et al._ (2016) Y. Li, W. Hu, and M. Takada, Phys. Rev. D 93, 063507 (2016), arXiv:1511.01454 [astro-ph.CO] . * Baldauf _et al._ (2016) T. Baldauf, U. Seljak, L. Senatore, and M. Zaldarriaga, JCAP 09, 007 (2016), arXiv:1511.01465 [astro-ph.CO] . * Stücker _et al._ (2021) J. Stücker, A. S. Schmidt, S. D. M. White, F. Schmidt, and O. Hahn, Mon. Not. Roy. Astron. Soc. 503, 1473 (2021), arXiv:2003.06427 [astro-ph.CO] . * Masaki _et al._ (2020) S. Masaki, T. Nishimichi, and M. Takada, Mon. Not. Roy. Astron. Soc. 496, 483 (2020), arXiv:2003.10052 [astro-ph.CO] . * Akitsu _et al._ (2021a) K. Akitsu, Y. Li, and T. Okumura, JCAP 04, 041 (2021a), arXiv:2011.06584 [astro-ph.CO] . * Dai _et al._ (2015a) L. Dai, E. Pajer, and F. Schmidt, JCAP 10, 059 (2015a), arXiv:1504.00351 [astro-ph.CO] . * Akitsu _et al._ (2017) K. Akitsu, M. Takada, and Y. Li, Phys. Rev. D 95, 083522 (2017), arXiv:1611.04723 [astro-ph.CO] . * Akitsu and Takada (2018) K. Akitsu and M. Takada, Phys. Rev. D 97, 063527 (2018), arXiv:1711.00012 [astro-ph.CO] . * Akitsu _et al._ (2019) K. Akitsu, N. S. Sugiyama, and M. Shiraishi, Phys. Rev. D 100, 103515 (2019), arXiv:1907.10591 [astro-ph.CO] . * Li _et al._ (2018) Y. Li, M. Schmittfull, and U. Seljak, Journal of Cosmology and Astroparticle Physics 2018, 022 (2018). * Taruya and Akitsu (2021) A. Taruya and K. Akitsu, JCAP 11, 061 (2021), arXiv:2106.04789 [astro-ph.CO] . * Hu _et al._ (2016) W. Hu, C.-T. Chiang, Y. Li, and M. LoVerde, Phys. Rev. D 94, 023002 (2016), arXiv:1605.01412 [astro-ph.CO] . * Chiang _et al._ (2016) C.-T. Chiang, Y. Li, W. Hu, and M. LoVerde, Phys. Rev. D 94, 123502 (2016), arXiv:1609.01701 [astro-ph.CO] . * Chiang _et al._ (2018) C.-T. Chiang, W. Hu, Y. Li, and M. Loverde, Phys. Rev. D 97, 123526 (2018), arXiv:1710.01310 [astro-ph.CO] . * Ade _et al._ (2016) P. A. R. Ade _et al._ (Planck), Astron. Astrophys. 594, A13 (2016), arXiv:1502.01589 [astro-ph.CO] . * Manasse and Misner (1963) F. K. Manasse and C. W. Misner, J. Math. Phys. 4, 735 (1963). * Dai _et al._ (2015b) L. Dai, E. Pajer, and F. Schmidt, JCAP 11, 043 (2015b), arXiv:1502.02011 [gr-qc] . * Crocce _et al._ (2006) M. Crocce, S. Pueblas, and R. Scoccimarro, Mon. Not. Roy. Astron. Soc. 373, 369 (2006), arXiv:astro-ph/0606505 . * Knollmann and Knebe (2009) S. R. Knollmann and A. Knebe, Astrophys. J. Suppl. 182, 608 (2009), arXiv:0904.3662 [astro-ph.CO] . * Pajer _et al._ (2013) E. Pajer, F. Schmidt, and M. Zaldarriaga, Phys. Rev. D 88, 083502 (2013), arXiv:1305.0824 [astro-ph.CO] . * Smith _et al._ (2007) R. E. Smith, R. Scoccimarro, and R. K. Sheth, Phys. Rev. D 75, 063512 (2007), arXiv:astro-ph/0609547 . * Hamaus _et al._ (2010) N. Hamaus, U. Seljak, V. Desjacques, R. E. Smith, and T. Baldauf, Phys. Rev. D 82, 043515 (2010), arXiv:1004.5377 [astro-ph.CO] . * Baldauf _et al._ (2013) T. Baldauf, U. Seljak, R. E. Smith, N. Hamaus, and V. Desjacques, Phys. Rev. D 88, 083507 (2013), arXiv:1305.2917 [astro-ph.CO] . * Catelan _et al._ (2001) P. Catelan, M. Kamionkowski, and R. D. Blandford, Mon. Not. Roy. Astron. Soc. 320, L7 (2001), arXiv:astro-ph/0005470 . * Hirata and Seljak (2004) C. M. Hirata and U. Seljak, Phys. Rev. D 70, 063526 (2004), [Erratum: Phys.Rev.D 82, 049901 (2010)], arXiv:astro-ph/0406275 . * Desjacques _et al._ (2018) V. Desjacques, D. Jeong, and F. Schmidt, Phys. Rept. 733, 1 (2018), arXiv:1611.09787 [astro-ph.CO] . * Biagetti and Orlando (2020) M. Biagetti and G. Orlando, JCAP 07, 005 (2020), arXiv:2001.05930 [astro-ph.CO] . * Okumura and Taruya (2020) T. Okumura and A. Taruya, Mon. Not. R. Astron. Soc. 493, L124 (2020), arXiv:1912.04118 [astro-ph.CO] . * Kurita _et al._ (2021) T. Kurita, M. Takada, T. Nishimichi, R. Takahashi, K. Osato, and Y. Kobayashi, Mon. Not. Roy. Astron. Soc. 501, 833 (2021), arXiv:2004.12579 [astro-ph.CO] . * Akitsu _et al._ (2021b) K. Akitsu, T. Kurita, T. Nishimichi, M. Takada, and S. Tanaka, Phys. Rev. D 103, 083508 (2021b), arXiv:2007.03670 [astro-ph.CO] . * Kurita and Takada (2022) T. Kurita and M. Takada, Phys. Rev. D 105, 123501 (2022), arXiv:2202.11839 [astro-ph.CO] . * Schmidt and Jeong (2012b) F. Schmidt and D. Jeong, Phys. Rev. D 86, 083527 (2012b), arXiv:1204.3625 [astro-ph.CO] . * Kaiser (1987) N. Kaiser, Mon. Not. Roy. Astron. Soc. 227, 1 (1987). * Endlich _et al._ (2013) S. Endlich, A. Nicolis, and J. Wang, JCAP 10, 011 (2013), arXiv:1210.0569 [hep-th] . * Domènech _et al._ (2017) G. Domènech, T. Hiramatsu, C. Lin, M. Sasaki, M. Shiraishi, and Y. Wang, JCAP 05, 034 (2017), arXiv:1701.05554 [astro-ph.CO] . * Jeong and Kamionkowski (2012) D. Jeong and M. Kamionkowski, Phys. Rev. Lett. 108, 251301 (2012), arXiv:1203.0302 [astro-ph.CO] . * Dimastrogiovanni _et al._ (2014) E. Dimastrogiovanni, M. Fasiello, D. Jeong, and M. Kamionkowski, JCAP 12, 050 (2014), arXiv:1407.8204 [astro-ph.CO] . * Masui _et al._ (2017) K. W. Masui, U.-L. Pen, and N. Turok, Phys. Rev. Lett. 118, 221301 (2017), arXiv:1702.06552 [astro-ph.CO] . * Dodelson (2010) S. Dodelson, Phys. Rev. D 82, 023522 (2010), arXiv:1001.5012 [astro-ph.CO] . * Alizadeh and Hirata (2012) E. Alizadeh and C. M. Hirata, Phys. Rev. D 85, 123540 (2012), arXiv:1201.5374 [astro-ph.CO] . * Chisari _et al._ (2014) N. E. Chisari, C. Dvorkin, and F. Schmidt, Phys. Rev. D 90, 043527 (2014), arXiv:1406.4871 [astro-ph.CO] . * Philcox and Johnson (2022) O. H. E. Philcox and M. C. Johnson, (2022), arXiv:2206.07054 [astro-ph.CO] . * Lazeyras and Schmidt (2018) T. Lazeyras and F. Schmidt, JCAP 09, 008 (2018), arXiv:1712.07531 [astro-ph.CO] . * Abidi and Baldauf (2018) M. M. Abidi and T. Baldauf, JCAP 07, 029 (2018), arXiv:1802.07622 [astro-ph.CO] .
# A discrete Darboux–Lax scheme for integrable difference equations X. Fisenko<EMAIL_ADDRESS>S. Konstantinou-Rizos <EMAIL_ADDRESS>P. Xenitidis<EMAIL_ADDRESS>School of Mathematics, Computer Science and Engineering, Liverpool Hope University, L19 9JD Liverpool, UK ###### Abstract We propose a discrete Darboux–Lax scheme for deriving auto-Bäcklund transformations and constructing solutions to quad-graph equations that do not necessarily possess the 3D consistency property. As an illustrative example we use the Adler–Yamilov type system which is related to the nonlinear Schrödinger (NLS) equation [19]. In particular, we construct an auto-Bäcklund transformation for this discrete system, its superposition principle, and we employ them in the construction of the one- and two-soliton solutions of the Adler–Yamilov system. PACS numbers: 02.30.Ik, 02.90.+p, 03.65.Fd. Mathematics Subject Classification 2020: 37K60, 39A36, 35Q55, 16T25. Keywords: Darboux transformations, Bäcklund transformations, quad-graph equations, partial Keywords: difference equations, integrable lattice equations, 3D-consistency, soliton solutions. ## 1 Introduction It has become understood over the past few decades that integrable systems of partial difference equations (P$\Delta$Es) are interesting in their own right, see for instance [15] and references therein. On the one hand, they may model various natural phenomena, as well as processes in industry and the IT sector. On the other hand, they have many interesting algebro-geometric properties [6, 15], and are related to many important equations of Mathematical Physics such as the Yang–Baxter equation and the tetrahedron equation [1, 9, 16, 25]. Moreover, they can be derived from the discretisation of nonlinear partial differential equations (PDEs). One such approach is provided by the Darboux transformations of integrable nonlinear PDEs of evolution type [19, 30]; namely, the resulting relations from the permutability of two Darboux transformations can be interpreted as a P$\Delta$E. In this paper, we focus on a special class of P$\Delta$Es, the so-called quad- graph equations, or systems thereof. Quad-graph systems are equations of P$\Delta$Es defined on an elementary quadrilateral of the two-dimensional lattice. In particular, they are systems of the form $Q(f_{00},f_{10},f_{01},f_{11};\alpha,\beta)=0,$ (1) where $Q$ is a function of its arguments $f_{ij}$, $i,j=0,1$, and may also depend on parameters $\alpha$ and $\beta$. Schematically, equation (2) can be represented on the square, where $f_{00}$, $f_{10}$, $f_{01}$ and $f_{11}$ are placed on vertices of the square and the parameters $\alpha$, $\beta$ are placed on the edges as in Figure 2. Equation (2) can be thought as a partial difference equation (P$\Delta$E) by identifying $f$ with a function of two discrete variables $n,m\in\mathbb{Z}$, and $f_{ij}$ with its shifts in the $n$ and $m$ direction, i.e. $f_{ij}=f(n+i,m+j)$. .8 (0 0) (1 0) (1 1) (0 1) (0 0) (0 0) f:0.0 r:0.05 (1 0) f:0.0 r:0.05 (0 1) f:0.0 r:0.05 (1 1) f:0.0 r:0.05 (-.1 -.3) $f$ (.9 -.3) $f_{10}$ (-.1 1.1) $f_{01}$ (.9 1.1) $f_{11}$ (0.44 -.2) $\alpha$ (-.15 .45) $\beta$ Figure 1: Quad–graph equation. Quad-graph equations have attracted the interest of many researchers in the field of discrete integrable systems, see [15] for a review. This led to the developement of methods for solving them, e.g. [2, 13, 15, 24, 26], classifying them, e.g. [3, 7, 28], analysing their integrability properties, e.g. [22, 27, 21, 29], and relating them to the theory of Yang–Baxter and tetrahedron maps, e.g. [11, 25]. Of particular interest are the so-called 3D-consistent equations which can be extended to the three-dimensional lattice in a consistent way. For such equations a Lax representation and a Bäcklund transformation can be constructed in a systematic manner by employing the 3D consistency of the system [15]. In this paper, we propose a discrete Darboux–Lax scheme for deriving Bäcklund transformations and constructing solutions for quad-graph equations which are not necessarily 3D consistent but have a Lax representation. More precisely, we employ gauge-type transformations of the Lax pair and demonstrate how they give rise to Bäcklund transformations for the discrete system. As an illustrative example we use the Adler–Yamilov system which was derived as a discretisation of the NLS equation via a Darboux transformation in [19]. The paper is organised as follows. In the next section we provide all the necessary definitions for the text to be self-contained. More specifically, after fixing our notation, we give the definitions of the integrability of a system of P$\Delta$Es in the sense of the existence of a Lax pair, and of the Bäcklund transformation. In Section 3 we present the discrete Darboux–Lax scheme for constructing Bäcklund transformations and solving integrable P$\Delta$Es which do not necessarily possess the 3D consistency property. In Section 4 we apply our results to the Adler–Yamilov system given in [19] and derive the one- and two-soliton solutions by using the associated Bäcklund transformation and the corresponding Bianchi diagram. Finally, the closing section contains a summary of the obtained results and a discussion on how they can be extended and generalised. ## 2 Integrability of difference equations Let us start this section by introducing our notation. In what follows we deal with systems of equations which involve unknown functions depending on the two discrete variables $n$ and $m$. The dependence of a function $f=f(n,m)$ on these variables will be denoted with indices in the following way. $f_{ij}=f(n+i,m+j),\quad{\mbox{for all }}i,j\in{\mathbb{Z}}$ We also denote by $\mathcal{S}$ and $\mathcal{T}$ the shift operators in the $n$ and $m$ direction, respectively. Their action on a function $f$ is defined as ${\mathcal{S}}^{k}(f_{00})=f_{k0},\quad{\mathcal{T}}^{\ell}(f_{00})=f_{0\ell}.$ We denote vectors with bold face letters, e.g. $\textbf{f}_{00}=(f_{00}^{(1)},f_{00}^{(2)},\ldots f_{00}^{(k)})$. For systems of equations we also use bold letters. In particular, a system of quad-graph equations will be denoted by $\textbf{Q}({\textbf{f}}_{00},\textbf{f}_{10},\textbf{f}_{01},\textbf{f}_{11})=0.$ (2) Finally, we denote matrices with roman uppercase letters. For instance $\rm{L}(\mathbf{f}_{00},\mathbf{f}_{10};\alpha,\lambda)$ denotes a matrix with elements depending on $\mathbf{f}_{00}$, $\mathbf{f}_{10}$, $\alpha$ and $\lambda$. The semicolon in the arguments of the matrix is used to separate the fields $(\mathbf{f}_{00},\mathbf{f}_{10})$ from the parameters $\alpha$, $\lambda$. By $\lambda$ we denote the spectral parameter throughout the text. With our notation, let ${\rm{{L}}}(\mathbf{f}_{00},\mathbf{f}_{10};\lambda)$ and ${\rm{M}}(\mathbf{f}_{00},\mathbf{f}_{01};\lambda)$ be two $k\times k$ invertible matrices which depend on a function $\bf{f}$ and the spectral parameter $\lambda$.111Matrices ${\rm{L}}$, ${\rm{M}}$ may also depend on parameters but, as they do not play any role in our discussion in this and the following section, we suppress this dependence. Let also $\Psi=\Psi(n,m)$ be an auxiliary $k\times k$ matrix, and consider the following overdetermined linear system. ${\cal{S}}(\Psi)={\rm{L}}(\mathbf{f}_{00},\mathbf{f}_{10};\lambda)\Psi,\qquad{\cal{T}}(\Psi)={\rm{M}}(\mathbf{f}_{00},\mathbf{f}_{01};\lambda)\Psi.$ (3) For given $\bf{f}$, this system has a solution $\Psi$ provided that the two equations are consistent, i.e. the compatibility condition ${\cal{T}}\left({\cal{S}}(\Psi)\right)={\cal{S}}\left({\cal{T}}(\Psi)\right)$ holds. The latter condition can be written explicitly as ${\rm{L}}(\mathbf{f}_{01},\mathbf{f}_{11};\lambda){\rm{M}}(\mathbf{f}_{00},\mathbf{f}_{01};\lambda)={\rm{M}}(\mathbf{f}_{10},\mathbf{f}_{11};\lambda){\rm{L}}(\mathbf{f}_{00},\mathbf{f}_{10};\lambda).$ (4) If the above equation holds if and only if $\bf{f}$ satisfies (2), then we say that system of P$\Delta$Es (2) is _integrable_ , system (3) is a _Lax pair_ for (2), and equation (4) is called a _Lax representation_ for (2). Moreover, matrices ${\rm{{L}}}(\mathbf{f}_{00},\mathbf{f}_{10};\lambda)$ and ${\rm{M}}(\mathbf{f}_{00},\mathbf{f}_{01};\lambda)$ are referred to as _Lax matrices_ , and without loss of generality we assume that they have constant determinants. We close this section by giving the definition of Bäcklund transformation for quad-graph equations. Such transformations are related to the notion of integrability as it will become evident in the next section where we explore their connection to Lax pairs via the Darboux–Lax scheme. ###### Definition 2.1. Let ${\bf{Q}}[{\bf{f}}]:={\bf{Q}}({\bf{f}}_{00},{\bf{f}}_{10},{\bf{f}}_{01},{\bf{f}}_{11})=0$ and ${\bf{P}}[{\bf{g}}]:={\bf{P}}({\bf{g}}_{00},{\bf{g}}_{10},{\bf{g}}_{01},{\bf{g}}_{11})=0$ be two systems of quad-graph equations. Let also $\mathcal{B}(\mathbf{f}_{00},\mathbf{f}_{10},\mathbf{f}_{01},\mathbf{g}_{00},\mathbf{g}_{10},\mathbf{g}_{01};\varepsilon)=0$ (5) be a system of P$\Delta$Es. If system $\mathcal{B}=0$ can be integrated for $\bf{g}$ provided that $\bf{f}$ is a solution of $\mathbf{Q}[{\bf{f}}]=0$, and the resulting ${\bf{g}}(n,m)$ is a solution to $\mathbf{P}[{\bf{g}}]=0$, and vice versa, then system (5) is called a (hetero-) Bäcklund transformation for equations ${\bf{Q}}[{\bf{f}}]=0$ and ${\bf{P}}[{\bf{g}}]=0$. If $\mathbf{Q}[{\bf{a}}]=\mathbf{P}[{\bf{a}}]$, then (5) is called an auto- Bäcklund transformation for equation ${\bf{Q}}[{\bf{f}}]=0$. ## 3 Discrete Darboux–Lax scheme It is well known that for quad-graph systems which possess the 3D consistency property a Lax representation can be derived algorithmically [23, 5, 8, 15] and a Bäcklund transformation can be constructed systematically [2, 15]. However, there do exist quad-graph systems which have a Lax pair but do not possess the 3D consistency property. For this kind of systems we propose here a scheme for constructing Darboux and Bäcklund transformations. It should be emphasized here that this scheme works for _any_ system of difference equations irrespectively of their 3D consistency. For the integrable quad-graph system $\mathbf{Q}(\mathbf{f}_{00},\mathbf{f}_{10},\mathbf{f}_{01},\mathbf{f}_{11})=0\quad\Longleftrightarrow\quad{\rm{L}}(\mathbf{f}_{01},\mathbf{f}_{11};\lambda){\rm{M}}(\mathbf{f}_{00},\mathbf{f}_{01};\lambda)={\rm{M}}(\mathbf{f}_{10},\mathbf{f}_{11};\lambda){\rm{L}}(\mathbf{f}_{00},\mathbf{f}_{10};\lambda),$ (6) we define the discrete Darboux transformation as follows. ###### Definition 3.1. A discrete Darboux transformation for the integrable P$\Delta$E (6) is a gauge-like, spectral parameter-dependent transformation that leaves Lax matrices $\rm{L}$ and $\rm{M}$ covariant. That is, a transformation which involves an invertible matrix $\rm{B}$ such that $\displaystyle{\rm{L}}(\mathbf{f}_{00},\mathbf{f}_{10};\lambda)\longmapsto{\rm{L}}(\tilde{\mathbf{f}}_{00},\tilde{\mathbf{f}}_{10};\lambda)={\cal{S}}({\rm{B}}){\rm{L}}(\mathbf{f}_{00},\mathbf{f}_{10};\lambda){\rm{B}}^{-1},$ (7a) $\displaystyle{\rm{M}}(\mathbf{f}_{00},\mathbf{f}_{01};\lambda)\longmapsto{\rm{M}}(\tilde{\mathbf{f}}_{00},\tilde{\mathbf{f}}_{01};\lambda)={\cal{T}}({\rm{B}}){\rm{M}}(\mathbf{f}_{00},\mathbf{f}_{01};\lambda){\rm{B}}^{-1}.$ (7b) A consequence of the above definition is the following proposition. ###### Proposition 3.2. The Darboux transformation maps fundamental solutions of the linear system ${\cal{S}}(\Psi)={\rm{L}}(\mathbf{f}_{00},\mathbf{f}_{10};\lambda)\Psi,\quad{\cal{T}}(\Psi)={\rm{M}}(\mathbf{f}_{00},\mathbf{f}_{01};\lambda)\Psi,$ (8) to fundamental solutions of the linear system ${\cal{S}}(\tilde{\Psi})={\rm{L}}(\tilde{\mathbf{f}}_{00},\tilde{\mathbf{f}}_{10};\lambda)\tilde{\Psi},\quad{\cal{T}}(\tilde{\Psi})={\rm{M}}(\tilde{\mathbf{f}}_{00},\tilde{\mathbf{f}}_{01};\lambda)\tilde{\Psi},$ (9) via the relation $\tilde{\Psi}={\rm{B}}\Psi$. ###### Proof. Let $\Psi=\Psi(n,m)$ be a fundamental solution of the linear problem (8). We set $\tilde{\Psi}={\rm{B}}\Psi$ and then we shift in the $n$ direction to find that ${\cal{S}}(\tilde{\Psi})={\cal{S}}({\rm{B}}){\cal{S}}(\Psi)\stackrel{{\scriptstyle\eqref{linear- sys-1}}}{{=}}{\cal{S}}({\rm{B}}){\rm{L}}(\mathbf{f}_{00},\mathbf{f}_{10};\lambda)\Psi\stackrel{{\scriptstyle\eqref{Darboux- def-L}}}{{=}}{\rm{L}}(\tilde{\mathbf{f}}_{00},\tilde{\mathbf{f}}_{10};\lambda){\rm{B}}\Psi={\rm{L}}(\tilde{\mathbf{f}}_{00},\tilde{\mathbf{f}}_{10};\lambda)\tilde{\Psi}.$ Similarly, starting with $\tilde{\Psi}={\rm{B}}\Psi$ we shift in the $m$ direction and employ (8) and (7b) to find that ${\cal{T}}(\tilde{\Psi})={\rm{M}}(\tilde{\mathbf{f}}_{00},\tilde{\mathbf{f}}_{01};\lambda)\tilde{\Psi}$. Moreover, the solution $\tilde{\Psi}$ is fundamental, since $\Psi$ is fundamental and $\det(\tilde{\Psi})=\det{\rm{B}}\det{\Psi}\neq 0$. ∎ Using the above definition of the discrete Darboux transformation and corresponding Darboux matrix, we propose the following approach for the construction of a Darboux matrix and Bäcklund transformation, as well as for the derivation of the superposition principle for the Bäcklund transformation. * • We start by assuming an initial form for matrix $\rm{B}$. The simplest assumption we can make is that matrix ${\rm{B}}$ depends linearly on the spectral parameter, i.e. ${\rm{B}}=\lambda{\rm{B}}^{(1)}+{\rm{B}}^{(0)},$ (10) where matrices ${\rm{B}}^{(1)}$ and ${\rm{B}}^{(0)}$ do not depend on $\lambda$. * • We determine the elements of these two matrices by employing equations (7) written as ${\rm{L}}(\tilde{\mathbf{f}},\tilde{\mathbf{f}}_{10};\lambda){\rm{B}}={\cal{S}}({\rm{B}}){\rm{L}}(\mathbf{f},\mathbf{f}_{10};\lambda),\quad{\rm{M}}(\tilde{\mathbf{f}},\tilde{\mathbf{f}}_{01};\lambda){\rm{B}}={\cal{T}}({\rm{B}}){\rm{M}}(\mathbf{f},\mathbf{f}_{01};\lambda).$ (11) In our calculations we also have to take into account that $\det({\rm{B}})$ is a constant, an obvious consequence of (11) and our assumption that the Lax matrices have constant determinants. * • The derived Darboux matrix will depend in general on the ‘old’ and the ‘new’ fields $\bf{f}$ and $\tilde{\bf{f}}$, as well as the spectral parameter $\lambda$, and a parameter $\varepsilon$. It may also depend on some auxilliary function, a potential, $g(n,m)$. That is, ${\rm{B}}=\lambda{\rm{B}}^{(1)}(\mathbf{f}_{00},\tilde{\mathbf{f}}_{00},g;\varepsilon)+{\rm{B}}^{(0)}(\mathbf{f}_{00},\tilde{\mathbf{f}}_{00},g;\varepsilon).$ * • In the construction of the Darboux matrix (10) there will be some algebraic relations that define the Darboux matrix elements, as well as some difference equations for its elements. These difference equations will be of the form $\mathcal{B}^{(n)}(\mathbf{f}_{00},\mathbf{f}_{10},\tilde{\mathbf{f}}_{00},\tilde{\mathbf{f}}_{10},g;\varepsilon)=0,\qquad\mathcal{B}^{(m)}(\mathbf{f}_{00},\mathbf{f}_{01},\tilde{\mathbf{f}}_{00},\tilde{\mathbf{f}}_{01},g;\varepsilon)=0,$ (12) and constitute the $n$\- and the $m$-part, respectively, of an auto-Bäcklund transformation that relates the ‘old’ and the ‘new’ fields. In what follows, we will denote this transformation simply with ${\mathcal{B}}(\mathbf{f},\tilde{\mathbf{f}},g;\varepsilon)=0$. * • The Bianchi commuting diagram, aka superposition principle, for the auto- Bäcklund transformation (12) follows from the permutation of four Darboux matrices according to the diagram in Figure 2. .6 0.01 t:F (0 0) (1.5 1.5) (3 0) (1.5 -1.5) (0 0) (0 0) f:0.0 r:0.08 (1.5 1.5) f:0.0 r:0.08 (3 0) f:0.0 r:0.08 (1.5 -1.5) f:0.0 r:0.08 (.8 .8) (.85 .85) (2.25 .75) (2.3 .7) (.8 -.8) (.85 -.85) (2.25 -.75) (2.3 -.7) (-0.3 -0.1) $\bf{f}$ (1.4 1.65) $\tilde{\bf{f}}$ (3.15 -.1) $\tilde{\hat{\bf{f}}}=\hat{\tilde{\bf{f}}}$ (1.4 -1.85) $\hat{\bf{f}}$ (-0.75 0.9)${\rm{B}}(\mathbf{f},\tilde{\mathbf{f}},g_{1};\varepsilon_{1},\lambda)$ (-0.78 -1.1)${\rm{B}}(\mathbf{f},\hat{\mathbf{f}},g_{2};\varepsilon_{2},\lambda)$ (2.2 0.9)${\rm{B}}(\tilde{\mathbf{f}},\hat{\tilde{\mathbf{f}}},g_{12};\varepsilon_{2},\lambda)$ (2.3 -1.1)${\rm{B}}(\hat{\mathbf{f}},\tilde{\hat{\mathbf{f}}},g_{21};\varepsilon_{1},\lambda)$ Figure 2: Bianchi commuting diagram. It should be noted that $g_{12}\neq g_{21}$. More precisely, starting with a solution $\bf{f}$ of (6) we can construct two new solutions $\tilde{\bf{f}}$ and $\hat{\bf{f}}$ using the Bäcklund transformations ${\mathcal{B}}(\mathbf{f},\tilde{\mathbf{f}},g_{1};\varepsilon_{1})=0$ and ${\mathcal{B}}(\mathbf{f},\hat{\mathbf{f}},g_{2};\varepsilon_{2})=0$, respectively. Then, we can use the Bäcklund transformation with initial solution $\tilde{\bf{f}}$, a new potential $g_{12}$ and parameter $\varepsilon_{2}$ to derive a new solution $\hat{\tilde{\mathbf{f}}}$, i.e. ${\mathcal{B}}(\tilde{\mathbf{f}},\hat{\tilde{\mathbf{f}}},g_{12};\varepsilon_{2})=0$. In the same fashion, we can start with $\hat{\bf{f}}$, a new potential $g_{21}$ and parameter $\varepsilon_{1}$ to derive solution $\tilde{\hat{\mathbf{f}}}$, i.e. ${\mathcal{B}}(\hat{\mathbf{f}},\tilde{\hat{\mathbf{f}}},g_{21};\varepsilon_{1})=0$. By requiring $\tilde{\hat{\mathbf{f}}}=\hat{\tilde{\bf{f}}}$, according to Figure 2, we can construct this solution algebraically and it follows from the commutativity of the corresponding Darboux matrices, ${\rm{B}}(\tilde{\mathbf{f}},\hat{\tilde{\mathbf{f}}},g_{12};\varepsilon_{2},\lambda){\rm{B}}(\mathbf{f},\tilde{\mathbf{f}},g_{1};\varepsilon_{1},\lambda)={\rm{B}}(\hat{\mathbf{f}},\tilde{\hat{\mathbf{f}}},g_{21};\varepsilon_{1},\lambda){\rm{B}}(\mathbf{f},\hat{\mathbf{f}},g_{2};\varepsilon_{2},\lambda).$ (13) ## 4 The Adler–Yamilov system From the discussion in the previous section it is already obvious that multidimensional consistency is not essential for this method of derivation of Bäcklund transformations. In this section, we demonstrate the application of our scheme using as an illustrative example the Adler–Yamilov system related to the nonlinear Schrödinger equation [19]. The Adler–Yamilov system can be written as $p_{10}-p_{01}-\frac{\alpha-\beta}{1+p_{00}q_{11}}p_{00}=0,\qquad q_{10}-q_{01}+\frac{\alpha-\beta}{1+p_{00}q_{11}}q_{11}=0,$ (14) where $\alpha$, $\beta$ are complex parameters. Moreover, using matrix ${\rm{L}}(f,g;a,\lambda)=\lambda{\rm{L}}^{(1)}+{\rm{L}}^{(2)}(f,g;a)=\lambda\begin{pmatrix}1&0\\\ 0&0\end{pmatrix}+\begin{pmatrix}a+fg&f\\\ g&1\end{pmatrix},$ a Lax pair for (14) can be written as $\displaystyle{\cal{S}}(\Psi)={\rm{L}}(p_{00},q_{10};\alpha,\lambda)\Psi=\left(\lambda\begin{pmatrix}1&0\\\ 0&0\end{pmatrix}+\begin{pmatrix}\alpha+p_{00}q_{10}&p_{00}\\\ q_{10}&1\end{pmatrix}\right)\Psi,$ (15a) $\displaystyle{\cal{T}}(\Psi)={\rm{L}}(p_{00},q_{01};\beta,\lambda)\Psi=\left(\lambda\begin{pmatrix}1&0\\\ 0&0\end{pmatrix}+\begin{pmatrix}\beta+p_{00}q_{01}&p_{00}\\\ q_{01}&1\end{pmatrix}\right)\Psi.$ (15b) ### 4.1 The discrete Darboux–Lax scheme for the Adler–Yamilov system We start with our choice (10) for the initial form of the Darboux matrix $\rm{B}$, ${\rm{B}}=\lambda{\rm{B}}^{(1)}+{\rm{B}}^{(0)}=\lambda\begin{pmatrix}f^{(1)}_{00}&f^{(2)}_{00}\\\ f^{(3)}_{00}&f^{(4)}_{00}\end{pmatrix}+\begin{pmatrix}g^{(1)}_{00}&g^{(2)}_{00}\\\ g^{(3)}_{00}&g^{(4)}_{00}\end{pmatrix},$ (16) and the determining equations (11), which now become $\displaystyle\left(\lambda{\rm{L}}^{(1)}+{\rm{L}}^{(2)}(\tilde{p}_{00},\tilde{q}_{10};\alpha)\right)\left(\lambda{\rm{B}}^{(1)}+{\rm{B}}^{(0)}\right)=\left(\lambda{\cal{S}}\left({\rm{B}}^{(1)}\right)+{\cal{S}}\left({\rm{B}}^{(0)}\right)\right)\left(\lambda{\rm{L}}^{(1)}+{\rm{L}}^{(2)}(p_{00},q_{10};\alpha)\right),$ (17a) $\displaystyle\left(\lambda{\rm{L}}^{(1)}+{\rm{L}}^{(2)}(\tilde{p}_{00},\tilde{q}_{01};\beta)\right)\left(\lambda{\rm{B}}^{(1)}+{\rm{B}}^{(0)}\right)=\left(\lambda{\cal{T}}\left({\rm{B}}^{(1)}\right)+{\cal{T}}\left({\rm{B}}^{(0)}\right)\right)\left(\lambda{\rm{L}}^{(1)}+{\rm{L}}^{(2)}(p_{00},q_{01};\beta)\right).$ (17b) Since all the matrices involved in (17) are independent of $\lambda$, we collect the coefficients of the different powers of the spectral parameter. The $\lambda^{2}$ terms yield equations ${\rm{L}}^{(1)}{\rm{B}}^{(1)}={\cal{S}}\left({\rm{B}}^{(1)}\right){\rm{L}}^{(1)},\quad{\rm{L}}^{(1)}{\rm{B}}^{(1)}={\cal{T}}\left({\rm{B}}^{(1)}\right){\rm{L}}^{(1)},$ which lead to $f^{(1)}_{00}=c_{1}\in{\mathbb{R}},\quad f^{(2)}_{00}=f^{(3)}_{00}=0.$ (18) The $\lambda$ terms in relations (17) are ${\rm{L}}^{(1)}{\rm{B}}^{(0)}+{\rm{L}}^{(2)}(\tilde{p}_{00},\tilde{q}_{10};\alpha){\rm{B}}^{(1)}={\cal{S}}\left({\rm{B}}^{(1)}\right){\rm{L}}^{(2)}(p_{00},q_{10};\alpha)+{\cal{S}}\left({\rm{B}}^{(0)}\right){\rm{L}}^{(1)},$ ${\rm{L}}^{(1)}{\rm{B}}^{(0)}+{\rm{L}}^{(2)}(\tilde{p}_{00},\tilde{q}_{01};\beta){\rm{B}}^{(1)}={\cal{T}}\left({\rm{B}}^{(1)}\right){\rm{L}}^{(2)}(p_{00},q_{01};\beta)+{\cal{T}}\left({\rm{B}}^{(0)}\right){\rm{L}}^{(1)},$ which in view of (18) imply $f^{(4)}_{00}=c_{2}\in{\mathbb{R}},\quad g^{(2)}_{00}=c_{1}p_{00}-c_{2}\tilde{p}_{00},\quad g^{(3)}_{00}=c_{1}\tilde{q}_{00}-c_{2}q_{00},$ (19) $\left({\cal{S}}-1\right)\left(g^{(1)}_{00}\right)=c_{1}\left(\tilde{p}_{00}\tilde{q}_{10}-p_{00}q_{10}\right),\quad\left({\cal{T}}-1\right)\left(g^{(1)}_{00}\right)=c_{1}\left(\tilde{p}_{00}\tilde{q}_{01}-p_{00}q_{01}\right).$ (20) The $\lambda$ independent terms, ${\rm{L}}^{(2)}(\tilde{p}_{00},\tilde{q}_{10};\alpha){\rm{B}}^{(0)}={\cal{S}}\left({\rm{B}}^{(0)}\right){\rm{L}}^{(2)}(p_{00},q_{10};\alpha),\quad{\rm{L}}^{(2)}(\tilde{p}_{00},\tilde{q}_{01};\beta){\rm{B}}^{(0)}={\cal{T}}\left({\rm{B}}^{(0)}\right){\rm{L}}^{(2)}(p_{00},q_{01};\beta),$ (21) determine $g^{(4)}_{00}$ and provide us with the corresponding auto-Bäcklund transformation. Specifically, the $(2,2)$-elements of the above relations yield $\left({\cal{S}}-1\right)\left(g^{(4)}_{00}\right)=c_{2}\left(p_{00}q_{10}-\tilde{p}_{00}\tilde{q}_{10}\right),\quad\left({\cal{T}}-1\right)\left(g^{(4)}_{00}\right)=c_{2}\left(p_{00}q_{01}-\tilde{p}_{00}\tilde{q}_{01}\right).$ (22) The remaining entries of (21) in view of (20) and (22) become $\displaystyle c_{1}\left(p_{10}+p_{00}(\alpha+p_{00}q_{10})\right)-c_{2}\left(\tilde{p}_{10}+\tilde{p}_{00}(\alpha+\tilde{p}_{00}\tilde{q}_{10})\right)=g^{(4)}_{00}\tilde{p}_{00}-g^{(1)}_{00}p_{00},$ (23a) $\displaystyle c_{1}\left(p_{01}+p_{00}(\beta+p_{00}q_{01})\right)-c_{2}\left(\tilde{p}_{01}+\tilde{p}_{00}(\beta+\tilde{p}_{00}\tilde{q}_{01})\right)=g^{(4)}_{00}\tilde{p}_{00}-g^{(1)}_{00}p_{00},$ (23b) $\displaystyle\tilde{q}_{10}=\frac{c_{1}\tilde{q}_{00}-c_{2}(q_{00}-\alpha q_{10})-g^{(4)}_{00}q_{10}}{c_{1}(\alpha+p_{00}q_{10})-c_{2}\tilde{p}_{00}q_{10}-g^{(1)}_{00}},\quad\tilde{q}_{01}=\frac{c_{1}\tilde{q}_{00}-c_{2}(q_{00}-\beta q_{01})-g^{(4)}_{00}q_{01}}{c_{1}(\beta+p_{00}q_{01})-c_{2}\tilde{p}_{00}q_{01}-g^{(1)}_{00}},$ (23c) which play the role of the Bäcklund transformation. Finally we require the determinant of the Darboux matrix $\rm{B}$, which in view of (18) and (19) can be written as $\det\left({\rm{B}}\right)=c_{1}c_{2}\lambda^{2}+\left(c_{2}g^{(1)}_{00}+c_{1}g^{(4)}_{00}\right)\lambda+g^{(1)}_{00}g^{(4)}_{00}-\left(c_{1}p_{00}-c_{2}\tilde{p}_{00}\right)\left(c_{1}\tilde{q}_{00}-c_{2}q_{00}\right)$, to be constant. This requirement implies the relations $c_{2}g^{(1)}_{00}+c_{1}g^{(4)}_{00}=\kappa,\quad g^{(1)}_{00}g^{(4)}_{00}-\left(c_{1}p_{00}-c_{2}\tilde{p}_{00}\right)\left(c_{1}\tilde{q}_{00}-c_{2}q_{00}\right)=\varepsilon,\quad\kappa,\varepsilon\in{\mathbb{R}}.$ (24) It should be noted that the first relation in (24) may also be viewed as a consequence of (20) and (22). Summarizing, so far we have shown that the Darboux matrix $\rm{B}$ has the form ${\rm{B}}=\lambda{\rm{B}}^{(1)}+{\rm{B}}^{(0)}=\lambda\begin{pmatrix}c_{1}&0\\\ 0&c_{2}\end{pmatrix}+\begin{pmatrix}g^{(1)}_{00}&c_{1}p_{00}-c_{2}\tilde{p}_{00}\\\ c_{1}\tilde{q}_{00}-c_{2}q_{00}&g^{(4)}_{00}\end{pmatrix},$ (25) where potentials $g^{(1)}$ and $g^{(4)}$ are determined by (20), (22) and (24), and the Bäcklund transformation is given by (23). We consider now two cases: (i) $c_{1}=0$ and $c_{2}\neq 0$, and (ii) $c_{1}\neq 0$ and $c_{2}=0$. #### First case: $c_{1}=0$ and $c_{2}\neq 0$ If $c_{1}=0$ and $c_{2}\neq 0$, then we can choose $c_{2}=1$ without loss of generality. Then equations (20) imply that $g^{(1)}_{00}$ is a constant and we choose $g^{(1)}_{00}=1$.222It means we choose $\kappa=1$ in the first relation of (24). Moreover, the second relation in (24) implies $g^{(4)}_{00}=\tilde{p}_{00}q_{00}+\varepsilon,\quad\varepsilon\in{\mathbb{R}}.$ (26) With these choices, the Darboux matrix becomes ${\rm{B}}(q_{00},\tilde{p}_{00};\varepsilon)=\lambda\begin{pmatrix}0&0\\\ 0&1\end{pmatrix}+\begin{pmatrix}1&-\tilde{p}_{00}\\\ -q_{00}&\varepsilon+\tilde{p}_{00}q_{00}\end{pmatrix},$ (27) and relations (23) become $\displaystyle\tilde{p}_{10}=p_{00}+\frac{\alpha-\varepsilon}{1+\tilde{p}_{00}q_{10}}\tilde{p}_{00},\quad\tilde{q}_{10}=q_{00}-\frac{\alpha-\varepsilon}{1+\tilde{p}_{00}q_{10}}q_{10},$ (28a) $\displaystyle\tilde{p}_{01}=p_{00}+\frac{\beta-\varepsilon}{1+\tilde{p}_{00}q_{01}}\tilde{p}_{00},\quad\tilde{q}_{01}=q_{00}-\frac{\beta-\varepsilon}{1+\tilde{p}_{00}q_{01}}q_{01}.$ (28b) In view of the above choices and system (28), relations (22) hold identically. System (28) is an auto-Bäcklund transformation for the Adler–Yamilov system (14). Indeed, if we shift equations (28a) in the $m$ direction, and equations (28b) in the $n$ direction, respectively, then it can be readily verified that the two expressions for $\tilde{p}_{11}$ and the two expressions for $\tilde{q}_{11}$ coincide modulo the Adler–Yamilov system (14). Conversely, we rearrange the above system for $p_{00}$, $q_{10}$ and $q_{01}$, $\displaystyle q_{10}=\frac{q_{00}-\tilde{q}_{10}}{\alpha-\varepsilon-\tilde{p}_{00}(q_{00}-\tilde{q}_{10})},\quad p_{00}=\tilde{p}_{10}-\tilde{p}_{00}\left(\alpha-\varepsilon-\tilde{p}_{00}(q_{00}-\tilde{q}_{10})\right),$ (29a) $\displaystyle q_{01}=\frac{q_{00}-\tilde{q}_{01}}{\beta-\varepsilon-\tilde{p}_{00}(q_{00}-\tilde{q}_{01})},\quad p_{00}=\tilde{p}_{01}-\tilde{p}_{00}\left(\beta-\varepsilon-\tilde{p}_{00}(q_{00}-\tilde{q}_{01})\right).$ (29b) If we shift the first equation in (29a) in the $m$ direction and the first equation in (29b) in the $n$ direction, then the resulting expressions for $q_{11}$ coincide provided that $\tilde{p}$ and $\tilde{q}$ satisfy system (14). Moreover, subtracting the two expressions for $p_{00}$ in (29) we end up with $\tilde{p}_{10}-\tilde{p}_{01}-\tilde{p}_{00}\left(\tilde{p}_{00}(\tilde{q}_{10}-\tilde{q}_{01})+\alpha-\beta\right)=0,$ which holds on solutions of (14). Finally, according to Figure 2 and relation (13), the superposition principle for the auto-Bäcklund transformation (28) follows from ${\rm{B}}(\tilde{q}_{00},\hat{\tilde{p}}_{00};\varepsilon_{2}){\rm{B}}(q_{00},\tilde{p}_{00};\varepsilon_{1})={\rm{B}}(\hat{q}_{00},\tilde{\hat{p}}_{00};\varepsilon_{1}){\rm{B}}(q_{00},\hat{p}_{00};\varepsilon_{2}),$ and can be written as $\hat{\tilde{p}}_{00}=-\,\frac{\tilde{p}_{00}-\hat{p}_{00}}{\varepsilon_{1}-\varepsilon_{2}+(\tilde{p}_{00}-\hat{p}_{00})q_{00}},\quad\tilde{q}_{00}-\hat{q}_{00}=\left(\varepsilon_{1}-\varepsilon_{2}+(\tilde{p}_{00}-\hat{p}_{00})q_{00}\right)q_{00}.$ (30) #### Second case: $c_{1}\neq 0$ and $c_{2}=0$ If $c_{2}=0$ and $c_{1}\neq 0$, we choose $c_{1}=1$, relations (22) imply that $g^{(4)}_{00}=1$, and the determinant of $\rm{B}$ implies that $g^{(1)}_{00}=p_{00}\tilde{q}_{00}+\varepsilon$. In view of these choices, we arrive at a Darboux matrix which can be written in terms of matrix (27) as ${\rm{C}}(p_{00},\tilde{q}_{00};\varepsilon)=(\lambda+\varepsilon)\left({\rm{B}}(\tilde{q}_{00},p_{00};\varepsilon)\right)^{-1},$ (31) whereas relations (23) yield actually system (29) with the roles of new and old fields interchanged, i.e. system (29) accompanied by the interchange $(p,q)\leftrightarrow(\tilde{p},\tilde{q})$. Thus in this case we end up with the inverse of the Darboux and Bäcklund transformations we derived previously. ### 4.2 Derivation of soliton solutions We employ the auto-Bäcklund transformation (28) and its superposition principle (30) in the derivation of soliton solutions of the Adler–Yamilov system (14). Figure 3: The one soliton solution of the Adler–Yamilov system and the potential $1/g_{00}$. In both cases $\alpha=8$, $\beta=4$, $\varepsilon=1$ and $c=-2$. More precisely, we start with the solution333This solution can be constructed starting with the zero solution $p_{00}=q_{00}=0$ and using the second transformation we discussed in subsection 4.1 with $\varepsilon=0$. $p_{00}=0,\quad q_{00}=\alpha^{-n}\beta^{-m}.$ (32) With this seed solution, the first equation in (28a) and the first one in (28b) become $\tilde{p}_{10}=\frac{\alpha-\varepsilon}{1+\tilde{p}_{00}\alpha^{-n-1}\beta^{-m}}\tilde{p}_{00},\quad\tilde{p}_{01}=\frac{\beta-\varepsilon}{1+\tilde{p}_{00}\alpha^{-n}\beta^{-m-1}}\tilde{p}_{00}.$ (33) We can linearise these Ricatti equations by setting $\tilde{p}_{00}=\alpha^{n}\beta^{m}/g_{00}$, $(\alpha-\varepsilon)g_{10}=\alpha g_{00}+1,\quad(\beta-\varepsilon)g_{01}=\beta g_{00}+1.$ The general solution of this linear system is $g_{00}=\frac{\alpha^{n}}{(\alpha-\varepsilon)^{n}}\frac{\beta^{m}}{(\beta-\varepsilon)^{m}}c-\frac{1}{\varepsilon},$ (34) where $c\in{\mathbb{R}}$ is the arbitrary constant of integration, and thus $\tilde{p}_{00}=\frac{\varepsilon\,\alpha^{n}\beta^{m}(\alpha-\varepsilon)^{n}(\beta-\varepsilon)^{m}}{c\,\varepsilon\,\alpha^{n}\beta^{m}-(\alpha-\varepsilon)^{n}(\beta-\varepsilon)^{m}}.$ (35a) Using the seed solution (32) and the updated potential (35a), we can use either the equation for $\tilde{q}_{10}$ in (28a) or the equation for $\tilde{q}_{01}$ in (28b) to determine $\tilde{q}_{00}$. Both ways lead to $\tilde{q}_{00}=\frac{c\,\varepsilon^{2}}{c\,\varepsilon\,\alpha^{n}\beta^{m}-(\alpha-\varepsilon)^{n}(\beta-\varepsilon)^{m}}.$ (35b) This two-parameter family of solutions yields the one-soliton solution of (14): even though both functions in (35) diverge, their product represents a soliton. This interpretation is motivated by the relation of the Adler–Yamilov system to the nonlinear Schrödinger equation [19], and it is evident from the plot of the product $|\tilde{p}_{00}\tilde{q}_{00}|$. We also plot the potential $1/g_{00}$ which is a kink. See Figure 3. Figure 4: The two-soliton solution of the Adler–Yamilov system: solution (37) with $\alpha=8$, $\beta=4$, $\varepsilon_{1}=1$, $c_{1}=-2$, $\varepsilon_{2}=3$ and $c_{2}=8$. It should be noted that this solution requires the combination of a non-singular solution, i.e. the pair $(\tilde{p},\tilde{q})$ which corresponds to the one-soliton solution (35), and a singular one which is $(\hat{p},\hat{q})$. Having constructed the two-parameter family of solutions (35), we may use it along with the superposition principle (30) to determine a third solution and in particular the two-soliton solution of system (14). Starting with the same seed solution, the two solutions $(\tilde{p}_{00},\tilde{q}_{00})$ and $(\hat{p}_{00},\hat{q}_{00})$ involved in (30) follow from (35) by replacing parameters $(\varepsilon,c)$ with $(\varepsilon_{1},c_{1})$ and $(\varepsilon_{2},c_{2})$, respectively. In order to make the presentation more comprehensible, let us introduce the shorthand notation $\delta_{0}:=\alpha^{n}\beta^{m},\quad\delta_{1}:=(\alpha-\varepsilon_{1})^{n}(\beta-\varepsilon_{1})^{m},\quad\delta_{2}:=(\alpha-\varepsilon_{2})^{n}(\beta-\varepsilon_{2})^{m}.$ In terms of this notation, the seed solution is $p_{00}=0$, $q_{00}=1/\delta_{0}$, and the two solutions we described can be written as $\tilde{p}_{00}=\frac{\varepsilon_{1}\,\delta_{0}\delta_{1}}{c_{1}\,\varepsilon_{1}\,\delta_{0}-\delta_{1}},\leavevmode\nobreak\ \leavevmode\nobreak\ \tilde{q}_{00}=\frac{c_{1}\,\varepsilon_{1}^{2}}{c_{1}\,\varepsilon_{1}\,\delta_{0}-\delta_{1}},\quad\mbox{and}\quad\hat{p}_{00}=\frac{\varepsilon_{2}\,\delta_{0}\delta_{2}}{c_{2}\,\varepsilon_{2}\,\delta_{0}-\delta_{2}},\leavevmode\nobreak\ \leavevmode\nobreak\ \hat{q}_{00}=\frac{c_{2}\,\varepsilon_{2}^{2}}{c_{2}\,\varepsilon_{2}\,\delta_{0}-\delta_{2}},$ (36) respectively. With these formulae at our disposal, the first relation in (30) yields $\hat{\tilde{p}}_{00}=\frac{\varepsilon_{1}\varepsilon_{2}\delta_{0}\left(c_{1}\delta_{2}-c_{2}\delta_{1}\right)+(\varepsilon_{1}-\varepsilon_{2})\delta_{1}\delta_{2}}{c_{1}c_{2}\varepsilon_{1}\varepsilon_{2}(\varepsilon_{1}-\varepsilon_{2})\delta_{0}-c_{1}\varepsilon_{1}^{2}\delta_{2}+c_{2}\varepsilon_{2}^{2}\delta_{1}}.$ (37a) To find $\hat{\tilde{q}}_{00}$ we work in the same way we derived (35b) and according to the Bianchi diagram. More precisely, we can use the second equation either in (28a) or in (28b) with $\tilde{q}$ replaced by $\hat{\tilde{q}}$, $(p,q)$ replaced by $(\tilde{p},\tilde{q})$ given in (36), and parameter $\varepsilon$ replaced by $\varepsilon_{2}$ (alternatively, we can replace $(p,q)$ with $(\hat{p},\hat{q})$ given in (36), and parameter $\varepsilon$ with $\varepsilon_{1}$), to find $\hat{\tilde{q}}_{00}=\frac{c_{1}c_{2}\varepsilon_{1}^{2}\varepsilon_{2}^{2}(\varepsilon_{1}-\varepsilon_{2})}{c_{1}c_{2}\varepsilon_{1}\varepsilon_{2}(\varepsilon_{1}-\varepsilon_{2})\delta_{0}-c_{1}\varepsilon_{1}^{2}\delta_{2}+c_{2}\varepsilon_{2}^{2}\delta_{1}}.$ (37b) This four-parameter family of solutions yields the two-soliton solution of (14) in the same way we interpreted (35) as the one-soliton solution of the Adler–Yamilov system. See the plots of the product $|\hat{\tilde{p}}_{00}\hat{\tilde{q}}_{00}|$ in Figure 4. ## 5 Conclusions In this paper we proposed a new method for deriving Bäcklund transformations and constructing solutions for nonlinear integrable P$\Delta$Es which admit Lax representation but do not necessarily possess the 3D consistency property. Specifically, in our approach we consider Darboux transformations which leave the given Lax pair covariant, and by construction lead to Bäcklund transformations for the corresponding discrete system. The permutability of four Darboux matrices according to the Bianchi diagram in Figure 2 leads to the nonlinear superposition principle of the related Bäcklund transformation. Moreover, the latter transformation and its superposition principle can be used in the construction of interesting solutions to P$\Delta$Es starting from some simple ones. As an illustrative example we used the Adler–Yamilov system (14). For this system we constructed Darboux and corresponding Bäcklund transformations. With the use of transformation (28) and its superposition principle (30) we constructed the one- and two-soliton solutions starting with the seed solution $p_{00}=0,q_{00}=\alpha^{-n}\beta^{-m}$. In the illustrative example we considered in Section 4, the Lax representation (6) involves matrices $\rm{L}$ and $\rm{M}$ which have the same form, see (15). The natural question arises as to whether our method can be employed to the case of integrable P$\Delta$Es with Lax representation (6) where matrices ${\rm{L}}$ and $\rm{M}$ do not have the same form. The answer to this question is positive, the corresponding transformations may involve auxiliary functions (potentials), and this derivation is similar to the generic construction we presented in subsection 4.1, see Darboux matrix (25) and relations (20), (22), (23)and (24). Moreover our considerations can be extended to the generalized symmetries of the discrete system. Generalized symmetries are (integrable) differential- difference equations involving shifts in one lattice direction and are compatible with the P$\Delta$E. Their Lax pair is semi-discrete and its discrete part coincides with the one of the two equations of the fully discrete Lax pair (3), see for instance [12]. This relation allows us to extend the Darboux and Bäcklund transformations for the P$\Delta$E to corresponding ones for the differential-difference equations and employ the Bäcklund transformation and its superposition principle in the construction of solutions for the symmetries. We will demonstrate this method and its extensions in our future work in details using the Hirota KdV equation as an illustrative example, as well as systems appeared in [4, 10, 22]. In fact our results can be used and extended in various ways. 1. 1. Apply our method to construct solutions to all the NLS type equations derived in [19]. In [19] we classified Darboux transformations related to NLS type equations and constructed integrable discretisations of the latter, namely integrable systems of nonlinear P$\Delta$Es. By employing the discrete Darboux–Lax scheme we proposed in Section 3, one could derive Bäcklund transformations and construct soliton solutions to these nonlinear P$\Delta$Es. 2. 2. Study the solutions of the associated PDEs. The Adler–Yamilov system (14) constitutes a discretisation of the NLS equation via its Darboux transformation [19]. In this paper, we constructed soliton solutions to this system, so one could consider the continuum limits of these solutions to construct solutions to the NLS equation. This procedure could be applied to other NLS type equations which appeared in [19]. 3. 3. Study the corresponding Yang–Baxter maps. In [18, 20] matrix refactorisation problems of Darboux matrices for integrable PDEs were considered in order to derive solutions to the Yang–Baxter equation and the entwining Yang–Baxter equation. Since the generator of Yang–Baxter maps is a matrix refactorisation problem (13) it makes sense to understand how Bäcklund transformations for P$\Delta$Es are related to Yang–Baxter maps. In our future work, we plan to show that Yang–Baxter and entwining Yang–Baxter maps are superpositions of Bäcklund transformations of P$\Delta$Es. 4. 4. Extend the results to the case of discrete systems on a 3D lattice. One can extend the results employed in this paper to the case of 3D lattice integrable systems. It is expected that the superposition of Bäcklund transformations related to these systems are solutions to the tetrahedron equation. 5. 5. Extend the results to the case of Grassmann algebras. Grassmann extensions of Darboux transformations were employed in the construction of noncommutative versions of discrete systems, see for instance [14, 30] . However it was realised in [17] that quad-graph systems may lose their 3D consistency property in the Grassmann extension. The method we presented here could be generalised and employed in the construction of Bäcklund transformations and the derivation of solutions to Grassmann extended quad-graph systems which appeared in the literature. ## 6 Acknowledgements Xenia Fisenko’s work was supported by the Ministry of Research and Higher Education (Regional Mathematical Centre “Centre of Integrable Systems,” Agreement No. 075-02-2021-1397). Sotiris Konstantinou-Rizos’s work was funded by the Russian Science Foundation (grant No. 20-71-10110). ## References * [1] V. Adler, A. Bobenko, and Y. Suris. 2005 Geometry of Yang-Baxter maps: pencils of conics and quadrirational mappings Comm. Anal. Geom. 12 967–1007. * [2] J. Atkinson, J. Hietarinta and F. Nijhoff. 2007 Seed and soliton solutions for Adler’s lattice equation J. Phys. A: Math. Theor. 40 F1–F8. * [3] J. Atkinson and M. Nieszporski. 2014 Multi-Quadratic Quad Equations: Integrable Cases from a Factorized-Discriminant Hypothesis Int. Math. Res. Not. 2014 4215–4240. * [4] G. Berkeley, A.V. Mikhailov, and P. Xenitidis. 2016 Darboux transformations with tetrahedral reduction group and related integrable systems J. Math. Phys. 57, 092701. * [5] A. Bobenko and Y. Suris. 2002 Integrable systems on quad-graphs Int. Math. Res. Notices 11 573–611. * [6] A. Bobenko and Y. Suris. 2008 _Discrete Differential Geometry. Integrable Structure._ Graduate Studies in Mathematics 98 American Mathematical Society. * [7] R. Boll. 2011 Classification of 3D Consistent Quad-Equations J. Nonlinear Math. Phys. 18 337–365 * [8] T. Bridgman, W. Hereman, G.R.W. Quispel and P. van der Kamp. 2013 Symbolic Computation of Lax Pairs of Partial Difference Equations using Consistency Around the Cube Found. Comput. Math. 13517–544. * [9] V. Caudrelier, N. Crampé, C. Zhang. 2014 Integrable Boundary for Quad-Graph Systems: Three-Dimensional Boundary Consistency SIGMA 014 (24 pp). * [10] A. Fordy and P. Xenitidis. 2017 $\mathbb{Z}_{N}$ graded discrete Lax pairs and integrable difference equations. J. Phys. A: Math. Theor. 50 165205\. * [11] A. Fordy and P. Xenitidis. 2017 $\mathbb{Z}_{N}$ graded discrete Lax pairs and Yang–Baxter maps. Proc. R. Soc. A. 473 20160946\. * [12] A. Fordy and P. Xenitidis. 2020 Symmetries of ${\mathbb{Z}}_{N}$ graded discrete integrable systems J Phys A: Math Theor 53 235201 (30pp). * [13] W. Fu. 2020 Direct linearization approach to discrete integrable systems associated with $Z_{N}$ graded Lax pairs. Proc. R. Soc. A. 476. * [14] G.G. Grahovski, A.V. Mikhailov. 2013 Integrable discretisations for a class of nonlinear Scrödinger equations on Grassmann algebras, Phys. Lett. A 377 3254–3259. * [15] J. Hietarinta, N. Joshi, F. Nijhoff. 2016 _Discrete Systems and Integrability_ Cambridge texts in applied mathematics, Cambridge University Press. * [16] P. Kassotakis, M. Nieszporski, V. Papageorgiou, A. Tongas 2019 Tetrahedron maps and symmetries of three dimensional integrable discrete equations, J. Math. Phys. 60 123503\. * [17] S. Konstantinou-Rizos. 2020 On the 3D consistency of a Grassmann extended lattice Boussinesq system, J. Nuc. Phys. B 951 114878\. * [18] S. Konstantinou-Rizos and A.V. Mikhailov. 2013 Darboux transformations, finite reduction groups and related Yang–Baxter maps. J. Phys. A: Math. Theor. 46 425201\. * [19] S. Konstantinou-Rizos, A.V. Mikhailov and P. Xenitidis. 2015 Reduction groups and related integrable difference systems of nonlinear Schrödinger type J. Math. Phys. 56 082701\. * [20] S. Konstantinou-Rizos, G. Papamikos. 2019 Entwining Yang–Baxter maps related to NLS type equations, J. Phys. A: Math. Theor. 52 485201\. * [21] S.B. Lobb, F.W. Nijhoff. 2009 Lagrangian multiforms and multidimensional consistency, J. Phys. A: Math. Theor. 42 454013\. * [22] A.V. Mikhailov and P. Xenitidis. 2014 Second Order Integrability Conditions for Difference Equations: An Integrable Equation. Lett. Math. Phys. 104 431–450. * [23] F. Nijhoff. 2002 Lax pair for the Adler (lattice Krichever-Novikov) system Phys. Lett. A 297 49–58. * [24] F. Nijhoff, J. Atkinson and J. Hietarinta. 2010 A Constructive Approach to the Soliton Solutions of Integrable Quadrilateral Lattice Equations Commun. Math. Phys. 297 283–304. * [25] V. Papageorgiou, A. Tongas. 2007 Yang–Baxter maps and multi-field integrable lattice equations J. Phys. A: Math. Theor. 40 12677\. * [26] Y. Shi 2020 $Z_{N}$ graded discrete integrable systems and Darboux transformations. J. Phys. A: Math. Theor. 53 044003\. * [27] P. Xenitidis. 2018 Determining the symmetries of difference equations. Proc. R. Soc. A. 474 20180340\. * [28] P. Xenitidis. 2019 On consistent systems of difference equations. J. Phys. A: Math. Theor. 52 455201. * [29] P. Xenitidis, F. Nijhoff and S. Lobb. 2011 On the Lagrangian formulation of multidimensionally consistent systems. Proc. R. Soc. A. 467 3295–3317. * [30] L.L. Xue, D. Levi and Q.P. Liu. 2013 Supersymmetric KdV equation: Darboux transformation and discrete systems J. Phys. A: Math. Theor. (FTC) 46 502001\.
# comp-syn: Perceptually Grounded Word Embeddings with Color Bhargav Srinivasa Desikan University of Chicago Knowledge Lab <EMAIL_ADDRESS> Tasker Hull Psiphon Inc, Toronto <EMAIL_ADDRESS> Ethan O. Nadler Stanford University KIPAC & Department of Physics <EMAIL_ADDRESS> Douglas Guilbeault University of California, Berkeley Haas Business School <EMAIL_ADDRESS> Aabir Abubaker Kar University of Chicago Knowledge Lab <EMAIL_ADDRESS> Mark Chu Columbia University <EMAIL_ADDRESS> Donald Ruggiero Lo Sardo Sony CSL Paris <EMAIL_ADDRESS> ###### Abstract Popular approaches to natural language processing create word embeddings based on textual co-occurrence patterns, but often ignore embodied, sensory aspects of language. Here, we introduce the Python package comp-syn, which provides grounded word embeddings based on the perceptually uniform color distributions of Google Image search results. We demonstrate that comp-syn significantly enriches models of distributional semantics. In particular, we show that (1) comp-syn predicts human judgments of word concreteness with greater accuracy and in a more interpretable fashion than word2vec using low-dimensional word–color embeddings, and (2) comp-syn performs comparably to word2vec on a metaphorical vs. literal word-pair classification task. comp-syn is open- source on PyPi and is compatible with mainstream machine-learning Python packages. Our package release includes word–color embeddings for over 40,000 English words, each associated with crowd-sourced word concreteness judgments. ## 1 Introduction The embodied cognition paradigm seeks to ground semantic processing in bodily, affective, and social experiences. This paradigm explains how sensory information contributes to semantic processing, either through metaphors involving references to sensory experience [Lakoff and Turner, 1989, Gallese and Lakoff, 2005] or through the simulation of sensory experience in mental imagery [Bergen, 2012]. For example, color is pervasive in linguistic metaphors (e.g., “her bank accounts are in the _red_ ”), and primes a range of affective and interpretative responses [Mehta and Zhu, 2009, Elliot and Maier, 2014]. Extant methods in computational linguistics are limited in their ability to advance the embodied cognition paradigm due to their focus on text, which often precludes multi-modal analyses of images and other forms of sensory data—including color—involved in human meaning-making activities. Distributional semantics is a particularly prominent approach, wherein textual co-occurrence patterns are used to embed words in a high-dimensional space; the resulting “distance” between word embeddings correlates with semantic similarity [Landauer and Dumais, 1997]. Although leveraging neural networks [Mikolov et al., 2013] and syntactic information [Levy and Goldberg, 2014] has led to recent progress, popular models like word2vec lack firm grounding in their use of sensory information. Moreover, high-dimensional word embeddings created by neural networks are difficult to interpret [Şenel et al., 2018] and require long training periods [Ji et al., 2019]. Thus, it remains difficult to reconcile models of distributional semantics with theories of embodied semantics. To address these limitations, _multi-modal_ approaches to distributional semantics combine text and image data. However, these models continue to infer semantic associations using high-dimensional word-plus-image embeddings; their interpretability therefore remains an issue [Bruni et al., 2014, Socher et al., 2014, Lu et al., 2019]. With respect to color, multi-modal models operate in standard colorspaces like RGB that are not perceptually uniform [International Commission On Illumination, 1978] and embed color in a manner that combines it with complex, multidimensional spatial information. Here, we introduce a novel word embedding method based on color that is explicitly interpretable with respect to theories of embodied cognition. We build on work which shows that color distributions in online images reflect both affective and semantic similarities among words in abstract domains (e.g., in the domain of academic disciplines), while also characterizing human judgments of concept concreteness [Brysbaert et al., 2014, Guilbeault et al., 2020]. We present the Python package comp-syn111Short for “Computational Synaesthesia.”, which allows users to explore word–color embeddings based on the perceptually uniform color distributions of Google Image search results. We provide embeddings for a set of 40,000 common English words, and we benchmark the performance of our model using crowd-sourced human concreteness ratings [Brysbaert et al., 2014]. We show that comp-syn complements the performance of text-based distributional semantics models by providing an interpretable embedding that both (1) predicts human judgments of concept concreteness, and (2) distinguishes metaphorical and literal word pairs. ## 2 Python Package: comp-syn comp-syn is an open-source Python package available on GitHub and downloadable through PyPi. The code follows Python best practices and uses industry standard packages for scientific computing, facilitating easy integration with Python code bases; we provide complete details in the Supplementary Material. comp-syn creates word–color embeddings by computing the perceptually uniform $J_{z}A_{z}B_{z}$ [Safdar et al., 2017] color distributions of their corresponding top $100$ Google Image search results. We represent these distributions by their mean and (optionally) standard deviation in $8$ evenly- segmented $J_{z}A_{z}B_{z}$ bins, yielding 8 to 16-dimensional word–color embeddings. The details of our code, image searches and methods for generating word–color embeddings are described in the Appendices. We use Google rather than a curated image database such as ImageNet [Deng et al., 2009] because popular datasets are heavily biased toward concrete objects, which limits their applicability in abstract semantic domains. Moreover, Google Image search results reflect content that users interact with most frequently [Jing and Baluja, 2008], underscoring their relevance for connecting distributional properties of words and images to human semantic processing. To ensure that our approach is entirely unsupervised, we do not select particular features of images when measuring their color distributions. This approach avoids importing pre-determined semantic notions into our analysis and connects our method to cognitive theories that attribute aesthetic relevance to recognizable features in both the background and foreground of images [Riley, 1995, Elliot and Maier, 2014, Guilbeault et al., 2020]. Model | Embedding based on | Dimension | Concreteness prediction | Metaphor task ---|---|---|---|--- comp-syn | Perceptually uniform color distributions | $8$–$16$ | Linear: $R^{2}=0.96$ Nonlinear: $\ln\mathcal{L}=86$ | $92\%$ test set accuracy word2vec | Textual co-occurrence | $\sim 300$ | Linear: $R^{2}=0.17$ Nonlinear: $\ln\mathcal{L}=76$ | $95\%$ test set accuracy Table 1: Comparison of comp-syn and word2vec and summary of our main results. ## 3 Results Table 1 summarizes our main results, which we now describe in turn. Figure 1: The semantic coherence of comp-syn. (A) Word pair color similarity in comp-syn, measured using the Jensen-Shannon divergence between their perceptually uniform color distributions, correlates with word2vec cosine similarity ($N>10^{6}$ pairs). This holds for both concrete (blue) and abstract (magenta) word pairs, rated according to crowd-sourced human judgments. Data represent average color similarity in $40$ equally-sized bins of word2vec similarity. (B) Word pair similarity vs. human concreteness judgments ($N>10^{7}$ pairs). comp-syn similarity monotonically decreases with word pair concreteness, while word2vec similarity is a nonlinear function of word pair concreteness. Data represent average word pair similarity in word2vec (gray) and comp-syn (black) in $40$ equally-sized bins of summed word pair concreteness. ### 3.1 Comparison to word2vec Similarity We begin by demonstrating that pairwise word similarity is highly correlated in comp-syn and word2vec, illustrating the semantic coherence of our color embeddings. We calculate pairwise distances between each of the 40,000 words in our dataset and 500 randomly-selected words from the same set, yielding over $10^{7}$ distinct pairs. We compare the cosine similarity between word pairs in the Mikolov et al. [Mikolov et al., 2013] word2vec model with the Jensen-Shannon (JS) divergence between $J_{z}A_{z}B_{z}$ distributions in comp-syn. Fig. 1A shows that JS divergence in comp-syn is significantly correlated with cosine similarity in word2vec ($p<0.00001$, $\mathrm{JT}=773$, Jonckheere-Terpstra test). This implies that our low-dimensional, interpretable embedding captures aspects of the key information contained in a widely-used distributional semantics model. ### 3.2 Relation to Human Concreteness Judgments Next, we examine the relation between our word–color embeddings and human concreteness judgments [Brysbaert et al., 2014]. We label word pairs with summed concreteness ratings in the highest (lowest) quartile of our data as “concrete” (“abstract”). Fig. 1A shows that abstract word pairs are more similar in colorspace than concrete word pairs, even at fixed word2vec similarity ($p=0.001$, $\mathrm{DF}=77$, Dickey-Fuller test). Moreover, as shown in Fig. 1B, comp-syn captures concreteness judgments in a more interpretable fashion than word2vec. In particular, because abstract words are more similar in colorspace (on average), there is a monotonic relationship between color similarity and concreteness; indeed, a linear model of comp-syn similarity accounts for nearly all of the variance in word pair concreteness ($R^{2}=0.96$). On the other hand, word2vec similarity is a nonlinear function of word pair concreteness. Although nonlinear functions of word2vec similarity also predict concreteness well ($R^{2}>0.99$), comp-syn versions of the same nonlinear models are significantly more accurate (log-likelihood difference $\Delta\ln\mathcal{L}\sim 10$ in favor of comp-syn) and less complex (Bayesian information criterion $\Delta\mathrm{BIC}\sim 20$, also in favor of comp-syn). To qualitatively explore the relationship between our word–color embeddings and human concreteness judgment, Fig. 3A shows the perceptually uniform color distributions associated with some of the most and least concrete words in our corpus. These _colorgrams_ illustrate that concrete words (e.g., “pyramid”) are often associated with color distributions that are peaked in specific regions of colorspace, while abstract words (e.g., “concept”) feature more variegated color distributions. The spatial and textural features of the images reflect these properties, and exploring the relationship between these aspects of _colorgrams_ , is an interesting avenue for future study. Figure 2: Classifying metaphorical vs. literal adjective-noun pairs with comp- syn. (A) Test set classification accuracy for comp-syn (magenta) and word2vec (gray) as a function of PCA embedding dimension. (B) Average similarity of metaphorical and literal adjective-noun pairs in word2vec (left panel) and comp-syn (right panel). Error bars indicate $95\%$ confidence intervals. ### 3.3 Metaphor Pair Analysis The results above suggest that our word–color embeddings encode complementary information about concept concreteness relative to purely textual embeddings. This raises the question of what can be learned from cases in which word2vec and comp-syn provide conflicting similarity predictions when evaluated on the same pair of words. Here, we address this question by demonstrating that comp- syn significantly enriches metaphorical word pair classification, which often requires extensive manual tagging due to subtle uses of both sensory and abstract features [Lakoff and Johnson, 2008, Bethard et al., 2009, Indurkhya and Ojha, 2013, Dodge et al., 2015, Winter, 2019]. We trained a gradient-boosted tree classifier implemented via XGBoost [Chen and Guestrin, 2016] to label adjective-noun pairs as either metaphorical or literal, using over $8000$ word pairs from Tsvetkov et al. [Tsvetkov et al., 2014] and Gutiérrez et al. [Gutiérrez et al., 2016]. This dataset encompasses a statistically representative range of metaphorical and literal contexts for each adjective [Gutiérrez et al., 2016]. To compare embeddings, we compressed word2vec’s 300-dimensional word vector differences using PCA to match the dimensionality of comp-syn, following Bolukbasi et al. [Bolukbasi et al., 2016]. Fig. 2A shows that a classifier trained using only word2vec achieves a limiting test set accuracy of $95\%$, compared to $92\%$ for comp-syn. Importantly, comp-syn outperforms word2vec at low embedding dimensions, indicating that it captures semantic content in an interpretable fashion. This analysis does not demonstrate either model’s best-case performance on this task; rather, it highlights the complementary information provided by comp- syn. Strikingly, word2vec and comp-syn distances provide qualitatively different information when distinguishing literal and metaphorical adjective-noun pairs. Fig. 2B shows that literal adjective-noun pairs are more similar than metaphorical pairs in word2vec ($p<0.001$, Wilcoxon rank sum); the reverse holds in comp-syn, where literal pairs are significantly _less_ similar than metaphorical pairs ($p<0.001$, Wilcoxon rank sum). This is a consequence of the fact that images associated with concrete words are more variable in colorspace [Guilbeault et al., 2020]. In this way, comp-syn reveals differences in color similarity between literal and metaphorical adjective- noun pairs that are of interest for cognitive theory [Indurkhya and Ojha, 2013]. Particularly, our findings suggest that metaphors can exploit color similarities between words that are dissimilar in textual embeddings, which may help facilitate cognitive processing of semantic relations among concepts from distinct domains [Guilbeault et al., 2020]. Qualitative inspection of specific adjective-noun word pairs highlights some notable differences between textual and word–color embeddings. Fig. 3B provides visual representations of the color distributions for word pairs in our metaphorical vs. literal dataset that are most and least similar in comp- syn. Interestingly, while metaphorical pairs are more similar than literal pairs in comp-syn, the _least_ similar metaphorical pairs explicitly invoke color, e.g., “deep orange”. Algorithmically, this is due to the fact that comp-syn embeddings associated with color terms are unusually coherent. On the other hand, the color distributions associated with the most similar pairs in word2vec (e.g., “bushy beard”) often noticeably contrast, while word2vec’s least similar pairs (e.g., “rough customer”) do not strongly invoke color. These findings point to an important direction for future research enabled by the grounded nature of comp-syn: how do linguistic metaphors leverage sensory information to characterize colorspace itself (e.g., in the use of spatial information in the popular metaphor “deep purple”)? Figure 3: (A) Examples of the most and least concrete terms, visualized using our word–color embedding method. (B) Examples of the most and least similar adjective-noun pairs according to comp-syn word–color embeddings. We represent each term using a _colorgram_ , i.e., a composite image produced by averaging the perceptually uniform colors of pixels across Google Image search results. ## 4 Conclusion and Future Work We have presented comp-syn, a new Python package that provides perceptually grounded color-based word embeddings. These embeddings are interpretable with respect to theories of embodied cognition because (1) comp-syn represents color in a fashion that emulates human perception, and (2) comp-syn leverages Google Image search results that human users dynamically interact with and produce. By linking comp-syn with human concreteness judgments for 40,000 common English words, our package provides a multi-modal playground for exploring grounded semantics. We demonstrated that comp-syn enriches popular distributional semantics models in both word concreteness prediction and metaphorical word-pair classification. A myriad of comp-syn applications await, including color-based classification of text genres and the characterization of sensory imagery in everyday language. ## Acknowledgements The authors gratefully acknowledge the support of the Complex Systems Summer School hosted at the Institute of American Indian Arts and the Santa Fe Institute, where this project was initiated. D.G. gratefully acknowledges financial support from the Institute on Research and Innovation in Science. This research received support from the National Science Foundation (NSF) under grant no. NSF DGE-1656518 through the NSF Graduate Research Fellowship received by E.O.N. D.G. and T.H. acknowledge intellectual support from the Institute for Advanced Learning (IAL) in Ontario, Canada. ## References * [Bergen, 2012] Benjamin K Bergen. 2012\. Louder than words: The new science of how the mind makes meaning. Basic Books. * [Bethard et al., 2009] Steven Bethard, Vicky Lai, and James Martin. 2009\. Topic model analysis of metaphor frequency for psycholinguistic stimuli. Proceedings of the Workshop on Computational Approaches to Linguistic Creativity, pages 9–16. * [Bolukbasi et al., 2016] Tolga Bolukbasi, Kai-Wei Chang, James Zou, Venkatesh Saligrama, and Adam Kalai. 2016\. Man is to computer programmer as woman is to homemaker? debiasing word embeddings. In Proceedings of the 30th International Conference on Neural Information Processing Systems, NIPS’16, page 4356–4364, Red Hook, NY, USA. Curran Associates Inc. * [Bruni et al., 2014] Elia Bruni, Nam-Khanh Tran, and Marco Baroni. 2014\. Multimodal distributional semantics. Journal of artificial intelligence research, 49:1–47. * [Brysbaert et al., 2014] Marc Brysbaert, Amy Beth Warriner, and Victor Kuperman. 2014\. Concreteness ratings for 40 thousand generally known english word lemmas. Behavior research methods, 46(3):904–911. * [Chen and Guestrin, 2016] Tianqi Chen and Carlos Guestrin. 2016\. XGBoost: A scalable tree boosting system. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD ’16, pages 785–794, New York, NY, USA. ACM. * [Clark, 2018] Alex Clark. 2018\. Pillow python imaging library. Pillow—Pillow (PIL Fork) 5.4. 1 documentation. * [Deng et al., 2009] Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. 2009\. Imagenet: A large-scale hierarchical image database. 2009 IEEE Conference on Computer Vision and Pattern Recognition. * [Dodge et al., 2015] Ellen Dodge, Jisup Hong, and Elise Stickles. 2015\. Metanet: Deep semantic automatic metaphor analysis. Proceedings of the Third Workshop on Metaphor in NLP, pages 40–49. * [Elliot and Maier, 2014] Andrew Elliot and Markus Maier. 2014\. Color psychology: Effects of perceiving color on psychological functioning in humans. Annual Review of Psychology, 65:95–120. * [Gallese and Lakoff, 2005] Vittorio Gallese and George Lakoff. 2005\. The brain’s concepts: The role of the sensory-motor system in conceptual knowledge. Cognitive neuropsychology, 22(3-4):455–479. * [Guilbeault et al., 2020] Douglas Guilbeault, Ethan O Nadler, Mark Chu, Donald Ruggiero Lo Sardo, Aabir Abubaker Kar, and Bhargav Srinivasa Desikan. 2020\. Color associations in abstract semantic domains. Cognition, 201:104306. * [Gutiérrez et al., 2016] E. Dario Gutiérrez, Ekaterina Shutova, Tyler Marghetis, and Benjamin Bergen. 2016\. Literal and metaphorical senses in compositional distributional semantic models. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 183–193, Berlin, Germany, August. Association for Computational Linguistics. * [Hunter, 2007] John D Hunter. 2007\. Matplotlib: A 2d graphics environment. Computing in science & engineering, 9(3):90–95. * [Indurkhya and Ojha, 2013] Bipin Indurkhya and Amitash Ojha. 2013\. An empirical study on the role of perceptual similarity in visual metaphors and creativity. Metaphor and Symbol, 28. * [International Commission On Illumination, 1978] International Commission On Illumination. 1978\. Recommendations on uniform color spaces, color-difference equations, psychometric color terms. Color Research & Application, 15. * [Ji et al., 2019] S. Ji, N. Satish, S. Li, and P. K. Dubey. 2019\. Parallelizing word2vec in shared and distributed memory. IEEE Transactions on Parallel and Distributed Systems, 30(9):2090–2100. * [Jing and Baluja, 2008] Yushi Jing and Shumeet Baluja. 2008\. Pagerank for product image search. In Proceedings of the 17th international conference on World Wide Web, pages 307–316. * [Jones et al., 2001] Eric Jones, Travis Oliphant, Pearu Peterson, et al. 2001\. Scipy: Open source scientific tools for python. * [Lakoff and Johnson, 2008] George Lakoff and Mark Johnson. 2008\. Metaphors we live by. University of Chicago press. * [Lakoff and Turner, 1989] George Lakoff and Mark Turner. 1989\. More than cool reason: A field guide to poetic metaphor. University of Chicago press. * [Landauer and Dumais, 1997] Thomas K Landauer and Susan T Dumais. 1997\. A solution to plato’s problem: The latent semantic analysis theory of acquisition, induction, and representation of knowledge. Psychological review, 104(2):211. * [Levy and Goldberg, 2014] Omer Levy and Yoav Goldberg. 2014\. Dependency-based word embeddings. In Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pages 302–308. * [Lu et al., 2019] Jiasen Lu, Dhruv Batra, Devi Parikh, and Stefan Lee. 2019\. Vilbert: Pretraining task-agnostic visiolinguistic representations for vision-and-language tasks. In Advances in Neural Information Processing Systems, pages 13–23. * [Mehta and Zhu, 2009] Ravi Mehta and Rui Zhu. 2009\. Blue or red? exploring the effect of color on cognitive task performances. Science, 323:1226–1229. * [Mikolov et al., 2013] Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg S Corrado, and Jeff Dean. 2013\. Distributed representations of words and phrases and their compositionality. In Advances in neural information processing systems, pages 3111–3119. * [Miller, 1998] George A Miller. 1998\. WordNet: An electronic lexical database. MIT press. * [Pérez and Granger, 2007] Fernando Pérez and Brian E Granger. 2007\. Ipython: a system for interactive scientific computing. Computing in science & engineering, 9(3):21–29. * [Riley, 1995] Charles Riley. 1995\. Color Codes: Modern Theories of Color in Philosophy, Painting and Architecture, Literature, Music, and Psychology. UPNE. * [Safdar et al., 2017] Muhammad Safdar, Guihua Cui, Youn Jin Kim, and Ming Ronnier Luo. 2017\. Perceptually uniform color space for image signals including high dynamic range and wide gamut. Optics express, 25(13):15131–15151. * [Socher et al., 2014] Richard Socher, Andrej Karpathy, Quoc V Le, Christopher D Manning, and Andrew Y Ng. 2014\. Grounded compositional semantics for finding and describing images with sentences. Transactions of the Association for Computational Linguistics, 2:207–218. * [Tsvetkov et al., 2014] Yulia Tsvetkov, Leonid Boytsov, Anatole Gershman, Eric Nyberg, and Chris Dyer. 2014\. Metaphor detection with cross-lingual model transfer. In Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 248–258, Baltimore, Maryland, June. Association for Computational Linguistics. * [Virtanen et al., 2020] Pauli Virtanen, Ralf Gommers, Travis E Oliphant, Matt Haberland, Tyler Reddy, David Cournapeau, Evgeni Burovski, Pearu Peterson, Warren Weckesser, Jonathan Bright, et al. 2020\. Scipy 1.0: fundamental algorithms for scientific computing in python. Nature methods, 17(3):261–272. * [Walt et al., 2011] Stéfan van der Walt, S Chris Colbert, and Gael Varoquaux. 2011\. The numpy array: a structure for efficient numerical computation. Computing in science & engineering, 13(2):22–30. * [Winter, 2019] Bodo Winter. 2019\. Sensory Linguistics. John Benjamins Publishing Company. * [Şenel et al., 2018] L. K. Şenel, İ. Utlu, V. Yücesoy, A. Koç, and T. Çukur. 2018\. Semantic structure and interpretability of word embeddings. IEEE/ACM Transactions on Audio, Speech, and Language Processing, 26(10):1769–1779. ## Appendix A comp-syn Code comp-syn is an open source Python package, available on GitHub and downloadable through PyPi at https://github.com/comp-syn/comp-syn. It uses NumPy [Walt et al., 2011] and scipy [Jones et al., 2001] for computational purposes, matplotlib [Hunter, 2007] for visualisations, and Pillow [Clark, 2018] to load and manipulate images in Python. We provide documentation and extensive examples of code use through Jupyter/IPython notebooks [Pérez and Granger, 2007]. With a focus on community-based development, our code is PEP8 compliant, heavily commented and documented, includes continuous code integration and testing via Travis CI, and uses GitHub as a platform for issue tracking and ideation. comp-syn is structured so that it is easily compatible with contemporary deep learning, natural language processing and computational linguistics Python packages, and particularly with NumPy and SciPy as standard tools for scientific computing [Virtanen et al., 2020]. comp-syn includes the following modules: * • datahelper.py \- loader functions for scraping and organising image data; * • analysis.py \- analysis functions for computing word–color embeddings; * • logger.py \- logger functions; * • visualisation.py \- visualisation functions; * • vectors.py \- word–color embedding loader and manager; * • wordnet-functions.py \- WordNet functions to facilitate web scraping and linking with concreteness measures. ## Appendix B Data Distribution Along with the software, we also release word–color embeddings for the nearly 40,000 English language words. Our corpus of word–color embeddings is linked to crowd-sourced human judgments of concept concreteness from Brysbaert et al. [Brysbaert et al., 2014], and terms used for analysis are also linked to WordNet [Miller, 1998]. These embeddings can easily be downloaded and loaded into Python pipelines for various natural language processing tasks, including jointly training or enhancing distributional semantics models. The data is stored in the json file concreteness-color-embeddings.json and can be loaded using the vectors.py module class LoadVectorsFromDisk. Once loaded, each Vector object has the following information based on the top $100$ Google Image results for each word: * • rgb-dist \- the average RGB distribution, computed in $8$ evenly-segmented bins; * • jzazbz-dist \- the average perceptually uniform $J_{z}A_{z}B_{z}$ distribution, computed in $8$ evenly-segmented bins; * • jzazbz-dist-std \- the standard deviation over the perceptually uniform $J_{z}A_{z}B_{z}$ distribution, computed in $8$ evenly-segmented bins; * • colorgram \- a composite image created by pixel-wise averaging over the top $100$ images, representing the word’s “average image”; * • rgb-vector \- the mean RGB coordinates, averaged over the top $100$ images; * • jzazbz-vector \- the mean $J_{z}A_{z}B_{z}$ coordinates, averged over the top $100$ images * • concreteness-mean \- the mean of the concreteness scores for the word from [Brysbaert et al., 2014]; * • concreteness-sd \- the standard deviation of the concreteness scores for the word from [Brysbaert et al., 2014]. ## Appendix C Methods ### C.1 Google Image Search To generate word–color embeddings, we collect the top $100$ Google Image search results for each of the 40,000 terms in our analysis. The Google Image searches were run from 10 servers running in a commercial datacenter in New York, USA. These servers were created and used only for this experiment, so that results would not be overly-personalized. Additional search parameters were included as query strings safe=off&site=&tbm=isch&source=hp&gs_l=img. ### C.2 Word–Color Embeddings We use the PIL Python module to convert each image into an $m\times n\times 3$ array of sRGB values, where $m$ and $n$ are the intrinsic image dimensions. For computational efficiency, we then compress each image into an anti-aliased $300\times 300\times 3$ array. Next, we transform sRGB pixel values into their counterparts in the perceptually uniform $J_{z}A_{z}B_{z}$ colorspace.222comp- syn efficiently computes color embeddings in multiple colorspaces if desired, including more traditional (non-perceptually uniform) options such as RGB and HSV. Unlike in standard colorspace, Euclidean distances in $J_{z}A_{z}B_{z}$ coordinates linearly correspond to differences in human color perception [Safdar et al., 2017]. Moreover, our use of the $J_{z}A_{z}B_{z}$ colorspace rather than a standard colorspace like RGB increases the semantic coherence of our word–color embeddings [Guilbeault et al., 2020]. We measure the color distribution of each image in 8 evenly-segmented $J_{z}A_{z}B_{z}$ subvolumes spanning the range of $J_{z}A_{z}B_{z}$ coordinates that maps to all possible RGB tuples: $J_{z}\in[0,0.167]$, $A_{z}\in[-0.1,0.11]$, $B_{z}\in[-0.156,0.115]$. Next, we average $J_{z}A_{z}B_{z}$ distributions over all $100$ images for each term to obtain an aggregate, 8-dimensional color mean embedding. We also compute the standard deviation over the $100$ images in each $J_{z}A_{z}B_{z}$ subvolumes to obtain an aggregate, 8-dimensional color variability embedding. These color mean and variability embeddings can be concatenated to create a 16-dimensional embedding. The details of our compression, binning, and averaging steps do not affect our results [Guilbeault et al., 2020]. To compare word–color embeddings, we use the Jensen-Shannon (JS) divergence to measure the similarity of aggregate $J_{z}A_{z}B_{z}$ distributions. In particular, for $J_{z}A_{z}B_{z}$ distributions $C_{i}$ and $C_{j}$ associated with terms $i$ and $j$, the JS divergence is given by $D_{\mathrm{JS}}(C_{1}~{}||~{}C_{2})\equiv\frac{1}{2}\left[D_{\mathrm{KL}}(C_{1}~{}||~{}\bar{C}_{12})+D_{\mathrm{KL}}(C_{2}~{}||~{}\bar{C}_{12})\right],$ (1) where $D_{\rm{KL}}$ is the Kullback-Leibler divergence and $\bar{C}_{12}\equiv(C_{1}+C_{2})/2$. The JS divergence is a measure of the distance between two color distributions, such that lower values correspond to more similar distributions in perceptually uniform colorspace. We choose this metric because it is a well-defined distance measure that satisfies the triangle inequality and allows us to avoid undefined values associated with empty $J_{z}A_{z}B_{z}$ bins. Terms with relatively high mutual JS divergences usually exhibit _colorgrams_ with perceptibly different average colors. ## Appendix D Best Practices for Usage We caution that our word–color embeddings are _distributions_ rather than _vectors_. Thus, their components are positive semidefinite, and different embeddings must be compared using similarity measures designed for distributions such as JS divergence. On the other hand, word2vec embeddings are vectors with components that can be positive or negative, and are compared using similarity measures designed for vectors such as cosine similarity. Importantly, unlike word2vec embeddings, our word–color embeddings cannot be composed by vector addition. We are exploring algebraic techniques for word composition in colorspace; these techniques must respect the underlying mathematical structure of $J_{z}A_{z}B_{z}$ (and other) colorspaces, which are not closed under standard binary operations like addition or multiplication.
* [73] Marius Leordeanu, Rahul Sukthankar and Martial Hebert “Unsupervised learning for graph matching” In _International Journal of Computer Vision_ 96, 2012, pp. 28–45 * [74] Kang Li, Xiaodong Wu, Danny Z Chen and Milan Sonka “Optimal surface segmentation in volumetric images-a graph-theoretic approach” In _IEEE Transactions on Pattern Analysis and Machine Intelligence_ 28.1, 2005, pp. 119–134 * [75] Jiangyu Liu and Jian Sun “Parallel Graph-cuts by Adaptive Bottom-up Merging” In _IEEE Conference on Computer Vision and Pattern Recognition_ , 2010, pp. 2181–2188 * [76] Lei Liu, Yun Sheng, Guixu Zhang and Hassan Ugail “Graph cut based mesh segmentation using feature points and geodesic distance” In _Proceedings of the International Conference on Cyberworlds (CW)_ , 2015, pp. 115–120 * [77] Eric Maire and Philip John Withers “Quantitative X-ray tomography” In _International materials reviews_ 59.1, 2014, pp. 1–43 * [78] Leland McInnes, John Healy and James Melville “Umap: Uniform manifold approximation and projection for dimension reduction” In _arXiv:1802.03426_ , 2018 * [79] Fausto Milletari, Nassir Navab and Seyed-Ahmad Ahmadi “V-net: Fully convolutional neural networks for volumetric medical image segmentation” In _International Conference on 3D Vision_ , 2016, pp. 565–571 * [80] Shervin Minaee et al. “Image segmentation using deep learning: A survey” In _IEEE Transactions on Pattern Analysis and Machine Intelligence_ , 2021 * [81] Rajmund Mokso et al. “GigaFRoST: the gigabit fast readout system for tomography” In _Journal of synchrotron radiation_ 24.6, 2017, pp. 1250–1259 * [82] Sebastian Nowozin et al. “Decision tree fields” In _International Conference on Computer Vision_ , 2011, pp. 1668–1675 * [83] F. Pedregosa et al. “Scikit-learn: Machine Learning in Python” In _Journal of Machine Learning and Research_ 12, 2011, pp. 2825–2830 * [84] Bo Peng, Lei Zhang and David Zhang “A survey of graph theoretical approaches to image segmentation” In _Pattern Recognition_ 46.3, 2013, pp. 1020–1038 * [85] Sida Peng et al. “Deep snake for real-time instance segmentation” In _IEEE Conference on Computer Vision and Pattern Recognition_ , 2020, pp. 8533–8542 * [86] Yi Peng et al. “JF-Cut: A parallel graph cut approach for large-scale image and video” In _IEEE Transactions on Image Processing_ 24.2, 2015, pp. 655–666 * [87] Marius Reichardt et al. “3D virtual Histopathology of Cardiac Tissue from Covid-19 Patients based on Phase-Contrast X-ray Tomography” In _eLife._ , 2021 * [88] Carsten Rother, Vladimir Kolmogorov, Victor Lempitsky and Martin Szummer “Optimizing binary MRFs via extended roof duality” In _IEEE Conference on Computer Vision and Pattern Recognition_ , 2007, pp. 1–8 * [89] Alexander Shekhovtsov and Václav Hlaváč “A distributed mincut/maxflow algorithm combining path augmentation and push-relabel” In _International Journal of Computer Vision_ 104.3, 2013, pp. 315–342 * [90] Amber L Simpson et al. “A large annotated medical image dataset for the development and evaluation of segmentation algorithms” In _arXiv:1902.09063_ , 2019 * [91] Petter Strandmark and Fredrik Kahl “Parallel and Distributed Graph Cuts by Dual Decomposition” In _IEEE Conference on Computer Vision and Pattern Recognition_ , 2010, pp. 2085–2092 * [92] Paul Swoboda et al. “A study of lagrangean decompositions and dual ascent solvers for graph matching” In _IEEE Conference on Computer Vision and Pattern Recognition_ , 2017, pp. 1607–1616 * [93] Nima Tajbakhsh et al. “Embracing imperfect datasets: A review of deep learning solutions for medical image segmentation” In _Medical Image Analysis_ 63, 2020, pp. 101693 * [94] Lorenzo Torresani, Vladimir Kolmogorov and Carsten Rother “Feature correspondence via graph matching: Models and global optimization” In _European Conference on Computer Vision_ , 2008, pp. 596–609 * [95] Tanmay Verma and Dhruv Batra “MaxFlow Revisited: An Empirical Comparison of Maxflow Algorithms for Dense Vision Problems” In _British Machine Vision Conference_ , 2012, pp. 1–12 * [96] Vibhav Vineet and P J Narayanan “CUDA cuts: Fast graph cuts on the GPU” In _IEEE Conference on Computer Vision and Pattern Recognition Workshops_ , 2008, pp. 1–8 * [97] Yao Wang and Reinhard Beichel “Graph-based segmentation of lymph nodes in CT data” In _International Symposium on Visual Computing_ , 2010, pp. 312–321 * [98] University Waterloo “Max-flow problem instances in vision” Accessed 2021-02-05, https://vision.cs.uwaterloo.ca/data/maxflow * [99] Udaranga Wickramasinghe, Edoardo Remelli, Graham Knott and Pascal Fua “Voxel2mesh: 3d mesh model generation from volumetric data” In _International Conference on Medical Image Computing and Computer Assisted Intervention_ , 2020, pp. 299–308 Springer * [100] Xiaodong Wu and Danny Z Chen “Optimal net surface problems with applications” In _International Colloquium on Automata, Languages, and Programming_ , 2002, pp. 1029–1042 * [101] Yin Yin et al. “LOGISMOS—layered optimal graph image segmentation of multiple objects and surfaces: cartilage segmentation in the knee joint” In _IEEE Transactions on Medical Imaging_ 29.12, 2010, pp. 2023–2037 * [102] Miao Yu, Shuhan Shen and Zhanyi Hu “Dynamic Graph Cuts in Parallel” In _IEEE Transactions on Image Processing_ 26.8 Institute of ElectricalElectronics Engineers Inc., 2017 * [103] Miao Yu, Shuhan Shen and Zhanyi Hu “Dynamic Parallel and Distributed Graph Cuts” In _IEEE Transactions on Image Processing_ 25.12, 2015, pp. 5511–5525 | Patrick M. Jensen was born in Copenhagen, Denmark, in 1994. He received his B.Sc.Eng degree in 2017 and M.Sc.Eng degree in 2019, both in applied mathematics, at the Technical University of Denmark (DTU), Kgs. Lyngby, Denmark. He is currently pursuing a Ph.D. in 3D image analysis at the Visual Computing group at the Department of Applied Mathematics and Computer Science, Technical University of Denmark. His research interests lie in 3D image segmentation. ---|--- | Niels Jeppesen is an image analysis and machine learning specialist at FORCE Technology with a Ph.D. degree in image analysis of 3D structures from the Department of Applied Mathematics and Computer Science, Technical University of Denmark (DTU), Kgs. Lyngby, Denmark. His research interests lie in min-cut/max-flow algorithms and quantitative analysis of structures in 3D images. He applies these methods for automated quality control of structures and materials, in particular, in the wind turbine industry. ---|--- | Anders Bjorholm Dahl is professor in 3D image analysis, and a head of the Section for Visual Computing at the Department of Applied Mathematics and Computer Science, Technical University of Denmark (DTU), Kgs. Lyngby, Denmark. He is heading The Center for Quantification of Imaging Data from MAX IV, focusing on quantitative analysis of 3D images. His research is focused on image segmentation and its applications. ---|--- | Vedrana Andersen Dahl is an associate professor at the Department of Applied Mathematics and Computer Science, Technical University of Denmark (DTU), Kgs. Lyngby, Denmark. Her primary research interest is in the use of geometric models for the analysis of volumetric data. This includes volumetric segmentation and methods based on deformable meshes. She developed analysis tools with applications in material science, industrial inspection, and biomedicine. ---|---
# Topological properties of P.A. random graphs with edge-step functions Caio Alves1 1 Institute of Mathematics, , University of Leipzig – Augustusplatz 10, 04109 Leipzig e-mail<EMAIL_ADDRESS>, Rodrigo Ribeiro2 2 PUC Chile, Av. Vicuña Mackenna 4860, Macul, La Florida, Región Metropolitana, Chile. e-mail<EMAIL_ADDRESS>and Rémy Sanchis3 3Departamento de Matemática, Universidade Federal de Minas Gerais, Av. Antônio Carlos 6627 C.P. 702 CEP 30123-970 Belo Horizonte-MG, Brazil e-mail<EMAIL_ADDRESS> (Date: 1 Institute of Mathematics, University of Leipzig 2 PUC Chile, Pontificia Universidad Católica de Chile. 3 Departamento de Matemática, Universidade Federal de Minas Gerais. ) ###### Abstract. In this work we investigate a preferential attachment model whose parameter is a function $f:\mathbb{N}\to[0,1]$ that drives the asymptotic proportion between the numbers of vertices and edges of the graph. We investigate topological features of the graphs, proving general bounds for the diameter and the clique number. Our results regarding the diameter are sharp when $f$ is a regularly varying function at infinity with strictly negative index of regular variation $-\gamma$. For this particular class, we prove a characterization for the diameter that depends only on $-\gamma$. More specifically, we prove that the diameter of such graphs is of order $1/\gamma$ with high probability, although its vertex set order goes to infinity polynomially. Sharp results for the diameter for a wide class of slowly varying functions are also obtained. The almost sure convergence for the properly normalized logarithm of the clique number of the graphs generated by slowly varying functions is also proved. _Keywords_ : complex networks; cliques; preferential attachment; concentration bounds; diameter; scale-free; small-world MSC 2010 subject classifications. Primary 05C82; Secondary 60K40, 68R10 ## 1\. Introduction P. Ërdos and A. Rényi in their seminal paper [17] introduced the random graph model that now carries their name in order to solve combinatorial problems. However, the theory of Random Graphs as a whole has proven to be a useful tool for treating concrete problems as well. Any discrete set of entities whose elements interact in a pairwise fashion may be seen as a graph: the vertices represent the entities, and the edges, the possible interactions. This approach is nowadays intuitive and very fruitful. In the scenario where there exists some randomness on the interactions among the entities, random graphs became the natural tool to represent abstract or real phenomena. From a mathematical/statistical point of view, the Ërdos-Rényi model – and many others related to it – is homogeneous, in the sense that its vertices are statistically indistinguishable. However, the empirical findings of the seminal work of A. Bárabasi and R. Álbert [5] suggested that many real-world networks are non-homogeneous. They observed that such graphs were scale-free, i.e., their degree sequence had a power-law distribution. The authors proposed a mechanism – known as preferential attachment – that could explain the emergence of such highly skewed distributions. Roughly speaking, the idea is that some sort of popularity drives the interaction among the entities. Motivated mainly by these empirical findings, nowadays Preferential Attachment models (PA-models for short) constitutes a well known class of random graph models investigated from both theoretical and applied perspectives. Recently, the preferential attachment mechanism has been generalized in many ways and combined with other rules of attachment, such as spatial proximity [19] and fitness of vertices [14]. It also arises naturally even in models where it is not entirely explicit such as the deletion-duplication models [4, 26], in which vertices’ degree still evolve according to the PA-rule. Furthermore, the PA-models provide an interesting and natural environment for other random processes, such as bootstrap percolation, contact process and random walks, see [3, 11, 20] for recent examples of random processes whose random media is sampled from some PA-model. When dealing with PA-models, there exists a set of natural questions that arises. They concern the empirical degree distribution, the order of the diameter and the robustness of the network. Their interest relies on modeling purposes and on the implications for the graph’s combinatorial structures. In this paper we address the two latter topics on a PA-model which is a modification of the BA-model. The kind of result we pursuit is to show that some graph properties hold asymptotically almost surely (a.a.s). Given a sequence of random graphs $\\{G_{t}\\}_{t\in\mathbb{N}}$, we say that a graph property $\mathcal{P}$ holds a.a.s, if ${\mathbb{P}}\left(G_{t}\in\mathcal{P}\right)=1-o(1)$ i.e., the probability of observing such property increases to $1$ as $t$ goes to infinity. For instance, $\mathcal{P}$ may be the set of graphs having diameter less than the logarithm of the total number of vertices. In order to offer a clearer discussion of our results, we introduce the model in the next subsection, then we discuss separately the properties which we want the graph to satisfy a.a.s, as well as the associated motivation. ### 1.1. Preferential attachment model with an edge-step function The model we investigate here in its generality was proposed in [2] and combines the traditional preferential attachment rule with a function called edge-step function that drives the growth rate of the vertex set. The model has one parameter $f$ which is a real non-negative function with domain given by $\mathbb{N}$ and bounded by one on the $L_{\infty}$-norm. Without loss of generality and to simplify the expressions we deal with, we start the process from an initial graph $G_{1}$ consisting in one vertex and one loop. The model evolves inductively and at each step the next graph is obtained by performing one of the two stochastic operations defined below on the previous one: * • Vertex-step \- Add a new vertex $v$ and add an edge $\\{u,v\\}$ by choosing $u\in G$ with probability proportional to its degree. More formally, conditionally on $G$, the probability of attaching $v$ to $u\in G$ is given by $P\left(v\rightarrow u\middle|G\right)=\frac{\text{degree}(u)}{\sum_{w\in G}\text{degree}(w)}.$ * • Edge-step \- Add a new edge $\\{u_{1},u_{2}\\}$ by independently choosing vertices $u_{1},u_{2}\in G$ according to the same rule described in the vertex-step. We note that both loops and parallel edges are allowed. The model alternates between the two types of operations according to a sequence $\\{Z_{t}\\}_{t\geq 1}$ of independent random variables such that $Z_{t}\stackrel{{\scriptstyle\tiny d}}{{=}}\mathrm{Ber}(f(t))$. We then define inductively a random graph process $\\{G_{t}(f)\\}_{t\geq 1}$ as follows: start with $G_{1}$. Given $G_{t}(f)$, obtain $G_{t+1}(f)$ by either performing a vertex-step on $G_{t}(f)$ when $Z_{t}=1$ or performing an edge-step on $G_{t}(f)$ when $Z_{t}=0$. Given $f$, its partial sum is an important quantity for us and we reserve the letter $F$ to denote it, i.e., $F$ is a function defined as (1.1) $F(t):=1+\sum_{s=2}^{t}f(s).$ Observe that the edge-step function $f$ is intimately related to the growth of the vertex set. If we let $V_{t}$ denote the number of vertices added up to time $t$, then (1.2) $V_{t}=1+\sum_{s=2}^{t}Z_{s}\approx F(t),$ since the sequence of random variables $(Z_{s})_{s\geq 1}$ is independent. Thus, abusing from the notation for a brief moment, we may write $\frac{\mathrm{d}V_{t}}{\mathrm{d}t}=f(t).$ When the proper machinery has been settled, we will discuss in Section 8 that some regularity should be imposed on $f$ in order to avoid some pathological behaviors. For now, we define a list of conditions we may impose on $f$ at different points of the paper in order to get the proper results. For instance, we say $f$ satisfies condition $(\mathrm{D})$ if it is non- increasing. We define the further conditions: (D0) $f\text{ is non-increasing and }\lim_{t\to\infty}f(t)=0;$ (S) $\sum_{s=1}^{\infty}\frac{f(s)}{s}<\infty;$ (Lκ) $\sum_{s=t^{1/13}}^{t}\frac{f(s)}{s}<(\log t)^{\kappa},\text{ for all }t\in\mathbb{N}\text{ and some }\kappa\in(0,1);$ (RVγ) $\exists\gamma,\text{ such that }\forall a>0,\;\lim_{t\to\infty}\frac{f(at)}{f(t)}=\frac{1}{a^{\gamma}}.$ We must point out that for modeling purposes, conditions (D) and (D0) may be desirable. For instance, in the context of social networks, these conditions assure that the rate at which new individuals join the network is decreasing as the size of the network increases. Whereas, conditions (S) and (Lκ) are related to the order of the maximum degree of $G_{t}(f)$. In [2], the authors point out that the maximum degree at time $t$ should be of order (1.3) $t\cdot\exp\left\\{-\frac{1}{2}\sum_{s=2}^{t}\frac{f(s)}{s-1}\right\\}.$ A function satisfying condition (RVγ) is called regularly varying at infinity and the exponent $\gamma$ is called the index of regular variation. Functions in this class are well-studied in mathematics in many contexts and a variety of asymptotic results for them and their integrals is known due mainly to the theory developed by Karamata, see [8] for a complete reference. In general, we may say that this paper investigates how sensitive some graph observables are to changes of $f$ and aims at a general characterization of such observables for a class of functions $f$ that is as wide as possible. ### 1.2. Robustness and large cliques As said before, one of the main questions concerns the robustness of the network or how vulnerable the network is to spread of a disease [6] or to deliberate/random attacks [10]. Regarding the spread of rumors of diseases some graph substructures play important roles, such as stars, triangles and cliques (complete subgraphs). The latter is also related to the robustness of the network, since the existence of a large clique may let the network less vulnerable to attacks aiming at edge-deletion. In [7], the authors give lower and upper bounds for the clique number in a uncorrelated PA-model. Their bounds are polynomial on the number of vertices. In the same direction, in [21] the authors prove some sort of phase transition on the clique number of a random graph $G(n,\alpha)$ on $n$ vertices and degree sequence obeying a power-law with exponent $\alpha$. For their model, when $\alpha>2$ the clique number is of constant size, whereas for heavy tails, when $0<\alpha<2$, $\omega(G(n,\alpha))$ is a power of $n$. In this matter, we prove the existence a.a.s of large cliques whose order depends essentially on the definite integral of $f$. The result is formally stated on the theorem below ###### Theorem 1 (Large cliques). Let $f$ be an edge-step function satisfying condition (D0). Then, for every $\delta\in(0,1)$, there exist a positive constant $C$ depending on $\delta$ only such that ${\mathbb{P}}\left(\exists K_{n}\subset G_{t}(f),\textit{ such that }n\geq CF(t^{\frac{1-\delta}{2}})\right)=1-o(1),$ where $K_{n}$ denotes a subgraph isomorphic to the complete graph with $n$ vertices. For edge-step functions which are also slowly varying varying the above theorem can provide the right order of largest clique of $G_{t}(f)$, i.e., the so-called clique number, denoted by $\omega(G_{t}(f))$. In the more general case of a regularly varying function $f$, a non-sharp result can be obtained. We summarize this in the Corollary below ###### Corollary 1.4 (Clique number for regular varying functions). let $f$ be an edge-step function satisfying conditions (D0) and (RVγ), for $\gamma\in[0,1)$. Then, 1. (a) for any $\varepsilon>0$ and $t$ sufficiently large ${\mathbb{P}}\left(t^{(1-\gamma)\frac{1-\varepsilon}{2}}\leq\omega(G_{t}(f))\leq 7t^{\frac{1}{2}}\right)\geq 1-\frac{1}{\log(t)};$ 2. (b) For $f$ under (RV0), we have $\lim_{t\to\infty}\frac{\log\omega(G_{t}(f))}{\log F(t^{\frac{1}{2}})}=1,\;\text{almost surely.}$ ### 1.3. Shaping the diameter Another topological property of graphs which is also related to spread of rumors and connectivity of networks is the diameter, that is, the maximal graph distance between two vertices of said graph. Originally, investigating the diameter of real-world networks, the authors in [25] observed that, although coming from different contexts, those networks usually have diameter of order less than the logarithm of the number of verices, the so-called small-world phenomena. In this paper we also address the issue of determining the order of the diameter. Our main goals in this subject are to obtain a characterization for the diameter imposing conditions on $f$ as weak as possible and also to obtain regimes for the diameter arbitrarily small but still preserving the scalefreeness of the graph. In order to slow the growth of the diameter of PA-models, two observables play important roles: the maximum degree and the proportion of vertices with low degree. The former tends to concentrate connections on vertices with very high degree which acts in the way of shortening the diameter, since they attract connections to them. Whereas the latter, acts in the opposite way. In [28] and [18], the authors have shown that in the configuration model with power-law distribution the diameter order is extremely sensitive to the proportion of vertices with degree 1 and 2. One way to reduce the effect of low degree vertices on the diameter is via affine preferential attachment rules, i.e., introducing a parameter $\delta$ and choosing vertices with probability proportional to their degree plus $\delta$. In symbols, conditionally on $G_{t}$, we connect a new vertex $v_{t+1}$ to an existing one $u$ with probability ${\mathbb{P}}\left(v_{t+1}\rightarrow u\middle|G_{t}\right)=\frac{\text{degree}(u)+\delta}{\sum_{w\in G_{t}}(\text{degree}(w)+\delta)}.$ By taking a negative $\delta$, the above rule increases the influence of high degree vertices and indeed decreases drastically the diameter’s order. For instance, for positive $\delta$ the diameter of $G_{t}$ is at least $\log(t)$, whereas for $\delta<0$ the diameter of $G_{t}$ is at most $\log\log t$. See [15] for several results on the diameter of different combinations for the affine preferential attachment rule. Diminishing the effect of low degree vertices is not enough to break the growth of the diameter completely. The reason for that is, despite their low degree, these vertices exist in large amount. Even the existence of a vertex with degree close to $t$ at time $t$ may not be enough to freeze the diameter’s growth. In [23] the authors have proven that the maximum degree of a modification of the BA-model is of order $t$ at time $t$. However, the authors believe that this is not enough to obtain a diameter of order $\log\log t$, the reason being that this large hub still has to compete with a large number of low degree vertices. #### 1.3.1. General bounds for the diameter As said before, our goal is to develop bounds for the diameter of $G_{t}(f)$ with $f$ as general as possible. Under the condition of monoticity, we prove the following lower bound ###### Theorem 2 (Lower bound on the diameter). Let $f$ be an edge-step function satisfying condition $(\mathrm{D})$. Then (1.5) ${\mathbb{P}}\left(\mathrm{diam}(G_{t}(f))\geq\frac{1}{3}\left(\frac{\log t}{\log\log t}\wedge\frac{\log t}{-\log f(t)}\right)\right)=1-o(1).$ For the upper bound, by a coupling argument, we are able to prove that the diameter of the Barabási-Albert random tree is the ceiling for the diameter generated by any $f$. Requiring more information on $f$, we prove upper bounds that, for a broad class of functions, are of the same order of the lower bounds given by the previous theorem. This is all summarized in the Theorem below. ###### Theorem 3 (Upper bound on the diameter). Let $f$ be an edge-step function. Then 1. (a) $\mathrm{diam}(G_{t}(f))$ is at most the diameter of the Barabási-Albert random tree, i.e., ${\mathbb{P}}\left(\mathrm{diam}(G_{t}(f))\leq\log t\right)=1-o(1).$ 2. (b) if $f$ also satisfies condition (S) then there exists a positive constant $C_{1}$ such that ${\mathbb{P}}\left(\mathrm{diam}(G_{t}(f))\leq 2+6\left(\frac{\log t}{-\log\left(\sum_{s=t^{\frac{1}{13}}}^{t}\frac{f(s)}{s-1}\right)}\wedge\frac{\log t}{\log\log t}\right)\right)\geq 1-C_{1}t^{-144^{-1}};$ 3. (c) if $f$ satisfies condition (Lκ) then there exists a positive constant $C_{2}$ such that ${\mathbb{P}}\left(\mathrm{diam}(G_{t}(f))\leq 2+\frac{6}{1-\kappa}\frac{\log t}{\log\log t}\right)\geq 1-C_{2}t^{-144^{-1}}.$ #### 1.3.2. The class regularly varying functions In [2], the authors prove a characterization of the empirical degree distribution of graphs generated by $f$ satisfying condition (RVγ), for $\gamma\in[0,1)$. More specifically, they prove that the degree distribution of such graphs obeys a power law distribution whose exponent depends only on the index of regular variation $-\gamma$. A byproduct of our general bounds is a similar characterization for the diameter. For edge-step functions satisfying conditions (D0) and (RVγ) for $\gamma\in(0,\infty)$ the graphs generated by such functions have constant diameter and its order depends only on the index of regular variation $-\gamma$. We state this result in the theorem below ###### Theorem 4 (Diameter of regularly varying functions). Let $f$ be an edge-step function satisfying conditions (D0) and (RVγ), for $\gamma\in(0,\infty)$. Then, ${\mathbb{P}}\left(\frac{1}{4\gamma}\leq\mathrm{diam}(G_{t}(f))\leq\frac{100}{\gamma}+2\right)=1-o(1).$ #### 1.3.3. The class of slowly varying functions The case when $\gamma=0$ is richer in terms of possible orders of the diameter and does not admit a nice characterization as the one we obtain for positive $\gamma$. In this settings, we present another consequence of our bounds for particular subclasses of the class of slowly varying functions. Let us first define the subclass of functions and later state how our results fit these specific classes. (1.6) $L:=\left\\{f\text{ is an edge-step function such that }f(t)=\frac{1}{\log^{\alpha}(t)},\text{ for some }\alpha>0\right\\};$ (1.7) $E:=\left\\{f\text{ is an edge-step function such that }f(t)=e^{-\log^{\alpha}(t)},\text{ for some }\alpha\in(0,1)\right\\}.$ It is straightforward to verify that functions belonging to the set above defined are slowly varying. For functions belonging to the two subclasses $L$ and $E$, our results have the following consequences, verifiable through elementary calculus, ###### Corollary 1.8. Let $f$ be an edge-step function. 1. (a) if $f$ belongs to $L$, with $\alpha\leq 1$, then ${\mathbb{P}}\left(\frac{1}{3}\frac{\log t}{\log\log t}\leq\mathrm{diam}(G_{t}(f))\leq\frac{8}{\alpha}\frac{\log t}{\log\log t}\right)=1-o(1);$ 2. (b) if $f$ belongs to $L$, with $\alpha>1$, then ${\mathbb{P}}\left(\frac{1}{3\alpha}\frac{\log t}{\log\log t}\leq\mathrm{diam}(G_{t}(f))\leq\frac{7}{\alpha-1}\frac{\log t}{\log\log t}\right)=1-o(1);$ 3. (c) if $f$ belongs to $E$, then ${\mathbb{P}}\left(C_{\alpha}^{-1}(\log t)^{1-\alpha}\leq\mathrm{diam}(G_{t}(f))\leq C_{\alpha}(\log t)^{1-\alpha}\right)=1-o(1),$ for some $C_{\alpha}\geq 1$. ### 1.4. Comparing $G_{t}(f)$ and $G_{t}(h)$ If two edge-step functions are “close” to each other in some sense, then one should expect that the random processes they generate should be “close” as well. We make precise this intuition in the theorem below. In words, if $f$ and $h$ are close in the $L_{1}(\mathbb{N})$-norm (denoted by $\|\cdot\|_{1}$), then $\mathrm{Law}(\\{G_{t}(f)\\}_{t\geq 1})$ is close to $\mathrm{Law}(\\{G_{t}(h)\\}_{t\geq 1})$ in the total variation distance, denoted by $\mathrm{dist}_{TV}(\cdot,\cdot)$. ###### Theorem 5. Consider $f$ and $h$ two edge-step functions. We have (1.9) $\mathrm{dist}_{TV}\left(\mathrm{Law}(\\{G_{t}(f)\\}_{t\geq 1}),\mathrm{Law}(\\{G_{t}(h)\\}_{t\geq 1})\right)\leq\|f-h\|_{1}.$ In particular, if $(f_{n})_{n\in\mathbb{N}}$ is a sequence of edge-step functions, then (1.10) $f_{n}\stackrel{{\scriptstyle L_{1}(\mathbb{N})}}{{\longrightarrow}}f\implies\mathcal{L}((G_{t}(f_{n}))_{t\geq 1})\stackrel{{\scriptstyle\mathrm{dist}_{TV}}}{{\longrightarrow}}\mathcal{L}((G_{t}(f))_{t\geq 1}).$ The above theorem may be read as a perturbative statement. It assures that we may add a small noise $\epsilon=\epsilon(t)$ to an edge-step function $f$ and still obtain the same process up to an error of at most $\|\epsilon\|_{1}$ in total variation distance. ### 1.5. Main technical ideas In order to prove the existence of some given subgraph in the (affine) BA- random graphs a key ingredient is usually to use the fact that two given vertices $v_{i}$ and $v_{j}$ may be connected only at one specific time-step, since (assuming $i<j$) the model’s dynamic only allows $v_{j}$ to connect to $v_{i}$ at the moment in which $v_{j}$ is created. This property facilitates the computation of the probability of the occurrence of a given subgraph and decreases the combinatorial complexity of the arguments. In [9, 16] the authors estimate the number of triangles and cherries (paths of length $3$) on the (affine) BA-model and their argument relies heavily on this feature of the model. In our case, however, the edge-step prevents an application of such arguments, since a specific subgraph may appear at any time after the vertices have been added. Another difficulty in our setup is the degree of generality we work with. Our case replaces the parameter $p\in(0,1]$ in the models investigated in [1, 12, 13] by any non-negative real function $f$ with $\|f\|_{\infty}\leq 1$. The introduction of such function naturally increases the complexity of any analytical argument one may expect to rely on and makes it harder to discover threshold phenomena. This is the reason why in our work the Karamata’s Theory of regularly varying functions is crucial in order to prove sharper results. In order to overcome the issues presented above, more specifically to prove Theorem 1 and Theorem 5, we construct an auxiliary process that we call the doubly-labeled random tree process, $\\{\mathcal{T}_{t}\\}_{t\geq 1}$. In essence, this process is a realization of the traditional BA-model (obtained in our settings choosing $f\equiv 1$) where each vertex has two labels attached to it. We then show in Proposition 2.1 how to generate $G_{t}(f)$ from $\mathcal{T}_{t}$ using the information on those labels. This procedure allows us to generate $G_{t}(f)$ and $G_{t}(h)$, for two distinct functions $f$ and $h$, from exactly the same source of randomness. The upshot is that an edge-step function may be seen as a map from the space of doubly-labeled trees to the space of (multi)graphs – therefore it makes sense to use the notation $f(\mathcal{T}_{t})$. Furthermore, this map has the crucial property of being monotonic (in a way we make precise latter). Roughly speaking, if $f\leq h$, then certain monotonic graph observables respect this order, so if $\zeta$ is such an observable than $\zeta(f(\mathcal{T}_{t}))\leq\zeta(h(\mathcal{T}_{t}))$. In Proposition 2.5 we give important examples of suitable monotonic graph observables, the diameter being one of them. Our machinery then allows us to transpose some results about graphs generated for functions in a particular regime to another just by comparing the functions themselves. We use known results about cliques when $f$ is taken to be a constant less than one to propagate this result down other regimes of functions. For the proof of Theorem 2, we apply the second moment method on the number of isolated paths. This approach demands correlation estimations for the existence of two such paths in $G_{t}(f)$, which we do only under the assumption of $f$ being monotonic. For Theorem 3, we apply a lower bound for the degree of earlier vertices which is obtained by estimation of negative moments of a given vertex’s degree, and then show that, under conditions (S) or (Lκ), long paths of younger vertices are unlikely and older vertices are all very close in graph distance. Finally, using results from the Karamata’s theory of regularly varying functions, we verify that this broad class of functions satisfies our assumptions, proving Theorem 4. ### 1.6. Organization In Section 2 we introduce the doubly-labeled random tree process and prove the main results about it and as a consequence of this results, we obtain Theorem 5. In Section 3 we explore the theory developed in the previous section to prove the existence of large cliques, i.e., Theorem 1. Then, in Section 4, we prove technical estimates for the degree of a given vertex, which is needed for the upper bound on the diameter. Section 5 is devoted to the general lower bound for the diameter, i.e., for the proof of Theorem 2. We prove the upper bound for the diameter, Theorem 3, in Section 6. Finally, in Section 7 we show how our results fit on the class of the regularly varying functions and subclasses of slowly varying functions. We end the paper at Section 8 with some comments on the affine version of our model and a brief discussion on what may happen to the model if some regularity conditions are dropped. ### 1.7. Notation We let $V(G_{t}(f))$ and $E(G_{t}(f))$ denote the set of vertices and edges of $G_{t}(f)$, respectively. Given a vertex $v\in V(G_{t}(f))$, we will denote by $D_{t}(v)$ its degree in $G_{t}(f)$. We will also denote by $\Delta D_{t}(v)$ the _increment_ of the discrete function $D_{t}(v)$ between times $t$ and $t+1$, that is, $\Delta D_{t}(v)=D_{t+1}(v)-D_{t}(v).$ When necessary in the context, we may use $D_{G}(v)$ to denote the degree of $v$ in the graph $G$. Given two sets $A,B\subseteq V(G_{t}(f))$, we let $\\{A\leftrightarrow B\\}$ denote the event where there exists an edge connecting a vertex from $A$ to a vertex from $B$. We denote the complement of this event by $\\{A\nleftrightarrow B\\}$. We let $\mathrm{dist}(A,B)$ denote the graph distance between $A$ and $B$, i.e. the minimum number of edges that a path that connects $A$ to $B$ must have. When one of these subsets consists of a single vertex, i.e. $A=\\{v\\}$, we drop the brackets from the definition and use $\\{v\leftrightarrow B\\}$ and $\mathrm{dist}(v,B)$, respectively. For $t\in\mathbb{N}$, we let $[t]$ denote the set $\\{1,\dots,t\\}$. Regarding constants, we let $C_{1},C_{2},\dots$ and $c,c_{1},c_{2},\dots$ be positive real numbers that do not depend on $t$ whose values may vary in different parts of the paper. The dependence on other parameters will be highlighted throughout the text. Since our model is inductive, we use the notation $\mathcal{F}_{t}$ to denote the $\sigma$-algebra generated by all the random choices made up to time $t$. In this way we obtain the natural filtration $\mathcal{F}_{0}\subset\mathcal{F}_{1}\subset\dots$ associated to the process. ## 2\. The doubly-labeled random tree process In this section we introduce a stochastic process $\\{\mathcal{T}_{t}\\}_{t\geq 1}$ that provides a grand coupling between the random graphs $\\{G_{t}(f)\\}_{t\geq 1}$ for every edge-step function $f$. The process $\\{\mathcal{T}_{t}\\}_{t\geq 1}$ is essentially a realization of the Barabási-Albert random tree where each vertex has two labels: an earlier vertex chosen according to the preferential attachment rule and an independent uniform random variable. The label consisting in the earlier vertex can be seen as a “ghost directed edge”, we later use these random labels to collapse subsets of vertices into a single vertex in order to obtain a graph with the same distribution as $G_{t}(f)$ for any prescribed function $f:\mathbb{N}\to[0,1]$. We begin our process with a graph $\mathcal{T}_{1}$ consisting as usual in a single vertex and a single loop connecting said vertex to itself. We then inductively construct the labeled graph $\mathcal{T}_{t+1}$ from $\mathcal{T}_{t}$ in the following way: Figure 1. A sample of the process $\\{\mathcal{T}_{t}\\}_{t\geq 1}$ up to time $6$ without the uniform labels. The dashed lines indicate the label $\ell(v_{j})$ taken by each vertex $v_{j}$. * (i) We add to $\mathcal{T}_{t}$ a vertex $v_{t+1}$; * (ii) We tag $v_{t+1}$ with a random label $\ell(v_{t+1})$ chosen from the set $V(\mathcal{T}_{t})$ with the preferential attachment rule, that is, with probability of choosing $u\in V(\mathcal{T}_{t})$ proportional to the degree of $u$ in $\mathcal{T}_{t}$; * (iii) Independently from the step above, we add an edge $\\{w,v_{t+1}\\}$ to $E(\mathcal{T}_{t})$ where $w\in V(\mathcal{T}_{t})$ is also randomly chosen according to the preferential attachment rule, see Figure 1. We then finish the construction by tagging each vertex $v_{j}\in V(\mathcal{T}_{t})$ with a second label consisting in an independent random variable $U_{j}$ with uniform distribution on the interval $[0,1]$, as shown in Figure 2. We note that only the actual edges contribute to the degree taken in consideration in the preferential attachment rule, the tags are not considered. Figure 2. The graph $\mathcal{T}_{6}$, now each vertex receives an independent uniform random variable. We now make precise the notation $f(\mathcal{T}_{t})$ which indicates that an edge-step function $f$ may also be seen as a function that maps a doubly- labeled tree to a (multi)graph. Our goal is to define this map in such way that $f(\mathcal{T}_{t})\stackrel{{\scriptstyle d}}{{=}}G_{t}(f)$. In order to do so, let us fix such a function $f$. Given $v_{j}\in V(\mathcal{T}_{t})$, we compare $U_{j}$ to $f(j)$. If $U_{j}\leq f(j)$, we do nothing. Otherwise, we collapse $v_{j}$ onto its label $\ell(v_{j})$, that is, we consider the set $\\{v_{j},\ell(v_{j})\\}$ to be a single vertex with the same labels as $\ell(v_{j})$. We then update the label of all vertices $v$ such that $\ell(v)=v_{j}$ to $\\{v_{j},\ell(v_{j})\\}$. This procedure is associative in the sense that the order of the vertices on which we perform this operation does not affect the final resulting graph, as long as we perform it for all the vertices of $\mathcal{T}_{t}$. We let then $f(\mathcal{T}_{t})$ be the (multi)graph obtained when this procedure has run over all the $t$ vertices of $\mathcal{T}_{t}$. We refer to Figure 3 as an illustration of the final outcome. In the next proposition we prove that $f(\mathcal{T}_{t})$ is indeed distributed as $G_{t}(f)$. Figure 3. The figure shows how one can sample the distribution of $G_{6}(f)$ using the labeled graph $\mathcal{T}_{6}$. ###### Proposition 2.1. Let $\mathcal{T}_{t}$ be the doubly-labeled tree above defined and $f$ an edge-step function. Then, $f(\mathcal{T}_{t})\stackrel{{\scriptstyle d}}{{=}}G_{t}(f).$ ###### Proof. We first observe that the associativity of the collapsing operation and the independence of the sequence $(U_{j})_{j\geq 1}$ from the previous operations imply that we can glue together the vertex $v_{t+1}$ to $\ell(v_{t+1})$ whenever $U_{t+1}>f(t+1)$ right after we complete step (iii) of the above construction by induction. The resulting graph has either a new vertex $v_{t+1}$ with an edge $\\{w(v_{t+1}),v_{t+1}\\}$ or an edge $\\{w(v_{t+1}),\ell(v_{t+1})\\}$ with the exact same probability distribution as the $(t+1)$-th step in the construction of the graph $(G_{t}(f))_{t\geq 1}$. By induction, both random graphs have the same distribution. In the light of the above discussion and Proposition 2.1, from now on, we will tacitly assume that all process $\\{G_{t}(f)\\}_{t\geq 1}$ for all edge-step functions are on the same probability space provided by our previous results. A straightforward consequence of Proposition 2.1 is Theorem 5 which roughly speaking states that edge-step functions close to each other in the $L_{1}(\mathbb{N})$-norm generates essentially the same processes. Once we have the machinery provided by Proposition 2.1, the proof of this fact becomes a simple application of the union bound. ###### Proof of Theorem 5. Fixed two edge-step functions $f$ and $h$, by Proposition 2.1 we have that (2.2) $\mathrm{Law}(\\{G_{t}(f)\\}_{t\geq 1})=\mathrm{Law}(\\{f(\mathcal{T}_{t})\\}_{t\geq 1}).$ Thus (2.3) $\begin{split}\mathrm{dist}_{TV}\left(\mathrm{Law}(\\{G_{t}(f)\\}_{t\geq 1}),\mathrm{Law}(\\{G_{t}(h)\\}_{t\geq 1})\right)&\leq{\mathbb{P}}\left(\\{f(\mathcal{T}_{t})\\}_{t\geq 1}\neq\\{h(\mathcal{T}_{t})\\}_{t\geq 1}\right)\\\ &\leq\sum_{i=1}^{\infty}{\mathbb{P}}\left(U_{i}\in(f(i)\wedge h(i),f(i)\vee h(i))\right)\\\ &=\sum_{i=1}^{\infty}|f(i)-g(i)|\\\ &=\|f-g\|_{1},\end{split}$ and the above equation immediately implies (1.10). The next step is to use the coupling provided by the auxiliary process $\\{\mathcal{T}_{t}\\}_{t\geq 1}$ to compare graph observable of graphs generated by different edge-step functions. This method will allow us to transport results we have obtained for a fixed $f$ to other edge-step functions by comparing them with $f$. To do this we introduce the notion of increasing (decreasing) graph observable. ###### Definition 1 (Monotone graph observable). We say that a (multi)graph observable $\zeta$ is increasing if, given two edge-step functions $f$ and $h$, we have that (2.4) $f(s)\leq h(s),\forall s\leq t\implies\zeta(f(\mathcal{T}_{t}))\leq\zeta(h(\mathcal{T}_{t}))\;\textit{a.s.}$ When the second inequality in (2.4) holds with “$\geq$”, we say $\zeta$ is decreasing. A first example of an increasing observable is the total number of vertices. Indeed, since every vertex $v_{j}$ of $\mathcal{T}_{t}$ remain preserved under $f$ whenever its $U_{j}$ label is less than $f(j)$, it follows that every $f$-preserved $v_{j}$ is $h$-preserved as well. The next proposition states that the maximum degree is a decreasing observable whereas the diameter, in which we are interested in, is an increasing one. ###### Proposition 2.5. Let $f$ and $h$ be two edge-step functions satisfying the relation $f(s)\leq h(s)$ for all $s\in(1,\infty)$. Then, 1. (a) the maximum degree is a _decreasing_ graph observable. More precisely, if we let $D_{\textit{max}}(G)$ denote the maximum degree of a (multi)graph $G$, then ${\mathbb{P}}\left(\forall t\in\mathbb{N},\;D_{\textit{max}}(G_{t}(f))\geq D_{\textit{max}}(G_{t}(h))\right)=1;$ 2. (b) the diameter is an _increasing_ graph observable. More precisely, ${\mathbb{P}}\left(\forall t\in\mathbb{N},\;\mathrm{diam}(G_{t}(f))\leq\mathrm{diam}(G_{t}(h))\right)=1$ ###### Proof. The proofs follow from an analysis of the action of $f$ over $\mathcal{T}_{t}$. Proof of part (a): We first point out that if $v_{s}$ is a vertex of $\mathcal{T}_{t}$ whose $U_{s}$-label is less than $f(s)$, then (2.6) $D_{G_{t}(f)}(v_{s})\geq D_{G_{t}(h)}(v_{s}).$ To see why the above inequality is true, notice that the degree of $v_{s}$ in $G_{t}(f)$ is the total number of solid edges (see Figure 2) incident on $v_{s}$ in $\mathcal{T}_{t}$ plus the total number of vertices whose $\ell$-label points to $v_{s}$ and whose $U$-label is greater than $f$. Formally, (2.7) $D_{G_{t}(f)}(v_{s})=\sum_{r=1}^{t}\mathbb{1}\left\\{v_{r}\to v_{s}\text{ in }\mathcal{T}_{t}\right\\}+\mathbb{1}\left\\{\ell(v_{r})=v_{s},U_{r}\geq f(r)\right\\}.$ The contribution of the first term in the RHS of the above identity remains stable under $f$ and $h$, however, being $f(r)\leq h(r)$ for all $r\leq t$, we automatically have $\mathbb{1}\left\\{\ell(v_{r})=v_{s},U_{r}\geq f(r)\right\\}\geq\mathbb{1}\left\\{\ell(v_{r})=v_{s},U_{r}\geq h(r)\right\\},$ which implies that $D_{G_{t}(f)}(v_{s})\geq D_{G_{t}(h)}(v_{s})$. To finish the proof of this part, notice that, if $v_{s}$ is truly a vertex in $G_{t}(h)$ but is glued to its $\ell$-label under $f$, this operation just increases the maximum degree in the sense that the degree of $\ell(v_{s})$ inherits all the contributions of $v_{s}$. Thus, if $v_{\textit{max}}$ is a vertex that achieves $D_{\textit{max}}(G_{t}(h))$, then if $v_{\textit{max}}$ is also a vertex of $G_{t}(f)$ by the above discussion, its degree is at least $D_{\textit{max}}(G_{t}(h))$. On the other hand, if it is identified to $\ell(v_{\textit{max}})$, then the degree of $\ell(v_{\textit{max}})$ is also at least the maximum degree of $G_{t}(h)$. Proof of part (b): Observe that if $u$ and $w$ are vertices whose $U$-label is less than $f$ and such that $u\leftrightarrow w$ in $G_{t}(f)$, then $u$ and $w$ are possibly not connected by a single edge in $G_{t}(h)$, in other words, if $\mathrm{dist}_{G_{t}(f)}(u,w)=1$ then $\mathrm{dist}_{G_{t}(h)}(u,w)\geq 1$. Therefore, if $v_{0}=u,v_{1},\cdots,v_{k}=w$ is a minimal path in $G_{t}(f)$, by the previous observation, this path induces a path in $G_{t}(h)$ whose length is equal to or greater than $k$. On the other hand, if $u=v_{0},v_{1},\cdots,v_{k}=w$ is a path realizing the minimal distance between $u$ and $w$, in $G_{t}(h)$ we have that this path induces another path in $G_{t}(f)$ of same length or less. To see this, just notice that, if $U_{j}>f(j)$ and $\ell(v_{j})$ points to a vertex outside the path for some $j\in\\{2,3,\cdots,k-1\\}$, then, after identifying $v_{j}$ to its $\ell$-label, this operation does not increase the path length. This is enough to conclude the proof. ## 3\. Large Cliques: proof of Theorem 1 In this section we keep exploring the machinery developed in previous sections. Here, we transpose the existence of large cliques using the edge- step function, i.e., if $h$ is such that $G_{t}(h)$ has a clique of order $K$ and $f$ another function satisfying $f(s)\leq h(s)$ for all $s\in(1,\infty)$, then the clique existence propagates to $G_{t}(f)$ too but with a possibly $f$-dependent order. Our strategy to prove Theorem 1 will be to apply the results obtained for the function identically equal to $p$, with $p\in(0,1)$ (which we denote by $\mathrm{cons}_{p}$) in [1] and propagate them to smaller functions. For the sake of the reader’s convenience, we state and comment the aforementioned results here. Given $m\in\mathbb{Z}_{+}$, we order the vertices of $G_{t}(\mathrm{cons}_{p})$ from oldest to earliest and then divide them into blocks of size $m$. We then denote by $d_{t,m}(j)$ the sum of all degrees of the vertices from the $j$-th block. The following theorem has a very involved notation, but in essence it provides with high probability an explicit polynomial lower bound for $d_{t,m}(j)$. ###### Theorem (Theorem 2 from [1] ). Given $p\in(0,1]$, let $\xi\in(0,2(2-p)^{-1}-1)$ and fix $m\in\mathbb{Z}_{+}$ sufficiently large. Define $\zeta_{m}:=\frac{(1-p)}{2(2-p)m^{\xi}}.$ and let $1<R<m(1-p/2)(1-\zeta_{m}).$ There exists a positive constant $c=c(m,R,p)$ such that, for $\beta\in(0,(1-p/2)(1-\zeta_{m}))\textit{ and }j\geq m^{\frac{2}{1-p}}+1,$ we have (3.1) $\mathbb{P}\left(d_{t,m}(j)<t^{\beta}\right)\leq c\frac{j^{R}}{t^{R-\beta R(1-p/2)^{-1}(1-\zeta_{m})^{-1}}}+\frac{m}{[(j-1)m]^{99}}.$ Using the above bound one can then prove ###### Theorem (Theorem $1$ from [1]). For any $\varepsilon>0$ and every $p\in(0,1)$, it follows that (3.2) ${\mathbb{P}}\left(\exists K_{n}\subset G_{t}(\mathrm{cons}_{p}),\textit{ such that }n=t^{(1-\varepsilon)\frac{(1-p)}{2-p}}\right)=1-o(1).$ We can now use the doubly-labeled tree together with the above results in order to prove Theorem 1. It will be useful to recall our special notation to the expected number of vertices, (3.3) $F(t):={\mathbb{E}}V(G_{t}(f))=1+\sum_{s=2}^{t}f(s).$ ###### Proof of Theorem 1. Given $\delta>0$ sufficiently small, let $\varepsilon,p=\delta/6$ so that $\alpha=\alpha(p,\varepsilon):=\frac{1-p}{2-p}(1-\varepsilon)>\frac{1}{2}\left(1-\frac{\delta}{2}\right).$ In the proof of Theorem $1$ of [1] one uses Theorem $2$ of [1] to show that there exist a fixed integer $m=m(\varepsilon,p)>0$ and a small number $\varepsilon^{\prime}\in(0,\alpha)$ with the following property: if one divides the set of vertices born between times $t^{\varepsilon^{\prime}}$ and $t^{\alpha}$ into disjoint subsets of $m$ vertices born _consecutively_ , then with high probability (at least $1$ minus a polynomial function of $t$) one can choose a vertex from each of these subsets in such a way that the subgraph induced by the set of chosen vertices is a complete subgraph of $G_{t}(\mathrm{cons}_{p})$. Having the above in mind and letting $m$ be the auxiliary $m$ that appears in proof of Theorem 1 of [1], define the following event (3.4) $A_{p,\varepsilon^{\prime},m,k}:=\left\\{\begin{array}[]{c}\exists\\{t_{1},\cdots,t_{k}\\}\subset[t]\text{ such that }t_{j}\in(t^{\varepsilon^{\prime}}+(j-1)m,t^{\varepsilon^{\prime}}+jm]\\\ \text{and }v_{t_{j}}\leftrightarrow v_{t_{i}}\textit{ in }G_{t}(\mathrm{cons}_{p})\text{ for all }i,j\in[k]\end{array}\right\\}.$ In the context of the doubly labeled tree process, the event $A_{p,\varepsilon^{\prime},m,k}$ says that we may find $k$ vertices of $\mathcal{T}_{t}$ with the property that the $j$-th vertex $v_{t_{j}}$ was added by the doubly-labeled tree process sometime in the interval $(t^{\varepsilon^{\prime}}+(j-1)m,t^{\varepsilon^{\prime}}+jm]$ and its $U_{t_{j}}$-label is less than $p$. Moreover, when we apply $\mathrm{cons}_{p}$ on $\mathcal{T}_{t}$ in the event $A_{p,\varepsilon^{\prime},m,k}$, all these $k$ vertices form a complete graph in the resulting graph $G_{t}(\mathrm{cons}_{p})$. Theorem $1$ of [1] states that setting $k=t^{\alpha}-t^{\varepsilon^{\prime}}$, the event $A_{p,\varepsilon^{\prime},m,k}$ occurs with probability at least $1-t^{-\eta}$, for some positive small $\eta$ depending on $p$, which, in our case, is a function of $\delta$. Thus, for $t$ large enough, we may simply use that (3.5) ${\mathbb{P}}\left(A_{p,\varepsilon^{\prime},m,t^{\alpha}}\right)\geq 1-\frac{1}{\log t},$ this bound will be useful for the proof of Corollary 1.4. This is why we are exchanging a polynomial decay by a $\log$ one, to get rid off the dependency on $\delta$ and consequently simplify latter arguments. Now, let $t$ be large enough so that $f(t^{\varepsilon^{\prime}})<p$. This is possible since $f$ decreases to zero. Also notice that if $u$ and $v$ are vertices of $G_{t}(f)$ added after time $t^{\varepsilon^{\prime}}$ and are connected in $G_{t}(\mathrm{cons}_{p})$, then they are connected in $G_{t}(f)$ as well. Moreover, since $f$ is non-increasing and the $U$-labels are assigned independently and according to a uniform distribution on $[0,1]$, we have for all $j\leq k$ (3.6) ${\mathbb{P}}\left(U_{t_{j}}\leq f(t_{j})\;\middle|\;A_{p,\varepsilon^{\prime},m,k}\right)\geq p^{-1}f(t^{\varepsilon^{\prime}}+jm).$ Setting $k=(t^{\alpha}-t^{\varepsilon^{\prime}})/m$ and using that $f$ is non- increasing, we have that (3.7) $m\sum_{j=0}^{k}f(t^{\varepsilon^{\prime}}+jm)\geq F(t^{\alpha})-F(t^{\varepsilon^{\prime}}+m)\geq F(t^{\alpha})/2$ for large enough $t$. By the independence of the $U$-labels, it follows that conditioned on $A_{p,\varepsilon^{\prime},m,k}$, the random variable that counts how many vertices of the clique in $G_{t}(\mathrm{cons}_{p})$ remain vertices of $G_{t}(f)$ as well is a sum of $k$ independent random variables taking values on $\\{0,1\\}$ and whose expected value is greater than $F(t^{\alpha})/2mp$. Finally, by Chernoff bounds, this random variable is at least half its expected value with probability at least $1-\exp\\{-F(t^{\alpha})/16mp\\}$. This and (3.5) gives us that there exists a constant $C_{\delta}$ such that (3.8) ${\mathbb{P}}\left(\nexists K_{n}\subset G_{t}(f),\textit{ such that }n\geq C_{\delta}F(t^{\frac{1-\delta}{2}})\right)\leq\frac{1}{\log t}+e^{-F(t^{\alpha})/16mp},$ which proves the Theorem. ## 4\. Technical estimates for the degree In this section we develop technical estimates related to the degree of a given vertex. We begin by stating one of the most fundamental identities in the study of preferential attachment models: the conditional distribution of the increment of the degree of a given vertex. Given $v\in V(G_{t}(f))$, we have (4.1) $\begin{split}\mathbb{P}\left(\Delta D_{t}(v)=0\middle|\mathcal{F}_{t}\right)&=f(t+1)\left(1-\frac{D_{t}(v)}{2t}\right)+(1-f(t+1))\left(1-\frac{D_{t}(v)}{2t}\right)^{2}\\\ \mathbb{P}\left(\Delta D_{t}(v)=1\middle|\mathcal{F}_{t}\right)&=f(t+1)\frac{D_{t}(v)}{2t}+2(1-f(t+1))\frac{D_{t}(v)}{2t}\left(1-\frac{D_{t}(v)}{2t}\right),\\\ \mathbb{P}\left(\Delta D_{t}(v)=2\middle|\mathcal{F}_{t}\right)&=(1-f(t+1))\frac{D_{t}(v)^{2}}{4t^{2}}.\end{split}$ To see why the above identities hold true, observe for example that in order for $\Delta D_{t}(v)=0$, either a vertex step was taken, and the vertex did not connect to $v$, or an edge step was taken and neither of the endpoints of the new edge connected to $v$. The other equations follow from analogous reasonings. As a direct consequence, we obtain (4.2) $\begin{split}{\mathbb{E}}\left[\Delta D_{t}(v)\middle|\mathcal{F}_{t}\right]&=1\cdot f(t+1)\cdot\frac{D_{t}(v)}{2t}+1\cdot 2(1-f(t+1))\frac{D_{t}(v)}{2t}\left(1-\frac{D_{t}(v)}{2t}\right)\\\ &\quad+2\cdot(1-f(t+1))\frac{D^{2}_{t}(v)}{4t^{2}}\\\ &=\left(1-\frac{f(t+1)}{2}\right)\frac{D_{t}(v)}{t}.\end{split}$ Using the above equation repeatedly one obtains, conditioned on the event where the vertex $v$ is born at time $t_{0}$, (4.3) $\begin{split}{\mathbb{E}}\left[D_{t}(v)\right]&={\mathbb{E}}\left[{\mathbb{E}}\left[D_{t}(v)\middle|\mathcal{F}_{t-1}\right]\right]\\\ &=\left(1+\frac{1}{t-1}-\frac{f(t)}{2(t-1)}\right){\mathbb{E}}\left[D_{t-1}(v)\right]\\\ &=\prod_{s=t_{0}}^{t-1}\left(1+\frac{1}{s}-\frac{f(s+1)}{2s}\right).\end{split}$ We state a lower bound for the degree, whose proof comes from a direct application of Theorem $2$ of [1] and Proposition 2.5, which assures that the maximum degree is a decreasing graph observable. ###### Lemma 1. Let $f$ be an edge-step function such that $f(t)\rightarrow 0$ as $t$ goes to infinity. Then, for any fixed $\varepsilon>0$ we have ${\mathbb{P}}\left(D_{\textit{max}}\left(G_{t}(f)\right)<t^{1-\varepsilon}\right)\leq t^{-2}.$ ###### Proof. By Theorem $2$ of [1], w.h.p, $D_{\text{max}}(G_{t}(p))$ is at least $t^{(1-\varepsilon)(1-p/2)}$, for fixed $\varepsilon>0$ and $p\in(0,1)$. Thus, taking $p_{0}$ small enough, we have that $D_{\text{max}}(G_{t}(p_{0}))$ is at least $t^{1-\varepsilon}$, w.h.p. Then, using Proposition 2.5, we have that for large enough $t$, $D_{\text{max}}(G_{t}(f))\geq D_{\text{max}}(G_{t}(p_{0}))$ which proves the Lemma. Our objective now is to obtain a polynomial lower bound for the degree of older vertices, which will be important in the proof of the upper bound for the diameter in Theorem 3. We begin with an upper bound for the expectation of the multiplicative inverse of the degree. Recall the definition of the process $(Z_{t})_{t\geq 1}$, consisting of independent Bernoulli variables that dictate whether a vertex-step or an edge-step is performed at time $t$. ###### Lemma 2. Given any edge-step function $f$, consider the process $\\{G_{t}(f)\\}_{t\geq 1}$. Denote by $v_{i}$ the vertex born at time $i\in\mathbb{N}$. We have (4.4) ${\mathbb{E}}\left[(D_{t}(v_{i}))^{-1}Z_{i}\right]\leq f(i)\left(\frac{t-1}{i}\right)^{-\frac{1}{6}}.$ And consequently (4.5) ${\mathbb{P}}\left(D_{t}(v_{i})\leq\left(\frac{t-1}{i}\right)^{\frac{1}{12}}\middle|Z_{i}=1\right)\leq\left(\frac{t-1}{i}\right)^{-\frac{1}{12}}.$ ###### Proof. If $Z_{i}=1$, then for every $s\geq 1$ we have that $\Delta D_{s}(v_{i})\geq 0$, that $D_{s+1}(v_{i})\leq D_{s}(v_{i})+2\leq 3D_{s}(v_{i})$, and that $D_{s}(v_{i})$ is $\mathcal{F}_{s}$ measurable. Together with (4.2), these facts imply. on the event $\\{Z_{i}=1\\}$, (4.6) $\begin{split}{\mathbb{E}}\left[\frac{1}{D_{s+1}(v_{i})}-\frac{1}{D_{s}(v_{i})}\middle|\mathcal{F}_{s}\right]&={\mathbb{E}}\left[-\frac{\Delta D_{s}(v_{i})}{D_{s+1}(v_{i})D_{s}(v_{i})}\middle|\mathcal{F}_{s}\right]\\\ &\leq-\frac{1}{3(D_{s}(v_{i}))^{2}}\left(1-\frac{f(s+1)}{2}\right)\frac{D_{s}(v_{i})}{s}\\\ &\leq-\frac{1}{6s}\frac{1}{D_{s}(v_{i})},\end{split}$ since $f(k)\leq 1$ for very $k\in\mathbb{N}$. Therefore, (4.7) ${\mathbb{E}}\left[Z_{i}\cdot(D_{s+1}(v_{i}))^{-1}\middle|\mathcal{F}_{s}\right]\leq Z_{i}(1-(6s)^{-1})(D_{s}(v_{i}))^{-1}.$ Iterating the above argument from $i$ until $t$, we obtain (4.8) ${\mathbb{E}}\left[Z_{i}\cdot(D_{t}(v_{i}))^{-1}\right]\leq f(i)\prod_{s=i}^{t-1}\left(1-\frac{1}{6s}\right)\leq f(i)\exp\left\\{-\frac{1}{6}\sum_{s=i}^{t-1}\frac{1}{s}\right\\}\leq f(i)\left(\frac{t-1}{i}\right)^{-\frac{1}{6}},$ proving (4.4). Equation (4.5) is then obtained by an elementary application of the Markov inequality: (4.9) $\begin{split}{\mathbb{P}}\left(D_{t}(v_{i})\leq\left(\frac{t-1}{i}\right)^{\frac{1}{12}}\middle|Z_{i}=1\right)&={\mathbb{P}}\left((D_{t}(v_{i}))^{-1}\geq\left(\frac{t-1}{i}\right)^{-\frac{1}{12}}\middle|Z_{i}=1\right)\\\ &\leq\left(\frac{t-1}{i}\right)^{\frac{1}{12}}{\mathbb{E}}\left[(D_{t}(v_{i}))^{-1}\middle|Z_{i}=1\right]\\\ &\leq\left(\frac{t-1}{i}\right)^{-\frac{1}{12}}.\end{split}$ We now provide an elementary consequence of the above result, which uses the union bound in order to show that, with high probability, every vertex born before time $t^{\frac{1}{12}}$ has degree at least $t^{\frac{1}{15}}$ by time $t$. ###### Lemma 3. Using the same notation as in Lemma 2 we have, for every edge-step function $f$ and for sufficiently large $t\in\mathbb{N}$, (4.10) ${\mathbb{P}}\left(\exists i\in\mathbb{N},1\leq i\leq t^{\frac{1}{12}},\text{ such that }Z_{i}=1\text{ and }D_{t}(v_{i})\leq t^{\frac{1}{15}}\right)\leq Ct^{-\frac{1}{144}}.$ ###### Proof. By the union bound and equation (4.5), we have that the probability in the left hand side of (4.10) is smaller than or equal to (4.11) $\begin{split}\sum_{i=1}^{t^{\frac{1}{12}}}{\mathbb{P}}\left(Z_{i}=1,D_{t}(v_{i})\leq t^{\frac{1}{15}}\right)&\leq\sum_{i=1}^{t^{\frac{1}{12}}}f(i){\mathbb{P}}\left(D_{t}(v_{i})\leq\left(\frac{t-1}{i}\right)^{\frac{1}{12}}\middle|Z_{i}=1\right)\\\ &\leq\sum_{i=1}^{t^{\frac{1}{12}}}\left(\frac{t-1}{i}\right)^{-\frac{1}{12}}\\\ &\leq Ct^{-\frac{1}{12}+\frac{1}{12}\left(1-\frac{1}{12}\right)}\\\ &\leq Ct^{-\frac{1}{144}},\end{split}$ finishing the proof of the Lemma. ## 5\. General lower bound for the diameter: proof of Theorem 2 The proof of Theorem 2 follows a second moment argument. The idea is to count the number of “long” (the specific size depending on $f$ and $t$) _isolated paths_ in $G_{t}(f)$. We begin by showing that the expected number of isolated paths goes to infinity with $t$ in Lemma 4. We then show in Lemma 5 that the presence of a specific isolated path is almost independent from the presence of some other given isolated path whenever said paths are disjoint. This “almost independence” makes the second moment of the number of such paths very close to the first moment squared. The proof is then completed via the Payley- Zygmund inequality. We start by defining precisely what we mean by an isolated path. ###### Definition 2 (Isolated path). Let $l$ be a positive integer. Let $\vec{t}=(t_{1},..,t_{l})$ be a vector of distinct positive integers. We say that this vector corresponds to an isolated path $\\{v_{t_{1}},\dots,v_{t_{l}}\\}$ in $G_{t}(f)$ if and only if: * • $t_{l}\leq t$; * • $t_{i}<t_{j}$ whenever $1\leq i<j\leq l$; * • during each time $t_{i}$, $i=1,\dots,l$, a vertex-step is performed; * • for every integer $k\leq l$, the subgraph induced by the vertices $\\{v_{t_{i}}\\}_{1\leq i\leq k}$ is connected in $G_{t_{k}}(f)$; * • for $i=1,\dots,l-1$, the degree of $v_{t_{i}}$ in $G_{t}(f)$ is $2$. The degree of $v_{t_{l}}$ in $G_{t}(f)$ is $1$. In other words, an isolated path $\\{v_{t_{i}}\\}_{1\leq i\leq l}$ is a path where each vertex $v_{t_{i}}$, for $i=2,\dots,l$, is born at time $t_{i}$ and makes its first connection to its predecessor $v_{t_{i-1}}$. Other than that, no other vertex or edge gets attached to $\\{v_{t_{i}}\\}_{1\leq i\leq l}$. We will denote $\\{v_{t_{i}}\\}_{1\leq i\leq l}$ by $v_{\vec{t}}$. Given $\xi\in(0,1)$, denote by $\mathcal{S}_{l,\xi}(t)$ the set of all isolated paths in $G_{t}(f)$ of size $l$ whose vertices were created between times $\xi t$ and $t$. Our first goal is to obtain lower bounds for ${\mathbb{E}}\left[|\mathcal{S}_{l,\xi}(t)|\right]$: ###### Lemma 4. Let $f$ be a non-increasing edge-step function, then, for any $0<\xi<1$ and any integer $l$, the following lower bound holds: (5.1) ${\mathbb{E}}\left[|\mathcal{S}_{l,\xi}(t)|\right]\geq\binom{(1-\xi)t}{l}\frac{f(t)^{l}}{(2t)^{l-1}}\left(1-\frac{2l}{\xi t}\right)^{t}.$ Furthermore, for (5.2) $l\leq\frac{1}{3}\left(\frac{\log t}{\log\log t}\wedge\frac{\log t}{-\log f(t)}\right),$ we have that, for sufficiently large $t$, (5.3) ${\mathbb{E}}\left[|\mathcal{S}_{l,\xi}(t)|\right]\geq t^{\frac{1}{4}}.$ ###### Proof. The random variable $|\mathcal{S}_{l,\xi}(t)|$ can be written as (5.4) $|\mathcal{S}_{l,\xi}(t)|=\sum_{t_{1}<t_{2}<\dots<t_{l}}\mathbb{1}\left\\{v_{\vec{t}}\in\mathcal{S}_{l,\xi}(t)\right\\}\quad\implies\quad{\mathbb{E}}\left[|\mathcal{S}_{l,\xi}(t)|\right]=\sum_{t_{1}<t_{2}<\dots<t_{l}}\mathbb{P}\left(v_{\vec{t}}\in\mathcal{S}_{l,\xi}(t)\right).$ So it will be important to obtain a proper lower bound for $\mathbb{P}\left(v_{\vec{t}}\in\mathcal{S}_{l,\xi}(t)\right)$. Given a time vector of an isolated path $\vec{t}=(t_{1},\dots,t_{l})$ such that $\xi t\leq t_{1}<t_{2}<\dots<t_{l}\leq t$, it follows that (5.5) $\mathbb{P}\left(v_{\vec{t}}\in\mathcal{S}_{l,\xi}(t)\right)\geq\frac{f(t)^{l}}{(2t)^{l-1}}\left(1-\frac{2l}{\xi t}\right)^{t},$ since in order for $v_{\vec{t}}$ to be in $\mathcal{S}_{l,\xi}(t)$, we need to assure that $l$ vertices are born exactly at times $t_{1},\dots,t_{l}$ (which happens with probability greater than $f(t)^{l}$, by the monotonicity of $f$), that $v_{t_{i}}$ connects to $v_{t_{i-1}}$ for every $i=2,\dots,l$ (which happens with probability greater than $(2t)^{-(l-1)}$), and that no other vertex or edge connects to $v_{\vec{t}}$ until time $t$ (which happens with probability greater than $\left(1-\frac{2l}{\xi t}\right)^{t}$). Finally, by counting the number of possible ways to choose $t_{1}<t_{2}<\dots<t_{l}$ so that $t_{i}\in[\xi t,t]$ for $1\leq i\leq l$, we obtain 5.1. We now assume $l$ to be such that 5.2 holds. Stirling’s formula gives us $\begin{split}\log\left(\binom{(1-\xi)t}{l}\right)&\geq c+(1-\xi)t\log((1-\xi)t)-(1-\xi)t+\frac{\log((1-\xi)t)}{2}\\\ &\quad-l\log l+l-\frac{\log l}{2}\\\ &\quad-((1-\xi)t-l)\log((1-\xi)t-l)+(1-\xi)t-l-\frac{\log((1-\xi)t-l)}{2},\end{split}$ Since $l\ll t$, we obtain (5.6) $\log\left(\binom{(1-\xi)t}{l}\right)\geq l\log t-l\log l-cl.$ Equation (5.5) then implies, again for sufficiently large $t$, (5.7) $\log\left(\mathbb{P}\left(v_{\vec{t}}\in\mathcal{S}_{l,\xi}(t)\right)\right)\geq l\log f(t)-(l-1)\log t-cl.$ Combining the above inequality with (5.1), (5.2), and (5.6) gives us that, for large enough $t$, $\begin{split}{\mathbb{E}}\left[|\mathcal{S}_{l,\xi}(t)|\right]\geq\exp\left\\{l\log t-l\log l+l\log f(t)-(l-1)\log t-cl\right\\}\geq t^{\frac{1}{4}},\end{split}$ since $l\log l\leq\frac{1}{3}\frac{\log t}{\log\log t}\log\log t\left(1-\frac{\log\log\log t+\log 3}{\log\log t}\right)\leq\frac{\log t}{3},$ which finishes the proof of the lemma. Some new notation will be useful throughout the proof of Theorem 2: ###### Definition 3 (Degree of an isolated path). Given an isolated path $v_{\vec{t}}=\\{v_{t_{j}}\\}_{1\leq j\leq l}$, we denote by $D_{r}(v_{\vec{t}})$ the sum of the degrees of each of its vertices at time $r$, i.e. : $D_{r}(v_{\vec{t}})=\sum_{t_{i}\in\vec{t}}D_{r}(v_{t_{i}}),$ where we assumed that $D_{r}(v_{t_{i}})=0$ if $t_{i}<r$. Note that if $r>t_{l}$ and $v_{\vec{t}}$ has size $l$, then $D_{r}(v_{\vec{t}})=2l-1$. Furthermore, by the same reasoning as in (4.1), (5.8) $\begin{split}{\mathbb{P}}\left(\Delta D_{s}(v_{\vec{t}})=0~{}\middle|\mathcal{F}_{s}\right)&=f(s+1)\left(1-\frac{D_{s}(v_{\vec{t}})}{2s}\right)+(1-f(s+1))\left(1-\frac{D_{s}(v_{\vec{t}})}{2s}\right)^{2}\\\ &=1-\left(1-\frac{f(s+1)}{2}-(1-f(s+1))\frac{D_{s}(v_{\vec{t}})}{4s}\right)\frac{D_{s}(v_{\vec{t}})}{s}.\end{split}$ Paley-Zigmund’s inequality (see e.g. section $5.5$ of [22]) assures us that, for any $0\leq\theta\leq 1$, (5.9) $\mathbb{P}\left(|\mathcal{S}_{l,\xi}(t)|>\theta{\mathbb{E}}\left[|\mathcal{S}_{l,\xi}(t)|\right]\right)\geq(1-\theta)^{2}\frac{\left({\mathbb{E}}\left[|\mathcal{S}_{l,\xi}(t)|\right]\right)^{2}}{{\mathbb{E}}\left[|\mathcal{S}_{l,\xi}(t)|^{2}\right]}.$ If we are able to guarantee that 1. (i) ${\mathbb{E}}\left[|\mathcal{S}_{l,\xi}(t)|\right]\rightarrow\infty$; 2. (ii) $\left({\mathbb{E}}\left[|\mathcal{S}_{l,\xi}(t)|\right]\right)^{2}=(1-o(1)){\mathbb{E}}\left[|\mathcal{S}_{l,\xi}(t)|^{2}\right]$; then by choosing $\theta=\theta(t)={\mathbb{E}}\left[|\mathcal{S}_{l,\xi}(t)|\right]^{-1/2}$, we see that $\mathbb{P}\left(|\mathcal{S}_{l,\xi}(t)|>{\mathbb{E}}\left[|\mathcal{S}_{l,\xi}(t)|\right]^{1/2}\right)\geq\left(1-{\mathbb{E}}\left[|\mathcal{S}_{l,\xi}(t)|\right]^{-1/2}\right)^{2}(1-o(1))=1-o(1),$ thus guaranteeing with probability $1-o(1)$ the existence of many isolated paths of length $l$, finishing the proof of the theorem. By Lemma 4, we know that item (i) is true. Therefore from now on we will focus on proving item (ii). In order for the isolated path $v_{\vec{t}}$ to appear the following must happen: * • a vertex $v_{t_{1}}$ must be be created at time $t_{1}$, which happens with probability $f(t_{1})$; * • between times $t_{1}+1$ and $t_{2}-1$ there can be no new connection to $v_{t_{1}}$, which, by (5.8), happens with probability $\prod_{r_{1}=t_{1}+1}^{t_{2}-1}\left(1-\left(1-\frac{f(r_{1})}{2}-(1-f(r_{1}))\frac{1}{4(r_{1}-1)}\right)\frac{1}{r_{1}-1}\right);$ * • In general, at time $t_{k}$ a vertex $v_{t_{k}}$ is created and makes its first connection to $v_{t_{k-1}}$, no new connection is then made to $\\{v_{t_{j}}\\}_{1\leq j\leq k}$ between times $t_{k}+1$ and $t_{k+1}-1$ for every $k=2,\dots,l-1$, all this happens with probability equal to $\displaystyle f(t_{k})\frac{1}{2(t_{k}-1)}\prod_{r_{k}=t_{k}+1}^{t_{k+1}-1}\left(1-\left(1-\frac{f(r_{k})}{2}-(1-f(r_{k}))\frac{D_{r_{k}}(v_{\vec{t}})}{4(r_{k}-1)}\right)\frac{D_{r_{k}}(v_{\vec{t}})}{r_{k}-1}\right)$ $\displaystyle=f(t_{k})\frac{1}{2(t_{k}-1)}\prod_{r_{k}=t_{k}+1}^{t_{k+1}-1}\left(1-\left(1-\frac{f(r_{k})}{2}-(1-f(r_{k}))\frac{2k-1}{4(r_{k}-1)}\right)\frac{2k-1}{r_{k}-1}\right);$ * • finally, a vertex $v_{t_{l}}$ is born at time $t_{l}$, connects to $v_{t_{l-1}}$ and no new connection is made to $\\{v_{t_{j}}\\}_{1\leq j\leq l}$ between times $t_{l}+1$ and $t$. This implies (5.10) $\begin{split}\@ADDCLASS{ltx_eqn_lefteqn}$\displaystyle{\mathbb{P}}\left(v_{\vec{t}}\in\mathcal{S}_{l,\xi}(t)\right)$\mbox{}\hfil\\\ &=f(t_{1})\prod_{r_{1}=t_{1}+1}^{t_{2}-1}\left(1-\left(1-\frac{f(r_{1})}{2}-(1-f(r_{1}))\frac{1}{4(r_{1}-1)}\right)\frac{1}{r_{1}-1}\right)\\\ &\quad\times\dots\times f(t_{k})\frac{1}{2(t_{k}-1)}\prod_{r_{k}=t_{k}+1}^{t_{k+1}-1}\left(1-\left(1-\frac{f(r_{k})}{2}-(1-f(r_{k}))\frac{2k-1}{4(r_{k}-1)}\right)\frac{2k-1}{r_{k}-1}\right)\\\ &\quad\times\dots\times f(t_{l})\frac{1}{2(t_{l}-1)}\prod_{r_{l}=t_{l}+1}^{t}\left(1-\left(1-\frac{f(r_{l})}{2}-(1-f(r_{l}))\frac{2l-1}{4(r_{l}-1)}\right)\frac{2l-1}{r_{l}-1}\right).\end{split}$ Given two time vectors $\vec{r}$ and $\vec{t}$, we note that ${\mathbb{P}}\left(v_{\vec{t}},v_{\vec{r}}\in\mathcal{S}_{l,\xi}(t)\right)$ is only nonzero if $\vec{r}$ and $\vec{t}$ have either disjoint or identical sets of entries. Our focus now is on proving the following lemma: ###### Lemma 5. Let $l$ be such that (5.2) is satisfied. For two isolated paths with disjoint time vectors $\vec{t}$ and $\vec{r}$, we have (5.11) $\frac{{\mathbb{P}}\left(v_{\vec{t}}\in\mathcal{S}_{l,\xi}(t)\right){\mathbb{P}}\left(v_{\vec{r}}\in\mathcal{S}_{l,\xi}(t)\right)}{{\mathbb{P}}\left(v_{\vec{t}},v_{\vec{r}}\in\mathcal{S}_{l,\xi}(t)\right)}=1+o(1).$ ###### Proof. To prove the above result we will make a comparison between the two probabilities terms ${\mathbb{P}}\left(v_{\vec{t}},v_{\vec{r}}\in\mathcal{S}_{l,\xi}(t)\right)$ and ${\mathbb{P}}\left(v_{\vec{t}}\in\mathcal{S}_{l,\xi}(t)\right){\mathbb{P}}\left(v_{\vec{r}}\in\mathcal{S}_{l,\xi}(t)\right)$. We can write both these terms as products in the manner of (5.10). We can then compare the terms from these products associated to each time $s\in[\xi t,t]$. There are two cases we must study. Case 1: $s\in\vec{t}$ but $s\notin\vec{r}$ ($s\notin\vec{t}$ but $s\in\vec{r}$.). The product term related to time $s$ in ${\mathbb{P}}\left(v_{\vec{t}},v_{\vec{r}}\in\mathcal{S}_{l,\xi}(t)\right)$ is (5.12) $\frac{f(s)}{2(s-1)},$ since a new vertex is created and then makes its first connection specifically to the latest vertex of $\vec{t}$. On the other hand, the term related to time $s$ in ${\mathbb{P}}\left(v_{\vec{t}}\in\mathcal{S}_{l,\xi}(t)\right){\mathbb{P}}\left(v_{\vec{r}}\in\mathcal{S}_{l,\xi}(t)\right)$ is $\frac{f(s)}{2(s-1)}\left(1-\left(1-\frac{f(s)}{2}-(1-f(s))\frac{D_{s-1}(v_{\vec{r}})}{4(s-1)}\right)\frac{D_{s-1}(v_{\vec{r}})}{s-1}\right),$ since the term related to $s$ in the product form of ${\mathbb{P}}\left(v_{\vec{t}}\in\mathcal{S}_{l,\xi}(t)\right)$ continues to be equal to (5.12), but the related term in ${\mathbb{P}}\left(v_{\vec{r}}\in\mathcal{S}_{l,\xi}(t)\right)$ is (5.13) $\left(1-\left(1-\frac{f(s)}{2}-(1-f(s))\frac{D_{s-1}(v_{\vec{r}})}{4(s-1)}\right)\frac{D_{s-1}(v_{\vec{r}})}{s-1}\right).$ The above expression is the term that will appear regarding the time $s$ in the fraction in the left hand side of (5.11). This case occurs $2l$ times since the isolated paths are disjoint. Thus, recalling that $s\in[\xi t,t]$, $l\leq 3^{-1}\log(t)/\log(\log(t))$, and that the degree of each isolated path is at most $2l-1$, we obtain that there exist constants $c_{1},c_{2}>0$ such that we can bound the product of all the terms of the form (5.13) from above by $\left(1-\frac{c_{1}}{t}\right)^{2l},$ and from below by $\left(1-\frac{c_{2}l}{t}\right)^{2l}.$ Observe that both products go to $1$ as $t$ goes to infinity. Case 2: $s\notin\vec{t}$ and $s\notin\vec{r}$. In ${\mathbb{P}}\left(v_{\vec{r}}\in\mathcal{S}_{l,\xi}(t)\right)$ as well as in ${\mathbb{P}}\left(v_{\vec{t}}\in\mathcal{S}_{l,\xi}(t)\right)$ we see terms of the form (5.13), since we must avoid the isolated paths in both events. But in the term related to ${\mathbb{P}}\left(v_{\vec{t}},v_{\vec{r}}\in\mathcal{S}_{l,\xi}(t)\right)$ we actually observe $\left(1-\left(1-\frac{f(s)}{2}-(1-f(s))\frac{(D_{s-1}(v_{\vec{t}})+D_{s-1}(v_{\vec{r}}))}{4(s-1)}\right)\frac{(D_{s-1}(v_{\vec{t}})+D_{s-1}(v_{\vec{r}}))}{s-1}\right),$ since we must guarantee that neither isolated path receives a connection. We note however that $\displaystyle\left(1-\left(1-\frac{f(s)}{2}-(1-f(s))\frac{D_{s-1}(v_{\vec{r}})}{4(s-1)}\right)\frac{D_{s-1}(v_{\vec{r}})}{s-1}\right)$ $\displaystyle\phantom{**********************}\times\left(1-\left(1-\frac{f(s)}{2}-(1-f(s))\frac{D_{s-1}(v_{\vec{t}})}{4(s-1)}\right)\frac{D_{s-1}(v_{\vec{t}})}{s-1}\right)\quad$ $\displaystyle=\left(1-\left(1-\frac{f(s)}{2}-(1-f(s))\frac{(D_{s-1}(v_{\vec{t}})+D_{s-1}(v_{\vec{r}}))}{4(s-1)}\right)\frac{(D_{s-1}(v_{\vec{t}})+D_{s-1}(v_{\vec{r}}))}{s-1}\right)$ $\displaystyle\phantom{**********************}\times\left(1+O\left(\frac{l^{2}}{t^{2}}\right)\right),$ since $D_{s-1}(v_{\vec{t}}),D_{s-1}(v_{\vec{r}})\leq 2l-1$ and $s\geq\xi t$. In the fraction in the left hand side of (5.11), we will then have $\Theta(t)$ terms of the form $\left(1+O\left(\frac{l^{2}}{t^{2}}\right)\right),$ But again, as in Case $1$, their product goes to 1 as $t\to\infty$ since $l^{2}=o(t)$. This finishes the proof of the Lemma. We can finally finish the proof of the theorem. ###### Proof of Theorem 2. Let again $l$ be such that (5.2) is satisfied, and consider $\xi\in(0,1)$. Since it is impossible for two non disjoint and non equal isolated paths to exist at the same time, we have that (5.14) $\begin{split}{\mathbb{E}}\left[|\mathcal{S}_{l,\xi}(t)|^{2}\right]&={\mathbb{E}}\left[\left(\sum_{\vec{t}}\mathbb{1}\left\\{v_{\vec{t}}\in\mathcal{S}_{l,\xi}(t)\right\\}\right)\left(\sum_{\vec{r}}\mathbb{1}\left\\{v_{\vec{r}}\in\mathcal{S}_{l,\xi}(t)\right\\}\right)\right]\\\ &=\sum_{\begin{subarray}{c}\vec{t},\vec{r}\\\ \text{disjoint}\end{subarray}}{\mathbb{P}}\left(v_{\vec{t}},v_{\vec{r}}\in\mathcal{S}_{l,\xi}(t)\right)+{\mathbb{E}}\left[|\mathcal{S}_{l,\xi}(t)|\right],\end{split}$ and that $\left({\mathbb{E}}\left[|\mathcal{S}_{l,\xi}(t)|\right]\right)^{2}=\sum_{\begin{subarray}{c}\vec{t},\vec{r}\\\ \text{disjoint}\end{subarray}}{\mathbb{P}}\left(v_{\vec{t}}\in\mathcal{S}_{l,\xi}(t)\right){\mathbb{P}}\left(v_{\vec{r}}\in\mathcal{S}_{l,\xi}(t)\right)+\sum_{\begin{subarray}{c}\vec{t},\vec{r}\\\ \vec{r}\cap\vec{t}\neq\emptyset\end{subarray}}{\mathbb{P}}\left(v_{\vec{t}}\in\mathcal{S}_{l,\xi}(t)\right){\mathbb{P}}\left(v_{\vec{r}}\in\mathcal{S}_{l,\xi}(t)\right).$ Therefore, by lemmas 4 and 5, we obtain (5.15) $\begin{split}\frac{{\mathbb{E}}\left[|\mathcal{S}_{l,\xi}(t)|^{2}\right]}{\left({\mathbb{E}}\left[|\mathcal{S}_{l,\xi}(t)|\right]\right)^{2}}&=\frac{\sum_{\begin{subarray}{c}\vec{t},\vec{r}\\\ \text{disjoint}\end{subarray}}{\mathbb{P}}\left(v_{\vec{t}},v_{\vec{r}}\in\mathcal{S}_{l}(t)\right)}{\left({\mathbb{E}}\left[|\mathcal{S}_{l,\xi}(t)|\right]\right)^{2}}+\frac{{\mathbb{E}}\left[|\mathcal{S}_{l,\xi}(t)|\right]}{\left({\mathbb{E}}\left[|\mathcal{S}_{l,\xi}(t)|\right]\right)^{2}}\\\ &\leq\frac{\sum_{\begin{subarray}{c}\vec{t},\vec{r}\\\ \text{disjoint}\end{subarray}}{\mathbb{P}}\left(v_{\vec{t}},v_{\vec{r}}\in\mathcal{S}_{l,\xi}(t)\right)}{\sum_{\begin{subarray}{c}\vec{t},\vec{r}\\\ \text{disjoint}\end{subarray}}{\mathbb{P}}\left(v_{\vec{t}}\in\mathcal{S}_{l,\xi}(t)\right){\mathbb{P}}\left(v_{\vec{r}}\in\mathcal{S}_{l,\xi}(t)\right)}+\frac{1}{{\mathbb{E}}\left[|\mathcal{S}_{l,\xi}(t)|\right]}\\\ &\leq\frac{\sum_{\begin{subarray}{c}\vec{t},\vec{r}\\\ \text{disjoint}\end{subarray}}(1+o(1)){\mathbb{P}}\left(v_{\vec{t}}\in\mathcal{S}_{l,\xi}(t)\right){\mathbb{P}}\left(v_{\vec{r}}\in\mathcal{S}_{l,\xi}(t)\right)}{\sum_{\begin{subarray}{c}\vec{t},\vec{r}\\\ \text{disjoint}\end{subarray}}{\mathbb{P}}\left(v_{\vec{t}}\in\mathcal{S}_{l,\xi}(t)\right){\mathbb{P}}\left(v_{\vec{r}}\in\mathcal{S}_{l,\xi}(t)\right)}+\frac{1}{{\mathbb{E}}\left[|\mathcal{S}_{l,\xi}(t)|\right]}\\\ &\leq 1+o(1),\end{split}$ which proves the desired result. ## 6\. General upper bound for the diameter: proof of Theorem 3 In this section we provide a proof for Theorem 3. The main idea is to use Lemmas 1 and 3 to show that, with high probability, all vertices born up to time $t^{\frac{1}{12}}$ are in a connected component with diameter $2$. We then use a first moment estimate (Lemma 7 below) to show that the lengths of the paths formed by newer vertices have the desired upper bound. Given $k\in\mathbb{N}$ and $k$ times $s_{1},\dots,s_{k}\in\mathbb{N}$ such that $s_{1}<\dots<s_{k}$, we say that the vector of times $\vec{s}=(s_{1},\dots,s_{k})$ is a _vertex path_ if in the process $\\{G_{t}(f)\\}_{t\geq 1}$, at each time $s_{j}$, for $j=2,\dots,k$, a vertex is born and makes its first (vertex-step) connection to the vertex born at time $s_{j-1}$. We denote by $\\{s_{1}\leftarrow s_{2}\leftarrow\dots\leftarrow s_{k}\\}$ the event where $\vec{s}$ is a vertex path. We will show: ###### Lemma 6. Using the notation above, we have, for each vector $\vec{s}=(s_{1},\dots,s_{k})$ and step function $f$, (6.1) $\displaystyle{\mathbb{P}}\left(s_{1}\leftarrow s_{2}\leftarrow\dots\leftarrow s_{k}\right)\leq f(s_{1})\frac{s_{k}-1}{s_{1}+1}\prod_{m=2}^{k}\frac{f(s_{m})}{2(s_{m}-1)}.$ ###### Proof. Consider the events $\\{s_{1}\leftarrow s_{2}\leftarrow\dots\leftarrow s_{k-1}\\}$ and $\\{s_{k-1}\leftarrow s_{k}\\}$, defined analogously as the event in the above equation, but for the vectors $(s_{1},\dots,s_{k-1})$ and $(s_{k-1},s_{k})$ respectively. We have (6.2) $\begin{split}{\mathbb{P}}\left(s_{1}\leftarrow s_{2}\leftarrow\dots\leftarrow s_{k}\right)&={\mathbb{E}}\left[\mathbb{1}\\{s_{1}\leftarrow s_{2}\leftarrow\dots\leftarrow s_{k-1}\\}{\mathbb{P}}\left(s_{k-1}\leftarrow s_{k}\middle|\mathcal{F}_{s_{k}-1}\right)\right]\\\ &={\mathbb{E}}\left[\mathbb{1}\\{s_{1}\leftarrow s_{2}\leftarrow\dots\leftarrow s_{k-1}\\}f(s_{k})\cdot\frac{D_{s_{k}-1}(v_{s_{k-1}})}{2(s_{k}-1)}\right].\end{split}$ But, crucially, conditioned on the event where a vertex is born at time $s_{k-1}$, the degree of said vertex at time $s_{k}-1$ depends only on the connections made after time $s_{k-1}$, and is therefore independent of the indicator function above. We then obtain, by (4.3), (6.3) $\begin{split}\@ADDCLASS{ltx_eqn_lefteqn}$\displaystyle{\mathbb{P}}\left(s_{1}\leftarrow s_{2}\leftarrow\dots\leftarrow s_{k}\right)$\mbox{}\hfil\quad\\\ &={\mathbb{E}}\left[\mathbb{1}\\{s_{1}\leftarrow s_{2}\leftarrow\dots\leftarrow s_{k-1}\\}f(s_{k})\cdot\frac{D_{s_{k}-1}(v_{s_{k-1}})}{2(s_{k}-1)}\middle|Z_{s_{k-1}}=1\right]f(s_{k-1})\\\ &=f(s_{k-1})f(s_{k}){\mathbb{P}}\left(s_{1}\leftarrow s_{2}\leftarrow\dots\leftarrow s_{k-1}\middle|Z_{s_{k-1}}=1\right){\mathbb{E}}\left[\frac{D_{s_{k}-1}(v_{s_{k-1}})}{2(s_{k}-1)}\middle|Z_{s_{k-1}}=1\right]\\\ &={\mathbb{P}}\left(s_{1}\leftarrow s_{2}\leftarrow\dots\leftarrow s_{k-1}\right)\frac{f(s_{k})}{2(s_{k}-1)}\prod_{m=s_{k-1}}^{s_{k}-2}\left(1+\frac{1}{m}-\frac{f(m+1)}{2m}\right)\\\ &\leq{\mathbb{P}}\left(s_{1}\leftarrow s_{2}\leftarrow\dots\leftarrow s_{k-1}\right)\frac{f(s_{k})}{2(s_{k}-1)}\exp\left\\{\sum_{m=s_{k-1}}^{s_{k}-2}\left(\frac{1}{m}-\frac{f(m+1)}{2m}\right)\right\\},\end{split}$ by elementary properties of the exponential function. Iterating the above argument and recalling that the vertex $s_{1}$ is born with probability $f(s_{1})$, we obtain (6.4) $\begin{split}{\mathbb{P}}\left(s_{1}\leftarrow s_{2}\leftarrow\dots\leftarrow s_{k}\right)&\leq f(s_{1})\exp\left\\{\sum_{m=s_{1}}^{s_{k}-2}\left(\frac{1}{m}-\frac{f(m+1)}{2m}\right)\right\\}\prod_{m=2}^{k}\frac{f(s_{m})}{2(s_{m}-1)}\\\ &\leq f(s_{1})\frac{s_{k}-1}{s_{1}+1}\prod_{m=2}^{k}\frac{f(s_{m})}{2(s_{m}-1)},\end{split}$ finishing the proof of the lemma. Given $k,t_{0},t\in\mathbb{N}$, denote by $\mathcal{V}_{k,t_{0}}(t)$ the set of all vertex-paths of length $k$ whose vertices were born between times $t_{0}$ and $t$. Our goal now is to bound from above the expectation of $|\mathcal{V}_{k,t_{0}}(t)|$, this will be accomplished in the next lemma. ###### Lemma 7. Using the notation above defined, we have, for $k,t_{0},t\in\mathbb{N}$, (6.5) ${\mathbb{E}}\left[|\mathcal{V}_{k,t_{0}}(t)|\right]\leq C_{1}\exp\left\\{2\log t-(k-2)\left(\log(k-2)+C_{2}+\log\left(\sum_{j=t_{0}}^{t}\frac{f(j)}{j-1}\right)\right)\right\\}.$ ###### Proof. We will use Lemma 6 and the an application of the union bound. First, fix $s_{1},s_{k}\in\mathbb{N}$ such that $t_{0}\leq s_{1}<s_{k}\leq t$. We have, by Stirling’s approximation formula and the positivity of the terms involved, (6.6) $\begin{split}\sum_{\begin{subarray}{c}s_{2},s_{3},\dots,s_{k-1}\\\ s_{1}<s_{2}<\dots<s_{k-1}<s_{k}\end{subarray}}\prod_{m=2}^{k-1}\frac{f(s_{m})}{s_{m}-1}&\leq\frac{1}{(k-2)!}\left(\sum_{j=t_{0}}^{t}\frac{f(j)}{j-1}\right)^{k-2}\\\ &\leq C\exp\left\\{-(k-2)\left(\log(k-2)-1-\log\left(\sum_{j=t_{0}}^{t}\frac{f(j)}{j-1}\right)\right)\right\\}.\end{split}$ We can then show, by the above equation, Lemma 6, and the union bound, (6.7) $\begin{split}{\mathbb{E}}\left[|\mathcal{V}_{k,t_{0}}(t)|\right]&\leq\sum_{\begin{subarray}{c}s_{1},\dots,s_{k}\\\ t_{0}\leq s_{1}<\dots<s_{k}\leq t\end{subarray}}{\mathbb{P}}\left(s_{1}\leftarrow s_{2}\leftarrow\dots\leftarrow s_{k}\right)\\\ &\leq\sum_{\begin{subarray}{c}s_{1},\dots,s_{k}\\\ t_{0}\leq s_{1}<\dots<s_{k}\leq t\end{subarray}}f(s_{1})\frac{s_{k}-1}{s_{1}+1}\prod_{m=2}^{k}\frac{f(s_{m})}{2(s_{m}-1)}\\\ &=\frac{1}{2^{k-1}}\sum_{\begin{subarray}{c}s_{1},s_{k}\\\ t_{0}\leq s_{1}<s_{k}\leq t\end{subarray}}\frac{f(s_{1})f(s_{k})}{s_{1}+1}\sum_{\begin{subarray}{c}s_{2},\dots,s_{k-1}\\\ s_{1}<s_{2}<\dots<s_{k-1}<s_{k}\end{subarray}}\prod_{m=2}^{k-1}\frac{f(s_{m})}{s_{m}-1}\\\ &\leq C\exp\left\\{-(k-2)\left(\log(k-2)+C_{2}-\log\left(\sum_{j=t_{0}}^{t}\frac{f(j)}{j-1}\right)\right)\right\\}\sum_{\begin{subarray}{c}s_{1},s_{k}\\\ t_{0}\leq s_{1}<s_{k}\leq t\end{subarray}}\frac{f(s_{1})f(s_{k})}{s_{1}+1}\\\ &\leq C\exp\left\\{2\log t-(k-2)\left(\log(k-2)+C_{2}-\log\left(\sum_{j=t_{0}}^{t}\frac{f(j)}{j-1}\right)\right)\right\\},\end{split}$ thus concluding the proof of the result. We can finally finish the proof of Theorem 3. ###### Proof of Theorem 3. We prove part (a) first. Proof of part (a): This part follows immediately from the fact that the B-A random tree has diameter bounded by $\log t$ a.a.s (see Theorem $7.1$ of [27]) combined with Proposition 2.5 which states that the diameter is an increasing graph observable. Proof of part (b): Recall that, in this part of the theorem, $f$ is under condition (S), which holds if $\sum_{s=2}^{\infty}f(s)/s$ is finite. For $t_{0},t\in\mathbb{N}$ and $\delta\in(0,1)$, let $A_{\delta}(t_{0},t)$ be the event where every vertex born before time $t_{0}$ has degree at least $t^{\delta}$ in $G_{t}(f)$. For $\varepsilon\in(0,1)$, let $B_{\varepsilon}(t)$ be the event where there exists a vertex $v$ in $V(G_{t}(f))$ such that $D_{t}(v)\geq t^{1-\varepsilon}$. Then, by Lemmas 1 and 3, we have (6.8) ${\mathbb{P}}\left(A_{\frac{1}{15}}(t^{\frac{1}{12}},t)\right)\geq 1-Ct^{-144^{-1}},\quad\quad{\mathbb{P}}\left(B_{\frac{1}{30}}(t)\right)\geq 1-t^{-2}.$ We have, on the event where two vertices $u_{1},u_{2}\in V(G_{t}(f))$ are such that $D_{t}(u_{1})\geq t^{15^{-1}}$ and $D_{t}(u_{2})\geq t^{1-30^{-1}}$, (6.9) $\begin{split}{\mathbb{P}}\left(u_{1}\nleftrightarrow u_{2}\text{ in }G_{2t}(f)\right)&\leq\prod_{s=t+1}^{2t}\left(1-\frac{f(s)}{2(s-1)}\right)\left(1-\frac{t^{15^{-1}}t^{1-30^{-1}}}{4(s-1)^{2}}\right)\\\ &\leq\exp\left\\{-\frac{t^{15^{-1}}t^{1-30^{-1}}}{16t^{2}}\cdot t\right\\}\\\ &=\exp\left\\{-\frac{t^{-30^{-1}}}{16}\right\\}.\end{split}$ Recall the notation $v_{i}$ symbolizing the vertex born at time $i\in\mathbb{N}$. Together with (6.8) and the union bound, the above equation implies $\begin{split}\@ADDCLASS{ltx_eqn_lefteqn}$\displaystyle{\mathbb{P}}\left(\exists i,j\in\mathbb{N},\text{ such that }1\leq i<j\leq t^{12^{-1}},Z_{i}=Z_{j}=1,\text{ and }\mathrm{dist}(v_{i},v_{j})>2\text{ in }G_{2t}(f)\right)$\mbox{}\hfil\phantom{***}\\\ &\leq Ct^{-144^{-1}}+t^{-2}+\sum_{1\leq i<j\leq t^{\frac{1}{12}}}{\mathbb{P}}\left(A_{\frac{1}{15}}(t^{\frac{1}{12}},t),B_{\frac{1}{30}}(t),\mathrm{dist}(v_{i},v_{j})>2\text{ in }G_{2t}(f)\right)\\\ &\leq Ct^{-144^{-1}}+t^{-2}+t^{-6^{-1}}\exp\left\\{-\frac{t^{-30^{-1}}}{16}\right\\}\\\ &\leq Ct^{-144^{-1}}.\end{split}$ This implies the existence of a constant $C_{1}>0$ such that the probability of there existing two vertices born before time $C_{1}t^{12^{-1}}$ such that the distance between said vertices is larger than $2$ in $G_{t}(f)$ is polynomially small in $t$. We now turn our attention to vertices born after $t^{\frac{1}{13}}$. We will use Lemma 7 in order to bound the probability of there existing long vertex-paths formed by vertices born after $t^{\frac{1}{13}}$, the notion of a “long” path being $f$-dependent. Let $V_{t}(t^{\frac{1}{13}})$ denote the set of all vertices of $V(G_{t}(f))$ born before time $t^{\frac{1}{13}}$, let ${\bf d}_{\mathrm{max}}(t^{\frac{1}{13}},t)$ be the length of a maximal vertex-path of vertices born between times $t^{\frac{1}{13}}$ and $t$. Let $u_{1},u_{2}\in V(G_{t}(f))$. Since $G_{t}(f)$ is connected, (6.10) $\begin{split}\mathrm{dist}(u_{1},u_{2})&\leq\mathrm{dist}(u_{1},V_{t}(t^{\frac{1}{13}}))+\mathrm{diam}(V_{t}(t^{\frac{1}{13}}))+\mathrm{dist}(u_{2},V_{t}(t^{\frac{1}{13}}))\\\ &\leq 2{\bf d}_{\mathrm{max}}(t^{\frac{1}{13}},t)+\mathrm{diam}(V_{t}(t^{\frac{1}{13}})).\end{split}$ But we know that $\mathrm{diam}(V_{t}(t^{\frac{1}{13}}))\leq 2$ with high probability. Bounding ${\bf d}_{\mathrm{max}}(t^{\frac{1}{13}},t)$ then gives us an a.a.s. upper bound for the diameter of $G_{t}(f)$. Now, given $t\in\mathbb{N}$, if (6.11) $k\geq 3\left(\frac{\log t}{-\log\left(\sum_{s=t^{\frac{1}{13}}}^{t}\frac{f(s)}{s-1}\right)}\wedge\frac{\log t}{\log\log t}\right),$ then, by Lemma 7, and since $\log\left(\sum_{j=t^{\frac{1}{13}}}^{t}\frac{f(j)}{j-1}\right)$ is eventually negative for large $t$ (recall again that $f$ satisfies (S)), we have, (6.12) $\begin{split}{\mathbb{P}}\left({\bf d}_{\mathrm{max}}>k\right)&\leq{\mathbb{E}}\left[|\mathcal{V}_{k,t^{1/13}}(t)|\right]\\\ &\leq C_{1}\exp\left\\{2\log t-(k-2)\left(\log(k-2)+C_{2}-\log\left(\sum_{j=t^{\frac{1}{13}}}^{t}\frac{f(j)}{j-1}\right)\right)\right\\}\\\ &\leq Ct^{-\frac{1}{2}}.\end{split}$ The above upper bound and (6.10) proves part (b). Proof of part (c): Recall that in this part, we are under condition (Lκ), which holds if, for $\kappa\in(0,1)$ and every $t\in\mathbb{N}$ sufficiently large, one has (6.13) $\sum_{s=t^{1/13}}^{t}\frac{f(s)}{s}<(\log t)^{\kappa}.$ Then, let $k$ be so that $k\geq\frac{3}{1-\kappa}\frac{\log t}{\log\log t}.$ We then obtain, again by Lemma 7, for sufficiently large $t$, (6.14) $\begin{split}\@ADDCLASS{ltx_eqn_lefteqn}$\displaystyle{\mathbb{P}}\left({\bf d}_{\mathrm{max}}>k\right)$\mbox{}\hfil\quad\quad\\\ &\leq{\mathbb{E}}\left[|\mathcal{V}_{k,t^{1/13}}(t)|\right]\\\ &\leq C_{1}\exp\left\\{2\log t-(k-2)\left(\log(k-2)+C_{2}-\log\left(\sum_{j=t^{\frac{1}{13}}}^{t}\frac{f(j)}{j-1}\right)\right)\right\\}\\\ &\leq C_{1}\exp\left\\{2\log t-\frac{3}{1-\kappa}\frac{\log t}{\log\log t}\left((1-\kappa)\log\log t-C-\log\log\log t\right)\right\\}\\\ &\leq Ct^{-\frac{1}{2}}.\end{split}$ Finally, the above upper bound together with (6.10) finishes the proof of the Theorem. ## 7\. The family of regularly varying functions In this section we explore our results when more information on $f$ is provided. In particular, we assume that $f$ satisfies condition (RVγ) for $\gamma\in[0,1)$. Recall that this condition holds whenever $\exists\gamma,\text{ such that }\forall a>0,\;\lim_{t\to\infty}\frac{f(at)}{f(t)}=\frac{1}{a^{\gamma}}.$ Also recall that when $f$ satisfies the above condition for $\gamma>0$ we say it is a regular varying function with index of regular variation $-\gamma$. The case $\gamma=0$ is said to be slowly varying. To prove the results under the assumption of regular variation our arguments rely on the theorems from Karamata’s theory. In particular, the Representation Theorem (Theorem 1.4.1 of [8]) and Karamata’s theorem (Proposition 1.5.8 of [8]). The former states that if $f$ is a regularly varying function with index $-\gamma$, then there exists a slowly varying function $\ell$ such that $f$ is of the form (7.1) $f(t)=\frac{\ell(t)}{t^{\gamma}}.$ Whereas, the latter states that if $\ell$ is a slowly varying function, then, for any $a<1$ (7.2) $\int_{1}^{t}\frac{\ell(x)}{x^{a}}\mathrm{d}x\sim\frac{\ell(t)t^{1-a}}{1-a},$ and for $a>1$, we have (7.3) $\int_{t}^{\infty}\frac{\ell(x)}{x^{a}}\mathrm{d}x\sim\frac{\ell(t)}{(a-1)t^{a-1}}.$ We begin by proving Corollary 1.4, which gives lower and upper bounds for the clique number $\omega(G_{t}(f))$. ###### Proof of Corollary 1.4. We prove part (a) first. Proof of part (a): The lower bound follows from (7.1), when $f$ is regularly varying, (7.2) and Theorem 1. Just observe that, if $f$ is regularly varying, then $F(t)=\Theta({t^{1-\gamma}})$, whereas if it is slowly, we have $F(t)=\Theta(f(t)t)$. In the latter case, we appeal to another result from Karamata’s Theory (Corollary A.6 of [2]) which assures that for all $\varepsilon>0$, (7.4) $\frac{f(t)}{t^{\varepsilon}}\stackrel{{\scriptstyle t\to\infty}}{{\longrightarrow}}0.$ Then, by Theorem 1, $G_{t}(f)$ has, with probability $1-o(1)$, a clique of order $C_{\varepsilon}F(t^{\frac{1}{2}(1-\frac{\varepsilon}{2})})$, which gives the desired lower bound for $\omega(G_{t}(f))$. The upper bound comes from the deterministic bound that says that a graph with $t$ (simple) edges has at most $t^{\frac{3}{2}}$ triangles (see e.g. Theorem $4$ of [24]), and the fact that a complete subgraph with $k$ vertices has $6^{-1}k(k-1)(k-2)$ triangles. Proof of part (b): First observe that for any small $\varepsilon>0$ by (7.2), $F(t^{\frac{1}{2}(1-\varepsilon)})\gg\log\log t$ for large enough $t$. Thus, by (3.8), we have that (7.5) ${\mathbb{P}}\left(\omega(G_{t}(f))\geq C_{\varepsilon}F(t^{\frac{1-\varepsilon}{2}})\right)\geq 1-\frac{2}{\log t}.$ Now, define the following sequence of deterministic times (7.6) $t_{k}:=(1+\varepsilon)^{k^{2}}.$ By the first Borel-Cantelli Lemma and (7.5) we have that, (7.7) $\liminf_{k\to\infty}\frac{\log\omega(G_{t_{k}}(f))}{\log(C_{\varepsilon}F(t_{k}^{\frac{1-\varepsilon}{2}}))}\geq 1,\;\text{a.s.}$ Using that $f$ is slowly varying, Karamata’s Theorem (7.2), and (7.4), we also have that (7.8) $\lim_{t\to\infty}\frac{\log(C_{\varepsilon}F(t^{\frac{1-\varepsilon}{2}}))}{\log F(t^{\frac{1}{2}})}=1-\frac{\varepsilon}{2}.$ For the same reason, it also follows that (7.9) $\lim_{t\to\infty}\frac{\log F(t^{\frac{1}{2}})}{\log(7t^{\frac{1}{2}})}=1.$ Then (7.7) yields (7.10) $\liminf_{k\to\infty}\frac{\log\omega(G_{t_{k}}(f))}{\log F(t_{k}^{\frac{1}{2}})}\geq 1-\varepsilon,\;\text{a.s.}.$ And using the upper bound given in part (a) and (7.9) we also obtain (7.11) $\limsup_{k\to\infty}\frac{\log\omega(G_{t_{k}}(f))}{\log F(t_{k}^{\frac{1}{2}})}\leq 1,\;\text{a.s.}$ Now, note that $\omega(G_{t}(f))\leq\omega(G_{t+1}(f))$, since $G_{t}(f)$ is contained in $G_{t+1}(f)$. Thus, for any time $t\in(t_{k},t_{k+1})$ it follows (7.12) $\begin{split}\frac{\log(\omega(G_{t_{k}}(f)))}{\log F(t_{k+1}^{\frac{1}{2}})}\leq\frac{\log(\omega(G_{t}(f)))}{\log F(t^{\frac{1}{2}})}\leq\frac{\log(\omega(G_{t_{k+1}}(f)))}{\log F(t_{k}^{\frac{1}{2}})}.\end{split}$ Using (7.10), (7.11) and that $\lim_{k\to\infty}\frac{\log F(t_{k+1}^{\frac{1}{2}})}{\log F(t_{k}^{\frac{1}{2}})}=1$ we finally obtain that (7.13) $1-\varepsilon\leq\liminf_{t\to\infty}\frac{\log(\omega(G_{t}(f)))}{\log F(t^{\frac{1}{2}})}\leq\limsup_{t\to\infty}\frac{\log(\omega(G_{t}(f)))}{\log F(t^{\frac{1}{2}})}\leq 1,\;\text{a.s.}.$ Since $\varepsilon$ was arbitrarily chosen, we conclude the proof. The next step is to prove the constant order of the diameter of the graphs generated by regularly varying functions and how it depends on the index of regular variation on infinity. This result is stated on Theorem 4 and we provide a proof for it below. ###### Proof of Theorem 4. We begin proving the lower bound. Lower bound: By the Representation Theorem (7.1) we have that there exists a slowly varying function $\ell$ such that $f(t)=\ell(t)/t^{\gamma}$. Moreover, (7.4) implies that (7.14) $\frac{\log\ell(t)}{\log t}\to 0.$ Thus, we have that (7.15) $-\frac{\log t}{\log f(t)}=\frac{\log t}{\gamma\log(t)+\log\ell(t)}=\frac{1-o(1)}{\gamma}.$ Applying Theorem 2 gives us a lower bound of order $\gamma^{-1}$. Upper bound: By the Representation Theorem and (7.4), we have that $f$ also satisfies condition (S). Just notice that (7.16) $\sum_{s=1}^{\infty}\frac{f(s)}{s}=\sum_{s=1}^{\infty}\frac{\ell(s)}{s^{1+\gamma}}<\infty.$ And by Karamata’s Theorem (Equation (7.3) in particular) we have that (7.17) $\sum_{s=t^{\frac{1}{13}}}^{t}\frac{\ell(s)}{s^{1+\gamma}}\sim\frac{\ell(t)}{t^{\gamma\frac{\gamma}{13}}}\implies-\frac{\log t}{\log\left(\sum_{s=t^{\frac{1}{13}}}^{t}\frac{\ell(s)}{s^{1+\gamma}}\right)}\sim\frac{\log t}{\frac{\gamma}{13}\log t+\log\ell(t)}=\frac{13(1-o(1))}{\gamma}$ And finally, applying Theorem 3 we prove the result. ## 8\. Final comments We end this paper with a brief discussion on the affine version of our model and how dropping some regularity conditions on $f$ may produce a sequence of graphs $\\{G_{t}\\}_{t\in\mathbb{N}}$ that has a subsequence which is essentially of complete graphs and another one which is close to the BA-model. ### Affine version At the introduction, we have discussed the affine version of the PA-rule, which we recall below. ${\mathbb{P}}\left(v_{t+1}\rightarrow u\middle|G_{t}\right)=\frac{\text{degree}(u)+\delta}{\sum_{w\in G_{t}}(\text{degree}(w)+\delta)}.$ In [2], the authors showed that the effect of the affine term $\delta$ vanishes in the long run when one is dealing with the empirical degree distribution. I.e., their results show that $\delta$ has no effect on the degree sequence of the graphs, what is not observed in the affine version of the BA-model, for which the exponent of the power-law distribution depends on $\delta$, see [13]. However, regarding the diameter, we believe $\delta$ may have an increasing/decreasing effect on the diameter’s order. One also may find interesting to consider $\delta=\delta(t)$ and investigate which one takes over: is it the edge-step function or the affine term? ### Dropping regularity conditions In this part we illustrate that dropping some assumptions on $f$ may produce a somewhat pathological sequence of graphs. For instance, if we drop the assumption of $f$ being non-increasing, we may obtain a sequence of graphs whose diameter sequence oscillates between $1$ and $\log t$. More generally, the sequence of graphs oscillates between graphs similar to the BA-random tree, $\\{\mathrm{BA}_{t}\\}_{t\in\mathbb{N}}$, and graphs close to complete graphs. Let $(t_{k})_{k\in\mathbb{N}}$ be the following sequence: $t_{0}=1$ and $t_{k+1}:=\exp\\{t_{k}\\}$, for $k>1$. Now, let $h$ be the edge-step function defined as follows (8.1) $h(t)=\begin{cases}1&\quad\text{ if }t\in[t_{2k},t_{2k+1}],\\\ 0&\quad\text{ if }t\in(t_{2k+1},t_{2k+2}).\end{cases}$ The idea behind such $h$ is that between times $[t_{2k},t_{2k+1}]$ the process behaves essentially as the traditional BA-model, whereas at interval $(t_{2k+1},t_{2k+2})$ the process “messes things up” connecting almost all vertices by only adding new edges. Moreover, in both regimes the process has time enough to “forget about what was built in the past”. Using Lemma 3 and reasoning as in (6.9), one may prove that $\mathrm{diam}_{G_{t_{2k+1}^{13}}(h)}G_{t_{2k+1}}(h)\leq 2,\text{ a.a.s.}$ However, the process does not add any new vertex in the interval $(t_{2k+1},t_{2k+2})$. Therefore, $\mathrm{diam}G_{t_{2k+2}}(h)\leq 2$, a.a.s. On the other hand, if we sample $G_{t_{2k+3}}(h)$ and $\mathrm{BA}_{t_{2k+3}}$ from the doubly-labeled tree $\mathcal{T}_{t_{2k+3}}$ it follows that the quantity $\max_{u>t_{2k+2}}\mathrm{dist}(u,\mathcal{T}_{t_{2k+2}})$ remains the same under $h$ and under $f\equiv 1$, since paths using vertices added after $t_{2k+2}$ belong to the graphs generated by both functions. Finally, noting that $\mathcal{T}_{t_{2k+2}}$ has diameter at most $\log t_{2k+2}=t_{2k+1}$ w.h.p, which means it is a very small graph when compared to $\mathcal{T}_{t_{2k+3}}$, it is not hard to see that the diameter of $G_{2k+3}(h)$ has the same order the diameter from $\mathrm{BA}_{t_{2k+3}}$. Observe that the function $h$ may be constructed considering it equal to any other edge-step function $f$ instead of the constant case $1$. Roughly speaking, when we sample both processes from the doubly-labeled random tree, up to a very small subgraph, the graph $G_{t_{2k+1}}(h)$ is similar to $G_{t_{2k+1}}(f)$, whereas, $G_{t_{2k}}(h)$ is a graph of diameter at most $2$ and whose vertices have degree at least a subpolynomial of $t_{2k}$. The conclusion is that if we drop some monoticity assumption on $f$, we may obtain a sequence of graphs having at least two subsequences that are completely different as graphs. Acknowledgements We are thankful to Roberto Imbuzeiro Oliveira for the many inspiring and pleasant conversations. C.A. was partially supported by Fundação de Amparo à Pesquisa do Estado de São Paulo (FAPESP), grants 2013/24928-2 and 2015/18930-0, and by the Deutsche Forschungsgemeinschaft (DFG). R.R. was partially supported by Conselho Nacional de Desenvolvimento Científico e Tecnológico (CNPq) and by the project Stochastic Models of Disordered and Complex Systems. The Stochastic Models of Disordered and Complex Systems is a Millennium Nucleus (NC120062) supported by the Millenium Scientific Initiative of the Ministry of Science and Technology (Chile). R.S. has been partially supported by Conselho Nacional de Desenvolvimento Científico e Tecnológico (CNPq) and by FAPEMIG (Programa Pesquisador Mineiro), grant PPM 00600/16. ## References * [1] C. Alves, R. Ribeiro, and R. Sanchis. Large communities in a scale-free network. Journal of Statistical Physics, 166(1):137–149, 2017. * [2] C. Alves, R. Ribeiro, and R. Sanchis. Preferential attachment random graphs with edge-step functions. 04 2017. * [3] M. Amin Abdullah and N. Fountoulakis. A phase transition in the evolution of bootstrap percolation processes on preferential attachment graphs. Random Structures & Algorithms, 52(3):379–418, 2018. * [4] A. Backhausz and T. F. Móri. Asymptotic properties of a random graph with duplications. J. Appl. Probab., 52(2):375–390, 06 2015. * [5] A-L. Barabási and R. Albert. Emergence of scaling in random networks. Science, 1999. * [6] N. Berger, B. Bollobás, C. Borgs, J. Chayes, and O. Riordan. Degree distribution of the fkp network model. Theoretical computer science, 379(3):306–316, 2007. * [7] G. Bianconi and M. Marsili. Emergence of large cliques in random scale-free networks. EPL (Europhysics Letters), 74(4):740, 2006. * [8] N. H. Bingham, C. M. Goldie, and J. L. Teugels. Regular variation, volume 27 of encyclopedia of mathematics and its applications, 1989. * [9] B. Bollobás and O. Riordan. Mathematical results on scale-free random graphs. in handbook of graphs and networks: From the genome to the internet. pages 1–34, 2003. * [10] B. Bollobás and O. Riordan. Robustness and vulnerability of scale-free random graphs. Internet Math., 1(1):1–35, 2003. * [11] S. Chatterjee and R. Durrett. Contact processes on random graphs with power law degree distributions have critical value 0. Ann. Probab., 37(6):2332–2356, 11 2009. * [12] F. Chung and L. Lu. Complex Graphs and Networks (Cbms Regional Conference Series in Mathematics). American Mathematical Society, Boston, MA, USA, 2006. * [13] C. Cooper and A. Frieze. A general model of web graphs. Random Struct. Algorithms, 22(3):311–335, 2003. * [14] S. Dereich and M. Ortgiese. Robust analysis of preferential attachment models with fitness. Combinatorics, Probability and Computing, 23(3):386–411, 2014\. * [15] S. Dommers, R. van der Hofstad, and G. Hooghiemstra. Diameters in preferential attachment models. Journal of Statistical Physics, 139(1):72–107, 2010. * [16] N. Eggemann and S.D. Noble. The clustering coefficient of a scale-free random graph. Discrete Applied Mathematics, 2011. * [17] P. Erdös and A. Rényi. On random graphs i. Publicationes Mathematicae Debrecen, 6:290, 1959. * [18] D. Fernholz and V. Ramachandran. The diameter of sparse random graphs. Random Structures and Algorithms, 30(4):482–516, 2007. * [19] E. Jacob and P. Mörters. Spatial preferential attachment networks: Power laws and clustering coefficients. Ann. Appl. Probab., 25(2):632–662, 04 2015. * [20] E. Jacob and P. Mörters. The contact process on scale-free networks evolving by vertex updating. Royal Society Open Science, 4(5):170081, 2017. * [21] S. Janson, T. Luczak, and I. Norros. Large cliques in a power-law random graph. J. App. Prob., 2010. * [22] Russell Lyons and Yuval Peres. Probability on Trees and Networks, volume 42 of Cambridge Series in Statistical and Probabilistic Mathematics. Cambridge University Press, New York, 2016. Available at http://pages.iu.edu/~rdlyons/. * [23] Y. Malyshkin and E. Paquette. The power of choice combined with preferential attachement. Electronic Communications in Probability, 19(44):1–13, 2014. * [24] I. Rivin. Counting cycles and finite dimensional lp norms. Advances in applied mathematics, 29(4):647–662, 2002. * [25] S.H Strogatz and D. J. Watts. Tcollective dynamics of ’small-world’ networks. Nature, 1998. * [26] E. Thörnblad. Asymptotic degree distribution of a duplication–deletion random graph model. Inter. Math., 11, 03 2014. * [27] R. Van Der Hofstad. Random graphs and complex networks volume 2. Available on http://www. win. tue. nl/ rhofstad/ NotesRGCNII. pdf, 2018. * [28] D. Znamenski, G. Hooghiemstra, and R. van der Hofstad. A phase transition for the diameter of the configuration model. Internet Mathematics, 4(1):113–128, 2007.
# Learning Dense Visual Correspondences in Simulation to Smooth and Fold Real Fabrics Aditya Ganapathi1, Priya Sundaresan1, Brijen Thananjeyan1, Ashwin Balakrishna1, Daniel Seita1, Jennifer Grannen1, Minho Hwang1, Ryan Hoque1, Joseph E. Gonzalez1, Nawid Jamali2, Katsu Yamane2, Soshi Iba2, Ken Goldberg1 1AUTOLab at the University of California, Berkeley, USA2Honda Research Institute, USACorrespondence to Aditya Ganapathi<EMAIL_ADDRESS> (112001) ###### Abstract Robotic fabric manipulation is challenging due to the infinite dimensional configuration space, self-occlusion, and complex dynamics of fabrics. There has been significant prior work on learning policies for specific deformable manipulation tasks, but comparatively less focus on algorithms which can efficiently learn many different tasks. In this paper, we learn visual correspondences for deformable fabrics across different configurations in simulation and show that this representation can be used to design policies for a variety of tasks. Given a single demonstration of a new task from an initial fabric configuration, the learned correspondences can be used to compute geometrically equivalent actions in a new fabric configuration. This makes it possible to robustly imitate a broad set of multi-step fabric smoothing and folding tasks on multiple physical robotic systems. The resulting policies achieve $80.3\%$ average task success rate across 10 fabric manipulation tasks on two different robotic systems, the da Vinci surgical robot and the ABB YuMi. Results also suggest robustness to fabrics of various colors, sizes, and shapes. See https://tinyurl.com/fabric-descriptors for supplementary material and videos. ## I Introduction Robot fabric manipulation has applications in folding laundry [46, 17, 4, 25], bed making [37], surgery [43, 42, 38, 8], and manufacturing [27, 45]. However, while robots are able to learn general purpose policies to manipulate a variety of rigid objects with increasing reliability [22, 6, 28, 14, 20], learning such policies for manipulating deformable objects remains an open problem due to difficulties in sensing and control. While there is significant prior work on geometric [35, 46, 1, 23] and learning based approaches [47, 36, 37] for fabric manipulation, these approaches often involve designing or learning task-specific manipulation policies, making it difficult to efficiently reuse information for different tasks. In this work, we leverage recent advances in dense keypoint learning [6] to learn visual point-pair correspondences across fabric in different configurations. Then, given a single offline demonstration of a fabric manipulation task from a given configuration, we utilize the learned correspondences to compute geometrically equivalent actions to complete the task on a similar fabric in a different configuration. For example, a human might provide a sequence of actions that would fold a T-shirt when it is placed neck up in a smoothed configuration. However, when a robot is operating at test time, it will likely encounter a different T-shirt whose color, size and pose may differ from the T-shirt used for the demonstration. We find that learning visual correspondences that are invariant across these fabric attributes provides a powerful representation for defining controllers that can generalize to the above variations. Figure 1: We use learned visual correspondences across different fabric configurations to perform a variety of fabric manipulation tasks on the ABB YuMi (top) and the da Vinci Research Kit (bottom). Given a single demonstration of smoothing or folding, the robot uses the learned correspondences to compute geometrically equivalent actions for fabric of different color and in different initial configurations. This enables robust one-shot imitation learning of tasks that involve smoothing then folding. We extend work by [41], which leverages dense object descriptors [6] to learn visual correspondences for rope using synthetic depth data in simulation. These correspondences are then used to learn new rope manipulation tasks such as rearrangement or knot tying given a single task demonstration. We find that similar visual correspondence learning methods are also effective for learning correspondences between different fabric configurations using task-agnostic RGB data collected entirely in simulation and can be used to perform fabric manipulation tasks. Precisely, given a user demonstration of the task from a given initial fabric configuration, we leverage the learned visual correspondences to perform the same task from different initial configurations by computing geometrically equivalent actions using the correspondences. This approach has a number of appealing properties. First, visual correspondences can be learned purely in simulation without task-specific data and widely applied to a variety of real fabric manipulation tasks with no further training. Second, training in simulation enables sufficient data variety through domain randomization, making it possible to learn correspondences that generalize to fabrics with different colors, shapes, and configurations. Third, since perception and control are decoupled, the same perception module can be used on different robots with no additional training. We contribute (1) a framework for learning dense visual correspondences of fabric in simulation using dense object descriptors from [6, 41] and applying them to manipulation tasks on real fabrics with unseen colors, scales, and textures, (2) a data generation pipeline for collecting images of fabrics and clothing in Blender [3] and a testbed to experiment with different manipulation policies on these fabrics in simulation and (3) physical experiments on both the da Vinci Research Kit (dVRK) [15] and the ABB YuMi suggesting that the learned descriptors transfer effectively on two different robotic systems. We experimentally validate the method on 10 different tasks involving 5 T-shirts and 5 square fabrics of varying dimensions and colors and achieve an average task success rate of $80.3\%$. ## II Related Work Fabric manipulation is an active area of robotics research [2, 33, 18, 11]. Over the past decade, the research has primarily been focused on three different categories: perception-based manipulation, learning-based algorithms in the real world, and learning-based algorithms in simulation which are then transferred to real robots. Traditional Vision-Based Algorithms for Fabric Manipulation: Much of the prior work on perception-based deformable object manipulation relies on traditional image processing and vision techniques to estimate the state of the fabric. This state estimation is then used to define geometric controllers which bring the fabric into some desired configuration. However, due to limitations in these traditional vision algorithms, most prior work makes specific assumptions on the fabric’s initial configurations or requires more complex robotic manipulators to bring the fabric into a desired starting configuration. For example, Miller _et al_. [25] demonstrate a robust folding pipeline for clothing by fitting a polygonal contour to the fabric and designing a geometric controller on top of it, but assume that the initial state of the fabric is flat. Sun _et al_. [39, 40] perform effective fabric smoothing by estimating the wrinkles in the fabric, but condition on a near- flat starting fabric. Other work relies on “vertically smoothing” fabrics using gravity [26, 16, 17, 4, 23] to standardize the initial configuration and to expose fabric corners before attempting the task, which is difficult for large fabrics or single-armed robots. Learning-Based Algorithms in the Real World: More recent approaches have transitioned to end to end learning of fabric manipulation directly on a real system, but these approaches have struggled to generalize to a variety of fabrics and tasks due to the high volume of training data required. For example, [5] use model-based reinforcement learning to learn fabric manipulation policies which generalize to many tasks, but require several days of continuous data collection on a real physical system and perform relatively low precision tasks. Jia _et al_. [12, 13] show impressive collaborative human-robot cloth folding under the assumption that fabric has already been grasped and is in a particular starting configuration, and [35] demonstrate deformable object manipulation while requiring task-specific kinesthetic demonstrations. In follow-up work, [19] consider many of the same tasks as in this paper and demonstrate that policies can be learned to fold fabric using reinforcement learning with only one hour of experience on a real robot. In contrast, we learn entirely in simulation and decouple perception from control, making it easier to generalize to different fabric colors and shapes and flexibly deploy the learned policies on different robots. Sim-to-Real Learning-Based Algorithms: Due to the recent success of sim-to- real transfer [32, 44], many recent papers leverage simulation to collect large amounts of training data which is used to learn fabric manipulation policies. Seita _et al_. [37, 36] and [47] followed-up on the smoothing task from [39] by generalizing to a wider range of initial fabric states using imitation learning (DAgger [31]), and reinforcement learning (Soft Actor- Critic [9]) respectively, but still tailor policies specifically for smoothing. Similarly, [24] learn fabric folding policies by using deep reinforcement learning augmented with task-specific demonstrations. These works use simulation to optimize fabric manipulation policies for specific tasks. In follow-up and concurrent work, [10] and [48] use simulation to train fabric manipulation policies using model-based reinforcement learning for multiple tasks. In contrast, we leverage simulation to learn visual representations of fabric to capture its geometric structure without task- specific data or a model of the environment and then use this representation to design intuitive controllers for several tasks from different starting configurations. Dense Object Descriptors: We learn visual representations for fabric by using dense object descriptors [6, 34], which were shown to enable task oriented manipulation of various rigid and slightly deformable objects [6]. This approach uses a deep neural network to learn a representation which encourages corresponding pixels in images of an object in different configurations to have similar representations in embedding space. Such descriptors can be used to design geometrically structured manipulation policies for grasping [6], assembly [49], or for learning from demonstrations [7]. [41] extend this idea to manipulation of ropes, and demonstrate that deformation-invariant dense object descriptors can be learned for rope using synthetic depth data in simulation and then transferred to a real physical system. [41] then use the learned descriptors to imitate offline demonstrations of various rope manipulation tasks. In this work, we apply the techniques from [41] to learn descriptors which capture geometric correspondence across different fabric configurations from synthetic RGB images and use them for 2D fabric manipulation. ## III Problem Definition ### III-A Assumptions We assume a deformable object is place on a planar workspace in an initial configuration $\xi_{1}$ observed by an overhead camera corresponding RGB image $I_{1}:=I_{1}(\xi_{1})\in\mathbb{R}^{W\times H\times 3}$. As in prior work [36, 47], we assume that fabric manipulation tasks can be completed by a sequence of actions where each includes grasping at a _pick point_ , pulling to a _place point_ without changing the orientation of the end-effector, and releasing at the place point. We additionally assume access to a single demonstration of each task in the form of a sequence of pick and place actions from some arbitrary initial fabric configuration $\xi_{1}$. These demonstrations can be collected offline, such as through a GUI where a user clicks on an image of fabric to indicate pick and place point pixels. However, the fabric used to create the instruction does not have to be of the same color, the same size or in the same initial configuration as the fabric the robot sees at test time. The only requirement is that the fabric be of a similar geometry. For example, T-shirts can be compared to other instances of T-shirts, but not to pants or long-sleeved shirts. ### III-B Task Definition Define the action at step $j$ as $\mathbf{a}_{j}=((x_{g},y_{g})_{j},(x_{p},y_{p})_{j})$ (III.1) where $(x_{g},y_{g})_{j}$ and $(x_{p},y_{p})_{j}$ are the pixel coordinates of a grasp point on the fabric and place point respectively in image $I_{j}$ at time $j$. The robot grasps the world coordinate associated with the grasp point and then moves to the world coordinate associated with the place point without changing the end effector orientation. This causes the fabric located at $(x_{g},y_{g})_{j}$ in the image to be placed on top of the world coordinate associated with $(x_{p},y_{p})_{j}$ with the same surface normals as before. In future work, we will investigate how to execute more complex actions that result in reversed surface normals, which requires a rotation motion during the action. We are given a sequence of actions $\left(\mathbf{a}_{j}\right)_{j=1}^{n}$ executed on a fabric starting in configuration $\xi_{1}$ and corresponding observations $\left(I_{j}\right)_{j=1}^{n}$. Then at test-time, a similar object is dropped onto the surface in a previously unseen configuration and the goal is to generate a corresponding sequence of actions for a fabric in some previously unseen configuration. Specifically, the robot generates a new sequence of actions: $\Big{(}\mathbf{a}_{j}^{\prime}\Big{)}_{j=1}^{n}=\Big{(}d_{I_{j}\rightarrow I_{j}^{\prime}}(x_{g},y_{g})_{j},\;\;d_{I_{j}\rightarrow I_{j}^{\prime}}(x_{p},y_{p})_{j}\Big{)}_{j=1}^{n}$ (III.2) for $j\in\\{1,\ldots,n\\}$ where $d_{I_{j}\rightarrow I_{j}^{\prime}}:\mathbb{R}^{2}\rightarrow\mathbb{R}^{2}$ is a function which estimates the corresponding point $(x^{\prime},y^{\prime})_{j}$ in $I_{j}^{\prime}$ given a point $(x,y)_{j}$ in $I_{j}$. This function is difficult to compute directly from images in general, and even more so difficult to compute for images of highly deformable objects due to their infinite degrees of freedom and tendency to self-occlude. Thus, we leverage dense object descriptors [6] to approximate $d_{I_{j}\rightarrow I_{j}^{\prime}}$ for any $I_{j}$ and $I_{j}^{\prime}$, as described in Sections V and VI. ## IV Simulator We use Blender 2.8, an open-source simulation and rendering engine [3] released in mid-2019, to both create large synthetic RGB training datasets and model the fabric dynamics for simulated experiments using its in-built fabric solver based on [30, 29]. We simulate T-shirts and square fabrics, each of which we model as a polygonal mesh made up of 729 vertices, a square number we experimentally tuned to trade-off between fine-grained deformations and reasonable simulation speed. These meshes can be easily constructed in Blender by starting with a planar grid of vertices and then removing vertices and edges to create desired shapes. See Figure 2 for an illustration. Each vertex on the mesh has a global coordinate which we can query directly through Blender’s API, allowing for easily available ground truth information about various locations on the mesh and their pixel counterparts. See Figure 2 in the supplement for examples of these meshes. We can also simulate finer- grained manipulation of the mesh including grasps, pulls, and folds. To implement the action space defined in Section III, we first deproject the pixel corresponding to the pick point and map it to the vertex whose global coordinates are closest in $\mathbb{R}^{3}$ to the pixel’s deprojected coordinates. We then directly manipulate this vertex by pinning it and translating it over a sequence of 30 frames. This generalized form of manipulation allows us to easily execute experiments in simulation. See the supplement for further details on how we pin vertices, the overall technique and experiments in simulation. ## V Dense Shape Descriptor Training ### V-A Dense Object Descriptor Training Procedure We consider an environment with a deformable fabric on a flat tabletop and learn policies that perform smoothing and folding tasks. The policies we train use point-pair correspondences that are generated between overhead images of the fabric in different configurations. We generate deformation-invariant correspondences by training dense object descriptors [41, 6] on synthetically generated images of the fabric in different configurations. Figure 2: Fabric Meshes: Examples of the meshes generated in Blender for both square cloth (left) and t-shirts (right). The ground-truth vertices are highlighted in the second and fourth columns. In [6], an input image $I$ is mapped to a descriptor volume $Z=f_{\theta}(I)\in\mathbb{R}^{W\times H\times D}$, where each pixel $(i,j)$ has a corresponding descriptor vector $Z_{i,j}\in\mathbb{R}^{D}$. Descriptors are generated by a Siamese network $f_{\theta}$ and are guided closer together for corresponding pixels in images and pushed apart by at least some margin $M$ for non-corresponding pairs by minimizing a pixel-wise contrastive loss function during training [6]. Corresponding pairs of pixels represent the same point on an object. In this work, we also train a Siamese network to output descriptors that are close for corresponding pairs of pixels and separated for non-corresponding pixel pairs for overhead images of fabric in different configurations. Since ground-truth pixel correspondences are difficult to obtain in images across deformations of a real fabric, we train the network on synthetic RGB data from Blender (see Section IV), where perfect information about the pixel correspondences is available through the global coordinates of the fabric mesh’s vertices. Note that during training, the sampled image inputs to the Siamese network are of the same fabric type to ensure valid correspondences. That is, two different images of T-shirts can be passed into the network, but not a T-shirt and square fabric. Figure 3 demonstrates the pipeline for predicting descriptors for correspondence generation. The learned descriptors can then be used to approximate the correspondence function $d_{I\rightarrow I^{\prime}}$ described in Section III: $\displaystyle\left((i^{\prime\prime}_{l},j^{\prime\prime}_{l})\right)_{l=1}^{k}$ $\displaystyle=\operatorname*{arg\,min}_{(i^{\prime}_{1},j^{\prime}_{1})...(i^{\prime}_{k},j^{\prime}_{k})}\sum_{l=1}^{k}\lVert f_{\theta}(I)_{i,j}-f_{\theta}(I^{\prime})_{i^{\prime}_{l},j^{\prime}_{l}}\rVert_{2}$ $\displaystyle\text{s.t. }(i^{\prime}_{n},j^{\prime}_{n})\neq(i^{\prime}_{m},j^{\prime}_{m})\;\forall m,n\in[k]$ $\displaystyle d_{I\rightarrow I^{\prime}}(i,j)$ $\displaystyle=\operatorname*{arg\,min}_{(i^{\prime},j^{\prime})}\sum_{l=1}^{k}\lVert(i^{\prime},j^{\prime})-(i^{\prime\prime}_{l},j^{\prime\prime}_{l})\rVert_{2}$ The first equation computes the top $k$ pixel matches based on their distance in descriptor space and the second equation computes the geometric median of these matches in pixel space. In the physical setup, we experimentally found $k=20$ to give us the most robust predictions. Figure 3: Learning Visual Correspondences: pipeline for training dense object nets for robot fabric manipulation. Left: we train a dense correspondence network on pairs of simulated fabric images to learn pixel-wise correspondences using a pixel-wise contrastive loss. Right: we use the learned descriptors for policy optimization. We can use correspondence to map a reference action to a new fabric configuration. For example, we show an image of a wrinkled fabric in “State 2,” and we can use descriptors to figure out the action needed to smooth the fabric from “State 2” to “State 1.” ### V-B Dataset Generation and Domain Randomization To enable generalization of the learned descriptors to a range of fabric manipulation tasks, we generate a diverse dataset of initial fabric configurations. The first step simulates dropping the fabric onto the planar workspace while executing similar pinning actions to those described in Section IV on an arbitrary subset of vertices, causing some vertices to fall due to gravity while others stay fixed. We then release the pinned vertices 30 frames later so that they collapse on top of the fabric. This allows us to create realistic deformations in the mesh. We then export RGB images which serve as inputs to the Siamese network, pixel-wise annotations which gives us correspondences, and segmentation masks which allow us to sample matches on the fabric. Simulating soft-body animations is in general a computationally time-consuming process which makes it difficult to render large datasets in short periods of time. We take steps toward mitigating this issue by rendering 10 images per drop, allowing us to collect 10x as much data in the same time period. In simulation, we found that the test time pixel match error was unaffected when including these unsettled images of the fabric in the dataset. We additionally make use of domain randomization [32, 44] by rendering images of the scene while randomizing parameters including mesh size, lighting, camera pose, texture, color and specularity (see supplement for further details). We also restrict the rotation about the z-axis to be between $(-\pi/4,\pi/4)$ radians to reduce ambiguity during descriptor training due to the natural symmetry of fabrics such as squares. To randomize the image background, we sample an image from MSCOCO [21] and “paste” the rendered fabric mask on top. For our experiments, we generated one (domain-randomized) dataset, including both T-shirts and square fabric, and train a single model which we used for the experiments in Section VII. For reference, generating a single dataset of 7,500 images, half T-shirts and half square cloth, with 729 annotations per image takes approximately 2 hours on a 2.6GHz 6-core Intel Core i7 MacBook Pro. ## VI Descriptor-Parameterized Controller As discussed in Section III-B, the robot receives a demonstration of the task consisting of actions $\left(\mathbf{a}_{j}\right)_{j=1}^{n}$ and observations $\left(I_{j}\right)_{j=1}^{n}$. At execution time, the robot starts with the fabric in a different configuration, and the fabric itself may have a different texture or color. At time $j\in[n]$, the robot observes $I^{\prime}_{j}$ then executes $\pi_{j}(I_{j}^{\prime})=\left(d_{I_{j}\rightarrow I_{j}^{\prime}}(x_{g},y_{g})_{j},\;\;d_{I_{j}\rightarrow I_{j}^{\prime}}(x_{p},y_{p})_{j}\right)$ where $d_{I_{j}\rightarrow I_{j}^{\prime}}$ is defined in Section V-A. $\pi_{j}$ uses the geometric structure learned by the descriptor network to identify semantically relevant pixels in $I_{j}^{\prime}$ to generate actions that manipulate these keypoints. Because the descriptor network learns about task-agnostic fabric geometry, we train one network for a variety of tasks and use it across different fabric configurations from the ones supplied in demonstrations. Additionally, we do not require separate networks for each fabric type we wish to train. For example, one step of a task could involve grasping the top-right corner of the fabric and taking an action to place it in alignment with the bottom-left corner, thereby folding the fabric. The robot could receive an offline demonstration of this task on an initially flat fabric, but then be asked to perform the same task on a crumpled, rotated fabric. To do this, the robot must be able to identify the corresponding points in the new fabric configuration (top-right and bottom-left corners) and define a new action to align them. $\pi_{j}$ computes correspondences for the grasp and place points across the demonstration frame and the new observation to generate a corresponding action for the new configuration. ### VI-A Fabric Smoothing In the square fabric smoothing task, the robot starts with a crumpled fabric and spreads it into a smooth configuration on a planar workspace as in [36]. To complete this task, we use the approach from [36] and iterate over fabric corners, pulling each one to their target locations on an underlying plane. However, while [36] design a policy to do this using ground-truth knowledge of the fabric in simulation, we alternatively locate corners on the crumpled fabric using a learned descriptor network and a source image of a flat fabric where the corners are labeled. For the T-shirt smoothing task, we apply a similar method but instead iterate over the corners of the sleeves and the base of the T-shirt. ### VI-B Fabric Folding The fabric folding task involves executing a sequence of folds on a fairly smooth starting configuration. For each folding task, we use a single offline demonstration containing up to 4 pick and place actions, each defined by a pick and drop pixel location and collected by a human via a GUI. The descriptor-parameterized controller is then executed in an open-loop manner. Figure 4: Fabric Specifications: Images and dimensions of the square fabrics and shirts we use in experiments. ## VII Experiments We experimentally evaluate (1) the quality of the learned descriptors and their sensitivity to training parameters and (2) the performance of the descriptor-based policies from Section VI on two physical robotic systems, the da Vinci Research Kit (dVRK) [15] and the ABB YuMi. Results suggest that the learned descriptors and the resulting policies are robust to changes in fabric configuration and color. ### VII-A Tasks We consider 10 fabric manipulation tasks executed on a set of 5 T-shirts and 5 square fabrics in the real world: 1. 1. Single Fold (SF): A single fold where one corner is pulled to its opposing corner. 2. 2. Double Inward Fold (DIF): Two opposing corners are folded to the center of the fabric. 3. 3. Double Triangle Fold (DTF): Two sets of opposing corners are aligned with each other. 4. 4. Double Straight Fold (DSF): The square cloth is folded in half twice, first along the horizontal bisector and then along the vertical bisector. 5. 5. Four Corners Inward Fold (FCIF): All four corners are sequentially folded to the center of the cloth. 6. 6. T-Shirt Sleeves Fold (TSF): The two sleeves of a t-shirt are folded to the center of the shirt. 7. 7. T-Shirt Sleeve to Sleeve Fold (TSTSF): The left sleeve of a T-shirt is folded to the right sleeve of the T-shirt. 8. 8. Smoothing (S): Fabric is flattened from a crumpled state. 9. 9. Smoothing + Double Triangle Fold (SDTF): Fabric is smoothed then the DTF is executed. 10. 10. Smoothing + Sleeve to Sleeve Fold (SSTSF): T-shirt is smoothed then TSTSF is executed. All fabrics are varied either in dimension or color according to Figure 4. Additionally, we execute a subset of these tasks in simulation. A single visual demonstration consisting of up to 4 actions is provided to generate a policy which the robot then tries to emulate in the same number of actions. Figure 5: Policy Rollouts: We visualize policy execution on the YuMi for tasks 2, 3, 4, 5, 6 and 7 as described in Section VII-A. The first four columns show the folding instructions on some initial fabric and the last four columns show the corresponding folds executed on novel starting configurations for a different fabric. Figure 6: Full Folding Sequence: The first and second row is a time-lapse of a sequence of 6 actions taken by the YuMi and dVRK respectively, and with actions overlaid by red arrows, to successively smooth a wrinkled fabric and then fold it according to task 3 in Section VII-A. The third row is a time-lapse of a sequence of 5 actions taken by the YuMi to complete task 10 in Section VII-A. Here, robot actions are overlaid with blue arrows. ### VII-B Experimental Setup We execute fabric folding and smoothing experiments on the dVRK [15] and ABB YuMi robot. The dVRK is equipped with the Zivid OnePlus RGBD sensor that outputs $1900\times 1200$ pixel images at 13 FPS at depth resolution $0.5$ mm. The workspace of the dVRK is only $5"\times 5"$, so we use only square fabric of the same dimension while varying the color according to Figure 4. Manipulating small pieces of fabric into folds is challenging due to the elasticity of the fabric, so we add weight to the fabric by dampening it with water. Additionally, we place a layer of 1 inch foam rubber below the fabric to avoid damaging the gripper. The YuMi has a $36"\times 24"$ workspace, and since only one arm is utilized resulting in a more limited range of motion, we only manipulate at most $12"\times 12"$ pieces of fabric which we do not dampen. In this setup we use a 1080p Logitech webcam to collect overhead color images. For the YuMi, we use both T-shirts and square fabric of varying dimension and color but go no lower than $9"\times 9"$ fabrics due to its larger gripper. Finally, for both robots, we use a standard pixel to world calibration procedure to get the transformation from pixel coordinates to planar workspace coordinates. For both robots, we follow the same experimental protocol. We manually place the fabric in configurations similar to those shown in Figure 2 and deform them by pulling at multiple locations on the fabric. To obtain image input for the descriptor networks, we crop and resize the overhead image to be $485\times 485$ such that the fabric is completely contained within the image. Although lighting conditions, camera pose and workspace dimensions are significantly different between the two robotic systems, no manual changes are made to the physical setup. We find that the learned descriptors are sufficiently robust to handle this environmental variability. We evaluate the smoothing task by computing the coverage of the cropped workspace before and after execution. For the folding tasks, as in [19], we consider an outcome a success if the final state is visually consistent with the goal image. Conventional quantitative metrics such as intersection of union between the final state and a target image provide limited diagnostic information when starting configurations are significantly different as in the presented experiments. ### VII-C Results We evaluate the smoothing and folding policies on both the YuMi and dVRK on square fabrics and T-shirts. Table II shows the success rates of our method on all proposed tasks in addition to a breakdown of the failure cases detailed in Table III. We observe that the descriptor-parameterized controller is able to successfully complete almost all folding tasks at least $75\%$ of the time, and the smoothing policies are able to increase coverage of the cloth to over $83\%$ (Table I). The execution of the smoothing policy followed by the double triangle folding policy results in successful task completion $6/10$ and $8/10$ times on the YuMi and dVRK respectively. We find that the most frequent failure mode is an unsuccessful grasp of the fabric which is compounded for tasks that require more actions. Though this is independent of the quality of the learned descriptors, it highlights the need for more robust methods to grasp highly deformable objects. Task | Robot | Avg. Start Coverage | Avg. End Coverage ---|---|---|--- S | YuMi | $71.4\pm 6.2$ | $83.2\pm 8.1$ S | dVRK | $68.4\pm 4.4$ | $86.4\pm 5.2$ Table I: Physical Fabric Smoothing Experiments: We test the smoothing policies designed in Section VI on the YuMi and the dVRK robots. Both robots are able to increase coverage during the smoothing task by $11-22$ percent on average. Task | Robot | # Actions | Success | Error A | Error B | Error C ---|---|---|---|---|---|--- SF | YuMi | 1 | 18/20 | 2 | 0 | 0 SF | dVRK | 1 | 20/20 | 0 | 0 | 0 DIF | YuMi | 2 | 16/20 | 3 | 0 | 1 DIF | dVRK | 2 | 20/20 | 0 | 0 | 0 DTF | YuMi | 2 | 14/20 | 3 | 2 | 1 DTF | dVRK | 2 | 18/20 | 0 | 2 | 0 TSF | YuMi | 2 | 15/20 | 3 | 0 | 2 SDTF | YuMi | 6 | 6/10 | 2 | 1 | 1 SDTF | dVRK | 6 | 8/10 | 0 | 2 | 0 DSF | YuMi | 3 | 15/20 | 1 | 1 | 3 DSF | dVRK | 3 | 17/20 | 1 | 0 | 2 FCIF | YuMi | 4 | 13/20 | 5 | 1 | 1 FCIF | dVRK | 4 | 18/20 | 0 | 1 | 1 TSTSF | YuMi | 1 | 17/20 | 2 | 0 | 1 SSTSF | YuMi | 5 | 6/10 | 2 | 0 | 2 Table II: Physical Fabric Folding Experiments: We test the folding policies from Section VI on the YuMi and the dVRK. We observe both robots are able to perform almost all folding tasks at least $75$ percent of the time. The YuMi is able to perform the smoothing then folding task $6/10$ times and the dVRK is able to do so $8/10$ times. Error | Description ---|--- A | Gripper picks up more than one layer of fabric or fabric slips out of gripper due to inaccurate depth of grasp B | Pick or drop correspondence error greater than 30 pixels (10% of cloth width) or pick correspondence not on fabric mask C | Unintended physics: resulting fold does not hold due to variable stiffness of the fabric, friction of the fabric, or friction of the underlying plane Table III: Failure Mode Categorization ## VIII Discussion and Future Work We present an approach for multi-task fabric manipulation by using visual correspondences learned entirely in simulation. Experiments suggest that the learned correspondences are robust to different fabric colors, shapes, textures, and sizes and make it possible to efficiently learn 10 different fabric smoothing and folding tasks on two different physical robotic systems with no training in the real world. In future work, we plan to explore hierarchical fabric manipulation policies, where visual correspondences can be used to define coarse action plans while a closed loop controller can be learned to realize these plans. We will also explore more complex fabric manipulation tasks, such as wrapping rigid objects, in which reasoning about fabric dynamics is critical or tasks involving manipulating multiple fabrics simultaneously. ## References * [1] Benjamin Balaguer and Stefano Carpin “Combining Imitation and Reinforcement Learning to Fold Deformable Planar Objects” In _Proc. IEEE/RSJ Int. Conf. on Intelligent Robots and Systems (IROS)_ , 2011 * [2] Julia Borras, Guillem Alenya and Carme Torras “A Grasping-centered Analysis for Cloth Manipulation” In _arXiv:1906.08202_ , 2019 * [3] Blender Online Community “Blender - a 3D modelling and rendering package”, 2018 Blender Foundation URL: http://www.blender.org * [4] Andreas Doumanoglou, Andreas Kargakos, Tae-Kyun Kim and Sotiris Malassiotis “Autonomous Active Recognition and Unfolding of Clothes Using Random Decision Forests and Probabilistic Planning” In _Proc. IEEE Int. Conf. Robotics and Automation (ICRA)_ , 2014 * [5] Frederik Ebert et al. “Visual Foresight: Model-Based Deep Reinforcement Learning for Vision-Based Robotic Control” In _arXiv:1812.00568_ , 2018 * [6] Peter R. Florence, Lucas Manuelli and Russ Tedrake “Dense Object Nets: Learning Dense Visual Object Descriptors By and For Robotic Manipulation” In _Conf. on Robot Learning (CoRL)_ , 2018 * [7] Peter Florence, Lucas Manuelli and Russ Tedrake “Self-Supervised Correspondence in Visuomotor Policy Learning” In _IEEE Robotics & Automation Letters_, 2020 * [8] Jennifer Grannen* et al. “Learning Robot Policies for Untangling Dense Knots in Linear Deformable Structures”, https://goldberg.berkeley.edu/pubs/CoRL-2020-Untangling-submitted.pdf, 2020 * [9] Tuomas Haarnoja, Aurick Zhou, Pieter Abbeel and Sergey Levine “Soft Actor-Critic: Off-Policy Maximum Entropy Deep Reinforcement Learning with a Stochastic Actor” In _Proc. Int. Conf. on Machine Learning (ICML)_ , 2018 * [10] Ryan Hoque et al. “VisuoSpatial Foresight for Multi-Step, Multi-Task Fabric Manipulation” In _Proc. Robotics: Science and Systems (RSS)_ , 2020 * [11] Rishabh Jangir, Guillem Alenya and Carme Torras “Dynamic Cloth Manipulation with Deep Reinforcement Learning” In _Proc. IEEE Int. Conf. Robotics and Automation (ICRA)_ , 2020 * [12] Biao Jia, Zhe Hu, Jia Pan and Dinesh Manocha “Manipulating Highly Deformable Materials Using a Visual Feedback Dictionary” In _Proc. IEEE Int. Conf. Robotics and Automation (ICRA)_ , 2018 * [13] Biao Jia et al. “Cloth Manipulation Using Random-Forest-Based Imitation Learning” In _Proc. IEEE Int. Conf. Robotics and Automation (ICRA)_ , 2019 * [14] Dmitry Kalashnikov et al. “QT-Opt: Scalable Deep Reinforcement Learning for Vision-Based Robotic Manipulation” In _Conf. on Robot Learning (CoRL)_ , 2018 * [15] P Kazanzides et al. “An Open-Source Research Kit for the da Vinci Surgical System” In _Proc. IEEE Int. Conf. Robotics and Automation (ICRA)_ , 2014 * [16] Yasuyo Kita, Toshio Ueshiba, Ee Sian Neo and Nobuyuki Kita “A Method For Handling a Specific Part of Clothing by Dual Arms” In _Proc. IEEE/RSJ Int. Conf. on Intelligent Robots and Systems (IROS)_ , 2009 * [17] Yasuyo Kita, Toshio Ueshiba, Ee Sian Neo and Nobuyuki Kita “Clothes State Recognition Using 3D Observed Data” In _Proc. IEEE Int. Conf. Robotics and Automation (ICRA)_ , 2009 * [18] Oliver Kroemer, Scott Niekum and George Konidaris “A Review of Robot Learning for Manipulation: Challenges, Representations, and Algorithms” In _arXiv:1907.03146_ , 2019 * [19] Robert Lee et al. “Learning Arbitrary-Goal Fabric Folding with One Hour of Real Robot Experience” In _Conf. on Robot Learning (CoRL)_ , 2020 * [20] Sergey Levine et al. “Learning hand-eye coordination for robotic grasping with deep learning and large-scale data collection” In _The International Journal of Robotics Research_ 37.4-5, 2018, pp. 421–436 DOI: 10.1177/0278364917710318 * [21] Tsung-Yi Lin et al. “Microsoft COCO: Common Objects in Context” In _arXiv:1405.0312_ , 2014 * [22] Jeffrey Mahler et al. “Learning Ambidextrous Robot Grasping Policies” In _Science Robotics_ , 2019 * [23] Jeremy Maitin-Shepard, Marco Cusumano-Towner, Jinna Lei and Pieter Abbeel “Cloth Grasp Point Detection Based on Multiple-View Geometric Cues with Application to Robotic Towel Folding” In _Proc. IEEE Int. Conf. Robotics and Automation (ICRA)_ , 2010 * [24] Jan Matas, Stephen James and Andrew J. Davison “Sim-to-Real Reinforcement Learning for Deformable Object Manipulation” In _Conf. on Robot Learning (CoRL)_ , 2018 * [25] Stephen Miller et al. “A Geometric Approach to Robotic Laundry Folding” In _Int. Journal of Robotics Research (IJRR)_ , 2012 * [26] Fumiaki Osawa, Hiroaki Seki and Yoshitsugu Kamiya “Unfolding of Massive Laundry and Classification Types by Dual Manipulator” In _Journal of Advanced Computational Intelligence and Intelligent Informatics_ 11.5, 2007 * [27] J.. Parker, R. Dubey, F.. Paul and R.. Becker “Robotic Fabric Handling for Automating Garment Manufacturing” In _Journal of Manufacturing Science and Engineering_ 105, 1983 * [28] Lerrel Pinto and Abhinav Gupta “Supersizing Self-supervision: Learning to Grasp from 50K Tries and 700 Robot Hours” In _Proc. IEEE Int. Conf. Robotics and Automation (ICRA)_ , 2016 * [29] Xavier Provot “Collision and self-collision handling in cloth model dedicated to design garments” In _Computer Animation and Simulation’97_ Springer, 1997, pp. 177–189 * [30] Xavier Provot “Deformation constraints in a mass-spring model to describe rigid cloth behaviour” In _Graphics interface_ , 1995, pp. 147–147 Canadian Information Processing Society * [31] Stephane Ross, Geoffrey J Gordon and J Andrew Bagnell “A Reduction of Imitation Learning and Structured Prediction to No-Regret Online Learning” In _International Conference on Artificial Intelligence and Statistics (AISTATS)_ , 2011 * [32] Fereshteh Sadeghi and Sergey Levine “CAD2RL: Real Single-Image Flight without a Single Real Image” In _Proc. Robotics: Science and Systems (RSS)_ , 2017 * [33] Jose Sanchez, Juan-Antonio Corrales, Belhassen-Chedli Bouzgarrou and Youcef Mezouar “Robotic Manipulation and Sensing of Deformable Objects in Domestic and Industrial Applications: a Survey” In _Int. Journal of Robotics Research (IJRR)_ , 2018 * [34] Tanner Schmidt, Richard A. Newcombe and Dieter Fox “Self-Supervised Visual Descriptor Learning for Dense Correspondence” In _IEEE Robotics and Automation Letters_ 2, 2017, pp. 420–427 * [35] John Schulman, Jonathan Ho, Cameron Lee and Pieter Abbeel “Learning from Demonstrations Through the Use of Non-Rigid Registration” In _Int. S. Robotics Research (ISRR)_ , 2013 * [36] Daniel Seita et al. “Deep Imitation Learning of Sequential Fabric Smoothing From an Algorithmic Supervisor” In _Proc. IEEE/RSJ Int. Conf. on Intelligent Robots and Systems (IROS)_ , 2020 * [37] Daniel Seita et al. “Deep Transfer Learning of Pick Points on Fabric for Robot Bed-Making” In _Int. S. Robotics Research (ISRR)_ , 2019 * [38] Changyeob Shin et al. “Autonomous Tissue Manipulation via Surgical Robot Using Learning Based Model Predictive Control” In _Proc. IEEE Int. Conf. Robotics and Automation (ICRA)_ , 2019 * [39] Li Sun et al. “A Heuristic-Based Approach for Flattening Wrinkled Clothes” In _Towards Autonomous Robotic Systems. TAROS 2013. Lecture Notes in Computer Science, vol 8069_ , 2014 * [40] Li Sun, Gerardo Aragon-Camarasa, Simon Rogers and J. Siebert “Accurate Garment Surface Analysis using an Active Stereo Robot Head with Application to Dual-Arm Flattening” In _Proc. IEEE Int. Conf. Robotics and Automation (ICRA)_ , 2015 * [41] Priya Sundaresan et al. “Learning Rope Manipulation Policies using Dense Object Descriptors Trained on Synthetic Depth Data” In _Proc. IEEE Int. Conf. Robotics and Automation (ICRA)_ , 2020 * [42] Brijen Thananjeyan et al. “Safety Augmented Value Estimation from Demonstrations (SAVED): Safe Deep Model-Based RL for Sparse Cost Robotic Tasks” In _IEEE Robotics & Automation Letters_, 2020 * [43] Brijen Thananjeyan et al. “Multilateral Surgical Pattern Cutting in 2D Orthotropic Gauze with Deep Reinforcement Learning Policies for Tensioning” In _Proc. IEEE Int. Conf. Robotics and Automation (ICRA)_ , 2017 * [44] Josh Tobin et al. “Domain Randomization for Transferring Deep Neural Networks from Simulation to the Real World” In _Proc. IEEE/RSJ Int. Conf. on Intelligent Robots and Systems (IROS)_ , 2017 * [45] Eric Torgerson and Fanget Paul “Vision Guided Robotic Fabric Manipulation for Apparel Manufacturing” In _Proc. IEEE Int. Conf. Robotics and Automation (ICRA)_ , 1987 * [46] Bryan Willimon, Stan Birchfield and Ian Walker “Model for Unfolding Laundry using Interactive Perception” In _Proc. IEEE/RSJ Int. Conf. on Intelligent Robots and Systems (IROS)_ , 2011 * [47] Yilin Wu et al. “Learning to Manipulate Deformable Objects without Demonstrations” In _RSS_ , 2020 * [48] Wilson Yan, Ashwin Vangipuram, Pieter Abbeel and Lerrel Pinto “Learning Predictive Representations for Deformable Objects Using Contrastive Estimation” In _Conf. on Robot Learning (CoRL)_ , 2020 * [49] Kevin Zakka, Andy Zeng, Johnny Lee and Shuran Song “Form2Fit: Learning Shape Priors for Generalizable Assembly from Disassembly” In _Proc. IEEE Int. Conf. Robotics and Automation (ICRA)_ , 2020 ## IX Appendix The appendix is organized as follows: * • Appendix A contains additional details on the fabric simulator * • Appendix B contains additional details on the experiments conducted in simulation * • Appendix C contains results for simulation experiments * • Appendix D shows visualizations of the learned descriptor mappings. * • Appendix E contains images from additional physical trials executed on the dVRK and YuMi. * • Appendix F conducts a detailed study on the effect of various hyperparameters on the quality of the learned visual correspondences. ### IX-A Fabric Simulator Details We use Blender 2.8 to both create dynamic cloth simulations and to render images of the fabric in different configurations. As can be seen in Figure 2, we are able to retrieve the world coordinates of each vertex via Blender’s API which we then use to find ground truth pixel correspondences through an inverse camera to world transformation. This allows us to create dense pixel- vertex annotations along the surface of the fabric which we feed to into the descriptor training procedure. Figure 9 is a visualization of the learned descriptors and Figure 7 contains examples of the domain randomized training data we generate through Blender. #### IX-A1 Fabric Model To generate the square cloth in Blender, we first import a default square mesh and subdivide it three times to create a grid of $27\times 27$ grid of vertices. We found that this number of square vertices resulted in a visually realistic animation in comparison to our real fabrics. We additionally add 0.02 meter thickness to the cloth to increase its weight which creates more realistic collision physics. In order to apply Blender’s in-built cloth physics to the mesh, we simply make use of the cloth physics modifier through which we are able modify the parameters shown in Table IV. Internally, Blender simulates fabric physics for polygonal meshes with gravitational forces, damping, stiffness, and by interconnecting the mesh vertices with four types of virtual springs: tension springs, compression springs, shear springs, and angular bending springs. Each vertex also exerts repulsive forces within a self-contained virtual sphere on vertices both within fabric and in surrounding objects, to simulate self-collisions and collisions with other objects. We visually tune the simulator by replaying a fabric folding action while varying parameters, most notably the friction coefficients and spring elasticity constants. From observing videos of the folding actions, we settle on the parameter values specified in Table IV. A visualization of these steps can be seen in the top row of Figure 8. To generate the t-shirt mesh, we similarly import a default square mesh and subdivide it three times, but also delete all vertices that do not lie in a predefined t-shirt cutout of the square mesh which results in the bottom right image of Figure 2. #### IX-A2 Manipulation with Hook Objects We utilize hook objects to take actions in the Blender simulator. A hook object attaches to a mesh vertex and exerts a proportional sphere of influence over the selected vertex and those in its vicinity, pulling the fabric in the direction of movement. We simulate a grasp, drag, and drop of the fabric by assigning a hook object to a fabric vertex, moving this hook over a series of frames to the deprojected pixel drop location, and removing the hook object assignment to release the cloth. Figure 7: Examples of domain-randomized images of the starting fabric states encountered in the dataset generation phase described in Section V-B. The first two columns show examples of images with a square fabric, and the last two columns show similar examples but with a t-shirt. Figure 8: The top row illustrates the the process of creating cloth in Blender from a default square mesh. The second row is an example of a starting configuration generated by dropping the cloth from a fixed height and pinning a single arbitrary vertex. The pinned vertex is labeled by the red circle. The third row illustrates frames from a folding action in the simulator and the last row shows the corresponding rendered images of the settled cloth before and after the action. Table IV: Blender Cloth Simulation Parameters Parameter | Explanation | Value ---|---|--- Quality Steps | quality of cloth stability and collision response | 5.0 Speed Multiplier | how fast simulation progresses | 1.0 Cloth Mass (kg) | – | 0.3 Air Viscosity | air damping | 1.0 Tension Springs | tension damping/stretching | 5.0 Compression Springs | compression damping/stretching | 5.0 Shear Springs | damping of shear behavior | 5.0 Bending Springs | damping of bending behavior | 0.5 Friction | friction with self-contact | 5 Self-Collision Distance (m) | per-vertex spherical radius for repulsive forces | 0.015 Figure 9: Visualization of the 3-dimensional descriptors learned via the training procedure described in Section V by mapping each pixel’s descriptor vector to an RGB vector. Thus, similar colors across the images of columns two and four represent corresponding points on the square cloth. Figure 10: Additional rollouts of the smoothing task from randomly chosen starting configurations. The learned descriptors are used to locate the corners of the fabric and successively pull them to a reference location in an image of the flat cloth. Figure 11: Additional rollouts of the smoothing task from randomly chosen starting configurations. The learned descriptors are used to locate the corners of the fabric and successively pull them to a reference location in an image of the flat cloth. Figure 12: Additional rollouts of the folding tasks described in Section VII-A from arbitrary starting configurations. The left column, center column and right column contain results for tasks 4, 3 and 2 respectively. #### IX-A3 Starting Configurations and Actions To generate varied starting configurations, we simulate dropping the fabric from 0.2 meters above the workspace while pinning it an arbitrary subset of the 729 vertices. After 30 frames in the animation, the pinned vertices are released and are allowed to settle for another 30 frames. This creates natural deformation in the cloth and introduces a wide range of starting configurations to the training dataset. A sequence of these steps is shown in the second row of Figure 8. When running simulated experiments, taking pick and place actions requires manipulating the cloth via hook objects as defined in Section IV. A sequence of frames throughout the course of an action using a hook object as well as the corresponding rendered frames are shown in the last two rows of Figure 8. ### IX-B Simulation Experiment Details In simulation, we conduct 50 trials of the first 4 folding tasks described in VII-A on a domain randomized test set generated as described in V-B. We consider an outcome a success if the final state is visually consistent with the target image. We additionally declared a failure when the planned pick and drop pixels were more than 50 pixels away from their correct ground truth locations which we had access to in Blender. Note that this is neither a sufficient nor necessary condition for a successful fold, but nevertheless serves as a decent heuristic. While we considered more quantitative metrics such as structural similarity between the target image and the final state and summed distance between corresponding vertices on the mesh, these metrics are insufficient when the test time starting configuration is significantly different from the demonstration configuration. ### IX-C Simulation Experiment Results We evaluate the folding policies designed in Section VI in the simulated fabric environment. The policies successfully complete the tasks $84$ to $96$ percent of the time (Table V). Task | Success Rate ---|--- Single Fold | 46/50 Double Inward Fold | 48/50 Double Triangle Fold | 42/50 T-Shirt Sleeves Fold | 44/50 Table V: Simulated Fabric Folding Experiments: We observe that the system is able to successfully complete the tasks $84$ to $96$ percent of the time in simulation. Success is determined by visual inspection of the cloth after the sequence of actions is executed. Example simulation rollouts are shown in Figure 14. ### IX-D Descriptor Mapping Visualizations We present the descriptor volumes produced by a model trained to output 3-dimensional descriptors. We coarsely visualize the volumes by presenting them as RGB images (Figure 9), and observe that corresponding pixels of the cloth map to similar colors in the descriptor volumes across configurations. ### IX-E Physical Trial Trajectories In this section, we present additional trials of the physical experiments conducted using the descriptor-based policies for smoothing (Figure 10, Figure 11) and folding (Figure 12). ### IX-F Descriptor Quality Ablations Figure 13: Ablation studies: We study the sensitivity of the learned dense object descriptors as described in Sections V and IX-F to training parameters. Starting from top left, and proceeding clockwise, we test the effect of testing on RGB vs depth images, on the descriptor dimension (either 3, 9, or 16), on the number of ground truth annotations, and whether domain randomization is used. All results are evaluated using pixel match error on a held-out set of image pairs. Figure 14: Simulation Policy Visualization: Visualization of the policy executed in simulation (with Blender) using learned descriptors for folding tasks 2 and 4 described in VII-A. The first two columns show the corresponding folding instructions from a web interface (pick-and-place actions shown with red arrows) for tasks 2 and 4. The third column shows images of the previously unseen initial configurations of fabrics before the actions, while the last two columns show the result of executing descriptor-parameterized actions. Results suggest that the learned descriptors can be used to successfully perform a variety of folding tasks from varying initial configurations. To investigate the quality of the learned descriptors with training process described in Section V, we perform four sets of ablation studies. We evaluate the quality of learned descriptors in a manner similar to [41] by evaluating the $\ell_{2}$ pixel distance of the pixel match error on a set of 100 pairs of held-out validation set images, where for each we sample 100 pixel pairs. We study the effect of training descriptors on (1) RGB or depth images, (2) using descriptor dimension 3, 9, or 16, (3) using 200, 450, or 700 ground- truth annotated images, and (4) whether domain randomization is used or not. Results suggest that the learned descriptors are best with RGB data, with descriptor dimension between 3 and 16 and with domain randomization, though the performance is generally insensitive to the parameter choices, suggesting a robust training procedure. Based on these results, we use RGB images with domain randomization, and with descriptor dimension 3 for all simulated experiments for both the t-shirt and square fabric. We use RGB, domain- randomized, 9-dimensional descriptors for real fabric experiments. See Figure 13 for plots.
# Path integral control under McKean-Vlasov dynamics Timothy Bennett111Corresponding author<EMAIL_ADDRESS>222Department of Mathematics and Statistics, University of South Alabama, Mobile, AL, 36688, United States. ###### Abstract We investigate the complexities of the McKean–Vlasov optimal control problem, exploring its various formulations such as the strong and weak formulations, as well as both Markovian and non-Markovian setups within financial markets. Furthermore, we examine scenarios where the law governing the control process impacts the dynamics of options. By conceptualizing controls as probability measures on a fitting canonical space with filtrations, we unlock the potential to devise classical measurable selection methods, conditioning strategies, and concatenation arguments within this innovative framework. These tools enable us to establish the dynamic programming principle under a wide range of conditions. ## Introduction: The Black-Scholes model stands as the quintessential example of financial models, embodying key principles that have influenced subsequent developments. Its widespread adoption reflects the common adjustments and trade-offs necessary to translate theoretical concepts into practical tools for financial analysis and decision-making. Let us dive deep into the core principles of this model. The main aim of creating the Black-Scholes model is to evaluate and manage options, operating on the premise that the volatility of the underlying asset remains constant (Fournié et al.,, 1997). Building on this assumption, it provides an analytical formula for calculating the value of a European option, which relies on the volatility parameter of the system. However, the model’s main drawback arises from the fact that in actuality, volatility fluctuates, and the market does not necessarily conform to the values predicted by this model (Pramanik,, 2020). However, practitioners in financial markets have widely embraced the Black- Scholes model, albeit with some modifications that deviate from the original theoretical expectations. These adjustments include anticipating negotiations for transactions in terms of implicit volatility rather than option prices, implementing theoretical delta hedging as recommended by the model, and recognizing that the implicit volatility tends to be higher for options further from the money and closer to maturity (Pramanik, 2021c, ; Pramanik and Polansky,, 2021, 2023). Consequently, the Black-Scholes model retains significant practical utility, although its role differs from the primary hypotheses outlined in the original theory. While one might assume that the compromises and adaptations made in utilizing the Black-Scholes model are unique to this particular context, we argue that they offer broader insights applicable across various fields beyond finance. So far in our initial discussion, we overlooked the inclusion of control theory. Control theory encompasses two primary forms: open loop and closed loop control. In open loop control, the decision of the policymaker remains uninfluenced by the state variable’s information. Conversely, closed loop control entails the policymaker’s decision being contingent upon all available state variable information (Pramanik, 2024a, ). Integrating control variables is crucial in understanding the dynamics of financial markets. For instance, consider an entrepreneur aiming to maximize long-term profits, assuming control over advertising expenditure. If they increase spending on commercials presently, they may afford a celebrity endorsement, thereby boosting sales probability in subsequent periods (Pramanik and Polansky,, 2020). Conversely, heightened sales in the current period provide more funds for future commercial investment. This exemplifies closed loop control within financial markets. Furthermore, all of theses financial models are not always linear stochastic differential equations (SDEs) in practice. Therefore, the importance of non-linear SDEs come into place. One such SDE is McKean-Vlasov (McKean Jr,, 1966) system. The next section gives an overview of it. ## Overview of McKean-Vlasov SDEs: McKean-Vlasov type SDEs are commonly known as nonlinear SDEs, with _non- linear_ highlighting the potential dependence of coefficients on the solutions’ marginal distributions rather than the state variable itself. The distinctive aspect of these coefficients, even when deterministic, leads to non-local feedback effects, precluding the conventional Markov property (Carmona and Delarue,, 2015). Incorporating the marginal distribution into the system’s state could restore the Markov property but at the expense of significantly increasing the complexity by introducing a probability measure, rendering the system infinite-dimensional. While analyzing the infinitesimal generator can utilize tools developed for infinite-dimensional differential operators, conventional differential calculus, even in infinite dimensions, struggles to ensure the second component of the state process matches the statistical distribution of the first component (Carmona and Delarue,, 2015). Nevertheless, pathwise analysis complexity remains manageable and can be effectively addressed using standard stochastic calculus tools. This points towards devising a wholly probabilistic approach to address optimal control in these nonlinear systems. The optimal control of systems governed by McKean-Vlasov SDEs appears to be a relatively novel challenge, curiously overlooked in stochastic control literature. Typically, solving a McKean-Vlasov SDE involves employing a fixed- point argument: Initially, a set of potential marginal distribution candidates is established, and then the resulting standard SDE is solved, with the fixed- point argument entailing the requirement that the solution’s marginal distributions match the initially chosen ones. Introducing stochastic control introduces an additional layer of optimization atop this fixed-point process. This formulation shares many similarities with the mean-field game (MFG) problem, as initially proposed by Lasry and Lions, (2007). In Carmona et al., (2013), the similarities and disparities between these two problems are identified and discussed. It is notably emphasized that optimizing first and subsequently seeking a fixed point yields the solution to a mean-field game problem, while first determining the fixed point and then optimizing leads to the resolution of the optimal control for McKean-Vlasov SDEs (Polansky and Pramanik,, 2021; Pramanik, 2024b, ). The solutions to both problems describe equilibrium states of large populations whose interactions and objectives follow mean-field patterns. The nuances between these equilibrium concepts are subtle and hinge on the optimization component’s formulation within the equilibrium model. Linear quadratic models are presented in Carmona et al., (2013) to offer illustrative examples of these distinctions. As McKean-Vlasov SDEs exhibit non-Markovian dynamics, it is logical to tackle optimization through a tailored version of the Pontryagin stochastic maximum principle, rather than forcing fit with the Hamilton-Jacobi-Bellman (HJB) framework. The stochastic Pontryagin method hinges on introducing and manipulating adjoint processes, which are solutions to adjoint backward stochastic differential equations (BSDEs). These equations involve partial derivatives of the Hamiltonian function concerning the state variable (Carmona and Delarue,, 2015). However, in the context of McKean-Vlasov SDEs, the marginal distributions of solutions serve as full-fledged variables of the Hamiltonian function, necessitating differentiation in the pursuit of critical points. This discrepancy likely contributes to the impasse in existing literature. Until now, only dynamics contingent on moments of the marginal distributions have been explored, where differentiability with respect to the measure can be achieved through conventional calculus chain rules (Carmona and Delarue,, 2015). The right notation of differentiability was introduced in Lasry and Lions, 2006a and Lasry and Lions, 2006b . ## Pontryagin principle and Path integral control: The stochastic Pontryagin principle stands as a potent tool, yet its illuminating insights often necessitate restrictive assumptions on the models. For instance, findings in Carmona and Delarue, (2015) lean on a set of technical prerequisites that confine the scope of models to those characterized by coefficients primarily linear in state, control, and measure variables, alongside costs that exhibit convexity concerning the state and control variables. While these conditions may seem limiting, they are typical in applications of the stochastic Pontryagin principle to control problems. Notably, assuming the convexity of the control space is done for simplicity’s sake. More extensive spaces could be accommodated, albeit at the expense of employing spike variation techniques and introducing an additional adjoint equation. However, Carmona and Delarue, (2015) opted against pursuing this broader generality, as it would entail added complexity in notation and technicalities, potentially obscuring the core focus of their research, especially without specific application-driven motivation (Pramanik, 2022a, ; Pramanik, 2022b, ; Pramanik, 2023b, ; Pramanik, 2023c, ). The essential aspect of the stochastic Pontryagin principle involves seeking within the control set for a potential minimizer of the Hamiltonian (often referred to as satisfying the Isaacs condition), while the complementary aspect suggests incorporating the minimizer’s formula into both the forward equation governing dynamics and the adjoint backward equation describing adjoint processes (Hua et al.,, 2019; Pramanik,, 2016; Pramanik, 2021b, ). The inclusion of the minimizer in these equations establishes a significant interconnection between the forward and backward equations, effectively reducing the control problem solution to that of a forward-backward stochastic differential equation (FBSDE). Applying this approach to the current scenario entails investigating FBSDEs where the marginal distributions of solutions emerge in the coefficients of the equations. Carmona and Delarue, (2015) term these equations “mean-field FBSDEs” or “FBSDEs of McKean-Vlasov type,” asserting that their study had not been undertaken prior to their research. They noted that while a general existence result was proposed in Carmona and Delarue, (2013), one of the assumptions therein prevents its application to the linear quadratic model. A specific variant of the stochastic Pontryagin principle, such as the Feynman-type path integral method introduced by Pramanik, (2020), can be utilized and further developed by Pramanik and Polansky, 2024a to obtain an analytical solution for the system in question (Pramanik and Polansky, 2024b, ). When the state variable is high-dimensional and the system dynamics are nonlinear, as seen in Merton-Garman-Hamiltonian SDEs, numerical construction of an HJB equation becomes extremely difficult. The Feynman-type path integral approach effectively overcomes this challenge of dimensionality and provides a localized analytical solution. To employ this approach, we first establish a stochastic Lagrangian for each continuous time point within the interval $s\in[0,t]$, where $t>0$. Subsequently, we divide this entire time span into $n$ equal-length subintervals and define a Riemann measure corresponding to the state variable for each subinterval. After constructing a Euclidean action function, we obtain a Schroedinger-type equation via Wick rotation. By enforcing the first-order conditions with respect to both the state and control variables, we determine the solution of the system. This methodology holds potential application in cancer research (Dasgupta et al.,, 2023; Hertweck et al.,, 2023; Kakkat et al.,, 2023; Khan et al.,, 2023; Vikramdeo et al.,, 2023; Khan et al.,, 2024). ## Further impact: We gain insight into the challenge of optimal control for mean-field stochastic differential equations, also known as McKean–Vlasov stochastic differential equations in academic literature. This problem presents a stochastic control scenario where the state process is driven by SDEs with coefficients influenced by the present time, the trajectories of the state process, and its distribution (or conditional distribution in cases involving shared noise). Likewise, the reward functionals can be influenced by the distribution of the state process (Pramanik, 2023a, ). A common limitation of the aforementioned results is their general requirement for some Markovian property of the system or its distribution, along with strong regularity assumptions on the coefficients and reward functions under consideration. This might come as a surprise to those familiar with the classical stochastic Pontryagin principle (Pramanik, 2021a, ). In stochastic control problems, it is feasible to utilize measurable selection arguments to derive the stochastic Pontryagin principle without necessitating more than mild measurability assumptions (Pramanik and Maity,, 2024). Typically, two crucial elements are needed to establish the dynamic programming principle: first, ensuring the stability of controls concerning conditioning and concatenation, and second, the measurability of the associated value function. Measurable selection arguments offer a framework to validate the conditioning, concatenation, and measurability requirements of the associated value function without relying on strong assumptions. Through the concept of relaxed control, interpreting a control as a probability measure on a canonical space, and leveraging martingale problem theory, researchers have demonstrated a stochastic Pontryagin principle using straightforward and transparent reasoning (Djete et al.,, 2022). In the future, we aim to explore scenarios where the drift and diffusion components, as well as the reward functions in option markets, can depend on the joint conditional distribution of the state process path and the closed-loop control. ## References * Carmona and Delarue, (2013) Carmona, R. and Delarue, F. (2013). Mean field forward-backward stochastic differential equations. Electronic Communications in Probability, 18:1–15. * Carmona and Delarue, (2015) Carmona, R. and Delarue, F. (2015). Forward–backward stochastic differential equations and controlled mckean–vlasov dynamics. The Annals of Probability, pages 2647–2700. * Carmona et al., (2013) Carmona, R., Delarue, F., and Lachapelle, A. (2013). Control of mckean–vlasov dynamics versus mean field games. Mathematics and Financial Economics, 7:131–166. * Dasgupta et al., (2023) Dasgupta, S., Acharya, S., Khan, M. A., Pramanik, P., Marbut, S. M., Yunus, F., Galeas, J. N., Singh, S., Singh, A. P., and Dasgupta, S. (2023). Frequent loss of cacna1c, a calcium voltage-gated channel subunit is associated with lung adenocarcinoma progression and poor prognosis. Cancer Research, 83(7_Supplement):3318–3318. * Djete et al., (2022) Djete, M. F., Possamaï, D., and Tan, X. (2022). Mckean–vlasov optimal control: the dynamic programming principle. The Annals of Probability, 50(2):791–833. * Fournié et al., (1997) Fournié, E., Lasry, J.-M., and Lions, P.-L. (1997). Some nonlinear methods for studying far-from-the-money contingent claims. Publ. Newton Inst, pages 115–45. * Hertweck et al., (2023) Hertweck, K. L., Vikramdeo, K. S., Galeas, J. N., Marbut, S. M., Pramanik, P., Yunus, F., Singh, S., Singh, A. P., and Dasgupta, S. (2023). Clinicopathological significance of unraveling mitochondrial pathway alterations in non-small-cell lung cancer. The FASEB Journal, 37(7):e23018. * Hua et al., (2019) Hua, L., Polansky, A., and Pramanik, P. (2019). Assessing bivariate tail non-exchangeable dependence. Statistics & Probability Letters, 155:108556. * Kakkat et al., (2023) Kakkat, S., Pramanik, P., Singh, S., Singh, A. P., Sarkar, C., and Chakroborty, D. (2023). Cardiovascular complications in patients with prostate cancer: Potential molecular connections. International Journal of Molecular Sciences, 24(8):6984. * Khan et al., (2023) Khan, M. A., Acharya, S., Anand, S., Sameeta, F., Pramanik, P., Keel, C., Singh, S., Carter, J. E., Dasgupta, S., and Singh, A. P. (2023). Myb exhibits racially disparate expression, clinicopathologic association, and predictive potential for biochemical recurrence in prostate cancer. Iscience, 26(12). * Khan et al., (2024) Khan, M. A., Acharya, S., Kreitz, N., Anand, S., Sameeta, F., Pramanik, P., Keel, C., Singh, S., Carter, J., Dasgupta, S., et al. (2024). Mp60-05 myb exhibits racially disparate expression and clinicopathologic association and is a promising predictor of biochemical recurrence in prostate cancer. The Journal of Urology, 211(5S):e1000. * (12) Lasry, J.-M. and Lions, P.-L. (2006a). Jeux à champ moyen. i–le cas stationnaire. Comptes Rendus Mathématique, 343(9):619–625. * (13) Lasry, J.-M. and Lions, P.-L. (2006b). Jeux à champ moyen. ii–horizon fini et contrôle optimal. Comptes Rendus. Mathématique, 343(10):679–684. * Lasry and Lions, (2007) Lasry, J.-M. and Lions, P.-L. (2007). Mean field games. Japanese journal of mathematics, 2(1):229–260. * McKean Jr, (1966) McKean Jr, H. P. (1966). A class of markov processes associated with nonlinear parabolic equations. Proceedings of the National Academy of Sciences, 56(6):1907–1911. * Polansky and Pramanik, (2021) Polansky, A. M. and Pramanik, P. (2021). A motif building process for simulating random networks. Computational Statistics & Data Analysis, 162:107263. * Pramanik, (2016) Pramanik, P. (2016). Tail non-exchangeability. Northern Illinois University. * Pramanik, (2020) Pramanik, P. (2020). Optimization of market stochastic dynamics. In SN Operations Research Forum, volume 1, page 31. Springer. * (19) Pramanik, P. (2021a). Consensus as a nash equilibrium of a stochastic differential game. arXiv preprint arXiv:2107.05183. * (20) Pramanik, P. (2021b). Effects of water currents on fish migration through a feynman-type path integral approach under 8/3 liouville-like quantum gravity surfaces. Theory in Biosciences, 140(2):205–223. * (21) Pramanik, P. (2021c). Optimization of dynamic objective functions using path integrals. PhD thesis, Northern Illinois University. * (22) Pramanik, P. (2022a). On lock-down control of a pandemic model. arXiv preprint arXiv:2206.04248. * (23) Pramanik, P. (2022b). Stochastic control of a sir model with non-linear incidence rate through euclidean path integral. arXiv preprint arXiv:2209.13733. * (24) Pramanik, P. (2023a). Optimal lock-down intensity: A stochastic pandemic control approach of path integral. Computational and Mathematical Biophysics, 11(1):20230110. * (25) Pramanik, P. (2023b). Path integral control in infectious disease modeling. arXiv preprint arXiv:2311.02113. * (26) Pramanik, P. (2023c). Path integral control of a stochastic multi-risk sir pandemic model. Theory in Biosciences, 142(2):107–142. * (27) Pramanik, P. (2024a). Dependence on tail copula. J, 7(2):127–152. * (28) Pramanik, P. (2024b). Estimation of optimal lock-down and vaccination rate of a stochastic sir model: A mathematical approach. European Journal of Statistics, 4:3–3. * Pramanik and Maity, (2024) Pramanik, P. and Maity, A. K. (2024). Bayes factor of zero inflated models under jeffereys prior. arXiv preprint arXiv:2401.03649. * Pramanik and Polansky, (2020) Pramanik, P. and Polansky, A. M. (2020). Motivation to run in one-day cricket. arXiv preprint arXiv:2001.11099. * Pramanik and Polansky, (2021) Pramanik, P. and Polansky, A. M. (2021). Optimal estimation of brownian penalized regression coefficients. arXiv preprint arXiv:2107.02291. * Pramanik and Polansky, (2023) Pramanik, P. and Polansky, A. M. (2023). Scoring a goal optimally in a soccer game under liouville-like quantum gravity action. In Operations Research Forum, volume 4, page 66. Springer. * (33) Pramanik, P. and Polansky, A. M. (2024a). Optimization of a dynamic profit function using euclidean path integral. SN Business & Economics, 4(1):8. * (34) Pramanik, P. and Polansky, A. M. (2024b). Semicooperation under curved strategy spacetime. The Journal of Mathematical Sociology, 48(2):172–206. * Vikramdeo et al., (2023) Vikramdeo, K. S., Anand, S., Sudan, S. K., Pramanik, P., Singh, S., Godwin, A. K., Singh, A. P., and Dasgupta, S. (2023). Profiling mitochondrial dna mutations in tumors and circulating extracellular vesicles of triple-negative breast cancer patients for potential biomarker development. FASEB BioAdvances, 5(10):412.
# Chiral valley phonons and flat phonon bands in moiré materials Indrajit Maity, Arash A. Mostofi Johannes Lischner<EMAIL_ADDRESS>Departments of Materials and Physics and the Thomas Young Centre for Theory and Simulation of Materials, Imperial College London, South Kensington Campus, London SW7 2AZ, UK ###### Abstract We investigate the chirality of phonon modes in twisted bilayer $\mathrm{WSe_{2}}$ and demonstrate distinct chiral behavior of the $K/K^{\prime}$ valley phonons for twist angles close to $0^{\circ}$ and close to $60^{\circ}$. In particular, multiple chiral non-degenerate $K/K^{\prime}$ valley phonons are found for twist angles near $60^{\circ}$ whereas no non- degenerate chiral modes are found for twist angles close to $0^{\circ}$. Moreover, we discover two sets of emergent chiral valley modes that originate from an inversion symmetry breaking at the moiré scale and find similar modes in moiré patterns of strain-engineered bilayers $\mathrm{WSe_{2}}$ and $\mathrm{MoSe_{2}/WSe_{2}}$ heterostructures. At the energy gap between acoustic and optical modes, the formation of flat phonon bands for a broad range of twist angles is observed in twisted bilayer $\mathrm{WSe_{2}}$. Our findings are relevant for understanding electron-phonon and exciton-phonon scattering in moiré materials and also for the design of phononic analogues of flat band electrons. Introduction. Chirality - the characteristic of an object that can be distinguished from its mirror image - plays a fundamental role in physics, chemistry, and biology. For example, different enantiomers of chiral molecules absorb different amounts of right-handed and left-handed polarized light, which enables their spectroscopic characterization through measurement of the circular dichroism [1]. In condensed matter physics, the chirality of electrons gives rise to many exotic effects, such as Klein tunneling [2] in graphene, and the chiral magnetic effect in three-dimensional semi-metals [3, 4]. Recently, the study of chirality in phonons has attracted significant interest. For example, Zhang and Niu [5] predicted the existence of chiral phonons at the $K$ and $K^{\prime}$ valleys of monolayer $\mathrm{WSe_{2}}$, which was subsequently verified by experiments that measured the circular dichroism of phonon-assisted intervalley transitions of holes [6]. The interaction of such chiral valley phonons with other quasiparticles is relevant for understanding and controlling many electronic and optical phenomena [7, 8, 9, 10, 11, 12, 13]. For instance, Li et al. demonstrated that the coupling between a chiral valley phonon and an intervalley exciton in monolayer $\mathrm{WSe_{2}}$ can lead to a long exciton lifetime maintaining the valley polarization, which is important for valley-excitonics [12]. Chen et al. observed the entanglement of chiral phonon modes in monolayer $\mathrm{WSe_{2}}$ and single photons emitted from an embedded quantum dot in the material, which promises to reveal new avenues for phonon-driven entanglement of quantum dots [14]. Besides monolayer $\mathrm{WSe_{2}}$, chiral phonons have been found in other materials [15, 16, 17, 18, 19]. A particularly promising platform for observing and manipulating chiral phonons are twisted bilayers of two- dimensional (2D) materials. Since the discovery of flat electronic bands in twisted bilayer graphene [20], such moiré materials have emerged as rich systems to investigate the properties of correlated electrons, excitons and phonons [21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42]. For example, at small twist angles phonon properties are significantly modified as a consequence of large atomic reconstructions [43, 38, 44] which result in the localization of incipient phason modes [45] and optical phonon modes [45, 46, 47, 48, 49, 50, 51]. To date, however, the chirality of phonons in twisted bilayer systems has not been studied. In this Letter, we study the chirality of phonon modes in moiré materials. Focusing on twisted bilayer $\mathrm{WSe_{2}}$, we first demonstrate qualitatively different phonon chiralities for twist angles close to $0^{\circ}$ and close to $60^{\circ}$ and explain these differences by analyzing the folding of the monolayer $K$ and $K^{\prime}$-valleys into the Brillouin zone of the twisted bilayer. For twist angles near $60^{\circ}$, we find two sets of emergent chiral phonons modes which originate from symmetry breaking at the moiré scale. Finally, we discover flat chiral phonon bands in the energy gap between acoustic and optical modes without the requirement of a magic angle. Such flat phonon bands have recently been reported in metamaterials [52, 53, 54, 55], but not yet in moiré systems. Methods. For twist angles near $0\degree$ ($\theta$) or $60\degree$ ($60^{\circ}-\theta$), the size of the moiré unit cell becomes very large making standard first-principles approaches for calculating phonon properties unfeasible. We therefore use simpler models for interatomic interactions. Specifically, intralayer interactions in $\mathrm{WSe_{2}}$ and $\mathrm{MoSe_{2}}$ are described using a Stillinger-Weber potential, which has been demonstrated to give accurate results for monolayers [56, 57]. For interlayer interactions, we use the Kolmogorov-Crespi potential that can correctly reproduce the interlayer binding energy landscape obtained using first-principles calculations [58]. The relaxed atomic positions of our twisted bilayer systems are determined using the implementation of these potentials within the LAMMPS code [59]. Phonon frequencies and polarization vectors are obtained by diagonalizing the dynamical matrix using a modified version of the PHONOPY code [60] (see Supplementary Information (SI), Sec. I, for additional details). For monolayer graphene, we use the REBO potential [61]. Following Zhang and Niu [5], we define the chirality of a phonon with crystal momentum $\mathbf{q}$ and band index $n$ along the $z$-direction (assuming that the 2D material lies in the $x$-$y$ plane) as $\begin{split}S^{z}_{n{\bf q}}=&\bra{\epsilon_{n{\bf q}}}\widehat{S}^{z}\ket{\epsilon_{n{\bf q}}}\\\ =&\bra{\epsilon_{n{\bf q}}}\sum_{j=1}^{N}\Big{[}\ket{R_{j}}\bra{R_{j}}-\ket{L_{j}}\bra{L_{j}}\Big{]}\ket{\epsilon_{n{\bf q}}},\end{split}$ (1) where $\ket{\epsilon_{n{\bf q}}}$ is the normalized polarization vector, $\widehat{S}^{z}$ denotes the phonon circular polarization operator and $N$ is the number of atoms in the moiré unit cell. $\ket{R_{j}}$ and $\ket{L_{j}}$ denote the right and left circularly polarized basis vectors for atom $j$. For example, for the first two atoms these basis vectors are given by $|R_{1}\rangle=1/\sqrt{2}\times{(1,i,0,0,\ldots)^{\mathrm{T}}}$, $|L_{1}\rangle=1/\sqrt{2}\times{(1,-i,0,0,\ldots)^{\mathrm{T}}}$, $|R_{2}\rangle=1/\sqrt{2}\times{(0,0,1,i,0,0,\ldots)^{\mathrm{T}}}$, and $|L_{2}\rangle=1/\sqrt{2}\times{(0,0,1,-i,0,0,\ldots)^{\mathrm{T}}}$. The phonon mode is linearly polarized when $S_{n\bf{q}}^{z}=0$, circularly polarized when $S_{n\bf{q}}^{z}=\pm 1$ and elliptically polarized when $0<|S_{n\bf{q}}^{z}|<1$. We refer to both the circularly and elliptically polarized modes as chiral phonons. Note that for systems with time-reversal symmetry, we have $S^{z}_{n{\bf q}}=-S^{z}_{n-{\bf q}}$, whereas systems with inversion-symmetry obey $S^{z}_{n{\bf q}}=S^{z}_{n-{\bf q}}$. Therefore, the existence of non-degenerate chiral phonon modes requires the breaking of one of these symmetries [5, 6]. As a test, we first study the chirality of phonons in monolayer graphene (which has both time-reversal and inversion symmetry) and monolayer $\mathrm{WSe_{2}}$ (which is not inversion symmetric). The results are shown in Sec. II of the SI [62]. As expected, all graphene phonons are achiral. For $\mathrm{WSe_{2}}$, we find multiple non-degenerate chiral phonon modes near the $K$ and $K^{\prime}$ valleys, consistent with previous first-principles calculations [6], and the signs of the chiralities in the two valleys are opposite. Figure 1: (a),(b): Chiral phonons in twisted bilayer $\mathrm{WSe_{2}}$ at twist angles of $21.78^{\circ}$ and $38.22^{\circ}$, respectively. Dark solid lines represent phonon bands and blue and red color indicates the associated chirality. (c),(d): Red circles denote the unfolded $K_{\mathrm{M}}$ point of the moiré BZ into the monolayer BZ for $21.78^{\circ}$ and $38.22^{\circ}$, respectively. The green and blue solid lines represent the unrotated and rotated monolayer BZs, whereas the black solid line represents the moiré BZ. Chiral valley phonons in twisted bilayer ${WSe_{2}}$. The naturally occuring bilayer $\mathrm{WSe_{2}}$ has 2H stacking and is inversion symmetric [63]. The untwisted bilayer, therefore, does not exhibit chiral valley phonons. However, the introduction of a twist between the layers results in the emergence of chiral phonons, as demonstrated in Figs. 1(a) and (b), which show the phonon chirality $S^{z}$ of each phonon mode at twist angles of $21.78^{\circ}$ and $38.22^{\circ}$, respectively. At $21.78^{\circ}$, all phonons near the $K_{\mathrm{M}}$ and $K^{\prime}_{\mathrm{M}}$ valleys are achiral (with the subscript $\mathrm{M}$ denoting that these k-points belong to the moiré Brillouin zone), but there are some chiral phonons between $\Gamma_{\mathrm{M}}$ and $M_{\mathrm{M}}$. In contrast, we find several chiral phonon modes in the $K_{\mathrm{M}}$/$K^{\prime}_{\mathrm{M}}$ valleys for the $38.22^{\circ}$ twist angle (but none between $\Gamma_{\mathrm{M}}$ and $M_{\mathrm{M}}$). Qualitatively similar results are found for other twist angles near $0\degree$ and $60\degree$, respectively. The different behaviour of the $K_{\mathrm{M}}$/$K^{\prime}_{\mathrm{M}}$ valley phonons for twist angles near $0^{\circ}$ and near $60^{\circ}$ can be understood by analyzing the folding of the monolayer phonon modes into the smaller moiré Brillouin zone. For the untwisted system (i.e., $\theta=0^{\circ}$), the $K$ point of the top layer (denoted $K_{\mathrm{t}}$) coincides with the $K$ point of the bottom layer (denoted $K_{\mathrm{b}}$) and similarly $K^{\prime}_{\mathrm{t}}$ and $K^{\prime}_{\mathrm{b}}$ coincide. When instead a small, but finite twist angle is considered, the BZs of the two layers are rotated relative to each other and a smaller moiré BZ is created, see Fig. 1(c). To understand which crystal momenta $\mathbf{q}_{i}$ of the monolayer BZs fold onto a specific point $\mathbf{Q}_{\mathrm{M}}$ in the moiré BZ, we “unfold” $\mathbf{Q}_{\mathrm{M}}$ by adding all possible reciprocal lattice vectors ${\bf G}_{i}$ corresponding to the moiré crystal according to [64] ${\bf q}_{i}={\bf Q}_{\mathrm{M}}+{\bf G}_{i}.$ (2) Figure 1(c) shows the set of points in the monolayer BZs that result from unfolding $K_{\mathrm{M}}$ for a twist angle of $21.78^{\circ}$. It can be seen that both $K^{\prime}_{\mathrm{t}}$ and $K_{\mathrm{b}}$ fold onto $K_{\mathrm{M}}$. Since the chirality of the phonons at $K^{\prime}_{\mathrm{t}}$ and $K_{\mathrm{b}}$ have equal magnitudes, but opposite signs, the phonons of the twisted system have a vanishing chirality. Instead, for the system with $60^{\circ}$ twist angle, $K_{\mathrm{t}}$ coincides with $K^{\prime}_{\mathrm{b}}$ and similarly $K^{\prime}_{\mathrm{t}}$ and $K_{\mathrm{b}}$ coincide. Fig. 1(d) shows that for a twist angle near $60^{\circ}$, $K_{\mathrm{M}}$ unfolds onto $K^{\prime}_{\mathrm{t}}$ and $K^{\prime}_{\mathrm{b}}$ which have phonons with the same chirality sign. As a consequence, the phonons of the twisted bilayer exhibit a non-vanishing chirality at the $K_{\mathrm{M}}$ valley (and also for the $K^{\prime}_{\mathrm{M}}$ valley). Figure 2: (a) Chiral $K$-valley phonon modes of monolayer $\mathrm{WSe_{2}}$ (first column) and twisted bilayer $\mathrm{WSe_{2}}$ at different twist angles (columns labelled $\beta$). For comparison, results are also shown for twisted bilayers without interlayer interactions (columns labelled $\alpha$). The colour of the symbols indicates the chirality of the phonon modes. The emergent chiral modes are highlighted by curly brackets. (b)-(e) Absolute value of the in-plane components of the normalized polarization vector for some of the chiral modes for a twist angle of $58^{\circ}$. In (b) and (c), the polarization vector of W atoms in the top layer is used, In (d) and (e), the polarization vector of Se atoms in the top layer is shown. Different stacking locations and associated chiralities within the moiré unit cell, which is denoted by dashed lines, are indicated. Figure 2(a) shows the chiralities of phonons in the $K_{\mathrm{M}}$-valley as a function of twist angle near $60\degree$ (columns labelled $\beta$) and compares them to the $\mathrm{WSe_{2}}$ monolayer result (the first column) and also to the results of _decoupled_ twisted bilayers at the same twist angles (columns labelled $\alpha$), in which no interlayer interactions between the layers are considered. Note that no atomic relaxations occur in decoupled twisted bilauers. For a twist angle of $38.22\degree$, the phonon chiralities of the twisted bilayer are very similar to those of the monolayer and also to those of the decoupled system. As the twist angle approaches $60\degree$, two novel sets of chiral valley phonons emerge [indicated by the curly brackets in Fig. 2(a)] in the twisted system which are absent in the monolayer and in the decoupled bilayer: one set with frequencies near 10 $\mathrm{cm^{-1}}$ and the other with frequencies near 160 $\mathrm{cm^{-1}}$. The frequencies of the first set soften as the twist angle approaches $60\degree$. To understand the origin of these emergent chiral valley phonons, we plot the magnitude of the polarization vector in the moiré unit cell for 58∘, see Figs. 2(b)-(e). Figs. 2(b)-(c) show results for the two emergent valley phonons with the largest chiralities from the low-energy set (near $10~{}\mathrm{cm^{-1}}$). In particular, the mode with $6.2~{}\mathrm{cm^{-1}}$ has a chirality of $-0.56$ and is localized in the $\mathrm{B^{Se/Se}}$ stacking regions while the one with $9.6~{}\mathrm{cm^{-1}}$ has a chirality of 0.35 and is localized in the $\mathrm{B^{W/W}}$ stacking regions. Analyzing the contributions to the total chirality of these modes, we find that both layers and both W and Se atoms contribute. Figs. 2(d)-(e) show the polarization vectors of the two valley phonons belonging to the second emergent set (with frequencies near $160~{}\mathrm{cm^{-1}}$) that have the largest chiralities. The mode with frequency $159.2~{}\mathrm{cm^{-1}}$ is localized in the $\mathrm{B^{Se/Se}}$ region while the one with frequency $159.4~{}\mathrm{cm^{-1}}$ is approximately localized around the $\mathrm{B^{W/W}}$ region. Interestingly, only the Se atoms contribute to the total chiralities of these modes. The emergence of these novel chiral valley phonons can be explained by an inversion symmetry breaking on the moiré scale that occurs as the twist angle approaches $60\degree$. Near this twist angle, the moiré unit cell contains three high-symmetry stackings: 2H, $\mathrm{B^{W/W}}$, and $\mathrm{B^{Se/Se}}$. Among these stackings 2H is most stable and for twist angles larger than $\sim~{}55\degree$ atomic relaxations result in a significant increase in the regions with 2H stacking as well as the formation of domain walls [43] (see SI, Sec. III). It can be shown [43] that this system can be mapped onto an equivalent system of two particles (with one particle representing $\mathrm{B^{W/W}}$ regions and the other $\mathrm{B^{Se/Se}}$ regions) that occupy different sublattices of a triangular lattice and are connected by springs. Since the energies of $\mathrm{B^{W/W}}$ and $\mathrm{B^{Se/Se}}$ regions are different, the two particles have different masses and therefore the effective system lacks inversion symmetry. This in turn results in the emergence of novel chiral phonon modes. In contrast, the twisted bilayer near $0\degree$ features three high-symmetry stackings: AA, $\mathrm{B^{Se/W}}$, and $\mathrm{B^{W/Se}}$. Among these stackings $\mathrm{B^{Se/W}}$ and $\mathrm{B^{W/Se}}$ are the most stable stackings. This system can be mapped onto a set of particles (representing $\mathrm{AA}$ regions) which occupy the sites of a triangular lattice and are connected by springs. This effective system exhibits inversion symmetry and therefore no emergent chiral phonon modes are observed. Our analysis shows that the polarization vectors of chiral phonons near $0^{\circ}$ and $60^{\circ}$ are qualitatively different from the those of the monolayer phonons. This suggests that commonly used approaches for calculating phonon properties in twisted bilayers that are based on a zone folding procedure of monolayer results are unreliable [65, 66]. Figure 3: (a) The phonon band structure of twisted bilayer $\mathrm{WSe_{2}}$ at $58^{\circ}$ in the vicinity of the energy gap between acoustic and optical modes. (b) Chirality $S^{z}$ (b) and band width (c) of flat acoustic ($\mathrm{A1^{\prime}}$) and optical ($\mathrm{O1^{\prime}}$) phonon bands as function of twist angle near $60^{\circ}$. (d) and (e): Polarization vectors of $\mathrm{A1^{\prime}}$ and $\mathrm{O1^{\prime}}$ modes at a twist angle of $58^{\circ}$. Different stackings are indicated by colored symbols and the moiré unit cell is outlined using dashed gray lines. In (d), the displacements of W atoms in the top layer is shown, while in (e) that of Se atoms in the top layer was used. Phononic band flattening in twisted bilayer $\mathrm{WSe_{2}}$. Figure 3(a) demonstrates the emergence of flat phonon bands in the vicinity of the energy gap between acoustic and optical modes for a $58\degree$ twisted bilayer. This phononic band gap is inherited from the monolayer which has an energy gap of 17.9 $\mathrm{cm^{-1}}$ between the acoustic and optical phonon modes with the highest-energy acoustic mode being located at $K$ and the lowest-energy optical mode at the $\Gamma$ point of the BZ. The highest-energy acoustic mode at $K$ is non-degenerate and chiral with $S^{z}=-0.5$, while the lowest-energy optical modes at $\Gamma$ are degenerate and achiral. Fig. 3(b) shows that the chirality associated with the highest-energy acoustic mode at $K_{\mathrm{M}}$ (denoted as $\mathrm{A1^{\prime}}$) changes dramatically and becomes achiral as the twist angle approaches $60^{\circ}$. This change in chirality is caused by the localization of the phonon mode in regions with inversion-symmetric 2H stacking (see discussion below). In contrast, the lowest-energy optical mode at $K_{\mathrm{M}}$ (denoted as $\mathrm{O1^{\prime}}$) becomes chiral as the twist angle approaches $60^{\circ}$. Fig. 3(c) shows the bandwidths of the A1′ and O1′ modes and demonstrates that these bands become extremely flat as the twist angle approaches $60^{\circ}$ and the size of the moiré cell increases. Similar to the electronic bandwidths in twisted transition metal dichalcogenide bilayers, the flattening does not occur at a specific magic angle, unlike the case of twisted bilayer graphene [20]. Interestingly, the widths of the flat phonon bands are significantly smaller than the width of the flat electron bands at the same twist angle. For example, at a twist angle of 56.5∘ of the twisted bilayer of TMDs, the width of the electronic valence band is 5 meV [27] and the width of the largest acoustic phonon band is 0.1 meV. The flattening of the phonon bands is a consequence of both zone folding and the localization of phonons in real space (see SI, Sec. IV for details). Figs. 3(d) and (e) show the polarization vectors (at $K_{\mathrm{M}}$) of the A1′ and O1′ modes for a twist angle of $58^{\circ}$. The A1′ mode is localized in regions of 2H stacking, while O1′ is localized in $\mathrm{B^{Se/Se}}$ regions as well as the domain walls. It is interesting to note that the localization of the highest acoustic phonons is strikingly similar to that of the highest electronic valence states in these systems [67]. Similar band flattening is also observed for twist angles near $0^{\circ}$ (see SI[62], Sec. IV), where the highest-energy acoustic mode is localized in the $\mathrm{B^{W/Se}/B^{Se/W}}$ stacking region and the lowest-energy optical mode is localized in regions of $\mathrm{AA}$ stacking and domain walls. The distinct localization of the acoustic and optical modes can be explained by analyzing the frequencies of these modes in the various untwisted bilayers with high-symmetry stacking arrangements. For example, the A1′ mode originates from the $K$ point of the monolayer BZ. For the untwisted bilayers, the phonon frequencies at $K$ are 141.08 cm-1 for $\mathrm{B^{Se/Se}}$ stacking, 141.1 cm-1 for $\mathrm{B^{W/W}}$ stacking and 141.8 cm-1 for $\mathrm{2H}$ stacking. Therefore, the highest-energy acoustic mode localizes on the $\mathrm{2H}$ regions as the moiré unit cell increases in size. Chiral valley phonons in strain-engineered moiré materials and heterostructures. Besides twisting, there are two approaches to create a moiré pattern in bilayer systems: (i) by applying strain to only one of two identical layers and (ii) by stacking two different 2D materials on top of each other. We investigate the existence of chiral valley phonons in both cases. In contrast to the twisted bilayer at small twist angles, we find that the strain-engineered bilayer exhibits chiral phonons in the $K_{\mathrm{M}}$ valley (see Sec. V of the SI for a summary of these modes and more details [62]). In addition, emergent chiral phonon modes with frequencies of $\sim~{}10~{}\mathrm{cm^{-1}}$ and $\sim~{}156~{}\mathrm{cm^{-1}}$ are found. Finally, we also observe emergent chiral phonon modes in a $\mathrm{WSe_{2}/MoSe_{2}}$ heterostructure at a twist angle of $3.14^{\circ}$ (see SI [62], Sec. V). Summary. In this paper, we have demonstrated the presence of chiral phonons in twisted bilayer WSe2 as well as other moiré materials and studied their properties. We have found that the phonon chirality depends on the twist angle with systems near $0^{\circ}$ exhibiting qualitatively different chiralities than systems near $60^{\circ}$. Very close to $60^{\circ}$, we observe emergent chiral modes as well as flat phonon bands. The predicted chiral properties of phonons in twisted bilayer materials can be measured with helicity-resolved Raman spectroscopy [15, 68]. While the resolution of such techniques is typically not high enough to access individual modes in systems near 0∘ and 60∘, they can potentially measure the k-point and helicity- resolved phonon density of states (see SI, Sec. VII). Future work should investigate the scattering of chiral phonons with other quasiparticles, such as electrons, excitons, and photons and the effects of such scattering processes on the electronic and optical properties of moiré materials. ###### Acknowledgements. Acknowledgments This project has received funding from the European Union’s Horizon 2020 research and innovation programme under the Marie Skłodowska- Curie grant agreement No 101028468. The authors acknowledge support from the Thomas Young Centre under grant TYC-101, and discussions with Rup Chowdhury, Nikita Tepliakov and Saurabh Srivastav. Supplementary Information (SI) : Chiral valley phonons and flat phonon bands in moiré materials ## I I: Simulation details ### I.1 Generation of structures #### Moiré pattern due to twist: The commensurate twist angles that give rise to the least-area moiré unit-cell for twisted bilayer of $\mathrm{WSe_{2}}$ are given by, $\cos(\theta_{i})=\frac{3i^{2}+3i+1/2}{3i^{2}+3i+1}$ (3) for an integer $i$ [69]. One of the set of moiré lattice vectors are- ${\bf A_{1}}=i{\bf a_{1}}+(i+1){\bf a_{2}}$ and ${\bf A_{2}}=-(i+1){\bf a_{1}}+(2i+1){\bf a_{2}}$ with ${\bf a_{1}}=(1/2,\sqrt{3}/2)a_{0}$ and ${\bf a_{2}}=(-1/2,\sqrt{3}/2)a_{0}$, where $a_{0}$ is the lattice constant of singe layer $\mathrm{WSe_{2}}$. For twist angles ($\theta_{i}$) close to $0^{\circ}$, we construct the moiré unit-cell starting from AA stacking using TWISTER [70]. The high-symmetry stackings present in the moiré patterns are- $\mathrm{AA}$ (with W and Se of the top layer are directly above W and Se of bottom layer), $\mathrm{B^{W/Se}}$ (Bernal stacking with W of the top layer directly above Se of bottom layer), and $\mathrm{B^{Se/W}}$ (Bernal stacking with Se of the top layer directly above W of bottom layer). These stackings are sometimes referred to as AA, AB, and BA, as well. For twist angles ($60^{\circ}-\theta_{i}$) close to $60^{\circ}$, we construct the moiré unit- cell starting from 2H stacking. The high-symmetry stackings present in the moiré patterns are- $\mathrm{2H}$ (with W and Se of the top layer are directly above Se and W of bottom layer), $\mathrm{B^{Se/Se}}$ (Bernal stacking with Se of the top layer directly above Se of bottom layer), and $\mathrm{B^{W/W}}$ (Bernal stacking with W of the top layer directly above W of bottom layer). These stackings are sometimes referred to as AA′, A′B, and AB′, respectively. Twist angles | Number of atoms | Moiré length (in Å) ---|---|--- 21.78∘ | 42 | 8.8 9.4∘ | 222 | 20.4 7.3∘ | 366 | 26.1 6∘ | 546 | 31.9 4.41∘ | 1626 | 54.6 3.14∘ | 1986 | 60.4 2∘ | 4902 | 94.9 38.22∘ | 42 | 8.8 50.6∘ | 222 | 20.4 52.7∘ | 366 | 26.1 54∘ | 546 | 31.9 55.59∘ | 1626 | 54.6 56.86∘ | 1986 | 60.4 58∘ | 4902 | 94.9 58.31∘ | 6846 | 113.3 Table 1: Moiré patterns of twisted bilayer $\mathrm{WSe_{2}}$ studied in this work. #### Moiré pattern due to strain We use the unit-cell lattice vectors of single layer of $\mathrm{WSe_{2}}$ as ${\bf a_{1}}=a_{0}(1,0,0)$, ${\bf a_{2}}=a_{0}(1/2,\sqrt{3}/2,0)$, and ${\bf a_{3}}=a_{0}(0,0,30)$, where $a_{0}$ is the lattice constant. We can create the moiré patterns in two ways: by straining only one of the layer of $\mathrm{WSe_{2}}$, and by straining both the layers. We use the first approach in this work. We apply a tensile strain to the bottom layer. We discuss the details of this approach below. The periodicity of the strained bilayer is computed in the following manner: $\begin{bmatrix}m&0\\\ 0&m\end{bmatrix}\begin{pmatrix}{\bf a_{1}}\\\ {\bf a_{2}}\end{pmatrix}=\begin{bmatrix}n&0\\\ 0&n\end{bmatrix}\begin{pmatrix}{\bf a_{1}^{b}}\\\ {\bf a_{2}^{b}}\end{pmatrix}$ (4) In the case of 3.3 % biaxial strain, we use $m=30$, and $n=29$ starting from AA stacking configuration. The moiré lattice constant of the strained bilayer is 98.9 Å and contains 5223 atoms. #### Moiré heterostructure We replace the top layer of bilayer of $\mathrm{WSe_{2}}$ by $\mathrm{MoSe_{2}}$ and construct moiré lattice by rotating the layer. Since the experimental lattice constants of these two materials are similar, we simply replace one of the $\mathrm{WSe_{2}}$ layer by $\mathrm{MoSe_{2}}$ in the twisted bilayer $\mathrm{WSe_{2}}$ and create the moiré heterostructure. ### I.2 Structural relaxations The moiré patterns are relaxed using the LAMMPS package with the Stillinger- Weber [56], and Kolmogorov-Crespi [58] potentials to capture the intralayer and interlayer interactions of the twisted bilayer of $\mathrm{WSe_{2}}$, respectively. The Kolmogorov-Crespi parameters used in this work can correctly reproduce the interlayer binding energy landscape, obtained using density functional theory. The atomic relaxations produced using these parameters are in excellent agreement with relaxations performed using density functional theory. The intralayer Stillinger-Weber potential quite accurately captures the full phonon dispersion, in particular the acoustic modes. The accurate potentials used in our simulations are expected to capture the phonon renormalization of the moiré materials. At first, we relax the simulation box to ensure the in-plane pressure is as small as possible. Next, we relax the atoms within a fixed simulation box with the force tolerance of $10^{-5}$ eV/Å for any atom along any direction. ### I.3 Force constants calculations For monolayer $\mathrm{WSe_{2}}$, we use a $6\times 6\times 1$ supercell to compute the force constants. For large twist angles of twisted bilayer $\mathrm{WSe_{2}}$, we use a $2\times 2\times 1$ supercell to compute the force constants. The moiré patterns contain thousands of atoms for small twist angles. Therefore, we use $1\times 1\times 1$ supercell to compute the force- constants. ## II II : Chiral phonons in 2D materials ### II.1 Absence of chiral valley-phonons in graphene Figure 4: Chirality of phonons in single layer of graphene. The solid lines represent the phonon frequencies, and hexagons (absent in the figure as $S^{z}=0$) represent the computed $S_{z}$ values after summing over all the basis atoms. There are no non-degenerate chiral modes in single layer of graphene due to inversion symmetry. $\omega$ | $S^{z}$ | $S^{z}_{C1}$ | $S^{z}_{C{2}}$ ---|---|---|--- (in $\mathrm{cm^{-1}}$) | | | 505.2 | 0.0 | 0.0 | 0.0 505.2 | 0.0 | 0.0 | 0.0 821.1 | 0.0 | 0.5 | -0.5 1201.5 | 0.0 | -0.5 | 0.5 1201.5 | 0.0 | -0.5 | 0.5 1724.5 | 0.0 | 0.5 | -0.5 Table 2: Phonon modes, total chirality, and chirality of the each types of atoms in the case of single-layer of graphene. Within the unit-cell, two atoms can have opposite chirality, but the total chirality is 0. $\omega=505.2$ cm-1 $\omega=505.2$ cm-1 $\omega=821.1$ cm-1 $\omega=1201.5$ cm-1 $\omega=1201.5$ cm-1 $\omega=1724.5$ cm-1 Eigenvectors associated with phonon modes in grapheneThe Blue circles denote the Carbon atoms ### II.2 Presence of chiral non-degenerate valley phonons in monolayer $\mathrm{WSe_{2}}$ Figure 5: Chirality of phonons in monolayer of $\mathrm{WSe_{2}}$. The solid lines represent the phonon frequencies, and hexagons represent the computed $S_{z}$ values after summing over all the basis atoms. There are multiple non-degenerate chiral modes in monolayer of $\mathrm{WSe_{2}}$ due to inversion symmetry breaking. $\omega$ | $S^{z}$ | $S^{z}_{W1}$ | $S^{z}_{Se{2}}$ | $S^{z}_{Se{3}}$ ---|---|---|---|--- (in $\mathrm{cm^{-1}}$) | | | | 118.9 | -0.38 | 0.0 | -0.19 | -0.19 121.2 | 0.76 | 0.76 | 0.0 | 0.0 141.4 | -0.49 | -0.75 | 0.13 | 0.13 200.66 | 0.0 | 0.0 | 0.0 | 0.0 212.06 | -1.0 | 0.0 | -0.5 | -0.5 216.52 | 1.0 | 0.0 | 0.5 | 0.5 240.28 | 0.49 | -0.25 | 0.37 | 0.37 249.2 | -0.62 | 0.0 | -0.31 | -0.31 255.38 | 0.24 | 0.24 | 0.0 | 0.0 Table 3: Chirality associated with all the valley phonon modes (at $K$) in single layer of $\mathrm{WSe_{2}}$. $\omega=118.9$ cm-1 $\omega=121.2$ cm-1 $\omega=141.4$ cm-1 $\omega=200.6$ cm-1 $\omega=212$ cm-1 $\omega=216.5$ cm-1 $\omega=240.3$ cm-1 $\omega=249.2$ cm-1 $\omega=255.4$ cm-1 Eigenvectors associated with phonon modes in single layer $\mathrm{WSe_{2}}$The Blue and red circles denote the Se and W atoms, respectively. ## III III : Interlayer separation for twist angles close to $60^{\circ}$ and $0^{\circ}$ Figure 6: (a),(b): Significant structural relaxations and inversion symmetry breaking at the moiré scale illustrated using interlayer separation for twist angles of $58^{\circ}$ and $2^{\circ}$, respectively. The associated color bar denotes the magnitude of interlayer separation and is in Å. The moiré unit cell and stackings are marked. ## IV IV : Flat phonon bands Figure 7: (a), (b) Evidence of band-flattening for a twist angle of 3.14∘ and $56.86^{\circ}$ twisted bilayer $\mathrm{WSe_{2}}$ near the acoustic-optic phonon band gap. The blue lines denote the calculations with twisted bilayers and the red dashed lines denote calculations with decoupled twisted bilayers. Importantly, the band flattens and degeneracies in the band-structure of the decoupled systems are significantly reduced. (c),(d) ((e),(f)) Absolute value of the polarization vectors at $\Gamma_{\mathrm{M}}$ of $\mathrm{A1}$ and $\mathrm{O1}$ modes at a twist angle of $3.14^{\circ}$ ($56.86^{\circ}$). Different stackings are indicated by colored symbols and the moiré unit cell is outlined using dashed gray lines. In (c),(e) the displacements of W atoms in the top layer is shown, while in (d),(f) that of Se atoms in the top layer was used. (g) The band-flattening is indicated by the reduction of band-width with twist angles. | Bandwidth in cm-1 ---|--- | Twisted bilayer | Decoupled twisted bilayer Twist angle | A1 | O1 | A1 | O1 55.59∘ | 1.029 | 0.185 | 1.321 | 0.327 56.86∘ | 0.427 | 0.069 | 0.672 | 0.110 Table 4: Comparison of bandwidths in the twisted bilayers that include all the relaxation effects and in the decoupled twisted bilayers due to zone-folding for $55.59^{\circ}$ and $56.86^{\circ}$ twisted bilayers. Note that no atomic relaxations occur in decoupled twisted bilayers. ## V V : Chiral valley phonons in strain engineered moiré and $\mathrm{MoSe_{2}/WSe_{2}}$ heterostructure Figure 8: (a),(b)-Chiralities at K point for all the phonon modes in 3.3$\%$ biaxially strained moiré of $\mathrm{WSe_{2}}$ and the $\mathrm{MoSe_{2}/WSe_{2}}$ heterostructure. The left column of (a) denote the non-interacting and the right column of (a) interacting bilayer calculations. Similarly, we show the results for (b). The emergent chiral modes are highlighted by curly brackets. We apply a bi-axial tensile strain to the bottom layer of an AA stacked bilayer $\mathrm{WSe_{2}}$. The resulting strained structure contains three unique high-symmetry stackings: $\mathrm{A\widetilde{A}}$, $\mathrm{B^{W/\widetilde{Se}}}$ and $\mathrm{B^{Se/\widetilde{W}}}$, where the tilde indicates a sublattice of the strained layer. These stackings are distinct from those of the twisted bilayer near $0^{\circ}$ as they break layer symmetry. ## VI VI: Achiral phonon modes at $K/K^{\prime}$ for twist angles close to $0^{\circ}$ Figure 9: Chiral phonons in twisted bilayer $\mathrm{WSe_{2}}$ at twist angle of (a) $4.41^{\circ}$, (b) $55.59^{\circ}$, (c) $3.14^{\circ}$ and (d) $56.86^{\circ}$). Black solid circles represent phonon bands (computed at a few discrete points) and blue and red color indicates the associated chirality (computed at the same set of discrete points). Moreover, the size of the colored circles are proportional to absolute value of the chirality. As this figure is somewhat difficult to interpret, we have also directly inspected the computed chirality values at $\mathrm{K_{M}}$ and find that there are non- degenerate chiral modes for twist angles close to 60∘ but none for twist abgles close to 0∘. ## VII VII: Chirality resolved and total density of states Figure 10: (a),(b),(c): Total and chirality resolved density of states (DOS). We have used $36\times 36\times 1$, $12\times 12\times 1$, and $3\times 3\times 1$ grids to compute the phonon DOS for monolayer $\mathrm{WSe_{2}}$, $38.22^{\circ}$ twisted bilayer $\mathrm{WSe_{2}}$, and $55.59^{\circ}$ twisted bilayer $\mathrm{WSe_{2}}$, respectively. The delta function in DOS calculation has been replaced by a Gaussian with a standard deviation of 4.8 cm-1. Figure 11: (a),(b),(c): $k$-point resolved density of states (DOS). We have used $36\times 36\times 1$, $12\times 12\times 1$, and $3\times 3\times 1$ k-point grids to compute the phonon DOS for monolayer $\mathrm{WSe_{2}}$, $38.22^{\circ}$ twisted bilayer $\mathrm{WSe_{2}}$, and $55.59^{\circ}$ twisted bilayer $\mathrm{WSe_{2}}$, respectively. A Gaussian broadening with a standard deviation of 4.8 cm-1 has been used. ## References * Kneer _et al._ [2018] L. M. Kneer, E.-M. Roller, L. V. Besteiro, R. Schreiber, A. O. Govorov, and T. Liedl, ACS Nano 12, 9110 (2018). * Katsnelson _et al._ [2006] M. I. Katsnelson, K. S. Novoselov, and A. K. Geim, Nature Physics 2, 620 (2006). * Li and Kharzeev [2016] Q. Li and D. E. Kharzeev, Nuclear Physics A 956, 107 (2016), the XXV International Conference on Ultrarelativistic Nucleus-Nucleus Collisions: Quark Matter 2015. * Li _et al._ [2016] Q. Li, D. E. Kharzeev, C. Zhang, Y. Huang, I. Pletikosifá, A. V. Fedorov, R. D. Zhong, J. A. Schneeloch, G. D. Gu, and T. Valla, Nature Physics 12, 550 (2016). * Zhang and Niu [2015] L. Zhang and Q. Niu, Phys. Rev. Lett. 115, 115502 (2015). * Zhu _et al._ [2018] H. Zhu, J. Yi, M.-Y. Li, J. Xiao, L. Zhang, C.-W. Yang, R. A. Kaindl, L.-J. Li, Y. Wang, and X. Zhang, Science 359, 579 (2018). * Carvalho _et al._ [2017] B. R. Carvalho, Y. Wang, S. Mignuzzi, D. Roy, M. Terrones, C. Fantini, V. H. Crespi, L. M. Malard, and M. A. Pimenta, Nature Communications 8, 14670 (2017). * Chen _et al._ [2015] S.-Y. Chen, C. Zheng, M. S. Fuhrer, and J. Yan, Nano Letters 15, 2526 (2015). * Kang _et al._ [2018] M. Kang, S. W. Jung, W. J. Shin, Y. Sohn, S. H. Ryu, T. K. Kim, M. Hoesch, and K. S. Kim, Nature Materials 17, 676 (2018). * Li _et al._ [2019a] Z. Li, T. Wang, C. Jin, Z. Lu, Z. Lian, Y. Meng, M. Blei, S. Gao, T. Taniguchi, K. Watanabe, T. Ren, S. Tongay, L. Yang, D. Smirnov, T. Cao, and S.-F. Shi, Nature Communications 10, 2469 (2019a). * Liu _et al._ [2019] E. Liu, J. van Baren, T. Taniguchi, K. Watanabe, Y.-C. Chang, and C. H. Lui, Phys. Rev. Research 1, 032007 (2019). * Li _et al._ [2019b] Z. Li, T. Wang, C. Jin, Z. Lu, Z. Lian, Y. Meng, M. Blei, M. Gao, T. Taniguchi, K. Watanabe, T. Ren, T. Cao, S. Tongay, D. Smirnov, L. Zhang, and S.-F. Shi, ACS Nano 13, 14107 (2019b). * Delhomme _et al._ [2020] A. Delhomme, D. Vaclavkova, A. Slobodeniuk, M. Orlita, M. Potemski, D. M. Basko, K. Watanabe, T. Taniguchi, D. Mauro, C. Barreteau, E. Giannini, A. F. Morpurgo, N. Ubrig, and C. Faugeras, 2D Materials 7, 041002 (2020). * Chen _et al._ [2019] X. Chen, X. Lu, S. Dubey, Q. Yao, S. Liu, X. Wang, Q. Xiong, L. Zhang, and A. Srivastava, Nature Physics 15, 221 (2019). * Chen _et al._ [2018] H. Chen, W. Zhang, Q. Niu, and L. Zhang, 2D Materials 6, 012002 (2018). * Juraschek and Spaldin [2019] D. M. Juraschek and N. A. Spaldin, Phys. Rev. Materials 3, 064405 (2019). * Du _et al._ [2019] L. Du, J. Tang, Y. Zhao, X. Li, R. Yang, X. Hu, X. Bai, X. Wang, K. Watanabe, T. Taniguchi, D. Shi, G. Yu, X. Bai, T. Hasan, G. Zhang, and Z. Sun, Advanced Functional Materials 29, 1904734 (2019). * Zhang _et al._ [2020a] W. Zhang, A. Srivastava, X. Li, and L. Zhang, Phys. Rev. B 102, 174301 (2020a). * Grissonnanche _et al._ [2020] G. Grissonnanche, S. Thériault, A. Gourgout, M.-E. Boulanger, E. Lefrancois, A. Ataei, F. Laliberté, M. Dion, J.-S. Zhou, S. Pyon, T. Takayama, H. Takagi, N. Doiron-Leyraud, and L. Taillefer, Nature Physics 16, 1108 (2020). * Bistritzer and MacDonald [2011] R. Bistritzer and A. H. MacDonald, Proceedings of the National Academy of Sciences 108, 12233 (2011). * Cao _et al._ [2018a] Y. Cao, V. Fatemi, S. Fang, K. Watanabe, T. Taniguchi, E. Kaxiras, and P. Jarillo-Herrero, Nature 556, 43 (2018a). * Cao _et al._ [2018b] Y. Cao, V. Fatemi, A. Demir, S. Fang, S. L. Tomarken, J. Y. Luo, J. D. Sanchez-Yamagishi, K. Watanabe, T. Taniguchi, E. Kaxiras, R. C. Ashoori, and P. Jarillo-Herrero, Nature 556, 80 (2018b). * Tran _et al._ [2020] K. Tran, J. Choi, and A. Singh, 2D Materials 8, 022002 (2020). * Wang _et al._ [2020] L. Wang, E.-M. Shih, A. Ghiotto, L. Xian, D. A. Rhodes, C. Tan, M. Claassen, D. M. Kennes, Y. Bai, B. Kim, K. Watanabe, T. Taniguchi, X. Zhu, J. Hone, A. Rubio, A. N. Pasupathy, and C. R. Dean, Nature Materials 19, 861 (2020). * Huang _et al._ [2020] M. Huang, Z. Wu, J. Hu, X. Cai, E. Li, L. An, X. Feng, Z. Ye, N. Lin, K. T. Law, _et al._ , arXiv preprint arXiv:2006.05615 (2020). * Scuri _et al._ [2020] G. Scuri, T. I. Andersen, Y. Zhou, D. S. Wild, J. Sung, R. J. Gelly, D. Bérubé, H. Heo, L. Shao, A. Y. Joe, A. M. Mier Valdivia, T. Taniguchi, K. Watanabe, M. Lončar, P. Kim, M. D. Lukin, and H. Park, Phys. Rev. Lett. 124, 217403 (2020). * Naik and Jain [2018] M. H. Naik and M. Jain, Phys. Rev. Lett. 121, 266401 (2018). * Zhang _et al._ [2020b] Z. Zhang, Y. Wang, K. Watanabe, T. Taniguchi, K. Ueno, E. Tutuc, and B. J. LeRoy, Nature Physics 16, 1093 (2020b). * An _et al._ [2020] L. An, X. Cai, D. Pei, M. Huang, Z. Wu, Z. Zhou, J. Lin, Z. Ying, Z. Ye, X. Feng, R. Gao, C. Cacho, M. Watson, Y. Chen, and N. Wang, Nanoscale Horiz. 5, 1309 (2020). * Xian _et al._ [2020] L. Xian, M. Claassen, D. Kiese, M. M. Scherer, S. Trebst, D. M. Kennes, and A. Rubio, arXiv preprint arXiv:2004.02964 (2020). * Devakul _et al._ [2021] T. Devakul, V. Crépel, Y. Zhang, and L. Fu, arXiv preprint arXiv:2106.11954 (2021). * Vitale _et al._ [2021] V. Vitale, K. Atalar, A. A. Mostofi, and J. Lischner, arXiv preprint arXiv:2102.03259 (2021). * Venkateswarlu _et al._ [2020] S. Venkateswarlu, A. Honecker, and G. Trambly de Laissardière, Phys. Rev. B 102, 081103 (2020). * Li _et al._ [2021] E. Li, J.-X. Hu, X. Feng, Z. Zhou, L. An, K. T. Law, N. Wang, and N. Lin, arXiv preprint arXiv:2103.06479 (2021). * Soriano and Lado [2021] D. Soriano and J. Lado, arXiv preprint arXiv:2105.04502 (2021). * Shabani _et al._ [2021] S. Shabani, D. Halbertal, W. Wu, M. Chen, S. Liu, J. Hone, W. Yao, D. N. Basov, X. Zhu, and A. N. Pasupathy, Nature Physics 17, 720 (2021). * Brem _et al._ [2020] S. Brem, K.-Q. Lin, R. Gillen, J. M. Bauer, J. Maultzsch, J. M. Lupton, and E. Malic, Nanoscale 12, 11088 (2020). * Weston _et al._ [2020] A. Weston, Y. Zou, V. Enaldiev, A. Summerfield, N. Clark, V. Zólyomi, A. Graham, C. Yelgel, S. Magorrian, M. Zhou, J. Zultak, D. Hopkinson, A. Barinov, T. H. Bointon, A. Kretinin, N. R. Wilson, P. H. Beton, V. I. Fal’ko, S. J. Haigh, and R. Gorbachev, Nature Nanotechnology 15, 592 (2020). * Sinha _et al._ [2020] S. Sinha, P. C. Adak, R. S. Surya Kanthi, B. L. Chittari, L. D. V. Sangani, K. Watanabe, T. Taniguchi, J. Jung, and M. M. Deshmukh, Nature Communications 11, 5548 (2020). * Goodwin _et al._ [2021] Z. A. Goodwin, L. Klebl, V. Vitale, X. Liang, V. Gogtay, X. van Gorp, D. M. Kennes, A. A. Mostofi, and J. Lischner, arXiv preprint arXiv:2105.12641 (2021). * Graham _et al._ [2020] A. J. Graham, J. Zultak, M. J. Hamer, V. Zolyomi, S. Magorrian, A. Barinov, V. Kandyba, A. Giampietri, A. Locatelli, F. Genuzio, N. C. Teutsch, C. Salazar, N. D. M. Hine, V. I. Falḱo, R. V. Gorbachev, and N. R. Wilson, 2D Materials 8, 015016 (2020). * Pizarro _et al._ [2019] J. M. Pizarro, M. J. Calderón, and E. Bascones, Journal of Physics Communications 3, 035024 (2019). * Maity _et al._ [2021] I. Maity, P. K. Maiti, H. R. Krishnamurthy, and M. Jain, Phys. Rev. B 103, L121102 (2021). * Carr _et al._ [2018] S. Carr, D. Massatt, S. B. Torrisi, P. Cazeaux, M. Luskin, and E. Kaxiras, Phys. Rev. B 98, 224102 (2018). * Maity _et al._ [2020] I. Maity, M. H. Naik, P. K. Maiti, H. R. Krishnamurthy, and M. Jain, Phys. Rev. Research 2, 013335 (2020). * Gadelha _et al._ [2021] A. C. Gadelha, D. A. A. Ohlberg, C. Rabelo, E. G. S. Neto, T. L. Vasconcelos, J. L. Campos, J. S. Lemos, V. Ornelas, D. Miranda, R. Nadas, F. C. Santana, K. Watanabe, T. Taniguchi, B. van Troeye, M. Lamparski, V. Meunier, V.-H. Nguyen, D. Paszko, J.-C. Charlier, L. C. Campos, L. G. Can√ßado, G. Medeiros-Ribeiro, and A. Jorio, Nature 590, 405 (2021). * Koshino and Son [2019] M. Koshino and Y.-W. Son, Phys. Rev. B 100, 075416 (2019). * Quan _et al._ [2021] J. Quan, L. Linhart, M.-L. Lin, D. Lee, J. Zhu, C.-Y. Wang, W.-T. Hsu, J. Choi, J. Embley, C. Young, T. Taniguchi, K. Watanabe, C.-K. Shih, K. Lai, A. H. MacDonald, P.-H. Tan, F. Libisch, and X. Li, Nature Materials (2021), 10.1038/s41563-021-00960-1. * Lamparski _et al._ [2020] M. Lamparski, B. V. Troeye, and V. Meunier, 2D Materials 7, 025050 (2020). * [50] K.-Q. Lin, J. Holler, J. M. Bauer, P. Parzefall, M. Scheuck, B. Peng, T. Korn, S. Bange, J. M. Lupton, and C. Schüller, Advanced Materials n/a, 2008333. * Debnath _et al._ [2020] R. Debnath, I. Maity, R. Biswas, V. Raghunathan, M. Jain, and A. Ghosh, Nanoscale 12, 17272 (2020). * Deng _et al._ [2020] Y. Deng, M. Oudich, N. J. Gerard, J. Ji, M. Lu, and Y. Jing, Phys. Rev. B 102, 180304 (2020). * Rosendo López _et al._ [2020] M. Rosendo López, F. Peñaranda, J. Christensen, and P. San-Jose, Phys. Rev. Lett. 125, 214301 (2020). * Gardezi _et al._ [2021] S. M. Gardezi, H. Pirie, S. Carr, W. Dorrell, and J. E. Hoffman, 2D Materials 8, 031002 (2021). * Martí-Sabaté and Torrent [2021] M. Martí-Sabaté and D. Torrent, Phys. Rev. Applied 15, L011001 (2021). * Mobaraki _et al._ [2018] A. Mobaraki, A. Kandemir, H. Yapicioglu, O. Gülseren, and C. Sevik, Computational Materials Science 144, 92 (2018). * Kandemir _et al._ [2016] A. Kandemir, H. Yapicioglu, A. Kinaci, T. Çağın, and C. Sevik, Nanotechnology 27, 055703 (2016). * Naik _et al._ [2019] M. H. Naik, I. Maity, P. K. Maiti, and M. Jain, The Journal of Physical Chemistry C 123, 9770 (2019). * Plimpton [1995] S. Plimpton, Journal of Computational Physics 117, 1 (1995). * Togo and Tanaka [2015] A. Togo and I. Tanaka, Scripta Materialia 108, 1 (2015). * Brenner _et al._ [2002] D. W. Brenner, O. A. Shenderova, J. A. Harrison, S. J. Stuart, B. Ni, and S. B. Sinnott, Journal of Physics: Condensed Matter 14, 783 (2002). * [62] See Supplementary Information (SI) for simulation details, chiral phonon in monolayer graphene, $\mathrm{WSe_{2}}$ and polarization vectors, interlayer separation for twist angles close to $0^{\circ}$ and close to $60^{\circ}$, flat phonon bands for twist angles close to 0∘, chiral valley phonons in strain engineered moiré and $\mathrm{MoSe_{2}/WSe_{2}}$ heterostructure, achiral phonons for twist angles close to $0^{\circ}$, chirality and k-point resolved density of states, and movies to illustrate distinct chiral valley phonons twist angles close to $0^{\circ}$ and close to $60^{\circ}$ (attached as mp4 files) . * Jones _et al._ [2014] A. M. Jones, H. Yu, J. S. Ross, P. Klement, N. J. Ghimire, J. Yan, D. G. Mandrus, W. Yao, and X. Xu, Nature Physics 10, 130 (2014). * Popescu and Zunger [2012] V. Popescu and A. Zunger, Phys. Rev. B 85, 085201 (2012). * Lin _et al._ [2018] M.-L. Lin, Q.-H. Tan, J.-B. Wu, X.-S. Chen, J.-H. Wang, Y.-H. Pan, X. Zhang, X. Cong, J. Zhang, W. Ji, P.-A. Hu, K.-H. Liu, and P.-H. Tan, ACS Nano 12, 8770 (2018). * Parzefall _et al._ [2021] P. Parzefall, J. Holler, M. Scheuck, A. Beer, K.-Q. Lin, B. Peng, B. Monserrat, P. Nagler, M. Kempf, T. Korn, and C. Schüller, 2D Materials 8, 035030 (2021). * Kundu _et al._ [2021] S. Kundu, M. H. Naik, H. Krishnamurthy, and M. Jain, arXiv preprint arXiv:2103.07447 (2021). * Tatsumi _et al._ [2018] Y. Tatsumi, T. Kaneko, and R. Saito, Phys. Rev. B 97, 195444 (2018). * Lopes dos Santos _et al._ [2007] J. M. B. Lopes dos Santos, N. M. R. Peres, and A. H. Castro Neto, Phys. Rev. Lett. 99, 256802 (2007). * Naik _et al._ [2022] S. Naik, M. H. Naik, I. Maity, and M. Jain, Computer Physics Communications 271, 108184 (2022).
# SYGMA: System for Generalizable Modular Question Answering Over Knowledge Bases Sumit Neelam, Udit Sharma, Hima Karanam, Shajith Ikbal, Pavan Kapanipathi Ibrahim Abdelaziz, Nandana Mihindukulasooriya, Young-Suk Lee, Santosh Srivastava Cezar Pendus, Saswati Dana, Dinesh Garg, Achille Fokoue, G P Shrivatsa Bhargav Dinesh Khandelwal, Srinivas Ravishankar, Sairam Gurajada, Maria Chang, Rosario Uceda-Sosa, Salim Roukos, Alexander Gray, Guilherme Lima Ryan Riegel, Francois Luus, L Venkata Subramaniam IBM Research ###### Abstract Knowledge Base Question Answering (KBQA) tasks that involve complex reasoning are emerging as an important research direction. However, most KBQA systems struggle with generalizability, particularly on two dimensions: (a) across multiple reasoning types where both datasets and systems have primarily focused on multi-hop reasoning, and (b) across multiple knowledge bases, where KBQA approaches are specifically tuned to a single knowledge base. In this paper, we present SYGMA, a modular approach facilitating generalizability across multiple knowledge bases and multiple reasoning types. Specifically, SYGMA contains three high level modules: 1) KB-agnostic question understanding module that is common across KBs 2) Rules to support additional reasoning types and 3) KB-specific question mapping and answering module to address the KB-specific aspects of the answer extraction. We demonstrate effectiveness of our system by evaluating on datasets belonging to two distinct knowledge bases, DBpedia and Wikidata. In addition, to demonstrate extensibility to additional reasoning types we evaluate on multi-hop reasoning datasets and a new Temporal KBQA benchmark dataset on Wikidata, named TempQA- WD111https://github.com/IBM/tempqa-wd, introduced in this paper. We show that our generalizable approach has better competetive performance on multiple datasets on DBpedia and Wikidata that requires both multi-hop and temporal reasoning. ## 1 Introduction The goal of Knowledge Base Question Answering (KBQA) systems is to answer natural language questions by retrieving and reasoning over facts in a Knowledge Base (KB). KBQA has gained significant popularity both due to its practical real-world applications and challenging research problems associated with it (Fu et al. 2020). However, most existing datasets and research in this area are primarily focused on single/multi-hop reasoning on a single knowledge base (Trivedi et al. 2017; Usbeck et al. 2017; Yih et al. 2016). Consequently, this has encouraged research in methodologies that are tuned to a restricted set of reasoning types on a single knowledge base, in turn lacking generalizability (Kapanipathi et al. 2021; Zou et al. 2014a; Sorokin and Gurevych 2018a). In this work, we propose a modular approach that is generalizable across knowledge bases and reasoning types. Figure 1: A description of how information about Nero’s predecessor is represented in DBpedia vs. Wikidata Open Knowledge Bases such as DBpedia (Auer et al. 2007), Wikidata (Vrandečić and Krötzsch 2014), and Freebase (Bollacker et al. 2008) form the basis of many KBQA datasets. Each of these knowledge Bases, however, have different representations. For instance: information associated to the question “Who was roman emperor before Nero?" from DBpedia and Wikidata are shown in Figure 1. We can see that Wikidata represents properties of facts such as replaces, temporal and spatial with reification.222Information about facts are explicitly mapped by making facts as primitive nodes. On the other hand, DBpedia manages to represent it as a simple fact with the relationship to existing entity nodes. Handling such different representations is an unexplored challenge and requires to be addressed for a generalizable KBQA approach.The second aspect where KBQA approaches fail to generalize is in handling complex reasoning types such as temporal and spatial reasoning. Table 1 shows examples of questions that require different types of reasoning to answer them. Most research have focused on multi-hop reasoning questions (Bordes et al. 2015a; Dubey et al. 2019; Berant et al. 2013), lacking both approaches and datasets that are adaptable to complex reasoning types. Particularly a single approach that can answer questions with different reasoning types. In this paper, we present a modular approach called SYGMA (System for Generalizable and Modular question Answering over knowledge bases), that is built on a framework adaptable to different KB representations and reasoning types. SYGMA introduces intermediate representations based on lambda calculus to adapt to new reasoning types and representations. SYGMA , following the recent advances in KBQA such as NSQA (Kapanipathi et al. 2021) and ReTraCk (Chen et al. 2021), is a modular approach with independent Abstract Meaning Representation, Entity Linking, and Relation Linking modules. In contrast to NSQA which is tuned specifically to DBpedia-based KBQA datasets, SYGMA is generalizable to multiple knowledge bases and multiple reasoning types. In order to evaluate our generalizable approach, we consider datasets that have: (a) DBpedia and Wikidata as knowledge bases, and (b) multi-hop and temporal as the reasoning types. In pursuit of this goal, we also address the lack of temporal reasoning datasets on knowledge bases that we are considering in this work, specifically Wikidata. We annotate and create TempQA-WD; a Wikidata-based dataset focused on temporal reasoning. TempQA-WD is a parallel dataset to TempQuestions dataset (Jia et al. 2018) which is a Freebase based temporal reasoning dataset derived from three other datasets Free917 (Cai and Yates 2013), WebQuestions (Berant et al. 2013) and ComplexQuestions (Bao et al. 2016). While the Freebase dataset consists of question answer pairs along with temporal signal classifications, our dataset is annotated with intermediate SPARQL queries enabling evaluation of different modules in a KBQA pipeline. | Category --- Example | Single-Hop --- | _Who directed Titanic Movie?_ --- SPARQL: select distinct ?a where { wd:Q44578 wdt:P57 ?a} | Multi-hop --- | _Which movie is directed by James Cameron_ --- _starring Leonardo DiCaprio?_ SPARQL: select distinct ?a where {?a wdt:P57 wd:Q42574. ?a wdt:P161 wd:Q38111. } | Temporal --- | _Who was the US President_ --- _during cold war?_ SPARQL: in Figure 3 Table 1: Examples of Single-hop, Multi-hop and Temporal reasoning questions on Wikidata. In summary, the main contributions of this paper are: * • A modular, generalizable approach, called SYGMA, for KBQA with lambda calculus based intermediate representations that enable adaptation to: (a) multiple knowledge bases, specifically DBpedia and Wikidata, and (b) multiple reasoning types such as multi-hop and temporal reasoning. * • A benchmark dataset, called TempQA-WD, for building temporal-aware KBQA systems on Wikidata with parallel annotations on Freebase. * • Experimental results show SYGMA achieves state-of-the-art numbers on LC-QuAD 1.0, WebQSP-WD, and comparable numbers on QALD-9 dataset. We also present baseline numbers for Simple WebQuestions and new dataset introduced in this paper; TempQA-WD. ## 2 Related Work Datasets | Knowledge Base | Emphasis | SPARQL | Intermediate | Natural questions ---|---|---|---|---|--- | | | | Evaluation | QALD-9 | DBpedia | Multi-hop | ✓ | ✓ | ✓ LC-QuAD 1.0 | DBpedia | Multi-hop | ✓ | ✓ | ✗ LC-QuAD 2 | DBpedia, Wikidata | Multi-hop | ✓ | ✓ | ✗ Simple Questions | Freebase, DBpedia, Wikidata | Single-hop | ✓ | ✓ | ✓ Web Questions | Freebase | Multi-hop | ✗ | ✗ | ✓ WebQSP | Freebase, Wikidata | Multi-hop | ✓ | ✓ | ✓ Complex Web Questions | Freebase | Multi-hop | ✗ | ✗ | ✓ TempQuestions | Freebase | Temporal | ✗ | ✗ | ✓ TempQA-WD | Freebase, Wikidata | Temporal | ✓ | ✓ | ✓ Table 2: This table compares most of the KBQA datasets based on features relevant to the work presented in this paper. KBQA Systems: KBQA approaches can be primarily categorized into two groups: (a) Question-to-entities: where techniques output the answer entities from the knowledge graph ignoring the SPARQL query (Saxena, Tripathi, and Talukdar 2020; Sun et al. 2018; Vakulenko et al. 2019), and (b) semantic parsing based: where approaches output intermediate SPARQL queries (or logic forms) that can retrieve the answer from the knowledge base (Singh et al. 2018; Kapanipathi et al. 2021; Zou et al. 2014b). First, in question-to-entities category, (Vakulenko et al. 2019) leverages a message passing technique on a two hop graph from the entities mentioned in the question to retrieve answers from the knowledge graph. (Saxena, Tripathi, and Talukdar 2020) encodes both text and entities, text based on language models and entities based on knowledge graph embeddings (Trouillon et al. 2016) and shows that text can help KBQA in an incomplete setting. In contrast to question-to-entities approach, the semantic parsing based approaches improves interpretability and facilitates evaluation of different modules as shown by Frankestein (Singh et al. 2018), and NSQA (Kapanipathi et al. 2021). Both, Frankestein (Singh et al. 2018), and NSQA (Kapanipathi et al. 2021) follow a pipeline based approach with the differentiating factor being the use of Abstract Meaning Representation as a starting point by NSQA. We build on top of these representational efforts and introduce $\lambda$-expressions with the required additional functions for adapting to new knowledge graphs and new reasoning types. KBQA Datasets: Over the years, many question answering datasets have been developed for KBQA, such as Free917 (Cai and Yates 2013), SimpleQuestions (Bordes et al. 2015b), WebQuestions (Berant et al. 2013), QALD-9 (Usbeck et al. 2017), LC-QuAD 1.0 (Trivedi et al. 2017), and LC-QuAD 2.0 (Dubey et al. 2019). In Table 2, we compare each of these datasets across the following features: (a) underlying KB, including subsequent extensions, e.g. Wikidata (Diefenbach et al. 2017) and DBpedia (Azmy et al. 2018) based versions of SimpleQuestions, as well as the Wikidata subset of WebQSP (Sorokin and Gurevych 2018a); (b) reasoning types that are emphasized in the dataset; (c) availability of SPARQL queries, entities, and relationships for intermediate evaluation; and (d) the use of question templates, which can often generate noisy, unnatural questions. As Table 2 shows, our dataset is distinguished from prior work in its emphasis on temporal reasoning, its application to both Freebase and Wikidata, and its annotation of intermediate representations and SPARQL queries. The most relevant KBQA dataset to our work is TempQuestions (Jia et al. 2018), upon which we base TempQA-WD, as described in dataset section. CRONQUESTIONS (Saxena, Chakrabarti, and Talukdar 2021) is another temporal QA dataset proposed recently over a small subset of Wikidata. It has template based questions targeted at temporal KG embeddings and QA. ## 3 SYGMA: System Description Motivated by Kapanipathi et al. (2021), SYGMA is designed as a neuro-symbolic system, avoiding the need for end-to-end training. Figure 2 shows the overall KBQA pipeline with independent modules and their intermediate representations. The approach is designed with the goal of generalization across multiple KBs and reasoning types. Supporting generalization is accomplished in two stages as shown in Figure 2, driven by an example in Figure 3. The first is a KB- agnostic question understanding stage, which takes in a natural language question as input and outputs an intermediate representation that can be common across different KBs. The second is question mapping and reasoning stage where, first, the necessary heuristics to introduce templates for new reasoning types are applied. Next, the KB-agnostic representation is mapped to the vocabulary of the KG to output KB-specific representation that has a deterministic mapping to SPARQL. We use Wikidata and DBpedia to show generalization across KBs, and use multi-hop and temporal reasoning to evaluate across reasoning types. Before we get into the details of these modules and intermediate representations, we give a brief description of the Lambda calculus and the temporal functions used by SYGMA to generate its intermediate logic representation. Figure 2: Architecture of SYGMA that shows the pipeline with independent modules and intermediate representations. The intermediate representations are supported by appropriate heuristics to facilitate generalizability. Figure 3: An illustration of the outputs at the intermediate stages of the pipeline in SYGMA Lambda Calculus $\lambda$-calculus, by definition, is considered the smallest universal programming language that expresses any computable function. In particular, we have adopted Typed $\lambda$-Calculus presented in (Zettlemoyer and Collins 2012) which support addition of new higher order functions necessary for handling various reasoning types. We use constants and logical connectives like AND, OR, NEGATION and functions like argmin, argmax, count, etc., presented in this work. Apart from this, we also added new temporal functions to demonstrate the system’s adaptability to support new reasoning types. For example, consider the following question and its corresponding logical form: Question: | When was Barack Obama born? ---|--- Logical Form: | | $\lambda$t. born(b,“Barack Obama") $\wedge$ --- interval(t, b) Here, b is instance variable for event born(b, “Barack Obama") and interval(t, b) finds time for the event denoted by b. Variable t is unknown which is marked as $\lambda$ variable. Temporal Functions: We introduce interval, overlap, before, after, teenager, year; where interval gets time interval associated with event and overlap, before, after are used to compare temporal events. teenager gets teenager age interval for a person, and year return year of a date.333Although there are many other possible temporal functions for specific English words, such as adult and tween, we defined teenager because it was featured in the datasets explored here. ### 3.1 KB-Agnostic Question Understanding The modules in this stage aim at deriving logical expressions of the natural language question that is common for all the knowledge bases. In particular, we use the formalism of $\lambda$-calculus for logical representation, i.e., to map questions to their corresponding $\lambda$-expressions representing the semantics. #### AMR In order to have a generic parse (logical expression) that is common across KBs, SYGMA performs initial language understanding using Abstract Meaning Representation (AMR) (Banarescu et al. 2013), a semantic representation language based on PropBank. AMR encodes the meaning of a sentence into a rooted directed acyclic graph where nodes represent concepts and edges represent relations. We adopt an action pointer transformer architecture of Zhou et al. (2021) for transition-based AMR parsing and self-training technique of Lee et al. (2020) for domain adaptation to KBQA. AMR provides generic representation that can be used for multi-hop reasoning (Kapanipathi et al. 2021). However for temporal expressions with :time relation, we have to augment the AMR annotations with implicit predicates that cannot be captured by ellipsis and/or re-entrancies. An example of AMR annotation for the question Who was US president during cold war? is given in Figure 3. The representation encodes time edge with cold war event as sub- event of time. Also in case of before/after constraints it explicitly captures the constraint as part of the time edge. If there are no explicit constraints like before/after or ordinal in the time edge, we treat them as overlap constraints. #### KB-Agnostic Lambda Expression AMR is a fairly granular representation. However, for a KBQA system, specifically on KGs such as DBpedia, Wikidata, and Freebase, such granularity can add noise. Therefore to construct KB-agnostic $\lambda$-expressions of the questions from their corresponding AMRs and their identified entity and relation mentions, we use the transformational heuristics described in Table 3. The table shows the transformation from AMR frame (high level) to the corresponding Lambda expression. The rule type describes where the rule is in general applied to all reasoning (base) vs temporal. We handle these as conjunction of multiple triples in the KB with some projection variable and/or numerical operation to get the final answer. We use AMR unknown/imperative constructs to identify the projection variables and AMR polarity/interrogative frames for queries which fall into boolean answers or ASK questions in SPARQL. The next set of rules present in Table 3 are few sample template rules used to capture temporal events and constraints to help in temporal reasoning. Figure 3 gives an example question that falls into the temporal overlap rule in the table and how the AMR constructs are used to split the events is highlighted with nodes being bold in the table. Complete set of rules used can be found in the supplementary material. Note that we covered temporal reasoning as an additional form of reasoning in the current system. However, more reasoning templates like spatial reasoning with additional operators like coordinates(), south() …etc can be added. Table 3 shows an example template for spatial questions such as "which states are to the south of California?". | Rule --- Type AMR A = (v/frame …) | Lambda Expression L = $\psi$(v) | Base --- | (v/frame :arg0(v0/frame0) … :argn(vn/framen)) --- | frame(v, v0, … vn) $\land$ $\psi$(v0) $\land$ … $\psi$(vn) --- | Base --- | (v/frame :arg1(a/amr-unknown) … :argn(vn/framen)) --- | $\lambda$a. $\psi$(v) --- | Numerical --- | (v/frame :arg0(v0/frame0 :quant(a/amr-unknown)) … --- :argn(vn/framen)) | count($\lambda$v0. $\psi$(v)) --- | Temporal --- | (v/frame :arg0(v0/frame0) … :argn(vn/framen) --- :time(a/amr-unknown)) | $\lambda$ev. $\psi$(v) $\land$ interval(ev, v) --- | Temporal --- | (v/frame :arg0(a/amr-unknown) … :argn(vn/framen) --- :time(b/before :op1(n/nested-frame))) | argmax($\lambda$a. $\psi$(v) , $\lambda$a. $\lambda$ev. $\psi$(n) $\land$ interval(ev, v) --- $\land$ interval(en, n) $\land$ before(ev, en), 0, 1) | Temporal --- | (v/frame :arg0(a/amr-unknown) … :argn(vn/framen) --- :time(n/nested-frame)) | $\lambda$a. $\psi$(v) $\land$ $\psi$(n) $\land$ interval(ev, v) $\land$ interval(en, n) --- $\land$ overlap(ev, en) | Spatial --- | (b/be-located-at-91 :arg0(a/amr-unknown), --- :mod(s/south) :op1(n/nested-frame)) | $\lambda$a. $\psi$(b)$\land$ $\psi$(n) $\land$ coordinate(cb, b) $\land$ coordinate(cn, n) --- $\land$ south(cb, cn) Table 3: AMR to Lambda Translation ### 3.2 Question Mapping and Reasoning This is the second stage which comprises of modules that transform the KB- agnostic $\lambda$-expression of the questions into KB-specific $\lambda$-expression. These modules are entity linking and relationship linking that are specific to the underlying KB along with any KB specific rules to handle special forms of transformation or reasoning. The modules are interchangeable to different knowledge graphs. Currently, we demonstrate the adaptability of these modules to DBpedia and Wikidata. #### KB-Specific Lambda Expression: Structurally, it is similar to $\lambda$-expression with all entities and relations mapped to KB entities and relations respectively. _Entity Linking:_ The goal of Entity Linking is to map entity mentions as captured in the $\lambda$-expression of the question to their corresponding KB-specific entities. We use a recently proposed zero-shot entity linking approach called BLINK (Wu et al. 2020). For a question where entity mentions are already identified, bi-encoder piece of the BLINK is used to predict top-K entities. For this prediction, we use an entity dictionary of 5.9M English Wikipedia entities, mapped to their corresponding Wikidata and DBpedia entities. _Relation Linking:_ SYGMA’s Relation linking component takes in the question text and an AMR graph as input and returns a ranked list of KG relationships. For instance, given a question such as "Who was the US President during cold war?" (see Fig 3) and the corresponding AMR, the goal of the relation linking component is to find the corresponding Wikidata KB relations position held (P39), start time (P580), end time (P582). For this task, we use state-of-the- art AMR-based relation linking approach with models built for both DBpedia and Wikidata (Naseem et al. 2021). SPARQL Query: This module maps KB-specific $\lambda$-expressions onto SPARQL queries through a deterministic approach. Each KB-specific $\lambda$-expression construct is mapped to an equivalent SPARQL construct, as rules given in Table 4. Each of lambda expressions contains one or more terms such that each term $T_{i}$ is comprised of one or more predicates connected via $\land$ or $\lor$. Translation of different predicates is present in Table 4. Some of the KB-spe cific aspects like handling reification in Wikidata are shown in the Table. To get the time interval in case of reified events, Start time(wdt:P580), end time(wdt:P582), or point in time(wdt:P585) connected to intermediate statement node are used. Type | Expression/Predicate E | SPARQL S = $\phi$(E) ---|---|--- | $\lambda$ abstraction --- | $\lambda$x.T --- | SELECT DISTINCT ?x WHERE { $\phi$(T) } --- | Count expression --- | count($\lambda$x.T) --- | SELECT (COUNT(?x) AS ?c) WHERE { $\phi$(T) } --- | Argmax expression --- | argmax($\lambda$x.T1, $\lambda$x.$\lambda$y. T2, O, L) --- | SELECT DISTINCT ?x WHERE { $\phi$(T1) $\phi$(T2) --- } ORDER BY DESC(?yStart) LIMIT L OFFSET O | KB Predicate --- | IRIp(r, s$|$IRIs, o$|$IRIo) --- | ?s$|$IRIs IRIp ?o$|$IRIo. --- | Interval predicate --- for reified facts | wdt:PID(e, s$|$wd:QID1, o$|$wd:QID2) --- $\land$ interval(ei, e) | ?s$|$wd:QID1 p:PID ?e. ?e ps:PID ?o$|$wd:QID2. --- ?e pq:P580 ?eiStart. ?e pq:P582 ?eiEnd. | Overlap predicate --- | overlap(e1i e2i) --- | FILTER(?e1iStart<=?e2iEnd && ?e2iStart<=?e1iEnd) --- | Before predicate --- | before(e1i, e2i) --- | FILTER(?e1iEnd<=?e2iStart) --- | After predicate --- | after(e1i, e2i) --- | FILTER(?e1iStart>=?e2iEnd) --- Table 4: Translation of KB Specific Lambda Expression ## 4 TempQA-WD Dataset In order to evaluate our approach on temporal reasoning, specifically on DBpedia and Wikidata as KGs, we require a dataset to be based on one of these two KGs. TempQuestions (Jia et al. 2018) is the first KBQA dataset intended to focus specifically on temporal reasoning. However, TempQuestions is based on Freebase.We adapt TempQuestions to Wikidata444Creating a parallel dataset on two KGs (DBpedia and Wikidata) is non trivial and labor intensive. Therefore, we opted to create a dataset based on Wikidata alone since it is the most up- to-date KG. to create a temporal QA dataset that has three desirable properties. First, we create a generalizable benchmark that has parallel answer annotations on two KBs. Second, we take advantage of Wikidata’s evolving, up-to-date knowledge. Lastly, we enhance TempQuestions with SPARQL queries which was missing in original dataset. We also add entity, relation, intermediate lambda expression annotations for a subset of the dataset that are used by the SYGMA. There has been two previous attempts at transferring Freebase-QA questions to Wikidata; namely WebQSP-WD (Sorokin and Gurevych 2018a) and SimpleQuestions- WD(SWQ-WD) (Diefenbach, Singh, and Maret 2017). SWQ-WD contains only single triple questions whereas WebQSP-WD have only the final question answers that map directly to corresponding entities in Wikidata. However, as stated (Sorokin and Gurevych 2018a), one challenge is that not all Freebase answers can be directly mapped to entities in Wikidata. For example, the Freebase answer annotation for the question “When did Moscow burn?” is “1812 Fire of Moscow”, despite the year being entangled with the event itself. In contrast, Wikidata explicitly represents the year of this event, with an entity for “Fire in Moscow” and an associated year of “1812”. Thus, a direct mapping between the two answers is not possible, as it would amount to a false equivalence between “1812 Fire of Moscow” and “1812”. In order to address such issues, we enlisted a team of annotators to manually create and verify SPARQL queries, ensuring not only that the SPARQL formulation was correct, but that the answers accurately reflected the required answer type (as in the “Fire in Moscow” example above) and the evolving knowledge in Wikidata. Having SPARQL queries also facilitates intermediate evaluation of the approaches that use semantic parsing to directly generate the query or the query graph, increasing interpretability and performance in some cases (Sorokin and Gurevych 2018a). Dataset Details Dataset | Size | Answer | Add’l Details ---|---|---|--- TempQuestions(Freebase) | 1271 | Freebase | A TempQA-WD (Wikidata) | 839 | Wikidata | A + B A Subset of TempQA-WD | 175 | Wikidata | A + B + C Table 5: Benchmark dataset details. TempQuestions denote original dataset with Freebase answers (Jia et al. 2018). Second row is the subset adapted to Wikidata and the third row is the devset taken out of it. Set-A denote TempQuestions with- {temporal signal, question type, data source}. Set-B -{Wikidata SPARQL , answer, category}. Set-C-{AMR, $\lambda$-expression, entities, relations, KB-specific $\lambda$-expression} Table 5 gives details of our new benchmark dataset. We took all the questions from TempQuestions dataset (of size $1271$) and chose a subset for which we could find Wikidata answers. This subset has 839 questions that constitute our new dataset, TempQA-WD. We annotated this set with their corresponding Wikidata SPARQL queries and Wikidata answers. We also retained the Freebase answers from the TempQuestions dataset effectively creating parallel answers from two KBs. Additionally, we added a question complexity label to each question, according to the reasoning complexity required to answer the question. Details of categorization are present in supplementary material. Within this dataset, we chose a subset of $175$ questions for detailed annotations as described in Set-C in the table description leaving $664$ question as test set. ## 5 Evaluation KB $\rightarrow$ | DBpedia | Wikidata ---|---|--- DataSet $\rightarrow$ | LC-QuAD 1.0 | QALD-9 | WebQSP-WD | SWQ-WD | TempQA-WD Reasoning $\rightarrow$ | multi-hop | multi-hop | multi-hop | single-hop | Temporal System $\downarrow$ | P | R | F1 | P | R | F1 | P | R | F1 | P | R | F1 | P | R | F1 WDAqua | 0.22 | 0.38 | 0.28 | 0.26 | 0.26 | 0.25 | - | - | - | - | - | - | - | - | - gAnswer | - | - | - | 0.29 | 0.32 | 0.29 | - | - | - | - | - | - | - | - | - QAMP | 0.25 | 0.50 | 0.33 | - | - | - | - | - | - | - | - | - | - | - | - NSQA | 0.44 | 0.45 | 0.44 | 0.31 | 0.32 | 0.31 | - | - | - | - | - | - | - | - | - GGNN | - | - | - | - | - | - | 0.27 | 0.32 | 0.26 | - | - | - | - | - | - SYGMA | 0.47 | 0.48 | 0.47 | 0.29 | 0.30 | 0.29 | 0.32 | 0.36 | 0.31 | 0.42 | 0.55 | 0.44 | 0.32 | 0.34 | 0.32 Table 6: SYGMA’s Performance across knowledge bases and reasoning types. Baseline numbers are taken from Kapanipathi et al. (2021) and Sorokin and Gurevych (2018b). P-Precision, R-Recall #### Implementation Details: We implemented our pipeline using a Flow Compiler (Chakravarti et al. 2019) stitching individual modules exposed as gRPC services. We defined the ANTLR grammar to define $\lambda$-expressions. KB-Specific $\lambda$-expression to SPARQL module is implemented in Java using Apache Jena SPARQL modules, to create SPARQL objects and generate the final SPARQL query that is run on target KB end point. The rest of the modules are implemented in Python and are exposed as gRPC services. #### Datasets: We considered two different KBs namely DBpedia and Wikidata to evaluate our system. We did not consider Freebase as its no longer actively maintained and is not up-to-date. Along with the new benchmark dataset TempQA-WD introduced in this paper, we also evaluate our baseline system on two other benchmark datasets adopted to Wikidata, to evaluate its effectiveness beyond temporal questions. They are SWQ-WD (Diefenbach et al. 2017), which consists of 14894 train and 5622 test set questions, and WebQSP-WD (Sorokin and Gurevych 2018b), which consists of 2880 train and 1033 test set questions. For DBpedia, we used QALD-9 (Usbeck et al. 2017), that has 408 training and 150 test questions, and LC-QuAD 1.0 (Trivedi et al. 2017) data set, that has 4k training set and 1k test set. Each of these datasets operate on their own version of the DBpedia and SPARQLs provided and we used the same instance of DBpedia for evaluation. The baseline system is tuned with dev sets of 175 from TempQA-WD and 200 from SWQ-WD train and 200 from LC-QuAD 1.0 and evaluated on all the the five test sets. #### Baselines We Compare SYGMA with various KBQA baseline systems as given in Table 6. To our knowledge there is no other system that works across both DBpedia and Wikidata. We took the numbers reported from NSQA (Kapanipathi et al. 2021) work, which is current state-of-the-art on both LC-QuAD 1.0 and QALD-9 datasets. We also compare our work with GGNN (Sorokin and Gurevych 2018a) which is the only known benchmark for WebQSP-WD dataset. We also compare our work with WDAqua (Diefenbach, Singh, and Maret 2017), gAnswer (Zou et al. 2014b) and QAMP (Vakulenko et al. 2019) systems. ### 5.1 Results & Discussion Table 6 shows performance of SYGMA on two different KBs with five different datasets described in Section 5. We also show the type of reasoning required for each dataset along with the precision/recall and F1 measure. We use GERBIL (Usbeck et al. 2019) to compute performance metrics from the pairs of gold answers and system generated answers from the pipeline. To our knowledge, ours is the first system reporting KBQA numbers across two different KBs and varying reasoning types. We get state of the art numbers beating NSQA (Kapanipathi et al. 2021) on LC-QuAD 1.0 and also achieve comparable numbers for QALD. For WebQSP-WD dataset we get state of the art numbers nearly $20\%$ gain over GGNN (Sorokin and Gurevych 2018a). The accuracy numbers for Wikidata datasets show that there is ample scope for improvement for different modules. This gets evident when we look at the module performances on a small set of development sets across datasets. For WebQSP-WD we did not have any ground truth for evaluating entity linking and relation linking. AMR: Table 7 show the performance of the AMR parser on the 5 development sets in Smatch (standard measure used to measure AMRs) and Exact Match ($\%$ of questions that match fully with ground truth AMR). Dataset | KB | Smatch | Exact Match ---|---|---|--- LC-QuAD 1.0 | DBpedia | 87.6 | 30.0 QALD-9 | DBpedia | 89.3 | 41.8 WebQSP-WD | Wikidata | 88.0 | 43.8 SWQ-WD | Wikidata | 83.0 | 37.8 TempQA-WD | Wikidata | 89.6 | 39.8 Table 7: AMR parser performances on the development sets Entity Linking: Table 8 below captures the performance of entity linking module on different datasets. The question level accuracy is computed by considering all the entities present in each question. However, mention level accuracy is computed by considering single mention at a time. We can see that entity linking performance for Wikidata datasets is low compared to DBpedia. One of the reasons for is all these datasets are adopted from Freebase dataset and the way entities represented in these two KBs is different. Also, very little research is done on entity linking for Wikidata and KBQA in general. Dataset | KB | Accuracy (%) ---|---|--- Mention Level | Question Level LC-QuAD 1.0 | DBpedia | $86.81\;(91.94)$ | $84.00\;(90.50)$ QALD-9 | DBpedia | $89.52\;(94.28)$ | $89.79\;(93.87)$ TempQA-WD | Wikidata | $74.01\;(82.47)$ | $57.14\;(69.71)$ SWQ | Wikidata | $72.27\;(83.18)$ | $70.14\;(81.59)$ Table 8: Entity linking performance on development sets when gold mentions are provided. The numbers inside parentheses denote the Hits@5 scores. Relation Linking: Table 9 captures the performance of the relation linking module. It demonstrates that relation linking is still a challenging tasks specially with multi-hop reasoning and temporal reasoning, where the query graph is disconnected across events, there is room for improvement. Dataset | KB | Precision | Recall | F1 ---|---|---|---|--- LC-QuAD 1.0 | DBpedia | 0.52 | 0.50 | 0.50 QALD-9 | DBpedia | 0.55 | 0.53 | 0.53 TempQA-WD | Wikidata | 0.43 | 0.43 | 0.42 SWQ | Wikidata | 0.67 | 0.68 | 0.67 Table 9: Relation linking performance on development sets. #### Ablation Study: To gain more insights on the performance, we also did an ablation study of SYGMA using TempQA-WD dev set for Wikidata and 100 question dev set from LC- QuAD 1.0 that we manually annotated with ground truth for all the modules. This shows the impact of individual module on overall performance is evaluated. Table 10 shows the results. For example GT-$AMR$ refers to the case where ground truth AMR is fed directly into $\lambda$-module. The table shows large jump in accuracy (in both the datasets) when fed with ground truth entities (GT-EL) and ground truth relations (GT-RL). This points to the need for improved entity linking and relation linking on both datasets across KBs. | TempQA-WD | LC-QuAD 1.0 ---|---|--- | P | R | F1 | P | R | F1 NO GT | 0.47 | 0.50 | 0.47 | 0.50 | 0.52 | 0.50 GT-AMR | 0.50 | 0.51 | 0.50 | 0.51 | 0.53 | 0.51 GT-$\lambda$ | 0.52 | 0.53 | 0.52 | 0.52 | 0.53 | 0.52 GT-EL | 0.60 | 0.62 | 0.60 | 0.63 | 0.66 | 0.64 GT-RL | 0.92 | 0.93 | 0.92 | 0.99 | 0.98 | 0.98 GT-KB-$\lambda$ | 0.93 | 0.93 | 0.93 | 1.0 | 0.99 | 0.99 GT-SPARQL | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 Table 10: Ablation Study on TempQA-WD and LC-QuAD 1.0 dev sets. P-Precision, R-Recall ## 6 Conclusion In this paper, we present SYGMA system that is generalizable across knowledge bases and reasoning types. We introduce KB-agnostic question understanding component that is common across KBs with AMR based intermediate $\lambda$-Calculus representation. Question Mapping and reasoning module on the specific KB is customized per KB. we also presented rule modules that can aid in adding new reasoning type. We introduced a new benchmark dataset TempQA-WD for temporal KBQA on Wikidata. Experimental results show that SYGMA indeed achieves its generalization goals with state of the art results on LC- QuAD 1.0 and WebQSP-WD. ## References * Auer et al. (2007) Auer, S.; Bizer, C.; Kobilarov, G.; Lehmann, J.; Cyganiak, R.; and Ives, Z. 2007\. Dbpedia: A nucleus for a web of open data. In _The semantic web_ , 722–735. Springer. * Azmy et al. (2018) Azmy, M.; Shi, P.; Lin, J.; and Ilyas, I. 2018. Farewell Freebase: Migrating the SimpleQuestions Dataset to DBpedia. In _Proceedings of the 27th International Conference on Computational Linguistics_ , 2093–2103. Santa Fe, New Mexico, USA: Association for Computational Linguistics. * Banarescu et al. (2013) Banarescu, L.; Bonial, C.; Cai, S.; Georgescu, M.; Griffitt, K.; Hermjakob, U.; Knight, K.; Koehn, P.; Palmer, M.; and Schneider, N. 2013. Abstract Meaning Representation for Sembanking. In _Proceedings of Linguistic Annotation Workshop 2013_. * Bao et al. (2016) Bao, J.; Duan, N.; Yan, Z.; Zhou, M.; and Zhao, T. 2016. Constraint-Based Question Answering with Knowledge Graph. In _Proceedings of COLING 2016, the 26th International Conference on Computational Linguistics: Technical Papers_ , 2503–2514. Osaka, Japan: The COLING 2016 Organizing Committee. * Berant et al. (2013) Berant, J.; Chou, A.; Frostig, R.; and Liang, P. 2013. Semantic Parsing on Freebase from Question-Answer Pairs. In _Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing_ , 1533–1544. Seattle, Washington, USA: Association for Computational Linguistics. * Bollacker et al. (2008) Bollacker, K.; Evans, C.; Paritosh, P.; Sturge, T.; and Taylor, J. 2008. Freebase: a collaboratively created graph database for structuring human knowledge. In _Proceedings of the 2008 ACM SIGMOD international conference on Management of data_ , 1247–1250. * Bordes et al. (2015a) Bordes, A.; Usunier, N.; Chopra, S.; and Weston, J. 2015a. Large-scale simple question answering with memory networks. _arXiv preprint arXiv:1506.02075_. * Bordes et al. (2015b) Bordes, A.; Usunier, N.; Chopra, S.; and Weston, J. 2015b. Large-scale Simple Question Answering with Memory Networks. _CoRR_ , abs/1506.02075. * Cai and Yates (2013) Cai, Q.; and Yates, A. 2013. Large-scale Semantic Parsing via Schema Matching and Lexicon Extension. In _Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)_ , 423–433. Sofia, Bulgaria: Association for Computational Linguistics. * Chakravarti et al. (2019) Chakravarti, R.; Pendus, C.; Sakrajda, A.; Ferritto, A.; Pan, L.; Glass, M.; Castelli, V.; Murdock, J. W.; Florian, R.; Roukos, S.; and Sil, A. 2019. CFO: A Framework for Building Production NLP Systems. In _Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP): System Demonstrations_ , 31–36. Hong Kong, China: Association for Computational Linguistics. * Chen et al. (2021) Chen, S.; Liu, Q.; Yu, Z.; Lin, C.-Y.; Lou, J.-G.; and Jiang, F. 2021. ReTraCk: A Flexible and Efficient Framework for Knowledge Base Question Answering. In _Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing: System Demonstrations_ , 325–336. Online: Association for Computational Linguistics. * Diefenbach, Singh, and Maret (2017) Diefenbach, D.; Singh, K.; and Maret, P. 2017. Wdaqua-core0: A question answering component for the research community. In _Semantic Web Evaluation Challenge_ , 84–89. Springer. * Diefenbach et al. (2017) Diefenbach, D.; Tanon, T. P.; Singh, K. D.; and Maret, P. 2017. Question Answering Benchmarks for Wikidata. In _Proceedings of the ISWC 2017 Posters & Demonstrations and Industry Tracks co-located with 16th International Semantic Web Conference (ISWC 2017), Vienna, Austria, October 23rd - to - 25th, 2017._ * Dubey et al. (2019) Dubey, M.; Banerjee, D.; Abdelkawi, A.; and Lehmann, J. 2019. _LC-QuAD 2.0: A Large Dataset for Complex Question Answering over Wikidata and DBpedia_ , 69–78. ISBN 978-3-030-30795-0. * Fu et al. (2020) Fu, B.; Qiu, Y.; Tang, C.; Li, Y.; Yu, H.; and Sun, J. 2020. A survey on complex question answering over knowledge base: Recent advances and challenges. _arXiv preprint arXiv:2007.13069_. * Jia et al. (2018) Jia, Z.; Abujabal, A.; Saha Roy, R.; Strötgen, J.; and Weikum, G. 2018. TempQuestions: A Benchmark for Temporal Question Answering. In _Companion Proceedings of the The Web Conference 2018_ , WWW ’18, 1057–1062. Republic and Canton of Geneva, CHE: International World Wide Web Conferences Steering Committee. ISBN 9781450356404. * Kapanipathi et al. (2021) Kapanipathi, P.; Abdelaziz, I.; Ravishankar, S.; Roukos, S.; Gray, A.; Astudillo, R.; Chang, M.; Cornelio, C.; Dana, S.; Fokoue, A.; et al. 2021. Leveraging Abstract Meaning Representation for Knowledge Base Question Answering. _Findings of the Association for Computational Linguistics: ACL_. * Lee et al. (2020) Lee, Y.-S.; Astudillo, R. F.; Naseem, T.; Reddy, R. G.; Florian, R.; and Roukos, S. 2020. Pushing the Limits of AMR Parsing with Self-Learning. In _Findings of the Association for Computational Linguistics: EMNLP 2020_. * Naseem et al. (2021) Naseem, T.; Ravishankar, S.; Mihindukulasooriya, N.; Abdelaziz, I.; Lee, Y.-S.; Kapanipathi, P.; Roukos, S.; Gliozzo, A.; and Gray, A. 2021. A Semantics-aware Transformer Model of Relation Linking for Knowledge Base Question Answering. In _Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 2: Short Papers)_ , 256–262. Online: Association for Computational Linguistics. * Saxena, Chakrabarti, and Talukdar (2021) Saxena, A.; Chakrabarti, S.; and Talukdar, P. P. 2021. Question Answering Over Temporal Knowledge Graphs. In Zong, C.; Xia, F.; Li, W.; and Navigli, R., eds., _Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, ACL/IJCNLP 2021, (Volume 1: Long Papers), Virtual Event, August 1-6, 2021_ , 6663–6676. Association for Computational Linguistics. * Saxena, Tripathi, and Talukdar (2020) Saxena, A.; Tripathi, A.; and Talukdar, P. 2020. Improving Multi-hop Question Answering over Knowledge Graphs using Knowledge Base Embeddings. In _Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics_ , 4498–4507. Online: Association for Computational Linguistics. * Singh et al. (2018) Singh, K.; Both, A.; Sethupat, A.; and Shekarpour, S. 2018. Frankenstein: a platform enabling reuse of question answering components. In _European Semantic Web Conference_ , 624–638. Springer. * Sorokin and Gurevych (2018a) Sorokin, D.; and Gurevych, I. 2018a. Modeling Semantics with Gated Graph Neural Networks for Knowledge Base Question Answering. In _Proceedings of the 27th International Conference on Computational Linguistics_ , 3306–3317. Association for Computational Linguistics. * Sorokin and Gurevych (2018b) Sorokin, D.; and Gurevych, I. 2018b. Modeling Semantics with Gated Graph Neural Networks for Knowledge Base Question Answering. In _Proceedings of the 27th International Conference on Computational Linguistics_ , 3306–3317. Association for Computational Linguistics. * Sun et al. (2018) Sun, H.; Dhingra, B.; Zaheer, M.; Mazaitis, K.; Salakhutdinov, R.; and Cohen, W. 2018. Open Domain Question Answering Using Early Fusion of Knowledge Bases and Text. In _Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing_ , 4231–4242. Brussels, Belgium: Association for Computational Linguistics. * Trivedi et al. (2017) Trivedi, P.; Maheshwari, G.; Dubey, M.; and Lehmann, J. 2017. Lc-quad: A corpus for complex question answering over knowledge graphs. In _Proceedings of the 16th International Semantic Web Conference (ISWC)_ , 210–218. Springer. * Trouillon et al. (2016) Trouillon, T.; Welbl, J.; Riedel, S.; Gaussier, E.; and Bouchard, G. 2016. Complex Embeddings for Simple Link Prediction. In _Proceedings of the 33rd International Conference on International Conference on Machine Learning - Volume 48_ , ICML’16, 2071–2080. JMLR.org. * Usbeck et al. (2017) Usbeck, R.; Ngonga Ngomo, A.-C.; Haarmann, B.; Krithara, A.; Röder, M.; and Napolitano, G. 2017. 7th Open Challenge on Question Answering over Linked Data (QALD-7). In _Semantic Web Evaluation Challenge_ , 59–69. Springer International Publishing. * Usbeck et al. (2019) Usbeck, R.; Röder, M.; Hoffmann, M.; Conrads, F.; Huthmann, J.; Ngomo, A. N.; Demmler, C.; and Unger, C. 2019. Benchmarking question answering systems. _Semantic Web_ , 10(2): 293–304. * Vakulenko et al. (2019) Vakulenko, S.; Fernández, J.; Polleres, A.; de Rijke, M.; and Cochez, M. 2019\. Message Passing for Complex Question Answering over Knowledge Graphs. In _Proceedings of the 28th ACM International Conference on Information and Knowledge Management (CIKM2019_ , 1431–1440. Beijing, China: ACM. * Vrandečić and Krötzsch (2014) Vrandečić, D.; and Krötzsch, M. 2014. Wikidata: a free collaborative knowledgebase. _Communications of the ACM_ , 57(10): 78–85. * Wu et al. (2020) Wu, L.; Petroni, F.; Josifoski, M.; Riedel, S.; and Zettlemoyer, L. 2020. Scalable Zero-shot Entity Linking with Dense Entity Retrieval. In _EMNLP_ , 6397–6407. * Yih et al. (2016) Yih, W.-t.; Richardson, M.; Meek, C.; Chang, M.-W.; and Suh, J. 2016. The Value of Semantic Parse Labeling for Knowledge Base Question Answering. In _Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers)_ , 201–206. Berlin, Germany: Association for Computational Linguistics. * Zettlemoyer and Collins (2012) Zettlemoyer, L. S.; and Collins, M. 2012. Learning to Map Sentences to Logical Form: Structured Classification with Probabilistic Categorial Grammars. arXiv:1207.1420. * Zhou et al. (2021) Zhou, J.; Naseem, T.; Astudillo, R. F.; and Florian, R. 2021. AMR Parsing with Action-Pointer Transformer. _https://arxiv.org/abs/2104.14674_. * Zou et al. (2014a) Zou, L.; Huang, R.; Wang, H.; Yu, J. X.; He, W.; and Zhao, D. 2014a. Natural language question answering over RDF: a graph data driven approach. In _Proceedings of the 2014 ACM SIGMOD international conference on Management of data_ , 313–324. * Zou et al. (2014b) Zou, L.; Huang, R.; Wang, H.; Yu, J. X.; He, W.; and Zhao, D. 2014b. Natural Language Question Answering over RDF: A Graph Data Driven Approach. In _Proceedings of the 2014 ACM SIGMOD International Conference on Management of Data_ , SIGMOD ’14, 313–324. New York, NY, USA: Association for Computing Machinery. ISBN 9781450323765.
MIDLMedical Imaging with Deep Learning 2022 Short Paper – MIDL 2022 submission – Under Review Haoyue ZhangHaoyue Zhang and Jennifer S Polson contributed equally.1,2<EMAIL_ADDRESS> Jennifer S Polson1,2<EMAIL_ADDRESS> Eric J<EMAIL_ADDRESS> Kambiz Nael3<EMAIL_ADDRESS> William Speier1,3<EMAIL_ADDRESS> Corey W. Arnold1,2,3,4<EMAIL_ADDRESS> 1 Computational Diagnostics Lab, University of California, Los Angeles, CA 90024 USA 2 Department of Bioengineering, University of California, Los Angeles, CA 90024 USA 3 Department of Radiology, University of California, Los Angeles, CA 90024, USA 4 Department of Pathology, University of California, Los Angeles, CA 90024, USA # Predicting Thrombectomy Recanalization from CT Imaging Using Deep Learning Models ###### Abstract For acute ischemic stroke (AIS) patients with large vessel occlusions, clinicians must decide if the benefit of mechanical thrombectomy (MTB) outweighs the risks and potential complications following an invasive procedure. Pre-treatment computed tomography (CT) and angiography (CTA) are widely used to characterize occlusions in the brain vasculature. If a patient is deemed eligible, a modified treatment in cerebral ischemia (mTICI) score will be used to grade how well blood flow is reestablished throughout and following the MTB procedure. An estimation of the likelihood of successful recanalization can support treatment decision-making. In this study, we proposed a fully automated prediction of a patient’s recanalization score using pre-treatment CT and CTA imaging. We designed a spatial cross attention network (SCANet) that utilizes vision transformers to localize to pertinent slices and brain regions. Our top model achieved an average cross-validated ROC-AUC of 77.33 $\pm$ 3.9%. This is a promising result that supports future applications of deep learning on CT and CTA for the identification of eligible AIS patients for MTB. ###### keywords: Acute Ischemic Stroke, Computed Tomography, Deep Learning, Vision Transformers ††editors: Under Review for MIDL 2022 ## 1 Introduction Stroke is the fifth leading cause of death and the leading cause of long-term disability; of the 795,000 new and recurrent strokes each year, acute ischemic stroke (AIS) accounts for 87% of cases [Tsao et al.(2022)]. Mechanical thrombectomy (MTB) is the leading treatment for patients with clots in large blood vessels. In this procedure, a blood clot is surgically removed from an artery to achieve recanalization, i.e., restored blood flow. As a standard measurement for recanalization achieved, a modified treatment in cerebral ischemia (mTICI) score [Tomsick(2007)] is assigned to patients post-treatment. This post-treatment score is clinically significant, as it has been shown that favorable scores, i.e., mTICI 2c or greater, are associated with better clinical outcomes in the long term [Ángel Chamorro et al.(2017)]. Unfavorable scores (mTICI less than 2c) indicate that the treatment did not effectively clear the blood vessel. Imaging has been identified as one modality to illustrate patient physiology that could influence the likelihood of a successful MTB procedure. Predicting final mTICI score prior to a procedure can provide doctors and with more more information when considering treatment options. Deep learning has been shown to leverage the amount of detail in images to improve prediction accuracy [LeCun et al.(2015)LeCun, Bengio, and Hinton]. Current literature presents models that perform semi-automated prediction of mTICI score based on pre-treatment CT imaging, with inconsistent performance [Hilbert et al.(2019), Siddiqui et al.(2021)]. We propose a fully automated model that uses both CT and CTA images to predict mTICI score post- treatment, incorporating attention modules into a deep learning network to effectively localize to informative stroke regions without requiring manual segmentation. ## 2 Data and Methods The cohort used for this study comprises patients treated from 2012-2019. A patient was included in the cohort if they had CT and CTA imaging, underwent thrombectomy for stroke, and were assigned an mTICI score post-MTB. Of the 254 eligible patients, 69 patients were excluded due to missing either CT or CTA series, and 8 were excluded to due unclear stroke location, leaving 177 patients total. The dataset matched demographic distributions seen in other stroke studies, and the target labels were approximately balanced. Patient images were processed using a previously published pipeline adapted for CT, which included brain extraction and registration to a CT template in MNI space [Zhang et al.(2021b)]. Figure 1: Overview of the proposed deep learning architecture. The figure details the entire architecture, neighborhood branch modules, spatial attention transformer module (SAT), and cross attention transformer module (CAT). Utilizing a ResNet backbone, CT and CTA images were used as a slice-wise input (input size 26x224x224) for a global 2D convolutional block, chosen because of the large slice thickness found in stroke protocols [He et al.(2016)]. The outputs from neighboring slices were then fed into ResNet34-based branches that shared weights across slice neighborhoods. The model leveraged two versions of transformer attention modules. A spatial attention transformer (SAT) utilized multi-head attention on each slice to focus on salient regions [Dosovitskiy et al.(2021)]. Within each neighborhood branch, a cross attention transformer (CAT) identified important slices. Finally, the branch outputs were subjected to a weighted softmax layer to ultimately generate binary predictions. The architecture and modules are summarized in Figure 1. The model was trained for 200 epochs with early stopping, using the Adam optimizer with weight decay, a batch size of 12, and a learning rate of 0.0001. Model | ROC-AUC | Accuracy | Precision | Sensitivity | Specificity ---|---|---|---|---|--- Siddiqui et al. | 0.717 | – | – | – | – Hilber et al. | 0.65 $\pm$ 0.10 | – | – | – | – Radiomics | 0.6859 $\pm$ 0.043 | 0.6877 $\pm$ 0.039 | 0.6417 $\pm$ 0.068 | 0.7425 $\pm$ 0.123 | 0.6421 $\pm$ 0.126 ResNet34 | 0.5840 $\pm$ 0.036 | 0.5656 $\pm$ 0.046 | 0.5410 $\pm$ 0.067 | 0.8500 $\pm$ 0.300 | 0.3253 $\pm$ 0.296 SCANet | 0.7732 $\pm$ 0.039 | 0.7523 $\pm$ 0.042 | 0.7424 $\pm$ 0.145 | 0.8250 $\pm$ 0.174 | 0.6905 $\pm$ 0.215 Table 1: Performance of our current model benchmarked against results from literature as well as previously published models applied to this cohort [Hilbert et al.(2019), Siddiqui et al.(2021)]. ## 3 Results and Discussion The results of our experiments are summarized in Table 1. The average ROC-AUC achieved by SCANet was 0.7732 $\pm$ 0.039. This is a significant improvement over the previously published fully automatic deep learning model[Siddiqui et al.(2021)]. Our method also demonstrates higher and more robust performance metrics than the state-of-the-art model requiring manual clot segmentation [Hilbert et al.(2019)]. In addition to the literature benchmarks, SCANet performs better than a radiomics-based model and standard deep learning architecture when trained on the same cohort [Zhang et al.(2021a)]. Clinicians decide to perform MTB based on likelihood of successful recanalization, but it is unknown what factors underlie MTB responses. Clinical images such as CT and CTA contain valuable information to predict procedure outcome, and deep learning models have the capability to learn representations from highly dimensional imaging data. This study sought to predict final MTB recanalization in a fully automatic manner, leveraging recent advances in vision transformers to localize to the stroke region. We showed that our proposed model outperforms prior fully- and semi-automated machine and deep learning models. The primary limitation of our study is the small sample size, which precludes more robust validation. A few future directions include experimenting on a larger dataset across several institutions, optimizing the preprocessing pipeline to more effectively preserve high resolution CTA, and correlation of the immediate treatment response with long-term outcomes. These steps can produce a model that more accurately predicts MTB recanalization, in turn helping doctors and patients in the treatment decision process. ## References * [Dosovitskiy et al.(2021)] Alexey Dosovitskiy et al. An image is worth 16x16 words: Transformers for image recognition at scale. _ICLR_ , 2021. * [He et al.(2016)] Kaiming He et al. Deep residual learning for image recognition. In _Proceedings of the IEEE conference on computer vision and pattern recognition_ , pages 770–778, 2016. * [Hilbert et al.(2019)] Adam Hilbert et al. Data-efficient deep learning of radiological image data for outcome prediction after endovascular treatment of patients with acute ischemic stroke. _Computers in Biology and Medicine_ , 115:103516, 2019. * [LeCun et al.(2015)LeCun, Bengio, and Hinton] Yann LeCun, Yoshua Bengio, and Geoffrey Hinton. Deep learning. _Nature_ , 521(7553):436–444, 2015. * [Siddiqui et al.(2021)] Fazeel M Siddiqui et al. Quantitative assessment of hyperdense sign measured by hounsfield units is associated with unsuccessful mechanical thrombectomy. _Clinical Neuroradiology_ , 31(4):1111–1119, 2021\. * [Tomsick(2007)] Thomas A. Tomsick. Timi, tibi, tici: I came, i saw, i got confused. _American Journal Of Neuroradiology_ , 28(2):382–384, 2007. * [Tsao et al.(2022)] Connie W. Tsao et al. Heart disease and stroke statistics-2022 update: A report from the american heart association. _Circulation_ , 145(8):e391–e426, 2022. * [Zhang et al.(2021a)] Haoyue Zhang et al. A machine learning approach to predict acute ischemic stroke thrombectomy reperfusion using discriminative mr image features. In _2021 IEEE EMBS International Conference on Biomedical and Health Informatics (BHI)_ , pages 1–4, 2021a. * [Zhang et al.(2021b)] Haoyue Zhang et al. Intra-domain task-adaptive transfer learning to determine acute ischemic stroke onset time. _Computerized Medical Imaging and Graphics_ , 90:101926, 2021b. * [Ángel Chamorro et al.(2017)] Ángel Chamorro et al. Complete reperfusion is required for maximal benefits of mechanical thrombectomy in stroke patients. _Scientific Reports_ , 7(1):11636, 2017.
We consider a combined system of regular delivery trucks and crowdsourced drones, available via a sharing economy platform, to provide a technology-assisted crowd-based last-mile delivery experience. We develop analytical models and methods for a system in which package delivery is performed by a big truck carrying many packages to a neighborhood or a town in a metropolitan area and then the packages are assigned to crowdsourced drone operators to deliver them to their final destinations. We develop several optimization models for various cases of the problem and use a combination of heuristic algorithms to solve this NP-hard problem. Finally, we present computational results for the models and algorithms, and conduct an exhaustive sensitivity analysis to check the influence of different parameters and assumptions on the final results. We also provide extensive managerial insights on the benefits of our proposed model if implemented in practice. § INTRODUCTION The number of deliveries and the revenue obtained from delivery operations have been growing continuously and rapidly during the last two decades, thanks to the exponential growth of e-commerce. However, the efficiency of delivery operations remains a big challenge. The last mile of delivery process has consistently been one of the most expensive (nearly or even more than 50% of the total cost), least efficient, and most polluting part of the entire parcel delivery supply chain <cit.>; the fact that Amazon Flex has been paying $18-$25 for Uber-like package delivery services <cit.>, while they have not increased their hourly wages to $15 up until just recently <cit.>, speaks to the expensiveness of the last-mile delivery operations. The expensiveness of last-mile delivery is due to several factors including the facts that it is a labor-intense operation, it is a scattered operation serving different individual customers at dispersed places, which often results in underutilized carrier capacity and the least economy of scale in the whole delivery operation, and that such deliveries are usually very time-consuming because of road congestion, accessibility of the destinations, and most importantly unattended deliveries. The rapidly increasing importance of same day and same hour delivery in our lives will make this operation even more inefficient. Such deliveries are mostly used for low-value high-frequency products such as grocery items for which the shipping cost could quickly become disproportionate in the eyes of consumers. The advancement of technology can revolutionize the conventional delivery practices and boost the efficiency. Among these advancements are the recent efforts to adopt autonomous vehicles, unmanned aerial vehicles (UAVs), automated guided vehicles (AGVs), and other droids in package delivery operations. The integration of autonomous and semi-autonomous technologies into the last-mile delivery operations in a centralized or decentralized manner have the potential to remove or mitigate the long-lasting factors such as pooling and routing inefficiencies that have been contributing to the expensiveness of last-mile delivery. Utilizing drones, for example, can lead to significant savings in the delivery operation and have been explored for a few years <cit.>. According to <cit.>, Amazon's shipping costs in the period between 2016 and 2019 was around $6,400M, $8,025M, $10,025M, $12,375M, and Prime Air operations could save about $1,044M, $2,677M, $5,147M, $8,549M from these costs. In an earlier work <cit.>, we have shown that for a centralized delivery system to be competitive with the decentralized household shopping model, a very large portion of the population have to adopt the centralized system and shows inefficient pooling as the primary cause of inefficient last-mile delivery. In this paper, we analyze the impact of decentralization, in particular crowdsourcing of the last part of the last-mile delivery operations when integrated with new technologies, on the efficiency of pooling and clustering customers. To reduce the cost of the last mile delivery operation and reducing the delivery time (customer's waiting time), we analyze the traditional last mile delivery operation when conducted in coordination with crowdsourced vehicles, bikes, or drones, as illustrated in Figure <ref>. These crowdsourced delivery operators can work in the context of sharing economy and form one side of a two-sided market platform. In fact, this provides a combination of centralized and decentralized systems. Sharing economy platforms, crowdsourced agents, and last-mile delivery. In this paper, we combine the crowdsourced autonomous delivery vessels with regular delivery trucks, vans, cars, and bikes to provide a technology-assisted last-mile delivery experience and a better and smoother transition to a fully autonomous parcel delivery ecosystem. We develop analytical models and methods for one of these intermediary systems, in which package delivery is performed by a big truck carrying a large number of packages to a neighborhood or a town in a metropolitan area and then assign the packages to crowdsourced delivery agents who operate drones[The technology around civil applications of drones is developing rapidly, that can be observed by a quick look into the websites of major companies in this field. Many companies have developed different types of drones for video recording or delivering products. DJI is a famous manufacturer of drones for customers to record videos or take photos. Phantom 4 Pro V2.0 is one of its recent products which costs $1599 and can travel 30 minutes with a full battery. Its speed is around 45 mph and weighs 1375 g. Amazon, Google, Domino, and DHL are all developing their own drones for delivering packages. Amazon's Prime Air can carry products with a weight of 2.26 kg and fly for 30 minutes with 50 mph speed, while it can be controlled 10 miles away from the base. One could compare this to the Google Wing that has a capacity of 1.5 kg and can fly at 75 mph within 8.7 miles from the base. DHL's commercial drones are the biggest ones with 2200 mm diagonal size and can fly at 43.5 mph within 7.5 miles from the base carrying packages up to 3 kg.] to deliver them to their final destinations. Other applications such as post disaster response operations in which trucks cannot access all customers and have to rely on drones in the neighborhood can also be considered for this problem. We introduced this problem in our preliminary work <cit.>. In this paper, we expand that and study and analyze this problem in a more comprehensive way. To the best of our knowledge, besides our earlier extended abstract, this is the first work studying this problem. A combination of heuristic algorithms is used to solve this NP-hard problem and an exhaustive sensitivity analysis is done to check the influence of different parameters and assumptions such as speed ratio of drones and trucks, the number of drones in the service region, and the distribution of the customers. The simulation results show significant savings in the total delivery cost under reasonable assumptions. §.§ Related Work Relevant work in the literature are mainly in two categories: drone-aided delivery and crowdsourced delivery, with the latter being more relevant to our work. The literature of drone-aided delivery is so young but has been expanding rapidly. It can be divided into four main categories; travelling salesman problem (TSP) with drones, vehicle routing problem (VRP) with drones, Done(-only) delivery problem, and carrier-vehicle with drones <cit.>. Murray and Chu <cit.> introduced routing problems for two combined systems of truck and drone, namely flying sidekick travelling salesman problem (FSTSP) and parallel drone scheduling TSP. The first problem and its variants have been called with several names including HorseFly Problem, TSPD, Truck-and-Drone TSP, etc. Since this first paper in 2015 the literature on such problems have been growing rapidly (see <cit.> for example). Although these problems resemble some similarities with the problem we study in this paper, but they are essentially solving a different problem. The main difference is that all of these problems are defined under a centralized delivery system, while our problem combines a centralized truck delivery system and decentralized drone delivery system, and to the best of our knowledge is the first to do so. We will show that the drone(-only) delivery problem is the centralized delivery version of a special case of our problem when the range and speed of drones are large enough, while the pure TSP model is the centralized alternative for the case where the flying range and speed of drones are rather small. For a review on the literature of centralized drone-aided delivery problems please see <cit.>. Crowdsourced delivery, in essence, can be categorized under pickup and delivery problem (PDP), where the goal is to transport parcels from origins to destinations while minimizing costs. The PDP has been studied for almost half a century; see <cit.> for overview on this body of literature. Unlike the traditional PDP setting, our problem has on-demand features, which is also somewhat related to the literature of dynamic PDP <cit.>, and the delivery operation is partially conducted on a sharing economy platform. Sharing economy indicates a system in which people share access to goods and services as opposed to ownership and it has been extensively studied <cit.>. However, many aspects of the application of sharing economy systems, especially those with (semi-)autonomous technologies, in logistics and delivery services has received less attention <cit.>. The paper <cit.> introduced the idea of “crowdphysics” – crowdsourcing tasks that require people to collaborate and synchronize both in time and physical space and discussed the concept using a crowd-powered delivery service. The paper <cit.> introduced a typology of the early startup businesses. A first framework for crowd logistics was proposed in <cit.>. The paper <cit.> developed a more comprehensive framework for crowd-based operations in city-level logistics. The authors discussed different kinds of crowd-based logistics including crowdsourced delivery, cargo-hitching (i.e., integration of freight and passenger transport), receiving packages through pickup locations, providing storage by the crowd, and returning (reverse flow) logistics. There have been a number of theoretical and experimental research related to each of these topics, out of which we review those relevant to crowdsourced delivery. In 2013 Wal-Mart started to consider the idea the idea of encouraging individuals/shoppers in a store who are willing to deliver packages for online customers on their way back home <cit.>. Later, Amazon started considering the crowd as couriers <cit.>. Many researchers explored the crowdsourced delivery service with crowdsourced drivers coming to the package center to pickup the packages and deliver them to their destination. Among the relevant theoretical research, the paper <cit.> studied this idea and formulated it as a vehicle routing problem with occasional drivers (VRPOD). The paper <cit.> studied this problem more in-depth under both static and dynamic settings, in which the latter incorporates probabilistic information about future online orders and in-store customer arrivals. The paper <cit.> proposed a stochastic optimization model for the same problem. A similar problem with stochastic supply of crowd vehicle and time windows was analyzed in <cit.>. The authors presented a two-stage stochastic model, an exact branch and price algorithm, a column generation heuristic. The paper <cit.> also considered a group of ad-hoc drivers for performing parcel delivery tasks and proposed a route planning system that automatically makes the matching between ad-hoc drivers and these tasks. The platform also operates a fleet of dedicated vehicles for delivery tasks that cannot be completed by the ad-hoc drivers. The authors present a rolling horizon framework and an exact solution approach based on a matching formulation to solve the problem. They also compared their results with the traditional delivery system and concluded that the use of crowdsourced drivers can significantly reduce the costs. On the experimental side, the paper <cit.> tried to leverage social networks to find available crowd couriers. They compared three delivery systems with respect to their greenhouse gas emission: the traditional delivery system, a pickup location delivery system (PLS) in which parcels are delivered to a pickup center and customers pick them up on their way to home, work, etc, and finally a socially networked PLS (SPLS) where the last-mile delivery from the pickup locations are done by agents socially connected to the customers. They showed that the PLS system is worse than the traditional delivery system but SPLS significantly reduces the emission. The paper <cit.> also studied the idea of creating a social network-supported last mile delivery and developed a simulation model for it, in which they exploit the social networks of the online customer to find potential couriers in their networks. A trial with crowd participants for book delivery in a Finish library was conducted in <cit.> that showed crowdsourced delivery reduces vehicle miles traveled. The paper <cit.> used a survey to analyze potential driver behavior in choosing to change status from pure commuters to traveler-shippers. Meanwhile, the paper <cit.> also created a survey to study the determinants of crowd-shipping acceptance among drivers. The paper <cit.> developed an agent-based simulation model for the crowdsourced last-mile delivery service with the existence of central pickup location/warehouse and identified the important factors influencing its performance. They set the packages to have three types of accepted delivery time windows and three kinds of crowdsourced agents, including driving, cycling, or walking, delivering one package each time. In contrast to an entirely crowdsourced fleet, the performance of a hybrid fleet of privately owned delivery vehicles and crowdsourced assets was simulated in <cit.>. Relatively fewer papers in the literature have considered a cooperative delivery system with cooperation between traditional delivery trucks and crowdsourced carriers or between the crowdsources themselves. The paper <cit.> evaluated the use of shared mobility for last mile delivery services in coordination with delivery trucks. They tried to minimize the combined transportation and outsourcing cost of the trucks and shared mobility. They also considered minimizing greenhouse gas emissions as one of their objectives. They used an analytical model and found that crowdsourced shared mobility is not as economically scalable as the conventional truck-only system with respect to the operating costs, due to the fact that the payment to crowdsourced drivers is route specific and accounts for the competition with the ride-share service market. However, they state that a transition towards this model can create other economic and operational benefits such as reducing the truck fleet size, avoiding high-demand areas and peak hours, and adjusting vehicle loading capacities. Cooperation is usually conducted through intermediary depots. A crowd-delivery problem with a number of pop stations as intermediary between trucks and crowdsource workers was modeled as a network min-cost flow problem in <cit.> in which workers are assigned to their closest pop station. The paper <cit.> considered cyclists and pedestrians as crowdsources who are close to customers and can perform the last-leg parcel delivery and the first-leg parcel pickup with relaying the parcels with a truck carrier at some relay points selected from a pre-defined set of such points. They formulated the problem as a mixed integer non-linear program to simultaneously select crowdsources and relay points and find the truck routes and schedule. The crowdsource routes and their feasibility are determined by the crowdsources themselves before making a bid (that includes the job they are willing to do and the price) and the platform only selects the the best offered bids. Pedestrians were also considered in <cit.> as crowd delivery partners assigned to their closest relay point determined by a circle packing scheme. A similar problem was discussed in <cit.> and was formulated as a vehicle routing problem (VRP) type model, in which customers can be served by either an operator's truck or a crowdsourcing partner that received the package at a relay point. The paper <cit.> also considered a VRPOD with time windows and intermediate depots. The authors also considered the scenario in which the crowdsourced drivers are forbidden at the central depot and can only visit the transshipment centers. Similar transshipment points with occasional drivers were considered in a pickup and delivery problem in <cit.>. The same problem was considered in <cit.> assuming the existence of pickup locations where the packages could be dropped off and the customers can walk and pick up their packages there. A multi-objective version was analyzed in <cit.>. Another collaborative system between delivery truck and crowdsourced drivers was considered in a recent paper <cit.> for which the authors present a heuristic algorithm that solves the involved routing and assignment problems separately. The idea of relaying packages have also been used in a setting that crowd agents could cooperate with each other. The paper <cit.> analyzed a multi-driver multi-parcel matching problem (MDMPMP) based on the multi-hop ride sharing problem, in which parcels may be transported by a single or by multiple drivers, being relayed between drivers on the path from parcels origin to its destination. Crowd drivers have preset routes from their origins to their destinations from which can deviate slightly to pick up or drop off a package. There is a lack of a study on the design of a cooperative delivery system with a truck and autonomous or semi-autonomous crowdsourced carriers. In this paper, we fill this gap. §.§ Regulations Before we get into the details of the problem and our solution approach, it is necessary to discuss the applicability of the studied problem. One important area that needs to be considered in drone-related operations is their regulations. In 2016, the U.S. government announced that commercial use of drones needs to be registered by the Federal Aviation Administration(FAA) and Department of Transportation(DOT) and need to follow the new regulations <cit.>. The person who operates the drones needs either to be at least 16 years old and get a remote pilot certificate with a small unmanned aircraft systems (UAS) rating or to be directly supervised by someone who already has the certificate <cit.>. The regulations state that all drones including their carried materials or products must weigh less than 55 lbs which is 25 kg, the limit speed of drones is 100 mph, and the maximum altitude is 400 feet above ground level. The drones need to be within the visual line-of-sight of the remote pilot during the flight and need to be close enough to receive the command from the pilot. Meanwhile, all the drones must only be operated in daylight which is from 30 minutes before official sunrise to 30 minutes after official sunset. They need to yield the way to other aircraft and the minimum weather visibility is 3 miles from a control station. One person can only operate one drone at a time and cannot manipulate a moving vehicle unless it is a sparsely populated area. Moreover, drones cannot carry hazardous materials <cit.>. Other regulations are mainly related to the safety issues that drones need to be checked before their launching and the object being carried by the drone is securely attached. On the other hand, most of the restrictions can be waived if the applicant demonstrates that the operation can safely be conducted under the terms of a certificate of waiver <cit.>. § PROBLEM STATEMENT Consider a region $R$ as a residential area in which one truck has to go through all the neighborhoods to deliver some packages. There are also private drone operators in the area that could deliver packages from the truck to their final destination (households). When the truck stops at a neighborhood corner, the crowd-based drones, after receiving an order from the courier, will fly from their base to that corner to pick up the packages, deliver them to the customers and go back to their base, i.e., operator's house, for recharging its battery. The objective is to design a coordinated system between the truck and these crowdsourced drones in a way that minimizes the total time spent on fulfilling the demand of all customers in that area. Figure <ref> shows an illustration of this cooperative delivery between a truck and crowd-sourced drones. In this problem we assume that: * Each drone can only carry one item at a time. * The charging time for drones at its home base is 0 (i.e., changing quickly to a new battery). * Each drone base location launches only one drone. * The speed of all drones is set to be fixed and equal to each other. * There is no weight limit for a drone to carry the package. So all packages of a customer will be delivered in one trip. * Only one truck is used and its capacity is sufficient for the entire demand in the region. * If there is no drone nearby, the truck will serve all the customers in that neighborhood. * Each drone can be only used at one truck's stop (cluster center). * The truck will wait at the pickup center to track the drones until the driver receives the confirmation that the packages are delivered to the customers by the drone and send necessary instructions to drone operators in case of delivery failures. * The compensation mechanism for drone delivery tasks is assumed to be set by the platform separately. A schematic representation of the proposed crowdsourced drone delivery. Depending on whether there are one or multiple pickup centers, or whether drones need to go back to their home bases after each delivery or not, we define four problems as followings: Problem I–(One Center with Recharging): Problem with one pickup center with the assumption that after each delivery, drones should stop at their home bases (for recharging/replacing battery) before revisiting the pickup center for another delivery. Problem II–(One Center with Revisiting): Problem with one pickup center with the assumption that after each delivery, drones could revisit the pickup center for another delivery before going back to their home bases. Problem III–(Multiple Centers with Recharging): Problem with multiple pickup centers with the assumption that after each delivery, drones should stop at their home base (for recharging/replacing battery) before revisiting the pickup centers for another delivery. Problem IV–(Multiple Centers with Revisiting): Problem with multiple pickup centers with the assumption that after each delivery, drones could revisit the pickup centers for another delivery before going back to their home base. The four considered problems. Recharging Revisiting One Center Problem I Problem II Multiple Centers Problem III Problem IV <Ref> summarizes the conditions of these four problems. In the followings we will model these four problems. In the rest of this section, we delve into each of these problems and model them mathematically. The following notations are commonly used in all of the presented models: we denote the set of customer nodes with $C$ and set of drone nodes with $D$ and let $v_D$ be the speed of drones, $v_T$ be the speed of trucks, $L$ be the longest distance (range) a drone can travel, and $M$ to be a sifficiently large number. Note that setting $L=0$ reduces problems III and IV to TSP. §.§ Problem I–(One Center with Recharging) In this problem we have one pickup center and assume that drones should go back to their home base for recharging after each delivery. There is no truck route and truck (or a store, warehouse, or distribution center) operates as a depot for a fleet of crowdsourced drones to pick up the package and deliver them to the customers. The problem with one center is important to study because it sets a a basis for the multiple centers problems and also it helps to understand the dynamics inside a cluster, introduced later, in a better way. The insight behind our algorithm is partly driven by this sub-problem. Before we present our model for this problem we define the sets, parameters, and variables in Table <ref>. Sets, parameters, and variables used in the one center model. Parameters/Variables Description $d_{ij}$ Route length going from node $i\in D$ to the center node (0) and then to node $j\in C$ and back to node $i$ $x_{ij}$ Binary decision variable. It is $1$ when a drone travels from node $i\in D$ to the center node 0 and then to node $j\in C$ and back to node $i$. It is $0$ otherwise. $q_{i}$ Total travel time of the drone with base at node $i$ $Q$ Maximum time spent by all drones (makespan) A mixed integer linear programming (MILP) model for this problem can be written as follows: \begin{eqnarray} \minimize \qquad Q & & \quad \qquad \st \nonumber \\ \sum_{i\in D} x_{ij} & = & 1\,, \qquad \forall j\in C \label{eq:Problem1SatisfyAll} \\ d_{ij}x_{ij} & \leq & L\,, \qquad \forall i\in D, \; \forall j\in C \label{eq:Problem1DroneRange} \\ \frac{1}{v_D} \sum_{j \in C} d_{ij}x_{ij} & \leq & q_i \,, \qquad \forall i \in D \label{eq:Problem1DroneTime} \\ 0 \;\; \leq \;\; q_i & \leq & Q\,, \qquad \forall i\in D \label{eq:Problem1TotalTime} \\ x_{ij} & \in & \{0, 1\}\,, \quad \forall i\in D, j\in C \nonumber \end{eqnarray} The objective is to minimize the maximum time of drone routes (makespan). Constraint (<ref>) makes sure all customers are visited once. Constraint (<ref>) ensures that the distance traveled by each drone does not exceed the maximum travel distance (range) of drones. Constraints (<ref>) finds the time spent by each drone and Constraint (<ref>) calculates the maximum time among all drones. An alternative and less efficient but more flexible (adjustable) model for this problem is provided in <ref> of the Online Supplement of this paper. §.§ Problem II–(One Center with Revisiting) In this problem we consider one pickup center (no truck route) and after each delivery, drones could revisit the pickup center for another delivery before going back to their home bases. We first note that a feasible drone route in this case is a route that starts from a drone node $i\in D$, satisfies the demand of one or multiple customer nodes after picking their packages up at the center node, and comes back to the starting drone node while not violating the drone range $L$. Let $\Pi_i$ denote the set of feasible drone routes starting from drone node $i\in D$ and let $\Pi_i(j) \subset \Pi_i$ be the set of routes that contain the customer node $j\in C$. We use $k$ to index these routes. For any route $\pi_{ik} \subset \Pi_i$ let $d_{\pi_{ik}}$ denote its total length. We define our decision variable as x_π_ik = { 1 if route π_ik ∈Π_i is selected; 0 . . Then we can form an MILP formulation: \begin{eqnarray} \minimize \qquad Q \qquad & & \qquad \qquad \st \nonumber \\ & & \nonumber \\ \sum_{i} \sum_{\pi_{ik} \in \Pi_i(j)} x_{\pi_{ik}} & = & 1\,, \quad \qquad \forall j\in C \label{eq:Problem2SatisfyAll} \\ d_{\pi_{ik}} x_{\pi_{ik}} & \leq & L \;\; \quad \qquad \forall i\in D,\; \pi_{ik} \in \Pi_i \label{eq:Problem2DroneRange} \\ \frac{1}{v_D} \sum_{\pi_{ik} \in \Pi_i} d_{\pi_{ik}} x_{\pi_{ik}} & \leq & q_i\,, \quad \qquad \forall i \in D \label{eq:Problem2DroneTime} \\ 0 \;\; \leq \;\; q_i & \leq & Q\,, \quad \qquad \forall i\in D \label{eq:Problem2TotalTime} \\ x_{\pi_{ik}} & \in & \{0, 1\}\,, \;\;\quad \forall i\in D, \; \pi_{ik} \in \Pi_i \nonumber %\\ \end{eqnarray} Constraint (<ref>) ensures all customers are visited once. Constraints (<ref>) and (<ref>) find the maximum time among all drone routes. An alternative and less efficient but more flexible (adjustable) model for this problem is provided in <ref> of the Online Supplement of this paper. §.§ Problem III–(Multiple Centers with Recharging) In this problem we can have multiple pickup centers and drones should go back to their home base for recharging after each delivery. All the customer and drone nodes are in a region $R$. We first assume that truck nodes or centers can be placed anywhere in $R$ (we will change this later). This implies that all deliveries are done by the drones and the truck carries package through different neighborhoods and stops at one center in each neighborhood where it passes the packages to the drones. Although the pickup centers are not necessarily the same as customers locations but if distance between pickup centers and some customers are less than or equal to a threshold, we assume that those customers are served by the truck. The sets, parameters, and variables used in our model are defined in Table <ref>. Note that some of the variables are infinite dimensional. Sets, parameters, and variables used in the general (multi-center) model. Parameters/Variables Description $d_{ij}$ Distance between node $i\in D$ and $j\in C$ $\bm{y}_i$ Vector of location of drone $i \in D$ $\bm{z}_j$ Vector of location of customer $j \in C$ $\tau$ Distance threshold for serving the customer by the truck (if the distance between truck's stop and a customer is less than or equal to this threshold) $N_T$ Total number of truck centers $\bm{x}_m$ Vector of location of center $m$ with $\bm{x}_m \in R$ and $m = 1, \dots, N_T$ $D \times R \times C \xrightarrow{} \{0,1\}$. It is $1$ when drone travels from node $i\in D$ to center $\bm{x}_m$ then to node $j\in C$ and back to node $i$. It is $0$ otherwise. $\zeta_{\bm{x}_mj}$ $R \times C \xrightarrow{} \{0,1\}$. It is $1$ when customer $j\in C$ is served by the truck that has stopped at location $\bm{x}_m$ and 0 otherwise $q_{i\bm{x}_m}$ Total distance travelled by drone node $i$ assigned to center $\bm{x}_m$ $Q_{\bm{x}_m}$ Maximum distance among all $q_{i\bm{x}_m}$ $R \times R \xrightarrow{} \{0,1\}$. It is $1$ if the path from center $\bm{x}_m$ to $\bm{x}_n$ is used by the truck $u_{\bm{x}_m}$ A dummy variable We can model this problem in its most general sense as an infinite dimensional optimization problem that follows: \begin{eqnarray} \minimize \; \frac{1}{v_T} \iint_{\bm{x}_m\,, \bm{x}_n \in R} \|\bm{x}_m-\bm{x}_n\| \gamma_{(\bm{x}_m\bm{x}_n)} \, dA & + & \iint_{\bm{x}_m \in R} Q_{\bm{x}_m} \, dA \qquad \qquad \st \nonumber \\ \sum_{i\in D}\iint_{\bm{x}_m \in R} \xi_{i\bm{x}_mj}\, dA & = & 1\,, \qquad \qquad \forall j\in C \label{eq:Problem3CentersAnywhere-AllCustomersServed} \\ \|\bm{x}_m - \bm{z}_j\| & \leq & \tau + M (1- \zeta_{\bm{x}_mj}) \,, \quad \forall j \in C\,, \; \forall \bm{x}_m \in R \label{eq:Problem3CentersAnywhere-TruckServing} \\ (\|\bm{x}_m - \bm{y}_i\| + \|\bm{x}_m - \bm{z}_j\| + d_{ij}) \, \xi_{i\bm{x}_mj} & \leq &L\,, \qquad \qquad \forall i \in D\,, \; \forall j \in C\,, \; \forall \bm{x}_m \in R \label{eq:Problem3CentersAnywhere-DroneRange} \\ \frac{1}{v_D} \sum_{j\in C} (\|\bm{x}_m - \bm{y}_i\| + \|\bm{x}_m - \bm{z}_j\| + d_{ij}) \, \xi_{i\bm{x}_mj} \, (1-\zeta_{\bm{x}_mj}) & \leq & q_{i\bm{x}_m} \,, \qquad \forall i \in D\,, \; \forall \bm{x}_m \in R \label{eq:Problem3CentersAnywhere-CenterDroneTime} \\ q_{i\bm{x}_m} - \frac{1}{v_D} \max_{j} d_{ji} \xi_{i\bm{x}_mj} & \leq & Q_{\bm{x}_m}\,, \qquad \forall i\in D\,, \; \forall \bm{x}_m \in R \label{eq:Problem3CentersAnywhere-CenterMakespan} \\ \iint_{\bm{x}_m \in R,\bm{x}_m\neq \bm{x}_n}{\gamma_{(\bm{x}_m\bm{x}_n)}} dA & = & 1\,, \qquad \qquad \forall \bm{x}_n \in R \label{eq:Problem3CentersAnywhere-TSPassignment1} \\ \iint_{\bm{x}_n \in R,\bm{x}_n\neq \bm{x}_m}\gamma_{(\bm{x}_m\bm{x}_n)} dA & = & 1\,, \qquad \qquad \forall \bm{x}_m \in R \label{eq:Problem3CentersAnywhere-TSPassignment2} \\ \iint_{\bm{x}_m\,, \bm{x}_n \in R} \gamma_{\bm{x}_m\bm{x}_n} dA & = & N_T\,, \label{eq:Problem3CentersAnywhere-CentersNumber} \\ u_{\bm{x}_m}-u_{\bm{x}_n} + N_T \; \gamma_{\bm{x}_m\bm{x}_n} & \leq & N_{T}-1 \,, \qquad \forall \; 2\leq u_{\bm{x}_m} \neq u_{\bm{x}_n} \leq N_T \label{eq:Problem3CentersAnywhere-TSPsubtour} \\ d_{i\bm{x}_mj}, \; d_{\bm{x}_m\bm{x}_n}, \; q_{i\bm{x}_m}, \; Q_{\bm{x}_m} & \geq & 0\,, \qquad \qquad \; \forall i\in D\,, \; \forall j\in C\,, \; \forall \bm{x}_m \in R \nonumber \\ \zeta_{\bm{x}_mj}, \; \xi_{i\bm{x}_mj}, \; \gamma_{\bm{x}_m\bm{x}_n} & \in & \{0,1\} \,, \qquad \forall j\in C\,, \; \forall \bm{x}_m\,, \bm{x}_n \in R \nonumber \\ N_T, \; u_{\bm{x}_m} & \in & \N \,, \qquad \qquad \forall \bm{x}_m \in R \nonumber %\\ \end{eqnarray} The objective function minimizes the sum of the time that truck takes to travel between the pickup centers on its route and the total time that it takes to serve clusters of customers with drones at these stopping points. Constraint (<ref>) ensures all customer are visited once. Constraint (<ref>) controls whether a customer is served by the truck. Constraint (<ref>) enforces the drone battery limit. Constraints (<ref>) and (<ref>) calculates the time cost in each center, excluding the time spent on the last edge of last route of each drone. Constraint (<ref>)-(<ref>) are travelling salesman problem (TSP) constraints for the truck as well as to calculate number of pickup centers. This is obviously a complex problem to solve and to make it easier we need some relaxations. To do this, we assume that the truck only stops at a customer location and while stopping there that location will serve as a pickup center for drones as well. Luckily, this simple and very reasonable relaxation, makes the problem much easier to model and solve. This assumption is confirmed to be valid and beneficial by findings of <cit.> that showed the benefits of having dedicated drivers in parallel to crowd drivers in the context of dynamic ride-sharing, where the dedicated drivers, in addition to satisfying part of the demand, could also serve riders for whom no ad-hoc drivers are available. The sets, parameters, and variables used in this new model are defined in Table <ref>. Sets, parameters, and variables used in the general (multi-center) model. Parameters/Variables Description $d_{pp'}$ Distance between node $p\in C$ and $p'\in C$ $d^p_{ij}$ Route length going from node $i\in D$ to the center node $p\in C$ and then to customer node $j\in C$ and back to node $i$ $\delta_{ji}$ Distance of customer node $j\in C$ to the drone node $i\in D$ $y_p$ Binary decision variable that is 1 if node $p\in C$ is served by the truck and 0 otherwise $x^p_{ij}$ Binary decision variable that is $1$ when a drone travels from node $i\in D$ to the center node $p\in C$ and then to customer node $j\in C$ and back to node $i$ and is $0$ otherwise $z_{ip}$ Binary decision variable that is $1$ when drone $i\in D$ is used for delivery at center node $p\in C$ and is $0$ otherwise $q_{ip}$ Total travel time of the drone with base at node $i\in D$ that is assigned to center $p\in C$ $Q_p$ Maximum time spent by all drones assigned to center $p\in C$ $\Delta_{ip}$ Maximum length of the last tour segment among all tours formed by drone $i \in D$ through center $p \in C$ $\eta^{p}_{ij}$ Binary decision variable equal to 1 if $d_{ji}x^p_{ij} = \Delta_{ip}$ and 0 otherwise $\gamma_{pp'}$ Binary decision variable equal to 1 if the path from $p$ to $p'$ has been used by the truck An MILP model for this problem can be written as following: \begin{eqnarray} \minimize \;\;\; \frac{1}{v_T}\sum_{p\in C}\sum_{p'\in C} d_{pp'}\gamma_{pp'} & + & \sum_{p} Q_p \qquad \qquad \qquad \st \nonumber \\ \sum_{p\in C} y_p & \geq & 1 \; , \label{eq:Problem3CentersOnCustomersMinOneTruckNode} \\ \sum_{i\in D}\sum_{p\in C} x^p_{ij} & = & 1-y_j \; , \quad \qquad \qquad \forall j\in C \label{eq:Problem3CentersOnCustomersAssignment} \\ \sum_{i\in D} z_{ip} & \leq & M y_p \; , \quad \qquad \qquad \forall p\in C \label{eq:Problem3CentersOnCustomersTruckThenDrone} \\ \sum_{p\in C} z_{ip} & \leq & 1 \; , \quad \qquad \qquad \qquad \forall i\in D \label{eq:Problem3CentersOnCustomersEachDroneOneCenter} \\ \sum_{j \in C} x^p_{ij} & \leq & M z_{ip} \; , \quad \qquad \qquad \forall i\in D, p\in C \label{eq:Problem3CentersOnCustomersIfDroneAtCenter} \\ d^p_{ij} x^p_{ij} & \leq & L \; , \quad \qquad \qquad \qquad \forall i\in D, j,p\in C \label{eq:Problem3CentersOnCustomersDroneRange} \\ \frac{1}{v_D}\sum_{j\in C} d^p_{ij}x^p_{ij} - M(1-z_{ip}) & \leq & q_{ip} \; , \qquad \qquad \qquad \forall i\in D, p\in C \label{eq:Problem3CentersOnCustomersCenterDroneTime} \\ q_{ip} - (1/v_D) \max_{j} \delta_{ji} x^p_{ij} & \leq & Q_p \; , \qquad \qquad \qquad \forall i\in D, p\in C \label{eq:Problem3CentersOnCustomersCenterMakespan} \\ 2\gamma_{pp'} & \leq & y_{p} + y_{p'} \; , \qquad \qquad \forall p\neq p' \in C \label{eq:Problem3CentersOnCustomersTSPtruckNodesMakePath} \\ \sum_{p\in C} \gamma_{pp'} & = & y_{p'} \; , \qquad \qquad \qquad \forall p'\in C \label{eq:Problem3CentersOnCustomersTSPassignment1} \\ \sum_{p'\in C} \gamma_{pp'} & = & y_p \; , \qquad \qquad \qquad \forall p \in C \label{eq:Problem3CentersOnCustomersTSPassignment2} \\ \sum_{p\in S} \; \sum_{p'\in S,\, p'\neq p} \gamma_{pp'} & \leq & |S| -1 \; , \qquad \qquad \forall S \subset C\,, \; 2 \leq |S| \leq |C|-2 \label{eq:Problem3CentersOnCustomersTSPsubtour} \\ M \gamma_{pp} + \sum_{j\in C} y_{j} -1 & \leq & M \; , \qquad \qquad \qquad \forall p \in C \label{eq:Problem3CentersOnCustomersTSPorStar} \\ x^p_{ij}, \; \gamma_{pp'}, \; y_p & \in & \{0,1\} \;, \qquad \qquad \forall i\in D, j,p,p' \in C \nonumber \\ q_{ip}, \; z_{ip}, \; Q_p & \geq & 0 \;, \qquad \qquad \qquad \forall i\in D, p\in C \nonumber %\\ \end{eqnarray} The objective function minimizes the sum of the time that truck takes to travel between the pickup centers on its route and the total time that it takes to serve customers with crowdsourced drones at these stopping points. Constraint (<ref>) ensures that at least one customer node is set to be truck node. Constraint (<ref>) ensures all customers are visited once, either by the truck or by one of the drones. Constraint (<ref>) ensures that a node can serve as pickup center for drones if it is visited by the truck. Constraint (<ref>) enforces the assumption that each drone can only be used at one truck center. Constraint (<ref>) ensures that customers are served by a drone through a center node if that drone is used at that center. Constraint (<ref>) ensures that a drone cannot fly more than its range (battery limit). Constraints (<ref>) and (<ref>) are used to find the time spent at each stop of the truck to serve a group of customers. Note that truck does not have to wait at a pickup center after all packages in that neighborhood are delivered so the time spent on the last segment of the last route for each drone is deducted. This implies that we may prefer to assign the drone to a farther customer. Constraint (<ref>) is to make sure that the truck tour is formed through truck nodes. Constraints (<ref>)–(<ref>) are TSP constraints for the truck. Note that we treat the set of constraints in (<ref>) as lazy constraints and add them to our model in the implementation in a lazy fashion only when they are needed. Finally, constraint (<ref>) allows the model to switch from having a TSP tour to a one center star shape (hub-and-spoke) model when the range and speed of drones are large enough. To linearize the $\max_{j} \delta_{ji}x^p_{ij}$ term in (<ref>), we replace this term with $\Delta_{ip}$ and constraint (<ref>) with \begin{eqnarray} \hspace{0.6in} q_{ip} - \frac{1}{v_D} \Delta_{ip} & \leq & Q_p \; , \qquad \qquad \qquad \forall i\in D, p\in C \label{eq:Problem3CentersOnCustomersCenterMakespanLinearized} \end{eqnarray} and add the following four constraints \begin{eqnarray} \hspace{1.7in} \delta_{ji}x^p_{ij} & \leq & \Delta_{ip} \;, \quad \qquad \qquad \qquad \forall i\in D, p\in C, j\in C \label{eq:Problem3CentersOnCustomersCenterMakespanLinearized1} \\ \Delta_{ip} & \leq & \delta_{ji}x^p_{ij} + M(1-\eta^p_{ij}) \;, \quad \forall i\in D, p\in C, j\in C \label{eq:Problem3CentersOnCustomersCenterMakespanLinearized2} \\ \sum_{j\in C} \eta^p_{ij} & \geq & z_{ip} \;, \qquad \qquad \qquad \qquad \forall i\in D, p\in C \label{eq:Problem3CentersOnCustomersCenterMakespanLinearized3} \\ \eta^p_{ij} & \leq & x^p_{ij} \qquad \qquad \qquad \qquad \forall i\in D, p\in C, j\in C \label{eq:Problem3CentersOnCustomersCenterMakespanLinearized4} \\ \eta^p_{ij} & \in & \{0,1\} \quad \qquad \qquad \qquad \forall i\in D, p\in C, j\in C \nonumber \end{eqnarray} to form an MILP model. We solved this model for small instances but for larger problems we rely on a heuristic algorithm. §.§ Problem IV–(Multiple Centers with Revisiting) In this problem we consider multiple pickup centers and assume that after each delivery, drones could revisit the pickup center for another delivery before going back to their home bases. Similar to the latter part of the previous section, we assume the pickup centers (truck nodes) are a subset of customer nodes. Also, similar to Problem II, we define a feasible drone route as a route that starts from a drone node $i\in D$, satisfies the demand of one or multiple customer nodes after picking their packages up at a center node, and comes back to the starting drone node while not violating the drone range $L$. We assume that the revisits happen to the same center and the drone could start serving through another center after going back to its base. Relaxing this assumption is doable but comes at huge computational costs. We consider no restriction on the number of revisits for each drone. We pre-calculate all such routes. The sets, parameters, and variables used in our model are defined in Table <ref>. Sets, parameters, and variables used in the general (multi-center) model. Sets/Parameters/Variables Description $\Pi^p_{i}$ Set of all feasible routes that start from drone node $i\in D$ and go through pickup center $p \in C$ $\Pi^p_{i}(j)$ Set of all feasible routes that start from drone node $i\in D$, go through pickup center $p \in C$, and traveled through customer node $j \in C$ $d_{pp'}$ Distance between node $p\in C$ and $p'\in C$ $d_{\pi^p_{ik}}$ Length of route $\pi^p_{ik} \in \Pi^p_i$ for drone node $i \in D$ $\delta_{\pi^p_{ik}}$ Length of the last line segment of route $\pi^p_{ik}$ $y_p$ Binary decision variable that is 1 if node $p\in C$ is served by the truck and 0 otherwise $x_{\pi^p_{ik}}$ Binary decision variable that is $1$ when a drone travels along the route $\pi^p_{ik}\in \Pi^p_{i}$ and is $0$ otherwise $z_{ip}$ Binary decision variable that is $1$ when drone $i\in D$ is used for delivery at center node $p\in C$ and is $0$ otherwise $q_{ip}$ Total travel time of the drone with base at node $i\in D$ that is assigned to center $p\in C$ $Q_p$ Maximum time spent by all drones assigned to center $p\in C$ $\Delta_{ip}$ Maximum length of the last tour segment among all routes $\pi^p_{ik}\in \Pi^p_{i}$ $\eta_{\pi^p_{ik}}$ Binary decision variable equal to 1 if $\delta_{\pi^p_{ik}}x_{\pi^p_{ik}} = \Delta_{ip}$ and 0 otherwise $\gamma_{pp'}$ Binary decision variable equal to 1 if the path from $p$ to $p'$ has been used by the truck An MILP model for this problem can be written as following: \begin{eqnarray} \minimize \;\;\; \frac{1}{v_T}\sum_{p\in C}\sum_{p'\in C} d_{pp'}\gamma_{pp'} & + & \sum_{p} Q_p \qquad \qquad \st \nonumber \\ \sum_{p\in C} y_p & \geq & 1 \; , \label{eq:Problem4CentersOnCustomersMinOneTruckNode} \\ \sum_{p\in C}\sum_{i\in D} \sum_{\pi^p_{ik}\in \Pi^p_i(j)}x_{\pi^p_{ik}} & = & 1-y_j \; , \quad \qquad \forall j\in C \label{eq:Problem4CentersOnCustomersAssignment} \\ \sum_{i\in D} z_{ip} & \leq & M y_p \; , \qquad \qquad \forall p\in C \label{eq:Problem4CentersOnCustomersTruckThenDrone} \\ \sum_{p\in C} z_{ip} & \leq & 1 \; , \quad \qquad \qquad \qquad \forall i\in D \label{eq:Problem4CentersOnCustomersEachDroneOneCenter} \\ \sum_{\pi^p_{ik}\in \Pi^p_i} x_{\pi^p_{ik}} & \leq & M z_{ip} \; , \qquad \qquad \forall i\in D, p\in C \label{eq:Problem4CentersOnCustomersIfDroneAtCenter} \\ \frac{1}{v_D} \sum_{\pi^p_{ik}\in \Pi^p_i} d_{\pi^p_{ik}} x_{\pi^p_{ik}} - M(1-z_{ip}) & \leq & q_{ip} \; , \quad \qquad \qquad \forall i\in D, \; p\in C \label{eq:Problem4CentersOnCustomersCenterDroneTime} \\ q_{ip} - (1/v_D) \max_{\pi^p_{ik}\in \Pi^p_i} \delta_{\pi^p_{ik}}x_{\pi^p_{ik}} & \leq & Q_p \; , \quad \qquad \qquad \forall i\in D, \; p\in C \label{eq:Problem4CentersOnCustomersCenterMakespan} \\ 2\gamma_{pp'} & \leq & y_{p} + y_{p'} \; , \qquad \qquad \forall p\neq p'\in C \label{eq:Problem4CentersOnCustomersTSPtruckNodesMakePath} \\ \sum_{p\in C} \gamma_{pp'} & = & y_{p'} \; , \qquad \qquad \qquad \forall p'\in C \label{eq:Problem4CentersOnCustomersTSPassignment1} \\ \sum_{p'\in C} \gamma_{pp'} & = & y_p \; , \qquad \qquad \qquad \forall p \in C \label{eq:Problem4CentersOnCustomersTSPassignment2} \\ \sum_{p\in S} \; \sum_{p'\in S,\, p'\neq p} \gamma_{pp'} & \leq & |S| -1 \; , \quad \qquad \forall S \subset C\,, \; 2 \leq |S| \leq |C|-2 \label{eq:Problem4CentersOnCustomersTSPsubtour} \\ M \gamma_{pp} + \sum_{j\in C} y_{j} -1 & \leq & M \; , \qquad \qquad \qquad \forall p \in C \label{eq:Problem4CentersOnCustomersTSPorStar} \\ x_{\pi^p_{ik}}, \; \gamma_{pp'},\; y_p & \in & \{0,1\} \; , \; \qquad \qquad \forall i\in D,\; p,p'\in C,\; \pi^p_{ik} \in \Pi^p_i \nonumber \\ q_{ip}, \; Q_p & \geq & 0 \;, \qquad \qquad \qquad \forall i\in D,\; p\in C \nonumber %\\ \end{eqnarray} The objective function minimizes the sum of the time that spent on the truck route and the drone routes. Constraint (<ref>) ensures that at least one customer node is set to be truck node (pickup center). Constraint (<ref>) ensures either the truck or one of the drones visits each of the customers. Constraint (<ref>) ensures that a customer nodes can only be used as pickup center for drones if that node is visited by the truck. Constraint (<ref>) enforces the assumption that each drone can only be used at one truck center. Constraint (<ref>) enforces drone routes to be formed at pickup centers. Constraints (<ref>) and (<ref>) are used to find the time spent at each stop of the truck to serve a group of customers. Note that at each center the truck may not wait until the last drone route is completed, which may, at times, prioritize a farther customer over a closer customer for a drone. Constraint (<ref>) is to make sure that the truck tour is formed through truck nodes. Constraints (<ref>)–(<ref>) are TSP constraints for the truck. Note that we treat the set of constraints in (<ref>) as lazy constraints and add them to our model in the implementation in a lazy fashion only when they are needed. Finally, constraint (<ref>) allows the model to switch from having a TSP tour to a one center star shape (hub-and-spoke) model when the range and speed of drones are large enough. To linearize the $\max_{\pi^p_{ik}\in \Pi^p_i} \delta_{\pi^p_{ik}}x_{\pi^p_{ik}}$ term in (<ref>), we replace this term with $\Delta_{ip}$ and constraint (<ref>) with \begin{eqnarray} \hspace{0.1in} q_{ip} - \frac{1}{v_D} \Delta_{ip} & \leq & Q_p \; , \qquad \qquad \forall i\in D, p\in C \label{eq:Problem4CentersOnCustomersCenterMakespanLinearized} \end{eqnarray} and add the following four constraints \begin{eqnarray} \hspace{1.7in} \delta_{\pi^p_{ik}}x_{\pi^p_{ik}} & \leq & \Delta_{ip} \;, \qquad \qquad \qquad \forall i\in D, p\in C, \, \pi^p_{ik}\in \Pi^p_i \label{eq:Problem4CentersOnCustomersCenterMakespanLinearized1} \\ \Delta_{ip} & \leq & \delta_{\pi^p_{ik}}x_{\pi^p_{ik}} + M(1-\eta_{\pi^p_{ik}}) \;, \;\; \forall i\in D, p\in C, \pi^p_{ik}\in \Pi^p_i \label{eq:Problem4CentersOnCustomersCenterMakespanLinearized2} \\ \sum_{\pi^p_{ik}\in \Pi^p_i} \eta_{\pi^p_{ik}} & \geq & z_{ip} \;, \label{eq:Problem4CentersOnCustomersCenterMakespanLinearized3} \\ \eta_{\pi^p_{ik}} & \leq & x_{\pi^p_{ik}} \;, \;\; \qquad \qquad \forall i\in D, p\in C, \pi^p_{ik}\in \Pi^p_i \label{eq:Problem4CentersOnCustomersCenterMakespanLinearized4} \\ \eta_{\pi^p_{ik}} & \in & \{0,1\} \;, \qquad \qquad \forall i\in D, p\in C, \, \pi^p_{ik}\in \Pi^p_i \nonumber \end{eqnarray} to form an MILP model. We also solved this model for small instances but for larger problems we rely on our heuristic algorithm. § ALGORITHM It is clear that the problem is NP-hard and we have to take a sub-optimal approach to solve the problem. Our algorithm consists of several sub-routines as explained in the following. The high-level steps of our algorithm are as follows: Step 1: Running a binary search and the $k$-means clustering algorithm to find the minimum feasible $k$ to cluster all customers into $k$ clusters with radius $L/4$ and finding the cluster centers. Step 2: Partitioning the region with the Voronoi tessellation generated by the $k$ cluster center points found in Step 1 and assigning customers and drone bases to their nearest centers. Step 3: Running a Tabu Search algorithm to solve the sub-problem in each cluster where parcels are assigned to drones to be delivered to customers in a way that minimizes total delivery time. Step 4: Solving the truck tour by the Lin-Kernighan-Helsgaun (LKH) algorithm to go through all cluster centers as well as all customer nodes that do not have any drone node in that cluster. We first cluster all customers into $k$ groups, where each group will have a center that will serve as the truck's stoping point, by solving a $k$-means clustering problem to ensure all customers are within radius $L/4$ of their closest centers. A similar idea for finding truck's stopping points for launching drones in the centralized system of combined truck and drone has been previously shown to be efficient <cit.>. To solve the $k$-means clustering problem we use Lloyd's algorithm <cit.>, which is described in <ref> of the Online Supplement of this paper. This minimizes the total squared Euclidean distances between customer nodes and their closest cluster centers. The choice of radius $L/4$ is to make sure that with one full battery the drones in each cluster can finish the delivery and return to their base. The number $k$ is the smallest integer that makes this feasible and will be found using a binary search. The feasibility here means that all customers are located in at least one circle that is centered at a cluster center with radius $L/4$. The clustering step is shown in <ref>. The final geographic partitioning, i.e., identifying the boundaries of each cluster, is done using a Voronoi partitioning scheme <cit.> with the cluster centers as the generators. This determines the assignment of customer nodes and drone nodes to their closest cluster centers. These nodes would be assigned to the nearest center only if their distance to the center is less than $L/4$. Clusters with no drone will be served by the truck. The partitioning step is illustrated in <ref>. In the next step, in each cluster (Voronoi cell), a Tabu Search algorithm <cit.> will be used to solve the one center problem. The steps of a generic Tabu Search algorithm are explained in in <ref> of the Online Supplement of this paper. In each cluster, several drones will fly from their origin (base), go to the truck center, deliver the packages to the customer and go back to their base. We represent a solution of the Tabu Search with a list of all nodes. As an example consider the following solution 8 1 5 7 1 2 9 1 3 10 1 4 9 1 6 for a cluster (one center problem) with 5 customer nodes and 4 drone nodes. In this solution, $1$ represents the truck node, $2$ to $6$ represents the customer nodes and $7$ to $10$ represents the drone nodes. Therefore, a solution representation like this indicates that a drone from location 8 goes to the cluster center (truck node) at location 1 and then goes to customer node 5, while drones from locations 7 and 10 also go to the cluster center and deliver the packages to customer nodes 2 and 4, respectively. Drone from location 9 first serves the customer node 3 and comes back to its home base to change the battery, then it goes to serve the customer node 6. In this solution, four drones will deliver packages for five customers, and the solution for one cluster is shown in <ref>, where the black square is the cluster center, blue stars are the customer-owned drones' location and small black circles are the customers' location. It is clear that each such list represents a unique solution of the problem but each solution can be represented with multiple such lists. The Tabu Search algorithm is applied to this solution to find a new solution. During each iteration, an action will be applied to the current solution. An action is either to randomly switch the position of customer nodes (switching 3 and 5 in the following example) 8 1 3 7 1 2 9 1 5 10 1 4 9 1 6 or randomly change a drone node among all the drone nodes in that cluster (changing 7 to 9 in the following example). 8 1 5 9 1 2 9 1 3 10 1 4 9 1 6 This gives us an action list that has a size of $|C|(|C|-1)/2+|D|$. In the final step of the algorithm, we solve a travelling salesman problem (TSP) to find the truck tour among all cluster centers and all customer nodes that do not have any drone node nearby. All customers in any cluster that has no drones in it, will be served by the truck. Lin-Kernighan-Helsgaun (LKH) algorithm <cit.> is used for this purpose. The LKH algorithm is a local search heuristic to solve TSP with swapping operations inside the tours to make a new tour. It is an iterative algorithm with two set of links, with each link connecting two nodes, and in each step it will decide pairs of links that need to be switched by deleting links in the first set from the tour and adding the links from the second set to the tour (a generalization of the well-known 2-opt and 3-opt exchanges). <ref> shows the TSP tour in our example. It can be seen that there are some cluster centers that are not on the truck route. This is because in those clusters all of the customers are served by the truck and thus there is no reason for the truck to go to the cluster center and wait there for the drones anymore. $k$-means clustering Voronoi Partition One cluster result Truck route Illustration of the main steps of the Algorithm for a problem with 40 customers and $k=9$ clusters. The $k$-means clustering step is shown in (fig:ClusteringStep) and the Voronoi partitioning step is illustrated in (fig:VoronoiStep). (fig:OneClusterTS) shows the routes for the drones in the upper left cluster and (fig:TruckRouteTSP) shows the truck route going through cluster centers and customer nodes that are not served by the drones. § IMPROVEMENTS OF THE ALGORITHM Next, we will show that some further improvements can be applied to the proposed model with sharing drones. In this paper, we analyze four such improvements. In the first three, we try to find the best location and the best number of truck centers by moving these centers to one of the customer nodes and merging these centers, if we could still cover the all customers using the new centers. In the last improvement, we try to utilize the battery of drones, that is if the battery remaining in the drones can deliver another package, instead of going back to the base, we let them to go back to the cluster center to pick up the next package immediately after delivering the previous one. §.§ Moving the Cluster Centers to Their Nearest Feasible Customer Nodes Both of our MILP models developed for Problems III and IV assume pickup centers (truck nodes) as a subset of customer nodes. We will also show in <ref> that in our algorithm moving the cluster centers to some well-selected customer nodes, provided that the new pickup center still could serve all of its assigned customers via the available assigned drones to that center, improves the performance of the algorithm. Therefore, to make the comparison between the optimization models and the algorithm valid, we incorporate this modification into our algorithm. In the first modification we move the cluster center to the nearest customer node to that center provided that the center could still cover all of its assigned customers using its assigned drones as illustrated in <ref>. While we can save one drone trip to the customer that is now served by the truck, the length of the truck route or the routes of the other drones could either decrease or increase. We will analyze this trade-off in Section <ref>. <ref> summarizes the steps of the algorithm with this modification. One could take an alternative approach by incorporating the $k$-medoids algorithm that automatically selects one of the central customers in each cluster as the cluster center <cit.>. Moving cluster centers §.§ Moving the Cluster Centers to Nearest Feasible Customer Node to Area Center The modification in Algorithm 1 may provide a slight improvement. However, we can do better by moving the cluster centers strategically having the goal of making the truck route shorter. The intuition behind this is due to the general understanding that the drones are faster than trucks especially in urban delivery operations. Also, in the centralized alternative (TSPD) it is known that for having any significant gain by combining a truck and a drone in delivery operations we must have the speed of the drone at least twice as much as the speed of the truck <cit.> and it is most often assumed to have drones faster than the truck. Thus, since drone has speed advantage, we expect to gain in delivery time by decreasing the length of truck route and increasing the length of drone routes. We will analyze this trade-off in Section <ref>. Therefore, we modify the algorithm as follows. We first find the centroid of the cluster centers (truck nodes). Let's call it $\bm{g}$. Then, for each cluster center we find one of its feasible assigned customer nodes closest to $\bm{g}$ among all other assigned customers to this center and closer to $\bm{g}$ than the cluster center itself. We choose this customer node, if existing, as the new cluster center (truck node). The steps of the algorithm with this modification are summarized in <ref>. §.§ Merging Centers We can still do a significant improvement to Algorithm <ref> by trying to merge any two centers if all customers assigned to one could be served via the other by using drones assigned to either of these two centers. This comes from our assumption that the drone is faster than the truck and making the truck route shorter by reducing the number of centers could benefit us in the total delivery time. We first sort centers found in Algorithm <ref> with respect to their distance from $\bm{g}$ in a descending order. We then start from the the first center in the list, which is the farthest from $\bm{g}$, and check the possibility of its merge into its closest center. If it is feasible, we assign this center, which itself is a customer node, and all customer nodes assigned to this center, to its closest center and we serve them using combined set of drones assigned to these two centers. The feasibility here is defined as whether drones in this combined set could serve all customers assigned to these two centers via the latter center while staying within their flying range $L$. This also limits the impact of the very restricting constraint of considering a range of $L/4$ in the initial clustering step. Once we have checked the possibility of a merge for all $k$ centers and if at least one merge is happened, we repeat this step with the remaining number of centers. This process continues until no more merge is possible. The pseudocode of the algorithm with this modification is presented in <ref>. $\noun{RechargingRouteFinderMC1}\:(R,C,D,L)$ Generates the truck route and all the drone routes after finding the best location for the truck centers in the continuous space and then moving those centers to the nearest customer's place. Input region $R$, set of customer nodes $C$, set of drone nodes $D$, and drone range $L$. Truck route $T_0$ and routes of drones $T_j\,, \; j\in D$ used for serving the customers. Let $T_0$ denote the truck route and $T_j$ denote the route for drone $j\in D$ Find a feasible $k$ and cluster all customer nodes in $C$ into $k$ clusters within distance $L/4$ using a binary search and $k$-means clustering algorithms Let $\bm{m}_1,...,\bm{m}_k$ be the cluster centers and generate their Voronoi tessellation Let $V_1,...,V_k$ be the $k$ Voronoi cells and use them to assign all customer and drone nodes to their corresponding centers if their distance to the center is less than $L/4$ Let $\bm{z}_{\ell} = \argmin_{\bm{z}_j \in V_{\ell}} \|\bm{z}_j - \bm{m}_{\ell}\|\,, \; \forall j\in C$ Let $\bm{m}_{\ell}^{\prime} = \bm{z}_{\ell}$ Let $C_{\ell}=\{\bm{z}_j \in V_{\ell} \,, \; j\in C\}$ $\bm{z}_j \in C_{\ell} \; \&\& \; \bm{z}_j \neq \bm{m}_{\ell}^{\prime}$ $\bm{z}_j$ can be served via $\bm{m}_{\ell}^{\prime}$ using drones in $V_{\ell}$ $|C_{\ell}\backslash \{\bm{m}_{\ell}^{\prime}\}| = 1$ && $2*\|\bm{z}_j - \bm{m}_{\ell}^{\prime}\| / v_{T} \leq (\|\bm{y}_i - \bm{m}_{\ell}^{\prime}\| + \|\bm{m}_{\ell}^{\prime} - \bm{z}_j\| + \|\bm{z}_j - \bm{y}_i\|)/v_{D} \,, \; \bm{y}_i \in V_{\ell}\,, \; i\in D$ Let $\bm{z}_j$ be a truck node, i.e., to be served by the truck Assign $\bm{z}_j$ to $\bm{m}_{\ell}^{\prime}$, i.e., to be served by a drone Let $\bm{z}_j$ be a truck node, i.e., to be served by the truck Update $k$ to be the updated number of truck nodes and reindex them as $\bm{m}_1^{\prime},...,\bm{m}_k^{\prime}$ Run the Tabu Search algorithm to find the drone routes in all Voronoi cells, i.e., $T_j\,, \; j \in D$ Set $T_0 = \{\bm{m}_1^{\prime}, \bm{m}_2^{\prime}\}$ $k\geq 3$ Run the Lin-Kernighan-Helsgaun algorithm to find the truck route $T_0$ through $\bm{m}_1^{\prime},...,\bm{m}_k^{\prime}$ $T_0$ and $T_j\,, \; j\in D$ $\noun{RechargingRouteFinderMC2}\:(R,C,D,L)$ Generates the truck route and all the drone routes. It first finds the best location for the truck centers and then moves those centers to the nearest customer's place. Finally, it finds the centroid of these centers and moves these centers to one of their assigned customers that is closest to this centroid and keeps the delivery feasible. Input region $R$, set of customer nodes $C$, set of drone nodes $D$, and drone range $L$. Truck route $T_0$ and routes of drones $T_j\,, \; j\in D$ used for serving the customers. Let $T_0$ denote the truck route and $T_j$ denote the route for drone $j\in D$ Find a feasible $k$ and cluster all customer nodes in $C$ into $k$ clusters within distance $L/4$ using a binary search and $k$-means clustering algorithms Let $\bm{m}_1,...,\bm{m}_k$ be the cluster centers and generate their Voronoi tessellation Let $V_1,...,V_k$ be the $k$ Voronoi cells and use them to assign all customer and drone nodes to their corresponding centers if their distance to the center is less than $L/4$ Let $\bm{z}_{\ell} = \argmin_{\bm{z}_j \in V_{\ell}} \|\bm{z}_j - \bm{m}_{\ell}\|\,, \; \forall j\in C$ Let $\bm{m}_{\ell}^{\prime} = \bm{z}_{\ell}$ Let $C_{\ell}=\{\bm{z}_j \in V_{\ell} \,, \; j\in C\}$ $\bm{z}_j \in C_{\ell} \; \&\& \; \bm{z}_j \neq \bm{m}_{\ell}^{\prime}$ $\bm{z}_j$ can be served via $\bm{m}_{\ell}^{\prime}$ using drones in $V_{\ell}$ $|C_{\ell}\backslash \{\bm{m}_{\ell}^{\prime}\}| = 1$ && $2*\|\bm{z}_j - \bm{m}_{\ell}^{\prime}\| / v_{T} \leq (\|\bm{y}_i - \bm{m}_{\ell}^{\prime}\| + \|\bm{m}_{\ell}^{\prime} - \bm{z}_j\| + \|\bm{z}_j - \bm{y}_i\|)/v_{D} \,, \; \bm{y}_i \in V_{\ell}\,, \; i\in D$ Let $\bm{z}_j$ be a truck node, i.e., to be served by the truck Assign $\bm{z}_j$ to $\bm{m}_{\ell}^{\prime}$, i.e., to be served by a drone Let $\bm{z}_j$ be a truck node, i.e., to be served by the truck Update $k$ to be the updated number of truck nodes and reindex them as $\bm{m}_1^{\prime},...,\bm{m}_k^{\prime}$ Let $\bm{g}$ be the centroid of $\bm{m}_1^{\prime},...,\bm{m}_k^{\prime}$ Set $d_j = \|\bm{z}_j - \bm{g}\|\,, \; \forall j\in C$ and $d_{\bm{m}_{\ell}^{\prime}} = \|\bm{m}_{\ell}^{\prime} - \bm{g}\|$ Let $A_{\ell}$ be an array of customer nodes in $V_{\ell}$ sorted in ascending order of $d_j\,, \; \forall j\in C$ Find the customer node $\bm{z}_{\ell}$ with the smallest index in $A_{\ell}$ and $d_{\bm{z}_{\ell}} \leq d_{\bm{m}_{\ell}^{\prime}}$ that has at least one drone within distance $L/4$ and if chosen as cluster center we can still serve all other customers in $V_{\ell}$ using drones in $V_{\ell}$ and let $\bm{m}_{\ell}^{\prime\prime} = \bm{z}_{\ell}$. If no such point exists let $\bm{m}_{\ell}^{\prime\prime} = \bm{m}_{\ell}^{\prime}$ Run the Tabu Search algorithm to find the drone routes in all Voronoi cells, i.e., $T_j\,, \; j \in D$ Set $T_0 = \{\bm{m}_1^{\prime\prime}, \bm{m}_2^{\prime\prime}\}$ $k\geq 3$ Run the Lin-Kernighan-Helsgaun algorithm to find the truck route $T_0$ through $\bm{m}_1^{\prime\prime},...,\bm{m}_k^{\prime\prime}$ $T_0$ and $T_j\,, \; j\in D$ $\noun{RechargingRouteFinderMC3}\:(R,C,D,L)$ – Generates the truck route and all the drone routes. It first finds the best location for the truck centers and then moves those centers to the nearest customer's place. It then finds the centroid of these centers and moves these centers to one of their assigned customers that is closest to this centroid and keeps the delivery feasible. Finally, it tries to merge these centers to shorten the truck tour. Input region $R$, set of customer nodes $C$, set of drone nodes $D$, and drone range $L$. Truck route $T_0$ and routes of drones $T_j\,, \; j\in D$ used for serving the customers. Let $T_0=\emptyset$ denote the truck route and $T_j=\emptyset$ denote the route for drone $j\in D$ Find a feasible $k$ and cluster all customer nodes in $C$ into $k$ clusters within distance $L/4$ using a binary search and $k$-means clustering Let $\bm{m}_1,...,\bm{m}_k$ be the cluster centers and generate their Voronoi tessellation Let $V_1,...,V_k$ be the $k$ Voronoi cells and use them to assign all customer and drone nodes to their corresponding centers within distance $L/4$ Let $\bm{z}_{\ell} = \argmin_{\bm{z}_j \in V_{\ell}} \|\bm{z}_j - \bm{m}_{\ell}\|\,, \; \forall j\in C$ Let $\bm{m}_{\ell}^{\prime} = \bm{z}_{\ell}$ Let $C_{\ell}=\{\bm{z}_j \in V_{\ell} \,, \; j\in C\}$ $\bm{z}_j \in C_{\ell} \; \&\& \; \bm{z}_j \neq \bm{m}_{\ell}^{\prime}$ $\bm{z}_j$ can be served via $\bm{m}_{\ell}^{\prime}$ using drones in $V_{\ell}$ $|C_{\ell}\backslash \{\bm{m}_{\ell}^{\prime}\}| = 1$ && $2*\|\bm{z}_j - \bm{m}_{\ell}^{\prime}\| / v_{T} \leq (\|\bm{y}_i - \bm{m}_{\ell}^{\prime}\| + \|\bm{m}_{\ell}^{\prime} - \bm{z}_j\| + \|\bm{z}_j - \bm{y}_i\|)/v_{D} \,, \; \bm{y}_i \in V_{\ell}\,, \; i\in D$ Let $\bm{z}_j$ be a truck node, i.e., to be served by the truck Assign $\bm{z}_j$ to $\bm{m}_{\ell}^{\prime}$, i.e., to be served by a drone Let $\bm{z}_j$ be a truck node, i.e., to be served by the truck Update $k$ to be the updated number of truck nodes and reindex them as $\bm{m}_1^{\prime},...,\bm{m}_k^{\prime}$ Let $\bm{g}$ be the centroid of $\bm{m}_1^{\prime},...,\bm{m}_k^{\prime}$ Set $d_j = \|\bm{z}_j - \bm{g}\|\,, \; \forall j\in C$ and $d_{\bm{m}_{\ell}^{\prime}} = \|\bm{m}_{\ell}^{\prime} - \bm{g}\|$ Let $A_{\ell}$ be an array of customer nodes in $V_{\ell}$ sorted in ascending order of $d_j\,, \; \forall j\in C$ Find the customer node $\bm{z}_{\ell}$ with the smallest index in $A_{\ell}$ and $d_{\bm{z}_{\ell}} \leq d_{\bm{m}_{\ell}^{\prime}}$ that has at least one drone within distance $L/4$ and if chosen as cluster center we can still serve all other customers in $V_{\ell}$ using drones in $V_{\ell}$ and let $\bm{m}_{\ell}^{\prime\prime} = \bm{z}_{\ell}$. If no such point exists let $\bm{m}_{\ell}^{\prime\prime} = \bm{m}_{\ell}^{\prime}$ Let $\Omega = [\bm{m}_1^{\prime\prime},...,\bm{m}_{k}^{\prime\prime}]$, $\Omega^{\prime} = \emptyset$, and $count=k$ $count > 0$ Set $count=0$ $k = 1$ Sort centers $\bm{m}_{\ell}^{\prime\prime} \in \Omega, \ell=1,...,k$ in a descending order of their distances to $\bm{g}$ and reindex them accordingly Reallocate customer and drone nodes to their nearest centers. For customer nodes pick the closest center that could serve it Let $\bm{\mu} = \argmin_{\bm{m}_{\ell}^{\prime\prime} \in \Omega \backslash \{\Omega^{\prime}\cup \{\bm{m}_{r}^{\prime\prime}\}\}} \|\bm{m}_{\ell}^{\prime\prime} - \bm{m}_{r}^{\prime\prime}\|\,, \; \ell=1,...,k$ and let $s$ be its index in $\Omega$, i.e., $\bm{\mu} = \bm{m}_{s}^{\prime\prime}$. Break a tie with closeness to $\bm{g}$ Let $feasibility = true$ $\bm{z}_j \in V_r \cup V_s\,, \; j\in C$ Find $\bm{y}_i \in V_{r} \cup V_{s}\,, \; i\in D$ such that $\|\bm{y}_i - \bm{\mu}\| + \|\bm{\mu} - \bm{z}_j\| + \|\bm{z}_j - \bm{y}_i\| \leq L$. If no such drone exists, let $feasibility = false$ $feasibility = false$ $feasibility = true$ $k = 2$ Merge the center with fewer customers (say $\bm{m}_{2}^{\prime\prime}$) into the center with more customers (say $\bm{m}_{1}^{\prime\prime}$) and reallocate its assigned customers to that center Set $\Omega^{\prime} = \Omega^{\prime} \cup \{\bm{m}_{2}^{\prime\prime}\}$ Merge the center $\bm{m}_{r}^{\prime\prime}$ and $V_r$ into $\bm{m}_{s}^{\prime\prime}$ and $V_s$ and allocate its assigned (customer and drone) nodes to $\bm{m}_{s}^{\prime\prime}$ Set $\Omega^{\prime} = \Omega^{\prime} \cup \{\bm{m}_{r}^{\prime\prime}\}$ Set $count = count +1$ Set $\Omega = \Omega \backslash \Omega^{\prime}$, $k = length(\Omega)$, and $\Omega^{\prime} = \emptyset$ Run the Tabu Search algorithm to find the drone routes in all Voronoi cells, i.e., $T_j\,, \; j \in D$ Set $T_0 = \{\bm{m}_1^{\prime\prime}, \bm{m}_2^{\prime\prime}\}$ $k\geq 3$ Run the Lin-Kernighan-Helsgaun algorithm to find the truck route $T_0$ through $\bm{m}_1^{\prime\prime},...,\bm{m}_k^{\prime\prime}$ $T_0$ and $T_j\,, \; j\in D$ §.§ Utilization of Battery In Problem IV, we assumed that drones can revisit the pickup center (truck's stopping place) as long as they do not violate their range (battery duration) limit. Therefore, we need to modify the algorithm to utilize the battery of the drones and allow drones to revisit their assigned pickup center before returning to their base. We assume no limit on the number of times a drone could go back to a pickup center to deliver another package as long as it remains within the range. By implementing this modification, obviously, we may expect some savings since if one drone can deliver another package without going back to the base, it can save time on that trip. However, this adds to the waiting time of the truck and makes the truck route costlier. This trade-off is analyzed in Section <ref>. We apply this modification to all three Algorithms <ref>, <ref>, and <ref>. We skip presenting the pseudocode here for brevity (the only difference is allowing revisiting in range conditions and in the TS step). The modification in the tabu search step is applied by changing the solution representation. Our original example of a solution in a cluster for a problem with 5 customer nodes and 4 drone nodes can change as following 8 1 5 7 1 2 9 1 3 1 6 10 1 4 In this solution, drone from location 9 first serves the customer node 3 and then revisits the truck node $1$ (without going to its base first) to serve the customer node 6. § COMPUTATIONAL RESULTS As problems III and IV are the generalized versions of Problems I and II, we only do our computational results for Problems III and IV. We will see that for certain values of input parameters we actually get solutions for the special cases of Problems I and II. We tested the MILP model of Problems III and IV on a set of two synthetic problem instances with 16 and 12 ($P1$) and 16 and 16 ($P2$) customer and drone nodes, respectively, distributed uniformly at random in a unit square. The speed of drones is set to be 2 times the speed of the truck in both models. For the drone flying range $L$ we took the maximum length of a drone in a unit box, i.e., $2+\sqrt{2} \simeq 3.4$, and with increments of 0.2 we generated 17 instance for each problem. In the tabu search step of the algorithms we have set |Tabu List| = 0.7 * |Action List| with a termination condition of 1000 iterations. We have also set the precision to be $\epsilon=10^{-5}$ for all solutions. We implemented our models in Julia 1.9.2 using Gurobi Optimizer 10.0.2 on an Intel Core i7-5600U @2.6GHz, 8GB DDR3 RAM computer. We ran our algorithms for the same problem instances on the same computer using MATLAB R2022b. We also ran our algorithms on larger problem instances for comparison with the results of traditional centralized delivery system (pure TSP model) and the coordinated truck and drone delivery (TSP-D model). §.§ Results for Multiple Centers with Recharging Table <ref> compares the computational results of the our optimization model for Problem III and the proposed algorithms on the 34 instances of $P1$ and $P2$. As it is evident from the table, our algorithms are much faster than the optimization model. The average running time of the final version of our algorithm (Algorithm <ref>) is 8.94 seconds versus 885.13 seconds of the optimization model for $P1$ and 9.56 seconds versus 172.23 seconds for $P2$. The average (maximum) optimality gap of the solutions provided by Algorithm <ref> for $P1$ and $P2$ is 11.92% (57.66%) and 16.61% (102.71%), respectively. Figure <ref> illustrates optimal solutions obtained by our optimization model and optimal/suboptimal solutions obtained by Algorithm <ref> for $P1$ and $P2$ for select instances of $L$. Comparison between the results of the optimization model for Problem III (Multiple Centers with Recharging) and the algorithms on two synthetic examples with 16 and 12 ($P1$) and 16 and 16 ($P2$) customers nodes and drone nodes, respectively, distributed randomly in a unit box and for different values of $L$. The column “Time (s)” shows the computational time in seconds. The column “Gap” presents the optimality gap of the solutions found by the algorithms. The instances for which the optimal solution is found by our algorithms are shown in bold (0.00% gap). 11cP1 (16-12) – Recharging 2-12 2[4]*L 2cOpt. Model 3cAlgorithm 1 3cAlgorithm 2 3cAlgorithm 3 2-12 Time (s) Obj. Value Time (s) Obj. Value Gap (%) Time (s) Obj. Value Gap (%) Time (s) Obj. Value Gap (%) 1c|0.2 2.80 3.3325 0.36 3.3325 0.00% 0.36 3.3325 0.00% 0.42 3.3325 0.00% 1c|0.4 21.93 3.2561 0.63 3.2837 0.85% 0.61 3.2837 0.85% 0.62 3.3385 2.53% 1c|0.6 384.27 3.0865 0.70 3.2659 5.81% 0.67 3.2868 6.49% 1.12 3.7672 22.05% 1c|0.8 7772.69 2.4655 0.89 3.2056 30.02% 0.81 3.2881 33.36% 1.65 2.7818 12.83% 1c|1 483.70 1.5588 0.96 3.6451 133.85% 0.92 3.6421 133.66% 1.62 2.4575 57.66% 1c|1.2 5445.16 1.4339 1.14 2.8359 97.78% 1.14 2.8080 95.84% 3.56 1.7979 25.39% 1c|1.4 11.45 0.5176 1.19 3.1534 509.28% 1.35 2.9181 463.82% 13.04 0.6887 33.07% 1c|1.6 97.43 0.4828 1.05 3.1534 553.13% 1.17 2.9218 505.17% 12.88 0.5029 4.16% 1c|1.8 43.86 0.4686 3.12 1.7029 263.40% 3.18 1.7107 265.07% 13.28 0.5849 24.82% 1c|2 100.87 0.4686 3.03 1.7029 263.40% 3.06 1.4503 209.51% 12.84 0.4772 1.84% 1c|2.2 92.24 0.4686 2.96 1.7029 263.40% 2.95 1.4503 209.51% 12.83 0.4869 3.90% 1c|2.4 97.30 0.4686 12.63 0.4737 1.09% 12.82 0.4869 3.90% 13.11 0.4832 3.12% 1c|2.6 100.90 0.4686 12.64 0.4694 0.18% 12.62 0.4755 1.47% 14.06 0.4734 1.03% 1c|2.8 96.93 0.4686 12.66 0.4836 3.21% 12.85 0.4832 3.12% 12.69 0.4775 1.91% 1c|3 98.45 0.4686 12.91 0.4832 3.12% 12.96 0.5029 7.32% 12.86 0.4828 3.03% 1c|3.2 98.74 0.4686 12.78 0.4836 3.21% 12.71 0.4734 1.03% 12.86 0.4846 3.41% 1c|3.4 98.49 0.4686 12.78 0.4755 1.47% 13.15 0.4832 3.12% 12.59 0.4775 1.91% mygray Average 885.13 - 5.44 - 125.48% 5.49 - 114.31% 8.94 - 11.92% 11cP2 (16-16) – Recharging 2-12 2[4]*L 2cOpt. Model 3cAlgorithm 1 3cAlgorithm 2 3cAlgorithm 3 2-12 Time (s) Obj. Value Time (s) Obj. Value Gap (%) Time (s) Obj. Value Gap (%) Time (s) Obj. Value Gap (%) 1c|0.2 2.63 3.5817 0.55 3.5817 0.00% 0.44 3.5817 0.00% 0.47 3.5817 0.00% 1c|0.4 82.08 3.5817 0.46 3.6717 2.51% 0.40 3.6717 2.51% 0.49 3.7663 5.15% 1c|0.6 19.84 3.2325 0.35 3.4684 7.30% 0.33 3.7302 15.40% 0.55 3.7508 16.03% 1c|0.8 90.52 2.7967 0.90 3.5360 26.43% 0.66 3.5360 26.43% 1.11 3.3767 20.74% 1c|1 209.00 2.2349 1.25 3.6115 61.59% 1.22 3.6115 61.59% 1.22 3.6115 61.59% 1c|1.2 631.79 1.7369 1.28 3.3403 92.31% 1.31 3.3403 92.31% 3.16 2.5199 45.08% 1c|1.4 1645.00 1.1662 1.17 3.3403 186.43% 1.20 3.3403 186.43% 3.26 2.3639 102.71% 1c|1.6 69.15 0.4673 1.38 3.2021 585.27% 1.36 2.9233 525.62% 15.25 0.4673 0.00% 1c|1.8 16.60 0.4438 1.19 3.5377 697.22% 1.25 3.3313 650.72% 15.41 0.4438 0.00% 1c|2 6.51 0.4438 1.13 3.0457 586.34% 1.18 3.0333 583.56% 15.05 0.5803 30.76% 1c|2.2 37.22 0.4438 3.78 1.6636 274.90% 3.44 1.6636 274.90% 15.66 0.4438 0.00% 1c|2.4 21.44 0.4438 3.56 1.6636 274.90% 3.40 1.6636 274.90% 15.02 0.4438 0.00% 1c|2.6 13.89 0.4438 3.41 1.6636 274.90% 3.43 1.6636 274.90% 14.99 0.4438 0.00% 1c|2.8 11.76 0.4438 14.77 0.4438 0.00% 15.72 0.4438 0.00% 15.50 0.4450 0.28% 1c|3 23.71 0.4438 15.00 0.4438 0.00% 14.89 0.4450 0.28% 15.15 0.4438 0.00% 1c|3.2 23.79 0.4438 14.71 0.4598 3.61% 14.75 0.4438 0.00% 15.26 0.4438 0.00% 1c|3.4 22.98 0.4438 14.87 0.4438 0.00% 14.88 0.4438 0.00% 14.94 0.4438 0.00% mygray Average 172.23 - 4.69 - 180.81% 4.70 - 174.68% 9.56 - 16.61% $P1,\;L = 0.2$ $P1,\;L = 0.6$ $P1,\;L = 0.8$ $P1,\;L = 0.2$ $P1,\;L = 0.6$ $P1,\;L = 0.8$ $P1,\;L = 1.0$ $P1,\;L = 1.4$ $P1,\;L = 1.8$ $P1,\;L = 1.0$ $P1,\;L = 1.4$ $P1,\;L = 1.8$ $P2,\;L = 0.2$ $P2,\;L = 0.6$ $P2,\;L = 0.8$ $P2,\;L = 0.2$ $P2,\;L = 0.6$ $P2,\;L = 0.8$ $P2,\;L = 1.0$ $P2,\;L = 1.4$ $P2,\;L = 1.8$ $P2,\;L = 1.0$ $P2,\;L = 1.4$ $P2,\;L = 1.8$ Examples of optimal solutions of Problem III (multiple centers with recharging) obtained by the optimization model for problems $P1$ with 16 customer nodes and 12 drone nodes and $P2$ with 16 customer nodes and 16 drone nodes for select values of drone range $L$ are provided on the left three columns. Their corresponding (optimal/suboptimal) solutions obtained by Algorithm <ref> is shown on the three columns on the right. It can be seen that drones are not necessarily assigned to their closest center or closest customer. This is due to the fact that the truck is not required to wait at a center while its assigned drones are traversing the last edge of their route (after delivery) to go back to their base. For small drone range such as $L = 0.2$ (with a 2-to-1 relative drone speed assumption) we are better off to not use the drones and for very large drone ranges such as $L\geq 1.8$ the problem reduces to a one center (drone-only) delivery problem. §.§ Results for Multiple Centers with Revisiting We also tested the MILP model of Problem IV and our algorithms on the same problem instances with the same relative drone speed of 2. <ref> summarizes the results of this experiment. We observe that the average computational time of the optimization model, when compared to that of Problem III, is significantly increased. However, this is mostly due to one to two values of $L$ and for the rest the run time is closer to, although still higher than, what we observed for Problem III. This increase is reasonable because the solution space of Problem IV is substantially larger than the solution space of Problem III. The revisiting model achieved better solutions than the recharging model in 13 instances of $P1$ and one instance of $P2$. The average running time of the final version of our algorithm (Algorithm <ref>) is 6.11 seconds versus 3397.84 seconds of the optimization model for $P1$ and 6.66 seconds versus 1202.10 seconds for $P2$. The average (maximum) optimality gap of the solutions provided by Algorithm <ref> for $P1$ and $P2$ is 28.01% (279.14%) and 17.82% (83.38%), respectively. <ref> illustrates solutions of the considered instances of Problem IV solved by our optimization model for select values of $L$ as well as the corresponding solutions obtained by Algorithm <ref>. Comparison between the results of the optimization model for Problem IV (Multiple Centers with Revisiting) and the algorithms on two synthetic examples with 16 and 12 and 16 and 16 customers nodes and drone nodes, respectively, distributed randomly in a unit box and for different values of $L$. The column “Time (s)” shows the computational time in seconds. The column “Gap” presents the optimality gap of the solutions found by the algorithms. The instances for which the optimal solution is found by our algorithms are shown in bold (0.00% gap). The underlined optimal objective values show the instances that the revisiting model achieved better solution than the recharging model. 11cP1 (16-12) – Revisiting 2-12 2[4]*L 2cOpt. Model 3cAlgorithm 1 3cAlgorithm 2 3cAlgorithm 3 2-12 Time (s) Obj. Value Time (s) Obj. Value Gap (%) Time (s) Obj. Value Gap (%) Time (s) Obj. Value Gap (%) 1c|0.2 2.99 3.3325 0.40 3.3325 0.00% 0.37 3.3325 0.00% 0.48 3.3325 0.00% 1c|0.4 18.87 3.2561 0.63 3.6008 10.59% 0.51 3.5956 10.42% 0.61 3.3176 1.89% 1c|0.6 954.06 3.0865 0.48 3.4071 10.39% 0.46 3.4071 10.39% 0.75 3.6581 18.52% 1c|0.8 12324.75 2.4651 0.52 3.4377 39.45% 0.61 3.4765 41.03% 1.25 2.8226 14.50% 1c|1 1758.83 1.5250 0.68 3.2693 114.38% 0.83 3.3081 116.92% 1.09 2.2191 45.51% 1c|1.2 40669.77 1.3796 0.95 3.2953 138.86% 0.87 2.7749 101.14% 2.39 1.9162 38.89% 1c|1.4 34.30 0.5029 0.99 2.7771 452.25% 0.91 2.7862 454.06% 2.03 1.9066 279.14% 1c|1.6 56.83 0.4742 1.11 2.9175 515.21% 1.09 2.6562 460.12% 9.84 0.5181 9.25% 1c|1.8 87.52 0.4563 2.23 1.6961 271.72% 2.21 1.6961 271.72% 9.48 0.6040 32.37% 1c|2 120.90 0.4563 2.35 1.6742 266.91% 2.40 1.4503 217.86% 9.44 0.4690 2.79% 1c|2.2 142.73 0.4563 2.42 1.6742 266.91% 2.31 1.4447 216.61% 9.55 0.4775 4.66% 1c|2.4 178.49 0.4563 9.54 0.4709 3.21% 9.74 0.4727 3.60% 9.61 0.4775 4.66% 1c|2.6 228.82 0.4563 9.47 0.4805 5.31% 9.32 0.4856 6.43% 9.55 0.4846 6.20% 1c|2.8 226.00 0.4563 9.60 0.4832 5.90% 9.59 0.4686 2.70% 9.53 0.4694 2.88% 1c|3 272.74 0.4563 9.39 0.4828 5.81% 9.49 0.4828 5.81% 9.38 0.4710 3.23% 1c|3.2 335.45 0.4563 9.61 0.4828 5.81% 9.40 0.4767 4.47% 9.42 0.4828 5.81% 1c|3.4 350.22 0.4563 9.39 0.4828 5.81% 9.45 0.4742 3.93% 9.51 0.4832 5.90% mygray Average 3397.84 - 4.10 - 124.62% 4.09 - 113.36% 6.11 - 28.01% 11cP2 (16-16) – Revisiting 2-12 2[4]*L 2cOpt. Model 3cAlgorithm 1 3cAlgorithm 2 3cAlgorithm 3 2-12 Time (s) Obj. Value Time (s) Obj. Value Gap (%) Time (s) Obj. Value Gap (%) Time (s) Obj. Value Gap (%) 1c|0.2 3.44 3.5817 0.31 3.5817 0.00% 0.33 3.5817 0.00% 0.33 3.5817 0.00% 1c|0.4 55.07 3.5817 0.28 3.5817 0.00% 0.30 3.5817 0.00% 0.38 3.7663 5.15% 1c|0.6 16.14 3.2325 0.38 3.7280 15.33% 0.34 3.7280 15.33% 0.43 3.7508 16.03% 1c|0.8 97.30 2.7967 0.58 3.5360 26.43% 0.65 3.5360 26.43% 0.85 3.3767 20.74% 1c|1 475.36 2.2349 0.86 3.5589 59.24% 1.10 3.5571 59.16% 0.93 3.5588 59.23% 1c|1.2 7490.24 1.7369 1.15 3.1890 83.60% 1.07 3.1868 83.47% 1.06 3.1851 83.38% 1c|1.4 10233.86 1.1598 1.13 3.1689 173.23% 1.09 3.1851 174.64% 2.16 2.1160 82.45% 1c|1.6 24.63 0.4673 1.09 3.1689 578.16% 1.04 2.9051 521.71% 10.62 0.5055 8.18% 1c|1.8 57.27 0.4438 0.95 3.0112 578.58% 1.06 2.9436 563.34% 10.80 0.4774 7.59% 1c|2 54.01 0.4438 0.98 3.4090 668.23% 0.96 3.1674 613.78% 10.66 0.4438 0.00% 1c|2.2 81.60 0.4438 5.38 1.8542 317.85% 5.40 1.8129 308.55% 10.80 0.4675 5.36% 1c|2.4 124.37 0.4438 2.16 2.1089 375.25% 2.12 2.0368 358.99% 10.83 0.4631 4.37% 1c|2.6 167.88 0.4438 2.13 2.1235 378.54% 2.06 1.7725 299.44% 10.55 0.4640 4.56% 1c|2.8 280.61 0.4438 10.69 0.4438 0.00% 10.65 0.4438 0.00% 10.78 0.4701 5.94% 1c|3 295.28 0.4438 10.61 0.4461 0.53% 10.65 0.4438 0.00% 10.56 0.4438 0.00% 1c|3.2 460.98 0.4438 10.63 0.4673 5.30% 10.78 0.4626 4.25% 10.71 0.4438 0.00% 1c|3.4 517.70 0.4438 10.79 0.4438 0.00% 10.77 0.4673 5.30% 10.77 0.4438 0.00% mygray Average 1202.10 - 3.54 - 191.78% 3.55 - 178.49% 6.66 - 17.82% $P1,\;L = 0.2$ $P1,\;L = 0.6$ $P1,\;L = 0.8$ $P1,\;L = 0.2$ $P1,\;L = 0.6$ $P1,\;L = 0.8$ $P1,\;L = 1.0$ $P1,\;L = 1.4$ $P1,\;L = 1.8$ $P1,\;L = 1.0$ $P1,\;L = 1.4$ $P1,\;L = 1.8$ $P2,\;L = 0.2$ $P2,\;L = 0.6$ $P2,\;L = 0.8$ $P2,\;L = 0.2$ $P2,\;L = 0.6$ $P2,\;L = 0.8$ $P2,\;L = 1.0$ $P2,\;L = 1.4$ $P2,\;L = 1.8$ $P2,\;L = 1.0$ $P2,\;L = 1.4$ $P2,\;L = 1.8$ Examples of solutions for Problem IV (multiple centers with revisiting) on $P1$ and $P2$ instances for select values of maximum drone range $L$ obtained by the optimization model (optimal solutions) on the three columns on the left and by the revisiting version of Algorithm <ref> (optimal/suboptimal solutions) on the three columns on the right. Similar to the results of Problem III, small drone range makes the problem closer to a pure TSP (truck-only) model and for reasonably large $L$ the problem gets closer and closer to the one center (drone-only) delivery problem. It can be seen that for some large enough values of $L$ in both model and algorithm solutions, e.g., in (fig:model-P1-L1.8) and (fig:MC3-P2-L1.4), some of the drones were able to revisit the pickup center after delivering a package at a customer node. Such incidents are distinguishable by hub-and-spoke type (non-triangular) drone routes. §.§ Similarities Between the Results of the Two Problems It is apparent in Tables <ref> and <ref> that the objective function value for the optimization model and the algorithms decreases in $L$ as expected. We also see in Tables <ref> and <ref> that starting from $L=1.8$ the optimal objective function value remains the same. This is because at this level, the range of the drone is long enough to serve all customers with drones and thus the truck would not be used. This means for large enough $L$ and large enough relative speed of drone to truck, the problem essentially reduces to a one center problem with recharging (Problem I). This is also visible in Figures <ref> and <ref>. Furthermore, it can be seen in the results from both Problems III and IV that the running time of the optimization model and the optimality gap of our algorithm are both very sensitive to the values of $L$. It is much easier to solve the problem when $L$ is either small or large. Optimality gap for the middle values of $L$ (between 1.0 to 1.4 in these instances) is also larger. If the drone range is very low, which means drones can only fulfill the demand of nearby customer nodes, the problem is a pure or nearly TSP problem which is much faster to solve since its feasible region is smaller. In contrast, if the drone maximum allowed distance is in a reasonable range (between 1.0 to 1.4 in these instances), drones can be a big help in delivering the packages while still leaving a good portion of the customers to be served by the truck. Finding both drone routes and the truck route takes considerable time in this case, although the lion share of total time is due to the drone routes. Therefore, the feasibility region is much bigger than a TSP which increases the run time and makes finding optimal or high quality solutions harder. On the other end of the range, if the drone range is large enough (e.g., larger than 1.8), we can do all deliveries with drones without using the truck. In this case, the solving time is reasonable but still a bit higher than pure TSP. This is because larger values of $L$ increase the number of feasible drone routes. Moreover, we still have to determine which node(s) will serve as the truck node(s), and what's the best truck route (if any). This explains why there is a little bit of increase in solving time at the right tail of the drone range compared to the left tail. Although the solution space might be even bigger than that of the mid-level range of $L$ (e.g., between 1.0 to 1.4), it is much faster than that case because good upper bounds are found quickly (with for example fewer (or even one) centers) which helps skipping a large number of feasible solutions that use bigger number of centers. For the same reasons we can see a similar behavior for the optimality gap and when $L$ is either small or large, the gap between the algorithm and the optimal solution is much smaller than the results of the mid-level cases of drone range. §.§ Comparison with Traditional Delivery System We also compared our approach with the traditional centralized truck-only delivery system in which a single truck would deliver all packages, i.e., a pure TSP model. We ran both recharging and revisiting versions of Algorithm <ref>, which provides a decentralized alternative, as well as the LKH algorithm for the pure TSP model on several problem instances with 60 customer nodes and 40 drone nodes distributed uniformly at random in a unit square. Again, we set the relative speed of drone to be 2. For drone range we chose $L=0.8$ to ensure having both truck and drone routes. Figure <ref> shows solutions obtained by our algorithm presented side-by-side with the corresponding TSP solution for each instance. It is easy to see the advantages of our proposed decentralized system over the conventional centralized delivery system. On these examples, the average delivery time when we combine a truck with crowdsourced drones was 4.6280, while the average delivery time for the same problem if the truck serves all the customers was 6.2035. This crude comparison suggests an almost 25.32% improvement in the efficiency of the last-mile delivery, in our randomly generated examples, if we combine truck delivery and drone delivery in the context of sharing economy platforms. Certainly, this would highly depend on the chosen input parameters that we investigate further in Sections <ref> and <ref>. Total Time = 6.1880 Total Time = 5.0042 Total Time = 5.8312 Total Time = 4.5983 Total Time = 6.5363 Total Time = 4.9806 Total Time = 6.2583 Total Time = 3.9290 Sample solutions for a multiple center with recharging problem with 60 customer nodes and 40 drone nodes for $L=0.8$ obtained by our Algorithm <ref>, representing a decentralized delivery system, for different random examples in a unit box with their objective function values are illustrated side by side to their corresponding TSP solutions (centralized truck-only delivery system) for the same instances obtained by the LKH algorithm. § SENSITIVITY ANALYSIS In the rest of the paper, we perform an extensive sensitivity analysis with respect to several factors to study their impact on the quality of the solution and savings of the shared delivery system compared to the traditional truck-only delivery. These factors include speed of drones, number of available drones, a measure combining speed and number of drones, and customer distribution. Moreover, we analyze the impact of improvements presented in <ref> for different values of the input parameters. Finally, a comparison is made between three models to measure the impact of shared delivery model on carbon footprint. These three models are the traditional truck-only delivery (TSP) model, delivery with a truck and a drone where the truck carries a drone and both deliver packages in a coordinated way (we call this model TSPD), and our truck and shared drone delivery model (TSP-SD). Almost all simulations are done on an instance of the recharging problem with 60 customer nodes and several drone nodes (the number varies based on the input parameters) distributed uniformly at random or according to a distribution (depending on the analysis) in a unit square assuming $L=0.8$. In most cases, the revisiting model delivers similar results but wherever we expect a significant difference we run the simulation on the revisiting problem as well. To analyze the sensitivity of the proposed models and algorithms that combines a truck with crowdsourced drones in the last-mile delivery operation to input parameters, we consider the impact of several factors including relative speed of drones, relative number of crowdsourced drones, and the customer and drone distribution on the solutions obtained by our algorithm. We expect similar behavior in the optimal solutions. Here we define two parameters as: $\rho_1 = \frac{V_d}{V_t}=\frac{\text{speed of drone}}{\text{speed of truck}}$ and $\rho_2 = \frac{m}{n}=\frac{\text{number of drones}}{\text{number of customers}}$. In terms of customer distributions, a multivariate normal distribution is assumed and its variance will be adjusted. We compare the results with respect to the savings in time by using the crowdsourced drones that is \[ \frac{\text{Time cost of TSP (truck only) - Time cost of TSP-SD (crowdsourced drones combined with a truck)}}{\text{Time cost of TSP (truck only)}} \,, \] i.e., we compare our solutions with those obtained by the LKH algorithm. When comparing Algorithm <ref> to the results of LKH for pure TSP, one should keep in mind that any savings suggested by our algorithm in despite the fact that for our algorithm with $L=0.8$, we usually have a considerable optimality gap, while the optimality gap of LKH for pure TSP is almost zero. This difference could be seen as additional savings if our problem is solved to optimality. §.§ Relative Speed of Drones To analyze the sensitivity with respect to $\rho_1$, we adjust this parameter in the range $0.75\leq \rho_1 \leq 3$ that means the drone speed can vary from 0.75 times the truck speed to three times the truck speed, while we fix the number of drones to be half or twice of the number of customers, i.e., $\rho_2\in\{0.5,2\}$. The choice of the range for $\rho_1$ is driven by the fact that it is known in the literature of the centralized alternative (TSPD) that for having any significant gain by combining a truck and a drone in delivery operations we must have $\rho_1 \geq 2$ <cit.> and it is most often assumed to have $\rho_1 \geq 1$. $\rho_2 = 0.5$ $\rho_2 = 2$ Sensitivity analysis of the results of Algorithm <ref> with respect to the relative speed of drones for two cases of drones supply. From <ref> we can find that, when there is considerable shortage of drones relative to the number of customers, only when the drone speed is less than 0.75 times the speed of the truck ($\rho_1 \leq 1.15$), the TSP model will have less time cost than the model with shared drones. Also, from <ref> we can observe that when relative drone supply increases to 2 the chart shifts to left, so to speak, and we see a positive and significant savings for the shared model for all values of $\rho_1$ in the considered range ($0.75 \leq \rho_1 \leq 3$). In both situations, the TSP with sharing drones model has a consistently better performance than the TSP model when drones are reasonably faster than the speed of the truck ($\rho_1 \geq 1.15$). Intuitively, the combined model gains help from the drones by working (potentially) in parallel to each other and to the truck, and on the other hand loses some time by idling the truck at the cluster centers to wait for the drones to fly back and forth and deliver packages. The overall, tradeoff between these two opposite forces shows that the combined model has a significant advantage over the pure TSP model and the proposed model of combined truck and shared drones performs significantly better. This advantage becomes bigger by increasing drones speed relative to the speed of truck ($\rho_1$) with a diminishing return. Both these properties are quite consistent with the results of <cit.> that shows the gain for a TSPD model over TSP is proportional to $\sqrt{\rho_1}$. Finally, we also see that increasing the supply of drones ($\rho_2$) increases the savings. This phenomena is similar to growing the supply side in other sharing economy (two-sided market) applications. §.§ Relative Number of Drones We analyze the sensitivity with respect to $\rho_2$, by varying this parameter in the range $0.3 \leq \rho_2 \leq 2$, while we fix the speed of drones to either be one or two times that of the truck, i.e., $\rho_1\in\{1,2\}$. The choice of the range for $\rho_2$ is motivated by the well-known importance of having a balance between supply and demand in two-sided markets, i.e., $\rho_2=1$. $\rho_1 = 1$ $\rho_1 = 2$ Sensitivity analysis of the results of Algorithm <ref> with respect to the relative number of drones. From <ref>, we can find that if the speed of the drones is equal to the speed of the truck, for the TSP-SD model it is sufficient to have $\rho_2 \geq 0.6$ to consistently outperform the pure TSP model. Increasing the relative speed of drones to 2 (<ref>) both increases the magnitude of savings and the enlarges the range of $\rho_2$ that deliver positive savings ($\rho_2 \geq 0.45$). The same increasing and concave pattern can be observed in $\rho_2$ as well. Comparing <ref> to <ref> we see that the sensitivity of the performance of TSP-SD to relative speed of drones is much more than its sensitivity to relative number of drones. It is reasonable to see that the number of drones is not as important as the speed of drones. This is because, if drones are fast enough, we only need a few drones in that area to make the performance of the model with drones better than the normal TSP model, while having huge number of slow moving drones may still force us to prefer truck over the drone on many of the deliveries. §.§ Combining In this section, we try to combine the analysis of $\rho_1$ and $\rho_2$ together to see how much we can improve the delivery performance by including the crowd-owned drones when we consider both parameters at the same time, i.e., comparing TSP-SD with TSP. We let the parameters to vary in the ranges $0.75\leq \rho_1 \leq 3$ and $0.3 \leq \rho_2 \leq 2$. The sensitivity analysis of the results of Algorithm <ref> with respect to the relative speed of drones combined with the relative number of drones. The positive savings in (fig:SA3CompleteRange) are shown in (fig:SA3PositiveRange). From <ref>, we can see that if both $\rho_1$ and $\rho_2$ are small, the normal TSP will have much better performance than the model with sharing drones (TSP-SD), which we can also observe from the two previous sections, especially in <ref>. On the other hand, from <ref>, which shows the range with positive savings, we can see that the performance of the TSP-SD model improves as $\rho_1$ and $\rho_2$ increase. It reaches its maximum percentage of savings around $63.47\%$ when both $\rho_1$ and $\rho_2$ are around our set upper bounds of 3 and 2, respectively, implying that the improvement could increase further by relaxing this upper end of the range. Meanwhile, the growth rate becomes smaller as $\rho_1$ and $\rho_2$ grow, which is apparent by the concavity of the curve and suggests a diminishing return phenomena. We expect even bigger savings for the revisiting model but the pattern stays the same. For the other centralized alternative (TSPD) it is shown that for $\rho_1 \leq 2$ the gain in delivery time by using a drone is insignificant <cit.>. However, our shared drone delivery system shows considerable gain for a wider range of $\rho_1$ that makes our model more applicable across a variety of available drone technologies and regulatory environments. §.§ Population Distribution We consider two cases with different population distributions. The first case has one population center of 60 customers and 40 drone service providers distributed inside a unit box according to a multivariate Gaussian distribution with mean $\mu=(0.5,0.5)$ and covariance matrices $\Sigma=\usebox{\smlmats}$, where $\sigma^2 \in [0.005,0.1]$ with increments of 0.005. The second case is a clustered data with four population centers of 60 customers and 40 drone providers (across the clusters) distributed inside a unit box according to an even mixture of four truncated multivariate Gaussian distributions with means $\mu_{1},\mu_{2},\mu_{3},\mu_{4}=(0.25.0.25),(0.25,0.75),(0.75,0.75),(0.25,0.75)$ and covariance matrices $\Sigma_{1}=\Sigma_{2}=\Sigma_{3}=\Sigma_{4}=\usebox{\smlmats}$, where $\sigma^2 \in [0.005,0.1]$ with increments of 0.005. In both cases we assume $L=0.8$ and $\rho_1 = 2$ and compare the results of Algorithm <ref> with the results of LKH for pure TSP for different values of $\sigma^2$. The <ref> shows examples of generated instances with the solutions found by the recharging version of our Algorithm <ref>. Complete results for these instances are presented in <ref>. One Population Center with $\sigma^2$ = 0.1 One Population Center with $\sigma^2$ = 0.005 Four Population Centers with $\sigma^2$ = 0.1 Four Population Centers with $\sigma^2$ = 0.005 Two cases of population distribution with two values of $\sigma^2$ and the solutions of Algorithm <ref> for them. We see that having a clustered population may make a TSP tour a necessity in our solution regardless of how large drone range is relative of the variance. However, we can expect shorter TSP tour when variance is smaller. One Population Center Four Population Centers Comparison of TSP and TSP-SD models for two cases of population distribution for varying values of variance. We can see from <ref> that the proposed combined model of truck and sharing economy drones (TSP-SD) consistently and significantly performs better than the pure TSP (truck-only) model in both cases and across the range of the variance. Its best performance usually happens when customers are concentered in one place, i.e., very small variance, whether we have one center or several centers of population. This is consistent with real-life applications as in the metropolitan areas population distribution is usually nonuniform and residences usually live near each other in a neighborhood or town while having several neighborhoods or towns in the metropolitan area with some distances from each other. When the variance is low, we can see savings of up to $80\%$ in the delivery time, when compared to TSP, by using drones in the once center case (<ref>). Although this savings becomes smaller when the distribution of the customers gets closer to a uniforms distribution, we can still expect at least $30\%$ savings in the delivery time. When there are four centers of population in a large city, we can still see between $17\%$ and $45\%$ savings in the delivery time (<ref>). Unlike the one population center case that shows a negative correlation between the savings and $\sigma^2$, in four population centers case we see that the savings stays more or less flat that is due to the clustered nature of the population. These observations on the sensitivity of our proposed model to the customer distribution is consistent with findings of <cit.> that suggests that the benefits of the crowdsourced delivery depends on the spatial characteristics of the network (of pickup/delivery/transfer locations and roads). §.§ Impact of Moving Cluster Centers and Merging Centers Here, we compare the performance of our original heuristic, presented in <ref> in which the cluster centers (truck nodes) could be anywhere in the continuous space, with the final improved <ref>, and Algorithm <ref> discussed in <ref> in which the cluster centers (truck nodes) are located at some customer nodes. As mentioned in Sections <ref> and <ref>, moving a cluster center to customer nodes needs to satisfy one condition that is the new center needs to be able to still cover all of its assigned customers using its assigned drones. When merging the centers, as discussed in <ref>, this feasibility is also satisfied in another way by reallocation of customers. <ref> compares an example solution for before and after moving the cluster centers customer nodes (i.e., the original heuristic vs. <ref>). Without Moving Centers After Moving Centers A comparison of solution of our original heuristic, presented in <ref>, in (fig:SA_OHMC3_OH) and the solution obtained by <ref> after moving cluster centers to customer nodes and merging centers in (fig:SA_OHMC3_MC3). In this instance we have 60 customer nodes, 30 drone nodes ($\rho_2=0.5$), distributed uniformly at random in a unit box with $\rho_1=3$ and $L=0.8$. Sensitivity analysis of the impact of moving cluster centers to customer nodes via <ref> and <ref> and further merging them together in <ref> on the performance when compared to the original heuristic presented in <ref>. (fig:SA_MCrho2Half) shows the results for $\rho_2=0.5$ and (fig:SA_MCrho2Two) presents the results for $\rho_2=2$. The range of relative speed in both cases is $0.75 \leq \rho_1 \leq 3$. The sensitivity of the impact of this improvement with respect to drone range in illustrated in fig:SA_MC_L which shows the maximum improvement for middle range $L$. To evaluate the overall impact of this change, we use assume $0.75 \leq \rho_1 \leq 3$ and $\rho_2 \in \{0.5,2\}$ for our simulation. The results of comparison between <ref> and our original heuristic of <ref> is presented in <ref>. In Figures <ref> and <ref> We see about 45% and 38% savings on average for two values of $\rho_2$, respectively, which suggest a slight decrease in savings by increasing $\rho_2$. This is because when we have an abundance of drone supply the original heuristic could also perform well making the gap between the two algorithms narrower. However, this average decrease is quite small. The savings are more or less flat across different values of $\rho$ as well. This suggests the robustness of the implemented improvements with respect to $\rho_1$ and $\rho_2$, providing more than roughly 40% savings in the delivery time when compared to the algorithm before improvement. However, the design of <ref> hints more sensitivity with respect to drone range $L$ for the purpose of this comparison. This analysis for increments of 0.4 in $L$ is shown in <ref>. We see the biggest improvement (up to 68%) for the middle range $L$ where we needed it the most as already seen in Tables <ref> and <ref>. For very large $L$ both problems can result to one truck node (serving all customers with drones). If the location of this center is the same in the solutions obtained by these two algorithms, the only remaining factor making a difference in the solution is the drone routes obtained by the tabu search step. This explains the small gap for $L=2.8$. §.§ Revisiting versus Recharging Comparing the revisiting and recharging modes using the results of the algorithm is tricky. When using the proposed algorithm we could see revisiting model as a battery utilization modification to the algorithm that allows the drones to revisit the truck node after a delivery to pick up another package if their battery limits allow that. However, one should keep in mind that achieving any improvement by this modification highly depends on the problem instance and the input parameters. When comparing the results of the optimization models we expect to see further improvement by allowing revisiting for the drones that was identified in <ref> in several instances. However, comparing the results of <ref> in Tables <ref> and <ref> we can see a mixed situation and in some instances the recharging model is providing a better solution. This is due to the additional complexity of the revisiting model that makes it much harder for the algorithm to find better solutions for certain instances of the problem, resulting in a larger optimality gap and potentially weaker solution than the recharging model. In other words, we are dealing with a larger feasibility set potentially containing a better optimal solution but depending on the complexity of the problem instance we may end up with a better or worse solution than the output of the recharging model. <ref> shows an example of implementing this modification that results to weaker solutions for the revisiting model. When revisiting is desired we suggest to run both recharging and revisiting versions of the algorithm and pick the best. For smaller size problems we can obviously solve the problem with the optimization model of Problem IV to expect better results. Further adjustment of the parameters of the algorithm to better handle the complexity of the revisiting model for such unfriendly instances may also help to achieve positive savings compared to the recharging version of <ref>. Savings analysis when $\rho_2=0.5$ Savings analysis when $\rho_2=2$ Illustration of the impact of battery utilization modification that shows savings, for various values of $\rho_1$ with $\rho_2=\{0.5,2\}$, by comparison of three algorithms including recharging and revisiting versions of <ref> and the original heuristic of <ref>. § THE IMPACT ON CARBON FOOTPRINT AND COMPARISON WITH TSPD It is clear that drone delivery systems such as our proposed combined system and the coordinated but centralized system of a truck and a drone in which the drone is controlled from the truck (the TSPD model[We follow the name used by <cit.>. This is also called horsefly problem, truck-and-drone TSP, etc.]) have less carbon footprint than a truck-only (pure TSP) delivery system since drones are much more energy-efficient than trucks. In order to evaluate the energy efficiency of our proposed shared drone delivery model, we compare it with the TSPD model. First, note that if in both models we assume the drones are capable of carrying more than one package, the shared delivery model would have less carbon footprint since it becomes very similar to the comparison of the generalized traveling salesman problem (GTSP) and the TSP done in the earlier paper <cit.> that showed the household-level transportation has its own economies of scale and most likely outperforms a centralized delivery system. Now assume the drones can only carry one package per trip (which is the case in most current delivery drone technologies) and have to return to their home base (for replacing battery or recharging) after each delivery. The comparison would obviously depend on the balance between the supply of drones and demand in deliveries. The <ref> shows a comparison between the two systems in two situations when there is only one drone in each cluster ($\rho_2=0.18$ for this instance with 60 customers), i.e., extreme shortage of drone supply, and <ref> illustrates the results of this comparison when we increase drone supply to more reasonable levels in sharing economy environments, which as we will discuss next makes this comparison more fair as well. We use the ratio $\frac{\text{Time cost of TSPD - Time cost of TSP-SD}}{\text{Time cost of TSPD}}$ to calculate the percentage of savings. For the TSPD model we assumed that drone can pickup a package at one stopping point and returns to the truck after delivering it at either the same point or at another stopping point which makes the model more flexible and more efficient. We also assume no restriction of the drone range. Both of these assumptions give an advantage to this model over our TSP-SD. We solved this problem via the algorithms developed in our working papers <cit.> and picked the best result. Figures <ref> and <ref> show the solution of the TSPD model and the sharing model TSP-SD, respectively, for a problem with 60 customers distributed uniformly at random in a unit box. We have set $\rho_1=3$ and $L=0.8$ for both models. Using the binary search and $k$-means step of <ref> to find a feasible $k$ we found that $k=11$ cluster centers are needed. The number of clusters is further reduced by <ref> to 9 via merging centers as visible in <ref>. Assuming one drone in each of the original 11 clusters gives $\rho_2 = 0.18$. It is clear from <ref> that with this restrictive assumption the TSPD model will have much better performance than our TSP-SD model. This is understandable as in the shared model, the truck needs to wait at the cluster centers for one drone to serve all the customers which looks very inefficient when drone supply is limited. The drone-traveling time is huge in this circumstance and the truck idles when it could serve customers. In contrast, in the TSPD model the truck can serve another customer while the drone is delivering the package as well. It is obvious that the TSPD model is faster than the shared model if there is only one drone in each cluster since there is no such huge waiting time for the truck. TSPD can be seen as a model that has one drone for each customer in the neighborhood. To see this more clearly, imagine a very large relative drone speed, i.e., letting $\rho_1 \rightarrow \infty$. In such extreme case the truck will stop at some point in the region and all the customers will be served by the drone, forming a star-shape combined route. This situation is equivalent to a situation where each customer node sends its own drone to the truck to pick up its package and bring it back to customer's home, i.e., $\rho_2=1$. The sharing economy model assumes reasonable drone supply (such as a perfect balance of supply-demand with $\rho_2=1$) with limited relative drone speed $\rho_1$. Obviously, this situation justifies using the truck for serving some of the customers and leaves the question of finding the optimal trade-off. Therefore, to have a fair comparison we have to increase the relative number of drones in the sharing model. Note that the assumption of unlimited drone range ($L \rightarrow \infty$) for the TSPD model also gives a significant advantage to that model. This is one of the reasons behind the fact that TSPD is showing less sensitivity to $\rho_1$ in its considered range as can be observed in <ref> and Figures <ref> – fig:TSPDvsSharing-Savings-rhoTwoTwo. Of course, this will change when $\rho_1$ increases significantly beyond this range. Solution of the TSPD model Solution of the TSP-SD model with one drone in each cluster Comparison of TSPD model and TSP-SD model with one drone in each cluster ($\rho_2=0.18$) Comparison of the TSPD model the shared delivery model (TSP-SD) assuming only one drone in each cluster of the shared model. To investigate the impact of the balance between the supply of drones and the demand in deliveries, we did a sensitivity analysis in terms of the number of drones. In <ref>, it is clear that when the relative number of drones increases to half of the customers, the average savings of the sharing model can be as much as about 6%, an increase by 137 percentage points compared to the case with one drone in each cluster. We see in <ref> that from $\rho_1=1.35$ we start to see positive savings in this case. Furthermore, when there is the same number of drones and customers in the neighborhood ($\rho_2 = 1$), the sharing model will have better performance compared to the TSPD model on all values of $\rho_1$, and can achieve about $30\%$ savings on average. If we further increase the drone supply to $\rho_2 = 2$ there is more than $46\%$ savings on average in the delivery time. This is consistent with other sharing economy application and shows that to use shared drones in a delivery system, we need to have enough supply (of drones) relative to the demand in deliveries in order to have a better performance than the centralized alternative of coordinated truck and drone delivery system. Clearly, the gap between the TSPD and our TSP-SD model when we allow drones to revisit the pickup centers becomes wider, at least when we compare the optimal solutions. Assuming a larger $L$ will also increase this gap significantly. Comparison of the TSPD model the shared delivery model (TSP-SD) for different values of relative number and speed of drones. § CONCLUSION In this paper we have developed a shared last-mile delivery model in which a truck carries packages to a neighborhood and then outsources the last piece of the trip to private drone operators whose service can be utilized on a sharing economy platform. We have developed several optimization models for different versions of the problem and proposed efficient algorithms to solve them. Our computational analysis show that the shared delivery model (decentralized model) is much more efficient than the traditional truck-only delivery model (centralized model) in almost all possible scenarios and offers significant savings in the delivery time. This is aligned with the results from <cit.>. The comparison between the shared delivery model and the coordinated delivery system, in which a truck carries and controls a drone during the delivery operation, depends on other factors such as number of available drones in the platform, their capacity and speed. Our analysis in this case also shows a considerable overall advantage for the decentralization. For future work, one may look into considering different factors such as time windows for delivery to customers, time windows for drone availability, ability of drones to carry multiple packages at the same time, weight capacity for the drones, and combination of the system with crowdsourced drivers. Another avenue for extending this work is to incorporate pickup stations into the problem. Finally, it would be interesting to consider stochastic models: one example is to assume that some of the drones with some probability could fail to deliver and have to bring the package back to the truck and that customer must be then served by the truck. § ACKNOWLEDGMENTS The authors gratefully acknowledge support from a Tier-2 grant from Northeastern University (NU) as well as an NSF Planning Grant for establishing an engineering research center (ERC). Authors would also like to thank NU's SHARE Group, led by Ozlem Ergun, for their support and for creating an environment in which this research could be accomplished. *Online Supplement to “Last Mile Delivery with Drones and Sharing Economy” § ALTERNATIVE MODEL FOR PROBLEM I For the alternative formulation of Problem I, we change some of the settings of the model as detailed in Table <ref>. Sets, parameters, and variables used in the one center model. Sets/Parameters/Variables Description $C$ Customer Nodes $D$ Drone Nodes $T$ Truck nodes, i.e., only one node here. We call it node $0$. $d_{ij}$ Distance between node $i$ and $j$ for $i,j\in C\cup D \cup T$ $v_D$ Speed of drones $L$ Longest distance a drone can travel without charging battery $x_{ij}$ $\in \N$. The number of trips from node $i$ to node $j$. $q_{i}$ Total travel time of the drone with base at node $i$ $Q$ Maximum time spent by all drones (makespan) Then, an alternative mixed integer linear programming (MILP) model for this problem can be written as follows: \begin{eqnarray} \minimize \qquad Q & & \quad \qquad \st \nonumber \\ & & \nonumber \\ \sum_{i\in D} x_{ji} & = & 1 \,, \qquad \forall j \in C \label{eq:Problem1AlternativeDroneSatisfyAll} \\ x_{0i} + \sum_{j\in C} x_{ij} & = & 0 \,, \qquad \forall i \in D \label{eq:Problem1AlternativeNoSelfNoDirect} \\ x_{0j} & = & 1 \,, \qquad \forall j \in C \label{eq:Problem1AlternativeCenterSatisfyAll} \\ x_{i0} - \sum_{j\in C}x_{ji} & = & 0 \,, \qquad \forall i \in D \label{eq:Problem1AlternativeDroneNodeFlowBalance} \\ \sum_{i \in D}x_{i0} - \sum_{j\in C}x_{0j} & = & 0\,, \qquad \label{eq:Problem1AlternativeCenterFlowBalance} \\ (d_{i0}+d_{0j}+ d_{ji})x_{ji} & \leq & L\,, \qquad \forall i\in D, \; \forall j\in C \label{eq:Problem1AlternativeDroneRange} \\ \frac{1}{v_D}\sum_{j\in C}(d_{i0}+d_{0j}+ d_{ji})x_{ji} & \leq & q_i \,, \qquad \forall i\in D \label{eq:Problem1AlternativeDroneTime} \\ q_i & \leq & Q \,, \qquad \forall i\in D \label{eq:Problem1AlternativeTotalTime} \\ x_{0,j}\, \; x_{ji} & \in & \{0,1\} \,, \qquad \forall i,j\in C\cup D \nonumber \\ x_{i0} & \in & \N \,, \qquad \forall i \in D \nonumber \\ q_i\,, \; Q & \geq & 0 \,, \qquad \forall i\in D \nonumber \end{eqnarray} Constraint (<ref>) ensures all customers are visited once by a drone. Constraints (<ref>) and (<ref>) ensure all customers are serviced via the center (truck node). Constraint (<ref>) and (<ref>) enforce the balance of outbound and inbound flows at each drone node and at the pickup center. Constraint (<ref>) is the drone range limit. Constraints (<ref>) sums the total time (distance) of all routes starting from each drone node. Constraint (<ref>) finds the maximum time (distance) among all drones. § ALTERNATIVE MODEL FOR PROBLEM II For the alternative formulation of Problem I, we change some of the settings of the model as detailed in Table <ref>. We know that each drone can make multiple trips and in each trip it can serve multiple customers. This is obviously much more complicated than the model for Problem I. To do this in an easier way, we hypothetically each trip initiated from a drone node is made with a new drone. Therefore, we index different trips as different drones. Since each drone node would need at most $|C|$ trips to serve all customers we define set $K$ as the set of all hypothetical drones in which the first $|C|$ members are associated with drone node 1, the second $|C|$ members are associated with the second drone node, and so on and so forth. Therefore, we have $|K|=|C|\times|D|$. Sets, parameters, and variables used in the one center model. Sets/Parameters/Variables Description $C$ Customer Nodes $D$ Drone Nodes $K$ Set of artificial drones representing trips for drones in $D$ $T$ Truck nodes, i.e., only one node here. We call it node $0$. $d_{ij}$ distance between node $i$ and $j$ for $i,j\in C\cup D \cup T$ $v_D$ Speed of drones $L$ Longest distance a drone can travel without charging battery $x_{ijk}$ Binary variable. It is $1$ when drone $k$ travels from node $i$ to node $j$ and $0$ otherwise. $Z_{ijk}$ Binary variable. It is $1$ if both $x_{i0k},i\in D$ and $x_{0jk},j\in C$ are 1 and 0 otherwise. $Y_{ijk}$ Binary variable. It is $1$ if both $x_{i0k},i\in D$ and $x_{j0k},j\in C$ are 1 and 0 otherwise. $q_{i}$ Total travel time of the drone with base at node $i$ $Q$ Maximum time spent by all drones (makespan) Then, an alternative mixed integer linear programming (MILP) model for this problem can be written as follows: \begin{eqnarray} \minimize \qquad Q & & \quad \qquad \st \nonumber \\ & & \nonumber \\ \sum_{i\in D\cup \{0\}}\, \sum_{k\in K} x_{jik} & = & 1\,, \qquad \forall j\in C \label{eq:Problem2AlternativeDroneSatisfyAll} \\ \sum_{k\in K} x_{0ik} + \sum_{j\in C} \sum_{k\in K} x_{ijk} & = & 0 \,, \qquad \forall i \in D \label{eq:Problem2AlternativeNoSelfNoDirect} \\ \sum_{k\in K} x_{0jk} & = & 1 \,, \qquad \forall j \in C \label{eq:Problem2AlternativeCenterSatisfyAll-1} \\ \sum_{i\in D\cup C}x_{i0k} - \sum_{j \in C} x_{0jk} & = & 0\,, \qquad \forall k\in K \label{eq:Problem2AlternativeCenterSatisfyAll-2} \\ \sum_{k\in K} x_{i0k} - \sum_{j \in C} \sum_{k\in K} x_{jik} & = & 0\,, \qquad \forall i\in D \label{eq:Problem2AlternativeDroneNodeFlowBalance} \\ \sum_{i \in D\cup C} \sum_{k\in K} x_{i0k} - \sum_{j\in D\cup C} \sum_{k\in K} x_{0jk} & = & 0\,, \qquad \label{eq:Problem2AlternativeCenterFlowBalance} \\ \sum_{i \in D} d_{i0}x_{i0k} + \sum_{j\in C}(d_{0j}x_{0jk} + d_{j0}x_{j0k}) + \sum_{j\in C} \sum_{i \in D} d_{ji} x_{jik} & \leq & L\,, \qquad \forall k \in K \label{eq:Problem2AlternativeDroneRange} \\
# Can mobility induce orders in active XY spins on a substrate? Astik Haldar<EMAIL_ADDRESS>Theory Division, Saha Institute of Nuclear Physics, HBNI, 1/AF Bidhannagar, Calcutta 700064, West Bengal, India Apurba Sarkar<EMAIL_ADDRESS>School of Mathematical & Computational Sciences, Indian Association for the Cultivation of Science, Kolkata-700032, West Bengal, India Swarnajit Chatterjee <EMAIL_ADDRESS>Center for Biophysics & Department for Theoretical Physics, Saarland University, 66123 Saarbrücken, Germany Abhik Basu<EMAIL_ADDRESS><EMAIL_ADDRESS>Theory Division, Saha Institute of Nuclear Physics, HBNI, 1/AF Bidhannagar, Calcutta 700064, West Bengal, India ###### Abstract We elucidate how the interplay between diffusive mobility and number conservation can make nearly phase-ordered active XY spins on a substrate stable. For wide-ranging model parameters, it has stable uniform phases with orientational order logarithmically stronger or weaker than in equilibrium, together with miniscule (i.e., hyperuniform) or giant number fluctuations, respectively, forming a new universality class. For other parameters, it has no stable uniformly ordered phase. Our theory sheds light on wide-ranging systems, e.g., active superfluids on substrates, synchronization of oscillators, active carpets of cilia and bacterial flagella, and sandblasting. Nonequilibrium systems can be strikingly different from their equilibrium analogs: they can display both ordered states or instabilities not found in their equilibrium analogs. For instance, a two-dimensional (2D) collection of self-propelled particles can be in orientationally long-range ordered states in the presence of finite noises Marchetti _et al._ (2013); Toner and Tu (1995), which is prohibited in equilibrium systems with continuous symmetries Mermin and Wagner (1966). Driven systems can show novel instabilities as well, e.g., driven 2D isotropic Bose systems, e.g., driven exciton polariton fluids on a substrate Carusotto and Ciuti (2013), cannot exhibit off-diagonal algebraic correlations (i.e., 2D superfluidity) on a substrate, but show only short-range order (SRO) for any weak noises Altman _et al._ (2015), unlike quasi-long-range order (QLRO) at low enough temperature ($T$) in its equilibrium counterpart Chaikin and Lubensky (1995). In this Letter, we study diffusively moving nearly phase-ordered active XY spins on a 2D solid substrate, as illustrated in Fig. 1. The “activity” of the model stems from the propensity of the spins to rotate in response to the local spin concentration and the magnitude of the local phase difference. Absence of self-propulsion of the spins makes this active model distinct from the celebrated “moving XY” Toner-Tu model Toner and Tu (1995) or the microscopic Vicsek model Vicsek _et al._ (1995) for flocks. We develop the hydrodynamic theory for this system, supplemented by a lattice-gas type microscopic agent-based model constructed by us. Surprisingly, this model encompasses both the striking features of nonequilibrium systems vis-a-vis their equilibrium counterparts, viz., novel order and instabilities in different parameter ranges, and belongs to a heretofore unstudied universality class. Figure 1: Schematic diagram of the model. Each lattice site contains an arbitrary number of XY spins, with the phase at a site being the average phases of all the spins there, marked by bluish thick arrows. The phase changes by local “active rotations” in response to the local spin concentration and magnitude of the nearest-neighbor phase differences, and by simple relaxation (thick yellow arrow). The density changes via currents having diffusive and local phase difference-dependent parts, denoted by the red broken horizontal arrow. Interestingly, this theory also describes the 2D Kardar-Parisi-Zhang (KPZ) equation Kardar _et al._ (1986); Barabási _et al._ (1995) for surface growth or erosion (“sandblasting”) with a conserved species on it. Our most striking result is that, in contrast to either non-number conserving or immobile XY spins on a substrate that admit only SRO Altman _et al._ (2015), the mobility and number conservation together in this system can lead to stable uniformly ordered phases; see supplemental movie MOV1 sup . These ordered phases are distinguished by varieties of stable orientational order for a wide range of model parameters. Although like QLRO, the variance of the orientational fluctuations grows with the system size $L$, it does so either logarithmically slower than QLRO (i.e. stronger order than QLRO or “SQLRO”), or faster than QLRO (i.e. weaker order than QLRO or “WQLRO”), depending upon the model parameters. In a surprising correspondence with phase fluctuations, the density fluctuations are either generically miniscule, or giant, respectively. Intriguingly, fluctuations, which are underdamped in these ordered states, are characterized by continuously varying scaling exponents. For other choices of the parameters, there are no stable uniform phase-ordered states. Without mobility or number conservation, our model predicts short- range phase-ordering; see supplemental movie MOV2 sup . In the equilibrium limit, the model shows QLRO phase order and normal number fluctuations. Our model provides a broad perspective concerning the impact of number fluctuations on the orientational order in wide classes of 2D driven systems with rotational invariance at low noises that exhibit QLRO in equilibrium. For instance, can an isotropic driven number-conserving mobile ordered 2D condensate exist, and if so, with what type of order? Likewise, generic conditions for synchronization in phase-locked states of mobile oscillators, having the same $U(1)$ internal symmetry as XY spins, is a question of paramount significance Pikovsky _et al._ (2003); Dörfler and Bullo (2014); Strogatz (2018); Uriu _et al._ (2010); Uriu and Morelli (2014); Zhou _et al._ (2016); Levis _et al._ (2017); Frasca _et al._ (2008); Peruani _et al._ (2010); Levis _et al._ (2017); Banerjee and Basu (2017). Other experimentally realizable systems motivating our theory include “active carpets” of cilia or bacterial flagella, modeled as active rotors grafted on the carpet and can order Uchida and Golestanian (2010), and take part in nutrient transport Mathijssen _et al._ (2018) or mucous in respiratory tract D. R. Brumley and Goldstein (2012), monolayers of spinning colloids Bruot and Cicuta (2016); Zhang _et al._ (2021), and in vitro magnetic cilia carpets et al (2020). Lastly, possible suppression of the thresholdless instability of 2D KPZ surfaces is an overarching theoretical issue, given the paradigmatic status of the KPZ equation in nonequilibrium physics Barabási _et al._ (1995). Below we outline our hydrodynamic theory that quantitatively explains indispensable role of the conserved density fluctuations to sustain order here, as dramatically manifested in movies MOV1 vis-à-vis MOV2 sup . Details can be found in the associated long paper (ALP) Haldar _et al._ (2021). Due to friction from the substrate, there is no momentum conservation, so the only conserved variable on the surface is the spin number density $c({\bf x},t)$. In addition, for nearly phase-ordered spins, the broken symmetry phase fluctuations $\theta({\bf x},t)$ about an arbitrary reference state are slow variables with relaxation rates diverging in the long wavelength limit, but the amplitude fluctuations relax fast. Therefore, $\theta({\bf x},t)$, and $c({\bf x},t)$ are the only hydrodynamic variables. For a driven system, we must write down the equations of motion by appealing to the general symmetries of the underlying microscopic dynamics (i.e., translation and rotation) and the phase-ordered state (here, arbitrariness of the reference state) and conservation laws. Additional equilibrium requirements like detailed balance do not apply to our nonequilibrium system. Retaining up to the lowest order symmetry-permitted nonlinear terms in $\theta$ and spatial gradients, the dynamical equation of $\theta({\bf x},t)$ has the form $\partial_{t}\theta=\kappa\nabla^{2}\theta+\frac{\lambda}{2}({\bm{\nabla}}\theta)^{2}+\Omega(c)+f_{\theta},$ (1) where $\Omega(c)$ is a general function of $c$. In case of fluctuating surfaces, $\theta({\bf x},t)$ is the local height with respect to an arbitrary base plane Barabási _et al._ (1995). Number density $c$ follows a conservation law with a current ${\bf J}_{c}$ given by ${\bf J}_{c}=-D{\bm{\nabla}}c-\lambda_{0}\tilde{\Omega}(c){\bm{\nabla}}\theta.$ (2) Again we have truncated up to the lowest order in $\theta$ and spatial gradients; $\tilde{\Omega}(c)$ is yet another function of $c$. While we have constructed (1) and (2) by writing down all the rotation-invariant leading order terms, each term actually carries a simple physical interpretation. The linear $\kappa\nabla^{2}\theta$ term in (1), and $D{\bm{\nabla}}c$ term in (2) are just the equilibrium spin relaxation and particle diffusion, respectively; $\kappa>0$ and $D>0$, respectively, are the spin stiffness and diffusivity. The $\lambda$\- and $\Omega(c)$-terms in (1) represent, respectively, active rotations of the spins in response to local gradients of $\theta$, and local spin concentration. The $\lambda_{0}$-term in (2) models particle currents in response to spatially nonuniform phases. Gaussian noises $f_{\theta}$ and $f_{c}$ are white and conserved noises, respectively, with zero mean, and variances $\displaystyle\langle f_{\theta}({\bf x},t)f_{\theta}(0,0)\rangle=2D_{\theta}\delta^{2}({\bf x})\delta(t),$ (3) $\displaystyle\langle f_{c}({\bf x},t)f_{c}(0,0)\rangle=2D_{c}(-\nabla^{2})\delta^{2}({\bf x})\delta(t),$ (4) consistent with $\theta$ and $c$ being, respectively, a non-conserved and a conserved variable. Writing $c({\bf x},t)=c_{0}+\delta c({\bf x},t),\,\langle\delta c\rangle=0$, we obtain $\displaystyle\frac{\partial\theta}{\partial t}=\kappa\nabla^{2}\theta+\Omega_{1}\delta c+\frac{\lambda}{2}({\bm{\nabla}}\theta)^{2}+\Omega_{2}(\delta c)^{2}+f_{\theta},$ (5) $\displaystyle\frac{\partial\delta c}{\partial t}=\lambda_{0}\tilde{\Omega}_{0}\nabla^{2}\theta+D\nabla^{2}\delta c+\lambda_{0}\tilde{\Omega}_{1}{\bm{\nabla}}\cdot(\delta c{\bm{\nabla}}\theta)+f_{c},$ (6) where we have retained the most relevant nonlinear terms in fields and gradients. Parameters $\Omega_{1}\equiv\partial\Omega/\partial c|_{c=c_{0}},\,\Omega_{2}\equiv\partial^{2}\Omega/\partial^{2}c|_{c=c_{0}},\,\tilde{\Omega}_{0}\equiv\tilde{\Omega}|_{c=c_{0}},\,\tilde{\Omega}_{1}\equiv\partial\tilde{\Omega}/\partial c|_{c=c_{0}}$. Couplings $\lambda,\Omega_{2},\lambda_{0}\tilde{\Omega}_{1}$ have arbitrary signs. For active carpets with grafted rotors, $c$ represents the concentration of nutrients or other chemical substance, moved around by the beating of cilia or bacterial flagella Uchida and Golestanian (2010); Mathijssen _et al._ (2018). Purely for symmetry reasons, Eqs. (5) and (6) describe the hydrodynamics of immobile active XY spins interacting with an incompressible binary fluid in its well-mixed phase on a substrate, where $c$ now is the binary fluid order parameter, and also chiral active hexatics on a substrate Maitra _et al._ (2020). One-dimensional versions of (5) and (6) are same as the hydrodynamic equations of a one-dimensional sedimenting crystal Lahiri and Ramaswamy (1997). It is useful to study Eqs. (5) and (6) in their linear limit first, by setting $\lambda=\Omega_{2}=\tilde{\Omega}_{1}=0$. When $\Omega_{1}\tilde{\Omega}_{0}\lambda_{0}>0$, uniform states are linearly stable. Small fluctuations around these linearly stable states travel non- dispersively, i.e., with a wavespeed independent of wavevector $k$. The correlators in the linear theory are exactly calculated conveniently in Fourier space. For instance, the phase and density autocorrelation functions show QLRO and “normal number fluctuations (NNF)”, indistinguishable from the 2D equilibrium XY model and a non-critical equilibrium system with short-range interactions. In particular, equal-time phase correlator $C^{0}_{\theta\theta}(k)\equiv\langle|\theta({\bf k},t)|^{2}\rangle_{0}\approx\frac{\overline{D}}{2\Gamma k^{2}},$ (7) in the long wavevlength limit, where $\overline{D}\equiv D_{\theta}+D_{c}$ and $\Gamma\equiv(\kappa+D)/2$; “0” refers to a linear theory result. This in turn gives for the variance $\Delta_{\theta}^{0}\equiv\langle\theta^{2}({\bf x},t)\rangle_{0}=\frac{\overline{D}}{4\pi\Gamma}\ln\left(\frac{L}{a_{0}}\right)$ (8) in 2D; $a_{0}$ is a small-scale cutoff. Equation (8) corresponds to a logarithmically rough Edward-Wilkinson (EW) Barabási _et al._ (1995) surface at 2D. Furthermore $C_{\theta\theta}^{0}(r)\equiv\langle\left[\theta({\bf x+r},t)-\theta({\bf x},t)\right]^{2}\rangle_{0}\approx\frac{\overline{D}}{2\pi\Gamma}\ln(r/a_{0})$ (9) for large $r$ Chaikin and Lubensky (1995). Equations (8) and (9) imply QLRO. The equal-time density correlator $C_{cc}^{0}(k)\equiv\langle|\delta c({\bf k},t)|^{2}\rangle_{0}\approx\frac{\overline{D}}{2\Gamma}$ (10) is independent of $k$ in the long wavelength limit. In real space, $C^{0}_{cc}(r)$ vanishes for $r\gg\zeta$, a microscopic length characterizing the short-range interactions. This further means that the standard deviation of the number fluctuations of $N$ spins contained in a fixed open area, $\sqrt{\langle N^{2}\rangle-\langle N\rangle^{2}}\equiv\sigma(N_{0})$ scales with the mean $\langle N\rangle\equiv N_{0}$ as $\sqrt{N_{0}}$, giving NNF as expected in a non-critical equilibrium system with short-range interactions. If $\Omega_{1}\tilde{\Omega}_{0}\lambda_{0}<0$, the uniform states are linearly unstable with a growth rate proportional to $k$. The Fluctuation-Dissipation-Theorem (FDT) Chaikin and Lubensky (1995) is broken in the linearized theory. This manifests in the non-vanishing cross- correlator $C_{\times}({\bf k})\equiv\langle\theta({\bf-k},t)\delta c({\bf k},t)\rangle$, which is a model parameter-dependent constant in the limit of small $k$. It now behooves us to find whether nonlinear effects are relevant (in the RG or renormalization group sense), and the scaling properties of any ordered states that are robust against finite noises. To study this, we perform one- loop perturbative RG analysis on Eqs. (5) and (6) at 2D, similar to the RG calculations on the KPZ equation Forster _et al._ (1977); Barabási _et al._ (1995), or the coupled Burgers-like equation for Magnetohydrodynamics Basu _et al._ (1999); Basu and Frey (2004, 2009). It turns out, as discussed below, that the nonlinearities either introduce logarithmic modulations to the scaling of the linearly stable states, or to destroy those states altogether. As usual, the RG is done by tracing over the short wavelength Fourier modes of the fields Forster _et al._ (1977); Barabási _et al._ (1995); Hohenberg and Halperin (1977), by expanding in the dimensionless coupling constant $g\sim\lambda^{2}\overline{D}/\Gamma^{3}$; see ALP for technical details. It predicts that not all but some of the linearly stable states are robust against noises. This is controlled by $\mu_{1}\equiv\Omega_{2}/\lambda,\,\mu_{2}\equiv\lambda_{0}\tilde{\Omega}_{1}/\lambda$, that are marginal (in the RG sense) at the one-loop order. We find that for wide ranges of $\mu_{1},\,\mu_{2}$, renormalized scale-dependent $g(k)=1/[\Delta_{1}(\mu_{1},\mu_{2})\ln(\Lambda/k)]$ flows to zero very slowly, where $\Delta_{1}>0$ as $k\rightarrow 0$ at 2D, under successive applications of the RG procedure, for stable ordered phases; $\Lambda=2\pi/a_{0}$ is an upper wavevector cutoff. This gives, as obtained from the RG flow equations, that renormalized, scale-dependent $\Gamma(k)=\Gamma[\ln(\Lambda/k)]^{\eta_{2}}$ and $\overline{D}(k)=\overline{D}[\ln(\Lambda/k)]^{\eta_{1}}$, both diverging logarithmically when $k\rightarrow 0$; $\eta_{1}(\mu_{1},\mu_{2}),\,\eta_{2}(\mu_{1},\mu_{2})>0$ are constants related to $\Delta_{1}(>0)$. These log-divergences due to the “slow” vanishing of the coupling constant are reminiscent of the logarithmic anomalous elasticity in three-dimensional equilibrium smectics Grinstein and Pelcovits (1981, 1982). The resulting renormalized theory, owing to $g(k)\rightarrow 0$ as $k\rightarrow 0$, is effectively linear, albeit with renormalized parameters. Straightforward calculations of the renormalized correlation functions show that the stable ordered states have scaling properties essentially same as in the linear theory, modulated only by logarithmic corrections, which either strengthen or weaken the linear theory order. For instance, the renormalized phase correlator $C_{\theta\theta}^{R}(k)$ reads $C^{R}_{\theta\theta}(k)\approx\frac{\overline{D}}{2\Gamma k^{2}[\ln(\Lambda/k)]^{\eta}},$ (11) for $k\rightarrow 0$; $\eta\equiv\eta_{2}-\eta_{1}$. Here and below, $R$ refers to renormalized quantities. Exponent $\eta$ varies continuously with $\mu_{1},\,\mu_{2}$ and can be positive or negative; detailed calculations show $\eta<1/3$ always Haldar _et al._ (2021). For $\eta>(<)0$, $C^{R}_{\theta\theta}(k)\ll(\gg)C^{0}_{\theta\theta}(k)$ as $k\rightarrow 0$, demonstrating strong suppression (enhancement) of fluctuations in the long wavelength limit. Next, the renormalized variance $\Delta_{\theta}^{R}\equiv\langle\theta^{2}({\bf x},t)\rangle_{R}$ now acquires a novel $L$-dependence: $\Delta_{\theta}^{R}\approx\frac{\overline{D}}{4\pi\Gamma}\left[\ln\left(\frac{L}{a_{0}}\right)\right]^{1-\eta},$ (12) This means $\Delta_{\theta}^{R}$ can grow, respectively, logarithmically slower or faster with $\ln L$ than QLRO, representing order stronger or weaker than QLRO, named, respectively, SQLRO, or WQLRO; for $\eta=0$, QLRO is retrieved. Equation (12) also implies a surface logarithmically smoother or rougher than the 2D EW surface. Likewise, the renormalized correlator $C_{\theta\theta}^{R}(r)$, related to the inverse Fourier transform of $C_{\theta\theta}^{R}(k)$, scales as $C_{\theta\theta}^{R}(r)\equiv\langle[\theta({\bf x+r},t)-\theta({\bf x},t)]^{2}\rangle_{R}\approx\frac{\overline{D}}{2\pi\Gamma}[\ln(r/a_{0})]^{1-\eta},$ (13) for large $r$. Related to $C^{R}_{\theta\theta}(r)$, the renormalized spin correlation function $C_{ZZ}^{R}(r)$ for large $r$ is $C^{R}_{ZZ}(r)\equiv\langle\cos[\theta({\bf x+r},t)-\theta({\bf x},t)]\rangle_{R}\approx(r/a_{0})^{-\tilde{\gamma}(r)}$ (14) where the $r$-dependent exponent $\tilde{\gamma}(r)$ has a complex form: $\tilde{\gamma}(r)\equiv\frac{\overline{D}}{4\pi\Gamma}[\ln(r/a_{0})]^{-\eta}$ (15) for large $r$. For $\eta>0(<0)$, clearly $C_{ZZ}(r)$ decays much slower (faster) for large $r$, giving SQLRO (WQLRO). Next in contrast to its linear theory analog, renormalized density correlator $C_{cc}^{R}(k)$ picks up a weak $k$-dependence: $C_{cc}^{R}(k)\approx\frac{\overline{D}}{2\Gamma}\left[\ln\left(\frac{\Lambda}{k}\right)\right]^{-\eta}$ (16) in the hydrodynamic limit $k\rightarrow 0$. Evidently, for $\eta>0$, the density fluctuations are strongly suppressed in the limit $k\rightarrow 0$, implying miniscule number fluctuations (MNF) or hyperuniformity Torquato and Stillinger (2003) vis-a-vis for $\eta=0$, same as the equilibrium result. In contrast, if $\eta<0$ the density fluctuations are hugely enhanced when $k\rightarrow 0$. This is giant number fluctuations (GNF), often encountered in orientationally ordered active fluids Marchetti _et al._ (2013), and also in equilibrium superfluids Chaikin and Lubensky (1995). Lastly, the equal-time renormalized density autocorrelator $C_{cc}^{R}(r)$, the inverse Fourier transform of $C_{cc}^{R}(k)$, is $C^{R}_{cc}(r)\approx\frac{\overline{D}}{4\pi\Gamma}\frac{-\eta}{r^{2}[\ln(r/a_{0})]^{(1+\eta)}},$ (17) for large $r$, $r/a_{0}\gg 1$. Thus for $\eta>(<)0$, $C^{R}_{cc}(r)$ falls off relatively faster (slower) for MNF (GNF). By using (17) we re-express MNF (GNF) for $\eta>(<)0$: $\sigma(N_{0})\propto\sqrt{N_{0}/(\ln N_{0})^{\eta}}<(>)\sqrt{N_{0}},$ (18) in the renormalized theory. For $\eta=0$, (15) and (18) reduce to the well- known results for the 2D equilibrium XY model and an equilibrium system with NNF having short-range interactions away from any critical point. We thus show SQLRO (WQLRO) phase order is necessarily accompanied by MNF (GNF). Like $C_{cc}^{R}(k)$, renormalized cross-correlation function $C_{\times}^{R}(k)$ also picks up a weak $k$-dependence: $C_{\times}^{R}(k)\propto\left[\ln(\frac{\Lambda}{k})\right]^{(1-3\eta)/2}\times{\cal O}(1)$ for small $k$. Thus FDT remains broken in the renormalized theory. The model shows breakdown of conventional dynamic scaling in the ordered phases: the form of $\Gamma(k)$ in the renormalized theory shows that the diffusive scaling of time $t$ with $r$ is modulated by logarithmic corrections as $t\propto r^{2}/[\ln(r/a_{0})]^{(1-\eta)/2}$ (19) for large $r$. Thus the fluctuations relax logarithmically faster than ordinary diffusion or in the linear theory. Also, they relax faster with WQLRO ($\eta<0$) than SQLRO ($\eta>0$). Equations (11)-(19) collectively define the new universality class, characterized by the SQLRO-MNF and WQLRO-GNF correspondence and continuously varying scaling exponents. For other choices of $\mu_{1},\,\mu_{2}$, uniform ordered states get unstable due to nonlinear effects. In the equilibrium limit, the nonlinearities in (5) and (6) vanish, and QLRO phase-order together with density NNF follows. A phase diagram of the model in the $\mu_{1}-\mu_{2}$ plane over limited ranges of $\mu_{1},\,\mu_{2}$ is shown in Fig. 2(a). Numerical results from our agent-based model indeed confirm the existence of stable order with SQLRO/MNF and WQLRO/GNF as generic, holding beyond any perturbative calculations, and that topological defects in the active XY model (not included in the hydrodynamic theory) do not proliferate and destroy order at low-enough non-zero noises. It consists of $N_{\text{tot}}$ XY spins on a 2D square lattice of size $L\times L$, with periodic boundary conditions and a mean spin number density $c_{0}\equiv N_{\text{tot}}/L^{2}$. The microscopic update rules for $\theta$ depend on $c$ and the lattice analogs of the phase difference-dependent active rotations of the spins along with equilibrium relaxation processes; $c$ changes via hopping to the nearest neighbor sites, provided either the density or the phase in the randomly chosen neighboring target site is different. Stochasticity is introduced in the update of $\theta$ by adding a (small) white noise $\xi$ to it Vicsek _et al._ (1995), and to $c$ through hopping of particles in randomly selected sites to randomly chosen nearest neighbors; see ALP for details. Our numerical studies reveal stable ordered states at finite noise $\xi$; see movie MOV1 sup . Furthermore, in these ordered states, depending upon certain parameters $g_{1},\,g_{2}$, which are the substitutes of $\mu_{2},\,\mu_{1}$ here (see ALP for precise correspondence) and a few other parameters that define the relative probabilities of the various microscopic dynamical processes Haldar _et al._ (2021), numerical analogs of $\Delta_{\theta}^{R}(L)$ can grow slower and faster than $\ln(L/a_{0})$, implying SQLRO and WQLRO respectively. In addition, $\sigma(N_{0})$ as a function of $\ln\,N_{0}$ in a given open area grows logarithmically slower and faster than $\sqrt{N_{0}}$, showing MNF and GNF, correspondingly with SQLRO and WQLRO, respectively. Generic dependence of $\eta$ on $g_{1},\,g_{2}$ are also found; see ALP Haldar _et al._ (2021). Two representative plots from the simulations of the agent-based model in Fig. 2(b) and (c) show the $\ln L$-dependence of $\langle\theta^{2}\rangle$ with SQLRO and WQLRO, and correspondingly, $\ln\,N_{0}$-dependence of $\sigma(N_{0})$ displaying MNF and GNF, respectively. These establish the SQLRO (WQLRO) and MNF (GNF) correspondence in the agent-based model, in agreement with the hydrodynamic theory 111The quantitative accuracy of these plots is restricted due to the limited range of $L$ used in the simulations.. Figure 2: (a) Schematic phase diagram in the $\mu_{1}-\mu_{2}$ plane in the linearly stable hydrodynamic theory. SQLRO/MNF, WQLRO/GNF and disordered regions are marked in different colors. QLRO/NNF is the (blue) line demarcating SQLRO/MNF and WQLRO/GNF regions. Plots from agent-based simulations: (b) $\langle\theta^{2}\rangle/\ln L$ versus $\ln L$ showing SQLRO and WQLRO, and (c) $\sigma(N_{0})/\sqrt{N_{0}}$ versus $\ln\,N_{0}$ showing MNF and GNF ($L=128$) with $c_{0}=5,\,\xi=0.1$ and (SQLRO/MNF): $g_{1}=1.0,\,g_{2}=0.02$, and (WQLRO/GNF): $g_{1}=2.0,\,g_{2}=0.03$, respectively. The red broken horizontal lines in (b) and (c) correspond to the equilibrium results. The SQLRO-MNF and WQLRO-GNF correspondences are clearly established in the agent-based model (see text). The simulations also detect two kinds of disordered states, corresponding to the two different routes to disorder (linear instability and nonlinear effects) in the hydrodynamic theory; see ALP Haldar _et al._ (2021) for details. For immobile spins with uniform density, $\delta c=0$, or if there is no number conservation, e.g., if the spins are created and destroyed, $\delta c$ relaxes fast, and is a non-hydrodynamic variable. Then Eq. (5) reduces to the KPZ equation with only SRO in 2D Kardar _et al._ (1986); Barabási _et al._ (1995); Altman _et al._ (2015), in agreement with our numerical studies in its immobile limit (movie MOV2 sup ). Heuristically, fluctuations of the conserved density $c$ create an effective long-range interactions between phases at distant locations, which for appropriately chosen parameters can suppress the instability of the 2D KPZ equation, ultimately sustaining order. We have, therefore, presented the hydrodynamic theory of phase-ordering in diffusively mobile active XY model on a substrate, supplemented by numerical studies of an equivalent agent-based model. Existence of stable SQLRO and WQLRO phase orders, correspondingly with MNF and GNF, respectively, uncovering a novel correspondence, are predicted in wide ranges of the parameter space. This is our central conclusion. Absence of ordered states elsewhere in the parameter space is also revealed. Acknowledgement:- AH and AB thank T. Banerjee and N. Sarkar for critical reading and helpful suggestions. AS thanks University Grants Commission (UGC), India for a research fellowship. SC is financially supported by the German Research Foundation (DFG) within the Collaborative Research Center SFB 1027 and Indian Association for the Cultivation of Science, Kolkata. AS and SC also acknowledges Prof. R. Paul (IACS, Kolkata) for some invaluable suggestions and discussions and are also thankful to him for providing the computational resources. AB thanks J. Toner for valuable discussions in the early stages of this work, and comments concerning derivations of Eqs. (17) and (18), and P. K. Mohanty, A. Maitra and D. Levis for helpful comments, and the SERB, DST (India) for partial financial support through the MATRICS scheme [file no.: MTR/2020/000406]. AB has designed the problem, AH has contributed to the analytical part of the work, AS and SC have contributed equally to the numerical part of the work. All four authors jointly wrote the manuscript. ## References * Marchetti _et al._ (2013) M. C. Marchetti, J. F. Joanny, S. Ramaswamy, T. B. Liverpool, J. Prost, M. Rao, and R. A. Simha, “Hydrodynamics of soft active matter,” Rev. Mod. Phys. 85, 1143–1189 (2013). * Toner and Tu (1995) J. Toner and Y. Tu, “Long-Range Order in a Two-Dimensional Dynamical $\mathrm{XY}$ Model: How Birds Fly Together,” Phys. Rev. Lett. 75, 4326–4329 (1995). * Mermin and Wagner (1966) N. D. Mermin and H. Wagner, “Absence of ferromagnetism or antiferromagnetism in one-or two-dimensional isotropic Heisenberg models,” Phys. Rev. Lett. 17, 1133 (1966). * Carusotto and Ciuti (2013) I. Carusotto and C. Ciuti, “Quantum fluids of light,” Rev. Mod. Phys. 85, 299–366 (2013). * Altman _et al._ (2015) E. Altman, L. M. Sieberer, L. Chen, S. Diehl, and J. Toner, “Two-dimensional superfluidity of exciton polaritons requires strong anisotropy,” Phys. Rev. X 5, 011017 (2015). * Chaikin and Lubensky (1995) P. M. Chaikin and T. C. Lubensky, _Principles of condensed matter physics_ (Cambridge university press Cambridge, 1995). * Vicsek _et al._ (1995) T. Vicsek, A. Czirók, E. Ben-Jacob, I. Cohen, and O. Shochet, “Novel type of phase transition in a system of self-driven particles,” Phys. Rev. Lett. 75, 1226 (1995). * Kardar _et al._ (1986) M. Kardar, G. Parisi, and Y.-C. Zhang, “Dynamic Scaling of Growing Interfaces,” Phys. Rev. Lett. 56, 889–892 (1986). * Barabási _et al._ (1995) A-L Barabási, H. E Stanley, _et al._ , _Fractal concepts in surface growth_ (Cambridge university press, 1995). * (10) “See Supplemental Material for movies showing the time evolution of the spin configurations in a typical ordered phase (MOV1) and disordered phase without mobility (MOV2), respectively, of the model,” . * Pikovsky _et al._ (2003) A. Pikovsky, J. Kurths, M. Rosenblum, and J. Kurths, _Synchronization: a universal concept in nonlinear sciences_ , 12 (Cambridge university press, 2003). * Dörfler and Bullo (2014) F. Dörfler and F. Bullo, “Synchronization in complex networks of phase oscillators: A survey,” Automatica 50, 1539–1564 (2014). * Strogatz (2018) S. H Strogatz, _Nonlinear dynamics and chaos with student solutions manual: With applications to physics, biology, chemistry, and engineering_ (CRC press, 2018). * Uriu _et al._ (2010) K. Uriu, Y. Morishita, and Y. Iwasa, “Random cell movement promotes synchronization of the segmentation clock,” Proc. Natl. Acad. Sci. U.S.A. 107, 4979–4984 (2010). * Uriu and Morelli (2014) K. Uriu and L. G. Morelli, “Collective cell movement promotes synchronization of coupled genetic oscillators,” Biophys. J. 107, 514–526 (2014). * Zhou _et al._ (2016) J. Zhou, Y. Zou, S. Guan, Z. Liu, and S. Boccaletti, “Synchronization in slowly switching networks of coupled oscillators,” Sc. Rep. 6, 35979 (2016). * Levis _et al._ (2017) Demian Levis, Ignacio Pagonabarraga, and Albert Díaz-Guilera, “Synchronization in Dynamical Networks of Locally Coupled Self-Propelled Oscillators,” Phys. Rev. X 7, 011028 (2017). * Frasca _et al._ (2008) M. Frasca, A. Buscarino, A. Rizzo, L. Fortuna, and S. Boccaletti, “Synchronization of moving chaotic agents,” Phys. Rev. Lett. 100, 044102 (2008). * Peruani _et al._ (2010) F. Peruani, E. M Nicola, and L. G. Morelli, “Mobility induces global synchronization of oscillators in periodic extended systems,” New J. Phys. 12, 093029 (2010). * Banerjee and Basu (2017) T. Banerjee and A. Basu, “Active hydrodynamics of synchronization and ordering in moving oscillators,” Phys. Rev. E 96, 022201 (2017). * Uchida and Golestanian (2010) N. Uchida and R. Golestanian, “Synchronization and Collective Dynamics in a Carpet of Microfluidic Rotors,” Phys. Rev. Lett. 104, 178103 (2010). * Mathijssen _et al._ (2018) A. J. T. M. Mathijssen, F. Guzmán-Lastra, A. Kaiser, and H. Löwen, “Nutrient Transport Driven by Microbial Active Carpets,” Phys. Rev. Lett. 121, 248101 (2018). * D. R. Brumley and Goldstein (2012) T. J. Pedley D. R. Brumley, M. Polin and R. E. Goldstein, “Hydrodynamic Synchronization and Metachronal Waves on the Surface of the Colonial Alga Volvox carteri,” Phys. Rev. Lett. 109, 268102 (2012). * Bruot and Cicuta (2016) N. Bruot and P. Cicuta, “Realizing the Physics of Motile Cilia Synchronization with Driven Colloids,” Annu. Rev. Condens. Matter Phys 7, 323–348 (2016). * Zhang _et al._ (2021) Z. Zhang, H. Yuan, Y. Dou, M. O. de la Cruz, and K. J. M. Bishop, “Quincke Oscillations of Colloids at Planar Electrodes,” Phys. Rev. Lett. 126, 258001 (2021). * et al (2020) H. Gu et al, “Magnetic cilia carpets with programmable metachronal waves,” Nat. Comm. 11, 2637 (2020). * Haldar _et al._ (2021) A. Haldar, A. Sarkar, S. Chatterjee, and A. Basu, “Active XY model on a substrate: Density fluctuations and phase ordering,” companion long paper; ArXiv: 2105.03919 (2021). * Maitra _et al._ (2020) A. Maitra, M. Lenz, and R. Voituriez, “Chiral Active Hexatics: Giant Number Fluctuations, Waves, and Destruction of Order,” Phys. Rev. Lett. 125, 238005 (2020). * Lahiri and Ramaswamy (1997) R. Lahiri and S. Ramaswamy, “Are Steadily Moving Crystals Unstable?” Phys. Rev. Lett. 79, 1150–1153 (1997). * Forster _et al._ (1977) D. Forster, D. R Nelson, and M. J Stephen, “Large-distance and long-time properties of a randomly stirred fluid,” Phys. Rev. A 16, 732 (1977). * Basu _et al._ (1999) A. Basu, J. K Bhattacharjee, and S. Ramaswamy, “Mean magnetic field and noise cross-correlation in magnetohydrodynamic turbulence: results from a one-dimensional model,” Eur. Phys. J. B 9, 725–730 (1999). * Basu and Frey (2004) A. Basu and E. Frey, “Novel universality classes of coupled driven diffusive systems,” Phys. Rev. E 69, 015101 (2004). * Basu and Frey (2009) A. Basu and E. Frey, “Scaling and universality in coupled driven diffusive models,” J. Stat. Mech.: Theory Exp. 2009, P08013 (2009). * Hohenberg and Halperin (1977) P. C. Hohenberg and B. I. Halperin, “Theory of dynamic critical phenomena,” Rev. Mod. Phys. 49, 435 (1977). * Grinstein and Pelcovits (1981) G. Grinstein and R. A. Pelcovits, “Anharmonic Effects in Bulk Smectic Liquid Crystals and Other ”One-Dimensional Solids”,” Phys. Rev. Lett. 47, 856–859 (1981). * Grinstein and Pelcovits (1982) G. Grinstein and R. A. Pelcovits, “Smectic-$A-C$ transition in three dimensions,” Phys. Rev. A 26, 2196–2217 (1982). * Torquato and Stillinger (2003) S. Torquato and F. H. Stillinger, “Local density fluctuations, hyperuniformity, and order metrics,” Phys. Rev. E 68, 041113 (2003). * Note (1) The quantitative accuracy of these plots is restricted due to the limited range of $L$ used in the simulations.
# A Zeroth-Order Momentum Method for Risk-Averse Online Convex Games Zifan Wang, Yi Shen, Zachary I. Bell, Scott Nivison, Michael M. Zavlanos, and Karl H. Johansson *This work is supported in part by AFOSR under award #FA9550-19-1-0169 and by NSF under award CNS-1932011.Zifan Wang and Karl H. Johansson are with Division of Decision and Control Systems, School of Electrical Enginnering and Computer Science, KTH Royal Institute of Technology, and also with Digital Futures, SE-10044 Stockholm, Sweden. Email: <EMAIL_ADDRESS>Shen and Michael M. Zavlanos are with the Department of Mechanical Engineering and Materials Science, Duke University, Durham, NC, USA. Email: {yi.shen478<EMAIL_ADDRESS>I. Bell and Scott Nivison are with the Air Force Research Laboratory, Eglin AFB, FL, USA. Email: {scott.nivison<EMAIL_ADDRESS> ###### Abstract We consider risk-averse learning in repeated unknown games where the goal of the agents is to minimize their individual risk of incurring significantly high cost. Specifically, the agents use the conditional value at risk (CVaR) as a risk measure and rely on bandit feedback in the form of the cost values of the selected actions at every episode to estimate their CVaR values and update their actions. A major challenge in using bandit feedback to estimate CVaR is that the agents can only access their own cost values, which, however, depend on the actions of all agents. To address this challenge, we propose a new risk-averse learning algorithm with momentum that utilizes the full historical information on the cost values. We show that this algorithm achieves sub-linear regret and matches the best known algorithms in the literature. We provide numerical experiments for a Cournot game that show that our method outperforms existing methods. ## I Introduction In online convex games [1, 2], agents interact with each other in the same environment in order to minimize their individual cost functions. Many applications, including traffic routing [3] and online marketing [4], among many others [5], can be modeled as online convex games. The individual cost functions depend on the joint decisions of all agents and are typically unknown but can be accessed via querying. Using this information, the agents can sequentially select the best actions that minimize their expected accumulated costs. This can be done using online learning algorithms whose performance is typically measured using the notions of regret [5] that quantify the gap between the agents’ online decisions and the best decisions in hindsight. Many such online learning algorithms have been recently proposed, and have been shown to achieve sub-linear regret, which indicates that optimal decisions are eventually selected; see e.g., [1, 2, 6]. In this paper, here we focus on online convex games for high-stakes applications, where decisions that minimize the expected cost functions are not necessarily desirable. For example, in portfolio management, a selection of assets with the highest expected return is not necessarily desirable since it may have high volatility that can lead to catastrophic losses. To capture unexpected and catastrophic outcomes, risk-averse criteria have been widely proposed in place of the expectation, such as the Sharpe Ratio [7] and Conditional Value at Risk (CVaR) [8]. In this paper, we consider online convex games with risk-averse agents that employ CVaR as risk measure. We assume that only cost function evaluations at selected decisions are accessible. Despite the practical utility of CVaR as a measure of risk in high-stakes applications, the theoretical analysis of risk- averse learning methods that employ the CVaR is limited, mainly since the CVaR values must be estimated from the distributions of the unknown cost functions. To avoid this problem, [9] proposes a variational definition of CVaR that allows to reformulate the computation of CVaR values into an optimization problem by introducing additional variables. Building on this reformulation, many risk-averse learning methods have been proposed, including [10, 11, 12], and their convergence rates have been analysed. Specifically, [10] provides performance guarantees for stochastic gradient descent-type algorithms in risk-averse statistical learning problems; [11] addresses risk-averse online convex optimization problems with bandit feedback; and [12] proposes an adaptive sampling strategy for CVaR learning, and transforms the supervised learning problem into a zero-sum game between the sampler and the learner. However, all the above works focus on single-agent learning problems. To the best of our knowledge, multi-agent risk-averse learning problems have not been analyzed in the literature, with only a few exceptions [13], [14]. The work in [13] solves multi-armed bandit problems with finite actions, which are different from the convex games with continuous actions considered here. Most closely related to this paper is our recent work in [14] proposing a new risk-averse learning algorithm for online convex games, where the agents utilize the cost evaluations at each time step to update their actions and achieve no-regret learning with high probability. To improve the learning rate of this algorithm, [14] also proposes two variance reduction methods on the CVaR estimates and gradient estimates, respectively. In this paper, we build on this work to significantly improve the performance of the algorithm using a notion of momentum that combines these variance reduction techniques. Specifically, the proposed new algorithm reuses samples of the cost functions from the full history of data while appropriately decreasing the weights of out-of-date samples. The idea is that, using more samples, the unknown distribution of the cost functions needed to estimate the CVaR values can be better estimated. To do so, at each time step, the proposed algorithm estimates the distribution of the cost functions using both the immediate empirical distribution estimate and the distribution estimate from the previous time step, which summarizes all past cost values. Then, to construct the gradient estimates, more weight is allocated to more recent information, similar to momentum gradient descent methods [15]. To reduce the variance of the gradient estimate, we employ residual feedback [16], which uses previous gradient estimates to update the current one. Although [14] showed that reusing samples from the last iteration or residual feedback can individually improve the performance of the algorithm, it also identified that the combination of sample reuse and residual feedback is an unexplored and theoretically nontrivial problem. In this work, we combine sample reuse and residual feedback by proposing a novel zeroth-order momentum method that reuses all historical samples, compared to samples from only the last iteration. An additional important contribution of this work is that the proposed zeroth-order momentum method subsumes Algorithm 3 in [14] as a special case by appropriately choosing the momentum parameter. We show that our proposed method theoretically matches the best known result in [14] but empirically outperforms other existing methods. The rest of the paper is organized as follows. In Section II, we formally define the problem and provide some preliminary results. In Section III, we summarize results in [14] that are needed to develop our proposed method. In Section IV, we present our proposed method and analyze its regret. In Section V, we numerically verify our method using a Cournot game example. Finally, in Section VI, we conclude the paper. ## II Problem definition We consider a repeated game $\mathcal{G}$ with $N$ risk-averse agents. Each agent selects an action $x_{i}$ from the convex set $\mathcal{X}_{i}\subseteq\mathbb{R}^{d_{i}}$, and then receives a stochastic cost $J_{i}(x_{i},x_{-i},\xi_{i}):\mathcal{X}\times\Xi_{i}\rightarrow\mathbb{R}$, where $x_{-i}$ represents all agents’ actions except for agent $i$, and $\mathcal{X}=\Pi_{i=1}^{N}\mathcal{X}_{i}$ is the joint action space. We sometimes instead write $J_{i}(x,\xi_{i})$ for ease of notation, where $x=(x_{i},x_{-i})$ is the concatenated vector of all agents’ actions. The cost $J_{i}(x_{i},x_{-i},\xi_{i})$ is stochastic, as $\xi_{i}\in\Xi_{i}$ is a stochastic variable. We assume that the diameter of the convex set $\mathcal{X}_{i}$ is bounded by $D_{x}$ for all $i=1,\ldots,N$. In what follows, we make the following assumptions on the class of games we consider in this paper. ###### Assumption 1. The function $J_{i}(x_{i},x_{-i},\xi_{i})$ is convex in $x_{i}$ for every $\xi_{i}\in\Xi_{i}$ and bounded by $U$, i.e., $|J_{i}(x,\xi_{i})|\leq U$, for all $i=1,\ldots,N$. ###### Assumption 2. $J_{i}(x,\xi_{i})$ is $L_{0}$-Lipschitz continuous in $x$ for every $\xi_{i}\in\Xi_{i}$, for all $i=1,\ldots,N$. Assumptions 1 and 2 are common assumptions in the analysis of online learning in games [17]. The goal of the risk-averse agents is to minimize the risk of incurring significantly high costs, for possibly different risk levels. In this work, we utilize CVaR as the risk measure. For a given risk level $\alpha_{i}\in[0,1]$, CVaR is defined as the average of the $\alpha_{i}$ percent worst-case cost. Specifically, we denote by $F_{i}^{x}(y)=\mathbb{P}\\{J_{i}(x,\xi_{i})\leq y\\}$ as the cumulative distribution function (CDF) of the random cost $J_{i}(x,\xi_{i})$ for agent $i$; and by $J^{\alpha_{i}}$ the cost value at the $1-\alpha_{i}$ quantile of the distribution, also called the Value at Risk (VaR). Then, for a given risk level $\alpha_{i}\in[0,1]$, CVaR of the cost function $J_{i}(x,\xi_{i})$ of agent $i$ is defined as $\displaystyle C_{i}(x):$ $\displaystyle={\rm{CVaR}}_{\alpha_{i}}[J_{i}(x,\xi_{i})]$ $\displaystyle:$ $\displaystyle=\mathbb{E}_{F}[J_{i}(x,\xi_{i})|J_{i}(x,\xi_{i})\geq J^{\alpha_{i}}].$ Notice that the CVaR value is determined by the distribution function $F_{i}^{x}(y)$ for a given $\alpha_{i}$, so we sometimes write CVaR as a function of the distribution function, i.e., ${\rm{CVaR}}_{\alpha_{i}}[F_{i}^{x}]:={\rm{CVaR}}_{\alpha_{i}}[J_{i}(x,\xi_{i})]$. In addition, we assume that the agents cannot observe other agents’ actions, but the only information they can observe is the cost evaluations of the jointly selected actions at each time step. Given Assumptions 1 and 2, the following lemma lists properties of CVaR, that will be important in the analysis that follows. The proof can be found in [11]. ###### Lemma 1. Given Assumptions 1 and 2, we have that $C_{i}(x_{i},x_{-i})$ is convex in $x_{i}$ and $L_{0}$-Lipschitz continuous in $x$, for all $i=1,\ldots,N$. The following additional assumption on the variation of the CDFs is needed for the analysis that follows. ###### Assumption 3. Let $F^{w}_{i}(y)=\mathbb{P}\\{J_{i}(w,\xi_{i})\leq y\\}$ and $F^{v}_{i}(y)=\mathbb{P}\\{J_{i}(v,\xi_{i})\leq y\\}$. There exist a constant $C_{0}>0$ such that $\displaystyle\mathop{\rm{sup}}_{y}|F^{w}_{i}(y)-F^{v}_{i}(y)|\leq C_{0}\left\|w-v\right\|.$ Assumption 3 states that the difference between two CDFs can be bounded by the difference between the corresponding two action profiles. It is less restrictive than Assumption 3 in [14] since it does not depend on the algorithm but only on the cost functions. A common measure of the ability of the agents to learn online optimal decisions that minimize their individual risk-averse objective functions $C_{i}(x)$ is the algorithm regret, which is defined as the difference between the actual rewards and best rewards that the agent could have achieved by playing the single best action in hindsight. Suppose the action sequences of agent $i$ and the other agents in the team are $\\{\hat{x}_{i,t}\\}_{t=1}^{T}$ and $\\{\hat{x}_{-i,t}\\}_{t=1}^{T}$, respectively. Then, we define the regret of agent $i$ as $\displaystyle{\rm{R}}_{C_{i}}(T)=\sum_{t=1}^{T}C_{i}(\hat{x}_{i,t},\hat{x}_{-i,t})-\mathop{\rm{min}}_{\tilde{x}_{i}\in\mathcal{X}_{i}}\sum_{t=1}^{T}C_{i}(\tilde{x}_{i},\hat{x}_{-i,t}).$ An algorithm is said to be no-regret if its regret grows sub-linearly with the number of episodes $T$. In this paper, we propose a no-regret learning algorithm, so that $\lim_{T\rightarrow\infty}\frac{{\rm{R}}_{C_{i}}(T)}{T}=0$, $i=1,\ldots,N$. ## III Preliminary Results In this section, we summarize some results from [14] that lay the foundation for the subsequent analysis. Note that the CVaR values depend on the distributions of the cost functions, which are generally unknown. To estimate the distribution of the cost functions, and subsequently the CVaR values with only a few samples, a sampling strategy is proposed in [14] that uses a decreasing number of samples with the number of iterations. Specifically, at episode $t$, the sampling strategy used by the agents is designed as follows $\displaystyle n_{t}=\lceil bU^{2}(T-t+1)^{a}\rceil,$ (1) where $\lceil\cdot\rceil$ is the regularized ceiling function, $T$ is the number of episodes, $U$ is the bound of $J_{i}$ as in Assumption 1, and $a,b\in(0,1)$ are parameters to be selected later. Using the cost evaluations at each episode, zeroth-order methods can be employed to estimate the CVaR gradients and then update the agents’ actions. Specifically, at episode $t$, the agents calculate the number of samples $n_{t}$ according to (1). Then, each agent perturbs the current action $x_{i,t}$ by an amount $\delta u_{i,t}$, where $u_{i,t}\in\mathbb{S}^{d_{i}}$ is a random perturbation direction sampled from the unit sphere $\mathbb{S}^{d_{i}}\subset\mathbb{R}^{d_{i}}$ and $\delta$ is the size of this perturbation. Next the agents play their perturbed actions $\hat{x}_{i,t}=x_{i,t}+\delta u_{i,t}$ for $n_{t}$ times, and obtain $n_{t}$ cost evaluations which are utilized to update their actions. To facilitate the theoretical analysis, we define the $\delta$-smoothed function $C_{i}^{\delta}(x)=\mathbb{E}_{w_{i}\sim\mathbb{B}_{i},u_{-i}\sim\mathbb{S}_{-i}}[C_{i}(x_{i}+\delta w_{i},x_{-i}+\delta u_{-i})]$, where $\mathbb{S}_{-i}=\Pi_{j\neq i}\mathbb{S}_{j}$, and $\mathbb{B}_{i}$, $\mathbb{S}_{i}$ denote the unit ball and unit sphere in $\mathbb{R}^{d_{i}}$, respectively. The size of the perturbation $\delta$ here is related to a smoothing parameter that controls how well $C_{i}^{\delta}(x)$ approximates $C_{i}(x)$. As shown in [14], the function $C_{i}^{\delta}(x)$ satisfies the following properties. ###### Lemma 2. Let Assumptions 1 and 2 hold. Then we have that 1. 1. $C_{i}^{\delta}(x_{i},x_{-i})$ is convex in $x_{i}$, 2. 2. $C_{i}^{\delta}(x)$ is $L_{0}$-Lipschitz continuous in $x$, 3. 3. $|C_{i}^{\delta}(x)$-$C_{i}(x)|\leq\delta L_{0}\sqrt{N}$, 4. 4. $\mathbb{E}[\frac{d_{i}}{\delta}C_{i}(\hat{x}_{t})u_{i,t}]=\nabla_{i}C_{i}^{\delta}(x_{t})$. The last property in Lemma 2 shows that the term $\frac{d_{i}}{\delta}C_{i}(\hat{x}_{t})u_{i,t}$ is an unbiased estimate of the gradient of the smoothed function $C_{i}^{\delta}(x)$. Next we present a lemma that helps bound the distance between two CVaR values by the distance between the two corresponding CDFs. The proof can be found in [14]. ###### Lemma 3. Let $F$ and $G$ be two CDFs of two random variables and suppose the random variables are bounded by $U$. Then we have that $\displaystyle|{\rm{CVaR}}_{\alpha}[F]-{\rm{CVaR}}_{\alpha}[G]|\leq\frac{U}{\alpha}\mathop{\rm{sup}}_{y}|F(y)-G(y)|.$ ## IV A zeroth-order momentum method In this section, we present a new one-point zeroth-order momentum method that combines sample reuse as in Algorithm 2 in [14] and residual feedback as in Algorithm 3 in [14], but unlike Algorithm 2 in [14] that reuses samples only from the last iteration, it uses all past samples to update the agents’ actions. The new algorithm is given as Algorithm 1. Algorithm 1 Risk-averse learning with momentum 0: Initial value $x_{0}$, step size $\eta$, parameters $a$, $b$, $\delta$, $T$, risk level $\alpha_{i}$, $i=1,\cdots,N$. 1: for ${\rm{episode}}\;t=1,\ldots,T$ do 2: Select $n_{t}=\lceil bU^{2}(T-t+1)^{a}\rceil$ 3: Each agent samples $u_{i,t}\in\mathbb{S}^{d_{i}}$, $i=1,\ldots,N$ 4: Each agent play $\hat{x}_{i,t}=x_{i,t}+\delta u_{i,t}$, $i=1,\ldots,N$ 5: for $j=1,\ldots,n_{t}$ do 6: Let all agents play $\hat{x}_{i,t}$ 7: Obtain $J_{i}(\hat{x}_{i,t},\hat{x}_{-i,t},\xi_{i}^{j})$ 8: end for 9: for agent $i=1,\ldots,N$ do 10: Build EDF $\bar{F}_{i,t}(y)$ 11: Estimate CVaR: ${\rm{CVaR}}_{\alpha_{i}}[\bar{F}_{i,t}]$ 12: Construct gradient estimate$\bar{g}_{i,t}=\frac{d_{i}}{\delta}\left({\rm{CVaR}}_{\alpha_{i}}[\bar{F}_{i,t}]-{\rm{CVaR}}_{\alpha_{i}}[\bar{F}_{i,t-1}]\right)u_{i,t}$ 13: Update $x$: $x_{i,t+1}\leftarrow\mathcal{P}_{\mathcal{X}_{i}^{\delta}}(x_{i,t}-\eta\bar{g}_{i,t})$ 14: end for 15: end for Using the sampling strategy in (1), each agent plays the perturbed action $\hat{x}_{i,t}$ for $n_{t}$ times and obtains $n_{t}$ samples at episode $t$. For agent $i$, we denote the CDF of the random cost $J_{i}(\hat{x}_{t},\xi_{i})$ that is returned by the perturbed action $\hat{x}_{t}$ as $F_{i,t}^{\hat{x}_{t}}(y)=\mathbb{P}\\{J_{i}(\hat{x}_{t},\xi_{i})\leq y\\}$. Since $\hat{x}_{t}$ depends on $t$, we write $F_{i,t}$ for $F_{i,t}^{\hat{x}_{t}}$ for ease of notation. Using bandit feedback in the form of finitely many cost evaluations, the agents cannot obtain the accurate CDF $F_{i,t}$. Instead, they construct an empirical distribution function (EDF) $\hat{F}_{i,t}$ of the cost $J_{i}(\hat{x}_{t},\xi_{i})$ using $n_{t}$ cost evaluations by $\displaystyle\hat{F}_{i,t}(y)=\frac{1}{n_{t}}\sum_{j=1}^{n_{t}}\mathbf{1}\\{J_{i}(\hat{x}_{t},\xi_{i}^{j})\leq y\\}.$ (2) To improve estimate of the CDF, we utilize past samples for all episodes $t\geq 2$, and construct a modified distribution estimate $\bar{F}_{i,t}$ by adding a momentum term: $\displaystyle\bar{F}_{i,t}(y)=\beta\bar{F}_{i,t-1}(y)+(1-\beta)\hat{F}_{i,t}(y),$ (3) where $\beta\in[0,1)$ is the momentum parameter. For $t=1$, we set $\bar{F}_{i,t}(y)=\hat{F}_{i,t}(y)$. The agents utilize the distribution estimate $\bar{F}_{i,t}$ to calculate the CVaR estimates and further construct the gradient estimate using residual feedback as $\displaystyle\bar{g}_{i,t}=\frac{d_{i}}{\delta}\left({\rm{CVaR}}_{\alpha_{i}}[\bar{F}_{i,t}]-{\rm{CVaR}}_{\alpha_{i}}[\bar{F}_{i,t-1}]\right)u_{i,t},$ (4) where $\delta$ is the size of the perturbation on the action $x_{i,t}$ defined above. Depending on the size of the perturbation, the played action $\hat{x}_{i,t}$ may be infeasible, i.e., $\hat{x}_{i,t}\not\in\mathcal{X}_{i}$. To handle this issue, we define the projection set $\mathcal{X}_{i}^{\delta}=\\{x_{i}\in\mathcal{X}_{i}|{\rm{dist}}(x_{i},\partial\mathcal{X}_{i})\geq\delta\\}$ that we use in the projected gradient-descent update $\displaystyle x_{i,t+1}=\mathcal{P}_{\mathcal{X}_{i}^{\delta}}(x_{i,t}-\eta\bar{g}_{i,t}).$ (5) Note that the agents are not able to obtain accurate values of $C_{i}(\hat{x}_{t})$ using finite samples. In fact, if we use the distribution estimate $\bar{F}_{i,t}$ in (3) to calculate the CVaR values, there will be a CVaR estimation error, which is defined as $\displaystyle\bar{\varepsilon}_{i,t}:={\rm{CVaR}}_{\alpha_{i}}[\bar{F}_{i,t}]-{\rm{CVaR}}_{\alpha_{i}}[F_{i,t}].$ As a result, the gradient estimate $\bar{g}_{i,t}$ in (4) is biased since $\mathbb{E}[\bar{g}_{i,t}]=\mathbb{E}\left[\frac{d_{i}}{\delta}(C_{i}(\hat{x}_{t})+\bar{\varepsilon}_{i,t})u_{i,t}\right]=\nabla_{i}C_{i}^{\delta}(x_{t})+\mathbb{E}\left[\frac{d_{i}}{\delta}\bar{\varepsilon}_{i,t}u_{i,t}\right]$, with the bias captured by the last term. In what follows, we present a lemma that bounds the sum of the CVaR estimation errors. ###### Lemma 4. Suppose that Assumption 3 holds. Then, the sum of the CVaR estimation errors satisfies $\displaystyle\sum_{t=1}^{T}\left\|\bar{\varepsilon}_{i,t}\right\|\leq\frac{U}{\alpha_{i}}\Big{(}\frac{\beta}{1-\beta}e_{i,1}+\frac{\beta\Sigma_{1}}{1-\beta}T+\sum_{t=1}^{T}r_{t}\Big{)}$ (6) with probability at least $1-\gamma$, where $e_{i,t}:=\mathop{\rm{sup}}_{y}|\bar{F}_{i,t}(y)-F_{i,t}(y)|$, $r_{t}:=\sqrt{\frac{{\rm{ln}}(2T/\gamma)}{2bU^{2}(T-t+1)^{a}}}$, and $\Sigma_{1}:=C_{0}\eta\frac{d_{i}}{\delta}U\sqrt{N}+2C_{0}\delta$. ###### Proof. To bound the CVaR estimation error, it suffices to bound the CDF differences using Lemma 3. Recall the definition of $\bar{F}_{i,t}$ in (3). Then, we have that $\displaystyle\mathop{\rm{sup}}_{y}|\bar{F}_{i,t}(y)-F_{i,t}(y)|$ $\displaystyle=$ $\displaystyle\mathop{\rm{sup}}_{y}|\beta\bar{F}_{i,t-1}(y)+(1-\beta)\hat{F}_{i,t}(y)-F_{i,t}(y)|$ $\displaystyle=$ $\displaystyle\mathop{\rm{sup}}_{y}|\beta(\bar{F}_{i,t-1}(y)-{F}_{i,t-1}(y))+\beta({F}_{i,t-1}(y)-{F}_{i,t}(y))$ $\displaystyle+(1-\beta)(\hat{F}_{i,t}(y)-{F}_{i,t}(y))|$ $\displaystyle\leq$ $\displaystyle\beta\mathop{\rm{sup}}_{y}|\bar{F}_{i,t-1}(y)-{F}_{i,t-1}(y)|+\beta\mathop{\rm{sup}}_{y}|{F}_{i,t-1}(y)-{F}_{i,t}(y)|$ $\displaystyle+(1-\beta)\mathop{\rm{sup}}_{y}|\hat{F}_{i,t}(y)-{F}_{i,t}(y)|.$ (7) We now bound the latter two terms in the right-hand-side of (IV) separately. Using Assumption 3 and Lemma 2, we have that $\displaystyle\mathop{\rm{sup}}_{y}|F_{i,t}(y)-F_{i,t-1}(y)|$ $\displaystyle\leq$ $\displaystyle C_{0}\left\|\hat{x}_{t}-\hat{x}_{t-1}\right\|\leq C_{0}\left\|x_{t}-x_{t-1}\right\|+2C_{0}\delta$ $\displaystyle\leq$ $\displaystyle C_{0}\eta\left\|\bar{g}_{t}\right\|+2C_{0}\delta\leq C_{0}\eta\frac{d_{i}}{\delta}U\sqrt{N}+2C_{0}\delta:=\Sigma_{1},$ (8) where the last inequality holds since $\left\|\bar{g}_{t}\right\|=\sqrt{\sum_{i=1}^{N}\left\|\bar{g}_{i,t}\right\|^{2}}\leq\frac{\sqrt{N}d_{i}U}{\delta}$. Then, applying the Dvoretzky–Kiefer–Wolfowitz (DKW) inequality, we obtain that $\displaystyle\mathbb{P}\left\\{\mathop{\rm{sup}}_{y}|F_{i,t}(y)-\hat{F}_{i,t}(y)|\geq\sqrt{\frac{{\rm{ln}}(2/\bar{\gamma})}{2n_{t}}}\right\\}\leq\bar{\gamma}.$ (9) Define the events in (9) as $A_{t}$ and denote $\gamma=\bar{\gamma}T$. Then, the following inequality holds $\displaystyle\mathop{\rm{sup}}_{y}|F_{i,t}(y)-\hat{F}_{i,t}(y)|\leq\sqrt{\frac{{\rm{ln}}(2T/\gamma)}{2n_{t}}},\forall t=1,\ldots,T,$ (10) with probability at least $1-\gamma$ since $1-\mathbb{P}\\{\bigcup_{t=1}^{T}A_{t}\\}\geq 1-\sum_{t=1}^{T}\mathbb{P}\\{A_{t}\\}\geq 1-T\frac{\gamma}{T}\geq 1-\gamma$. Substituting the bounds in (IV) and (10) into (IV), and using the definition of $e_{i,t}$, we get that $\displaystyle e_{i,t}\leq\beta e_{i,t-1}+\beta\Sigma_{1}+(1-\beta)r_{t}.$ (11) Note that (11) holds for all $t\in\\{2,\cdots,T\\}$. By iteratively using this inequality, we have that $\displaystyle e_{i,t}\leq\frac{\beta}{1-\beta}\Sigma_{1}+(1-\beta)\sum_{k=0}^{k-2}\beta^{k}r_{t-k}+\beta^{t-1}e_{1,i}.$ (12) Taking the sum over $t=1,\ldots,T$ of both sides of (12), we have that $\displaystyle\sum_{t=1}^{T}e_{i,t}\leq\frac{\beta}{1-\beta}e_{i,1}+\frac{\beta\Sigma_{1}}{1-\beta}T+\sum_{t=1}^{T}r_{t},$ (13) with probability at least $1-\gamma$. Finally, using Lemma 3, we can obtain the desired result. ∎ Lemma 4 bounds the CVaR estimation error caused by using past samples. The next lemma quantifies the error due to residual feedback in the gradient estimate (4). Specifically, it bounds the second moment of the gradient estimate. ###### Lemma 5. Let Assumptions 1, 2 and 3 hold. Suppose that the action is updated as in (5). Then, the second moment of the gradient estimate satisfies $\displaystyle\sum_{t=1}^{T}\left\|\bar{g}_{t}\right\|^{2}\leq\frac{1}{1-\sigma}\left\|\bar{g}_{1}\right\|^{2}+\frac{1}{1-\sigma}\Sigma_{2}T,$ (14) where $\bar{g}_{t}:=(\bar{g}_{i,t},\bar{g}_{-i,t})$ is the concatenated vector of all agents’ gradient estimates, $\sigma=\frac{4d_{i}^{2}L_{0}^{2}N\eta^{2}}{\delta^{2}}$, $\Sigma_{2}=16d_{i}^{2}N^{2}L_{0}^{2}+\frac{48C_{0}^{2}d_{i}^{4}U^{4}N\eta^{2}\beta^{2}}{\delta^{4}(1-\beta)^{2}}(\sum_{i}\frac{1}{\alpha_{i}^{2}})+\frac{192C_{0}^{2}d_{i}^{2}U^{2}\beta^{2}}{(1-\beta)^{2}}(\sum_{i}\frac{1}{\alpha_{i}^{2}})+\frac{12\ln(2T/\gamma)d_{i}^{2}}{b\delta^{2}}(\sum_{i}\frac{1}{\alpha_{i}^{2}})+\frac{24d_{i}^{2}U^{2}\beta^{2}e_{m}^{2}}{\delta^{2}}(\sum_{i}\frac{1}{\alpha_{i}^{2}})$, and $e_{m}:=\mathop{\max}_{i}\\{e_{i,1}\\}$. ###### Proof. Using the fact that $r_{t}\leq\sqrt{\frac{{\rm{ln}}(2T/\gamma)}{2bU^{2}}}:=r$ and (12), we have that $\displaystyle\left\|e_{i,t}\right\|^{2}$ $\displaystyle\leq\left\|\frac{\beta}{1-\beta}\Sigma_{1}+r+\beta e_{m}\right\|^{2}$ $\displaystyle\leq 3\left\|\frac{\beta}{1-\beta}\Sigma_{1}\right\|^{2}+3\left\|r\right\|^{2}+3\left\|\beta e_{m}\right\|^{2},$ (15) where the last inequality holds since $\left\|a+b+c\right\|^{2}\leq 3\left\|a\right\|^{2}+3\left\|b\right\|^{2}+3\left\|c\right\|^{2}$. Then, applying Lemma 3 to (IV), we obtain that $\displaystyle\left\|\bar{\varepsilon}_{i,t}\right\|^{2}\leq\frac{U^{2}}{\alpha_{i}^{2}}\left(\frac{3\beta^{2}\Sigma_{1}^{2}}{(1-\beta)^{2}}+3\left\|r\right\|^{2}+3\beta^{2}\left\|e_{m}\right\|^{2}\right).$ (16) By the definition of $\bar{g}_{i,t}$ in (4), we have that $\displaystyle\left\|\bar{g}_{i,t}\right\|^{2}=\frac{d_{i}^{2}}{\delta^{2}}\left(({\rm{CVaR}}_{\alpha_{i}}[\bar{F}_{i,t}]-{\rm{CVaR}}_{\alpha_{i}}[\bar{F}_{i,t-1}])u_{i,t}\right)^{2}$ $\displaystyle\leq\frac{d_{i}^{2}}{\delta^{2}}\left({\rm{CVaR}}_{\alpha_{i}}[\bar{F}_{i,t}]-{\rm{CVaR}}_{\alpha_{i}}[\bar{F}_{i,t-1}]\right)^{2}\left\|u_{i,t}\right\|^{2}$ $\displaystyle\leq\frac{d_{i}^{2}}{\delta^{2}}\Big{(}2({\rm{CVaR}}_{\alpha_{i}}[F_{i,t}]-{\rm{CVaR}}_{\alpha_{i}}[F_{i,t-1}])^{2}$ $\displaystyle\quad+2(\bar{\varepsilon}_{i,t}-\bar{\varepsilon}_{i,t-1})^{2}\Big{)}\left\|u_{i,t}\right\|^{2}$ $\displaystyle\leq\frac{d_{i}^{2}}{\delta^{2}}\Big{(}2L_{0}^{2}\left\|\hat{x}_{t}-\hat{x}_{t-1}\right\|^{2}+4|\bar{\varepsilon}_{i,t}|^{2}+4|\bar{\varepsilon}_{i,t-1}|^{2}\Big{)},$ (17) where the last inequality is due to the fact that the CVaR function is Lipschitz continuous as shown in Lemma 1. We now bound the term $\left\|\hat{x}_{t}-\hat{x}_{t-1}\right\|^{2}$ in (IV). Recalling the update rule $x_{i,t}=\mathcal{P}_{\mathcal{X}_{i}^{\delta}}(x_{i,t-1}-\eta_{i}\bar{g}_{i,t-1})$, we have that $\displaystyle\left\|\hat{x}_{t}-\hat{x}_{t-1}\right\|^{2}=\left\|x_{t}-x_{t-1}+\delta u_{t}-\delta u_{t-1}\right\|^{2}$ $\displaystyle\leq$ $\displaystyle 2\left\|x_{t}-x_{t-1}\right\|^{2}+2\delta^{2}(2\left\|u_{t}\right\|^{2}+2\left\|u_{t-1}\right\|^{2})$ $\displaystyle\leq$ $\displaystyle 2\left\|x_{t}-x_{t-1}\right\|^{2}+8N\delta^{2}\leq 2\eta^{2}\left\|\bar{g}_{t-1}\right\|^{2}+8N\delta^{2}.$ Substituting the above inequality and the inequality in (16) into the right hand side of (IV) and summing both sides over all agents $i=1,\ldots,N$, we have that $\displaystyle\left\|\bar{g}_{t}\right\|^{2}=\sum_{i=1}^{N}\left\|\bar{g}_{i,t}\right\|^{2}$ $\displaystyle\leq\frac{d_{i}^{2}N}{\delta^{2}}\Big{(}4L_{0}^{2}\eta^{2}\left\|\bar{g}_{t-1}\right\|^{2}+16L_{0}^{2}N\delta^{2}\Big{)}$ $\displaystyle\quad+\frac{8d_{i}^{2}}{\delta^{2}}\sum_{i=1}^{N}\frac{U^{2}}{\alpha_{i}^{2}}\left(\frac{3\beta^{2}\Sigma_{1}^{2}}{(1-\beta)^{2}}+3\left\|r\right\|^{2}+3\beta^{2}\left\|e_{m}\right\|^{2}\right)$ $\displaystyle\leq\sigma\left\|\bar{g}_{t-1}\right\|^{2}+16d_{i}^{2}N^{2}L_{0}^{2}+\frac{12\ln(2T/\gamma)d_{i}^{2}}{b\delta^{2}}(\sum_{i}\frac{1}{\alpha_{i}^{2}})$ $\displaystyle\quad+\frac{24d_{i}^{2}U^{2}\beta^{2}}{\delta^{2}(1-\beta)^{2}}(C_{0}\eta\frac{d_{i}}{\delta}U\sqrt{N}+2C_{0}\delta)^{2}(\sum_{i}\frac{1}{\alpha_{i}^{2}})$ $\displaystyle\quad+\frac{24d_{i}^{2}U^{2}\beta^{2}e_{m}^{2}}{\delta^{2}}(\sum_{i}\frac{1}{\alpha_{i}^{2}})$ $\displaystyle\leq\sigma\left\|\bar{g}_{t-1}\right\|^{2}+\Sigma_{2},$ (18) where the last inequality is due to the fact that $(C_{0}\eta\frac{d_{i}}{\delta}U\sqrt{N}+2C_{0}\delta)^{2}\leq 2C_{0}^{2}\eta^{2}\frac{d_{i}^{2}}{\delta^{2}}U^{2}N+8C_{0}^{2}\delta^{2}$. Summing (IV) over all episode $t=1,\ldots,T$ completes the proof. ∎ Before proving the main theorem, we present a regret decomposition lemma that links the regret to the errors bounds on the CVaR estimates and gradient estimates as provided in Lemmas 4 and 5. ###### Lemma 6. Let Assumptions 1, 2 and 3 hold. Then, the regret of Algorithm 1 satisfies $\displaystyle{\rm{R}}_{C_{i}}(T)\leq$ $\displaystyle\frac{D_{x}^{2}}{2\eta}+\frac{\eta}{2}\mathbb{E}\big{[}\sum_{t=1}^{T}\left\|\bar{g}_{i,t}\right\|^{2}\big{]}+(4\sqrt{N}+\Omega)L_{0}\delta T$ $\displaystyle+\frac{d_{i}D_{x}}{\delta}\sum_{t=1}^{T}|\bar{\varepsilon}_{i,t}|,$ (19) where $\Omega>0$ is a constant that represents the error from projection $\mathcal{P}$. ###### Proof. The proof can be adapted from Lemma 5 in [14] and is omitted due to the space limitations. ∎ ###### Theorem 1. Let Assumptions 1, 2 and 3 hold and select $\eta=\frac{D_{x}}{d_{i}L_{0}N}T^{-\frac{3a}{4}}$, $\delta=\frac{D_{x}}{N^{\frac{1}{6}}}T^{-\frac{a}{4}}$, $\beta=\frac{1}{U^{2}T^{\frac{a}{4}}}$. Suppose that $n_{t}$ is chosen as in (1) with $a\in(0,1)$, and the EDF and the gradient estimate are defined as in (2) and (4), respectively. Then, when $T\geq(8N^{\frac{2}{3}})^{\frac{1}{a}}$, Algorithm 1 achieves regret ${\rm{R}}_{C_{i}}(T)=\mathcal{O}(D_{x}d_{i}L_{0}NS(\alpha)\ln(T/\gamma)T^{1-\frac{a}{4}}),$ with probability at least $1-\gamma$ and where $S(\alpha):=\sum_{i=1}^{N}\frac{1}{\alpha_{i}^{2}}$. ###### Proof. Substituting the bounds in Lemmas 4 and 5 into Lemma 6, we have that $\displaystyle{\rm{R}}_{C_{i}}(T)\leq$ $\displaystyle\frac{D_{x}^{2}}{2\eta}+\frac{\eta}{2}\left(\frac{1}{1-\sigma}\left\|\bar{g}_{1}\right\|^{2}+\frac{1}{1-\sigma}\Sigma_{2}T\right)$ $\displaystyle+\frac{d_{i}D_{x}U}{\delta\alpha_{i}}\Big{(}\frac{\beta}{1-\beta}e_{i,1}+\frac{\beta\Sigma_{1}}{1-\beta}T+\sum_{t=1}^{T}r_{t}\Big{)}$ $\displaystyle+(4\sqrt{N}+\Omega)L_{0}\delta T.$ (20) Substituting $\delta$, $\eta$ and $\beta$ into the above inequality, we can obtain the desired result. The detailed proof is straightforward and is omitted due to the space limitations. ∎ Note that Algorithm 1 in fact achieves the regret ${\rm{R}}_{C_{i}}(T)=\tilde{\mathcal{O}}(T^{1-\frac{a}{4}})$, in which we use the notation $\tilde{\mathcal{O}}$ to hide constant factors and poly- logarithmic factors of $T$. This result matches the best result in [14]. Compared to [14], the analysis of Algorithm 1 that combines historical information and residual feedback is nontrivial and requires an appropriate selection of the momentum parameter to show convergence. Note that Algorithm 3 in [14] is a special case of Algorithm 1 for the momentum parameter $\beta=0$. ## V Numerical Experiments In this section, we illustrate the proposed algorithm on a Cournot game. Specifically, we consider two risk-averse agents $i=1,2$ that compete with each other. Each agent determines the production level $x_{i}$ and receives an individual cost feedback which is given by $J_{i}=1-(2-\sum_{j}x_{j})x_{i}+0.2x_{i}+\xi_{i}x_{i}$, where $\xi_{i}\sim U(0,1)$ is a uniform random variable. Here we utilize the term $\xi_{i}x_{i}$ to represent the uncertainty occurred in the market, which is proportional to the production level $x_{i}$. The agents have their own risk levels $\alpha_{i}$ and they aim to minimize the risk of incurring high costs, i.e., the CVaR of their cost functions. Note that the regret for agent $i$ depends on the sequence of the other agent’s actions $\\{x_{-i,t}\\}_{t=1}^{T}$, and this sequence depends on the algorithm. Therefore, it is not appropriate to compare different algorithms in terms of the regret. Instead, we use the empirical performance to evaluate the algorithms. Specifically, we compute the CVaR values at each episode, and observe how fast the agents can minimize the CVaR values during the learning process. To compute the distribution function $\bar{F}_{i,t}$, we divide the interval $[0,U]$ into $n$ bins of equal width, and we approximate the expectation by the sum of finite terms. We compare our zeroth-order momentum method with the two algorithms in [14], which we term as the algorithm with sample reuse and the algorithm with residual feedback, respectively. The number of samples $n_{t}$ at each episode for Algorithm 1 is shown in Figure 1. We set the momentum parameter $\beta=0.5$. Each algorithm is run for 20 trials and the parameters of these algorithms are separately optimally tuned. The CVaR values of these algorithms are presented in Figure 2. We observe that all the algorithms finally converge to the same CVaR values, but the zeroth-order momentum method proposed here converges faster. This is because our method uses all past samples to get a better estimate of the CVaR values and thus allows for a larger step size. Moreover, we also observe that the standard deviation of both our proposed algorithm and the algorithm with residual feedback is sufficiently small, which verifies the variance reduction effect of using residual feedback. Figure 1: The number of samples of Algorithm 1. Figure 2: Comparative results among our proposed zeroth-order momentum method, and the three algorithms in [14]. Shaded areas represent $\pm$ one standard deviation over 20 runs. Figure 3: CVaR values achieved by Algorithm 1 with different $\beta$ values. Shaded areas represent $\pm$ one standard deviation over 20 runs. Moreover, we explore the effect of different momentum parameters $\beta$. In Figure 3, we present the achieved CVaR values of our method for various $\beta$ values and all other parameters unchanged. Recalling the definition of the distribution estimate in (3), large values of $\beta$ place more weight on outdated cost values and thus reduce the convergence speed. ## VI Conclusion In this work, we proposed a zeroth-order momentum method for online convex games with risk-averse agents. The use of momentum that employs past samples allowed to improve the ability of the algorithm to estimate the CVaR values. Zeroth-order estimation of the CVaR gradients using residual feedback allowed us to reduce the variance of the CVaR gradient estimates. We showed that the proposed algorithm theoretically achieves no-regret learning with high probability, matching the best known result in literature. Moreover, we provided numerical simulations that demonstrated the superior performance of our method in practice. While sample reuse and residual feedback had been both separately shown to improve the performance of online learning, here we showed that their combination yields yet better performance. ## References * [1] S. Shalev-Shwartz and Y. Singer, “Convex repeated games and fenchel duality,” in _NIPS_ , vol. 6. Citeseer, 2006, pp. 1265–1272. * [2] G. J. Gordon, A. Greenwald, and C. Marks, “No-regret learning in convex games,” in _Proceedings of the 25th international conference on Machine learning_ , 2008, pp. 360–367. * [3] P. G. Sessa, I. Bogunovic, M. Kamgarpour, and A. Krause, “No-regret learning in unknown games with correlated payoffs,” _Advances in Neural Information Processing Systems_ , vol. 32, pp. 13 624–13 633, 2019. * [4] Y. Shi and B. Zhang, “No-regret learning in cournot games,” _arXiv preprint arXiv:1906.06612_ , 2019. * [5] E. Hazan, “Introduction to online convex optimization,” _arXiv preprint arXiv:1909.05207_ , 2019. * [6] S. Shalev-Shwartz _et al._ , “Online learning and online convex optimization,” _Foundations and trends in Machine Learning_ , vol. 4, no. 2, pp. 107–194, 2011. * [7] W. F. Sharpe, “The sharpe ratio,” _Journal of portfolio management_ , vol. 21, no. 1, pp. 49–58, 1994. * [8] P. Artzner, F. Delbaen, J.-M. Eber, and D. Heath, “Coherent measures of risk,” _Mathematical finance_ , vol. 9, no. 3, pp. 203–228, 1999. * [9] R. T. Rockafellar, S. Uryasev _et al._ , “Optimization of conditional value-at-risk,” _Journal of risk_ , vol. 2, pp. 21–42, 2000. * [10] T. Soma and Y. Yoshida, “Statistical learning with conditional value at risk,” _arXiv preprint arXiv:2002.05826_ , 2020. * [11] A. R. Cardoso and H. Xu, “Risk-averse stochastic convex bandit,” in _The 22nd International Conference on Artificial Intelligence and Statistics_. PMLR, 2019, pp. 39–47. * [12] S. Curi, K. Levy, S. Jegelka, A. Krause _et al._ , “Adaptive sampling for stochastic risk-averse learning,” _arXiv preprint arXiv:1910.12511_ , 2019\. * [13] A. Tamkin, R. Keramati, C. Dann, and E. Brunskill, “Distributionally-aware exploration for cvar bandits,” in _NeurIPS 2019 Workshop on Safety and Robustness on Decision Making_ , 2019. * [14] Z. Wang, Y. Shen, and M. Zavlanos, “Risk-averse no-regret learning in online convex games,” in _International Conference on Machine Learning_. PMLR, 2022, pp. 22 999–23 017. * [15] N. Qian, “On the momentum term in gradient descent learning algorithms,” _Neural networks_ , vol. 12, no. 1, pp. 145–151, 1999. * [16] Y. Zhang, Y. Zhou, K. Ji, and M. M. Zavlanos, “Boosting one-point derivative-free online optimization via residual feedback,” _arXiv preprint arXiv:2010.07378_ , 2020. * [17] M. Bravo, D. Leslie, and P. Mertikopoulos, “Bandit learning in concave n-person games,” _Advances in Neural Information Processing Systems_ , vol. 31, 2018.
# Measuring the magnetic origins of solar flares, CMEs and Space Weather Philip Judge High Altitude Observatory, National Center for Atmospheric Research, Boulder CO 80307-3000, USA Matthias Rempel High Altitude Observatory, National Center for Atmospheric Research, Boulder CO 80307-3000, USA Rana Ezzeddine Department of Astronomy, University of Florida, 211 Bryant Space Sciences Center, Gainesville, Florida, 32611, USA Lucia Kleint Université de Genève, Centre Universitaire d’Informatique, 7, route de Drize, 1227 Carouge, Switzerland Ricky Egeland High Altitude Observatory, National Center for Atmospheric Research, Boulder CO 80307-3000, USA Svetlana V. Berdyugina Leibniz-Institut fuer Sonnenphysik (KIS), Schöneckstrasse 6, 79104, Freiburg, Germany Thomas Berger Space Weather Technology, Research and Education Center, 3775 Discovery Drive N433 Boulder, CO 80309 Paul Bryans High Altitude Observatory, National Center for Atmospheric Research, Boulder CO 80307-3000, USA Joan Burkepile High Altitude Observatory, National Center for Atmospheric Research, Boulder CO 80307-3000, USA Rebecca Centeno High Altitude Observatory, National Center for Atmospheric Research, Boulder CO 80307-3000, USA Giuliana de Toma High Altitude Observatory, National Center for Atmospheric Research, Boulder CO 80307-3000, USA Mausumi Dikpati High Altitude Observatory, National Center for Atmospheric Research, Boulder CO 80307-3000, USA Yuhong Fan High Altitude Observatory, National Center for Atmospheric Research, Boulder CO 80307-3000, USA Holly Gilbert High Altitude Observatory, National Center for Atmospheric Research, Boulder CO 80307-3000, USA Daniela A. Lacatus High Altitude Observatory, National Center for Atmospheric Research, Boulder CO 80307-3000, USA ###### Abstract We take a broad look at the problem of identifying the magnetic solar causes of space weather. With the lackluster performance of extrapolations based upon magnetic field measurements in the photosphere, we identify a region in the near UV part of the spectrum as optimal for studying the development of magnetic free energy over active regions. Using data from SORCE, Hubble Space Telescope, and SKYLAB, along with 1D computations of the near-UV (NUV) spectrum and numerical experiments based on the MURaM radiation-MHD and HanleRT radiative transfer codes, we address multiple challenges. These challenges are best met through a combination of near UV lines of bright Mg II, and lines of Fe II and Fe I (mostly within the $4s-4p$ transition array) which form in the chromosphere up to $2\times 10^{4}$ K. Both Hanle and Zeeman effects can in principle be used to derive vector magnetic fields. However, for any given spectral line the $\tau=1$ surfaces are generally geometrically corrugated owing to fine structure such as fibrils and spicules. By using multiple spectral lines spanning different optical depths, magnetic fields across nearly-horizontal surfaces can be inferred in regions of low plasma $\beta$, from which free energies, magnetic topology and other quantities can be derived. Based upon the recently-reported successful suborbital space measurements of magnetic fields with the CLASP2 instrument, we argue that a modest space-borne telescope will be able to make significant advances in the attempts to predict solar eruptions. Difficulties associated with blended lines are shown to be minor in an Appendix. Sun: atmosphere - Sun: magnetic fields ## 1 Introduction Commenting on the origin of solar flares, in 1960 Gold & Hoyle wrote > The requirements of the theory can therefore be stated quite definitely. > Magnetic field configurations must be found that are capable of storing > energy densities hundreds of times greater than occur in any other form, and > that are stable most of the time. A situation that occurs only a small > fraction of the time must be able to lead to instability in which this > energy can rapidly be dissipated into heat and mass motion… These requirements have remained unchanged in the intervening six decades. We now have a “standard model” of flares (Carmichael, 1964; Sturrock, 1966; Hirayama, 1974; Kopp & Pneuman, 1976), involving a loss of equilibrium of arcades of closed magnetic flux systems, particle acceleration and, frequently, ejection of plasmoids. The instability can only arise when the free magnetic energy stored in the coronal magnetic fields exceeds certain thresholds. This is, however, a necessary but not sufficient condition for flaring and plasmoid ejection, because other global and local properties of the magnetic field, notably the topology, also determine the stabilility of coronal MHD systems (Low, 1994). Enormous interest in solar eruptions is driven by society’s need to understand origins of “space weather” (e.g. Eddy, 2009) as well as its intrinsic scientific challenges. However, much like weather prediction before the advent of modern instrumentation and computers, predicting the properties of eruptions and flares has remained a difficult challenge primarily due to the lack of quantitative data available for analysis of the pre-flare state and flare-triggering mechanism(s). In particular, the only high-resolution, high cadence, data available to date has been the magnetic field state in the photosphere, which has been used to define “boundary conditions” to the overlying coronal magnetic field, where most of the free energy is stored prior to eruption. Herein lies the basic problem: magnetic fields measured in the photosphere are not ideally suited to the task. We elaborate upon this below. The situation described by Gold & Hoyle (1960) draws analogy with trying to predict the timing and location of lightning strikes (Judge, 2020). By measuring updrafts and/or observing cloud buildup one can identify the causes of the buildup of electrostatic energy, but lightning occurs in a predictable way only in a statistical sense. In-situ measurements of electric fields are required even then to make probabilistic forecasts of lightning strikes to the ground, without reference to intra- or inter- cloud discharges. In this sense the weaknesses found in previous prediction methods is perhaps unsurprising. The difficulty of the task has been summarized in the write-up of an inter- agency “All Clear Workshop” (Barnes et al., 2016) in which the authors conclude with the sobering statement that > For M-class flares and above, the set of methods tends toward a weakly > positive skill score (as measured with several distinct metrics), with no > participating method proving substantially better than climatological > forecasts. Steady improvements in the performance of flare forecasts has been documented in more recent years (e.g. Leka et al., 2019). However, these authors make explicit the need to augment existing photospheric magnetic data in order to proceed, with a heterogeneous mix of theory and observation: > quantitative “modern” forecasts incorporate …physical understanding as they > often characterize coronal magnetic energy storage by proxy, with the > parameterizations of photospheric magnetograms. In other words, measurements of the photospheric magnetic field can be related only indirectly and through certain ad-hoc parameterizations, to the magnetic free energy of the overlying corona, which is the ultimate origin of instabilities leading to flares and CMEs. With the availability of a decade of high-resolution photospheric magnetic field data from NASA’s Solar Dynamics Observatory (SDO) mission, recent efforts have focused on statistical pattern recognition methods for flare prediction that generally fall under the “machine learning (ML)” sub-field of artificial intelligence research. For example, a study by Bobra & Couvidat (2015) used photospheric vector magnetic data of 2071 active regions from the Helioseismic and Magnetic Imager (HMI) instrument on SDO. They predicted strong flares using a Support Vector Machine (SVM) algorithm applied to a set of 25 parameters derived from the vector magnetic field measured in SHARPs (Space-weather HMI Active Region Patches), which are cut-outs around sunspot active regions. They achieved a high True Skill Statistic (TSS) of $\approx$ 0.8, but at the cost of getting a significant number of false positives, a common characteristic of all ML flare prediction algorithms optimized for TSS in the strongly unbalanced training sets which naturally contain many more “no flare” examples than “flare” examples. Later studies (Park et al., 2018; Liu et al., 2019; Chen et al., 2019; Li et al., 2020) applied recurrent architectures such as Long Short-Term Memory (LSTM) systems and deep learning systems including Convolution Neural Network (CNN) and autoencoder architectures to explore feature sets that extend beyond the magnetic field- derived features of earlier ML flare prediction systems. However these studies do not demonstrate significant increases in skill scores relative to earlier approaches. These studies have used only photospheric measurements of magnetic fields. Investigations have also been carried out that include information from higher atmospheric layers, e.g., by incorporating SDO Atmospheric Imaging Array (AIA) data into existing ML flare prediction systems (Nishizuka et al., 2018; Jonas et al., 2018), or by using extrapolated coronal magnetic field models (Korsós et al., 2020). Chromospheric UV spectral lines from IRIS have also been scrutinized for additional flare prediction utility (Panos & Kleint, 2020). These studies all show some degree of improvement in flare prediction metrics when including chromospheric or derived coronal information. More recently, Deshmukh et al. (2020) showed that combining Topological Data Analysis (TDA) of the radial magnetic field structure in active regions with the SHARPs vector magnetic field metadata results in similar or slightly higher skill scores compared to using selected SHARPs metadata alone. This result implies that inclusion of high-resolution imaging of active region and flows may be an important complement to spectroscopic measurements when deducing atmospheric conditions that evolve to a flaring state. Physically, however, solar flare prediction based upon photospheric data is ultimately confronted with basic difficulties, beyond the well-known disambiguation problem associated with native symmetries in the Zeeman effect. The photosphere is a dense radiating layer of high density ($10^{-7}$ g cm-3) plasma. Outside of sunspot umbrae the photospheric magnetic field is unable to suppress convective motions which therefore control the magnetic fields threading the fluid, a case where the plasma-$\beta=$ gas pressure / magnetic pressure, is generally $>1$. In traversing the chromosphere, emerging magnetic fields drop in strength algebraically but the plasma densities and energy densities drop exponentially (by 7 and 5 orders of magnitude, respectively), so that the tenuous plasma at the coronal base is in a low-$\beta$ state. One immediate practical problem is that tiny dynamical changes which may be unobservable in the dense photosphere can have enormous consequences for the tenuous plasma and magnetic fields above: the problem is observationally ill- posed. As a result of these different photospheric and coronal plasma regimes, the time scales for significant photospheric magnetic field evolution are typically measured in minutes to hours whereas the dynamic evolution and reconnection of the coronal magnetic field configurations leading to flares, after a quasi-static buildup phase, is measured in seconds. This fundamental mismatch in dynamical time scales evidently cannot be overcome through any type of analysis of photospheric magnetic field data and leads to the conclusion that quantitative analysis of the magnetic field and flows in the chromosphere and corona are necessary to catalyze significant breakthroughs. Figure 1: Contribution functions are shown as a function of temperature from the model C of Fontenla et al. (1993). The emergent intensities are over- plotted as solid lines for a ray from close to disk center ($\cos\theta=0.887$, $\theta$ the angular distance from disk center as measured from Sun center). The Ca II lines at 393 and 854 nm span from 3.8 to 4.0 in log$T$ (the latter forming in deeper cooler layers than the former), and the 1083 nm line of He I forms within the last pressure scale height of the chromosphere. Notice the extension of contributions for the Mg II resonance lines beyond all others, to $\log T_{e}\geq 4.3$, and the multiple lines of Fe II between 259 and 261 nm whose cores form close to 3.9-4.0 in log$T$, and weaker lines in the 275 nm region with stronger contributions from lower temperatures. A modest atomic model for Fe II was adopted to examine the line cores formed in non LTE in the upper chromosphere. The calculations used the RH computer program Uitenbroek (2001). Far more complete LTE calculations are made in the Appendix where spectral line blends are discussed. In order to address this and other questions, our purpose here is to explore the entire solar spectrum to attempt to identify more direct ways to identify the free energy and topology of magnetic fields above active regions. In this way we hope to put the prediction of solar magnetic eruptions, the source of all major space weather events, on a firm observational basis. The specific problem we address is reviewed in section 2. It will involve spectro- polarimetry. Before proceeding, we point out that our goals may on first reading appear similar to efforts measuring only intensities of EUV and UV emission lines (e.g. De Pontieu et al., 2020), an endeavor with a history of 6+ decades yet able to study only the effects and not causes of magnetic energy storage and release. Instead our work is closer to an important study by Trujillo Bueno et al. (2017), but the latter focuses more on the physical processes behind the magnetic imprints across the solar UV spectrum, including the near-UV spectral region of the Mg II $h$ & $k$ lines. We focus instead on finding an optimal set of observations which target the problem of the build-up and release of magnetic energy. We argue that spectro-polarimetry of the near-UV solar spectrum is indeed a profitable avenue for research. ## 2 Observational signatures of energy build-up and release ### 2.1 Measuring magnetic fields in and above chromospheric plasma Two approaches are in principle possible: in-situ measurements and remote sensing. Clearly in-situ methods are far beyond the capabilities of known technology. The state-of-the-art mission Parker Solar Probe (Fox et al., 2016) has a perihelion distance of just below ten solar radii, far above active regions where most of the free magnetic energy exists. Therefore we limit our discussion to measuring magnetic fields remotely, an area which has a firm physical basis (Landi degl’Innocenti & Landolfi, 2004). Briefly, two types of remote-sensing measurements have been made: direct measurements of light emitted from within coronal plasma, and measurements close to the corona’s lower boundary, some 9 pressure scale heights above the photosphere. Coronal measurements include (see, for example, reviews by Judge et al., 2001; Casini et al., 2017; Kleint & Gandorfer, 2017, for more complete references): * • Measurements of line-of-sight (LOS) field strengths using differences between ordinary and extraordinary rays of magneto-ionic theory at GHz frequencies Gelfreikh (1994). * • Measurements of field strengths across surfaces determined typically by a harmonic of gyro-resonant motions of electrons at frequencies of several GHz, frequencies uncontaminated by bremsstrahlung opacity only when the field strength exceeds $\approx$200 G (White, 2004). * • Serendipitous level-crossings in atomic ions produces changes in EUV spectra through level mixing by the Zeeman effect (Li et al., 2017), yielding magnetic field strengths. * • A long history of measurements of magnetic dipole lines during eclipses or with coronagraphs hold great promise for measuring vector magnetic fields in the corona. The infrared, coronagraphic DKIST facility offers significant new opportunities for advancements using such lines (e.g. Judge et al., 2013). * • Occultations of background radio sources by the corona have revealed integrated properties of coronal magnetic fields, weighted by electron density through the Faraday effect (Bird & Edenhofer, 1990), a technique also applied to signals from spacecraft away from the Sun-Earth line (Bird, 1982). Of the second type, we can list: * • Several strong lines measured using spectro-polarimetric measurements of the Zeeman effect from the ground can yield LOS components of chromospheric magnetic fields. Very high photometric precision is however required for vector fields, beyond most current capabilities (Uitenbroek, 2010), owing to the second-order $\epsilon^{2}$ signature of the Zeeman effect in the small ratio $\epsilon$ of magnetic splitting to the line Doppler width (defined below). These include Na I D lines, Ca II H, K and infrared triplet lines. Synthetic Ca II line profiles are shown in the non-LTE calculations using the RH non-LTE program including partial redistribution (Uitenbroek, 2001) in Figure 1. * • The He I 1083 nm triplet has a proven record of magnetic field determinations in a variety of cool structures such as spicules (Centeno et al., 2010), filaments (Kuckein et al., 2012; Díaz Baso et al., 2019) and cool loops (Solanki et al., 2003), in the vicinity of the base of the corona before, during and after flares (Judge et al., 2014) even extending far into the corona (Judge et al., 2012; Schad et al., 2016). Both the Zeeman and Hanle effects contribute to the formation of this line, owing to the large contribution of anisotropic radiation incident on the helium atoms to the statistical equilibrium in the triplet states (e.g. Trujillo Bueno & Asensio Ramos, 2007). * • Recently, sub-orbital rockets have carried sensitive spectropolarimeters to observe the resonance lines of hydrogen (Kano et al., 2017) and singly-ionized magnesium (Trujillo Bueno et al., 2018; Ishikawa et al., 2021). Both Zeeman and Hanle effects have revealed information on magnetic fields near the very top of the solar chromosphere. * • Measurements at 30 GHz and above made with the ALMA interferometer facility can in principle yield line-of-sight magnetic field strengths using Gelfreikh’s (1994) theory of continuum polarization (Loukitcheva, 2014, 2020), but we have been unable to find reports of solar polarization measurements with ALMA. It can observe from 0.4 to 8.8 mm wavelengths (750 and 34 Ghz respectively), thereby sampling heights from 600 km to 2000 km in the chromosphere. Transition region plasma under non-flaring conditions is tenuous and thin. It contributes very little to the far brighter chromospheric emission at the high frequencies measured by ALMA. Next we assess the ability of these techniques to address the specific problem of interest. Figure 2: Images are shown to highlight the corrugated surfaces at which the centers of UV lines are formed (close to $\tau=1$) in MURaM calculations of the upper solar chromosphere. The calculations are from a numerical experiment where a magnetic flux system is emerging from beneath the solar atmosphere. The emergent fields carry with them plasma, which is revealed by the extended heights of the $\tau=1$ surfaces, which can exceed 10 Mm. Typically the line shifts in the lower transition region are smaller than the line widths (Athay & Dere, 1991). Therefore the $\tau=1$ surfaces were computed without taking into account Doppler shifts, at the centers of lines of Mg II $k$, and several lines of Fe II. The $x-y$ images of height of formation are plotted in order of decreasing opacity from Mg II to Fe II 2769.75 Å. In the bottom right panel, slices of these $\tau=1$ surfaces are shown, taken along the black line in the five other panels. But the black lines in this final panel are magnetic lines of force in the $y-z$ plane. Evidently, a combination of these and other lines of iron with a wide range of opacities spans the ranges of excursions of the surfaces away from a horizontal plane. ### 2.2 Establishing methods to measure free magnetic energy and topology While information on magnetism above the photosphere is contained in all the above strategies, few offer ways to assess the necessary quantities concerning magnetic free energy as magnetic fields emerge, build and change topology as it evolves and is suddenly released. If we consider measurements within the corona, we would have to probe the 3D magnetic structure. However, the corona is optically thin (with the exception of gyro-resonance emission which yields just the magnetic field strength within resonant surfaces, whose geometry is not known from one line of sight), which means that recovery of the 3D coronal field would require measurements from at least two different lines of sight (e.g. Kramar et al., 2006). Until suitable instruments are flown on spacecraft away from the Sun-Earth line, the required stereoscopic views will be unavailable. Therefore we must seek alternatives. When chromospheric, transition region and coronal plasmas are connected by magnetic lines of force (Judge, 2021), the plasma pressure near the top of the chromosphere is close to that of the overlying corona. Accurate measurements of magnetic fields using spectral lines formed in plasma up to a few times $10^{4}$ K can therefore be used to inform us on the evolving magnetic state of the overlying corona (Trujillo Bueno et al., 2017). Selected spectral line profiles formed across the upper chromosphere are shown in Figure 1, intended to illustrate near UV lines in comparison to familiar lines commonly observed from the ground. The solar spectrum between 250 and 281 nm is dominated by line transitions of Mg II and many lines belonging to the $4s-4p$ transition array in neutral and singly ionized iron. In the figure, contributions to the intensities of Mg II and typical Fe II lines are shown as a function of electron temperature. This abscissa was chosen instead of height and continuum optical depth because in the particular 1-dimensional model adopted, the temperature gradient in the transition region is very large, making lines appear to originate from the same heights. The regime of plasma $\beta=8\pi p/B^{2}<1$ in upper chromospheric, transition region and coronal lines is the same, if the photospheric magnetic field has expanded to fill the volume. In active region plasma this appears to be the case (Ishikawa et al., 2021). When measurements are made close to the force- free state ($\beta\ll 1$), then theory can be reliably invoked in at least two ways: application of the magnetic virial theorem (Chandrasekhar, 1961), and extrapolation of fields with observational boundary conditions compatible with the magnetostatic conditions (Wiegelmann & Sakurai, 2021). Any waves and/or tangential discontinuities measured will also be of importance to assessing the free magnetic energy and its evolution. Fleishman et al. (2017) have confirmed these theoretical concepts quantitatively using model simulations. Skeptics of our statements might rightfully recall that the chromosphere is finely structured (see, e.g., the remarkable structure in narrow-band images of Cauzzi et al., 2008; Wang et al., 2016; Robustini et al., 2018). It is also known, for example from limb observations, that transition region plasma at intermediate temperatures has contributions from structures far from a level surface which might otherwise be amenable to mathematical techniques to estimate free energies and extrapolated magnetic fields. A non-level, “corrugated” surface of unit optical depth in a given spectral line can lead to spurious inferences of magnetic structure even within low-$\beta$ plasmas (see, e.g., section 4.3 of the paper by Trujillo Bueno et al., 2017, and our Figure 2 below). The question then arises, is the interface between chromosphere and corona just too finely structured to make measurements of magnetic fields there of little use? Hints from two pieces of work suggest that this is not the case. Firstly, the CLASP2 measurements reported by Ishikawa et al. (2021) have shown that, at an angular resolution of 1” which is far smaller than the scales of large coronal structures leading to flares, the fields measured in the core of the Mg II $h$ & $k$ lines across active regions are unipolar and spatially smooth over large areas. Any mixed polarity fields undetected by CLASP2 must lie below 1” scales and will therefore not penetrate more than a few hundred km above the photosphere. But the fields measured by CLASP2 may well contain unseen tangential discontinuities arising from small-scale angular displacements across flux surfaces (Judge et al., 2011). Such features would constitute an additional source of free magnetic energy whose consequences might then be explored using the very techniques advocated for below. Secondly, Judge (2021) has shown that the attention to the dynamics and fine structure of transition region plasma has been over-emphasized, statistically. Most of the chromosphere-corona transition region on the Sun is indeed in a thermal interface connected by magnetic lines of force, and not isolated, as was claimed by Hansteen et al. (2014) and references therein, all of them based on circumstantial evidence accumulated using spectral line intensities alone. One example is shown in Figure 2. This figure is based on a MURaM flux emergence simulation in a 98 Mm wide domain, reaching from -8 to 75 Mm in the vertical direction. The snapshot corresponds to a highly energized corona, in which a strong bipolar field has emerged into a previously quiet region. Ten minutes after this, a flare and a mass ejection occurred. The simulation is based on the Rempel (2017) version of MURaM, which uses a simplified description of the chromosphere. The $\tau=1$ surfaces were computed using parameterized li¡ne opacities at line center and integrating downwards from the corona. The opacities were derived using a function, which for the Mg II $k$-line for example, are of the form $\kappa_{\nu=\nu_{0}}=\frac{\pi e^{2}f_{ij}}{m_{e}c}\frac{n_{\mathrm{i}}}{n_{\mathrm{Mg^{+}}}}\frac{n_{\mathrm{Mg}^{+}}}{n_{\mathrm{Mg}}}\frac{n_{\mathrm{Mg}}}{n_{\mathrm{H}}}n_{\mathrm{H}}\phi_{\nu_{0}},$ (1) in which the lower level population $n_{\mathrm{i}}$ is expanded as a product of ionization fraction $\frac{n_{\mathrm{Mg}^{+}}}{n_{\mathrm{Mg}}}$ (a function of electron temperature $T_{e}$), the abundance $\frac{n_{\mathrm{Mg}}}{n_{\mathrm{H}}}$, and the hydrogen density $n_{\mathrm{H}}$ from the model. The first factor is the frequency-integrated cross section of the $k$ line from fundamental constants and the absorption oscillator strength $f_{ij}$ between levels $i$ and $j$, and $\phi_{\nu}$ the line profile at frequency $\nu$, normalized so that $\int_{0}^{\infty}\phi_{\nu}d\nu=1$. The profile is dominated for these ions by an assumed microturbulence of 10 km s-1. The term $\frac{n_{\mathrm{Mg}^{+}}}{n_{\mathrm{Mg}}}$ is computed assuming that neutral Mg is negligible, but allowing for electron impact ionization to Mg2+ and higher ions and radiative recombination using data from Allen (1973). This figure illustrates approximate line formation heights, sufficient for the present paper. The excursions in height corresponding to the emerging flux system vary from a maximum near 10 Mm for Mg II $k$, to a few Mm for a line of Fe II at 276.9 nm. As can be seen, these corrugations can be probed with multiple spectral features with different opacities. Thus measurements at the top of the chromosphere, properly sampling the potential excursions of the isotherms and isobars in the transition between chromosphere and corona, appear to be a promising approach to the problem. At the very least, such new magnetic measurements could be compared in detail with advanced simulations of the Sun’s atmosphere (Gudiksen et al., 2011; Rempel, 2017). Below we will suggest one approach to this problem based upon machine learning techniques. Figure 3: A low-resolution UV irradiance spectrum from the Whole Heliospheric Interval in 2008 is shown (from https://lasp.colorado.edu/lisird/data/whi_ref_spectra). The near UV region has a photon flux density far higher than the majority of vacuum UV lines below 200 nm. The exception is H L$\alpha$ which roughly 10x smaller than the flux density at 260 nm and longer wavelengths. ### 2.3 The near UV solar spectrum and the chromosphere-corona interface A standard, low resolution solar UV spectrum is shown in Figure 3. After an extensive search, we have identified a spectral range of 256-281 nm as a promising part of the solar spectrum for measurement of vector magnetic fields at the base of the corona. Our search, driven by the need to probe in detail magnetic structure in height above the solar surface and/or depth along the LOS of prominences, led to the choice of important transitions in the 4s-4p transition array of Fe II. While there are strong Fe II 4s-4p transitions in the region near 234 nm, the solar fluxes are weaker than in the 256-281 nm range. Our wavelength selection is significantly broader than the 279.2-280.7 nm spectral region of CLASP2, which includes the Mg II $h$ & $k$ line cores and magnetically-sensitive wings (Auer et al., 1980; Belluzzi & Trujillo Bueno, 2012; Alsina Ballester et al., 2016; del Pino Alemán et al., 2016) the 3p-3d lines (del Pino Alemán et al., 2020), and photospheric lines of Mn I (Ishikawa et al., 2021). Both Hanle and Zeeman effects can be brought to bear on the problem depending on the strength of the magnetic fields found across the solar atmosphere. The Hanle effect can be important to remove the 180o ambiguity present in Zeeman effect measurements. The rationale for this spectral region is based upon requirements demanded by the desire to measure vector magnetic fields within force-free ($\approx$ low $\beta$) plasma at the coronal base, over an area covering a typical active region. An angular resolution sufficient to resolve the thermodynamic fine structure is also a constraint. Such features are most clearly seen only in images of features within the narrow spectral lines chromospheric images. This constraint poses difficulties for current observations at radio and sub-mm wavelengths, and in hard X-rays. Soft X-rays ($\leq 1$ keV) have no known sensitivity to magnetic fields of solar magnitude, and along with EUV photons cannot escape from or penetrate deep into the chromosphere. We are then left to consider vacuum UV to infrared wavelengths. We further eliminate wavelengths above the atmospheric cutoff at 310 nm because of the lack of spectral lines that can span many decades in optical depths above the chromosphere. The Balmer and higher line series of hydrogen are poor choices for Zeeman diagnostic work, the lightness of hydrogen makes lines broad and patterns overlapping. For Ba-$\alpha$, seven anomalous Zeeman patterns will lie within half a wave number for typical solar field strengths even in the photosphere, thereby making the observation of anomalous Zeeman patterns extremely difficult. The Hanle effect has been explored by Štěpán & Trujillo Bueno (2010) in their comparison of 1D radiative transfer calculations with limb observations of the quiet Sun by Gandorfer (2000). The value of H$\alpha$ as a diagnostic of magnetic field remains a subject for further research. The strongest line in the chromospheric visible spectrum is Ca II $K$ has just $1/20$ the opacity of the Mg II $k$ line (see Figures 1 and 5). Also shown in Figure 1 is the well-known line of He I at 1083 nm which is present in active region plasma. However, the line, belonging to the triplet system, populated largely by recombination following EUV photoionization, tends to form where coronal ionizing photons (and perhaps even hot particles) cannot penetrate the chromosphere. Our non-LTE calculations (Figure 1) reveal that significant opacity in the line is narrowly confined between surfaces of column mass well below the height of the NUV Mg II lines, because of the depth of penetration of vacuum UV radiation from the overlying corona in the continua of H and He. When included along with the Ca II resonance and infrared triplet lines, the optical- near UV regions accessible from the ground extend neither high enough, nor do they sample opacity with the fineness of UV lines (Figures 1, 2 and 5. Lines of Mn I are sufficiently close to the strong Mg II lines for the wings of the latter to contribute to opacity, but this was not accounted for in the plot. Therefore heights of formation of Mn I lines are lower limits. ). The solar irradiance spectrum at vacuum UV wavelengths (below 200 nm, and above 91 nm below which the Lyman continuum absorbs line radiation), including H L$\alpha$ and many chromospheric and transition region lines, is shown in Figure 3. All are substantially weaker than the Mg II lines. The polarimetric measurements necessary to infer magnetic fields require large photon fluxes. The VUV region is a factor of at least 10-100 times dimmer than the NUV. Also, the Zeeman signatures vary in proportion to the wavelengths observed. Further, most of the transition region lines are optically thin, so that there is no single “depth of formation” of a given line. The exceptions are the Lyman lines of hydrogen and 58.4 nm resonance line of He I. But using only this region of the spectrum would also fail to sample surfaces of multiple optical depths, and thus fail to address the need to span the anticipated corrugated isothermal and isobaric surfaces (Figure 2), particularly in active regions, the sources of most solar eruptions. Figure 4: A smoothed UV spectrum of $\alpha$ Cen A, obtained by the Hubble Space Telescope, is displayed as a proxy for the mean solar intensity spectrum. The large number of lines of Fe I and Fe II are indicated and annotated with their multiplet terms. The Mg II spectrum is also annotated including the $3s-3p$ $h$ and $k$ lines, and the $3p-3d$ lines. Notice that the permitted Fe II lines (shown with solid bars with downward ticks) consist of the sextet-D resonance lines $(4s-4p)$ near 2600 Å, and quartet transitions at other wavelengths that occur between metastable (even parity) quartet lower levels and odd-parity quartets of the same transition array ($4s-4p)$. In contrast, the NUV region (from say 200 to 310 nm, the atmospheric cutoff) is comparatively brighter, containing lines (Mg II) whose cores form within lower transition region. Compared with lines accessible from the ground, lines have smaller Zeeman signatures, but have increased linear polarization generated by the higher radiation anisotropies at UV wavelengths (much stronger limb-darkening). Thus, the Hanle signals will statistically be stronger than at visible wavelengths. Further, the region contains a plethora of lines of neutral and singly ionized iron within the $4s-4p$ transition array, including resonance lines near 260 nm (see Figure 4). Their complex spectra span and sample multiple depths across the chromosphere, extending into lower transition region plasma (see in particular the contours of unit optical depth surfaces for several lines in the lower right panel of Figure 2, and the plot of relative optical depths in Figure 5). The richness of the Fe II spectrum itself has been known for decades in astronomy (e.g. Viotti et al., 1988). Figure 4 identifies lines and multiplets of Fe plotted above a smoothed Hubble Space Telescope spectrum of $\alpha$ Cen A, which is a good proxy for disk-averaged intensity spectrum of the Sun. (There is no disk-averaged solar spectrum of comparable quality). All of the strong transitions of iron lie near or below the Mg II lines near 280 nm. NUV spectra obtained at the solar limb during the SKYLAB era revealed a spectacular array of Fe II lines in emission against the darkness of space (Doschek et al., 1977). Particularly strong in the limb spectra (Figure 10) are multiplets in the sextet system near 259-263 nm, and quartet transitions in the 273 nm region. These Fe II lines were readily detected up to 8” above the limb in the SKYLAB data, the limb itself is already far higher than the photosphere. Such emission is very likely associated with the kind of structures modeled in Figure 2. A downside of the richness of the spectra of iron is the possibility of blended lines. The important issues of blended spectra lines at NUV wavelengths and effects of missing opacity are addressed in the Appendix. They are shown to be a minor issue, affecting only a few lines in the regions between 256 and 281 nm, owing to the overwhelming strength of the Fe II lines formed high enough in the chromosphere to aid in the diagnosis of magnetism there. Between 281 and 310 nm, there are just 10 “wide” lines, i.e. strong enough to exhibit wing absorption profiles, compared with 38 listed by Moore et al. (1982) between 256 and 281 nm. None of the lines above 281 nm form high enough in the Sun’s atmosphere to contribute significantly beyond measurements of the chromospheric magnetic field at shorter wavelengths. Figure 5: Properties of spectral lines calculated using line center opacities as a function of wavelength are shown. Circles show lines of singly-charged ions, squares show lines of neutral species. Symbol sizes reflect critical Hanle field strengths. Relative optical depths of the lines are shown on the ordinate. The Mn I lines should include wing opacity from the Mg II line, but this was omitted here. The Mn I optical depths are thus a lower limit. The “?” for the He I 1083 nm line refers to the unknown relative optical depth of this line because of peculiarities in the formation of this multiplet in the Sun (e.g. Judge et al., 2015, who discuss this multiplet in the context of a well- observed flare). Relative optical depths at the centers of the lines in this region, and “Hanle field strengths” are shown in Figure 5, including lines of He, Mg, Ca, Mn and Fe. These Hanle field strengths (see equations 2 and 3 below) are those near which the emergent polarized spectra are strongly sensitive to magnetic field strength and direction, (Landi degl’Innocenti & Landolfi, 2004, see below). ### 2.4 Spectral signatures of magnetic fields The Zeeman effect is well-known in laboratory and astrophysics as a characteristic splitting in intensity ($I$) spectra. But in the Sun and stars, the Stokes parameters $I,Q,U$ and $V$ are used to measure magnetic fields in commonly encountered situations where the widths of the lines exceed the Zeeman splitting (this is assumed in the following discussion). As noted above, Zeeman signals depend upon the quantity $\varepsilon=\omega_{B}/\Delta\omega_{0}$, where $\omega_{B}$ is the Larmor frequency of the gyrating ion: $\hbar\omega_{B}=\mu_{0}B.$ (2) Here $\mu_{0}$ is the Bohr magneton, and $B$ the magnetic field strength in G. $\Delta\omega_{0}$ is the line width (Doppler width for chromospheric and coronal lines). The amplitude of circular polarization measured by Stokes $V$ varies linearly with $\epsilon$, the linear polarization profiles $Q$ and $U$ have amplitudes varying as $\varepsilon^{2}$. While $B$ is measured in Gauss, it is important to remember that unless contributions from non-magnetic regions to $I,Q,U$ and $V$ can be determined, the polarized components can be used to determine only the average flux density per unit area in Mx cm-2, not field strength. Table 1: Prominent lines and blends in the solar chromospheric spectrum $\lambda$ nm | Landé g | log $\omega_{L}$ | log $\sum_{i}A_{ji}$ | $B_{HANLE}$ | log $\tau_{0}$ | Ion | blend ---|---|---|---|---|---|---|--- lab. (air) | factor | rad sec-1 | sec-1 | G | | | severity∗ Ion $\lambda$ 256.254 | 1.21 | 7.03 | 8.49 | 28.8 | -2.54 | Fe II | 256.348 | 1.10 | 6.99 | 8.49 | 31.9 | -2.87 | Fe II | 1 Fe I .341 256.691 | 0.83 | 6.86 | 8.49 | 42.2 | -3.30 | Fe II | 257.437 | 1.30 | 7.06 | 8.51 | 28.4 | -4.33 | Fe II | 1 Co I .351 257.792 | 1.33 | 7.07 | 8.49 | 26.4 | -3.32 | Fe II | 1 Mg I .888 258.258 | 1.47 | 7.11 | 8.49 | 23.9 | -3.19 | Fe II | 258.588 | 1.50 | 7.12 | 8.46 | 21.7 | -2.03 | Fe II | 2 Fe I .588 259.055 | 1.50 | 7.12 | 8.48 | 22.8 | -4.47 | Fe II | 1 Co I .059 259.154 | 1.49 | 7.12 | 8.49 | 23.5 | -3.24 | Fe II | 259.373 | 2.17 | 7.28 | 8.49 | 16.2 | -4.02 | Fe II | 1 Mn I .372 259.837 | 1.50 | 7.12 | 8.46 | 21.7 | -1.93 | Fe II | 259.940 | 1.56 | 7.14 | 8.46 | 21.1 | -1.39 | Fe II | 1 Fe I .957 260.709 | 1.50 | 7.12 | 8.66 | 34.8 | -1.99 | Fe II | 261.107 | 1.90 | 7.22 | 8.49 | 18.4 | -4.23 | Fe II | 1 Cr II? .104 261.187 | 1.59 | 7.15 | 8.46 | 20.5 | -1.79 | Fe II | 261.382 | 1.50 | 7.12 | 8.46 | 21.7 | -2.20 | Fe II | 261.762 | 1.66 | 7.16 | 8.46 | 19.6 | -2.36 | Fe II | 262.041 | 1.87 | 7.22 | 8.66 | 27.9 | -3.75 | Fe II | 262.167 | 3.34 | 7.47 | 8.46 | 9.8 | -2.76 | Fe II | 262.567 | 1.50 | 7.12 | 8.46 | 21.9 | -2.08 | Fe II | 262.829 | 1.50 | 7.12 | 8.66 | 34.8 | -2.21 | Fe II | 263.105 | 1.50 | 7.12 | 8.46 | 21.7 | -2.02 | Fe II | 1 Fe II .045† 263.132 | 1.50 | 7.12 | 8.46 | 21.7 | -1.97 | Fe II | 264.112 | 1.87 | 7.22 | 8.51 | 19.8 | -4.62 | Fe II | 1 Ti I .089 268.351 | 1.84 | 7.21 | 8.51 | 20.1 | -4.99 | Fe II | 2 Cr II .345 269.283 | 1.93 | 7.23 | 8.47 | 17.2 | -4.40 | Fe II | 1 Fe I .265, Fe II .260 270.938 | 2.10 | 7.27 | 8.47 | 15.8 | -4.30 | Fe II | 1 Cr II .931 271.441 | 1.50 | 7.12 | 8.60 | 30.3 | -3.09 | Fe II | 271.670 | 1.33 | 7.07 | 8.46 | 24.8 | -3.11 | Fe II | 1 Mn II .680 271.903 | 1.25 | 7.04 | 8.25 | 16.3 | -4.45 | Fe I | 2 Fe I .903 272.090 | 1.17 | 7.01 | 8.28 | 18.4 | -4.76 | Fe I | 272.358 | 1.00 | 6.94 | 8.27 | 21.1 | -5.22 | Fe I | 272.488 | 1.20 | 7.02 | 8.47 | 27.7 | -3.05 | Fe II | 1 Fe I .495 272.754 | 1.50 | 7.12 | 8.60 | 30.4 | -3.05 | Fe II | 273.073 | 0.80 | 6.85 | 8.47 | 41.6 | -3.19 | Fe II | 273.245 | 1.45 | 7.10 | 8.46 | 22.8 | -4.92 | Fe II | 1 Fe II .233 273.697 | 1.50 | 7.12 | 8.60 | 30.3 | -3.22 | Fe II | 1 Fe I .696 273.731 | 2.00 | 7.25 | 8.27 | 10.5 | -5.12 | Fe I | 273.955 | 1.43 | 7.10 | 8.61 | 32.1 | -2.31 | Fe II | 274.241 | 1.67 | 7.17 | 8.28 | 12.9 | -5.01 | Fe I | 1 Ti I .230, V II .243 274.320 | 0.50 | 6.64 | 8.47 | 66.6 | -2.81 | Fe II | 274.407 | 2.50 | 7.34 | 8.27 | 8.4 | -5.51 | Fe I | 274.648 | 0.90 | 6.90 | 8.47 | 37.0 | -2.59 | Fe II | 274.698 | 1.37 | 7.08 | 8.60 | 33.2 | -2.66 | Fe II | 1 Fe I .698 274.918 | 1.20 | 7.02 | 8.60 | 38.0 | -3.02 | Fe II | 274.932 | 1.07 | 6.97 | 8.46 | 30.9 | -2.38 | Fe II | 274.949 | 0.00 | 4.31 | 8.60 | … | -3.24 | Fe II | 275.014 | 1.58 | 7.14 | 8.25 | 12.8 | -5.04 | Fe I | 275.574 | 1.17 | 7.01 | 8.46 | 28.1 | -2.17 | Fe II | 275.633 | 2.00 | 7.25 | 8.28 | 10.7 | -5.52 | Fe I | 2 Fe I .633, Cr II .630 276.181 | 1.50 | 7.12 | 8.60 | 30.4 | -3.25 | Fe II | 1 Fe I .178, Cr I .174 276.893 | 1.50 | 7.12 | 8.60 | 30.3 | -3.08 | Fe II | 277.273 | 1.50 | 7.12 | 8.61 | 30.6 | -3.14 | Fe II | 1 Fe I .283 277.534 | 1.34 | 7.07 | 8.46 | 24.2 | -6.06 | Fe II | 1 ? 279.565 | 1.17 | 7.01 | 8.43 | 25.3 | 0.00 | Mg II | 279.482 | 1.98 | 7.24 | 8.57 | 21.3 | -5.65 | Mn I | 279.827 | 1.70 | 7.17 | 8.57 | 24.8 | -5.79 | Mn I | 280.108 | 0.84 | 6.87 | 8.57 | 50.3 | -5.96 | Mn I | 280.270 | 1.33 | 7.07 | 8.43 | 21.9 | -0.30 | Mg II | 285.213 | 1.00 | 6.94 | 8.69 | 55.8 | -2.20 | Mg I | 393.366 | 1.17 | 7.01 | 8.17 | 14.3 | -1.55 | Ca II | 854.200 | 1.10 | 6.99 | 7.00 | 1.0 | -2.50 | Ca II | Landé g-factors are computed using LS coupling. The final column lists the severity of the blend (as assessed by Moore et al., 1982), and the wavelength of the blended line (decimal nm). ∗The severity is: 0=unblended; 1=moderate blend; 2=severe blend. The data in Figure 5 are taken directly from this table. †The strong line of Fe II multiplet 1 is blended with a line of multiplet 171. The Hanle effect has a quite different and complementary behavior independent of the Doppler width $\Delta\omega_{0}$. It requires a pre-existing state of polarization induced by symmetry breaking effects such as anisotropic and/or polarized incident radiation. When the radiation is anisotropic, such as radiation emerging into the upper chromosphere from the photosphere below, the atom develops atomic polarization (a particular imbalance of populations of magnetic substates called the atomic “alignment”). In the presence of a magnetic field, these populations and associated atomic coherences (essentially representing the entanglement of neighboring atomic states) are altered by the magnetic field in the regime where $g_{ji}\omega_{B}\mathcal{T}_{j}\sim 1,$ (3) where $g_{ji}$ is the effective Landé $g-$factor for the transition from upper level $j$ to lower $i$, and $\mathcal{T}_{j}=1/\sum_{i}A_{ji}$ is the inverse lifetime of level $j$, the sum taken over all lower levels $i$. Figure 5 shows spectral lines of Fe and Mg, their wavelengths, optical depths, with symbol sizes indicating the magnetic field for which equality holds in equation (3), using data compiled in the NIST spectroscopy database. These Hanle field strengths vary between roughly 1 G and 70 G for the different lines, field strengths that are anticipated in quiet and active regions of the solar chromosphere. The data shown in Figure 5 are listed in Table 1, the final column of which which also addresses the potentially serious problem of spectral line blends. This problem is shown in the Appendix to be easily dealt with. In a classical picture, when $g_{ji}\omega_{B}\mathcal{T}_{j}\approx 1$, the ion on average gyrates through 1 radian before it emits photons sharing the common upper level. When $\omega_{B}\mathcal{T}_{j}\gg 1$, the “strong-field limit” of the Hanle effect, the radiating ions exhibit many gyrations before emitting a photon and all “memory” of the particular orientation of the radiating ions is statistically averaged out. This condition is the same as that found for coronal forbidden lines where $\mathcal{T}_{j}$ is typically between 1 and 0.01 seconds, not $10^{-8}$ seconds for the permitted lines studied here (Casini et al., 2017). The only signature of the magnetic field is in its direction projected on to the plane-of-the-sky, modulo 90o. If $\omega_{B}\mathcal{T}_{i}\ll 1$ (the weak field limit) the magnetic field enters as a small perturbation in expressions for the emergent radiation and the magnetic effects cannot easily be observed. But near the Hanle regime (equation 3) the magnetic field is encoded as a rotation and de-polarization of the emergent linearly polarized radiation. The symmetry broken by the mis- alignment between the directed radiation vector and magnetic field vector resolves ambiguities associated with Zeeman measurements. This effect may be useful for identifying tangential discontinuities (changes of magnetic field direction not strength, across a flux surface). In practice, both the Zeeman and Hanle effects can be used in a complementary fashion to find signatures of magnetism in the Sun’s atmosphere using such spectra (Trujillo Bueno et al., 2018; Ishikawa et al., 2021). Finally, we have investigated whether molecular transitions are significant in this spectral region, using the line list complied by R. Kurucz111 http://kurucz.harvard.edu/linelists.html and the synthesis codes by Berdyugina et al. (2003). In particular, we have found that OH transitions from the electronic A–X system are abundant in this spectral region, with stronger lines toward the red. They form below most chromospheric lines owing to chromospheric stratification and the rapid drop of molecular number density with gas pressure, but remarkably, can also form above deeper photospheric lines. A few moderately strong OH features may be visible at wavelengths between strong atomic lines in the quiet sun spectrum. They become especially important in models of sunspot umbrae. Numerous molecular CO lines in this spectral region are quite weak, but they are so many that the continuum level is effectively reduced by a few percent, with stronger lines toward the blue. If observed, both OH and CO lines are potentially interesting for sensing magnetic fields using both the Zeeman and Hanle effects (Berdyugina & Solanki, 2002; Berdyugina et al., 2003; Berdyugina & Fluri, 2004), as well as temperature and pressure near the interface between the photosphere and chromosphere. (The reader can refer to the paper by Landi Degl’Innocenti, 2007 for elementary theory of molecules in magnetic fields applied to solar conditions). We conclude that, based on known atomic and molecular data, there are many lines of Fe II whose cores are not strongly blended. The case of blended lines from sunspot umbrae requires further work outside of the scope of this paper. ### 2.5 Diagnostic Sensitivity of Mg II k Figure 6: Observing geometry for uniform field computations with HanleRT used in the panels of Figure 7. Note the convention for negative LOS $\theta$. Figure 7: _Top:_ “Heat Map” of the Mg II k core polarization amplitude for a uniform magnetic field inclined at $\theta_{B}=-30^{\circ}$ from vertical. Gray regions are where the amplitude falls below 0.1% of the intensity, here taken as a typical instrumental SNR. _Bottom:_ Derivative of polarization amplitude data from the top panel with respect to magnetic field to produce a “heat map” of the response. The color scale gives the relative change in amplitude of the spectral feature in response to a 20% change in $|B|$. Regions of the map brighter than the 0.1% value on the color scale (highlighted by black box) indicate an observing geometry that could infer $|B|$ to within 20% assuming an instrumental polarization SNR of 0.1% of the intensity. See text for further details. In Figures 1 and 2 we saw how the Mg II $k$ line has contributions from plasma that is highest in the Sun’s atmosphere. The $h$ line is a factor of two less opaque, but it has a higher Landé g-factor than $k$ and so is more sensitive to the Zeeman effect (Ishikawa et al., 2021). However, its upper $J=1/2$ level is unpolarizable and so it can have no Hanle effect in the line core. Both the Zeeman and Hanle effects can and should be used to probe chromospheric magnetism. The Zeeman effect produces the largest polarimetric signals when fields are relatively strong (average flux densities of a few hundred Mx cm-2 were measured using the weak field limit applied to the Mg lines by Ishikawa et al., 2021). However, for weaker fields, the Hanle effect becomes the dominant source of polarized light variations in response to the solar magnetic fields. Following early theoretical work by Belluzzi & Trujillo Bueno (2012), the scattering polarization, Hanle and Zeeman effects across the $h$ and $k$ lines was investigated in semi-empirical 1D models (Alsina Ballester et al., 2016; del Pino Alemán et al., 2016, 2020). The Hanle field of the $k$ line is 23 G (see Figure 5). These authors demonstrated that, for field strengths of the order of 10 G, measurable variations of the order of $\sim$1% in the fractional linear polarization of Mg II $k$ are produced by the Hanle effect. Here, we extend this work with a larger grid of LOS angles and magnetic field strengths (see Figure 6 for the geometry) and show polarization amplitude and magnetic response “heat maps” in Figure 7. As in del Pino Alemán et al. (2020) we used HanleRT to compute the intensity and polarization from a Mg II 3-term atom including PRD effects in atmosphere C of Fontenla et al. (1993). We computed profiles for a grid of uniform magnetic field vectors inclined at $\theta_{B}=-30^{\circ}$ from vertical, with fixed azimuth, and with $|B|$ ranging from 1 to 200 G at 1 G intervals. These values of $|B|$ are about 20 and 10 times smaller and larger, respectively, than the Hanle field. We computed the emergent Stokes profiles for 13 LOS viewing angles $\theta$ ranging from $+75^{\circ}{}$ to $-75^{\circ}{}$ at 15∘ intervals, as well as $\mu=\cos(\theta)=0.1$, which corresponds to $\theta\sim\pm 84.5$∘ or a limb distance of about 5′′. We examine the relative variation in the amplitude of the fractional polarization (hereafter simply “response“), $|B|\cdot dA_{i}/dB$, where $i$ indicates the polarization $Q$, $U$, or $V$, and $A_{i}$ is the amplitude, $A_{i}=({\rm max}(i)-{\rm min}(i))/I\cdot 100\%$, within the Mg II k core region. The top panel of 7 shows the amplitude $A_{i}$, while the bottom panel shows the response with respect to magnetic field over the full range of $|B|$ and LOS $\theta$. The units of the response are $\%I/\%B$, and we have scaled the data such that the denominator represents a 20% change in $|B|$. Under this normalization, a response of 1 corresponds to a 1% amplitude change for a 20% change in $|B|$. In order to assess the practical application of these calculations, we consider noise at the level of $\sim 0.1\%$ of intensity ($10^{-3}I$). Those regions in the response heat map brighter than that value on the color scale are where the instrument would be able to infer components of the magnetic field to within 20%. We conclude that we would be able to reasonably estimate this vector orientation for magnetic field strengths of $\sim$5 to 50 G at observation angles from about $\theta=\pm 45^{\circ}$ to the limb due to the Hanle effect in $Q$ and/or $U$, and stronger line-of-sight fields at nearly any observation angle from the Zeeman effect in Stokes $V$. The largest sensitivity in $Q$ and $U$ occurs for observations near the limb and field strengths below the critical Hanle field. (Dark vertical strips in the response of Stokes $U$ are due to local maxima in the amplitude curve near the critical Hanle field for Mg II $k$; compare to the top panel of Figure 7). The $V$ amplitude is constant and zero at $\theta=60^{\circ}$ as this is the angle where the magnetic field vector is orthogonal to the line-of-sight. Measurable linear polarization is produced at that geometry, however its variation with magnetic field is too low to be used as a diagnostic. “Heat map” diagrams for other field inclination angles are qualitatively similar, with the regions of maximum response due to the Hanle and Zeeman effects shifting in relation to the field & LOS geometry. We conclude that diagnosing magnetic field from the Hanle effect in Mg II $k$ is feasible for a large range of geometries and with achievable signal-to- noise ratios. The range of field strengths to which the Hanle effect is sensitive complement those higher field strengths ($|B|\geq 50$G) to which the Zeeman effect is sensitive, for the case of the Mg II $k$ line. Additional work on the Hanle effect in the lines of Fe II is underway and will be reported elsewhere. Importantly, there is at least one line (at 2744.01 Å) with a Hanle field above 60 G, and several where the critical field exceeds that of the Mg II $k$ line. ### 2.6 Addressing the “corrugation problem” Figure 2 shows how state-of-the-art numerical models lead to contours in optical depth which can be far from horizontal. The free energy in these MHD simulations generated by convection and emerging flux leads to highly dynamic evolution in the lower density regions of most interest to us at the top of the chromosphere. The geometric distortions of the $\tau=1$ surfaces in these tenuous regions naturally reflect the low inertia of the plasma, subjected to Lorentz, pressure gradient and gravitational forces, evolving in response to the forcing from below. But, as noted earlier, the extreme corrugations of order 10 Mm are not negligible compared with the scale of the emerging flux system itself. The question then arises, how can we infer magnetic field at a constant geometric height, if we are to apply mathematical techniques to estimate the free energy and topology within the corona? Progress can be anticipated without such techniques, by direct comparison of numerical prediction and measurements. But this question must be addressed if we are to measure the development of free energy in the overlying corona in time, in general. Let us assume that the calculations are a reasonable, “first order” approximation to the kind of geometric distortions of the $\tau=1$ surfaces of sets of lines. The advantage of observing the 259-281 nm region is that we know the order of the opacities of each line in a sequence: the Mg II lines are the most opaque, next come the sextet system transitions of Fe II and then the quartet systems, each with its spread of oscillator strengths. Thus, given a calculation, we can (as in Figure 2) find the $\tau=1$ surfaces. We could compute Stokes profiles and associate a magnetic field from such synthetic data, knowing the heights of each line’s formation. But this is not the situation we face with the remotely-sensed Stokes data, we have no knowledge other than the Stokes profiles of many lines, on a two dimensional geometric image of the Sun at each wavelength. Fortunately, this problem fits well into the picture of machine learning algorithms (e.g. Bobra & Mason, 2021). Synthetic data can be used to train an algorithm to find level geometric surfaces (Asensio Ramos & Díaz Baso, 2019) of direct use for mathematics to derive quantities of interest. The large coverage of optical depths of the 256-281 nm region and the variety of Hanle field strengths strongly suggest that, if the models are sufficiently complete to represent the atmospheric dynamics and temperature structure, then such an algorithm should have a non-zero probability of yielding the desired results. The combined use of Hanle and Zeeman effects enhances this probability, at least in principle. We are exploring ways in which this program of research might be implemented. ## 3 Discussion In this paper we have examined the general problem of identifying the sources of magnetic energy within coronal plasma, which leads to flaring, plasmoid and CME ejections. By a process of elimination, we arrived at a wavelength range between 2560 and 2810 Å, which appears to be the most promising range in which to measure conditions at the coronal base. Our conclusions build upon the important study of Trujillo Bueno et al. (2017), not only by seeking answers to questions concerning the magnetic free energy and topology, aimed at understanding the origins of phenomena responsible for Space Weather. Our conclusions also differ by proposing a potentially ideal, relatively narrow spectral region between 256 and 281 nm, expanding significantly the 279.2-280.7 region selected by the CLASP2 mission, to serve as a basis for future, more targeted research. To achieve measurements in this region clearly will require a space-based platform for spectropolarimetry. These conclusions are based upon several criteria: The spectrum must be bright enough; spectral features must extend as high as possible; the spectral features must contain information on magnetic fields through both the Zeeman and Hanle effects; the spectra must be demonstrably unblended; lines with multiple opacities are needed to span across the corrugated surfaces that are present at the coronal base. Thus we suggest that a novel spectropolarimeter, building upon a heritage from SKYLAB, SMM (UVSP), SoHO (SUMER), IRIS, CLASP and CLASP2, be considered for flight to address the problems of interest to science and society. Such a mission would complement the unique capabilties of the Daniel K. Inouye Solar Telescope which can observe important chromospheric lines (include all those shown in the top row of Figure 1) from the ground, at a sub-arcsecond spatial resolution and/or high cadence. Finally, it is important to note that the CLASP2 measurements of Ishikawa et al. (2021) achieved signal-to-noise ratios of over 1000 in the Mg II line cores. This demonstrates that a modest, D=25 cm telescope operating at NUV wavelengths is sufficient for delivering the weak signals required for diagnosis of magnetic fields, over active regions at least, where the magnetic flux density exceeds 100 Mx cm-2. Thanks to Roberto Casini and Alfred de Wijn for the many fruitful discussions regarding the Hanle effect and requirements to achieve new results using diverse spectral lines. NCAR is funded by the National Science Foundation. PGJ is grateful to the NSF for funding his research at a national center without any specific goal. The authors thank an anonymous referee for their helpful comments. ## Line blends and missing opacity The NUV region is well-known as a crowded region of the solar spectrum, and also a region where some opacity appears to be “missing”. The effects of this missing opacity are examined below using comparisons of detailed simulations and observations. It can be significant at certain wavelengths, but does it not affect the cores of strong chromospheric lines. (See also discussions by Fontenla et al., 2011; Peterson et al., 2017). Line blending is however a concern. We have addressed blending using data from the Hubble Space Telescope, from the SO82B spectrograph on SKYLAB, and using sophisticated radiative transfer models. The models are based on Local Thermodynamic Equilibrium (LTE) radiative transfer models computed using TURBOSPECTRUM (Plez, 2012), and adopting a custom calculated Solar 1D, spherical MARCS model atmosphere (Gustafsson et al., 2008). A complete linelist in the corresponding spectral region was adopted in the computations from the VALD database222http://vald.astro.uu.se/ (Ryabchikova et al., 2011). The linelist includes up-to-date atomic data, including line-broadening parameters from collisions with neutral hydrogen atoms, calculated ab-initio following Barklem et al. (1998) for all neutral atoms including dominant Mg II lines and iron group elements. A comparison of high resolution Hubble spectra of $\alpha$ Cen A with the models is informative. A typical example in an interesting region of the NUV spectrum is shown in Figure 8. The blue line represents the flux spectrum computed for the Sun, the red line the spectrum of $\alpha$ Cen A. The flux densities shown are computed at the respective stellar surfaces, with no ad- hoc adjustments of either scale. Figure 8: Flux spectra computed for a Solar atmospheric model (blue) using 1D, LTE radiative transfer calculations (see Section Line blends and missing opacity) are shown along with spectra of the Sun-like star $\alpha$ Cen A (red) obtained with the STIS instrument on the HST. The data are a composite of several exposures from the E230H grating, assembled and processed by Ayres (2014). The STIS data were shifted according to the redshift of 22.4 km s-1, and converted to flux at the stellar surface using the angular diameter of 9.09 milli-arcsec. The spectra are compared with no adjustment of flux or wavelength scale. The qualitative agreement is remarkable. The units are Å (wavelength) and erg cm-2 s-1 Å-1 (flux, $F=\pi\overline{I}$, where $\overline{I}$ is the mean disk intensity). The comparison is remarkable, particularly in the cores of some of the strong lines, most of which belong to Fe II. Wavelengths are evident (near 2623 Å for example) where opacity appears to be missing (blue line above red line), but in the Fe II lines identified in the figure there is no evidence for missing opacity, if anything, the observed line cores in red lie above the blue line, indicating emission from the star’s chromosphere. The model calculations also permit us to identify possible blends (Figure 9). This figure shows two computations, black is the full model calculation (also shown in blue in Figure 8), and in this figure blue is a calculation using opacity only from the Fe II lines. When significant differences between black and blue lines exist, these are wavelengths where opacities are present in lines other than Fe II. Conversely, when they agree, the Fe II spectrum dominates and these are essentially unblended wavelengths. As noted in the main text, the wavelength region above 281 and below 310 nm does not contain strong lines formed in the middle-high chromosphere, therefore these are not shown here. Above 281 nm the lack of such lines makes this a less desirable region because blending becomes a bigger problem. Figure 9: Two calculations are shown for the solar NUV spectrum: one (black) is made with the full set of lines, and the other (red) is made where lines of Fe II are the only sources of opacity. Strong lines from the quartet system of Fe II are annotated, along with lines of Fe I whose opacity is much smaller. The cores of the strong lines of Fe II are dominated by Fe II itself, the region above $\approx$ 265 nm is, aside from narrow lines, dominated by wings of the Mg II $h$ and $k$ lines. Lastly we can, to some extent, check for blends independently of the model computations. Rutten & Stencel (1980) were able to find weak lines in the wings of the Ca II $H$ and $K$ lines looking to data obtained near the solar limb. In Figure 10 we show (with an arbitrary flux calibration), spectra of $\alpha$ Cen A with spectra from the SO82B spectrograph on SKYLAB. The latter are taken from the scans across a quiet region of solar limb obtained on August 27 1973, the slit being tangential to the limb and stepped by 2” between pointings. These are the photographic data analyzed by Doschek et al. (1977). The “intensities” plotted here are photographic densities. But these densities suffice for us to find significant blends, which will potentially be revealed by changes in the spectra as the spectrograph slit was repointed from 2” inside the limb to 6” outside, in 2” steps. Figure 10: The left panel shows the spectral region close to the sextet resonance lines of Fe II is shown from $\alpha$ Cen A (top), and 5 spectra representative of the spectra as a function of position relative to the solar limb. The right panel shows the 2765-2780 Å region. Blue lines mark wavelengths of Fe II transitions. The three other lines seen in emission in the right panel are from Cr II at 2766.52 Å, and questionable identifications of Fe II 2767.50, a transition between high lying states, and of Mg I at 2779.8 Å (Doschek et al., 1977). Different behaviors are seen depending on the line of interest. In the wavelength range from 2595 to 2615 Å (left panel, showing resonance lines of the sextet system) there is no evidence of blended features. But the right panel of Figure 10 shows a spectral region containing lines of the quartet system of Fe II. In contrast to the resonance sextet transitions, here we find that essentially no Fe II lines are strong enough to be unblended. However, a plot of similar transitions (not shown) between 2750 and 2765 Å has stronger lines at 2756 and 2762 Å which are not strongly blended. These conclusions have been verified using NRL report 8653 (Moore et al., 1982), from which the last column of Table 1 has been constructed. There are only four lines which are classified as severely blended of all the quartets, sextets of Fe II and quintets of Fe I. ## References * Allen (1973) Allen, C. W. 1973, Astrophysical quantities (Athlone Press, Univ. London) * Alsina Ballester et al. (2016) Alsina Ballester, E., Belluzzi, L., & Trujillo Bueno, J. 2016, ApJL, 831, L15, doi: 10.3847/2041-8205/831/2/L15 * Asensio Ramos & Díaz Baso (2019) Asensio Ramos, A., & Díaz Baso, C. J. 2019, A&A, 626, A102, doi: 10.1051/0004-6361/201935628 * Athay & Dere (1991) Athay, R. G., & Dere, K. P. 1991, ApJ, 379, 776, doi: 10.1086/170553 * Auer et al. (1980) Auer, L. H., Rees, D. E., & Stenflo, J. O. 1980, A&A, 88, 302 * Ayres (2014) Ayres, T. R. 2014, in Astronomical Society of India Conference Series, Vol. 11, Astronomical Society of India Conference Series, 1 * Barklem et al. (1998) Barklem, P. S., Anstee, S. D., & O’Mara, B. J. 1998, PASA, 15, 336, doi: 10.1071/AS98336 * Barnes et al. (2016) Barnes, G., Leka, K. D., Schrijver, C. J., et al. 2016, ApJ, 829, 89, doi: 10.3847/0004-637X/829/2/89 * Belluzzi & Trujillo Bueno (2012) Belluzzi, L., & Trujillo Bueno, J. 2012, ApJL, 750, L11, doi: 10.1088/2041-8205/750/1/L11 * Berdyugina & Fluri (2004) Berdyugina, S. V., & Fluri, D. M. 2004, A&A, 417, 775, doi: 10.1051/0004-6361:20034452 * Berdyugina & Solanki (2002) Berdyugina, S. V., & Solanki, S. K. 2002, A&A, 385, 701, doi: 10.1051/0004-6361:20020130 * Berdyugina et al. (2003) Berdyugina, S. V., Solanki, S. K., & Frutiger, C. 2003, A&A, 412, 513, doi: 10.1051/0004-6361:20031473 * Bird (1982) Bird, M. K. 1982, Space Sci. Rev., 33, 99, doi: 10.1007/BF00213250 * Bird & Edenhofer (1990) Bird, M. K., & Edenhofer, P. 1990, Physics and Chemistry in Space, Vol. 20, Physics of the Inner Heliosphere I. Large-Scale Phenomena. (Berlin Heidelberg: Springer-Verlag), 13–97 * Bobra & Mason (2021) Bobra, M., & Mason, J. 2021, ”Machine Learning, Statistics, and Data Mining for Heliophysics”, doi: https://doi.org/10.5281/zenodo.1412824. * Bobra & Couvidat (2015) Bobra, M. G., & Couvidat, S. 2015, The Astrophysical Journal, 798, 135, doi: 10.1088/0004-637x/798/2/135 * Carmichael (1964) Carmichael, H. 1964, NASA Special Publication, 50, 451 * Casini et al. (2017) Casini, R., White, S. M., & Judge, P. G. 2017, SSR, 210, 145 * Cauzzi et al. (2008) Cauzzi, G., Reardon, K. P., Uitenbroek, H., et al. 2008, A&A, 480, 515, doi: 10.1051/0004-6361:20078642 * Centeno et al. (2010) Centeno, R., Trujillo Bueno, J., & Asensio Ramos, A. 2010, ApJ, 708, 1579, doi: 10.1088/0004-637X/708/2/1579 * Chandrasekhar (1961) Chandrasekhar, S. 1961, Hydrodynamic and hydromagnetic stability * Chen et al. (2019) Chen, Y., Manchester, W. B., Hero, A. O., et al. 2019, Space Weather, 17, 1404, doi: 10.1029/2019SW002214 * De Pontieu et al. (2020) De Pontieu, B., Martínez-Sykora, J., Testa, P., et al. 2020, ApJ, 888, 3, doi: 10.3847/1538-4357/ab5b03 * del Pino Alemán et al. (2016) del Pino Alemán, T., Casini, R., & Manso Sainz, R. 2016, ApJL, 830, L24 * del Pino Alemán et al. (2020) del Pino Alemán, T., Trujillo Bueno, J., Casini, R., & Manso Sainz, R. 2020, ApJ, 891, 91 * Deshmukh et al. (2020) Deshmukh, V., Berger, T., Meiss, J., & Bradley, E. 2020, arXiv e-prints, arXiv:2012.14405 * Díaz Baso et al. (2019) Díaz Baso, C. J., Martínez González, M. J., & Asensio Ramos, A. 2019, A&A, 625, A128, doi: 10.1051/0004-6361/201834790 * Doschek et al. (1977) Doschek, G. A., Feldman, U., & Cohen, L. 1977, ApJS, 33, 101, doi: 10.1086/190421 * Eddy (2009) Eddy, J. A. 2009, The Sun, the Earth and Near-Earth Space: A Guide to the Sun-Earth System (NASA) * Fleishman et al. (2017) Fleishman, G. D., Anfinogentov, S., Loukitcheva, M., Mysh’yakov, I., & Stupishin, A. 2017, ApJ, 839, 30, doi: 10.3847/1538-4357/aa6840 * Fontenla et al. (1993) Fontenla, J. M., Avrett, E. H., & Loeser, R. 1993, ApJ, 406, 319 * Fontenla et al. (2011) Fontenla, J. M., Harder, J., Livingston, W., Snow, M., & Woods, T. 2011, Journal of Geophysical Research (Atmospheres), 116, D20108, doi: 10.1029/2011JD016032 * Fox et al. (2016) Fox, N. J., Velli, M. C., Bale, S. D., et al. 2016, Space. Sci. Rev., 204, 7, doi: 10.1007/s11214-015-0211-6 * Gandorfer (2000) Gandorfer, A. 2000, The Second Solar Spectrum: A high spectral resolution polarimetric survey of scattering polarization at the solar limb in graphical representation. Volume I: 4625 Å to 6995 Å * Gelfreikh (1994) Gelfreikh, G. B. 1994, in IAU Colloq. 144, Solar Coronal Structures, ed. V. Rusin, P. Heinzel, & J.-C. Vial (Bratislava, Slovakia: VEDA Publishing Company), 21–28 * Gold & Hoyle (1960) Gold, T., & Hoyle, F. 1960, MNRAS, 120, 89 * Gudiksen et al. (2011) Gudiksen, B. V., Carlsson, M., Hansteen, V. H., et al. 2011, A&A, 531, A154 * Gustafsson et al. (2008) Gustafsson, B., Edvardsson, B., Eriksson, K., et al. 2008, A&A, 486, 951, doi: 10.1051/0004-6361:200809724 * Hansteen et al. (2014) Hansteen, V., De Pontieu, B., Carlsson, M., et al. 2014, Science, 346, 1255757, doi: 10.1126/science.1255757 * Hirayama (1974) Hirayama, T. 1974, SoPh, 34, 323 * Ishikawa et al. (2021) Ishikawa, R., Bueno, J. T., del Pino Alemán, T., et al. 2021, Science Advances, 7, doi: 10.1126/sciadv.abe8406 * Jonas et al. (2018) Jonas, E., Bobra, M. G., Shankar, V., Hoeksema, J. T., & Recht, B. 2018, Solar Physics, 293, 1 * Judge et al. (2001) Judge, P., Casini, R., Tomczyk, S., Edwards, D. P., & Francis, E. 2001, Coronal Magnetogmetry: a feasibility study, Tech. Rep. NCAR/TN-446-STR, National Center for Atmospheric Research * Judge (2020) Judge, P. G. 2020, The Sun: A Very Short Introduction (Oxford University Press) * Judge (2021) Judge, P. G. 2021, ApJ, in press * Judge et al. (2013) Judge, P. G., Habbal, S., & Landi, E. 2013, SoPh, 288, 467 * Judge et al. (2012) Judge, P. G., Kleint, L., Casini, R., & Schad, T. 2012, in American Astronomical Society Meeting Abstracts, Vol. 220, American Astronomical Society Meeting Abstracts #220, 521.19 * Judge et al. (2014) Judge, P. G., Kleint, L., Donea, A., Sainz Dalda, A., & Fletcher, L. 2014, ApJ, 796, 85 * Judge et al. (2015) Judge, P. G., Kleint, L., & Sainz Dalda, A. 2015, ApJ, 814, 100, doi: 10.1088/0004-637X/814/2/100 * Judge et al. (2011) Judge, P. G., Tritschler, A., & Low, B. C. 2011, ApJL, 730, L4, doi: 10.1088/2041-8205/730/1/L4 * Kano et al. (2017) Kano, R., Trujillo Bueno, J., Winebarger, A., et al. 2017, ApJL, 839, L10, doi: 10.3847/2041-8213/aa697f * Kleint & Gandorfer (2017) Kleint, L., & Gandorfer, A. 2017, Space Science Reviews, 210, 397, doi: 10.1007/s11214-015-0208-1 * Kopp & Pneuman (1976) Kopp, R., & Pneuman, G. 1976, SoPh, 50, 85 * Korsós et al. (2020) Korsós, M. B., Georgoulis, M. K., Gyenge, N., et al. 2020, ApJ, 896, 119, doi: 10.3847/1538-4357/ab8fa2 * Kramar et al. (2006) Kramar, M., Inhester, B., & Solanki, S. K. 2006, A&A, 456, 665, doi: 10.1051/0004-6361:20064865 * Kuckein et al. (2012) Kuckein, C., Martínez Pillet, V., & Centeno, R. 2012, A&A, 539, A131 * Landi Degl’Innocenti (2007) Landi Degl’Innocenti, E. 2007, A&A, 461, 1, doi: 10.1051/0004-6361:20042232 * Landi degl’Innocenti & Landolfi (2004) Landi degl’Innocenti, E. L., & Landolfi, M. 2004, Astrophysics and Space Library, Vol. 307, Polarization in Spectral Lines (Dordrecht: Kluwer Academic Publishers) * Leka et al. (2019) Leka, K. D., Park, S.-H., Kusano, K., et al. 2019, ApJs, 243, 36, doi: 10.3847/1538-4365/ab2e12 * Li et al. (2017) Li, W., Jönsson, P., Brage, T., & Hutton, R. 2017, Phys. Rev. A., 96, 052508, doi: 10.1103/PhysRevA.96.052508 * Li et al. (2020) Li, X., Zheng, Y., Wang, X., & Wang, L. 2020, ApJ, 891, 10, doi: 10.3847/1538-4357/ab6d04 * Liu et al. (2019) Liu, H., Liu, C., Wang, J. T. L., & Wang, H. 2019, ApJ, 877, 121, doi: 10.3847/1538-4357/ab1b3c * Loukitcheva (2014) Loukitcheva, M. 2014, in 40th COSPAR Scientific Assembly, Vol. 40, E2.2–28–14 * Loukitcheva (2020) Loukitcheva, M. 2020, Frontiers in Astronomy and Space Sciences, 7, 45, doi: 10.3389/fspas.2020.00045 * Low (1994) Low, B. C. 1994, Physics of Plasmas, 1, 1684, doi: 10.1063/1.870671 * Moore et al. (1982) Moore, C. E., Tousey, R., & Brown, C. M. 1982, The solar spectrum 3069 A - 2095 A. From the Echelle Spectrograph flown in 1961 and 1964 (Naval Research Laboratory) * Nishizuka et al. (2018) Nishizuka, N., Sugiura, K., Kubo, Y., Den, M., & Ishii, M. 2018, ApJ, 858, 113, doi: 10.3847/1538-4357/aab9a7 * Panos & Kleint (2020) Panos, B., & Kleint, L. 2020, ApJ, 891, 17, doi: 10.3847/1538-4357/ab700b * Park et al. (2018) Park, E., Moon, Y.-J., Shin, S., et al. 2018, ApJ, 869, 91, doi: 10.3847/1538-4357/aaed40 * Peterson et al. (2017) Peterson, R. C., Kurucz, R. L., & Ayres, T. R. 2017, ApJs, 229, 23, doi: 10.3847/1538-4365/aa6253 * Plez (2012) Plez, B. 2012, Turbospectrum: Code for spectral synthesis. http://ascl.net/1205.004 * Rempel (2017) Rempel, M. 2017, ApJ, 834, 10, doi: 10.3847/1538-4357/834/1/10 * Robustini et al. (2018) Robustini, C., Leenaarts, J., & de la Cruz Rodríguez, J. 2018, A&A, 609, A14, doi: 10.1051/0004-6361/201731504 * Rutten & Stencel (1980) Rutten, R. J., & Stencel, R. E. 1980, A&AS, 39, 415 * Ryabchikova et al. (2011) Ryabchikova, T. A., Pakhomov, Y. V., & Piskunov, N. E. 2011, Kazan Izdatel Kazanskogo Universiteta, 153, 61 * Schad et al. (2016) Schad, T. A., Penn, M. J., Lin, H., & Judge, P. G. 2016, ApJ, 833, 5, doi: 10.3847/0004-637X/833/1/5 * Solanki et al. (2003) Solanki, S. K., Lagg, A., Woch, J., Krupp, N., & Collados, M. 2003, Nature, 425, 692, doi: 10.1038/nature02035 * Sturrock (1966) Sturrock, P. A. 1966, Nature, 211, 695 * Trujillo Bueno & Asensio Ramos (2007) Trujillo Bueno, J., & Asensio Ramos, A. 2007, ApJ, 655, 642, doi: 10.1086/509795 * Trujillo Bueno et al. (2017) Trujillo Bueno, J., Landi Degl’Innocenti, E., & Belluzzi, L. 2017, Space Science Reviews, 210, 183, doi: 10.1007/s11214-016-0306-8 * Trujillo Bueno et al. (2018) Trujillo Bueno, J., Štěpán, J., Belluzzi, L., et al. 2018, ApJL, 866, L15, doi: 10.3847/2041-8213/aae25a * Uitenbroek (2001) Uitenbroek, H. 2001, ApJ, 557, 389 * Uitenbroek (2010) —. 2010, Mem. Soc. Astron. Italia, 81, 701 * Viotti et al. (1988) Viotti, R., Vittone, A., & Friedjung, M. 1988, Physics of formation of FE II lines outside LTE, Vol. 138, doi: 10.1007/978-94-009-4023-9 * Štěpán & Trujillo Bueno (2010) Štěpán, J., & Trujillo Bueno, J. 2010, ApJL, 711, L133, doi: 10.1088/2041-8205/711/2/L133 * Wang et al. (2016) Wang, Y., Su, Y., Hong, Z., et al. 2016, ApJ, 833, 250, doi: 10.3847/1538-4357/833/2/250 * White (2004) White, S. M. 2004, Coronal Magnetic Field Measurements Through Gyroresonance Emission, ed. D. E. Gary & C. U. Keller, Vol. 314, 89, doi: 10.1007/1-4020-2814-8_5 * Wiegelmann & Sakurai (2021) Wiegelmann, T., & Sakurai, T. 2021, Living Reviews in Solar Physics, 18, 1, doi: 10.1007/s41116-020-00027-4
# A construction of deformations to general algebras Dave Bowman, Dora Puljić, Agata Smoktunowicz ###### Abstract One of the questions investigated in deformation theory is to determine to which algebras can a given associative algebra be deformed. In this paper we investigate a different but related question, namely: for a given associative finite-dimensional $\mathbb{C}$-algebra $A$, find algebras $N$ which can be deformed to $A$. We develop a simple method which produces associative and flat deformations to investigate this question. As an application of this method we answer a question of Michael Wemyss about deformations of contraction algebras. ## Introduction Michael Wemyss and Will Donovan developed a method for characterising commutative rings which appear in algebraic geometry, by using methods from noncommutative ring theory. In [DW16] they introduced contraction algebras which provide important insight into resolutions of noncommutative singularities and invariants of flops. Contraction algebras can be described in a purely algebraic way, by using generators and relations. They appear in many questions which can be investigated by noncommutative ring theorists who are not familiar with advanced methods of algebraic geometry. Some of these questions are related to nilpotent rings. In 2022, Gavin Brown and Michael Wemyss described all finite dimensional contraction algebras which are local [BW22]. Notice that the Jacobson radical of a finite dimensional algebra is nilpotent, and hence contraction algebras which are local are very close to being nilpotent. Some other questions involve characterisation of contraction algebras which are not local. Deformations of contraction algebras give insight into invariants of flops. In this context Michael Wemyss asked questions about deformations of contraction algebras which are local. Using geometric methods he was able to conjecture which rings can be obtained as deformations of these contraction algebras. Since contraction algebras are noncommutative, the usual methods using derivations as in [CGG89] cannot be applied. Another approach can be to use Hochschild cohomology and the Gerstenhaber bracket, however this often leads to complicated calculations as seen in [SW13]. In this paper, we develop a method to calculate deformations of noncommutative local algebras. In particular we answer a question of Michael Wemyss regarding deformations of contraction algebras. For information about contraction algebras we refer reader to [Wem23, DW16]. In [HT18, Conjecture 4.3] Hua and Toda conjectured that there exists an algebraic flat deformation of the contraction algebra. As an application we confirm this conjecture for one of the contraction algebras constructed in [BW22]. Note that our method can be applied to other contraction algebras described in [BW22]. Observe that Theorem $4.2$ of [HT18] gives important information about deformations of contraction algebras which were obtained by using geometric methods. More information related to the existence of geometric deformations is given on pages 7-9 in [Tod15]. In [SW13] deformations of graded Hecke algebras were constructed using novel methods. It is known that deformations of graded Hecke algebras give important information about the Hecke algebras. Our method could be also used as another method of choice for constructing deformations of graded Hecke algebras. Some other methods of constructing deformations of associative algebras were described in [Wit19]. ## 1 Preliminaries ###### Definition 1.1. Let $S$ be a subset of a $k$-algebra $A$ for a commutative unital ring $k$. Then $S$ is a generating set if all elements of $A$ can be written as a $k$-linear sum of products of elements of $S$ using the operations in $A$. ###### Definition 1.2 ([Wit19]). A formal deformation $(A_{t},\ast)$ of a $k$-algebra $A$ is an associative $k$-bilinear multiplication $\ast$ on the $k[t]$-module $A[t]$, such that in the specification at $t=0$, the multiplication corresponds to that on $A$; this multiplication is required to be determined by a multiplication on elements of $A$ and extended to $A[t]$ by the Cauchy product rule. Any multiplication $\ast$ as in Definition 1.2 is determined by products of pairs of elements of $A$: for $a,b\in A$, write $a\ast b=\sum_{i\geq 0}\mu_{i}(a\otimes b)t^{i},$ where $\mu_{0}$ denotes the usual product in $A$ and $\mu_{i}$ for $i\geq 1$ are $k$-linear functions from $A\otimes A$ to $A$. ###### Definition 1.3 ([FO19]). Let $A$ be a $k$-algebra and for $t\in\mathbb{C}$ let $\\{A_{t}\\}$ be the family of algebras arising from a deformation $*$ of $A$. Then $*$ is a flat deformation if each $A_{t}$ has the same dimension. ###### Notation 1.4. We will denote by $\mathbb{C}[x,y]$ the polynomial ring over $\mathbb{C}$ in two non-commuting variables, $x$ and $y$. ###### Notation 1.5. Consider monomials of the ring $\mathbb{C}[x,y]$. We can order the monomials using shortlex ordering by specification $1<x<y$. We denote the monomials of $\mathbb{C}[x,y]$ by $p_{1},p_{1},p_{2}\dots$ where $p_{i}<p_{j}$ for $i<j$. ###### Notation 1.6. We will denote by $\mathbb{C}[x,y][t]$ the polynomial ring in variable $t$ with coefficients from the polynomial ring in two non-commuting variables $x$ and $y$, with coefficients from $\mathbb{C}$. ###### Notation 1.7. For $a_{i},b_{i}\in A$ we define $f:\mathbb{C}[x,y][t]\rightarrow A[t]$ to be a homomorphism of $\mathbb{C}$-algebras such that $f:x\mapsto\sum_{i=1}^{n}a_{i}t^{i},\quad f:y\mapsto\sum_{i=1}^{n}b_{i}t^{i},\quad f:t\mapsto t.$ We will denote $\ker f$ by $I$. ###### Notation 1.8. For an element $p_{i}$ in $\mathbb{C}[x,y]$ we will denote by $\overline{p_{i}}$ the same element $p_{i}$, but with all instances of $x$ replaced by $\sum_{i=1}^{n}a_{i}t^{i}$, and all instances of $y$ replaced by $\sum_{i=1}^{n}b_{i}t^{i}$. We will denote by $\overline{\overline{p_{i}}}$ the same element as $p_{i}$, but with all instances of $x$ replaced by $\sum_{i=1}^{n}a_{i}$, and all instances of $y$ replaced by $\sum_{i=1}^{n}b_{i}$. ###### Notation 1.9. Let $R$ be a ring and $I$ be an ideal of $R$, then elements of the factor ring $R/I$ will be denoted as $r+I$, where $r\in R$. Notice that $r+I=s+I$ for $r,s\in R$ if and only if $r-s\in I$. ## 2 Algorithm In this section we shall describe the simplest case of our proposed method. This requires the data of a $\mathbb{C}$-algebra $A$ and a two element generating set $\\{a,b\\}$. We shall state the steps of this algorithm and then provide notes regarding important components. Then some examples will be displayed. ###### Method 1. We fix a finite dimensional $\mathbb{C}$-algebra $A$ with two generators $a,b$. Then: 1. 1. We consider $A[t]$, the polynomial algebra over $A$. We will use $t$ as the parameter of our deformation. We define: $\displaystyle x$ $\displaystyle:=ta,$ $\displaystyle y$ $\displaystyle:=tb.$ 2. 2. We calculate: $\displaystyle x^{2}$ $\displaystyle=t^{2}a^{2},$ $\displaystyle xy$ $\displaystyle=t^{2}ab,$ $\displaystyle yx$ $\displaystyle=t^{2}ba,$ $\displaystyle y^{2}$ $\displaystyle=t^{2}b^{2},$ and continue to calculate larger products of $x,y$. We shall then proceed with a Diamond Lemma111See section 1 of [Ber78] like decomposition of large products of $a,b$ in terms of smaller products of $a,b$. This will cause all elements of sufficiently large length to have a power of $t$ as a factor. In doing so we will obtain relations on $x,y$ and products thereof. We terminate this process when we have enough relations to decompose any large product into only multiples of $x,y$ and powers of $t$. Our relations will be given by polynomials $p_{1},\dots,p_{m}\in\mathbb{C}[x,y][t]$ for some $m\in\mathbb{N}.$ This terminates finitely because $A$ finite dimensional. 3. 3. We present the algebra: $\mathcal{N}:={\mathbb{C}[x,y][t]}/{\left<p_{1},\dots,p_{m}\right>}.$ We note that sufficiently large products of $x,y$ will obtain a factor of $t$. Now we evaluate $t$ at various values in $T=[0,\infty)$. We denote by $N_{s}$ the algebra that arises from $\mathcal{N}$ by evaluating $t$ at $s$ and in particular we write $N=N_{0}$. By step (2) $N$ is local, in Section 5 we consider a more general method where $N$ is not necessarily local. The algebra $\mathcal{N}$ and the family of algebras $\\{N_{t}\\}$ are the output of this method. By Theorem 3.8 the family $\\{N_{t}\\}$ is a deformation of $N$ and by Proposition 3.9 it holds that $A\in\\{N_{t}\\}$. Changing the chosen generators $a,b$ can result in different algebras $N_{0}$ that have $A$ as an associative deformation. ### 2.1 Notes on the method We give some notes before the examples: * • In step (1) we are trying to capture the behaviour of the elements in $A$, but in a way that is controlled by $t$. This enables us to make use of a presentation while still having access to our deformation parameter $t$. * • We note that since $a,b$ generate $A$ some linear combination of products of $a,b$ will be the identity. In step (2) it is important we find a relation on $x,y,t$ and $1\in A$. * • In step (3) we discard $A$ entirely. However, because of the relations $p_{i}$, $x$ and $y$ “remember” that they came from $A$ and that they form a generating set. This step results in something subtle: the monomials $x,y\in N$ are no longer dependent on $t$ (and thus do not vanish when we set $t=0$) but their products maintain their dependence on $t$. We have loosened our grip on $x,y$ just enough for the deformation to take place. * • We note that our deformation will be associative since the multiplication of $N_{t}$ will inherit from the associative multiplication $*$ of the algebra $\mathcal{N}$. * • This algorithm is easily generalised to algebras that require a larger generating set. Suppose a finite dimensional $\mathbb{C}$-algebra $A$ has $n$ generators $\\{a_{1},\dots,a_{n}\\}$. Then in step (1) we would define $x_{i}:=ta_{i}$ for $1\leq i\leq n$ and proceed as described in Method 1. The resulting family of algebras will also be a flat and associative deformation of $A$. ### 2.2 Examples Here we shall work through some examples. We shall start with a simple example where $A=\mathbb{C}\oplus\mathbb{C}$ and then give a more complicated one where $A=M_{2}(\mathbb{C})$, the algebra of $2\times 2$ matrices over $\mathbb{C}$. Practically it can be useful to fix a basis for the algebra $A$ as this helps find the decompositions in step (2). ###### Example 2.1. We consider $A=\mathbb{C}\oplus\mathbb{C}$ and fix $a=(i,0),\;b=(0,1)$. Clearly $\mathbb{C}$-linear combinations of $a$ and $b$ span $A$ so our basis is just $\\{a,b\\}$. We write $\mathbbm{1}:=(1,1)$. Now we consider $A[t]$ and define: $\displaystyle x:=ta=(ti,0),$ $\displaystyle y:=tb=(0,t).$ We calculate: $\displaystyle x^{2}=t^{2}a^{2}=(-t^{2},0)=tix,$ $\displaystyle xy=t^{2}ab=(0,0)=0,$ $\displaystyle yx=t^{2}ab=(0,0)=0,$ $\displaystyle y^{2}=t^{2}b^{2}=(0,t^{2})=ty,$ $\displaystyle t\mathbbm{1}=-ix+y.$ Thus we have relations: $\displaystyle p_{1}$ $\displaystyle=x^{2}-tix=0,$ $\displaystyle p_{2}$ $\displaystyle=xy=0,$ $\displaystyle p_{3}$ $\displaystyle=yx=0,$ $\displaystyle p_{4}$ $\displaystyle=y^{2}-ty=0,$ $\displaystyle p_{5}$ $\displaystyle=t\mathbbm{1}-(-ix+y).$ Since $xy=yx=0$ and squares of $x,y$ can be reduced, we have enough relations to decompose any large product into only multiples of $x,y$ and powers of $t$. Thus we present: $N:={\mathbb{C}[x,y][t]}/{\left<p_{1},p_{2},p_{3},p_{4},p_{5}\right>}.$ We observe: $N_{0}={\mathbb{C}[x,y]}/{\left<x^{2},xy,yx,y^{2},-ix+y\right>},$ and $N_{1}={\mathbb{C}[x,y]}/{\left<x^{2}-ix,xy,yx,y^{2}-y,1+ix-y\right>.}$ We see that $x,y$ behave exactly as $a,b\in A$ and by Proposition 3.9. It is easily checked that the isomorphism is given by: $\displaystyle\phi:N_{1}$ $\displaystyle\to A$ $\displaystyle 1$ $\displaystyle\mapsto\mathbbm{1}$ $\displaystyle x$ $\displaystyle\mapsto(i,0)$ $\displaystyle y$ $\displaystyle\mapsto(0,1).$ Thus we have produced an algebra $N_{0}$ that has $A=\mathbb{C}\oplus\mathbb{C}$ as a deformation. ###### Example 2.2. We consider $A=M_{2}(\mathbb{C})$ and fix: $a=\begin{pmatrix}1&0\\\ 0&-1\end{pmatrix},\quad b=\begin{pmatrix}0&1\\\ 1&0\end{pmatrix}.$ We denote by $\mathbbm{1}$ the multiplicative unit in $A$. Next we compute: $\displaystyle\begin{pmatrix}1&0\\\ 0&0\end{pmatrix}=\frac{1}{2}(a+b^{2})=e_{1},$ $\displaystyle\begin{pmatrix}0&1\\\ 0&0\end{pmatrix}=\frac{1}{2}(ab+b)=e_{2},$ $\displaystyle\begin{pmatrix}0&0\\\ 1&0\end{pmatrix}=\frac{1}{2}(b-ab)=e_{3},$ $\displaystyle\begin{pmatrix}0&0\\\ 0&1\end{pmatrix}=\frac{1}{2}(b^{2}-a)=e_{4}$ and thus $a,b$ generate $A$ as an algebra. Now we consider $A[t]$ and define: $\displaystyle x:=ta,$ $\displaystyle y:=tb.$ We compute: $\displaystyle x^{2}=t^{2}\begin{pmatrix}1&0\\\ 0&1\end{pmatrix}=t^{2}\mathbbm{1},$ $\displaystyle xy=t^{2}\begin{pmatrix}0&1\\\ -1&0\end{pmatrix}=t^{2}(e_{2}-e_{3}),$ $\displaystyle yx=t^{2}\begin{pmatrix}0&-1\\\ 1&0\end{pmatrix}=t^{2}(e_{3}-e_{2}),$ $\displaystyle y^{2}=t^{2}\begin{pmatrix}1&0\\\ 0&1\end{pmatrix}=t^{2}\mathbbm{1}.$ We have obtained the relations: $\displaystyle p_{1}=x^{2}-t^{2}\mathbbm{1},$ $\displaystyle p_{2}=x^{2}-y^{2}=0,$ $\displaystyle p_{3}=xy+yx=0,$ $\displaystyle p_{4}=y^{2}-t^{2}\mathbbm{1}.$ Now we consider larger products to produce more relations: $\displaystyle x^{3}=t^{3}\begin{pmatrix}1&0\\\ 0&-1\end{pmatrix}=t^{2}x,$ $\displaystyle y^{3}=t^{3}\begin{pmatrix}0&1\\\ 1&0\end{pmatrix}=t^{2}y.$ Hence we have also obtained: $\displaystyle p_{5}=x^{3}-t^{2}x=0,$ $\displaystyle p_{6}=y^{3}-t^{2}y=0.$ These allow us to derive the following: $\displaystyle x^{2}y=y^{3}=yx^{2}=y^{3}=t^{2}y,$ $\displaystyle y^{2}x=x^{3}=xy^{2}=x^{3}=t^{2}x,$ which are enough to reduce an arbitrary product of $x,y$. Thus we present: $N:={\mathbb{C}[x,y][t]}/{\left<p_{1},p_{2},p_{3},p_{4},p_{5},p_{6}\right>}.$ We observe: $N_{0}={\mathbb{C}[x,y]}/{\left<x^{2},y^{2},xy+yx\right>},$ and $N_{1}={\mathbb{C}[x,y]}/{\left<x^{2}-1,y^{2}-1,xy+yx,x^{3}-x,y^{3}-y\right>}.$ By Proposition 3.9 we have $N_{1}\cong A$. It is easily checked that the isomorphism is given by: $\displaystyle\phi:N_{1}$ $\displaystyle\to A$ $\displaystyle 1$ $\displaystyle\mapsto\mathbbm{1}$ $\displaystyle x$ $\displaystyle\mapsto\begin{pmatrix}1&0\\\ 0&-1\end{pmatrix}$ $\displaystyle y$ $\displaystyle\mapsto\begin{pmatrix}0&1\\\ 1&0\end{pmatrix}.$ Thus we have produced an algebra $N_{0}$ that has $A=M_{2}(\mathbb{C})$ as a deformation. ## 3 Deformations to $A$ We now consider the general case in which $A$ is an associative finite dimensional $\mathbb{C}$-algebra generated by some elements $\sum_{i=1}^{n}a_{i},\sum_{i=1}^{n}b_{i}$ for some $a_{i},b_{i}\in A$. ### 3.1 Specification at $t=0$ In this section we define a multiplication on $\mathbb{C}[x,y][t]$ which gives the algebra $N$ at specification $t=0$. ###### Proposition 3.1. Let $\mathbb{C}[t]$ denote the polynomial ring in variable $t$. Further, assume that for every $r\in A$ there exists $i=i(r)$ such that $rt^{i}\in\text{im}f$. There exist elements $q_{1},q_{2},\dots,q_{n}\in{\mathbb{C}}[x,y][t]$ such that for every $k$ we have $f(p_{k})\in\sum_{i=1}^{n}\mathbb{C}[t]f(q_{i}).$ (1) Moreover $n=\dim A.$ ###### Proof. We let $\\{e_{1},\dots,e_{n}\\}$ be a basis of $A$. Let $k_{1}$ be the smallest possible integer such that for some integer $j_{1}$ we have $e_{j_{1}}t^{k_{1}}+\sum_{\begin{subarray}{c}i=1\\\ i\neq j_{1}\end{subarray}}^{n}\alpha_{i}(t)e_{i}\in\text{im}f$ for some $\alpha_{i}(t)\in\mathbb{C}[t]$. Let $k_{2}$ be the smallest possible integer such that for some integer $j_{2}\neq j_{1}$ we have $e_{j_{2}}t^{k_{2}}+\sum_{\begin{subarray}{c}i=1\\\ i\neq j_{1},j_{2}\end{subarray}}^{n}\alpha^{\prime}_{i}(t)e_{i}\in\text{im}f$ for some $\alpha^{\prime}_{i}(t)\in\mathbb{C}[t]$. Continuing in this way, we define a basis $\\{q_{1},\dots,q_{n}\\}$ such that $f(q_{i})=e_{j_{m}}t^{k_{m}}+\sum_{\begin{subarray}{c}i=1\\\ i\neq j_{1},\dots,j_{m}\end{subarray}}^{n}\alpha_{i}(t)e_{i}.$ Now notice that $f(q_{1}),f(q_{2}),\dots,f(q_{n})$ are linearly independent over $\mathbb{C}$ as $e_{1},e_{2},\dots,e_{n}$ are linearly independent over $\mathbb{C}$. We now prove that $f(q_{1})+t\text{im}f,\dots,f(q_{n})+t\text{im}f$ span the $\mathbb{C}$-algebra $\frac{\text{im}f}{t\text{im}f}$ over $\mathbb{C}$, and hence they are a basis for $\frac{\text{im}f}{t\text{im}f}$. Let $z_{1}\in\text{im}f$ and write $z_{1}=\sum_{i=1}^{n}r_{i}(t)e_{j_{i}}t^{s_{i}}$ for some $r_{i}\in\mathbb{C}[t]-t\mathbb{C}[t]$ and some $j_{i},s_{i}\in\mathbb{N}$. Notice that for the smallest $i$ such that $r_{i}(t)\neq 0$, $s_{i}\geq k_{i}$ for $k_{i}$ as above. Let $z_{2}=z_{1}-f(q_{1})t^{s_{1}-k_{1}}r_{1}(t).$ Note that we have $z_{2}\in\sum_{i=2}^{n}\mathbb{C}[t]e_{j_{i}}$ so we can write $z_{2}=\sum_{i=2}^{n}r^{2}_{i}(t)e_{j_{i}}t^{s^{2}_{i}}$ for some $r^{2}_{i}(t)\in\mathbb{C}[t]-t\mathbb{C}[t]$ and $s^{2}_{i}\in\mathbb{N}$. Notice that $q_{2}\in\text{im}f$. Now let $z_{3}=z_{2}-f(q_{2})t^{s^{2}_{2}-k_{2}}r^{2}_{2}(t).$ Continuing in this way we eventually arrive at $q_{n+1}=0.$ Hence by summing the following equations $\displaystyle 0$ $\displaystyle=z_{n}-f(q_{n})t^{s^{n}_{n}-k_{n}}r^{n}_{n}(t),$ $\displaystyle\vdots$ $\displaystyle z_{3}$ $\displaystyle=z_{2}-f(q_{2})t^{s^{2}_{2}-k_{2}}r^{2}_{2}(t)$ and $z_{2}=z_{1}-f(q_{1})t^{s_{1}-k_{1}}r_{1}(t)$ we arrive at $z_{1}=\sum_{l=1}^{n}f(q_{l})t^{s^{\prime}_{l}-k_{l}}r^{l}_{l}(t)\in\sum_{l=1}^{n}\mathbb{C}[t]f(q_{l})$ for some $s^{\prime}_{l}\in\mathbb{N}$ as required. ∎ Remark. Notice that for $k\in\mathbb{N}$ and $\alpha_{j}\in\mathbb{C}[t]$ $f(p_{k})-\sum_{j=1}^{n}\alpha_{j}f(q{{}_{j}})=0$ if and only if $p_{k}-\sum_{j=1}^{n}\alpha_{j}q_{j}\in I.$ ###### Proposition 3.2. For any $k,m\in\\{1,2,\dots,n\\}$ there exist $\zeta_{i,k}\in\mathbb{C}$, $\xi_{i,k}(t)\in{\mathbb{C}}[t]$ such that $p_{k}-\sum_{i=1}^{n}(\zeta_{i,k}q_{i}+t\xi_{i,k}(t)q_{i})\in I.$ In particular, there exist $\zeta_{i,k,m}\in\mathbb{C}$, $\xi_{i,k,m}(t)\in{\mathbb{C}}[t]$ such that $q_{k}q_{m}-\sum_{i=1}^{n}(\zeta_{i,k,m}q_{i}+t\xi_{i,k,m}(t)q_{i})\in I.$ ###### Proof. For each element $p_{k}\in{\mathbb{C}}[x,y][t]$ we have $p_{k}\in\sum_{i=1}^{n}{\mathbb{C}}[t]q_{i}+I.$ This follows by taking $f^{-1}$ of equation (1). Hence, for each $k\in\mathbb{N}$ there exist $\sigma_{i,k}\in\mathbb{C}[t]$ such that $p_{k}-\sum_{i=1}^{n}\sigma_{i,k}q_{i}\in I,$ where $I=\ker f$. Notice that we can write $\sigma_{i,k}(t)=\zeta_{i,k}+t\xi_{i,k}(t)$ for some $\zeta_{i,k}\in\mathbb{C}$, $\xi_{i,k}(t)\in{\mathbb{C}}[t]$, as to separate the terms dependent on $t$. The above then becomes $p_{k}-\sum_{i=1}^{n}(\zeta_{i,k}q_{i}+t\xi_{i,k}(t)q_{i})\in I.$ ∎ Reasoning similarly as in the proof of Proposition 3.1 we get the following corollary. ###### Corollary 3.3. Suppose $\sum_{i=1}^{n}\alpha_{i}q_{i}\in I$ for some $\alpha_{i}\in{\mathbb{C}}[t]$. Then $f(\sum_{i}^{n}\alpha_{i}q_{i})=0$ and consequently $\alpha_{i}=0$ for every $i\in\\{1,2,\dots,n\\}$. Moreover, all elements of $I$ are $\mathbb{C}[t]$-linear combinations of elements $p_{k}-\sum_{i=1}^{n}(\zeta_{i,k}q_{i}+t\xi_{i,k}(t)q_{i})$ for some $k\in\mathbb{N},\zeta_{i,k}\in\mathbb{C}$, $\xi_{i,k}(t)\in{\mathbb{C}}[t]$. ###### Notation 3.4. We denote by $J$ be the set consisting of $\mathbb{C}$-linear combinations of elements $p_{k}-\sum_{i=1}^{n}\zeta_{i,k}q_{i}\in{\mathbb{C}}[x,y],$ (2) where $\zeta_{i,k}$ are as in Proposition 3.2. Let $\langle J\rangle$ be the ideal of ${\mathbb{C}}[x,y]$ generated by elements from $J$. Further, let $N$ be the quotient algebra $\mathbb{C}[x,y]/\langle J\rangle$. ###### Corollary 3.5. We have $e\in J$ if and only if $e+e^{\prime}\in I$ for some $e^{\prime}\in t{\mathbb{C}}[x,y][t]$. In particular $J\subseteq I+t{\mathbb{C}}[x,y][t]$, and hence $\langle J\rangle\subseteq I+t{\mathbb{C}}[x,y][t]$. ###### Proposition 3.6. The dimension of $N$ equals $n$. Moreover, elements $q_{k}+J\in N$ are a basis of $N$ as a $\mathbb{C}$-vector space. ###### Proof. Notice that expression (2) implies that ${\mathbb{C}}[x,y]\subseteq\sum_{k=1}^{n}\mathbb{C}q_{k}+\langle J\rangle,$ and hence $\\{q_{k}+\langle J\rangle\mid k\in\\{1,2,\ldots,n\\}\\}$ span $N$ as a $\mathbb{C}$-vector space. Therefore, the dimension of $N$ does not exceed $n$. We now show that the elements $q_{k}+\langle J\rangle\in N$ for $k\in\\{1,2,\ldots,n\\}$ are linearly independent over $\mathbb{C}$. Suppose on the contrary that we have $\sum_{i=1}^{n}\xi_{i}q_{i}\in\langle J\rangle$ for some $\xi_{i}\in\mathbb{C}$. By Corollary 3.5 and Proposition 3.1 we have $\langle J\rangle\subseteq I+t{\mathbb{C}}[x,y][t]\subseteq I+t\sum_{i=1}^{n}{\mathbb{C}}[t]q_{i}.$ It follows that there is $e^{\prime}\in t\sum_{i=1}^{n}{\mathbb{C}}[t]q_{i}$ such that $\sum_{i=1}^{n}\xi_{i}q_{i}-e^{\prime}\in I$, contradicting Corollary 3.3. ∎ Observe now that the following holds: ###### Proposition 3.7. We can present $N$ as an $\mathbb{C}$-algebra with generators $d_{1},\dots,d_{n}$, which span $N$ as a $\mathbb{C}$-vector space, subject to relations $d_{k}d_{m}=\sum_{i=1}^{n}\zeta_{i,k,m}d_{i}$ where $\zeta_{i,k,m}$ are as in Proposition 3.2. ###### Proof. We let ${\mathbb{C}}[q_{1},\ldots,q_{n}]$ be the subalgebra of ${\mathbb{C}}[x,y]$ generated by elements $q_{1},\ldots,q_{n}\in{\mathbb{C}}[x,y]$. Similarly, let ${\mathbb{C}}[d_{1},\ldots,d_{n}]$ be the free $\mathbb{C}$-algebra generated by elements $d_{1},\ldots,d_{n}$. Let $\zeta:{\mathbb{C}}[d_{1},\ldots,d_{n}]\rightarrow{\mathbb{C}}[x,y]/\langle J\rangle$ be defined as $\zeta(d_{i})=q_{i}+\langle J\rangle$. We now show that $x+\langle J\rangle,y+\langle J\rangle\in\text{im}\zeta$. Notice that since $f(x)\in\text{im}f$ we have $f(x)\in\sum_{i=0}^{\infty}{\mathbb{C}}[t]f(q_{i}).$ Hence we have $f(x)=f(x^{\prime})$ for some $x^{\prime}\in\sum_{i=0}^{\infty}{\mathbb{C}}[t]f(q_{i}).$ Then $x-x^{\prime}\in I$, and so $x-x^{\prime\prime}\in J$ for some $x^{\prime\prime}\in\sum_{i=0}^{\infty}{\mathbb{C}}q_{i}.$ Then $x^{\prime\prime}+\langle J\rangle\in\text{im}\zeta$, and so $x+\langle J\rangle\in\text{im}\zeta$. An analogous argument shows $y+\langle J\rangle\in\text{im}\zeta$. Let $J^{\prime}=\ker(h)\subseteq{\mathbb{C}}[d_{1},\ldots,d_{n}].$ Then by the First Isomorphism Theorem for rings ${\mathbb{C}}[d_{1},\ldots,d_{n}]/J^{\prime}$ is isomorphic to ${\mathbb{C}}[x,y]/\langle J\rangle=N$. Let $J^{\prime\prime}$ be the ideal of ${\mathbb{C}}[d_{1},\ldots,d_{n}]$ generated by elements $d_{k}d_{m}-\sum_{i=1}^{n}\zeta_{i,k,m}d_{i}$ where $\zeta_{i,k,m}$ are as in Proposition 3.2. Observe that $\zeta\left(d_{k}d_{m}-\sum_{i=1}^{n}\zeta_{i,k,m}d_{i}\right)=q_{k}q_{m}-\sum_{i=1}^{n}\zeta_{i,k,m}q_{i}+\langle J\rangle=0+\langle J\rangle.$ Therefore $J^{\prime\prime}\subseteq J^{\prime}$. It follows that the dimension of ${\mathbb{C}}[d_{1},\dots,d_{n}]/J^{\prime}$ does not exceed the dimension of ${\mathbb{C}}[d_{1},\dots,d_{n}]/J^{\prime\prime}$. We now show that $J^{\prime}=J^{\prime\prime}$. Notice that ${\mathbb{C}}[d_{1},\ldots,d_{n}]/J^{\prime\prime}$ is at most $n$-dimensional, since every element can be presented as a linear combination of elements $d_{i}+J^{\prime\prime}$ for $i=1,2,\ldots,n$. On the other hand ${\mathbb{C}}[d_{1},\ldots,d_{n}]/J^{\prime}$ is isomorphic to ${\mathbb{C}}[x,y]/\langle J\rangle$, and hence is $n$-dimensional (by Proposition 3.6). Hence by comparing dimensions we get that the dimension of ${\mathbb{C}}[d_{1},\dots,d_{n}]/J^{\prime\prime}$ does not exceed the dimension of ${\mathbb{C}}[d_{1},\dots,d_{n}]/J^{\prime}$. Previously we showed that ${\mathbb{C}}[d_{1},\dots,d_{n}]/J^{\prime}$ does not exceed the dimension of ${\mathbb{C}}[d_{1},\dots,d_{n}]/J^{\prime\prime}$. It follows that $J^{\prime}=J^{\prime\prime}$, as required. ∎ We now define a multiplication on $N$ which gives a formal deformation to the algebra $A$. ###### Theorem 3.8. Let $d_{1},\dots,d_{n}$ be free generators of the $\mathbb{C}$-algebra $\mathbb{C}[d_{1},\dots,d_{n}]$ and suppose $\zeta_{i,k,m}\in\mathbb{C}$, $\xi_{i,k,m}(t)\in{\mathbb{C}}[t]$ are as in Proposition 3.2. Then the multiplication rule $d_{k}\ast_{t}d_{m}=\sum_{i=1}^{n}(\zeta_{i,k,m}d_{i}+t\xi_{i,k,m}(t)d_{i}).$ gives a formal deformation such that $\ast_{0}$ gives the multiplication on an algebra isomorphic to $N$. ###### Proof. Recall relations from Proposition 3.2 that for any $k,m\in\\{1,2,\dots,n\\}$ there exist $\zeta_{i,k,m}\in\mathbb{C}$, $\xi_{i,k,m}(t)\in{\mathbb{C}}[t]$ such that $q_{k}q_{m}-\sum_{i=1}^{n}(\zeta_{i,k,m}q_{i}+t\xi_{i,k,m}(t)q_{i})\in I.$ (3) We introduce notation $[q_{i}]:=q_{i}+I$ for elements of ${\mathbb{C}}[x,y]/I$. Hence we obtain the following relations corresponding to (3) which give the multiplicative table on ${\mathbb{C}}[x,y]/I$. We have $[q_{k}][q_{m}]-\sum_{i=1}^{n}(\zeta_{i,k,m}[q_{i}]+t\xi_{i,k,m}(t)[q_{i}]).$ Notice that these relations give a multiplication $d_{k}\ast_{t}d_{m}=\sum_{i=1}^{n}(\zeta_{i,k,m}d_{i}+t\xi_{i,k,m}(t)d_{i})$ (4) for $d_{i}\in\mathbb{C}[x,y].$ Now recall that $N$ is isomorphic to ${\mathbb{C}}[d_{1},\ldots,d_{n}]/J^{\prime\prime}$ where $J^{\prime\prime}$ is the ideal of ${\mathbb{C}}[d_{1},\ldots,d_{n}]$ generated by relations $d_{k}d_{m}-\sum_{i=1}^{n}\zeta_{i,k,m}d_{i}.$ Therefore by setting $t=0$ in (4) we obtain the multiplication rule for ${\mathbb{C}}[d_{1},\ldots,d_{n}]/J^{\prime\prime}$. ∎ ### 3.2 Specification at $t=1$ In this section we define a multiplication on $\mathbb{C}[x,y][t]$ which gives a formal deformation of $N$, such that at specification $t=1$ we get an algebra isomorphic to $A$. ###### Proposition 3.9. Let $\xi:\mathbb{C}[x,y][t]\rightarrow A$ be a homomorphism of $\mathbb{C}$-algebras such that $\xi(x)=\sum_{i=0}^{n}a_{i},\quad\xi(y)=\sum_{i=0}^{n}b_{i},\quad\xi(t)=1.$ Further, assume that for every $r\in A$ there exists $i=i(r)$ such that $rt^{i}\in\text{im}f$. Then $\ker\xi=\langle I,t-1\rangle.$ ###### Proof. We let $e=\sum_{i}\alpha_{i}p_{i}t^{\beta{i}}\in\ker\xi$ where $p_{i}$ are monomials in $\mathbb{C}[x,y]$, $\alpha_{i}\in\mathbb{C}$ and $\beta_{i}\in\mathbb{N}$. We will show there exists $\gamma_{i}\in\mathbb{N}$ such that $\hat{e}:=\sum_{i}\alpha_{i}p_{i}t^{\gamma_{i}}\in I$, so that $e\in I+\langle t-1\rangle$ as $e-\hat{e}\in\langle t-1\rangle.$ This follows since $e-\hat{e}=\sum_{i}\alpha_{i}p_{i}t^{\beta_{i}}-\sum_{i}\alpha_{i}p_{i}t^{\gamma_{i}}=\sum_{i}\alpha_{i}p_{i}(t^{\beta_{i}}-t^{\gamma_{i}})=\sum_{i}\alpha_{i}p_{i}t^{m}(t^{n}-1)\in\langle t-1\rangle$ for some $m,n\in\mathbb{N}$. Recall Notation 1.8 and note that $\xi(e)=\sum_{i}\alpha_{i}\overline{\overline{p_{i}}}=0$. Notice that by assumption there exist $j_{i}\in\mathbb{N}$ such that $\overline{\overline{p_{i}}}t^{j_{i}}\in\text{im}f$. We let $f(c_{i})=\overline{\overline{p_{i}}}t^{j_{i}}$ for some $c_{i}\in\mathbb{C}[x,y][t]$. Hence for large enough $k\in\mathbb{N}$ we have $f\left(\sum_{i}\alpha_{i}c_{i}t^{k-j_{i}}\right)=\sum_{i}\alpha_{i}\overline{\overline{p_{i}}}t^{k}=0.$ ∎ ###### Proposition 3.10. Let notation be as in Proposition 3.9, and suppose that $\sum_{i=0}^{n}a_{i}$ and $\sum_{i=0}^{n}b_{i}$ generate $A$ as a $\mathbb{C}$-algebra. We have $\frac{{\mathbb{C}}[x,y][t]/I}{\langle t-1+I\rangle}\cong A.$ ###### Proof. Note that the ideal $\langle t-1+I\rangle$ of ${\mathbb{C}}[x,y][t]/I$ equals the set $\frac{{\mathbb{C}}[x,y][t](t-1)+I}{I}=\frac{\langle I,t-1\rangle}{I}=\\{g(t-1)+I\mid g\in{\mathbb{C}}[x,y][t]\\}.$ By the Third Isomorphism Theorem we have $\frac{{\mathbb{C}}[x,y][t]/I}{\langle I,t-1\rangle/I}\cong\frac{{\mathbb{C}}[x,y][t]}{\langle I,t-1\rangle}.$ By Proposition 3.9 we have $\langle I,t-1\rangle=\ker\xi$ and as $\xi$ is onto, $\frac{{\mathbb{C}}[x,y][t]}{\ker\xi}\cong A$ by the First Isomorphism Theorem. ∎ We now define a multiplication on ${\mathbb{C}}[x,y][t]/\langle t-1,I\rangle$ which gives a deformation to $A$ at $t=1$. ###### Theorem 3.11. Let notation be as in Proposition 3.9, and suppose that $\sum_{i=0}^{n}a_{i}$ and $\sum_{i=0}^{n}b_{i}$ generate $A$ as a $\mathbb{C}$-algebra. Let $d_{1},\dots,d_{n}$ be free generators of the $\mathbb{C}$-algebra $\mathbb{C}[d_{1},\dots,d_{n}]$ and suppose $\zeta_{i,k,m}\in\mathbb{C}$, $\xi_{i,k,m}(t)\in{\mathbb{C}}[t]$ are as in Proposition 3.2. Then the multiplication rule $d_{k}\ast_{t}d_{m}=\sum_{i=1}^{n}(\zeta_{i,k,m}d_{i}+t\xi_{i,k,m}(t)d_{i}).$ gives a formal deformation such that $\ast_{1}$ gives the multiplication on an algebra generated by $d_{1},\dots,d_{n}$ isomorphic to ${\mathbb{C}}[x,y][t]/\langle I,t-1\rangle$. ###### Proof. We show that the algebra ${\mathbb{C}}[x,y][t]/\langle I,t-1\rangle$ is isomorphic to the algebra ${\mathbb{C}}[d_{1},d_{2},\ldots d_{n}]/I^{\prime}$ where $I^{\prime}$ is the ideal of ${\mathbb{C}}[d_{1},d_{2},\ldots d_{n}]$ generated by elements $d_{k}\ast_{1}d_{m}-\sum_{i=1}^{n}(\zeta_{i,k,m}d_{i}+\xi_{i,k,m}(1)d_{i})$ for $\zeta_{i,k,m}\in\mathbb{C}$, $\xi_{i,k,m}(t)\in{\mathbb{C}}[t]$ as in Proposition 3.2. We let $\delta:{\mathbb{C}}[d_{1},\ldots,d_{n}]\rightarrow{\mathbb{C}}[x,y][t]/\langle I,t-1\rangle$ be such that $\delta(d_{i})=q_{i}+\langle I,t-1\rangle,$ for $i\in\\{1,2,\ldots,n\\}$. Observe that $I^{\prime}\subseteq\ker(\delta)$ since $\displaystyle\delta\left(d_{k}*d_{m}-\sum_{i=1}^{n}(\zeta_{i,k,m}d_{i}+\xi_{i,k,m}(1)d_{i})\right)$ $\displaystyle=q_{k}*q_{m}-\sum_{i=1}^{n}(\zeta_{i,k,m}q_{i}+\xi_{i,k,m}(1)q_{i})$ $\displaystyle\quad+\langle I,t-1\rangle$ $\displaystyle=\langle I,t-1\rangle$ since $q_{k}\ast q_{m}-\sum_{i=1}^{n}(\zeta_{i,k,m}q_{i}+\xi_{i,k,m}(1)q_{i})\in I+\langle t-1\rangle$. Therefore the dimension of ${\mathbb{C}}[d_{1},\ldots,d_{n}]/\ker(\delta)$ does not exceed the dimension of ${\mathbb{C}}[d_{1},\ldots,d_{n}]/I^{\prime}$. Notice that ${\mathbb{C}}[d_{1},\ldots,d_{n}]/I^{\prime}$ is spanned by elements $d_{i}+I$ as a vector space, and hence has dimension at most $n$. On the other hand, by the First Isomorphism Theorem for rings: ${\mathbb{C}}[d_{1},\ldots,d_{n}]/ker(\delta)\cong\text{im}(\delta)={\mathbb{C}}[x,y][t]/\langle I,t-1\rangle,$ (5) which in turn is isomorphic to $A$ by Proposition 3.9, and hence has dimension $n$. It follows that $I^{\prime}=\ker(\delta)$ and hence ${\mathbb{C}}[d_{1},\ldots,d_{n}]/ker(\delta)$ is isomorphic to ${\mathbb{C}}[d_{1},\ldots,d_{n}]/I^{\prime}$. Where the equality in Equation 5 is true by the proof of Proposition 3.7. ∎ ###### Remark 3.12. Deformations at $t\neq 0$ The results from this section, hold with analogous proofs for a specialisation at other $t\neq 0$. Let $z\in\mathbb{C}$ and $z\neq 0$, the the deformation at $t=z$ is also isomorphic to $A$, provided that $\sum_{i=0}^{n}a_{i}z^{i}$ and $\sum_{i=0}^{n}b_{i}z^{i}$ generate $A$ as a $\mathbb{C}$-algebra. ###### Remark 3.13. Since the dimension of all the algebras in $\\{N_{t}\\}_{t\in[0,1]}$ remains $n=\dim(A)$ the deformation arising from $*$ is flat. ## 4 Application In this section we prove a conjecture of M. Wemyss using the method in section 3. The statement is that of the following theorem. ###### Theorem 4.1. Let $N=\frac{\mathbb{C}[x,y]}{\langle xy+yx,x^{3}+y^{2},y^{3}\rangle}.$ Then $N$ deforms into $A=M_{2}(\mathbb{C})\oplus\mathbb{C}^{\oplus 5}$. We prove this theorem through a series of lemmas. Firstly, we fix notation in addition to the notation of Theorem 4.1. ###### Notation 4.2. Let $i_{1},i_{2}$ be roots of the polynomial $x^{2}-x+1$. ###### Notation 4.3. Let $e\in\mathbb{C}$ be such that for $c=\frac{i_{1}-i_{2}}{\sqrt{1+e^{2}}}$ the polynomial $g(z)=(2z-1)^{2}z^{3}+c^{2}$ has $5$ distinct roots $\alpha_{1},\dots,\alpha_{5}\in\mathbb{C}$. Further, we assume that $e\neq 0$, $c^{2}\neq 3$ and $\alpha_{i}\neq\frac{1}{2}.$ Notice that $g(z)$ and $z^{3}+1$ have no common roots, so $\alpha_{i}\neq-1$ and $\alpha_{j}\notin\\{i_{1},i_{2}\\}$. ###### Notation 4.4. Let $\alpha=(\alpha_{1},\alpha_{2},\alpha_{3},\alpha_{4},\alpha_{5})$ and $\beta=(\beta_{1},\beta_{2},\beta_{3},\beta_{4},\beta_{5})$ for $\beta_{i}\in\mathbb{C}$ such that $(\beta_{1},\beta_{2},\beta_{3},\beta_{4},\beta_{5})(2\alpha_{1}-1,2\alpha_{2}-1,2\alpha_{3}-1,2\alpha_{4}-1,2\alpha_{5}-1)=c(1,1,1,1,1).$ Notice that such $\beta_{i}$ exist since $\alpha_{i}\neq\frac{1}{2}$. ###### Lemma 4.5. Let $\xi:\mathbb{C}[x,y][t]\rightarrow A$ be a homomorphism of $\mathbb{C}$-algebras such that $\displaystyle\xi(x)$ $\displaystyle=\begin{pmatrix}i_{1}&0\\\ 0&i_{2}\end{pmatrix}+\alpha,\quad\xi(y)=\begin{pmatrix}1&e\\\ e&-1\end{pmatrix}+\beta,\quad\xi(t)=1.$ Then $\xi(x),\xi(y)$ and $1$ generate $A$ as an algebra. ###### Proof. Observe that $\xi(x^{j}(x^{2}-x+1))=\alpha^{j}(\alpha^{2}-\alpha+1)$ (where $\alpha^{j}(\alpha^{2}-\alpha+1)=(\alpha_{1}^{j}(\alpha_{1}^{2}-\alpha_{1}+1),\dots,\alpha_{5}^{j}(\alpha_{5}^{2}-\alpha_{5}+1))$). Now let $W$ be the matrix with entries $W_{ij}=\alpha_{i}^{j}(\alpha_{i}^{2}-\alpha_{i}+1)$. Let $V$ be the Vandermonde matrix with entries $V_{ij}=\alpha_{i}^{j}$. Let $D$ be a diagonal matrix with entries $D_{ij}=\delta_{ij}(\alpha_{i}^{2}-\alpha_{i}+1)$. It follows that $W=DV.$ Now since $V$ is nonsingular, $W$ is nonsingular. Now this implies that any $5$-dimensional vector in $A$ is in $\text{im}\xi$. Then $\xi(x)-\alpha\in\text{im}\xi$ and hence $\begin{pmatrix}i_{1}&0\\\ 0&i_{2}\end{pmatrix}\in\text{im}\xi.$ Similarly, $\xi(y)-\beta\in\text{im}\xi$ and hence $\begin{pmatrix}1&e\\\ e&-1\end{pmatrix}\in\text{im}\xi.$ Now recall that $i_{1},i_{2}$ are roots of the polynomial $z^{3}+1$, so we have $\begin{pmatrix}i_{1}&0\\\ 0&i_{2}\end{pmatrix}^{3}=\begin{pmatrix}-1&0\\\ 0&-1\end{pmatrix}.$ Notice that every diagonal matrix is a linear combination of $\begin{pmatrix}i_{1}&0\\\ 0&i_{2}\end{pmatrix}$ and $\begin{pmatrix}-1&0\\\ 0&-1\end{pmatrix}$. Now, it can be checked that these matrices are sufficient to generate all matrices in $M_{2}(\mathbb{C})$. ∎ ###### Notation 4.6. Let $f:\mathbb{C}[x,y][t]\rightarrow A[t]$ be a homomorphism of $\mathbb{C}$-algebras such that $\displaystyle f(x)$ $\displaystyle=t^{2}\begin{pmatrix}i_{1}&0\\\ 0&i_{2}\end{pmatrix}+t^{2}\alpha,\quad f(y)=\frac{1}{\sqrt{1+e^{2}}}t^{3}\begin{pmatrix}1&e\\\ e&-1\end{pmatrix}+t^{3}\beta,\quad f(t)=t\mathbbm{1},$ where $\mathbbm{1}$ is the identity element in $A$. We write $\tilde{1}:=(1,1,1,1,1)$. As usual we denote by $1$ the identity element in $\mathbb{C}$ and thus in $\mathbb{C}[x,y][t]$. We now prove Theorem 4.1. ###### Proof. Referring to Notation 4.4 and 4.6 we calculate: $\displaystyle f(x^{2})$ $\displaystyle=t^{4}\Big{(}\begin{pmatrix}i_{1}^{2}&0\\\ 0&i_{2}^{2}\end{pmatrix}+\alpha^{2}\Big{)},$ $\displaystyle f(y^{2})$ $\displaystyle=t^{6}\Big{(}\begin{pmatrix}1&0\\\ 0&1\end{pmatrix}+\beta^{2}\Big{)},$ $\displaystyle f(xy)$ $\displaystyle=\frac{t^{5}}{\sqrt{1+e^{2}}}\Big{(}\begin{pmatrix}i_{1}&i_{1}e\\\ i_{2}e&-i_{2}\end{pmatrix}+\alpha\beta\Big{)},$ $\displaystyle f(yx)$ $\displaystyle=\frac{t^{5}}{\sqrt{1+e^{2}}}\Big{(}\begin{pmatrix}i_{1}&i_{2}e\\\ i_{1}e&-i_{2}\end{pmatrix}+\alpha\beta\Big{)}.$ Note that $f(x^{2}-t^{2}x+t^{4})=t^{4}(\alpha^{2}-\alpha+\tilde{1})$ and hence $\displaystyle f(x(x^{2}-t^{2}x+t^{4}))$ $\displaystyle=t^{6}\alpha(\alpha^{2}-\alpha+\tilde{1}),$ $\displaystyle f(x^{2}(x^{2}-t^{2}x+t^{4}))$ $\displaystyle=t^{8}\alpha^{2}(\alpha^{2}-\alpha+\tilde{1}),$ $\displaystyle f(y(x^{2}-t^{2}x+t^{4}))$ $\displaystyle=t^{7}\beta(\alpha^{2}-\alpha+\tilde{1}).$ Observe that $\displaystyle f(xy+yx)$ $\displaystyle=\frac{t^{5}}{\sqrt{1+e^{2}}}\begin{pmatrix}2i_{1}&e\\\ e&-2i_{2}\end{pmatrix}+t^{5}2\alpha\beta,$ $\displaystyle f(t^{2}y)$ $\displaystyle=\frac{t^{5}}{\sqrt{1+e^{2}}}\begin{pmatrix}1&e\\\ e&-1\end{pmatrix}+t^{5}\beta,$ $\displaystyle f(t^{5}\frac{i_{2}-i_{1}}{\sqrt{1+e^{2}}})$ $\displaystyle=\frac{t^{5}}{\sqrt{1+e^{2}}}\begin{pmatrix}i_{2}-i_{1}&0\\\ 0&i_{2}-i_{1}\end{pmatrix}+\frac{t^{5}(i_{2}-i_{1})}{\sqrt{1+e^{2}}}\tilde{1}.$ Note that $2i_{1}-1+(i_{2}-i_{1})=0$ and $-2i_{2}+1+(i_{2}-i_{1})=0$, so $f(xy+yx-t^{2}y+\frac{t^{5}(i_{2}-i_{1})}{\sqrt{1+e^{2}}})=2\alpha\beta-\beta+\frac{i_{2}-i_{1}}{\sqrt{1+e^{2}}}\tilde{1}.$ Now notice that $2\alpha\beta-\beta+\frac{i_{2}-i_{1}}{\sqrt{1+e^{2}}}\tilde{1}=0$ by assumptions of Notation 4.4. It follows that $xy+yx-t^{2}y+\frac{t^{5}(i_{2}-i_{1})}{\sqrt{1+e^{2}}}\in I$ and so $xy+yx\in\langle J\rangle$ is a relation in $N$ as defined in Notation 3.4. Notice that by Notation 4.4 we have $\beta^{2}(2\alpha-1)^{2}=c^{2}\tilde{1}$. Now it follows by Notation 4.3 that $\alpha^{3}+\beta^{2}=0,$ (6) and thus $x^{3}+y^{2}\in\langle J\rangle$ is a relation in $N$. We now proceed to show that $y^{3}\in\langle J\rangle$. Notice that: $f(y^{3})=\frac{t^{9}}{\sqrt{1+e^{2}}}\begin{pmatrix}1&e\\\ e&-1\end{pmatrix}+t^{9}\beta^{3}.$ Hence $y^{3}-t^{6}y=t^{9}(\beta^{3}-\beta).$ Therefore it is sufficient to show that $y^{3}-t^{6}y+q\in I$ for some $q\in t\mathbb{C}[x,y][t]$. We denote: $\displaystyle r_{1}$ $\displaystyle:=t^{5}(x^{2}-t^{2}x+t^{4})$ $\displaystyle r_{2}$ $\displaystyle:=t^{3}x(x^{2}-t^{2}x+t^{4})$ $\displaystyle r_{3}$ $\displaystyle:=tx^{2}(x^{2}-t^{2}x+t^{4})$ $\displaystyle r_{4}$ $\displaystyle:=t^{2}y(x^{2}-t^{2}x+t^{4})$ Notice that the elements $\displaystyle f(r_{1})$ $\displaystyle=t^{9}(\alpha^{2}-\alpha+\tilde{1})=:t^{9}e_{1},$ $\displaystyle f(r_{2})$ $\displaystyle=t^{9}\alpha(\alpha^{2}-\alpha+\tilde{1})=:t^{9}e_{2},$ $\displaystyle f(r_{3})$ $\displaystyle=t^{9}\alpha^{2}(\alpha^{2}-\alpha+\tilde{1})=:t^{9}e_{3},$ $\displaystyle f(r_{4})$ $\displaystyle=t^{9}\beta(\alpha^{2}-\alpha+\tilde{1})=:t^{9}e_{4}$ are in $t\mathbb{C}[x,y][t]$. Note that it suffices to show that $y^{3}-t^{6}y+q\in I$ for some $q\in\mathbb{C}r_{1}+\mathbb{C}r_{2}+\mathbb{C}r_{3}+\mathbb{C}r_{4}$. Therefore it suffices to show that $\beta^{3}-\beta\in\mathbb{C}e_{1}+\mathbb{C}e_{2}+\mathbb{C}e_{3}+\mathbb{C}e_{4}.$ (7) Now relation (7) is equivalent to $(2\alpha-1)\beta(\beta^{2}-1)\in\mathbb{C}(2\alpha-1)e_{1}+\mathbb{C}(2\alpha-1)e_{2}+\mathbb{C}(2\alpha-1)e_{3}+\mathbb{C}(2\alpha-1)e_{4},$ which in turn is equivalent to $-c(\alpha^{3}+1)\in\mathbb{C}(2\alpha-1)e_{1}+\mathbb{C}(2\alpha-1)e_{2}+\mathbb{C}(2\alpha-1)e_{3}+\mathbb{C}(2\alpha-1)e_{4}.$ (8) Now, notice that $\alpha^{3}+1=(\alpha+1)(\alpha^{2}+\alpha+1)$ and that $\alpha+1$ has non-zero entries, so multiplying relation (8) by $\alpha+1$ we get an equivalent relation $\displaystyle(\alpha^{3}+1)(\alpha+1)\in\mathbb{C}$ $\displaystyle(2\alpha-1)(\alpha^{3}+1)+\mathbb{C}(2\alpha-1)(\alpha^{3}+1)\alpha$ $\displaystyle+\mathbb{C}(2\alpha-1)(\alpha^{3}+1)\alpha^{2}+\mathbb{C}(\alpha^{3}+1).$ Notice that the right hand side can be rewritten as $\mathbb{C}(\alpha^{3}+1)+\mathbb{C}(\alpha^{3}+1)\alpha+\mathbb{C}(\alpha^{3}+1)\alpha^{2}+\mathbb{C}(\alpha^{3}+1)\alpha^{3}$. ∎ ## 5 A more general method In this section we consider a more general method that results in algebras $N$ that are not necessarily local and have more generators. Furthermore, we consider $A$ to have finitely many generators. Let $A$ be a finitely generated $\mathbb{C}$-algebra. 1. 1. We consider a free $\mathbb{C}$-algebra $F$ with $j$ generators $x_{1},\dots,x_{j}$ for some $j\in\mathbb{N}$. We define: $\displaystyle f:F[t]$ $\displaystyle\to A[t]$ $\displaystyle x_{i}$ $\displaystyle\mapsto\sum\limits_{l=0}^{m}a_{i,l}t^{l}$ for some $m\in\mathbb{N}$ and some $a_{i,l}\in A$. We also assume that for each $z\in\mathbb{C}$, $z\neq 0$ elements $\sum\limits_{l=0}^{m}a_{i,l}z^{l}$ for $i=1,2,\ldots j$ generate $A$. In other words, images of the generators generate A when $t$ is set to non-zero numbers. 2. 2. We define $\mathcal{N}:=\frac{\mathbb{C}[x_{1},\dots,x_{j}][t]}{I},$ where $I:=\ker(f)$. 3. 3. We assume that for every $r\in A$ there exists $i=i(r)$ such that $rt^{i}\in\text{im}f$. 4. 4. Our algebra $N=N_{0}$ is isomorphic to $\frac{\mathcal{N}}{\langle t+I\rangle}$ and deforms flatly and associatively to $A$. Moreover $N$ is isomorphic to $\frac{\mathbb{C}[x_{1},\dots,x_{j}][t]}{\langle I,t\rangle},$ where $\langle I,t\rangle$ is the ideal of $\mathbb{C}[x_{1},\dots,x_{j}][t]$ generated by elements of $I$ and by $t$. If $j=2$ then this method corresponds to the method described in section 3 provided that $a_{i,0}=0$ for all $i$. Proof. When reasoning as in the previous chapters we will use at all places $\mathbb{C}[x_{1},\ldots,x_{j}]$ instead of $\mathbb{C}[x,y]$, and we will use $\mathbb{C}[x_{1},\ldots,x_{j}][t]$ instead of $\mathbb{C}[x,y][t]$. Let $A$ have dimension $n$ over $\mathbb{C}$. Recall that we $\mathcal{N}:=\frac{\mathbb{C}[x_{1},\dots,x_{j}][t]}{I},$ where $I:=\ker(f)$. * • Reasoning analogously as in Proposition 3.1 we obtain that there are $q_{1},\ldots,q_{n}\in F$ such that $f(F[t])\subseteq\sum_{i=1}^{n}{\mathbb{C}}f(q_{i}),$ and hence, since $I=\ker f$, we get $F(t)\subseteq\sum_{i=1}^{n}{\mathbb{C}}q_{i}+I.$ In particular, there are $\zeta_{i,k,m}\in\mathbb{C}$, $\xi_{i,k,m}(t)\in{\mathbb{C}}[t]$ such that $q_{k}\cdot q_{m}-(\sum_{i=1}^{n}\zeta_{i,k,m}q_{i}+t\xi_{i,k,m}q_{i})\in I.$ * • Reasoning similarly as in Theorem 3.8 we obtain that if $d_{1},\ldots,d_{n}$ are free generators then $d_{k}*d_{m}-(\sum_{i=1}^{n}\zeta_{i,k,m}d_{i}+t\xi_{i,k,m}d_{i}),$ gives a formal deformation such that $*_{0}$ gives a multiplication on $N$ where $N$ is the $\mathbb{C}$-algebra with generators $d_{1},\ldots,d_{n}$ subject to relations $d_{k}*_{0}d_{m}=\sum_{i=1}^{n}\zeta_{i,k,m}d_{i}.$ Observe that $*$ defines a multiplication on the algebra $\mathbb{C}[d_{1},\ldots,d_{n}][t]/I^{\prime}$, where $I^{\prime}$ is the ideal of $\mathbb{C}[d_{1},\ldots,d_{n}][t]$ generated by elements $d_{k}*d_{m}-(\sum_{i=1}^{n}\zeta_{i,k,m}d_{i}+t\xi_{i,k,m}d_{i}).$ Notice that $d_{1},\ldots,d_{n}$ are free generators of the free algebra $\mathbb{C}[d_{1},\ldots,d_{n}]$. Notice that the algebra $\mathbb{C}[d_{1},\ldots,d_{n}][t]/I^{\prime}$ is isomorphic to algebra $\mathcal{N}$. To see this consider the map $\sigma\mathbb{C}[d_{1},\ldots,d_{n}][t]\rightarrow{\mathcal{N}}$ given by $d_{i}\rightarrow q_{i}+I.$ Notice that $I^{\prime}\subseteq\ker\sigma$ since $(q_{k}+I)*(q_{m}+I)=\sum_{i=1}^{n}\zeta_{i,k,m}(q_{i}+I)+t\xi_{i,k,m}(q_{i}+I).$ Recall that $\ker\sigma=I^{\prime}$ because if $a\in ker\sigma$ then by using relations from $I^{\prime}$ we can present $a$ as $a^{\prime}+i$ where $i^{\prime}\in I$ and where $a^{\prime}=\sum_{i=1}^{n}\alpha_{i}(t)d_{i}$ for some $\alpha_{i}(t)\in{\mathbb{C}}[t]$. Notice that then $\sigma(i^{\prime})=0$ and hence $\sigma(a)=\sigma(a^{\prime})=\sigma(\sum_{i=1}^{n}\alpha_{i}(t)d_{i})=\sum_{i=1}^{n}\alpha_{i}(t)q_{i}.$ Reasoning as in Corollary $3.3$ we get that this implies that all $\alpha_{i}=0$, hence $a^{\prime}=0$, hence $a=i^{\prime}$. Therefore, $\ker\sigma\subseteq I$ and consequently $\ker\sigma=I^{\prime}$. Therefore, by the First Isomorphism Theorem for algebras ${\mathbb{C}}[d_{1},\ldots,d_{n}][t]/I^{\prime}$ is isomorphic to $Im(\sigma)$ which is equal to ${\mathcal{N}}$. To see that $\sigma$ is surjective, notice that $q_{1}+I,\ldots,q_{n}+I$ span $\mathcal{N}$ as a linear space (as mentioned before this can be proved similarly as Proposition 3.1). * • Reasoning as in Proposition 3.6 we obtain that $N$ is $n$ dimensional $\mathbb{C}$-algebra. * • Reasoning similarly as in Proposition 3.7 we obtain that the algebra $N$ is isomorphic to the algebra $\frac{\mathbb{C}[x_{1},\dots,x_{j}][t]}{\langle I,t\rangle}$ where ${\langle I,t\rangle}$ is the ideal of $\mathbb{C}[x_{1},\dots,x_{j}][t]$ generated by elements from $I$ and by $t$. By the Third Isomorphism Theorem $N$ is isomorphic to $\frac{\mathcal{N}}{\langle t+I\rangle}.$ This can be seen by reasoning similarly as in Proposition 3.10 but taking $t$ instead of $t-1$ in the proof. * • We will now consider deformations $*_{1}$ at $t=1$. Consider first the homomorphism of $\mathbb{C}$-algebras $\xi:F[t]\rightarrow A$ given by $\xi(x_{i})=\sum_{l=0}^{m}a_{i,l}$ for $i=1,2,\ldots,j$. Reasoning similarly as in Proposition 3.9 we obtain that $\ker\xi=\langle I,t-1\rangle$. Reasoning analogously as in the last two lines of Proposition 3.10 we obtain that $F[t]/\ker\xi$ is isomorphic to $A$. * • Reasoning as in Theorem 3.11 we see that the algebra ${\mathbb{C}}[d_{1},\ldots,d_{n}][t]/I$ deforms at $t=1$ to the algebra $F[t]/{\langle I,t-1\rangle}=F[t]/{\ker\xi}$ which is isomorphic to $A$. * • We will now consider deformations $*_{z}$ at $t=z$ where $z\in\mathbb{C},z\neq 0,$ $z\neq 1$. Consider first the homomorphism of $\mathbb{C}$-algebras $\xi:F[t]\rightarrow A$ given by $\xi(x_{i})=\sum_{l=0}^{m}a_{i,l}z^{l}$ for $i=1,2,\ldots,j$. Reasoning similarly as in Proposition 3.9 we obtain that $\ker\xi=\langle I,t-z\rangle$ (this can be done by using $t^{\prime}=t/z$ at the place of $t$ in the proof of Proposition 3.9, and in the first line declaring that $p_{i}\in F$ instead of $p_{i}\in{\mathbb{C}}[x,y]$). Next, reasoning analogously as in the last two lines of Proposition 3.10 we obtain that $F[t]/\ker\xi$ is isomorphic to $A$ (this follows since elements $\xi(x_{i})$ generate $A$). Next, reasoning as in Theorem 3.11 we see that the algebra ${\mathbb{C}}[d_{1},\ldots,d_{n}][t]/I$ deforms at $t=z$ to the algebra $F[t]/{\langle I,t-z\rangle}=F[t]/{\ker\xi}$ which is is isomorphic to $A$. ∎ ## 6 Future Work In this section we pose some questions about Method 1 and suggest some generalisations. ###### Question 6.1. Let $A$ be a $\mathbb{C}$-algebra and suppose $G_{1}=\\{a,b\\}$ and $G_{2}=\\{a^{\prime},b^{\prime}\\}$ are two distinct generating sets of $A$. Under what conditions does Method 1 using $A,G_{1}$ and $A,G_{2}$ result in the same algebra $N_{0}$ that has $A$ as a deformation? ###### Question 6.2. Let $A$ be a $\mathbb{C}$-algebra. Under what conditions does Method 1 only produce one algebra $N$ that has $A$ as a deformation. ###### Question 6.3. Let $A$ be a $\mathbb{C}$-algebra. Does their exist a finitely terminating algorithm that will produce all the algebras $N$ arising from Method 1 that have $A$ as a deformation? ###### Question 6.4. Let $A$ be a $\mathbb{C}$-algebra. Do ring theoretic properties of the generators $a,b$ (for example being idempotent, being irreducible idempotent, prime) determine any properties of the algebra $N$ arising from Method 1 using $A,\\{a,b\\}$? ###### Question 6.5. Let $A$ be a $\mathbb{C}$-algebra. Do ring theoretic properties of $A$ (for example being semi-simple, local) determine any properties of the algebra $N$ arising from Method 1 using $A$? ###### Question 6.6. Can Method 1 be generalised to be used for algebras over a general commutative ring? #### Acknowledgements We are grateful to Michael Wemyss for suggesting his interesting questions which inspired this paper, and for useful comments about contraction algebras and their applications in geometry. The third author acknowledges support from the EPSRC programme grant EP/R034826/1 and from the EPSRC research grant EP/V008129/1. ## References * [Ber78] George M Bergman. The diamond lemma for ring theory. Advances in Mathematics, 29(2):178–218, February 1978. * [BW22] Gavin Brown and Michael Wemyss. Local normal forms of noncommutative functions, 2022. * [CGG89] Vincent Coll, Murray Gerstenhaber, and Anthony Giaquinto. An explicit deformation formula with noncommuting derivations. Israel Math. Conf. Proc., 1:396–403, 01 1989. * [DW16] Will Donovan and Michael Wemyss. Noncommutative deformations and flops. Duke Mathematical Journal, 165(8), June 2016. arXiv:1309.0698 [math]. * [FO19] Boris Feigin and Alexandre Odesski. Flat deformations of algebras and functional equations. Journal of Combinatorial Algebra, 3(3):215–236, September 2019\. * [HT18] Zheng Hua and Yukinobu Toda. Contraction algebra and invariants of singularities. Int. Math. Res. Not. IMRN, (10):3173–3198, 2018. * [SW13] Anne V. Shepler and Sarah Witherspoon. PBW deformations of skew group algebras in positive characteristic, December 2013. arXiv:1312.3616 [math]. * [Tod15] Yukinobu Toda. Non-commutative width and Gopakumar-Vafa invariants. Manuscripta Math., 148(3-4):521–533, 2015. * [Wem23] Michael Wemyss. A lockdown survey on cdv singularities, 2023. * [Wit19] Sarah J. Witherspoon. Hochschild cohomology for algebras, volume 204 of Graduate Studies in Mathematics. American Mathematical Society, Providence, RI, [2019] ©2019.
# Extracting non-local inter-polaron interactions from collisional dynamics Junichi Takahashi Department of Electronic and Physical Systems, Waseda University, Tokyo 169-8555, Japan Hiroyuki Tajima Department of Mathematics and Physics, Kochi University, Kochi 780-8520, Japan Quantum Hadron Physics Laboratory, RIKEN Nishina Center, Wako, Saitama 351-0198, Japan Eiji Nakano Department of Mathematics and Physics, Kochi University, Kochi 780-8520, Japan Kei Iida Department of Mathematics and Physics, Kochi University, Kochi 780-8520, Japan ###### Abstract This study develops a novel experimental method of deducing the profile of interaction induced between impurities in a trapped gas of ultracold Fermi/Bose atoms, which are often referred to as Fermi/Bose polarons. In this method, we consider a two-body Fermi/Bose polaron collision experiment in which impurities and atoms interact only weakly. Numerical simulations of the quantum dynamics reveal the possibility to obtain information regarding the non-local induced interaction between two polarons from a measured profile of the polaron wave packet at several snapshots. This is because the potential of the induced interaction is well balanced by the quantum potential whenever the WKB approximation for the relevant Schrödinger equation is applicable. Introduction— To describe interparticle interactions is indispensable in various fields of physics. At extremely high energy, fundamental interactions are mediated by gauge bosons Book_Weinberg1995 , while in atomic nuclei, the nuclear force at a distance occurs through exchange of mesons Taketani1951 . In conventional superconductors, attractive electron–electron interactions are induced by phonons Bardeen1957 . In high-$T_{\rm c}$ superconductors, furthermore, background spin and/or charge fluctuations are expected to play a role Scalapino1995 . Ultracold atomic systems can be used as a platform for studying medium-induced interactions. Particularly, such systems can exhibit high tunability and pureness, thereby serving as excellent quantum simulators for studying quantum many-body theories Bloch2008 ; Giorgini2008 ; Dalfovo1999 . For example, cold atomic Fermi systems with a long scattering length close to the unitary limit are good simulators with regard to low density neutron matter Horikoshi2019 . They also possess the unique feature of having the real-time dynamics and momentum distribution measured Pethick-Smith . Moreover, atomic Fermi/Bose polarons, which are quasiparticles that consist of impurities (minority atoms) immersed in a degenerate Fermi gas Schirotzek2009 ; Frohlich2011 ; Kohstall2012 or a Bose-Einstein condensate Jorgensen2016 ; Hu2016 ; Rentrop2016 , have been realized. Since the impurities have the mass and the interaction with each other modified from those in vacuum by the impurity- medium interaction, a great deal of research has been conducted both from theoretical and experimental sides. Theoretically, the Ruderman-Kittel-Kasuya- Yoshida (RKKY) type Ruderman1 ; Kasuya1 ; Yoshida1 and Yukawa/Efimov type interactions have been proposed as typical examples of non-local induced interactions between heavy Fermi and Bose polarons Nishida2009 ; Suchet2017 ; Heiselberg2000 ; Yu2012 ; Enss2020 ; Naidon2018 , respectively. In connection with such induced interactions, moreover, bi-polaron formation for Bose polarons Naidon2018 ; Camacho-Guardian2018 ; Dehkharghani2018 ; Volosniev2015 ; Mistakidis2019 and the energy shift and broadening of Fermi polarons Giraud2012 ; Mistakidis2019c ; Tajima2018 have been actively discussed. Recently, the presence of non-local interactions between Fermi polarons has been experimentally confirmed DeSalvo2019 ; Edri2020 . Here, the formation of solitons through the corruption of a Bose gas was found to be different from the behavior predicted when only local interactions occur in a boson-fermion mixture, which is an evidence for the existence of a non-local interaction mediated by fermions. The non-local nature of such interactions, however, remains to be investigated experimentally. Given that similar induced interactions occur in various fields of physics, therefore, real cold atom experiments in this direction would serve as a cornerstone. We now address the question “What is enough information to find induced interactions between particles in a medium?” For concreteness, we consider the case in which two polarons are approximately described by a two-body time- dependent Schrödinger equation (TDSE) with a mediated interaction instead of dealing with the medium explicitly. In this case, the simplest answer is a two-body polaron wavefunction. That is, the interaction can be obtained from the wavefunction by transforming the TDSE as $U(r)=\Psi^{-1}\left[i\hbar\frac{\partial}{\partial t}-h_{1}({\bf x}_{1})-h_{2}({\bf x}_{2})\right]\Psi$, where $\Psi$ is the wavefunction, $h_{i}$ denotes the one-particle Hamiltonian, $r=|{\bf x}_{1}-{\bf x}_{2}|$, and $U$ denotes the interaction between two polarons. This means that one can in principle constitute the inter-polaron interaction $U$ as function of the relative coordinate if the $\Psi$ keeps nonvanishing in the range of $U(r)\not=0$ throughout such experiments as the collisional ones Sommer2011 ; Joseph2011 ; Valtolina2017 ; Reynold2020 . In reality, however, it is difficult to simultaneously obtain the phase and amplitude of the wavefunction by experiments. In fact, at a given snapshot, only the amplitude of the wavefunction can be measured via the square root of the probability density of the wavefunction. For trapped cold atomic systems, one can use, e.g., a wave packet of a non-interacting Bose gas as an impurity and then identify the number density of the Bose gas with the probability density of the impurity. In this study, we propose a method of deducing the non-local interaction between two such impurities in a medium from the density of the Bose gas at several snapshots. It is noteworthy that the idea of deriving bulk thermodynamic quantities from the density of a single-species cold atomic gas has been already proposed Ho2010 . Despite similarity in strategy, there is a crucial difference from our proposal. Authors in Ref. Ho2010 use the density distribution to obtain the thermodynamic properties of the corresponding homogeneous system within local density approximation (LDA), while information about the non-local interaction between atoms in a trap, which is of interest here, is out of the reach of LDA. Formulation— For future possible experimental realizations, we focus on a two- polaron collision in a harmonic trap filled with a Fermi/Bose atomic gas. Each impurity, which is a Bose condensate wave packet, is initially localized by a separate confining harmonic potential, released at $t=0$, and allowed to collide with another impurity. Note that experimental setups that involve a scenario highly similar to that considered in this study have been proposed Mistakidis2019b ; Magierski2019 ; Kwasniok2020 ; Tajima2020 , while a part of them have been realized in experiment by using optical tweezers Reynold2020 . We assume that at $t=0$, each impurity has already been immersed in the majority gas long enough to form a polaron and that at $t\geq 0$ a medium- induced interaction occurs between the polarons. Furthermore, we consider the polarons to be distinguishable particles that do not interact directly with each other and to remain robust during the time evolution owing to coupling between the impurity and the majority gas that is too weak for the entire system to be relaxed. In a regime of strong coupling, on the other hand, a hydrodynamic description looks more relevant. In fact, a hydrodynamic analysis of collisions of two polaronic clouds in a similar setup has been performed Tajima2020 ; in this case, induced inter-polaron interactions have been incorporated even inside a cloud but only within the LDA. In the weak coupling regime of interest here, the dynamics of the two polarons essentially obeys the TDSE: $i\hbar\partial_{t}\Psi=[h({\bf x}_{1})+h({\bf x}_{2})+U_{\rm med}(r)]\Psi$, where $h({\bf x}_{i})=-\hbar^{2}\nabla^{2}_{{\bf x}_{i}}/2m+V_{\rm HO}({\bf x}_{i})$ with the trap potential $V_{\rm HO}({\bf x}_{i})=m\omega^{2}{\bf x}_{i}^{2}/2$ and $U_{\rm med}(r)$ denotes an interaction mediated by the majority gas. Here, for simplicity, we assume that the trap frequency $\omega$ and the bare mass $m$ of an impurity atom are the same between the polarons and that the dynamics is 1D. We next rewrite the TDSE in the center-of-mass frame by using the relative coordinate $x=x_{1}-x_{2}$ as $i\hbar\frac{\partial\varphi}{\partial t}=\left[-\frac{\hbar^{2}}{2m_{\rm r}}\frac{\partial^{2}}{\partial x^{2}}+V_{\rm r,HO}(x)+U_{\rm med}(|x|)\right]\varphi$ where $m_{\rm r}=m/2$ is the reduced mass of the impurities, $V_{\rm r,HO}(x)=\frac{1}{2}m_{\rm r}\omega^{2}x^{2}$, and $\varphi(x,t)$ is the relative wavefunction of the polarons. We then substitute the wavefunction $\psi=\sqrt{\rho(x,t)}\exp[iS(x,t)/\hbar]$ into the TDSE, thereby obtaining the coupled equations that are the equation of continuity and the quantum Hamilton–Jacobi equation Book_Tannor2007 $\displaystyle C(S;x,t)+Q(\rho;x,t)+U_{\rm med}(|x|)=0,$ (1) with $C(S;x,t)=\frac{\partial S}{\partial t}+\frac{1}{2m_{\rm r}}\left(\frac{\partial S}{\partial x}\right)^{2}+V_{\rm r,HO}$ and $Q(\rho:x,t)=-\frac{\hbar^{2}}{2m_{\rm r}\sqrt{\rho}}\frac{\partial^{2}\sqrt{\rho}}{\partial x^{2}}$. Here, $C$ is the potential given by the phase fluctuations, while $Q$ is the potential given by the density fluctuaions, which is often referred to as the quantum potential. Equation (1) shows that the interaction $U_{\rm med}$ is determined by the balance of those potentials. Analysis and Results — We proceed to solve the TDSE for a particular interaction and obtain density profiles of a polaron to calculate the quantum potential. Since each polaron is initially trapped in the respective confining harmonic potential, we set the initial two-polaron wavefunction as a pair of the Gaussian wave packets via $\Psi(x_{1},x_{2},t=0)=\phi_{+}(x_{1})\phi_{-}(x_{2})$, with $\phi_{\pm}(x)=\frac{1}{(2\pi\eta^{2})^{\frac{1}{4}}}\exp\left[-\frac{(x\pm x_{0})^{2}}{4\eta^{2}}\right],$ where the two parameters $x_{0}$ and $\eta$ controls the initial position and the width of the respective wave packet. For such nodeless wave packets, the WKB approximation is expected to be good. We solve the TDSE under the corresponding initial condition $\varphi(x,t=0)=\frac{1}{(4\pi\eta^{2})^{\frac{1}{4}}}\exp\left[-\frac{(x-2x_{0})^{2}}{8\eta^{2}}\right],$ by using the second-order split-step Fourier method. We adopt (i) RKKY, (ii) Yukawa, and (iii) Efimov type interactions as $U_{\rm med}$. We note that the interactions (i)–(iii), which are derived for a 3D homogeneous medium, are used for simplicity and that the concrete form of the interactions is not important for the present purpose of finding a way of capturing the mediated interactions through experiments unless the range of the interactions is significantly long. (i) The RKKY type interaction is as follows: $\displaystyle U_{\rm med}(|x|)=\frac{\hbar^{2}Ma^{2}}{2\pi M_{\rm r}^{2}}\frac{2k_{{\rm F}}|x|\cos(2k_{\rm F}|x|)-\sin(2k_{\rm F}|x|)}{|x|^{4}},$ (2) where $a$ is the $s$-wave scattering length between a majority atom of mass $M$ and an impurity, taken to be independent of which impurity the majority atom interacts with, $k_{\rm F}$ is the Fermi wavenumber of the majority gas, and $M_{\rm r}=mM/(m+M)$. Figure 1 presents the probability densities for the relative two-polaron wave packet before and after the collision of a pair of the polaron wave packets that evolve from various initial conditions under the influence of $U_{\rm med}$. We set the number of atoms in the majority gas as a typical one, $N=10^{5}$, while assuming $(k_{\rm F}a)^{2}=2$ and $M=m$. It is interesting to note that before the collision, the width of the two-polaron relative wave packet increases (decreases) with $t$ in the case of Fig. 1(d) (Fig. 1(a)), while remaining almost unchanged with $t$ in the cases of Figs. 1(b) and 1(c). This behavior stems from the fact that the Gaussian wave packet for a free particle with mass $m_{\rm r}$ has the width increased with time like $\sqrt{2}\eta\sqrt{1+(\hbar t/4m_{\rm r}\eta^{2})^{2}}$, while, as two polarons approach each other, the confining potential acts to reduce the width of the wave packet like $\hbar/\sqrt{2m_{\rm r}(E-V_{\rm r,HO})}$ with the relative energy $E$. Incidentally, the collision time is of order $t_{c}$ (quarter of the dipole oscillation period) in the cases of Figs. 1(a)–(c), while being well before $t_{c}$ in the case of Fig. 1(d). Figure 1: Probability density of the two-polaron relative wave packet (time vs. relative position) for (a) $x_{0}/R_{\rm F}=0.5,\,\eta/R_{\rm F}=0.1$, (b) $x_{0}/R_{\rm F}=0.5,\,\eta/R_{\rm F}=0.05$, (c) $x_{0}/R_{\rm F}=0.25,\,\eta/R_{\rm F}=0.05$, and (d) $x_{0}/R_{\rm F}=0.25,\,\eta/R_{\rm F}=0.01$. Here, $R_{\rm F}$ and $E_{\rm F}$ denote the Thomas–Fermi radius and Fermi energy, respectively. The Fermi energy is given by $E_{\rm F}=\frac{1}{2}m\omega^{2}R_{\rm F}^{2}=(6N)^{\frac{1}{3}}\hbar\omega$. The time $t_{c}$ at which the center of the relative wave packet reaches $x=0$ can be estimated from $t_{c}\omega_{\rm F}\simeq\frac{1}{4}\frac{2\pi}{\omega}\omega_{\rm F}\simeq 125,$ where $\omega_{\rm F}=E_{\rm F}/\hbar$. Figure 2: Relative density profile $\rho(x,t)$ at three values of $t\omega_{\rm F}$ of order or greater than the collision time. The distinction between (a)–(d) corresponds to the difference in initial parameters as in Fig. 1. Figure 3: Same as Fig. 2 for minus the quantum potential. For comparison, the RKKY interaction employed to solve the TDSE is also plotted in dotted line. For detailed movies, see Animation . Figure 2 depicts the relative density $\rho(x,t)$ at three different times, most of which are close to the collision time extracted from Fig. 1. Numerical data in the regions of $x<0$ and $x>0$ represent the behavior of the transmitted component and of the incident and reflected components of the relative wave packet, respectively. Once this kind of profiles are obtaind by experiments, one can use them as input data in deriving the quantum potential. We proceed to exhibit, in Fig. 3, minus the quantum potential $-Q(\rho;x,t)$ that can be derived from pseudo data for $\rho(x,t)$ shown in Fig. 2. One can observe from Fig. 3 that the quantum potential in the transmitted wave regime ($x<0$) well reproduces the oscillating pattern of the RKKY interaction for any initial condition, while the reproducibility of the RKKY interaction itself depends on the initial condition. Comparing Figs. 3 (a)–(c), in which cases the width of the transmitted relative wave packet just after the collision is different, we find it advantageous for the transmitted wave packet to be sufficiently spread. Remarkably, Fig. 3 (d) shows an even better reproducibility of the interaction for the initial condition that allows the two respective wave packets to spread and merge well before the centers of these packets come together at the origin at a timescale $t_{c}$ of the dipole oscillation. This supports the tendency that the wider the transmitted wave packet, the higher the reproducibility of the interaction from the quantum potential. This tendency in turn ensures $C(S;x,t)\simeq 0$ and $U_{\rm med}\simeq-Q(\rho;x,t)$ in the region of $x<0$ where the WKB approximation is a good approximation because of the sufficiently spread wave packet and gentle spatial dependence of the confining potential. In the region of $x>0$ in each panel of Figs. 1 and 2, on the other hand, a density oscillation pattern emerges after the collision. This pattern stems naturally from interference between the incident and reflected components of the relative wave packet, while the presence of the reflected component arises mainly from a repulsive part of $U_{\rm med}$. As can be seen from Fig. 3, however, the periodicity of this density oscillation is totally different from that of $U_{\rm med}$. In other words, such quantum interference leads inevitably to $C\not\simeq 0$. Then, the WKB approximation is no longer valid, and hence the mediated interaction cannot be reproduced by $-Q$. Let us turn to the (ii) Yukawa and (iii) Efimov type interactions: $\displaystyle U_{\rm med}(|x|)=\begin{cases}U_{0}\frac{\exp(-\kappa_{0}|x|)}{\kappa_{0}|x|}&({\rm Yukawa}\,{\rm type}),\\\ U_{1}\frac{\kappa_{1}^{-2}}{|x|^{2}}&({\rm Efimov}\,{\rm type}).\end{cases}$ (3) We again confine ourselves to sufficiently spread relative wave packets for the WKB approximation to hold at least in the region of $x<0$. Figure 4 shows minus the relative quantum potentials obtained by numerically simulating collision dynamics of two Bose polarons that interact with (ii) or (iii). We can observe that the quantum potential tends to well reproduce the respective mediated interaction in the regime ($x<0$), just like the RKKY case shown in Fig. 3. Comparing Figs. 4 (a) and (b), in which cases the initial position of the relative wave packet is different under the same Yukawa interaction, one can see that $U_{\rm med}$ is more reproducible in the former than in the latter. On the other hand, comparing Figs. 4 (c) and (d), which are the Efimov counterparts to Figs. 4 (a) and (b), one can realize that $U_{\rm med}$ is more reproducible in the latter than in the former. These results imply that tuning of the initial energy would be desirable for better reproduction of $U_{\rm med}$ in a manner that depends on the range of the mediated interaction. We remark in passing that in contrast to the RKKY case, almost no density oscillation emerges in the region of $x>0$ for the Yukawa and Efimov type interactions, whose purely attractive nature produces almost no reflected wave packet. Then, the WKB approximation is expected to be valid even for $x>0$. A significant deviation of $-Q$ from $U_{\rm med}$ that can be observed in this region for the Efimov type interaction, therefore, suggests that the long-range nature distorts the incident component of the relative wave packet even at a semiclassical level. Figure 4: Minus the quantum potential at three different times as plotted for the Yukawa (upper panels) and Efimov (lower panels) type interactions with $U_{0}/\hbar\omega=U_{1}/\hbar\omega=-0.1$ and $\kappa_{0}R_{\rm HO}=\kappa_{1}R_{\rm HO}=20$, where $R_{\rm HO}=\sqrt{\hbar/m_{r}\omega}$. In panels (a) and (c) [(b) and (d)], the initial parameters are set to $x_{0}/R_{\rm HO}=0.25$ and $\eta/R_{\rm HO}=0.05$ ($x_{0}/R_{\rm HO}=1$ and $\eta/R_{\rm HO}=0.05$). The dotted line in each panel denotes the mediated interaction involved. For detailed movies, see Animation . Summary— This study has developed a novel practical method of deducing the induced interactions between two impurities from a measured profile of the polaron wave packet at several snapshots of a two-polaron collisional dynamics in the case in which impurities and medium atoms interact only weakly. The key to success in this method is the validity of the WKB approximation, which is satisfied for sufficiently spread polaron wave packets. We have successfully demonstrated by solving the TDSE the possibility of reproducing such interactions as the RKKY, Yukawa, and Efimov type from the measured profile of the wave packets via the quantum potential. In real possible experiments, one would be prepared to confine a degenerate Fermi gas or Bose condensate with a sufficiently large trap and treat two non- interacting Bose gases with different internal states as impurities by putting each of them in a sufficiently narrow trap that is located at a symmetrical position with respect to the large trap of a majority gas and then by releasing the two Bose gases from the trap at the same time. We expect that observation of transmitted wave packets after the collision would be possible owing to the different internal states, which would help to distinguish between two impurities. We emphasize that our strategy to deduce the mediated interaction from the probability density of the two-polaron wavefunction is different from the method of estimating the potential from a phase shift as used in an inverse scattering method InverseScattering . In general, the full profile of the medium-induced non-local interaction that occurs between particles is theoretically unknown. In this regard, we believe that the method proposed here could help to easily and quantitatively obtain information regarding mediated interactions from experiments. It is nevertheless foreseeable that in actual experiments the interaction cannot be described by the type of mediated interactions used in this study due to various effects such as the finite volume one. If the experiment is carried out under appropriate initial conditions, however, the resultant deviation between the empirically deduced interactions and the full interaction could be duly reduced. Also, our proposed method might open an opportunity to study nuclear interactions that are in principle microscopically known from QCD Ishii2007 ; Ishii2012 ; Iritani2019 , from a different viewpoint that utilizes a cold atomic system as a quantum simulator. As a next step, we will attempt to test our method for 3D collision systems and investigate the applicability to general interactions. Acknowledgments— We would like to thank K. Nishimura, T. Hata, K. Ochi, and Y. Yamanaka for useful discussions. This work was supported in part by Grants-in- Aid for Scientific Research from JSPS (Nos. 17K05445, 18K03501, 18H05406, 18H01211, and 19K14619). ## References * (1) S. Weinberg, The Quantum Theory of Fields, Vol. 1 (Cambridge University Press, New York, 1995). * (2) M. Taketani, S. Nakamura, and M. Sasaki, Prog. Theor. Phys. 6, 581 (1951). * (3) J. Bardeen, L. N. Cooper, and J. R. Schrifeer, Phys. Rev. 108, 1175 (1957). * (4) D. Scalapino, Physics Reports 250, 329 (1995). * (5) I. Bloch, J. Dalibard, and W. Zwerger, Rev. Mod. Phys. 80, 885 (2008). * (6) S. Giorgini, L. P. Pitaevskii, and S. Stringari, Rev. Mod. Phys. 80, 1215 (2008). * (7) F. Dalfovo, S. Giorgini, L. P. Pitaevskii, and S. Stringari, Rev. Mod. Phys. 71, 463 (1999). * (8) M. Horikoshi, M. Kuwata-Gonokami, Int. J. Mod. Phys. E 28, 1930001 (2019). * (9) See, for instance, C. J. Pethick and H. Smith, Bose-Einstein Condensation in Dilute Gases (Cambridge University Press, Cambridge, 2008). * (10) A. Schirotzek, C.-H. Wu, A. Sommer, and M. W. Zwierlein, Phys. Rev. Lett. 102, 230402 (2009). * (11) B. Fröhlich, M. Feld, E. Vogt, M. Koschorreck, W. Zwerger, and M. Köhl, Phys. Rev. Lett. 106, 105301 (2011). * (12) C. Kohstall, M. Zaccanti, M. Jag, A. Trenkwalder, P. Massignan, G. M. Bruun, F. Schreck, and R. Grimm, Nature(London) 485, 615 (2012). * (13) N. B. Jørgensen, L. Wacker, K. T. Skalmstang, M. M. Parish, J. Levinsen, R. S. Christensen, G. M. Bruun, and J. J. Arlt, Phys. Rev. Lett. 117, 055302 (2016). * (14) M.-G. Hu, M. J. Van de Graaff, D. Kedar, J. P. Corson, E. A. Cornell, and D. S. Jin, Phys. Rev. Lett. 117, 055301 (2016). * (15) T. Rentrop, A. Trautmann, F. A. Olivares, F. Jendrzejewski, A. Komnik, and M. K. Oberthaler, Phys. Rev. X 6, 041041 (2016). * (16) M. A. Ruderman and C.¬ Kittel, Phys. Rev. 96, 99 (1954). * (17) T. Kasuya, Prog. Theor. Phys. 16, 45 (1956). * (18) K. Yosida, Phys. Rev. 106, 893 (1957). * (19) Y. Nishida, Phys. Rev. A 79, 013629 (2009). * (20) D. Suchet, Z. Wu, F. Chevy, and G. M. Bruun, Phys. Rev. A 95, 043643 (2017). * (21) H. Heiselberg, C. J. Pethick, H. Smith, L. Viverit, Phys. Rev. Lett. 85, 2418 (2000). * (22) Z. Yu and C. J. Pethick Phys. Rev. A 85, 063616 (2012). * (23) T. Enss, B. Tran, M. Rautenberg, M. Gerken, E. Lippi, M. Drescher, B. Zhu, M. Weidemuller, and M. Salmhofer, arXiv:2009.03029. * (24) P. Naidon, J. Phys. Soc. Jpn. 87, 043002 (2018). * (25) A. Camacho-Guardian, L. A. Peña Ardila, T. Pohl, and G. M. Bruun, Phys. Rev. Lett. 121, 013401 (2018). * (26) A. S. Dehkharghani, A. G. Volosniev, and N. T. Zinner, Phys. Rev. Lett. 121, 080405 (2018). * (27) A. G. Volosniev, H.-W. Hammer, and N. T. Zinner, Phys. Rev. A 92, 023623 (2015). * (28) S. I. Mistakidis, A. G. Volosniev, N. T. Zinner, and P. Schmelcher Phys. Rev. A 100, 013619 (2019). * (29) S. Giraud and R. Combescot, Phys. Rev. A 85, 013605 (2012). * (30) S. I. Mistakidis, G. C. Katsimiga, G. M. Koutentakis, and P. Schmelcher, New J. Phys. 21, 043032 (2019). * (31) H. Tajima and S. Uchino, New J. Phys. 20, 073048 (2018). * (32) B. J. DeSalvo, K. Patel, G. Cai, and C. Chin, Nature 568, 61 (2019). * (33) H. Edri, B. Raz, N. Matzliah, N. Davidson, and R. Ozeri, Phys. Rev. Lett. 124, 163401 (2020). * (34) A. Sommer, M. Ku, G. Roati, M. W. Zwierlein, Nature 472, 201 (2011). * (35) J. A. Joseph, J. E. Thomas, M. Kulkarni, A. G. Abanov, Phys. Rev. Lett. 106, 150401 (2011). * (36) G. Valtolina, F. Scazza, A. Amico, A. Burchianti, A. Recati, T. Enss, M. Inguscio, M. Zaccanti, G. Roati, Nat. Phys. 13, 704 (2017). * (37) L. A. Reynolds, E. Schwartz, U. Ebling, M. Weyland, J. Brand, M. F. Andersen, Phys. Rev. Lett. 124, 073401 (2020). * (38) T.-L. Ho and Q. Zhou, Nat. Phys 6, 131 (2010). * (39) S. I. Mistakidis, G. C. Katsimiga, G. M. Koutentakis, Th. Busch, and P. Schmelcher, Phys. Rev. Lett. 122, 183001 (2019). * (40) P. Magierski, B. Tüzemen, and G. Wlazłowski, Phys. Rev. A 100, 033613 (2019) * (41) J. Kwasniok, S. I. Mistakidis, and P. Schmelcher, Phys. Rev. A 101, 053619 (2020). * (42) H. Tajima, J. Takahashi, E. Nakano, and K. Iida, Phys. Rev. A 102, 051302(R) (2020). * (43) D. J. Tannor, Introduction to Quantum Mechanics: A Time-Dependent Perspective, (University Science Books, 2007). * (44) See Supplemental Material at {URL will be provided by the publisher}. * (45) R. G. Newton, Scattering Theory of Waves and Particles (Springer-Verlag, New York, 1982). * (46) N. Ishii, S. Aoki and T. Hatsuda, Phys. Rev. Lett. 99, 022001 (2007). * (47) N. Ishii, S. Aoki, T. Doi, T. Hatsuda, Y. Ikeda, Y. Inoue, K. Murano, H. Nemura, and K. Sasaki, Phys. Lett. B 712, 437 (2012). * (48) T. Iritani, S. Aoki, T. Doi, S. Gongyo, T. Hatsuda, Y. Ikeda, T. Inoue, N. Ishii, H. Nemura, K. Sasaki, HAL QCD Collaboration, Phys. Rev. D 99, 014514 (2019). Supplemental online material for: “Extracting non-local inter-polaron interactions from collisional dynamics” In the supplimental material, we show the list of movies. List of movies: * 1. File: `RKKY_x0_050_eta_010.mp4` Dynamics of the relative density and minus the quantum potential that corresponds to Figs. 2(a) and 3(a). Youtube: `https://youtu.be/By_ZALuI6A8` * 2. File: `RKKY_x0_050_eta_005.mp4` Dynamics of the relative density and minus the quantum potential that corresponds to Figs. 2(b) and 3(b). Youtube: `https://youtu.be/dTKHjs05SfQ` * 3. File: `RKKY_x0_025_eta_005.mp4` Dynamics of the relative density and minus the quantum potential that corresponds to Figs. 2(c) and 3(c). Youtube: `https://youtu.be/9iXiWWM9H6o` * 4. File: `RKKY_x0_025_eta_001.mp4` Dynamics of the relative density and minus the quantum potential that corresponds to Figs. 2(d) and 3(d). Youtube: `https://youtu.be/Q74kZCrYWrY` * 5. File: `Yukawa_x0_025_eta_005.mp4` Dynamics of the relative density and minus the quantum potential that corresponds to Fig. 4(a). Youtube: `https://youtu.be/nTRTT_8gGME` * 6. File: `Yukawa_x0_100_eta_005.mp4` Dynamics of the relative density and minus the quantum potential that corresponds to Fig. 4(b). Youtube: `https://youtu.be/tJkNVh91SrY` * 7. File: `Efimov_x0_025_eta_005.mp4` Dynamics of the relative density and minus the quantum potential that corresponds to Fig. 4(c). Youtube: `https://youtu.be/fK-G-StErAM` * 8. File: `Efimov_x0_100_eta_005.mp4` Dynamics of the relative density and minus the quantum potential that corresponds to Fig. 4(d). Youtube: `https://youtu.be/2RqUnCu-Shg`
# An Exploration of Mars Colonization with Agent-Based Modeling Edgar Arguello 111Authors are listed in alphabetical order. Sam Carter Cristina Grieg Michael Hammer Chris Prather Clark Petri Anamaria Berea222 Department of Computational Social Sciences, George Mason University ## 1 Introduction Establishing a human settlement on Mars is an incredibly complex engineering problem. The inhospitable nature of the Martian environment requires any habitat to be largely self-sustaining. Beyond mining a few basic minerals and water [4], the colonizers will be dependent on earth resupply and replenishment of necessities via technological means, i.e., splitting martian water into oxygen for breathing and hydrogen for fuel [6]. Our goal is to better understand the interactions of future Martian colonists through an Agent-Based Modeling (ABM) approach. Accounting for engineering and technological limitations, we draw on research regarding high performing teams in isolated and high stress environments (ex: submarines, Arctic exploration, war) to include psychological components within the ABM. Interactions between agents with different psychological profiles are modeled at the individual level while global events such as accidents or delays in earth resupply affect the colony as a whole. ## 2 Space Settlements Literature Review ### 2.1 Use of Agent-Based Models to Simulate Small-Scale Communities Given the many facets to the problem, a complex systems approach is appropriate to extend our insights and avoid uninformed predictions later [10]. ABMs may prove useful to help understand the ecological and demographic conditions needed for sustainable resource-management in a constrained human habitat [10]. It has been demonstrated that integrating Geographic Information System (GIS) overlays and multi-agent simulation provides a means to explore group behaviors which may be affected by an agent’s spatial knowledge of their terrain [16]. ABMs provide useful “laboratories” for competing hypotheses to identify rules of behavior that lead to group dynamics [16], which we would then like to apply to our simulation of a potential Mars settlement. ### 2.2 Team and Interpersonal Psychology The modern era of teams working in extreme conditions commenced with the advent of nuclear submarines in the 1950s, the space race of the 1960s, and expanded Arctic exploration of the 1900s [14]. These pursuits specifically buoyed interest in how teams perform in extreme environments, what makes them successful, and how individuals cope with the challenges presented by such conditions. The technical sophistication necessary for persistent survival in submarine missions or space exploration, coupled with their isolation and inherent risks, position those endeavors as proxies for initializing a martian colony ABM. In a study of Antarctic exploration teams, researchers found that psychological discomfort that occurred was limited to the individual or between individuals levels and not the group as a whole [8]. Furthermore, the researchers identified a steady linear increase in group tension over the course of the mission. The interplay of individual interactions as well as a broad and continual drain on psychological well-being, albeit minor, are concepts which underpin much of our psychological modeling. It should be noted that specific personality profiles as determined by psychological examination were predictive of interpersonal cohesion in some instances [8], though these considerations are an area for future research. Attempts to connect quantitative biomarkers such as cortisol to psychological examination performance while operating in extreme environments have shown inconclusive results [18]. However, the same study did identify overarching coping mechanisms of a submarine crew. While these mechanisms showed no correlation to the personality profiles, they were beneficial. Traits such as strong achievement motivation and interpersonal orientation were associated with superior coping during submarine missions [18]. In another study of submarine service, optimism, humor accompanied by cynicism, and a strong perception of the submarine service were the primary coping mechanisms exhibited by the crew [9]. Additionally, having some semblance of occasional privacy with their own bed was a highly valuable coping mechanism, a consideration for future research with more granular spatial modeling. Research conducted on Soldiers and Marines during their deployment to Iraq demonstrated that the intensity of combat situations is the primary determinant of behavioral health challenges [2]. Other deployment-issues, such as their length and frequency, also impacted the psychological health of the studied Soldiers and Marines, albeit to a lesser extent than direct combat. The present literature demonstrates that team and individual psychological success in extreme environments can be broadly attributed to coping capacity, which we define as the ability of people, organizations and systems, using available skills and resources, to manage adverse conditions, risk or disasters. Within the model, coping capacity is aggregated from a series of sub-categories and quantitatively incorporated on a Cronbach alpha scale for each individual agent. Cronbach’s alpha scoring is drawn from the psychological changes in a 100-Day remote field groups study, where researchers examined two Australian teams traversing across the Lambert Glacier Basin, paying close attention to personality characteristics, environmental factors, and interpersonal factors as predictors of Group Tensions, Personal Morale, Emotional State, Cognitive Readiness, and the Team’s Work-Life [8]. The agents within the model are categorized into four separate personality categories: Agreeable, Social, Reactive, and Neurotic. Their coping ability and resilience are linked to personality profiles typified by strong instrumentality and achievement motivation combined with interpersonal sensitivity. The four categories are drawn from the Positive Instrumental- Expressive, the Hostile, Competitive Interpersonal orientation, and the low achievement motivation combined with passive-aggressive characteristics linked to Personality Characteristics Inventory (PCI) profiles which NASA has used for screening, monitoring, and evaluating subjects in isolated and confined extreme (ICE) environments [18]. The categorical measurements are as follows: * • Agreeables - Individuals with the lowest degree of competitiveness, low aggressiveness, and not fixated of stringent routine. Cronbach Alpha score starting at 0.97 * • Socials - Individuals with a medium degree of competitiveness, extroverted, require social interaction, but are not fixated on stringent routines. Cronbach Alpha score starting at 0.94 * • Reactives - Individuals with a medium degree of competitiveness, competitive interpersonal orientation, and fixated on stringent routines. Cronbach Alpha score starting at 0.89 * • Neurotics - Individuals with a high degree of competitiveness, highly aggressive interpersonal characteristics, and challenged ability to adapt to boredom or a change in routine. Cronbach Alpha score starting at 0.84 Additionally, each agent is granted skills associated with their civilian and military occupational specialties consistent with NASA’s Human Factors and Behavioral Performance Element research, which analyzed the abilities that are generalizable across circumstances and crew roles and those that will be required by all crew members during a 30-month expedition to Mars [1]. The skills are broken up into two categories; Management (Skill 1) and Engineer (Skill 2), which is extrapolated from NASA’s more comprehensive chart of Leader, Pilot, Physician, Biologist, Geologist, Computer, Electrician, Mechanic, and Combined [1]. The psychological structure of the model is explored in more detail in section 3. ### 2.3 Martian Environment and Local Economy Our challenge is to examine the unique institutional conditions required to manage a settlement at such distance from Earth. In particular, we sought to address the following questions: What conditions are needed to maintain a stable colony on Mars? What combination of personality types would do best in this hostile environment? How many resources are needed for two years between resupplies and assuming occasional accidents? To tackle these questions, we examined research relating to similar questions in four contexts: existing ABMs, economic and related literature, ISS and Antarctica outpost data, and agricultural research. Collectively, the research suggests important design principles for Martian economic modeling. Economic modeling of a remote settlement involves modeling the external dependencies (food, etc.) and the internal labor and trade economy of the settlement itself. Food is the single largest external dependency. For example, food shipments arrive at the McMurtry and other Antarctica settlements only once annually, so this provides us with a starting point for calculating the estimated bi-annual shipments to Mars. Consequently, the frequency and size of external food shipment, as well as the degree to which a Mars settlement would attempt to grow food, are essential variables to calibrate to model a sustainable Mars settlement. Second, waste disposal and recycling (including carbon dioxide from humans and other types of emissions from plants) are costs that must be included. Our current focus is on supporting a stable population on Mars, so waste disposal, recycling, economic gain from mining minerals, and other second order considerations were left for future work. Lastly, the degree of task differentiation in a settlement will directly correlate to the extent that labor and food shipments generate an internal bartering economy. Task differentiation is itself largely driven by the size of the settlement, which is also critical to other parts of the economic model. Developing a Martian economy is also left for future work. ### 2.4 External Martian Economy To examine how a Martian economy might function outside of a colony, we looked at other harsh environments with possible mineral deposits; the oceans and Antarctica. Various countries have exploited mineral deposits in the oceans before there were laws in place to govern mining or for the protection of marine vegetation and wildlife. There is a call now to bring together the fragmented efforts and explicitly address the protection of ocean ecosystems.[12] Exploration of Antarctica has gone differently given the lack of obvious mineral deposits and the harsher environment. After the initial exploration period during colonial times, there was a scientific period when all activity on Antarctica was frozen and all signers of the Antarctic Treaty agreed to collaborative, scientific exploration and experiments were allowed.[17] With the possibility of mineral discovery, efforts are now underway to expand on the original treaty to address additional uses.[3] ### 2.5 Energy Sources in Space Energy sources are another key consideration for colonization. Solar power is limited by daylight hours, seasonal variability, and dust accumulation on solar panels.[11] We chose to model our colony on the nuclear power generator employed by NASA and the U.S. Department of Energy for the Perseverance rover in 2011. The Perseverance uses a Multi-Mission Radioisotope Thermoelectric Generator (MMRTG) power system to provide a steady supply of electricity. The MMRTG is expected to operate for at least 14 years, which is more than seven times longer than the prime mission.[11] Nuclear fission can be utilized to provide surface power on the moon and Mars, and small, light fission reactors could provide up to 10 kilowatts of electrical power for at least 10 years, according to a NASA website. This would provide enough electricity to power several households. Together NASA and DOE plan to design and test one of these low-enriched uranium systems on the moon by the late 2020s.[5] For our Mars settlement we will assume a safe, stable nuclear power generator. ### 2.6 Radiation International space scientists have written on the threat particle radiation poses to future human colonists of Mars. After four years on Mars, colonists would be exposed to a level of radiation above what is assessed safe for humans, according to the article.[19]Additionally the mission should depart for Mars when solar activity is at its peak (the solar maximum), because the increased solar activity deflects the most dangerous particles from distant galaxies. ## 3 ABM description and ODD ### 3.1 KEY ASSUMPTIONS Due to the complexity of the model, a variety of base assumptions were made: 1) Assume the Mars colony has already been constructed, that food, air and water are being produced locally and will not require a start-up time to create. 2) Assume a nuclear generator, similar to the Perseverance power source, has been installed and the habitat has a steady electricity source for at least 7 years. ### 3.2 PROBLEM FORMULATION The main goal of the model is to simulate a theoretical habitat environment on Mars for a colony carrying out the mining of minerals to be sent back to Earth. There is an emphasis on the mental state of each Martian agent, since that directly impacts the success of the mission. ### 3.3 MODEL DESCRIPTION 1\. Purpose The agent-based model MARS-COLONY is intended to provide insights into the possible scenarios that might develop from varying initial population sizes, personality type and skill levels of colonists sent to run a mining colony for minerals. The goal is to aid decision-makers with planning by highlighting potential pit falls. To make the model as realistic as possible, research is used on the studied coping capacities of common personality types, and random accidents are introduced both transporting food from Earth, as well habitat disasters on Mars that also lower food supplies. Accidents not only threaten food supplies, but considerably increase stress levels of colony members. The model was developed using NetLogo 6.2. 2\. Entities, State Variables, Scales There are two collectives and two agent types in the model. The first agent type is the Martian, which represents a single human member of the settlement. The entire group of Martian agents form the collective “Martians”. The second agent type is the stressor, which can be triggered by a settlement habitat accident or an earth shipping disaster. When triggered, the stressor interacts with the Martians in the settlement to reduce their coping capacity and health. A stressor can dissipate over time, or can be removed by the Martian agents if they succeed a skills check to recover from the habitat accident. Table 1 lists the attributes and starting attribute values for Martian agents, and Table 2 lists the attribute and starting values for stressor agents. Attribute Name | Description | Starting Value (or Range) ---|---|--- coping-capacity | psychological coping capacity of agents | 0.84 – 0.98 resilience | coping category | “neurotic”, “reactive”, “social”, or “agreeable” skill-1 | value of skill 1 | 0 - 100 skill-2 | value of skill 2 | 0 - 100 health | agent health value | 100 food | food available for the given agent | 1 air | air available for the given agent | 1 water | water available for the given agent | 1 waste | waste produced by the given agent | 1 partner | the martian with which the agent is partnered, or ”nobody” if not partnered | nobody taskmate-food | taskmate for food production | nobody taskmate-water | taskmate for water production | nobody taskmate-air | taskmate for air production | nobody taskmate-accident | taskmate for habitat accident recovery | nobody taskmate-waste | taskmate for waste removal | nobody Table 1: Agent Attributes: Martian Attribute Name | Description | Starting Value (or Range) ---|---|--- identity | two types of stressor identity externalities | “Shipping” or “Habitat” Table 2: Entity Attributes: Stressor Figure 1: Martian Attributes The settlement acts as additional entity, capable of storing air, food, water, and waste, and having a set production capacity for these resources. Martian agents interact with the settlement by working together to produce resources, store resources, and consume settlement resources. The settlement is a 50x50 grid, with each grid cell having an individual resource production capacity, but collective and unlimited storage. The geospatial location of each Martian agent is randomly assigned during model initialization, and agents move randomly around the grid. Table 3 lists the attributes and starting values for the Martian settlement. Attribute Name | Description | Starting Value (or Range) ---|---|--- p-food | the rate at which food, on average, can be grown per time step in each patch of the settlement | 0.5 p-air | air recycled per patch of the settlement | 5.88 p-water | water recycled per patch of the settlement | 28 p-waste | non-recyclable waste (i.e., fecal matter, plant emissions) processed per patch of the settlement | 0 settlement-air | total air available for settlement | 5.88 * number-of-martians * 156 settlement-water | total water available for settlement | 28 * number-of-martians * 156 settlement-food | total food available for settlement | 10.5 * number-of-martians * 156 settlement-waste | total waste produced by settlement | 0 settlement-minerals | total minerals available for settlement | 0 Table 3: Entity Attributes: Settlement The environment is the final entity, which determines the probability of a habitat accident, earth shipment frequency and success probability, time scale, technological efficiency, and production skill requirements. Table 4 lists the environmental attributes and starting values. Attribute Name | Description | Starting Value (or Range) ---|---|--- ticks | time step for model | 1 tick = 1 week minerals-shipment | amount of minerals received in each shipment from Earth | 100 food-shipment | amount of food received in each shipment from Earth | 10.5 * number-of-martians * 78 shipment-frequency | frequency of shipments from Earth | 78 ticks technology | technological efficiency | 0.5 energy | energy (currently unused) | 1.0 food-s1 | amount of skill 1 required to produce food | 0 - 100 food-s2 | amount of skill 2 required to produce food | 0 - 100 water-s1 | amount of skill 1 required to produce water | 0 - 100 water-s2 | amount of skill 2 required to produce water | 0 - 100 air-s1 | amount of skill 1 required to produce air | 0 - 100 air-s2 | amount of skill 2 required to produce air | 0 - 100 waste-s1 | amount of skill 1 required to produce waste | 0 - 100 waste-s2 | amount of skill 2 required to produce waste | 0 - 100 accident-s1 | amount of skill 1 required to recover from a habitat accident | 0 - 100 accident-s2 | amount of skill 2 required to recover from a habitat accident | 0 - 100 Table 4: Entity Attributes: Environment 3\. Process overview and scheduling The only user input parameter currently in the model is the initial number of Martians, which ranges from 4 – 152, in increments of 4. When the model is initialized, the total number of Martians are divided equally between four possible psychological categories: neurotic, reactive, social, and agreeable. Martians, Stressors, the Settlement, and the Environment are initialized with the values listed in Tables 1 – 4. At each tick, Martians are able to move, sleep, team up with other Martians, produce resources, consume resources, and engage socially. To move, the Martians face a random direction and move forward one grid cell. Next, they receive additional health from sleeping, with total health capped at 100. Martians are then able to partner with other Martians to perform a skills check for production. If the skills check is successful, they produce food, water, and air, each with a unique skills check. Waste is produced as a by- product of successful production. After resources are produced, they are then consumed. Each Martian has a weekly requirement for food, water, and air. If an individual Martian produces enough resources to meet this requirement, they consume the defined amount and any remainder is contributed to the settlement resource pool. If the Martian did not produce enough resources to meet the weekly requirement, they consume the required resource amount from the settlement resource pool. If the settlement resource pool does not have enough resources to meet a Martian’s weekly requirement, the Martian’s health level is reduced. Once the Martian’s health reaches zero, the Martian dies and is removed from the model. Additionally, there is a small random chance of death for each Martian to represent unforeseen mortality. The settlement restores its production capacity for each grid cell and may consume resources at each tick. Martian production is limited by the total capacity of each settlement grid cell, and reduces the amount of producible resources on that cell during the time step. At the next time step, the resources are renewed to maintain a set weekly production capacity. The settlement may also consume minerals, which are received from earth shipment, to improve technological efficiency, which then improves Martian production capability. At each time step, the environment has a chance to add new Martians, experience an earth shipping disaster, receive an earth shipment, and to experience a habitat accident. Every 78 weeks, there is a chance that 4 new Martians may be added to the colony from thr resupply shuttle, with a random psychological category assigned upon addition. Also at 78 weeks, an earth shipment may be received or there is a small random chance that an earth shipping disaster may occur. If the earth shipment is received, the settlement receives food and minerals, and improves technological efficiency. If an earth shipping disaster occurs, the shipment is not received, and an earth shipping disaster stressor is generated. There is also a small random chance of a habitat accident occurring at each time step. If a habitat accident does occur, it randomly affects either the food, water, air, or mineral settlement resource supplies, and reduces the selected resource by half. It also generates a habitat accident stressor. If a stressor is generated, it has a chance to interact with Martians or to dissipate over time. Habitat accident stressors persist until they dissipate or are removed by a successful Martian skill check. While active, they continue to reduce the settlement resource supplies. After 4 time steps, the habitat accident stressor will dissipate and no longer affect the settlement resource supplies. 4\. Design Concept _Basic Principles_ The model was developed to explore the psychological, social, technological, economic, and logistical factors that would influence the long-term viability of a human Martian settlement. Martian settlers interact with each other while producing and consuming resources, which affects their overall health. They are divided into four personality types that determine the degree and direction of this effect. Each Martian also possesses a skillset that determines their success in producing resources for themselves and the settlement. The settlement is supplied by regular shipments from earth, which are susceptible to shipping disasters, and the settlement can experience a habitat accident, both of which negatively influence settlement supplies and overall Martian health. _Emergence_ The primary observed emergent phenomenon occurs in the decline of the Martian population. While the members of the settlement have an equal probability of being affected by lack of settlement resources, habitat accidents, or earth shipping disasters, Martians with the “neurotic” psychology die at a much higher rate than those of other psychologies, and once their population reaches a low enough level, the settlement population stabilizes. Martians with the neurotic psychology and a high coping capacity benefit the least from interaction with other Martians, and are penalized the most if they have a low coping capacity. Our results suggest that this effect is a driver of the Martian population decline, and once minimized or removed, can produce a stable settlement. _Adaptation_ The Martian agents do not have explicit adaptation, but do have implicit adaptation in the use of their assigned skills. During each time step, the model seeks to partner two Martian agents whose skills sum to the required production threshold for resources or for habitat accident recovery in the case of a habitat accident. This represents the idea that Martian settlers would work together based on their actual skills to accomplish various tasks needed to provide for the settlement, and would adapt as needed for these tasks. _Objectives_ The overall objective of the model is to produce a long-term stable Martian population. Each Martian’s objective is to partner with other Martians to produce resources and recover from habitat accidents. If Martians are unable to produce sufficient resources, the population will begin to decline, making future production more difficult due to a decline in the available skills required for production. _Sensing_ Martian agents move around the grid space and are able to sense other agents within 3 grid spaces of themselves. If there are one or more Martian agents within this range, a random agent is selected as a partner, and the agents interact to increase or decrease overall health and coping capacity, depending on current coping capacity and each agent’s psychological category. _Interaction_ In addition to interacting with agents in close spatial proximity, as described in the section above, Martian agents also interact to produce resources, recover from habitat accidents, and remove waste from the settlement. Pairs of agents are selected by the model based on their skill levels to achieve a set productivity threshold, which results in successful achievement of these activities. If a pair cannot be made to reach the threshold, the unpaired agents are not successful in production, recovery, or waste removal. Stressor agents interact with the settlement to reduce resource supplies or prevent an earth shipment from resupplying food and minerals. _Stochasticity_ Stochasticity appears in the assignment of agent skill levels and coping capacity, Martian agent movement, Martian death, Martian agent addition, earth shipping disaster, habitat accidents, and Martian agent teaming for production and interaction. When Martian agents are generated, levels for their skills are assigned from a random uniform distribution for the first skill, and the second skill then sums with the first to equal one hundred. Martian agents move randomly around the grid space, and face a small random probability of death at each time step. There are also random probabilities of Martian agents being added to the model, the occurrence of an earth shipping disaster when a shipment is due to arrive, or a habitat accident occurring, with the type of resource affected by the habitat accident also randomly selected. When agents pair up for resource production, habitat accident recovery, or waste removal, the pairing is random given the pair reaches the production threshold requirement. If agents are within three grid cells of each other, they can randomly be assigned to interact to affect health and coping capacity. _Collectives_ Martian agents form a settlement collective, drawing on a pool of shared resources and producing an overall amount of settlement waste. If an individual agent is unable to produce enough resources for their own needs in a given time step, they may draw on the collective settlement resources. If they produce excess resources, they are contributed to the settlement resource pool. _Observation_ Outputs collected from the model for observation include coping capacity statistics, number of Martian agents, total settlement resources, shipments received, shipping disasters and habitat accidents experienced, population coping and resilience, overall Martian health, and time elapsed. 5\. Initialization The setup procedure imports a Martian landscape of geographic files. The global variables are then checked for total water, food and air for the habitat. Also total energy and technology levels needed to run the habitat and mine minerals. Additional parameters are assigned for the frequency of shipment and habitat accidents. Every 78 ticks there is also a slight chance of a birth or death in the community. The individual Martian attributes are then calculated by dividing resources equally, and randomly assigning personality types and skill. 6\. Input Data Geographic Information System (GIS) files of the Mars landscape have been loaded for awareness, but no input data is used in this model. Mars exploration is still in the early stages and applicable data is not yet available. ### 3.4 SETTING PARAMETERS One of the challenges of agent-based modeling is determining how to set the initial parameters for the model. Empirical data are often a source of initial parameters, even in far-flung scenarios such as past civilizations or extra- planetary settlements. Our model draws upon a number of sources to set our parameters. Global Variables. The Earth shipment and shipment frequency variables are discussed above. This discussion also excludes shipping and accident counter variables, which are discussed under results. The other global variables include the settlement’s total resource stockpiles (settlement-air, settlement-food, and settlement-water). The values of the parameters are primarily derived from a review of the resource storage and production capabilities of the International Space Station (ISS). NASA states that a single ISS resident consumes 5.88kg of air, 28L of water, and 10.5kg of food in a given week.[7][13][15] Our initial settlement-food, settlement-air, and settlement-water assume a ramp-up, pre-settlement series of deliveries that stockpile supplies equal to two earth shipment cycles (156 weeks/time steps). However, there are proven technologies from the ISS to allow for the sustainable production of air and water. Conversely, there is as yet no method for the sustainable production of food in low gravity. As a result, we assume that earth shipments will deliver food but not water or air. Earth shipments deliver sufficient food for one additional earth shipment cycle (10.5kg of food per settler per week). Our global variables also define several other rates necessary to the function of the settlement. The technology parameter sets a rate of aggregate technological efficiency. This term serves as a modifier to the amount of production generated by labor in the settlement. Technology is initially set at 0.5 For this early version of the model, energy serves as a similar term that specifies the amount of resource consumption required for the settlement to function. This placeholder value is set at 1.0. Lastly, the global variables set the skill levels necessary for settlement production functions. An assumed score of 100 is needed across two sets of skills (-s1 and -s2) to successfully accomplish each of four tasks (food, water, air production, and accident recovery). Each pair of scores related to resource production is set with the same values. For example, food-s1 is set as a random integer between 0 and 100. Food-s2 is set as the difference between 100 and food-s1 (100 - food-s1). Accident recovery sets both both accident-s1 and accident-s2 as a random integer from 0 to 100. This represents the ex ante uncertainty about what types of skills may be needed for a given emergency. It is important to remember that these values set the required skill score that must be met for successful production; they do not guarantee that settlers will have these skills. Settler Variables. Settlers are assigned two skills. Skill 1 is set as a random number from 0 to 100. Skill 2 is set as 100 - Skill 1, such that each settler has a total skill level of 100. Settlers have a partner variable (all settlers begin unpartnered) and indicator variables that store their task assignment. Each settler is created with one of four resilience types: nuerotic, reactive, social, and agreeable. Each of these resilience types are assigned a coping score based on our research in team function and group theory. These coping scores are 0.84, 0.89, 0.94, and 0.98. Coping scores determine how settlers continue to produce and consume resources after adverse events such as a habitat accident. Settlement (Patch) Variables. Patch variables in this model represent the modest estimated amount of food production that could be generated in a garden or farm (0.5 kg per patch per time step), the WRPS + WTS air production system (5.88kg per patch per time step), and OPS water recycling capacity (28L per patch per time step). These production levels are activated in the production functions only with the requisite levels of settler skill levels, as discussed above. Exogenous Variables. This model contains only two exogenous variables: the starting number of settlers and the number of time steps in a given model run. ### 3.5 IMPLEMENTATION VERIFICATION Our code shows cyclical variability in the population sizes and is able to stabilize. The code appears to run correctly, the colonists have plenty of resources, and we are able to focus on the interaction between the different personality types. However we are unable to validate the values since there has not been an actual Mars colony we can compare data to so model data serves largely to highlight problems for consideration. ## 4 Scenarios for a Stable Mars Settlement ### 4.1 A Stable Population Size Our goal was to see which initial population sizes lead to a stable colony size. We did 5 runs of our model, for 28 Earth years, varying the initial population size from 10 to 170 by steps of 10. Given that there are 4 critical tasks that are needed continuously (air, water, food production and waste removal) in addition to handling disasters, and two skills needed for each task, we chose a population size of 10 as the minimum needed for a ”stable” colony size. The population is allowed to dip below 10 as long as it bounces back within 1.5 years, or the amount of time between Earth resupply shuttles. A plot of our runs below shows that all initial population sizes above 50 were able to maintain a population above 10 for all time steps. We will therefore only consider initial populations below 50 as we try to determine a minimum initial population which is able to maintain, or bounce back quickly to, a stable colony size equal to or greater than 10 for all 28 years. Figure 2: Variation in number of agents at each time Step (one week) in the simulation. Since our 4 personality types are equally distributed at setup, for our focus on smaller initial populations we started at 10 (the minimum for a stable colony), then increased to 50 with a step size of 4. Each time tick in our model represents a week, so 1.5 years is equivalent to 84 ticks. For an initial population of 10, the plot below shows the population sizes over 28 years with prolonged time periods when the population drops below 10 Figure 3: Population sustainability with a start of population = 10. For an initial population size of 22, the plot below show the population sizes over 28 years remains above 10, meeting our definition of a stable colony. Figure 4: Population sustainability with a start of population = 22. A summary of which small initial population sizes resulted in stable colonies over 28 years, in 5 runs, is given in the table below. For the purpose of this model, an initial population of 22 was the minimum required to maintain a viable colony size over the long run. Initial Population Size | Does Population Size Bounce Back Above 10 or Not ---|--- 10 | No Bounce Back 14 | No Bounce Back 18 | No Bounce Back 22 | Successful Bounce Back 26 | No Bounce Back 30 | Successful Bounce Back 34 | Successful Bounce Back 38 | No Bounce Back 42 | Successful Bounce Back 46 | Successful Bounce Back 50 | Successful Bounce Back Table 5: Small Initial Populations Which Do and Do Not Lead to Stable Colonies ### 4.2 An Agreeable Personality Type Does Best In all runs, the Agreeable personality type was the only one to survive the full duration of model runs. This is likely because it has the highest coping capability, and after long periods of time every agent has been exposed to a series of stressor interactions, as well as space and habitat accidents. The picture of the MARS-COLONY NetLogo model below includes a visualization of total personality types over time, and the yellow Agreeable personality type is most resilient, the Neurotic personality type was the most likely to fail, the Reactive and Social personality types alternated in between. While this model assigns equal numbers of each personality type, future work could try adjusting the proportion of each to possibly lead to a lower required minimum initial population. For example, a crew of all Agreeable personalities may be more successful. Figure 5: Screenshot of the simulation in NetLogo. ### 4.3 Maintaining Sufficient Resources Despite Accidents Our data analysis shows that our food supplies are more than sufficiently resupplied with earth shipments, for the majority of time steps, even in scenarios where there are shipping disasters. Since the focus of this study was to examine the interplay of the 4 personality types we did not assume any production of air, water or food on the colony. Instead we provided multiple years of initial supplies, with resupplies on incoming shuttles. Similarly we were not concerned with the amount of waste generated by the colonists. Future work can focus on this area of the model and explore the trade off between resources needed from Earth, and amounts that could be generated on Mars. This is a vital component to establishing a stable colony which adds to the complexity of the model. ## 5 Conclusion In summary, the MARS-COLONY Agent-Based Model offers a preliminary look at necessary conditions to establish a stable mining colony on Mars. The main focus is on the personality types of colonists selected and how they perform throughout their time on Mars, using their skills to mine minerals and react to random resupply shuttle accidents or habitat disasters. The stress caused by accidents, as well as from interacting with other colonists, takes a toll and Agreeable personality types were assessed to be the most enduring for the long term, whereas Neurotics showed least adaptation capacity. This study assumed a steady available nuclear energy supply and sufficient initial air, water and food with sufficient resupply in order to focus on personality and mental health of colonists, however those are recommended areas of future work. ## References * [1] Jack Stuster Jurine Adolf Vicky Bryne and Maya Greene “Generalizable Skills and Knowledge for Exploration Missions” In _NTRS - NASA Technical Reports Server_ , 2019 DOI: https://ntrs.nasa.gov/citations/20190023004 * [2] Carl A. Castro and Dennis McGurk “The Intensity of Combat and Behavioral Health Status” In _Traumatology_ Vol.13, No.4, 2007, pp. 6–23 DOI: 10.1177/1534765607309950 * [3] Yen-Chiang Chang “Chinese legislation in the exploration of marine mineral resources and its adoption in the Arctic Ocean” In _Ocean and Coastal Management_ 168, 2019, pp. 265–273 DOI: https://doi.org/10.1016/j.ocecoaman.2018.11.012 * [4] Daniel Clery “Liquid water spied deep below polar ice cap on Mars” URL: https://www.science.org/content/article/liquid-water-spied-deep-below-polar-ice-cap-mars * [5] Emma Derr “Nuclear Taking Us Faster & Farther Into Space” URL: https://www.nei.org/news/2021/nuclear-taking-us-faster-and-farther-into-space * [6] Charles Dunnill “Method of Making Oxygen from Water in Zero Gravity Raises Hope for Long-Distance Space Travel” URL: https://www.scientificamerican.com/article/method-of-making-oxygen-from-water-in-zero-gravity-raises-hope-for-long-distance-space-travel/ * [7] National Geographic “SIMOC” URL: https://www.nasa.gov/mission_pages/station/research/benefits/water_in_space * [8] Joanna Wood Desmond J. Lugg Sylvia J. Hysong and Deborah L. Harm “Psychological Changes in Hundred-Day Remote Antarctic Field Groups” In _Environment and Behavior_ 31, 1999, pp. 299–337 * [9] Shaul Kimhi “Understanding Good Coping: A Submarine Crew Coping with Extreme Environmental Conditions” In _Psychology_ Vol.2, No.9, 2011, pp. 961–967 DOI: 10.4236/psych.2011.29145 * [10] Timothy A. Kohler and George J. Gumerman “Dynamics in Human and Primate Societies”, Santa Fe Institute Studies in the Sciences of Complexity Oxford University Press, 2000 * [11] NASA Jet Propulsion Laboratory “Mars 2020 Perseverance Launch Press Kit” URL: https://www.jpl.nasa.gov/news/press_kits/mars_2020/launch/mission/spacecraft/power/ * [12] Kathryn J. Mengerink Cindy L. Van Dover Jeff Ardron Maria BakerΩElva Escobar-Briones Kristina Gjerde J. Anthony Koslow Eva Ramirez-Llodra AnaΩLara-Lopez Dale Squires Tracey Sutton Andrew K. Sweetman Lisa A. Levin “A Call for Deep-Ocean Stewardship” In _Science_ 344, 2014, pp. 6185 DOI: DOI: 10.1126/science.1251458 * [13] NASA “Water Production in Space: Thirsting for a Solution”, 2015 URL: https://www.nasa.gov/mission_pages/station/research/benefits/water_in_space * [14] Tripp Driskell Eduardo Salas and James E. Driskell “Teams in extreme environments: alterations in team development and teamwork” In _Human Resource Management Review_ 28, 2018, pp. 434–449 DOI: http://dx.doi.org/10.1016/j.hrmr.2017.01.002 * [15] Michelle Starr “SIMOIC” URL: https://www.nationalgeographic.org/media/simoc * [16] Gillian D. Triggs “The Antarctic Treaty regime: Law, Environment and Resources” Cambridge University Press, 1989 * [17] Gillian D. Triggs “The Antarctic Treaty regime: Law, Environment and Resources” In _Military Psychology_ 11:4, 1999, pp. 381–404 DOI: https://doi.org/10.1207/s15327876mp1104˙3 * [18] Gro M. Sandal Inger M. Endresen Ragnar Vaernes and Holger Ursin “Personality and Coping Strategies During Submarine Missions” In _Military Psychology_ 11:4, 1999, pp. 381–404 DOI: https://doi.org/10.1207/s15327876mp1104˙3 * [19] Chris Young “Mars Is Safe for Humans, But Only for Four-Year Missions” URL: https://interestingengineering.com/mars-is-safe-for-humans-but-only-for-four-years
# Game of arrivals at a two queue network with heterogeneous customer routes Agniv Bandyopadhyay School of Technology and Computer Science Tata Institute of Fundamental Research Mumbai, India &Sandeep Juneja School of Technology and Computer Science Tata Institute of Fundamental Research Mumbai, India ###### Abstract We consider a queuing network that opens at a specified time, where customers are non-atomic and belong to different classes. Each class has its own route, and as is typical in the literature, the costs are a linear function of waiting and service completion time. We restrict ourselves to a two class, two queue network: this simplification is well motivated as the diversity in solution structure as a function of problem parameters is substantial even in this simple setting (e.g., a specific routing structure involves eight different regimes), suggesting a combinatorial blow up as the number of queues, routes and customer classes increase. We identify the unique Nash equilibrium customer arrival profile when the customer linear cost preferences are different. This profile is a function of problem parameters including the size of each class, service rates at each queue, and customer cost preferences. When customer cost preferences match, under certain parametric settings, the equilibrium arrival profiles may not be unique and may lie in a convex set. We further make a surprising observation that in some parametric settings, customers in one class may arrive in disjoint intervals. Further, the two classes may arrive in contiguous intervals or in overlapping intervals, and at varying rates within an interval, depending upon the problem parameters. _K_ eywords strategic arrivals $\cdot$ queuing network games $\cdot$ population games. ## 1 Introduction Queueing games or games where strategic customers, served by a queue or a queuing network, decide on actions such as which queue to join, whether to join, when to join, what level of priority to select and so on, are well studied in the literature (see [6, 7, 9] for surveys). In this paper we focus on the queuing arrival game where customers decide on when to join a queuing network. This is in contrast to much of the literature that considers arrival games to a single queue (see, e.g., [2, 5, 8, 9, 12, 13, 15]). Applications of such arrival games to multiple queues are many: Customers arriving to a health facility where they queue up to meet a doctor and then may queue to get tests done and/or procure medicines; in banks where different customers may need to go to different and multiple counters; in cafeteria where customers queue up for different and multiple food items, and so on. In consonance with much of the existing literature, we model customers as non-atomic ‘infinitesimal’ fluid particles with costs that are linear in waiting time and in time to service, customers are served in a first-come-first-serve manner and service facility opens at time zero (see [11, 12, 15]). In [13], uniqueness of equilibrium solution was proven in a single queue setting with stochastic service rates and large no of users. Moreover, [13] showed that, the equilibrium solution with a large no of users is well approximated by the corresponding fluid system and thus lending support to the fluid analysis. This fluid setting has resulted in unique and elegant, easily calculable customer equilibrium profiles for single queues as well as certain symmetric queuing networks where customers have homogeneous travel routes (see [11, 12]). To keep the discussion simple we focus on a two queue, two class customer setting where each class of customers has a distinct route, and customers in each queue are served in a first-come-first-served manner. While this set-up may be practically interesting, our main contributions are theoretical: Our key aim is to test whether the elegance and simplicity of customer equilibrium behaviour to a single queue extends to more general queuing networks in presence of heterogeneous routing. Even in the simple two queue setting, we see that, unlike for the single queue, here the solution structure and order of arrivals in equilibrium is a function of all the problem parameters, i.e., linear coefficients of the cost function, the queue service rates and the population size of each class of customers. For one set of customer travel routes we observe that depending upon problem parameters, there exist eight distinct solution structures. This suggests that as the number of queues increase there may be a rapid blow-up in the solution structures. This may make the problem of identifying and learning the correct structure computationally prohibitive. In this paper, we do not address the issue of customers learning the equilibrium profile by repeatedly playing the game. Limiting behaviour of players repeatedly updating their action in a game using a simple rule, often called a ‘no-regret’ policy, are studied in [1] and we refer the reader to [3] for a comprehensive exposition of this literature. Our other broad contributions/observations are: 1) We find that similar to the single queue setting, the equilibrium profile of arriving customers is unique for a wide set of parameters. However, interestingly, when customer cost preferences across classes are identical up to a constant, there may be multiple equilibrium arrival profiles, all lying in a convex set that we identify. Although there are many arrival profiles in equilibrium in this case, they all have identical social cost. 2) In [12], the equilibrium profile is determined for the case when multiple classes of customers with linear costs are arriving at a single queue again. They find that different classes of customers arrive in non-overlapping and contiguous intervals. In our two queue setting we find that depending upon the problem parameters, in equilibrium, arrivals may come in non-overlapping and contiguous intervals, in overlapping intervals, or under certain parametric settings, a class of customers may even arrive in disjoint intervals. Moreover, we show that whether the classes will arrive over overlapping sets or not is independent of the population sizes and decided entirely by the queue service rates and customer preferences. Related literature: The arrival games to queues were first considered by [5]. The concert queueing game is the fluid setting was introduced in [12]. The arrival game in a fluid network of bottleneck queues including tandem, Trellis, and general feed-forward networks was considered in [11], where they characterized the equilibrium arrival profile in each of these topologies. Transportation modelling community has extensively studied arrival games. Vickrey in [16], introduced the morning commute problem. Unlike the concert queuing game, in these transportation problems, the service facility has no predetermined opening time. Instead, the customers have a preferred time to complete service and a cost for arriving too early or too late (see [10]). This led to a huge literature on arrival games to a bottleneck queue, the impact of tolls, etc. (see [14] for an extensive list of references). Much of the transportation literature considers single queue settings. Lindsey, in an influential work [15], establishes the existence of equilibrium arrival profile for multiple classes of customers with general non-linear cost functions arriving at a bottleneck queue with a constant service rate, through intricate fixed point arguments. Our work differs from transportation literature in that we consider a two queue network with heterogeneous arrival routes, linear costs, and in this setting we are able characterize the equilibrium user arrival profiles in a closed form, and for a large class of parameters, show that these profiles are unique. Outline of the paper: In Section 2 we provide the background to the arrival queueing game and overview the two-class, two queue, two-route networks that we consider. We emphasize on two heterogeneous routes networks 1) where the departures of the two classes are through different queues (Heterogeneous Departure System or HDS) and 2) where the arrivals enter at different queues (Heterogeneous Arrival System or HAS). In Section 3, we identify the equilibrium arrival profile for all possible parameters for arriving customers for HDS. In particular, we see that these parameters can be partitioned into four distinct regions each having a separate solution structure, when the two customer classes have unequal preferences. In Section 4 we similarly analyze HAS. Here we discover that the parameter space can be partitioned into eight regions based on the solution structure, when the two customer classes have unequal preferences. Moreover, for both HDS and HAS, when the groups have identical preference, we identify a parametric regime where unique equilibrium exists, as well as a parametric regime where the equilibrium is non-unique and the set of equilibrium profiles is convex. We end with a brief conclusion in Section 5. In the main body, we have confined our discussion to the main proof ideas behind our results and have kept the detailed proofs in the appendix. ## 2 Preliminaries ### 2.1 Fluid Model We consider a fluid model having two classes of customers or users. The size of each class $i=1,2$ is given by a positive quantity $\Lambda_{i}>0$. In every class $i=1,2$ individual users are infinitesimal and the set of all users in class $i$ is given by the points in the interval $[0,\Lambda_{i}]$. We define functions $F_{i}:\mathbb{R}\to[0,\Lambda_{i}]$ for $i=1,2$ such that, $F_{i}(t)$ denotes the amount of users of class $i$ that arrive by time $t$. We call $F_{i}$ the arrival profile of class $i$ users. Therefore, each $F_{i}$ is non-decreasing and satisfies $F_{i}(-\infty)=0$ and $F_{i}(+\infty)=\Lambda_{i}$. We consider $F_{i}$ that are right-continuous and can be expressed as a sum of a non-decreasing absolutely continuous function and a non-decreasing discrete function. We call the pair $\mathbf{F}=\\{F_{1},F_{2}\\}$ as the joint arrival profile of the two classes. We consider a network comprising of two queues, both starting service at time $t=0$. Let $\mu_{1}$ and $\mu_{2}$, respectively, denote the deterministic fixed service rates at the two queues after they start service. We consider four routes of the two arriving classes to the two queues. These are displayed in Table 1. Instance I Instance II Instance III (Heterogeneous Departure System or HDS) Instance IV (Heterogeneous Arrival System or HAS) Table 1: Various instances with two groups traveling through two queues connected in tandem Instance I is equivalent to two groups of users arriving at a two-layer tandem network to travel by the same path. By Theorem 5 of [11], the instance is equivalent to the case where the two groups are arriving at a single queue of capacity $\min\\{\mu_{1},\mu_{2}\\}$. Instance II is equivalent to the case where the two queues independently serve the two groups and therefore is equivalent to two independent instances of a single queue with a single class customer arrivals. Hence, the first two instances are reducible to instances with just one queue studied in [12]. In this paper we study the arrival equilibrium behaviour in the other two instances III and IV. We refer to them as Heterogeneous Departure (HDS) and Heterogeneous Arrival Systems (HAS), respectively. ### 2.2 Waiting and Departure Times To specify the waiting and departure times in a system, first consider a single queue setting where $A(t)$ denotes the total mass of users of all classes that have arrived at the queue by time $t$. Let $\mu$ denote the service rate. Then at time $t$, the length of the waiting queue developed in that queue will be (see Theorem 6.5 in [4]): $\displaystyle Q(t)$ $\displaystyle=A(t)-\mu\cdot\max\\{t,0\\}+\sup_{s\in[0,t]}\max\left\\{\mu s-A(s),0\right\\}.$ (1) We assume that if there is a jump in the arrival profile $A$ at time $t$, the arrivals are positioned in the queue at uniformly random order. As a result, a user arriving at that queue at time $t$ will suffer an expected waiting time of $\displaystyle W(t)$ $\displaystyle=\frac{Q(t+)+Q(t-)}{2\mu}+\max\\{0,-t\\},~{}\text{and departs at time}~{}~{}\tau(t)=W(t)+t,$ (2) where $Q(t+)$ and $Q(t-)$ respectively denote the right and left limits of $Q$ at time $t$. Note that if the queue length process $Q(\cdot)$ is continuous (which is the case if $A(\cdot)$ is absolutely continuous), waiting time as a function of time $t$ will be $W(t)=\frac{Q(t)}{\mu}+\max\\{0,-t\\}$. If the arrival profile $A(\cdot)$ is absolutely continuous, by (1) and (2), the departure time as a function of time $t$ will be: $\tau(t)=\frac{A(t)}{\mu}+\sup_{s\in[0,t]}\max\left\\{0,s-\frac{A(s)}{\mu}\right\\}$. Whenever $Q(t)>0$, the term $\sup_{s\in[0,u]}\max\\{\mu s-A(s),0\\}$ is independent of the choice of $u\in[t-\delta,t+\delta]$ for $\delta>0$ and sufficiently small, and when $t<0$, $\tau_{1}(t)=\frac{A(t)}{\mu}$. If $A(\cdot)$ is absolutely continuous, its derivative $A^{\prime}(\cdot)$ will exist a.e. As a result, $\displaystyle\tau^{\prime}(t)$ $\displaystyle=\frac{A^{\prime}(t)}{\mu}~{}~{}\text{a.e. in the closure of the set of times}~{}\\{s~{}|~{}s<0~{}\text{or}~{}Q(s)>0\\}.$ (3) The above observation will be useful in our analysis of HDS and HAS in the later sections. ###### Definition 2.1. We say the queue is _engaged_ at time $t$ if $t$ lies in the closure of the set $\\{s~{}|~{}Q(s)>0\\}$. By (3), after the queue starts serving, users depart at rate $\mu$ whenever the queue is engaged. We introduce the following notation: * • $A_{i}(t)$ be the total mass of customers of both the groups who have arrived at queue $i=1,2$ till time $t$ (Note that while $F_{i}$ denotes arrival profile corresponding to user class $i$, $A_{i}$ denotes overall arrival profile to queue $i$.). * • $Q_{i}(t)$ and $W_{i}(t)$ be the length of the waiting queue, and the waiting time that a customer arriving in queue $i$ at time $t$ will observe. Let $\tau_{i}(t)$ denote the time that customer will depart the system. * • $W_{\mathbf{F}}^{(j)}(t)$ and $\tau_{\mathbf{F}}^{(j)}(t)$ be the waiting and departure times from the network suffered by a class $j$ user arriving at time $t$ for $j\in\\{1,2\\}$. Explicit dependence on $\mathbf{F}$ in this notation is useful to our analysis later. For both the queues $i=1,2$ upon defining the arrival profile $A_{i}(\cdot)$, using (1) and (2), $Q_{i}(\cdot)$, $W_{i}(\cdot)$ and $\tau_{i}(\cdot)$ are well-defined. Now we specify the waiting and departure times of both the queues in HDS and HAS as functions of time, under the assumption that the joint arrival profile $\mathbf{F}=\\{F_{1},F_{2}\\}$ is absolutely continuous (we later argue by Lemma 2.1 that considering absolutely continuous joint arrival profiles are sufficient for identifying equilibrium behavior). 1. HDS : Arrival profiles at individual queues are $A_{1}(t)=F_{1}(t)+F_{2}(t)$ and $A_{2}(t)=F_{2}(\tau_{1}^{-1}(t))$, where $\tau_{1}^{-1}(t)=\sup\\{s~{}|~{}\tau_{1}(s)\leq t\\}$. Both $A_{1}(\cdot)$ and $A_{2}(\cdot)$ are absolutely continuous. With this, $W_{\mathbf{F}}^{(1)}(t)=W_{1}(t)$, $\tau_{\mathbf{F}}^{(1)}(t)=\tau_{1}(t)$, $W_{\mathbf{F}}^{(2)}(t)=W_{1}(t)+W_{2}(\tau_{1}(t))$, and $\tau_{\mathbf{F}}^{(2)}(t)=\tau_{1}(t)+W_{2}(\tau_{1}(t))$. 2. HAS : Arrival profile at individual queues are $A_{1}(t)=F_{1}(t)$ and $A_{2}(t)=F_{1}(\tau_{1}^{-1}(t))+F_{2}(t)$ where $\tau_{1}^{-1}(t)=\sup\\{s~{}|~{}\tau_{1}(s)\leq t\\}$. Both $A_{1}(\cdot)$ and $A_{2}(\cdot)$ are absolutely continuous. With this, $W_{\mathbf{F}}^{(1)}(t)=W_{1}(t)+W_{2}(\tau_{1}(t))$, $\tau_{\mathbf{F}}^{(1)}(t)=\tau_{1}(t)+W_{2}(\tau_{1}(t))$, $W_{\mathbf{F}}^{(2)}(t)=W_{2}(t)$, and $\tau_{\mathbf{F}}^{(2)}(t)=\tau_{2}(t)$. ### 2.3 Solution Concept We assume that every user in group $i$ ($i\in\\{1,2\\}$) has a cost function linear in her waiting and departure times from the network given by: $C_{\mathbf{F}}^{(i)}(t)=\alpha_{i}\cdot W_{\mathbf{F}}^{(i)}(t)+\beta_{i}\cdot\tau_{\mathbf{F}}^{(i)}(t)$, where $\alpha_{i}$ and $\beta_{i}$ are positive constants quantifying the cost suffered by a class $i$ user for unit waiting time and delay in departure. ###### Definition 2.2 (Support of arrival profile). Given an arrival profile $t\mapsto B(t)$ such that $B(+\infty)<\infty$, the support of $B$, denoted by $\mathcal{S}(B)$, is defined as the smallest closed set having a $B$-measure equal to $B(+\infty)$. ###### Definition 2.3 (Equilibrium Arrival Profile (EAP)). The joint arrival profile $\mathbf{F^{\star}}=\\{F^{\star}_{1},F^{\star}_{2}\\}$ is an Equilibrium Arrival Profile of this game if for both the groups $i\in\\{1,2\\}$: $t\in\mathcal{S}(F_{i}^{\star})$ and $\tilde{t}\in\mathbb{R}$ implies $C_{\mathbf{F^{\star}}}^{(i)}(t)\leq C_{\mathbf{F^{\star}}}^{(i)}(\tilde{t})$. In particular, the arrival profile is iso-cost along its support. Note that, the EAP doesn’t change upon normalizing the cost function of both the classes $i=1,2$ by multiplying $1/(\alpha_{i}+\beta_{i})$. For simplicity, and without loss of generality, we assume both the classes $i=1,2$ have their normalized cost function, which are: $C_{\mathbf{F}}^{(i)}(t)=\gamma_{i}W_{\mathbf{F}}^{(i)}(t)+(1-\gamma_{i})\tau_{\mathbf{F}}^{(i)}(t)$ where $\gamma_{i}=\frac{\alpha_{i}}{\alpha_{i}+\beta_{i}}$ quantifies the preference of every class $i$ user. A value of $\gamma_{i}$ close to $1$ indicates users in group $i$ prefer late departure compared to waiting a long time in the network and $\gamma_{i}$ close to $0$ implies the opposite. So, we use $\gamma_{i}$ to quantify the cost preference of every group $i$ user. ###### Remark 1. _EAP captures the aggregate equilibrium behavior of the group. We can equivalently define Nash equilibrium at individual level where under it no individual has unilateral incentive to deviate. As is well known and discussed in more detail in[12], the two concepts are equivalent._ ###### Lemma 2.1. In every EAP, $\mathbf{F}=\\{F_{1},F_{2}\\}$ of the HDS and HAS, the arrival profiles $F_{1}$ and $F_{2}$ are absolutely continuous. Proof of the above lemma is similar to proof of statement (ii) of Lemma 1 in [12]. On assuming contradiction, if any of the arrival profiles have a jump, any user arriving in that jump will be strictly better off arriving slightly early and as a result the arrival profile cannot be an EAP. We argued before that Instances I and II in Table 1 are reducible to instances where one or more groups of users having distinct preferences are arriving to a single queue. [12] show that when two classes of customers having cost preferences $\gamma_{1}$ and $\gamma_{2}$ arrive at a single queue with service rate $\mu$, the EAP has a simple structure. The class with smaller $\gamma_{i}$ comes first at arrival rate $\mu\cdot\min\\{\gamma_{1},\gamma_{2}\\}$ over an interval, while the next class arrives at a contiguous but non-overlapping interval, at rate $\mu\cdot\max\\{\gamma_{1},\gamma_{2}\\}$. Fig 1 illustrates this EAP and the resulting queue length with the assumption $\gamma_{1}<\gamma_{2}$ and is useful to contrast with the various EAP structures that we find for HDS and HAS in Sections 3 and 4 below. The queue length process is constructed assuming that in the EAP, class 2 users start arriving from a positive time, which is equivalent to saying, masses of the two classes satisfy $\Lambda_{1}>\left(\frac{1}{\gamma_{2}}-1\right)\Lambda_{2}$. Figure 1: EAP structure (left) and resulting queue length process (right) when two classes of users with cost preferences $\gamma_{1}$ and $\gamma_{2}$ (assuming $\gamma_{1}<\gamma_{2}$) are arriving at a queue of capacity $\mu$. The support boundaries are $T_{2,f}=\frac{\Lambda_{1}+\Lambda_{2}}{\mu},~{}T_{2,a}=\frac{\Lambda_{1}}{\mu}-\left(\frac{1}{\gamma_{2}}-1\right)\frac{\Lambda_{2}}{\mu}$ and $T_{1,a}=-\left(\frac{1}{\gamma_{1}}-1\right)\frac{\Lambda_{1}}{\mu}-\left(\frac{1}{\gamma_{2}}-1\right)\frac{\Lambda_{2}}{\mu}$. The queue length process is illustrated only for the situation $T_{2,a}>0$, or equivalently $\Lambda_{1}>\left(\frac{1}{\gamma_{2}}-1\right)\Lambda_{2}$. Red and blue, respectively, represents class 1 and 2 populations. The black dashed line represents the total waiting mass of the two classes in the plot for queue length. ## 3 Heterogeneous Departure Systems (HDS) In this section, we consider the situation where the two classes arrive at the first queue and depart from different queues, as illustrated in Table 1. If $\mu_{1}\leq\mu_{2}$, class 2 users arrive at queue 2 at a maximum rate of $\mu_{1}$ and as a result, queue 2 remains empty and the cost of class 2 is unaffected by the second queue. Thus, if $\mu_{1}\leq\mu_{2}$, the instance becomes equivalent to both the groups arriving at a queue of capacity $\mu_{1}$. The problem is identical to the two-class, single queue case studied in [12]. Therefore, in subsequent discussion, we restrict ourselves to HDS with $\mu_{1}>\mu_{2}$. We further consider the case $\gamma_{1}\neq\gamma_{2}$ separately from $\gamma_{1}=\gamma_{2}$ since the latter displays different behaviour. ### 3.1 Unequal Preferences: $\gamma_{1}\neq\gamma_{2}$ Structural properties of EAP. We identify several structural properties that every EAP of HDS satisfies. We will exploit these properties later to narrow our search of an EAP. Many of these properties are true even when the two groups have equal preference, i.e., $\gamma_{1}=\gamma_{2}$, and we use them later in Section 3.2. As mentioned earlier, when $\mu_{1}\leq\mu_{2}$, the second queue is not relevant to equilibrium behaviour, and the two classes arrive in disjoint, contiguous intervals in the order of increasing $\gamma$’s. Lemma 3.1 shows that the EAP has a similar arrival pattern up to a threshold for $\mu_{1}>\mu_{2}$. ###### Lemma 3.1 (Threshold Behavior). If $\gamma_{1}\neq\gamma_{2}$, in the EAP, the two classes arrive over contiguous intervals with disjoint interiors if and only if $\mu_{1}\leq\mu_{2}\cdot\max\left\\{1,\frac{\gamma_{2}}{\gamma_{1}}\right\\}$. Below we sketch the proof of sufficiency of the condition stated in Lemma 3.1. Proving the other direction requires exploiting the behavior of the two queues in EAP and also the structure of $\mathcal{S}(F_{1})$ and $\mathcal{S}(F_{2})$. So, we prove that after stating Lemma 3.5. Proof sketch of sufficiency in Lemma 3.1: The detailed proof of sufficiency of the condition $\mu_{1}\leq\mu_{2}\cdot\max\left\\{1,\frac{\gamma_{2}}{\gamma_{1}}\right\\}$ is in Lemma A.3 and is similar to the proof sketch we will present here, but instead uses some other supplementary lemmas and tools. First we show via contradiction that if $\mu_{1}\leq\mu_{2}\cdot\max\left\\{1,\frac{\gamma_{2}}{\gamma_{1}}\right\\}$, $\mathcal{S}(F_{1})$ and $(\mathcal{S}(F_{2}))^{o}$ cannot overlap. We later argue via Lemma 3.2 (stated later) that in every EAP $\mathbf{F}=\\{F_{1},F_{2}\\}$, $\mathcal{S}(F_{1})$ and $\mathcal{S}(F_{2})$ are intervals. As a result, by the previous two statements, sufficiency of the stated condition will follow. If $\mu_{1}\leq\mu_{2}\cdot\max\left\\{1,\frac{\gamma_{2}}{\gamma_{1}}\right\\}$, and $\mathcal{S}(F_{1}),(\mathcal{S}(F_{2}))^{o}$ overlap, we can find $t_{1},t_{2}\in\mathcal{S}(F_{1})$ such that $[t_{1},t_{2}]\subseteq\mathcal{S}(F_{2})$. Note that queue 1 must be engaged in $(t_{1},t_{2})$, otherwise the class 1 user arriving at $t_{2}$ is strictly better off arriving at the time when queue 1 is empty in $(t_{1},t_{2})$. Hence using 3, $(C_{\mathbf{F}}^{(1)})^{\prime}(t)=\tau_{1}^{\prime}(t)-\gamma_{1}=\frac{F_{1}^{\prime}(t)+F_{2}^{\prime}(t)}{\mu_{1}}-\gamma_{1}$ in $[t_{1},t_{2}]$. Now $t_{1},t_{2}\in\mathcal{S}(F_{1})$ implies $C_{\mathbf{F}}^{(1)}(t_{2})=C_{\mathbf{F}}^{(1)}(t_{1})$, which by integrating $(C_{\mathbf{F}}^{(1)})^{\prime}(t)=\frac{F_{1}^{\prime}(t)+F_{2}^{\prime}(t)}{\mu_{1}}-\gamma_{1}$ in $[t_{1},t_{2}]$ gives $F_{1}(t_{2})+F_{2}(t_{2})-(F_{1}(t_{1})+F_{2}(t_{1}))=\mu_{1}\gamma_{1}(t_{2}-t_{1}).$ Therefore $\mu_{1}\gamma_{1}(t_{2}-t_{1})$ is the total mass of the two groups that have arrived in $[t_{1},t_{2}]$. Since $[t_{1},t_{2}]\subseteq\mathcal{S}(F_{2})$, we must have $(C_{\mathbf{F}}^{(2)})^{\prime}(t)=0$ and hence $(\tau_{\mathbf{F}}^{(2)})^{\prime}(t)=\gamma_{2}$ in $[t_{1},t_{2}]$. As $\tau_{\mathbf{F}}^{(2)}(t)=\tau_{2}(\tau_{1}(t))$, we have $(\tau_{\mathbf{F}}^{(2)})^{\prime}(t)=\tau_{2}^{\prime}(\tau_{1}(t))\tau_{1}^{\prime}(t)$ in $[t_{1},t_{2}]$, assuming the derivatives exist. Later we argue in Lemma 3.3 that, queue 2 must remain engaged at $\tau_{1}(t)$ for every $t\in[t_{1},t_{2}]$. Therefore, using (3), $\tau_{2}^{\prime}(\tau_{1}(t))=\frac{A_{2}^{\prime}(\tau_{1}(t))}{\mu_{2}}$, and hence $(\tau_{\mathbf{F}}^{(2)})^{\prime}(t)=\frac{A_{2}^{\prime}(\tau_{1}(t))\tau_{1}^{\prime}(t)}{\mu_{2}}$ in $[t_{1},t_{2}]$. Since $A_{2}(\tau_{1}(t))=F_{2}(t)$, the previous statement implies $(\tau_{\mathbf{F}}^{(2)})^{\prime}(t)=\frac{F_{2}^{\prime}(t)}{\mu_{2}}$ in $[t_{1},t_{2}]$. Combining this with the observation that $(\tau_{\mathbf{F}}^{(2)})^{\prime}(t)=\gamma_{2}$ a.e. in $[t_{1},t_{2}]$, we have $F_{2}^{\prime}(t)=\mu_{2}\gamma_{2}$ a.e. in $[t_{1},t_{2}]$. Therefore, $F_{2}(t_{2})-F_{2}(t_{1})=\mu_{2}\gamma_{2}(t_{2}-t_{1}).$ Since a positive mass of both the classes arrive in $[t_{1},t_{2}]$, we must have $F_{1}(t_{2})+F_{2}(t_{2})-(F_{1}(t_{1})+F_{2}(t_{1}))>F_{2}(t_{2})-F_{2}(t_{1})$, which implies $\mu_{1}\gamma_{1}>\mu_{2}\gamma_{2}$ by the previous arguments. This contradicts our assumption that $\mu_{1}\leq\mu_{2}\max\left\\{1,\frac{\gamma_{2}}{\gamma_{1}}\right\\}$. ∎ Lemma 3.2 and 3.4 imply EAPs can only be piece-wise linear arrival profiles with intervals as support. Proof of Lemma 3.2 (in A.1) is done via contradiction by arguing if there is a gap, we can identify a user who can improve her cost by arriving at a different time. ###### Lemma 3.2 (Structure of supports). If $\mu_{1}>\mu_{2}$ and $\mathbf{F}=\\{F_{1},F_{2}\\}$ is an EAP, then $\mathcal{S}(F_{1})$ and $\mathcal{S}(F_{1})\cup\mathcal{S}(F_{2})$ must be intervals. Additionally, if $\gamma_{1}\neq\gamma_{2}$, $\mathcal{S}(F_{2})$ must also be an interval. For a joint arrival profile $\mathbf{F}=\\{F_{1},F_{2}\\}$, we define the quantities $T_{i,a}\overset{def.}{=}\inf\mathcal{S}(F_{i})$ and $T_{i,f}\overset{def.}{=}\sup\mathcal{S}(F_{i})$ for the two classes $i\in\\{1,2\\}$. For every EAP $\mathbf{F}$, $\mathcal{S}(F_{1}),~{}\mathcal{S}(F_{2})$ must be compact, as cost of the two classes over their support must be finite. As a result, the support boundaries $T_{1,a},T_{1,f},T_{2,a},T_{2,f}$ must be finite. By Lemma 3.2, we can further say, $\mathcal{S}(F_{1})=[T_{1,a},T_{1,f}]$ and $\mathcal{S}(F_{2})=[T_{2,a},T_{2,f}]$, such that union of the two intervals is also an interval. Lemma 3.3 is about the behavior of queue 2 in equilibrium. ###### Lemma 3.3. If $\mu_{1}>\mu_{2}$ and $\gamma_{1}\neq\gamma_{2}$, then under EAP, for every $t\in(T_{2,a},T_{2,f})$, $\tau_{1}(t)$ belongs to the closure of the set $\\{s~{}|~{}Q_{2}(s)>0\\}$. Proof sketch of Lemma 3.3: We consider two separate cases: 1. 1. If $t\in(T_{2,a},T_{2,f})/(T_{1,a},T_{1,f})$, two possibilities are there. If queue 1 is engaged at $t$, class 2 users arrive from queue 1 to 2 at rate $\mu_{1}>\mu_{2}$ at $\tau_{1}(t)$, making queue 2 engaged. Otherwise, if queue 1 is empty at $t$, in EAP, the network cannot be empty at $t$ with a positive mass of class 2 users arriving in $(t,T_{2,f})$. As a result, queue 2 must be engaged at $t=\tau_{1}(t)$. 2. 2. If $t\in(T_{1,a},T_{1,f})\cap(T_{2,a},T_{2,f})$, we assume a contradiction, i.e., queue 2 is empty in some neighbourhood of $\tau_{1}(t)$. As a result, by continuity of $\tau_{1}(\cdot)$, only queue 1 will be serving in some neighbourhood of $t$. Therefore $\tau_{\mathbf{F}}^{(1)}(\cdot)=\tau_{\mathbf{F}}^{(2)}(\cdot)=\tau_{1}(\cdot)$ in that neighbourhood. Also, for both the classes $i=1,2$ to be iso-cost in that neighbourhood, we need $(C_{\mathbf{F}}^{(i)})^{\prime}(s)=\tau_{1}^{\prime}(s)-\gamma_{i}=0$. This gives us $\tau_{1}^{\prime}(s)=\gamma_{1}=\gamma_{2}$, contradicting $\gamma_{1}\neq\gamma_{2}$. ∎ Lemma 3.4 states conditions on the arrival rates necessary for the two classes to have constant cost over their support in any EAP. ###### Lemma 3.4 (Rates of arrival). If $\mu_{1}>\mu_{2}$, $\gamma_{1}\neq\gamma_{2}$, and $\mathbf{F}=\\{F_{1},F_{2}\\}$ is an EAP, the following properties must be true almost everywhere: $\displaystyle F_{1}^{\prime}(t)$ $\displaystyle=\begin{cases}\mu_{1}\gamma_{1}~{}~{}&\text{if}~{}t\in\mathcal{S}(F_{1})/\mathcal{S}(F_{2}),\\\ \mu_{1}\gamma_{1}-\mu_{2}\gamma_{2}~{}~{}&\text{if}~{}t\in\mathcal{S}(F_{1})\cap\mathcal{S}(F_{2}),\end{cases}~{}~{}\text{and,}~{}F_{2}^{\prime}(t)=\mu_{2}\gamma_{2}~{}~{}\text{if}~{}t\in\mathcal{S}(F_{2}),$ where from Lemma 3.1, $\mathcal{S}(F_{1})\cap\mathcal{S}(F_{2})$ has zero measure if $\mu_{1}\gamma_{1}\leq\mu_{2}\gamma_{2}$. In the proof of Lemma 3.4 (in A.1), we first observe that both the classes $i=1,2$ have $C_{\mathbf{F}}^{(i)}(t)=\tau_{\mathbf{F}}^{(i)}(t)-\gamma_{i}t$ constant in $[T_{i,a},T_{1,f}]$, causing $(\tau_{\mathbf{F}}^{(i)})^{\prime}(t)=\gamma_{i}$. The rest of the proof relies on relating $(\tau_{\mathbf{F}}^{(i)})^{\prime}(t)$ with the arrival rates $F_{1}^{\prime}(t)$ and $F_{2}^{\prime}(t)$ of the two classes. Towards that, we use (3) and leverage the facts that queue 1 has positive waiting time in $(T_{1,a},T_{1,f})$, (otherwise, if $Q_{1}(t)=0$ at some $t\in(T_{1,a},T_{1,f})$, every class 1 user arriving in $(t,T_{1,f}]$ is strictly better off arriving at time $t$) and queue 2 stays engaged in $[\tau_{1}(T_{2,a}),\tau_{1}(T_{2,f})]$ (by Lemma 3.3). Lemma 3.5 is about the state of the two queues at support boundaries $T_{1,f}$ and $T_{2,f}$. Proof of Lemma 3.5 is done via a contradiction by showing that if the specified properties do not hold, a user can reduce her cost by arriving at a different time. ###### Lemma 3.5. If $\mu_{1}>\mu_{2}$ and $\gamma_{1}\neq\gamma_{2}$, then under EAP, queue length at the second queue is zero at $\tau_{1}(T_{2,f})$. If, in addition, we have $\mu_{1}\gamma_{1}>\mu_{2}\gamma_{2}$, queue length at the first queue equals zero at $T_{1,f}$. Proof of necessity of the condition in Lemma 3.1: We prove via contradiction that $\mu_{1}\leq\mu_{2}\cdot\max\left\\{1,\frac{\gamma_{2}}{\gamma_{1}}\right\\}$ is necessary for the intervals $[T_{1,a},T_{1,f}]$ and $[T_{2,a},T_{2,f}]$ to overlap. This forms the other direction of Lemma 3.1. On assuming a contradiction, we must have $\mu_{1}>\mu_{2}\cdot\max\left\\{1,\frac{\gamma_{2}}{\gamma_{1}}\right\\}$ and interiors of $[T_{1,a},T_{1,f}]$, $[T_{2,a},T_{2,f}]$ are disjoint. Now by Lemma 3.2, union of $[T_{1,a},T_{1,f}]$ and $[T_{2,a},T_{2,f}]$ must be an interval. This leaves us two possibilities: 1. 1. If $T_{1,f}=T_{2,a}$, by Lemma 3.5, queue 1 will be empty at $T_{1,f}$, making the whole network empty at $T_{1,f}$. As a result, every class 2 user will be strictly better off arriving at $T_{1,f}$. 2. 2. If $T_{2,f}=T_{1,a}$, queue 1 must have a positive waiting time in $(T_{2,a},T_{2,f}]$ and as a result, class 2 users arrive at queue 2 at rate $\mu_{1}>\mu_{2}$ in $[0,\tau_{1}(T_{2,f})]$, causing $Q_{2}(\tau_{1}(T_{2,f}))=(\mu_{1}-\mu_{2})\cdot\tau_{1}(T_{2,f})>0$, contradicting Lemma 3.5. ∎ Specification of the EAP. Theorem 3.1 specifies the unique EAP of this regime and we mention below the support boundaries of the unique EAP, which we will refer to later in Theorem 3.1. 1. 1. If $\gamma_{1}\leq\frac{\mu_{2}}{\mu_{1}}\gamma_{2}$, then $\displaystyle T_{1,a}$ $\displaystyle=-\left(\frac{1}{\gamma_{1}}-1\right)\frac{\Lambda_{1}}{\mu_{1}}-\left(\frac{1}{\gamma_{2}}-1\right)\frac{\Lambda_{2}}{\mu_{2}},~{}T_{1,f}=T_{2,a}=\frac{\Lambda_{1}}{\mu_{1}}-\left(\frac{1}{\gamma_{2}}-1\right)\frac{\Lambda_{2}}{\mu_{2}},~{}\text{and}~{}T_{2,f}=\frac{\Lambda_{1}}{\mu_{1}}+\frac{\Lambda_{2}}{\mu_{2}}.$ (4) 2. 2. If $\gamma_{1}>\frac{\mu_{2}}{\mu_{1}}\gamma_{2}$ and $\Lambda_{1}\geq\min\left\\{\frac{1-\gamma_{2}}{1-\gamma_{1}}\left(\frac{\mu_{1}\gamma_{1}}{\mu_{2}\gamma_{2}}-1\right),\frac{\mu_{1}}{\mu_{2}}-1\right\\}\cdot\Lambda_{2}$, then 1. a) if $\gamma_{1}<\gamma_{2}$, then $\displaystyle\vspace{-0.05in}T_{1,a}$ $\displaystyle=-\frac{1-\gamma_{1}}{\mu_{1}\gamma_{1}}\left[\Lambda_{1}+\frac{1-\gamma_{2}}{1-\gamma_{1}}\Lambda_{2}\right],~{}T_{1,f}=\frac{1}{\mu_{1}}\left[\Lambda_{1}+\frac{1-\gamma_{2}}{1-\gamma_{1}}\Lambda_{2}\right],$ $\displaystyle T_{2,a}$ $\displaystyle=\frac{1}{\mu_{1}}\left[\Lambda_{1}-\frac{\mu_{1}-\mu_{2}\gamma_{2}}{\mu_{2}\gamma_{2}}\frac{1-\gamma_{2}}{1-\gamma_{1}}\Lambda_{2}\right],~{}\text{and}~{}T_{2,f}=\frac{1}{\mu_{1}}\left[\Lambda_{1}+\frac{\mu_{1}(\gamma_{2}-\gamma_{1})+\mu_{2}\gamma_{2}(1-\gamma_{2})}{\mu_{2}\gamma_{2}(1-\gamma_{1})}\Lambda_{2}\right].$ (5) 2. b) if $\gamma_{1}>\gamma_{2}$, then $\displaystyle\vspace{-0.05in}T_{1,a}$ $\displaystyle=\frac{\gamma_{1}-\gamma_{2}}{\gamma_{1}}\frac{\Lambda_{2}}{\mu_{1}\gamma_{1}-\mu_{2}\gamma_{2}}-\frac{1-\gamma_{1}}{\gamma_{1}}\frac{\Lambda_{1}+\Lambda_{2}}{\mu_{1}},~{}T_{1,f}=\frac{\Lambda_{1}+\Lambda_{2}}{\mu_{1}},$ $\displaystyle T_{2,a}$ $\displaystyle=-\frac{1-\gamma_{1}}{\gamma_{1}}\frac{\Lambda_{1}}{\mu_{1}}-\left(\frac{\gamma_{1}}{\gamma_{2}}+(1-\gamma_{1})\frac{\mu_{2}}{\mu_{1}}-1\right)\frac{\Lambda_{2}}{\mu_{2}\gamma_{1}},~{}\text{and}~{}T_{2,f}=-\frac{1-\gamma_{1}}{\gamma_{1}}\frac{\Lambda_{1}}{\mu_{1}}+\left(1-(1-\gamma_{1})\frac{\mu_{2}}{\mu_{1}}\right)\frac{\Lambda_{2}}{\mu_{2}\gamma_{1}}.$ (6) 3. 3. If $\gamma_{1}>\frac{\mu_{2}}{\mu_{1}}\gamma_{2}$ and $\Lambda_{1}<\min\left\\{\frac{1-\gamma_{2}}{1-\gamma_{1}}\left(\frac{\mu_{1}\gamma_{1}}{\mu_{2}\gamma_{2}}-1\right),\frac{\mu_{1}}{\mu_{2}}-1\right\\}\cdot\Lambda_{2}$, then $\displaystyle~{}T_{1,a}$ $\displaystyle=\frac{1-\gamma_{2}}{\mu_{1}-\mu_{2}\gamma_{2}}\left[\Lambda_{2}-\frac{(1-\gamma_{1})\mu_{1}}{(1-\gamma_{2})(\mu_{1}\gamma_{1}-\mu_{2}\gamma_{2})}\Lambda_{1}\right],~{}T_{1,f}=\frac{\Lambda_{1}+(1-\gamma_{2})\Lambda_{2}}{\mu_{1}-\mu_{2}\gamma_{2}}$ $\displaystyle T_{2,a}$ $\displaystyle=-\frac{1-\gamma_{2}}{\gamma_{2}}\frac{\Lambda_{2}}{\mu_{2}},~{}\text{and}~{}T_{2,f}=\frac{\Lambda_{2}}{\mu_{2}}.$ (7) ###### Theorem 3.1. If $\mu_{1}>\mu_{2}$ and $\gamma_{1}\neq\gamma_{2}$, HDS has a unique EAP. The arrival rates in the EAP are given below along with the support boundaries: 1. 1. If $\gamma_{1}\leq\frac{\mu_{2}}{\mu_{1}}\gamma_{2}$, $F_{1}^{\prime}(t)=\mu_{1}\gamma_{1}$ for $t\in[T_{1,a},T_{1,f}]$ and $F_{2}^{\prime}(t)=\mu_{2}\gamma_{2}$ for $t\in[T_{2,a},T_{2,f}]$ where $T_{1,a},T_{1,f},T_{2,a}$ and $T_{2,f}$ are given in (4). 2. 2. If $\frac{\mu_{2}}{\mu_{1}}\gamma_{2}<\gamma_{1}<\gamma_{2}$, 1. 2a. when $\Lambda_{1}\geq\frac{1-\gamma_{2}}{1-\gamma_{1}}\left(\frac{\mu_{1}\gamma_{1}}{\mu_{2}\gamma_{2}}-1\right)\Lambda_{2}$, $\displaystyle F_{1}^{\prime}(t)$ $\displaystyle=\begin{cases}\mu_{1}\gamma_{1}~{}~{}&\text{if}~{}t\in[T_{1,a},T_{2,a}],\\\ \mu_{1}\gamma_{1}-\mu_{2}\gamma_{2}~{}~{}&\text{if}~{}t\in[T_{2,a},T_{1,f}],\end{cases}~{}~{}\text{and}~{}F_{2}^{\prime}(t)=\mu_{2}\gamma_{2}~{}~{}\text{if}~{}t\in[T_{2,a},T_{2,f}]$ where $T_{1,a},T_{1,f},T_{2,a}$ and $T_{2,f}$ are given in ( ‣ 2.). 2. 2b. when $\Lambda_{1}<\frac{1-\gamma_{2}}{1-\gamma_{1}}\left(\frac{\mu_{1}\gamma_{1}}{\mu_{2}\gamma_{2}}-1\right)\Lambda_{2}$, $F_{1}^{\prime}(t)=\mu_{1}\gamma_{1}-\mu_{2}\gamma_{2}$ for $t\in[T_{1,a},T_{1,f}]$, and $F_{2}^{\prime}(t)=\mu_{2}\gamma_{2}$ for $t\in[T_{2,a},T_{2,f}]$, where $T_{1,a},T_{1,f},T_{2,a}$ and $T_{2,f}$ are given in (3.). 3. 3. If $\gamma_{2}<\gamma_{1}$, 1. 3a. when $\Lambda_{1}\geq\left(\frac{\mu_{1}}{\mu_{2}}-1\right)\Lambda_{2}$, $\displaystyle F_{1}^{\prime}(t)$ $\displaystyle=\begin{cases}\mu_{1}\gamma_{1}-\mu_{2}\gamma_{2}~{}~{}&\text{if}~{}t\in[T_{1,a},T_{2,f}],\\\ \mu_{1}\gamma_{1}~{}~{}&\text{if}~{}t\in[T_{2,f},T_{1,f}],\end{cases}~{}~{}\text{and}~{}F_{2}^{\prime}(t)=\mu_{2}\gamma_{2}~{}~{}\text{if}~{}t\in[T_{2,a},T_{2,f}]$ where $T_{1,a},T_{1,f},T_{2,a}$ and $T_{2,f}$ are given in ( ‣ 2.). 2. 3b. when $\Lambda_{1}<\left(\frac{\mu_{1}}{\mu_{2}}-1\right)\Lambda_{2}$, the EAP has a closed form same as case 2b. ###### Remark 2. _Figures 2, 3, 4, and 5 show the illustrative EAPs and resulting queue length processes, respectively, for cases 1, 2a, 2b and 3a of Theorem 3.1. EAP structure in case 3b of Theorem 3.1 is similar to case 2b and is illustrated in Figure 4. The structure of the queue length processes in certain regimes depend on the support boundaries and may vary accordingly. In the figures referred above, we have illustrated only one possible structure of the queue length process. Moreover, we have mentioned the conditions to be satisfied by the support boundaries for the attainment of that structure in the caption. In all these figures, red and blue, respectively, represent class 1 and 2 populations, and the black dashed line represents the total mass of the two populations waiting in queue 1._ ###### Remark 3. _The illustrative EAPs referred to in the previous remark also displays the threshold behavior stated in Lemma 3.1. Since $\mu_{1}\leq\mu_{2}\cdot\max\left\\{1,\frac{\gamma_{2}}{\gamma_{1}}\right\\}$ in case 1 of Theorem 3.1, we can see the support intervals of the two classes to have disjoint interiors in Figure 2. On the other hand, since cases 2 and 3 of Theorem 3.1 have $\mu_{1}>\mu_{2}\cdot\left\\{1,\frac{\gamma_{2}}{\gamma_{1}}\right\\}$, two classes arrive over overlapping intervals, as can be seen in Figure 3, 4, and 5._ The following lemma specifies the arrival order of the two classes in EAP and is necessary for proving Theorem 3.1. ###### Lemma 3.6. If $\mu_{1}>\mu_{2}$ and $\gamma_{1}\neq\gamma_{2}$, the support boundaries of EAP of HDS satisfy: 1. 1. $T_{1,f}=T_{2,a}$ if $\gamma_{1}\leq\frac{\mu_{2}}{\mu_{1}}\gamma_{2}$, 2. 2. $T_{1,f}\leq T_{2,f}$ if $\frac{\mu_{2}}{\mu_{1}}\gamma_{2}<\gamma_{1}<\gamma_{2}$, and 3. 3. $T_{1,a}>T_{2,a}$ if $\gamma_{1}>\gamma_{2}$. Proof of Lemma 3.6 (in A.1) is done via contradiction by showing that if the stated property doesn’t hold, a user can improve her cost by arriving at a different time. ###### Remark 4. _The arrival orders anticipated by Lemma 3.6 can be observed in the illustrative EAPs referred earlier. By Lemma 3.6, in case 2 ($\gamma_{2}>\gamma_{1}>\frac{\mu_{2}}{\mu_{1}}\gamma_{2}$) of Theorem 3.1, class 2 population finishes arrival after class 1, as can be observed in Figure 3 and 4. Similarly for case 3 ($\gamma_{1}>\gamma_{2}$) of Theorem 3.1, class 1 population starts arrival after class 2, as can be observed in Figure 5 and 4. The class arriving first in case 2 and finishing late in case 3 is the one having a population size significantly larger among the two classes. In case 2 (or case 3): 1) when $\Lambda_{1}>c\Lambda_{2}$, class 1 starts arriving before class 2 (or finishes arriving after class 2), 2) when $\Lambda_{1}=c\Lambda_{2}$, both classes start arriving from the same time (or finishes arriving at the same time), 3) when $\Lambda_{1}<c\Lambda_{2}$, class 1 starts arriving after class 2 starts (or finishes arriving before class 2), where $c=\min\left\\{\frac{1-\gamma_{2}}{1-\gamma_{1}}\left(\frac{\mu_{1}\gamma_{1}}{\mu_{2}\gamma_{2}}-1\right),\frac{\mu_{1}}{\mu_{2}}-1\right\\}$ for both case 2 and 3. _ Key steps in the proof of Theorem 3.1 (details in A.1.1): By Lemma 3.2 we only consider candidates $\mathbf{F}=\\{F_{1},F_{2}\\}$ which are absolutely continuous with supports of the form $\mathcal{S}(F_{1})=[T_{1,a},T_{1,f}]$, $\mathcal{S}(F_{2})=[T_{2,a},T_{2,f}]$, such that their union is an interval and the arrival rates $F_{1}^{\prime}(\cdot),~{}F_{2}^{\prime}(\cdot)$ satisfy the property in Lemma 3.4. We eliminate candidates not satisfying structural properties desired of an EAP and will be eventually left with only one candidate, which is the only candidate that can qualify as an EAP. We then prove that the only remaining candidate is indeed an EAP. Thus, existence and uniqueness of EAP follows. We follow this agenda for the three cases: 1) $\gamma_{1}\leq\frac{\mu_{2}}{\mu_{1}}\gamma_{2}$, 2) $\frac{\mu_{2}}{\mu_{1}}\gamma_{2}<\gamma_{1}<\gamma_{2}$, and 3) $\gamma_{2}<\gamma_{1}$. The final candidate we get in these cases have their arrival rates and support boundaries same as the joint arrival profiles mentioned in Theorem 3.1 under the respective cases. Case 1 $\mathbf{\gamma_{1}\leq\frac{\mu_{2}}{\mu_{1}}\gamma_{2}}$: By Lemma 3.1, every EAP must have $[T_{1,a},T_{1,f}]$ and $[T_{2,a},T_{2,f}]$ disjoint. Lemma 3.7 helps us substantially narrow our search for an EAP. ###### Lemma 3.7. Under case 1 $\gamma_{1}\leq\frac{\mu_{2}}{\mu_{1}}\gamma_{2}$, every EAP has (4) as support boundaries $T_{1,a},T_{1,f},T_{2,a},T_{2,f}$. Proof Sketch: By Lemma 3.6, we must have $T_{1,f}=T_{2,a}$. By Lemma 3.4, $F_{1}^{\prime}(t)=\mu_{1}\gamma_{1}$ in $[T_{1,a},T_{1,f}]$ and $F_{2}^{\prime}(t)=\mu_{2}\gamma_{2}$ in $[T_{1,f},T_{2,f}]$. Therefore, $T_{1,a},T_{1,f},T_{2,f}$ must satisfy: $\displaystyle T_{1,f}$ $\displaystyle=T_{1,a}+\frac{\Lambda_{1}}{\mu_{1}\gamma_{1}}~{}\text{and}~{}T_{2,f}=T_{1,f}+\frac{\Lambda_{2}}{\mu_{2}\gamma_{2}}.$ (8) Now, queue 2 must be empty at $\tau_{1}(T_{2,f})$ (by Lemma 3.5) and must stay engaged in the image of $[T_{1,f},T_{2,f}]$ under $\tau_{1}(\cdot)$ (by Lemma 3.3), which is $[\tau_{1}(T_{1,f}),\tau_{1}(T_{2,f})]$. Since, the first and last class 2 users arrive at queue 2, respectively at times $\tau_{1}(T_{1,f})$ and $\tau_{1}(T_{2,f})$, the previous statement implies, $\displaystyle\mu_{2}\cdot(\tau_{1}(T_{2,f})-\tau_{1}(T_{1,f}))$ $\displaystyle=\Lambda_{2}$ (9) For queue 2 to be empty at $\tau_{1}(T_{2,f})$, queue 1 must also be empty at $T_{2,f}$. Otherwise if queue 1 has a positive waiting time at $T_{2,f}$, class 2 users will be arriving at queue 2 from queue 1 at rate $\mu_{1}>\mu_{2}$ in $[\tau_{1}(T_{2,f}-\delta),\tau_{1}(T_{2,f})]$, where $\delta>0$ picked sufficiently small such that queue 1 has positive waiting time in $[T_{2,f}-\delta,T_{2,f}]$. By (3), $\tau_{1}^{\prime}(t)=\frac{F_{2}^{\prime}(t)}{\mu_{1}}=\frac{\mu_{2}\gamma_{2}}{\mu_{1}}>0$ in $[T_{2,f}-\delta,T_{2,f}]$, giving us $\tau_{1}(T_{2,f})>\tau_{1}(T_{2,f}-\delta)$. As a result, $Q_{2}(\tau_{1}(T_{2,f}))\geq(\mu_{1}-\mu_{2})\cdot(\tau_{1}(T_{2,f})-\tau_{1}(T_{2,f}-\delta))>0$, which contradicts Lemma 3.5. With queue 1 empty at $T_{2,f}$, we have $\tau_{1}(T_{2,f})=T_{2,f}$. Note that queue 1 must have a positive waiting time in $(T_{1,a},T_{1,f}]$, otherwise, every class 2 user arriving after $T_{1,f}$ is strictly better off arriving at the time queue 1 is empty. Applying (3), we have $\tau_{1}(T_{1,f})-\tau_{1}(T_{1,a})=\frac{F_{1}(T_{1,f})-F_{1}(T_{1,a})}{\mu_{1}}=\frac{\Lambda_{1}}{\mu_{1}}$. Since the network cannot be empty at time zero, we have $T_{1,a}<0$. This gives us $\tau_{1}(T_{1,a})=0$ and hence, $\tau_{1}(T_{1,f})=\frac{\Lambda_{1}}{\mu_{1}}$. Putting $\tau_{1}(T_{2,f})=T_{2,f}$ and $\tau_{1}(T_{1,f})=\frac{\Lambda_{1}}{\mu_{1}}$ in (9), we get $T_{2,f}=\frac{\Lambda_{1}}{\mu_{1}}+\frac{\Lambda_{2}}{\mu_{2}}$. Using this in (8) and by the fact $T_{2,a}=T_{1,f}$, we get the values of $T_{1,a},T_{1,f},T_{2,a},T_{2,f}$ mentioned in (4). ∎ The only candidate having arrival rates satisfying Lemma 3.4 and support boundaries from Lemma 3.7 is the joint arrival profile under case 1 of Theorem 3.1. Proving that this unique remaining candidate is an EAP requires analyzing the behavior of the two queues induced by this candidate.s Details of it are in A.1. Figure 2: Illustrative EAP (left) and resulting queue length process (right) of HDS with $\mu_{1}>\mu_{2}$ under case 1: $\gamma_{1}\leq\frac{\mu_{2}}{\mu_{1}}\cdot\gamma_{2}$ and the condition $T_{1,f}>0$ in (4). $T\in(T_{1,f},T_{2,f})$ is the time at which queue 1 empties after time zero. In queue 1, since class 1 users stop arriving after $T_{1,f}$, class 1 waiting mass decreases faster in $[T_{1,f},\tau_{1}(T_{1,f})]$ than in $[0,T_{1,f}]$. Case 2 $\frac{\mu_{2}}{\mu_{1}}\cdot\gamma_{2}<\gamma_{1}<\gamma_{2}$: By Lemma 3.1, $[T_{1,a},T_{1,f}]$ and $[T_{2,a},T_{2,f}]$ must overlap, and by Lemma 3.6, class 2 shouldn’t finish arriving before class 1, i.e., $T_{1,f}\leq T_{2,f}$. Therefore, every EAP must have its support boundaries $T_{1,a},T_{1,f},T_{2,a},T_{2,f}$ ordered by one of the following two ways: 1. Type I: $T_{1,a}\leq T_{2,a}<T_{1,f}\leq T_{2,f}$ with $T_{1,a}<0$, and, 2. Type II: $T_{2,a}<T_{1,a}<T_{1,f}\leq T_{2,f}$ with $T_{2,a}<0$. Note that, we need $\min\\{T_{1,a},T_{2,a}\\}<0$, otherwise, any user will be strictly better off arriving at time zero when the network is empty. The following lemma now specifies the necessary and sufficient conditions for existence of an EAP under these two types. Existence of unique EAP in case 2 of Theorem 3.1 follows trivially from this lemma. ###### Lemma 3.8. If $\gamma_{2}>\gamma_{1}>\frac{\mu_{2}}{\mu_{1}}\gamma_{2}$, the following statements are true: * • There exists an EAP under Type I if and only if $\Lambda_{1}\geq\frac{1-\gamma_{2}}{1-\gamma_{1}}\left(\frac{\mu_{1}\gamma_{1}}{\mu_{2}\gamma_{2}}-1\right)\Lambda_{2}$ and if it exists, it will be unique with a closed form same as the joint arrival profile under case 2a of Theorem 3.1. * • There exists an EAP under Type II if and only if $\Lambda_{1}<\frac{1-\gamma_{2}}{1-\gamma_{1}}\left(\frac{\mu_{1}\gamma_{1}}{\mu_{2}\gamma_{2}}-1\right)\Lambda_{2}$ and if it exists, it will be unique with a closed form same as the joint arrival profile under case 2b of Theorem 3.1. Proof Sketch of Lemma 3.8: Identifying the support boundaries: For every EAP under Type I, we identify the following system of equations to be satisfied by its support boundaries $T_{1,a},T_{1,f},T_{2,a},T_{2,f}$: 1) By Lemma 3.4, $F_{2}^{\prime}(t)=\mu_{2}\gamma_{2}$ in $[T_{2,a},T_{2,f}]$, giving us: $T_{2,f}=T_{2,a}+\frac{\Lambda_{2}}{\mu_{2}\gamma_{2}}$. 2) Since $T_{1,a}<0$, queue 1 starts serving at time zero and must have positive waiting time in $(0,T_{1,f})$. By Lemma 3.5, queue 1 must be empty at time $T_{1,f}$. Therefore, $F_{1}(T_{1,f})+F_{2}(T_{1,f})=\mu_{1}T_{1,f}$. Now $F_{1}(T_{1,f})=\Lambda_{1}$ and $F_{2}(T_{1,f})=\mu_{2}\gamma_{2}(T_{1,f}-T_{2,a})$ (by Lemma 3.4). Plugging this in, we get: $\Lambda_{1}+\mu_{2}\gamma_{2}(T_{1,f}-T_{2,a})=\mu_{1}T_{1,f}$. 3) By definition of EAP, $C_{\mathbf{F}}^{(1)}(T_{1,a})=C_{\mathbf{F}}^{(1)}(T_{1,f})$. Queue 1 empties at $T_{1,f}$ (by Lemma 3.5), giving us $C_{\mathbf{F}}^{(1)}(T_{1,f})=(1-\gamma_{1})T_{1,f}$. The first class 1 user arriving at time $T_{1,a}$ gets served by queue 1 at time zero and waits for $-T_{1,a}$ time, giving us $C_{\mathbf{F}}^{(1)}(T_{1,a})=-\gamma_{1}T_{1,a}$. Equating these two costs, we get: $T_{1,a}=-\left(\frac{1}{\gamma_{1}}-1\right)T_{1,f}$. 4) By Lemma 3.3, queue 2 stays engaged in $[\tau_{1}(T_{2,a}),\tau_{1}(T_{2,f})]$ and by Lemma 3.5, queue 2 empties at $\tau_{1}(T_{2,f})$. In this way, queue 2 serves the class 2 population in $[\tau_{1}(T_{2,a}),\tau_{1}(T_{2,f})]$, giving us $\mu_{2}\cdot(\tau_{1}(T_{2,f})-\tau_{1}(T_{2,a}))=\Lambda_{2}$. Queue 1 empties at $T_{1,f}$ (by Lemma 3.5) and after $T_{1,f}$, since class 2 users arrive at rate $\mu_{2}\gamma_{2}<\mu_{1}\gamma_{1}<\mu_{1}$, queue 1 remains empty at $T_{2,f}$, giving us $\tau_{1}(T_{2,f})=T_{2,f}$. Since queue 1 has positive waiting time in $(T_{1,a},T_{2,a}]$, we have $\tau_{1}(T_{2,a})-\tau_{1}(T_{1,a})=\frac{F_{1}(T_{2,a})-F_{1}(T_{1,a})}{\mu_{1}}=\gamma_{1}(T_{2,a}-T_{1,a})$ (by (3) and Lemma 3.4), which upon putting $\tau_{1}(T_{1,a})=0$ gives us $\tau_{1}(T_{2,a})=\gamma_{1}(T_{2,a}-T_{1,a})$. Placing the expressions obtained for $\tau_{1}(T_{2,a})$ and $\tau_{1}(T_{2,f})$ into $\mu_{2}(\tau_{1}(T_{2,f})-\tau_{1}(T_{2,a}))=\Lambda_{2}$, we get, $\mu_{2}(T_{2,f}-\gamma_{1}(T_{2,a}-T_{1,a}))=\Lambda_{2}$. Getting the necessary condition: Solution of the identified system of equations is ( ‣ 2.) and therefore, every EAP under Type I must have ( ‣ 2.) as support boundaries. Now ( ‣ 2.) must be ordered by $T_{1,a}\leq T_{2,a}<T_{1,f}\leq T_{2,f}$ to represent an EAP of Type I. Note that ( ‣ 2.) satisfies $T_{2,a}-T_{1,a}=\frac{1}{\mu_{1}\gamma_{1}}\left[\Lambda_{1}-\frac{1-\gamma_{2}}{1-\gamma_{1}}\left(\frac{\mu_{1}\gamma_{1}}{\mu_{2}\gamma_{2}}-1\right)\Lambda_{2}\right]$. As a result, for $T_{2,a}\geq T_{1,a}$, we need $\Lambda_{1}\geq\frac{1-\gamma_{2}}{1-\gamma_{1}}\left(\frac{\mu_{1}\gamma_{1}}{\mu_{2}\gamma_{2}}-1\right)\Lambda_{2}$ and therefore this is a necessary condition for existence of an EAP under Type I. Proving sufficiency of the obtained necessary condition: Now, if $\Lambda_{1}\geq\frac{1-\gamma_{2}}{1-\gamma_{1}}\left(\frac{\mu_{1}\gamma_{1}}{\mu_{2}\gamma_{2}}-1\right)\Lambda_{2}$, it is easy to verify that ( ‣ 2.) satisfies $T_{1,a}\leq T_{2,a}<T_{1,f}\leq T_{2,f}$. Therefore, with the necessary condition satisfied and upon plugging in arrival rates using Lemma 3.4, we get a candidate having arrival rates and support boundaries same as the joint arrival profile mentioned under case 2a of Theorem 3.1. This candidate satisfies $F_{1}(T_{1,f})=\Lambda_{1},~{}F_{2}(T_{2,f})=\Lambda_{2}$ and will be the only candidate under Type I to qualify as an EAP. Proving that this candidate is an EAP requires analyzing the behavior of the two queues. Details of this argument is in A.1. Therefore, it follows that the obtained necessary condition is also sufficient for existence of an EAP under Type I, and once it is satisfied, there is a unique EAP under Type I which has a closed form same as the one mentioned under case 2a of Theorem 3.1. Thus the first statement of the lemma follows. The second statement is proved via an argument similar to the one used for proving the first statement via identifying a system of equations satisfied by the support boundaries and getting the necessary condition by imposing $T_{2,a}<T_{1,a}$ on the solution to that system. Details of the argument in A.1.1. ∎ Case 3: $\gamma_{1}>\gamma_{2}$: By Lemma 3.1, $[T_{1,a},T_{1,f}]$ and $[T_{2,a},T_{2,f}]$ must overlap, and by Lemma 3.6, $T_{1,a}>T_{2,a}$. Therefore, the support boundaries of any EAP must be ordered by the following two orderings: 1. Type I: $T_{2,a}<T_{1,a}<T_{2,f}\leq T_{1,f}$ , and 2. Type II: $T_{2,a}<T_{1,a}<T_{1,f}<T_{2,f}$. For both the above types, every EAP must have $T_{2,a}<0$. The following lemma specifies the necessary and sufficient condition for existence of an EAP under the two types mentioned above. Existence of unique EAP under case 3 of Theorem 3.1 follows trivially from this lemma. ###### Lemma 3.9. If $\gamma_{1}>\gamma_{2}$, the following statements are true: 1. 1. There exists an EAP under Type I if and only if $\Lambda_{1}\geq\left(\frac{\mu_{1}}{\mu_{2}}-1\right)\Lambda_{2}$ and if it exists, it has a closed form same as the one mentioned under case 3a of Theorem 3.1. 2. 2. There exists an EAP under Type II if and only if $\Lambda_{1}<\left(\frac{\mu_{1}}{\mu_{2}}-1\right)\Lambda_{2}$ and if it exists, it has a closed form same as the one mentioned under case 3b of Theorem 3.1. Proof of the above lemma follows an argument similar to the proof of Lemma 3.8 by identifying a system of linear equations to be satisfied by the support boundaries of any EAP under the two types. Details are in A.1.1. ∎ The proof of the main results in Sections 3.2, 4.1 and 4.2 will follow a similar sequence of arguments as used in the proof of Theorem 3.1, through elimination of candidates not satisfying structural properties of every EAP in that regime. Figure 3: Illustrative EAP (left) and resulting queue length process (right) of HDS with $\mu_{1}>\mu_{2}$ under case 2a of Theorem 3.1 with the assumption $T_{2,a}<0$ in ( ‣ 2.). Queue 1 divides its capacity between the two classes from $\tau_{1}(T_{2,a})$ proportionally to their arrival rate. As a result, queue 1 serves class 1 users at a slower rate in $[\tau_{1}(T_{2,a}),T_{1,f}]$ than in $[0,\tau_{1}(T_{2,a})]$, causing class 1 mass in queue 1 to decrease at a slower rate in the former interval than in the latter one. Figure 4: Illustrative EAP (left) of HDS with $\mu_{1}>\mu_{2}$ under case 2b and 3b of Theorem 3.1. The resulting queue length process (right) is for case 2b with the assumption $T_{2,a}<0$ in (3.). Queue 1 divides its capacity between the two classes from $\tau_{1}(T_{1,a})$ proportionally to their service rates. As a result, queue 1 serves class 2 users at a lesser rate in $[\tau_{1}(T_{1,a}),T_{1,f}]$ than in $[0,\tau_{1}(T_{1,a})]$, causing class 2 waiting mass to decrease at a lesser rate in the former interval than in the latter one. In queue 2, waiting mass increases in $[\tau_{1}(T_{1,a}),T_{1,f}]$ because class 2 arrival rate to queue 2 is $\mu_{2}\gamma_{1}/\gamma_{2}>\mu_{2}$. However the waiting mass increases in $[\tau_{1}(T_{1,a}),T_{1,f}]$ at a rate slower than in $[0,\tau_{1}(T_{1,a})]$, since the arrival rate in the formal interval is $\mu_{2}\gamma_{2}/\gamma_{1}<\mu_{1}=$ arrival rate in the latter interval. For case 3b of Theorem 3.1, with the assumption $T_{1,a}<0$ on (3.), queue 1 length process has the same structure. Queue 2 length process of case 3b is also same as case 2b, except for one region: queue 2 length decreases in $[\tau_{1}(T_{1,a}),T_{1,f}]$, since class 2 arrival rate to queue 2 is $\mu_{2}\gamma_{2}/\gamma_{1}<\mu_{2}$ in case 3b. Figure 5: Illustrative EAP (left) and resulting queue length process (right) of HDS with $\mu_{1}>\mu_{2}$ under case 3a of Theorem 3.1 with the assumption $T_{1,a}<0$ and $\tau_{1}(T_{1,a})=\frac{\mu_{2}\gamma_{2}}{\mu_{1}}(T_{1,a}-T_{2,a})<T_{2,f}$ in ( ‣ 2.). In $[\tau_{1}(T_{1,a}),T_{1,f}]$, queue 1 divides its capacity between the two classes proportionally to their arrival rates. As a result, class 2 mass in queue 1 decreases at a lesser rate in $[\tau_{1}(T_{1,a}),T_{2,f}]$ than in $[0,\tau_{1}(T_{1,a})]$. After $T_{2,f}$, as class 2 users stop arriving, the class 2 mass in queue 1 decreases to zero at a higher rate in $[T_{2,f},\tau_{1}(T_{2,f})]$ than in $[\tau_{1}(T_{1,a}),T_{2,f}]$. Class 1 mass in queue 1 initially decreases in $[\tau_{1}(T_{1,a}),T_{2,f}]$. Since class 1 users increase their arrival rate after $T_{2,f}$, slope of the red line increases and the class 1 mass in queue 1 can even increase in $[T_{2,f},\tau_{1}(T_{2,f})]$ if $\mu_{1}\gamma_{1}>\mu_{1}-\mu_{2}\gamma_{2}/\gamma_{1}$. After $\tau_{1}(T_{2,f})$, queue 1 allocates the entire capacity to class 1 and the class 1 mass in queue 1 decreases to zero in $[\tau_{1}(T_{2,f}),T_{1,f}]$. ### 3.2 Equal Preferences: $\gamma_{1}=\gamma_{2}=\gamma$ Structural properties of any EAP. We first identify the structural properties of any EAP in this regime and later we exploit them to refine our search of EAP. By Lemma 3.2, we can restrict our search to joint arrival profiles having supports satisfying $\mathcal{S}(F_{1})=[T_{1,a},T_{1,f}]$ and $\mathcal{S}(F_{1})\cup\mathcal{S}(F_{2})=[T_{a},T_{f}]$ for some $T_{f}>T_{a}$ and $T_{1,f}>T_{1,a}$. Lemma 3.10 states the arrival rates of the two classes in EAP and is proved (in A.2) via an argument similar to the one used for proving Lemma 3.4. Lemma 3.11 is proved via contradiction, by showing that, if the stated condition doesn’t hold, users of one of the classes can improve by arriving at a different time. ###### Lemma 3.10 (Rate of arrival). If $\mu_{1}>\mu_{2}$ and $\gamma_{1}=\gamma_{2}=\gamma$, the following properties must be true almost everywhere for every EAP $\mathbf{F}=\\{F_{1},F_{2}\\}$, $\displaystyle F_{1}^{\prime}(t)+F_{2}^{\prime}(t)$ $\displaystyle=\mu_{1}\gamma~{}~{}~{}\text{if $t\in\mathcal{S}(F_{1})$, and,}$ $\displaystyle F_{2}^{\prime}(t)$ $\displaystyle\leq\mu_{2}\gamma~{}~{}~{}\text{if $t\in\mathcal{S}(F_{2})$, with equality when $Q_{2}(\tau_{1}(t))>0$}.$ ###### Lemma 3.11. If $\mu_{1}>\mu_{2}$ and $\gamma_{1}=\gamma_{2}=\gamma$, in the EAP, 1. 1. If $\mathcal{S}(F_{2})$ has a gap $(t,t+\delta]$ at $t\in\mathcal{S}(F_{2})$ for some $\delta>0$ sufficiently small, queue 2 must be empty at $\tau_{1}(t)$. 2. 2. Queue 1 must be empty time at $T_{1,f}$. Specification of the EAP. ###### Theorem 3.2. If $\mu_{1}>\mu_{2}$ and $\gamma_{1}=\gamma_{2}=\gamma$, EAP of HDS exists and: 1. 1. If $\Lambda_{1}<\left(\frac{\mu_{1}}{\mu_{2}}-1\right)\Lambda_{2}$, the EAP is unique and its arrival rates and support boundaries are: $\displaystyle F_{1}^{\prime}(t)$ $\displaystyle=(\mu_{1}-\mu_{2})\gamma~{}\text{for}~{}t\in[T_{1,a},T_{1,f}]~{}\text{and},~{}F_{2}^{\prime}(t)=\mu_{2}\gamma~{}\text{for}~{}t\in[T_{a},T_{f}],$ $\displaystyle\text{where},~{}T_{1,a}$ $\displaystyle=\frac{1-\gamma}{\mu_{1}-\mu_{2}\gamma}\left[\Lambda_{2}-\frac{\mu_{1}}{(\mu_{1}-\mu_{2})\gamma}\Lambda_{1}\right],~{}T_{1,f}=\frac{\Lambda_{1}+(1-\gamma)\Lambda_{2}}{\mu_{1}-\mu_{2}\gamma},~{}T_{a}=-\frac{1-\gamma}{\gamma}\frac{\Lambda_{2}}{\mu_{2}},~{}\text{and}~{}T_{f}=\frac{\Lambda_{2}}{\mu_{2}}.$ 2. 2. If $\Lambda_{1}\geq\left(\frac{\mu_{1}}{\mu_{2}}-1\right)\Lambda_{2}$, the set of EAPs is the convex set of joint arrival profiles $\mathbf{F}=\\{F_{1},F_{2}\\}$ satisfying: 1) $\mathcal{S}(F_{1})=[T_{a},T_{f}]$, $\mathcal{S}(F_{2})\subseteq[T_{a},T_{f}]$, for $i=1,2$ $F_{i}(T_{f})=\Lambda_{i}$, 2) $F_{1}^{\prime}(t)+F_{2}^{\prime}(t)=\mu_{1}\gamma$, and 3) $F_{2}^{\prime}(t)\leq\mu_{2}\gamma$ for every $t\in[T_{a},T_{f}]$, where $T_{a}=-\left(\frac{1}{\gamma}-1\right)\frac{\Lambda_{1}+\Lambda_{2}}{\mu_{1}}$ and $T_{f}=\frac{\Lambda_{1}+\Lambda_{2}}{\mu_{1}}$. We only provide the key steps of the proof of Theorem 3.2 below. The details are given in A.2.1. Following two supporting lemmas are necessary for proving Theorem 3.2. ###### Lemma 3.12. In every EAP, the cost of the class 2 users remains constant over $[T_{a},T_{f}]$. ###### Lemma 3.13. In every EAP, if $t\in\mathcal{S}(F_{1})\cap\mathcal{S}(F_{2})$, then class 2 users will arrive at queue 2 at a maximum rate of $\mu_{2}$ at time $\tau_{1}(t)$. Proofs of Lemma 3.12 and Lemma 3.13 are in A.2. Lemma 3.12 is proved via exploiting Lemma 3.11 and the fact that the two classes have equal cost preferences. Proof of Lemma 3.13 is based on the observation that $A_{2}(\tau_{1}(t))=F_{2}(t)$. As a result, at time $\tau_{1}(t)$, class 2 users arrive at queue 2 at rate $A_{2}^{\prime}(\tau_{1}(t))=\frac{F_{2}^{\prime}(t)}{\tau_{1}^{\prime}(t)}$, assuming that the derivatives exist. Following this, we use Lemma 3.10 and (3) to upper bound the rate of arrival. Key steps of proof of Theorem 3.2: By Lemma 3.2, we consider candidates $\mathbf{F}=\\{F_{1},F_{2}\\}$ which are absolutely continuous with $\mathcal{S}(F_{1})=[T_{1,a},T_{1,f}],~{}\mathcal{S}(F_{1})\cup\mathcal{S}(F_{2})=[T_{a},T_{f}]$ and have arrival rates $F_{1}^{\prime}(\cdot),~{}F_{2}^{\prime}(\cdot)$ given by Lemma 3.10. We eliminate candidates which do not satisfy the structural properties of any EAP and narrow our search of an EAP to a smaller set of candidates. We follow this agenda separately over two subsets of candidates with: 1) $T_{f}>T_{1,f}$ and 2) $T_{f}=T_{1,f}$. Lemma 3.14 and 3.15 provides the necessary and sufficient condition for existence of EAPs of the mentioned two types. The statement of Theorem 3.2 follows from these two lemmas. ###### Lemma 3.14. If $\gamma_{1}=\gamma_{2}=\gamma$, there exists an EAP with $T_{f}>T_{1,f}$ if and only if $\Lambda_{1}<\left(\frac{\mu_{1}}{\mu_{2}}-1\right)\Lambda_{2}$, and if it exists, it will be unique with a closed form same as the joint arrival profile mentioned under case 1 of Theorem 3.2. ###### Lemma 3.15. If $\gamma_{1}=\gamma_{2}=\gamma$, there exists an EAP with $T_{f}=T_{1,f}$ if and only if $\Lambda_{1}\geq\left(\frac{\mu_{1}}{\mu_{2}}-1\right)\Lambda_{2}$, and if it exists, the set of all EAPs with $T_{1,f}=T_{f}$ is the set of joint arrival profiles mentioned under case 2 of Theorem 3.2. Below we sketch the proof of Lemma 3.14 and 3.15. The remaining details are presented in A.2.1. Proof sketch of Lemma 3.14: Identifying the support boundaries: First we identify the support boundaries $T_{a},T_{f}$ of any EAP with $T_{f}>T_{1,f}$: 1. 1) Exploiting Lemma 3.13 and 3.11, we argue that queue 2 has a positive waiting time in $(0,T_{f})$ and empties at $T_{f}$. As a result, queue 2 serves the whole class 2 population in $(0,T_{f})$, giving us $T_{f}=\frac{\Lambda_{2}}{\mu_{2}}$. 2. 2) Using Lemma 3.12 we have, $C_{\mathbf{F}}^{(2)}(T_{a})=C_{\mathbf{F}}^{(2)}(T_{f})$. The class 2 user arriving at $T_{a}$ gets served at time zero causing $C_{\mathbf{F}}^{(2)}(T_{a})=-\gamma T_{a}$. Since the network is empty at $T_{f}$, $C_{\mathbf{F}}^{(2)}(T_{f})=(1-\gamma)T_{f}$. Therefore, $C_{\mathbf{F}}^{(2)}(T_{a})=C_{\mathbf{F}}^{(2)}(T_{f})$ implies $T_{a}=-\left(\frac{1}{\gamma}-1\right)T_{f}=-\left(\frac{1}{\gamma}-1\right)\frac{\Lambda_{2}}{\mu_{2}}$. 3. 3) The class 2 population of mass $\Lambda_{2}$ has to arrive in the interval $\left[T_{a},T_{f}\right]$ of length $\frac{\Lambda_{2}}{\mu_{2}\gamma_{2}}$ at maximum rate of $\mu_{2}\gamma_{2}$ (by Lemma 3.10). This leaves us with only one possible arrival rate for class 2 users: $F_{2}^{\prime}(t)=\mu_{2}\gamma_{2}$ in $[T_{a},T_{f}]$. As a result, by Lemma 3.10, we must have $F_{1}^{\prime}(t)=(\mu_{1}-\mu_{2})\gamma$ in $[T_{1,a},T_{1,f}]$, giving us $T_{1,f}=T_{1,a}+\frac{\Lambda_{1}}{(\mu_{1}-\mu_{2})\gamma}$. 4. 4) Since $T_{a}<0$, queue 1 starts serving from time zero and has positive waiting time in $[0,T_{1,f})$. By Lemma 3.11, queue 1 empties at $T_{1,f}$. Therefore, $\mu_{1}T_{1,f}=F_{1}(T_{1,f})+F_{2}(T_{1,f})=\Lambda_{1}+\newline \mu_{2}\gamma\cdot(T_{1,f}-T_{a})$. Getting the necessary condition: Solving the above system of equations, we get $T_{a}=-\left(\frac{1}{\gamma}-1\right)\frac{\Lambda_{2}}{\mu_{2}}$, $T_{f}=\frac{\Lambda_{2}}{\mu_{2}}$, $T_{1,a}=\frac{1-\gamma}{\mu_{1}-\mu_{2}\gamma}\left[\Lambda_{2}-\frac{\mu_{1}}{(\mu_{1}-\mu_{2})\gamma}\Lambda_{1}\right]$, $T_{1,f}=\frac{\Lambda_{1}+(1-\gamma)\Lambda_{2}}{\mu_{1}-\mu_{2}\gamma}$ and as a result, every EAP with $T_{f}>T_{1,f}$ has $T_{a},T_{f},T_{1,a},T_{1,f}$ same as the ones we identified. For $T_{f}>T_{1,f}$, we need $\Lambda_{1}<\left(\frac{\mu_{1}}{\mu_{2}}-1\right)\Lambda_{2}$ and therefore, this is a necessary condition for existence of an EAP with $T_{f}>T_{1,f}$. Identifying the unique EAP: If the necessary condition $\Lambda_{1}<\left(\frac{\mu_{1}}{\mu_{2}}-1\right)\Lambda_{2}$ is satisfied, it is easy to verify that, the support boundaries obtained above follow the ordering $T_{a}<T_{1,a}<T_{1,f}<T_{f}$. With this ordering satisfied, the only candidate with $T_{f}>T_{1,f}$, which qualifies to be an EAP is: $F_{1}^{\prime}(t)=(\mu_{1}-\mu_{2})\gamma$ in $[T_{1,a},T_{1,f}]$ and $F_{2}^{\prime}(t)=\mu_{2}\gamma$ in $[T_{a},T_{f}]$, with $T_{a},T_{f},T_{1,a},T_{1,f}$ same as we identified before. Moreover, the candidate obtained is same as the joint arrival profile under case 1 of Theorem 3.2. Proving sufficiency of the obtained necessary condition: Another interesting observation is, if $\Lambda_{1}<\left(\frac{\mu_{1}}{\mu_{2}}-1\right)\Lambda_{2}$, the obtained candidate with $T_{f}>T_{1,f}$ is same as the EAPs in cases 2b and 3b of Theorem 3.1, respectively, upon taking limits $\gamma_{1}=\gamma,~{}\gamma_{2}\to\gamma+$ and $\gamma_{2}=\gamma,~{}\gamma_{1}\to\gamma+$. Proving that this candidate is an EAP follows by the argument used for proving the unique Type II candidate in Lemma 3.8 is an EAP (in A.1.1), except replacing $T_{2,a},T_{2,f}$ with $T_{a},T_{f}$ and taking $\gamma_{1}=\gamma_{2}=\gamma$. Therefore, the statement of Lemma 3.14 stands proved. ∎ Proof sketch of Lemma 3.15: Identifying $\mathbf{T_{a}}$: Since $[0,T_{f}]\subseteq\mathcal{S}(F_{1})$, queue 1 must have a positive waiting time in $[0,T_{f})$. By Lemma 3.11, queue 1 empties at time $T_{f}$. Therefore we get $\mu_{1}T_{f}=F_{1}(T_{f})+F_{2}(T_{f})=\Lambda_{1}+\Lambda_{2}$ implying $T_{f}=\frac{\Lambda_{1}+\Lambda_{2}}{\mu_{1}}$. Identifying $\mathbf{T_{f}}$: By Lemma 3.12, we have $C_{\mathbf{F}}^{(2)}(T_{a})=C_{\mathbf{F}}^{(2)}(T_{f})$. Using the argument used in the proof sketch of Lemma 3.14 for finding $T_{a}$, we get $T_{a}=-\left(\frac{1}{\gamma}-1\right)T_{f}=-\left(\frac{1}{\gamma}-1\right)\frac{\Lambda_{1}+\Lambda_{2}}{\mu_{1}}$. Getting the necessary condition: By Lemma 3.10, all class 2 users arrive in $[T_{a},T_{f}]$ at a maximum rate of $\mu_{2}\gamma$. Therefore for existence of an EAP with $T_{f}=T_{1,f}$, we need the necessary condition $\mu_{2}\gamma\cdot(T_{f}-T_{a})\geq\Lambda_{2}$, which after some manipulation gives us $\Lambda_{1}\geq\left(\frac{\mu_{1}}{\mu_{2}}-1\right)\Lambda_{2}$. Identifying the set of EAPs: By Lemma 3.10, $\mu_{1}\gamma$ is the maximum rate at which the entire population of mass $\Lambda_{1}+\Lambda_{2}$ can arrive within the time interval $[T_{a},T_{f}]$ of length $\frac{\Lambda_{1}+\Lambda_{2}}{\mu_{1}\gamma}$. Therefore, we must have $F_{1}^{\prime}(t)+F_{2}^{\prime}(t)=\mu_{1}\gamma$ in $[T_{a},T_{f}]$. Also by Lemma 3.10, we have $F_{1}^{\prime}(t)+F_{2}^{\prime}(t)=\mu_{1}\gamma$ only in $\mathcal{S}(F_{1})=[T_{1,a},T_{1,f}]\subseteq[T_{a},T_{f}]$, which implies $T_{1,a}=T_{a}$. By Lemma 3.13, class 2 users arrive at queue 2 from queue 1 in $[0,T_{f}]$ at a maximum rate of $\mu_{2}$. Hence queue 2 stays empty and as a result, by Lemma 3.10, $F_{2}^{\prime}(t)\leq\mu_{2}\gamma$ for every $t\in[T_{a},T_{f}]$. Therefore, our reduced set of candidates with $T_{f}=T_{1,f}$ will be the set of all joint arrival profiles $\mathbf{F}=\\{F_{1},F_{2}\\}$ satisfying: 1) $\mathcal{S}(F_{1})=[T_{a},T_{f}]$ and $\mathcal{S}(F_{2})\subseteq[T_{a},T_{f}]$, for $i=1,2$ $F_{i}(T_{f})=\Lambda_{i}$, 2) $F_{1}^{\prime}(t)+F_{2}^{\prime}(t)=\mu_{1}\gamma$, and 3) $F_{2}^{\prime}(t)\leq\mu_{2}\gamma$ for every $t\in[T_{a},T_{f}]$ where $T_{a}=-\frac{1-\gamma}{\gamma}\frac{\Lambda_{1}+\Lambda_{2}}{\mu_{1}}$ and $T_{f}=\frac{\Lambda_{1}+\Lambda_{2}}{\mu_{1}}$, which is exactly the set mentioned under the case 2 of Theorem 3.2. Proving sufficiency of the obtained necessary condition: It is easy to verify that, if the necessary condition $\Lambda_{1}\geq\left(\frac{\mu_{1}}{\mu_{2}}-1\right)\Lambda_{2}$ holds, the set of candidates with $T_{f}=T_{1,f}$ mentioned earlier is non-empty. Two elements of the set are the limits of the EAPs in cases 2a and 3a of Theorem 3.1, respectively, when $\gamma_{1}=\gamma,~{}\gamma_{2}\to\gamma+$ and $\gamma_{2}=\gamma,~{}\gamma_{1}\to\gamma+$. After this, we prove that, if $\Lambda_{1}\geq\left(\frac{\mu_{1}}{\mu_{2}}-1\right)\cdot\Lambda_{2}$, every joint arrival profile $\mathbf{F}=\\{F_{1},F_{2}\\}$ in the obtained set of candidates is an EAP. This step requires exploiting Lemma 3.12 and the fact that queue 2 stays empty when users are arriving by such a candidate and the details of it are in A.2.1. Hence, the statement of Lemma 3.15 stands proved. ∎ ## 4 Heterogeneous Arrival System (HAS) In this section, we consider the two groups arrive at different queues and depart from the same queue as illustrated in Table 1. Namely, the $i$-th group arrive at $i$-th queue and both the groups depart from the 2nd queue. We consider the case $\gamma_{1}\neq\gamma_{2}$ and $\gamma_{1}=\gamma_{2}$ separately, since the later displays different behaviors. ### 4.1 Unequal Preferences $\gamma_{1}\neq\gamma_{2}$ As for HDS, here too we narrow down our search for EAP by exploiting its structural properties. The detailed account of these structural properties as well as detailed proofs of results in this section are given in B.1. The main results in this section are proved by arguments similar to those used for proving Theorem 3.1 for HDS, except that we now exploit a different set of structural properties and the no. of different regimes is significantly larger and diverse for HAS. In this section, we state the key results and after Theorem 4.2 in Remark 7, we discuss two interesting cases: 1) where the EAP corresponds to arrivals in a stream coming in disjoint intervals and 2) where all the arrivals for class 2 arrive before time zero. The following lemma identifies an important threshold property of an EAP. Proof of Lemma 4.1 is in B.1 and follows a sequence of arguments similar to the one used for proving Lemma 3.1. ###### Lemma 4.1 (Threshold Behavior). If $\gamma_{1}\neq\gamma_{2}$, in EAP, queue 2 serves the two classes over disjoint sets of time if and only if $\mu_{1}\geq\mu_{2}\gamma_{2}$. Proof sketch: Proof of Lemma 4.1 is similar to that of Lemma 3.1. First we argue via a contradiction that if $\mu_{1}\geq\mu_{2}\gamma_{2}$, then queue 2 cannot serve the two classes together in an EAP. If queue 2 is serving the two classes together, we argue that, queue 1 must stay engaged in a neighbourhood of that time for class 1 to be iso cost (by Lemma B.1 in B.1). As a result, class 1 users arrive from queue 1 to 2 at rate $\mu_{1}$. For class 2 have constant cost in that neighbourhood, we argue that, users of the two classes arrive at queue 2 at a combined rate of $\mu_{2}\gamma_{2}$. Since queue 2 serves a positive mass of both the classes together, the combined arrival rate must be strictly larger than the arrival rate of class 1 to queue 2, implying $\mu_{2}\gamma_{2}>\mu_{1}$ and contradicting $\mu_{1}\geq\mu_{2}\gamma_{2}$. The other direction is proved via contradiction by exploiting the structures of $\mathcal{S}(F_{1}),~{}\mathcal{S}(F_{2})$ in an EAP and showing that, if $\mu_{1}<\mu_{2}\gamma_{2}$ and queue 2 is serving the two classes over disjoint sets of times, we can find a user who can decrease her cost by arriving at a different time. ∎ Specification of the EAP. The two regimes $\mu_{1}<\mu_{2}\gamma_{2}$ and $\mu_{1}\geq\mu_{2}\gamma_{2}$ exhibit substantially different EAP structures, owing to the threshold behavior given by Lemma 4.1. For conciseness, we specify the unique EAP in these two regimes separately in Theorem 4.1 (for $\mu_{1}<\mu_{2}\gamma_{2}$) and 4.2 (for $\mu_{1}\geq\mu_{2}\gamma_{2}$). Proofs of these theorems (in B.1) have a structure similar to the proof of Theorem 3.1. For the two classes $i=1,2$, we define the quantities $T_{i,a}=\inf\mathcal{S}(F_{i})$ and $T_{i,f}=\sup\mathcal{S}(F_{i})$. In an EAP, both the sets $\mathcal{S}(F_{1}),~{}\mathcal{S}(F_{2})$ are compact and therefore $T_{1,a},T_{1,f},T_{2,a},T_{2,f}$ are all finite. Regime I $\mu_{1}<\mu_{2}\gamma_{2}$. We mention below the support boundaries of the unique EAP, which we will refer later in Theorem 4.1: 1. 1. If $\mu_{1}<\mu_{2}\gamma_{2}$ and $\Lambda_{1}\geq\max\left\\{\frac{1-\gamma_{2}}{1-\gamma_{1}},1\right\\}\cdot\frac{\mu_{1}}{\mu_{2}-\mu_{1}}\Lambda_{2}$, then $\displaystyle T_{1,a}$ $\displaystyle=-\frac{1-\gamma_{1}}{\gamma_{1}}\frac{\Lambda_{1}}{\mu_{1}}+\frac{1-\gamma_{2}}{\gamma_{1}}\frac{\Lambda_{2}}{\mu_{2}-\mu_{1}},~{}T_{1,f}=\frac{\Lambda_{1}}{\mu_{1}},~{}T_{2,a}=-\frac{1-\gamma_{2}}{\gamma_{2}}\frac{\Lambda_{2}}{\mu_{2}-\mu_{1}},~{}\text{and}~{}T_{2,f}=\frac{\Lambda_{2}}{\mu_{2}-\mu_{1}}.$ (10) 2. 2. If $\mu_{1}<\mu_{2}\gamma_{2}$ and $\Lambda_{1}<\max\left\\{\frac{1-\gamma_{2}}{1-\gamma_{1}},1\right\\}\cdot\frac{\mu_{1}}{\mu_{2}-\mu_{1}}\Lambda_{2}$, then 1. a) if $\gamma_{1}>\gamma_{2}$, then $\displaystyle T_{1,a}$ $\displaystyle=\frac{1}{\mu_{2}}\left(\Lambda_{2}-\frac{1-\gamma_{1}}{1-\gamma_{2}}\frac{\mu_{2}-\mu_{1}}{\mu_{1}}\Lambda_{1}\right),~{}T_{1,f}=\frac{1}{\mu_{2}}\left(\Lambda_{2}+\frac{(\gamma_{1}-\gamma_{2})\mu_{2}+(1-\gamma_{1})\mu_{1}}{(1-\gamma_{2})\mu_{1}}\Lambda_{1}\right),$ $\displaystyle T_{2,a}$ $\displaystyle=-\frac{1}{\mu_{2}\gamma_{2}}\left((1-\gamma_{1})\Lambda_{1}+(1-\gamma_{2})\Lambda_{2}\right),~{}\text{and}~{}T_{2,f}=\frac{1}{\mu_{2}}\left(\Lambda_{2}+\frac{1-\gamma_{1}}{1-\gamma_{2}}\Lambda_{1}\right).$ (11) 2. b) if $\gamma_{1}<\gamma_{2}$, then $\displaystyle T_{1,a}$ $\displaystyle=-\left(\frac{\gamma_{2}}{\gamma_{1}}-1\right)\frac{\Lambda_{1}}{\mu_{1}},~{}T_{1,f}=\frac{\Lambda_{1}}{\mu_{1}},~{}T_{2,a}=-\frac{1-\gamma_{2}}{\gamma_{2}}\frac{\Lambda_{1}+\Lambda_{2}}{\mu_{2}},~{}\text{and}~{}T_{2,f}=\frac{\Lambda_{1}+\Lambda_{2}}{\mu_{2}}.$ (12) ###### Theorem 4.1. If $\gamma_{1}\neq\gamma_{2}$ and $\mu_{1}<\mu_{2}\gamma_{2}$, HAS has a unique EAP which has the following arrival rates and support boundaries: 1. 1. If $\gamma_{1}>\gamma_{2}$: 1. 1a. when $\Lambda_{1}\geq\frac{1-\gamma_{2}}{1-\gamma_{1}}\frac{\mu_{1}}{\mu_{2}-\mu_{1}}\Lambda_{2}$: $\displaystyle F_{1}^{\prime}(t)$ $\displaystyle=\begin{cases}\frac{\mu_{1}\gamma_{1}}{\gamma_{2}}~{}~{}&\text{for}~{}t\in[T_{1,a},\tau_{1}^{-1}(T_{2,f})],\\\ \mu_{1}\gamma_{1}~{}~{}&\text{for}~{}t\in[\tau_{1}^{-1}(T_{2,f}),T_{1,f}],\end{cases}~{}\text{and}~{}F_{2}^{\prime}(t)=\begin{cases}\mu_{2}\gamma_{2}~{}~{}&\text{for}~{}t\in[T_{2,a},0],\\\ \mu_{2}\gamma_{2}-\mu_{1}~{}~{}&\text{for}~{}t\in[0,T_{2,f}],\\\ \end{cases}$ where $T_{1,a},T_{1,f},T_{2,a}$ and $T_{2,f}$ are given by (10). 2. 1b. when $\Lambda_{1}<\frac{1-\gamma_{2}}{1-\gamma_{1}}\frac{\mu_{1}}{\mu_{2}-\mu_{1}}\Lambda_{2}$: $\displaystyle F_{1}^{\prime}(t)$ $\displaystyle=\begin{cases}\frac{\mu_{1}\gamma_{1}}{\gamma_{2}}~{}~{}&\text{for}~{}t\in[T_{1,a},\tau_{1}^{-1}(T_{2,f})],{}\\\ \mu_{1}\gamma_{1}~{}~{}&\text{for}~{}t\in[\tau_{1}^{-1}(T_{2,f}),T_{1,f}],\\\ \end{cases}~{}\text{and}~{}F_{2}^{\prime}(t)=\begin{cases}\mu_{2}\gamma_{2}~{}~{}&\text{for}~{}t\in[T_{2,a},T_{1,a}],\\\ \mu_{2}\gamma_{2}-\mu_{1}~{}~{}&\text{for}~{}t\in[T_{1,a},T_{2,f}],\\\ \end{cases}$ where $T_{1,a},T_{1,f},T_{2,a}$ and $T_{2,f}$ are given by ( ‣ 2.). 2. 2. If $\gamma_{1}<\gamma_{2}$: 1. 2a. when $\Lambda_{1}\geq\frac{\mu_{1}}{\mu_{2}-\mu_{1}}\Lambda_{2}$, the EAP has closed form same as case 1a. 2. 2b. when $\Lambda_{1}<\frac{\mu_{1}}{\mu_{2}-\mu_{1}}\Lambda_{2}$: $\displaystyle F_{1}^{\prime}(t)$ $\displaystyle=\frac{\mu_{1}\gamma_{1}}{\gamma_{2}}~{}~{}\text{for}~{}t\in[T_{1,a},T_{1,f}],~{}\text{and}~{}F_{2}^{\prime}(t)=\begin{cases}\mu_{2}\gamma_{2}~{}~{}&\text{for}~{}t\in[T_{2,a},0]\cup[T_{1,f},T_{2,f}],\\\ \mu_{2}\gamma_{2}-\mu_{1}~{}~{}&\text{for}~{}t\in[0,T_{1,f}],\end{cases}$ where $T_{1,a},T_{1,f},T_{2,a}$ and $T_{2,f}$ are given by (12). ###### Remark 5. _Illustrative EAPs and resulting queue length processes of cases 1a, 1b, 2a and 2b of Theorem 4.1 are illustrated, respectively, in Figures 6, 7, 8 and 9. Structure of the queue length processes in some regimes might vary depending on the support boundaries and related quantities. We have illustrated only one possible structure of the queue length process in those regimes and mentioned the conditions on the support boundaries for the attainment of that structure in the caption. Red and blue respectively denote class 1 and 2 populations. In the plot on right-top, the black dashed line represents the total waiting mass of the two classes in queue 2. The same is true for the illustrative EAPs and resulting queue length processes referred to in Theorem 4.2._ Figure 6: Illustrative EAP (left) and queue length process (right) for HAS with $\mu_{1}<\mu_{2}\gamma_{2}$ under case 1a of Theorem 4.1 and the assumption $T=T_{1,a}+\frac{\gamma_{2}}{\gamma_{1}}T_{2,f}>0$ on (10). $T\overset{def.}{=}\tau_{1}^{-1}(T_{2,f})$. Figure 7: Illustrative EAP (left) and queue length process (right) for HAS with $\mu_{1}<\mu_{2}\gamma_{2}$ under case 1b of Theorem 4.1. $T\overset{def.}{=}\tau_{1}^{-1}(T_{2,f})$. Figure 8: Illustrative EAP (left) and queue length process (right) for HAS with $\mu_{1}<\mu_{2}\gamma_{2}$ under case 2a of Theorem 4.2 and the assumption $T=T_{1,a}+\frac{\gamma_{2}}{\gamma_{1}}T_{2,f}>0$ on (10). $T\overset{def.}{=}\tau_{1}^{-1}(T_{2,f})$. Figure 9: Illustrative EAP for HAS with $\mu_{1}<\mu_{2}\gamma_{2}$ under case 2b of Theorem 4.2 and the assumption $\tau_{2}(0)=-\gamma_{2}T_{2,a}<T_{1,f}$ on (12). Mass of class 2 waiting in queue 2 can increase in $[T_{1,f},\tau_{2}(T_{1,f})]$ if $\mu_{2}\gamma_{2}>\mu_{2}-\mu_{1}/\gamma_{2}$. Regime II $\mu_{1}\geq\mu_{2}\gamma_{2}$. The quantity $T\overset{def}{=}\inf\\{t>0~{}|~{}Q_{2}(t)=0\\}$, representing the time at which queue 2 empties for the first time, is necessary to describe the unique EAP in Theorem 4.2. We mention below the support boundaries and $T$ of the unique EAP, which we will refer later in the statement of Theorem 4.2: 1. 1. If $\mu_{2}\gamma_{1}>\mu_{1}\geq\mu_{2}\gamma_{2}$ and $\Lambda_{1}\geq\frac{\mu_{1}}{(1-\gamma_{1})\mu_{2}}\Lambda_{2}$; or $\mu_{1}\geq\mu_{2}\cdot\max\\{\gamma_{1},\gamma_{2}\\}$ and $\left(\frac{\mu_{2}}{\mu_{1}}-1\right)\Lambda_{1}>\Lambda_{2}$, then $\displaystyle T_{1,a}$ $\displaystyle=\frac{\Lambda_{2}}{\mu_{2}\gamma_{1}}-\left(\frac{1}{\gamma_{1}}-1\right)\frac{\Lambda_{1}}{\mu_{1}},~{}T_{1,f}=\frac{\Lambda_{1}}{\mu_{1}},~{}T=\frac{\Lambda_{2}}{\mu_{2}-\mu_{1}},~{}T_{2,a}=-\frac{\Lambda_{2}}{\mu_{2}\gamma_{2}},~{}\text{and}~{}T_{2,f}=0.$ (13) 2. 2. If $\mu_{2}\gamma_{1}>\mu_{1}\geq\mu_{2}\gamma_{2}$ and $\Lambda_{1}<\frac{\mu_{1}}{(1-\gamma_{1})\mu_{2}}\Lambda_{2}$, then $\displaystyle T_{1,a}=T_{2,f}$ $\displaystyle=\frac{1}{\mu_{2}}\left(\Lambda_{2}-\frac{(1-\gamma_{1})\mu_{2}}{\mu_{1}}\Lambda_{1}\right),~{}T_{1,f}=\gamma_{1}\frac{\Lambda_{1}}{\mu_{1}}+\frac{\Lambda_{2}}{\mu_{2}},$ $\displaystyle T$ $\displaystyle=\frac{\Lambda_{2}}{\mu_{2}}+\frac{(1-\gamma_{1})\Lambda_{1}}{\mu_{2}-\mu_{1}},~{}\text{and}~{}T_{2,a}=-\left(\frac{1}{\gamma_{2}}-1\right)\frac{\Lambda_{2}}{\mu_{2}}-(1-\gamma_{1})\frac{\Lambda_{1}}{\mu_{1}}.$ (14) 3. 3. If $\mu_{1}\geq\mu_{2}\cdot\max\\{\gamma_{1},\gamma_{2}\\}$ and $\left(\frac{\mu_{2}}{\mu_{1}}-1\right)\Lambda_{1}\leq\Lambda_{2}\leq\left(\frac{1}{\max\\{\gamma_{1},\gamma_{2}\\}}-1\right)\Lambda_{1}$: $\displaystyle T_{1,a}$ $\displaystyle=\frac{1}{\mu_{2}}\left(\Lambda_{2}-\left(\frac{1}{\gamma_{1}}-1\right)\Lambda_{1}\right),~{}T=T_{1,f}=\frac{\Lambda_{1}+\Lambda_{2}}{\mu_{2}},~{}T_{2,a}=-\frac{\Lambda_{2}}{\mu_{2}\gamma_{2}},~{}\text{and}~{}T_{2,f}=0.$ (15) 4. 4. If $\mu_{1}\geq\mu_{2}\gamma_{1}>\mu_{2}\gamma_{2}$ and $\Lambda_{2}>\left(\frac{1}{\gamma_{1}}-1\right)\Lambda_{1}$: $\displaystyle T_{1,a}=T_{2,f}$ $\displaystyle=-\left(\frac{1}{\gamma_{1}}-1\right)\frac{\Lambda_{1}}{\mu_{2}}+\frac{\Lambda_{2}}{\mu_{2}},~{}T_{1,f}=T=\frac{\Lambda_{1}+\Lambda_{2}}{\mu_{2}},~{}\text{and}$ $\displaystyle T_{2,a}$ $\displaystyle=-\left(\frac{1}{\gamma_{1}}-1\right)\frac{\Lambda_{1}}{\mu_{2}}-\left(\frac{1}{\gamma_{2}}-1\right)\frac{\Lambda_{2}}{\mu_{2}}.$ (16) 5. 5. If $\mu_{1}\geq\mu_{2}\gamma_{2}>\mu_{2}\gamma_{1}$ and $\Lambda_{2}>\left(\frac{1}{\gamma_{2}}-1\right)\Lambda_{1}$: $\displaystyle T_{1,a}$ $\displaystyle=-\left(\frac{1}{\gamma_{1}}-\frac{1}{\gamma_{2}}\right)\frac{\Lambda_{1}}{\mu_{2}},~{}T_{1,f}=\frac{\Lambda_{1}}{\mu_{2}\gamma_{2}},~{}T=T_{2,f}=\frac{\Lambda_{1}+\Lambda_{2}}{\mu_{2}},~{}\text{and}~{}T_{2,a}=-\frac{1-\gamma_{2}}{\gamma_{2}}\frac{\Lambda_{1}+\Lambda_{2}}{\mu_{2}}.$ (17) ###### Theorem 4.2. If $\mu_{1}\geq\mu_{2}\gamma_{2}$ and $\gamma_{1}\neq\gamma_{2}$, HAS has a unique EAP. Moreover, the EAP has the following arrival rates and support boundaries: 1. 1. If $\mu_{2}\gamma_{1}>\mu_{1}\geq\mu_{2}\gamma_{2}$: 1. 1a. when $\Lambda_{1}\geq\frac{\mu_{1}}{(1-\gamma_{1})\mu_{2}}\Lambda_{2}$: $\displaystyle F_{1}^{\prime}(t)$ $\displaystyle=\begin{cases}\mu_{2}\gamma_{1}~{}~{}&\text{for}~{}t\in[T_{1,a},\tau_{1}^{-1}(T)],\\\ \mu_{1}\gamma_{1}~{}~{}&\text{for}~{}t\in[\tau_{1}^{-1}(T),T_{1,f}],\end{cases}~{}\text{and}~{}F_{2}^{\prime}(t)=\mu_{2}\gamma_{2}~{}~{}\text{for}~{}t\in\left[T_{2,a},T_{2,f}\right]$ where $T_{1,a},T_{1,f},T_{2,a},T_{2,f}$ and $T$ are in (13). 2. 1b. when $\Lambda_{1}<\frac{\mu_{1}}{(1-\gamma_{1})\mu_{2}}\Lambda_{2}$, $\displaystyle F_{1}^{\prime}(t)$ $\displaystyle=\begin{cases}\mu_{2}\gamma_{1}~{}~{}&\text{for}~{}t\in[T_{1,a},\tau_{1}^{-1}(T)],\\\ \mu_{1}\gamma_{1}~{}~{}&\text{for}~{}t\in[\tau_{1}^{-1}(T),T_{1,f}],\end{cases}~{}\text{and}~{}F_{2}^{\prime}(t)=\mu_{2}\gamma_{2}~{}~{}~{}\text{for}~{}t\in[T_{2,a},T_{2,f}]$ where $T_{1,a},T_{1,f},T_{2,a},T_{2,f}$ and $T$ are in (2.). 2. 2. If $\mu_{1}\geq\mu_{2}\gamma_{1}>\mu_{2}\gamma_{2}$: 1. 2a. when $\left(\frac{\mu_{2}}{\mu_{1}}-1\right)\Lambda_{1}>\Lambda_{2}$, the EAP has a closed form same as case 1a. 2. 2b. when $\left(\frac{1}{\gamma_{1}}-1\right)\Lambda_{1}\geq\Lambda_{2}\geq\left(\frac{\mu_{2}}{\mu_{1}}-1\right)\Lambda_{1}$, $\displaystyle F_{1}^{\prime}(t)$ $\displaystyle=\mu_{2}\gamma_{1}~{}~{}\text{for}~{}t\in[T_{1,a},T_{1,f}]~{}\text{and}~{}F_{2}^{\prime}(t)=\mu_{2}\gamma_{2}~{}~{}\text{for}~{}t\in\left[T_{2,a},T_{2,f}\right]$ where $T_{1,a},T_{1,f},T_{2,a},T_{2,f}$ and $T$ are in (15). 3. 2c. when $\Lambda_{2}>\left(\frac{1}{\gamma_{1}}-1\right)\Lambda_{1}$, $\displaystyle F_{1}^{\prime}(t)$ $\displaystyle=\mu_{2}\gamma_{1}~{}~{}~{}\text{for}~{}t\in[T_{1,a},T_{1,f}]~{}\text{and}~{}F_{2}^{\prime}(t)=\mu_{2}\gamma_{2}~{}~{}~{}\text{for}~{}t\in[T_{2,a},T_{2,f}]$ where $T_{1,a},T_{1,f},T_{2,a},T_{2,f}$ and $T$ are in (4.). 3. 3. If $\mu_{1}\geq\mu_{2}\gamma_{2}>\mu_{2}\gamma_{1}$: 1. 3a. when $\left(\frac{\mu_{2}}{\mu_{1}}-1\right)\Lambda_{1}>\Lambda_{2}$, the EAP has a closed form same as case 1a. 2. 3b. when $\left(\frac{1}{\gamma_{2}}-1\right)\Lambda_{1}\geq\Lambda_{2}\geq\left(\frac{\mu_{2}}{\mu_{1}}-1\right)\Lambda_{1}$, the EAP has a closed form same as case 2b. 3. 3c. when $\Lambda_{2}>\left(\frac{1}{\gamma_{2}}-1\right)\Lambda_{1}$, $\displaystyle F_{1}^{\prime}(t)$ $\displaystyle=\mu_{2}\gamma_{1}~{}~{}\text{for}~{}t\in[T_{1,a},T_{1,f}]~{}\text{and}~{}F_{2}^{\prime}(t)=\mu_{2}\gamma_{2}~{}~{}\text{for}~{}t\in[T_{2,a},0]\cup[T_{1,f},T_{2,f}]$ where $T_{1,a},T_{1,f},T_{2,a},T_{2,f}$ and $T$ are in (17). ###### Remark 6. _Illustrative EAPs and resulting queue length processes of cases 1a, 1b, 2a, 2b, 2c, 3a, 3b, and 3c of Theorem 4.2 are illustrated, respectively, in Figures 10, 11, 12, 13, 14, 15, 16 and 17._ ###### Remark 7. _We show by Lemma B.6 and B.12 in B.1 that in every EAP of HAS under case 3 ($\mu_{1}\geq\mu_{2}\gamma_{2}>\mu_{2}\gamma_{1}$) of Theorem 4.2, class 1 users arrive from queue 1 to queue 2 in $[0,T_{1,f}]$ at rate $\mu_{1}$. As a result, by Lemma 4.1, class 2 users cannot arrive in $[0,T_{1,f}]$. However, in every EAP, class 2 users start arriving before time zero (otherwise queue 2 stays empty at time zero and every class 2 user can be strictly better off arriving at time zero). Therefore there are two possible ways in which class 2 users can arrive in any EAP: 1) the whole class 2 population arrives before time zero, or, 2) a fraction of the class 2 population arrives after $T_{1,f}$ and the rest arrives before time zero. While proving the existence of a unique EAP in case 3 of Theorem 4.2, we argue that, if $\Lambda_{2}$ is under an identified threshold of $\left(\frac{1}{\gamma_{2}}-1\right)\Lambda_{1}$, the whole class 2 population arrives before time zero and waits in queue 2 at time zero, as can be seen for cases 3a and 3b, respectively, in Figure 15 and 16. After $\Lambda_{2}$ crosses that threshold, a fraction of the class 2 population of mass $\gamma_{2}\left[\Lambda_{2}-\left(\frac{1}{\gamma_{2}}-1\right)\Lambda_{1}\right]$ arrives after $T_{1,f}$ and the rest arrives before time zero, as can be seen for case 3c in Figure 17._ Figure 10: Illustrative EAP (left) and resulting queue length process (right) of HAS with $\mu_{1}\geq\mu_{2}\gamma_{2}$ under case 1a of Theorem 4.2 with the assumption $T^{0}>0$, where $T^{0}\overset{def.}{=}\tau_{1}^{-1}(T)$. Figure 11: Illustrative EAP (left) and resulting queue length process (right) of HAS with $\mu_{2}\gamma_{1}>\mu_{1}\geq\mu_{2}\gamma_{2}$ under case 1b of Theorem 4.2. $T^{0}\overset{def.}{=}\tau_{1}^{-1}(T)$. . Figure 12: Illustrative EAP (left) and resulting queue length process (right) of HAS with $\mu_{1}\geq\mu_{2}\gamma_{1}>\mu_{2}\gamma_{2}$ under case 2a of Theorem 4.2 with the assumption $T^{0}>0$ where $T^{0}\overset{def.}{=}\tau_{1}^{-1}(T)$. Figure 13: Illustrative EAP (left) and queue length process (right) of HAS with $\mu_{1}\geq\mu_{2}\gamma_{1}>\mu_{2}\gamma_{2}$ under case 2b of Theorem 4.2 with assumption $\tau_{2}(0)<T_{1,idle}$, where $T_{1,idle}$ is the time at which queue 1 empties after the network starts at time zero. In the illustration, we assumed $\mu_{1}<\mu_{2}$, causing mass of class 1 users waiting in queue 2 to decrease in $[\tau_{2}(0),T_{1,idle}]$, but it can be non-decreasing if $\mu_{1}\geq\mu_{2}$. For the same assumption, the overall mass of the two classes is decreasing in $[0,T_{1,idle}]$ and can be non- decreasing if $\mu_{1}\geq\mu_{2}$. Figure 14: Illustrative EAP (left) and resulting queue length process (right) of HAS with $\mu_{1}\geq\mu_{2}\gamma_{1}>\mu_{2}\gamma_{2}$ under case 2c of Theorem 4.2. Figure 15: Illustrative EAP of HAS (left) and queue length process (right) with $\mu_{1}\geq\mu_{2}\gamma_{2}>\mu_{2}\gamma_{1}$ under case 3a of Theorem 4.2 with the assumption $T^{0}=\tau_{1}^{-1}(T)>0$. Figure 16: Illustrative EAP (left) and resulting queue length process (right) of HAS with $\mu_{1}\geq\mu_{2}\gamma_{2}>\mu_{2}\gamma_{1}$ under case 3b of Theorem 4.2 with the assumption $\tau_{2}(0)<T_{1,idle}$, where $T_{1,idle}$ is the time at which queue 1 empties after the network starts at time zero. Figure 17: Illustrative EAP (left) and queue length process (right) of HAS with $\mu_{1}\geq\mu_{2}\gamma_{2}>\mu_{2}\gamma_{1}$ under case 3c of Theorem 4.2 with the assumption $T_{1,f}>\tau_{2}(0)>T_{1,idle}$, where $T_{1,idle}$ is the time at which queue 1 empties after time zero. ### 4.2 Equal Preferences $\gamma_{1}=\gamma_{2}=\gamma$ We state the main result of this section separately for two regimes as Theorem 4.3 (for $\mu_{1}<\mu_{2}\gamma$), and 4.4 (for $\mu_{1}\geq\mu_{2}\gamma$), since they have different EAP structures. Proofs of these two theorems (in B.2) are similar in structure to that of Theorem 3.2. $T_{1,a},T_{1,f},T_{2,a},T_{2,f}$ are as in the unequal preference case. Let $T_{a}=\inf\mathcal{S}(F_{1})\cup\mathcal{S}(F_{2})$ and $T_{f}=\sup\mathcal{S}(F_{1})\cup\mathcal{S}(F_{2})$. ###### Theorem 4.3. If $\mu_{1}<\mu_{2}\gamma$, 1. 1. when $\Lambda_{1}>\frac{\mu_{1}}{\mu_{2}-\mu_{1}}\Lambda_{2}$, the EAP is unique and has the following arrival rates and support boundaries: $\displaystyle F_{1}^{\prime}(t)$ $\displaystyle=\begin{cases}\mu_{1}~{}~{}&\text{if}~{}~{}t\in[T_{1,a},\tau_{1}^{-1}(T_{2,f})],\\\ \mu_{1}\gamma~{}~{}&\text{if}~{}~{}t\in[\tau_{1}^{-1}(T_{2,f}),T_{1,f}].\end{cases},~{}\text{and}~{}F_{2}^{\prime}(t)=\begin{cases}\mu_{2}\gamma~{}~{}&\text{if}~{}t\in[T_{2,a},0],\\\ \mu_{2}\gamma-\mu_{1}~{}~{}&\text{if}~{}t\in[0,T_{2,f}],\end{cases}$ $\displaystyle\text{where}~{}T_{1,a}$ $\displaystyle=-\frac{1-\gamma}{\gamma}\left(\frac{\Lambda_{1}}{\mu_{1}}-\frac{\Lambda_{2}}{\mu_{2}-\mu_{1}}\right),~{}T_{1,f}=\frac{\Lambda_{1}}{\mu_{1}},~{}T_{2,a}=-\frac{1-\gamma}{\gamma}\frac{\Lambda_{2}}{\mu_{2}-\mu_{1}},~{}\text{and}~{}T_{2,f}=\frac{\Lambda_{2}}{\mu_{2}-\mu_{1}}.$ 2. 2. when $\Lambda_{1}\leq\frac{\mu_{1}}{\mu_{2}-\mu_{1}}\Lambda_{2}$, set of EAPs is given by the convex set of joint customer arrival profiles $\\{F_{1},F_{2}\\}$ satisfying: 1) $F_{1}^{\prime}(t)=0,~{}F_{2}^{\prime}(t)=\mu_{2}\gamma$ for all $t\in[T_{a},0]$, 2) $F_{1}^{\prime}(t)\leq\mu_{1},~{}F_{1}^{\prime}(t)+F_{2}^{\prime}(t)=\mu_{2}\gamma$ for all $t\in[0,T_{f}]$, and 3) $F_{1}(T_{f})=\Lambda_{1},~{}F_{2}(T_{f})=\Lambda_{2}$, where $T_{a}=-\left(\frac{1}{\gamma}-1\right)\frac{\Lambda_{1}+\Lambda_{2}}{\mu_{2}}$ and $T_{f}=\frac{\Lambda_{1}+\Lambda_{2}}{\mu_{2}}$. ###### Remark 8. _If $\Lambda_{1}>\frac{\mu_{1}}{\mu_{2}-\mu_{1}}\Lambda_{2}$, the EAPs in case 1a and 2a of Theorem 4.1, respectively, upon taking limits $\gamma_{1}\to\gamma+,~{}\gamma_{2}=\gamma$ and $\gamma_{1}=\gamma,~{}\gamma_{2}\to\gamma+$ converges to the EAP under case 1 of Theorem 4.3. On the other hand, if $\Lambda_{1}\leq\frac{\mu_{1}}{\mu_{2}-\mu_{1}}\Lambda_{2}$, the EAPs in case 1b and 2b of Theorem 4.1, respectively, upon taking limits $\gamma_{1}\to\gamma+,~{}\gamma_{2}=\gamma$ and $\gamma_{1}=\gamma,~{}\gamma_{2}\to\gamma+$ might converge to different limits, but both the limits will be contained in the convex set of EAPs under case 2 of 4.1._ The quantity $T\overset{def.}{=}\inf\left\\{t>0~{}|~{}Q_{2}(t)=0\right\\}$ will be necessary to describe the structure of the EAP in the regime $\mu_{1}\leq\mu_{2}\gamma$. ###### Theorem 4.4. If $\mu_{1}\geq\mu_{2}\gamma$, 1. 1. when $\left(\frac{\mu_{2}}{\mu_{1}}-1\right)\Lambda_{1}>\Lambda_{2}$, the EAP is unique and has the following arrival rates and support boundaries: $\displaystyle F_{1}^{\prime}(t)$ $\displaystyle=\begin{cases}\mu_{2}\gamma~{}~{}&\text{if}~{}t\in[T_{1,a},\tau_{1}^{-1}(T)],\\\ \mu_{1}\gamma~{}~{}&\text{if}~{}t\in[\tau_{1}^{-1}(T),T_{1,f}],\end{cases},~{}\text{and}~{}F_{2}^{\prime}(t)=\mu_{2}\gamma~{}~{}\text{if}~{}t\in\left[-\frac{\Lambda_{2}}{\mu_{2}\gamma},0\right],$ $\displaystyle\text{where}~{}T_{1,a}$ $\displaystyle=\frac{\Lambda_{2}}{\mu_{2}\gamma}-\left(\frac{1}{\gamma}-1\right)\frac{\Lambda_{1}}{\mu_{1}},~{}T=\frac{\Lambda_{2}}{\mu_{2}-\mu_{1}},~{}\text{and}~{}T_{1,f}=\frac{\Lambda_{1}}{\mu_{1}}.$ 2. 2. when $\left(\frac{1}{\gamma}-1\right)\Lambda_{1}\geq\Lambda_{2}\geq\left(\frac{\mu_{2}}{\mu_{1}}-1\right)\Lambda_{1}$, the EAP is unique and has the following arrival rates and support boundaries: $\displaystyle F_{1}^{\prime}(t)$ $\displaystyle=\mu_{2}\gamma~{}~{}\text{if}~{}t\in[T_{1,a},T_{1,f}],~{}\text{and}~{}F_{2}^{\prime}(t)=\mu_{2}\gamma~{}~{}\text{if}~{}t\in\left[-\frac{\Lambda_{2}}{\mu_{2}\gamma},0\right],$ $\displaystyle\text{where}~{}T_{1,a}$ $\displaystyle=\frac{1}{\mu_{2}}\left(\Lambda_{2}-\left(\frac{1}{\gamma}-1\right)\Lambda_{1}\right),~{}\text{and}~{}T=T_{1,f}=\frac{\Lambda_{1}+\Lambda_{2}}{\mu_{2}}.$ 3. 3. If $\Lambda_{2}>\left(\frac{1}{\gamma}-1\right)\Lambda_{1}$, set of EAPs is given by the convex set of joint customer arrival profiles $\\{F_{1},F_{2}\\}$ satisfying: 1) $F_{1}^{\prime}(t)=0$ and $F_{2}^{\prime}(t)=\mu_{2}\gamma$ for all $t\in[T_{a},0]$, 2) $F_{1}^{\prime}(t)+F_{2}^{\prime}(t)=\mu_{2}\gamma$ for all $t\in[0,T_{f}]$, and 3) $F_{1}(T_{f})=\Lambda_{1},~{}F_{2}(T_{f})=\Lambda_{2}$, where $T_{a}=-\left(\frac{1}{\gamma}-1\right)\frac{\Lambda_{1}+\Lambda_{2}}{\mu_{2}}$ and $T_{f}=\frac{\Lambda_{1}+\Lambda_{2}}{\mu_{2}}$. ###### Remark 9. _If $\left(\frac{\mu_{2}}{\mu_{1}}-1\right)\Lambda_{1}>\Lambda_{2}$, the EAPs in case 2a and 3a of Theorem 4.2, respectively upon taking limits $\gamma_{1}\to\gamma+,\gamma_{2}=\gamma$ and $\gamma_{1}=\gamma,\gamma_{2}\to\gamma+$, converges to the EAP under case 1 of Theorem 4.4. If $\left(\frac{1}{\gamma}-1\right)\Lambda_{1}\geq\Lambda_{2}\geq\left(\frac{\mu_{2}}{\mu_{1}}-1\right)\Lambda_{1}$, upon taking the same limits, the EAPs in case 2b and 3b of Theorem 4.2 converges to the EAP under case 2 of Theorem 4.4. If $\Lambda_{2}>\left(\frac{1}{\gamma}-1\right)\Lambda_{1}$, upon taking the same limits, the EAPs in case 2c and 3c of Theorem 4.2 converges to different limits and both of them are contained in the convex set of EAPs under case 3 of Theorem 4.4._ ## 5 Conclusion Arrival games to a single queue have been well studied in the literature, but the important area of arrival games to a general queuing network, has limited literature. As we discovered, this may be because the problem complexity may increase substantially with the network size. We studied a simple two queue, two class network with a linear cost structure, where we analyzed in detail two heterogeneous routing configurations: (1) where the customer classes arrived at the same queue but departed in different queues, and (2) where the customer classes arrived at different queues but departed from the same queue. We discovered non-trivial customer behaviour not apparent in a single queue or in networks studied thus far in the literature. In a specific setting we found that the parameter space had to be partitioned into eight distinct regions, where each region had its own closed form parametric representation of the arrival equilibrium profile. We found that although for most set of parameters, the equilibrium profile is unique, there exists settings where the collection of equilibrium profiles is not unique but a convex set. While in a single queue, multi-class customers arrive in contiguous, non-overlapping intervals, in our two queue setting there are regions where, in equilibrium, different class arrival times may overlap. Further, there exist regions, where a single class customer may arrive in disjoint intervals. The broad message of the paper is mixed and motivates further research - it suggests that even for more complex networks one may expect a unique equilibrium profile for most parameter settings. It also suggests that the number of different solution structures may blow up with the network size, so that learning the structure in any instance may be a difficult task. ## Acknowledgement Our work is supported by Department of Atomic Energy, Government of India, under project no. RTI4001. ## References * [1] Avrim Blum and Yishay Mansour. Learning, Regret Minimization, and Equilibria, page 79–102. Cambridge University Press, 2007. * [2] Jesper Breinbjerg. Equilibrium arrival times to queues with general service times and non-linear utility functions. European Journal of Operational Research, 261(2):595–605, 2017\. * [3] Nicolo Cesa-Bianchi and Gábor Lugosi. Prediction, learning, and games. Cambridge university press, 2006. * [4] Hong Chen and David D Yao. Fundamentals of queueing networks: Performance, asymptotics, and optimization, volume 4. Springer, 2001. * [5] Amihai Glazer and Refael Hassin. ?/m/1: On the equilibrium distribution of customer arrivals. European Journal of Operational Research, 13(2):146–150, 1983. * [6] Refael Hassin. Rational Queueing. Chapman and Hall/CRC, 2016. * [7] Refael Hassin and Moshe Haviv. To queue or not to queue: Equilibrium behavior in queueing systems, volume 59. Springer Science & Business Media, 2003. * [8] Moshe Haviv and Binyamin Oz. A busy period approach to some queueing games. Queueing Systems, 97:261–277, 2021. * [9] Moshe Haviv and Liron Ravner. A survey of queueing systems with strategic timing of arrivals. Queueing Systems, 99:163–198, 2021. * [10] Chris Hendrickson and George Kocur. Schedule delay and departure time decisions in a deterministic model. Transportation Science, 15(1):62–77, 1981. * [11] Harsha Honnappa and Rahul Jain. Strategic arrivals into queueing networks: The network concert queueing game. Operations Research, 63(1):247–259, 2015. * [12] Rahul Jain, Sandeep Juneja, and Nahum Shimkin. The concert queueing game: to wait or to be late. Discrete Event Dynamic Systems, 21:103–138, 2011. * [13] Sandeep Juneja and Nahum Shimkin. The concert queueing game: strategic arrivals with waiting and tardiness costs. Queueing Systems, 74:369–402, 2013. * [14] Zhi-Chun Li, Hai-Jun Huang, and Hai Yang. Fifty years of the bottleneck model: A bibliometric review and future research directions. Transportation research part B: methodological, 139:311–342, 2020\. * [15] Robin Lindsey. Existence, uniqueness, and trip cost function properties of user equilibrium in the bottleneck model with multiple user classes. Transportation Science, 38:293–314, 2004. * [16] William S. Vickrey. Congestion theory and transport investment. The American Economic Review, 59(2):251–260, 1969. ## Appendix For the ease of analysis, we define the set $E_{j}\overset{def.}{=}\\{t~{}|~{}t<0~{}\text{or}~{}Q_{j}(t)>0\\}$ to be the set of all times when queue $j$ has a positive waiting time for $j=1,2$. For any set $S\subseteq\mathbb{R}$, we use the notations $\overline{S}$ and $S^{o}$ to respectively denote the closure and interior of $S$. By Lemma 2.1, we consider only absolutely continuous arrival profiles $F_{1}$ and $F_{2}$ and as a result, $\mathcal{S}(F_{1})$ and $\mathcal{S}(F_{2})$ cannot have isolated points. ## Appendix A Proofs of Lemmas in Section 3 As argued in Section 3, we assume $\mu_{1}>\mu_{2}$ for the following discussion. ### A.1 Proofs of Lemmas in Section 3.1 (HDS with $\gamma_{1}\neq\gamma_{2}$) Several of the lemmas proved below will extend to the case where $\gamma_{1}=\gamma_{2}$. So, whenever some lemma is applicable only for the case of unequal preferences, we mention it explicitly in the lemma statement. Lemma A.1 specifies the state of queue 2 when users of both classes are arriving together in queue 1 and is a special case of Lemma 3.3. This lemma helps us prove Lemma 3.1 and together with Lemma A.2, it will help us calculate the arrival rate of the two classes in equilibrium. ###### Lemma A.1. If $\gamma_{1}\neq\gamma_{2}$ and $\mu_{1}>\mu_{2}$, in the EAP, $t\in(\mathcal{S}(F_{1}))^{o}\cap(\mathcal{S}(F_{2}))^{o}$ implies $\tau_{1}(t)\in\overline{E_{2}}$. ###### Proof. Assuming contradiction, i.e., $\tau_{1}(t)\notin\overline{E_{2}}$, there is a neighbourhood $[\tau_{1}(t)-\delta,\tau_{1}(t)+\delta]$ of $\tau_{1}(t)$, where queue 2 is empty. Since the arrival profiles $F_{1},~{}F_{2}$ are absolutely continuous, $\tau_{1}(\cdot)$ will be continuous. As a result, we have $[t-\epsilon,t+\epsilon]\subseteq(\mathcal{S}(F_{1}))^{o}\cap(\mathcal{S}(F_{2}))^{o}$ such that, $Q_{2}(\tau_{1}(s))=0$ for all $s\in[t-\epsilon,t+\epsilon]$. Hence every class 2 user arriving in $[t-\epsilon,t+\epsilon]$ waits only at queue 1. For both the classes $i=1,2$ departure time $\tau_{\mathbf{F}}^{(i)}(s)=\tau_{1}(s)$ in $[t-\epsilon,t+\epsilon]$ and hence cost $C_{\mathbf{F}}^{(i)}(s)=\tau_{1}(s)-\gamma_{i}s$. Since, $[t-\epsilon,t+\epsilon]\subseteq(\mathcal{S}(F_{1}))^{o}\cap(\mathcal{S}(F_{2}))^{o}$, by definition of EAP, for $i=1,2$ $C_{\mathbf{F}}^{(i)}(\cdot)$ is constant in $[t-\epsilon,t+\epsilon]$. Therefore, $(C_{\mathbf{F}}^{(i)})^{\prime}(s)=\tau_{1}^{\prime}(s)-\gamma_{i}=0$ in $[t-\epsilon,t+\epsilon]$, giving us $\tau_{1}^{\prime}(s)=\gamma_{1}=\gamma_{2}$, contradicting the assumption $\gamma_{1}\neq\gamma_{2}$. ∎ Lemma A.2 helps us relate the arrival rate of class 2 users and will be referred to ubiquitously while proving the structural properties. ###### Lemma A.2. For every absolutely continuous joint user arrival profile $\\{F_{1},F_{2}\\}$, $\tau_{\mathbf{F}}^{(2)}(\cdot)$ is differentiable a.e. in the set of times $\tau_{1}^{-1}(\overline{E_{2}})\overset{def.}{=}\\{t~{}|~{}\tau_{1}(t)\in\overline{E_{2}}\\}$ with derivative $(\tau_{\mathbf{F}}^{(2)})^{\prime}(t)=\frac{F_{2}^{\prime}(t)}{\mu_{2}}$ a.e. As a consequence, $(C_{\mathbf{F}}^{(2)})^{\prime}(t)=\frac{F_{2}^{\prime}(t)}{\mu_{2}}-\gamma_{2}$ a.e. in $\tau_{1}^{-1}(\overline{E_{2}})$. ###### Proof. We divide the proof into two separate cases: 1. 1. In $\overline{E_{1}}^{c}$, we will have $\tau_{1}(t)=t$ and as a result $\overline{E_{1}}^{c}\cap\tau_{1}^{-1}(\overline{E_{2}})=\overline{E_{1}}^{c}\cap\overline{E_{2}}$. Therefore, using (3) $(\tau_{\mathbf{F}}^{(2)})^{\prime}(t)=\tau_{2}^{\prime}(t)=\frac{F_{2}^{\prime}(t)}{\mu_{2}}$ a.e. in $\overline{E_{1}}^{c}\cap\tau_{1}^{-1}(\overline{E_{2}})$. 2. 2. In $\overline{E_{1}}$, by (3), $\tau_{1}^{\prime}(\cdot)$ exists a.e. Again using (3), since $\tau_{1}(t)\in\overline{E_{2}}$, we have $\tau_{2}^{\prime}(\tau_{1}(t))=\frac{A_{2}^{\prime}(\tau_{1}(t))}{\mu_{2}}$. Therefore, $(\tau_{\mathbf{F}}^{(2)})^{\prime}(t)=\frac{A_{2}^{\prime}(\tau_{1}(t))\tau_{1}^{\prime}(t)}{\mu_{2}}$ a.e. in $\overline{E_{1}}\cap\tau_{1}^{-1}(\overline{E_{2}})$. Since $A_{2}(\tau_{1}(t))=F_{2}(t)$, by chain rule, $(\tau_{\mathbf{F}}^{(2)})^{\prime}(t)=\frac{F_{2}^{\prime}(t)}{\mu_{2}}$ a.e. in $\overline{E_{1}}\cap\tau_{1}^{-1}(\overline{E_{2}})$. Since $C_{\mathbf{F}}^{(2)}(t)=\tau_{\mathbf{F}}^{(2)}(t)-\gamma_{2}t$, the second statement of the lemma follows trivially. ∎ Lemma B.5 implies that $\mu_{1}\leq\mu_{2}\cdot\max\left\\{1,\frac{\gamma_{2}}{\gamma_{1}}\right\\}$ is sufficient for $\mathcal{S}(F_{1})$ and $(\mathcal{S}(F_{2}))^{o}$ to be disjoint, and is helpful for proving sufficiency of the condition in Lemma 3.1. ###### Lemma A.3. If $\mu_{1}\gamma_{1}\leq\mu_{2}\gamma_{2}$, in every EAP, $\mathcal{S}(F_{1})$ and $(\mathcal{S}(F_{2}))^{o}$ are disjoint. ###### Proof. Proof of Lemma A.3 follows the argument mentioned after Lemma 3.1 for proving sufficiency of $\mu_{1}\leq\mu_{2}\cdot\max\left\\{1,\frac{\gamma_{2}}{\gamma_{1}}\right\\}$ for $\mathcal{S}(F_{1}),\mathbf{S}(F_{2})$ to be non-overlapping, except while proving $(C_{\mathbf{F}}^{(2)})^{\prime}(t)=\frac{F_{2}^{\prime}(t)}{\mu_{2}}-\gamma_{2}$ in $[t_{1},t_{2}]$, we use Lemma B.1 to argue queue 2 stays engaged at $\tau_{1}(t)$ and then use Lemma A.2 to have $(\tau_{\mathbf{F}}^{(2)})^{\prime}(t)=\frac{F_{2}^{\prime}(t)}{\mu_{2}}$ in $[t_{1},t_{2}]$. ∎ As we stated in the main body, $T_{i,a}=\inf\mathcal{S}(F_{i})$ and $T_{i,f}=\sup\mathcal{S}(F_{i})$ for the two classes $i=1,2$. Note that $T_{1,a},T_{1,f},T_{2,a},T_{2,f}$ are all finite if $\mathbf{F}=\\{F_{1},F_{2}\\}$ is an EAP. Lemma A.4 and A.5 are respectively the first and second statements of Lemma 3.5. ###### Lemma A.4. (First statement of Lemma 3.5) In the EAP, if $\mathcal{S}(F_{2})=[T_{2,a},T_{2,f}]$, then queue 2 will have zero waiting time at $\tau_{1}(T_{2,f})$. ###### Proof. Assume contradiction, i.e., $\tau_{1}(T_{2,f})\in E_{2}$. For $\delta>0$ chosen sufficiently small, image of $[T_{2,f},T_{2,f}+\delta]$ under $\tau_{1}(\cdot)$ is contained in $E_{2}$. Therefore by Lemma A.2, $(\tau_{\mathbf{F}}^{(2)})^{\prime}(t)=\frac{F_{2}^{\prime}(t)}{\mu_{2}}$ in $[T_{2,f},T_{2,f}+\delta]$. Since $F_{2}^{\prime}(t)=0$ in $(T_{2,f},T_{2,f}+\delta]$, we get $\tau_{\mathbf{F}}^{(2)}(T_{2,f}+\delta)=\tau_{\mathbf{F}}^{(2)}(T_{2,f})$. As a result, the class 2 user arriving at $T_{2,f}$ will be strictly better off arriving at $T_{2,f}+\delta$, contradicting that $\\{F_{1},F_{2}\\}$ is an EAP. ∎ We now prove below Lemma 3.2 about the structure of the supports of the two classes in EAP. Proof of Lemma 3.2: (1) On assuming contradiction, there exists $t_{1},t_{2}\in\mathcal{S}(F_{1})\cup\mathcal{S}(F_{2})$ such that $t_{2}>t_{1}$ and $F_{1}(t_{1})+F_{2}(t_{1})=F_{1}(t_{2})+F_{2}(t_{2})$. Note that $t_{2}<\max\\{T_{1,f},T_{2,f}\\}$. Otherwise, $\max\\{T_{1,f},T_{2,f}\\}$ will be an isolated point of $\mathcal{S}(F_{1})\cup\mathcal{S}(F_{2})$, contradicting absolute continuity of $F_{1}$ and $F_{2}$. With $t_{2}<\max\\{T_{1,f},T_{2,f}\\}$, a positive mass of users must arrive after $t_{2}$. Therefore, none of the classes can have zero waiting time in $[t_{1},t_{2}]$, implying $[t_{1},t_{2}]\subseteq E_{1}\cup E_{2}$. There can be two situations now: * • If $t_{1}\in E_{1}$: For $\delta>0$ picked sufficiently small, $[t_{1},t_{1}+\delta]\subseteq E_{1}$. Since $A_{1}^{\prime}(s)=0$ for every $s\in[t_{1},t_{1}+\delta]$, by (3), $\tau_{1}(t_{1})=\tau_{1}(t_{1}+\delta)$. As a result, for both classes $i=1,2$ $\tau_{\mathbf{F}}^{(i)}(t_{1})=\tau_{\mathbf{F}}^{(i)}(t_{1}+\delta)$. Therefore, the user arriving at $t_{1}$ will be strictly better off arriving at $t_{1}+\delta$. * • If $t_{1}\notin E_{1}$: Since no users arrive in $[t_{1},t_{2}]$, queue 1 has zero waiting time in $[t_{1},t_{2}]$ (by (3)) implying, $[t_{1},t_{2}]\subseteq E_{2}$. Therefore some positive mass of class 2 users must have arrived before $t_{1}$. Let $\tilde{t}=\sup\\{t\leq t_{1}~{}|~{}t\in\mathcal{S}(F_{2})\\}=$ the last time before $t_{1}$ some class 2 user has arrived. Since no class 2 users arrives in $[\tilde{t},t_{2}]$, $Q_{2}(\cdot)$ cannot increase in $[\tau_{1}(\tilde{t}),\tau_{1}(t_{2})]$. Therefore, $t_{2}=\tau_{1}(t_{2})\in E_{2}$ implies $[\tau_{1}(\tilde{t}),\tau_{1}(t_{2})]\subseteq E_{2}$. With this observation, since $F_{2}^{\prime}(s)=0$ in $(\tilde{t},t_{2})$, we conclude $\tau_{\mathbf{F}}^{(2)}(\tilde{t})=\tau_{\mathbf{F}}^{(2)}(t_{2})$ (by Lemma A.2). Therefore, the class 2 user arriving at $\tilde{t}$ will be strictly better off arriving at $t_{2}$ instead. In both cases, $\mathbf{F}=\\{F_{1},F_{2}\\}$ will not be an EAP, contradicting our assumption. (2) On assuming contradiction, there exists $t_{1},t_{2}\in\mathcal{S}(F_{2})$ such that $F_{2}(t_{1})=F_{2}(t_{2})$. We must have $t_{2}<T_{2,f}$, otherwise, $T_{2,f}$ will be an isolated point in $\mathcal{S}(F_{2})$. Since $\mathcal{S}(F_{1})\cup\mathcal{S}(F_{2})$ is an interval, we must have $[t_{1},t_{2}]\subseteq\mathcal{S}(F_{1})$. Two situations might arise: * • If $\tau_{1}(t_{1})\in E_{2}$: By continuity of $\tau_{1}(\cdot)$, we can find $\delta\in(0,t_{2}-t_{1})$ sufficiently small such that image of $[t_{1},t_{1}+\delta]$ under $\tau_{1}(\cdot)$ is contained in $E_{2}$. Since $F_{2}^{\prime}(t)=0$ in $[t_{1},t_{2}]$, by Lemma A.2, the previous statement implies $\tau_{\mathbf{F}}^{(2)}(t_{1})=\tau_{\mathbf{F}}^{(2)}(t_{1}+\delta)$. Therefore the class 2 user arriving at $t_{1}$ will be strictly better off arriving at $t_{1}+\delta$. * • If $\tau_{1}(t_{1})\notin E_{2}$: With no class 2 user arriving at queue 2 in $[\tau_{1}(t_{1}),\tau_{1}(t_{2})]$, we have $[\tau_{1}(t_{1}),\tau_{1}(t_{2})]\subseteq E_{2}^{c}$. Therefore for $t\in[t_{1},t_{2}]$, $\tau_{\mathbf{F}}^{(1)}(t)=\tau_{\mathbf{F}}^{(2)}(t)=\tau_{1}(t)$. Now $[t_{1},t_{2}]\subseteq\mathcal{S}(F_{1})$ implies $C_{\mathbf{F}}^{(1)}(\cdot)$ is constant in $[t_{1},t_{2}]$. Hence $(C_{\mathbf{F}}^{(1)})^{\prime}(t)=\tau_{1}^{\prime}(t)-\gamma_{1}=0$ in $(t_{1},t_{2})$, implying $\tau_{1}^{\prime}(t)=\gamma_{1}$. As a result, in $(t_{1},t_{2})$, $(C_{\mathbf{F}}^{(2)})^{\prime}(t)=\tau_{1}^{\prime}(t)-\gamma_{2}=\gamma_{1}-\gamma_{2}\neq 0$. Now if $\gamma_{1}>\gamma_{2}$, class 2 user arriving at time $t_{2}$ will be better off arriving at time $t_{1}$ and vice-versa when $\gamma_{1}<\gamma_{2}$. $\\{F_{1},F_{2}\\}$ cannot be an EAP in both the situations. (3) We consider two separate situations for proving $\mathcal{S}(F_{1})$ is an interval: * • If $\mu_{1}\gamma_{1}\neq\mu_{2}\gamma_{2}$: On assuming contradiction, there exists $t_{1},t_{2}\in\mathcal{S}(F_{1})$ such that, $F_{1}(t_{1})=F_{1}(t_{2})$. We must have $t_{2}<T_{1,f}$, otherwise, $T_{1,f}$ will be an isolated point in $\mathcal{S}(F_{1})$. Since $\mathcal{S}(F_{1})\cup\mathcal{S}(F_{2})$ is an interval, we must have $[t_{1},t_{2}]\subseteq\mathcal{S}(F_{2})$. Also, we must have $[t_{1},t_{2}]\subseteq E_{1}$, since a positive mass of class 1 users arrive after $t_{2}$. This implies, class 2 users will be arriving at queue 2 at a rate $\mu_{1}>\mu_{2}$ in $[\tau_{1}(t_{1}),\tau_{1}(t_{2})]$ and therefore $[\tau_{1}(t_{1}),\tau_{1}(t_{2})]\subseteq\overline{E_{2}}$. Using Lemma A.2 and the definition of EAP, in $[t_{1},t_{2}]$, $(C_{\mathbf{F}}^{(2)})^{\prime}(t)=(\tau_{\mathbf{F}}^{(2)})^{\prime}(t)-\gamma_{2}=\frac{F_{2}^{\prime}(t)}{\mu_{2}}-\gamma_{2}=0$ a.e. implying $F_{2}^{\prime}(t)=\mu_{2}\gamma_{2}$ a.e. Since $[t_{1},t_{2}]\subseteq E_{1}$, by (3), $\tau_{1}^{\prime}(t)=\frac{F_{2}^{\prime}(t)}{\mu_{1}}=\frac{\mu_{2}\gamma_{2}}{\mu_{1}}$ a.e. in $[t_{1},t_{2}]$. As a result, $(C_{\mathbf{F}}^{(1)})^{\prime}(t)=\tau_{1}^{\prime}(t)-\gamma_{1}=\frac{\mu_{2}\gamma_{2}}{\mu_{1}}-\gamma_{1}$ a.e. in $[t_{1},t_{2}]$. Now if $\mu_{1}\gamma_{1}>\mu_{2}\gamma_{2}$, class 1 user arriving at time $t_{1}$ will be strictly better off arriving at time $t_{2}$ and vice-versa if $\mu_{1}\gamma_{1}<\mu_{2}\gamma_{2}$. Therefore $\\{F_{1},F_{2}\\}$ will not be an EAP. * • If $\mu_{1}\gamma_{1}=\mu_{2}\gamma_{2}$: Since $\mu_{1}>\mu_{2}$, $\mu_{1}\gamma_{1}=\mu_{2}\gamma_{2}$ implies $\gamma_{1}<\gamma_{2}$. By 2nd statement of the lemma, $\mathcal{S}(F_{2})$ must be the interval $[T_{2,a},T_{2,f}]$. By Lemma A.3, $\mathcal{S}(F_{1})$ and $(T_{2,a},T_{2,f})$ are disjoint and their union is an interval. As a result, $\mathcal{S}(F_{1})$ is union of two intervals, one ending at $T_{2,a}$ and the other one starting from $T_{2,f}$. We now show that $\mathcal{S}(F_{1})$ has no element larger than $T_{2,f}$. This implies that $\mathcal{S}(F_{1})$ is an interval ending at $T_{2,a}$. For this, we again assume a contradiction and suppose that there exists $T>T_{2,f}$ such that $[T_{2,f},T]\subseteq\mathcal{S}(F_{1})$. By Lemma A.4, queue 2 must be empty at $\tau_{1}(T_{2,f})$. Therefore for $t\in[T_{2,f},T]$, $\tau_{\mathbf{F}}^{(1)}(t)=\tau_{\mathbf{F}}^{(2)}(t)=\tau_{1}(t)$. Since $[T_{2,f},T]\subseteq\mathcal{S}(F_{1})$, we will have $(C_{\mathbf{F}}^{(1)})^{\prime}(t)=\tau_{1}^{\prime}(t)-\gamma_{1}=0$ implying $\tau_{1}^{\prime}(t)=\gamma_{1}$ in $[T_{2,f},T]$. Using this, in $[T_{2,f},T]$, $(C_{\mathbf{F}}^{(2)})^{\prime}(t)=\tau_{1}^{\prime}(t)-\gamma_{2}=\gamma_{1}-\gamma_{2}<0$. Hence, the class 2 user arriving at $T_{2,f}$ is strictly better off arriving at $T$, implying that $\\{F_{1},F_{2}\\}$ is not an EAP. ∎ By Lemma 3.2, for every EAP $\\{F_{1},F_{2}\\}$, we must have $\mathcal{S}(F_{1})=[T_{1,a},T_{1,f}]$, $\mathcal{S}(F_{2})=[T_{2,a},T_{2,f}]$ and their union must be an interval. Proof of Lemma 3.3: If $t\in(T_{1,a},T_{1,f})\cap(T_{2,a},T_{2,f})$, the statement follows from Lemma A.1. Otherwise, we can obtain $\delta>0$ such that either $[t,t+\delta]$ or $[t-\delta,t]\subseteq(T_{2,a},T_{2,f})/(T_{1,a},T_{1,f})$. Now if $t\in E_{1}$, we can choose $\delta>0$ small enough such that $[t-\delta,t+\delta]\subseteq E_{1}$ As a result, $\tau_{1}(t-\delta)<\tau_{1}(t)<\tau_{1}(t+\delta)$ (by (3)) and class 2 users arrive in queue 2 at a rate $\mu_{1}>\mu_{2}$ in $[\tau_{1}(t),\tau_{1}(t+\delta)]$ or $[\tau_{1}(t-\delta),\tau_{1}(t)]$ implying that $\tau_{1}(t)\in\overline{E_{2}}$. Otherwise, if $t\in E_{1}^{c}$, since a positive mass of class 2 users arrive in $(t,T_{2,f}]$, we must have $t=\tau_{1}(t)\in E_{2}$. Therefore, the lemma stands proved. ∎ Proof of Lemma 3.4: By Lemma 3.3, if $t\in\mathcal{S}(F_{2})$, $\tau_{1}(t)\in\overline{E_{2}}$, which by Lemma A.2 implies $(C_{\mathbf{F}}^{(2)})^{\prime}(t)=\frac{F_{2}^{\prime}(t)}{\mu_{2}}-\gamma_{2}$ a.e. in $\mathcal{S}(F_{2})$. Since $C_{\mathbf{F}}^{(2)}(\cdot)$ is constant in $\mathcal{S}(F_{2})$, we must have $F_{2}^{\prime}(t)=\mu_{2}\gamma_{2}$ a.e. in $\mathcal{S}(F_{2})$. Note that $t\in\mathcal{S}(F_{1})=[T_{1,a},T_{1,f}]$ implies $t\in\overline{E_{1}}$. Therefore, by (3), $(C_{\mathbf{F}}^{(1)})^{\prime}(t)=(\tau_{\mathbf{F}}^{(1)})^{\prime}(t)-\gamma_{1}=\frac{F_{1}^{\prime}(t)+F_{2}^{\prime}(t)}{\mu_{1}}-\gamma_{1}$ a.e. in $\mathcal{S}(F_{1})$. Since $C_{\mathbf{F}}^{(1)}(\cdot)$ is constant in $\mathcal{S}(F_{1})$, we must have $F_{1}^{\prime}(t)+F_{2}^{\prime}(t)=\mu_{1}\gamma_{1}$ a.e. in $\mathcal{S}(F_{1})$. As $F_{2}^{\prime}(t)=\mu_{2}\gamma_{2}$ a.e. in $\mathcal{S}(F_{2})$, we must have, $\displaystyle F_{1}^{\prime}(t)$ $\displaystyle=\begin{cases}\mu_{1}\gamma_{1}~{}~{}&\text{if}~{}t\in\mathcal{S}(F_{1})/\mathcal{S}(F_{2}),~{}\text{and},\\\ \mu_{1}\gamma_{1}-\mu_{2}\gamma_{2}~{}~{}&\text{if}~{}t\in\mathcal{S}(F_{1})\cap\mathcal{S}(F_{2}).\end{cases}$ Note that, the above arrival rates are positive a.e. because, by Lemma A.3 and 3.2, if $\mu_{1}\gamma_{1}\leq\mu_{2}\gamma_{2}$, $\mathcal{S}(F_{1})\cap\mathcal{S}(F_{2})$ has zero Lebesgue measure. ∎ ###### Lemma A.5. (Second statement of Lemma 3.5) In the EAP, if $\mu_{1}\gamma_{1}>\mu_{2}\gamma_{2}$ and $\mathcal{S}(F_{1})=[T_{1,a},T_{1,f}]$, then queue 1 will have zero waiting time at $T_{1,f}$. ###### Proof. Assume a contradiction, i.e., $T_{1,f}\in E_{1}$ with $\mu_{1}\gamma_{1}>\mu_{2}\gamma_{2}$. Two situations are possible: * • If $(T_{1,f},\infty)\cap\mathcal{S}(F_{2})=\emptyset$, by (3), $\tau_{\mathbf{F}}^{(1)}(T_{1,f})=\tau_{\mathbf{F}}^{(1)}(T_{idle})$, where $T_{idle}>T_{1,f}$ is the time queue 1 empties after $T_{1,f}$. Hence, the class 1 user arriving at $T_{1,f}$ is strictly better off arriving at $T_{idle}$. * • If a positive mass of class 2 users arrive after $T_{1,f}$, by Lemma 3.2 and 3.4, they will arrive over an interval $[T_{1,f},T]$ at rate $\mu_{2}\gamma_{2}$, where $T>T_{1,f}$. Therefore, picking $\delta\in(0,T-T_{1,f})$ sufficiently small, we can have $[T_{1,f},T_{1,f}+\delta]\subseteq E_{1}$. By (3), for $t\in[T_{1,f},T_{1,f}+\delta]$, $(C_{\mathbf{F}}^{(1)})^{\prime}(t)=\tau_{1}^{\prime}(t)-\gamma_{1}=\frac{F_{2}^{\prime}(t)}{\mu_{1}}-\gamma_{1}=\frac{\mu_{2}\gamma_{2}}{\mu_{1}}-\gamma_{1}<0$ a.e. Hence the class 1 user arriving at $T_{1,f}$ will be strictly better off arriving at $T_{1,f}+\delta$. For both the situations above, we get $\\{F_{1},F_{2}\\}$ will not be an EAP. ∎ Proof of Lemma 3.5: Lemma 3.5 follows from Lemma A.4 and A.5. ∎ Proof of Lemma 3.1: By Lemma A.3 and 3.2, we get $\mu_{1}\leq\mu_{2}\cdot\max\\{1,\frac{\gamma_{2}}{\gamma_{1}}\\}$ is sufficient for $[T_{1,a},T_{1,f}]$ and $[T_{2,a},T_{2,f}]$ to have disjoint interiors. Proving that the condition is necessary follows from the argument after Lemma 3.5 in Section 3.1. Hence Lemma 3.1 stands proved. ∎ Proof of Lemma 3.6: (1) If $T_{1,f}=T_{2,a}$, queue 1 is empty at $T_{1,f}$ (by Lemma 3.5), making the network empty at $T_{1,f}$. As a result, every class 2 user will be strictly better off arriving at $T_{1,f}$ and $\mathbf{F}$ will not be an EAP. Therefore, the only other possibility is $T_{2,f}=T_{1,a}$. (2) On assuming a contradiction, if $T_{2,f}<T_{1,f}$, by Lemma 3.1, we must have $T_{2,f}\in(T_{1,a},T_{1,f})$. Since group 1 users will arrive in $[T_{2,f},T_{1,f}]$, we must have $[T_{2,f},T_{1,f}]\subseteq\overline{E_{1}}$. Using Lemma A.4, queue 2 has a zero waiting time for $t\geq\tau_{1}(T_{2,f})$ and as a consequence $\tau_{\mathbf{F}}^{(2)}(t)=\tau_{1}(t)$ for $t\geq T_{2,f}$. Therefore, for $t\in[T_{2,f},T_{1,f}]$, $\displaystyle(C^{(2)}_{\mathbf{F}})^{\prime}(t)$ $\displaystyle=(\tau_{F}^{(2)})^{\prime}(t)-\gamma_{2}=\tau_{1}^{\prime}(t)-\gamma_{2}$ $\displaystyle=\frac{F_{1}^{\prime}(t)}{\mu_{1}}-\gamma_{2}~{}~{}$ (using (3) and the fact that $[T_{2,f},T_{1,f}]\in\overline{E_{1}}$) $\displaystyle=\gamma_{1}-\gamma_{2}<0.~{}~{}$ (using Lemma 3.4). Thus, the group 2 user arriving at $T_{2,f}$ will be strictly better off by arriving at $T_{1,f}$. Hence $\mathbf{F}$ will not be an EAP. (3) Let $T=\min\left\\{T_{1,f},T_{2,f}\right\\}$. By Lemma 3.1, $\max\\{T_{1,a},T_{2,a}\\}<T$. If we assume contradiction to the statement, we must have $T_{1,a}\leq T_{2,a}<T$. As a result, class 2 users are arriving in $[T_{2,a},T]$ alongside class 1 users. Using Lemma 3.4, for every $t\in[T_{2,a},T]$, $\displaystyle F_{2}^{\prime}(t)$ $\displaystyle=\mu_{2}\gamma_{2},~{}\text{and},$ $\displaystyle F_{1}^{\prime}(t)$ $\displaystyle=\mu_{1}\gamma_{1}-\mu_{2}\gamma_{2}.$ Moreover, for $t\in[T_{2,a},T]$, we have $A_{2}(\tau_{1}(t))=F_{2}(t)$. Again $[T_{2,a},T]\subseteq\mathcal{S}(F_{1})$ implies $[T_{2,a},T]\subseteq\overline{E_{1}}$ and therefore by (3) $\tau_{1}^{\prime}(t)=\frac{F_{1}^{\prime}(t)+F_{2}^{\prime}(t)}{\mu_{1}}=\gamma_{1}$ a.e. in $[T_{2,a},T]$. Hence, $A_{2}^{\prime}(\tau_{1}(t))=\frac{F_{2}^{\prime}(t)}{\tau_{1}^{\prime}(t)}=\mu_{2}\frac{\gamma_{2}}{\gamma_{1}}<\mu_{2}$ a.e. in $[T_{2,a},T]$. This implies that class 2 users are arriving at queue 2 at a rate $<\mu_{2}$ from the time when queue 2 starts serving them, i.e., $\tau_{1}(T_{2,a})$. Therefore $Q_{2}(t)=0$ for all $t\in[\tau_{1}(T_{2,a}),\tau_{1}(T)]$, which contradicts Lemma A.1. ∎ #### A.1.1 Proof of Theorem 3.1 Below we mention those parts of the proof of Theorem 3.1 which were not covered in the proof sketch mentioned in Section 3.1. Of the skipped portion is the proof of existence and uniqueness of EAP for case 3, $\gamma_{1}>\gamma_{2}$, owing to its similarity in arguments with case 2 $\gamma_{2}>\gamma_{1}>\frac{\mu_{2}}{\mu_{1}}\cdot\gamma_{2}$. Here, we cover those portions of the proof for each of the three cases in which the theorem statement is divided. Case 1 $\gamma_{1}\leq\frac{\mu_{2}}{\mu_{1}}\gamma_{2}$: For convenience of the reader, we mention the unique candidate we obtained by Lemma 3.4 and 3.7: $\displaystyle F_{1}^{\prime}(t)$ $\displaystyle=\mu_{1}\gamma_{1}~{}\text{if}~{}t\in[T_{1,a},T_{1,f}]~{}\text{and}~{}F_{2}^{\prime}(t)=\mu_{2}\gamma_{2}~{}\text{if}~{}t\in[T_{1,f},T_{2,f}].$ (18) We first verify that, when the two classes arrive by the unique remaining candidate profile, for both the classes $i=1,2$ cost derivative satisfies 1) $(C_{\mathbf{F}}^{(i)})^{\prime}(t)\leq 0$ in $(-\infty,T_{i,a})$, 2) $(C_{\mathbf{F}}^{(i)})^{\prime}(t)=0$ in $[T_{i,a},T_{i,f}]$, and 3) $(C_{\mathbf{F}}^{(i)})^{\prime}(t)\geq 0$ in $(T_{i,f},\infty)$. The preceding statement implies both the classes have their cost constant on the support interval and higher outside of it, implying the candidate is an EAP. Following our agenda, we go by the following sequence of arguments. * • In $(-\infty,T_{1,a}]$: $\tau_{\mathbf{F}}^{(i)}(t)=0$ for both classes $i\in\\{1,2\\}$ in $(-\infty,T_{1,a}]$. As a result, $(C_{\mathbf{F}}^{(i)})^{\prime}(t)=-\gamma_{i}<0$ in $(-\infty,T_{1,a}]$. * • In $[T_{1,a},T_{1,f}]$: Queue 1 stays engaged in $[T_{1,a},T_{1,f}]$, since $A_{1}(t)=F_{1}(t)>\mu_{1}\cdot\max\\{t,0\\}$ in $[T_{1,a},T_{1,f}]$. Hence by (3), for both the classes $i\in\\{1,2\\}$, $(C_{\mathbf{F}}^{(i)})^{\prime}(t)=\frac{F_{1}^{\prime}(t)}{\mu_{1}}-\gamma_{i}=\gamma_{1}-\gamma_{i}$. As a result, for $i=1$, $(C_{\mathbf{F}}^{(1)})^{\prime}(t)=0$ and for $i=2$, $(C_{\mathbf{F}}^{(2)})^{\prime}(t)=\gamma_{1}-\gamma_{2}<0$. * • In $[T_{1,f},T_{2,f}]$: Since $\mu_{1}\cdot T_{2,f}=\Lambda_{1}+\frac{\mu_{1}}{\mu_{2}}\Lambda_{2}>\Lambda_{1}+\Lambda_{2}=A_{1}(T_{2,f})$, queue 1 empties at some time $T\in(T_{1,f},T_{2,f}]$. As class 2 users arrive at a constant rate of $\mu_{2}\gamma_{2}$ in $[T_{1,f},T_{2,f}]$, queue 1 stays engaged in $[T_{1,f},T]$ and empty in $[T,T_{2,f}]$. Therefore, in $[T_{1,f},T]$, by (3), $(C_{\mathbf{F}}^{(1)})^{\prime}(t)=\frac{F_{2}^{\prime}(t)}{\mu_{1}}-\gamma_{1}=\frac{\mu_{2}\gamma_{2}-\mu_{1}\gamma_{1}}{\mu_{1}}\geq 0$ and in $(T,T_{2,f}]$, since queue 1 is empty, $(C_{\mathbf{F}}^{(1)})^{\prime}(t)=1-\gamma_{1}\geq 0$. Now we argue that queue 2 remains engaged in $[\tau_{1}(T_{1,f}),T_{2,f}]$ by considering two parts separately: * – In $[\tau_{1}(T_{1,f}),T]$, class 2 users arrive at queue 2 from queue 1 at rate $\mu_{1}>\mu_{2}$ in that period, making queue 2 engaged. * – In $[T,T_{2,f}]$, since $A_{2}^{\prime}(t)=\mu_{2}\gamma_{2}<\mu_{2}$, $A_{2}(t)-\mu_{2}\cdot(t-\tau_{1}(T_{1,f}))$ is a decreasing function. Combining this with the fourth equation $A_{2}(T_{2,f})=\mu_{2}\cdot(T_{2,f}-\tau_{1}(T_{1,f}))$ in the identified linear system, we get $A_{2}(t)>\mu_{2}\cdot(t-\tau_{1}(T_{1,f}))$ in $[T,T_{2,f})$, implying queue 2 stays engaged. Therefore, by (3), for $t\in[T_{1,f},T_{2,f}]$, $\tau_{\mathbf{F}}^{(2)}(t)=\tau_{2}(\tau_{1}(t))=\tau_{1}(T_{1,f})+\frac{A_{2}(\tau_{1}(t))}{\mu_{2}}=\frac{\Lambda_{1}}{\mu_{1}}+\frac{F_{2}(t)}{\mu_{2}}$. Hence, $(C_{\mathbf{F}}^{(2)})^{\prime}(t)=(\tau_{\mathbf{F}}^{(2)})^{\prime}(t)-\gamma_{2}=\frac{F_{2}^{\prime}(t)}{\mu_{2}}-\gamma_{2}=0$ in $[T_{1,f},T_{2,f}]$. Upon summarizing, for $t\in[T_{1,f},T_{2,f}]$, we get $(C_{\mathbf{F}}^{(1)})^{\prime}(t)\geq 0$ and $(C_{\mathbf{F}}^{(2)})^{\prime}(t)=0$. * • In $[T_{2,f},\infty)$: By the last step, both the queues stay empty after $T_{2,f}$. As a result, for both the classes $i\in\\{1,2\\}$, $(C_{\mathbf{F}}^{(i)})^{\prime}(t)=1-\gamma_{i}>0$ in $[T_{2,f},\infty)$. Case 2 $\frac{\mu_{2}}{\mu_{1}}\gamma_{2}<\gamma_{1}<\gamma_{2}$: Skipped parts of the proof of Lemma 3.8: Identifying the unique candidate for Type II: For every EAP under Type II, we obtain the following system of equations to be satisfied by the support boundaries $T_{1,a},T_{1,f},T_{2,a},T_{2,f}$: 1. 1. By Lemma 3.4, $F_{1}^{\prime}(t)=\mu_{1}\gamma_{1}-\mu_{2}\gamma_{2}$ in $[T_{1,a},T_{1,f}]$, giving us: $T_{1,f}=T_{1,a}+\frac{\Lambda_{1}}{\mu_{1}\gamma_{1}-\mu_{2}\gamma_{2}}$. 2. 2. By Lemma 3.4, $F_{2}^{\prime}(t)=\mu_{2}\gamma_{2}$ in $[T_{2,a},T_{2,f}]$, giving us: $T_{2,f}=T_{2,a}+\frac{\Lambda_{2}}{\mu_{2}\gamma_{2}}$. 3. 3. Applying the argument used for getting the third equation of Type I, we must have $F_{1}(T_{1,f})+F_{2}(T_{1,f})=\mu_{1}T_{1,f}$. Plugging in $F_{1}(T_{1,f})=\Lambda_{1}$ and $F_{2}(T_{1,f})=\mu_{2}\gamma_{2}(T_{1,f}-T_{2,a})$ (by Lemma 3.4), we get: $\Lambda_{1}+\mu_{2}\gamma_{2}(T_{1,f}-T_{2,a})=\mu_{1}T_{1,f}$. 4. 4. By an argument similar to the one used for getting the fourth equation of Type I, we have $\mu_{2}\cdot(\tau_{1}(T_{2,f})-\tau_{1}(T_{2,a}))=\Lambda_{2}$. Since queue 1 is empty at $T_{1,f}$ (by Lemma 3.5) and class 2 users arrive at rate $\mu_{2}\gamma_{2}<\mu_{1}$ (by Lemma 3.4), queue 1 stays empty at $T_{2,f}$, causing $\tau_{1}(T_{2,f})=T_{2,f}$. Queue 1 serves the first class 2 user at time zero, causing $\tau_{1}(T_{2,a})=0$. Plugging these in, we get: $T_{2,f}=\frac{\Lambda_{2}}{\mu_{2}}$. ∎ The solution to the above system of equations is in (3.). The support boundaries in (3.) must satisfy $T_{2,a}<T_{1,a}<T_{1,f}\leq T_{2,f}$ to represent an EAP under Type II. Imposing $T_{1,a}>T_{2,a}$ on (3.), we get the condition $\Lambda_{1}<\frac{1-\gamma_{2}}{1-\gamma_{1}}\left(\frac{\mu_{1}\gamma_{1}}{\mu_{2}\gamma_{2}}-1\right)\Lambda_{2}$. Moreover, if $\Lambda_{1}<\frac{1-\gamma_{2}}{1-\gamma_{1}}\left(\frac{\mu_{1}\gamma_{1}}{\mu_{2}\gamma_{2}}-1\right)\Lambda_{2}$, (3.) satisfies $T_{2,a}<T_{1,a}<T_{1,f}\leq T_{2,f}$ and upon plugging in arrival rates from Lemma 3.4, we get a candidate having arrival rates and support boundaries same as the joint arrival profile mentioned under case 2b of Theorem 3.1. This candidate satisfies $F_{1}(T_{1,f})=\Lambda_{1},~{}F_{2}(T_{2,f})=\Lambda_{2}$ and will be the only Type II candidate to qualify as an EAP. Upon proving that this identified Type II candidate is an EAP, the second statement of Lemma 3.8 will follow. Proving that the unique candidate obtained is an EAP: For convenience of the reader we mention the candidates we were left with for both the types: 1. 1. If $\Lambda_{1}\geq\frac{1-\gamma_{2}}{1-\gamma_{1}}\left(\frac{\mu_{1}\gamma_{1}}{\mu_{2}\gamma_{2}}-1\right)\Lambda_{2}$, the unique Type I candidate is: $\displaystyle F_{1}^{\prime}(t)$ $\displaystyle=\begin{cases}\mu_{1}\gamma_{1}~{}~{}&\text{if}~{}t\in[T_{1,a},T_{2,a}),\\\ \mu_{1}\gamma_{1}-\mu_{2}\gamma_{2}~{}~{}&\text{if}~{}t\in[T_{2,a},T_{1,f}],\end{cases}~{}\text{and}~{}F_{2}^{\prime}(t)=\mu_{2}\gamma_{2}~{}~{}\text{if}~{}t\in[T_{2,a},T_{2,f}]$ (19) where $T_{1,a},T_{1,f},T_{2,a},T_{2,f}$ are in ( ‣ 2.). 2. 2. If $\Lambda_{1}<\frac{1-\gamma_{2}}{1-\gamma_{1}}\left(\frac{\mu_{1}\gamma_{1}}{\mu_{2}\gamma_{2}}-1\right)\Lambda_{2}$, the unique Type II candidate is: $\displaystyle F_{1}^{\prime}(t)$ $\displaystyle=\mu_{1}\gamma_{1}-\mu_{2}\gamma_{2}~{}~{}\text{if}~{}t\in[T_{2,a},T_{1,f}],~{}\text{and}~{}F_{2}^{\prime}(t)=\mu_{2}\gamma_{2}~{}~{}\text{if}~{}t\in[T_{2,a},T_{2,f}]$ (20) where $T_{1,a},T_{1,f},T_{2,a},T_{2,f}$ are in (3.). By the following sequence of arguments, for both the types, we argue that if users of the two classes arrive by the unique remaining candidate, both classes $i=1,2$ have their cost derivative satisfying: $(C_{\mathbf{F}}^{(i)})^{\prime}(t)\leq 0$ in $(-\infty,T_{i,a})$, $(C_{\mathbf{F}}^{(i)})^{\prime}(t)=0$ in $[T_{i,a},T_{i,f}]$, and $(C_{\mathbf{F}}^{(i)})^{\prime}(t)\geq 0$ in $(T_{i,f},\infty)$. This will imply, for both the classes, cost is constant in the support interval and higher outside and hence the candidate is an EAP. 1. 1. In $(-\infty,\min\\{T_{1,a},T_{2,a}\\}]$: Note that for $t\in(-\infty,\min\\{T_{1,a},T_{2,a}\\}]$, $\tau_{\mathbf{F}}^{(i)}(t)=0$ for both the classes $i\in\\{1,2\\}$. As a result, $(C_{\mathbf{F}}^{(i)})^{\prime}(t)=-\gamma_{i}<0$ for $i\in\\{1,2\\}$. 2. 2. In $[\min\\{T_{1,a},T_{2,a}\\},T_{1,f}]$: Obtained candidates of both the types satisfy $A_{1}^{\prime}(t)=F_{1}^{\prime}(t)+F_{2}^{\prime}(t)<\mu_{1}$ in $[\min\\{T_{1,a},T_{2,a}\\},T_{1,f}]$, making $A_{1}(t)-\mu_{1}t$ decreasing in $[0,T_{1,f}]$. Since both the types satisfy $A_{1}(T_{1,f})=\mu_{1}T_{1,f}$, the preceding statement imply $A_{1}(t)>\mu_{1}t$ in $[0,T_{1,f})$ and therefore, queue 1 is engaged in $[\min\\{T_{1,a},T_{2,a}\\},T_{1,f}]$ and empty after $T_{1,f}$. Using (3) and the arrival rates from (19) and (20), we get $(C_{\mathbf{F}}^{(1)})^{\prime}(t)=\tau_{1}^{\prime}(t)-\gamma_{1}=\frac{F_{1}^{\prime}(t)+F_{2}^{\prime}(t)}{\mu_{1}}-\gamma_{1}=0$ for $t\in[\min\\{T_{1,a},T_{2,a}\\},T_{1,f}]$. For analyzing the cost of the second class, we consider the two types separately: * • In Type I ($T_{1,a}\leq T_{2,a}<T_{1,f}$): In $[T_{1,a},T_{2,a}]$, $(C_{\mathbf{F}}^{(2)})^{\prime}(t)=\tau_{1}^{\prime}(t)-\gamma_{2}=\frac{F_{1}^{\prime}(t)+F_{2}^{\prime}(t)}{\mu_{1}}-\gamma_{2}=\gamma_{1}-\gamma_{2}<0$. Upon observing $A_{2}(\tau_{1}(t))=F_{2}(t)$, we can write using chain rule that $A_{2}^{\prime}(\tau_{1}(t))=\frac{(A_{2}\circ\tau_{1})^{\prime}(t)}{\tau_{1}^{\prime}(t)}=\frac{F_{2}^{\prime}(t)}{\tau_{1}^{\prime}(t)}$, whenever the derivatives exists. Now for $t\in[T_{2,a},T_{1,f}]$, using (3), rate of arrival of class 2 users in queue 2 at $\tau_{1}(t)$ is $A_{2}^{\prime}(\tau_{1}(t))=\frac{F_{2}^{\prime}(t)}{\tau_{1}^{\prime}(t)}=\mu_{1}\cdot\frac{F_{2}^{\prime}(t)}{F_{1}^{\prime}(t)+F_{2}^{\prime}(t)}=\frac{\mu_{2}\gamma_{2}}{\gamma_{1}}>\mu_{2}$. As a result, queue 2 remains engaged in $[\tau_{1}(T_{2,a}),T_{1,f}]$. With this, the time of service of class 2 users arriving in $[T_{2,a},T_{1,f}]$ is $\tau_{\mathbf{F}}^{(2)}(t)=\tau_{2}(\tau_{1}(t))=\tau_{1}(T_{2,a})+\frac{A_{2}(\tau_{1}(t))}{\mu_{2}}=\tau_{1}(T_{2,a})+\frac{F_{2}(t)}{\mu_{2}}$. Using this and the arrival rates from (19), $(C_{\mathbf{F}}^{(2)})^{\prime}(t)=(\tau_{\mathbf{F}}^{(2)})^{\prime}(t)-\gamma_{2}=\frac{F_{2}^{\prime}(t)}{\mu_{2}}-\gamma_{2}=0$ in $[T_{2,a},T_{1,f}]$. * • In Type II ($T_{2,a}<T_{1,a}<T_{1,f}$): In $[T_{2,a},T_{1,a}]$, rate of arrival of class 2 users at queue 2 at time $\tau_{1}(t)$ is $\mu_{1}>\mu_{2}$, and $\tau_{1}(T_{2,a})=0$. As a result, queue 2 has a positive waiting queue at $\tau_{1}(T_{1,a})$. Following a calculation similar to the one used in the previous step for Type I, rate of arrival of class 2 users at queue 2 at time $\tau_{1}(t)$ is $\frac{\mu_{2}\gamma_{2}}{\gamma_{1}}\geq\mu_{2}$ when $t\in[T_{1,a},T_{1,f}]$. The last two statements imply queue 2 has positive waiting time in $(0,T_{1,f}]$. Using this, time of service of class 2 users arriving in $[T_{2,a},T_{1,f}]$ is $\tau_{\mathbf{F}}^{(2)}(t)=\tau_{2}(\tau_{1}(t))=\tau_{1}(T_{2,a})+\frac{A_{2}(\tau_{1}(t))}{\mu_{2}}=\frac{F_{2}(t)}{\mu_{2}}$ and as a result, $(C_{\mathbf{F}}^{(2)})^{\prime}(t)=\frac{F_{2}^{\prime}(t)}{\mu_{2}}-\gamma_{2}=0$. 3. 3. In $[T_{1,f},T_{2,f}]$: Queue 1 remains empty in this region. On the other hand, queue 2 remains engaged in $[T_{1,f},T_{2,f}]$, since
# TIT/HEP-689 May 2022 Analytic continuation for giant gravitons Yosuke Imamura111E-mail<EMAIL_ADDRESS> Department of Physics, Tokyo Institute of Technology, Tokyo 152-8551, Japan We investigate contributions of giant gravitons to the superconformal index. We concentrate on coincident giant gravitons wrapped around a single cycle, and each contribution is obtained by a certain variable change for fugacities from the index of the worldvolume theory on the giant gravitons. Because we treat the index as a series of fugacities and the variable change relates different convergence regions, we need an analytic continuation before summing up such contributions. We propose a systematic prescription for the continuation. Although our argument is based on some unproved assumptions, it passes non-trivial numerical checks for some examples. With the prescription we can calculate the indices of the M5-brane theories from those of the M2-brane theories, and vice versa. ###### Contents 1. 1 Introduction 2. 2 Plethystic exponential and plethystic logarithm 3. 3 Variable changes and analytic continuation 4. 4 Numerical tests 1. 4.1 Double expansion of the Schur index for $T=D3$ 2. 4.2 Simple expansion of the Schur index for $T=D3$ 3. 4.3 Superconformal index for $T=D3$ 4. 4.4 M5 from M2 5. 4.5 M2 from M5 5. 5 Conclusions and Discussion 6. A Full index of $6$-dim $A_{N-1}$ SCFT ## 1 Introduction For the last few decades the AdS/CFT correspondence [1, 2, 3] has been playing an important role in the progress of string theory and quantum field theories. It provides novel approaches for investigation of different physical quantities in various situations. It is very powerful for the analysis of large $N$ gauge theories, and even in the finite $N$ region, where the Planck length is not negligible, it is possible to calculate supersymmetry protected quantities like $R$-charges of gauge invariant operators by using string theory in the AdS background. Generating functions of such quantities have been calculated on the gravity side of the AdS/CFT corresponsence. For example, the superconformal index of the ${\cal N}=4$ $U(N)$ SYM in the large $N$ limit was reproduced as the index of the bulk supergravity modes [4]. Extended branes are important when $N$ is finite. It is known that not only branes wrapped around topologically non- trivial cycles [5, 6] but also branes wrapped around trivial cycles can be stable and BPS. Such branes carry the same quantum numbers with point-like gravitons, and are called giant gravitons [7, 8, 9, 10]. The BPS partition function for scalar fields in 4d ${\cal N}=4$ SYM was reproduced by geometric quantization of giant gravitons in [11], and the same quantity was also obtained by using the dual giants in [12]. Similar analysis was done in [13] for the 3d and 6d theories realized on M2 and M5-branes. The contribution of giant gravitons with different wrapping numbers to the superconformal index was suggested by a characteristic form of the analytic result in [14] for the unrefined Schur index of ${\cal N}=4$ SYM. (See also [15] for generalization to ${\cal N}=2$ quiver gauge theories.) It has been confirmed by studying fluctuation modes of giant gravitons in [16, 17, 18] for the ${\cal N}=4$ SYM, and for more general theories in [19, 20, 21, 22, 23]. In all cases the index $I_{N}^{(T)}$ of a theory $T$ with finite $N$ whose holographic dual is string/M theory in $AdS\times M_{T}$ is given by a multiple expansion of the form $\displaystyle\frac{I_{N}^{(T)}}{I_{\infty}^{(T)}}=\sum_{m_{1}=0}^{\infty}\cdots\sum_{m_{d}=0}^{\infty}x_{1}^{m_{1}N}\cdots x_{d}^{m_{d}N}F^{(T)}_{m_{1},\ldots,m_{d}},$ (1) where $m_{1},\ldots,m_{d}$ are wrapping numbers of giant gravitons around appropriately chosen $d$ supersymmetric cycles in $M_{T}$. The number of cycles $d$ depends on the theory $T$. For each set of the wrapping numbers the function $F^{(T)}_{m_{1},\ldots,m_{d}}$ is the index of the field theory realized on the corresponding system of giant gravitons and $F_{0,\ldots,0}^{(T)}=1$. A typical structure of the theory is shown in Figure 1. Figure 1: A typical structure of the theory realized on a system of giant gravitons. Each vertex represents the theory $T_{i}$ with rank $m_{i}$ realized on $m_{i}$ coincident giant gravitons wrapped around cycle $i$. Edges connecting vertices are degrees of freedom arising on intersections of cycles. $d=3$ case is shown. It is the direct product of theories $T_{i}$ realized on cycles $i=1,\ldots,d$ coupling to the degrees of freedom arising on their intersections. $F^{(T)}_{m_{1},\ldots,m_{k}}$ are $N$-independent, and the $N$-dependence appears only through the prefactor $x_{1}^{m_{1}N}\cdots x_{d}^{m_{d}N}$. Similar expansion of the index is also proposed in [24] based on a complementary analysis on the gauge theory side, and such expansions were named “the giant graviton expansions”. See also [25]. In their analysis the correspondents for giant gravitons and fluctuations on them are baryon operators and their modifications, respectively. Although there are some technical subtleties about the relation between these expansions obtained on the two sides of the AdS/CFT correspondence they are expected to be essentially the same. In this work we focus on the contribution of giant gravitons wrapped around a single cycle $i$. Let $F^{(T)}_{i,m_{i}}=F^{(T)}_{0,\ldots,m_{i},\ldots,0}$ be such contributions. It is essentially the index $I_{m_{i}}^{(T_{i})}$ of the theory $T_{i}$ realized on the worldvolume of $m_{i}$ coincident giant gravitons. More precisely, $F_{i,m}^{(T)}$ and $I_{m}^{(T_{i})}$ are related by a simple variable change of the fugacities as we will explain shortly. A purpose of this work is to study this relation in detail. For concreteness let us consider the Schur index [26] of the ${\cal N}=4$ $U(N)$ SYM. Let $J_{1}$ and $J_{2}$ be the angular momenta and $R_{x}$, $R_{y}$, and $R_{z}$ be the $R$ charges. The Schur index is defined by $\displaystyle I_{N}^{(D3)}(x,y)=\mathop{\rm tr}\nolimits[(-1)^{F}q^{J_{1}}x^{R_{x}}y^{R_{y}}],\quad q=xy,$ (2) where we denote the ${\cal N}=4$ $U(N)$ SYM by $T=D3$. The constraint $q=xy$ is necessary to preserve the supersymmetry. We eliminate $q$ by the constraint and treat the index as a function of $x$ and $y$. The giant graviton expansion for the Schur index is given in [17] as the double expansion $\displaystyle\frac{I_{N}^{(D3)}}{I^{(D3)}_{\infty}}=\sum_{m_{x}=0}^{\infty}\sum_{m_{y}=0}^{\infty}x^{m_{x}N}y^{m_{y}N}F^{(D3)}_{m_{x},m_{y}}.$ (3) The two non-negative integers $m_{x}$ and $m_{y}$ are wrapping numbers over the cycles $i=x$ and $i=y$, respectively, which are respectively defined to be the $R_{x}$ and $R_{y}$ fixed roci in $M_{D3}=S^{5}$.222We use $x$ and $y$ for the labels of cycles. These should not be confused with the fugacities. A special property of the Schur index is that the functions $F_{m_{x},m_{y}}^{(D3)}$ are factorized [17]: $\displaystyle F_{m_{x},m_{y}}^{(D3)}=(xy)^{m_{x}m_{y}}F^{(D3)}_{x,m_{x}}F^{(D3)}_{y,m_{y}}.$ (4) Therefore, we only need to determine $F_{i,m}^{(D3)}$ to write down the expansion (3). $F_{i,m}^{(D3)}$ is the index of the ${\cal N}=4$ $U(m)$ SYM realized on $m$ coincident giant gravitons wrapped around the cycle $i(=x,y)$. The theories $T$, $T_{x}$, and $T_{y}$ are accidentally the same in this case, and the corresponding indices $I_{N}^{(D3)}$, $F_{x,N}^{(D3)}$, and $F_{y,N}^{(D3)}$ are also essentially the same functions. Careful analysis of the action of superconformal algebra on the $AdS$ boundary and that on giant gravitons wrapped around the cycle $x$ shows that they are related by the involution [16] $\displaystyle(H,J_{1},J_{2},R_{x},R_{y},R_{z})\rightarrow(H-2R_{x},R_{y},R_{z},-R_{x},J_{1},J_{2}).$ (5) A similar relation also holds for cycle $y$. This means that $I_{N}^{(D3)}(x,y)$ and $F_{i,N}^{(D3)}(x,y)$ are related by the variable changes [16] $\displaystyle F_{x,N}^{(D3)}(x,y)=I_{N}^{(D3)}(x^{-1},xy),\quad F_{y,N}^{(D3)}(x,y)=I_{N}^{(D3)}(xy,y^{-1}).$ (6) See also [24] for a derivation of (6) on the gauge theory side. Despite the simplicity of (6), it is not straightforward to calculate $F_{i,N}^{(D3)}$ from $I_{N}^{(D3)}$, and vice versa, because we usually do not have analytic form of the indices, and only series expansions are available. The variable changes relate different convergence regions of the functions, and it is a non-trivial problem to generate one from another. In this paper, we propose a simple prescription to obtain the functions $F_{i,N}^{(T)}$ from $I_{N}^{(T_{i})}$. To understand the relation, we need to carefully analyze the dependence of the series expansion on the choice of expansion variables. Different choices of expansion variables give different series expansions. Indeed, the expansion found in [24] is the simple expansion over a single non-negative integer $m$: $\displaystyle\frac{I_{N}^{(T)}}{I_{\infty}^{(T)}}=\sum_{m=0}^{\infty}x^{mN}F_{i,m}^{(T)}.$ (7) This looks different from the multiple expansion (1). Once we have understood the relation between $I_{N}^{(T_{i})}$ and $F_{i,N}^{(T)}$, we can give a partial explanation to this difference. In the process of confirming the relation (7) the analytic continuation has been necessarily done in [24] for functions appearing in (7) based on a pole structure commonly found in the functions. Our prescription is useful to clarify the pole structure and enable us to study the relations like (6) in a systematic way. The paper is organized as follows. In the next section we prepare tools used in the following sections. In particular, we define “a domain” specifying the expansion variables. We also define the plethystic exponential and the plethystic logarithm with emphasis on the dependence on domains. In section 3 we explain the necessity of the analytic continuation, and propose a simple prescription to realize it. We numerically test the proposed method in Section 4 for the maximally supersymmetric theories realized on D3, M5, and M2-branes. The last section is devoted to conclusions and discussion. ## 2 Plethystic exponential and plethystic logarithm The calculation of the index $I$ of a free field theory usually starts from the analysis of single-particle states. The corresponding index is called the letter index. It is obtained by summing up the contributions from all one- particle states. If there are two fugacities $x$ and $y$ it is given by333We mainly consider the case with two variables. Generalization is straightforward. $\displaystyle i=\sum_{(n_{x},n_{y})\in R}c_{n_{x},n_{y}}x^{n_{x}}y^{n_{y}},$ (8) where $R$ is a certain region on the $2$-dimensional charge lattice specifying a set of monomials appearing in the expansion. We call such a region “a domain.” We usually adopt $R$ such that it covers all one-particle states. We call such a domain “a physical domain”. For the Schur index of $T=D3$ we can take $\displaystyle R_{0}=\\{(n_{x},n_{y})|n_{x},n_{y}\geq 0\\}\backslash(0,0)$ (9) as a physical domain. Later we will discuss other choices of $R$. We call the series associated with a domain $R$ the $R$-series. For a free field theory the series sums up into a simple rational function. For example, the letter index of the ${\cal N}=4$ $U(1)$ vector multiplet is [4] $\displaystyle i^{(D3)}_{1}=1-\frac{(1-x)(1-y)}{1-xy},$ (10) and that of the type IIB supergravity multiplet in $AdS_{5}\times S^{5}$, which is dual to the large $N$ SYM, is [4] $\displaystyle i^{(D3)}_{\infty}=\frac{1}{1-x}+\frac{1}{1-y}-\frac{1}{1-xy}-1.$ (11) By expanding these functions into terms appearing in the domain $R_{0}$ we can obtain the set of coefficients in (8). In the following we use the notation $i|_{R}$ when we want to emphasize that $i$ is given as an $R$-series. We graphically express a series by plotting the coefficients in the two- dimensional lattice as shown in Figure 2. Figure 2: (a) The letter index $i_{1}^{(D3)}$ of the ${\cal N}=4$ vector multiplet. (b) The letter index $i_{\infty}^{(D3)}$ of the type IIB supergravity multiplet in $AdS_{5}\times S^{5}$. Once we have obtained the letter index (8), the index for multi-particle states is uniquely determined by considering the combinatorics of the letters. The solution to this problem is $\displaystyle I=\mathop{\rm Pexp}\nolimits i,$ (12) where $\mathop{\rm Pexp}\nolimits$ is the plethystic exponential defined by $\displaystyle\mathop{\rm Pexp}\nolimits f(x,y)=\exp\left(\sum_{m=1}^{\infty}\frac{1}{m}f(x^{m},y^{m})\right).$ (13) For the definition (13) to make sense the condition $|x^{n_{x}}y^{n_{y}}|<1$ must hold for all terms appearing in the sum (8). This means all terms in $f$ must be contained in the domain $R_{(x,y)}$ defined by $\displaystyle R_{(x,y)}:=\\{(n_{x},n_{y})|n_{x}\log|x|+n_{y}\log|y|<0\\}.$ (14) For example, if we take $x$ and $y$ satisfying $|x|=|y|<1$, (14) gives the upper half plane with the horizontal boundary excluded, and all points of $i_{1}^{(D3)}$ and $i_{\infty}^{(D3)}$ shown in Figure 2 are contained in the domain. (Namely, we can use $R_{(x,y)}$ as a physical domain.) If $i$ is expanded with such a domain $R$, $I$ given by (12) can also be expanded in a similar form; $\displaystyle I=1+\sum_{(n_{x},n_{y})\in R}c^{\prime}_{n_{x},n_{y}}x^{n_{x}}y^{n_{y}}$ (15) with the same domain $R$. We can show that the map from $\\{c_{n_{x},n_{y}}\\}$ to $\\{c^{\prime}_{n_{x},n_{y}}\\}$ is one to one if * • $R$ is a sector with the center angle $\theta\leq\pi$. * • $R\cap(-R)=\phi$. The latter is the condition for the boundary of $R$. We do not include the origin, and if $\theta=\pi$ we can include at most one of two boundary rays. We always require $R$ to satisfy these conditions, and then we can define the inverse operation to $\mathop{\rm Pexp}\nolimits$: $\displaystyle i=\mathop{\rm Plog}\nolimits I.$ (16) This is called the plethystic logarithm. It is important that we can introduce the group structure for $\mathop{\rm Pexp}\nolimits i$ under the multiplication, which corresponds to the addition for $i$. Even for an interacting theory we can calculate the index by using localization method if it is a Lagrangian theory. For the ${\cal N}=4$ SYM with $U(N)$ gauge group the index is given by $\displaystyle I^{(D3)}_{N}|_{R_{(x,y)}}=\frac{1}{N!}\int\prod_{a=1}^{N}\frac{dz_{a}}{2\pi iz_{a}}\prod_{a\neq b}\left(1-\frac{z_{a}}{z_{b}}\right)\mathop{\rm Pexp}\nolimits\left(i_{1}^{(D3)}|_{R_{(x,y)}}\sum_{a,b}\frac{z_{a}}{z_{b}}\right),$ (17) where $z_{a}$ ($a=1,\ldots,N$) are gauge fugacities. As is explicitly shown we suppose $i_{1}^{(D3)}$ is given as an $R_{(x,y)}$-series. Usually $R_{(x,y)}$ is assumed to be a physical domain. This means that $|x^{n_{x}}y^{n_{y}}|<1$ is satisfied for all terms in $i_{1}^{(D3)}|_{R_{(x,y)}}$. If we take the unit circle $|z_{a}|=1$ for the integration contours $|x^{n_{x}}y^{n_{y}}\frac{z_{a}}{z_{b}}|<1$ is also satisfied, and $\mathop{\rm Pexp}\nolimits$ in (17) makes sense. The contour integrals extract the contribution of gauge invariant states and give $I_{N}$ as an $R_{(x,y)}$-series. ## 3 Variable changes and analytic continuation In the following we frequently use the variable changes in (6), and it is convenient to introduce $\sigma_{x}$ and $\sigma_{y}$ to represent these variable changes as follows. $\displaystyle\sigma_{x}(x,y,q)=(x^{-1},q,y),\quad\sigma_{y}(x,y,q)=(q,y^{-1},x).$ (18) If we use the triangular lattice to express the series expansion as in Figure 2, $\sigma_{x}$ and $\sigma_{y}$ are the reflections through the lines perpendicular to $n_{x}$ and $n_{y}$ axes, respectively. With the maps $\sigma_{i}$ ($i=x,y$) the relations in (6) are expressed in the simple form $\displaystyle F_{i,N}^{(D3)}=\sigma_{i}I_{N}^{(D3)}\quad(i=x,y).$ (19) We want to confirm that $F_{i,N}^{(D3)}$ given by (19) correctly reproduce $I_{N}^{(D3)}$ via (3). Let $R$ be a physical domain used for the expansion of $I_{N}^{(D3)}$. To confirm the relation (3) holds we need $F_{i,N}^{(D3)}|_{R}$. However, the variable change $\sigma_{i}$ transforms $I_{N}^{(D3)}|_{R}$ into $F_{i,N}^{(D3)}|_{\sigma_{i}R}$, where $\sigma_{i}R$ is the image of $R$ under the map $\sigma_{i}$. Therefore, to compare the left hand side and the right hand side of (3), we need to resum or analytically continue $F_{i,N}^{(D3)}|_{\sigma_{i}R}$ to $F_{i,N}^{(D3)}|_{R}$. In general, if we have $i|_{R}$ for a domain $R$, it can also be regarded as $i|_{R^{\prime}}$ for another domain $R^{\prime}$ which contains $R$ as a subset. Therefore, it is convenient to define “a maximal domain” which cannot be enlarged any more. $R$ is a maximal domain if $\displaystyle R\cup(-R)\cup O=L,$ (20) where $O$ is the origin and $L$ is the whole plane. A maximal domain $R$ is a sector with center angle $\pi$, and only one of the boundary rays is included in $R$. Let us suppose that $f|_{R^{\prime}}$ for a maximal domain $R^{\prime}$ is given and we want to obtain $(\mathop{\rm Pexp}\nolimits f)|_{R}$ for another maximal domain $R\neq R^{\prime}$. We divide $R^{\prime}$ into the following two parts. $\displaystyle R_{1}=R^{\prime}\backslash R,\quad R_{2}=R^{\prime}\cap R.$ (21) By definition, $R^{\prime}=R_{1}\cup R_{2}$ and $R=(-R_{1})\cup R_{2}$. Correspondingly, we divide $f$ into two parts $f_{1}$ and $f_{2}$ so that $f|_{R^{\prime}}=f_{1}|_{R_{1}}+f_{2}|_{R_{2}}$. The plethystic exponential of $f|_{R^{\prime}}$ is factorized into two factors: $\displaystyle\mathop{\rm Pexp}\nolimits f|_{R^{\prime}}=(\mathop{\rm Pexp}\nolimits f_{1}|_{R_{1}})(\mathop{\rm Pexp}\nolimits f_{2}|_{R_{2}}).$ (22) The factor $\mathop{\rm Pexp}\nolimits f_{2}|_{R_{2}}$ can be regarded as an $R$-series. What we need to do to obtain $(\mathop{\rm Pexp}\nolimits f)|_{R}$ is to rewrite the other factor $\mathop{\rm Pexp}\nolimits f_{1}|_{R_{1}}$ as a $(-R_{1})$-series. This can be done by using analytic continuation as follows. Let us suppose that $f_{1}$ is given by $\displaystyle f_{1}(x,y)=\sum_{(n_{x},n_{y})\in R_{1}}c_{n_{x},n_{y}}x^{n_{x}}y^{n_{y}}.$ (23) For values of $x$ and $y$ such that $R_{1}\subset R_{(x,y)}$ we can rewrite the plethystic exponential in the product form $\displaystyle\mathop{\rm Pexp}\nolimits f_{1}|_{R_{1}}=\prod_{(n_{x},n_{y})\in R_{1}}\frac{1}{(1-x^{n_{x}}y^{n_{y}})^{c_{n_{x},n_{y}}}}.$ (24) Once we have obtained this expression, we can analytically continue this function to values of $x$ and $y$ such that $-R_{1}\subset R_{(x,y)}$. Then, we can rewrite (24) as $\displaystyle\prod_{(n_{x},n_{y})\in R_{1}}\frac{(-x^{n_{x}}y^{n_{y}})^{-c_{n_{x},n_{y}}}}{(1-x^{-n_{x}}y^{-n_{y}})^{c_{n_{x},n_{y}}}}=P(x,y)\mathop{\rm Pexp}\nolimits f_{1}(x^{-1},y^{-1})|_{-R_{1}},$ (25) where $P(x,y)$ is the monomial function $\displaystyle P(x,y)=\prod_{(n_{x},n_{y})\in R_{1}}(-x^{n_{x}}y^{n_{y}})^{-c_{n_{x},n_{y}}}.$ (26) It is important that if $f_{1}$ has infinitely many terms (26) may not be well-defined, and then our prescription does not work. By combining (25) with the factor $\mathop{\rm Pexp}\nolimits f_{2}$ we obtain the analytically continued plethystic exponential associated with the domain $R$: $\displaystyle\mathop{\rm Pexp}\nolimits_{R}f\equiv P\mathop{\rm Pexp}\nolimits f^{\rm mod}|_{R}$ (27) where $f^{\rm mod}$ is defined by $\displaystyle f^{\rm mod}(x,y)=f_{1}(x^{-1},y^{-1})+f_{2}(x,y).$ (28) Each point $(n_{x},n_{y})\in R_{1}$ in $f$ is moved to the opposite point $(-n_{x},-n_{y})\in-R_{1}$ in the modified function $f^{\rm mod}$, and all terms of $f^{\rm mod}$ are contained in $R$. In fact, essentially the same prescription has been used in previous works. Because giant gravitons are wrapped around topologically trivial cycles, there are unwrapping modes with negative excitation energies. Such modes are handled in [16, 17, 18] by applying the prescription explained above to the integrand of the gauge fugacity integrals. The variable change of the index after the gauge fugacity integrals has also been considered in [24]. Although detailed explanation is not given, the pole structure caused by the factor (24) was pointed out, and essentially the same method seems to be used. The prescription explained above can be summarized in the following equations. $\displaystyle F_{i,N}^{(T)}|_{R}=\mathop{\rm Pexp}\nolimits_{R}\sigma_{i}i_{N}^{(T_{i})},\quad i_{N}^{(T_{i})}=\mathop{\rm Plog}\nolimits I_{N}^{(T_{i})}.$ (29) We first calculate $i_{N}^{(T)}$ as the plethystic logarithm of $I_{N}^{(T_{i})}$, and then we calculate the analytically continued plethystic exponential. Unfortunately, it will turn out that this prescription works only in limited cases. If $\sigma_{i}i_{N}^{(T_{i})}$ has infinitely many points outside the domain $R$, then the factor (26) becomes an infinite product which in general does not converge. Even so, we will find that (29) is quite effective for simple expansion (7). ## 4 Numerical tests In this section we numerically calculate $F_{i,N}^{(T)}$ from $I_{N}^{(T_{i})}$ by (29) in some examples and confirm the consistency with known results. The main results in 4.2 and 4.3 have been already given in [24]. ### 4.1 Double expansion of the Schur index for $T=D3$ We first consider the double giant graviton expansion of the Schur index of $T=D3$ studied in [17]. Let us define ${\mathfrak{q}}$ and $u$ by $\displaystyle x={\mathfrak{q}}u,\quad y={\mathfrak{q}}u^{-1},\quad q={\mathfrak{q}}^{2}.$ (30) In [17] functions $F_{i,N}^{(D3)}$ are treated as ${\mathfrak{q}}$-series and the coefficients are given as rational functions of $u$.444The fugacity ${\mathfrak{q}}$ is denoted by $q$ in [17]. Here we do the further $u^{-1}$ expansion ($u$ expansion around $u=\infty$) after the ${\mathfrak{q}}$-expansion. This corresponds to the maximal physical domain (Figure 3 (a)) $\displaystyle R_{qu}$ $\displaystyle=\lim_{{\mathfrak{q}}\rightarrow 0,|u|>1}R_{(x,y)}$ $\displaystyle=\\{(n_{x},n_{y})|n_{x}+n_{y}>0\\}\cup\\{(n_{x},n_{y})|n_{x}+n_{y}=0,n_{y}>0\\}.$ (31) Figure 3: The maximal domains $R_{qu}$ and $R_{yx}$ Let us first consider $F_{x,1}^{(D3)}=\mathop{\rm Pexp}\nolimits(\sigma_{x}i_{1}^{(D3)})$ and $F_{y,1}^{(D3)}=\mathop{\rm Pexp}\nolimits(\sigma_{y}i_{1}^{(D3)})$. We find $\sigma_{x}i_{1}^{(D3)}$ has one term $x^{-1}={\mathfrak{q}}^{-1}u^{-1}$ that is not contained in $R_{qu}$, and we have to move it to the opposite point to define the modified letter index. For $\sigma_{y}i_{1}^{(D3)}$ we need to move two points corresponding to $xy^{-1}=u^{2}$ and $y^{-1}={\mathfrak{q}}^{-1}u$. See Figure 4. Figure 4: The modification of letter indices $\sigma_{x}i_{1}^{(D3)}$ and $\sigma_{y}i_{1}^{(D3)}$ Following the prescription explained in the previous section, we can calculate $F_{x,1}^{(D3)}|_{R_{qu}}$ and $F_{y,1}^{(D3)}|_{R_{qu}}$. The results are shown in Figure 5. Figure 5: (a) $F_{x,1}^{(D3)}|_{R_{qu}}$ (b) $F_{y,1}^{(D3)}|_{R_{qu}}$ These agree with the results in [17]:555$F_{x,N}(x,y)$ is denoted by $F_{N}(q,u)$ in [17] and $F_{y,N}(x,y)$ corresponds to $F_{N}(q,u^{-1})$. $F_{N}$ are given in [17] as ${\mathfrak{q}}$-series. We need further $u^{-1}$ expansion to obtain (32), (33), (34), and (35). $\displaystyle F_{x,1}^{(D3)}$ $\displaystyle=(-u-u^{-1}-u^{-3}-\cdots){\mathfrak{q}}$ $\displaystyle+(-u^{2}+1){\mathfrak{q}}^{2}+(-u^{3}+u^{-3}){\mathfrak{q}}^{3}+(-u^{4}+1-u^{-2}+u^{-6}){\mathfrak{q}}^{4}+\cdots,$ $\displaystyle F^{(D3)}_{y,1}$ $\displaystyle=(u^{-3}+u^{-5}+u^{-7}+\cdots){\mathfrak{q}}$ $\displaystyle+(1-u^{-2}){\mathfrak{q}}^{2}+(u^{3}-u^{-3}){\mathfrak{q}}^{3}+(u^{6}-u^{2}+1-u^{-4}){\mathfrak{q}}^{4}+\cdots.$ (32) This is not surprising because essentially the same prescription is used in [17] for the single giant graviton sector. Next, let us apply (29) to $F_{i,N}^{(D3)}$ with $N\geq 2$. We first calculate $I_{N}^{(D3)}$ by (17), and calculate their plethystic logarithms $i_{N}^{(D3)}=\mathop{\rm Plog}\nolimits I_{N}^{(D3)}$. $i_{N}^{(D3)}$ have information not only about letters generating the operator spectra but also about non-trivial constraints among letters called “syzygies” [27]. The results for $N=2,3,4$ are shown in Figure 6. Figure 6: $i_{N}^{(D3)}$ for $N=2,3,4$ are shown. The distributions are left- right symmetric, and we show only the left half ($n_{x}\geq n_{y}$) of each plot. The dotted lines show a numerical cut-off $n_{x}+n_{y}\leq 20$. The asymptotic behavior of the distribution of dots is important for the following analysis. We find in the plots in Figure 6 that except for finite number of dots near the origin almost all dots are contained in the sectors bounded by $E_{1}:n_{y}=Nn_{x}$ and $E_{2}:n_{x}=Nn_{y}$. For $N=2$ the image of $E_{2}$ under $\sigma_{x}$, $\sigma_{x}E_{2}$, coincides with the right half of the horizontal line, which is the boundary of $R_{qu}$ (Figure 7). Figure 7: The modification of $\sigma_{x}i_{2}^{(D3)}$ for $R_{qu}$. The dotted line shows a numerical cut-off. There are three points outside $R_{qu}$ corresponding to terms $x^{-1}={\mathfrak{q}}^{-1}u$, $x^{-2}={\mathfrak{q}}^{-2}u^{2}$, and $x^{-2}y={\mathfrak{q}}^{-1}u^{3}$. These give the factor $P=-{\mathfrak{q}}^{4}u^{-6}=-x^{5}y^{-1}$, and the analytically continued plethystic exponential is shown in Figure 8. Figure 8: $F_{x,2}^{(D3)}|_{R_{qu}}$ is shown. The dotted line shows a numerical cut-off. Again, we find agreement with the corresponding result in [17]: $\displaystyle F_{x,2}^{(D3)}$ $\displaystyle=(-u^{6}+u^{4}+2+u^{-2}+3u^{-4}+2u^{-6}+\cdots){\mathfrak{q}}^{4}$ $\displaystyle+(-u^{9}+2u^{5}){\mathfrak{q}}^{5}+(-u^{12}+2u^{6}+2){\mathfrak{q}}^{6}+\cdots.$ (33) Next, let us consider $F_{y,2}^{(D3)}$. This time, $\sigma_{y}E_{1}$ is the semi-infinite horizontal line, and unlike the previous case the line is not contained in $R_{qu}$. This means that there are infinitely many points outside $R_{qu}$, and the factor (26) becomes infinite product. Therefore, (29) does not work. In fact, this is also expected from $F_{y,2}^{(D3)}$ in [17]: $\displaystyle F_{y,2}^{(D3)}({\mathfrak{q}},u)$ $\displaystyle=(2u^{-10}+u^{-12}+3u^{-14}+\cdots){\mathfrak{q}}^{4}+\cdots.$ (34) As we see, the leading term has coefficient $2$, and it is impossible to obtain such a series by the plethystic exponential. For $N=3$ and $4$ the situation is worse. Both $\sigma_{x}$ and $\sigma_{y}$ maps one of $E_{1}$ and $E_{2}$ below the boundary of $R_{qu}$, and there are infinitely many points outside $R_{qu}$, which make the factor (26) ill- defined. We can also see that the leading term coefficients of the functions in [17] are not $\pm 1$. $\displaystyle F_{x,3}^{(D3)}$ $\displaystyle=(-2u^{15}+u^{13}+\cdots){\mathfrak{q}}^{9}+\cdots,$ $\displaystyle F_{y,3}^{(D3)}$ $\displaystyle=(5u^{-21}+2u^{-23}+\cdots){\mathfrak{q}}^{9}+\cdots,$ $\displaystyle F_{x,4}^{(D3)}$ $\displaystyle=(-5u^{28}+2u^{26}+\cdots){\mathfrak{q}}^{16}+\cdots,$ $\displaystyle F_{y,4}^{(D3)}$ $\displaystyle=(14u^{-36}+5u^{-38}+\cdots){\mathfrak{q}}^{16}+\cdots.$ (35) These expansions cannot be reproduced by the relation (29). ### 4.2 Simple expansion of the Schur index for $T=D3$ Let us consider the simple giant graviton expansion investigated in [24]. The index is treated in [24] as a series obtained by the successive expansions in which $y$ expansion is carried out first and then $x$ expansion is done for the coefficients of the first expansion. The associated maximal physical domain is (Figure 3 (b)) $\displaystyle R_{yx}=\lim_{y\rightarrow 0,|x|<1}R_{(x,y)}=\\{(n_{x},n_{y})|n_{y}>0\\}\cup\\{(n_{x},0)|n_{x}>0\\}.$ (36) Let us start with the analysis of $F_{x,1}^{(D3)}|_{R_{yx}}$ and $F_{y,1}^{(D3)}|_{R_{yx}}$. Figure 9: (a) The modification of $\sigma_{x}i_{1}$. (b) $\sigma_{y}i_{1}$ For $\sigma_{x}i_{1}^{(D3)}$, there is one dot outside $R_{yx}$ (Figure 9 (a)). This is the same as in Figure 4 (a), and hence we obtain the same result for $F_{x,1}^{(D3)}$ as the previous subsection (Figure 5 (a)). However, for $\sigma_{y}i_{1}^{(D3)}$, there are infinitely many points outside $R_{yx}$ corresponding to terms $y^{-1}x^{k}$ ($k=0,1,2,\ldots$) (Figure 9 (b)), and (26) becomes $\displaystyle P=\prod_{k=0}^{\infty}(-yx^{-k}).$ (37) Although this is an infinite product, we can give it significance. $P$ includes the factor $y^{+\infty}$, and the contribution decouples. We can simply treat $F_{y,1}^{(D3)}|_{R_{yx}}$ as $0$. Namely, as far as single- wrapping giant graviton contributions are concerned, the difference between the multiple expansion (1) and the simple expansion (7) comes from the different choice of the domains. Let us proceed to $N\geq 2$ contributions. For $\sigma_{y}i_{N}^{(D3)}$, we find infinitely many points outside $R_{qu}$, just like the $N=1$ case. Because both positive and negative coefficients appear, it is not clear whether they decouple like $F_{y,1}^{(D3)}$. Here, based on the analysis in [24], we simply assume their decoupling and focus only on $F_{x,N}^{(D3)}$. In $i_{N}^{(D3)}$ shown in Figure 6 we find $N$ dots on the positive part of the $n_{x}$ axis corresponding to the $1/2$ BPS operators $\mathop{\rm tr}\nolimits X^{k}$ ($k=1,2,\ldots,N$). By $\sigma_{x}$ they are mapped to the $N$ points on the negative part of the $n_{x}$ axis. Because they are not contained in $R_{yx}$ we need to move them back to the opposite points on the positive part of the $n_{x}$ axis (Figure 10). Figure 10: The modification of $i_{N}^{(D3)}$ for $R_{yx}$ for $N=2,3,4$. (a) $\sigma_{x}i_{2}$. (b) $\sigma_{x}i_{3}$. (c) $\sigma_{x}i_{4}$. The dotted lines show a numerical cut-off The factorization corresponding to (22) is $\displaystyle I_{N}^{(D3)}(x,y)=\frac{1}{\prod_{k=1}^{N}(1-x^{k})}I^{\prime(D3)}_{N}(x,y),$ (38) The fractional factor coming from the $N$ points on the $n_{x}$ axis and produces poles on the $x$-plane at $k$-roots of unity with $1\leq k\leq N$. As is pointed out in [24] these are only poles we need to take care of in the analytic continuation. The function $I^{\prime(D3)}_{N}$ corresponds to all other points with $n_{y}\geq 1$ in the range $0\leq n_{x}\leq(N+1)n_{y}$. The coefficient of $y$-expansion $I^{\prime(D3)}_{N}$ at order $n_{y}$ is a Laurant polynomial of $x$ consisting of terms in this range. Therefore, we can safely perform the variable change $\sigma_{x}$ for the function $I^{\prime(D3)}_{N}$. After the variable change we obtain $\displaystyle F_{x,N}^{(D3)}=\frac{(-1)^{N}x^{\frac{N(N+1)}{2}}}{\prod_{k=1}^{N}(1-x^{k})}\sigma_{x}I^{\prime(D3)}_{N}$ (39) where $\sigma_{x}I^{\prime(D3)}_{N}$ has terms in the range $-Nn_{y}\leq n_{x}\leq n_{y}$ for each order of its $y$-expansion. Including the fractional factor with the numerator $(-1)^{N}x^{\frac{N(N+1)}{2}}$, we obtain $F_{x,N}^{(D3)}|_{R_{yx}}$ consisting of terms in the region (Figure 11) $\displaystyle\frac{N(N+1)}{2}-Nn_{y}\leq n_{x}.$ (40) Figure 11: (a) $F_{x,2}^{(D3)}|_{R_{yx}}$. (b) $F_{x,3}^{(D3)}|_{R_{yx}}$. (c) $F_{x,4}^{(D3)}|_{R_{yx}}$. The dotted lines show a numerical cut-off. As is confirmed in [24], the simple expansion (7) correctly reproduces the index $I_{N}^{(D3)}$ for different values of $N$. See Figure 12. Figure 12: The result of numerical check for $N=3$. (a) The ratio $I_{3}^{(D3)}/I_{\infty}^{(D3)}$. (b) $(I_{3}^{(D3)}/I_{\infty}^{(D3)})-\sum_{m=0}^{4}x^{mN}F_{x,m}^{(D3)}$. The dotted lines show numerical cut-offs. The red lines labeled by $m$ show thresholds for respective wrapping numbers. Up to expected errors due to the $m=5$ contribution indicated by the line labeled by $m=5$, all terms are correctly canceled. ### 4.3 Superconformal index for $T=D3$ In the analysis of the Schur index we could focus on $F_{i,N}^{(D3)}|_{R}$ thanks to the factorization (4). This is not the case for the superconformal index without the Schur limit taken. The superconformal index of the ${\cal N}=4$ SYM is defined by [4] $\displaystyle I^{(D3)}=\mathop{\rm tr}\nolimits[(-1)^{F}q^{J_{1}}p^{J_{2}}x^{R_{x}}y^{R_{y}}z^{R_{z}}],\quad qp=xyz.$ (41) The giant graviton expansion investigated in [18] is a triple expansion with functions $F_{m_{x},m_{y},m_{z}}^{(D3)}$ in the summand, and the involution (5) gives the following relation between $F_{x,N}^{(D3)}=F_{N,0,0}^{(D3)}$ and $I_{N}^{(D3)}$ [16, 24]: $\displaystyle F_{x,N}^{(D3)}(p,q,x,y,z)=I_{N}^{(D3)}(y,z,x^{-1},p,q).$ (42) Let $\sigma_{x}$ denote this variable change. $F_{0,N,0}^{(D3)}$ and $F_{0,0,N}^{(D3)}$ are also related to $I_{N}^{(D3a)}$ via similar variable changes, which we denote by $\sigma_{y}$ and $\sigma_{z}$, respectively. Due to the absence of the factorization, we need to calculate functions $F_{m_{x},m_{y},m_{z}}^{(D3)}$ directly by using localization formula similar to (17). The letter index appearing in the integrand of the localization formula contains $\sigma_{x}i_{1}^{(D3)}$, $\sigma_{y}i_{1}^{(D3)}$, and $\sigma_{z}i_{1}^{(D3)}$ at the same time, where $i_{1}^{(D3)}$ is given by [4] $\displaystyle i_{1}^{(D3)}=1-\frac{(1-x)(1-y)(1-z)}{(1-q)(1-p)}.$ (43) Even if we use a maximal domain we cannot cover three functions $\sigma_{i}i_{1}^{(D3)}$ with a single domain, and we need to use deformed contours for the gauge fugacity integrals [18, 25]. Despite the absence of the factorization, the analysis in [24] suggests that only $F_{x,m}^{(D3)}$ contribute to the index if we use an appropriate domain, and then our relation (29) is still useful to obtain $F_{x,N}^{(D3)}$ from $I_{N}^{(D3)}$. Corresponding to four independent variables we need to use a four-dimensional lattice to express full expansion of the index. To make two- dimensional plot possible we project the four-dimensional lattice onto the two-dimensional lattice corresponding to the unrefinement $y=z$ and $p=q$. Then, the index becomes function of $x$ and $y$ just like the Schur index.666The projection is used only when we show functions as two-dimensional plots, and we do not take any unrefinement in the calculation. We regard the index as a function of $x$, $\sqrt{yz}$, $y/z$, and $q/p$, and treat $\sqrt{yz}$ as if $y$ in the previous subsection. Although we neglect $y/z$ and $q/p$ in the graphical expression, we keep all variables in the calculation. The letter indices $i_{1}^{(D3)}$ and $\sigma_{i}i_{1}^{(D3)}$ ($i=x,y,z$) after the projection are shown in Figure 13. Figure 13: (a) The letter index $i_{1}^{(D3)}$ (b) The modification of $\sigma_{x}i_{1}^{(D3)}$ for $R_{yx}$. (c) $\sigma_{y}i_{2}^{(D3)}$ and $\sigma_{z}i_{2}^{(D3)}$ are identical after the projection to the two- dimensional lattice. We see that if we use the domain $R_{yx}$ $F_{y,1}^{(D3)}$ and $F_{z,1}^{(D3)}$ decouple just like the Schur index, and $F_{x,1}^{(D3)}$ is the only contribution with $m_{x}+m_{y}+m_{z}=1$. Although we cannot prove the decoupling of all $F_{m_{x},m_{y},m_{z}}^{(D3)}$ with $m_{y}+m_{z}\geq 1$, the analysis in [24] shows that we can reproduce the finite $N$ index by the simple expansion containing only $F_{x,m}^{(D3)}$. See Figure 14 for the result of a numerical test. Figure 14: (a) The ratio of the superconformal indices $I_{N}^{(D3)}/I_{\infty}^{(D3)}$ for $N=3$. (b) $(I_{N}^{(D3)}/I_{\infty}^{(D3)})-\sum_{m=0}^{4}x^{mN}F_{x,N}^{(D3)}$ for $N=3$. The dotted lines show numerical cut-offs. The red lines labeled by $m$ show thresholds for respective wrapping numbers. Only small number of dots, which are consistent with the higher order contributions with $m\geq 5$, are left. ### 4.4 M5 from M2 Let us apply our method to the theory realized on $N$ coincident M5-branes, which we denote by $T=M5$. It is the $6$-dim ${\cal N}=(2,0)$ theory of $A_{N-1}$ type together with a free tensor multiplet. The superconformal index is defined by [28] $\displaystyle I_{N}^{(M5)}=\mathop{\rm tr}\nolimits[(-1)^{F}\check{q}_{1}^{\check{J}_{12}}\check{q}_{2}^{\check{J}_{34}}\check{q}_{3}^{\check{J}_{56}}\check{x}_{1}^{\check{R}_{12}}\check{x}_{2}^{\check{R}_{34}}]\quad\check{q}_{1}\check{q}_{2}\check{q}_{3}=\check{x}_{1}\check{x}_{2},$ (44) where $\check{J}_{12}$, $\check{J}_{34}$, and $\check{J}_{56}$ are Cartan generators of $SO(6)_{\rm spin}$ and $\check{R}_{12}$ and $\check{R}_{34}$ are Cartan generators of $SO(5)_{R}$. The index is given as a double giant graviton expansion associated with two two-cycles in [21], and the single giant graviton sector has been studied. The cycles are defined as the fixed loci of $\check{R}_{12}$ and $\check{R}_{34}$ in $M_{M5}=S^{4}$. We call $\check{R}_{12}$ ($\check{R}_{34}$) fixed locus “12-cycle” (“34-cycle”). M2-branes wrapped around these cycles contribute to the index. The superconformal index of the $3$d ${\cal N}=8$ SCFT realized on coincident M2-branes is defined by [28] $\displaystyle I_{N}^{(M2)}=\mathop{\rm tr}\nolimits[(-1)^{F}\hat{q}^{\hat{J}_{12}}\hat{x}_{1}^{\hat{R}_{12}}\hat{x}_{2}^{\hat{R}_{34}}\hat{x}_{3}^{\hat{R}_{56}}\hat{x}_{4}^{\hat{R}_{78}}]\quad\hat{q}=\hat{x}_{1}\hat{x}_{2}\hat{x}_{3}\hat{x}_{4},$ (45) where $\hat{J}_{12}$ is the spin and $\hat{R}_{12}$, $\hat{R}_{34}$, $\hat{R}_{56}$, and $\hat{R}_{78}$ are $SO(8)_{R}$ Cartan generators. We can calculate this index for an arbitrary $N$ by applying the localization method [29] to the ABJM theory with Chern-Simons level $k=1$ [30]. The Cartan generators of the $3$d ${\cal N}=8$ superconformal algebra acting on M2-branes wrapped around $12$-cycle and those of the $6$d ${\cal N}=(2,0)$ superconformal algebra acting on the $AdS_{7}$ boundary are related by [21] $\displaystyle(\hat{H},\hat{J}_{12},\hat{R}_{12},\hat{R}_{34},\hat{R}_{56},\hat{R}_{78})=(\tfrac{1}{2}\check{H}-\tfrac{3}{2}\check{R}_{12},\check{R}_{34},\check{J}_{12},\check{J}_{34},\check{J}_{56},-\check{R}_{12}).$ (46) Correspondingly, the fugacities in (44) and those in (45) are related by $\displaystyle\sigma_{12}(\hat{q},\hat{x}_{1},\hat{x}_{2},\hat{x}_{3},\hat{x}_{4})=(\check{x}_{2},\check{q}_{1},\check{q}_{2},\check{q}_{3},\check{x}_{1}^{-1}).$ (47) The variable change associated with $34$-cycle, $\sigma_{34}$, is given by swapping $\check{x}_{1}$ and $\check{x}_{2}$ after $\sigma_{12}$. the M2-giant contributions $F_{12,N}^{(M5)}$ are obtained from the index $I_{N}^{(M2)}$ by the relation $\displaystyle F_{12,N}^{(M5)}=\sigma_{12}I_{N}^{(M2)}.$ (48) Again, there are four independent variables and the lattice is four- dimensional. We want to define a projection to two-dimensional lattice to show an expansion as a two-dimensional plot. We introduce variables $x$, $y$, and $u_{i}$ ($i=1,2,3$) by $\displaystyle\check{q}_{i}=yu_{i}\quad(u_{1}u_{2}u_{3}=1),\quad\check{x}_{1}=x,\quad\check{x}_{2}=x^{-1}y^{3}.$ (49) We focus on $x$ and $y$ to show series in figures. We also rewrite the fugacities for M2-branes as follows: $\displaystyle\hat{q}=xy^{3},\quad\hat{x}_{i}=yu_{i}\quad(u_{1}u_{2}u_{3}=1),\quad\hat{x}_{4}=x.$ (50) Then, the variable change (47) becomes $\displaystyle\sigma_{12}(x,y,u_{i})=(x^{-1},y,u_{i}).$ (51) Namely, (48) becomes $\displaystyle F_{12,N}^{(M5)}(x,y,u_{i})=I_{N}^{(M2)}(x^{-1},y,u_{i}).$ (52) As in the previous examples, let us first look at the letter index for a single giant graviton. The theory on a single M2-brane is the free theory of an ${\cal N}=8$ scalar multiplet with the letter index [28] $\displaystyle i_{1}^{(M2)}=\frac{\hat{x}_{1}+\hat{x}_{2}+\hat{x}_{3}+\hat{x}_{4}-\hat{q}(\hat{x}_{1}^{-1}+\hat{x}_{2}^{-1}+\hat{x}_{3}^{-1}+\hat{x}_{4}^{-1})}{1-\hat{q}}.$ (53) We show $i_{1}^{(M2)}$, $\sigma_{12}i_{1}^{(M2)}$, and $\sigma_{34}i_{1}^{(M2)}$ projected on the two-dimensional plane in Figure 15. Figure 15: (a) The letter index $i_{1}^{(M2)}$. (b) The modification of $\sigma_{1}i_{1}^{(M2)}$ for $R_{yx}$. (c) Infinitely many dots of $\sigma_{2}i_{1}^{(M2)}$ are out of $R_{yx}$. If we take the domain $R_{yx}$ defined in (36), $\sigma_{34}i_{1}^{(M2)}$ has infinitely many points outside $R_{yx}$. They give the factor $\displaystyle P=\prod_{k=1}^{\infty}(-y^{3}x^{-k})\prod_{k=1}^{\infty}(-yx^{-k})^{-3}.$ (54) If we take the product for each $k$ first, we obtain $\prod_{k=1}^{\infty}x^{2k}=x^{\infty}$ and the contribution decouples. Just like the previous examples, we simply assume that $F_{m_{12},m_{34}}^{(M5)}$ with $m_{34}\geq 1$ decouple, and let us consider the simple expansion associated with the $12$-cycle $\displaystyle\frac{I_{N}^{(M5)}}{I_{\infty}^{(M5)}}=\sum_{m=0}^{\infty}x^{mN}F_{12,m}^{(M5)}.$ (55) Let us numerically confirm (55) holds for small $N$. $I_{N}^{(M5)}$ with $N=-1,0,1$ and $\infty$ are given by $\displaystyle I_{-1}^{(M5)}=0,\quad I_{0}^{(M5)}=1,\quad I_{1}^{(M5)}=\mathop{\rm Pexp}\nolimits i_{1}^{(M5)},\quad I_{\infty}^{(M5)}=\mathop{\rm Pexp}\nolimits i_{\infty}^{(M5)},$ (56) where $i_{1}^{(M5)}$ and $i_{\infty}^{(M5)}$ are the letter indices of the tensor multiplet and the supergravity multiplet in $AdS_{7}\times S^{4}$. They are given by [28] $\displaystyle i_{1}^{(M5)}=\frac{\check{x}_{1}+\check{x}_{2}-\check{q}_{1}\check{q}_{2}-\check{q}_{2}\check{q}_{3}-\check{q}_{3}\check{q}_{1}+\check{x}_{1}\check{x}_{2}}{(1-\check{q}_{1})(1-\check{q}_{2})(1-\check{q}_{3})},$ (57) and $\displaystyle i_{\infty}^{(M5)}=\frac{\check{x}_{1}+\check{x}_{2}-\check{q}_{1}\check{q}_{2}-\check{q}_{2}\check{q}_{3}-\check{q}_{3}\check{q}_{1}+\check{x}_{1}\check{x}_{2}(\check{q}_{1}+\check{q}_{2}+\check{q}_{3}-\check{x}_{1}-\check{x}_{2})}{(1-\check{x}_{1})(1-\check{x}_{2})(1-\check{q}_{1})(1-\check{q}_{2})(1-\check{q}_{3})}.$ (58) By using these we can calculate the left hand side of (55). In the following we calculate the right hand side of (55). We first calculate the index $I_{N}^{(M2)}$ for different $N$ with the physical domain $R_{yx}$. This is done by applying localization method [29] to the ABJM theory [30]. The plethystic logarithms $i_{N}^{(M2)}$ for small $N$ are shown in Figure 16. Figure 16: $i_{N}^{(M2)}=\mathop{\rm Plog}\nolimits I_{N}^{(M2)}$ for $N=1,2,3,4$ are shown. As shown by the dotted lines the $y$ expansions are cut off at $n_{y}=3$. They have similar structure to $i_{N}^{(D3)}$. In particular, for each $N$ there are $N$ dots on the positive part of the vertical axis corresponding to $N$ $1/2$ BPS operators. After the variable change (51), these dots have to be moved to the opposite points to obtain the modified letter index. The expansion of the analytically continued plethystic exponential $F_{12,N}^{(M5)}|_{R_{yx}}=\mathop{\rm Pexp}\nolimits_{R_{yx}}\sigma_{12}i_{N}^{(M2)}$ for $N=1,2,3,4$ are shown in Figure 17. Figure 17: M2-giant contributions $F_{12,m}^{(M5)}$ with wrapping number $m=1,2,3,4$ are shown. As shown by the dotted lines the $y$ expansions are cut off at $n_{y}=3$. For $N\geq 2$ the dots are in the region above the line $n_{x}=\frac{m(m+1)}{2}+mN-mn_{y}$. ($N=1$ case is exceptional and the line is $n_{x}=1-\frac{2}{3}n_{y}$.) With the functions $F_{12,m}^{(M5)}$ obtained above we can confirm that (55) holds for $N=-1,0,1$ up to expected errors. See Figure 18. Figure 18: The results of consistency check for $N=-1,0,1$. $I_{N}^{(M5)}/I_{\infty}^{(M5)}$ for $N=0$ and $N=1$ are shown in (b) and (d), respectively. Trivial one, $I_{-1}^{(M5)}/I_{\infty}^{(M5)}=0$ is not shown. $(I_{N}^{(M5)}/I_{\infty}^{(M5)})-\sum_{m=0}^{4}x^{mN}F_{12,m}^{(M5)}$ for $N=-1$, $0$, and $1$ are shown in (a), (c), and (e),respectively. As shown by the dotted lines the $y$ expansions are cut off at $n_{y}=3$. In (a), (c), and (e), all terms below the $m=5$ lines are correctly canceled, and only expected errors due to $m\geq 5$ contributions are left. This strongly suggests that (55) correctly gives $I_{N}^{(M5)}$ for an arbitrary $N$. Although we only show the two-dimensional plots, we can calculate the full superconformal index. See Appendix A. We can also compare our results with the analytic result for the Schur-like limit [31, 32] obtained from the analysis of five-dimensional SYM. It is known that by setting $\check{q}_{1}=\check{x}_{1}$, the index $I_{N}^{(M5)}$ reduces to the following function of the single variable $\check{x}_{1}$: $\displaystyle I_{N}^{(M5)}=\mathop{\rm Pexp}\nolimits\frac{\check{x}_{1}+\check{x}_{1}^{2}+\cdots+\check{x}_{1}^{N}}{1-\check{x}_{1}}.$ (59) If we take the same limit in our results shown in Appendix A we obtain $\displaystyle I_{2}^{(M5)}$ $\displaystyle=1+\check{x}_{1}+3\check{x}_{1}^{2}+5\check{x}_{1}^{3}+10\check{x}_{1}^{4}+\underline{15\check{x}_{1}^{5}}+{\cal O}(\check{x}_{1}^{6}),$ $\displaystyle I_{3}^{(M5)}$ $\displaystyle=1+\check{x}_{1}+3\check{x}_{1}^{2}+6\check{x}_{1}^{3}+12\check{x}_{1}^{4}+\underline{20\check{x}_{1}^{5}}+{\cal O}(\check{x}_{1}^{6}),$ $\displaystyle I_{4}^{(M5)}$ $\displaystyle=1+\check{x}_{1}+3\check{x}_{1}^{2}+6\check{x}_{1}^{3}+13\check{x}_{1}^{4}+\underline{22\check{x}_{1}^{5}}+{\cal O}(\check{x}_{1}^{6}),$ $\displaystyle I_{5}^{(M5)}$ $\displaystyle=1+\check{x}_{1}+3\check{x}_{1}^{2}+6\check{x}_{1}^{3}+13\check{x}_{1}^{4}+\underline{23\check{x}_{1}^{5}}+{\cal O}(\check{x}_{1}^{6}).$ (60) We only showed terms independent of $\check{q}_{2}$ and $\check{q}_{3}$. The underlines indicate the terms are incorrect. These are consistent with (59) because in our numerical results the $y$ expansion is cut off at $n_{y}=3$ and the expected order of errors is ${\cal O}(\check{x}_{1}^{4})$ or higher. The $\check{q}_{2}$ or $\check{q}_{3}$-dependent terms which we did not show in (60) are also consistent with the expected errors. Note that the Schur-like limit is ill-defined for each M2-giant contribution because the limit corresponds to $\hat{x}_{1}=\hat{x}_{4}^{-1}$ via (47). The terms $\hat{x}_{1}$ and $\hat{x}_{1}^{-1}$ appear at the same time in the letter index $i_{1}^{(M2)}$ in the limit, and there is no domain covering both of them. We need to take the limit after summing up the contributions of M2-giants. ### 4.5 M2 from M5 It is possible to interchange the roles of M2 and M5 in the previous subsection. Namely, we can calculate the finite $N$ index of M2 theory by summing up M5-giant contributions by the relation $\displaystyle\frac{I_{N}^{(M2)}|_{R_{yx}}}{I_{\infty}^{(M2)}|_{R_{yx}}}=\sum_{m=0}^{\infty}x^{mN}(\sigma_{12}^{-1}I_{m}^{(M5)})|_{R_{yx}},$ (61) where the large $N$ index is given by $I_{\infty}^{(M2)}=\mathop{\rm Pexp}\nolimits i_{\infty}^{(M2)}$ with the letter index [28] $\displaystyle i_{\infty}^{(M2)}=\frac{(1-\hat{q}\hat{x}_{1}^{-1})(1-\hat{q}\hat{x}_{2}^{-1})(1-\hat{q}\hat{x}_{3}^{-1})(1-\hat{q}\hat{x}_{4}^{-1})}{(1-\hat{x}_{1})(1-\hat{x}_{2})(1-\hat{x}_{3})(1-\hat{x}_{4})(1-\hat{q})^{2}}-\frac{1-\hat{q}+\hat{q}^{2}}{(1-\hat{q})^{2}}.$ (62) The calculation to test (61) is parallel to the previous example, and we only give a brief explanation. In [21] four five-cycles in $M_{M5}=S^{7}$ were taken into account. As the previous examples we can show for a single M5 giant that if we take $R_{yx}$ only one of the four cycles, $\hat{R}_{78}$ fixed locus, gives non-trivial contribution. Hence, we focus on $F_{78,N}^{(M2)}$ associated with the cycle. By using the results of previous subsection for $I_{N}^{(M5)}$ we can obtain the plethystic logarithms $i_{N}^{(M5)}=\mathop{\rm Plog}\nolimits I_{N}^{(M5)}$ for small $N$ shown in Figure 19. Figure 19: $i_{N}^{(M5)}$ for $N=1,2,3,4,5$ are shown. As shown by the dotted lines the $y$ expansions are cut off at $n_{y}=3$. As the previous examples we find $N$ dots associated with the $1/2$ BPS operators on the positive part of the vertical axis for each $N$. The functions $F_{78,N}^{(M2)}=\mathop{\rm Pexp}\nolimits_{R_{yx}}\sigma_{12}^{-1}i_{N}^{(M5)}$ are shown in Figure 20. Figure 20: M5 giants contributions $F_{78,N}^{(M2)}$ for $N=1,2,3,4,5$ are shown. As shown by the dotted lines the $y$ expansions are cut off at $n_{y}=3$. The results of a numerical check of (61) for $N=-1,0,1$ are shown in Figure 21. Figure 21: $(I_{N}^{(M2)}/I_{\infty}^{(M2)})$ for $N=0,1$ are shown in (b) and (d), respectively. The trivial one $I_{-1}^{(M2)}/I_{\infty}^{(M2)}=0$ is not shown. $(I_{N}^{(M2)}/I_{\infty}^{(M2)})-\sum_{m=0}^{5}x^{mN}F_{78,m}^{(M2)}$ for $N=-1,0,1$ are shown in (a), (c), and (e), respectively. As shown by the dotted lines the $y$ expansions are cut off at $n_{y}=3$. In (a), (c), and (e) all terms below $m=6$ lines are correctly canceled, and only expected errors due to $m\geq 6$ contributions are left. ## 5 Conclusions and Discussion We investigated the giant graviton expansions. In particular, we concentrated on coincident giant gravitons wrapped around a single cycle. Because the expansion domain for giant gravitons and that of the boundary theory are different, we need to perform an analytic continuation to relate them. We proposed a simple prescription (29) to realize it. We explicitly showed for the Schur index of $T=D3$ that different choices of the expansion domains give different functions for single giant graviton contributions $F_{i,1}^{(D3)}$ ($i=x,y$). If we use $R_{qu}$ in (31) both $F_{x,1}^{(D3)}$ and $F_{y,1}^{(D3)}$ contribute to the Schur index, while if we use $R_{yx}$ in (36), $F_{y,1}^{(D3)}$ does not contribute. Although we have not proved $F_{y,N\geq 2}^{(D3)}$ also vanish, this partially explains why there are two different expansions: the simple expansion found in [24] and the multiple expansion in [17]. It is surprising that they give the same result even though some contributions are lost in the simple expansion. This may be related to the fact that the set of functions $F_{m_{x},m_{y}}$ are strongly constrained. For example, we can formally substitute negative $N$ to expansion (1) or (7), and find that the result is vanishing. This gives very strong constraints on the functions. This implies that the functions share common information, and only small subset appearing in the simple expansion may be sufficient to give the complete answer. It would be interesting to investigate the structure of the constrained set of the functions. An important application of our method is the calculation of the full superconformal index of the 6d ${\cal N}=(2,0)$ $A_{N-1}$ SCFT. It is in principle possible to calculate $I_{N}^{(M5)}$ for an arbitrary $N$ up to an arbitrary order starting from the indices $I_{m}^{(M2)}$ of the M2-brane theories with different $m$. Some results are shown in Appendix A. An important merit of our method is that the theory on giant gravitons does not have to be Lagrangian theories unlike the method adopted in [17, 18, 25], which uses deformed contours in the gauge fugacity integrals to realize the analytic continuation. We only need the final expression of the index of the giant graviton theory. It enables us to apply the method to M5-giants, as was demonstrated in 4.5. Another example with non-Lagrangian giant gravitons is $AdS_{5}\times S^{5}$ with a $7$-brane insertion [33, 34]. As is demonstrated in [23], the giant graviton expansion works well for this system at least for the leading contribution. There are higher order contributions coming from giant gravitons coincident with the $7$-brane, on which a non-Lagrangian theory is realized. We expect our method is useful for the analysis of such contributions. We studied two expansion domains $R_{yx}$ and $R_{qu}$ for the Schur index of $T=D3$. An advantage of $R_{qu}$ over $R_{yx}$ is that for each value of $m=m_{x}+m_{y}$ the symmetry between $x$ and $y$ is manifest. Unfortunately, it turned out that our method has limited applicability for the calculation with the domain $R_{qu}$ due to the ill-defined factor $P$. It would be very nice if we can somehow regularize the product in $P$ when it is infinite and make it possible to apply the method to an arbitrary contribution $F_{i,m}^{(T)}$. In general, the functions $F_{m_{1},\ldots,m_{d}}^{(T)}$ do not factorize into $F_{i,m}^{(T)}$. In the case of D3-giants it is possible to write $F_{m_{1},\ldots,m_{d}}^{(T)}$ in the matrix integral form. Even so, it is not clear how we should choose the integration contours. Because we cannot take a physical domain of expansion, using unit circles for integration contours is not justified. Although some rules for contours have been proposed [18, 25], it is desirable to find more efficient method of calculation applicable to non-Lagrangian giant gravitons. Another important problem is to clarify to what extent the giant graviton expansion works. In this work we focused only on the maximally supersymmetric theories: 4d ${\cal N}=4$ SYM on D3-branes, 3d ${\cal N}=8$ SCFT on M2-branes, and 6d ${\cal N}=(2,0)$ SCFT on M5-branes. The analysis on the gauge theory side [24, 25] found the structure of the giant graviton expansions in variety of theories. It would be interesting to study the applicability of our method to more general class of theories which have holographic dual description. In addition to the giant graviton expansions (1) and (7), another similar expansion was proposed in [35]. It may be interesting to study the relation among them. ## Acknowledgments The author would like to thank S. Murayama and D. Yokoyama for valuable discussions. The work of Y. I. was partially supported by Grand-in-Aid for Scientific Research (C) (No.21K03569), Ministry of Education, Science and Culture, Japan. This work used computational resources TSUBAME3.0 supercomputer provided by Tokyo Institute of Technology. ## Appendix A Full index of $6$-dim $A_{N-1}$ SCFT Let $\chi_{[a,b]}$ be the $SU(3)$ character of the representation with Dynkin label $[a,b]$ defined such that $\chi_{[1,0]}=u_{1}+u_{2}+u_{3}$. The superconformal indices $I_{N}^{(M5)}$ for $N=2,3,4,5$ are given as follows. $\displaystyle I^{(M5)}_{2}$ $\displaystyle=(1+\check{x}_{1}+2\check{x}_{1}^{2}+2\check{x}_{1}^{3}+3\check{x}_{1}^{4}+3\check{x}_{1}^{5}+4\check{x}_{1}^{6}+4\check{x}_{1}^{7}+5\check{x}_{1}^{8}+5\check{x}_{1}^{9}+6\check{x}_{1}^{10}+\cdots)$ $\displaystyle+y\chi_{[1,0]}(\check{x}_{1}+2\check{x}_{1}^{2}+3\check{x}_{1}^{3}+4\check{x}_{1}^{4}+5\check{x}_{1}^{5}+6\check{x}_{1}^{6}+7\check{x}_{1}^{7}+8\check{x}_{1}^{8}+9\check{x}_{1}^{9}+10\check{x}_{1}^{10}+\cdots)$ $\displaystyle+y^{2}[\chi_{[0,1]}(-1-2\check{x}_{1}-3\check{x}_{1}^{2}-3\check{x}_{1}^{3}-4\check{x}_{1}^{4}-4\check{x}_{1}^{5}-5\check{x}_{1}^{6}-5\check{x}_{1}^{7}-6\check{x}_{1}^{8}-6\check{x}_{1}^{9}-7\check{x}_{1}^{10}+\cdots)$ $\displaystyle\quad+\chi_{[2,0]}(\check{x}_{1}+3\check{x}_{1}^{2}+5\check{x}_{1}^{3}+8\check{x}_{1}^{4}+10\check{x}_{1}^{5}+13\check{x}_{1}^{6}+15\check{x}_{1}^{7}+18\check{x}_{1}^{8}+20\check{x}_{1}^{9}+23\check{x}_{1}^{10}+\cdots)]$ $\displaystyle+y^{3}[(\check{x}_{1}^{-1}+2+\check{x}_{1}-\check{x}_{1}^{3}-2\check{x}_{1}^{4}-3\check{x}_{1}^{5}-4\check{x}_{1}^{6}-5\check{x}_{1}^{7}-6\check{x}_{1}^{8}-7\check{x}_{1}^{9}-8\check{x}_{1}^{10}+\cdots)$ $\displaystyle\quad+\chi_{[1,1]}(-1-3\check{x}_{1}-5\check{x}_{1}^{2}-6\check{x}_{1}^{3}-6\check{x}_{1}^{4}-6\check{x}_{1}^{5}-6\check{x}_{1}^{6}-6\check{x}_{1}^{7}-6\check{x}_{1}^{8}-6\check{x}_{1}^{9}-6\check{x}_{1}^{10}+\cdots)$ $\displaystyle\quad+\chi_{[3,0]}(\check{x}_{1}+3\check{x}_{1}^{2}+7\check{x}_{1}^{3}+11\check{x}_{1}^{4}+16\check{x}_{1}^{5}+21\check{x}_{1}^{6}+26\check{x}_{1}^{7}+31\check{x}_{1}^{8}+36\check{x}_{1}^{9}+41\check{x}_{1}^{10}+\cdots)]$ $\displaystyle+\cdots$ (63) $\displaystyle I^{(M5)}_{3}$ $\displaystyle=(1+\check{x}_{1}+2\check{x}_{1}^{2}+3\check{x}_{1}^{3}+4\check{x}_{1}^{4}+5\check{x}_{1}^{5}+7\check{x}_{1}^{6}+8\check{x}_{1}^{7}+10\check{x}_{1}^{8}+12\check{x}_{1}^{9}+14\check{x}_{1}^{10}+\cdots)$ $\displaystyle+y\chi_{[1,0]}(\check{x}_{1}+2\check{x}_{1}^{2}+4\check{x}_{1}^{3}+6\check{x}_{1}^{4}+9\check{x}_{1}^{5}+12\check{x}_{1}^{6}+16\check{x}_{1}^{7}+20\check{x}_{1}^{8}+25\check{x}_{1}^{9}+30\check{x}_{1}^{10}+\cdots)$ $\displaystyle+y^{2}[\chi_{[0,1]}(-1-2\check{x}_{1}-4\check{x}_{1}^{2}-5\check{x}_{1}^{3}-7\check{x}_{1}^{4}-8\check{x}_{1}^{5}-10\check{x}_{1}^{6}-11\check{x}_{1}^{7}-13\check{x}_{1}^{8}-14\check{x}_{1}^{9}-16\check{x}_{1}^{10}+\cdots)$ $\displaystyle\quad+\chi_{[2,0]}(\check{x}_{1}+3\check{x}_{1}^{2}+6\check{x}_{1}^{3}+11\check{x}_{1}^{4}+17\check{x}_{1}^{5}+25\check{x}_{1}^{6}+34\check{x}_{1}^{7}+45\check{x}_{1}^{8}+57\check{x}_{1}^{9}+71\check{x}_{1}^{10}+\cdots)]$ $\displaystyle+y^{3}[(\check{x}_{1}^{-1}+2+2\check{x}_{1}+\check{x}_{1}^{2}-\check{x}_{1}^{3}-4\check{x}_{1}^{4}-8\check{x}_{1}^{5}-12\check{x}_{1}^{6}-18\check{x}_{1}^{7}-24\check{x}_{1}^{8}-31\check{x}_{1}^{9}-39\check{x}_{1}^{10}+\cdots)$ $\displaystyle\quad+\chi_{[1,1]}(-1-3\check{x}_{1}-6\check{x}_{1}^{2}-10\check{x}_{1}^{3}-13\check{x}_{1}^{4}-16\check{x}_{1}^{5}-18\check{x}_{1}^{6}-19\check{x}_{1}^{7}-19\check{x}_{1}^{8}-19\check{x}_{1}^{9}-17\check{x}_{1}^{10}+\cdots)$ $\displaystyle\quad+\chi_{[3,0]}(\check{x}_{1}+3\check{x}_{1}^{2}+8\check{x}_{1}^{3}+15\check{x}_{1}^{4}+26\check{x}_{1}^{5}+40\check{x}_{1}^{6}+58\check{x}_{1}^{7}+79\check{x}_{1}^{8}+105\check{x}_{1}^{9}+133\check{x}_{1}^{10}+\cdots)]$ $\displaystyle+\cdots$ (64) $\displaystyle I^{(M5)}_{4}$ $\displaystyle=(1+\check{x}_{1}+2\check{x}_{1}^{2}+3\check{x}_{1}^{3}+5\check{x}_{1}^{4}+6\check{x}_{1}^{5}+9\check{x}_{1}^{6}+11\check{x}_{1}^{7}+15\check{x}_{1}^{8}+18\check{x}_{1}^{9}+23\check{x}_{1}^{10}+\cdots)$ $\displaystyle+y\chi_{[1,0]}(\check{x}_{1}+2\check{x}_{1}^{2}+4\check{x}_{1}^{3}+7\check{x}_{1}^{4}+11\check{x}_{1}^{5}+16\check{x}_{1}^{6}+23\check{x}_{1}^{7}+31\check{x}_{1}^{8}+41\check{x}_{1}^{9}+53\check{x}_{1}^{10}+\cdots)$ $\displaystyle+y^{2}[\chi_{[0,1]}(-1-2\check{x}_{1}-4\check{x}_{1}^{2}-6\check{x}_{1}^{3}-9\check{x}_{1}^{4}-11\check{x}_{1}^{5}-15\check{x}_{1}^{6}-17\check{x}_{1}^{7}-21\check{x}_{1}^{8}-23\check{x}_{1}^{9}-27\check{x}_{1}^{10}+\cdots)$ $\displaystyle\quad+\chi_{[2,0]}(\check{x}_{1}+3\check{x}_{1}^{2}+6\check{x}_{1}^{3}+12\check{x}_{1}^{4}+20\check{x}_{1}^{5}+32\check{x}_{1}^{6}+47\check{x}_{1}^{7}+68\check{x}_{1}^{8}+92\check{x}_{1}^{9}+124\check{x}_{1}^{10}+\cdots)]$ $\displaystyle+y^{3}[(\check{x}_{1}^{-1}+2+2\check{x}_{1}+2\check{x}_{1}^{2}-4\check{x}_{1}^{4}-10\check{x}_{1}^{5}-18\check{x}_{1}^{6}-30\check{x}_{1}^{7}-44\check{x}_{1}^{8}-62\check{x}_{1}^{9}-84\check{x}_{1}^{10}+\cdots)$ $\displaystyle\quad+\chi_{[1,1]}(-1-3\check{x}_{1}-6\check{x}_{1}^{2}-11\check{x}_{1}^{3}-17\check{x}_{1}^{4}-23\check{x}_{1}^{5}-30\check{x}_{1}^{6}-36\check{x}_{1}^{7}-40\check{x}_{1}^{8}-43\check{x}_{1}^{9}-43\check{x}_{1}^{10}+\cdots)$ $\displaystyle\quad+\chi_{[3,0]}(\check{x}_{1}+3\check{x}_{1}^{2}+8\check{x}_{1}^{3}+16\check{x}_{1}^{4}+30\check{x}_{1}^{5}+50\check{x}_{1}^{6}+79\check{x}_{1}^{7}+117\check{x}_{1}^{8}+168\check{x}_{1}^{9}+231\check{x}_{1}^{10}+\cdots)]$ $\displaystyle+\cdots$ (65) $\displaystyle I^{(M5)}_{5}$ $\displaystyle=(1+\check{x}_{1}+2\check{x}_{1}^{2}+3\check{x}_{1}^{3}+5\check{x}_{1}^{4}+7\check{x}_{1}^{5}+10\check{x}_{1}^{6}+13\check{x}_{1}^{7}+18\check{x}_{1}^{8}+23\check{x}_{1}^{9}+30\check{x}_{1}^{10}+\cdots)$ $\displaystyle+y\chi_{[1,0]}(\check{x}_{1}+2\check{x}_{1}^{2}+4\check{x}_{1}^{3}+7\check{x}_{1}^{4}+12\check{x}_{1}^{5}+18\check{x}_{1}^{6}+27\check{x}_{1}^{7}+38\check{x}_{1}^{8}+53\check{x}_{1}^{9}+71\check{x}_{1}^{10}+\cdots)$ $\displaystyle+y^{2}[\chi_{[0,1]}(-1-2\check{x}_{1}-4\check{x}_{1}^{2}-6\check{x}_{1}^{3}-10\check{x}_{1}^{4}-13\check{x}_{1}^{5}-18\check{x}_{1}^{6}-22\check{x}_{1}^{7}-28\check{x}_{1}^{8}-32\check{x}_{1}^{9}-38\check{x}_{1}^{10}+\cdots)$ $\displaystyle\quad+\chi_{[2,0]}(\check{x}_{1}+3\check{x}_{1}^{2}+6\check{x}_{1}^{3}+12\check{x}_{1}^{4}+21\check{x}_{1}^{5}+35\check{x}_{1}^{6}+54\check{x}_{1}^{7}+81\check{x}_{1}^{8}+116\check{x}_{1}^{9}+163\check{x}_{1}^{10}+\cdots)]$ $\displaystyle+y^{3}[(\check{x}_{1}^{-1}+2+2\check{x}_{1}+2\check{x}_{1}^{2}+\check{x}_{1}^{3}-3\check{x}_{1}^{4}-10\check{x}_{1}^{5}-20\check{x}_{1}^{6}-36\check{x}_{1}^{7}-57\check{x}_{1}^{8}-86\check{x}_{1}^{9}-123\check{x}_{1}^{10}+\cdots)$ $\displaystyle\quad+\chi_{[1,1]}(-1-3\check{x}_{1}-6\check{x}_{1}^{2}-11\check{x}_{1}^{3}-18\check{x}_{1}^{4}-27\check{x}_{1}^{5}-37\check{x}_{1}^{6}-48\check{x}_{1}^{7}-59\check{x}_{1}^{8}-69\check{x}_{1}^{9}-75\check{x}_{1}^{10}+\cdots)$ $\displaystyle\quad+\chi_{[3,0]}(\check{x}_{1}+3\check{x}_{1}^{2}+8\check{x}_{1}^{3}+16\check{x}_{1}^{4}+31\check{x}_{1}^{5}+54\check{x}_{1}^{6}+89\check{x}_{1}^{7}+138\check{x}_{1}^{8}+208\check{x}_{1}^{9}+300\check{x}_{1}^{10}+\cdots)]$ $\displaystyle+\cdots$ (66) We can also calculate the index of $A_{N-1}$ SCFT $I_{N}^{(A)}$ by removing the contribution of the free tensor multiplet by $\displaystyle I_{N}^{(A)}=\frac{I_{N}^{(M5)}}{I_{1}^{(M5)}}.$ (67) The numerical results for $N=2,3,4,5$ are as follows: $\displaystyle I^{(A)}_{2}$ $\displaystyle=(1+\check{x}_{1}^{2}+\check{x}_{1}^{4}+\check{x}_{1}^{6}+\check{x}_{1}^{8}+\check{x}_{1}^{10}+\cdots)$ $\displaystyle+y\chi_{[1,0]}(\check{x}_{1}^{2}+\check{x}_{1}^{4}+\check{x}_{1}^{6}+\check{x}_{1}^{8}+\check{x}_{1}^{10}+\cdots)$ $\displaystyle+y^{2}[\chi_{[0,1]}(-\check{x}_{1}-\check{x}_{1}^{3}-\check{x}_{1}^{5}-\check{x}_{1}^{7}-\check{x}_{1}^{9}+\cdots)$ $\displaystyle\quad+\chi_{[2,0]}(\check{x}_{1}^{2}+2\check{x}_{1}^{4}+2\check{x}_{1}^{6}+2\check{x}_{1}^{8}+2\check{x}_{1}^{10}+\cdots)]$ $\displaystyle+y^{3}[(1-\check{x}_{1}+\check{x}_{1}^{2}-x^{3}+\check{x}_{1}^{4}-\check{x}_{1}^{5}+\check{x}_{1}^{6}-\check{x}_{1}^{7}+\check{x}_{1}^{8}-\check{x}_{1}^{9}+\check{x}_{1}^{10}+\cdots)$ $\displaystyle\quad+\chi_{[1,1]}(-\check{x}_{1}-2\check{x}_{1}^{3}+\check{x}_{1}^{4}-2\check{x}_{1}^{5}+\check{x}_{1}^{6}-2\check{x}_{1}^{7}+\check{x}_{1}^{8}-2\check{x}_{1}^{9}+\check{x}_{1}^{10}+\cdots)$ $\displaystyle\quad+\chi_{[3,0]}(\check{x}_{1}^{2}+2\check{x}_{1}^{4}+3\check{x}_{1}^{6}+3\check{x}_{1}^{8}+3\check{x}_{1}^{10}+\cdots)]$ $\displaystyle+\cdots$ (68) $\displaystyle I^{(A)}_{3}$ $\displaystyle=(1+\check{x}_{1}^{2}+\check{x}_{1}^{3}+\check{x}_{1}^{4}+\check{x}_{1}^{5}+2\check{x}_{1}^{6}+\check{x}_{1}^{7}+2\check{x}_{1}^{8}+2\check{x}_{1}^{9}+2\check{x}_{1}^{10}+\cdots)$ $\displaystyle+y\chi_{[1,0]}(\check{x}_{1}^{2}+\check{x}_{1}^{3}+\check{x}_{1}^{4}+2\check{x}_{1}^{5}+2\check{x}_{1}^{6}+2\check{x}_{1}^{7}+3\check{x}_{1}^{8}+3\check{x}_{1}^{9}+3\check{x}_{1}^{10}+\cdots)$ $\displaystyle+y^{2}[\chi_{[0,1]}(-\check{x}_{1}-\check{x}_{1}^{2}-\check{x}_{1}^{3}-2\check{x}_{1}^{4}-\check{x}_{1}^{5}-2\check{x}_{1}^{6}-2\check{x}_{1}^{7}-2\check{x}_{1}^{8}-2\check{x}_{1}^{9}-3\check{x}_{1}^{10}+\cdots)$ $\displaystyle\quad+\chi_{[2,0]}(\check{x}_{1}^{2}+\check{x}_{1}^{3}+2\check{x}_{1}^{4}+3\check{x}_{1}^{5}+4\check{x}_{1}^{6}+4\check{x}_{1}^{7}+6\check{x}_{1}^{8}+6\check{x}_{1}^{9}+7\check{x}_{1}^{10}+\cdots)]$ $\displaystyle+y^{3}[(1-\check{x}_{1}^{4}-\check{x}_{1}^{5}-\check{x}_{1}^{6}-2\check{x}_{1}^{7}-2\check{x}_{1}^{8}-2\check{x}_{1}^{9}-3\check{x}_{1}^{10}+\cdots)$ $\displaystyle\quad+\chi_{[1,1]}(-\check{x}_{1}-\check{x}_{1}^{2}-2\check{x}_{1}^{3}-3\check{x}_{1}^{4}-2\check{x}_{1}^{5}-3\check{x}_{1}^{6}-3\check{x}_{1}^{7}-2\check{x}_{1}^{8}-3\check{x}_{1}^{9}-3\check{x}_{1}^{10}+\cdots)$ $\displaystyle\quad+\chi_{[3,0]}(\check{x}_{1}^{2}+\check{x}_{1}^{3}+2\check{x}_{1}^{4}+4\check{x}_{1}^{5}+5\check{x}_{1}^{6}+6\check{x}_{1}^{7}+9\check{x}_{1}^{8}+10\check{x}_{1}^{9}+11\check{x}_{1}^{10}+\cdots)]$ $\displaystyle+\cdots$ (69) $\displaystyle I^{(A)}_{4}$ $\displaystyle=(1+\check{x}_{1}^{2}+\check{x}_{1}^{3}+2\check{x}_{1}^{4}+\check{x}_{1}^{5}+3\check{x}_{1}^{6}+2\check{x}_{1}^{7}+4\check{x}_{1}^{8}+3\check{x}_{1}^{9}+5\check{x}_{1}^{10}+\cdots)$ $\displaystyle+y\chi_{[1,0]}(\check{x}_{1}^{2}+\check{x}_{1}^{3}+2\check{x}_{1}^{4}+2\check{x}_{1}^{5}+4\check{x}_{1}^{6}+4\check{x}_{1}^{7}+6\check{x}_{1}^{8}+6\check{x}_{1}^{9}+9\check{x}_{1}^{10}+\cdots)$ $\displaystyle+y^{2}[\chi_{[0,1]}(-\check{x}_{1}-\check{x}_{1}^{2}-2\check{x}_{1}^{3}-2\check{x}_{1}^{4}-3\check{x}_{1}^{5}-3\check{x}_{1}^{6}-4\check{x}_{1}^{7}-4\check{x}_{1}^{8}-5\check{x}_{1}^{9}-5\check{x}_{1}^{10}+\cdots)$ $\displaystyle\quad+\chi_{[2,0]}(\check{x}_{1}^{2}+\check{x}_{1}^{3}+3\check{x}_{1}^{4}+3\check{x}_{1}^{5}+7\check{x}_{1}^{6}+7\check{x}_{1}^{7}+12\check{x}_{1}^{8}+12\check{x}_{1}^{9}+19\check{x}_{1}^{10}+\cdots)]$ $\displaystyle+y^{3}[(1+\check{x}_{1}^{2}-\check{x}_{1}^{3}-3\check{x}_{1}^{5}-2\check{x}_{1}^{6}-6\check{x}_{1}^{7}-5\check{x}_{1}^{8}-9\check{x}_{1}^{9}-9\check{x}_{1}^{10}+\cdots)$ $\displaystyle\quad+\chi_{[1,1]}(-\check{x}_{1}-\check{x}_{1}^{2}-3\check{x}_{1}^{3}-3\check{x}_{1}^{4}-6\check{x}_{1}^{5}-5\check{x}_{1}^{6}-8\check{x}_{1}^{7}-6\check{x}_{1}^{8}-10\check{x}_{1}^{9}-6\check{x}_{1}^{10}+\cdots)$ $\displaystyle\quad+\chi_{[3,0]}(\check{x}_{1}^{2}+\check{x}_{1}^{3}+3\check{x}_{1}^{4}+4\check{x}_{1}^{5}+9\check{x}_{1}^{6}+10\check{x}_{1}^{7}+17\check{x}_{1}^{8}+20\check{x}_{1}^{9}+30\check{x}_{1}^{10}+\cdots)]$ $\displaystyle+\cdots$ (70) $\displaystyle I^{(A)}_{5}$ $\displaystyle=(1+\check{x}_{1}^{2}+\check{x}_{1}^{3}+2\check{x}_{1}^{4}+2\check{x}_{1}^{5}+3\check{x}_{1}^{6}+3\check{x}_{1}^{7}+5\check{x}_{1}^{8}+5\check{x}_{1}^{9}+7\check{x}_{1}^{10}+\cdots)$ $\displaystyle+y\chi_{[1,0]}(\check{x}_{1}^{2}+\check{x}_{1}^{3}+2\check{x}_{1}^{4}+3\check{x}_{1}^{5}+4\check{x}_{1}^{6}+6\check{x}_{1}^{7}+8\check{x}_{1}^{8}+10\check{x}_{1}^{9}+13\check{x}_{1}^{10}+\cdots)$ $\displaystyle+y^{2}[\chi_{[0,1]}(-\check{x}_{1}-\check{x}_{1}^{2}-2\check{x}_{1}^{3}-3\check{x}_{1}^{4}-3\check{x}_{1}^{5}-5\check{x}_{1}^{6}-5\check{x}_{1}^{7}-7\check{x}_{1}^{8}-7\check{x}_{1}^{9}-9\check{x}_{1}^{10}+\cdots)$ $\displaystyle\quad+\chi_{[2,0]}(\check{x}_{1}^{2}+\check{x}_{1}^{3}+3\check{x}_{1}^{4}+4\check{x}_{1}^{5}+7\check{x}_{1}^{6}+10\check{x}_{1}^{7}+15\check{x}_{1}^{8}+19\check{x}_{1}^{9}+27\check{x}_{1}^{10}+\cdots)]$ $\displaystyle+y^{3}[(1+\check{x}_{1}^{2}-\check{x}_{1}^{4}-2\check{x}_{1}^{5}-4\check{x}_{1}^{6}-7\check{x}_{1}^{7}-10\check{x}_{1}^{8}-14\check{x}_{1}^{9}-19\check{x}_{1}^{10}+\cdots)$ $\displaystyle\quad+\chi_{[1,1]}(-\check{x}_{1}-\check{x}_{1}^{2}-3\check{x}_{1}^{3}-4\check{x}_{1}^{4}-6\check{x}_{1}^{5}-9\check{x}_{1}^{6}-10\check{x}_{1}^{7}-13\check{x}_{1}^{8}-15\check{x}_{1}^{9}-16\check{x}_{1}^{10}+\cdots)$ $\displaystyle\quad+\chi_{[3,0]}(\check{x}_{1}^{2}+\check{x}_{1}^{3}+3\check{x}_{1}^{4}+5\check{x}_{1}^{5}+9\check{x}_{1}^{6}+14\check{x}_{1}^{7}+21\check{x}_{1}^{8}+30\check{x}_{1}^{9}+42\check{x}_{1}^{10}+\cdots)]$ $\displaystyle+\cdots$ (71) ## References * [1] J. M. Maldacena, “The Large N limit of superconformal field theories and supergravity,” Int. J. Theor. Phys. 38, 1113 (1999) [Adv. Theor. Math. Phys. 2, 231 (1998)] doi:10.1023/A:1026654312961, 10.4310/ATMP.1998.v2.n2.a1 [hep-th/9711200]. * [2] S. S. Gubser, I. R. Klebanov and A. M. Polyakov, “Gauge theory correlators from noncritical string theory,” Phys. Lett. B 428, 105-114 (1998) doi:10.1016/S0370-2693(98)00377-3 [arXiv:hep-th/9802109 [hep-th]]. * [3] E. Witten, “Anti-de Sitter space and holography,” Adv. Theor. Math. Phys. 2, 253-291 (1998) doi:10.4310/ATMP.1998.v2.n2.a2 [arXiv:hep-th/9802150 [hep-th]]. * [4] J. Kinney, J. M. Maldacena, S. Minwalla and S. Raju, “An Index for 4 dimensional super conformal theories,” Commun. Math. Phys. 275, 209 (2007) doi:10.1007/s00220-007-0258-7 [hep-th/0510251]. * [5] E. Witten, “Baryons and branes in anti-de Sitter space,” JHEP 9807, 006 (1998) doi:10.1088/1126-6708/1998/07/006 [hep-th/9805112]. * [6] S. S. Gubser and I. R. Klebanov, “Baryons and domain walls in an N=1 superconformal gauge theory,” Phys. Rev. D 58, 125025 (1998) doi:10.1103/PhysRevD.58.125025 [arXiv:hep-th/9808075 [hep-th]]. * [7] J. McGreevy, L. Susskind and N. Toumbas, “Invasion of the giant gravitons from Anti-de Sitter space,” JHEP 0006, 008 (2000) doi:10.1088/1126-6708/2000/06/008 [hep-th/0003075]. * [8] M. T. Grisaru, R. C. Myers and O. Tafjord, “SUSY and goliath,” JHEP 0008, 040 (2000) doi:10.1088/1126-6708/2000/08/040 [hep-th/0008015]. * [9] A. Hashimoto, S. Hirano and N. Itzhaki, “Large branes in AdS and their field theory dual,” JHEP 0008, 051 (2000) doi:10.1088/1126-6708/2000/08/051 [hep-th/0008016]. * [10] A. Mikhailov, “Giant gravitons from holomorphic surfaces,” JHEP 0011, 027 (2000) doi:10.1088/1126-6708/2000/11/027 [hep-th/0010206]. * [11] I. Biswas, D. Gaiotto, S. Lahiri and S. Minwalla, “Supersymmetric states of N=4 Yang-Mills from giant gravitons,” JHEP 0712, 006 (2007) doi:10.1088/1126-6708/2007/12/006 [hep-th/0606087]. * [12] G. Mandal and N. V. Suryanarayana, “Counting 1/8-BPS dual-giants,” JHEP 0703, 031 (2007) doi:10.1088/1126-6708/2007/03/031 [hep-th/0606088]. * [13] S. Bhattacharyya and S. Minwalla, “Supersymmetric states in M5/M2 CFTs,” JHEP 12, 004 (2007) doi:10.1088/1126-6708/2007/12/004 [arXiv:hep-th/0702069 [hep-th]]. * [14] J. Bourdier, N. Drukker and J. Felix, “The exact Schur index of $\mathcal{N}=4$ SYM,” JHEP 1511, 210 (2015) doi:10.1007/JHEP11(2015)210 [arXiv:1507.08659 [hep-th]]. * [15] J. Bourdier, N. Drukker and J. Felix, “The $\mathcal{N}=2$ Schur index from free fermions,” JHEP 1601, 167 (2016) doi:10.1007/JHEP01(2016)167 [arXiv:1510.07041 [hep-th]]. * [16] R. Arai and Y. Imamura, “Finite $N$ Corrections to the Superconformal Index of S-fold Theories,” PTEP 2019, no.8, 083B04 (2019) doi:10.1093/ptep/ptz088 [arXiv:1904.09776 [hep-th]]. * [17] R. Arai, S. Fujiwara, Y. Imamura and T. Mori, “Schur index of the ${\cal N}=4$ $U(N)$ supersymmetric Yang-Mills theory via the AdS/CFT correspondence,” Phys. Rev. D 101, no.8, 086017 (2020) doi:10.1103/PhysRevD.101.086017 [arXiv:2001.11667 [hep-th]]. * [18] Y. Imamura, “Finite-N superconformal index via the AdS/CFT correspondence,” PTEP 2021, no.12, 123B05 (2021) doi:10.1093/ptep/ptab141 [arXiv:2108.12090 [hep-th]]. * [19] R. Arai, S. Fujiwara, Y. Imamura and T. Mori, “Finite $N$ corrections to the superconformal index of orbifold quiver gauge theories,” JHEP 1910, 243 (2019) doi:10.1007/JHEP10(2019)243 [arXiv:1907.05660 [hep-th]]. * [20] R. Arai, S. Fujiwara, Y. Imamura and T. Mori, “Finite $N$ corrections to the superconformal index of toric quiver gauge theories,” PTEP 2020, no.4, 043B09 (2020) doi:10.1093/ptep/ptaa023 [arXiv:1911.10794 [hep-th]]. * [21] R. Arai, S. Fujiwara, Y. Imamura, T. Mori and D. Yokoyama, “Finite-$N$ corrections to the M-brane indices,” JHEP 11, 093 (2020) doi:10.1007/JHEP11(2020)093 [arXiv:2007.05213 [hep-th]]. * [22] S. Fujiwara, Y. Imamura and T. Mori, “Flavor symmetries of six-dimensional ${\cal N}=(1,0)$ theories from AdS/CFT correspondence,” JHEP 05, 221 (2021) doi:10.1007/JHEP05(2021)221 [arXiv:2103.16094 [hep-th]]. * [23] Y. Imamura and S. Murayama, “Holographic index calculation for Argyres-Douglas and Minahan-Nemeschansky theories,” [arXiv:2110.14897 [hep-th]]. * [24] D. Gaiotto and J. H. Lee, “The Giant Graviton Expansion,” [arXiv:2109.02545 [hep-th]]. * [25] J. H. Lee, “Exact Stringy Microstates from Gauge Theories,” [arXiv:2204.09286 [hep-th]]. * [26] A. Gadde, L. Rastelli, S. S. Razamat and W. Yan, “Gauge Theories and Macdonald Polynomials,” Commun. Math. Phys. 319, 147 (2013) doi:10.1007/s00220-012-1607-8 [arXiv:1110.3740 [hep-th]]. [27] * [27] S. Benvenuti, B. Feng, A. Hanany and Y. H. He, “Counting BPS Operators in Gauge Theories: Quivers, Syzygies and Plethystics,” JHEP 11, 050 (2007) doi:10.1088/1126-6708/2007/11/050 [arXiv:hep-th/0608050 [hep-th]]. * [28] J. Bhattacharya, S. Bhattacharyya, S. Minwalla and S. Raju, “Indices for Superconformal Field Theories in 3,5 and 6 Dimensions,” JHEP 02, 064 (2008) doi:10.1088/1126-6708/2008/02/064 [arXiv:0801.1435 [hep-th]]. * [29] S. Kim, “The Complete superconformal index for N=6 Chern-Simons theory,” Nucl. Phys. B 821, 241-284 (2009) doi:10.1016/j.nuclphysb.2009.06.025 [arXiv:0903.4172 [hep-th]]. * [30] O. Aharony, O. Bergman, D. L. Jafferis and J. Maldacena, “N=6 superconformal Chern-Simons-matter theories, M2-branes and their gravity duals,” JHEP 10, 091 (2008) doi:10.1088/1126-6708/2008/10/091 [arXiv:0806.1218 [hep-th]]. * [31] H. C. Kim, S. Kim, S. S. Kim and K. Lee, “The general M5-brane superconformal index,” [arXiv:1307.7660 [hep-th]]. * [32] C. Beem, L. Rastelli and B. C. van Rees, “$\mathcal{W}$ symmetry in six dimensions,” JHEP 05, 017 (2015) doi:10.1007/JHEP05(2015)017 [arXiv:1404.1079 [hep-th]]. * [33] A. Fayyazuddin and M. Spalinski, “Large N superconformal gauge theories and supergravity orientifolds,” Nucl. Phys. B 535, 219-232 (1998) doi:10.1016/S0550-3213(98)00545-8 [arXiv:hep-th/9805096 [hep-th]]. * [34] O. Aharony, A. Fayyazuddin and J. M. Maldacena, “The Large N limit of N=2, N=1 field theories from three-branes in F theory,” JHEP 07, 013 (1998) doi:10.1088/1126-6708/1998/07/013 [arXiv:hep-th/9806159 [hep-th]]. * [35] S. Murthy, “Unitary matrix models, free fermion ensembles, and the giant graviton expansion,” [arXiv:2202.06897 [hep-th]].