text
stringlengths
6
128k
In this paper we explain how the notion of ''weak Dirichlet process'' is the suitable generalization of the one of semimartingale with jumps. For such a process we provide a unique decomposition which is new also for semimartingales: in particular we introduce ''characteristics'' for weak Dirichlet processes. We also introduce a weak concept (in law) of finite quadratic variation. We investigate a set of new useful chain rules and we discuss a general framework of (possibly path-dependent with jumps) martingale problems with a set of examples of SDEs with jumps driven by a distributional drift.
One of the main features of superfluids is the presence of topological defects with quantised circulation. These objects are known as quantum vortices and exhibit a hydrodynamic behaviour. Nowadays, particles are the main experimental tool used to visualise quantum vortices and to study their dynamics. We use a self-consistent model based on the three-dimensional Gross-Pitaevskii (GP) equation to explore theoretically and numerically the attractive interaction between particles and quantised vortices at very low temperature. Particles are described as localised potentials depleting the superfluid and following Newtonian dynamics. We are able to derive analytically a reduced central-force model that only depends on the classical degrees of freedom of the particle. Such model is found to be consistent with the GP simulations. We then generalised the model to include deformations of the vortex filament. The resulting long-range mutual interaction qualitatively reproduces the observed generation of a cusp on the vortex filament during the particle approach. Moreover, we show that particles can excite Kelvin waves on the vortex filament through a resonance mechanism even if they are still far from it.
It is usually assumed that the inflationary fluctuations start from the Bunch-Davies (BD) vacuum and the $i\varepsilon$ prescription is used when interactions are calculated. We show that those assumptions can be verified explicitly by calculating the loop corrections to the inflationary two-point and three-point correlation functions. Those loop corrections can be resumed to exponential factors, which suppress non-BD coefficients and behave as the $i\varepsilon$ factor for the case of the BD initial condition. A new technique of loop chain diagram resummation is developed for this purpose. For the non-BD initial conditions which is setup at finite time and has not fully decayed, explicit correction to the two-point and three-point correlation functions are calculated. Especially, non-Gaussianity in the folded limit is regularized due to the interactions.
The physics of strongly correlated quantum particles within a flat band was originally explored as a route to itinerant ferromagnetism and, indeed, a celebrated theorem by Lieb rigorously establishes that the ground state of the repulsive Hubbard model on a bipartite lattice with unequal number of sites in each sublattice must have nonzero spin S at half-filling. Recently, there has been interest in Lieb geometries due to the possibility of novel topological insulator, nematic, and Bose-Einstein condensed (BEC) phases. In this paper, we extend the understanding of the attractive Hubbard model on the Lieb lattice by using Determinant Quantum Monte Carlo to study real space charge and pair correlation functions not addressed by the Lieb theorems.
The newly developed "strongly constrained and appropriately normed" (SCAN) meta-generalized-gradient approximation (meta-GGA) can generally improve over the non-empirical Perdew-Burke-Ernzerhof (PBE) GGA not only for strong chemical bonding, but also for the intermediate-range van der Waals (vdW) interaction. However, the long-range vdW interaction is still missing. To remedy this, we propose here pairing SCAN with the non-local correlation part from the rVV10 vdW density functional, with only two empirical parameters. The resulting SCAN+rVV10 yields excellent geometric and energetic results not only for molecular systems, but also for solids and layered-structure materials, as well as the adsorption of benzene on coinage metal surfaces. Especially, SCAN+rVV10 outperforms all current methods with comparable computational efficiencies, accurately reproducing the three most fundamental parameters---the inter-layer binding energies, inter-, and intra-layer lattice constants---for 28 layered-structure materials. Hence, we have achieved with SCAN+rVV10 a promising vdW density functional for general geometries, with minimal empiricism.
The advent of Transformers marked a significant breakthrough in sequence modelling, providing a highly performant architecture capable of leveraging GPU parallelism. However, Transformers are computationally expensive at inference time, limiting their applications, particularly in low-resource settings (e.g., mobile and embedded devices). Addressing this, we (1) begin by showing that attention can be viewed as a special Recurrent Neural Network (RNN) with the ability to compute its \textit{many-to-one} RNN output efficiently. We then (2) show that popular attention-based models such as Transformers can be viewed as RNN variants. However, unlike traditional RNNs (e.g., LSTMs), these models cannot be updated efficiently with new tokens, an important property in sequence modelling. Tackling this, we (3) introduce a new efficient method of computing attention's \textit{many-to-many} RNN output based on the parallel prefix scan algorithm. Building on the new attention formulation, we (4) introduce \textbf{Aaren}, an attention-based module that can not only (i) be trained in parallel (like Transformers) but also (ii) be updated efficiently with new tokens, requiring only constant memory for inferences (like traditional RNNs). Empirically, we show Aarens achieve comparable performance to Transformers on $38$ datasets spread across four popular sequential problem settings: reinforcement learning, event forecasting, time series classification, and time series forecasting tasks while being more time and memory-efficient.
The spin dynamics of the ferromagnetic Kondo lattice CeRuPO is investigated by Electron Spin Resonance (ESR) at microwave frequencies of 1, 9.4, and 34~GHz. The measured resonance can be ascribed to a rarely observed bulk Ce3+ resonance in a metallic Ce compound and can be followed below the ferromagnetic transition temperature Tc=14 K. At T>Tc the interplay between the RKKY-exchange interaction and the crystal electric field anisotropy determines the ESR parameters. Near Tc the spin relaxation rate is influenced by the critical fluctuations of the order parameter.
As the complexity of quantum systems such as quantum bit arrays increases, efforts to automate expensive tuning are increasingly worthwhile. We investigate machine learning based tuning of gate arrays using the CMA-ES algorithm for the case study of Majorana wires with strong disorder. We find that the algorithm is able to efficiently improve the topological signatures, learn intrinsic disorder profiles, and completely eliminate disorder effects. For example, with only 20 gates, it is possible to fully recover Majorana zero modes destroyed by disorder by optimizing gate voltages.
We introduce a scheme for constructing partly occupied, maximally localized Wannier functions (WFs) for both molecular and periodic systems. Compared to the traditional occupied WFs the partly occupied WFs posses improved symmetry and localization properties achieved through a bonding-antibonding closing procedure. We demonstrate the equivalence between bonding-antibonding closure and the minimization of the average spread of the WFs in the case of a benzene molecule and a linear chain of Pt atoms. The general applicability of the method is demonstrated through the calculation of WFs for a metallic system with an impurity: a Pt wire with a hydrogen molecular bridge.
We study the prospects of detecting continuous gravitational waves (CGWs) from spinning neutron stars (NSs), gravitationally lensed by the galactic supermassive black hole. Assuming various astrophysically motivated spatial distributions of galactic NSs, we find that CGW signals from a few ($\sim 0-6$) neutron stars should be strongly lensed. Lensing will produce two copies of the signal (with time delays of seconds to minutes) that will interfere with each other. The relative motion of the NS with respect to the lensing optical axis will change the interference pattern, which will help us to identify a lensed signal. Accounting for the magnifications and time delays of the lensed signals, we investigate their detectability by ground-based detectors. Modelling the spin distribution of NSs based on that of known pulsars and assuming an ellipticity of $\epsilon = 10^{-7}$, lensed CGWs are unlikely to be detectable by LIGO and Virgo in realistic searches involving $\mathcal{O}(10^{12})$ templates. However, third generation detectors have a $\sim 2-51\%$ probability of detecting at least one lensed CGW signal. For an ellipticity of $\epsilon = 10^{-8}$, the detection probability reduces to $\sim 0-18 \, \% $. Though rare, such an observation will enable interesting probes of the supermassive black hole and its environment.
In this article we are describing a new algorithm for detecting and validating partial horizontal gene transfers (HGT). The presented algorithm is based on a sliding window procedure which analyzes fragments of the given multiple sequence alignment. A bootstrap procedure incorporated in our method can be used to estimate the support of each inferred partial HGT. The new algorithm can be also applied to confirm or discard complete (i.e., traditional) horizontal gene transfers detected by any HGT inferring algorithm. While working on a full-genome scale, the introduced algorithm can be used to assess the level of mosaicism of the whole species genomes as well as the rates of complete and partial HGT underlying the evolution of the considered set of species.
We show that if $n\geq 1$, $\Omega\subset \mathbb R^{n+1}$ is a connected domain with porous boundary, and $E\subset \partial\Omega$ is a set of finite and positive Hausdorff $H^{n}$-measure upon which the harmonic measure $\omega$ is absolutely continuous with respect to $H^{n}$, then $\omega|_E$ is concentrated on an $n$-rectifiable set.
Matrix representations of the Maxwell equations are well-known. However, all these representations lack an exactness or/and are given in terms of a {\em pair} of matrix equations. We present a matrix representation of the Maxwell equation in presence of sources in a medium with varying permittivity and permeability. It is shown that such a representation necessarily requires $8 \times 8$ matrices and an explicit representation for them is presented.
High-throughput ab-initio calculations, cluster expansion techniques and thermodynamic modeling have been synergistically combined to characterize the binodal and the spinodal decompositions features in the pseudo-binary lead chalcogenides PbSe-PbTe, PbS-PbTe, and PbS-PbSe. While our results agree with the available experimental data, our consolute temperatures substantially improve with respect to previous computational modeling. The computed phase diagrams corroborate that the formation of spinodal nanostructures causes low thermal conductivities in these alloys. The presented approach, making a rational use of online quantum repositories, can be extended to study thermodynamical and kinetic properties of materials of technological interest.
We report on the current-induced magnetization switching of a three-terminal perpendicular magnetic tunnel junction by spin-orbit torque and the read-out using the tunnelling magnetoresistance (TMR) effect. The device is composed of a perpendicular Ta/FeCoB/MgO/FeCoB stack on top of a Ta current line. The magnetization of the bottom FeCoB layer can be switched reproducibly by the injection of current pulses with density $5\times10^{11}$ A/m$^2$ in the Ta layer in the presence of an in-plane bias magnetic field, leading to the full-scale change of the TMR signal. Our work demonstrates the proof of concept of a perpendicular spin-orbit torque magnetic memory cell.
We develop two N=2 superfield formulations of free equations of motion for the joint model of all D=4 massless higher-superspin fields in generating form. The explicit Osp(2|4) supersymmetry is achieved without exploiting the harmonic superspace, and with adding no auxiliary component fields to those of N=1 superfields. The formulations are developed in two different Osp(2|4) homogeneous superspaces which have a structure of a fibre bundle over the standard D=4 AdS superspace, with dimensions (7|4) and (7|8). The N=2 covariant derivatives in these spaces are expressed in terms of N=1 ones which gives simple rules for component analysis.
Introductory lectures on the Kraichnan model of passive advection
Emerging trends in smartphones, online maps, social media, and the resulting geo-located data, provide opportunities to collect traces of people's socio-economical activities in a much more granular and direct fashion, triggering a revolution in empirical research. These vast mobile data offer new perspectives and approaches for measurements of economic dynamics and are broadening the research fields of social science and economics. In this paper, we explore the potential of using mobile big data for measuring economic activities of China. Firstly, We build indices for gauging employment and consumer trends based on billions of geo-positioning data. Secondly, we advance the estimation of store offline foot traffic via location search data derived from Baidu Maps, which is then applied to predict revenues of Apple in China and detect box-office fraud accurately. Thirdly, we construct consumption indicators to track the trends of various industries in service sector, which are verified by several existing indicators. To the best of our knowledge, we are the first to measure the second largest economy by mining such unprecedentedly large scale and fine granular spatial-temporal data. Our research provides new approaches and insights on measuring economic activities.
We present the analysis of the muon events with all muon multiplicities collected during 21804 hours of operation of the first LVD tower. The measured angular distribution of muon intensity has been converted to the `depth -- vertical intensity' relation in the depth range from 3 to 12 km w.e.. The analysis of this relation allowed to derive the power index, $\gamma$, of the primary all-nucleon spectrum: $\gamma=2.78 \pm 0.05$. The `depth -- vertical intensity' relation has been converted to standard rock and the comparison with the data of other experiments has been done. We present also the derived vertical muon spectrum at sea level.
For the first time, baryon-antibaryon photoproduction in the reaction $\gamma p \to \Lambda \bar{\Lambda} p$ has been observed at photon energies from threshold near 4.9 GeV to 11.6 GeV. The measurements are in progress with the GlueX spectrometer in Hall D at Jefferson Lab. We describe here the apparatus and methods used to make these measurements and outline the physics goals of the work. Some of the newly-seen reaction phenomenology is presented.
To date publish of a giant social network jointly from different parties is an easier collaborative approach. Agencies and researchers who collect such social network data often have a compelling interest in allowing others to analyze the data. In many cases the data describes relationships that are private and sharing the data in full can result in unacceptable disclosures. Thus, preserving privacy without revealing sensitive information in the social network is a serious concern. Recent developments for preserving privacy using anonymization techniques are focused on relational data only. Preserving privacy in social networks against neighborhood attacks is an initiation which uses the definition of privacy called k-anonymity. k-anonymous social network still may leak privacy under the cases of homogeneity and background knowledge attacks. To overcome, we find a place to use a new practical and efficient definition of privacy called ldiversity. In this paper, we take a step further on preserving privacy in collaborative social network data with algorithms and analyze the effect on the utility of the data for social network analysis.
We provide groupoid models for Toeplitz and Cuntz-Krieger algebras of topological higher-rank graphs. Extending the groupoid models used in the theory of graph algebras and topological dynamical systems to our setting, we prove results on essential freeness and amenability of the groupoids which capture the existing theory, and extend results involving group crossed products of graph algebras.
Compression schemes have been extensively used in Federated Learning (FL) to reduce the communication cost of distributed learning. While most approaches rely on a bounded variance assumption of the noise produced by the compressor, this paper investigates the use of compression and aggregation schemes that produce a specific error distribution, e.g., Gaussian or Laplace, on the aggregated data. We present and analyze different aggregation schemes based on layered quantizers achieving exact error distribution. We provide different methods to leverage the proposed compression schemes to obtain compression-for-free in differential privacy applications. Our general compression methods can recover and improve standard FL schemes with Gaussian perturbations such as Langevin dynamics and randomized smoothing.
We report the first laboratory and interstellar detection of the alpha-cyano vinyl radical (H2CCCN). This species was produced in the laboratory by an electric discharge of a gas mixture of vinyl cyanide, CH2CHCN, and Ne, and its rotational spectrum was characterized using a Balle-Flygare narrowband-type Fourier-transform microwave spectrometer operating in the frequency region of 8-40 GHz. The observed spectrum shows a complex structure due to tunneling splittings between two torsional sublevels of the ground vibronic state, 0+ and 0-, derived from a large-amplitude inversion motion. In addition, the presence of two equivalent hydrogen nuclei makes necessary to discern between ortho- and para-H2CCCN. A least squares analysis reproduces the observed transition frequencies with a standard deviation of ca. 3 kHz. Using the laboratory predictions, this radical is detected in the cold dark cloud TMC-1 using the Yebes 40m telescope and the QUIJOTE line survey. The 404-303 and 505-404 rotational transitions, composed of several hyperfine components, were observed in the 31.0-50.4 GHz range. Adopting a rotational temperature of 6K we derive a column density of (1.4+/-0.2)e11 cm-2 and (1.1+/-0.2)e11 cm-2 for ortho-H2CCCN and para-H2CCCN, respectively. The reactions C + CH3CN, and perhaps also N + CH2CCH, emerge as the most likely routes to H2CCCN in TMC-1.
Odometry forms an important component of many manned and autonomous systems. In the rail industry in particular, having precise and robust odometry is crucial for the correct operation of the Automatic Train Protection systems that ensure the safety of high-speed trains in operation around the world. Two problems commonly encountered in such odometry systems are miscalibration of the wheel encoders and slippage of the wheels under acceleration and braking, resulting in incorrect velocity estimates. This paper introduces an odometry system that addresses these problems. It comprises of an Extended Kalman Filter that tracks the calibration of the wheel encoders as state variables, and a measurement pre-processing stage called Sensor Consensus Analysis (SCA) that scales the uncertainty of a measurement based on how consistent it is with the measurements from the other sensors. SCA uses the statistical z-test to determine when an individual measurement is inconsistent with the other measurements, and scales the uncertainty until the z-test passes. This system is demonstrated on data from German Intercity-Express high-speed trains and it is shown to successfully deal with errors due to miscalibration and wheel slip.
A common problem in various applications is the additive decomposition of the output of a function with respect to its input variables. Functions with binary arguments can be axiomatically decomposed by the famous Shapley value. For the decomposition of functions with real arguments, a popular method is the pointwise application of the Shapley value on the domain. However, this pointwise application largely ignores the overall structure of functions. In this paper, axioms are developed which fully preserve functional structures and lead to unique decompositions for all Borel measurable functions.
Simple models are constructed for "acceleressence" dark energy: the latent heat of a phase transition occurring in a hidden sector governed by the seesaw mass scale v^2/M_Pl, where v is the electroweak scale and M_Pl the gravitational mass scale. In our models, the seesaw scale is stabilized by supersymmetry, implying that the LHC must discover superpartners with a spectrum that reflects a low scale of fundamental supersymmetry breaking. Newtonian gravity may be modified by effects arising from the exchange of fields in the acceleressence sector whose Compton wavelengths are typically of order the millimeter scale. There are two classes of models. In the first class the universe is presently in a metastable vacuum and will continue to inflate until tunneling processes eventually induce a first order transition. In the simplest such model, the range of the new force is bounded to be larger than 25 microns in the absence of fine-tuning of parameters, and for couplings of order unity it is expected to be \approx 100 microns. In the second class of models thermal effects maintain the present vacuum energy of the universe, but on further cooling, the universe will "soon" smoothly relax to a matter dominated era. In this case, the range of the new force is also expected to be of order the millimeter scale or larger, although its strength is uncertain. A firm prediction of this class of models is the existence of additional energy density in radiation at the eV era, which can potentially be probed in precision measurements of the cosmic microwave background. An interesting possibility is that the transition towards a matter dominated era has occurred in the very recent past, with the consequence that the universe is currently decelerating.
Graph coloring is one of the most famous computational problems with applications in a wide range of areas such as planning and scheduling, resource allocation, and pattern matching. So far coloring problems are mostly studied on static graphs, which often stand in stark contrast to practice where data is inherently dynamic and subject to discrete changes over time. A temporal graph is a graph whose edges are assigned a set of integer time labels, indicating at which discrete time steps the edge is active. In this paper we present a natural temporal extension of the classical graph coloring problem. Given a temporal graph and a natural number $\Delta$, we ask for a coloring sequence for each vertex such that (i) in every sliding time window of $\Delta$ consecutive time steps, in which an edge is active, this edge is properly colored (i.e. its endpoints are assigned two different colors) at least once during that time window, and (ii) the total number of different colors is minimized. This sliding window temporal coloring problem abstractly captures many realistic graph coloring scenarios in which the underlying network changes over time, such as dynamically assigning communication channels to moving agents. We present a thorough investigation of the computational complexity of this temporal coloring problem. More specifically, we prove strong computational hardness results, complemented by efficient exact and approximation algorithms. Some of our algorithms are linear-time fixed-parameter tractable with respect to appropriate parameters, while others are asymptotically almost optimal under the Exponential Time Hypothesis (ETH).
We describe a combined halo model to constrain the distribution of neutral hydrogen (HI) in the post-reionization universe. We combine constraints from the various probes of HI at different redshifts: the low-redshift 21-cm emission line surveys, intensity mapping experiments at intermediate redshifts, and the Damped Lyman-Alpha (DLA) observations at higher redshifts. We use a Markov Chain Monte Carlo (MCMC) approach to combine the observations and place constraints on the free parameters in the model. Our best-fit model involves a relation between neutral hydrogen mass $M_{\rm HI}$ and halo mass $M$ with a non-unit slope, and an upper and a lower cutoff. We find that the model fits all the observables but leads to an underprediction of the bias parameter of DLAs at $z \sim 2.3$. We also find indications of a possible tension between the HI column density distribution and the mass function of HI-selected galaxies at $z\sim 0$. We provide the central values of the parameters of the best-fit model so derived. We also provide a fitting form for the derived evolution of the concentration parameter of HI in dark matter haloes, and discuss the implications for the redshift evolution of the HI-halo mass relation.
The Meridional Overturning Circulation (MOC) is a system of surface and deep currents encompassing all ocean basins, crucial to the Earth's climate. Detecting potential climatic changes in the MOC first requires a careful characterisation of its inherent variability. Key components of the MOC are the Atlantic MOC (AMOC) and the Antarctic Circumpolar Current (ACC). The role of boundary properties in determining the AMOC and ACC is investigated, as a function of cross-sectional coordinate and depth, using a hierarchy of general circulation models. The AMOC is decomposed as the sum of near-surface Ekman, depth-independent bottom velocity and eastern and western boundary density components. The decomposition proves a useful low-dimensional characterisation of the full 3-D overturning circulation. The estimated total basin-wide AMOC overturning streamfunction, reconstructed using only boundary information, is in good agreement with direct calculations of the overturning using meridional velocities. The time-mean maximum overturning streamfunction is relatively constant with latitude, despite its underlying boundary contributions varying considerably, especially in the northern hemisphere. Applying a similar decomposition diagnostic to the ACC provides insight into the differences between model simulations, revealing spatial resolution dependence of ACC transport through the Drake Passage. All models exhibit a weak ACC compared to observations. Density maps along sloping ocean boundaries are produced, incorporating the western and eastern Atlantic boundaries and the Antarctic coastline. Isopycnals are flat over long portions of eastern ocean boundaries, and slope linearly on western boundaries. Results of this thesis serve to indicate the importance of boundary information in characterising the AMOC and the ACC, and the relative simplicity of along-sloping-boundary density structure.
In this work a comparison between different galaxy luminosity function estimators by means of Monte-Carlo simulations is presented. The simulations show that the C- method of Lynden-Bell (1971) and the STY method derived by Sandage, Tammann & Yahil (1979) are the best estimators to measure the shape of the luminosity function. The simulations also show that the STY estimator has a bias such that the faint-end slope is underestimated for steeper inclinations of the Schechter Function, and that this bias becomes quite severe when the sample contains only a few hundred objects. Overall, the C- is the most robust estimator, being less affected by different values of the faint end slope of the Schechter parameterization and sample size. The simulations are also used to compare different estimators of the luminosity function normalization. They demonstrate that most methods bias the recovered mean density towards values which are about 20% lower than the input value.
Optical Character Recognition (OCR), the task of extracting textual information from scanned documents is a vital and broadly used technology for digitizing and indexing physical documents. Existing technologies perform well for clean documents, but when the document is visually degraded, or when there are non-textual elements, OCR quality can be greatly impacted, specifically due to erroneous detections. In this paper we present an improved detection network with a masking system to improve the quality of OCR performed on documents. By filtering non-textual elements from the image we can utilize document-level OCR to incorporate contextual information to improve OCR results. We perform a unified evaluation on a publicly available dataset demonstrating the usefulness and broad applicability of our method. Additionally, we present and make publicly available our synthetic dataset with a unique hard-negative component specifically tuned to improve detection results, and evaluate the benefits that can be gained from its usage
In this note we analyse the relation between the triple-pomeron and Good-Walker formalisms for diffractive excitation in DIS and hadronic collisions. In both approaches gap events are interpreted as the shadow of absorption into inelastic channels. We here argue that the two formalisms are just different views of the same phenomenon. We first demonstrate how this relation works in a simple toy model, and then show how the relevant features of the toy model are also realized in real perturbative QCD.
Detecting and analyzing various defect types in semiconductor materials is an important prerequisite for understanding the underlying mechanisms as well as tailoring the production processes. Analysis of microscopy images that reveal defects typically requires image analysis tasks such as segmentation and object detection. With the permanently increasing amount of data that is produced by experiments, handling these tasks manually becomes more and more impossible. In this work, we combine various image analysis and data mining techniques for creating a robust and accurate, automated image analysis pipeline. This allows for extracting the type and position of all defects in a microscopy image of a KOH-etched 4H-SiC wafer that was stitched together from approximately 40,000 individual images.
Mixtures of polarised fermions of two different masses can form weakly-bound clusters, such as dimers and trimers, that are universally described by the scattering length between the heavy and light fermions. We use the resonating group method to investigate the low-energy scattering processes involving dimers or trimers. The method reproduces approximately the known particle-dimer and dimer-dimer scattering lengths. We use it to estimate the trimer-trimer scattering length, which is presently unknown, and find it to be positive.
Fontanari et al introduced [Phys. Rev. Lett. 91, 218101 (2003)] a model for studying the Muller's ratchet phenomenon in growing asexual populations. They studied two situations, either including or not a death probability for each newborn, but were able to find analytical (recursive) expressions only in the no-decay case. In this paper a branching process formalism is used to find recorrence equations that generalize the analytical results of the original paper besides confirming the interesting effects their simulations revealed.
Few-shot Named Entity Recognition (NER) exploits only a handful of annotations to identify and classify named entity mentions. Prototypical network shows superior performance on few-shot NER. However, existing prototypical methods fail to differentiate rich semantics in other-class words, which will aggravate overfitting under few shot scenario. To address the issue, we propose a novel model, Mining Undefined Classes from Other-class (MUCO), that can automatically induce different undefined classes from the other class to improve few-shot NER. With these extra-labeled undefined classes, our method will improve the discriminative ability of NER classifier and enhance the understanding of predefined classes with stand-by semantic knowledge. Experimental results demonstrate that our model outperforms five state-of-the-art models in both 1-shot and 5-shots settings on four NER benchmarks. We will release the code upon acceptance. The source code is released on https: //github.com/shuaiwa16/OtherClassNER.git.
Collective plasmon excitations in solids that result from the process of photoemission are an important area of fundamental research. In this study, we identify a significant number ($n$) of multiple bulk plasmons ($n\omega_p$) in the hard x-ray photoelectron spectra of the core levels and valence bands (VBs) of two well-known, nearly free electron metals, aluminum (Al) and magnesium (Mg). On the basis of earlier theoretical works, we estimate the contributions of extrinsic, intrinsic, and interference processes to the intensities of 1$s$ to 2$s$ core level plasmons. The intrinsic contribution diminishes from 22% for 1$\omega_p$, to 4.4% for 2$\omega_p$, and becomes negligible thereafter (0.5% for 3$\omega_p$). The extrinsic and intrinsic plasmon contributions do not vary significantly across a broad range of photoelectron kinetic energies, and also between the two metals (Al and Mg). The interference contribution varies from negative to zero as $n$ increases. An asymmetric line shape is observed for the bulk plasmons, which is most pronounced for 1$\omega_p$. Signature of the surface plasmon is detected in normal emission, and it exhibits a significantly increased intensity in the grazing emission. The VB spectra of Al and Mg, which are dominated by $s$-like states, exhibit excellent agreement with the calculated VB based on density functional theory. The VB exhibits four multiple bulk plasmon peaks in the loss region, which are influenced by an intrinsic process in addition to the extrinsic process. On a completely oxidized aluminum surface, the relative intensity of the Al metal bulk plasmon remains nearly unaltered, while the surface plasmon is completely attenuated.
We reconstruct dark energy properties from two complementary supernova datasets -- the newly released Gold+HST sample and SNLS. The results obtained are consistent with standard $\Lambda$CDM model within $2\sigma$ error bars although the Gold+HST data favour evolving dark energy slightly more than SNLS. Using complementary data from baryon acoustic oscillations and the cosmic microwave background to constrain dark energy, we find that our results in this case are strongly dependent on the present value of the matter density $\Omega_m$. Consequently, no firm conclusions regarding constancy or variability of dark energy density can be drawn from these data alone unless the value of $\Omega_m$ is known to an accuracy of a few percent. However, possible variability is significantly restricted if this data is used in conjunction with supernova data.
A central issue in the design of tokamaks or stellarators is the coils that produce the external magnetic fields. The freedom that remains unstudied in the design of coils is enormous. This freedom could be quickly studied computationally at low cost with high reliability. In particular, the space between toroidal field or modular coils that block access to the plasma chamber could be increased by a large factor. This paper explains how this could be done using the concept of current-potential patches that was developed in Todd Elder's thesis.
We present a scheme for symmetric multiparty quantum state sharing of an arbitrary $m$-qubit state with $m$ Greenberger-Horne-Zeilinger states following some ideas from the controlled teleportation [Phys. Rev. A \textbf{72}, 02338 (2005)]. The sender Alice performs $m$ Bell-state measurements on her $2m$ particles and the controllers need only to take some single-photon product measurements on their photons independently, not Bell-state measurements, which makes this scheme more convenient than the latter. Also it does not require the parties to perform a controlled-NOT gate on the photons for reconstructing the unknown $m$-qubit state and it is an optimal one as its efficiency for qubits approaches the maximal value.
We study multiple zeta values (MZVs) from the viewpoint of zeta-functions associated with the root systems which we have studied in our previous papers. In fact, the $r$-ple zeta-functions of Euler-Zagier type can be regarded as the zeta-function associated with a certain sub-root system of type $C_r$. Hence, by the action of the Weyl group, we can find new aspects of MZVs which imply that the well-known formula for MZVs given by Hoffman and Zagier coincides with Witten's volume formula associated with the above sub-root system of type $C_r$. Also, from this observation, we can prove some new formulas which especially include the parity results of double and triple zeta values. As another important application, we give certain refinement of restricted sum formulas, which gives restricted sum formulas among MZVs of an arbitrary depth $r$ which were previously known only in the cases of depth $2,3,4$. Furthermore, considering a sub-root system of type $B_r$ analogously, we can give relevant analogues of the Hoffman-Zagier formula, parity results and restricted sum formulas.
Presence-only records may provide data on the distributions of rare species, but commonly suffer from large, unknown biases due to their typically haphazard collection schemes. Presence-absence or count data collected in systematic, planned surveys are more reliable but typically less abundant. We proposed a probabilistic model to allow for joint analysis of presence-only and survey data to exploit their complementary strengths. Our method pools presence-only and presence-absence data for many species and maximizes a joint likelihood, simultaneously estimating and adjusting for the sampling bias affecting the presence-only data. By assuming that the sampling bias is the same for all species, we can borrow strength across species to efficiently estimate the bias and improve our inference from presence-only data. We evaluate our model's performance on data for 36 eucalypt species in southeastern Australia. We find that presence-only records exhibit a strong sampling bias toward the coast and toward Sydney, the largest city. Our data-pooling technique substantially improves the out-of-sample predictive performance of our model when the amount of available presence-absence data for a given species is scarce. If we have only presence-only data and no presence-absence data for a given species, but both types of data for several other species that suffer from the same spatial sampling bias, then our method can obtain an unbiased estimate of the first species' geographic range.
Relativistic Riemannian superfluid hydrodynamics used in general relativity to investigate superfluids in pulsars is extended to non-Riemannian background spacetime endowed with Cartan torsion. From the Gross-Pitaeviskii (GP) it is shown that in the weak field Cartan torsion approximation, the torsion vector is orthogonal to the superfluid plane wave velocity. Torsion vector is also shown to be aligned along the vortex direction in the superfluid. The background torsion is shown to induce rotation on the fluid as happens with the acoustic torsion in the analogue non-Riemannian non-relativistic superfluid models. The torsion part of the current would be connected to the normal part of the superfluid velocity while the Riemannian part of the velocity would be connected to the superfluid velocity itself. Magnus effect and the rotation of the superfluid are analysed. Since the Kalb-Ramond field is easily associated with torsion our method seems to be equivalent to the vortex-cosmic string relativistic superfluid method developed by Carter and Langlois to investigate rotating neutron stars.
We present a semi-parametric approach to photographic image synthesis from semantic layouts. The approach combines the complementary strengths of parametric and nonparametric techniques. The nonparametric component is a memory bank of image segments constructed from a training set of images. Given a novel semantic layout at test time, the memory bank is used to retrieve photographic references that are provided as source material to a deep network. The synthesis is performed by a deep network that draws on the provided photographic material. Experiments on multiple semantic segmentation datasets show that the presented approach yields considerably more realistic images than recent purely parametric techniques. The results are shown in the supplementary video at https://youtu.be/U4Q98lenGLQ
Some aspects of light-like compactifications of superstring theory and their implications for the matrix model of M-theory are discussed.
We provide a generalization of the Lie algebra of conformal Killing vector fields to conformal Killing-Yano forms. A new Lie bracket for conformal Killing-Yano forms that corresponds to slightly modified Schouten-Nijenhuis bracket of differential forms is proposed. We show that conformal Killing-Yano forms satisfy a graded Lie algebra in constant curvature manifolds. It is also proven that normal conformal Killing-Yano forms in Einstein manifolds also satisfy a graded Lie algebra. The constructed graded Lie algebras reduce to the graded Lie algebra of Killing-Yano forms and the Lie algebras of conformal Killing and Killing vector fields in special cases.
We study the applicability of composite fermion theory to electrons in two-dimensional parabolically-confined quantum dots in a strong perpendicular magnetic field in the limit of low Zeeman energy. The non-interacting composite fermion spectrum correctly specifies the primary features of this system. Additional features are relatively small, indicating that the residual interaction between the composite fermions is weak. \footnote{Published in Phys. Rev. B {\bf 52}, 2798 (1995).}
The present article analyses the impact on cosmology, in particular on the evolution of cosmological perturbations, of the existence of extra-dimensions. The model considered here is that of a five-dimensional Anti-de Sitter spacetime where ordinary matter is confined to a brane-universe. The homogeneous cosmology is recalled. The equations governing the evolution of cosmological perturbations are presented in the most transparent way: they are rewritten in a form very close to the equations of standard cosmology with two types of corrections: a. corrections due to the unconventional evolution of the homogeneous solution, which change the background-dependent coefficients of the equations; b. corrections due to the curvature along the fifth dimension, which act as source terms in the evolution equations.
Determining the poverty levels of various regions throughout the world is crucial in identifying interventions for poverty reduction initiatives and directing resources fairly. However, reliable data on global economic livelihoods is hard to come by, especially for areas in the developing world, hampering efforts to both deploy services and monitor/evaluate progress. This is largely due to the fact that this data is obtained from traditional door-to-door surveys, which are time consuming and expensive. Overhead satellite imagery contain characteristics that make it possible to estimate the region's poverty level. In this work, I develop deep learning computer vision methods that can predict a region's poverty level from an overhead satellite image. I experiment with both daytime and nighttime imagery. Furthermore, because data limitations are often the barrier to entry in poverty prediction from satellite imagery, I explore the impact that data quantity and data augmentation have on the representational power and overall accuracy of the networks. Lastly, to evaluate the robustness of the networks, I evaluate them on data from continents that were absent in the development set.
We present Paper II of the Eccentric Debris Disc Morphologies series to explore the effects that significant free and forced eccentricities have on high-resolution millimetre-wavelength observations of debris discs, motivated by recent ALMA images of HD53143's disc. In this work, we explore the effects of free eccentricity, and by varying disc fractional widths and observational resolutions, show for a range of narrow eccentric discs, orbital overlaps result in dust emission distributions that have either one or two radial peaks at apocentre and/or pericentre. The narrowest discs contain two radial peaks, whereas the broadest discs contain just one radial peak. For fixed eccentricities, as fractional disc widths are increased, we show that these peaks merge first at apocentre (producing apocentre glow), and then at pericentre (producing pericentre glow). Our work thus demonstrates that apocentre/pericentre glows in models with constant free and forced eccentricities can be both width and resolution dependent at millimetre wavelengths, challenging the classical assertion that apocentre/pericentre glows are purely wavelength dependent. We discuss future high-resolution observations that can distinguish between competing interpretations of underlying debris disc eccentricity distributions.
We discuss an extension of the scalar auxiliary variable approach, which was originally introduced by Shen et al. ([Shen, Xu, Yang, J. Comput. Phys., 2018]) for the discretization of deterministic gradient flows. By introducing an additional scalar auxiliary variable, this approach allows to derive a linear scheme, while still maintaining unconditional stability. Our extension augments the approximation of the evolution of this scalar auxiliary variable with higher order terms, which enables its application to stochastic partial differential equations. Using the stochastic Allen--Cahn equation as a prototype for nonlinear stochastic partial differential equations with multiplicative noise, we propose an unconditionally energy stable, linear, fully discrete finite element scheme based on our augmented scalar auxiliary variable method. Recovering a discrete version of the energy estimate and establishing Nikolskii estimates with respect to time, we are able to prove convergence of discrete solutions towards pathwise unique martingale solutions by applying Jakubowski's generalization of Skorokhod's theorem. A generalization of the Gy\"ongy--Krylov characterization of convergence in probability to quasi-Polish spaces finally provides convergence of fully discrete solutions towards strong solutions of the stochastic Allen--Cahn equation. Finally, we present numerical simulations underlining the practicality of the scheme and the importance of the introduced augmentation terms.
We present analytical results for the $O(\alpha _s ^2)$ contributions to the functions $\eta _A$ and $\eta _V$ which parameterize QCD corrections to semileptonic $b \to c$ transitions at zero recoil. Previously obtained approximate results are confirmed. The methods of computing the relevant two-loop diagrams with two mass scales are discussed in some detail.
The newly discovered topological Dirac semimetals host the possibilities of various topological phase transitions through the control of spin-orbit coupling as well as symmetries and dimensionalities. Here, we report a magnetotransport study of high-mobility (Cd1-xZnx)3As2 films, where the topological Dirac semimetal phase can be turned into a trivial insulator via chemical substitution. By high-field measurements with a Hall-bar geometry, magnetoresistance components ascribed to the chiral charge pumping have been distinguished from other extrinsic effects. The negative magnetoresistance exhibits a clear suppression upon Zn doping, reflecting decreasing Berry curvature of the band structure as the topological phase transition is induced by reducing the spin-orbit coupling.
Given a graph $G=(V,E)$ on $n$ vertices and an assignment of colours to its edges, a set of edges $S \subseteq E$ is said to be rainbow if edges from $S$ have pairwise different colours assigned to them. In this paper, we investigate rainbow spanning trees in randomly coloured random $G_{k-out}$ graphs.
In this article, we adapted five recent SSL methods to the task of audio classification. The first two methods, namely Deep Co-Training (DCT) and Mean Teacher (MT), involve two collaborative neural networks. The three other algorithms, called MixMatch (MM), ReMixMatch (RMM), and FixMatch (FM), are single-model methods that rely primarily on data augmentation strategies. Using the Wide-ResNet-28-2 architecture in all our experiments, 10% of labeled data and the remaining 90% as unlabeled data for training, we first compare the error rates of the five methods on three standard benchmark audio datasets: Environmental Sound Classification (ESC-10), UrbanSound8K (UBS8K), and Google Speech Commands (GSC). In all but one cases, MM, RMM, and FM outperformed MT and DCT significantly, MM and RMM being the best methods in most experiments. On UBS8K and GSC, MM achieved 18.02% and 3.25% error rate (ER), respectively, outperforming models trained with 100% of the available labeled data, which reached 23.29% and 4.94%, respectively. RMM achieved the best results on ESC-10 (12.00% ER), followed by FM which reached 13.33%. Second, we explored adding the mixup augmentation, used in MM and RMM, to DCT, MT, and FM. In almost all cases, mixup brought consistent gains. For instance, on GSC, FM reached 4.44% and 3.31% ER without and with mixup. Our PyTorch code will be made available upon paper acceptance at https:// github. com/ Labbe ti/ SSLH.
We consider an $\varepsilon$-periodic ($\varepsilon\to 0$) tubular structure, modelled as a magnetic Laplacian on a metric graph, which is periodic along a single axis. We show that the corresponding Hamiltonian admits norm-resolvent convergence to an ODE on $\mathbb{R}$ which is fourth order at a discrete set of values of the magnetic potential (\emph{critical points}) and second-order generically. In a vicinity of critical points we establish a mixed-order asymptotics. The rate of convergence is also estimated. This represents a physically viable model of a phase transition as the strength of the (constant) magnetic field increases.
Falling oil revenues and rapid urbanization are putting a strain on the budgets of oil producing nations which often subsidize domestic fuel consumption. A direct way to decrease the impact of subsidies is to reduce fuel consumption by reducing congestion and car trips. While fuel consumption models have started to incorporate data sources from ubiquitous sensing devices, the opportunity is to develop comprehensive models at urban scale leveraging sources such as Global Positioning System (GPS) data and Call Detail Records. We combine these big data sets in a novel method to model fuel consumption within a city and estimate how it may change due to different scenarios. To do so we calibrate a fuel consumption model for use on any car fleet fuel economy distribution and apply it in Riyadh, Saudi Arabia. The model proposed, based on speed profiles, is then used to test the effects on fuel consumption of reducing flow, both randomly and by targeting the most fuel inefficient trips in the city. The estimates considerably improve baseline methods based on average speeds, showing the benefits of the information added by the GPS data fusion. The presented method can be adapted to also measure emissions. The results constitute a clear application of data analysis tools to help decision makers compare policies aimed at achieving economic and environmental goals.
We determine a class of ringed space X, for which the category of locally free sheaves of bounded rank is equivalent to the category of finitely generated projective A(X)-modules, where A(X) denote the ring of global sections of X. The well-known Serre-Swan theorems for affine schemes, differentiable manifolds, Stein spaces, etc., are then derived.
We present in detail a calculation of the next-to-leading order QCD corrections to the process $e^+e^-\to 3$ jets with massive quarks. To isolate the soft and collinear divergencies of the four parton matrix elements, we modify the phase space slicing method to account for masses. Our computation allows for the prediction of oriented three jet events involving heavy quarks, both on and off the Z resonance, and of any event shape variable which is dominated by three jet configurations. We show next-to-leading order results for the three jet fraction, the differential two jet rate, and for the thrust distribution at various c.m. energies.
We benchmark contemporary action recognition models (TSN, TRN, and TSM) on the recently introduced EPIC-Kitchens dataset and release pretrained models on GitHub (https://github.com/epic-kitchens/action-models) for others to build upon. In contrast to popular action recognition datasets like Kinetics, Something-Something, UCF101, and HMDB51, EPIC-Kitchens is shot from an egocentric perspective and captures daily actions in-situ. In this report, we aim to understand how well these models can tackle the challenges present in this dataset, such as its long tail class distribution, unseen environment test set, and multiple tasks (verb, noun and, action classification). We discuss the models' shortcomings and avenues for future research.
Numerical simulations of vesicle suspensions are performed in two dimensions to study their dynamical and rheological properties. An hybrid method is adopted, which combines a mesoscopic approach for the solvent with a curvature-elasticity model for the membrane. Shear flow is induced by two counter-sliding parallel walls, which generate a linear flow profile. The flow behavior is studied for various vesicle concentrations and viscosity ratios between the internal and the external fluid. Both the intrinsic viscosity and the thickness of depletion layers near the walls are found to increase with increasing viscosity ratio.
Many factors influence speech yielding different renditions of a given sentence. Generative models, such as variational autoencoders (VAEs), capture this variability and allow multiple renditions of the same sentence via sampling. The degree of prosodic variability depends heavily on the prior that is used when sampling. In this paper, we propose a novel method to compute an informative prior for the VAE latent space of a neural text-to-speech (TTS) system. By doing so, we aim to sample with more prosodic variability, while gaining controllability over the latent space's structure. By using as prior the posterior distribution of a secondary VAE, which we condition on a speaker vector, we can sample from the primary VAE taking explicitly the conditioning into account and resulting in samples from a specific region of the latent space for each condition (i.e. speaker). A formal preference test demonstrates significant preference of the proposed approach over standard Conditional VAE. We also provide visualisations of the latent space where well-separated condition-specific clusters appear, as well as ablation studies to better understand the behaviour of the system.
We review some recent progress on applications of Cluster Expansions. We focus on a system of classical particles living in a continuous medium and interacting via a stable and tempered pair potential. We review the cluster expansion in both the canonical and the grand canonical ensemble and compute thermodynamic quantities such as the pressure, the free energy as well as various correlation functions. We derive the equation of state either by performing inversion of the density-activity series or directly in the canonical ensemble. Further applications to the liquid state expansions and the relevant closures are discussed, in particular their convergence in the gas regime.
The radio-wavelength detection of extensive air showers (EAS) initiated by cosmic-ray interactions in the Earth's atmosphere is a promising technique for investigating the origin of these particles and the physics of their interactions. The Low Frequency Array (LOFAR) and the Owens Valley Long Wavelength Array (OVRO-LWA) have both demonstrated that the dense cores of low frequency radio telescope arrays yield detailed information on the radiation ground pattern, which can be used to reconstruct key EAS properties and infer the primary cosmic-ray composition. Here, we demonstrate a new observation mode of the Murchison Widefield Array (MWA), tailored to the observation of the sub-microsecond coherent bursts of radiation produced by EAS. We first show how an aggregate 30.72 MHz bandwidth (3072x 10 kHz frequency channels) recorded at 0.1 ms resolution with the MWA's voltage capture system (VCS) can be synthesised back to the full bandwidth Nyquist resolution of 16.3 ns. This process, which involves `inverting' two sets of polyphase filterbanks, retains 90.5% of the signal-to-noise of a cosmic ray signal. We then demonstrate the timing and positional accuracy of this mode by resolving the location of a calibrator pulse to within 5 m. Finally, preliminary observations show that the rate of nanosecond radio-frequency interference (RFI) events is 0.1 Hz, much lower than that found at the sites of other radio telescopes that study cosmic rays. We conclude that the identification of cosmic rays at the MWA, and hence with the low-frequency component of the Square Kilometre Array, is feasible with minimal loss of efficiency due to RFI.
MoTe2 is a paradigmatic van der Waals layered semimetal with two energetically close electronic phases, the topologically trivial 1Tprime and the low-temperature Td type-II Weyl semimetal phase. The ability to manipulate this phase transition, perhaps towards occurring near room temperature, would open new avenues for harnessing the full potential of Weyl semimetals for high-efficiency electronic and spintronic applications. Here, we show that potassium dosing on 1Tprime-MoTe2 induces a Lifshitz transition by a combination of angle-resolved photoemission spectroscopy, scanning tunneling microscopy, x-ray spectroscopy and density functional theory. While the electronic structure shifts rigidly for small concentrations of K, MoTe2 undergoes significant band structure renormalization for larger concentrations. Our results demonstrate that the origin of this electronic structure change stems from alkali metal intercalation. We show that these profound changes are caused by effectively decoupling the 2D sheets, bringing K-intercalated 1Tprime-MoTe2 to the quasi-2D limit, but do not cause a topological phase transition.
We consider the problem of explaining the decisions of deep neural networks for image recognition in terms of human-recognizable visual concepts. In particular, given a test set of images, we aim to explain each classification in terms of a small number of image regions, or activation maps, which have been associated with semantic concepts by a human annotator. This allows for generating summary views of the typical reasons for classifications, which can help build trust in a classifier and/or identify example types for which the classifier may not be trusted. For this purpose, we developed a user interface for "interactive naming," which allows a human annotator to manually cluster significant activation maps in a test set into meaningful groups called "visual concepts". The main contribution of this paper is a systematic study of the visual concepts produced by five human annotators using the interactive naming interface. In particular, we consider the adequacy of the concepts for explaining the classification of test-set images, correspondence of the concepts to activations of individual neurons, and the inter-annotator agreement of visual concepts. We find that a large fraction of the activation maps have recognizable visual concepts, and that there is significant agreement between the different annotators about their denotations. Our work is an exploratory study of the interplay between machine learning and human recognition mediated by visualizations of the results of learning.
We review various theoretical methods for measuring dark matter properties at the Large Hadron Collider.
We report on the first analysis of AstroSat observation of the Z-source GX 5- 1 on February 26-27, 2017. The hardness-intensity plot reveals that the source traced out the horizontal and normal branches. The 0.8-20 keV spectra from simultaneous SXT and LAXPC data at different locations of the hardness-intensity plot can be well described by a disk emission and a thermal Comptonized component. The ratio of the disk flux to the total i.e. the disk flux ratio increases monotonically along the horizontal to the normal one. Thus, the difference between the normal and horizontal branches is that in the normal branch, the disk dominates the flux while in the horizontal one it is the Comptonized component which dominates. The disk flux scales with the inner disk temperature as T_{in}^{5.5} and not as T_{in}{4} suggesting that either the inner radii changes dramatically or that the disk is irradiated by the thermal component changing its hardness factor. The power spectra reveal a Quasi Periodic Oscillation whose frequency changes from \sim 30 Hz to 50 Hz. The frequency is found to correlate well with the disk flux ratio. In the 3-20 keV LAXPC band the r.m.s of the QPO increases with energy (r.m.s \prop E0.8), while the harder X-ray seems to lag the soft ones with a time-delay of a milliseconds. The results suggest that the spectral properties of the source are characterized by the disk flux ratio and that the QPO has its origin in the corona producing the thermal Comptonized component.
The electroproduction of J/psi and psi(2S) mesons is studied in elastic, quasi-elastic and inclusive reactions for four momentum transfers 2 < Q^2 < 80 GeV^2 and photon-proton centre of mass energies 25 < W < 180 GeV. The data were taken with the H1 detector at the electron proton collider HERA in the years 1995 to 1997. The total virtual photon-proton cross section for elastic J/psi production is measured as a function of Q^2 and W. The dependence of the production rates on the square of the momentum transfer from the proton (t) is extracted. Decay angular distributions are analysed and the ratio of the longitudinal and transverse cross sections is derived. The ratio of the cross sections for quasi-elastic psi(2S) and J/psi meson production is measured as a function of Q^2. The results are discussed in terms of theoretical models based upon perturbative QCD. Differential cross sections for inclusive and inelastic production of J/psi mesons are determined and predictions within two theoretical frameworks are compared with the data, the non-relativistic QCD factorization approach including colour octet and colour singlet contributions, and the model of Soft Colour Interactions.
We provide the last missing piece of the complete non-perturbative description of the low energy effective action emerging from Calabi-Yau compactifications of type II string theory --- NS5-brane instanton corrections to the hypermultiplet moduli space $M_H$. We find them using S-duality symmetry of the type IIB formulation. The result is encoded in a set of holomorphic functions on the twistor space of $M_H$ and includes all orders of the instanton expansion.
Face clustering plays an essential role in exploiting massive unlabeled face data. Recently, graph-based face clustering methods are getting popular for their satisfying performances. However, they usually suffer from excessive memory consumption especially on large-scale graphs, and rely on empirical thresholds to determine the connectivities between samples in inference, which restricts their applications in various real-world scenes. To address such problems, in this paper, we explore face clustering from the pairwise angle. Specifically, we formulate the face clustering task as a pairwise relationship classification task, avoiding the memory-consuming learning on large-scale graphs. The classifier can directly determine the relationship between samples and is enhanced by taking advantage of the contextual information. Moreover, to further facilitate the efficiency of our method, we propose a rank-weighted density to guide the selection of pairs sent to the classifier. Experimental results demonstrate that our method achieves state-of-the-art performances on several public clustering benchmarks at the fastest speed and shows a great advantage in comparison with graph-based clustering methods on memory consumption.
Quantum trajectories describe the stochastic evolution of an open quantum system conditioned on continuous monitoring of its output, such as by an ideal photodetector. In practice an experimenter has access to an output filtered through various electronic devices, rather than the microscopic states of the detector. This introduces several imperfections into the measurement process, of which only inefficiency has previously been incorporated into quantum trajectory theory. However, all electronic devices have finite bandwidths, and the consequent delay in conveying the output signal to the observer implies that the evolution of the conditional state of the quantum system must be non-Markovian. We present a general method of describing this evolution and apply it to avalanche photodiodes (APDs) and to photoreceivers. We include the effects of efficiency, dead time, bandwidth, electronic noise, and dark counts. The essential idea is to treat the quantum system and classical detector jointly, and to average over the latter to obtain the conditional quantum state. The significance of our theory is that quantum trajectories for realistic detection are necessary for sophisticated approaches to quantum feedback, and our approach could be applied in many areas of physics.
Consider a graph with a rotation system, namely, for every vertex, a circular ordering of the incident edges. Given such a graph, an angle cover maps every vertex to a pair of consecutive edges in the ordering -- an angle -- such that each edge participates in at least one such pair. We show that any graph of maximum degree 4 admits an angle cover, give a poly-time algorithm for deciding if a graph with no degree-3 vertices has an angle-cover, and prove that, given a graph of maximum degree 5, it is NP-hard to decide whether it admits an angle cover. We also consider extensions of the angle cover problem where every vertex selects a fixed number $a>1$ of angles or where an angle consists of more than two consecutive edges. We show an application of angle covers to the problem of deciding if the 2-blowup of a planar graph has isomorphic thickness 2.
The Smoothed Particles Hydrodynamics (SPH) is a particle-based, meshfree, Lagrangian method used to simulate multidimensional fluids with arbitrary geometries, most commonly employed in astrophysics, cosmology, and computational fluid-dynamics (CFD). It is expected that these computationally-demanding numerical simulations will significantly benefit from the up-and-coming Exascale computing infrastructures, that will perform 10 18 FLOP/s. In this work, we review the status of a novel SPH-EXA mini-app, which is the result of an interdisciplinary co-design project between the fields of astrophysics, fluid dynamics and computer science, whose goal is to enable SPH simulations to run on Exascale systems. The SPH-EXA mini-app merges the main characteristics of three state-of-the-art parent SPH codes (namely ChaNGa, SPH-flow, SPHYNX) with state-of-the-art (parallel) programming, optimization, and parallelization methods. The proposed SPH-EXA mini-app is a C++14 lightweight and flexible header-only code with no external software dependencies. Parallelism is expressed via multiple programming models, which can be chosen at compilation time with or without accelerator support, for a hybrid process+thread+accelerator configuration. Strong and weak-scaling experiments on a production supercomputer show that the SPH-EXA mini-app can be efficiently executed with up 267 million particles and up to 65 billion particles in total on 2,048 hybrid CPU-GPU nodes.
We propose a method to account for model error due to unresolved scales in the context of the ensemble transform Kalman filter (ETKF). The approach extends to this class of algorithms the deterministic model error formulation recently explored for variational schemes and extended Kalman filter. The model error statistic required in the analysis update is estimated using historical reanalysis increments and a suitable model error evolution law. Two different versions of the method are described; a time-constant model error treatment where the same model error statistical description is time-invariant, and a time-varying treatment where the assumed model error statistics is randomly sampled at each analysis step. We compare both methods with the standard method of dealing with model error through inflation and localization, and illustrate our results with numerical simulations on a low order nonlinear system exhibiting chaotic dynamics. The results show that the filter skill is significantly improved through the proposed model error treatments, and that both methods require far less parameter tuning than the standard approach. Furthermore, the proposed approach is simple to implement within a pre-existing ensemble based scheme. The general implications for the use of the proposed approach in the framework of square-root filters such as the ETKF are also discussed.
We numerically investigate density perturbations generated in the smooth hybrid new inflation model, a kind of double inflation model that is designed to reproduce the running spectral index suggested by the WMAP results. We confirm that this model provides the running spectral index within 1sigma range of the three year WMAP result. In addition, we find a sharp and strong peak on the spectrum of primordial curvature perturbation at small scales. This originates from amplification of fluctuation in the first inflaton fields due to parametric resonance, which takes place in the oscillatory phase between two inflationary regime. Formation probability of primordial black holes (PBHs) is discussed as a consequence of such peak.
In this note we focus on three independent problems on Okounkov bodies for projective varieties. The main goal is to present a geometric version of the classical Fujita Approximation Theorem, a Jow-type theorem and a cardinality formulae for Minkowski bases on a certain class of smooth projective surfaces.
Over the last decade, simultaneous wireless information and power transfer (SWIPT) has become a practical and promising solution for connecting and recharging battery-limited devices, thanks to significant advances in low-power electronics technology and wireless communications techniques. To realize the promised potentials, advanced resource allocation design plays a decisive role in revealing, understanding, and exploiting the intrinsic rate-energy tradeoff capitalizing on the dual use of radio frequency (RF) signals for wireless charging and communication. In this paper, we provide a comprehensive tutorial overview of SWIPT from the perspective of resource allocation design. The fundamental concepts, system architectures, and RF energy harvesting (EH) models are introduced. In particular, three commonly adopted EH models, namely the linear EH model, the nonlinear saturation EH model, and the nonlinear circuit-based EH model are characterized and discussed. Then, for a typical wireless system setup, we establish a generalized resource allocation design framework which subsumes conventional resource allocation design problems as special cases. Subsequently, we elaborate on relevant tools from optimization theory and exploit them for solving representative resource allocation design problems for SWIPT systems with and without perfect channel state information (CSI) available at the transmitter, respectively. The associated technical challenges and insights are also highlighted. Furthermore, we discuss several promising and exciting future research directions for resource allocation design for SWIPT systems intertwined with cutting-edge communication technologies, such as intelligent reflecting surfaces, unmanned aerial vehicles, mobile edge computing, federated learning, and machine learning.
The frequency shift of light in the gravitational field generated by a rotating body is investigated. We consider the scenario in which both the light source and the observer are in motion. The frequency shift is calculated up to the second-order post-Minkowskian approximation via two different methods and the same result is achieved. The higher-order effects of the gravitational source's rotation on the frequency shift is obtained. Especially, when both the light source and the observer are located in the asymptotically flat region, an elegant formula is obtained, which can be easily used in the astronomical observations to determine the rotating gravitational source's mass and angular momentum.
The first author in recent work with D. Gay developed the notion of a Morse structure on an open book as a tool for studying closed contact 3-manifolds. We extend the notion of Morse structure to extendable partial open books in order to study contact 3-manifolds with convex boundary.
We introduce a lattice model of interacting spins and bosons that leads to Luttinger-liquid physics, and allows for quantitative tests of the theory of bosonization by means of trapped-ion or superconducting-circuit experiments. By using a variational bosonization ansatz, we calculate the power-law decay of spin and boson correlation functions, and study their dependence on a single tunable parameter, namely a bosonic driving. For small drivings, Matrix-Product-States (MPS) numerical methods are shown to be efficient and validate our ansatz. Conversely, even static MPS become inefficient for large-driving regimes, such that the experiment can potentially outperform classical numerics, achieving one of the goals of quantum simulations.
Providing secure communications over the physical layer with the objective of achieving perfect secrecy without requiring a secret key has been receiving growing attention within the past decade. The vast majority of the existing studies in the area of physical layer security focus exclusively on the scenarios where the channel inputs are Gaussian distributed. However, in practice, the signals employed for transmission are drawn from discrete signal constellations such as phase shift keying and quadrature amplitude modulation. Hence, understanding the impact of the finite-alphabet input constraints and designing secure transmission schemes under this assumption is a mandatory step towards a practical implementation of physical layer security. With this motivation, this article reviews recent developments on physical layer security with finite-alphabet inputs. We explore transmit signal design algorithms for single-antenna as well as multi-antenna wiretap channels under different assumptions on the channel state information at the transmitter. Moreover, we present a review of the recent results on secure transmission with discrete signaling for various scenarios including multi-carrier transmission systems, broadcast channels with confidential messages, cognitive multiple access and relay networks. Throughout the article, we stress the important behavioral differences of discrete versus Gaussian inputs in the context of the physical layer security. We also present an overview of practical code construction over Gaussian and fading wiretap channels, and we discuss some open problems and directions for future research.
We have performed transverse-field muon spin relaxation (TF-$\mu$SR) measurements on ambient-pressure-grown polycrystalline $\mathrm{LaO_{0.5}F_{0.5}BiS_{2}}$. From these measurements, no signature of magnetic order is found down to 25 mK. The value of the magnetic penetration depth extrapolated to 0 K is 0.89 (5) $\mu$m. The temperature dependence of superconducting penetration depth is best described by either a multigap s + s-wave model with $\Delta_{1}$ = 0.947 (7) meV and $\Delta_{2}$ = 0.22 (4) meV or the ansiotropic s-wave model with $\Delta(0)$ = 0.776 meV and anisotropic gap amplitude ratio $\Delta_{min}/\Delta_{max}$ = 0.34. Comparisons with other potentially multigap $\mathrm{BiS_{2}}$-based superconductors are discussed. We find that these $\mathrm{BiS_{2}}$-based superconductors, including $\mathrm{Bi_{4}O_{4}S_3}$ and the high-pressure synthesized $\mathrm{LaO_{0.5}F_{0.5}BiS_{2}}$, generally conform to the Uemura relation.
Autonomous driving is a complex task which requires advanced decision making and control algorithms. Understanding the rationale behind the autonomous vehicles' decision is crucial to ensure their safe and effective operation on highway driving. This study presents a novel approach, HighwayLLM, which harnesses the reasoning capabilities of large language models (LLMs) to predict the future waypoints for ego-vehicle's navigation. Our approach also utilizes a pre-trained Reinforcement Learning (RL) model to serve as a high-level planner, making decisions on appropriate meta-level actions. The HighwayLLM combines the output from the RL model and the current state information to make safe, collision-free, and explainable predictions for the next states, thereby constructing a trajectory for the ego-vehicle. Subsequently, a PID-based controller guides the vehicle to the waypoints predicted by the LLM agent. This integration of LLM with RL and PID enhances the decision-making process and provides interpretability for highway autonomous driving.
Existence of shear horizontal (SH) surface waves in 2D-periodic phononic crystals with an asymmetric depth-dependent profile is theoretically reported. Examples of dispersion spectra with band gaps for subsonic and supersonic SH surface waves are demonstrated. The link between the effective (quasistatic) speeds of the SH bulk and surface waves is established. Calculation and analysis is based on the integral form of projector on the subspace of evanescent modes which means no need for their explicit finding. This new method can be extended to the vector waves and the 3D case.
In this paper we develope a categorical theory of relations and use this formulation to define the notion of quantization for relations. Categories of relations are defined in the context of symmetric monoidal categories. They are shown to be symmetric monoidal categories in their own right and are found to be isomorphic to certain categories of $A-A$ bicomodules. Properties of relations are defined in terms of the symmetric monoidal structure. Equivalence relations are shown to be commutative monoids in the category of relations. Quantization in our view is a property of functors between monoidal categories. This notion of quantization induce a deformation of all algebraic structures in the category, in particular the ones defining properties of relations like transitivity and symmetry.
We study the generation and evolution of entanglement between two qubits coupled through one-dimensional waveguide modes. By using a complete quantum electrodynamical formalism we go beyond the Markovian approximation. The diagonalization of the hamiltonian is carried out, and a set of quasi-localized eigenstates is found. We show that when the qubit-waveguide coupling is increased, the Markov approximation is not anymore valid, and the generation of entanglement is worsened.
We solve numerically the Boltzmann equation in the early universe in the presence of a constant electric field and find the electrical conductivity $\sigma$ in the range $1\MeV\lsim T\lsim M_W$. The main contribution to $\sigma$ is shown to be due to leptonic interactions. For $T\lsim 100\MeV$ we find $\sigma\simeq 0.76T$ while at $T\simeq M_W$ we obtain $\sigma\simeq 6.7T$
Nonlinear couplings between photons and electrons in new materials give rise to a wealth of interesting nonlinear phenomena. This includes frequency mixing, optical rectification or nonlinear current generation, which are of particular interest for generating radiation in spectral regions that are difficult to access, such as the terahertz gap. Owing to its specific linear dispersion and high electron mobility at room temperature, graphene is particularly attractive for realizing strong nonlinear effects. However, since graphene is a centrosymmetric material, second-order nonlinearities a priori cancel, which imposes to rely on less attractive third-order nonlinearities. It was nevertheless recently demonstrated that dc-second-order nonlinear currents as well as ultrafast ac-currents can be generated in graphene under optical excitation. The asymmetry is introduced by the excitation at oblique incidence, resulting in the transfer of photon momentum to the electron system, known as the photon drag effect. Here, we show broadband coherent terahertz emission, ranging from about 0.1-4 THz, in epitaxial graphene under femtosecond optical excitation, induced by a dynamical photon drag current. We demonstrate that, in contrast to most optical processes in graphene, the next-nearest-neighbor couplings as well as the distinct electron-hole dynamics are of paramount importance in this effect. Our results indicate that dynamical photon drag effect can provide emission up to 60 THz opening new routes for the generation of ultra-broadband terahertz pulses at room temperature.
This paper presents an extension of the recently introduced planewave density interpolation (PWDI) method to the electric field integral equation (EFIE) formulation of problems of scattering and radiation by perfect electric conducting (PEC) objects. Relying on Kirchhoff integral formula and local interpolation of surface current densities that regularize the kernel singularities, the PWDI method enables off- and on-surface EFIE operators to be re-expressed in terms of integrands that are globally bounded (or even more regular) over the whole domain of integration, regardless of the magnitude of the distance between target and source points. Surface integrals resulting from the application of the method-of-moments (MoM) using Rao-Wilton-Glisson (RWG) basis functions, can then be directly and easily evaluated by means of elementary quadrature rules irrespective of the singularity location. The proposed technique can be applied to simple and composite surfaces comprising two or more simply-connected overlapping components. The use of composite surfaces can significantly simplify the geometric treatment of complex structures, as the PWDI method enables the use of separate non-conformal meshes for the discretization of each of the surface components that make up the composite surface. A variety of examples, including multi-scale and intricate structures, demonstrate the effectiveness of the proposed methodology.
In this work we have studied the scattering of scalar field around an extended black hole in F(R) gravity using WKB method. We have obtained the wave function in different regions such as near the horizon region, away from horizon and far away from horizon and the absorption cross section are calculated. We find that the absorption cross section is inversely proportional to the cube of Hawking temperature. We have also evaluated the Hawking temperature of the black hole via tunneling method.
Let $X$ be a smooth compact complex surface with the canonical divisor $K_X$ ample and let $\Theta_X$ be its holomorphic tangent bundle. Bridgeland stability conditions are used to study the space $H^1 (\Theta_X)$ of infinitesimal deformations of complex structures of $X$ and its relation to the geometry/topology of $X$. The main observation is that for $X$ with $H^1 (\Theta_X)$ nonzero and the Chern numbers $(c_2 (X), K^2_X)$ subject to $$ \tau_X :=2ch_2 (\Theta_X)=K^2_X -2c_2(X) >0 $$ the object $\Theta_X [1]$ of the derived category of bounded complexes of coherent sheaves on $X$ is Bridgeland unstable in a certain part of the space of Bridgeland stability conditions. The Harder-Narasimhan filtrations of $\Theta_X [1]$ for those stability conditions are expected to provide new insights into geometry of surfaces of general type and the study of their moduli. The paper provides a certain body of evidence that this is indeed the case.
We describe the origins of recurrence relations between field theory amplitudes in terms of the construction of Feynman diagrams. In application we derive recurrence relations for the amplitudes of QED which hold to all loop orders and for all combinations of external particles. These results may also be derived from the Schwinger-Dyson equations.
We theoretically investigate photonic time-crystalline behaviour initiated by optical excitation above the electronic gap of the excitonic insulator candidate $\rm{Ta_2 Ni Se_5}$. We show that after electron photoexcitation, electron-phonon coupling leads to an unconventional squeezed phonon state, characterised by periodic oscillations of phonon fluctuations. Squeezing oscillations lead to photonic time crystalline behaviour. The key signature of the photonic time crystalline behaviour is THz amplification of reflectivity in a narrow frequency band. The theory is supported by experimental results on $\rm{Ta_2 Ni Se_5}$ where photoexcitation with short pulses leads to enhanced terahertz reflectivity with the predicted features. We explain the key mechanism leading to THz amplification in terms of a simplified Hamiltonian whose validity is supported by ab-initio DFT calculations. Our theory suggests that the pumped $\rm{Ta_2 Ni Se_5}$ is a gain medium, demonstrating that squeezed phonon noise may be used to create THz amplifiers in THz communication applications.
Reliable offroad autonomy requires low-latency, high-accuracy state estimates of pose as well as velocity, which remain viable throughout environments with sub-optimal operating conditions for the utilized perception modalities. As state estimation remains a single point of failure system in the majority of aspiring autonomous systems, failing to address the environmental degradation the perception sensors could potentially experience given the operating conditions, can be a mission-critical shortcoming. In this work, a method for integration of radar velocity information in a LiDAR-inertial odometry solution is proposed, enabling consistent estimation performance even with degraded LiDAR-inertial odometry. The proposed method utilizes the direct velocity-measuring capabilities of an Frequency Modulated Continuous Wave (FMCW) radar sensor to enhance the LiDAR-inertial smoother solution onboard the vehicle through integration of the forward velocity measurement into the graph-based smoother. This leads to increased robustness in the overall estimation solution, even in the absence of LiDAR data. This method was validated by hardware experiments conducted onboard an all-terrain vehicle traveling at high speed, ~12 m/s, in demanding offroad environments.
This paper proposes FREEtree, a tree-based method for high dimensional longitudinal data with correlated features. Popular machine learning approaches, like Random Forests, commonly used for variable selection do not perform well when there are correlated features and do not account for data observed over time. FREEtree deals with longitudinal data by using a piecewise random effects model. It also exploits the network structure of the features by first clustering them using weighted correlation network analysis, namely WGCNA. It then conducts a screening step within each cluster of features and a selection step among the surviving features, that provides a relatively unbiased way to select features. By using dominant principle components as regression variables at each leaf and the original features as splitting variables at splitting nodes, FREEtree maintains its interpretability and improves its computational efficiency. The simulation results show that FREEtree outperforms other tree-based methods in terms of prediction accuracy, feature selection accuracy, as well as the ability to recover the underlying structure.
One of the primary goals of nuclear physics is to understand the force between nucleons, which is a necessary step for understanding the structure of nuclei and how nuclei interact with each other. Rutherford discovered the atomic nucleus in 1911, and the large body of knowledge about the nuclear force since acquired was derived from studies made on nucleons or nuclei. Although antinuclei up to antihelium-4 have been discovered and their masses measured, we have no direct knowledge of the nuclear force between antinucleons. Here, we study antiproton pair correlations among data taken by the STAR experiment at the Relativistic Heavy Ion Collider and show that the force between two antiprotons is attractive. In addition, we report two key parameters that characterize the corresponding strong interaction: namely, the scattering length (f0) and effective range (d0). As direct information on the interaction between two antiprotons, one of the simplest systems of antinucleons, our result provides a fundamental ingredient for understanding the structure of more complex antinuclei and their properties.
We extend two rigorous results of Aizenman, Lebowitz, and Ruelle in their pioneering paper of 1987 on the Sherrington-Kirkpatrick spin-glass model without external magnetic field to the quantum case with a "transverse field" of strength $b$. More precisely, if the Gaussian disorder is weak in the sense that its standard deviation $v>0$ is smaller than the temperature $1/\beta$, then the (random) free energy almost surely equals the annealed free energy in the macroscopic limit and there is no spin-glass phase for any $b/v\geq0$. The macroscopic annealed free energy (times $\beta$) turns out to be non-trivial and given, for any $\beta v>0$, by the global minimum of a certain functional of square-integrable functions on the unit square according to a Varadhan large-deviation principle. For $\beta v<1$ we determine this minimum up to the order $(\beta v)^4$ with the Taylor coefficients explicitly given as functions of $\beta b$ and with a remainder not exceeding $(\beta v)^6/16$. As a by-product we prove that the so-called static approximation to the minimization problem yields the wrong $\beta b$-dependence even to lowest order. Our main tool for dealing with the non-commutativity of the spin-operator components is a probabilistic representation of the Boltzmann-Gibbs operator by a Feynman-Kac (path-integral) formula based on an independent collection of Poisson processes in the positive half-line with common rate $\beta b$. Its essence dates back to Kac in 1956, but the formula was published only in 1989 by Gaveau and Schulman.
In this paper, we study several aspects of solitary wave solutions of the rotation Benjamin-Ono equation. By solving a minimization problem on the line, we construct a family of even travelling waves $\psi_{c,\gamma}$. We also study the strong convergence of this family and we establish the uniqueness of $\psi_{c,\gamma}$ for $\gamma$ small enough. Note that this improves the results in [5] where the stability of the set of ground states is proven.