text,summary " We show that the increasingly popular nonlinear optical technique of time-domain coherent anti-Stokes Raman scattering (CARS), which is usually understood in terms of the semiclassical time-dependent third-order polarization, can be equally explained in terms of the time-delayed version of the Yuratich equation so popular in traditional frequency-domain CARS. The method brings out the strong dependence of CARS time traces and time-delayed CARS lineshapes on the spectral envelope of the probe laser electric field. Examples are analytically shown for experimental results that are otherwise treated by means of numerical methods only. ",Spectral model of time-domain coherent anti-Stokes Raman scattering " We obtain closed form expressions for the expected conditional degree distribution and the joint degree distribution of the linear preferential attachment model for network growth in the steady state. We consider the multiple-destination preferential attachment growth model, where incoming nodes at each timestep attach to $\beta$ existing nodes, selected by degree-proportional probabilities. By the conditional degree distribution $p(\ell| k)$, we mean the degree distribution of nodes that are connected to a node of degree $k$. By the joint degree distribution $p(k,\ell)$, we mean the proportion of links that connect nodes of degrees $k$ and $\ell$. In addition to this growth model, we consider the shifted-linear preferential growth model and solve for the same quantities, as well as a closed form expression for its steady-state degree distribution. ",Degree Correlation in Scale-Free Graphs " Generalizing visual recognition models trained on a single distribution to unseen input distributions (i.e. domains) requires making them robust to superfluous correlations in the training set. In this work, we achieve this goal by altering the training images to simulate new domains and imposing consistent visual attention across the different views of the same sample. We discover that the first objective can be simply and effectively met through visual corruptions. Specifically, we alter the content of the training images using the nineteen corruptions of the ImageNet-C benchmark and three additional transformations based on Fourier transform. Since these corruptions preserve object locations, we propose an attention consistency loss to ensure that class activation maps across original and corrupted versions of the same training sample are aligned. We name our model Attention Consistency on Visual Corruptions (ACVC). We show that ACVC consistently achieves the state of the art on three single-source domain generalization benchmarks, PACS, COCO, and the large-scale DomainNet. ",Attention Consistency on Visual Corruptions for Single-Source Domain Generalization " The thermodynamic properties of an ideal gas of charged vector bosons (with mass m and charge e) is studied in a strong external homogeneous magnetic field no greater than the critical value B_{cr}=m^2/e. The thermodynamic potential, after appropriate analytic continuation, is then used in the study of the spontaneous production of charged spin-one boson pairs from vacuum in the presence of a supercritical homogeneous magnetic field at finite temperature. ",Pair Production of charged vector bosons in supercritical magnetic fields at finite temperatures " A slope $\frac pq$ is called a characterizing slope for a given knot $K_0\subset S^3$ if whenever the $\frac pq$--surgery on a knot $K\subset S^3$ is homeomorphic to the $\frac pq$--surgery on $K_0$ via an orientation preserving homeomorphism, then $K=K_0$. In a previous paper, we showed that, outside a certain finite set of slopes, only the negative integers could possibly be non-characterizing slopes for the torus knot $T_{5,2}$. Applying recent work of Baldwin--Hu--Sivek, we improve our result by showing that a nontrivial slope $\frac pq$ is a characterizing slope for $T_{5,2}$ if $\frac pq>-1$ and $\frac pq\notin \{0,1, \pm\frac12,\pm\frac13\}$. In particular, every nontrivial L-space slope of $T_{5,2}$ is characterizing for $T_{5,2}$. As a consequence, if a nontrivial $\frac pq$-surgery on a non-torus knot in $S^3$ yields a manifold of finite fundamental group, then $|p|>9$. ","Characterizing slopes for torus knots, II" " Let $G$ be an affine algebraic group with a reductive identity component $G^{0}$ acting regularly on an affine Krull scheme $X = {Spec} (R)$ over an algebraically closed field. Let $T$ be an algebraic subtorus of $G$ and suppose that ${Q}(R)^{T}= {Q}(R^{T})$ of quotient fields. We will show: If $G$ is the centralizer of $T$ in $G$, then the pseudo-reflections of the action of $G$ on $R^{T}$ can be lifted to those on $R$. ",Liftings of pseudo-reflection groups of toric quotients of Krull schemes " We study the phase diagram of a system of soft-core dipolar bosons confined to a two-dimensional optical lattice layer. We assume that dipoles are aligned perpendicular to the layer such that the dipolar interactions are purely repulsive and isotropic. We consider the full dipolar interaction and perform Path Integral Quantum Monte Carlo simulations using the Worm Algorithm. Besides a superfluid phase, we find various solid and supersolid phases. We show that, unlike what was found previously for the case of nearest-neighboring interaction, supersolid phases are stabilized not only by doping the solids with particles but with holes as well. We further study the stability of these quantum phases against thermal fluctuations. Finally, we discuss pair formation and the stability of the pair checkerboard phase formed in a bilayer geometry, and suggest experimental conditions under which the pair checkerboard phase can be observed. ",Quantum Phases of Soft-Core Dipolar Bosons in Optical Lattices " In the quest for applicable quantum information technology miniaturised, compact and scalable sources are of paramount importance. Here, we present the concept for the generation of 2-photon N00N states without further post-processing in a single non-linear optical element. Based upon a periodically poled waveguide coupler, we present the principle of state generation via type-0 parametric down-conversion inside this type of devices. With the eigenmode description of the linear optical element, we utilise the delocalised photon pair generation to generate a N00N state in the measurement basis. We show, that we are able to eliminate the need for narrow-band spectral filtering, as well as for phase-stabilisation of the pump light, making this approach an elegant way to produce 2-photon N00N states. ",N00N states from a single non-linear directional coupler " A comparison of structural features of quantum and classical physical theories, such as the information capacity of systems subject to these theories, requires a common formal framework for the presentation of corresponding concepts (such as states, observables, probability, entropy). Such a framework is provided by the notion of statistical model developed in the convexity approach to statistical physical theories. Here we use statistical models to classify and survey all possible types of embedding and extension of quantum probabilistic theories subject to certain reasonable constraints. It will be shown that the so-called canonical classical extension of quantum mechanics is essentially the only `good' representation of the quantum statistical model in a classical framework. All quantum observables are thus identified as fuzzy classical random variables. ",Less (precision) is more (information): quantum information in fuzzy probability theory " We study the influence of thermal fluctuations in the phase diagram of a recently introduced two-dimensional phase field crystal model with an external pinning potential. The model provides a continuum description of pinned lattice systems allowing for both elastic deformations and topological defects. We introduce a non-conserved version of the model and determine the ground-state phase diagram as a function of lattice mismatch and strength of the pinning potential. Monte Carlo simulations are used to determine the phase diagram as a function of temperature near commensurate phases. The results show a rich phase diagram with commensurate, incommensurate and liquid-like phases with a topology strongly dependent on the type of ordered structure. A finite-size scaling analysis of the melting transition for the $c(2 \times 2)$ commensurate phase shows that the thermal correlation length exponent $\nu$ and specific heat behavior are consistent with the Ising universality class as expected from analytical arguments. ",Thermal fluctuations and phase diagrams of the phase field crystal model with pinning " The high-Reynolds number stratified wake of a slender body is studied using a high-resolution hybrid simulation. The wake generator is a 6:1 prolate spheroid with a tripped boundary layer, the diameter-based body Reynolds number is $Re= U_\infty D/\nu = 10^5$, and the body Froude numbers are $Fr=U_\infty/ND=\{2,10,\infty\}$. The wake defect velocity ($U_d$) decays following three stages with different wake decay rates \citep{Spedding1997} as for a bluff body. However, the transition points among stages do not follow the expected $Nt = Nx/U_\infty$ values. Comparison with the wake of a circular disk in similar conditions \citep{Chongsiripinyo2020} quantifies the influence of the wake generator - bluff versus slender - in stratified flow. The strongly stratified $Fr=2$ wake is in a resonant state. The steady lee waves strongly modulate the mean flow and, relative to the disk, the $Fr=2$ spheroid wake shows an earlier transition from the non-equilibrium (NEQ) stage to the quasi-two-dimensional (Q2D) stage. The NEQ-Q2D transition is followed by a sharp increase in the turbulent kinetic energy and horizontal wake meanders. At $Fr=10$, the start of the NEQ stage is delayed. Transfers between kinetic energy and potential energy reservoirs (both mean and turbulence) are analyzed and the flows are compared in phase space (local Froude and Reynolds number as coordinates). Overall, the results of this study point to the difficulty of finding a universal framework for stratified wake evolution, independent of the features of the body, and provide insights into how buoyancy effects depend on the wake generator. ",The high-Reynolds-number stratified wake of a slender body and its comparison with a bluff-body wake " We say that a finite group $G$ satisfies the independence property if, for every pair of distinct elements $x$ and $y$ of $G$, either $\{x,y\}$ is contained in a minimal generating set for $G$ or one of $x$ and $y$ is a power of the other. We give a complete classification of the finite groups with this property, and in particular prove that every such group is supersoluble. A key ingredient of our proof is a theorem showing that all but three finite almost simple groups $H$ contain an element $s$ such that the maximal subgroups of $H$ containing $s$, but not containing the socle of $H$, are pairwise non-conjugate. ",Finite groups satisfying the independence property " Industrial robots play an increasingly important role in a growing number of fields. For example, robotics is used to increase productivity while reducing costs in various aspects of manufacturing. Since robots are often set up in production lines, the breakdown of a single robot has a negative impact on the entire process, in the worst case bringing the whole line to a halt until the issue is resolved, leading to substantial financial losses due to the unforeseen downtime. Therefore, predictive maintenance systems based on the internal signals of robots have gained attention as an essential component of robotics service offerings. The main shortcoming of existing predictive maintenance algorithms is that the extracted features typically differ significantly from the learnt model when the operation of the robot changes, incurring false alarms. In order to mitigate this problem, predictive maintenance algorithms require the model to be retrained with normal data of the new operation. In this paper, we propose a novel solution based on transfer learning to pass the knowledge of the trained model from one operation to another in order to prevent the need for retraining and to eliminate such false alarms. The deployment of the proposed unsupervised transfer learning algorithm on real-world datasets demonstrates that the algorithm can not only distinguish between operation and mechanical condition change, it further yields a sharper deviation from the trained model in case of a mechanical condition change and thus detects mechanical issues with higher confidence. ",Domain Adaptation for Robot Predictive Maintenance Systems " In nonlinear state-space models, sequential learning about the hidden state can proceed by particle filtering when the density of the observation conditional on the state is available analytically (e.g. Gordon et al., 1993). This condition need not hold in complex environments, such as the incomplete-information equilibrium models considered in financial economics. In this paper, we make two contributions to the learning literature. First, we introduce a new filtering method, the state-observation sampling (SOS) filter, for general state-space models with intractable observation densities. Second, we develop an indirect inference-based estimator for a large class of incomplete-information economies. We demonstrate the good performance of these techniques on an asset pricing model with investor learning applied to over 80 years of daily equity returns. ",State-Observation Sampling and the Econometrics of Learning Models " Photosystem 0 concerns a primitive mechanism for free energy gain as ATP from fluctuating light during early evolution. The PS0 reaction centers had no reducing power: charge transport was only temporary. Light induced within the reaction centers metastable dipoles that generated a membrane potential. This in turn drove ATP synthesis by protons moving through the ATP synthase enzyme. After the decay of the dipole potential in the dark, the protons either (1) returned across the membrane by conduction or (2) were pumped back by ATP synthase, backwards active as ATPase at a higher H+/ATP ratio. PS0 constitutes a link to previously proposed free energy sources for early evolution that worked on thermal cycling. Several contemporary photosynthetic phenomena may be relics of PS0. ","Photosystem 0, a proposed ancestral photosystem without reducing power that synthesized ATP during light-dark cycling" " In this paper we report the effect of the jet-medium interplay as implemented in EPOS 3 on the ridge like structure observed in high multiplicity p-Pb collisions at $\sqrt{s_{NN}} = $ 5.02 TeV. EPOS 3 takes into account hydrodynamically expanding bulk matter, jets and the jet-medium interaction. The basis of this model is multiple scatterings where each scattering finally produces flux tube / string. In the higher multiplicity event classes where the flux tube/string density is higher, there is a finite probability that the strings will pick up quarks and antiquarks (or diquarks) from the bulk (core) for flux tube breaking to produce jet hadrons (corona) instead of producing them via usual Schwinger mechanism. This will eventually create a correlation between core and corona and also influence the corona-corona correlation as the corona particles containing quarks and antiquarks (or diquarks) from the bulk also carry the fluid information. We report the relative contributions of the core-core, core-corona, corona-core and corona-corona correlations towards the ridge in the high and low multiplicity p-Pb collisions at $\sqrt{s_{NN}} = $ 5.02 TeV using the data generated by EPOS 3. The multiplicity evolution of the ridges in all the cases is also reported. ",Ridge from jet-medium interaction in p-Pb collisions at $\sqrt{s_{NN} }$ = 5.02 TeV " Taking into account the mixing effects between left- and right-handed top-squarks, we calculate the genuine supersymmetric eletroweak correction to top quark production at the Tevatron in the minimal supersymmetric model. The analytic expressions of the corrections to both the parton level cross section and the total hadronic cross section are presented. Some numerical examples are also given to show the size of the corrections. ",Top-squark mixing effects in the supersymmetric electroweak corrections to top quark production at the Tevatron " Atomically thin layers of two-dimensional (2D) materials such as graphene, MoS2 and h-BN have immense potential as sensors and electronic devices thanks to their highly desirable electronic, mechanical, optical and heat transport properties. In particular their extreme stiffness, tensile strength and low density allows for high frequency electronic devices, resonators and ultra-sensitive detectors providing realistic avenues for down-scaling electronic devices and nanoelectromechanical systems (NEMS). Whilst nanoscale morphology and electronic properties of 2D materials can be studied using existing electron or scanning probe microscopy approaches, time-dependant phenomena on the ns and shorter time-scales cannot be readily explored. Here we use the heterodyne principle to reach into this ns time-scale and create a local nanoscale probe for electrostatically induced actuation of a graphene resonator, with amplitude sensitivity down to pm range and time sensitivity in the ns range. We experimentally observed response times of 20-120 ns for resonators with beam lengths of 180 nm to 2.5 um in line with the theoretical predictions for such NEMS devices. ",Nanoscale Mapping of Nanosecond Time-scale Electro-Mechanical Phenomena in Graphene NEMS " High-speed spectroscopy of two pulsating subdwarf B stars, KPD 2109+4401 and PB 8783, is presented. Radial motions are detected with the same frequencies as reported from photometric observations and with amplitudes of ~2 km/sec in two or more independent modes. These represent the first direct observations of surface motion due to multimode non-radial oscillations in subdwarf B stars. In the case of the sdB+F binary PB 8783, the velocities of both components are resolved; high-frequency oscillations are found only in the sdB star and not the F star. There also appears to be evidence for mutual motion of the binary components. If confirmed, it implies that the F-type companion is >~1.2 times more massive than the sdB star, while the amplitude of the F star acceleration over 4 hours would constrain the orbital period to lie between 0.5 and 3.2d. ",Radial velocities of pulsating subdwarf B stars: KPD 2109+4401 and PB 8783 We report on the deterministic fabrication of sub-um mesa structures containing single quantum dots by in-situ electron-beam lithography. The fabrication method is based on a two-step lithography process using a low-temperature cathodoluminescence (CL) spectroscopy setup. In the first step the position and spectral features of single InGaAs quantum dots (QDs) are detected by CL. Then circular sub-um mesa-structures are exactly defined by high-resolution electron-beam lithography and subsequent etching in the second step. CL spectroscopy and micro-photoluminscence spectroscopy demonstrate the high optical quality of the single-QD mesa-structures with emission linewidths below 15 ueV and g(2)(0) = 0.04. Our lithography method allows for an alignment precision better than 100 nm which paves the way for a fully-deterministic device technology using in-situ CL lithography. ,In-situ electron-beam lithography of deterministic single-quantum-dot mesa-structures using low-temperature cathodoluminescence spectroscopy " For a multivariate random walk with i.i.d. jumps satisfying the Cramer moment condition and having a mean vector with at least one negative component, we derive the exact asymptotics of the probability of ever hitting the positive orthant that is being translated to infinity along a fixed vector with positive components. This problem is motivated by and extends results from a paper by F. Avram et al. (2008) on a two-dimensional risk process. Our approach combines the large deviation techniques from a recent series of papers by A. Borovkov and A. Mogulskii with new auxiliary constructions, which enable us to extend their results on hitting remote sets with smooth boundaries to the case of boundaries with a ""corner"" at the ""most probable hitting point"". We also discuss how our results can be extended to the case of more general target sets. ",The exact asymptotics of the large deviation probabilities in the multivariate boundary crossing problem " We study the absorption of scalar fields by extreme/exotic compact objects (ECOs) -- horizonless alternatives to black holes -- via a simple model in which dissipative mechanisms are encapsulated in a single parameter. Trapped modes, localized between the ECO core and the potential barrier at the photosphere, generate Breit-Wigner-type spectral lines in the absorption cross section. Absorption is enhanced whenever the wave frequency resonates with a trapped mode, leading to a spectral profile which differs qualitatively from that of a black hole. We introduce a model based on Nariai spacetime, in which properties of the spectral lines are calculated in closed form. We present numerically calculated absorption cross sections and transmission factors for example scenarios, and show how the Nariai model captures the essential features. We argue that, in principle, ECOs can be distinguished from black holes through their absorption spectra. ",Spectral lines of extreme compact objects " We update the bounds on fermions with electric charge $\epsilon e$ and mass $m_\epsilon$. For $m_\epsilon\lsim m_e$ we find $10^{-15}\lsim\epsilon<1$ is excluded by laboratory experiments, astrophysics and cosmology. For larger masses, the limits are less restrictive and depend on $m_\epsilon$. For milli-charged neutrinos, the limits are stronger, especially if the different flavors mix as suggested by current experimental evidence. ",Updated Bounds on Milli-Charged Particles " Interferences are not positive-definite and therefore they can change sign over the phase space. If the contributions of the regions where the interference is positive and negative nearly cancel each other, interference effects are hard to measure. In this paper, we propose a method to quantify the ability of an observable to separate an interference positive and negative contributions and therefore to revive the interference effects in measurements. We apply this method to the anomalous gluon operator in the SMEFT for which the interference suppression is well-known. We show that we can get contraints on its coefficient, using the interference only, similar to those obtained by including the square of the new physics amplitude. ",Reviving the interference: framework and proof-of-principle for the anomalous gluon self-interaction in the SMEFT " Suppose that $m$ drivers each choose a preferred parking space in a linear car park with $n$ spots. In order, each driver goes to their chosen spot and parks there if possible, and otherwise takes the next available spot if it exists. If all drivers park successfully, the sequence of choices is called a parking function. Classical parking functions correspond to the case $m=n$; we study here combinatorial and probabilistic aspects of this generalized case. We construct a family of bijections between parking functions $\text{PF}(m, n)$ with $m$ cars and $n$ spots and spanning forests $\mathscr{F}(n+1, n+1-m)$ with $n+1$ vertices and $n+1-m$ distinct trees having specified roots. This leads to a bijective correspondence between $\text{PF}(m, n)$ and monomial terms in the associated Tutte polynomial of a disjoint union of $n-m+1$ complete graphs. We present an identity between the ""inversion enumerator"" of spanning forests with fixed roots and the ""displacement enumerator"" of parking functions. The displacement is then related to the number of graphs on $n+1$ labeled vertices with a fixed number of edges, where the graph has $n+1-m$ disjoint rooted components with specified roots. We investigate various probabilistic properties of a uniform parking function, giving a formula for the law of a single coordinate. As a side result we obtain a recurrence relation for the displacement enumerator. Adapting known results on random linear probes, we further deduce the covariance between two coordinates when $m=n$. ",Parking functions: From combinatorics to probability " In real-world dialogue systems, the ability to understand the user's emotions and interact anthropomorphically is of great significance. Emotion Recognition in Conversation (ERC) is one of the key ways to accomplish this goal and has attracted growing attention. How to model the context in a conversation is a central aspect and a major challenge of ERC tasks. Most existing approaches are generally unable to capture both global and local contextual information efficiently, and their network structures are too complex to design. For this reason, in this work, we propose a straightforward Dual-stream Recurrence-Attention Network (DualRAN) based on Recurrent Neural Network (RNN) and Multi-head ATtention network (MAT). The proposed model eschews the complex network structure of current methods and focuses on combining recurrence-based methods with attention-based methods. DualRAN is a dual-stream structure mainly consisting of local- and global-aware modules, modeling a conversation from distinct perspectives. To achieve the local-aware module, we extend the structure of RNN, thus enhancing the expressive capability of the network. In addition, we develop two single-stream network variants for DualRAN, i.e., SingleRANv1 and SingleRANv2. We conduct extensive experiments on four widely used benchmark datasets, and the results reveal that the proposed model outshines all baselines. Ablation studies further demonstrate the effectiveness of each component. ",A Dual-Stream Recurrence-Attention Network with Global-Local Awareness for Emotion Recognition in Textual Dialogue " In high-energy nuclear collisions, light nuclei can be regarded as a cluster of baryons and their yields are sensitive to the baryon density fluctuations. Thus, the production of light nuclei can be used to study the QCD phase transition, at which the baryon density fluctuation will be enhanced. The yield ratio of light nuclei, defined as $N(t)$$\times$$N(p)$/$N^2(d)$, is predicted to be sensitive observable to search for the 1st-order phase transition and/or QCD critical point in heavy-ion collisions. In this paper, we present the energy and centrality dependence of (anti)deuteron and triton production in Au+Au collisions at $\sqrt{s_{\mathrm{NN}}}$ = 7.7, 11.5, 14.5, 19.6, 27, 39, 54.4, 62.4, and 200 GeV measured by the STAR experiment at RHIC. We show beam-energy dependence for the coalescence parameter, $B_2(d)$ and $B_3(t)$, particle ratios, $d/p$, $t/p$, and $t/d$, and the yield ratio of $N(t)$$\times$$N(p)$/$N^2(d)$. More importantly, non-monotonic energy dependence is observed for the yield ratio, $N(t)$$\times$$N(p)$/$N^2(d)$, in 0-10\% central Au+Au collisions with a peak around 20-30 GeV. Their physics implications on QCD critical point search and change of the equation of state will be discussed. ","Light Nuclei ($d$, $t$) Production in Au + Au Collisions at $\sqrt{s_{NN}}$ = 7.7 - 200 GeV" " We have used the Goddard IRAM 2-Millimeter Observer (GISMO) with the 30 m IRAM telescope to carry out a 2 mm survey of the Galaxy's central molecular zone (CMZ). These observations detect thermal emission from cold ISM dust, thermal free-free emission from ionized gas, and nonthermal synchrotron emission from relatively flat-spectrum sources. Archival data sets spanning $3.6 \mu$m to 90 cm are used to distinguish different emission mechanisms. After the thermal emission of dust is modeled and subtracted, the remaining 2 mm emission is dominated by free-free emission, with the exception of the brightest nonthermal filament (NTF) that runs though the middle of the bundle of filaments known as the Radio Arc. This is the shortest wavelength at which any NTF has been detected. The GISMO observations clearly trace this NTF over a length of ~0.2$^\circ$, with a mean 2 mm spectral index which is steeper than at longer wavelengths. The 2 mm to 6 cm (or 20 cm) spectral index steepens from $\alpha \approx -0.2$ to $-0.7$ as a function distance from the Sickle H II region, suggesting that this region is directly related to the NTF. A number of unresolved (at $21''$) 2 mm sources are found nearby. One appears to be thermal dust emission from a molecular cloud that is associated with an enigmatic radio point source whose connection to the Radio Arc is still debated. The morphology and colors at shorter IR wavelengths indicate other 2 mm unresolved sources are likely to be compact H II regions. ",2 mm GISMO Observations of the Galactic Center. II. A Nonthermal Filament in the Radio Arc and Compact Sources " The qualities of electron refrigeration by means of tunnel junctions between superconducting and normal--metal electrodes are studied theoretically. A suitable approximation of the basic expression for the heat current across those tunnel junctions allows the investigation of several features of the device such as its optimal bias voltage, its maximal heat current, its optimal working point, and the maximally gained temperature reduction. Fortunately, the obtained results can be compared with those of a recent experiment. ",Electron Refrigeration in the Tunneling Approach " Photocatalytic water splitting reaction on TiO2 surface is one of the fundamental issues that bears significant implication in hydrogen energy technology and has been extensively studied. However, the existence of the very first reaction step, the direct photo-dissociation of water, has been disregarded. Here, we provide unambiguously experimental evidence to demonstrate that adsorbed water molecules on reduced rutile TiO2(110)-1\times1 surface can be dissociated under UV irradiation using low temperature scanning tunneling microscopy. It is identified that a water molecule at fivefold coordinated Ti (Ti5c) site can be photocatalytically dissociated, resulting in a hydroxyl at Ti5c and another hydroxyl at bridge oxygen row. Our findings reveal a missing link in the photocatalytic water splitting reaction chain, which greatly contribute to the detailed understanding of underlying mechanism. ",Evidence of Photocatalytic Dissociation of Water on TiO2 with Atomic Resolution " We investigate the observational viability of a class of $\alpha$-attractors inflationary models in light of the most recent Cosmic Microwave Background (CMB) and Large-Scale Structure (LSS) data. By considering a double-well potential we perform a slow-roll analysis to study the behavior of this class of models, which is a continuous interpolation between the chaotic inflation for large values of $\alpha$ and the universal attractor, i.e., $n_s=1- 2/N$ and $r=\alpha 12/N^2$ for small $\alpha$, where $n_s$ is the scalar spectral index, $r$ is the tensor-to-scalar ratio, and $N$ is the e-fold number. In order to explore the parameter space of the model, we also perform a MCMC analysis and find $\alpha=7.56\pm 5.15$ ($1\sigma$). ",Observational constraints on $\alpha$-attractor inflationary models with a Higgs-like potential " We study the thermodynamics of a crystalline solid by applying intermediate statistics manifested by q-deformation. We based part of our study on both Einstein and Debye models, exploring primarily deformed thermal and electrical conductivities as a function of the deformed Debye specific heat. The results revealed that the q-deformation acts in two different ways but not necessarily as independent mechanisms. It acts as a factor of disorder or impurity, modifying the characteristics of a crystalline structure, which are phenomena described by q-bosons, and also as a manifestation of intermediate statistics, the B-anyons (or B-type systems). For the latter case, we have identified the Schottky effect, normally associated with high-Tc superconductors in the presence of rare-earth-ion impurities, and also the increasing of the specific heat of the solids beyond the Dulong-Petit limit at high temperature, usually related to anharmonicity of interatomic interactions. Alternatively, since in the q-bosons the statistics are in principle maintained the effect of the deformation acts more slowly due to a small change in the crystal lattice. On the other hand, B-anyons that belong to modified statistics are more sensitive to the deformation. ",Intermediate statistics in thermoelectric properties of solids " Multimodal semantic understanding often has to deal with uncertainty, which means the obtained messages tend to refer to multiple targets. Such uncertainty is problematic for our interpretation, including inter- and intra-modal uncertainty. Little effort has studied the modeling of this uncertainty, particularly in pre-training on unlabeled datasets and fine-tuning in task-specific downstream datasets. In this paper, we project the representations of all modalities as probabilistic distributions via a Probability Distribution Encoder (PDE) by utilizing sequence-level interactions. Compared to the existing deterministic methods, such uncertainty modeling can convey richer multimodal semantic information and more complex relationships. Furthermore, we integrate uncertainty modeling with popular pre-training frameworks and propose suitable pre-training tasks: Distribution-based Vision-Language Contrastive learning (D-VLC), Distribution-based Masked Language Modeling (D-MLM), and Distribution-based Image-Text Matching (D-ITM). The fine-tuned models are applied to challenging downstream tasks, including image-text retrieval, visual question answering, visual reasoning, and visual entailment, and achieve state-of-the-art results. ",MAP: Multimodal Uncertainty-Aware Vision-Language Pre-training Model In this paper we discuss how the question about the rationality of L^2-Betti numbers is related to the Isomorphism Conjecture in algebraic K-theory and why in this context noncommutative localization appears as an important tool. ,"L^2-Betti numbers, isomorphism conjectures and noncommutative localization" " We study torsion subgroups of elliptic curves with complex multiplication (CM) defined over number fields which admit a real embedding. We give a complete classification of the groups which arise up to isomorphism as the torsion subgroup of a CM elliptic curve defined over a number field of odd degree: there are infinitely many. Restricting to the case of prime degree, we show that there are only finitely many isomorphism classes. More precisely, there are six ""Olson groups"" which arise as torsion subgroups of CM elliptic curves over number fields of every degree, and there are precisely 17 ""non-Olson"" CM elliptic curves defined over a prime degree number field. ",Torsion Points on CM Elliptic Curves Over Real Number Fields " In this reply we answer the comment by A. Dhar (cond-mat/0203077) on our Letter ""Simple one dimensional model of heat conduction which obeys Fourier's law"" (Phys. Rev. Lett. 86, 5486 (2001), cond-mat/0104453) ","Reply to comment on ""Simple one-dimensional model of heat conduction which obeys Fourier's law""" " We investigate the critical curve of the string tension sigma(T) as a function of temperature in quenched gauge invariant SU(3) lattice gauge theory. We extract sigma(T) from the colour averaged free energy of a static quark-antiquark pair. To compute the free energy, we utilize a pair of gauge invariant Polyakov loop and antiloop correlations, and apply the multihit procedure to enhance the signal to noise ratio. We find that the string tension departs from the zero temperature sigma(0) at T close to 0.5 Tc. We cover the relevant temperature range from 0.5 Tc up to the confinement temperature Tc using 57 different sets of pure gauge lattice configurations with four temporal extensions (4,6,8,12), different beta and a spatial volume of 48^3 in lattice units. ",Lattice QCD computation of the SU(3) String Tension critical curve " The Kepler object KIC 12557548 shows irregular eclipsing behaviour with a constant 15.685 hr period, but strongly varying transit depth. In this paper we fit individual eclipses, in addition to fitting binned light curves, to learn more about the process underlying the eclipse depth variation. Additionally, we put forward observational constraints that any model of this planet-star system will have to match. We find two quiescent spells of ~30 orbital periods each where the transit depth is <0.1%, followed by relatively deep transits. Additionally, we find periods of on-off behaviour where >0.5% deep transits are followed by apparently no transit at all. Apart from these isolated events we find neither significant correlation between consecutive transit depths nor a correlation between transit depth and stellar intensity. We find a three-sigma upper limit for the secondary eclipse of 4.9*10^-5, consistent with a planet candidate with a radius of less than 4600 km. Using the short cadence data we find that a 1-D exponential dust tail model is insufficient to explain the data. We improved our model to a 2-D, two-component dust model with an opaque core and an exponential tail. Using this model we fit individual eclipses observed in short cadence mode. We find an improved fit of the data, quantifying earlier suggestions by Budaj (2013) of the necessity of at least two components. We find that deep transits have most absorption in the tail, and not in a disk-shaped, opaque coma, but the transit depth and the total absorption show no correlation with the tail length. ",Analysis and interpretation of 15 quarters of Kepler data of the disintegrating planet KIC 12557548 b " A potential crewed mission to Mars would require us to solve a number of problems, including how to protect astronauts against the devastating effects of energetic charged particles from Solar and Galactic sources. The radiation environment on Mars is of particular interest, since maintaining optimal absorbed doses by astronauts is crucial to their survival. Here, we give an overview of the conditions on Mars, as determined by theoretical models and in-situ measurements, and present the main proposed strategies to mitigate radiation exposure while on Mars. Specifically, we focus on the passive shielding technique. Several widely used materials, along with some innovative ones and combinations of those, are studied for their behavior against Solar Energetic Particle Events and Galactic Cosmic Rays in the Martian environment. For that purpose, we implement the GEANT4 package, a Monte-Carlo numerical model developed by CERN, which is specifically applied to simulate interactions of radiation with matter. A description of our model will be given, followed by outputs of the numerical model. We conclude that hydrogen-rich materials act as better attenuators, as expected, but other materials can be helpful against cosmic rays too. ",Radiation protection and shielding materials for crewed missions on the surface of Mars " The CP-violating phases in the soft supersymmetry-breaking sector in orbifold compactifications with a continuous Wilson line are investigated. In this case the modular symmetry is the Siegel modular group $Sp(4,Z)$ of genus two. In particular, we study the case that the hidden sector non-perturbative superpotential is determined by the Igusa cusp form ${\cal C}_{12}$ of modular weight 12. The effect of large non-perturbative corrections to the dilaton K\""ahler potential on the resulting CP-violating phases is also investigated. ",The effect of Wilson line moduli on CP-violation by soft supersymmetry breaking terms " The compress-and-forward relay scheme developed by (Cover and El Gamal, 1979) is improved with a modification on the decoding process. The improvement follows as a result of realizing that it is not necessary for the destination to decode the compressed observation of the relay; and even if the compressed observation is to be decoded, it can be more easily done by joint decoding with the original message, rather than in a successive way. An extension to multiple relays is also discussed. ",An Improvement of Cover/El Gamal's Compress-and-Forward Relay Scheme " This survey article is concerned with the study of bifurcations of piecewise-smooth maps. We review the literature in circle maps and quasi-contractions and provide paths through this literature to prove sufficient conditions for the occurrence of two types of bifurcation scenarios involving rich dynamics. The first scenario consists of the appearance of periodic orbits whose symbolic sequences and ""rotation"" numbers follow a Farey tree structure; the periods of the periodic orbits are given by consecutive addition. This is called the {\em period adding} bifurcation, and its proof relies on results for maps on the circle. In the second scenario, symbolic sequences are obtained by consecutive attachment of a given symbolic block and the periods of periodic orbits are incremented by a constant term. It is called the {\em period incrementing} bifurcation, in its proof relies on results for maps on the interval. We also discuss the expanding cases, as some of the partial results found in the literature also hold when these maps lose contractiveness. The higher dimensional case is also discussed by means of {\em quasi-contractions}. We also provide applied examples in control theory, power electronics and neuroscience where these results can be applied to obtain precise descriptions of their dynamics. ",The Period adding and incrementing bifurcations: from rotation theory to applications " Hierarchical multi-label text classification aims to classify the input text into multiple labels, among which the labels are structured and hierarchical. It is a vital task in many real world applications, e.g. scientific literature archiving. In this paper, we survey the recent progress of hierarchical multi-label text classification, including the open sourced data sets, the main methods, evaluation metrics, learning strategies and the current challenges. A few future research directions are also listed for community to further improve this field. ",Recent Advances in Hierarchical Multi-label Text Classification: A Survey " Coulomb interaction between charged particles is a well-known phenomenon in many areas of researches. In general the Coulomb repulsion force broadens the pulse width of an electron bunch and limits the temporal resolution of many scientific facilities such as ultrafast electron diffraction and x-ray free-electron lasers. Here we demonstrate a scheme that actually makes use of Coulomb force to compress a relativistic electron beam. Furthermore, we show that the Coulomb-driven bunch compression process does not introduce additional timing jitter, which is in sharp contrast to the conventional radio-frequency buncher technique. Our work not only leads to enhanced temporal resolution in electron beam based ultrafast instruments that may provide new opportunities in probing material systems far from equilibrium, but also opens a promising direction for advanced beam manipulation through self-field interactions. ",Coulomb-driven relativistic electron beam compression " A weighted digraph is a digraph such that every arc is assigned a nonnegative number, called the weight of the arc. The weighted outdegree of a vertex $v$ in a weighted digraph $D$ is the sum of the weights of the arcs with $v$ as their tail, and the weight of a directed cycle $C$ in $D$ is the sum of the weights of the arcs of $C$. In this note we prove that if every vertex of a weighted digraph $D$ with order $n$ has weighted outdegree at least 1, then there exists a directed cycle in $D$ with weight at least $1/\log_2 n$. This proves a conjecture of Bollob\'{a}s and Scott up to a constant factor. ",A note on heavy cycles in weighted digraphs " Electrically driven spin resonance (EDSR) is an established tool for controlling semiconductor spin qubits. Here, we theoretically study a frequency-mixing variant of EDSR, where two driving tones with different drive frequencies are applied, and the resonance condition connects the spin Larmor frequency with the sum of the two drive frequencies. Focusing on flopping-mode operation, we calculate the parameter dependence of the Rabi frequency and the Bloch-Siegert shift. A shared-control spin qubit architecture could benefit from this bichromatic EDSR scheme, as it enables simultaneous single-qubit gates. ",Electrically driven spin resonance with bichromatic driving " Understanding relationships between feature variables is one important way humans use to make decisions. However, state-of-the-art deep learning studies either focus on task-agnostic statistical dependency learning or do not model explicit feature dependencies during prediction. We propose a deep neural network framework, dGAP, to learn neural dependency Graph and optimize structure-Aware target Prediction simultaneously. dGAP trains towards a structure self-supervision loss and a target prediction loss jointly. Our method leads to an interpretable model that can disentangle sparse feature relationships, informing the user how relevant dependencies impact the target task. We empirically evaluate dGAP on multiple simulated and real datasets. dGAP is not only more accurate, but can also recover correct dependency structure. ",Relate and Predict: Structure-Aware Prediction with Jointly Optimized Neural DAG " We summarise status and recent results of the European Twisted Mass collaboration (ETMC). The collaboration has generated gauge configurations for three different values of the lattice spacing smaller or equal 0.1 fm and values of the charged pseudo scalar mass as low as 300 MeV with two flavours of maximally twisted mass quarks. We provide evidence that O(a) improvement works very well with maximally twisted mass fermions and that also higher order lattice artifacts appear to be small. The currently only quantity in the light meson and baryon sector where cut-off effects are visible is the neutral pseudo scalar meson mass and we present an attempt to understand this from a theoretical point of view. We describe finite size effects and quark mass dependence of the mass and decay constant of the (charged) pseudo scalar meson with chiral perturbation theory formulae and our current estimate for the low energy constants l_{3,4} is l_3=3.44(8)(35) and l_4=4.61(4)(11). Results for the average up-down, the strange and the charm quark mass and the chiral condensate are also presented. ",Lattice QCD with two light Wilson quarks and maximally twisted mass " We calculate the radiated energy to $O(\hbar)$ from a charged wave-packet in the uniform magnetic field. In the high-speed and weak-field limit, while the non-commutativity of the system reduces the classical radiation, the additional corrections originated from the velocity uncertainty of the wave-packet leads to an enhancement of the radiation. ",Quantum Corrections to Synchrotron Radiation from Wave-Packet " We apply Monte Carlo Renormalization group to the crumpling transition in random surface models of fixed connectivity. This transition is notoriously difficult to treat numerically. We employ here a Fourier accelerated Langevin algorithm in conjunction with a novel blocking procedure in momentum space which has proven extremely successful in $\lambda\phi^4$. We perform two successive renormalizations in lattices with up to $64^2$ sites. We obtain a result for the critical exponent $\nu$ in general agreement with previous estimates and similar error bars, but with much less computational effort. We also measure with great accuracy $\eta$. As a by-product we are able to determine the fractal dimension $d_H$ of random surfaces at the crumpling transition. ",M.C.R.G. Study of Fixed-connectivity Surfaces " We consider a family of slightly extended version of the Raynaud's surfaces X over the field of positive characteristic with Mumford-Szpiro type polarizations Z, which have Kodaira non-vanishing H^1(X, Z^{-1})\ne 0. The surfaces are at least normal but smooth under a special condition. We compute the cohomologies H^i(X, Z^n), for intergers i and n, and study their (non-)vanishing. Finally, we give a fairly large family of non Mumford-Szpiro type polarizations Z_{a,b} with Kodaira non-vanishing. ",On non-vanishing of cohomologies of generalized Raynaud polarized surfaces " Keyphrase extraction from a given document is the task of automatically extracting salient phrases that best describe the document. This paper proposes a novel unsupervised graph-based ranking method to extract high-quality phrases from a given document. We obtain the contextualized embeddings from pre-trained language models enriched with topic vectors from Latent Dirichlet Allocation (LDA) to represent the candidate phrases and the document. We introduce a scoring mechanism for the phrases using the information obtained from contextualized embeddings and the topic vectors. The salient phrases are extracted using a ranking algorithm on an undirected graph constructed for the given document. In the undirected graph, the nodes represent the phrases, and the edges between the phrases represent the semantic relatedness between them, weighted by a score obtained from the scoring mechanism. To demonstrate the efficacy of our proposed method, we perform several experiments on open source datasets in the science domain and observe that our novel method outperforms existing unsupervised embedding based keyphrase extraction methods. For instance, on the SemEval2017 dataset, our method advances the F1 score from 0.2195 (EmbedRank) to 0.2819 at the top 10 extracted keyphrases. Several variants of the proposed algorithm are investigated to determine their effect on the quality of keyphrases. We further demonstrate the ability of our proposed method to collect additional high-quality keyphrases that are not present in the document from external knowledge bases like Wikipedia for enriching the document with newly discovered keyphrases. We evaluate this step on a collection of annotated documents. The F1-score at the top 10 expanded keyphrases is 0.60, indicating that our algorithm can also be used for 'concept' expansion using external knowledge. ",Topic Aware Contextualized Embeddings for High Quality Phrase Extraction " The KKLT construction of dS vacua relies on an uplift term that arises from an anti-D3-brane. It was argued by Kachru, Pearson and Verlinde that this anti-D3-brane is an excited state in a supersymmetric theory since it can decay to a supersymmetric ground state. Hence the anti-D3-brane breaks supersymmetry spontaneously and one should be able to package all the world-volume fields on the anti-D3-brane into a four dimensional $\cal{N}=1$ supersymmetric action. Here we extend previous results and identify the constrained superfields that correspond to all the degrees of freedom on the anti-D3-brane. In particular, we show explicitly that the four 4D worldvolume spinors give rise to constrained chiral multiplets $S$ and $Y^i$, $i=1,2,3$ that satisfy $S^2=SY^i=0$. We also conjecture (and provide evidence in a forthcoming publication) that the vector field $A_\mu$ and the three scalars $\phi^i$ give rise to a field strength multiplet $W_\alpha$ and three chiral multiplets $H^i$ that satisfy the constraints $S W_\alpha= \bar{D}_{\dot \alpha} (S \bar H^i)=0$. This is the first time that such constrained multiplets appear in string theory constructions. ",Constrained superfields from an anti-D3-brane in KKLT " In this paper we study the set of digit frequencies that are realised by elements of the set of $\beta$-expansions. The main result of this paper demonstrates that as $\beta$ approaches $1,$ the set of digit frequencies that occur amongst the set of $\beta$-expansions fills out the simplex. As an application of our main result, we obtain upper bounds for the local dimension of certain biased Bernoulli convolutions. ",Exceptional digit frequencies and expansions in non-integer bases " We identify the class of elementary groups: the smallest class of totally disconnected locally compact second countable (t.d.l.c.s.c.) groups that contains the profinite groups and the discrete groups, is closed under group extensions of profinite groups and discrete groups, and is closed under countable increasing unions. We show this class enjoys robust permanence properties. In particular, it is closed under group extension, taking closed subgroups, taking Hausdorff quotients, and inverse limits. A characterization of elementary groups in terms of well-founded descriptive-set-theoretic trees is then presented. We conclude with three applications. We first prove structure results for general t.d.l.c.s.c. groups. In particular, we show a compactly generated t.d.l.c.s.c. group decomposes into elementary groups and topologically characteristically simple groups via group extension. We then prove two local-to-global structure theorems: Locally solvable t.d.l.c.s.c. groups are elementary and [A]-regular t.d.l.c.s.c. groups are elementary. ",Elementary totally disconnected locally compact groups " The Kondo effect, a hallmark of strong correlation physics, is characterized by the formation of an extended cloud of singlet states around magnetic impurities at low temperatures. While many implications of the Kondo cloud's existence have been verified, the existence of the singlet cloud itself has not been directly demonstrated. We suggest a route for such a demonstration by considering an observable that has no classical analog, but is still experimentally measurable: ""singlet weights"", or projections onto particular entangled two-particle states. Using approximate theoretical arguments, we show that it is possible to construct highly specific energy- and position-resolved probes of Kondo correlations. Furthermore, we consider a quantum transport setup that can be driven away from equilibrium by a bias voltage. There, we show that singlet weights are enhanced by voltage even as the Kondo effect is weakened by it. This exposes a patently nonequilibrium mechanism for the generation of Kondo-like entanglement that is inherently different from its equilibrium counterpart. ",Resolving the nonequilibrium Kondo singlet in energy- and position-space using quantum measurements " Cell spreading requires a major reorganisation of the actin cytoskeleton, from a cortical structure to a lamellipodium where filaments are mostly parallel to the substrate. We propose a model inspired by the physics of nematic liquid crystals and fluid membranes, in which the coupling between actin mechanics, filaments orientation, and the local curvature of the cell membrane naturally yields the collective reorientation of actin filaments at the highly curved edge of a spreading cell. Filament orientation increases the traction force exerted by the frictional flow of polymerising actin on the substrate, creating a positive feedback loop between edge curvature, filament alignment, and traction force that promotes cell spreading. We establish the condition under which this feedback leads to a full wetting transition, which we interpret as the initiation of a lamellipodium, and we uncover the existence of bi-stability between partial and full spreading which could trigger spontaneous cell polarization and lead to migration. ",Model of lamellipodium initiation during cell spreading " Using the formalism of the conditional amplitude, we study the response part of the exchange-correlation potential in the strong-coupling limit of density functional theory, analysing its peculiar features and comparing it with the response potential averaged over the coupling constant for small atoms and for the hydrogen molecule. We also use a simple one-dimensional model of a stretched heteronuclear molecule to derive exact properties of the response potential in the strong-coupling limit. The simplicity of the model allows us to unveil relevant features also of the exact Kohn-Sham potential and its different components, namely the appearance of a second peak in the correlation kinetic potential on the side of the most electronegative atom. ",Response potential in the strong-interaction limit of DFT: Analysis and comparison with the coupling-constant average " We construct a formalism for evolving spherically symmetric black hole initial data sets within a canonical approach to quantum gravity. This problem can be formulated precisely in quantum reduced loop gravity, a framework which has been successfully applied to give a full theory derivation of loop quantum cosmology. We extend this setting by implementing a particular choice of partial gauge which is then used to select a kinematical Hilbert space where the symmetry reduction is imposed through semiclassical states. The main result of this investigation is an effective Hamiltonian that can be used to solve for quantum black hole geometries by evolving classical black hole initial data sets. ",Quantum evolution of black hole initial data sets: Foundations " A detailed study of event by event fluctuation of maximum particle density of the produced particles in narrow pseudo-rapidity interval in terms of the scaled variance {\omega} has been carried out for 16O-AgBr, 28Si-AgBr and 32S-AgBr interactions at an incident momentum of 4.5 AGeV/c. For all the interactions the values of scaled variance are found to be greater than zero indicating the presence of strong event by event fluctuation of maximum particle density values in the multiparticle production process. The event by event fluctuations are found to decrease with the increase of pseudo-rapidity interval. Experimental analysis has been compared with the results obtained from the analysis of events simulated by the Ultra Relativistic Quantum Molecular Dynamics (UrQMD) model. UrQMD model could not replicate the experimental results. ",Event-By-Event Fluctuation of Maximum Particle Density in Narrow Pseudo-Rapidity Interval at a Few AGeV/c " The microscopic model of atomic diffusion is considered to describe the short-range order relaxation kinetics within the f.c.c.-Ni-Fe Permalloys. The model takes into account both the discrete and anisotropic characters of atomic jumps within the long-range field of concentration heterogeneities of the interacting atoms. The diffusion coefficients and activation energies for the disordered Ni-Fe permalloy are estimated with the evaluated probabilities of atomic jumps. As shown, the increasing of a temperature with a fixed composition influences on the 'potential' field of interatomic interaction ambiguously: the field 'potential' increases for defined coordination shells and decreases for some of other ones. Although the temperature increasing promotes the increasing of any atomic-probabilities jumps generally, but decreasing of the action of 'potential' field generated by the atoms of defined element and caused by its concentration heterogeneities onto the distant sites results in increasing of the atomic-jumps' probabilities of just this element, first of all, into the sites, which are more distant from the 'source' of heterogeneity. Within the framework of the static concentration waves' method along with the self-consistent field approximation, the Onsager-type kinetics equation is obtained to describe the long-range order relaxation by the L12-type superstructure. To calculate diffusivities for the ordered Ni3Fe permalloy, the independent, diffraction experimental data of the long-range order parameter relaxation are used. Theoretical curves of the long-range order time evolution for the non-stoichiometric f.c.c.-Ni-Fe permalloys are plotted. Decreasing of the concentration of alloying element results in decelerating of the long-range order parameter change and in increasing of its relaxation time. ",Diffusivities and kinetics of short-range and long-range orderings in Ni-Fe permalloys " We search for possible correlations between neutron star observables and thermodynamic quantities that characterize high density nuclear matter. We generate a set of model-independent equations of state describing stellar matter from a Taylor expansion around saturation density. Each equation of state which is a functional of the nuclear matter parameters is thermodynamically consistent, causal and compatible with astrophysical observations. We find that the neutron star tidal deformability and radius are strongly correlated with the pressure, the energy density and the sound velocity at different densities. Similar correlations are also exhibited by a large set of mean-field models based on non-relativistic and relativistic nuclear energy density functionals. These model independent correlations can be employed to constrain the equation of state at different densities above saturation from measurements of NS properties with multi-messenger observations. In particular, precise constraints on the radius of PSR J0030+0451 thanks to NICER observations would allow to better infer the properties of matter around two times the nuclear saturation density. ",Empirical constraints on the high-density equation of state from multi-messenger observables " We study the single and double lepton polarization asymmetries in the semileptonic $B$ meson decays $B \to K_1 (1270) \ell^+ \ell^-$ $\ell \equiv e$, $\mu$, $\tau$), where the strange $P$-wave meson, $K_1(1270)$, is the mixtures of the $K_{1A}$ and $K_{1B}$, which are the $1^3P_1$ and $1^1P_1$ states, respectively. The lepton polarization asymmetries show relatively strong dependency in the various region of dileptonic invariant mass. The lepton polarization asymmetries can also be used for determining the $K_1(1270)$--$K_1(1400)$ mixing angle, $\theta_{K_1}$ and new physics effects. Furthermore, it is shown that these asymmetries in $B\to K_1(1270)\ell^+\ell^-$ decay compared with those of $B\to K^*\ell^+\ell^-$ decay are more sensitive to the dileptonic invariant mass. ",Lepton polarization in $B \to K_1 \ell^+ \ell^-$ Decays " Large language models (LLMs) have played a pivotal role in revolutionizing various facets of our daily existence. Solving attention regression is a fundamental task in optimizing LLMs. In this work, we focus on giving a provable guarantee for the one-layer attention network objective function $L(X,Y) = \sum_{j_0 = 1}^n \sum_{i_0 = 1}^d ( \langle \langle \exp( \mathsf{A}_{j_0} x ) , {\bf 1}_n \rangle^{-1} \exp( \mathsf{A}_{j_0} x ), A_{3} Y_{*,i_0} \rangle - b_{j_0,i_0} )^2$. Here $\mathsf{A} \in \mathbb{R}^{n^2 \times d^2}$ is Kronecker product between $A_1 \in \mathbb{R}^{n \times d}$ and $A_2 \in \mathbb{R}^{n \times d}$. $A_3$ is a matrix in $\mathbb{R}^{n \times d}$, $\mathsf{A}_{j_0} \in \mathbb{R}^{n \times d^2}$ is the $j_0$-th block of $\mathsf{A}$. The $X, Y \in \mathbb{R}^{d \times d}$ are variables we want to learn. $B \in \mathbb{R}^{n \times d}$ and $b_{j_0,i_0} \in \mathbb{R}$ is one entry at $j_0$-th row and $i_0$-th column of $B$, $Y_{*,i_0} \in \mathbb{R}^d$ is the $i_0$-column vector of $Y$, and $x \in \mathbb{R}^{d^2}$ is the vectorization of $X$. In a multi-layer LLM network, the matrix $B \in \mathbb{R}^{n \times d}$ can be viewed as the output of a layer, and $A_1= A_2 = A_3 \in \mathbb{R}^{n \times d}$ can be viewed as the input of a layer. The matrix version of $x$ can be viewed as $QK^\top$ and $Y$ can be viewed as $V$. We provide an iterative greedy algorithm to train loss function $L(X,Y)$ up $\epsilon$ that runs in $\widetilde{O}( ({\cal T}_{\mathrm{mat}}(n,n,d) + {\cal T}_{\mathrm{mat}}(n,d,d) + d^{2\omega}) \log(1/\epsilon) )$ time. Here ${\cal T}_{\mathrm{mat}}(a,b,c)$ denotes the time of multiplying $a \times b$ matrix another $b \times c$ matrix, and $\omega\approx 2.37$ denotes the exponent of matrix multiplication. ","A Fast Optimization View: Reformulating Single Layer Attention in LLM Based on Tensor and SVM Trick, and Solving It in Matrix Multiplication Time" " Taxonomies are semantic hierarchies of concepts. One limitation of current taxonomy learning systems is that they define concepts as single words. This position paper argues that contextualized word representations, which recently achieved state-of-the-art results on many competitive NLP tasks, are a promising method to address this limitation. We outline a novel approach for taxonomy learning that (1) defines concepts as synsets, (2) learns density-based approximations of contextualized word representations, and (3) can measure similarity and hypernymy among them. ",Learning Taxonomies of Concepts and not Words using Contextualized Word Representations: A Position Paper " Normative modelling is an emerging method for understanding the underlying heterogeneity within brain disorders like Alzheimer Disease (AD) by quantifying how each patient deviates from the expected normative pattern that has been learned from a healthy control distribution. Since AD is a multifactorial disease with more than one biological pathways, multimodal magnetic resonance imaging (MRI) neuroimaging data can provide complementary information about the disease heterogeneity. However, existing deep learning based normative models on multimodal MRI data use unimodal autoencoders with a single encoder and decoder that may fail to capture the relationship between brain measurements extracted from different MRI modalities. In this work, we propose multi-modal variational autoencoder (mmVAE) based normative modelling framework that can capture the joint distribution between different modalities to identify abnormal brain structural patterns in AD. Our multi-modal framework takes as input Freesurfer processed brain region volumes from T1-weighted (cortical and subcortical) and T2-weighed (hippocampal) scans of cognitively normal participants to learn the morphological characteristics of the healthy brain. The estimated normative model is then applied on Alzheimer Disease (AD) patients to quantify the deviation in brain volumes and identify the abnormal brain structural patterns due to the effect of the different AD stages. Our experimental results show that modeling joint distribution between the multiple MRI modalities generates deviation maps that are more sensitive to disease staging within AD, have a better correlation with patient cognition and result in higher number of brain regions with statistically significant deviations compared to a unimodal baseline model with all modalities concatenated as a single input. ",Normative Modeling using Multimodal Variational Autoencoders to Identify Abnormal Brain Structural Patterns in Alzheimer Disease We propose a novel mechanism of the Kondo effect driven by a chirality imbalance (or chiral chemical potential) of relativistic light fermions. This effect is realized by the mixing between a right- or left-handed fermion and a heavy impurity in the chirality imbalanced matter even at zero density. This is different from the usual Kondo effect induced by finite density. We derive the Kondo effect from both a perturbative calculation and a mean-field approach. We also discuss the temperature dependence of the Kondo effect. The Kondo effect at nonzero chiral chemical potential can be tested by future lattice simulations. ,Kondo effect driven by chirality imbalance " The second-generation (2G) mobile systems were developed in response to the growing demand for a system that met mobile communication demands while also providing greater interoperability with other systems. International organizations were crucial in the development of a system that would offer better services, be more transparent, and be more interoperable with other networks. The aim of having a single set of standards for networks worldwide was sadly not realized by the 2G network standards. The third generation (3G) was born. It was called the universal terrestrial mobile system (UMTS), which is European telecommunications standards institute (ETSI) driven. IMT-2000 is the international telecommunication union-telecommunication standardization sector (ITU-T) name for the 3G network. Wide-band code division multiple access (WCDMA) is the air interface technology for the UMTS. This platform offers many services that are based on the Internet, along with video calling, imaging, etc. Further advancements to mobile network technology led to long term evolution (LTE), a technology referred to as 4G. The primary goal of LTE was to improve the speed and capacity of mobile networks while lowering latency. As we move to an ALL-IP system, mobile networks' design becomes much simpler. LTE uses orthogonal frequency division multiplexing (OFDM) in its air interface. This paper details all mentioned mobile generations, as well as all the differences between them in terms of hardware and software architectures. ",From 2G to 4G Mobile Network: Architecture and Key Performance Indicators " CRExplorer version 1.6.7 was released on July 5, 2016. This version includes the following new features and improvements: Scopus: Using ""File"" - ""Import"" - ""Scopus"", CRExplorer reads files from Scopus. The file format ""CSV"" (including citations, abstracts and references) should be chosen in Scopus for downloading records. Export facilities: Using ""File"" - ""Export"" - ""Scopus"", CRExplorer exports files in the Scopus format. Using ""File"" - ""Export"" - ""Web of Science"", CRExplorer exports files in the Web of Science format. These files can be imported in other bibliometric programs (e.g. VOSviewer). Space bar: Select a specific cited reference in the cited references table, press the space bar, and all bibliographic details of the CR are shown. Internal file format: Using ""File"" - ""Save"", working files are saved in the internal file format ""*.cre"". The files include all data including matching results and manual matching corrections. The files can be opened by using ""File"" - ""Open"". ",New features of CitedReferencesExplorer (CRExplorer) " We report the first measurement of the parity-violating asymmetry A_PV in the elastic scattering of polarized electrons from 208Pb. A_PV is sensitive to the radius of the neutron distribution (Rn). The result A_PV = 0.656 \pm 0.060 (stat) \pm 0.014 (syst) ppm corresponds to a difference between the radii of the neutron and proton distributions Rn - Rp = 0.33 +0.16 -0.18 fm and provides the first electroweak observation of the neutron skin which is expected in a heavy, neutron-rich nucleus. ",Measurement of the Neutron Radius of 208Pb Through Parity-Violation in Electron Scattering " We show how to compute analytically time and space dependent correlations in one dimensional quantum integrable systems with an impurity. Our approach is based on a description of these systems in terms of massless scattering of quasiparticles. Correlators follow then from matrix elements of local operators between multiparticle states, the ``massless form factors''. Although an infinite sum of these form factors has to be considered in principle, we find that for current, spin, and energy operators, only a few (typically two or three) are necessary to obtain an accuracy of more than $1\%$, for {\bf arbitrary coupling strength}, that is all the way from short to large distances. As examples we compute, at zero temperature, the frequency dependent conductance in a Luttinger liquid with impurity, the spectral function in the double well problem of dissipative quantum mechanics and part of the space dependent succeptibility in the Kondo model . ",Form factors approach to current correlations in one dimensional systems with impurities " We consider the joint SPX-VIX calibration within a general class of Gaussian polynomial volatility models in which the volatility of the SPX is assumed to be a polynomial function of a Gaussian Volterra process defined as a stochastic convolution between a kernel and a Brownian motion. By performing joint calibration to daily SPX-VIX implied volatility surface data between 2012 and 2022, we compare the empirical performance of different kernels and their associated Markovian and non-Markovian models, such as rough and non-rough path-dependent volatility models. In order to ensure an efficient calibration and a fair comparison between the models, we develop a generic unified method in our class of models for fast and accurate pricing of SPX and VIX derivatives based on functional quantization and Neural Networks. For the first time, we identify a \textit{conventional one-factor Markovian continuous stochastic volatility model} that is able to achieve remarkable fits of the implied volatility surfaces of the SPX and VIX together with the term structure of VIX futures. What is even more remarkable is that our conventional one-factor Markovian continuous stochastic volatility model outperforms, in all market conditions, its rough and non-rough path-dependent counterparts with the same number of parameters. ",Joint SPX-VIX calibration with Gaussian polynomial volatility models: deep pricing with quantization hints Dynamics of a one-dimensional growing front with an unstable straight profile are analyzed. We argue that a coarsening process occurs if and only if the period \lambda of the steady state solution is an increasing function of its amplitude A. This statement is rigorously proved for two important classes of conserved and nonconserved models by investigating the phase diffusion equation of the steady pattern. We further provide clear numerical evidences for the growth equation of a stepped crystal surface. ,When does coarsening occur in the dynamics of one-dimensional fronts ? " Building general-purpose robots to perform a diverse range of tasks in a large variety of environments in the physical world at the human level is extremely challenging. It requires the robot learning to be sample-efficient, generalizable, compositional, and incremental. In this work, we introduce a systematic learning framework called SAGCI-system towards achieving these above four requirements. Our system first takes the raw point clouds gathered by the camera mounted on the robot's wrist as the inputs and produces initial modeling of the surrounding environment represented as a file of Unified Robot Description Format (URDF). Our system adopts a learning-augmented differentiable simulation that loads the URDF. The robot then utilizes the interactive perception to interact with the environment to online verify and modify the URDF. Leveraging the differentiable simulation, we propose a model-based learning algorithm combining object-centric and robot-centric stages to efficiently produce policies to accomplish manipulation tasks. We apply our system to perform articulated object manipulation tasks, both in the simulation and the real world. Extensive experiments demonstrate the effectiveness of our proposed learning framework. Supplemental materials and videos are available on https://sites.google.com/view/egci. ","SAGCI-System: Towards Sample-Efficient, Generalizable, Compositional, and Incremental Robot Learning" " Context. Apertif is a multi-beam receiver system for the Westerbork Synthesis Radio Telescope that operates at 1.1-1.5 GHz, which overlaps with various radio services, resulting in contamination of astronomical signals with radio-frequency interference (RFI). Aims. We analyze approaches to mitigate Apertif interference and design an automated detection procedure for its imaging mode. Using this approach, we present long-term RFI detection results of over 300 Apertif observations. Methods. Our approach is based on the AOFlagger detection approach. We introduce several new features, including ways to deal with ranges of invalid data (e.g. caused by shadowing) in both the SumThreshold and scale-invariant rank operator steps; pre-calibration bandpass calibration; auto-correlation flagging; and HI flagging avoidance. These methods are implemented in a new framework that uses the Lua language for scripting, which is new in AOFlagger version 3. Results. Our approach removes RFI fully automatically, and is robust and effective enough for further calibration and (continuum) imaging of these data. Analysis of 304 observations show an average of 11.1% of lost data due to RFI with a large spread. We observe 14.6% RFI in auto-correlations. Computationally, AOFlagger achieves a throughput of 370 MB/s on a single computing node. Compared to published machine learning results, the method is one to two orders of magnitude faster. ",An interference detection strategy for Apertif based on AOFlagger 3 " Air free-cooled data centers (DCs) have not existed in the tropical zone due to the unique challenges of year-round high ambient temperature and relative humidity (RH). The increasing availability of servers that can tolerate higher temperatures and RH due to the regulatory bodies' prompts to raise DC temperature setpoints sheds light upon the feasibility of air free-cooled DCs in tropics. However, due to the complex psychrometric dynamics, operating the air free-cooled DC in tropics generally requires adaptive control of supply air condition to maintain the computing performance and reliability of the servers. This paper studies the problem of controlling the supply air temperature and RH in a free-cooled tropical DC below certain thresholds. To achieve the goal, we formulate the control problem as Markov decision processes and apply deep reinforcement learning (DRL) to learn the control policy that minimizes the cooling energy while satisfying the requirements on the supply air temperature and RH. We also develop a constrained DRL solution for performance improvements. Extensive evaluation based on real data traces collected from an air free-cooled testbed and comparisons among the unconstrained and constrained DRL approaches as well as two other baseline approaches show the superior performance of our proposed solutions. ",Deep Reinforcement Learning for Tropical Air Free-Cooled Data Center Control " Riemann surface carries a natural line bundle, the determinant bundle. The space of sections of this line bundle (or its multiples) constitutes a natural non-abelian generalization of the spaces of theta functions on the Jacobian. There has been much progress in the last few years towards a better understanding of these spaces, including a rigorous proof of the celebrated Verlinde formula which gives their dimension. This survey paper tries to explain what is now known and what remains open. ",Vector bundles on curves and generalized theta functions: recent results and open problems " Interpreting the inner function of neural networks is crucial for the trustworthy development and deployment of these black-box models. Prior interpretability methods focus on correlation-based measures to attribute model decisions to individual examples. However, these measures are susceptible to noise and spurious correlations encoded in the model during the training phase (e.g., biased inputs, model overfitting, or misspecification). Moreover, this process has proven to result in noisy and unstable attributions that prevent any transparent understanding of the model's behavior. In this paper, we develop a robust interventional-based method grounded by causal analysis to capture cause-effect mechanisms in pre-trained neural networks and their relation to the prediction. Our novel approach relies on path interventions to infer the causal mechanisms within hidden layers and isolate relevant and necessary information (to model prediction), avoiding noisy ones. The result is task-specific causal explanatory graphs that can audit model behavior and express the actual causes underlying its performance. We apply our method to vision models trained on classification tasks. On image classification tasks, we provide extensive quantitative experiments to show that our approach can capture more stable and faithful explanations than standard attribution-based methods. Furthermore, the underlying causal graphs reveal the neural interactions in the model, making it a valuable tool in other applications (e.g., model repair). ",Causal Analysis for Robust Interpretability of Neural Networks " In the presence of strong magnetic fields the electronic bandstructure of graphene drastically changes. The Dirac cone collapses into discrete non-equidistant Landau levels, which can be externally tuned by changing the magnetic field. In contrast to conventional materials, specific Landau levels are selectively addressable using circularly polarized light. Exploiting these unique properties, we propose the design of a tunable laser operating in the technologically promising terahertz spectral range. To uncover the many-particle physics behind the emission of light, we perform a fully quantum mechanical investigation of the non-equilibrium dynamics of electrons, phonons, and photons in optically pumped Landau-quantized graphene embedded into an optical cavity. The gained microscopic insights allow us to predict optimal experimental conditions to realize a technologically promising terahertz laser. ",Proposal for a tunable graphene-based terahertz Landau-level laser We show that every abelian Polish group is the topological factor-group of a closed subgroup of the full unitary group of a separable Hilbert space with the strong operator topology. It follows that all orbit equivalence relations induced by abelian Polish group actions are Borel reducible to some orbit equivalence relations induced by actions of the unitary group. ,On a universality property of some abelian Polish groups " This paper explores the grounding issue concerning multimodal semantic representation from a computational cognitive-linguistic view. Five perceptual properties of groundedness are annotated and analyzed: Affordance, Perceptual salience, Object number, Gaze cueing, and Ecological Niche Association (ENA). We annotated selected images from the Flickr30k dataset with exploratory analyses and statistical modeling of their captions. Our findings suggest that a comprehensive understanding of an object or event requires cognitive attention, semantic distinctions in linguistic expression, and multimodal construction. During this construction process, viewers integrate situated meaning and affordance into multimodal semantics, which is consolidated into image captions used in the image-text dataset incorporating visual and textual elements. Our findings suggest that situated meaning and affordance grounding are critical for grounded natural language understanding systems to generate appropriate responses and show the potential to advance the understanding of human construal in diverse situations. ",Exploring the Grounding Issues in Image Caption " In this paper we discuss four problems regarding Markov equivalences for subclasses of loopless mixed graphs. We classify these four problems as finding conditions for internal Markov equivalence, which is Markov equivalence within a subclass, for external Markov equivalence, which is Markov equivalence between subclasses, for representational Markov equivalence, which is the possibility of a graph from a subclass being Markov equivalent to a graph from another subclass, and finding algorithms to generate a graph from a certain subclass that is Markov equivalent to a given graph. We particularly focus on the class of maximal ancestral graphs and its subclasses, namely regression graphs, bidirected graphs, undirected graphs, and directed acyclic graphs, and present novel results for representational Markov equivalence and algorithms. ",Markov Equivalences for Subclasses of Loopless Mixed Graphs " Fluorodeoxyglucose Positron Emission Tomography (FDG-PET) combined with Computed Tomography (CT) scans are critical in oncology to the identification of solid tumours and the monitoring of their progression. However, precise and consistent lesion segmentation remains challenging, as manual segmentation is time-consuming and subject to intra- and inter-observer variability. Despite their promise, automated segmentation methods often struggle with false positive segmentation of regions of healthy metabolic activity, particularly when presented with such a complex range of tumours across the whole body. In this paper, we explore the application of the nnUNet to tumour segmentation of whole-body PET-CT scans and conduct different experiments on optimal training and post-processing strategies. Our best model obtains a Dice score of 69\% and a false negative and false positive volume of 6.27 and 5.78 mL respectively, on our internal test set. This model is submitted as part of the autoPET 2023 challenge. Our code is available at: https://github.com/anissa218/autopet\_nnunet ",Autopet Challenge 2023: nnUNet-based whole-body 3D PET-CT Tumour Segmentation " We analyze the data and discuss their implications for the microscopic origin of the low frequency flux noise in superconducting circuits. We argue that this noise is produced by spins at the superconductor insulator boundary whose dynamics is due to RKKY interaction. We show that this mechanism explains size independence of the noise, different frequency dependences of the spectra reported in large and small SQUIDs and gives the correct intensity for realistic parameters. ",Microscopic origin of low frequency flux noise in Josephson circuits " This paper introduces a new type of unsupervised learning algorithm, based on the alignment of sentences and Harris's (1951) notion of interchangeability. The algorithm is applied to an untagged, unstructured corpus of natural language sentences, resulting in a labelled, bracketed version of the corpus. Firstly, the algorithm aligns all sentences in the corpus in pairs, resulting in a partition of the sentences consisting of parts of the sentences that are similar in both sentences and parts that are dissimilar. This information is used to find (possibly overlapping) constituents. Next, the algorithm selects (non-overlapping) constituents. Several instances of the algorithm are applied to the ATIS corpus (Marcus et al., 1993) and the OVIS (Openbaar Vervoer Informatie Systeem (OVIS) stands for Public Transport Information System.) corpus (Bonnema et al., 1997). Apart from the promising numerical results, the most striking result is that even the simplest algorithm based on alignment learns recursion. ",Bootstrapping Syntax and Recursion using Alignment-Based Learning " This work introduces efficient symbolic algorithms for quantitative reactive synthesis. We consider resource-constrained robotic manipulators that need to interact with a human to achieve a complex task expressed in linear temporal logic. Our framework generates reactive strategies that not only guarantee task completion but also seek cooperation with the human when possible. We model the interaction as a two-player game and consider regret-minimizing strategies to encourage cooperation. We use symbolic representation of the game to enable scalability. For synthesis, we first introduce value iteration algorithms for such games with min-max objectives. Then, we extend our method to the regret-minimizing objectives. Our benchmarks reveal that our symbolic framework not only significantly improves computation time (up to an order of magnitude) but also can scale up to much larger instances of manipulation problems with up to 2x number of objects and locations than the state of the art. ",Efficient Symbolic Approaches for Quantitative Reactive Synthesis with Finite Tasks " Kernel methods form a theoretically-grounded, powerful and versatile framework to solve nonlinear problems in signal processing and machine learning. The standard approach relies on the \emph{kernel trick} to perform pairwise evaluations of a kernel function, leading to scalability issues for large datasets due to its linear and superlinear growth with respect to the training data. Recently, we proposed \emph{no-trick} (NT) kernel adaptive filtering (KAF) that leverages explicit feature space mappings using data-independent basis with constant complexity. The inner product defined by the feature mapping corresponds to a positive-definite finite-rank kernel that induces a finite-dimensional reproducing kernel Hilbert space (RKHS). Information theoretic learning (ITL) is a framework where information theory descriptors based on non-parametric estimator of Renyi entropy replace conventional second-order statistics for the design of adaptive systems. An RKHS for ITL defined on a space of probability density functions simplifies statistical inference for supervised or unsupervised learning. ITL criteria take into account the higher-order statistical behavior of the systems and signals as desired. However, this comes at a cost of increased computational complexity. In this paper, we extend the NT kernel concept to ITL for improved information extraction from the signal without compromising scalability. Specifically, we focus on a family of fast, scalable, and accurate estimators for ITL using explicit inner product space (EIPS) kernels. We demonstrate the superior performance of EIPS-ITL estimators and combined NT-KAF using EIPS-ITL cost functions through experiments. ",Fast Estimation of Information Theoretic Learning Descriptors using Explicit Inner Product Spaces " The self-adjoint and $m$-sectorial extensions of coercive Sturm-Liouville operators are characterised, under minimal smoothness conditions on the coefficients of the differential expression. ",Selfadjoint and $m$ sectorial extensions of Sturm-Liouville operators " Exoplanet detection in the past decade by efforts including NASA's Kepler and TESS missions has discovered many worlds that differ substantially from planets in our own Solar system, including more than 400 exoplanets orbiting binary or multi-star systems. This not only broadens our understanding of the diversity of exoplanets, but also promotes our study of exoplanets in the complex binary and multi-star systems and provides motivation to explore their habitability. In this study, we analyze orbital stability of exoplanets in non-coplanar circumbinary systems using a numerical simulation method, with which a large number of circumbinary planet samples are generated in order to quantify the effects of various orbital parameters on orbital stability. We also train a machine learning model that can quickly determine the stability of the circumbinary planetary systems. Our results indicate that larger inclinations of the planet tend to increase the stability of its orbit, but change in the planet's mass range between Earth and Jupiter has little effect on the stability of the system. In addition, we find that Deep Neural Networks (DNNs) have higher accuracy and precision than other machine learning algorithms. ",Analyzing the Stability of Non-coplanar Circumbinary Planets using Machine Learning " The mysteries of sunspot penumbrae have been under an intense scrutiny for the past 10 years. During this time, some models have been proposed and refuted, while the surviving ones had to be modified, adapted and evolved to explain the ever-increasing array of observational constraints. In this contribution I will review two of the present models, emphasizing their contributions to this field, but also pinpointing some of their inadequacies to explain a number of recent observations at very high spatial resolution. To help explaining these new observations I propose some modifications to each of them. These modifications bring those two seemingly opposite models closer together into a general picture that agrees well with recent 3D magneto-hydrodynamic simulations. ",Models and Observations of Sunspot Penumbrae " Within the NRQCD factorization framework, we compute the next-to-leading-order QCD corrections to the gluon fragmentation into the ${}^1S_0^{(1,8)}$ Fock components of a quarkonium, at the lowest order in velocity expansion. We follow the operator definition of the fragmentation function advanced by Collins and Soper. The key technique underpinning our calculation is the sector decomposition method widely used in the area of multi-loop computation. It is found that the NLO QCD corrections have significant effects, and qualitatively modify the profiles of the corresponding leading-order fragmentation functions. ","Next-to-leading-order QCD corrections to gluon fragmentation into ${}^1S_0^{(1,8)}$ quarkonia" " Maintenance of existing software requires a large amount of time for comprehending the source code. The architecture of a software, however, may not be clear to maintainers if up to date documentations are not available. Software clustering is often used as a remodularisation and architecture recovery technique to help recover a semantic representation of the software design. Due to the diverse domains, structure, and behaviour of software systems, the suitability of different clustering algorithms for different software systems are not investigated thoroughly. Research that introduce new clustering techniques usually validate their approaches on a specific domain, which might limit its generalisability. If the chosen test subjects could only represent a narrow perspective of the whole picture, researchers might risk not being able to address the external validity of their findings. This work aims to fill this gap by introducing a new approach, Explaining Software Clustering for Remodularisation, to evaluate the effectiveness of different software clustering approaches. This work focuses on hierarchical clustering and Bunch clustering algorithms and provides information about their suitability according to the features of the software, which as a consequence, enables the selection of the most optimum algorithm and configuration from our existing pool of choices for a particular software system. The proposed framework is tested on 30 open source software systems with varying sizes and domains, and demonstrates that it can characterise both the strengths and weaknesses of the analysed software clustering algorithms using software features extracted from the code. The proposed approach also provides a better understanding of the algorithms behaviour through the application of dimensionality reduction techniques. ",E-SC4R: Explaining Software Clustering for Remodularisation " We consider the problem of designing an Ansatz for the fermion-photon vertex function, using three-dimensional quantum electrodynamics as a test case. In many existing studies, restrictions have been placed on the form of the vertex Ansatz by making the unsubstantiated assumption that in the quenched, massless limit the Landau gauge Dyson-Schwinger equations admit a trivial solution. We demonstrate, without recourse to this assumption, the existence of a non-local gauge in which the fermion propagator is the bare propagator. This result is used to provide a viable Ansatz for part of the vertex function. ",Deconstructing the vertex Ansatz in three dimensional quantum electrodynamics " Motivated by recent proposal by Potter et al. [Phys. Rev. X 6, 031026 (2016)] concerning possible thermoelectric signatures of Dirac composite fermions, we perform a systematic experimental study of thermoelectric transport of an ultrahigh-mobility GaAs/AlxGa1-xAs two dimensional electron system at filling factor v = 1/2. We demonstrate that the thermopower Sxx and Nernst Sxy are symmetric and anti-symmetric with respect to B = 0 T, respectively. The measured properties of thermopower Sxx at v = 1/2 are consistent with previous experimental results. The Nernst signals Sxy of v = 1/2, which have not been reported previously, are non-zero and show a power law relation with temperature in the phonon-drag dominant region. In the electron-diffusion dominant region, the Nernst signals Sxy of v = 1/2 are found to be significantly smaller than the linear temperature dependent values predicted by Potter et al., and decreasing with temperature faster than linear dependence. ",Thermopower and Nernst measurements in a half-filled lowest Landau level " Capsule Networks have great potential to tackle problems in structural biology because of their attention to hierarchical relationships. This paper describes the implementation and application of a Capsule Network architecture to the classification of RAS protein family structures on GPU-based computational resources. The proposed Capsule Network trained on 2D and 3D structural encodings can successfully classify HRAS and KRAS structures. The Capsule Network can also classify a protein-based dataset derived from a PSI-BLAST search on sequences of KRAS and HRAS mutations. Our results show an accuracy improvement compared to traditional convolutional networks, while improving interpretability through visualization of activation vectors. ",Capsule Networks for Protein Structure Classification and Prediction " This paper considers the cooperative output regulation problem for linear multi-agent systems with a directed communication graph, heterogeneous linear subsystems, and an exosystem whose output is available to only a subset of subsystems. Both the cases with nominal and uncertain linear subsystems are studied. For the case with nominal linear subsystems, a distributed adaptive observer-based controller is designed, where the distributed adaptive observer is implemented for the subsystems to estimate the exogenous signal. For the case with uncertain linear subsystems, the proposed distributed observer and the internal model principle are combined to solve the robust cooperative output regulation problem. Compared with the existing works, one main contribution of this paper is that the proposed control schemes can be designed and implemented by each subsystem in a fully distributed fashion for general directed graphs. For the special case with undirected graphs, a distributed output feedback control law is further presented. ",Fully Distributed Adaptive Controllers for Cooperative Output Regulation of Heterogeneous Linear Multi-agent Systems with Directed Graphs " Particle acceleration and heating at mildly relativistic magnetized shocks in electron-ion plasma are investigated with unprecedentedly high-resolution two-dimensional particle-in-cell simulations that include ion-scale shock rippling. Electrons are super-adiabatically heated at the shock, and most of the energy transfer from protons to electrons takes place at or downstream of the shock. We are the first to demonstrate that shock rippling is crucial for the energization of electrons at the shock. They remain well below equipartition with the protons. The downstream electron spectra are approximately thermal with a limited supra-thermal power-law component. Our results are discussed in the context of wakefield acceleration and the modelling of electromagnetic radiation from blazar cores. ",Mildly relativistic magnetized shocks in electron-ion plasmas -- II. Particle acceleration and heating " The Neptune Trojans are the most recent addition to the panoply of Solar system small body populations. The orbit of the first discovered member, 2001 QR$_{322}$, was investigated shortly after its discovery, based on early observations of the object, and it was found to be dynamically stable on timescales comparable to the age of the Solar system. As further observations were obtained of the object over the following years, the best-fit solution for its orbit changed. We therefore carried out a new study of 2001 QR$_{322}$'s orbit in 2010, finding that it lay on the boundary between dynamically stable and unstable regions in Neptune's Trojan cloud, and concluding that further observations were needed to determine the true stability of the object's orbit. Here we follow up on that earlier work, and present the preliminary results of a dynamical study using an updated fit to 2001 QR$_{322}$'s orbit. Despite the improved precision with which the orbit of 2001 QR$_{322}$ is known, we find that the best-fit solution remains balanced on a knife-edge, lying between the same regions of stability and instability noted in our earlier work. In the future, we intend to carry out new observations that should hopefully refine the orbit to an extent that its true nature can finally be disentangled. ",2001 QR$_{322}$ - an update on Neptune's first unstable Trojan companion " The continuous imaginary-time quantum Monte Carlo method with the worm update algorithm is applied to explore the ground state properties of the spin-1/2 Heisenberg model with antiferromagnetic (AF) coupling $J>0$ and ferromagnetic (F) coupling $J^{\prime}<0$ along zigzag and armchair directions, respectively, on honeycomb lattice. It is found that by enhancing the F coupling $J^{\prime}$ between zigzag AF chains, the system is smoothly crossover from one-dimensional zigzag spin chains to a two-dimensional magnetic ordered state. In absence of an external field, the system is in a stripe order phase. In presence of uniform and staggered fields, the uniform and staggered out-of-plane magnetizations appear while the stripe order keeps in $xy$ plane, and a second-order quantum phase transition (QPT) at a critical staggered field is observed. The critical exponents of correlation length for QPTs induced by a staggered field for the cases with $J>0$, $J^{\prime}<0$ and $J<0$, $J^{\prime}>0$ are obtained to be $\nu=0.677(2)$ and $0.693(0)$, respectively, indicating that both cases belong to O(3) universality. The scaling behavior in a staggered field is analyzed, and the ground state phase diagrams in the plane of coupling ratio and staggered field are presented for two cases. The temperature dependence of susceptibility and specific heat of both systems in external magnetic fields is also discussed. ",Quantum Monte Carlo Study on the Spin-1/2 Honeycomb Heisenberg Model with Mixing Antiferromagnetic and Ferromagnetic Interactions in External Magnetic Fields " High spectral purity frequency agile room temperature sources in the terahertz spectrum are foundational elements for imaging, sensing, metrology, and communications. Here we present a chip scale optical parametric oscillator based on an integrated nonlinear microresonator that provides broadly tunable single frequency and multi frequency oscillators in the terahertz regime. Through optical to terahertz down conversion using a plasmonic nanoantenna array, coherent terahertz radiation spanning 2.8 octaves is achieved from 330 GHz to 2.3 THz, with 20 GHz cavity mode limited frequency tuning step and 10 MHz intracavity mode continuous frequency tuning range at each step. By controlling the microresonator intracavity power and pump resonance detuning, tunable multi frequency terahertz oscillators are also realized. Furthermore, by stabilizing the microresonator pump power and wavelength, sub 100 Hz linewidth of the terahertz radiation with 10-15 residual frequency instability is demonstrated. The room temperature generation of both single frequency, frequency agile terahertz radiation and multi frequency terahertz oscillators in the chip scale platform offers unique capabilities in metrology, sensing, imaging and communications. ",Coherent terahertz radiation with 2.8-octave tunability through chip-scale photomixed microresonator optical parametric oscillation " Right-handed neutrinos with MeV to GeV mass are very promising candidates for dark matter (DM). Not only can they solve the missing satellite puzzle, the cusp-core problem of inner DM density profiles, and the too-big-to fail problem, {\it i.e.} that the unobserved satellites are too big to not have visible stars, but they can also account for the Standard Model (SM) neutrino masses at one loop. We perform a comprehensive study of the right-handed neutrino parameter space and impose the correct observed relic density and SM neutrino mass differences and mixings. We find that the DM masses are in agreement with bounds from big-bang nucleosynthesis, but that these constraints induce sizeable DM couplings to the charged SM leptons. We then point out that previously overlooked limits from current and future lepton flavour violation experiments such as MEG and SINDRUM heavily constrain the allowed parameter space. Since the DM is leptophilic, we also investigate electron recoil as a possible direct detection signal, in particular in the XENON1T experiment. We find that despite the large coupling and low backgrounds, the energy thresholds are still too high and the predicted cross sections too low due to the heavy charged mediator, whose mass is constrained by LEP limits. ","MeV neutrino dark matter: Relic density, lepton flavour violation and electron recoil" " The concept of intrinsic credibility has been recently introduced to check the credibility of ""out of the blue"" findings without any prior support. A significant result is deemed intrinsically credible if it is in conflict with a sceptical prior derived from the very same data that would make the effect non-significant. In this paper I propose to use Bayesian prior-predictive tail probabilities to assess intrinsic credibility. For the standard 5% significance level, this leads to a new p-value threshold that is remarkably close to the recently proposed p<0.005 standard. I also introduce the credibility ratio, the ratio of the upper to the lower limit of a standard confidence interval for the corresponding effect size. I show that the credibility ratio has to be smaller than 5.8 such that a significant finding is also intrinsically credible. Finally, a p-value for intrinsic credibility is proposed that is a simple function of the ordinary p-value and has a direct frequentist interpretation in terms of the probability of replicating an effect. ",The Assessment of Intrinsic Credibility and a New Argument for p<0.005 " This is the continuation of previous article. For subspaces $M^n(t)$ and $M^{n-m}(t)$ which are invariant manifolds of the differential equation under consideration we build a change of variables which splits this equation into a system of two independent equations. A notion of equivalence of linear differential equations of different orders is introduced. Necessary and sufficient conditions of this equivalence are given. These results are applied to the Flocke-Lyapunov theory for linear equations with periodic coefficients with a period T. In the case when monodromy matrix of the equation has negative eigenvalues, thus reduction in $R^m$ to an equation with constant coeficcients is possible only with doubling of reduction matrix period, we prove the possibility of splitting off in $R^m$ of equations with negative eigenvalues of monodromy matrix with the help of a real matrix without period doubling. For the fundamental matrix of solutions of an equation with periodic coefficients $X(t), X(t)=E$, we find representation $X(t)=\Phi(t)e^{Ht}\Phi^{+}(0)$ with real rectangular matrices $H$ and $\Phi(t), \Phi(t)=\Phi(t+T)$. We bring two applications of these results: 1) reduction of nonlinear differential equation in $R^n$ with distinguished linear part which is periodic with period T to the equation in $R^m, m>n$, with a constiant matrix of coefficients of the linear part; 2) for introdusing of amplitude-phase coordinates in the neigbourhood of periodic orbit of autonomous differential equation with separation of the linear part with constant matrix of coefficients. ",On invariant manifolds of linear differential equations. II " The hydrodynamic limit of a kinetic Cucker-Smale model is investigated. In addition to the free-transport of individuals and the Cucker-Smale alignment operator, the model under consideration includes a strong local alignment term. This term was recently derived as the singular limit of an alignment operator due to Motsch and Tadmor. The model is enhanced with the addition of noise and a confinement potential. The objective of this work is the rigorous investigation of the singular limit corresponding to strong noise and strong local alignment. The proof relies on the relative entropy method and entropy inequalities which yield the appropriate convergence results. The resulting limiting system is an Euler-type flocking system. ",Hydrodynamic limit of the kinetic Cucker-Smale flocking model " In conjunction with huge recent progress in camera and computer vision technology, camera-based sensors have increasingly shown considerable promise in relation to tactile sensing. In comparison to competing technologies (be they resistive, capacitive or magnetic based), they offer super-high-resolution, while suffering from fewer wiring problems. The human tactile system is composed of various types of mechanoreceptors, each able to perceive and process distinct information such as force, pressure, texture, etc. Camera-based tactile sensors such as GelSight mainly focus on high-resolution geometric sensing on a flat surface, and their force measurement capabilities are limited by the hysteresis and non-linearity of the silicone material. In this paper, we present a miniaturised dome-shaped camera-based tactile sensor that allows accurate force and tactile sensing in a single coherent system. The key novelty of the sensor design is as follows. First, we demonstrate how to build a smooth silicone hemispheric sensing medium with uniform markers on its curved surface. Second, we enhance the illumination of the rounded silicone with diffused LEDs. Third, we construct a force-sensitive mechanical structure in a compact form factor with usage of springs to accurately perceive forces. Our multi-modal sensor is able to acquire tactile information from multi-axis forces, local force distribution, and contact geometry, all in real-time. We apply an end-to-end deep learning method to process all the information. ",A Miniaturised Camera-based Multi-Modal Tactile Sensor " The large variation in seed mass among species inspired a vast array of theoretical and empirical research attempting to explain this variation. So far, seed mass variation was investigated by two classes of studies: one class focuses on species varying in seed mass within communities, while the second focuses on variation between communities, most often with respect to resource gradients. Here, we develop a model capable of simultaneously explaining variation in seed mass within and between communities. The model describes resource competition (for both soil and light resources) in annual communities and incorporates two fundamental aspects: light asymmetry (higher light acquisition per unit biomass for larger individuals) and growth allometry (negative dependency of relative growth rate on plant biomass). Results show that both factors are critical in determining patterns of seed mass variation. In general, growth allometry increases the reproductive success of small-seeded species while light asymmetry increases the reproductive success of large-seeded species. Increasing availability of soil resources increases light competition, thereby increasing the reproductive success of large-seeded species and ultimately the community (weighted) mean seed mass. An unexpected prediction of the model is that maximum variation in community seed mass (a measure of functional diversity) occurs under intermediate levels of soil resources. Extensions of the model incorporating size-dependent seed survival and disturbance also show patterns consistent with empirical observations. These overall results suggest that the mechanisms captured by the model are important in determining patterns of species and functional diversity. ",Seed mass diversity along resource gradients: the role of allometric growth rate and size-asymmetric competition " In this work we study families of pairs of window functions and lattices which lead to Gabor frames which all possess the same frame bounds. To be more precise, for every generalized Gaussian $g$, we will construct an uncountable family of lattices $\lbrace \Lambda_\tau \rbrace$ such that each pairing of $g$ with some $\Lambda_\tau$ yields a Gabor frame, and all pairings yield the same frame bounds. On the other hand, for each lattice we will find a countable family of generalized Gaussians $\lbrace g_i \rbrace$ such that each pairing leaves the frame bounds invariant. Therefore, we are tempted to speak about ""Gabor Frame Sets of Invariance"". ",Gabor Frame Sets of Invariance - A Hamiltonian Approach to Gabor Frame Deformations " X-ray observations of solar flares routinely reveal an impulsive high-energy and a gradual low-energy emission component, whose relationship is one of the key issues of solar flare study. The gradual and impulsive emission components are believed to be associated with, respectively, the thermal and nonthermal components identified in spectral fitting. In this paper, a prominent about 50 second hard X-ray (HXR) pulse of a simple GOES class C7.5 flare on 20 February 2002 is used to study the association between high energy, non-thermal and impulsive evolution, and low energy, thermal and gradual evolution. We use regularized methods to obtain time derivatives of photon fluxes to quantify the time evolution as a function of photon energy, obtaining a break energy between impulsive and gradual behavior. These break energies are consistent with a constant value of about 11 keV in agreement with those found spectroscopically between thermal and non-thermal components, but the relative errors of the former are greater than 15% and much greater than the a few percent errors found from the spectral fitting. These errors only weakly depend on assuming an underlying spectral model for the photons, pointing to the current data being inadequate to reduce the uncertainties rather than there being a problem associated with an assumed model. The time derivative method is used to test for the presence of a 'pivot energy' in this flare. Although these pivot energies are marginally consistent with a constant value of about 9 keV, its values in the HXR rise phase appear to be lower than those in the decay phase. ",Relationship between Hard and Soft X-ray Emission Components of a Solar Flare To a rational homology sphere graph manifold one can associate a weighted tree invariant called splice diagram. It was shown earlier that the splice diagram determines the universal abelian cover of the manifold. We will in this article turn the proof of this in to an algorithm to explicitly construct the universal abelian cover from the splice diagram. ,Constructing Universal Abelian Covers of Graph Manifolds " We propose a method for lossy image compression based on recurrent, convolutional neural networks that outperforms BPG (4:2:0 ), WebP, JPEG2000, and JPEG as measured by MS-SSIM. We introduce three improvements over previous research that lead to this state-of-the-art result. First, we show that training with a pixel-wise loss weighted by SSIM increases reconstruction quality according to several metrics. Second, we modify the recurrent architecture to improve spatial diffusion, which allows the network to more effectively capture and propagate image information through the network's hidden state. Finally, in addition to lossless entropy coding, we use a spatially adaptive bit allocation algorithm to more efficiently use the limited number of bits to encode visually complex image regions. We evaluate our method on the Kodak and Tecnick image sets and compare against standard codecs as well recently published methods based on deep neural networks. ",Improved Lossy Image Compression with Priming and Spatially Adaptive Bit Rates for Recurrent Networks " The competent programmer hypothesis states that most programmers are competent enough to create correct or almost correct source code. Because this implies that bugs should usually manifest through small variations of the correct code, the competent programmer hypothesis is one of the fundamental assumptions of mutation testing. Unfortunately, it is still unclear if the competent programmer hypothesis holds and past research presents contradictory claims. Within this article, we provide a new perspective on the competent programmer hypothesis and its relation to mutation testing. We try to re-create real-world bugs through chains of mutations to understand if there is a direct link between mutation testing and bugs. The lengths of these paths help us to understand if the source code is really almost correct, or if large variations are required. Our results indicate that while the competent programmer hypothesis seems to be true, mutation testing is missing important operators to generate representative real-world bugs. ",A new perspective on the competent programmer hypothesis through the reproduction of bugs with repeated mutations " We prove that any $\ell$ positive definite $d \times d$ matrices, $M_1,\ldots,M_\ell$, of full rank, can be simultaneously spectrally balanced in the following sense: for any $k < d$ such that $\ell \leq \lfloor \frac{d-1}{k-1} \rfloor$, there exists a matrix $A$ satisfying $\frac{\lambda_1(A^T M_i A) }{ \mathrm{Tr}( A^T M_i A ) } < \frac{1}{k}$ for all $i$, where $\lambda_1(M)$ denotes the largest eigenvalue of a matrix $M$. This answers a question posed by Peres, Popov and Sousi and completes the picture described in that paper regarding sufficient conditions for transience of self-interacting random walks. Furthermore, in some cases we give quantitative bounds on the transience of such walks. ",How many matrices can be spectrally balanced simultaneously? " We introduce and analyze a discontinuous Galerkin method for the numerical modelling of the equations of Multiple-Network Poroelastic Theory (MPET) in the dynamic formulation. The MPET model can comprehensively describe functional changes in the brain considering multiple scales of fluids. Concerning the spatial discretization, we employ a high-order discontinuous Galerkin method on polygonal and polyhedral grids and we derive stability and a priori error estimates. The temporal discretization is based on a coupling between a Newmark $\beta$-method for the momentum equation and a $\theta$-method for the pressure equations. After the presentation of some verification numerical tests, we perform a convergence analysis using an agglomerated mesh of a geometry of a brain slice. Finally we present a simulation in a three dimensional patient-specific brain reconstructed from magnetic resonance images. The model presented in this paper can be regarded as a preliminary attempt to model the perfusion in the brain. ",Numerical Modelling of the Brain Poromechanics by High-Order Discontinuous Galerkin Methods " We study contraction under a Markov semi-group and influence bounds for functions in $L^2$ tail spaces, i.e. functions all of whose low level Fourier coefficients vanish. It is natural to expect that certain analytic inequalities are stronger for such functions than for general functions in $L^2$. In the positive direction we prove an $L^{p}$ Poincar\'{e} inequality and moment decay estimates for mean $0$ functions and for all $1$ 1.5--2 bar radius) can be explained by the swing amplification theory and that the effects of the bar are not negligible in the inner regions ($R<$ 1.5--2 bar radius). These results suggest that, although grand-design spiral arms in barred galaxies are affected by the stellar bar, the grand-design spiral arms essentially originate not as bar-driven stationary density waves, but rather as self-excited dynamic patterns. We imply that a rigidly rotating grand-design spiral could not be a reasonable dynamical model for investigating gas flows and cloud formation even in barred spiral galaxies. ",Short-term dynamical evolution of grand-design spirals in barred galaxies " This paper is concerned with the existence of transition fronts for a one-dimensional twopatch model with KPP reaction terms. Density and flux conditions are imposed at the interface between the two patches. We first construct a pair of suitable super-and subsolutions by making full use of information of the leading edges of two KPP fronts and gluing them through the interface conditions. Then, an entire solution obtained thanks to a limiting argument is shown to be a transition front moving from one patch to the other one. This propagating solution admits asymptotic past and future speeds, and it connects two different fronts, each associated with one of the two patches. The paper thus provides the first example of a transition front for a KPP-type two-patch model with interface conditions. ",KPP transition fronts in a one-dimensional two-patch habitat " This work is concerned with the study of the adaptivity properties of nonparametric regression estimators over the $d$-dimensional sphere within the global thresholding framework. The estimators are constructed by means of a form of spherical wavelets, the so-called needlets, which enjoy strong concentration properties in both harmonic and real domains. The author establishes the convergence rates of the $L^p$-risks of these estimators, focussing on their minimax properties and proving their optimality over a scale of nonparametric regularity function spaces, namely, the Besov spaces. ",Adaptive global thresholding on the sphere " Most of today's popular deep architectures are hand-engineered to be generalists. However, this design procedure usually leads to massive redundant, useless, or even harmful features for specific tasks. Unnecessarily high complexities render deep nets impractical for many real-world applications, especially those without powerful GPU support. In this paper, we attempt to derive task-dependent compact models from a deep discriminant analysis perspective. We propose an iterative and proactive approach for classification tasks which alternates between (1) a pushing step, with an objective to simultaneously maximize class separation, penalize co-variances, and push deep discriminants into alignment with a compact set of neurons, and (2) a pruning step, which discards less useful or even interfering neurons. Deconvolution is adopted to reverse 'unimportant' filters' effects and recover useful contributing sources. A simple network growing strategy based on the basic Inception module is proposed for challenging tasks requiring larger capacity than what the base net can offer. Experiments on the MNIST, CIFAR10, and ImageNet datasets demonstrate our approach's efficacy. On ImageNet, by pushing and pruning our grown Inception-88 model, we achieve more accurate models than Inception nets generated during growing, residual nets, and popular compact nets at similar sizes. We also show that our grown Inception nets (without hard-coded dimension alignment) clearly outperform residual nets of similar complexities. ",Grow-Push-Prune: aligning deep discriminants for effective structural network compression " The Klein group contains only four elements. Nevertheless this little group contains a number of remarkable entry points to current highways of modern representation theory of groups. In this paper, we shall describe all possible ways in which the Klein group can act on vector spaces over a field of two elements. These are called representations of the Klein group. This description involves some powerful visual methods of representation theory which builds on the work of generations of mathematicians starting roughly with the work of K. Weiestrass. We also discuss some applications to properties of duality and Heller shifts of the representations of the Klein group. ",Representations of The miraculous Klein group " In the present work, we explore the existence, stability and dynamics of single and multiple vortex ring states that can arise in Bose-Einstein condensates. Earlier works have illustrated the bifurcation of such states, in the vicinity of the linear limit, for isotropic or anisotropic three-dimensional harmonic traps. Here, we extend these states to the regime of large chemical potentials, the so-called Thomas-Fermi limit, and explore their properties such as equilibrium radii and inter-ring distance, for multi-ring states, as well as their vibrational spectra and possible instabilities. In this limit, both the existence and stability characteristics can be partially traced to a particle picture that considers the rings as individual particles oscillating within the trap and interacting pairwise with one another. Finally, we examine some representative instability scenarios of the multi-ring dynamics including breakup and reconnections, as well as the transient formation of vortex lines. ","Single and Multiple Vortex Rings in Three-Dimensional Bose-Einstein Condensates: Existence, Stability and Dynamics" " Nonperturbative model of glueball is studied. The model is based on the nonperturbative quantization technique suggested by Heisenberg. 2- and 4-point Green functions for a gauge potential are expressed in terms of two scalar fields. The first scalar field describes quantum fluctuations of the subgroup $SU(n) \subset SU(N)$, and the second one describes quantum fluctuations of the coset $SU(N) / SU(n)$. An effective Lagrangian for the scalar fields is obtained. The coefficients for all terms in the Lagrangian are calculated, and it is shown that they depend on $\dim SU(n), \dim SU(N)$. It is demonstrated that a spherically symmetric solution describing the glueball does exist. ",Scalar model of SU(N) glueball \`a la Heisenberg " Light extra U(1) gauge bosons, so called hidden photons, which reside in a hidden sector have attracted much attention since they are a well motivated feature of many scenarios beyond the Standard Model and furthermore could mediate the interaction with hidden sector dark matter. We review limits on hidden photons from past electron beam dump experiments including two new limits from such experiments at KEK and Orsay. In addition, we study the possibility of having dark matter in the hidden sector. A simple toy model and different supersymmetric realisations are shown to provide viable dark matter candidates in the hidden sector that are in agreement with recent direct detection limits. ",Hidden Photons in connection to Dark Matter " Currently, there are about 3 dozen known super-Earth (M < 10 MEarth), of which 8 are transiting planets suitable for atmospheric follow-up observations. Some of the planets are exposed to extreme temperatures as they orbit close to their host stars, e.g., CoRot-7b, and all of these planets have equilibrium temperatures significantly hotter than the Earth. Such planets can develop atmospheres through (partial) vaporization of their crustal and/or mantle silicates. We investigated the chemical equilibrium composition of such heated systems from 500 - 4000 K and total pressures from 10-6 to 10+2 bars. The major gases are H2O and CO2 over broad temperature and pressure ranges, and Na, K, O2, SiO, and O at high temperatures and low pressures. We discuss the differences in atmospheric composition arising from vaporization of SiO2-rich (i.e., felsic) silicates (like Earth's continental crust) and MgO-, FeO-rich (i.e., mafic) silicates like the bulk silicate Earth. The computational results will be useful in planning spectroscopic studies of the atmospheres of Earth-like exoplanets. ",Vaporization of the Earth: Application to Exoplanet Atmospheres " We consider the ability of deep neural networks to represent data that lies near a low-dimensional manifold in a high-dimensional space. We show that deep networks can efficiently extract the intrinsic, low-dimensional coordinates of such data. We first show that the first two layers of a deep network can exactly embed points lying on a monotonic chain, a special type of piecewise linear manifold, mapping them to a low-dimensional Euclidean space. Remarkably, the network can do this using an almost optimal number of parameters. We also show that this network projects nearby points onto the manifold and then embeds them with little error. We then extend these results to more general manifolds. ",Efficient Representation of Low-Dimensional Manifolds using Deep Networks The formulation of quasi-local conformal Killling horizons(CKH) is extended to include rotation. This necessitates that the horizon be foliated by 2-spheres which may be distorted. Matter degrees of freedom which fall through the horizon is taken to be a real scalar field. We show that these rotating CKHs also admit a first law in differential form. ,Quasilocal rotating conformal Killing horizons " We study function spaces consisting of analytic functions with fast decay on horizontal strips of the complex plane with respect to a given weight function. Their duals, so called spaces of (ultra)hyperfunctions of fast growth, generalize the spaces of Fourier hyperfunctions and Fourier ultrahyperfunctions. An analytic representation theory for their duals is developed and applied to characterize the non-triviality of these function spaces in terms of the growth order of the weight function. In particular, we show that the Gelfand-Shilov spaces of Beurling type $\mathcal{S}^{(p!)}_{(M_p)}$ and Roumieu type $\mathcal{S}^{\{p!\}}_{\{M_p\}}$ are non-trivial if and only if $$ \sup_{p \geq 2}\frac{(\log p)^p}{h^pM_p} < \infty, $$ for all $h > 0$ and some $h > 0$, respectively. We also study boundary values of holomorphic functions in spaces of ultradistributions of exponential type, which may be of quasianalytic type. ",On the non-triviality of certain spaces of analytic functions. Hyperfunctions and ultrahyperfunctions of fast growth " Open Science has been a rising theme in the landscape of science policy in recent years. The goal is to make research that emerges from publicly funded science to become findable, accessible, interoperable and reusable (FAIR) for use by other researchers. Knowledge utilization policies aim to efficiently make scientific knowledge beneficial for society at large. This paper demonstrates how Astronomy aspires to be open and transparent given their criteria for high research quality, which aim at pushing knowledge forward and clear communication of findings. However, the use of quantitative metrics in research evaluation puts pressure on the researcher, such that taking the extra time for transparent publishing of data and results is difficult, given that astronomers are not rewarded for the quality of research papers, but rather their quantity. This paper explores the current mode of openness in Astronomy and how incentives due to funding, publication practices and indicators affect this field. The paper concludes with some recommendations on how policies such as making science more open have the potential to contribute to scientific quality in Astronomy. ",Knowledge Utilization and Open Science Policies: Noble aims that ensure quality research or Ordering discoveries like a pizza? " Determinantal point processes (DPPs) offer an elegant tool for encoding probabilities over subsets of a ground set. Discrete DPPs are parametrized by a positive semidefinite matrix (called the DPP kernel), and estimating this kernel is key to learning DPPs from observed data. We consider the task of learning the DPP kernel, and develop for it a surprisingly simple yet effective new algorithm. Our algorithm offers the following benefits over previous approaches: (a) it is much simpler; (b) it yields equally good and sometimes even better local maxima; and (c) it runs an order of magnitude faster on large problems. We present experimental results on both real and simulated data to illustrate the numerical performance of our technique. ",Fixed-point algorithms for learning determinantal point processes We discuss the hypotheses that cosmological baryon asymmetry and entropy were produced in the early Universe by phase transition of the scalar fields in the framework of spontaneous baryogenesis scenario. We show that annihilation of the matter-antimatter clouds during the cosmological hydrogen recombination could distort of the CMB anisotropies and polarization by delay of the recombination. After recombination the annihilation of the antibaryonic clouds (ABC) and baryonic matter can produce peak-like reionization at the high redshifts before formation of quasars and early galaxy formation. We discuss the constraints on the parameters of spontaneous baryogenesis scenario by the recent WMAP CMB anisotropy and polarization data and on possible manifestation of the antimatter clouds in the upcoming PLANCK data. ,Antimatter from the cosmological baryogenesis and the anisotropies and polarization of the CMB radiation " We study the effects of coupling between layers of stochastic neural field models with laminar structure. In particular, we focus on how the propagation of waves of neural activity in each layer is affected by the coupling. Synaptic connectivities within and between each layer are determined by integral kernels of an integrodifferential equation describing the temporal evolution of neural activity. Excitatory neural fields, with purely positive connectivities, support traveling fronts in each layer, whose speeds are increased when coupling between layers is considered. Studying the effects of noise, we find coupling also serves to reduce the variance in the position of traveling fronts, as long as the noise sources to each layer are not completely correlated. Neural fields with asymmetric connectivity support traveling pulses whose speeds are decreased by interlaminar coupling. Again, coupling reduces the variance in traveling pulse position, when noise is considered that is not totally correlated between layers. To derive our stochastic results, we employ a small-noise expansion, also assuming interlaminar connectivity scales similarly. Our asymptotic results agree reasonably with accompanying numerical simulations. ",Coupling layers regularizes wave propagation in laminar stochastic neural fields " We consider K-mouflage models which are K-essence theories coupled to matter. We analyse their quantum properties and in particular the quantum corrections to the classical Lagrangian. We setup the renormalisation programme for these models and show that K-mouflage theories involve a recursive construction whereby each set of counter-terms introduces new divergent quantum contributions which in turn must be subtracted by new counter-terms. This tower of counter-terms can be constructed by recursion and allows one to calculate the finite renormalised action of the model. In particular, the classical action is not renormalised and the finite corrections to the renormalised action contain only higher derivative operators. We establish an operational criterion for classicality, where the corrections to the classical action are negligible, and show that this is satisfied in cosmological and astrophysical situations for (healthy) K-mouflage models which pass the solar system tests. We also find that these models are quantum stable around astrophysical and cosmological backgrounds. We then consider the possible embedding of the K-mouflage models in an Ultra-Violet completion. We find that the healthy models which pass the solar system tests all violate the positivity constraint which would follow from the unitarity of the putative UV completion, implying that these healthy K-mouflage theories have no UV completion. We then analyse their behaviour at high energy and we find that the classicality criterion is satisfied in the vicinity of a high energy collision implying that the classical K-mouflage theory can be applied in this context. Moreover, the classical description becomes more accurate as the energy increases, in a way compatible with the classicalisation concept. ",The Quantum Field Theory of K-mouflage " We introduce a notion of freeness for $RO$-graded equivariant generalized homology theories, considering spaces or spectra $E$ such that the $R$-homology of $E$ splits as a wedge of the $R$-homology of induced virtual representation spheres. The full subcategory of these spectra is closed under all of the basic equivariant operations, and this greatly simplifies computation. Many examples of spectra and homology theories are included along the way. We refine this to a collection of spectra analogous to the pure and isotropic spectra considered by Hill--Hopkins--Ravenel. For these spectra, the $RO$-graded Bredon homology is extremely easy to compute, and if these spaces have additional structure, then this can also be easily determined. In particular, the homology of a space with this property naturally has the structure of a co-Tambara functor (and compatibly with any additional product structure). We work this out in the example of $BU_{\mathbb R}$ and coinduced versions of this. We finish by describing a readily computable bar and twisted bar spectra sequence, giving Bredon homology for various $E_{\infty}$ pushouts, and we apply this to describe the homology of $BBU_{\mathbb R}$. ",Freeness and equivariant stable homotopy " An asymptotic expansion is established for time averages of translation flows on flat surfaces. This result, which extends earlier work of A.Zorich and G.Forni, yields limit theorems for translation flows. The argument, close in spirit to that of G.Forni, uses the approximation of ergodic integrals by holonomy-invariant Hoelder cocycles on trajectories of the flows. The space of holonomy-invariant Hoelder cocycles is finite-dimensional, and is given by an explicit construction. First, a symbolic representation for a uniquely ergodic translation flow is obtained following S.Ito and A.M. Vershik, and then, the space of cocycles is constructed using a family of finitely-additive complex-valued holonomy-invariant measures on the asymptotic foliations of a Markov compactum. ",Finitely-additive measures on the asymptotic foliations of a Markov compactum " In the context of the spectral action and the noncommutative geometry approach to the standard model, we build a model based on a larger symmetry. With this ""grand symmetry"" it is natural to have the scalar field necessary to obtain the Higgs mass in the vicinity of 126 GeV. This larger symmetry mixes gauge and spin degrees of freedom without introducing extra fermions. Requiring the noncommutative space to be an almost commutative geometry (i.e. the product of manifold by a finite dimensional internal space) gives conditions for the breaking of this grand symmetry to the standard model. ","Grand Symmetry, Spectral Action, and the Higgs mass" " The main goal of this paper is to discuss how to integrate the possibilities of crowdsourcing platforms with systems supporting workflow to enable the engagement and interaction with business tasks of a wider group of people. Thus, this work is an attempt to expand the functional capabilities of typical business systems by allowing selected process tasks to be performed by unlimited human resources. Opening business tasks to crowdsourcing, within established Business Process Management Systems (BPMS) will improve the flexibility of company processes and allow for lower work-load and greater specialization among the staff employed on-site. The presented conceptual work is based on the current international standards in this field, promoted by Workflows Management Coalition. To this end, the functioning of business platforms was analysed and their functionality was presented visually, followed by a proposal and a discussion of how to implement crowdsourcing into workflow systems. ",Deploying Crowdsourcing for Workflow Driven Business Process " Keeping the two fundamental postulates of the special theory of relativity, the principle of relativity and the constancy of the one-way velocity of light in all inertial frames of reference, and assuming two generalized Finslerian structures of gravity-free space and time in the usual inertial coordinate system, we can modify the special theory of relativity. The modified theory is still characterized by the localized Lorentz transformation between any two usual inertial coordinate systems. It together with the quantum mechanics theory features a convergent and invariant quantum field theory. The modified theory also involves a new velocity distribution for free particles that is different from the Maxwell distribution. It is claimed that the deviation of the new distribution from its previous formula will provide experimental means of judging the modified special relativity theory. ",Velocity distribution of free particles in the modified special relativity theory " While solid-state devices offer naturally reliable hardware for modern classical computers, thus far quantum information processors resemble vacuum tube computers in being neither reliable nor scalable. Strongly correlated many body states stabilized in topologically ordered matter offer the possibility of naturally fault tolerant computing, but are both challenging to engineer and coherently control and cannot be easily adapted to different physical platforms. We propose an architecture which achieves some of the robustness properties of topological models but with a drastically simpler construction. Quantum information is stored in the symmetry-protected degenerate ground states of spin-1 chains, while quantum gates are performed by adiabatic non-Abelian holonomies using only single-site fields and nearest-neighbor couplings. Gate operations respect the symmetry, and so inherit some protection from noise and disorder from the symmetry-protected ground states. ",Holonomic quantum computing in symmetry-protected ground states of spin chains " The design of better automated dialogue evaluation metrics offers the potential of accelerate evaluation research on conversational AI. However, existing trainable dialogue evaluation models are generally restricted to classifiers trained in a purely supervised manner, which suffer a significant risk from adversarial attacking (e.g., a nonsensical response that enjoys a high classification score). To alleviate this risk, we propose an adversarial training approach to learn a robust model, ATT (Adversarial Turing Test), that discriminates machine-generated responses from human-written replies. In contrast to previous perturbation-based methods, our discriminator is trained by iteratively generating unrestricted and diverse adversarial examples using reinforcement learning. The key benefit of this unrestricted adversarial training approach is allowing the discriminator to improve robustness in an iterative attack-defense game. Our discriminator shows high accuracy on strong attackers including DialoGPT and GPT-3. ",An Adversarially-Learned Turing Test for Dialog Generation Models The Cauchy problem for the Yang-Mills system in two space dimensions is treated for data with minimal regularity assumptions. In the classical case of data in $L^2$-based Sobolev spaces we have to assume that the number of derivatives is more than $3/4$ above the critical regularity with respect to scaling. For data in $L^r$-based Fourier-Lebesgue spaces this result can be improved by $1/4$ derivative in the sense of scaling as $r \to 1$ . ,Low regularity well-posedness for the Yang-Mills system in 2D " Most of the existing works for dialogue generation are data-driven models trained directly on corpora crawled from websites. They mainly focus on improving the model architecture to produce better responses but pay little attention to considering the quality of the training data contrastively. In this paper, we propose a multi-level contrastive learning paradigm to model the fine-grained quality of the responses with respect to the query. A Rank-aware Calibration (RC) network is designed to construct the multi-level contrastive optimization objectives. Since these objectives are calculated based on the sentence level, which may erroneously encourage/suppress the generation of uninformative/informative words. To tackle this incidental issue, on one hand, we design an exquisite token-level strategy for estimating the instance loss more accurately. On the other hand, we build a Knowledge Inference (KI) component to capture the keyword knowledge from the reference during training and exploit such information to encourage the generation of informative words. We evaluate the proposed model on a carefully annotated dialogue dataset and the results suggest that our model can generate more relevant and diverse responses compared to the baseline models. ",Enhancing Dialogue Generation via Multi-Level Contrastive Learning We prove that elliptic tubes over properly convex domains of the real projective space are C-convex and complete Kobayashi-hyperbolic. We also study a natural construction of complexification of convex real projective manifolds. ,Convexity properties and complete hyperbolicity of Lempert's elliptic tubes " Quasar-galaxy pairs at small separations are important probes of gas flows in the disk-halo interface in galaxies. We study host galaxies of 198 MgII absorbers at $0.39\le z_{abs}\le1.05$ that show detectable nebular emission lines in the SDSS spectra. We report measurements of impact parameter (5.9$\le D[kpc]\le$16.9) and absolute B-band magnitude ($-18.7\le {\rm M_B}\le -22.3$ mag) of host galaxies of 74 of these absorbers using multi-band images from the DESI Legacy Imaging Survey, more than doubling the number of known host galaxies with $D\le17$ kpc. This has allowed us to quantify the relationship between MgII rest equivalent width($W_{2796}$) and D, with best-fit parameters of $W_{2796}(D=0) = 3.44\pm 0.20$ Angstrom and an exponential scale length of 21.6$^{+2.41}_{-1.97}$ $kpc$. We find a significant anti-correlation between $M_B$ and D, and $M_B$ and $W_{2796}$, consistent with the brighter galaxies producing stronger MgII absorption. We use stacked images to detect average emissions from galaxies in the full sample. Using these images and stacked spectra, we derive the mean stellar mass ($9.4\le log(M_*/M_\odot) \le 9.8$), star formation rate ($2.3\le{\rm SFR}[M_\odot yr^{-1}] \le 4.5$), age (2.5$-$4 Gyr), metallicity (12+log(O/H)$\sim$8.3) and ionization parameter (log~q[cm s$^{-1}$]$\sim$ 7.7) for these galaxies. The average $M_*$ found is less compared to those of MgII absorbers studied in the literature. The average SFR and metallicity inferred are consistent with that expected in the main sequence and the known stellar mass-metallicity relation, respectively. High spatial resolution follow-up spectroscopic and imaging observations of this sample are imperative for probing gas flows close to the star-forming regions of high-$z$ galaxies. ",Nature of the Galaxies On Top Of Quasars producing MgII absorption " The collapse of a massive star with low angular momentum content is commonly argued to result in the formation of a black hole without an accompanying bright transient. Our goal in this Letter is to understand the flow in and around a newly-formed black hole, involving accretion and rotation, via general relativistic hydrodynamics simulations aimed at studying the conditions under which infalling material can accrete without forming a centrifugally supported structure and, as a result, generate no effective feedback. If the feedback from the black hole is, on the other hand, significant, the collapse would be halted and we suggest that the event is likely to be followed by a bright transient. We find that feedback is only efficient if the specific angular momentum of the infalling material at the innermost stable circular orbit exceeds that of geodesic circular flow at that radius by at least $\approx 20\%$. We use the results of our simulations to constrain the maximal stellar rotation rates of the disappearing massive progenitors PHL293B-LBV and N6946-BH1, and to provide an estimate of the overall rate of disappearing massive stars. We find that about a few percent of single O-type stars with measured rotational velocities are expected to spin below the critical value before collapse and are thus predicted to vanish without a trace. ",On the maximum stellar rotation to form a black hole without an accompanying luminous transient " Relativistic runaway electron avalanches (RREAs) are generally accepted as a source of thunderstorms gamma-ray radiation. Avalanches can multiply in the electric field via the relativistic feedback mechanism based on processes with gamma-rays and positrons. This paper shows that a non-uniform electric field geometry can lead to the new RREAs multiplication mechanism - ""reactor feedback"", due to the exchange of high-energy particles between different accelerating regions within a thundercloud. A new method for the numerical simulation of RREA dynamics within heterogeneous electric field structures is proposed. The developed analytical description and the numerical simulation enables us to derive necessary conditions for TGF occurrence in the system with the reactor feedback Observable properties of TGFs influenced by the proposed mechanism are discussed. ",Relativistic runaway electron avalanches within complex thunderstorm electric field structures " An oscillating universe model is discussed, in which the singularity of the initial state of the universe is avoided by postulating an upper limit on spacetime curvature. This also results in the devising of the simplest possible structure - a primitive particle (usually called preon), which can be considered as the basic constituent of matter that have no properties except for the property of carrying a unit chromoelectric charge. The SU(3)xU(1)-symmetry of its field results in the emergence of a unique set of structures reproducing exactly the observed variety of the fundamental fermions and gauge bosons. The discussed scheme allows finding answers to many fundamental questions of the standard particle physics and cosmology based on a very few primary constituents. ",An oscillating universe model based on chromoelectric fields " Vehicle to Vehicle (V2V) communication has a great potential to improve reaction accuracy of different driver assistance systems in critical driving situations. Cooperative Adaptive Cruise Control (CACC), which is an automated application, provides drivers with extra benefits such as traffic throughput maximization and collision avoidance. CACC systems must be designed in a way that are sufficiently robust against all special maneuvers such as cutting-into the CACC platoons by interfering vehicles or hard braking by leading cars. To address this problem, a Neural- Network (NN)-based cut-in detection and trajectory prediction scheme is proposed in the first part of this paper. Next, a probabilistic framework is developed in which the cut-in probability is calculated based on the output of the mentioned cut-in prediction block. Finally, a specific Stochastic Model Predictive Controller (SMPC) is designed which incorporates this cut-in probability to enhance its reaction against the detected dangerous cut-in maneuver. The overall system is implemented and its performance is evaluated using realistic driving scenarios from Safety Pilot Model Deployment (SPMD). ",A Learning-based Stochastic MPC Design for Cooperative Adaptive Cruise Control to Handle Interfering Vehicles " This paper studies the optimal mechanisms for the vertically integrated utility to dispatch and incentivize the third-party demand response (DR) providers in its territory. A framework is proposed, with three-layer coupled Stackelberg and simultaneous games, to study the interactions and competitions among the profit-seeking process of the utility, the third-party DR providers, and the individual end users (EUs) in the DR programs. Two coupled single-leader-multiple-followers Stackelberg games with a three-layer structure are proposed to capture the interactions among the utility (modeled in the upper layer), the third-party DR providers (modeled in the middle layer), and the EUs in each DR program (modeled in the lower layer). The competitions among the EUs in each DR program is captured through a non-cooperative simultaneous game. An inconvenience cost function is proposed to model the DR provision willingness and capacity of different EUs. The Stackelberg game between the middle-layer DR provider and the lower-layer EUs is solved by converting the original bi-level programming to a singlelevel programming. This converted single-level programming is embedded in an iterative algorithm toward solving the entire coupled games framework. Case studies are performed on IEEE 34-bus and IEEE 69-bus test systems to illustrate the application of the proposed framework. ",Optimal Utilization of Third-Party Demand Response Resources in Vertically Integrated Utilities: A Game Theoretic Approach " We study the early work scheduling problem on identical parallel machines in order to maximize the total early work, i.e., the parts of non-preemptive jobs executed before a common due date. By preprocessing and constructing an auxiliary instance which has several good properties, we propose an efficient polynomial time approximation scheme with running time $O(n)$, which improves the result in [Gy\""{o}rgyi, P., Kis, T. (2020). A common approximation framework for early work, late work, and resource leveling problems. {\it European Journal of Operational Research}, 286(1), 129-137], and a fully polynomial time approximation scheme with running time $O(n)$ when the number of machines is a fixed number, which improves the result in [Chen, X., Liang, Y., Sterna, M., Wang, W., B{\l}a\.{z}ewicz, J. (2020b). Fully polynomial time approximation scheme to maximize early work on parallel machines with common due date. {\it European Journal of Operational Research}, 284(1), 67-74], where $n$ is the number of jobs, and the hidden constant depends on the desired accuracy. ",Improved approximation schemes for early work scheduling on identical parallel machines with common due date " For a graph $H$, a graph $G$ is $H$-saturated if $G$ does not contain $H$ as a subgraph but for any $e \in E(\overline{G})$, $G+e$ contains $H$. In this note, we prove a sharp lower bound for the number of paths and walks on length $2$ in $n$-vertex $K_{r+1}$-saturated graphs. We then use this bound to give a lower bound on the spectral radii of such graphs which is asymptotically tight for each fixed $r$ and $n\to\infty$. ",$K_{r+1}$-saturated graphs with small spectral radius " Eta Carinae is considered to be a massive colliding wind binary system with a highly eccentric (e \sim 0.9), 5.54-yr orbit. However, the companion star continues to evade direct detection as the primary dwarfs its emission at most wavelengths. Using three-dimensional (3-D) SPH simulations of Eta Car's colliding winds and radiative transfer codes, we are able to compute synthetic observables across multiple wavebands for comparison to the observations. The models show that the presence of a companion star has a profound influence on the observed HST/STIS UV spectrum and H-alpha line profiles, as well as the ground-based photometric monitoring. Here, we focus on the Bore Hole effect, wherein the fast wind from the hot secondary star carves a cavity in the dense primary wind, allowing increased escape of radiation from the hotter/deeper layers of the primary's extended wind photosphere. The results have important implications for interpretations of Eta Car's observables at multiple wavelengths. ",Multi-Wavelength Implications of the Companion Star in Eta Carinae " It is known that the Kadison-Singer Problem (KS) and the Paving Conjecture (PC) are equivalent to the Bourgain-Tzafriri Conjecture (BT). Also, it is known that (PC) fails for $2$-paving projections with constant diagonal $1/2$. But the proofs of this fact are existence proofs. We will use variations of the discrete Fourier Transform matrices to construct concrete examples of these projections and projections with constant diagonal $1/r$ which are not $r$-pavable in a very strong sense. In 1989, Bourgain and Tzafriri showed that the class of zero diagonal matrices with small entries (on the order of $\le 1/log^{1+\epsilon}n$, for an $n$-dimensional Hilbert space) are {\em pavable}. It has always been assumed that this result also holds for the BT-Conjecture - although no one formally checked it. We will show that this is not the case. We will show that if the BT-Conjecture is true for vectors with small coefficients (on the order of $\le C/\sqrt{n}$) then the BT-Conjecture is true and hence KS and PC are true. ",The Bourgain-Tzafriri conjecture and concrete constructions of non-pavable projections " We report the detection of the cyclotron(CR) and magneto-plasmon(MPR) resonances near the Fermi surface of high-mobility 2DES, by microwave photoresistance measurements. We observe large amplitude photoresistance oscillations originating from higher-order CR, transitions between non-adjacent Landau levels. Such transitions are drastically enhanced in low magnetic field as compared to those previously known in the high field limit. The scattering time of the CR is found to be nearly one order of magnitude higher than that of Shubnikov-de-Haas oscillations. Finally, distinct photoresistance peaks are observed in addition to the CR features. These are identified as resonances of the low-frequency MP modes at a cut-off wavelength determined by the width of the 2DES sample. ",Microwave Photoresistance Measurements of Magneto-excitations near a 2D Fermi Surface " Expanding a lower-dimensional problem to a higher-dimensional space and then projecting back is often beneficial. This article rigorously investigates this perspective in the context of finite mixture models, namely how to improve inference for mixture models by using auxiliary variables. Despite the large literature in mixture models and several empirical examples, there is no previous work that gives general theoretical justification for including auxiliary variables in mixture models, even for special cases. We provide a theoretical basis for comparing inference for mixture multivariate models with the corresponding inference for marginal univariate mixture models. Analytical results for several special cases are established. We show that the probability of correctly allocating mixture memberships and the information number for the means of the primary outcome in a bivariate model with two Gaussian mixtures are generally larger than those in each univariate model. Simulations under a range of scenarios, including misspecified models, are conducted to examine the improvement. The method is illustrated by two real applications in ecology and causal inference. ",Improving Inference of Gaussian Mixtures Using Auxiliary Variables " Solar coronal mass ejections (CMEs) produce adverse space weather effects at Earth. Planets in the close habitable zone of magnetically active M dwarfs may experience more extreme space weather than at Earth, including frequent CME impacts leading to atmospheric erosion and leaving the surface exposed to extreme flare activity. Similar erosion may occur for hot Jupiters with close orbits around solar-like stars. We have developed a model, Forecasting a CME's Altered Trajectory (ForeCAT), which predicts a CME's deflection. We adapt ForeCAT to simulate CME deflections for the mid-type M dwarf V374 Peg and hot Jupiters with solar-type hosts. V374 Peg's strong magnetic fields can trap CMEs at the M dwarfs's Astrospheric Current Sheet, the location of the minimum in the background magnetic field. Solar-type CMEs behave similarly, but have much smaller deflections and do not get trapped at the Astrospheric Current Sheet. The probability of planetary impact decreases with increasing inclination of the planetary orbit with respect to the Astrospheric Current Sheet - 0.5 to 5 CME impacts per day for M dwarf exoplanets, 0.05 to 0.5 CME impacts per day for solar-type hot Jupiters. We determine the minimum planetary magnetic field necessary to shield a planet's atmosphere from the CME impacts. M dwarf exoplanets require values between tens and hundreds of Gauss. Hot Jupiters around a solar-type star, however, require a more reasonable <30 G. These values exceed the magnitude required to shield a planet from the stellar wind, suggesting CMEs may be the key driver of atmospheric losses. ",Probability of CME Impact on Exoplanets Orbiting M Dwarfs and Solar-Like Stars " Wereportonanewmultiscalemethodapproachforthestudyofsystemswith wide separation of short-range forces acting on short time scales and long-range forces acting on much slower scales. We consider the case of the Poisson-Boltzmann equation that describes the long-range forces using the Boltzmann formula (i.e. we assume the medium to be in quasi local thermal equilibrium). We developed a new approach where fields and particle information (mediated by the equations for their moments) are solved self-consistently. The new approach is implicit and numerically stable, providing exact energy conservation. We tested different implementations all leading to exact energy conservation. The new method requires the solution of a large set of non-linear equations. We considered three solution strategies: Jacobian Free Newton Krylov, an alternative, called field hiding, based on hiding part of the residual calculation and replacing them with direct solutions and a Direct Newton Schwarz solver that considers simplified single particle-based Jacobian. The field hiding strategy proves to be the most efficient approach. ",Implicit temporal discretization and exact energy conservation for particle methods applied to the Poisson-Boltzmann equation " Cosmic chronometers may be used to measure the age difference between passively evolving galaxy populations to calculate the Hubble parameter H(z). The age estimator emerges from the relationship between the amplitude of the rest frame Balmer break at 4000 angstroms and the age of a galaxy, assuming that there is one single stellar population within each galaxy. However, recent literature has shown possible contamination (up to 2.4% of the stellar mass in a high redshift sample) of a young component embedded within the predominantly old population of the quiescent galaxy. We compared the data with the predictions of each model, using a new approach of distinguishing between systematic and statistical errors (in previous works, these had incorrectly been added in quadrature) and evaluating the effects of contamination by a young stellar component. The ages inferred using cosmic chronometers represent a galaxy-wide average rather than a characteristic of the oldest population alone. The average contribution from the young component to the rest luminosity at 4000 angstroms may constitute a third of the luminosity in some samples, which means that this is far from negligible. This ratio is significantly dependent on stellar mass, proportional to M^{-0.7}. Consequently, the measurements of the absolute value of the age or the differential age between different redshifts are incorrect and make the previous calculations of H(z) very inaccurate. Some cosmological models, such as the Einstein-de Sitter model or quasi-steady state cosmology, which are rejected under the assumption of a purely old population, can be made compatible with the predicted ages of the Universe as a function of redshift if we take this contamination into account. However, the static Universe models are rejected by these H(z) measurements, even when this contamination is taken into account. ",Impact of young stellar components on quiescent galaxies: deconstructing cosmic chronometers " In this paper, we study locally strongly convex affine hyperspheres in the unimodular affine space $\mathbb{R}^{n+1}$ which, as Riemannian manifolds, are locally isometric to the Riemannian product of two Riemannian manifolds both possessing constant sectional curvatures. As the main result, a complete classification of such affine hyperspheres is established. Moreover, as direct consequences, affine hyperspheres of dimensions 3 and 4 with parallel Ricci tensor are also classified. ",On product affine hyperspheres in $\mathbb{R}^{n+1}$ " Fully homomorphic encryption is an encryption method with the property that any computation on the plaintext can be performed by a party having access to the ciphertext only. Here, we formally define and give schemes for quantum homomorphic encryption, which is the encryption of quantum information such that quantum computations can be performed given the ciphertext only. Our schemes allows for arbitrary Clifford group gates, but become inefficient for circuits with large complexity, measured in terms of the non-Clifford portion of the circuit (we use the ""$\pi/8$"" non-Clifford group gate, which is also known as the $T$-gate). More specifically, two schemes are proposed: the first scheme has a decryption procedure whose complexity scales with the square of the number of $T$-gates (compared with a trivial scheme in which the complexity scales with the total number of gates); the second scheme uses a quantum evaluation key of length given by a polynomial of degree exponential in the circuit's $T$-gate depth, yielding a homomorphic scheme for quantum circuits with constant $T$-depth. Both schemes build on a classical fully homomorphic encryption scheme. A further contribution of ours is to formally define the security of encryption schemes for quantum messages: we define quantum indistinguishability under chosen plaintext attacks in both the public and private-key settings. In this context, we show the equivalence of several definitions. Our schemes are the first of their kind that are secure under modern cryptographic definitions, and can be seen as a quantum analogue of classical results establishing homomorphic encryption for circuits with a limited number of multiplication gates. Historically, such results appeared as precursors to the breakthrough result establishing classical fully homomorphic encryption. ",Quantum homomorphic encryption for circuits of low $T$-gate complexity " In this paper, the power flow solution of the two bus network is used to analytically characterise maximum power transfer limits of distribution networks, when subject to both thermal and voltage constraints. Traditional analytic methods are shown to reach contradictory conclusions on the suitability of reactive power for increasing power transfer. Therefore, a more rigorous analysis is undertaken, yielding two solutions, both fully characterised by losses. The first is the well-known thermal limit. The second we define as the `marginal loss-induced maximum power transfer limit'. This is a point at which the marginal increases in losses are greater than increases in generated power. The solution is parametrised in terms of the ratio of resistive to reactive impedance, and yields the reactive power required. The accuracy and existence of these solutions are investigated using the IEEE 34 bus distribution test feeder, and show good agreement with the two bus approximation. The work has implications for the analysis of reactive power interventions in distribution networks, and for the optimal sizing of distributed generation. ",Loss Induced Maximum Power Transfer in Distribution Networks " We show that the Lipschitz-free space with the Radon--Nikod\'{y}m property and a Daugavet point recently constructed by Veeorg is in fact a dual space isomorphic to $\ell_1$. Furthermore, we answer an open problem from the literature by showing that there exists a superreflexive space, in the form of a renorming of $\ell_2$, with a $\Delta$-point. Building on these two results, we are able to renorm every infinite-dimensional Banach space with a $\Delta$-point. Next, we establish powerful relations between existence of $\Delta$-points in Banach spaces and their duals. As an application, we obtain sharp results about the influence of $\Delta$-points for the asymptotic geometry of Banach spaces. In addition, we prove that if $X$ is a Banach space with a shrinking $k$-unconditional basis with $k < 2$, or if $X$ is a Hahn--Banach smooth space with a dual satisfying the Kadets--Klee property, then $X$ and its dual $X^*$ fail to contain $\Delta$-points. In particular, we get that no Lipschitz-free space with a Hahn--Banach smooth predual contains $\Delta$-points. Finally we present a purely metric characterization of the molecules in Lipschitz-free spaces that are $\Delta$-points, and we solve an open problem about representation of finitely supported $\Delta$-points in Lipschitz-free spaces. ",Delta-points and their implications for the geometry of Banach spaces " We establish a spectral multiplier theorem associated with a Schr\""odinger operator H=-\Delta+V(x) in \mathbb{R}^3. We present a new approach employing the Born series expansion for the resolvent. This approach provides an explicit integral representation for the difference between a spectral multiplier and a Fourier multiplier, and it allows us to treat a large class of Schr\""odinger operators without Gaussian heat kernel estimates. As an application to nonlinear PDEs, we show the local-in-time well-posedness of a 3d quintic nonlinear Schr\""odinger equation with a potential. ","A Spectral Multiplier Theorem associated with a Schr\""odinger Operator" " A logic is defined that allows to express information about statistical probabilities and about degrees of belief in specific propositions. By interpreting the two types of probabilities in one common probability space, the semantics given are well suited to model the influence of statistical information on the formation of subjective beliefs. Cross entropy minimization is a key element in these semantics, the use of which is justified by showing that the resulting logic exhibits some very reasonable properties. ",A Logic for Default Reasoning About Probabilities " We report a search for muon neutrino disappearance in the $\Delta m^{2}$ region of 0.5-40 $eV^2$ using data from both SciBooNE and MiniBooNE experiments. SciBooNE data provides a constraint on the neutrino flux, so that the sensitivity to $\nu_{\mu}$ disappearance with both detectors is better than with just MiniBooNE alone. The preliminary sensitivity for a joint $\nu_\mu$ disappearance search is presented. ",Search for Muon Neutrino Disappearance in a Short-Baseline Accelerator Neutrino Beam " Individual laser cooled atoms are delivered on demand from a single atom magneto-optic trap to a high-finesse optical cavity using an atom conveyor. Strong coupling of the atom with the cavity field allows simultaneous cooling and detection of individual atoms for time scales exceeding 15 s. The single atom scatter rate is studied as a function of probe-cavity detuning and probe Rabi frequency, and the experimental results are in good agreement with theoretical predictions. We demonstrate the ability to manipulate the position of a single atom relative to the cavity mode with excellent control and reproducibility. ",Deterministic loading of individual atoms to a high-finesse optical cavity " We study nonintersecting Brownian motions with two prescribed starting and ending positions, in the neighborhood of a tacnode in the time-space plane. Several expressions have been obtained in the literature for the critical correlation kernel $K\tac(x,y)$ that describes the microscopic behavior of the Brownian motions near the tacnode. One approach, due to Kuijlaars, Zhang and the author, expresses the kernel (in the single time case) in terms of a $4\times 4$ matrix valued Riemann-Hilbert problem. Another approach, due to Adler, Ferrari, Johansson, van Moerbeke and Vet\H o in a series of papers, expresses the kernel in terms of resolvents and Fredholm determinants of the Airy integral operator acting on a semi-infinite interval $[\sigma,\infty)$, involving some objects introduced by Tracy and Widom. In this paper we prove the equivalence of both approaches. We also obtain a rank-2 property for the derivative of the tacnode kernel. Finally, we find a Riemann-Hilbert expression for the multi-time extended tacnode kernel. ",The tacnode kernel: equality of Riemann-Hilbert and Airy resolvent formulas " Localization transitions as a function of temperature require a many-body mobility edge in energy, separating localized from ergodic states. We argue that this scenario is inconsistent because local fluctuations into the ergodic phase within the supposedly localized phase can serve as mobile bubbles that induce global delocalization. Such fluctuations inevitably appear with a low but finite density anywhere in any typical state. We conclude that the only possibility for many-body localization to occur are lattice models that are localized at all energies. Building on a close analogy with a model of assisted two-particle hopping, where interactions induce delocalization, we argue why hot bubbles are mobile and do not localize upon diluting their energy. Numerical tests of our scenario show that previously reported mobility edges cannot be distinguished from finite-size effects. ",Absence of many-body mobility edges " This paper presents a data-driven receding horizon fault estimation method for additive actuator and sensor faults in unknown linear time-invariant systems, with enhanced robustness to stochastic identification errors. State-of-the-art methods construct fault estimators with identified state-space models or Markov parameters, but they do not compensate for identification errors. Motivated by this limitation, we first propose a receding horizon fault estimator parameterized by predictor Markov parameters. This estimator provides (asymptotically) unbiased fault estimates as long as the subsystem from faults to outputs has no unstable transmission zeros. When the identified Markov parameters are used to construct the above fault estimator, zero-mean stochastic identification errors appear as model uncertainty multiplied with unknown fault signals and online system inputs/outputs (I/O). Based on this fault estimation error analysis, we formulate a mixed-norm problem for the offline robust design that regards online I/O data as unknown. An alternative online mixed-norm problem is also proposed that can further reduce estimation errors when the online I/O data have large amplitudes, at the cost of increased computational burden. Based on a geometrical interpretation of the two proposed mixed-norm problems, systematic methods to tune the user-defined parameters therein are given to achieve desired performance trade-offs. Simulation examples illustrate the benefits of our proposed methods compared to recent literature. ",Data-Driven Robust Receding Horizon Fault Estimation " Axisymmetric three-dimensional solitary waves in uniform two-component mixture Bose-Einstein condensates are obtained as solutions of the coupled Gross-Pitaevskii equations with equal intracomponent but varying intercomponent interaction strengths. Several families of solitary wave complexes are found: (1) vortex rings of various radii in each of the components, (2) a vortex ring in one component coupled to a rarefaction solitary wave of the other component, (3) two coupled rarefaction waves, (4) either a vortex ring or a rarefaction pulse coupled to a localised disturbance of a very low momentum. The continuous families of such waves are shown in the momentum-energy plane for various values of the interaction strengths and the relative differences between the chemical potentials of two components. Solitary wave formation, their stability and solitary wave complexes in two-dimensions are discussed. ",Solitary wave complexes in two-component mixture condensates " Spherically symmetric simulations of stellar core collapse and post-bounce evolution are used to test the sensitivity of the supernova dynamics to different variations of the input physics. We consider a state-of-the-art description of the neutrino-nucleon interactions, possible lepton-number changing neutrino reactions in the neutron star, and the potential impact of hydrodynamic mixing behind the supernova shock. ",Core-collapse supernova simulations: Variations of the input physics " We study the magnetization dynamics in a ferromagnet-insulator-superconductor tunnel junction and the associated buildup of the electrical polarization. We show that for an open circuit, the induced voltage varies strongly and nonmonotonically with the precessional frequency, and can be enhanced significantly by the superconducting correlations. For frequencies much smaller or much larger than the superconducting gap, the voltage drops to zero, while when these two energy scales are comparable, the voltage is peaked at a value determined by the driving frequency. We comment on the potential utilization of the effect for the low-temperature spatially-resolved spectroscopy of magnetic dynamics. ",Dynamic Magnetoelectric Effect in Ferromagnet-Superconductor Tunnel Junctions " We prove that for renormalizable Yang-Mills gauge theory with arbitrary compact gauge group (of at most a single abelian factor) and matter coupling, the absence of gauge anomalies can be established at the one-loop level. This proceeds by relating the gauge anomaly to perturbative agreement, which formalizes background independence. ",Background independence and the Adler-Bardeen theorem " Thus, the results of our studies lie in developing and implementing the basic principles of digital sorting the Laguerre-Gauss modes by radial numbers both for a non-degenerate and a degenerate state of a vortex beam subject to perturbations in the form of a hard-edged aperture of variable radius. The digital sorting of LG beams by the orthogonal basis involves the use of higher-order intensity moments, and subsequent scanning of the modulated beam images at the focal plane of a spherical lens. As a result, we obtain a system of linear equations for the squared mode amplitudes and the cross amplitudes of the perturbed beam. The solution of the equations allows one to determine the amplitudes of each LG mode and restore both the real mode array and the combined beam as a whole. First, we developed a digital sorting algorithm, and then two types of vortex beams were experimentally studied on its basis: a single LG beam and a composition of single LG beams with the same topological charges(azimuthal numbers) and different radial numbers . The beam was perturbed by means of a circular hard-edged aperture with different radii R. As a result of the perturbation, a set of secondary LG modes with different radial numbers k is appeared that is characterized by an amplitude spectrum . The spectrum obtained makes it possible to restore both the real array of LG modes and the perturbed beam itself with a degree of correlation not lower than. As a measure of uncertainty induced by the perturbation we measured the informational entropy (Shannon's entropy) ",Digital sorting perturbed laguerre-gaussian beams by radial numbers via high order intensity moments " We prove local Lipschitz regularity for bounded minimizers of functionals with nonstandard $p,q$-growth with the source term in the Lorentz space $L(N,1)$ under the restriction $q80%) lasting over the last few decades. ",Helioseismology: Observations and Space Missions " Consider a remote estimation problem where a sensor wants to communicate the state of an uncertain source to a remote estimator over a finite time horizon. The uncertain source is modeled as an autoregressive process with bounded noise. Given that the sensor has a limited communication budget, the sensor must decide when to transmit the state to the estimator who has to produce real-time estimates of the source state. In this paper, we consider the problem of finding a scheduling strategy for the sensor and an estimation strategy for the estimator to jointly minimize the worst-case maximum instantaneous estimation error over the time horizon. This leads to a decentralized minimax decision-making problem. We obtain a complete characterization of optimal strategies for this decentralized minimax problem. In particular, we show that an open loop communication scheduling strategy is optimal and the optimal estimate depends only on the most recently received sensor observation. ",Worst-case Guarantees for Remote Estimation of an Uncertain Source " In this work we explore many directions in the framework of gauge-gravity dualities. In type IIB theory we give an explicit derivation of the local metric for five branes wrapped on rigid two-cycles. Our derivation involves various interplays between warp factors, dualities and fluxes and the final result confirms our earlier predictions. We also find a novel dipole-like deformation of the background due to an inherent orientifold projection in the full global geometry. The supergravity solution for this deformation takes into account various things like the presence of a non-trivial background topology and fluxes as well as branes. Considering these, we manage to calculate the precise local solution using equations of motion. We also show that this dipole-like deformation has the desired property of decoupling the Kaluza-Klein modes from the IR gauge theory. Finally, for the heterotic theory we find new non-Kahler complex manifolds that partake in the full gauge-gravity dualities and study the mathematical structures of these manifolds including the torsion classes, Betti numbers and other topological data. ","Gauge-Gravity Dualities, Dipoles and New Non-Kahler Manifolds" " We present a detailed characterization of the recently discovered i-$R$-Cd ($R$ = Y, Gd-Tm) binary quasicrystals by means of x-ray diffraction, temperature-dependent dc and ac magnetization, temperature-dependent resistance and temperature-dependent specific heat measurements. Structurally, the broadening of x-ray diffraction peaks found for i-$R$-Cd is dominated by frozen-in phason strain, which is essentially independent of $R$. i-Y-Cd is weakly diamagnetic and manifests a temperature-independent susceptibility. i-Gd-Cd can be characterized as a spin-glass below 4.6 K via dc magnetization cusp, a third order non-linear magnetic susceptibility peak, a frequency-dependent freezing temperature and a broad maximum in the specific heat. i-$R$-Cd ($R$ = Ho-Tm) is similar to i-Gd-Cd in terms of features observed in thermodynamic measurements. i-Tb-Cd and i-Dy-Cd do not show a clear cusp in their zero-field-cooled dc magnetization data, but instead show a more rounded, broad local maximum. The resistivity for i-$R$-Cd is of order 300 $\mu \Omega$ cm and weakly temperature-dependent. The characteristic freezing temperatures for i-$R$-Cd ($R$ = Gd-Tm) deviate from the de Gennes scaling, in a manner consistent with crystal electric field splitting induced local moment anisotropy. ","Magnetic and transport properties of i-$R$-Cd icosahedral quasicrystals ($R$ = Y, Gd-Tm)" " We show that the equivariant small quantum $K$-group of a partial flag manifold is a quotient of that of the full flag manifold as a based ring. This yields a variant of the $K$-theoretic analogue of the parabolic version of Peterson's theorem [Lam-Shimozono, Acta Math. {\bf 204} (2010)] that exhibits different shape from the case of quantum cohomology. This note can be seen as an addendum to [K, arXiv:1805.01718 and arXiv:1810.07106]. ",On quantum $K$-groups of partial flag manifolds " We consider a class of domain-wall black hole solutions in the dilaton gravity with a Liouville-type dilaton potential. Using the surface counterterm approach we calculate the stress-energy tensor of quantum field theory (QFT) corresponding to the domain-wall black hole in the domain-wall/QFT correspondence. A brane universe is investigated in the domain-wall black hole background. When the tension term of the brane is equal to the surface counterterm, we find that the equation of motion of the brane can be mapped to the standard form of FRW equations, but with a varying gravitational constant on the brane. A Cardy-Verlinde-like formula is found, which relates the entropy density of the QFT to its energy density. At the moment when the brane crosses the black hole horizon of the background, the Cardy-Verlinde-like formula coincides with the Friedmann equation of the brane universe, and the Hubble entropy bound is saturated by the entropy of domain-wall black holes. ",Holography and Brane Cosmology in Domain Wall Backgrounds " The goal of this paper is to establish singular Adams type inequality for biharmonic operator on Heisenberg group. As an application, we establish the existence of a solution to \begin{equation*} \Delta_{\mathbb{H}^n}^2 u=\frac{f(\xi,u)}{\rho(\xi)^a}\,\,\text{ in }\Omega,\,\, u|_{\partial\Omega}=0=\left.\frac{\partial u}{\partial \nu}\right|_{\partial\Omega}, \end{equation*} where $0\in \Omega \subseteq \mathbb{H}^4$ is a smooth bounded domain, $0\leq a1000) temperature and polarization anisotropies in the CMB are a Rosetta stone for understanding the evolution of the universe. Through detailed measurements one may address everything from the physics of the birth of the universe to the history of star formation and the process by which galaxies formed. One may in addition track the evolution of the dark energy and discover the net neutrino mass. We are at the dawn of a new era in which hundreds of square degrees of sky can be mapped with arcminute resolution and sensitivities measured in microKelvin. Acquiring these data requires the use of special purpose telescopes such as the Atacama Cosmology Telescope (ACT), located in Chile, and the South Pole Telescope (SPT). These new telescopes are outfitted with a new generation of custom mm-wave kilo-pixel arrays. Additional instruments are in the planning stages. ",Observing the Evolution of the Universe " Space and time resolved concentrations of Ar ($^{3}P_2$) metastable atoms at the exit of an atmospheric pressure radio-frequency micro-plasma jet were measured using tunable diode laser absorption spectroscopy. The discharge features a coaxial geometry with a hollow capillary as an inner electrode and a ceramic tube with metal ring as outer electrode. Absorption profiles of metastable atoms as well as optical emission measurements reveal the dynamics and the filamentary structure of the discharge. The average spatial distribution of Ar metastables is characterized with and without a target in front of the jet, showing that the target potential and therewith the electric field distribution substantially changes the filaments' expansion. Together with the detailed analysis of the ignition phase and the discharge's behavior under pulsed operation, the results give an insight into the excitation and de-excitation mechanisms. ",Argon metastable dynamics in a filamentary jet micro-discharge at atmospheric pressure " We report the discovery of a large-amplitude oscillation in the hot subdwarf B star CS 1246 and present multi-colour photometry and time-resolved spectroscopy supporting this discovery. We used the 0.41-m PROMPT array to acquire data in the u', g', r', and i' filters simultaneously over 3 consecutive nights in 2009 April. These data reveal a single oscillation mode with a period of 371.707 +/- 0.002 s and an amplitude dependent upon wavelength, with a value of 34.5 +/- 1.6 mma in the u' filter. We detected no additional frequencies in any of the light curves. Subsequently, we used the 4.1-m SOAR telescope to acquire a time-series of 248 low-resolution spectra spanning 6 hrs to look for line profile variations. Models fits to the spectra give mean atmospheric values of Teff = 28450 +/- 700 K and log g = 5.46 +/- 0.11 undergoing variations with semi-amplitudes of 507 +/- 55 K and 0.034 +/- 0.009 dex, respectively. We also detect a radial velocity oscillation with an amplitude of 8.8 +/- 1.1 km/s. The relationship between the angular and physical radii variations shows the oscillation is consistent with a radial mode. Under the assumption of a radial pulsation, we compute the stellar distance, radius, and mass as d = 460 +/- 140 pc, R = 0.19 +/- 0.08 Rsun, and M = 0.39 +/- 0.22 Msun, respectively, using the Baade-Wesselink method. ",Photometry and spectroscopy of the new sdBV CS 1246 " We present results of interface-resolved simulations of heat transfer in suspensions of finite-size neutrally-buoyant spherical particles for solid volume fractions up to 35% and bulk Reynolds numbers from 500 to 5600. An Immersed Boundary-Volume of Fluid method is used to solve the energy equation in the fluid and solid phase. We relate the heat transfer to the regimes of particle motion previously identified, i.e. a viscous regime at low volume fractions and low Reynolds number, particle-laden turbulence at high Reynolds and moderate volume fraction and particulate regime at high volume fractions. We show that in the viscous dominated regime, the heat transfer is mainly due to thermal diffusion with enhancement due to the particle-induced fluctuations. In the turbulent-like regime, we observe the largest enhancement of the global heat transfer, dominated by the turbulent heat flux. In the particulate shear-thickening regime, however, the heat transfer enhancement decreases as mixing is quenched by the particle migration towards the channel core. As a result, a compact loosely-packed core region forms and the contribution of thermal diffusion to the total heat transfer becomes significant once again. The global heat transfer becomes, in these flows at volume fractions larger than 25%, lower than in single-phase turbulence. ",Regimes of heat transfer in particle suspensions " Multi-stage forceful manipulation tasks, such as twisting a nut on a bolt, require reasoning over interlocking constraints over discrete as well as continuous choices. The robot must choose a sequence of discrete actions, or strategy, such as whether to pick up an object, and the continuous parameters of each of those actions, such as how to grasp the object. In forceful manipulation tasks, the force requirements substantially impact the choices of both strategy and parameters. To enable planning and executing forceful manipulation, we augment an existing task and motion planner with controllers that exert wrenches and constraints that explicitly consider torque and frictional limits. In two domains, opening a childproof bottle and twisting a nut, we demonstrate how the system considers a combinatorial number of strategies and how choosing actions that are robust to parameter variations impacts the choice of strategy. ",Planning for Multi-stage Forceful Manipulation " Exceptional points are singularities of eigenvalues and eigenvectors for complex values of, say, an interaction parameter. They occur universally and are square root branch point singularities of the eigenvalues in the vicinity of level repulsions. The intricate connection between the distribution of exceptional points and particular fluctuation properties of level spacing is discussed. The distribution of the exceptional points of the problem $H_0+\lambda H_1$ is given for the situation of hard chaos. Theoretical predictions of local properties of exceptional points have recently been confirmed experimentally. This relates to the specific topological structure of an exceptional point as well as to the chiral properties of the wave functions associated with exceptional points. ",Global and Local Aspects of Exceptional Point " A broad family of phase transitions in the closed as well as open quantum systems is known to be mediated by a non-Hermitian degeneracy (a.k.a. exceptional point, EP) of the Hamiltonian. In the EP limit, in general, the merger of an $N-$plet of the energy eigenvalues is accompanied by a parallel (though not necessarily complete) degeneracy of eigenstates (forming an EP-asociated $K-$plet; in mathematics, $K$ is called the geometric multiplicity of the EP). In the literature, unfortunately, only the benchmark matrix models with $K=1$ can be found. In our paper the gap is filled: the EP-mediated quantum phase transitions with $K>1$ are called ""clustered"", and a family of benchmark models admitting such a clustering phenomenon is proposed and described. For the sake of maximal simplicity our attention is restricted to the real perturbed-harmonic-oscillator-type N by N matrix Hamiltonians which are exactly solvable and in which the perturbation is multiparametric (i.e., maximally variable) and antisymmetric (i.e., maximally non-Hermitian). A labeling (i.e., an exhaustive classification) of these models is provided by a specific partitioning of N. ",Quantum phase transitions mediated by clustered non-Hermitian degeneracies " Hierarchical triple stars are ideal laboratories for studying the interplay between orbital dynamics and stellar evolution. Both stellar wind mass loss and three-body dynamics cooperate to destabilise triples, which can lead to a variety of astrophysical exotica. So far our understanding of their evolution was mainly built upon results from extensive binary-single scattering experiments. Starting from generic initial conditions, we evolve an extensive set of hierarchical triples using a combination of the triple evolution code TRES and an N-body code. We find that the majority of triples preserve their hierarchy throughout their evolution, which is in contradiction with the commonly adopted picture that unstable triples always experience a chaotic, democratic resonant interaction. The duration of the unstable phase is much longer than expected, so that stellar evolution cannot be neglected. Typically an unstable triple dissolve into a single star and a binary; sometimes democratically (the initial hierarchy is lost and the lightest body usually escapes), but also in a hierarchical way (the tertiary is ejected in a slingshot, independent of its mass). Collisions are common, and mostly involve the two original inner binary components still on the main-sequence. This contradicts the idea that collisions with a giant during democratic encounters dominate. Together with collisions in stable triples, we find that triple evolution is the dominant mechanism for stellar collisions in the Milky Way. Furthermore, our simulations produce runaway and walk-away stars with speeds up to several tens km/s, with a maximum of a few 100km/s. We suggest that destabilised triples can alleviate the tension behind the origin of the observed run-away stars. Lastly, we present a promising indicator to make general predictions for the fate of a specific triple, based on the initial inclination of the system. ",Stellar triples on the edge; Comprehensive overview of the evolution of destabilised triples leading to stellar and binary exotica " We study the full counting statistics of heterostructures consisting of normal metal parts connected to a superconducting terminal. Assuming that coherent superconducting correlations are suppressed in the normal metals we show, using Keldysh-Nambu Green's functions, that the system can be mapped onto a purely normal system with twice the number of elements. For a superconducting beam splitter with several normal terminals we obtain general results for the counting statistics. ",Full counting statistics of incoherent Andreev transport " In this paper, we propose and develop an optimal nonconforming finite element method for the Stokes equations approximated by the Crouzix-Raviart element for velocity and the continuous linear element for pressure. Previous result in using the stabilization method for this finite element pair is improved and then proven to be stable. Then, optimal order error estimate is obtained and numerical results show the accuracy and robustness of the method. ",An optimal nonconforming finite element method for the Stokes equations " This paper develops a new class of exponential-type integrators where all the matrix exponentiations are performed in a single Krylov space of low dimension. The new family, called Lightly Implicit Krylov-Exponential (LIKE), is well suited for solving large scale systems of ODEs or semi-discrete PDEs. The time discretization and the Krylov space approximation are treated as a single computational process, and the Krylov space properties are an integral part of the new LIKE order condition theory developed herein. Consequently, LIKE methods require a small number of basis vectors determined solely by the temporal order of accuracy. The subspace size is independent of the ODE under consideration, and there is no need to monitor the errors in linear system solutions at each stage. Numerical results illustrate the favorable properties of new family of methods. ",Exponential-Krylov methods for ordinary differential equations " High resolution molecular line observations of CS, HCO+, C18O and N2H+ were obtained toward the starless globule FeSt 1-457 in order to investigate its kinematics and chemistry. The HCO+ and CS spectra show clear self-reversed and asymmetric profiles across the face of the globule. The sense of the observed asymmetry is indicative of the global presence of expansion motions in the outer layers of the globule. These motions appear to be subsonic and significantly below the escape velocity of the globule. Comparison of our observations with near-infrared extinction data indicate that the globule is gravitationally bound. Taken together these considerations lead us to suggest that the observed expansion has its origin in an oscillatory motion of the outer layers of the globule which itself is likely in a quasi-stable state near hydrostatic equilibrium. Analysis of the observed linewidths of CO and N2H+ confirm that thermal pressure is the dominant component of the cloud's internal support. A simple calculation suggests that the dominant mode of pulsation would be an l = 2 mode with a period of 0.3 Myr. Deformation of the globule due to the large amplitude l = 2 oscillation may be responsible for the double-peaked structure of the core detected in high resolution extinction maps. Detailed comparison of the molecular-line observations and extinction data provides evidence for significant depletion of C18O and perhaps HCO+ while N2H+ may be undepleted to a cloud depth of about 40 magnitudes of visual extinction. ",The Dynamical State of the Starless Dense Core FeSt 1-457: A Pulsating Globule? " Block modeling is widely used in studies on complex networks. The cornerstone model is the stochastic block model (SBM), widely used over the past decades. However, the SBM is limited in analyzing complex networks as the model is, in essence, a random graph model that cannot reproduce the basic properties of many complex networks, such as sparsity and heavy-tailed degree distribution. In this paper, we provide an edge exchangeable block model that incorporates such basic features and simultaneously infers the latent block structure of a given complex network. Our model is a Bayesian nonparametric model that flexibly estimates the number of blocks and takes into account the possibility of unseen nodes. Using one synthetic dataset and one real-world stock ownership dataset, we show that our model outperforms state-of-the-art SBMs for held-out link prediction tasks. ",Nondiagonal Mixture of Dirichlet Network Distributions for Analyzing a Stock Ownership Network " Recent work has demonstrated substantial gains on many NLP tasks and benchmarks by pre-training on a large corpus of text followed by fine-tuning on a specific task. While typically task-agnostic in architecture, this method still requires task-specific fine-tuning datasets of thousands or tens of thousands of examples. By contrast, humans can generally perform a new language task from only a few examples or from simple instructions - something which current NLP systems still largely struggle to do. Here we show that scaling up language models greatly improves task-agnostic, few-shot performance, sometimes even reaching competitiveness with prior state-of-the-art fine-tuning approaches. Specifically, we train GPT-3, an autoregressive language model with 175 billion parameters, 10x more than any previous non-sparse language model, and test its performance in the few-shot setting. For all tasks, GPT-3 is applied without any gradient updates or fine-tuning, with tasks and few-shot demonstrations specified purely via text interaction with the model. GPT-3 achieves strong performance on many NLP datasets, including translation, question-answering, and cloze tasks, as well as several tasks that require on-the-fly reasoning or domain adaptation, such as unscrambling words, using a novel word in a sentence, or performing 3-digit arithmetic. At the same time, we also identify some datasets where GPT-3's few-shot learning still struggles, as well as some datasets where GPT-3 faces methodological issues related to training on large web corpora. Finally, we find that GPT-3 can generate samples of news articles which human evaluators have difficulty distinguishing from articles written by humans. We discuss broader societal impacts of this finding and of GPT-3 in general. ",Language Models are Few-Shot Learners " Expository paper discussing AG or Goppa codes arising from curves, first from an abstract general perspective then turning to concrete examples associated to modular curves. We will try to explain these extremely technical ideas using a special case at a level to a typical graduate student with some background in modular forms, number theory, group theory, and algebraic geometry. Many examples using MAGMA are included. ",Remarks on codes from modular curves: MAGMA applications " A mimetic spectral element discretization, utilizing a novel Galerkin projection Hodge star operator, of the macroscopic Maxwell equations in Hamiltonian form is presented. The idea of splitting purely topological and metric dependent quantities is natural in the Hamiltonian modeling framework as the Poisson bracket is metric free with the Hamiltonian containing all metric information. This idea may be incorporated into the mimetic spectral element method by directly discretizing the Poincar\'e duality structure. This ""split exterior calculus mimetic spectral element method"" yields spatially discretized Maxwell's equations which are Hamiltonian and exactly and strongly conserve Gauss's laws. Moreover, the new discrete Hodge star operator is itself of interest as a partition of the purely topological and metric dependent portions of the Hodge star operator. As a simple test case, the numerical results of applying this method to a one-dimensional version of Maxwell's equations are given. ",A mimetic discretization of the macroscopic Maxwell equations in Hamiltonian form " Topological edge states are the core of topological photonics. Here we introduce the antihelical edge states of time-reversal symmetric topological metals and propose a photonic realization in an anisotropic square lattice of coupled ring resonators, where the clockwise and counterclockwise modes play the role of pseudospins. The antihelical edge states robustly propagate across the corners toward the diagonal of the square lattice: The same (opposite) pseudospins copropagate in the same (opposite) direction on the parallel lattice boundaries; the different pseudospins separate and converge at the opposite corners. The antihelical edge states in the topological metallic phase alter to the helical edge states in the topological insulating phase under a metal-insulator phase transition. The antihelical edge states provide a unique manner of topologically-protected robust light transport applicable for topological purification. Our findings create new opportunities for topological photonics and metamaterials. ",Antihelical Edge States in Two-dimensional Photonic Topological Metals " The concept of the $M$-decomposition was introduced by Cockburn et al.\ in Math. Comp.\ vol.\ 86 (2017), pp.\ 1609-1641 {to provide criteria to guarantee optimal convergence rates for the Hybridizable Discontinuous Galerkin (HDG) method for coercive elliptic problems}. In that paper they systematically constructed superconvergent hybridizable discontinuous Galerkin (HDG) methods to approximate the solutions of elliptic PDEs on unstructured meshes. In this paper, we use the $M$-decomposition to construct HDG methods for the Maxwell's equations on unstructured meshes in two dimension. In particular, we show the any choice of spaces having an $M$-decomposition, together with sufficiently rich auxiliary spaces, has an optimal error estimate and superconvergence even though the problem is not in general coercive. Unlike the elliptic case, we obtain a superconvergent rate for the curl of the solution, not the solution, and this is confirmed by our numerical experiments. ",Superconvergent HDG methods for Maxwell's equations via the $M$-decomposition " Several approximations are made to study the microcanonical formalism that are valid in the thermodynamics limit. Usually it is assumed that: 1)Stirling approximation can be used to evaluate the number of microstates; 2) the surface entropy can be replace by the volumen entropy; and 3)derivatives can be used even if the energy is not a continuous variable. It is also assumed that the results obtained from the microcanonical formalism agree with those from the canonical one. However, it is not clear if these assumptions are right for very small systems (10-100 particles). To answer this questions, two systems with exact solutions (the Einstein model of a solid and the two-level system)have been solve with and without these approximations. ",Thermostatistics of small systems: Exact results in the microcanonical formalism " We prove the vanishing of certain low degree cohomologies of some induced representations. As an application, we determine certain low degree cohomologies of congruence groups. ",Low degree cohomologies of congruence groups " The present paper studies non-uniform plastic deformations of crystals undergoing anti-plane constrained shear. The asymptotically exact energy density of crystals containing a moderately large density of excess dislocations is found by the averaging procedure. This energy density is extrapolated to the cases of extremely small or large dislocation densities. By incorporating the configurational temperature and the density of redundant dislocations, we develop the thermodynamic dislocation theory for non-uniform plastic deformations and use it to predict the stress-strain curves and the dislocation densities. ",Non-uniform plastic deformations of crystals undergoing anti-plane constrained shear " A quantity which measures total intrinsic spin along the z axis is constructed for planar gravity (fields dependent on z and t only), in both the Ashtekar complex connection formalism and in geometrodynamics. The total spin is conserved but (surprisingly) is not a surface term. This constant of the motion coincides with one of four observables previously discovered by Husain and Smolin. Two more of those observables can be interpreted physically as raising and lowering operators for total spin. ",Total intrinsic spin for plane gravity waves " As Deep Learning continues to yield successful applications in Computer Vision, the ability to quantify all forms of uncertainty is a paramount requirement for its safe and reliable deployment in the real-world. In this work, we leverage the formulation of variational inference in function space, where we associate Gaussian Processes (GPs) to both Bayesian CNN priors and variational family. Since GPs are fully determined by their mean and covariance functions, we are able to obtain predictive uncertainty estimates at the cost of a single forward pass through any chosen CNN architecture and for any supervised learning task. By leveraging the structure of the induced covariance matrices, we propose numerically efficient algorithms which enable fast training in the context of high-dimensional tasks such as depth estimation and semantic segmentation. Additionally, we provide sufficient conditions for constructing regression loss functions whose probabilistic counterparts are compatible with aleatoric uncertainty quantification. ",Scalable Uncertainty for Computer Vision with Functional Variational Inference " While traditional HPC has and continues to satisfy most workflows, a new generation of researchers has emerged looking for sophisticated, scalable, on-demand, and self-service control of compute infrastructure in a cloud-like environment. Many also seek safe harbors to operate on or store sensitive and/or controlled-access data in a high capacity environment. To cater to these modern users, the Minnesota Supercomputing Institute designed and deployed Stratus, a locally-hosted cloud environment powered by the OpenStack platform, and backed by Ceph storage. The subscription-based service complements existing HPC systems by satisfying the following unmet needs of our users: a) on-demand availability of compute resources, b) long-running jobs (i.e., $> 30$ days), c) container-based computing with Docker, and d) adequate security controls to comply with controlled-access data requirements. This document provides an in-depth look at the design of Stratus with respect to security and compliance with the NIH's controlled-access data policy. Emphasis is placed on lessons learned while integrating OpenStack and Ceph features into a so-called ""walled garden"", and how those technologies influenced the security design. Many features of Stratus, including tiered secure storage with the introduction of a controlled-access data ""cache"", fault-tolerant live-migrations, and fully integrated two-factor authentication, depend on recent OpenStack and Ceph features. ",Leveraging OpenStack and Ceph for a Controlled-Access Data Cloud " We present a uniform catalog of accurate distances to local molecular clouds informed by the Gaia DR2 data release. Our methodology builds on that of Schlafly et al. (2014). First, we infer the distance and extinction to stars along sightlines towards the clouds using optical and near-infrared photometry. When available, we incorporate knowledge of the stellar distances obtained from Gaia DR2 parallax measurements. We model these per-star distance-extinction estimates as being caused by a dust screen with a 2-D morphology derived from Planck at an unknown distance, which we then fit for using a nested sampling algorithm. We provide updated distances to the Schlafly et al. (2014) sightlines towards the Dame et al. (2001) and Magnani et al. (1985) clouds, finding good agreement with the earlier work. For a subset of 27 clouds, we construct interactive pixelated distance maps to further study detailed cloud structure, and find several clouds which display clear distance gradients and/or are comprised of multiple components. We use these maps to determine robust average distances to these clouds. The characteristic combined uncertainty on our distances is approximately 5-6%, though this can be higher for clouds at farther distances, due to the limitations of our single-cloud model. ",A Large Catalog of Accurate Distances to Local Molecular Clouds: The Gaia DR2 Edition " The inverse problem method is tested for a class of mean field statistical mechanics models representing a mixture of particles of different species. The robustness of the inversion is investigated for different values of the physical parameters, system sizes and independent samples. We show how to reconstruct the parameter values with a precision of a few percentages. ",Inverse problem robustness for multi-species mean field spin models " Motivated by 5d rank 2 SCFTs, we construct a smooth, non-compact Calabi-Yau 3-fold $X$ containing a rank 2 shrinkable surface $S=S_1\cup S_2$ glued over a smooth curve. This construction will be a generalization of the construction of a local surface for a smooth surface $S$ ",Local Calabi-Yau 3-folds for some rank 2 shrinkable surfaces " We report the use of plasma-enhanced atomic layer deposition (PEALD) to fabricate single-electron transistors (SETs) featuring ultra-thin (~1 nm) tunnel-transparent SiO2 in Ni-SiO2-Ni tunnel junctions. We show that as a result of the O2 plasma steps in PEALD of SiO2, the top surface of the underlying Ni electrode is oxidized. Additionally, the bottom surface of the upper Ni layer is also oxidized where it is in contact with the deposited SiO2, most likely as a result of oxygen-containing species on the surface of the SiO2. Due to the presence of these surface parasitic layers of NiO, which exhibit features typical of thermally activated transport, the resistance of Ni-SiO2-Ni tunnel junctions is drastically increased. Moreover, the transport mechanism is changed from quantum tunneling through the dielectric barrier to one consistent with thermally activated resistors in series with tunnel junctions. The reduction of NiO to Ni is therefore required to restore the metal-insulator-metal (MIM) structure of the junctions. Rapid thermal annealing in a forming gas ambient at elevated temperatures is presented as a technique to reduce both parasitic oxide layers. This method is of great interest for devices that rely on MIM tunnel junctions with ultra-thin barriers. Using this technique, we successfully fabricated MIM SETs with minimal trace of parasitic NiO component. We demonstrate that the properties of the tunnel barrier in nanoscale tunnel junctions can be evaluated by electrical characterization of SETs. ",Experimental Demonstration of Single Electron Transistors Featuring SiO2 PEALD in Ni-SiO2-Ni Tunnel Junctions " In this sequel to [1], we take up a second approach in bending the Bruhat-Tits tree. Inspired by the BTZ black hole connection, we demonstrate that one can transplant it to the Bruhat-Tits tree, at the cost of defining a novel ""exponential function"" on the p-adic numbers that is hinted by the BT tree. We demonstrate that the PGL$(2,Q_p)$ Wilson lines [2] evaluated on this analogue BTZ connection is indeed consistent with correlation functions of a CFT at finite temperatures. We demonstrate that these results match up with the tensor network reconstruction of the p-adic AdS/CFT with a different cutoff surface at the asymptotic boundary, and give explicit coordinate transformations that relate the analogue p-adic BTZ background and the ""pure"" Bruhat-Tits tree background. This is an interesting demonstration that despite the purported lack of descendents in p-adic CFTs, there exists non-trivial local Weyl transformations in the CFT corresponding to diffeomorphism in the Bruhat-Tits tree. ",Bending the Bruhat-Tits Tree II: the p-adic BTZ Black hole and Local Diffeomorphism on the Bruhat-Tits Tree " High-Entropy Materials are composed of multiple elements on comparatively simpler lattices. Due to the multicomponent nature of such materials, the atomic scale sampling is computationally expensive due to the combinatorial complexity. We propose a genetic algorithm based methodology for sampling such complex chemically-disordered materials. Genetic Algorithm based Atomistic Sampling Protocol (GAASP) variants can generate low and well as high-energy structures. GAASP low-energy variant in conjugation with metropolis criteria avoids the premature convergence as well as ensures the detailed balance condition. GAASP can be employed to generate the low-energy structures for thermodynamic predictions as well as diverse structures can be generated for machine learning applications. ",GAASP: Genetic Algorithm Based Atomistic Sampling Protocol for High-Entropy Materials " NU Ori is a massive spectroscopic and visual binary in the Orion Nebula Cluster, with 4 components: Aa, Ab, B, and C. The B0.5 primary (Aa) is one of the most massive B-type stars reported to host a magnetic field. We report the detection of a spectroscopic contribution from the C component in high-resolution ESPaDOnS spectra, which is also detected in a Very Large Telescope Interferometer (VLTI) dataset. Radial velocity (RV) measurements of the inner binary (designated Aab) yield an orbital period of 14.3027(7) d. The orbit of the third component (designated C) was constrained using both RVs and interferometry. We find C to be on a mildly eccentric 476(1) d orbit. Thanks to spectral disentangling of mean line profiles obtained via least-squares deconvolution we show that the Zeeman Stokes $V$ signature is clearly associated with C, rather than Aa as previously assumed. The physical parameters of the stars were constrained using both orbital and evolutionary models, yielding $M_{\rm Aa} = 14.9 \pm 0.5 M_\odot$, $M_{\rm Ab} = 3.9 \pm 0.7 M_\odot$, and $M_{\rm C} = 7.8 \pm 0.7 M_\odot$. The rotational period obtained from longitudinal magnetic field $\langle B_z \rangle$ measurements is $P_{\rm rot} = 1.09468(7)$ d, consistent with previous results. Modeling of $\langle B_z \rangle$ indicates a surface dipole magnetic field strength of $\sim 8$ kG. NU Ori C has a magnetic field strength, rotational velocity, and luminosity similar to many other stars exhibiting magnetospheric H$\alpha$ emission, and we find marginal evidence of emission at the expected level ($\sim$1% of the continuum). ",NU Ori: a hierarchical triple system with a strongly magnetic B-type star " A moir\'{e} pattern occurs when two periodic structures in a system have a slight mismatch period, resulting the coexistence of distinct phases in different large-scale spacial regions of the same system. Two periodic structures can arise from periodic electric and magnetic fields, respectively. We investigated the moir\'{e} pattern via a dimerized Kitaev spin chain with a periodic transverse field, which can be mapped onto the system of dimerized spinless fermions with p-wave superconductivity. The exact solution for staggered field demonstrated that the ground state has two distinct phases: (i) Neel magnetic phase for nonzero field, (ii) Spin liquid phase due to the emergence of isolated flat Bogoliubov--de Gennes band for vanishing field. We computed the staggered magnetization and local density of states (\textrm{LDOS}) for the field with a slight difference period to the chain lattice. Numerical simulation demonstrated that such two phases appear alternatively along the chain with a long beat period. Additionally, we proposed a dynamic scheme to detect the Moir\'{e} fringes based on the measurement of Loschmidt echo (\textrm{LE}) in the presence of local perturbation. ",Moir\'{e} pattern of spin liquid and Neel magnet in a Kitaev chain " Managing shared mutable states in high concurrency state access operations is a persistent challenge in Network Functions Virtualization (NFV). This is particularly true when striving to meet chain output equivalence (COE) requirements. This paper presents TransNFV, an innovative NFV framework that incorporates transactional semantics to optimize NFV state management. The TransNFV integrates VNF state access operations as transactions, resolves transaction dependencies, schedules transactions dynamically, and executes transactions efficiently. Initial findings suggest that TransNFV maintains shared VNF state consistency, meets COE requirements, and skillfully handles complex cross-flow states in dynamic network conditions. TransNFV thus provides a promising solution to enhance state management and overall performance in future NFV platforms. ",TransNFV: Integrating Transactional Semantics for Efficient State Management in Virtual Network Functions " The effect of scalar and vector meson mixing on the dilepton radiation from hot and dense hadronic matter is estimated in different isospin channels. In particular, we study the effect of $\sigma$-$\omega$ and $\rho-a_0$ mixing and calculate the corresponding rates. Effects are found to be significant compared to standard $\pi$-$\pi$ and $K$-${\bar K}$ annihilations. While the mixing in the isoscalar channel mostly gives a contribution in the invariant mass range between the two-pion threshold and the $\omega$ peak, the isovector channel mixing induces an additional peak just below that of the $\phi$. Experimentally, the dilepton signals from $\rho$-$a_0$ mixing seem to be more tractable than those from $\sigma$-$\omega$ mixing. ",The effects of meson mixing on dilepton spectra " Given a point A in the real Grassmannian, it is well-known that one can construct a soliton solution u_A(x,y,t) to the KP equation. The contour plot of such a solution provides a tropical approximation to the solution when the variables x, y, and t are considered on a large scale and the time t is fixed. In this paper we use several decompositions of the Grassmannian in order to gain an understanding of the contour plots of the corresponding soliton solutions. First we use the positroid stratification of the real Grassmannian in order to characterize the unbounded line-solitons in the contour plots at y>>0 and y<<0. Next we introduce a refinement of the positroid stratification -- the Deodhar decomposition of the Grassmannian -- which is defined to be the projection of Deodhar's decomposition of the complete flag variety. We index the components of the Deodhar decomposition of the Grassmannian by certain tableaux which we call Go-diagrams, and then use these Go-diagrams to characterize the contour plots of solitons solutions when t<<0. Finally we use these results to show that a soliton solution u_A(x,y,t) is regular for all times t if and only if A comes from the totally non-negative part of the Grassmannian. ",The Deodhar decomposition of the Grassmannian and the regularity of KP solitons " We consider four dimensional quantum field theories which have a continuous manifold of inequivalent exact ground states -- a moduli space of vacua. Classically, the singular points on the moduli space are associated with extra massless particles. Quantum mechanically these singularities can be smoothed out. Alternatively, new massless states appear there. These may be the elementary massless particles or new massless bound states. ",Exact Results on the Space of Vacua of Four Dimensional SUSY Gauge Theories " We constructed 70 SuperNova IDentification (SNID; Blondin & Tonry 2007) supernova (SN) templates using 640 spectra of stripped-envelope core-collapse SNe (SESNe) published by Modjaz et al. (2014). Fifty-six SN templates which are constructed from 458 spectra are brand new, increasing the number of SESNe and the number of SESNe spectra in the current SNID database by a factor of 2.6 and 2.0, respectively. We also made some type and phase refinements to templates in the current SNID database. ",SuperNova IDentification spectral templates of 70 stripped-envelope core-collapse supernovae " We consider the problem of learning a forest of nonlinear decision rules with general loss functions. The standard methods employ boosted decision trees such as Adaboost for exponential loss and Friedman's gradient boosting for general loss. In contrast to these traditional boosting algorithms that treat a tree learner as a black box, the method we propose directly learns decision forests via fully-corrective regularized greedy search using the underlying forest structure. Our method achieves higher accuracy and smaller models than gradient boosting (and Adaboost with exponential loss) on many datasets. ",Learning Nonlinear Functions Using Regularized Greedy Forest " We propose a novel scheme to generate entanglement among quantum-dot-based charge qubits via sequential electron transfer in an auxiliary quantum dot structure whose transport properties are conditioned by qubit states. The transfer protocol requires the utilization of resonant optical pulses combined with an appropriate voltage gate pattern. As an example illustrating the application of this scheme, we examine the nine-qubit Shor code state preparation together with the error syndrome measurement. ",Charge qubit entanglement via conditional single-electron transfer in an array of quantum dots " Due to the fact that fully supervised semantic segmentation methods require sufficient fully-labeled data to work well and can not generalize to unseen classes, few-shot segmentation has attracted lots of research attention. Previous arts extract features from support and query images, which are processed jointly before making predictions on query images. The whole process is based on convolutional neural networks (CNN), leading to the problem that only local information is used. In this paper, we propose a TRansformer-based Few-shot Semantic segmentation method (TRFS). Specifically, our model consists of two modules: Global Enhancement Module (GEM) and Local Enhancement Module (LEM). GEM adopts transformer blocks to exploit global information, while LEM utilizes conventional convolutions to exploit local information, across query and support features. Both GEM and LEM are complementary, helping to learn better feature representations for segmenting query images. Extensive experiments on PASCAL-5i and COCO datasets show that our approach achieves new state-of-the-art performance, demonstrating its effectiveness. ",Boosting Few-shot Semantic Segmentation with Transformers " This paper analyzes the new affiliation feature available in Google-Scholar Citations revealing that the affiliation-tool works well for most-institutions, it is unable to detect all existing institutions in database, and it is not always able to create unique-standardized entry for each-institution. ",The lost academic home: institutional affiliation links in Google Scholar Citations " Defining P* to be the complete lattice of upsets (ordered by reverse inclusion) of a poset P we give necessary and sufficient conditions on a subset S of P* for P to admit a meet-completion e from P to Q where e preserves the infimum of an upwardly closed set from P if and only if it is in S. We show that given S satisfying these conditions the set M of these completions forms a topped weakly lower semimodular lattice. In particular, when P is finite M is a lower semimodular lattice, and a lower bounded homomorphic image of a free lattice. We provide an example where M does not have a bottom element. ",Preserving meets in meet-dense poset completions " CVXPY is a domain-specific language for convex optimization embedded in Python. It allows the user to express convex optimization problems in a natural syntax that follows the math, rather than in the restrictive standard form required by solvers. CVXPY makes it easy to combine convex optimization with high-level features of Python such as parallelism and object-oriented design. CVXPY is available at http://www.cvxpy.org/ under the GPL license, along with documentation and examples. ",CVXPY: A Python-Embedded Modeling Language for Convex Optimization " We prove the super-exponential decay of probabilities that there exist $n$ crossings of a given quadrilateral in a simple $\text{CLE}_\kappa(\Omega)$, $\frac{8}{3}<\kappa\le 4$, as $n$ goes to infinity. Besides being of independent interest, this also provides the missing ingredient in arXiv:1809.00690 for proving the convergence of probabilities of cylindrical events for the double-dimer loop ensembles to those for the nested $\text{CLE}_4(\Omega)$. ",On the crossing estimates for simple conformal loop ensembles " Objective: For real-time markerless tumour tracking in stereotactic lung radiotherapy, we propose a different approach which uses patient-specific deep learning (DL) using a personalized data generation strategy, avoiding the need for collection of a large patient data set. We validated our strategy with digital phantom simulation and epoxy phantom studies. Methods: We developed lung tumour tracking for radiotherapy using a convolutional neural network trained for each phantom's lesion by using multiple digitally reconstructed radiographs (DRRs) generated from each phantom's treatment planning 4D-CT. We trained tumour-bone differentiation using large numbers of training DRRs generated with various projection geometries to simulate tumour motion. We solved the problem of using DRRs for training and X-ray images for tracking by using the training DRRs with random contrast transformation and random noise addition. Results: We defined adequate tracking accuracy as the % frames satisfying < 1 mm tracking error of the isocentre. In the simulation study, we achieved 100% tracking accuracy in 3-cm spherical and 1.5 x 2.25 x 3-cm ovoid masses. In the phantom study, we achieved 100% and 94.7% tracking accuracy in 3- and 2-cm spherical masses, respectively. This required 32.5 ms/frame (30.8 fps) real-time processing. Conclusions: We proved the potential feasibility of a real-time markerless tumour tracking framework for stereotactic lung radiotherapy based on patient-specific DL with personalized data generation with digital phantom and epoxy phantom studies. Advances in Knowledge: Using DL with personalized data generation is an efficient strategy for real-time lung tumour tracking. ",Real-time markerless tumour tracking with patient-specific deep learning using a personalized data generation strategy: Proof of concept by phantom study " We measure resistively the mean-field superconducting-normal phase boundaries of both kagome and honeycomb wire networks immersed in a transverse magnetic field. In addition to their agreement with theory about the overall shapes of phase diagrams, they show striking one-to-one correspondence between the cusps in the honeycomb and kagome phase boundaries. This correspondence is due to their geometric arrangements and agrees with Lin and Nori's recent calculation. We also find that for the frustrated honeycomb network at f=1/2, the current patterns in the superconducting phase differ between the low-temperature London regime and the higher-temperature Ginzburg-Landau regime near T_c. ",Comparison of Phase Boundaries between Kagome and Honeycomb Superconducting Wire Networks " Generalizing the 't Hooft and Veltman method of unitary regulators, we demonstrate for the first time the existence of local, Lorentz-invariant, physically motivated Lagrangians of quantum-electrodynamic phenomena such that: (i) Feynman diagrams are finite and equal the diagrams of QED but with regularized propagators. (ii) N-point Green functions are C-, P-, and T-invariant up to a phase factor, Lorentz-invariant and causal. (iii) No auxiliary particles or parameters are introduced. ",Realistic regularization of the QED Green functions " This paper focuses on the problem of controlling self-interested drivers in ride-sourcing applications. Each driver has the objective of maximizing its profit, while the ride-sourcing company focuses on customer experience by seeking to minimizing the expected wait time for pick-up. These objectives are not usually aligned, and the company has no direct control on the waiting locations of the drivers. In this paper, we provide two indirect control methods to optimize the set of waiting locations of the drivers, thereby minimizing the expected wait time of the customers: 1) sharing the location of all drivers with a subset of drivers, and 2) paying the drivers to relocate. We show that finding the optimal control for each method is NP-hard and we provide algorithms to find near-optimal control in each case. We evaluate the performance of the proposed control methods on real-world data and show that we can achieve between 20% to 80% improvement in the expected response. ",On Re-Balancing Self-Interested Agents in Ride-Sourcing Transportation Networks " A ""continuous measurement"" Quantum Zeno Effect (QZE) in the context of trapped ions is predicted. We describe the physical system and study its exact time evolution showing the appearance of Zeno Phenomena. New indicators for the occurrence of QZE in oscillatory systems are proposed and carefully discussed. ",Quantum Zeno Effect in Trapped Ions " We present integral-field spectroscopy of 27 galaxies in the Coma cluster observed with the Oxford SWIFT spectrograph, exploring the kinematic morphology-density relationship in a cluster environment richer and denser than any in the ATLAS3D survey. Our new data enables comparison of the kinematic morphology relation in three very different clusters (Virgo, Coma and Abell 1689) as well as to the field/group environment. The Coma sample was selected to match the parent luminosity and ellipticity distributions of the early-type population within a radius 15' (0.43 Mpc) of the cluster centre, and is limited to r' = 16 mag (equivalent to M_K = -21.5 mag), sampling one third of that population. From analysis of the lambda-ellipticity diagram, we find 15+-6% of early-type galaxies are slow rotators; this is identical to the fraction found in the field and the average fraction in the Virgo cluster, based on the ATLAS3D data. It is also identical to the average fraction found recently in Abell 1689 by D'Eugenio et al.. Thus it appears that the average slow rotator fraction of early type galaxies remains remarkably constant across many different environments, spanning five orders of magnitude in galaxy number density. However, within each cluster the slow rotators are generally found in regions of higher projected density, possibly as a result of mass segregation by dynamical friction. These results provide firm constraints on the mechanisms that produce early-type galaxies: they must maintain a fixed ratio between the number of fast rotators and slow rotators while also allowing the total early-type fraction to increase in clusters relative to the field. A complete survey of Coma, sampling hundreds rather than tens of galaxies, could probe a more representative volume of Coma and provide significantly stronger constraints, particularly on how the slow rotator fraction varies at larger radii. ",Fast and Slow Rotators in the Densest Environments: a SWIFT IFS study of the Coma Cluster " We develop a new master equation as a unified description of the effects of both quantum noise (system-bath interaction) and classical noise on a system's dynamics, using a two-dimensional series expansion method. When quantum and classical noises are both present, their combined effect on a system's dynamics is not necessarily a simple sum of the two individual effects. Thus previous master equations for open systems and those for classical noise, even when jointly used, may not capture the full physics. Our formalism can determine whether there is interference between quantum and classical noises and will be able to capture and describe such interference if there is any (in a perturbative manner). We find that, interestingly, second-order interference between quantum and classical noises vanishes identically. This work thus also serves to justify simple additive treatments of quantum and classical noises, especially in the weak coupling regime. For a Zeeman-splitted atom in a stochastic magnetic field interacting with an optical cavity, we use the formalism developed herein to find the overall decoherence rate between the atom's energy levels. ",Unifying the dynamical effects of quantum and classical noises " It can be insightful to extend qualitative studies with a secondary quantitative analysis (where the former suggests insightful questions that the latter can answer). Documenting developer beliefs should be the start, not the end, of Software Engineering research. Once prevalent beliefs are found, they should be checked against real-world data. For example, this paper finds several notable discrepancies between empirical evidence and the developer beliefs documented in Wan et al.'s recent TSE paper ""Perceptions, expectations, and challenges in defect prediction"". By reporting these discrepancies we can stop developers (a) wasting time on inconsequential matters or (b) ignoring important effects. For the future, we would encourage more ""extension studies"" of prior qualitative results with quantitative empirical evidence. ","Assessing Developer Beliefs: A Reply to ""Perceptions, Expectations, and Challenges in Defect Prediction""" " By assuming an aspherical stellar wind with an equatorial disk from a red giant, we investigate the production of Type Ia supernovae (SNe Ia) via symbiotic channel. We estimate that the Galactic birthrate of SNe Ia via symbiotic channel is between $1.03\times 10^{-3}$ and $2.27\times 10^{-5}$ yr$^{-1}$, the delay time of SNe Ia has wide range from $\sim$ 0.07 to 5 Gyr. The results are greatly affected by the outflow velocity and mass-loss rate of the equatorial disk. Using our model, we discuss the progenitors of SN 2002ic and SN 2006X. ",An Alternative Symbiotic Channel to Type Ia Supernovae " This thesis assesses the influence of astronomical phenomena on the Earth's biosphere and climate. I examine in particular the relevance of both the path of the Sun through the Galaxy and the evolution of the Earth's orbital parameters in modulating non-terrestrial mechanisms. I build models to predict the extinction rate of species, the temporal variation of the impact cratering rate and ice sheet deglaciations, and then compare these models with other models within a Bayesian framework. I find that the temporal distribution of mass extinction events over the past 550 Myr can be explained just as well by a uniform random distribution as by other models, such as variations in the stellar density local to the Sun arising from the Sun's orbit. Given the uncertainties in the Galaxy model and the Sun's current phase space coordinates, as well as the errors in the geological data, it is not possible to draw a clear connection between terrestrial extinction and the solar motion. In a separate study, I find that the solar motion, which modulates the Galactic tidal forces imposed on Oort cloud comets, does not significantly influence this cratering rate. My dynamical models, together with the solar apex motion, can explain the anisotropic perihelia of long period comets without needing to invoke the existence of a Jupiter-mass solar companion. Finally, I find that variations in the Earth's obliquity play a dominant role in triggering terrestrial deglaciations over the past 2 Myr. The precession of the equinoxes, in contrast, only becomes important in pacing large deglaciations after the transition from the 100-kyr dominant periodicity in the ice coverage to a 41-kyr dominant periodicity, which occurred 0.7 Myr ago. ",Investigations into the impact of astronomical phenomena on the terrestrial biosphere and climate " Visual Attention Networks (VAN) with Large Kernel Attention (LKA) modules have been shown to provide remarkable performance, that surpasses Vision Transformers (ViTs), on a range of vision-based tasks. However, the depth-wise convolutional layer in these LKA modules incurs a quadratic increase in the computational and memory footprints with increasing convolutional kernel size. To mitigate these problems and to enable the use of extremely large convolutional kernels in the attention modules of VAN, we propose a family of Large Separable Kernel Attention modules, termed LSKA. LSKA decomposes the 2D convolutional kernel of the depth-wise convolutional layer into cascaded horizontal and vertical 1-D kernels. In contrast to the standard LKA design, the proposed decomposition enables the direct use of the depth-wise convolutional layer with large kernels in the attention module, without requiring any extra blocks. We demonstrate that the proposed LSKA module in VAN can achieve comparable performance with the standard LKA module and incur lower computational complexity and memory footprints. We also find that the proposed LSKA design biases the VAN more toward the shape of the object than the texture with increasing kernel size. Additionally, we benchmark the robustness of the LKA and LSKA in VAN, ViTs, and the recent ConvNeXt on the five corrupted versions of the ImageNet dataset that are largely unexplored in the previous works. Our extensive experimental results show that the proposed LSKA module in VAN provides a significant reduction in computational complexity and memory footprints with increasing kernel size while outperforming ViTs, ConvNeXt, and providing similar performance compared to the LKA module in VAN on object recognition, object detection, semantic segmentation, and robustness tests. ",Large Separable Kernel Attention: Rethinking the Large Kernel Attention Design in CNN " Exotic spinor fields arise from inequivalent spin structures on non-trivial topological manifolds, $M$. This induces an additional term in the Dirac operator, defined by the cohomology group $H^1(M,\mathbb{Z}_2)$ that rules a Cech cohomology class. This formalism is extended for manifolds of any finite dimension, endowed with a metric of arbitrary signature. The exotic corrections to heat kernel coefficients, relating spectral properties of exotic Dirac operators to the geometric invariants of $M$, are derived and scrutinized. ",Hearing the shape of inequivalent spin structures and exotic Dirac operators " This report summarizes the second International Verification of Neural Networks Competition (VNN-COMP 2021), held as a part of the 4th Workshop on Formal Methods for ML-Enabled Autonomous Systems that was collocated with the 33rd International Conference on Computer-Aided Verification (CAV). Twelve teams participated in this competition. The goal of the competition is to provide an objective comparison of the state-of-the-art methods in neural network verification, in terms of scalability and speed. Along this line, we used standard formats (ONNX for neural networks and VNNLIB for specifications), standard hardware (all tools are run by the organizers on AWS), and tool parameters provided by the tool authors. This report summarizes the rules, benchmarks, participating tools, results, and lessons learned from this competition. ",The Second International Verification of Neural Networks Competition (VNN-COMP 2021): Summary and Results " With recent developments in high-precision quantum measurements, the question of whether observations of decoherence from spacetime fluctuations are accessible experimentally arises. Here we investigate the dynamics of bound states interacting with an environment of gravitons under the Markov approximation. The corresponding Lindblad master equation is presented that enables gravitational decoherence and dissipation due to zero-point spacetime fluctuations to be analyzed. Specifically, we consider a one-dimensional cavity of massless scalar particles that models a light beam with negligible spin polarizations being reflected between two free masses. Numerical simulations have been performed to illustrate the wave-modal dependent decoherence and dissipation of such a configuration. We further demonstrate the existence of nontrivial collective effects akin to superradiance, providing amplifications of gravitational decoherence for a large number of identical bosonic particles. ",Quantum dynamics of bound states under spacetime fluctuations " Deep levels control the space charge in electrically compensated semi-insulating materials. They limit the performance of radiation detectors but their interaction with free carriers can be favorably exploited in these devices to manipulate the spatial distribution of the electric field by optical beams. By using semi-insulating CdTe diodes as a case study, our results show that optical doping functionalities are achieved. As such, a highly stable, flux-dependent, reversible and spatially localized space charge is induced by a line-shaped optical beam focused on the cathode contact area. Real-time non-invasive imaging of the electric field is obtained through the Pockels effect. A simple and convenient method to retrieve the two-dimensional electric field components is presented. Numerical simulations involving just one deep level responsible for the electrical compensation confirm the experimental findings and help to identify the underlying mechanism and critical parameters enabling the optical writing functionalities. ",Optical Writing and Electro-Optic Imaging of Reversible Space Charges in Semi-Insulating CdTe Diodes " We analyzed spatial and spectral characteristics of quasi-periodic pulses (QPP) for the 18 August 2012 limb are, using new data from a complex of spectral and imaging instruments developed by the Siberian Solar Radio Telescope team and the Wind/Konus gamma-ray spectrometer. A sequence of broadband pulses with periods of approximately ten seconds were observed in X-rays at energies between 25 keV and 300 keV, and in microwaves at frequencies from a few GHz up to 34 GHz during an interval of one minute. The QPP X-ray source was located slightly above the limb where the south legs of large and small EUV loop systems were close to each other. Before the QPPs the soft X-ray emission and the Ramaty High Energy Solar Spectroscopic Imager signal from the energy channels below 25 keV were gradually arising for several minutes at the same location. It was found that each X-ray pulse showed the soft-hard-soft behavior. The 17 and 34 GHz microwave source were at footpoints of the small loop system and the source emitting in the 4.2 {7.4 GHz band in the large one. The QPPs were probably generated by modulation of acceleration processes in the energy release site. Analyzing radio spectra we determined the plasma parameters in the radio sources. The microwave pulses could be explained by relatively weak variations of the spectrum hardness of emitting electrons. ",Sources of Quasi-Periodic Pulses in the 18 August 2012 Flare " Starting from a graded Frobenius superalgebra $B$, we consider a graphical calculus of $B$-decorated string diagrams. From this calculus we produce algebras consisting of closed planar diagrams and of closed annular diagrams. The action of annular diagrams on planar diagrams can be used to make clockwise (or counterclockwise) annular diagrams into an inner product space. Our main theorem identifies this space with the space of symmetric functions equipped with the Jack inner product at Jack parameter $\operatorname{dim} B_\mathrm{even} - \operatorname{dim} B_\mathrm{odd}$. In this way, we obtain a graphical realization of that inner product space. ",A graphical calculus for the Jack inner product on symmetric functions " Recent experiments have established a type of nonsymmorphic symmetry protected nodal lines in the family of two-dimensional (2D) composition tunable materials NbSi$_x$Te$_2$. Here, we theoretically study the plasmonic properties of such nonsymmorphic nodal-line semimetals. We show that the nonsymmorphic character endows the plasmons with extremely strong anisotropy. There exist both intraband and interband plasmon branches. The intraband branch is gapless and has a $q^{1/2}$ dispersion. It is most dispersive and is independent of carrier density in direction normal to the nodal line, whereas along the nodal line, its dispersion is largely suppressed and its frequency scales linearly with carrier density. The interband branches are gapped and their long wavelength limits are connected with van Hove singularities of the band structure. We find that the single particle excitations are strongly suppressed in such systems, which decreases the Landau damping of plasmons. These characters are further verified by first-principles calculations on 2D NbSi$_x$Te$_2$. Interesting features in static screening of charged impurity are also discussed. Our result reveals characteristic plasmons in a class of nonsymmorphic topological semimetals and offers guidance for its experimental detection and possible applications. ",Plasmons in a two-dimensional nonsymmorphic nodal-line semimetal " We propose using the effective spin-orbit interaction of light in Bragg-modulated cylindrical waveguides for the effcient separation of spin-up and spin-down photons emitted by a single photon emitter. Due to the spin and directional dependence of photonic stopbands in the waveguides, spin-up (down) photon propagation in the negative (positive) direction along the waveguide axis is blocked while the same photon freely propagates in the opposite direction. ",The Design for a Nanoscale Single-Photon Spin Splitter " [Abridged] We present a robust measurement of the rest-frame UV luminosity function (LF) and its evolution during the peak epoch of cosmic star formation at 1 50% completeness in order to minimize these systematics, we still find that the faint-end slope is steep and getting steeper with redshift, though with slightly shallower (less negative) values ($\alpha=-1.55\pm0.06$, $-1.69\pm0.07$ and $-1.79\pm0.08$ for $z\sim1.3$, 1.9 and 2.6, respectively). Finally, we conclude that the faint star-forming galaxies with UV magnitudes of $-18.5 n$ ``samples''. An amplification procedure is valid if no algorithm can distinguish the set of $m$ samples produced by the amplifier from a set of $m$ independent draws from $D$, with probability greater than $2/3$. Perhaps surprisingly, in many settings, a valid amplification procedure exists, even when the size of the input dataset, $n$, is significantly less than what would be necessary to learn $D$ to non-trivial accuracy. Specifically we consider two fundamental settings: the case where $D$ is an arbitrary discrete distribution supported on $\le k$ elements, and the case where $D$ is a $d$-dimensional Gaussian with unknown mean, and fixed covariance. In the first case, we show that an $\left(n, n + \Theta(\frac{n}{\sqrt{k}})\right)$ amplifier exists. In particular, given $n=O(\sqrt{k})$ samples from $D$, one can output a set of $m=n+1$ datapoints, whose total variation distance from the distribution of $m$ i.i.d. draws from $D$ is a small constant, despite the fact that one would need quadratically more data, $n=\Theta(k)$, to learn $D$ up to small constant total variation distance. In the Gaussian case, we show that an $\left(n,n+\Theta(\frac{n}{\sqrt{d}} )\right)$ amplifier exists, even though learning the distribution to small constant total variation distance requires $\Theta(d)$ samples. In both the discrete and Gaussian settings, we show that these results are tight, to constant factors. Beyond these results, we formalize a number of curious directions for future research along this vein. ",Sample Amplification: Increasing Dataset Size even when Learning is Impossible " In this work we show that in a microfluidic network and in low Reynolds numbers a system can be irreversible because of hysteresis effects.The network, which is employed in our simulations, is taken from recent experiments. The network consists of one loop connected to input and output pipes. A train of droplets enter the system at a uniform rate, but they may leave it in different patterns, e.g. periodic or even chaotic. The out put pattern depends on the time interval among the incoming droplets as well as the network geometry and for some parameters the system is not reversible. ",Simulation of Droplet Trains in Microfluidic Networks " We propose and demonstrate a new magneto-optical trap (MOT) for alkaline-earth-metal-like (AEML) atoms where the narrow $^{1}S_{0}\rightarrow$$^{3}P_{1}$ transition and the broad $^{1}S_{0}\rightarrow$$^{1}P_{1}$ transition are spatially arranged into a core-shell configuration. Our scheme resolves the main limitations of previously adopted MOT schemes, leading to a significant increase in both the loading rate and the steady state atom number. We apply this scheme to $^{174}$Yb MOT, where we show about a hundred-fold improvement in the loading rate and ten-fold improvement in the steady state atom number compared to reported cases that we know of to date. This technique could be readily extended to other AEML atoms to increase the statistical sensitivity of many different types of precision experiments. ",Core-Shell Magneto-Optical Trap for Alkaline-Earth-Metal-Like Atoms " We give explicit criteria for the reflectionlessness, transparency, and invisibility of a finite-range potential in the presence of an incoherent (intensity-dependent) nonlinearity that is confined to the range of the potential. This allows us to conduct a systematic study of the effects of such a nonlinearity on a locally periodic class of finite-range potentials that display perturbative unidirectional invisibility. We use our general results to examine the effects of a weak Kerr nonlinearity on the behavior of these potentials and show that the presence of nonlinearity destroys the unidirectional invisibility of these potentials. If the strength of the Kerr nonlinearity is so weak that the first-order perturbation theory is reliable, the presence of nonlinearity does not affect the unidirectional reflectionlessness and transmission reciprocity of the potential. We show that the expected violation of the latter is a second order perturbative effect. ",Unidirectional Reflection and Invisibility in Nonlinear Media with an Incoherent Nonlinearity " Given a countable sofic group $\Gamma$, a finite alphabet $A$, a subshift $X \subseteq A^\Gamma$, and a potential $\phi: X \to \mathbb{R}$, we give sufficient conditions on $X$ and $\phi$ for expressing, in the uniqueness regime, the sofic entropy of the associated Gibbs measure $\mu$ as the limit of the Shannon entropies of some suitable finite systems approximating $\Gamma \curvearrowright (X,\mu)$. Next, we prove that if $\mu$ satisfies strong spatial mixing, then the sofic pressure admits a formula in terms of the integral of a random information function with respect to any $\Gamma$-invariant Borel probability measure with nonnegative sofic entropy. As a consequence of our results, we provide sufficient conditions on $X$ and $\phi$ for having independence of the sofic approximation for sofic pressure and sofic entropy, and for having locality of pressure in some relevant families of systems, among other applications. These results complement and unify those of Marcus and Pavlov (2015), Alpeev (2017), and Austin and Podder (2018). ",Kieffer-Pinsker type formulas for Gibbs measures on sofic groups " We present an approach to the dynamics of interacting particle systems, which allows to derive path integral formulas from purely stochastic considerations. We show that the resulting field theory is a dual version of the standard theory of Doi and Peliti. This clarify both the origin of the Cole-Hopf map between the two approaches and the occurence of imaginary noises in effective Langevin equations for reaction-diffusion systems. The advantage of our approach is that it focuses directly on the density field. We show some applications, in particular on the Zero Range Process, hydrodynamic limits and large deviation functional. ",Dynamics of interacting particle systems: stochastic process and field theory " In recent years participatory budgeting (PB) in Scotland has grown from a handful of community-led processes to a movement supported by local and national government. This is epitomized by an agreement between the Scottish Government and the Convention of Scottish Local Authorities (COSLA) that at least 1% of local authority budgets will be subject to PB. This ongoing research paper explores the challenges that emerge from this 'scaling up' or 'mainstreaming' across the 32 local authorities that make up Scotland. The main objective is to evaluate local authority use of the digital platform Consul, which applies Natural Language Processing (NLP) to address these challenges. This project adopts a qualitative longitudinal design with interviews, observations of PB processes, and analysis of the digital platform data. Thematic analysis is employed to capture the major issues and themes which emerge. Longitudinal analysis then explores how these evolve over time. The potential for 32 live study sites provides a unique opportunity to explore discrete political and social contexts which materialize and allow for a deeper dive into the challenges and issues that may exist, something a wider cross-sectional study would miss. Initial results show that issues and challenges which come from scaling up may be tackled using NLP technology which, in a previous controlled use case-based evaluation, has shown to improve the effectiveness of citizen participation. ",Evaluating the application of NLP tools in mainstream participatory budgeting processes in Scotland " Recently, graph-based planning algorithms have gained much attention to solve goal-conditioned reinforcement learning (RL) tasks: they provide a sequence of subgoals to reach the target-goal, and the agents learn to execute subgoal-conditioned policies. However, the sample-efficiency of such RL schemes still remains a challenge, particularly for long-horizon tasks. To address this issue, we present a simple yet effective self-imitation scheme which distills a subgoal-conditioned policy into the target-goal-conditioned policy. Our intuition here is that to reach a target-goal, an agent should pass through a subgoal, so target-goal- and subgoal- conditioned policies should be similar to each other. We also propose a novel scheme of stochastically skipping executed subgoals in a planned path, which further improves performance. Unlike prior methods that only utilize graph-based planning in an execution phase, our method transfers knowledge from a planner along with a graph into policy learning. We empirically show that our method can significantly boost the sample-efficiency of the existing goal-conditioned RL methods under various long-horizon control tasks. ",Imitating Graph-Based Planning with Goal-Conditioned Policies " Certain patterns of symmetry fractionalization in topologically ordered phases of matter are anomalous, in the sense that they can only occur at the surface of a higher dimensional symmetry-protected topological (SPT) state. An important question is to determine how to compute this anomaly, which means determining which SPT hosts a given symmetry-enriched topological order at its surface. While special cases are known, a general method to compute the anomaly has so far been lacking. In this paper we propose a general method to compute relative anomalies between different symmetry fractionalization classes of a given (2+1)D topological order. This method applies to all types of symmetry actions, including anyon-permuting symmetries and general space-time reflection symmetries. We demonstrate compatibility of the relative anomaly formula with previous results for diagnosing anomalies for $\mathbb{Z}_2^{\bf T}$ space-time reflection symmetry (e.g. where time-reversal squares to the identity) and mixed anomalies for $U(1) \times \mathbb{Z}_2^{\bf T}$ and $U(1) \rtimes \mathbb{Z}_2^{\bf T}$ symmetries. We also study a number of additional examples, including cases where space-time reflection symmetries are intertwined in non-trivial ways with unitary symmetries, such as $\mathbb{Z}_4^{\bf T}$ and mixed anomalies for $\mathbb{Z}_2 \times \mathbb{Z}_2^{\bf T}$ symmetry, and unitary $\mathbb{Z}_2 \times \mathbb{Z}_2$ symmetry with non-trivial anyon permutations. ",Relative Anomalies in (2+1)D Symmetry Enriched Topological States " In this paper we study algebraic structures of the classes of the $L_2$ analytic Fourier-Feynman transforms on Wiener space. To do this we first develop several rotation properties of the generalized Wiener integral associated with Gaussian processes. We then proceed to analyze the $L_2$ analytic Fourier-Feynman transforms associated with Gaussian processes. Our results show that these $L_2$ analytic Fourier--Feynman transforms are actually linear operator isomorphisms from a Hilbert space into itself. We finally investigate the algebraic structures of these classes of the transforms on Wiener space, and show that they indeed are group isomorphic. ",Algebraic structure of the $L_2$ analytic Fourier-Feynman transform associated with Gaussian processes on Wiener space " We present $\Delta$-UQ -- a novel, general-purpose uncertainty estimator using the concept of anchoring in predictive models. Anchoring works by first transforming the input into a tuple consisting of an anchor point drawn from a prior distribution, and a combination of the input sample with the anchor using a pretext encoding scheme. This encoding is such that the original input can be perfectly recovered from the tuple -- regardless of the choice of the anchor. Therefore, any predictive model should be able to predict the target response from the tuple alone (since it implicitly represents the input). Moreover, by varying the anchors for a fixed sample, we can estimate uncertainty in the prediction even using only a single predictive model. We find this uncertainty is deeply connected to improper sampling of the input data, and inherent noise, enabling us to estimate the total uncertainty in any system. With extensive empirical studies on a variety of use-cases, we demonstrate that $\Delta$-UQ outperforms several competitive baselines. Specifically, we study model fitting, sequential model optimization, model based inversion in the regression setting and out of distribution detection, & calibration under distribution shifts for classification. ",$\Delta$-UQ: Accurate Uncertainty Quantification via Anchor Marginalization " Small inhibitory neuronal circuits have long been identified as key neuronal motifs to generate and modulate the coexisting rhythms of various motor functions. Our paper highlights the role of a cellular switching mechanism to orchestrate such circuits. The cellular switch makes the circuits reconfigurable, robust, adaptable, and externally controllable. Without this cellular mechanism, the circuits rhythms entirely rely on specific tunings of the synaptic connectivity, which makes them rigid, fragile, and difficult to control externally. We illustrate those properties on the much studied architecture of a small network controlling both the pyloric and gastric rhythms of crabs. The cellular switch is provided by a slow negative conductance often neglected in mathematical modeling of central pattern generators. We propose that this conductance is simple to model and key to computational studies of rhythmic circuit neuromodulation. ",Cellular switches orchestrate rhythmic circuits " The resolution function of a spectrometer based on a strongly bent single crystal (bending radius of 10 cm or less) is evaluated. It is shown that the resolution is controlled by two parameters, (i) the ratio of the lattice spacing of the chosen reflection to the crystal thickness and (ii) a single parameter comprising crystal thickness, its bending radius, and anisotropic elastic constants of the chosen crystal. Diamond, due to its unique elastic properties, can provide notably higher resolution than silicon. The results allow to optimize the parameters of bent crystal spectrometers for the hard X-ray free electron laser sources. ",Resolution of a bent-crystal spectrometer for X-ray free electron laser pulses: diamond vs. silicon " Density functional theory was used to study the nonmagnetic (NM) and ferromagnetic (FM) phases of face-centered cubic cerium. Functionals of four levels of approximations for the exchange-correlation energy were used: LDA, PBE, LDA/PBE+$U$, and YS-PBEh. The latter two contain an adjustable parameter, the onsite Coulomb repulsion parameter $U$ for LDA/PBE+$U$ and the fraction $\alpha_x$ of Hartree-Fock exchange for YS-PBEh, which were varied in order to study their influence on the results. By supposing that, as a first approximation, the NM and FM solutions can be identified to the observed $\alpha$ and $\gamma$ phases, respectively, it is concluded that while a small value of $U$ or $\alpha_x$ leads to the correct trend for the stability ordering of the two phases, larger values are necessary for a more appropriate (but still not satisfying) description of the electronic structure. ",Nonmagnetic and ferromagnetic fcc cerium studied with one-electron methods " Numerical MHD codes have become extraordinarily powerful tools with which to study accretion turbulence. They have been used primarily to extract values for the classical $\alpha$ parameter, and to follow complex evolutionary development. Energy transport, which is at the heart of classical disk theory, has yet to be explored in any detail. Further topics that should be explored by simulation include nonideal MHD, radiation physics, and outburst behavior related to the temperature sensitivity of the resistivity. ",Numerical Simulations of the MRI and Real Disks " Following the recent discovery of X-ray quasi-periodic eruptions (QPEs) coming from the nucleus of the galaxy GSN 069, here we report on the detection of QPEs in the active galaxy named RX J1301.9+2747. QPEs are rapid and recurrent increases of the X-ray count-rate by more than one order of magnitude with respect to a stable quiescent level. During a XMM-Newton observation lasting 48 ks that was performed on 30 and 31 May 2019, three strong QPEs lasting about half an hour each were detected in the light curves of RX J1301.9+2747. The first two QPEs are separated by a longer recurrence time (about 20 ks) compared to the second and third (about 13 ks). This pattern is consistent with the alternating long-short recurrence times of the GSN 069 QPEs, although the difference between the consecutive recurrence times is significantly smaller in GSN 069. Longer X-ray observations will better clarify the temporal pattern of the QPEs in RX J1301.9+2747 and will allow a detailed comparison with GSN 069 to be performed. The X-ray spectral properties of QPEs in the two sources are remarkably similar, with QPEs representing fast transitions from a relatively cold and likely disk-dominated state to a state that is characterized by a warmer emission similar to the so-called soft X-ray excess, a component that is almost ubiquitously seen in the X-ray spectra of unobscured, radiatively efficient active galaxies. Previous X-ray observations of RX J1301.9+2747 in 2000 and 2009 strongly suggest that QPEs have been present for at least the past 18.5 years. The detection of QPEs from a second galactic nucleus after GSN 069 rules out contamination by a Galactic source in both cases, such that QPEs ought to be considered a novel extragalactic phenomenon associated with accreting supermassive black holes. ",X-ray quasi-periodic eruptions from the galactic nucleus of RX J1301.9+2747 " We here discuss the emergence of Quasi Stationary States (QSS), a universal feature of systems with long-range interactions. With reference to the Hamiltonian Mean Field (HMF) model, numerical simulations are performed based on both the original $N$-body setting and the continuum Vlasov model which is supposed to hold in the thermodynamic limit. A detailed comparison unambiguously demonstrates that the Vlasov-wave system provides the correct framework to address the study of QSS. Further, analytical calculations based on Lynden-Bell's theory of violent relaxation are shown to result in accurate predictions. Finally, in specific regions of parameters space, Vlasov numerical solutions are shown to be affected by small scale fluctuations, a finding that points to the need for novel schemes able to account for particles correlations. ",Exploring the thermodynamic limit of Hamiltonian models: convergence to the Vlasov equation " In 2020, Yamakawa and Okuno proposed a stabilized sequential quadratic semidefinite programming (SQSDP) method for solving, in particular, degenerate nonlinear semidefinite optimization problems. The algorithm is shown to converge globally without a constraint qualification, and it has some nice properties, including the feasible subproblems, and their possible inexact computations. In particular, the convergence was established for approximate-Karush-Kuhn-Tucker (AKKT) and trace-AKKT conditions, which are two sequential optimality conditions for the nonlinear conic contexts. However, recently, complementarity-AKKT (CAKKT) conditions were also consider, as an alternative to the previous mentioned ones, that is more practical. Since few methods are shown to converge to CAKKT points, at least in conic optimization, and to complete the study associated to the SQSDP, here we propose a revised version of the method, maintaining the good properties. We modify the previous algorithm, prove the global convergence in the sense of CAKKT, and show some preliminary numerical experiments. ",A revised sequential quadratic semidefinite programming method for nonlinear semidefinite optimization " This paper shows how to apply memoization (caching of subgoals and associated answer substitutions) in a constraint logic programming setting. The research is is motivated by the desire to apply constraint logic programming (CLP) to problems in natural language processing that involve (constraint) interleaving or coroutining, such as GB and HPSG parsing. ",Memoization in Constraint Logic Programming " We report the observation of charmless hadronic decays of charged B mesons to the final state K+K-pi+. Using a data sample of 347.5 fb^-1 collected at the Y(4S) resonance with the BABAR detector, we observe 429+/-43 signal events with a significance of 9.6 sigma. We measure the inclusive branching fraction BF(B+ --> K+K-pi+) = [5.0+/-0.5(stat)+/-0.5(syst)]x10^-6. Inspection of the Dalitz plot of signal candidates shows a broad structure peaking near 1.5 GeV/c^2 in the K+K- invariant mass distribution. We find the direct CP asymmetry to be consistent with zero. ",Observation of the Decay B+ --> K+K-pi+ " The tunnel conductance in normal-metal / insulator / PrOs$_4$Sb$_{12}$ junctions is theoretically studied, where skutterudite PrOs$_4$Sb$_{12}$ is considered to be an unconventional superconductor. The conductance are calculated for several pair potentials which have been proposed in recent works. The results show that the conductance is sensitive to the relation between the direction of electric currents and the position of point nodes. We also show that the conductance spectra often deviate from the shape of the bulk density of states and that the sub gap spectra have peak structures in the case of the spin-triplet pair potentials. The results indicate that the tunnel conductance is a useful tool to obtain an information of the pairing symmetry. ",Tunneling Spectra of Skutterudite PrOs_4Sb_{12} " The analysis of perturbative quantities is a powerful tool to distinguish between different Dark Energy models and gravity theories degenerated at the background level. In this work, we generalise the integral solution of the matter density contrast for General Relativity gravity to a wide class of Modified Gravity (MG) theories. To calculate this solution is necessary prior knowledge of the Hubble rate, the density parameter at the present epoch ($\Omega_{m0}$) and the functional form of the effective Newton's constant that characterises the gravity theory. We estimate in a model-independent way the Hubble expansion rate by applying a non-parametric reconstruction method to model-independent cosmic chronometer data and high-$z$ quasar data. In order to compare our generalised solution of the matter density contrast, using the non-parametric reconstruction of $H(z)$ from observational data, with purely theoretical one, we choose a parameterisation of the Screened MG and the $\Omega_{m0}$ from WMAP-9 collaborations. Finally, we calculate the growth index for the analysed cases, finding very good agreement between theoretical values and the obtained ones using the approach presented in this work. ",Reconstruction of cosmological matter perturbations in Modified Gravity " We report equilibrium geometric structures of CuO2, CuO3, CuO6, and CuO clusters obtained by an all-electron linear combination of atomic orbitals scheme within the density-functional theory with generalized gradient approximation to describe the exchange-correlation effects. The vibrational stability of all clusters is examined on the basis of the vibrational frequencies. A structure with Cs symmetry is found to be the lowest-energy structure for CuO2, while a -shaped structure with C2v symmetry is the most stable structure for CuO3. For the larger CuO6 and CuO clusters, several competitive structures exist with structures containing ozonide units being higher in energy than those with O2 units. The infrared and Raman spectra are calculated for the stable optimal geometries. ~ ","Molecular structures and vibrations of neutral and anionic CuOx (x = 1-3,6) clusters" " Quantum metrology holds the promise of an early practical application of quantum technologies, in which measurements of physical quantities can be made with much greater precision than what is achievable with classical technologies. In this review, we collect some of the key theoretical results in quantum parameter estimation by presenting the theory for the quantum estimation of a single parameter, multiple parameters, and optical estimation using Gaussian states. We give an overview of results in areas of current research interest, such as Bayesian quantum estimation, noisy quantum metrology, and distributed quantum sensing. We address the question how minimum measurement errors can be achieved using entanglement as well as more general quantum states. This review is presented from a geometric perspective. This has the advantage that it unifies a wide variety of estimation procedures and strategies, thus providing a more intuitive big picture of quantum parameter estimation. ",A Geometric Perspective on Quantum Parameter Estimation " Pretrained language models have been suggested as a possible alternative or complement to structured knowledge bases. However, this emerging LM-as-KB paradigm has so far only been considered in a very limited setting, which only allows handling 21k entities whose single-token name is found in common LM vocabularies. Furthermore, the main benefit of this paradigm, namely querying the KB using a variety of natural language paraphrases, is underexplored so far. Here, we formulate two basic requirements for treating LMs as KBs: (i) the ability to store a large number facts involving a large number of entities and (ii) the ability to query stored facts. We explore three entity representations that allow LMs to represent millions of entities and present a detailed case study on paraphrased querying of world knowledge in LMs, thereby providing a proof-of-concept that language models can indeed serve as knowledge bases. ","Language Models as Knowledge Bases: On Entity Representations, Storage Capacity, and Paraphrased Queries" " We discuss two ways in which one can study two-charge supertubes as components of generic three-charge, three-dipole charge supergravity solutions. The first is using the Born-Infeld action of the supertubes, and the second is via the complete supergravity solution. Even though the Born-Infeld description is only a probe approximation, we find that it gives exactly the same essential physics as the complete supergravity solution. Since supertubes can depend on arbitrary functions, our analysis strengthens the evidence for the existence of three-charge black-hole microstate geometries that depend on an infinite set of parameters, and sets the stage for the computation of the entropy of these backgrounds. We examine numerous other aspects of supertubes in three-charge, three-dipole charge supergravity backgrounds, including chronology protection during mergers, the contribution of supertubes to the charges and angular momenta, and the enhancement of their entropy. In particular, we find that entropy enhancement affects supertube fluctuations both along the internal and the spacetime directions, and we prove that the charges that give the enhanced entropy can be much larger than the asymptotic charges of the solution. We also re-examine the embedding of five-dimensional black rings in Taub-NUT, and show that in different coordinate patches a ring can correspond to different four-dimensional black holes. Last, but not least, we show that all the three-charge black hole microstate geometries constructed so far can be embedded in AdS_3 x S^3, and hence can be related to states of the D1-D5 CFT. ",Supertubes in Bubbling Backgrounds: Born-Infeld Meets Supergravity " In this contribution a method is introduced that allows for a linkage between the process-induced structural damage and the fracture behaviour. Based on an anisotropic elastic material model, different modelling approaches for initial damage effects are introduced and compared. The approaches are applied to remote laser cut carbon fibre reinforced polymers in order to model various thermally induced damage effects like chemical decomposition, micro-cracks and delamination. The dimensions of this heat affected zone are calculated with 1D-heat conduction. In experiment and simulation milled and laser cut specimens with different process parameters are compared in order to quantify the impact of the cutting technology on the fracture behaviour. For this purpose open hole specimens were used. ",Analysis of process-induced damage in remote laser cut carbon fibre reinforced polymers " We predict the dwarf galaxy detection limits for the upcoming Chinese Space Station Telescope (CSST) survey that will cover 17,500 deg$^{2}$ of the sky with a wide field of view of 1.1 deg$^2$. The point-source depth reaches 26.3 mag in the $g$ band and 25.9 mag in the $i$ band. Constructing mock survey data based on the designed photometric bands, we estimate the recovery rate of artificial dwarf galaxies from mock point-source photometric catalogues. The detection of these artificial dwarf galaxies is strongly dependent on their distance, magnitude and size, in agreement with searches in current surveys. We expect CSST to enable the detection of dwarf galaxies with $M_V = -3.0$ and $\mu_{250} = 32.0$ mag/arcsec$^2$ (surface-brightness limit for a system of half-light radius $r_{\rm h}$ = 250 pc at 400 kpc, and $M_V = -4.9$ and $\mu_{250} = 30.5$ mag/arcsec$^2$ around the Andromeda galaxy. Beyond the Local Group, the CSST survey will achieve $M_V = -5.8$, and $\mu_{250}$ = 29.7 mag/arcsec$^2$ in the distance range of 1--2 Mpc, opening up an exciting discovery space for faint field dwarf galaxies. With its optical bands, wide survey footprint, and space resolution, CSST will undoubtedly expand our knowledge of low-mass dwarf galaxies to an unprecedented volume. ",Local Group Dwarf Galaxy Detection Limit in the CSST survey " We propose a Markovian quantum master equation that can describe the Fano effect directly, by assuming a standard cavity quantum electrodynamics system. The framework allows us to generalize the Fano formula, applicable over the weak and strong coupling regimes with pure dephasing. A formulation of its emission spectrum is also given in a consistent manner. We then find that the interference responsible for the Fano effect is robust against pure dephasing. This is counterintuitive because the impact of interference is, in general, severely reduced by decoherence processes. Our approach thus provides a basis for theoretical treatments of the Fano effect and new insights into the quantum interference in open quantum systems. ",Theory of Fano effect in cavity quantum electrodynamics " Integral field spectroscopy of 11 type-Ib/c supernova explosion sites in nearby galaxies has been obtained using UH88/SNIFS and Gemini-N/GMOS. The use of integral field spectroscopy enables us to obtain both spatial and spectral information of the explosion site, allowing the identification of the parent stellar population of the supernova progenitor star. The spectrum of the parent population provides metallicity determination via strong-line method and age estimation obtained via comparison with simple stellar population (SSP) models. We adopt this information as the metallicity and age of the supernova progenitor, under the assumption that it was coeval with the parent stellar population. The age of the star corresponds to its lifetime, which in turn gives the estimate of its initial mass. With this method we were able to determine both the metallicity and initial (ZAMS) mass of the progenitor stars of the type Ib and Ic supernovae. We found that on average SN Ic explosion sites are more metal-rich and younger than SN Ib sites. The initial mass of the progenitors derived from parent stellar population age suggests that SN Ic have more massive progenitors than SN Ib. In addition, we also found indication that some of our SN progenitors are less massive than ~25 Msun, indicating that they may have been stars in a close binary system that have lost their outer envelope via binary interactions to produce Ib/c supernovae, instead of single Wolf-Rayet stars. These findings support the current suggestions that both binary and single progenitor channels are in effect in producing type Ib/c supernovae. This work also demonstrates the power of integral field spectroscopy in investigating supernova environments and active star forming regions. ",Integral field spectroscopy of supernova explosion sites: constraining mass and metallicity of the progenitors - I. Type Ib and Ic supernovae " In this work, we study strong and radiative decays of S-wave D\Xi molecular state, which is related to the \Omega^*_c states newly observed at LHCb. The coupling between the D\Xi molecular state and its constituents D and \Xi is calculated by using the compositeness condition. With the obtained coupling, the partial decay widths of the D\Xi molecular state into the \Xi_c^{+}K^{-}, \Xi^{'+}_cK^{-} and \Omega^{*}_c(2695)\gamma final states through hadronic loop are calculated with the help of the effective Lagrangians. By comparison with the LHCb observation, the current results of total decay width support the \Omega^{*}_c(3119) or \Omega^{*}_c(3050) as D\Xi molecule while the the decay width of the \Omega^{*}_c(3000), \Omega^{*}_c(3066) and \Omega^{*}_c(3090) can not be well reproduced in the molecular state picture. The partial decay widths are also presented and helpful to further understand the internal structures of \Omega^{*}_c(3119) and \Omega^{*}_c(3050). ",Strong and radiative decays of D\Xi molecular state and newly observed $\Omega_c$ states " We analyze stability and generation of discrete gap solitons in weakly coupled optical waveguides. We demonstrate how both stable and unstable solitons can be observed experimentally in the engineered binary waveguide arrays, and also reveal a connection between the gap-soliton instabilities and limitations on the mutual beam focusing in periodic photonic structures. ",Generation and stability of discrete gap solitons " A main puzzle of deep neural networks (DNNs) revolves around the apparent absence of ""overfitting"", defined in this paper as follows: the expected error does not get worse when increasing the number of neurons or of iterations of gradient descent. This is surprising because of the large capacity demonstrated by DNNs to fit randomly labeled data and the absence of explicit regularization. Recent results by Srebro et al. provide a satisfying solution of the puzzle for linear networks used in binary classification. They prove that minimization of loss functions such as the logistic, the cross-entropy and the exp-loss yields asymptotic, ""slow"" convergence to the maximum margin solution for linearly separable datasets, independently of the initial conditions. Here we prove a similar result for nonlinear multilayer DNNs near zero minima of the empirical loss. The result holds for exponential-type losses but not for the square loss. In particular, we prove that the weight matrix at each layer of a deep network converges to a minimum norm solution up to a scale factor (in the separable case). Our analysis of the dynamical system corresponding to gradient descent of a multilayer network suggests a simple criterion for ranking the generalization performance of different zero minimizers of the empirical loss. ",Theory IIIb: Generalization in Deep Networks " Safe interaction with the environment is one of the most challenging aspects of Reinforcement Learning (RL) when applied to real-world problems. This is particularly important when unsafe actions have a high or irreversible negative impact on the environment. In the context of network management operations, Remote Electrical Tilt (RET) optimisation is a safety-critical application in which exploratory modifications of antenna tilt angles of base stations can cause significant performance degradation in the network. In this paper, we propose a modular Safe Reinforcement Learning (SRL) architecture which is then used to address the RET optimisation in cellular networks. In this approach, a safety shield continuously benchmarks the performance of RL agents against safe baselines, and determines safe antenna tilt updates to be performed on the network. Our results demonstrate improved performance of the SRL agent over the baseline while ensuring the safety of the performed actions. ",A Safe Reinforcement Learning Architecture for Antenna Tilt Optimisation " We review recent finite opacity approaches (GLV, WW, WOGZ) to the computation of the induced gluon radiative energy loss and their application to the tomographic studies of the density evolution in ultra-relativistic nuclear collisions. ",Jet Quenching and Radiative Energy Loss in Dense Nuclear Matter " We establish that in Quantum Chromodynamics (QCD) at zero temperature, SU_{L+R}(N_F) exhibits the vector mode conjectured by Georgi and SU_{L-R}(N_F) is realized in either the Nambu-Goldstone mode or else Q_5^a is also screened from view at infinity. The Wigner-Weyl mode is ruled out unless the beta function in QCD develops an infrared stable zero. ","The Georgi ""Avatar"" of Broken Chiral Symmetry in Quantum Chromodynamics" " In this talk we discuss an improvement of the Diakonov-Petrov QCD Effective Action. We propose the Improved Effective Action, which is derived on the basis of the Lee-Bardeen results for the quark determinant in the instanton field. The Improved Effective Action provides proper account of the current quark masses, which is particularly important for strange quarks. This Action is successfully tested by the calculations of the quark condensate, the masses of the pseudoscalar meson octet and by axial-anomaly low-energy theorems. ",Light Quarks Beyond Chiral Limit " We show that a Reissner-Nordstr\""{o}m (RN) black hole can be formed by dropping a charged thin dust shell onto a RN naked singularity. This is in contrast to the fact that a RN naked singularity is prohibited from forming by dropping a charged thin dust shell onto a RN black hole. This implies the strong tendency of the RN singularity to be covered by a horizon in favour of cosmic censorship. We show that an extreme RN black hole can also be formed from a RN naked singularity by the same process in a finite advanced time. We also discuss the evolution of the charged thin dust shells and the causal structure of the resultant spacetimes. ",Dynamical Transition from a Naked Singularity to a Black Hole " We make use of the S=1 pseudospin formalism to describe the charge degree of freedom in a model high-$T_c$ cuprate with the on-site Hilbert space reduced to the three effective valence centers, nominally Cu$^{1+,\,2+,\,3+}$. Starting with a parent cuprate as an analogue of the quantum paramagnet ground state and using the Schwinger boson technique we found the pseudospin spectrum and conditions for the pseudomagnon condensation with phase transition to a superconducting state. ",Superconductivity in model cuprate as an S=1 pseudomagnon condensation " Following Cui et al. 2018 (hereafter Paper I) on the classification of large-scale environments (LSE) at z = 0, we push our analysis to higher redshifts and study the evolution of LSE and the baryon distributions in them. Our aim is to investigate how baryons affect the LSE as a function of redshift. In agreement with Paper I, the baryon models have negligible effect on the LSE over all investigated redshifts. We further validate the conclusion obtained in Paper I that the gas web is an unbiased tracer of total matter -- even better at high redshifts. By separating the gas mainly by temperature, we find that about 40 per cent of gas is in the so-called warm-hot intergalactic medium (WHIM). This fraction of gas mass in the WHIM decreases with redshift, especially from z = 1 (29 per cent) to z = 2.1 (10 per cent). By separating the whole WHIM gas mass into the four large-scale environments (i.e. voids, sheets, filaments, and knots), we find that about half of the WHIM gas is located in filaments. Although the total gas mass in WHIM decreases with redshift, the WHIM mass fractions in the different LSE seem unchanged. ",The large-scale environment from cosmological simulations II: The redshift evolution and distributions of baryons " The spectrum of higher harmonics in atoms calculated with a uniformized semiclassical propagator is presented and it is shown that higher harmonic generation is an interference phenomenon which can be described semiclassically. This can be concluded from the good agreement with the quantum spectrum. Moreover, the formation of a plateau in the spectrum is specifically due to the interference of irregular, time delayed, trajectories with regular orbits without a time-delay. This is proven by the absence of the plateau in an artificial semiclassical spectrum generated from a sample of trajectories from which the irregular trajectories (only a few percent) have been discarded. ",Irregular orbits generate higher harmonics " In this paper, we study in-depth the problem of online self-calibration for robust and accurate visual-inertial state estimation. In particular, we first perform a complete observability analysis for visual-inertial navigation systems (VINS) with full calibration of sensing parameters, including IMU and camera intrinsics and IMU-camera spatial-temporal extrinsic calibration, along with readout time of rolling shutter (RS) cameras (if used). We investigate different inertial model variants containing IMU intrinsic parameters that encompass most commonly used models for low-cost inertial sensors. The observability analysis results prove that VINS with full sensor calibration has four unobservable directions, corresponding to the system's global yaw and translation, while all sensor calibration parameters are observable given fully-excited 6-axis motion. Moreover, we, for the first time, identify primitive degenerate motions for IMU and camera intrinsic calibration. Each degenerate motion profile will cause a set of calibration parameters to be unobservable and any combination of these degenerate motions are still degenerate. Extensive Monte-Carlo simulations and real-world experiments are performed to validate both the observability analysis and identified degenerate motions, showing that online self-calibration improves system accuracy and robustness to calibration inaccuracies. We compare the proposed online self-calibration on commonly-used IMUs against the state-of-art offline calibration toolbox Kalibr, and show that the proposed system achieves better consistency and repeatability. Based on our analysis and experimental evaluations, we also provide practical guidelines for how to perform online IMU-camera sensor self-calibration. ","Online Self-Calibration for Visual-Inertial Navigation Systems: Models, Analysis and Degeneracy" " Electronegativity is shown to control charge transfer, energy level alignments, and electron currents in single molecule tunnel junctions, all of which are governed by correlations contained within the density matrix. This is demonstrated by the fact that currents calculated from the one-electron reduced density matrix to second order in electron correlation are identical to the currents obtained from the Green's function corrected to second order in electron self-energy. ",Electronegativity in quantum electronic transport " We present a numerically efficient technique to evaluate the Green's function for extended two dimensional systems without relying on periodic boundary conditions. Different regions of interest, or `patches', are connected using self energy terms which encode the information of the extended parts of the system. The calculation scheme uses a combination of analytic expressions for the Green's function of infinite pristine systems and an adaptive recursive Green's function technique for the patches. The method allows for an efficient calculation of both local electronic and transport properties, as well as the inclusion of multiple probes in arbitrary geometries embedded in extended samples. We apply the Patched Green's function method to evaluate the local densities of states and transmission properties of graphene systems with two kinds of deviations from the pristine structure: bubbles and perforations with characteristic dimensions of the order of 10-25 nm, i.e. including hundreds of thousands of atoms. The strain field induced by a bubble is treated beyond an effective Dirac model, and we demonstrate the existence of both Friedel-type oscillations arising from the edges of the bubble, as well as pseudo-Landau levels related to the pseudomagnetic field induced by the nonuniform strain. Secondly, we compute the transport properties of a large perforation with atomic positions extracted from a TEM image, and show that current vortices may form near the zigzag segments of the perforation. ",Patched Green's function techniques for two dimensional systems: Electronic behaviour of bubbles and perforations in graphene " We study the production of isolated photons in $e^+e^-$ annihilation and give the proof of the all-order factorization of the collinear singularities. These singularities are absorbed in the standard fragmentation functions of partons into a photon, while the effects of the isolation are consistently included in the short-distance cross section. We compute this cross section at order $\as$ and show that it contains large double logarithms of the isolation parameters. We explain the physical origin of these logarithms and discuss the possibility to resum them to all orders in $\as$. ",Factorization and soft-gluon divergences in isolated-photon cross sections " We consider a superlattice of parallel metal tunnel junctions with a spatially non-homogeneous probability for electrons to tunnel. In such structures tunneling can be accompanied by electron scattering that conserves energy but not momentum. In the special case of a tunneling probability that varies periodically with period $a$ in the longitudinal direction, i.e., perpendicular to the junctions, electron tunneling is accompanied by ""umklapp"" scattering, where the longitudinal momentum changes by a multiple of $h/a$. We predict that as a result a sequence of metal-insulator transitions can be induced by an external electric- or magnetic field as the field strength is increased. ",Umklapp-Assisted Electron Transport Oscillations in Metal Superlattices In the first part we summarize the status of the nucleon-nucleon (NN) problem in the context of Hamiltonian based constituent quark models and present results for the l=0 phase shifts obtained from the Goldstone-boson exchange model by applying the resonationg group method. The second part deals with the construction of local shallow and deep equivalent potentials based on a Supersymmetric Quantum Mechanics approach. ,The Nucleon-Nucleon Problem in Quark Models " We study the effect of the exchange interaction on the Coulomb blockade peak height statistics in chaotic quantum dots. Because exchange reduces the level repulsion in the many body spectrum, it strongly affects the fluctuations of the peak conductance at finite temperature. We find that including exchange substantially improves the description of the experimental data. Moreover, it provides further evidence of the presence of high spin states (S>1) in such systems. ",Exchange and the Coulomb blockade: Peak height statistics in quantum dots " We demonstrate the reconstruction of the exciton-polariton condensate loaded in a single active miniband in one-dimensional microcavity wires with a complex-valued periodic potentials. The effect appears due to strong polariton-polariton repulsion and it depends on the type of the single-particle dispersion of the miniband, which can be fine tuned by the real and imaginary components of the potential. As a result, the condensate can be formed in a $0$-state, $\pi$-state, or mixed state of spatiotemporal intermittency, depending on the shape of the miniband, strength of interparticle interaction, and distribution of losses in the system. The reconstruction of the condensate wave function takes place by proliferation of nuclei of the new condensate phase in the form of dark solitons. We show that, in general, the interacting polaritons are not condensed in the state with minimal losses, neither they accumulate in the state with a well-defined wave vector. ",Reconstruction of Exciton-Polariton Condensates in 1D Periodic Structures " Given a forcing notion $P$ that forces certain values to several classical cardinal characteristics of the reals, we show how we can compose $P$ with a collapse (of a cardinal $\lambda>\kappa$ to $\kappa$) such that the composition still forces the previous values to these characteristics. We also show how to force distinct values to $\mathfrak m$, $\mathfrak p$ and $\mathfrak h$ and also keeping all the values in Cicho\'n's diagram distint, using the Boolean Ultrapower method of arXiv:1708.03691 . (In arXiv:2006.09826 , the same was done for the newer Cicho\'n's Maximum construction, which avoids large cardinals.) ",Controlling classical cardinal characteristics while collapsing cardinals Order estimates for the Kolmogorov widths of an intersection of two finite-dimensional balls in a mixed norm under some conditions on the parameters are obtained. ,Estimates for the Kolmogorov widths of an intersection of two balls in a mixed norm " Counterfactual Explanations (CEs) are an important tool in Algorithmic Recourse for addressing two questions: 1. What are the crucial factors that led to an automated prediction/decision? 2. How can these factors be changed to achieve a more favorable outcome from a user's perspective? Thus, guiding the user's interaction with AI systems by proposing easy-to-understand explanations and easy-to-attain feasible changes is essential for the trustworthy adoption and long-term acceptance of AI systems. In the literature, various methods have been proposed to generate CEs, and different quality measures have been suggested to evaluate these methods. However, the generation of CEs is usually computationally expensive, and the resulting suggestions are unrealistic and thus non-actionable. In this paper, we introduce a new method to generate CEs for a pre-trained binary classifier by first shaping the latent space of an autoencoder to be a mixture of Gaussian distributions. CEs are then generated in latent space by linear interpolation between the query sample and the centroid of the target class. We show that our method maintains the characteristics of the input sample during the counterfactual search. In various experiments, we show that the proposed method is competitive based on different quality measures on image and tabular datasets -- efficiently returns results that are closer to the original data manifold compared to three state-of-the-art methods, which are essential for realistic high-dimensional machine learning applications. ",Counterfactual Explanation via Search in Gaussian Mixture Distributed Latent Space The lifetimed of the Bbar0 and B- meson lifetimes are measured using data recorded on the Z peak with the ALEPH detector at LEP. An improved analysis based on partially reconstructed Bbar0 -> D*+l-nubar and B- -> D0l-nubar decays is presented. ,Measurement of the B0 and B- meson lifetimes in ALEPH " A heterogeneous brittle material characterized by a random field of local toughness Kc(x) can be represented by an equivalent homogeneous medium of toughness, Keff. Homogenization refers to a process of estimating Keff from the local field Kc(x). An approach based on a perturbative expansion of the stress intensity factor along a rough crack front shows the occurrence of different regimes depending on the correlation length of the local toughness field in the direction of crack propagation. A `""weak pinning"" regime takes place for long correlation lengths, where the effective toughness is the average of the local toughness. For shorter correlation lengths, a transition to ""strong pinning"" occurs leading to a much higher effective toughness, and characterized by a propagation regime consisting in jumps between pinning configurations. ",Effective toughness of heterogeneous brittle materials " We present a simple model to study L\'{e}vy-flight foraging in a finite landscape with countable targets. In our approach, foraging is a step-based exploratory random search process with a power-law step-size distribution $P(l) \propto l^{-\mu}$. We find that, when the termination is regulated by a finite number of steps $N$, the optimum value of $\mu$ that maximises the foraging efficiency can vary substantially in the interval $\mu \in (1,3)$, depending on the landscape features (landscape size and number of targets). We further demonstrate that subjective returning can be another significant factor that affects the foraging efficiency in such context. Our results suggest that L\'{e}vy-flight foraging may arise through an interaction between the environmental context and the termination of exploitation, and particularly that the number of steps can play an important role in this scenario which is overlooked by most previous work. Our study not only provides a new perspective on L\'{e}vy-flight foraging, but also opens new avenues for investigating the interaction between foraging dynamics and environment as well as offers a realistic framework for analysing animal movement patterns from empirical data. ",Optimal L\'{e}vy-flight foraging in a finite landscape An s-tuple of positive integers are k-wise relatively prime if any k of them are relatively prime. Exact formula is obtained for the probability that s positive integers are k-wise relatively prime. ,The probability that random positive integers are k-wise relatively prime " We investigate the asymptotic rates of length-$n$ binary codes with VC-dimension at most $dn$ and minimum distance at least $\delta n$. Two upper bounds are obtained, one as a simple corollary of a result by Haussler and the other via a shortening approach combining Sauer-Shelah lemma and the linear programming bound. Two lower bounds are given using Gilbert-Varshamov type arguments over constant-weight and Markov-type sets. ",On the VC-Dimension of Binary Codes " Most depression assessment tools are based on self-report questionnaires, such as the Patient Health Questionnaire (PHQ-9). These psychometric instruments can be easily adapted to an online setting by means of electronic forms. However, this approach lacks the interacting and engaging features of modern digital environments. With the aim of making depression screening more available, attractive and effective, we developed Perla, a conversational agent able to perform an interview based on the PHQ-9. We also conducted a validation study in which we compared the results obtained by the traditional self-report questionnaire with Perla's automated interview. Analyzing the results from this study we draw two significant conclusions: firstly, Perla is much preferred by Internet users, achieving more than 2.5 times more reach than a traditional form-based questionnaire; secondly, her psychometric properties (Cronbach's alpha of 0.81, sensitivity of 96% and specificity of 90%) are excellent and comparable to the traditional well-established depression screening questionnaires. ","Perla: A Conversational Agent for Depression Screening in Digital Ecosystems. Design, Implementation and Validation" " We demonstrate that the new single crystal of YCu$_3$[OH(D)]$_{6.5}$Br$_{2.5}$ (YCOB) is a kagome Heisenberg antiferromagnet (KHA) without evident orphan spins ($\ll$ 0.8\%). The site mixing between polar OH$^-$ and non-polar Br$^-$ causes local distortions of Cu-O-Cu exchange paths, and gives rise to 70(2)\% of randomly distributed hexagons of alternate bonds ($\sim$ $J_1-\Delta J$ and $J_1+\Delta J$) and the rest of almost uniform hexagons ($\sim$ $J_1$) on the kagome lattice. Simulations of the random exchange model with $\Delta J$/$J_1$ = 0.7(1) show good agreement with the experimental observations, including the weak upturn seen in susceptibility and the slight polarization in magnetization. Despite the average antiferromagnetic coupling of $J_1$ $\sim$ 60 K, no conventional freezing is observed down to $T$ $\sim$ 0.001$J_1$, and the raw specific heat exhibits a nearly quadratic temperature dependence below 1 K $\sim$ 0.02$J_1$, phenomenologically consistent with a gapless (spin gap $\leq$ 0.025$J_1$) Dirac quantum spin liquid (QSL). Our result sheds new light on the theoretical understanding of the randomness-relevant gapless QSL behavior in YCOB, as well as in other relevant materials. ",Gapless Spin Liquid Behavior in A Kagome Heisenberg Antiferromagnet with Randomly Distributed Hexagons of Alternate Bonds " The surface tension of living cells and tissues originates from the generation of nonequilibrium active stresses within the cell cytoskeleton. Here, using laser ablation, we generate gradients in the surface tension of cellular aggregates as models of simple tissues. These gradients of active surface stress drive large-scale and rapid toroidal motion. Subsequently, the motions spontaneously reverse as stresses reaccumulate and cells return to their original positions. Both forward and reverse motions resemble Marangoni flows in viscous fluids. However, the motions are faster than the timescales of viscoelastic relaxation, and the surface tension gradient is proportional to mechanical strain at the surface. Further, due to active stress, both the surface tension gradient and surface strain are dependent upon the volume of the aggregate. These results indicate that surface tension can induce rapid and highly correlated elastic deformations in the maintenance of tissue shape and configuration. ",Gradients in solid surface tension drive Marangoni-like motions in cell aggregates " The one-loop Higgs coupling to two gluons has been invoked in the past to estimate that the fraction of the nucleon mass which is due to the Higgs is rather small but calculable (approximately 8 percent). To test the veracity of this hypothesis, we employ the same mechanism to compute the Higgs coupling to an arbitrary stable nucleus $A$ and its anti-nucleus $\bar{A}$. We find that the physical decay rate of a Higgs into a spin zero $A\bar{A}$ pair near the threshold corresponding to the Higgs mass is quite substantial, once we include the final state Coulomb corrections as well as possible form factor effects. If true, observation of even a few such decay events would be truly spectacular (with no competing background) since we are unaware of any other interaction which might lead to the production of a very heavy nucleus accompanied by its anti nucleus in nucleon-(anti-) nucleon scattering. ",Production and detection of heavy matter anti-matter from Higgs decays " We introduce a double quantum (DQ) 4-Ramsey measurement protocol that enables wide-field magnetic imaging using nitrogen vacancy (NV) centers in diamond, with enhanced homogeneity of the magnetic sensitivity relative to conventional single quantum (SQ) techniques. The DQ 4-Ramsey protocol employs microwave-phase alternation across four consecutive Ramsey (4-Ramsey) measurements to isolate the desired DQ magnetic signal from any residual SQ signal induced by microwave pulse errors. In a demonstration experiment employing a 1-$\mu$m-thick NV layer in a macroscopic diamond chip, the DQ 4-Ramsey protocol provides volume-normalized DC magnetic sensitivity of $\eta^\text{V}=34\,$nTHz$^{-1/2} \mu$m$^{3/2}$ across a $125\,\mu$m$ \,\times\,125\,\mu $m field of view, with about 5$\times$ less spatial variation in sensitivity across the field of view compared to a SQ measurement. The improved robustness and magnetic sensitivity homogeneity of the DQ 4-Ramsey protocol enable imaging of dynamic, broadband magnetic sources such as integrated circuits and electrically-active cells. ",NV-Diamond Magnetic Microscopy using a Double Quantum 4-Ramsey Protocol " Traditionally, networks operate at a small fraction of their capacities; however, recent technologies, such as Software-Defined Networking, may let operators run their networks harder (i.e., at higher utilization levels). Higher utilization can increase the network operator's revenue, but this gain comes at a cost: daily traffic fluctuations and failures might occasionally overload the network. We call such situations Resource Crunch. Dealing with Resource Crunch requires certain types of flexibility in the system. We focus on scenarios with flexible bandwidth requirements, e.g., some connections can tolerate lower bandwidth allocation. This may free capacity to provision new requests that would otherwise be blocked. For that, the network operator needs to make an informed decision, since reducing the bandwidth of a high-paying connection to allocate a low-value connection is not sensible. We propose a strategy to decide whether or not to provision a request (and which other connections to degrade) focusing on maximizing profits during Resource Crunch. To address this problem, we use an abstraction of the network state, called a Connection Adjacency Graph (CAG). We propose PROVISIONER, which integrates our CAG solution with an efficient Linear Program (LP). We compare our method to existing greedy approaches and to LP-only solutions, and show that our method outperforms them during Resource Crunch. ",Running the Network Harder: Connection Provisioning under Resource Crunch " Based on the isospin-dependent quantum molecular dynamics model, finite-size scaling effects on nuclear liquid--gas phase transition probes are investigated by studying the de-excitation processes of six thermal sources of different sizes with the same initial density and similar $N/Z$. Using several probes including the total multiplicity derivative ($dM_{tot}/dT$), second moment parameter ($M_2$), intermediate mass fragment (IMF) multiplicity ($N_{IMF}$), Fisher's power-law exponent ($\tau$), and Ma's nuclear Zipf's law exponent ($\xi$), the relationship between the phase transition temperature and the source size has been established. It is observed that the phase transition temperatures obtained from the IMF multiplicity, Fisher's exponent, and Ma's nuclear Zipf's law exponent have a strong correlation with the source size. Moreover, by employing the finite-size scaling law, the critical temperature $T_c$ and the critical exponent $\nu$ have been obtained for infinite nuclear matter. ",Finite-size scaling phenomenon of nuclear liquid--gas phase transition probes " Dynamic parallelism on GPUs allows GPU threads to dynamically launch other GPU threads. It is useful in applications with nested parallelism, particularly where the amount of nested parallelism is irregular and cannot be predicted beforehand. However, prior works have shown that dynamic parallelism may impose a high performance penalty when a large number of small grids are launched. The large number of launches results in high launch latency due to congestion, and the small grid sizes result in hardware underutilization. To address this issue, we propose a compiler framework for optimizing the use of dynamic parallelism in applications with nested parallelism. The framework features three key optimizations: thresholding, coarsening, and aggregation. Thresholding involves launching a grid dynamically only if the number of child threads exceeds some threshold, and serializing the child threads in the parent thread otherwise. Coarsening involves executing the work of multiple thread blocks by a single coarsened block to amortize the common work across them. Aggregation involves combining multiple child grids into a single aggregated grid. Our evaluation shows that our compiler framework improves the performance of applications with nested parallelism by a geometric mean of 43.0x over applications that use dynamic parallelism, 8.7x over applications that do not use dynamic parallelism, and 3.6x over applications that use dynamic parallelism with aggregation alone as proposed in prior work. ",A Compiler Framework for Optimizing Dynamic Parallelism on GPUs " In this talk the question of what is the upper bound on the lightest supersymmetric Higgs mass, m_h is addressed. This question is relevant since experimental lower bounds on m_h might implement, in the near future, exclusion of supersymmetry. By imposing (perturbative) unification of the gauge couplings at some high scale \simgt 10^{17} GeV, we have found that for a top-quark mass M_t=175 GeV, and depending on the supersymmetric parameters, this bound can be as high as 205 GeV. ",What is the upper limit on the lightest supersymmetric Higgs mass? " We study some kinematical aspects of quantum fields on causal sets. In particular, we are interested in free scalar fields on a fixed background causal set. We present various results building up to the study of the entanglement entropy of de Sitter horizons using causal sets. We begin by obtaining causal set analogs of Green functions for this field. First we construct the retarded Green function in a Riemann normal neighborhood (RNN) of an arbitrary curved spacetime. Then, we show that in de Sitter and patches of anti-de Sitter spacetimes the construction can be done beyond the RNN. This allows us to construct the QFT vacuum on the causal set using the Sorkin-Johnston construction. We calculate the SJ vacuum on a causal set approximated by de Sitter spacetime, using numerical techniques. We find that the causal set SJ vacuum does not correspond to any of the known Mottola-Allen vacua of de Sitter spacetime. This has potential phenomenological consequences for early universe physics. Finally, we study the spacetime entanglement entropy for causal set de Sitter horizons. The entanglement entropy of de Sitter horizons is of particular interest. As in the case of nested causal diamonds in 2d Minkowski spacetime, we find that the causal set naturally gives a volume law of entropy, both for nested causal diamonds in 4d Minkowski spacetime as well as 2d and 4d de Sitter spacetimes. However, an area law emerges when the high frequency modes in the SJ spectrum are truncated. The choice of truncation turns out to be non-trivial and we end with several interesting questions. ",Aspects of Quantum Fields on Causal Sets " Magnesium aluminate scandium oxide (ScAlMgO4) is a promising lattice-matched substrate material for GaN- and ZnO-based optoelectronic devices. Yet, despite its clear advantages over substrates commonly used in heteroepitaxial growth, several fundamental properties of ScAlMgO4 remain unsettled. Here, we provide a comprehensive picture of its optical, electronic and structural properties by studying ScAlMgO4 single crystals grown by the Czochralski method. We use variable angle spectroscopic ellipsometry to determine complex in-plane and out-of-plane refractive indices in the range from 193 to 1690 nm. An oscillator-based model provides a phenomenological description of the ellipsometric spectra with excellent agreement over the entire range of wavelengths. For convenience, we supply the reader also with Cauchy formulas describing the real part of the anisotropic refractive index for wavelengths above 400 nm. Ab initio many-body perturbation theory modeling provides information about the electronic structure of ScAlMgO4, and successfully validated experimentally obtained refractive index values. Simulations also show exciton binding energy as large as a few hundred of meV, indicating ScAlMgO4 as a promising material for implementation in low-threshold, deep-UV lasing devices operating at room temperature. X-ray diffraction measurements confirm lattice constants of ScAlMgO4 previously reported, but in addition, reveal that dominant crystallographic planes (001) are mutually inclined by about 0.009{\deg}. In view of our work, ScAlMgO4 is a highly transparent, low refractive index, birefringent material similar to a sapphire, but with a much more favorable lattice constant and simpler processing. ","Optical, electronic and structural properties of ScAlMgO4" " We show that the discrete time quantum walk on the Boolean hypercube of dimension $n$ has a strong dispersion property: if the walk is started in one vertex, then the probability of the walker being at any particular vertex after $O(n)$ steps is of an order $O(1.4818^{-n})$. This improves over the known mixing results for this quantum walk which show that the probability distribution after $O(n)$ steps is close to uniform but do not show that the probability is small for every vertex. A rigorous proof of this result involves an intricate argument about analytic properties of Bessel functions. ",Strong dispersion property for the quantum walk on the hypercube " The Reynolds-averaged Navier-Stokes (RANS) equations for steady-state assessment of incompressible turbulent flows remain the workhorse for practical computational fluid dynamics (CFD) applications. Consequently, improvements in speed or accuracy have the potential to affect a diverse range of applications. We introduce a machine learning framework for the {surrogate modeling of steady-state turbulent eddy viscosities for RANS simulations, given the initial conditions. This modeling strategy} is assessed for parametric interpolation, while numerically solving for the pressure and velocity equations to steady state, thus representing a framework that is hybridized with machine learning. We achieve {competitive} steady-state results with a significant reduction in solution time when compared to those obtained by the Spalart-Allmaras one-equation model. This is because the proposed methodology allows for considerably larger relaxation factors for the steady-state velocity and pressure solvers. Our assessments are made for a backward-facing step with considerable mesh anisotropy and separation to represent a practical CFD application. For test experiments with \textcolor{black}{either} varying inlet velocity conditions or step heights we see time-to-solution reductions around a factor of 5. The results represent an opportunity for the rapid exploration of parameter spaces that prove prohibitive when utilizing turbulence closure models with multiple coupled partial differential equations. \blfootnote{Code available publicly at \texttt{https://github.com/argonne-lcf/TensorFlowFoam}}. ",A turbulent eddy-viscosity surrogate modeling framework for Reynolds-Averaged Navier-Stokes simulations " Aims: We probe the radiatively-efficient, hot wind feedback mode in two nearby luminous unobscured (type 1) AGN from the Close AGN Reference Survey (CARS), which show intriguing kpc-scale arc-like features of extended [OIII] ionized gas as mapped with VLT-MUSE. We aimed to detect hot gas bubbles that would indicate the existence of powerful, galaxy-scale outflows in our targets, HE 0227-0931 and HE 0351+0240, from deep (200 ks) Chandra observations. Methods: By measuring the spatial and spectral properties of the extended X-ray emission and comparing with the sub kpc-scale IFU data, we are able to constrain feedback scenarios and directly test if the ionized gas is due to a shocked wind. Results: No extended hot gas emission on kpc-scales was detected. Unless the ambient medium density is low ($n_{H}\sim~1$ cm$^{-3}$ at 100 pc), the inferred upper limits on the extended X-ray luminosities are well below what is expected from theoretical models at matching AGN luminosities. Conclusions: We conclude that the highly-ionized gas structures on kpc scales are not inflated by a hot outflow in either target, and instead are likely caused by photo-ionization of pre-existing gas streams of different origins. Our non-detections suggest that extended X-ray emission from an AGN-driven wind is not universal, and may lead to conflicts with current theoretical predictions. ",The Close AGN Reference Survey (CARS): No evidence of galaxy-scale hot outflows in two nearby AGN " For a simple model of price-responsive demand, we consider a deregulated electricity marketplace wherein the grid (ISO, retailer-distributor) accepts bids per-unit supply from generators (simplified herein neither to consider start-up/ramp-up expenses nor day-ahead or shorter-term load following) which are then averaged (by supply allocations via an economic dispatch) to a common ""clearing"" price borne by customers (irrespective of variations in transmission/distribution or generation prices), i.e., the ISO does not compensate generators based on their marginal costs. Rather, the ISO provides sufficient information for generators to sensibly adjust their bids. Notwithstanding our idealizations, the dispatch dynamics are complex. For a simple benchmark power system, we find a price-symmetric Nash equilibrium through numerical experiments. ",Generation bidding game with flexible demand " In this paper, we study complete Vacuum Static Spaces. A complete classification of 3-dimensional complete Vacuum Static Spaces with non-negative scalar curvature and constant squared norm of Ricci curvature tensor is given by making use of the generalized maximum principle. ",$3$-dimensional complete vacuum static spaces " The $e^+ e^- \to K^0_{S}K^0_{L}$ cross section has been measured in the center-of-mass energy range 1004--1060 MeV at 25 energy points using $6.1 \times 10^5$ events with $K^0_{S}\to \pi^+\pi^-$ decay. The analysis is based on 5.9 pb$^{-1}$ of an integrated luminosity collected with the CMD-3 detector at the VEPP-2000 $e^+ e^-$ collider. To obtain $\phi(1020)$ meson parameters the measured cross section is approximated according to the Vector Meson Dominance model as a sum of the $\rho, \omega, \phi$-like amplitudes and their excitations. This is the most precise measurement of the $e^+ e^- \to K^0_{S}K^0_{L}$ cross section with a 1.8\% systematic uncertainty. ",Study of the process $e^+ e^- \to K^0_{S}K^0_{L}$ in the center-of-mass energy range 1004--1060 MeV with the CMD-3 detector at the VEPP-2000 $e^+ e^-$ collider " Tuning topological and magnetic properties of materials by applying an electric field is widely used in spintronics. In this work, we find a topological phase transition from topologically trivial to nontrivial states at an external electric field of about 0.1 V/A in MnBi$_2$Te$_4$ monolayer that is a topologically trivial ferromagnetic semiconductor. It is shown that when electric field increases from 0 to 0.15 V/A, the magnetic anisotropy energy (MAE) increases from about 0.1 to 6.3 meV, and the Curie temperature Tc increases from 13 to about 61 K. The increased MAE mainly comes from the enhanced spin-orbit coupling due to the applied electric field. The enhanced Tc can be understood from the enhanced $p$-$d$ hybridization and decreased energy difference between $p$ orbitals of Te atoms and $d$ orbitals of Mn atoms. Moreover, we propose two novel Janus materials MnBi$_2$Se$_2$Te$_2$ and MnBi$_2$S$_2$Te$_2$ monolayers with different internal electric polarizations, which can realize quantum anomalous Hall effect (QAHE) with Chern numbers $C$=1 and $C$=2, respectively. Our study not only exposes the electric field induced exotic properties of MnBi2Te4 monolayer, but also proposes novel materials to realize QAHE in ferromagnetic Janus semiconductors with electric polarization. ",Electric field induced topological phase transition and large enhancements of spin-orbit coupling and Curie temperature in two-dimensional ferromagnetic semiconductors " Several applications in astrophysics require adequately resolving many physical and temporal scales which vary over several orders of magnitude. Adaptive mesh refinement techniques address this problem effectively but often result in constrained strong scaling performance. The ParalleX execution model is an experimental execution model that aims to expose new forms of program parallelism and eliminate any global barriers present in a scaling-impaired application such as adaptive mesh refinement. We present two astrophysics applications using the ParalleX execution model: a tabulated equation of state component for neutron star evolutions and a cosmology model evolution. Performance and strong scaling results from both simulations are presented. The tabulated equation of state data are distributed with transparent access over the nodes of the cluster. This allows seamless overlapping of computation with the latencies introduced by the remote access to the table. Because of the expected size increases to the equation of state table, this type of table partitioning for neutron star simulations is essential while the implementation is greatly simplified by ParalleX semantics. ",Adaptive Mesh Refinement for Astrophysics Applications with ParalleX " The experimental results relevant for the understanding of the microscopic dynamics in liquid metals are reviewed, with special regards to the ones achieved in the last two decades. Inelastic Neutron Scattering played a major role since the development of neutron facilities in the sixties. The last ten years, however, saw the development of third generation radiation sources, which opened the possibility of performing Inelastic Scattering with X rays, thus disclosing previously unaccessible energy-momentum regions. The purely coherent response of X rays, moreover, combined with the mixed coherent/incoherent response typical of neutron scattering, provides enormous potentialities to disentangle aspects related to the collectivity of motion from the single particle dynamics. If the last twenty years saw major experimental developments, on the theoretical side fresh ideas came up to the side of the most traditional and established theories. Beside the raw experimental results, therefore, we review models and theoretical approaches for the description of microscopic dynamics over different length-scales, from the hydrodynamic region down to the single particle regime, walking the perilous and sometimes uncharted path of the generalized hydrodynamics extension. Approaches peculiar of conductive systems, based on the ionic plasma theory, are also considered, as well as kinetic and mode coupling theory applied to hard sphere systems, which turn out to mimic with remarkable detail the atomic dynamics of liquid metals. Finally, cutting edges issues and open problems, such as the ultimate origin of the anomalous acoustic dispersion or the relevance of transport properties of a conductive systems in ruling the ionic dynamic structure factor are discussed. ",Microscopic dynamics in liquid metals: the experimental point of view " In this report we discuss the organization of different levels of nature and the corresponding space-time structures by the consideration of a particular problem of time irreversibility. The fundamental time irreversibility problem consists in the following: how to reconcile the time-reversible microscopic dynamics and the irreversible macroscopic one. The recently proposed functional formulation of mechanics is aimed to solve this problem. The basic concept of this formulation is not a material point and a trajectory, like in the traditional formulation of mechanics, but a probability density function. Even if we deal with a single particle (not with an ensemble of particles), we describe its state as a probability density function. We justify this approach using measurement theory. A particular problem in the framework of the irreversibility problem is the derivation of the Boltzmann kinetic equation from the equations of microscopic dynamics. We propose a procedure for obtaining the Boltzmann equation from the Liouville equation based on the BBGKY hierarchy, the recently proposed functional formulation of classical mechanics, and the distinguishing between two scales of space-time, i.e., macro- and microscale. The notion of a space-time structure is introduced. It takes into account not only the space-time itself (i.e., a pseudo-Riemannian manifold), but also a characteristic length and time. The space-time structures form a hierarchy in sense that the initial values for the processes on the microscopic space-time structure (interactions of the particles) are assigned from the processes on the macroscopic one (kinetic phenomena). ","Hierarchy of space-time structures, Boltzmann equation, and functional mechanics" " We present the results of determining the parameters of the spiral arms of the Galaxy using the stars Gaia DR3, whose absolute magnitude is $M_G$ < 4, and which allow tracing spiral arms at large distances from the Sun. As tracers of spiral arms, we use the centroids of stellar spherical regions with a radius of 0.5 kpc, in which the deformation velocities along the coordinate axis R are insignificant. These kinematic tracers cover the Galactic plane within the Galactocentric coordinate ranges 140{\deg} < ${\theta}$ < 220{\deg} and 4 kpc < R < 14 kpc. The numerical values of the pitch angles of the spirals and their Galactocentric distances to the point of intersection of the spiral with the direction of the Galactic center - the Sun are in good agreement with the results of other authors. By extrapolating beyond the data we have, we present a schematic four-arm global pattern, consisting of the Scutum-Centaurus, Sagitarius-Carina, Perseus, Norma-Outer arms, as well as the local arm Orion. The uncertainties of the determined spiral parameters confirm that the structures identified are not false, but are reliable from the statistical point of view. ",Determining the parameters of the spiral arms of the Galaxy from kinematic tracers based on Gaia DR3 data " Using 2-d U(1) lattice gauge theory we study two definitions of the topological charge constructed from a generalized Villain action and analyze the implementation of the index theorem based on the overlap Dirac operator. One of the two definitions expresses the topological charge as a sum of the Villain variables and treats charge conjugation symmetry exactly, making it particularly useful for studying related physics. Our numerical analysis establishes that for both topological charge definitions the index theorem becomes exact quickly towards the continuum limit. ",Topology and index theorem with a generalized Villain lattice action -- a test in 2d " The two-dimensional q-state Potts model is subjected to a Z_q symmetric disorder that allows for the existence of a Nishimori line. At q=2, this model coincides with the +/- J random-bond Ising model. For q>2, apart from the usual pure and zero-temperature fixed points, the ferro/paramagnetic phase boundary is controlled by two critical fixed points: a weak disorder point, whose universality class is that of the ferromagnetic bond-disordered Potts model, and a strong disorder point which generalizes the usual Nishimori point. We numerically study the case q=3, tracing out the phase diagram and precisely determining the critical exponents. The universality class of the Nishimori point is inconsistent with percolation on Potts clusters. ",Phase diagram and critical exponents of a Potts gauge glass " The transition quadrupole moments, $Q_{\rm t}$, of four weakly populated collective bands up to spin $\sim$ $65\hbar$ in $^{157,158}$Er have been measured to be ${\sim}11 {\rm eb}$ demonstrating that these sequences are associated with large deformations. However, the data are inconsistent with calculated values from cranked Nilsson-Strutinsky calculations that predict the lowest energy triaxial shape to be associated with rotation about the short principal axis. The data appear to favor either a stable triaxial shape rotating about the intermediate axis or, alternatively, a triaxial shape with larger deformation rotating about the short axis. These new results challenge the present understanding of triaxiality in nuclei. ",Quadrupole Moments of Collective Structures up to Spin $\sim$ $65\hbar$ in $^{157}$Er and $^{158}$Er: A Challenge for Understanding Triaxiality in Nuclei " We present griz light curves of 146 spectroscopically confirmed Type Ia Supernovae ($0.03 < z <0.65$) discovered during the first 1.5 years of the Pan-STARRS1 Medium Deep Survey. The Pan-STARRS1 natural photometric system is determined by a combination of on-site measurements of the instrument response function and observations of spectrophotometric standard stars. We find that the systematic uncertainties in the photometric system are currently 1.2\% without accounting for the uncertainty in the HST Calspec definition of the AB system. A Hubble diagram is constructed with a subset of 113 out of 146 SNe Ia that pass our light curve quality cuts. The cosmological fit to 310 SNe Ia (113 PS1 SNe Ia + 222 light curves from 197 low-z SNe Ia), using only SNe and assuming a constant dark energy equation of state and flatness, yields $w=-1.120^{+0.360}_{-0.206}\textrm{(Stat)} ^{+0.269}_{-0.291}\textrm{(Sys)}$. When combined with BAO+CMB(Planck)+$H_0$, the analysis yields $\Omega_{\rm M}=0.280^{+0.013}_{-0.012}$ and $w=-1.166^{+0.072}_{-0.069}$ including all identified systematics (see also Scolnic et al. 2014). The value of $w$ is inconsistent with the cosmological constant value of $-1$ at the 2.3$\sigma$ level. Tension endures after removing either the BAO or the $H_0$ constraint, though it is strongest when including the $H_0$ constraint. If we include WMAP9 CMB constraints instead of those from Planck, we find $w=-1.124^{+0.083}_{-0.065}$, which diminishes the discord to $<2\sigma$. We cannot conclude whether the tension with flat $\Lambda$CDM is a feature of dark energy, new physics, or a combination of chance and systematic errors. The full Pan-STARRS1 supernova sample with $\sim\!\!$3 times as many SNe should provide more conclusive results. ",Cosmological Constraints from Measurements of Type Ia Supernovae discovered during the first 1.5 years of the Pan-STARRS1 Survey " This paper presents 1.4-GHz radio continuum observations of 15 very extended radio galaxies. These sources are so large that most interferometers lose partly their structure and total flux density. Therefore, single-dish detections are required to fill in the central (u,v) gap of interferometric data and obtain reliable spectral index patterns across the structures, and thus also an integrated radio continuum spectrum. We have obtained such 1.4-GHz maps with the 100-m Effelsberg telescope and combined them with the corresponding maps available from the NVSS. The aggregated data allow us to produce high-quality images, which can be used to obtain physical parameters of the mapped sources. The combined images reveal in many cases extended low surface-brightness cocoons. ",1.4-GHz observations of extended giant radio galaxies Perturbative probability conservation provides a strong constraint on the presence of new interactions of the Higgs boson. In this work we consider CP violating Higgs interactions in conjunction with unitarity constraints in the gauge-Higgs and fermion-Higgs sectors. Injecting signal strength measurements of the recently discovered Higgs boson allows us to make concrete and correlated predictions of how CP-violation in the Higgs sector can be directly constrained through collider searches for either characteristic new states or tell-tale enhancements in multi-Higgs processes. ,"Perturbative Higgs CP violation, unitarity and phenomenology" " The structural and magnetic phase transitions of the ternary iron arsenides SrFe2As2 and EuFe2As2 were studied by temperature-dependent x-ray powder diffraction and 57-Fe Moessbauer spectroscopy. Both compounds crystallize in the tetragonal ThCr2Si2-type structure at room temperature and exhibit displacive structural transitions at 203 K (SrFe2As2) or 190 K (EuFe2As2) to orthorhombic lattice symmetry in agreement with the group-subgroup relationship between I4/mmm and Fmmm. 57-Fe Moessbauer spectroscopy experiments with SrFe2 As2 show full hyperfine field splitting below the phase transition temperature (8.91(1) T at 4.2 K). Order parameters were extracted from detailed measurements of the lattice parameters and fitted to a simple power law. We find a relation between the critical exponents and the transition temperatures for AFe2As2 compounds, which shows that the transition of BaFe2As2 is indeed more continuous than the transition of SrFe2As2 but it remains second order even in the latter case. ",Structural and magnetic phase transitions in the ternary iron arsenides SrFe2As2 and EuFe2As2 " Abundances and energy spectra of cosmic ray nuclei are being measured with high accuracy by the AMS experiment. These observations can provide tight constraints to the propagation models of galactic cosmic rays. In the view of the release of these data, I present an evaluation of the model uncertainties associated to the cross-sections for secondary production of Li-Be-B nuclei in cosmic rays. I discuss the role of cross section uncertainties in the calculation of the boron-to-carbon and beryllium-to-boron ratios, as well as their impact in the determination of the cosmic-ray transport parameters. ",Fragmentation cross-sections and model uncertainties in Cosmic Ray propagation physics " We study the edge and surface theories of topological insulators from the perspective of anomalies and identify a novel Z2-anomaly associated with charge conservation. The anomaly is manifested through a 2-point correlation function involving creation and annihilation operators on two decoupled boundaries. Although charge conservation on each boundary requires this quantity to vanish, we find that it diverges. A corollary result is that under an insertion of a flux quantum the ground state evolves to an exactly orthogonal state independent of the rate at which the flux is inserted. The anomaly persists in the presence of disorder and imposes sharp restrictions on possible low energy theories. Being formulated in a many-body, field theoretical language, the anomaly allows to test the robustness of topological insulators to interactions in a concise way. ",The Z2-anomaly and boundaries of topological insulators " The plasma equilibrium in a linear trap at $\beta\approx 1$ (or above the mirror-instability threshold) under the topology-conservation constraint evolves into a kind of diamagnetic ""bubble"". This can take two forms: either the plasma body greatly expands in radius while containing the same magnetic flux, or, if the plasma radius is limited, the plasma distribution across flux-tubes changes, so that the same cross-section contains a greatly reduced flux. If the magnetic field of the trap is quasi-uniform around its minimum, the bubble can be made roughly cylindrical, with radius much larger than the radius of the corresponding vacuum flux-tube, and with non-paraxial ends. Then the effective mirror ratio of the diamagnetic trap becomes very large, but the cross-field transport increases. The confinement time can be found from solution of the system of equilibrium and transport equations and is shown to be $\tau_E\approx\sqrt{\tau_\parallel\tau_\perp}$. If the cross-field confinement is not too degraded by turbulence, this estimate in principle allows construction of relatively compact fusion reactors with lengths in the range of a few tens of meters. In many ways the described here diamagnetic confinement and the corresponding reactor parameters are similar to those claimed by the FRCs. ","Diamagnetic ""bubble"" equilibria in linear traps" " This paper is concerned with strong convergence of the truncated Euler-Maruyama scheme for neutral stochastic differential delay equations driven by Brownian motion and pure jumps respectively. Under local Lipschitz condition, convergence rates of the truncated EM scheme are given. ",Convergence rates of truncated EM scheme for NSDDEs " The microturbulent approximation of turbulent motions is widely used in radiative transfer calculations. Mainly motivated by its simple computational application it is probably in many cases an oversimplified treatment of the dynamical processes involved. This aspect is in particular important in the analysis of maser lines, since the strong amplification of radiation leads to a sensitive dependence of the radiation field on the overall velocity structure. To demonstrate the influence of large scale motions on the formation of maser lines we present a simple stochastic model which takes velocity correlations into account. For a quantitative analysis of correlation effects, we generate in a Monte Carlo simulation individual realizations of a turbulent velocity field along a line of sight. Depending on the size of the velocity correlation length we find huge deviations between the resulting random profiles in respect of line shape, intensity and position of single spectral components. Finally, we simulate the emission of extended maser sources. A qualitative comparison with observed masers associated with star forming regions shows that our model can reproduce the observed general spectral characteristics. We also investigate shortly, how the spectra are effected when a systematic velocity field (simulating expansion) is superposed on the fluctuations. Our results convincingly demonstrate that hydrodynamical motions are of great importance for the understanding of cosmic masers. ",Effects of correlated turbulent velocity fields on the formation of maser lines " We introduce a class of models for multidimensional control problems which we call skip-free Markov decision processes on trees. We describe and analyse an algorithm applicable to Markov decision processes of this type that are skip-free in the negative direction. Starting with the finite average cost case, we show that the algorithm combines the advantages of both value iteration and policy iteration -- it is guaranteed to converge to an optimal policy and optimal value function after a finite number of iterations but the computational effort required for each iteration step is comparable with that for value iteration. We show that the algorithm can also be used to solve discounted cost models and continuous time models, and that a suitably modified algorithm can be used to solve communicating models. ",Models and algorithms for skip-free Markov decision processes on trees " We analyze the dynamics of finite width effects in wide but finite feature learning neural networks. Starting from a dynamical mean field theory description of infinite width deep neural network kernel and prediction dynamics, we provide a characterization of the $\mathcal{O}(1/\sqrt{\text{width}})$ fluctuations of the DMFT order parameters over random initializations of the network weights. Our results, while perturbative in width, unlike prior analyses, are non-perturbative in the strength of feature learning. In the lazy limit of network training, all kernels are random but static in time and the prediction variance has a universal form. However, in the rich, feature learning regime, the fluctuations of the kernels and predictions are dynamically coupled with a variance that can be computed self-consistently. In two layer networks, we show how feature learning can dynamically reduce the variance of the final tangent kernel and final network predictions. We also show how initialization variance can slow down online learning in wide but finite networks. In deeper networks, kernel variance can dramatically accumulate through subsequent layers at large feature learning strengths, but feature learning continues to improve the signal-to-noise ratio of the feature kernels. In discrete time, we demonstrate that large learning rate phenomena such as edge of stability effects can be well captured by infinite width dynamics and that initialization variance can decrease dynamically. For CNNs trained on CIFAR-10, we empirically find significant corrections to both the bias and variance of network dynamics due to finite width. ",Dynamics of Finite Width Kernel and Prediction Fluctuations in Mean Field Neural Networks " Multisensor track-to-track fusion for target tracking involves two primary operations: track association and estimation fusion. For estimation fusion, lossless measurement transformation of sensor measurements has been proposed for single target tracking. In this paper, we investigate track association which is a fundamental and important problem for multitarget tracking. First, since the optimal track association problem is a multi-dimensional assignment (MDA) problem, we demonstrate that MDA-based data association (with and without prior track information) using linear transformations of track measurements is lossless, and is equivalent to that using raw track measurements. Second, recent superior scalability and performance of belief propagation (BP) algorithms enable new real-time applications of multitarget tracking with resource-limited devices. Thus, we present a BP-based multisensor track association method with transformed measurements and show that it is equivalent to that with raw measurements. Third, considering communication constraints, it is more beneficial for local sensors to send in compressed data. Two analytical lossless transformations for track association are provided, and it is shown that their communication requirements from each sensor to the fusion center are less than those of fusion with raw track measurements. Numerical examples for tracking an unknown number of targets verify that track association with transformed track measurements has the same performance as that with raw measurements and requires fewer communication bandwidths. ",On Communication-Efficient Multisensor Track Association via Measurement Transformation (Extended Version) " In many regular cases, there exists a (properly defined) limit of iterations of a function in several real variables, and this limit satisfies the functional equation (1-z)f(x)=f(f(xz)(1-z)/z); here z is a scalar and x is a vector. This is a special case of a well-known translation equation. In this paper we present a complete solution to this functional equation in case f is a continuous function on a single point compactification of a 2-dimensional real vector space. It appears that, up to conjugation by a homogeneous continuous function, there are exactly four solutions. Further, in a 1-dimensional case we present a solution with no regularity assumptions on f. ",Multi-variable translation equation which arises from homothety We study the relation between neutron removal cross section ($\sigma_{-N}$) and neutron skin thickness for finite neutron rich nuclei using the statistical abrasion ablation (SAA) model. Different sizes of neutron skin are obtained by adjusting the diffuseness parameter of neutrons in the Fermi distribution. It is demonstrated that there is a good linear correlation between $\sigma_{-N}$ and the neutron skin thickness for neutron rich nuclei. Further analysis suggests that the relative increase of neutron removal cross section could be used as a quantitative measure for the neutron skin thickness in neutron rich nuclei. ,Neutron removal cross section as a measure of neutron skin " The spin of a single electron confined in a semiconductor quantum dot is a natural qubit candidate. Fundamental building blocks of spin-based quantum computing have been demonstrated in double quantum dots with significant spin-orbit coupling. Here, we show that spin-orbit-coupled double quantum dots can be categorised in six classes, according to a partitioning of the multi-dimensional space of their $g$-tensors. The class determines physical characteristics of the double dot, i.e., features in transport, spectroscopy and coherence measurements, as well as qubit control, shuttling, and readout experiments. In particular, we predict that the spin physics is highly simplified due to pseudospin conservation, whenever the external magnetic field is pointing to special directions (`magic directions'), where the number of special directions is determined by the class. We also analyze the existence and relevance of magic loops in the space of magnetic-field directions, corresponding to equal local Zeeman splittings. These results present an important step toward precise interpretation and efficient design of spin-based quantum computing experiments in materials with strong spin-orbit coupling. ",Classification and magic magnetic-field directions for spin-orbit-coupled double quantum dots " We study several properties of blazars detected in the gamma-ray energy range by comparing the EGRET sources with a sample of radio blazars which can be considered possible gamma-ray candidates. We define three classes: non-gamma-ray blazars, blazars with quasi-steady gamma-ray emission, and gamma-ray blazars with substantial activity level. By combining the information of detected and candidate AGNs, we characterise the blazar activity, including the discovery of a region of consistency between the gamma-ray flaring duty-cycle and the recurrence time between flares. We also find a possible relation between the activity index of FSRQs and their black hole mass. ","The Duty-cycle of Gamma-ray Blazars: a New Approach, New Results" In this paper we prove transference inequalities for regular and uniform Diophantine exponents in the weighted setting. Our results generalize the corresponding inequalities that exist in the `non-weighted' case. ,Transference theorems for Diophantine approximation with weights " The CP Violating asymmetry in Bs mixing (beta_s) is one of the most promising measurements where physics beyond the Standard Model could be revealed. As such, analyses need to be subjected to great scrutiny. The mode Bs -> J/psi\phi has been used, and the mode Bs -> \phi \phi proposed for future measurements. These modes both have two vector particles in the final state and thus angular analyses must be used to disentangle the contributions from CP+ and CP- configurations. The angular distributions, however, could be distorted by the presence of S-waves masquerading as low mass K+K- pairs, that could result in erroneous values of beta_s. The S-waves could well be the result of a final state formed from an s-quark anti-s-quark pair in a 0+ spin-parity state, such as the f0(980) meson. Data driven and theoretical estimates of the Bs decay rate into the CP+ final state J/psi f0(980) are given, when f0 -> pi+pi-. The S-wave contribution in J\psi\phi should be taken into account when determining beta_s by including a K+K- S-wave amplitude in the fit. This may change the central value of current results and will also increase the statistical uncertainty. Importantly, the J/psi f0(980) mode has been suggested as an alternative channel for measuring beta_s. ",S-waves and the extraction of beta_s " In this work we explore the problem of answering a set of sum queries under Differential Privacy. This is a little understood, non-trivial problem especially in the case of numerical domains. We show that traditional techniques from the literature are not always the best choice and a more rigorous approach is necessary to develop low error algorithms. ",Answering Summation Queries for Numerical Attributes under Differential Privacy " Right-handed (RH) Majorana neutrinos play a crucial role in understanding the origin of neutrino mass, the nature of dark matter and the mechanism of matter-antimatter asymmetry. In this work, we investigate the observability of heavy Majorana neutrino through the top quark neutrinoless double beta decay process $t \to b \ell^+ \ell^+ j j$ at hadron colliders. By performing detector level simulation, we demonstrate that our method can give stronger limits on the light-heavy neutrino mixing parameters $|V_{eN, \mu N}|$ in the mass range of 15 GeV $< m_N <$ 80 GeV than other existing collider bounds. ",Top quark as a probe of heavy Majorana neutrino at the LHC and future collider " An ansatz is proposed for heptagon relation, that is, algebraic imitation of five-dimensional Pachner move 4--3. Our relation is realized in terms of matrices acting in a direct sum of one-dimensional linear spaces corresponding to 4-faces. ",Heptagon relation in a direct sum " We study the origin of the stellar $\alpha$-element-to-iron abundance ratio, $[\alpha/\mathrm{Fe}]_{\ast}$, of present-day central galaxies, using cosmological, hydrodynamical simulations from the Evolution and Assembly of GaLaxies and their Environments (EAGLE) project. For galaxies with stellar masses of $M_{\ast} > 10^{10.5}$ M$_{\odot}$, $[\alpha/\mathrm{Fe}]_{\ast}$ increases with increasing galaxy stellar mass and age. These trends are in good agreement with observations of early-type galaxies, and are consistent with a `downsizing' galaxy formation scenario: more massive galaxies have formed the bulk of their stars earlier and more rapidly, hence from an interstellar medium that was mostly $\alpha$-enriched by massive stars. In the absence of feedback from active galactic nuclei (AGN), however, $[\alpha/\mathrm{Fe}]_{\ast}$ in $M_{\ast} > 10^{10.5}$ M$_{\odot}$ galaxies is roughly constant with stellar mass and decreases with mean stellar age, extending the trends found for lower-mass galaxies in both simulations with and without AGN. We conclude that AGN feedback can account for the $\alpha$-enhancement of massive galaxies, as it suppresses their star formation, quenching more massive galaxies at earlier times, thereby preventing the iron from longer-lived intermediate-mass stars (supernova Type Ia) from being incorporated into younger stars. ",The origin of the $\alpha$-enhancement of massive galaxies " A key challenge in training neural networks for a given medical imaging task is often the difficulty of obtaining a sufficient number of manually labeled examples. In contrast, textual imaging reports, which are often readily available in medical records, contain rich but unstructured interpretations written by experts as part of standard clinical practice. We propose using these textual reports as a form of weak supervision to improve the image interpretation performance of a neural network without requiring additional manually labeled examples. We use an image-text matching task to train a feature extractor and then fine-tune it in a transfer learning setting for a supervised task using a small labeled dataset. The end result is a neural network that automatically interprets imagery without requiring textual reports during inference. This approach can be applied to any task for which text-image pairs are readily available. We evaluate our method on three classification tasks and find consistent performance improvements, reducing the need for labeled data by 67%-98%. ",Contrastive Cross-Modal Pre-Training: A General Strategy for Small Sample Medical Imaging " We have analyzed the observed image separations of the gravitationally lensed images of QSOs for a possible correlation with the source redshift. Contrary to the previously noted anti-correlation based on a smaller data set, no correlation is found for the currently available data. We have calculated the average image separations of the lensed QSOs as a function of source redshifts, for isothermal spheres with cores in a flat universe, taking into account the amplification bias caused by lensing. The shape of the distribution of average image separation as a function of redshift is very robust and is insensitive to most model parameters. Observations are found to be roughly consistent with the theoretical results for models which assume the lens distribution to be (i) Schechter luminosity function which, however, can not produce images with large separation and (ii) the mass condensations in a cold dark matter universe, as given by the Press-Schechter theory if an upper limit of 1-7$\times 10^{13}$ M$\odot$ is assumed on the mass of the condensations. ",Angular separations of the lensed QSO images " We study the problem of detecting adverse drug events in electronic healthcare records. The challenge in this work is to aggregate heterogeneous data types involving diagnosis codes, drug codes, as well as lab measurements. An earlier framework proposed for the same problem demonstrated promising predictive performance for the random forest classifier by using only lab measurements as data features. We extend this framework, by additionally including diagnosis and drug prescription codes, concurrently. In addition, we employ a recursive feature selection mechanism on top, that extracts the top-k most important features. Our experimental evaluation on five medical datasets of adverse drug events and six different classifiers, suggests that the integration of these additional features provides substantial and statistically significant improvements in terms of AUC, while employing medically relevant features. ",Aggregate-Eliminate-Predict: Detecting Adverse Drug Events from Heterogeneous Electronic Health Records " We discuss the occurrence of cracking in charged anisotropic polytropes with generalized polytropic equation of state through two different assumptions; (i) by carrying out local density perturbations under conformally flat condition (ii) by perturbing anisotropy, polytropic index and charge parameters. For this purpose, we consider two different definitions of polytropes exist in literature. We conclude that under local density perturbations scheme cracking does not appears in both types of polytropes and stable configuration are observed, while with second kind of perturbation cracking appears in both types of polytropes under certain conditions. ",Cracking of Charged Polytropes with Generalized Polytropic Equation of State " The nucleon properties and structure should be modified by short-range correlations (SRC) among nucleons. By analyzing SRC ratio data, we extract the mass of nucleon in the SRC pair and the expected number of pn-SRC pairs in deuterium, under the assumption that the SRC nucleon mass is universal in different nuclei. The nucleon mass of a two-nucleon SRC pair is $m_{\text{SRC}}=852\pm 18$ MeV, and the number of pn-SRC pairs in deuterium is $n^{\text{d}}_{\text{SRC}}=0.021\pm 0.005$. The mass deficit of the strongly overlapping nucleon can be explained by the trace anomaly contribution to the mass in QCD or alternatively by the vacuum energy in the MIT bag model. ",The Mass of Short-Range Correlated Nucleon We show that the study of conductivity nonlinearity gives a possibility to determine the condition when the diffusion conductivity changes to the hopping one with increasing disorder. It is experimentally shown that the conductivity of single quantum well GaAs/InGaAs/GaAs heterostructures behaves like diffusive one down to value of order $10^{-2}e^2/h$. ,Nonohmic conductivity as a probe of crossover from diffusion to hopping in two dimensions " An extended thermal particle boundary condition is devised to more efficiently and accurately model laser-plasma interactions in overdense plasmas. Particle-in-cell simulations of such interactions require many particles per cell, and a large region of background plasma is often necessary to correctly mimic a semi-infinite plasma and avoid electron refluxing from a truncated plasma. For long-pulse lasers of many picoseconds, such constraints can become prohibitively expensive. Here, an extended particle boundary condition (absorber) is designed that instantaneously stops and re-emits energetic particles streaming toward the simulation boundary over a defined region, allowing sufficient time and space for a suitably cool return current to develop in the background plasma. Tunable parameters of the absorber are explained, and simulations using the absorber with a 3-ps laser are shown to accurately reproduce those of a causally separated boundary while requiring only 20% the number of particles. ",Extended particle absorber for efficient modeling of intense laser-solid interactions " A solution for Smale's 17th problem, for the case of systems with bounded degree was recently given. This solution, an algorithm computing approximate zeros of complex polynomial systems in average polynomial time, assumed infinite precision. In this paper we describe a finite-precision version of this algorithm. Our main result shows that this version works within the same time bounds and requires a precision which, on the average, amounts to a polynomial amount of bits in the mantissa of the intervening floating-point numbers. ",Fast Computation of Zeros of Polynomial Systems with Bounded Degree under Finite-precision " We present a crowdsourcing workflow to collect image annotations for visually similar synthetic categories without requiring experts. In animals, there is a direct link between taxonomy and visual similarity: e.g. a collie (type of dog) looks more similar to other collies (e.g. smooth collie) than a greyhound (another type of dog). However, in synthetic categories such as cars, objects with similar taxonomy can have very different appearance: e.g. a 2011 Ford F-150 Supercrew-HD looks the same as a 2011 Ford F-150 Supercrew-LL but very different from a 2011 Ford F-150 Supercrew-SVT. We introduce a graph based crowdsourcing algorithm to automatically group visually indistinguishable objects together. Using our workflow, we label 712,430 images by ~1,000 Amazon Mechanical Turk workers; resulting in the largest fine-grained visual dataset reported to date with 2,657 categories of cars annotated at 1/20th the cost of hiring experts. ",Scalable Annotation of Fine-Grained Categories Without Experts " We study the effect of superconducting fluctuations on the longitudinal and the transverse (Hall) conductivity in homogeneously disordered films. Our calculation is based on the Usadel equation in the real-time formulation. We adjust this approach to derive analytic expressions for the fluctuation corrections in the entire metallic part of the temperature-magnetic field phase diagram, including the effects of both classical and quantum fluctuations. This method allows us to obtain fluctuation corrections in a compact and effective way, establishing a direct connection between phenomenological and microscopic calculations. ",Fluctuation conductivity in disordered superconducting films " Bi-directional LSTMs are a powerful tool for text representation. On the other hand, they have been shown to suffer various limitations due to their sequential nature. We investigate an alternative LSTM structure for encoding text, which consists of a parallel state for each word. Recurrent steps are used to perform local and global information exchange between words simultaneously, rather than incremental reading of a sequence of words. Results on various classification and sequence labelling benchmarks show that the proposed model has strong representation power, giving highly competitive performances compared to stacked BiLSTM models with similar parameter numbers. ",Sentence-State LSTM for Text Representation " We propose a new flexible force field for water. The model in addition to the Lennard-Jones and electrostatic parameters, includes the flexibility of the OH bonds and angles. The parameters are selected to give the experimental values of the density and dielectric constant of water at at 1 bar at 240K and the dipole moment of minimum density. The FBA/epsilon reproduces the experimental values of structural, thermodynamic and the phase behavior of water in a wide range of temperatures with better accuracy than atomistic and other flexible models. We expect that this new approach would be suitable for studying water solutions. ","Flexible Bond and Angle, FBA/epsilon model of water" " Josephson junctions containing ferromagnetic layers have generated interest for application in cryogenic memory. In a junction containing both a magnetically hard fixed layer and soft free layer with carefully chosen thicknesses, the ground-state phase difference of the junction can be controllably switched between 0 and {\pi} by changing the relative orientation of the two ferromagnetic layers from antiparallel to parallel. This phase switching has been observed in junctions using Ni fixed layers and NiFe free layers. We present phase-sensitive measurements of such junctions in low-inductance symmetric SQUID loops which simplify analysis relative to our previous work. We confirm controllable 0 - {\pi} switching in junctions with 2.0 nm Ni fixed layers and 1.25 nm NiFe free layers across multiple devices and using two SQUID designs, expanding the phase diagram of known thicknesses that permit phase control. ",Phase Controllable Josephson Junctions for Cryogenic Memory " Polaris the Cepheid has been observed for centuries, presenting surprises and changing our view of Cepheids and stellar astrophysics, in general. Specifically, understanding Polaris helps anchor the Cepheid Leavitt law, but the distance must be measured precisely. The recent debate regarding the distance to Polaris has raised questions about its role in calibrating the Leavitt law and even its evolutionary status. In this work, I present new stellar evolution models of Cepheids to compare with previously measured CNO abundances, period change and angular diameter. Based on the comparison, I show that Polaris cannot be evolving along the first crossing of the Cepheid instability strip and cannot have evolved from a rapidly-rotating main sequence star. As such, Polaris must also be at least 118 pc away and pulsates in the first overtone, disagreeing with the recent results of Turner et al. (2013). ",Revisiting the fundamental properties of Cepheid Polaris using detailed stellar evolution models " Shear-induced vertical mixing in a stratified flow is a key ingredient of thermohaline circulation. We experimentally determine the vertical flux of momentum and density of a forced gravity current using high-resolution velocity and density measurements. A constant eddy viscosity model provides a poor description of the physics of mixing, but a Prandtl mixing length model relating momentum and density fluxes to mean velocity and density gradients works well. For $ \approx 0.08$ and $Re_\lambda \approx 100$, the mixing lengths are fairly constant, about the same magnitude, comparable to the turbulent shear length. ",Mixing in stratified gravity currents: Prandtl mixing length " We describe a methodology for the automatic identification of communities of practice from email logs within an organization. We use a betweeness centrality algorithm that can rapidly find communities within a graph representing information flows. We apply this algorithm to an email corpus of nearly one million messages collected over a two-month span, and show that the method is effective at identifying true communities, both formal and informal, within these scale-free graphs. This approach also enables the identification of leadership roles within the communities. These studies are complemented by a qualitative evaluation of the results in the field. ",Email as Spectroscopy: Automated Discovery of Community Structure within Organizations " Fermi can measure the spectral properties of gamma-ray bursts over a very large energy range and is opening a new window on the prompt emission of these energetic events. Localizations by the instruments on Fermi in combination with follow-up by Swift provide accurate positions for observations at longer wavelengths leading to the determination of redshifts, the true energy budget, host galaxy properties and facilitate comparison with pre-Fermi bursts. Multi-wavelength follow-up observations were performed on the afterglows of four bursts with high energy emission detected by Fermi/LAT : GRB090323, GRB090328, GRB090510 and GRB090902B. They were obtained in the optical/near-infrared bands with GROND mounted at the MPG/ESO 2.2m telescope and additionally of GRB090323 in the optical with the 2 m telescope in Tautenburg, Germany. Three of the events are classified as long bursts while GRB090510 is a well localized short GRB with GeV emission. In addition, host galaxies were detected for three of the four bursts. Spectroscopic follow-up was initiated with the VLT for GRB090328 and GRB090510. The afterglow observations in 7 bands are presented for all bursts and their host galaxies are investigated. Knowledge of the distance and the local dust extinction enables comparison of the afterglows of LAT-detected GRBs with the general sample. The spectroscopic redshifts of GRB090328 and GRB090510 were determined to be z=0.7354+/-0.0003 and z=0.903 +/- 0.001 and dust corrected star-formation rates of 4.8 Mdot yr^-1 and 0.60 M_dot yr^-1 were derived for their host galaxies, respectively. The afterglows of long bursts exhibit power-law decay indices alpha from less than 1 to ~2.3 and spectral indices (beta) values from 0.65 to ~1.2 which are fairly standard for GRB afterglows. Constraints are placed on the jet half opening angles of less than 2.1 deg to greater than 6.4 deg which allows limits to be placed on the beaming corrected energies. These range from less than 5x10^50 erg to the one of the highest values ever recorded, greater than 2.2x10^52 erg for GRB090902B, and are not consistent with a standard candle. The extremely energetic long Fermi bursts have optical afterglows which lie in the top half of the brightness distribution of all optical afterglows detected in the Swift era or even in the top 5% if incompleteness is considered. The properties of the host galaxies of these LAT detected bursts in terms of extinction, star formation rates and masses do not appear to differ from previous samples. ","Optical and near-infrared follow-up observations of four Fermi/LAT GRBs : Redshifts, afterglows, energetics and host galaxies" " Maximally symmetric curved-brane solutions are studied in dilatonic braneworld models which realise the self-tuning of the effective four-dimensional cosmological constant. It is found that no vacua in which the brane has de Sitter or anti-de Sitter geometry exist, unless one modifies the near-boundary asymptotics of the bulk fields. In the holographic dual picture, this corresponds to coupling the UV CFT to a curved metric (possibly with a defect). Alternatively, the same may be achieved in a flat-space QFT with suitable variable scalar sources. With these ingredients, it is found that maximally symmetric, positive and negative curvature solutions with a stabilised brane position generically exist. The space of such solutions is studied in two different types of realisations of the self-tuning framework. In some regimes we observe a large hierarchy between the curvature on the brane and the boundary UV CFT curvature. This is a dynamical effect due to the self-stabilisation mechanism. This setup provides an alternative route to realising de Sitter space in string theory. ",De Sitter and Anti-de Sitter branes in self-tuning models " We study explained variation under the additive hazards regression model for right-censored data. We consider different approaches for developing such a measure, and focus on one that estimates the proportion of variation in the failure time explained by the covariates. We study the properties of the measure both analytically, and through extensive simulations. We apply the measure to a well-known survival data set as well as the linked Surveillance, Epidemiology and End Results (SEER)-Medicare database for prediction of mortality in early-stage prostate cancer patients using high dimensional claims codes. ",Explained Variation under the Additive Hazards Model " A new shrinkage-based construction is developed for a compressible vector $\boldsymbol{x}\in\mathbb{R}^n$, for cases in which the components of $\xv$ are naturally associated with a tree structure. Important examples are when $\xv$ corresponds to the coefficients of a wavelet or block-DCT representation of data. The method we consider in detail, and for which numerical results are presented, is based on increments of a gamma process. However, we demonstrate that the general framework is appropriate for many other types of shrinkage priors, all within the L\'{e}vy process family, with the gamma process a special case. Bayesian inference is carried out by approximating the posterior with samples from an MCMC algorithm, as well as by constructing a heuristic variational approximation to the posterior. We also consider expectation-maximization (EM) for a MAP (point) solution. State-of-the-art results are manifested for compressive sensing and denoising applications, the latter with spiky (non-Gaussian) noise. ",Multiscale Shrinkage and L\'evy Processes " The manual implementation of distributed systems is an error-prone task because of the asynchronous interplay of components and the environment. Bounded synthesis automatically generates an implementation for the specification of the distributed system if one exists. So far, bounded synthesis for distributed systems does not utilize their asynchronous nature. Instead, concurrent behavior of components is encoded by all interleavings and only then checked against the specification. We close this gap by identifying true concurrency in synthesis of asynchronous distributed systems represented as Petri games. This defines when several interleavings can be subsumed by one true concurrent trace. Thereby, fewer and shorter verification problems have to be solved in each iteration of the bounded synthesis algorithm. For Petri games, experimental results show that our implementation using true concurrency outperforms the implementation based on checking all interleavings. ",Efficient Trace Encodings of Bounded Synthesis for Asynchronous Distributed Systems " Logo detection in real-world scene images is an important problem with applications in advertisement and marketing. Existing general-purpose object detection methods require large training data with annotations for every logo class. These methods do not satisfy the incremental demand of logo classes necessary for practical deployment since it is practically impossible to have such annotated data for new unseen logo. In this work, we develop an easy-to-implement query-based logo detection and localization system by employing a one-shot learning technique. Given an image of a query logo, our model searches for it within a given target image and predicts the possible location of the logo by estimating a binary segmentation mask. The proposed model consists of a conditional branch and a segmentation branch. The former gives a conditional latent representation of the given query logo which is combined with feature maps of the segmentation branch at multiple scales in order to find the matching position of the query logo in a target image, should it be present. Feature matching between the latent query representation and multi-scale feature maps of segmentation branch using simple concatenation operation followed by 1x1 convolution layer makes our model scale-invariant. Despite its simplicity, our query-based logo retrieval framework achieved superior performance in FlickrLogos-32 and TopLogos-10 dataset over different existing baselines. ",A Deep One-Shot Network for Query-based Logo Retrieval " In this paper, for sufficiently large $n$ we determine the Ramsey number $R(G,nH)$ where $G$ is a $k$-uniform hypergraph with the maximum independent set that intersects each of the edges in $k-1$ vertices and $H$ is a $k$-uniform hypergraph with a vertex so that the hypergraph induced by the edges containing this vertex is a star. There are several examples for such $G$ and $H$, among them are any disjoint union of $k$-uniform hypergraphs involving loose paths, loose cycles, tight paths, tight cycles with a multiple of $k$ edges, stars, Kneser hypergraphs and complete $k$-uniform $k$-partite hypergraphs for $G$ and linear hypergraphs for $H$. As an application, $R(mG,nH)$ is determined where $m$ or $n$ is large and $G$ and $H$ are either loose paths, loose cycles, tight paths, or stars. Also, $R(G,nH)$ is determined when $G$ is a bipartite graph with a matching saturating one of its color classes and $H$ is an arbitrary graph for sufficiently large $n$. Moreover, some bounds are given for $R(mG,nH)$ which allow us to determine this Ramsey number when $m\geq n$ and $G$ and $H$, $(|V(G)|\geq |V(H)|)$, are 3-uniform loose paths or cycles, $k$-uniform loose paths or cycles with at most 4 edges and $k$-uniform stars with 3 edges. ",Ramsey numbers for multiple copies of hypergraphs " We present a study of 9 galaxy groups with evidence for non-Gaussianity in their velocity distributions out to 4R200. This sample is taken from 57 groups selected from the 2PIGG catalog of galaxy groups. Statistical analysis indicates that non-Gaussian groups have masses significantly higher than Gaussian groups. We also have found that all non-Gaussian systems seem to be composed of multiple velocity modes. Besides, our results indicate that multimodal groups should be considered as a set of individual units with their own properties. In particular, we have found that the mass distribution of such units are similar to that of Gaussian groups. Our results reinforce the idea of non-Gaussian systems as complex structures in the phase space, likely corresponding to secondary infall aggregations at a stage before virialization. The understanding of these objects is relevant for cosmological studies using groups and clusters through the mass function evolution. ",Non-Gaussian velocity distributions - The effect on virial mass estimates of galaxy groups " Master equations in the Lindblad form describe evolution of open quantum systems that is completely positive and simultaneously has a semigroup property. We analyze a possibility to derive this type of master equations from an intrinsically discrete dynamics that is modelled as a sequence of collisions between a given quantum system (a qubit) with particles that form the environment. In order to illustrate our approach we analyze in detail how a process of an exponential decay and a process of decoherence can be derived from a collision-like model in which particular collisions are described by SWAP and controlled-NOT interactions, respectively. ",Description of quantum dynamics of open systems based on collision-like models " There are numerous examples of societies with extremely stable mix of contrasting opinions. We argue that this stability is a result of an interplay between society network topology adjustment and opinion changing processes. To support this position we present a computer model of opinion formation based on some novel assumptions, designed to bring the model closer to social reality. In our model, the agents, in addition to changing their opinions due to influence of the rest of society and external propaganda, have the ability to modify their social network, forming links with agents sharing the same opinions and cutting the links with those they disagree with. To improve the model further we divide the agents into `fanatics' and `opportunists', depending on how easy is to change their opinions. The simulations show significant differences compared to traditional models, where network links are static. In particular, for the dynamical model where inter-agent links are adjustable, the final network structure and opinion distribution is shown to resemble real world observations, such as social structures and persistence of minority groups even when most of the society is against them and the propaganda is strong. ",Studies of opinion stability for small dynamic networks with opportunistic agents " In this paper we study finite interaction range corrections to the mosaic picture of the glass transition as emerges from the study of the Kac limit of large interaction range for disordered models. To this aim we consider point to set correlation functions, or overlaps, in a one dimensional random energy model as a function of the range of interaction. In the Kac limit, the mosaic length defines a sharp first order transition separating a high overlap phase from a low overlap one. Correspondingly we find that overlap curves as a function of the window size and different finite interaction ranges cross roughly at the mosaic lenght. Nonetheless we find very slow convergence to the Kac limit and we discuss why this could be a problem for measuring the mosaic lenght in realistic models. ",Mosaic length and finite interaction-range effects in a one dimensional random energy model " Mixture models have been widely used in modeling of continuous observations. For the possibility to estimate the parameters of a mixture model consistently on the basis of observations from the mixture, identifiability is a necessary condition. In this study, we give some results on the identifiability of multivariate logistic mixture models. ",Identifiability of multivariate logistic mixture models " [Abridged] We present a large, new set of stellar evolution models and isochrones for an alpha-enhanced metal distribution typical of Galactic halo and bulge stars; it represents a homogeneous extension of our stellar model library for a distribution already presented in Pietrinferni et al.(2004). The effect of the alpha-element enhancement has been properly taken into account in the nuclear network, opacity, equation of state and, for the first time, the bolometric corrections, and color transformations. This allows us to avoid the inconsistent use - common to all alpha-enhanced model libraries currently available - of scaled-solar bolometric corrections and color transformations for alpha-enhanced models and isochrones. We show how bolometric corrections to magnitudes obtained for the U,B portion of stellar spectra for T_{eff}<=6500K, are significantly affected by the metal mixture, especially at the higher metallicities. We also provide complete sets of evolutionary models for low-mass, He-burning stellar structures covering the whole metallicity range, to enable synthetic horizontal branch simulations. We compare our database with several widely used stellar model libraries from different authors, as well as with various observed color magnitude and color-color diagrams (Johnson-Cousins BVI and near infrared magnitudes, Stromgren colors) of Galactic field stars and globular clusters. We also test our isochrones comparing integrated optical colors and Surface Brightness Fluctuation magnitudes with selected globular cluster data. We find a general satisfactory agreement with the empirical constraints. ",A large stellar evolution database for population synthesis studies. II. Stellar models and isochrones for an alpha-enhanced metal distribution " We study local (also referred to as small-signal) stability of a network of identical DC/AC converters having a rotating degree of freedom. We develop a stability theory for a class of partitioned linear systems with symmetries that has natural links to classical stability theories of interconnected systems. We find stability conditions descending from a particular Lyapunov function involving an oblique projection onto the complement of the synchronous steady state set and enjoying insightful structural properties. Our sufficient and explicit stability conditions can be evaluated in a fully decentralized fashion, reflect a parametric dependence on the converter's steady-state variables, and can be one-to-one generalized to other types of systems exhibiting the same behavior, such as synchronous machines. Our conditions demand for sufficient reactive power support and resistive damping. These requirements are well aligned with practitioners' insights. ",Parametric local stability condition of a multi-converter system " We discuss some contradictions found in the literature concerning the problem of stability of collisionless spherical stellar systems which are the simplest anisotropic generalization of the well-known polytrope models. Their distribution function $F(E,L)$ is a product of power-low functions of the energy $E$ and the angular momentum $L$, i.e. $F\propto L^{-s}(-E)^q$. On the one hand, calculation of the growth rates in the framework of linear stability theory and N-body simulations show that these systems become stable when the parameter $s$ characterizing the velocity anisotropy of the stellar distribution is lower than some finite threshold value, $s1$ in order to study the large-scale structure of the universe and test cosmological models including Dark Energy. In addition, eROSITA is expected to yield a sample of a few million AGNs, including obscured objects, revolutionizing our view of the evolution of supermassive black holes. The survey will also provide new insights into a wide range of astrophysical phenomena, including X-ray binaries, active stars, and diffuse emission within the Galaxy. Results from early observations, some of which are presented here, confirm that the performance of the instrument is able to fulfil its scientific promise. With this paper, we aim to give a concise description of the instrument, its performance as measured on ground, its operation in space, and also the first results from in-orbit measurements. ",The eROSITA X-ray telescope on SRG " Data-driven models for audio source separation such as U-Net or Wave-U-Net are usually models dedicated to and specifically trained for a single task, e.g. a particular instrument isolation. Training them for various tasks at once commonly results in worse performances than training them for a single specialized task. In this work, we introduce the Conditioned-U-Net (C-U-Net) which adds a control mechanism to the standard U-Net. The control mechanism allows us to train a unique and generic U-Net to perform the separation of various instruments. The C-U-Net decides the instrument to isolate according to a one-hot-encoding input vector. The input vector is embedded to obtain the parameters that control Feature-wise Linear Modulation (FiLM) layers. FiLM layers modify the U-Net feature maps in order to separate the desired instrument via affine transformations. The C-U-Net performs different instrument separations, all with a single model achieving the same performances as the dedicated ones at a lower cost. ",Conditioned-U-Net: Introducing a Control Mechanism in the U-Net for Multiple Source Separations We propose a new reverse time migration method for reconstructing extended obstacles in the planar waveguide using acoustic waves at a fixed frequency. We prove the resolution of the reconstruction method in terms of the aperture and the thickness of the waveguide. The resolution analysis implies that the imaginary part of the cross-correlation imaging function is always positive and thus may have better stability properties. Numerical experiments are included to illustrate the powerful imaging quality and to confirm our resolution results. ,Reverse Time Migration for Reconstructing Extended Obstacles in Planar Acoustic Waveguides " Adiabatic and periodic variation of the lattice parameters can make it possible to transport charge through a system even without net external electric or magnetic fields, known as Thouless charge pumping. The amount of charge pumped in a cycle is quantized and entirely determined by the system's topology, which is robust against perturbations such as disorder and interactions. However, coupling to the environment may play a vital role in topological transport in many-body systems. We study the topological Thouless pumping, where the charge carriers interact with local optical phonons. The semi-classical multi-trajectory Ehrenfest method is employed to treat the phonon trajectories classically and charge carriers quantum mechanically. We find a breakdown of the quantized charge transport in the presence of phonons. It happens for any finite electron-phonon coupling strength at the resonance condition when the pumping frequency matches the phonon frequency, and it takes finite phonon coupling strength away from the resonance. Moreover, there exist parameter regimes with non-quantized negative and positive charge transport. The modified effective pumping path due to electron-phonon coupling accurately explains the underlying physics. In the large coupling regime where the pumping disappears, the phonons are found to eliminate the staggering of the onsite potentials, which is necessary for the pumping protocol. Finally, we present a stability diagram of quantized pumping as a function of time period of pumping and phonon coupling strength. ",Phonon-induced breakdown of Thouless pumping in the Rice-Mele-Holstein model " We survey the methods proposed in the literature for detecting moons of extrasolar planets in terms of their ability to distinguish between prograde and retrograde moon orbits, an important tracer of moon formation channel. We find that most moon detection methods, in particular, sensitive methods for detecting moons of transiting planets, cannot observationally distinguishing prograde and retrograde moon orbits. The prograde and retrograde cases can only be distinguished where dynamical evolution of the orbit due to e.g. three body effects is detectable, where one of the two cases is dynamically unstable or where new observational facilities which can implement a technique capable of differentiating the two cases, come on line. In particular, directly imaged planets are promising targets as repeated spectral and photometric measurements, required to determine moon orbit direction, could also be conducted with the primary interest of characterising the planet itself. ",Next Generation of Telescopes or Dynamics Required to Determine if Exo-Moons have Prograde or Retrograde Orbits " Radiative effects in the electroproduction of photons in polarized $ep$-scattering are calculated with the next-to-leading (NLO) accuracy. The contributions of loops and two photon emission were presented in analytical form. The covariant approach of Bardin and Shumeiko was used to extract the infrared divergence. All contributions to the radiative correction were presented in the form of the correction to the leptonic tensor thus allowing for further applications in other experiments, e.g., deep inelastic scattering. The radiative corrections (RC) to the cross sections and polarization asymmetries were analyzed numerically for kinematical conditions of the current measurement at Jefferson Lab. Specific attention was paid on analyzing kinematical conditions for the process with large radiative effect when momenta of two photons in the final state are collinear to momenta of initial and final electrons, respectively. ",QED radiative effects in the processes of exclusive photon electroproduction from polarized protons with the next-to-leading accuracy " For systems in an externally controllable time-dependent potential, the optimal protocol minimizes the mean work spent in a finite-time transition between given initial and final values of a control parameter. For an initially thermalized ensemble, we consider both Hamiltonian evolution for classical systems and Schr\""odinger evolution for quantum systems. In both cases, we show that for harmonic potentials, the optimal work is given by the adiabatic work even in the limit of short transition times. This result is counter-intuitive because the adiabatic work is substantially smaller than the work for an instantaneous jump. We also perform numerical calculations of the optimal protocol for Hamiltonian dynamics in an anharmonic quartic potential. For a two-level spin system, we give examples where the adiabatic work can be reached in either a finite or an arbitrarily short transition time depending on the allowed parameter space. ","Optimal protocols for Hamiltonian and Schr\""odinger dynamics" " We demonstrate the effect of the depletion force in experiments and simulations of vertically vibrated mixtures of large and small steel spheres. The system exhibits size segregation and a large increase in the pair correlation function of the large spheres for short distances that can be accurately described using a combination of the depletion potential derived for equilibrium colloidal systems and a Boltzmann factor. The Boltzmann factor defines an effective temperature for the system, which we compare to other measures of the temperature. ",The depletion force in a bi-disperse granular layer " We analyze interstellar absorption observed towards two subdwarf O stars, JL 9 and LSS 1274, using spectra taken by the Far Ultraviolet Spectroscopic Explorer (FUSE). Column densities are measured for many atomic and molecular species (H I, D I, C I, N I, O I, P II, Ar I, Fe II, and H2), but our main focus is on measuring the D/H ratios for these extended lines of sight, as D/H is an important diagnostic for both cosmology and Galactic chemical evolution. We find D/H=(1.00+/-0.37)e-5 towards JL 9, and D/H=(0.76+/-0.36)e-5 towards LSS 1274 (2 sigma uncertainties). With distances of 590+/-160 pc and 580+/-100 pc, respectively, these two lines of sight are currently among the longest Galactic lines of sight with measured D/H. With the addition of these measurements, we see a significant tendency for longer Galactic lines of sight to yield low D/H values, consistent with previous inferences about the deuterium abundance from D/O and D/N measurements. Short lines of sight with H I column densities of log N(H I)<19.2 suggest that the gas-phase D/H value within the Local Bubble is (D/H)_LBg=(1.56+/-0.04)e-5. However, the four longest Galactic lines of sight with measured D/H, which have d>500 pc and log N(H I)>20.5, suggest a significantly lower value for the true local-disk gas-phase D/H value, (D/H)_LDg=(0.85+/-0.09)e-5. One interpretation of these results is that D is preferentially depleted onto dust grains relative to H and that longer lines of sight that extend beyond the Local Bubble sample more depleted material. In this scenario, the higher Local Bubble D/H ratio is actually a better estimate than (D/H)_LDg for the true local-disk D/H, D/H)_LD. However, if (D/H)_LDg is different from (D/H)_LBg simply because of variable astration and incomplete ISM mixing, then (D/H)_LD=(D/H)_LDg. ",Two New Low Galactic D/H Measurements from FUSE " We study, using simulations the dynamical properties of complex ferromagnetic granular materials. The system of grains is modeled by a disordered two-dimensional lattice in which the grains are embedded, while the magnitude and direction of the easy axis are random. Using the monte-carlo method we track the dynamics of the magnetic moments of the grains. We observe a transition of the system from a macroscopic blocked (ferromagnetic) phase at low temperature in which the grain's magnetic moment do not flip to the other direction to an unblocked (superparamagnetic) phase at high temperature in which the magnetic moment is free to rotate. Our results suggest that this transition exhibits the characteristics of a second order phase transition such as the appearance of a giant cluster of unblocked grains which is fractal at the critical temperature, a peak in the size of the second largest cluster at the same temperature and a power law distribution of cluster sizes near the criticality. ",A dynamical phase transition in ferromagnetic granular materials " We present the results obtained from linear stability analysis and 2.5-dimensional magnetohydrodynamic (MHD) simulations of the magnetorotational instability (MRI), including the effects of cosmic rays (CRs). We took into account of the CR diffusion along the magnetic field but neglect the cross-field-line diffusion. Two models are considered in this paper: shearing box model and differentially rotating cylinder model. We studied how MRI is affected by the initial CR pressure (i.e., energy) distribution. In the shearing box model, the initial state is uniform distribution. Linear analysis shows that the growth rate of MRI does not depend on the value of CR diffusion coefficient. In the differentially rotating cylinder model, the initial state is a constant angular momentum polytropic disk threaded by weak uniform vertical magnetic field. Linear analysis shows that the growth rate of MRI becomes larger if the CR diffusion coefficient is larger. Both results are confirmed by MHD simulations. The MHD simulation results show that the outward movement of matter by the growth of MRI is not impeded by the CR pressure gradient, and the centrifugal force which acts to the concentrated matter becomes larger. Consequently, the growth rate of MRI is increased. On the other hand, if the initial CR pressure is uniform, then the growth rate of the MRI barely depends on the value of the CR diffusion coefficient. ",Analysis of the magneto-rotational instability with the effect of cosmic-ray diffusion " Accent recognition with deep learning framework is a similar work to deep speaker identification, they're both expected to give the input speech an identifiable representation. Compared with the individual-level features learned by speaker identification network, the deep accent recognition work throws a more challenging point that forging group-level accent features for speakers. In this paper, we borrow and improve the deep speaker identification framework to recognize accents, in detail, we adopt Convolutional Recurrent Neural Network as front-end encoder and integrate local features using Recurrent Neural Network to make an utterance-level accent representation. Novelly, to address overfitting, we simply add Connectionist Temporal Classification based speech recognition auxiliary task during training, and for ambiguous accent discrimination, we introduce some powerful discriminative loss functions in face recognition works to enhance the discriminative power of accent features. We show that our proposed network with discriminative training method (without data-augment) is significantly ahead of the baseline system on the accent classification track in the Accented English Speech Recognition Challenge 2020, where the loss function Circle-Loss has achieved the best discriminative optimization for accent representation. ",Deep Discriminative Feature Learning for Accent Recognition " In many situations, the decision maker observes items in sequence and needs to determine whether or not to retain a particular item immediately after it is observed. Any decision rule creates a set of items that are selected. We consider situations where the available information is the rank of a present observation relative to its predecessors. Certain ``natural'' selection rules are investigated. Theoretical results are presented pertaining to the evolution of the number of items selected, measures of their quality and the time it would take to amass a group of a given size. ",Select sets: Rank and file " I present an introduction to the field of Quantum Chromodynamics (QCD) with emphasis on nucleon spin structure and perturbative methods. After a somewhat comprehensive overview of perturbative QCD, including the systematics of renormalization, I introduce deeply-inelastic scattering and deeply-virtual Compton scattering along with the nonperturbative distribution functions appearing in these processes. I show how these processes can be used to extract the distribution of longitudinal spin of nucleons, and briefly discuss the complications associated with transverse spin components. There are five appendices which accompany this work, discussing topics ranging from canonical field quantization to dimensional regularization to the structure of several important Lie groups. These appendices are meant to give the reader a certain basic understanding of some of the mathematical details underlying the main text, as well as provide a useful reference. This work was submitted in a slightly different form to the University of Maryland at College Park in partial completion of the requirements for a doctor of philosophy in physics, received August 2000. ",A Study of Nucleon Spin Struture from Quantum Chromodynamics " Equations of motion of low-energy effective theories of quantum electrodynamics include infinitely many interaction terms, which make them difficult to solve. The self-duality property has facilitated research on the solutions to these equations. In this paper, equations of motion of systems of non-Abelian gauge fields on even-dimensional spheres are considered. It is demonstrated that the Cremmer-Scherk configuration, which satisfies certain generalized self-duality equations, becomes the classical solution for the class of systems that are given by arbitrary functions of class C^1 of 2m+1 quantities. For instance, Lagrangians consisting of multi-trace terms are included in this class. This result is likely to generate several new and interesting directions of research, including the classification of actions with respect to the stability condition against the Cremmer-Scherk configuration. ",Equations of Motion Solved by the Cremmer-Scherk Configuration on Even-Dimensional Spheres " The dynamics of a single hole (or electron) in the two dimensional Hubbard model is investigated. The antiferromagnetic background is described by a N\`eel state, and the hopping of the carrier is analyzed within a configuration interaction approach. Results are in agreement with other methods and with experimental data when available. All data are compatible with the opening of a mean field gap in a Fermi liquid of spin polarons, the so called Slater type of transition. In particular, this hypothesis explains the unusual dispersion relation of the quasiparticle bands near the transition. Recent photoemission data for Ca$_2$CuO$_2$Cl$_2$ are analyzed within this context. ",Dynamics of Holes and Universality Class of the Antiferromagnetic Transition in the Two Dimensional Hubbard Model " We establish a phase transition for the parking process on critical Galton--Watson trees. In this model, a random number of cars with mean $m$ and variance $\sigma^{2}$ arrive independently on the vertices of a critical Galton--Watson tree with finite variance $\Sigma^{2}$ conditioned to be large. The cars go down the tree towards the root and try to park on empty vertices as soon as possible. We show a phase transition depending on $$ \Theta:= (1-m)^2- \Sigma^2 (\sigma^2+m^2-m).$$ Specifically, when $m \leq 1$, if $ \Theta>0,$ then all but (possibly) a few cars will manage to park, whereas if $\Theta<0$, then a positive fraction of the cars will not find a spot and exit the tree through the root. This confirms a conjecture of Goldschmidt and Przykucki. ",The phase transition for parking on Galton--Watson trees " We consider the general problem of enumerating branched covers of the projective line from a fixed general curve subject to ramification conditions at possibly moving points. Our main computations are in genus 1; the theory of limit linear series allows one to reduce to this case. We first obtain a simple formula for a weighted count of pencils on a fixed elliptic curve E, where base-points are allowed. We then deduce, using an inclusion-exclusion procedure, formulas for the numbers of maps E->P^1 with moving ramification conditions. A striking consequence is the invariance of these counts under a certain involution. Our results generalize work of Harris, Logan, Osserman, and Farkas-Moschetti-Naranjo-Pirola. ",Enumerating pencils with moving ramification on curves A new mixing layer can be generated if the rotation of either of the two cylinders in a Taylor--Couette apparatus varies discontinuously along the symmetry axis. The mixing zone between the two resulting co-current streams gives rise to radial vorticity in addition to the primary axial vorticity. An analytic solution for the azimuthal velocity has been derived from which we show that the width of the mixing zone varies only with radial position. ,Mixing layer between two co-current Taylor-Couette flows " We study gravitational perturbations of the Kerr-AdS_5 x S^5 spacetime with equal angular momenta. In this spacetime, we found the two kinds of classical instabilities, superradiant and Gregory-Laflamme instabilities. The superradiant instability is caused by the wave amplification via superradiance, and by wave reflection due to the potential barrier of the AdS spacetime. The Gregory-Laflamme instability appears in Kaluza-Klein modes of the internal space S^5 and breaks the symmetry SO(6). Taken into account these instabilities, the phase structure of Kerr-AdS_5 x S^5 spacetime is revealed. The implication for the AdS/CFT correspondence is also discussed. ",Instabilities of Kerr-AdS_5 x S^5 Spacetime " We introduce a novel approach for the characterization of the quality of a laser beam that is not based on particular criteria for beam width definition. The Lorenz curve of a light beam is a sophisticated version of the so-called power-in-the-bucket curve, formed by the partial sums of discretized joint intensity distribution in the near and far fields sorted in decreasing order. According to majorization theory, a higher Lorenz curve implies that all measures of spreading in phase space, and, in particular, all R\'enyi (and Shannon) entropy-based measures of the beam width products in near and far fields, are unanimously smaller, providing a strong assessment of a better beam quality. Two beams whose Lorenz curves intersect can only be considered of relatively better or lower quality according to specific criteria, which can be inferred from the plot of the respective Lorenz curves. ",Lorenz curve of a light beam: evaluating beam quality from a mayorization perspective " We study the outcome of deferred acceptance when prospective medical residents can only apply to a limited set of hospitals. This limitation requires residents to make a strategic choice about the quality of hospitals they apply to. Through a mix of theoretical and experimental results, we study the effect of this strategic choice on the preferences submitted by participants, as well as on the overall welfare. We find that residents' choices in our model mimic the behavior observed in real systems where individuals apply to a mix of positions consisting mostly of places where they are reasonably likely to get accepted, as well as a few ""reach"" applications to hospitals of very high quality, and a few ""safe"" applications to hospitals of lower than their expected level. Surprisingly, the number of such ""safe"" applications is not monotone in the number of allowed applications. We also find that selfish behavior can hurt social welfare, but the deterioration of overall welfare is very minimal. ",Effect of selfish choices in deferred acceptance with short lists " In order to better understand the possibility of coronal heating by MHD waves, we analyze Fe xii 195.12{\AA} data observed with EUV Imaging Spectrometer (EIS) onboard Hinode. We performed a Fourier analysis of EUV intensity and Doppler velocity time series data in the active region corona. Notable intensity and Doppler velocity oscillations were found for two moss regions out of the five studied, while only small oscillations were found for five apexes of loops. The amplitudes of the oscillations were 0.4 - 5.7% for intensity and 0.2 - 1.2 kms-1 for Doppler velocity. In addition, oscillations of only Doppler velocity were seen relatively less often in the data. We compared the amplitudes of intensity and those of Doppler velocity in order to identify MHD wave modes, and calculated the phase delays between Fourier components of intensity and those of Doppler velocity. The results are interpreted in terms of MHD waves as follows: (1) few kink modes or torsional Alfv'en mode waves were seen in both moss regions and the apexes of loops; (2) upwardly propagating and standing slowmode waves were found inmoss regions; and (3) consistent with previous studies, estimated values of energy flux of the waves were several orders of magnitude lower than that required for heating active regions. ",Mode identification of MHD waves in an active region observed with Hinode/EIS " This paper studies operator forms of the nonhomogeneous associative classical Yang-Baxter equation (nhacYBe), extending and generalizing such studies for the classical Yang-Baxter equation and associative Yang-Baxter equation that can be tracked back to the works of Semonov-Tian-Shansky and Kupershmidt on Rota-Baxter Lie algebras and $\mathcal{O}$-operators. In general, solutions of the nhacYBe are characterized in terms of generalized $\mathcal{O}$-operators. The characterization can be given by the classical $\mathcal{O}$-operators precisely when the solutions satisfy an invariant condition. When the invariant condition is compatible with a Frobenius algebra, such solutions have close relationships with Rota-Baxter operators on the Frobenius algebra. In general, solutions of the nhacYBe can be produced from Rota-Baxter operators, and then from $\mathcal{O}$-operators when the solutions are taken in semi-direct product algebras. In the other direction, Rota-Baxter operators can be obtained from solutions of the nhacYBe in unitizations of algebras. Finally a classifications of solutions of the nhacYBe satisfying the mentioned invariant condition in all unital complex algebras of dimensions two and three are obtained. All these solutions are shown to come from Rota-Baxter operators. ",Operator forms of nonhomogeneous associative classical Yang-Baxter equation " In this work spin transport in corrugated armchair graphene nanoribbons (AGNR) is studied. We survey combined effects of spin-orbit interaction and surface roughness, employing the non-equilibrium Green's function formalism and four orbitals tight-binding model. We modify hopping parameters regarding bending and distance of corrugated carbon atoms. The effects of surface roughness parameters, such as roughness amplitude and correlation length, on the spin transport of the graphene nanoribbons are studied. We show that increasing surface roughness amplitude breaks the AGNR symmetry and hybridize $\mathit{\sigma}$ and $\mathit{\pi}$ orbitals, leading to more spin flipping and therefore decrease in polarization. Unlike the roughness amplitude, the longer correlation length makes AGNR surface smoother and increases polarization. Moreover, the Spin diffusion length of carriers is extracted and its dependency on the roughness parameters is investigated. We find the spin diffusion length for various surface corrugation amplitudes in order of 1 to 80 micrometers. ",Spin Relaxation at Graphene Nanoribbons in the presence of Substrate Surface Roughness " This paper completely solves the controllability problems of two-dimensional multi-input discrete-time bilinear systems with and without drift. Necessary and sufficient conditions for controllability, which cover the existing results, are obtained by using an algebraic method. Furthermore, for the uncontrollable systems, near-controllability is studied and necessary and sufficient conditions for the systems to be nearly controllable are also presented. Examples are provided to demonstrate the conceptions and results of the paper. ",On Controllability and Near-controllability of Multi-input Discrete-time Bilinear Systems in Dimension Two " An autonomous variational inference algorithm for arbitrary graphical models requires the ability to optimize variational approximations over the space of model parameters as well as over the choice of tractable families used for the variational approximation. In this paper, we present a novel combination of graph partitioning algorithms with a generalized mean field (GMF) inference algorithm. This combination optimizes over disjoint clustering of variables and performs inference using those clusters. We provide a formal analysis of the relationship between the graph cut and the GMF approximation, and explore several graph partition strategies empirically. Our empirical results provide rather clear support for a weighted version of MinCut as a useful clustering algorithm for GMF inference, which is consistent with the implications from the formal analysis. ",Graph partition strategies for generalized mean field inference " A common approach to metric-affine, local Poincar\'e, special-relativistic and Galilei spacetime geometry is developed. Starting from an affine composite bundle, we introduce local reference frames and their evolution along worldlines and we study both, absolute and relative simultaneity postulates, giving rise to alternative concepts of spacetime. In particular, the construction of the Minkowski metric, and its required invariance, allows either to reorganize the original affine bundle as a metric-affine geometry with explicit Lorentz symmetry, or to restrict it to a Poincar\'e geometry, both of them constituting the background of a wide class of gauge theories of gravity. ",A proposal of foundation of spacetime geometry " Iron-based superconductors have received much attention as a new family of high-temperature superconductors owing to their unique properties and distinct differences from cuprates and conventional superconductors. This paper reviews progress in thin film research on iron-based superconductors since their discovery for each of five material systems with an emphasis on growth, physical properties, device fabrication, and relevant bulk material properties. ",Thin Film Growth and Device Fabrication of Iron-Based Superconductors " The occurrence of phase fluctuations due to thermal excitations in Bose-Einstein condensates (BECs) is studied for a variety of temperatures and trap geometries. We observe the statistical nature of the appearence of phase fluctuations and characterize the dependence of their average value on temperature, number of particles and the trapping potential. We find pronounced phase fluctuations for condensates in very elongated traps in a broad temperature range. The results are of great importance for the realization of BEC in quasi 1D geometries, for matter wave interferometry with BECs, as well as for coherence properties of guided atom laser beams. ",Observation of Phase Fluctuations in Bose-Einstein Condensates We explain the experimentally observed Aharonov-Bohm (AB) resonance patterns of an antidot cluster by means of quantum and classical simulations and Feynman path integral theory. We demonstrate that the observed behavior of the AB period signals the crossover from a low B regime which can be understood in terms of electrons following classical orbits to an inherently quantum high B regime where this classical picture and semiclassical theories based on it do not apply. ,Resonance Patterns of an Antidot Cluster: From Classical to Quantum Ballistics " We present the first statistical analysis of 27 UVOT optical/ultra-violet lightcurves of GRB afterglows. We have found, through analysis of the lightcurves in the observer's frame, that a significant fraction rise in the first 500s after the GRB trigger, that all lightcurves decay after 500s, typically as a power-law with a relatively narrow distribution of decay indices, and that the brightest optical afterglows tend to decay the quickest. We find that the rise could either be produced physically by the start of the forward shock, when the jet begins to plough into the external medium, or geometrically where an off-axis observer sees a rising lightcurve as an increasing amount of emission enters the observers line of sight, which occurs as the jet slows. We find that at 99.8% confidence, there is a correlation, in the observed frame, between the apparent magnitude of the lightcurves at 400s and the rate of decay after 500s. However, in the rest frame a Spearman Rank test shows only a weak correlation of low statistical significance between luminosity and decay rate. A correlation should be expected if the afterglows were produced by off-axis jets, suggesting that the jet is viewed from within the half-opening angle theta or within a core of uniform energy density theta_c. We also produced logarithmic luminosity distributions for three rest frame epochs. We find no evidence for bimodality in any of the distributions. Finally, we compare our sample of UVOT lightcurves with the XRT lightcurve canonical model. The range in decay indices seen in UVOT lightcurves at any epoch is most similar to the range in decay of the shallow decay segment of the XRT canonical model. However, in the XRT canonical model there is no indication of the rising behaviour observed in the UVOT lightcurves. ",A statistical study of gamma-ray burst afterglows measured by the Swift Ultra-violet Optical Telescope " We use Cox's description for sheaves on toric varieties and results about the local cohomology with respect to monomial ideals to give a characteristic free approach to vanishing results on arbitrary toric varieties. As an application, we give a proof of a strong form of Fujita's conjecture in the case of toric varieties. We also prove that every sheaf on a toric variety corresponds to a module over the homogeneous coordinate ring, generalizing Cox's result for the simplicial case. ",Vanishing Theorems on Toric Varieties " Hybrid systems, such as bipedal walkers, are challenging to control because of discontinuities in their nonlinear dynamics. Little can be predicted about the systems' evolution without modeling the guard conditions that govern transitions between hybrid modes, so even systems with reliable state sensing can be difficult to control. We propose an algorithm that allows for determining the hybrid mode of a system in real-time using data-driven analysis. The algorithm is used with data-driven dynamics identification to enable model predictive control based entirely on data. Two examples---a simulated hopper and experimental data from a bipedal walker---are used. In the context of the first example, we are able to closely approximate the dynamics of a hybrid SLIP model and then successfully use them for control in simulation. In the second example, we demonstrate gait partitioning of human walking data, accurately differentiating between stance and swing, as well as selected subphases of swing. We identify contact events, such as heel strike and toe-off, without a contact sensor using only kinematics data from the knee and hip joints, which could be particularly useful in providing online assistance during walking. Our algorithm does not assume a predefined gait structure or gait phase transitions, lending itself to segmentation of both healthy and pathological gaits. With this flexibility, impairment-specific rehabilitation strategies or assistance could be designed. ",Data-Driven Gait Segmentation for Walking Assistance in a Lower-Limb Assistive Device " The automatic detection of changes or anomalies between multispectral and hyperspectral images collected at different time instants is an active and challenging research topic. To effectively perform change-point detection in multitemporal images, it is important to devise techniques that are computationally efficient for processing large datasets, and that do not require knowledge about the nature of the changes. In this paper, we introduce a novel online framework for detecting changes in multitemporal remote sensing images. Acting on neighboring spectra as adjacent vertices in a graph, this algorithm focuses on anomalies concurrently activating groups of vertices corresponding to compact, well-connected and spectrally homogeneous image regions. It fully benefits from recent advances in graph signal processing to exploit the characteristics of the data that lie on irregular supports. Moreover, the graph is estimated directly from the images using superpixel decomposition algorithms. The learning algorithm is scalable in the sense that it is efficient and spatially distributed. Experiments illustrate the detection and localization performance of the method. ",Online Graph-Based Change Point Detection in Multiband Image Sequences " Methods. The northern-eastern fringe of the Chameleon I dark cloud was observed with XMM-Newton, revisiting a region observed with ROSAT 15 years ago. Centered on the extended X-ray source CHXR49 we are able to resolve it into three major contributing components and to analyse their spectral properties. Furthermore, the deep exposure allows not only the detection of numerous, previously unknown X-ray sources, but also the investigation of variability and the study of the X-ray properties for the brighter targets in the field. We use EPIC spectra, to determine X-ray brightness, coronal temperatures and emission measures for these sources, compare the properties of classical and weak-line T Tauri stars and make a comparison with results from the ROSAT observation. Results. X-ray properties of T Tauri stars in Cha I are presented. The XMM-Newton images resolve some previously blended X-ray sources, confirm several possible ones and detect many new X-ray targets, resulting in the most comprehensive list with 71 X-ray sources in the northern Cha I dark cloud. The analysis of medium resolution spectra shows an overlapping distribution of spectral properties for classical and weak-line T Tauri stars, with the X-ray brighter stars having hotter coronae and a higher L_X/L_bol ratio. X-ray luminosity correlates with bolometric luminosity, whereas the L_X/L_bol ratio is slightly lower for the classical T Tauri stars. Large flares as well as a low iron and a high neon abundance are found in both types of T Tauri stars. Abundance pattern, plasma temperatures and emission measure distributions during quiescent phases are attributed toa high level of magnetic activity as the dominant source of their X-ray emission. ",A deep XMM-Newton X-ray observation of the Chamaeleon I dark cloud " With the advent of the 5th generation of wireless standards and an increasing demand for higher throughput, methods to improve the spectral efficiency of wireless systems have become very important. In the context of cognitive radio, a substantial increase in throughput is possible if the secondary user can make smart decisions regarding which channel to sense and when or how often to sense. Here, we propose an algorithm to not only select a channel for data transmission but also to predict how long the channel will remain unoccupied so that the time spent on channel sensing can be minimized. Our algorithm learns in two stages - a reinforcement learning approach for channel selection and a Bayesian approach to determine the optimal duration for which sensing can be skipped. Comparisons with other learning methods are provided through extensive simulations. We show that the number of sensing is minimized with negligible increase in primary interference; this implies that lesser energy is spent by the secondary user in sensing and also higher throughput is achieved by saving on sensing. ",Spectrum Access In Cognitive Radio Using A Two Stage Reinforcement Learning Approach " Nanoparticles grown in a plasma are used to visualize the process of film deposition in a pulsed radio-frequency (rf) atmospheric pressure glow discharge. Modulating the plasma makes it possible to successfully prepare porous TiO2 films. We study the trapping of the particles in the sheath during the plasma-on phase and compare it with numerical simulations. During the plasma-off phase, the particles are driven to the substrate by the electric field generated by residual ions, leading to the formation of porous TiO2 film. Using video microscopy, the collective dynamics of particles in the whole process is revealed at the most fundamental ""kinetic"" level. ",Particle dynamics in deposition of porous films with a pulsed radio-frequency atmospheric pressure glow discharge " Security attacks present unique challenges to self-adaptive system design due to the adversarial nature of the environment. However, modeling the system as a single player, as done in prior works in security domain, is insufficient for the system under partial compromise and for the design of fine-grained defensive strategies where the rest of the system with autonomy can cooperate to mitigate the impact of attacks. To deal with such issues, we propose a new self-adaptive framework incorporating Bayesian game and model the defender (i.e., the system) at the granularity of components in system architecture. The system architecture model is translated into a Bayesian multi-player game, where each component is modeled as an independent player while security attacks are encoded as variant types for the components. The defensive strategy for the system is dynamically computed by solving the pure equilibrium to achieve the best possible system utility, improving the resiliency of the system against security attacks. ",System Component-Level Self-Adaptations for Security via Bayesian Games " The YN results are presented from the Extended-soft-core (ESC) interactions. They consist of local- and non-local-potentials due to (i) One-boson-exchange (OBE), with pseudoscalar-, vector-, scalar-, and axial-vector-nonets, (ii) Diffractive exchanges, (iii) Two-pseudoscalar exchange, and (iv) Meson-pair-exchange (MPE). This model, called ESC04, describes NN and YN in a unified way using broken flavor SU(3)-symmetry. Novel ingredients are the inclusion of (i) the axial-vector-mesons, (ii) a zero in the scalar- and axial-vector meson form factors. We describe simultaneous fits to the NN- and YN-data, using four options in the ESC-model. Very good fits were obtained. G-matrix calculations with these four options are also reported. The obtained well depths (U_\Lambda, U_\Sigma, U_\Xi) reveal distinct features of ESC04a-d. The \Lambda\Lambda-interactions are demonstrated to be consistent with the observed data of_{\Lambda\Lambda}^6He. The possible three-body effects are investigated by considering phenomenologically the changes of the vector-meson masses in a nuclear medium. ",Extended-soft-core Baryon-Baryon Model II. Hyperon-Nucleon Interaction " We show that the delivery of fracture work to the tip of an advancing planar crack is strongly reduced by surface phonon emission, leading to forbidden ranges of crack speed. The emission can be interpreted through dispersion of the group velocity, and Rayleigh and Love branches contribute as well as other high frequency branches of the surface wave dispersion relations. We also show that the energy release rate which enters the Griffith criterion for the crack advance can be described as the product of the continuum solution with a function that only depends on the lattice geometry and describes the lattice influence on the phonon emission. Simulations are performed using a new finite element model for simulating elasticity and fractures. The model, built to allow fast and very large three-dimensional simulations, is applied to the simplified case of two dimensional samples. ",Role of surface waves on the relation between crack speed and the work of fracture " The $l$-th stopping redundancy $\rho_l(\mathcal C)$ of the binary $[n, k, d]$ code $\mathcal C$, $1 \le l \le d$, is defined as the minimum number of rows in the parity-check matrix of $\mathcal C$, such that the smallest stopping set is of size at least $l$. The stopping redundancy $\rho(\mathcal C)$ is defined as $\rho_d(\mathcal C)$. In this work, we improve on the probabilistic analysis of stopping redundancy, proposed by Han, Siegel and Vardy, which yields the best bounds known today. In our approach, we judiciously select the first few rows in the parity-check matrix, and then continue with the probabilistic method. By using similar techniques, we improve also on the best known bounds on $\rho_l(\mathcal C)$, for $1 \le l \le d$. Our approach is compared to the existing methods by numerical computations. ",Refined Upper Bounds on Stopping Redundancy of Binary Linear Codes " One of the key questions in Astrophysics concerns the issue of whether there exists an upper-mass limit to stars, and if so, what physical mechanism sets this limit, which might also determine if the upper-mass limit is metallicity (Z) dependent. We argue that mass loss by radiation-driven winds mediated by line opacity is one of the prime candidates setting the upper-mass limit. We present mass-loss predictions (dM/dt_wind) from Monte Carlo radiative transfer models for relatively cool (Teff = 15kK) inflated very massive stars (VMS) with large Eddington Gamma factors in the mass range 100-1000 Msun as a function of metallicity down to 1/100 Z/Zsun. We employ a hydrodynamic version of our Monte Carlo method, allowing us to predict the rate of mass loss (dM/dt_wind) and the terminal wind velocity (vinf) simultaneously. Interestingly, we find wind terminal velocities (vinf) that are low (100-500 km/s) over a wide Z-range, and we propose that the slow winds from VMS are an important source of self-enrichment in globular clusters. We also find mass-loss rates (dM/dt_wind), exceeding the typical mass-accretion rate (dM/dt_accr) of 0.001 Msun/yr during massive-star formation. We express our mass-loss predictions as a function of mass and Z, finding log dM/dt = -9.13 + 2.1 log(M/Msun) + 0.74 log(Z/Zsun) (Msun/yr). Even if stellar winds would not directly halt & reverse mass accretion during star formation, if the most massive stars form by stellar mergers stellar wind mass loss may dominate over the rate at which stellar growth takes place. We therefore argue that the upper-mass limit is effectively Z-dependent due to the nature of radiation-driven winds. This has dramatic consequences for the most luminous supernovae, gamma-ray bursts, and other black hole formation scenarios at different Cosmic epochs. ","Very Massive Stars: a metallicity-dependent upper-mass limit, slow winds, and the self-enrichment of Globular Clusters" " A catalogue of astrometric (positions, proper motions) and photometric (B, V, R, r', J) data of stars in fields with ICRF objects has been compiled at the Observatory of the National Academy of Sciences of Ukraine and the Kyiv University Observatory. All fields are located in the declination zone from 0 to +30 degrees; the nominal field size is 46' (right ascension) x 24' (declination). The observational basis of this work is 1100 CCD scans down to V=17 mag which were obtained with the Kyiv meridian axial circle in 2001-2003. The catalogue is presented in two versions. The version KMAC1-T contains 159 fields (104 796 stars) and was obtained with reduction to the Tycho2 catalogue. For another 33 fields, due to a low sky density of Tycho2 stars, the reduction was found to be unreliable. Transformation to the ICRF system in the second version of the catalogue (KMAC1-CU) was performed using the UCAC2 and CMC13 catalogues as a reference; it contains 115 032 stars in 192 fields and is of slightly better accuracy. The external accuracy of one catalogue position is about 50-90 mas for V < 15 mag stars. The average error of photometry is better than 0.1 mag for stars down to 16 mag. ",The Kyiv Meridian Axial Circle Catalogue of stars in fields with extragalactic radio sources " In this paper, we find the internal stratification for terrestrial planets with given mass-radius pairs, and use the core size and density to estimate their maximum dipolar magnetic moment. We also comment on the temporal evolution, although more information (e.g., core composition, mantle rheology and history) is crucial in determining the state of the dynamo with planetary age. ",Predicted dynamos for terrestrial extra-solar planets and their influence in habitability " We comment and discuss the findings and conclusions of a recent theoretical study of the diffraction of He atoms from a monolayer of Xe atoms adsorbed on the graphite (0001) surface [Khokonov et al., Surf. Sci. 496(2002)L13]. By revisiting the problem we demonstrate that all main conclusions of Khokonov et al. that pertain to the studied system are at variance with the available experimental and theoretical evidence and the results of multiple scattering calculations presented in this comment. ",Diffraction of He atoms from Xe monolayer adsorbed on the graphite (0001) revisited: The importance of multiple scattering processes " Pairing between spinless fermions can generate Majorana fermion excitations that exhibit intriguing properties arising from non-local correlations. But simple models indicate that non-local correlation between Majorana fermions becomes unstable at non-zero temperatures. We address this issue by showing that anisotropic interactions between dipolar fermions in optical lattices can be used to significantly enhance thermal stability. We construct a model of oriented dipolar fermions in a square optical lattice. We find that domains established by strong interactions exhibit enhanced correlation between Majorana fermions over large distances and long times even at finite temperatures, suitable for stable redundancy encoding of quantum information. Our approach can be generalized to a variety of configurations and other systems, such as quantum wire arrays. ",Enhancing the Thermal Stability of Majorana Fermions with Redundancy Using Dipoles in Optical Lattices " Chinese dynastic histories form a large continuous linguistic space of approximately 2000 years, from the 3rd century BCE to the 18th century CE. The histories are documented in Classical (Literary) Chinese in a corpus of over 20 million characters, suitable for the computational analysis of historical lexicon and semantic change. However, there is no freely available open-source corpus of these histories, making Classical Chinese low-resource. This project introduces a new open-source corpus of twenty-four dynastic histories covered by Creative Commons license. An original list of Classical Chinese gender-specific terms was developed as a case study for analyzing the historical linguistic use of male and female terms. The study demonstrates considerable stability in the usage of these terms, with dominance of male terms. Exploration of word meanings uses keyword analysis of focus corpora created for genderspecific terms. This method yields meaningful semantic representations that can be used for future studies of diachronic semantics. ",Corpus of Chinese Dynastic Histories: Gender Analysis over Two Millennia " We investigate the zero-temperature phase diagram and spin-wave properties of a double exchange magnet with on-site Hubbard repulsion. It is shown that even within a simple Hartree -- Fock approach this interaction (which is often omitted in theoretical treatments) leads to qualitatively important effects which are highly relevant in the context of experimental data for the colossal magnetoresistance compounds. These include the asymmetry of the doping dependence of spin stiffness, and the zone-boundary ``softening'' of spin wave dispersion. Effects of Hubbard repulsion on phase separation are analyzed as well. We also show that in the ferromagnetic phase, an unusual temperature-dependent effective electron-electron interaction arises at finite T. The mean-field scheme, however, does not yield the experimentally observed density of states depletion near the Fermi level. We speculate that proper treatment of electron-electron interactions may be necessary for understanding both this important feature and more generally the physics of colossal magnetoresistance phenomenon. ",Effects of the On-Site Coulomb Repulsion in Double Exchange Magnets " We present a representative set of analytic stationary state solutions of the Nonlinear Schr\""odinger equation for a symmetric double square well potential for both attractive and repulsive nonlinearity. In addition to the usual symmetry preserving even and odd states, nonlinearity introduces quite exotic symmetry breaking solutions - among them are trains of solitons with different number and sizes of density lumps in the two wells. We use the symmetry breaking localized solutions to form macroscopic quantum superpositions states and explore a simple model for the exponentially small tunneling splitting. ","Bose-Einstein condensates in a one-dimensional double square well: Analytical solutions of the Nonlinear Schr\""odinger equation and tunneling splittings" " The very recent Boomerang results give an estimate of unprecedented precision of the Cosmic Microwave Background anisotropies on sub--degree scales. A puzzling feature for theoretical cosmology is the low amplitude of the second acoustic peak. Through a qualitative discussion, we argue that a scarcely considered category of flat models, with a leptonic asymmetry, a high baryon density and a low cosmological constant seems to be in very good agreement with the data, while still being compatible with big bang nucleosynthesis and some other observational constraints. Although this is certainly not the only way to explain the data, we believe that these models deserve to be included in forthcoming likelihood analyses. ","Remarks on the Boomerang results, the baryon density and the leptonic asymmetry" " In a previous paper, using ergodic theory, Lo [1] derived a simple definite integral that provided an estimate of the view periods of ground stations to satellites. This assumes the satellites are in circular orbits with non-repeating ground tracks under linear $J_2$ perturbations. The novel feature is that this is done without the propagation of the trajectory by employing ergodic theory. This accelerated the telecommunications mission design and analysis by several orders of magnitude and greatly simplified the process. In this paper, we extend the view period integral to elliptical orbits. ",The Long-Term Forecast of Station View Periods for Elliptical Orbits " The massive upload of text on the internet creates a huge inverted index in information retrieval systems, which hurts their efficiency. The purpose of this research is to measure the effect of the Multi-Layer Similarity model of the automatic text summarization on building an informative and condensed invert index in the IR systems. To achieve this purpose, we summarized a considerable number of documents using the Multi-Layer Similarity model, and we built the inverted index from the automatic summaries that were generated from this model. A series of experiments were held to test the performance in terms of efficiency and relevancy. The experiments include comparisons with three existing text summarization models; the Jaccard Coefficient Model, the Vector Space Model, and the Latent Semantic Analysis model. The experiments examined three groups of queries with manual and automatic relevancy assessment. The positive effect of the Multi-Layer Similarity in the efficiency of the IR system was clear without noticeable loss in the relevancy results. However, the evaluation showed that the traditional statistical models without semantic investigation failed to improve the information retrieval efficiency. Comparing with the previous publications that addressed the use of summaries as a source of the index, the relevancy assessment of our work was higher, and the Multi-Layer Similarity retrieval constructed an inverted index that was 58% smaller than the main corpus inverted index. ",The Effect of the Multi-Layer Text Summarization Model on the Efficiency and Relevancy of the Vector Space-based Information Retrieval " The purpose of the present research is to investigate model mixed boundary value problems for the Helmholtz equation in a planar angular domain $\Omega_\alpha\subset\mathbb{R}^2$ of magnitude $\alpha$. The BVP is considered in a non-classical setting when a solution is sought in the Bessel potential spaces $\mathbb{H}^s_p(\Omega_\alpha)$, $s>1/p$, $1 0. From our simulations we find at low magnetic fields that the Nernst signal displays a characteristic ""tilted hill"" profile similar to experiments and a non-monotonic temperature dependence of the heat conductivity. We also investigate the effects of granularity and randomness, which become important for higher magnetic fields. In this regime geometric frustration strongly influences the results in both regular and random systems and leads to highly non-trivial magnetic field dependencies of the studied transport coefficients. ",Influence of vortices and phase fluctuations on thermoelectric transport properties of superconductors in a magnetic field " In this work, we have investigated a 2D model of band-to-band tunneling based on 2-band model and implemented it using 2D NEGF formalism. Being 2D in nature, this model better addresses the variation in the directionality of the tunneling process occurring in most practical TFET device structures. It also works as a compromise between semi-classical and multiband quantum simulation of TFETs. In this work, we have presented a sound step by step mathematical development of the numerical model. We have also discussed how this model can be implemented in simulators and pointed out a few optimizations that can be made to reduce complexity and to save time. Finally, we have performed elaborate simulations for a practical TFET design and compared the results with commercially available TCAD simulations, to point out the limitations of the simplistic models that are frequently used, and how our model overcomes these limitations. ",An Improved Physics Based Numerical Model of Tunnel FET Using 2D NEGF Formalism " We aim to improve our picture of the low chromosphere in the quiet-Sun internetwork by investigating the intensity, horizontal velocity, size and lifetime variations of small bright points (BPs; diameter smaller than 0.3 arcsec) observed in the Ca II H 3968 {\AA} passband along with their magnetic field parameters, derived from photospheric magnetograms. Several high-quality time series of disc-centre, quiet-Sun observations from the Sunrise balloon-borne solar telescope, with spatial resolution of around 100 km on the solar surface, have been analysed to study the dynamics of BPs observed in the Ca II H passband and their dependence on the photospheric vector magnetogram signal. Parameters such as horizontal velocity, diameter, intensity and lifetime histograms of the isolated internetwork and magnetic Ca II H BPs were determined. Mean values were found to be 2.2 km/s, 0.2 arcsec (150 km), 1.48 average Ca II H quiet-Sun and 673 sec, respectively. Interestingly, the brightness and the horizontal velocity of BPs are anti-correlated. Large excursions (pulses) in horizontal velocity, up to 15 km/s, are present in the trajectories of most BPs. These could excite kink waves travelling into the chromosphere and possibly the corona, which we estimate to carry an energy flux of 310 W/m^2, sufficient to heat the upper layers, although only marginally. The stable observing conditions of Sunrise and our technique for identifying and tracking BPs have allowed us to determine reliable parameters of these features in the internetwork. Thus we find, e.g., that they are considerably longer lived than previously thought. The large velocities are also reliable, and may excite kink waves. Although these wave are (marginally) energetic enough to heat the quiet corona, we expect a large additional contribution from larger magnetic elements populating the network and partly also the internetwork. ",Structure and Dynamics of Isolated Internetwork Ca II H Bright Points Observed by Sunrise " The direct imaging from the ground of extrasolar planets has become today a major astronomical and biological focus. This kind of imaging requires simultaneously the use of a dedicated high performance Adaptive Optics [AO] system and a differential imaging camera in order to cancel out the flux coming from the star. In addition, the use of sophisticated post-processing techniques is mandatory to achieve the ultimate detection performance required. In the framework of the SPHERE project, we present here the development of a new technique, based on Maximum A Posteriori [MAP] approach, able to estimate parameters of a faint companion in the vicinity of a bright star, using the multi-wavelength images, the AO closed-loop data as well as some knowledge on non-common path and differential aberrations. Simulation results show a 10^-5 detectivity at 5sigma for angular separation around 15lambda/D with only two images. ",Post processing of differential images for direct extrasolar planet detection from the ground " Mitigating the risk arising from extreme events is a fundamental goal with many applications, such as the modelling of natural disasters, financial crashes, epidemics, and many others. To manage this risk, a vital step is to be able to understand or generate a wide range of extreme scenarios. Existing approaches based on Generative Adversarial Networks (GANs) excel at generating realistic samples, but seek to generate typical samples, rather than extreme samples. Hence, in this work, we propose ExGAN, a GAN-based approach to generate realistic and extreme samples. To model the extremes of the training distribution in a principled way, our work draws from Extreme Value Theory (EVT), a probabilistic approach for modelling the extreme tails of distributions. For practical utility, our framework allows the user to specify both the desired extremeness measure, as well as the desired extremeness probability they wish to sample at. Experiments on real US Precipitation data show that our method generates realistic samples, based on visual inspection and quantitative measures, in an efficient manner. Moreover, generating increasingly extreme examples using ExGAN can be done in constant time (with respect to the extremeness probability $\tau$), as opposed to the $\mathcal{O}(\frac{1}{\tau})$ time required by the baseline approach. ",ExGAN: Adversarial Generation of Extreme Samples " Charge density wave (CDW) correlations have recently been shown to universally exist in cuprate superconductors. However, their nature at high fields inferred from nuclear magnetic resonance is distinct from that measured by x-ray scattering at zero and low fields. Here we combine a pulsed magnet with an x-ray free electron laser to characterize the CDW in YBa2Cu3O6.67 via x-ray scattering in fields up to 28 Tesla. While the zero-field CDW order, which develops below T ~ 150 K, is essentially two-dimensional, at lower temperature and beyond 15 Tesla, another three-dimensionally ordered CDW emerges. The field-induced CDW onsets around the zero-field superconducting transition temperature, yet the incommensurate in-plane ordering vector is field-independent. This implies that the two forms of CDW and high-temperature superconductivity are intimately linked. ",Three-Dimensional Charge Density Wave Order in YBa2Cu3O6.67 at High Magnetic Fields " The (heterotic) double field theories and the exceptional field theories are recently developed for manifestly duality covariant formulation of various supergravity theories, describing low-energy limit of various (heterotic) superstring and M-theory compactifications. These field theories are known to be reduced to the standard descriptions by introducing appropriately parameterized generalized metric and by applying suitably chosen section conditions. We generalize this development to non-geometric backgrounds by utilizing dual fields pertinent to non-geometric fluxes. We introduce different parameterizations for the generalized metric, in terms of the conventional supergravity fields or the dual fields. Under certain simplifying assumptions, we construct new effective action for non-geometric backgrounds. We then obtain the non-geometric backgrounds sourced by exotic branes. From them, we construct their $U$-duality monodromy matrices. The charge of exotic branes obtained from these monodromy matrices agrees perfectly with the charge obtained from the non-geometric flux integral. ",Effective Action for Non-Geometric Fluxes from Duality Covariant Actions " Smart power grids are one of the most complex cyber-physical systems, delivering electricity from power generation stations to consumers. It is critically important to know exactly the current state of the system as well as its state variation tendency; consequently, state estimation and state forecasting are widely used in smart power grids. Given that state forecasting predicts the system state ahead of time, it can enhance state estimation because state estimation is highly sensitive to measurement corruption due to the bad data or communication failures. In this paper, a hybrid deep learningbased method is proposed for power system state forecasting. The proposed method leverages Convolutional Neural Network (CNN) for predicting voltage magnitudes and a Deep Recurrent Neural Network (RNN) for predicting phase angels. The proposed CNN-RNN model is evaluated on the IEEE 118-bus benchmark. The results demonstrate that the proposed CNNRNN model achieves better results than the existing techniques in the literature by reducing the normalized Root Mean Squared Error (RMSE) of predicted voltages by 10%. The results also show a 65% and 35% decrease in the average and maximum absolute error of voltage magnitude forecasting. ",A Hybrid Deep Learning-Based State Forecasting Method for Smart Power Grids " As language models (LMs) become increasingly powerful, it is important to quantify and compare them for sociodemographic bias with potential for harm. Prior bias measurement datasets are sensitive to perturbations in their manually designed templates, therefore unreliable. To achieve reliability, we introduce the Comprehensive Assessment of Language Model bias (CALM), a benchmark dataset to quantify bias in LMs across three tasks. We integrate 16 existing datasets across different domains, such as Wikipedia and news articles, to filter 224 templates from which we construct a dataset of 78,400 examples. We compare the diversity of CALM with prior datasets on metrics such as average semantic similarity, and variation in template length, and test the sensitivity to small perturbations. We show that our dataset is more diverse and reliable than previous datasets, thus better capture the breadth of linguistic variation required to reliably evaluate model bias. We evaluate 20 large language models including six prominent families of LMs such as Llama-2. In two LM series, OPT and Bloom, we found that larger parameter models are more biased than lower parameter models. We found the T0 series of models to be the least biased. Furthermore, we noticed a tradeoff between gender and racial bias with increasing model size in some model series. The code is available at https://github.com/vipulgupta1011/CALM. ",CALM : A Multi-task Benchmark for Comprehensive Assessment of Language Model Bias " Alphanumeric authentication routinely fails to regulate access to resources with the required stringency, primarily due to usability issues. Initial deployment did not reveal the problems of passwords, deep and profound flaws only emerged once passwords were deployed in the wild. The need for a replacement is widely acknowledged yet despite over a decade of research into knowledge-based alternatives, few, if any, have been adopted by industry. Alternatives are unconvincing for three primary reasons. The first is that alternatives are rarely investigated beyond the initial proposal, with only the results from a constrained lab test provided to convince adopters of their viability. The second is that alternatives are seldom tested realistically where the authenticator mediates access to something of value. The third is that the testing rarely varies the device or context beyond that initially targeted. In the modern world different devices are used across a variety of contexts. What works well in one context may easily fail in another. Consequently, the contribution of this paper is an ""in the wild"" evaluation of an alternative authentication mechanism that had demonstrated promise in its lab evaluation. In the field test the mechanism was deployed to actual users to regulate access to an application in a context beyond that initially proposed. The performance of the mechanism is reported and discussed. We conclude by reflecting on the value of field evaluations of alternative authentication mechanisms. ",Alternative Authentication in the Wild " We study a discrete-time model where each packet has a cost of not being sent -- this cost might depend on the packet content. We study the tradeoff between the age and the cost where the sender is confined to packet-based strategies. The optimal tradeoff is found by an appropriate formulation of the problem as a Markov Decision Process (MDP). We show that the optimal tradeoff can be attained with finite-memory policies and we devise an efficient policy iteration algorithm to find these optimal policies. We further study a related problem where the transmitted packets are subject to erasures. We show that the optimal policies for our problem are also optimal for this new setup. Allowing coding across packets significantly extends the packet-based strategies. We show that when the packet payloads are small, the performance can be improved by coding. ",Optimal Policies for Age and Distortion in a Discrete-Time Model " In this work we explore the intriguing connections between searches for long-lived particles (LLPs) at the LHC and early universe cosmology. We study the non-thermal production of ultra-relativistic particles (i.e. dark radiation) in the early universe via the decay of weak-scale LLPs and show that the cosmologically interesting range $\Delta N_\text{eff} \sim 0.01-0.1$ corresponds to LLP decay lengths in the mm to cm range. These decay lengths lie at the boundary between prompt and displaced signatures at the LHC and can be comprehensively explored by combining searches for both. To illustrate this point, we consider a scenario where the LLP decays into a charged lepton and a (nearly) massless invisible particle. By reinterpreting searches for promptly decaying sleptons and for displaced leptons at both ATLAS and CMS we can then directly compare LHC exclusions with cosmological observables. We find that the CMB-S4 target value of $\Delta N_\text{eff}=0.06$ is already excluded by current LHC searches and even smaller values can be probed for LLP masses at the electroweak scale. ",Searching for dark radiation at the LHC " Firstly, we study the final masses of giant planets growing in protoplanetary disks through capture of disk gas, by employing an empirical formula for the gas capture rate and a shallow disk gap model, which are both based on hydrodynamical simulations. The shallow disk gaps cannot terminate growth of giant planets. For planets less massive than 10 Jupiter masses, their growth rates are mainly controlled by the gas supply through the global disk accretion, rather than their gaps. The insufficient gas supply compared with the rapid gas capture causes a depletion of the gas surface density even at the outside of the gap, which can create an inner hole in the protoplanetary disk. Our model can also predict the depleted gas surface density in the inner hole for a given planet mass. Secondly, our findings are applied to the formation of our solar system. For the formation of Jupiter, a very low-mass gas disk with a few or several Jupiter masses is required at the beginning of its gas capture because of the non-stopping capture. Such a low-mass gas disk with sufficient solid material can be formed through viscous evolution from an initially $\sim$10AU-sized compact disk with the solar composition. By the viscous evolution with a moderate viscosity of $\alpha \sim 10^{-3}$, most of disk gas accretes onto the sun and a widely spread low-mass gas disk remains when the solid core of Jupiter starts gas capture at $t \sim 10^7$ yrs. The depletion of the disk gas is suitable for explaining the high metallicity in giant planets of our solar system. A very low-mass gas disk also provides a plausible path where type I and II planetary migrations are both suppressed significantly. In particular, we also show that the type II migration of Jupiter-size planets becomes inefficient because of the additional gas depletion due to the rapid gas capture by themselves. ",Final Masses of Giant Planets II: Jupiter Formation in a Gas-Depleted Disk " (Abridged) We have used the VLA to search for neutral atomic hydrogen in the circumstellar envelopes of five AGB stars. We have detected HI 21-cm emission coincident in both position and velocity with the semi-regular variable RS Cnc. The emission comprises a compact, slightly elongated feature centered on the star with a mean diameter ~82'' (1.5e17 cm), plus an additional filament extending ~6' to the NW. This filament suggests that a portion of the mass loss is highly asymmetric. We estimate MHI=1.5e-3 Msun and M_dot~1.7e-7 Msun/yr. Toward R Cas, we detect weak emission that peaks at the stellar systemic velocity and overlaps with the location of its circumstellar dust shell and thus is probably related to the star. In the case of IRC+10216, we were unable to confirm the detection of HI in absorption against the cosmic background previously reported by Le Bertre & Gerard. However, we detect arcs of emission at projected distances of r~14'-18' (~2e18 cm) to the NW. A large separation of the emission from the star is plausible given its advanced evolutionary status, although it is unclear if the asymmetric distribution and complex velocity structure are consistent with a circumstellar origin. For EP Aqr, we detected HI emission comprising multiple clumps redward of the systemic velocity, but we are unable to determine unambiguously whether the emission arises from the circumstellar envelope or from interstellar clouds along the line-of-sight. Regardless of the adopted distance for the clumps, their inferred HI masses are at least an order of magnitude smaller than their individual binding masses. We detected our fifth target, R Aqr (a symbiotic binary), in the 1.4 GHz continuum, but did not detect any HI emission from the system. ",VLA Observations of HI in the Circumstellar Envelopes of Asymptotic Giant Branch Stars " The azimuthal correlations of D mesons and charged particles were measured with the ALICE detector in pp collisions at $\sqrt{s}=7$ TeV and p-Pb collisions at $\sqrt{s_{\rm NN}}=5.02$ TeV at the Large Hadron Collider. D$^0$, D$^+$, and D$^{*+}$ mesons and their charge conjugates with transverse momentum $30.3$ Gev/$c$. The properties of the correlation peak induced by the jet containing the D meson, described in terms of the yield of charged particles in the peak and peak width, are compatible within uncertainties between the two collision systems, and described by Monte-Carlo simulations based on the PYTHIA, POWHEG and EPOS 3 event generators. ",Measurement of azimuthal correlations of D mesons and charged particles in pp collisions at $\sqrt{s}=7$ TeV and p-Pb collisions at $\sqrt{s_{\rm NN}}=5.02$ TeV " We present a potential-density pair designed to model nearly isothermal star clusters (and similar self-gravitating systems) with a central core and an outer turnover radius, beyond which density falls off as $r^{-4}$. In the intermediate zone, the profile is similar to that of an isothermal sphere (density $\rho \propto r^{-2}$), somewhat less steep than the King 62 profile, and with the advantage that many dynamical quantities can be written in a simple closed form. We derive new analytic expressions for the cluster binding energy and velocity dispersion, and apply these to create toy models for cluster core collapse and evaporation. We fit our projected surface brightness profiles to observed globular and open clusters, and find that the quality of the fit is generally at least as good as that for the surface brightness profiles of King 62. This model can be used for convenient computation of the dynamics and evolution of globular and nuclear star clusters. ",A Dynamical Potential-Density Pair for Star Clusters With Nearly Isothermal Interiors " Recently, convolutional neural network (CNN) has demonstrated significant success for image restoration (IR) tasks (e.g., image super-resolution, image deblurring, rain streak removal, and dehazing). However, existing CNN based models are commonly implemented as a single-path stream to enrich feature representations from low-quality (LQ) input space for final predictions, which fail to fully incorporate preceding low-level contexts into later high-level features within networks, thereby producing inferior results. In this paper, we present a deep interleaved network (DIN) that learns how information at different states should be combined for high-quality (HQ) images reconstruction. The proposed DIN follows a multi-path and multi-branch pattern allowing multiple interconnected branches to interleave and fuse at different states. In this way, the shallow information can guide deep representative features prediction to enhance the feature expression ability. Furthermore, we propose asymmetric co-attention (AsyCA) which is attached at each interleaved node to model the feature dependencies. Such AsyCA can not only adaptively emphasize the informative features from different states, but also improves the discriminative ability of networks. Our presented DIN can be trained end-to-end and applied to various IR tasks. Comprehensive evaluations on public benchmarks and real-world datasets demonstrate that the proposed DIN perform favorably against the state-of-the-art methods quantitatively and qualitatively. ",Learning Deep Interleaved Networks with Asymmetric Co-Attention for Image Restoration " We solve the problem of expressing the Weyl scalars $\psi $ that describe gravitational perturbations of a Kerr black hole in terms of Cauchy data. To do so we use geometrical identities (like the Gauss-Codazzi relations) as well as Einstein equations. We are able to explicitly express $\psi $ and $\partial _t\psi $ as functions only of the extrinsic curvature and the three-metric (and geometrical objects built out of it) of a generic spacelike slice of the spacetime. These results provide the link between initial data and $\psi $ to be evolved by the Teukolsky equation, and can be used to compute the gravitational radiation generated by two orbiting black holes in the close limit approximation. They can also be used to extract waveforms from spacetimes completely generated by numerical methods. ",The imposition of Cauchy data to the Teukolsky equation III: The rotating case " Recent advances in classical machine learning have shown that creating models with inductive biases encoding the symmetries of a problem can greatly improve performance. Importation of these ideas, combined with an existing rich body of work at the nexus of quantum theory and symmetry, has given rise to the field of Geometric Quantum Machine Learning (GQML). Following the success of its classical counterpart, it is reasonable to expect that GQML will play a crucial role in developing problem-specific and quantum-aware models capable of achieving a computational advantage. Despite the simplicity of the main idea of GQML -- create architectures respecting the symmetries of the data -- its practical implementation requires a significant amount of knowledge of group representation theory. We present an introduction to representation theory tools from the optics of quantum learning, driven by key examples involving discrete and continuous groups. These examples are sewn together by an exposition outlining the formal capture of GQML symmetries via ""label invariance under the action of a group representation"", a brief (but rigorous) tour through finite and compact Lie group representation theory, a reexamination of ubiquitous tools like Haar integration and twirling, and an overview of some successful strategies for detecting symmetries. ",Representation Theory for Geometric Quantum Machine Learning " We investigate the many-body properties of a two-dimensional electron gas constrained to the surface of a sphere, a system which is physically realized in multielectron bubbles in liquid helium. A second-quantization formalism, suited for the treatment of a spherical two-dimensional electron gas (S2DEG), is introduced. Within this formalism, the dielectric response properties of the S2DEG are derived, and we identify both collective excitations and a spectrum of single-particle excitations. We find that the single-particle excitations are constrained to a well-defined region in the angular momentum - energy plane. The collective excitations differ in two important aspects from those of a flat 2DEG: on a sphere, the 'spherical plasmons' have a discrete frequency spectrum and the lowest frequency is nonzero. ",On the Spherical Two-Dimensional Electron Gas " We introduce a code generator that converts unoptimized C++ code operating on sparse data into vectorized and parallel CPU or GPU kernels. Our approach unrolls the computation into a massive expression graph, performs redundant expression elimination, grouping, and then generates an architecture-specific kernel to solve the same problem, assuming that the sparsity pattern is fixed, which is a common scenario in many applications in computer graphics and scientific computing. We show that our approach scales to large problems and can achieve speedups of two orders of magnitude on CPUs and three orders of magnitude on GPUs, compared to a set of manually optimized CPU baselines. To demonstrate the practical applicability of our approach, we employ it to optimize popular algorithms with applications to physical simulation and interactive mesh deformation. ",Sparsity-Specific Code Optimization using Expression Trees A coloring of the $\ell$-dimensional faces of $Q_n$ is called $d$-polychromatic if every embedded $Q_d$ has every color on at least one face. Denote by $p^\ell(d)$ the maximum number of colors such that any $Q_n$ can be colored in this way. We provide a new lower bound on $p^\ell(d)$ for $\ell > 1$. ,Linear polychromatic colorings of hypercube faces " The development of renewable energy generation empowers microgrids to generate electricity to supply itself and to trade the surplus on energy markets. To minimize the overall cost, a microgrid must determine how to schedule its energy resources and electrical loads and how to trade with others. The control decisions are influenced by various factors, such as energy storage, renewable energy yield, electrical load, and competition from other microgrids. Making the optimal control decision is challenging, due to the complexity of the interconnected microgrids, the uncertainty of renewable energy generation and consumption, and the interplay among microgrids. The previous works mainly adopted the modeling-based approaches for deriving the control decision, yet they relied on the precise information of future system dynamics, which can be hard to obtain in a complex environment. This work provides a new perspective of obtaining the optimal control policy for distributed energy trading and scheduling by directly interacting with the environment, and proposes a multiagent deep reinforcement learning approach for learning the optimal control policy. Each microgrid is modeled as an agent, and different agents learn collaboratively for maximizing their rewards. The agent of each microgrid can make the local scheduling decision without knowing others' information, which can well maintain the autonomy of each microgrid. We evaluate the performances of our proposed method using real-world datasets. The experimental results show that our method can significantly reduce the cost of the microgrids compared with the baseline methods. ",Distributed Energy Trading and Scheduling among Microgrids via Multiagent Reinforcement Learning " We consider an inverse source two-parameter sub-diffusion model subject to a nonlocal initial condition. The problem models several physical processes, among them are the microwave heating and light propagation in photoelectric cells. A bi-orthogonal pair of bases is employed to construct a series representation of the solution and a Volterra integral equation for the source term. We develop a numerical algorithm for approximating the unknown time-dependent source term. Due to the singularity of the solution near $t=0$, a graded mesh is used to improve the convergence rate. Numerical experiments are provided to illustrate the expected analytical order of convergence. ","Inverse source in two-parameter anomalous diffusion, numerical algorithms and simulations over graded time-meshes" " Fe3-xGeTe2 is an itinerant ferromagnet composed of two-dimensional layers weakly connected by van der Waals bonding that shows a variety of intriguing phenomena. Inelastic neutron scattering measurements on bulk single crystals of Fe2.75GeTe2 were performed to quantify the magnetic exchange interaction energies and anisotropy. The observed inelastic excitations are indicative of dominant in-plane correlations with negligible magnetic interactions between the layers. A spin-gap of 3.9 meV is observed allowing a measure of the magnetic anisotropy. As the excitations disperse to their maximum energy (~65 meV) they become highly damped, reflective of both the magnetic site occupancy reduction of 25{\%} on one Fe sublattice and the itinerant interactions. A minimal model is employed to describe the excitation spectra and extract nearest neighbor magnetic exchange interaction values. The temperature evolution of the excitations are probed and correlations shown to persist above Tc, indicative of low dimensional magnetism. ",Magnetic excitations in the quasi-2D ferromagnet Fe3-xGeTe2 measured with inelastic neutron scattering " We establish a strong-weak coupling duality between two types of free matrix models. In the large-N limit, the real-symmetric matrix model is dual to the quaternionic-real matrix model. Using the large-N conformal invariant collective field formulation, the duality is displayed in terms of the generators of the conformal group. The conformally invariant master Hamiltonian is constructed and we conjecture that the master Hamiltonian corresponds to the hermitian matrix model. ",Matrix-model dualities in the collective field formulation " Many radio pulsars have stable pulse profiles, but some exhibit mode changing where the profile switches between two or more quasi-stable modes of emission. So far, these effects had only been seen in relatively slow pulsars, but we show here that the pulse profile of PSR B1957+20, a millisecond pulsar, switches between two modes, with a typical time between mode changes of only $1.7$ s (or $\sim\!1000$ rotations), the shortest observed so far. The two modes differ in both intensity and polarization, with relatively large differences in the interpulse and much more modest ones in the main pulse. We find that the changes in the interpulse precede those in the main pulse by $\sim\!25$ ms, placing an empirical constraint on the timescale over which mode changes occurs. We also find that the properties of the giant pulses emitted by PSR B1957+20 are correlated with the mode of the regular emission: their rate and the rotational phase at which they are emitted both depend on mode. Furthermore, the energy distribution of the giant pulses emitted near the main pulse depends on mode as well. We discuss the ramifications for our understanding of the radio emission mechanisms as well as for pulsar timing experiments. ",Mode changing and giant pulses in the millisecond pulsar PSR B1957+20 " Heisenberg Hamiltonian was employed to describe the variation of energy of thick ferromagnetic films with second and fourth order anisotropies. At angle of 1.36 degrees and anisotropies of 1.25, energy is minimum for thick film of sc(001) with 1000 layers. Energy becomes minimum at angle of 1.18 degrees and fourth order anisotropy of 1.15 for thick film of bcc(001) with the same thickness. According to these simulations, these lattices can be easily oriented in some certain directions under the influence of some particular values of anisotropies. Energy varies with second and fourth order anisotropies in similar passion for both types of lattices. The energy gradually decreases with second and fourth order anisotropy for both types of lattices in the range described here. ",Magnetic anisotropy dependence of the energy of oriented thick ferromagnetic films " This paper presents an algorithm for estimating the weight of a maximum weighted matching by augmenting any estimation routine for the size of an unweighted matching. The algorithm is implementable in any streaming model including dynamic graph streams. We also give the first constant estimation for the maximum matching size in a dynamic graph stream for planar graphs (or any graph with bounded arboricity) using $\tilde{O}(n^{4/5})$ space which also extends to weighted matching. Using previous results by Kapralov, Khanna, and Sudan (2014) we obtain a $\mathrm{polylog}(n)$ approximation for general graphs using $\mathrm{polylog}(n)$ space in random order streams, respectively. In addition, we give a space lower bound of $\Omega(n^{1-\varepsilon})$ for any randomized algorithm estimating the size of a maximum matching up to a $1+O(\varepsilon)$ factor for adversarial streams. ",Sublinear Estimation of Weighted Matchings in Dynamic Data Streams " Considering the coupling of color $1 \otimes 1$ and $8 \otimes 8$ structures, we calculate the energy of the newly observed $X$(3915) as $S-$wave $D^*\bar{D^*}$ state in the Bhaduri, Cohler, and Nogami quark model by the Gaussian Expansion Method. Due to the color coupling, the bound state of $D^*\bar{D^*}$ with $J^{PC}=0^{++}$ is found, which is well consonant with the experimental data of the $X$(3915). The bound state of $B^*\bar{B^*}$ with $J^{PC}=0^{++}$ and $2^{++}$ are also predicted in this work. ",Dynamical study of the $X$(3915) as a molecular $D^*\bar{D^*}$ state in a quark model " We consider the six-vertex model with Domain Wall Boundary Conditions. Our main interest is the study of the fluctuations of the extremal lattice path about the arctic curves. We address the problem through Monte Carlo simulations. At $\Delta = 0$, the fluctuations of the extremal path along any line parallel to the square diagonal were rigorously proven to follow the Tracy-Widom distribution. We provide strong numerical evidence that this is true also for other values of the anisotropy parameter $\Delta$ ($0\leq \Delta < 1$). We argue that the typical width of the fluctuations of the extremal path about the arctic curves scales as $N^{1/3}$ and provide a numerical estimate for the parameters of the scaling random variable. ",Fluctuation of the phase boundary in the six-vertex model with Domain Wall Boundary Conditions: a Monte Carlo study " Many transformations in deep learning architectures are sparsely connected. When such transformations cannot be designed by hand, they can be learned, even through plain backpropagation, for instance in attention mechanisms. However, during learning, such sparse structures are often represented in a dense form, as we do not know beforehand which elements will eventually become non-zero. We introduce the adaptive, sparse hyperlayer, a method for learning a sparse transformation, paramatrized sparsely: as index-tuples with associated values. To overcome the lack of gradients from such a discrete structure, we introduce a method of randomly sampling connections, and backpropagating over the randomly wired computation graph. To show that this approach allows us to train a model to competitive performance on real data, we use it to build two architectures. First, an attention mechanism for visual classification. Second, we implement a method for differentiable sorting: specifically, learning to sort unlabeled MNIST digits, given only the correct order. ",Learning sparse transformations through backpropagation " Self-supervised representation learning solves auxiliary prediction tasks (known as pretext tasks) without requiring labeled data to learn useful semantic representations. These pretext tasks are created solely using the input features, such as predicting a missing image patch, recovering the color channels of an image from context, or predicting missing words in text; yet predicting this \textit{known} information helps in learning representations effective for downstream prediction tasks. We posit a mechanism exploiting the statistical connections between certain {\em reconstruction-based} pretext tasks that guarantee to learn a good representation. Formally, we quantify how the approximate independence between the components of the pretext task (conditional on the label and latent variables) allows us to learn representations that can solve the downstream task by just training a linear layer on top of the learned representation. We prove the linear layer yields small approximation error even for complex ground truth function class and will drastically reduce labeled sample complexity. Next, we show a simple modification of our method leads to nonlinear CCA, analogous to the popular SimSiam algorithm, and show similar guarantees for nonlinear CCA. ",Predicting What You Already Know Helps: Provable Self-Supervised Learning " We have obtained high-resolution spectra and carried out a detailed elemental abundance analysis for a new sample of 899 F and G dwarf stars in the Solar neighbourhood. The results allow us to, in a multi-dimensional space consisting of stellar ages, detailed elemental abundances, and full kinematic information for the stars, study and trace their respective origins. Here we briefly address selection criteria and discuss how to define a thick disc star. The results are discussed in the context of galaxy formation. ",The Galactic thin and thick discs in the context of galaxy formation " This paper presents a critical review of particle production in an uniform electric field and Schwarzchild-like spacetimes. Both problems can be reduced to solving an effective one-dimensional Schrodinger equation with a potential barrier. In the electric field case, the potential is that of an inverted oscillator -x^2 while in the case of Schwarchild-like spacetimes, the potential is of the form -1/x^2 near the horizon. The transmission and reflection coefficients can easily be obtained for both potentials. To describe particle production, these coefficients have to be suitably interpreted. In the case of the electric field, the standard Bogoliubov coefficients can be identified and the standard gauge invariant result is recovered. However, for Schwarzchild-like spacetimes, such a tunnelling interpretation appears to be invalid. The Bogoliubov coefficients cannot be determined by using an identification process similar to that invoked in the case of the electric field. The reason for such a discrepancy appears to be that, in the tunnelling method, the effective potential near the horizon is singular and symmetric. We also provide a new and simple semi-classical method of obtaining Hawking's result in the (t,r) co-ordinate system of the usual standard Schwarzchild metric. We give a prescription whereby the singularity at the horizon can be regularised with Hawking's result being recovered. This regularisation prescription contains a fundamental asymmetry that renders both sides of the horizon dissimilar. Finally, we attempt to interpret particle production by the electric field as a tunnelling process between the two sectors of the Rindler metric. ",Facets of Tunneling: Particle production in external fields " Forwarding table verification consists in checking the distributed data-structure resulting from the forwarding tables of a network. A classical concern is the detection of loops. We study this problem in the context of software-defined networking (SDN) where forwarding rules can be arbitrary bitmasks (generalizing prefix matching) and where tables are updated by a centralized controller. Basic verification problems such as loop detection are NP-hard and most previous work solves them with heuristics or SAT solvers. We follow a different approach based on computing a representation of the header classes, i.e. the sets of headers that match the same rules. This representation consists in a collection of representative header sets, at least one for each class, and can be computed centrally in time which is polynomial in the number of classes. Classical verification tasks can then be trivially solved by checking each representative header set. In general, the number of header classes can increase exponentially with header length, but it remains polynomial in the number of rules in the practical case where rules are constituted with predefined fields where exact, prefix matching or range matching is applied in each field (e.g., IP/MAC addresses, TCP/UDP ports). We propose general techniques that work in polynomial time as long as the number of classes of headers is polynomial and that do not make specific assumptions about the structure of the sets associated to rules. The efficiency of our method rely on the fact that the data-structure representing rules allows efficient computation of intersection, cardinal and inclusion. Finally, we propose an algorithm to maintain such representation in presence of updates (i.e., rule insert/update/removal). We also provide a local distributed algorithm for checking the absence of black-holes and a proof labeling scheme for locally checking the absence of loops. ",Forwarding Tables Verification through Representative Header Sets " Improving the film quality in the synthesis of large-area hexagonal boron nitride films (h-BN) for two-dimensional material devices remains a great challenge. The measurement of electrical breakdown dielectric strength (EBD) is one of the most important methods to elucidate the insulating quality of h-BN. In this work, the EBD of high quality exfoliated single-crystal h-BN was investigated using three different electrode structures under different environmental conditions to determine the ideal electrode structure and environment for EBD measurement. A systematic investigation revealed that EBD is not sensitive to contact force or electrode area but strongly depends on the relative humidity during measurement. Once the measurement environment is properly managed, it was found that the EBD values are consistent within experimental error regardless of the electrode structure, which enables the evaluation of the crystallinity of synthesized h-BN at the microscopic and macroscopic level by utilizing the three different electrode structures properly for different purposes. ",Comparison of device structures for the dielectric breakdown measurement of hexagonal boron nitride " I have proposed a measure for the cage effect in glass forming systems. A binary mixture of hard disks is numerically studied as a model glass former. A network is constructed on the basis of the colliding pairs of disks. A rigidity matrix is formed from the isostatic (rigid) sub--network, corresponding to a cage. The determinant of the matrix changes its sign when an uncaging event occurs. Time evolution of the number of the uncaging events is determined numerically. I have found that there is a gap in the uncaging timescales between the cages involving different numbers of disks. Caging of one disk by two neighboring disks sustains for a longer time as compared with other cages involving more than one disk. This gap causes two--step relaxation of this system. ",Binary mixture of hard disks as a model glass former: Caging and uncaging " A multispeckle technique for efficiently measuring correctly ensemble-averaged intensity autocorrelation functions of scattered light from non-ergodic and/or non-stationary systems is described. The method employs a CCD camera as a multispeckle light detector and a computer-based correlator, and permits the simultaneous calculation of up to 500 correlation functions, where each correlation function is started at a different time. The correlation functions are calculated in real time and are referenced to a unique starting time. The multispeckle nature of the CCD camera detector means that a true ensemble average is calculated; no time averaging is necessary. The technique thus provides a ""snapshot"" of the dynamics, making it particularly useful for non-stationary systems where the dynamics are changing with time. Delay times spanning the range from 1 ms to 1000 s are readily achieved with this method. The technique is demonstrated in the multiple scattering limit where diffusing-wave spectroscopy theory applies. The technique can also be combined with a recently-developed two-cell technique that can measure faster decay times. The combined technique can measure delay times from 10 ns to 1000 s. The method is peculiarly well suited for studying aging processes in soft glassy materials, which exhibit both short and long relaxation times, non-ergodic dynamics, and slowly-evolving transient behavior. ",Multispeckle diffusing-wave spectroscopy: a tool to study slow relaxation and time-dependent dynamics " This paper describes the dialog robot system designed by Team Irisapu for the preliminary round of the Dialogue Robot Competition 2022 (DRC2022). Our objective was to design a hospitable travel agent robot. The system we developed was ranked 8th out of 13 systems in the preliminary round of the competition, but our robot received high marks for its naturalness and likeability.Our next challenge is to create a system that can provide more useful information to users. ",Hospitable Travel Agent Dialogue Robot: Team Irisapu Project Description for DRC2022 " In this article we prove explicit formulae for the number of non-isomorphic cluster-tilted algebras of type \tilde{A}_n in the derived equivalence classes. In particular, we obtain the number of elements in the mutation classes of quivers of type \tilde{A}_n. As a by-product, this provides an alternative proof for the number of quivers of Dynkin type D_n which was first determined by Buan and Torkildsen. ",Counting the number of elements in the mutation classes of \tilde{A}_n-quivers " We show that a coupling of non-colliding simple random walkers on the complete graph on $n$ vertices can include at most $n - \log n$ walkers. This improves the only previously known upper bound of $n-2$ due to Angel, Holroyd, Martin, Wilson, and Winkler ({\it Electron.~Commun.~Probab.~18}, 2013). The proof considers couplings of i.i.d.~sequences of Bernoulli random variables satisfying a similar avoidance property, for which there is separate interest. Our bound in this setting should be closer to optimal. ",An upper bound on the size of avoidance couplings " The efficient and accurate simulation of material systems with defects using atomistic- to-continuum (a/c) coupling methods is a topic of considerable interest in the field of computational materials science. To achieve the desired balance between accuracy and computational efficiency, the use of a posteriori analysis and adaptive algorithms is critical. In this work, we present a rigorous a posteriori error analysis for three typical blended a/c coupling methods: the blended energy-based quasi-continuum (BQCE) method, the blended force-based quasi-continuum (BQCF) method, and the atomistic/continuum blending with ghost force correction (BGFC) method. We employ first and second-order finite element methods (and potentially higher-order methods) to discretize the Cauchy-Born model in the continuum region. The resulting error estimator provides both an upper bound on the true approximation error and a lower bound up to a theory-based truncation indicator, ensuring its reliability and efficiency. Moreover, we propose an a posteriori analysis for the energy error. We have designed and implemented a corresponding adaptive mesh refinement algorithm for two typical examples of crystalline defects. In both numerical experiments, we observe optimal convergence rates with respect to degrees of freedom when compared to a priori error estimates. ",A Posteriori Analysis and Adaptive Algorithms for Blended Type Atomistic-to-Continuum Coupling with Higher-Order Finite Elements " Subsampling is effective in Knowledge Graph Embedding (KGE) for reducing overfitting caused by the sparsity in Knowledge Graph (KG) datasets. However, current subsampling approaches consider only frequencies of queries that consist of entities and their relations. Thus, the existing subsampling potentially underestimates the appearance probabilities of infrequent queries even if the frequencies of their entities or relations are high. To address this problem, we propose Model-based Subsampling (MBS) and Mixed Subsampling (MIX) to estimate their appearance probabilities through predictions of KGE models. Evaluation results on datasets FB15k-237, WN18RR, and YAGO3-10 showed that our proposed subsampling methods actually improved the KG completion performances for popular KGE models, RotatE, TransE, HAKE, ComplEx, and DistMult. ",Model-based Subsampling for Knowledge Graph Completion " We study the generation of quantum entanglement between two giant atoms coupled to a one-dimensional waveguide. Since each giant atom interacts with the waveguide at two separate coupling points, there exist three different coupling configurations in the two-atom waveguide system: separated, braided, and nested couplings. Within the Wigner-Weisskopf framework for single coupling points, the quantum master equations governing the evolution of the two giant atoms are obtained. For each coupling configuration, the entanglement dynamics of the two giant atoms is studied, including the cases of two different atomic initial states: single- and double-excitation states. It is shown that the generated entanglement depends on the coupling configuration, phase shift, and atomic initial state. For the single-excitation initial state, there exists steady-state entanglement for these three couplings due to the appearance of the dark state. For the double-excitation initial state, an entanglement sudden birth is observed via adjusting the phase shift. In particular, the maximal entanglement for the nested coupling is about one order of magnitude larger than those of separate and braided couplings. In addition, the influence of the atomic frequency detuning on the entanglement generation is studied. This work can be utilized for the generation and control of atomic entanglement in quantum networks based on giant-atom waveguide-QED systems, which have wide potential applications in quantum information processing. ",Generation of two-giant-atom entanglement in waveguide-QED systems " Recent self-supervised learning methods are able to learn high-quality image representations and are closing the gap with supervised approaches. However, these methods are unable to acquire new knowledge incrementally -- they are, in fact, mostly used only as a pre-training phase over IID data. In this work we investigate self-supervised methods in continual learning regimes without any replay mechanism. We show that naive functional regularization, also known as feature distillation, leads to lower plasticity and limits continual learning performance. Instead, we propose Projected Functional Regularization in which a separate temporal projection network ensures that the newly learned feature space preserves information of the previous one, while at the same time allowing for the learning of new features. This prevents forgetting while maintaining the plasticity of the learner. Comparison with other incremental learning approaches applied to self-supervision demonstrates that our method obtains competitive performance in different scenarios and on multiple datasets. ",Continually Learning Self-Supervised Representations with Projected Functional Regularization " Following the observational evidence for cosmic acceleration which may exclude a possibility for the universe to recollapse to a second singularity, we review alternative scenarios of its future evolution. Although the de Sitter asymptotic state is still an option, some other asymptotic states which allow new types of singularities such as Big-Rip (due to a phantom matter) and sudden future singularities are also admissible and are reviewed in detail. The reality of these singularities which comes from the relation to observational characteristics of the universe expansion are also revealed and widely discussed. ",Future state of the Universe " To study spacelike surfaces of codimension two in the Lorentz-Minkowski space $\Bbb R^{n+1}_1,$ we construct a pair of maps whose values are in $HS_r:=H_+^n(\textbf v,1)\cap \{x_{n+1}=r\},$ called $\textbf n_r^{\pm}$-Gauss maps. It is showed that they are well-defined and useful to study practically flat as well as umbilic spacelike surfaces of codimension two in $\Bbb R^{n+1}_1.$ ",$HS_{r}$-valued Gauss maps and umbilic spacelike surfaces of codimension two " Joret, Micek, Milans, Trotter, Walczak, and Wang recently asked if there exists a constant $d$ such that if $P$ is a poset with cover graph of $P$ of pathwidth at most $2$, then $\dim(P)\leq d$. We answer this question in the affirmative by showing that $d=17$ is sufficient. We also show that if $P$ is a poset containing the standard example $S_5$ as a subposet, then the cover graph of $P$ has treewidth at least $3$. ",Posets with cover graph of pathwidth two have bounded dimension " We show how light can be controllably transported by light at microscale dimensions. We design a miniature device which consists of a short segment of an optical fiber coupled to transversely-oriented input-output microfibers. A whispering gallery soliton is launched from the first microfiber into the fiber segment and slowly propagates along its mm-scale length. The soliton loads and unloads optical pulses at designated input-output microfibers. The speed of the soliton and its propagation direction is controlled by the dramatically small, yet feasible to introduce permanently or all-optically, nanoscale variation of the effective fiber radius. ",Controlled transportation of light by light at the microscale " Multi-agent control problems constitute an interesting area of application for deep reinforcement learning models with continuous action spaces. Such real-world applications, however, typically come with critical safety constraints that must not be violated. In order to ensure safety, we enhance the well-known multi-agent deep deterministic policy gradient (MADDPG) framework by adding a safety layer to the deep policy network. In particular, we extend the idea of linearizing the single-step transition dynamics, as was done for single-agent systems in Safe DDPG (Dalal et al., 2018), to multi-agent settings. We additionally propose to circumvent infeasibility problems in the action correction step using soft constraints (Kerrigan & Maciejowski, 2000). Results from the theory of exact penalty functions can be used to guarantee constraint satisfaction of the soft constraints under mild assumptions. We empirically find that the soft formulation achieves a dramatic decrease in constraint violations, making safety available even during the learning procedure. ",Safe Deep Reinforcement Learning for Multi-Agent Systems with Continuous Action Spaces " We consider the problem of recovering a complete (i.e., square and invertible) matrix $\mathbf A_0$, from $\mathbf Y \in \mathbb R^{n \times p}$ with $\mathbf Y = \mathbf A_0 \mathbf X_0$, provided $\mathbf X_0$ is sufficiently sparse. This recovery problem is central to the theoretical understanding of dictionary learning, which seeks a sparse representation for a collection of input signals, and finds numerous applications in modern signal processing and machine learning. We give the first efficient algorithm that provably recovers $\mathbf A_0$ when $\mathbf X_0$ has $O(n)$ nonzeros per column, under suitable probability model for $\mathbf X_0$. In contrast, prior results based on efficient algorithms provide recovery guarantees when $\mathbf X_0$ has only $O(n^{1-\delta})$ nonzeros per column for any constant $\delta \in (0, 1)$. Our algorithmic pipeline centers around solving a certain nonconvex optimization problem with a spherical constraint, and hence is naturally phrased in the language of manifold optimization. To show this apparently hard problem is tractable, we first provide a geometric characterization of the high-dimensional objective landscape, which shows that with high probability there are no ""spurious"" local minima. This particular geometric structure allows us to design a Riemannian trust region algorithm over the sphere that provably converges to one local minimizer with an arbitrary initialization, despite the presence of saddle points. The geometric approach we develop here may also shed light on other problems arising from nonconvex recovery of structured signals. ",Complete Dictionary Recovery over the Sphere " For many spin-0 target nuclei neutron capture measurements yield information on level densities at the neutron separation energy. Also the average photon width has been determined from capture data as well as Maxwellian average cross sections for the energy range of unresolved resonances. Thus it is challenging to use this data set for a test of phenomenological prescriptions for the prediction of radiative processes. An important ingredient for respective calculations is the photon strength function for which a parameterization was proposed using a fit to giant dipole resonance shapes on the basis of theoretically determined ground state deformations including triaxiality. Deviations from spherical and axial symmetry also influence level densities and it is suggested to use a combined parameterization for both, level density and photon strength. The formulae presented give a good description of the data for low spin capture into 124 nuclei with 72 10 Gyr) populations, but show modest systematic differences at younger ages, and that [Fe/H] is fairly consistent, with small systematic differences related to the age systematics. Overall, EZ_Ages provides accurate estimates of fundamental parameters from medium resolution spectra of unresolved stellar populations in the old and intermediate-age regime, for the first time allowing quantitative estimates of the abundances of C, N, and Ca in these unresolved systems. The EZ_Ages code can be downloaded at http://www.ucolick.org/~graves/EZ_Ages.html ","Measuring Ages and Elemental Abundances from Unresolved Stellar Populations: Fe, Mg, C, N, and Ca" We show that a 2-subset-regular self-complementary 3-uniform hypergraph with $n$ vertices exists if and only if $n\ge 6$ and $n$ is congruent to 2 modulo 4. ,A note on 2-subset-regular self-complementary 3-uniform hypergraphs " This paper deals with the possible motion of nucleons in the nucleus, which is due to realistic inter-nucleonic forces. This approach provides new or more substantiated conclusions about the nuclear structure than those based on the effective interaction of nucleons, while the shell model of the nucleus may lead to questionable conclusions regarding the nuclear structure and nuclear reaction mechanisms. ",On the nucleons motion in the nucleus as being due to realistic inter-nucleonic forces " Parkinson's disease (PD) is a common neurological disorder characterized by gait impairment. PD has no cure, and an impediment to developing a treatment is the lack of any accepted method to predict disease progression rate. The primary aim of this study was to develop a model using clinical measures and biomechanical measures of gait and postural stability to predict an individual's PD progression over two years. Data from 160 PD subjects were utilized. Machine learning models, including XGBoost and Feed Forward Neural Networks, were developed using extensive model optimization and cross-validation. The highest performing model was a neural network that used a group of clinical measures, achieved a PPV of 71% in identifying fast progressors, and explained a large portion (37%) of the variance in an individual's progression rate on held-out test data. This demonstrates the potential to predict individual PD progression rate and enrich trials by analyzing clinical and biomechanical measures with machine learning. ",Prediction of individual progression rate in Parkinson's disease using clinical measures and biomechanical measures of gait and postural stability " We consider the problem of updating the SVD when augmenting a ""tall thin"" matrix, i.e., a rectangular matrix $A \in \RR^{m \times n}$ with $m \gg n$. Supposing that an SVD of $A$ is already known, and given a matrix $B \in \RR^{m \times n'}$, we derive an efficient method to compute and efficiently store the SVD of the augmented matrix $[ A B ] \in \RR^{m \times (n+n')}$. This is an important tool for two types of applications: in the context of principal component analysis, the dominant left singular vectors provided by this decomposition form an orthonormal basis for the best linear subspace of a given dimension, while from the right singular vectors one can extract an orthonormal basis of the kernel of the matrix. We also describe two concrete applications of these concepts which motivated the development of our method and to which it is very well adapted. ",SVD update methods for large matrices and applications " Subspace codes form the appropriate mathematical setting for investigating the Koetter-Kschischang model of fault-tolerant network coding. The Main Problem of Subspace Coding asks for the determination of a subspace code of maximum size (proportional to the transmission rate) if the remaining parameters are kept fixed. We describe a new approach to finding good subspace codes, which surpasses the known size limit of lifted MRD codes and is capable of yielding an alternative construction of the currently best known binary subspace code of packet length 7, constant dimension 3 and minimum subspace distance 4. ",A New Approach to the Main Problem of Subspace Coding " Transverse NMR relaxation in a macroscopic sample is shown to be extremely sensitive to the structure of mesoscopic magnetic susceptibility variations. Such a sensitivity is proposed as a novel kind of contrast in the NMR measurements. For suspensions of arbitrary shaped paramagnetic objects, the transverse relaxation is found in the case of a small dephasing effect of an individual object. Strong relaxation rate dependence on the objects' shape agrees with experiments on whole blood. Demonstrated structure sensitivity is a generic effect that arises in NMR relaxation in porous media, biological systems, as well as in kinetics of diffusion limited reactions. ",Transverse NMR relaxation as a probe of mesoscopic structure " The explicit formulas for the maps interconnecting the sets of solutions of the special double confluent Heun equation and the equation of the RSJ model of overdamped Josephson junction in case of shifted sinusoidal bias are given. The approach these are based upon leans on the extensive application of eigenfunctions of certain linear operator acting on functions holomorphic on the universal cover of the punctured complex plane. The functional equation the eigenfunctions noted obey is derived, the matrix form of the monodromy transformation they infer is given. ",The interrelation of the special double confluent Heun equation and the equation of RSJ model of Josephson junction revisited " Within the frame of quantum optics we analyze the properties of spontaneous emission of two-level atom in media with indefinite permittivity tensor where the geometry of the dispersion relation is characterized by an ellipsoid or a hyperboloid(hyperbolic medium). The decay rate is explicitly given with the orientation of the dipole transition matrix element taken into account. It indicates that for the ellipsoid case the intensity of the photons coupled into different modes can be tuned by changing the direction of the matrix element and for the hyperboloid case it is found that spontaneous emission in hyperbolic medium can be dramatically enhanced compared to the dielectric background. Moreover, spontaneous emission exhibit the strong directivity and get the maximum in the asymptote direction. ",Controlling spontaneous emission of a two-level atom by hyperbolic metamaterials " Flavour oscillations of sub-GeV atmospheric neutrinos and antineutrinos, traversing different distances inside the Earth, are a promising source of information on the leptonic CP phase $\delta$. In that energy range, the oscillations are very fast, far beyond the resolution of modern neutrino detectors. However, the necessary averaging over the experimentally typical energy and azimuthal angle bins does not wash out the CP violation effects. In this paper we derive very accurate analytic compact expressions for the averaged oscillations probabilities. Assuming spherically symmetric Earth, the averaged oscillation probabilities are described in terms of two analytically calculable effective parameters. Based on those expressions, we estimate maximal magnitude of CP-violation effects in such measurements and propose optimal observables best suited to determine the value of the CP phase in the PMNS mixing matrix. ",Analytical description of CP violation in oscillations of atmospheric neutrinos traversing the Earth " The Projected Dynamics method was originally developed to study metastable decay in ferromagnetic discrete spin models. Here, we apply it to a classical, continuous Heisenberg model with anisotropic ferromagnetic interactions, which evolves under a Monte Carlo dynamic. The anisotropy is sufficiently large to allow comparison with the Ising model. We describe the Projected Dynamics method and how to apply it to this continuous-spin system. We also discuss how to extract metastable lifetimes and how to extrapolate from small systems to larger systems. ",Application of the Projected Dynamics Method to an Anisotropic Heisenberg Model We consider quantum analogs of the relativistic Toda lattices and give new $2\times 2$ $L$-operators for these models. Making use of the variable separation the spectral problem for the quantum integrals of motion is reduced to solving one-dimensional separation equations. ,Separation of variables for the quantum relativistic Toda lattices " In the absence of acceleration, the velocity formula gives ""distance travelled equals speed multiplied by time"". For a broad class of Markov chains such as circulant Markov chains or random walk on complete graphs, we prove a probabilistic analogue of the velocity formula between entropy and hitting time, where distance is the entropy of the Markov trajectories from state $i$ to state $j$ in the sense of [L. Ekroot and T. M. Cover. The entropy of Markov trajectories. IEEE Trans. Inform. Theory 39(4): 1418-1421.], speed is the classical entropy rate of the chain, and the time variable is the expected hitting time between $i$ and $j$. This motivates us to define new entropic counterparts of various hitting time parameters such as average hitting time or commute time, and prove analogous velocity formulae and estimates between these quantities. ",Velocity formulae between entropy and hitting time for Markov chains " Violation of Mermin's and Svetlichny's inequalities can rule out the predictions of local hidden variable theory and can confirm the existence of true nonlocal correlation for n-particle pure quantum systems. Here we demonstrate the experimental violation of the above inequalities for W- and GHZ-class of states. We use IBM's five-qubit quantum computer for experimental implementation of these states and illustration of inequalities' violations. Our results clearly show the violations of both Mermin's and Svetlichny's inequalities for W and GHZ states respectively. Being a superconducting qubit-based quantum computer, the platform used here opens up the opportunity to explore multipartite inequalities which is beyond the reach of other existing technologies. ",Experimental demonstration of the violations of Mermin's and Svetlichny's inequalities for W- and GHZ-class of states " The ability to achieve near-unity light extraction efficiency is necessary for a truly deterministic single photon source. The most promising method to reach such high efficiencies is based on embedding single photon emitters in tapered photonic waveguides defined by top-down etching techniques. However, light extraction efficiencies in current top-down approaches are limited by fabrication imperfections and etching induced defects. The efficiency is further tempered by randomly positioned off-axis quantum emitters. Here, we present perfectly positioned single quantum dots on the axis of a tailored nanowire waveguide using bottom-up growth. In comparison to quantum dots in nanowires without waveguide, we demonstrate a 24-fold enhancement in the single photon flux, corresponding to a light extraction efficiency of 42 %. Such high efficiencies in one-dimensional nanowires are promising to transfer quantum information over large distances between remote stationary qubits using flying qubits within the same nanowire p-n junction. ",Bright single-photon sources in bottom-up tailored nanowires " In this paper, we propose a methodology for early recognition of human activities from videos taken with a first-person viewpoint. Early recognition, which is also known as activity prediction, is an ability to infer an ongoing activity at its early stage. We present an algorithm to perform recognition of activities targeted at the camera from streaming videos, making the system to predict intended activities of the interacting person and avoid harmful events before they actually happen. We introduce the novel concept of 'onset' that efficiently summarizes pre-activity observations, and design an approach to consider event history in addition to ongoing video observation for early first-person recognition of activities. We propose to represent onset using cascade histograms of time series gradients, and we describe a novel algorithmic setup to take advantage of onset for early recognition of activities. The experimental results clearly illustrate that the proposed concept of onset enables better/earlier recognition of human activities from first-person videos. ",Early Recognition of Human Activities from First-Person Videos Using Onset Representations " The recent Planck Legacy 2018 release has confirmed the presence of an enhanced lensing amplitude in CMB power spectra compared to that predicted in the standard $\Lambda$CDM model. A closed universe can provide a physical explanation for this effect, with the Planck CMB spectra now preferring a positive curvature at more than $99 \%$ C.L. Here we further investigate the evidence for a closed universe from Planck, showing that positive curvature naturally explains the anomalous lensing amplitude and demonstrating that it also removes a well-known tension within the Planck data set concerning the values of cosmological parameters derived at different angular scales. We show that since the Planck power spectra prefer a closed universe, discordances higher than generally estimated arise for most of the local cosmological observables, including BAO. The assumption of a flat universe could, therefore, mask a cosmological crisis where disparate observed properties of the Universe appear to be mutually inconsistent. Future measurements are needed to clarify whether the observed discordances are due to undetected systematics, or to new physics, or simply are a statistical fluctuation. ",Planck evidence for a closed Universe and a possible crisis for cosmology " We report the results of Cu and Cl nuclear magnetic resonance experiments (NMR) and thermal expansion measurements in magnetic fields in the coupled dimer spin system TlCuCl3. We found that the field-induced antiferromagnetic transition as confirmed by the splitting of NMR lines is slightly discontinuous. The abrupt change of the electric field gradient at the Cl sites, as well as the sizable change of the lattice constants, across the phase boundary indicate that the magnetic order is accompanied by simultaneous lattice deformation. ",Field-Induced Magnetic Order and Simultaneous Lattice Deformation in TlCuCl3 " We study the symmetry resolved entanglement entropies in gapped integrable lattice models. We use the corner transfer matrix to investigate two prototypical gapped systems with a U(1) symmetry: the complex harmonic chain and the XXZ spin-chain. While the former is a free bosonic system, the latter is genuinely interacting. We focus on a subsystem being half of an infinitely long chain. In both models, we obtain exact expressions for the charged moments and for the symmetry resolved entropies. While for the spin chain we found exact equipartition of entanglement (i.e. all the symmetry resolved entropies are the same), this is not the case for the harmonic system where equipartition is effectively recovered only in some limits. Exploiting the gaussianity of the harmonic chain, we also develop an exact correlation matrix approach to the symmetry resolved entanglement that allows us to test numerically our analytic results. ",Symmetry resolved entanglement in gapped integrable systems: a corner transfer matrix approach " There is a large change in surface rotation rates of sun-like stars on the pre-main sequence and early main sequence. Since these stars have dynamo driven magnetic fields, this implies a strong evolution of their magnetic properties over this time period. The spin-down of these stars is controlled by interactions between stellar winds and magnetic fields, thus magnetic evolution in turn plays an important role in rotational evolution. We present here the second part of a study investigating the evolution of large-scale surface magnetic fields in this critical time period. We observed stars in open clusters and stellar associations with known ages between 120 and 650 Myr, and used spectropolarimetry and Zeeman Doppler Imaging to characterize their large-scale magnetic field strength and geometry. We report 15 stars with magnetic detections here. These stars have masses from 0.8 to 0.95 Msun, rotation periods from 0.326 to 10.6 days, and we find large-scale magnetic field strengths from 8.5 to 195 G with a wide range of geometries. We find a clear trend towards decreasing magnetic field strength with age, and a power-law decrease in magnetic field strength with Rossby number. There is some tentative evidence for saturation of the large-scale magnetic field strength at Rossby numbers below 0.1, although the saturation point is not yet well defined. Comparing to younger classical T Tauri stars, we support the hypothesis that differences in internal structure produce large differences in observed magnetic fields, however for weak lined T Tauri stars this is less clear. ",The evolution of surface magnetic fields in young solar-type stars II: The early main sequence (250-650 Myr) " We study the quantum and classical scattering of Hamiltonian systems whose chaotic saddle is described by binary or ternary horseshoes. We are interested in parameters of the system for which a stable island, associated with the inner fundamental periodic orbit of the system exists and is large, but chaos around this island is well developed. In this situation, in classical systems, decay from the interaction region is algebraic, while in quantum systems it is exponential due to tunneling. In both cases, the most surprising effect is a periodic response to an incoming wave packet. The period of this self-pulsing effect or scattering echoes coincides with the mean period, by which the scattering trajectories rotate around the stable orbit. This period of rotation is directly related to the development stage of the underlying horseshoe. Therefore the predicted echoes will provide experimental access to topological information. We numerically test these results in kicked one dimensional models and in open billiards. ",Self-pulsing effect in chaotic scattering " A diabatic (configuration-fixed) constrained approach to calculate the potential energy surface (PES) of the nucleus is developed in the relativistic mean field model. {As an example}, the potential energy surfaces of $^{208}$Pb obtained from both adiabatic and diabatic constrained approaches are investigated and compared. {It is shown that} the diabatic constrained approach enables one to decompose the segmented PES obtained in usual adiabatic approaches into separate parts uniquely characterized by different configurations, {to follow the evolution of single-particle orbits till very deformed region}, and to obtain several well defined deformed excited states which can hardly be expected from the adiabatic PES's. ",Constrained relativistic mean field approach with fixed configurations " Inflationary models can correlate small-scale density perturbations with the long-wavelength gravitational waves (GW) in the form of the Tensor-Scalar-Scalar (TSS) bispectrum. This correlation affects the mass-distribution in the Universe and leads to the off-diagonal correlations of the density field modes in the form of the quadrupole anisotropy. Interestingly, this effect survives even after the tensor mode decays when it re-enters the horizon, known as the fossil effect. As a result, the off-diagonal correlation function between different Fourier modes of the density fluctuations can be thought as a way to probe the large-scale GW and the mechanism of inflation behind the fossil effect. Models of single field slow roll inflation generically predict a very small quadrupole anisotropy in TSS while in models of multiple fields inflation this effect can be observable. Therefore this large scale quadrupole anisotropy can be thought as a spectroscopy for different inflationary models. In addition, in models of anisotropic inflation there exists quadrupole anisotropy in curvature perturbation power spectrum. Here we consider TSS in models of anisotropic inflation and show that the shape of quadrupole anisotropy is different than in single field models. In addition in these models the quadrupole anisotropy is projected into the preferred direction and its amplitude is proportional to $g_* N_e$ where $N_e$ is the number of e-folds and $g_*$ is the amplitude of quadrupole anisotropy in curvature perturbation power spectrum. We use this correlation function to estimate the large scale GW as well as the preferred direction and discuss the detectability of the signal in the galaxy surveys like Euclid and 21 cm surveys. ",Clustering Fossil from Primordial Gravitational Waves in Anisotropic Inflation " We prove that the Crisp and Gow's quiver operation on a finite quiver Q produces a new quiver Q' with fewer vertices, such that the finite dimensional algebras kQ/J^2 and kQ'/J^2 are singularly equivalent. This operation is a general quiver operation which includes as specific examples some operations which arise naturally in symbolic dynamics (e.g., (elementary) strong shift equivalent, (in-out) splitting, source elimination, etc.). ",Singular equivalence of finite dimensional algebras with radical square zero " Cluster algebras were introduced by S. Fomin and A. Zelevinsky in connection with dual canonical bases. To a cluster algebra of simply laced Dynkin type one can associate the cluster category. Any cluster of the cluster algebra corresponds to a tilting object in the cluster category. The cluster tilted algebra is the algebra of endomorphisms of that tilting object. Viewing the cluster tilted algebra as a path algebra of a quiver with relations, we prove in this paper that the quiver of the cluster tilted algebra is equal to the cluster diagram. We study also the relations. As an application of these results, we answer several conjectures on the connection between cluster algebras and quiver representations. ",Quivers with relations and cluster tilted algebras Combinatorial characterisations of minimal rigidity are obtained for symmetric 2-dimensional bar-joint frameworks with either $\ell^1$ or $\ell^\infty$ distance constraints. The characterisations are expressed in terms of symmetric tree packings and the number of edges fixed by the symmetry operations. The proof uses new Henneberg-type inductive construction schemes. ,Symmetric isostatic frameworks with $\ell^1$ or $\ell^\infty$ distance constraints " We show that if an ample line bundle L on a nonsingular toric 3-fold satisfies h^0(L+2K)=0, then L is normally generated. As an application, we show that the anti-canonical divisor on a nonsingular toric Fano 4-fold is normally generated. ",Projective normality of nonsingular toric varieties of dimension three A distinctive property of human and animal intelligence is the ability to form abstractions by neglecting irrelevant information which allows to separate structure from noise. From an information theoretic point of view abstractions are desirable because they allow for very efficient information processing. In artificial systems abstractions are often implemented through computationally costly formations of groups or clusters. In this work we establish the relation between the free-energy framework for decision making and rate-distortion theory and demonstrate how the application of rate-distortion for decision-making leads to the emergence of abstractions. We argue that abstractions are induced due to a limit in information processing capacity. ,Abstraction in decision-makers with limited information processing capabilities " At present, the task of searching for compounds with a high superconducting transition temperature is a very relevant scientific direction. Usually, the calculation of is carried out by numerically solving the system of Eliashberg equations. In this paper, a set of programs for solving this system written in various forms on the imaginary axis is presented. As an example of the developed methods applications, calculations results of and thermodynamic properties of the metallic hydrogen I41/AMD phase and some other substances under high pressure are presented. ",Software Complex for the Numerical Solution of the Isotropic Imaginary-Axis Eliashberg Equations " Integration between magnetism and topology is an exotic phenomenon in condensed-matter physics. Here, we propose an exotic phase named topological crystalline antiferromagnetic state, in which antiferromagnetism intrinsically integrates with nontrivial topology, and we suggest such a state can be realized in tetragonal FeS. A combination of first-principles calculations and symmetry analyses shows that the topological crystalline antiferromagnetic state arises from band reconstruction induced by pair checker-board antiferromagnetic order together with band-gap opening induced by intrinsic spin-orbit coupling in tetragonal FeS. The topological crystalline antiferromagnetic state is protected by the product of fractional translation symmetry, mirror symmetry, and time-reversal symmetry, and present some unique features. In contrast to strong topological insulators, the topological robustness is surface-dependent. These findings indicate that non-trivial topological states could emerge in pure antiferromagnetic materials, which sheds new light on potential applications of topological properties in fast-developing antiferromagnetic spintronics. ",Topological crystalline antiferromagnetic state in tetragonal FeS " Splines and subdivision curves are flexible tools in the design and manipulation of curves in Euclidean space. In this paper we study generalizations of interpolating splines and subdivision schemes to the Riemannian manifold of shell surfaces in which the associated metric measures both bending and membrane distortion. The shells under consideration are assumed to be represented by Loop subdivision surfaces. This enables the animation of shells via the smooth interpolation of a given set of key frame control meshes. Using a variational time discretization of geodesics efficient numerical implementations can be derived. These are based on a discrete geodesic interpolation, discrete geometric logarithm, discrete exponential map, and discrete parallel transport. With these building blocks at hand discrete Riemannian cardinal splines and three different types of discrete, interpolatory subdivision schemes are defined. Numerical results for two different subdivision shell models underline the potential of this approach in key frame animation. ",Smooth Interpolation of Key Frames in a Riemannian Shell Space " Fine-grained visual classification (FGVC) is much more challenging than traditional classification tasks due to the inherently subtle intra-class object variations. Recent works mainly tackle this problem by focusing on how to locate the most discriminative parts, more complementary parts, and parts of various granularities. However, less effort has been placed to which granularities are the most discriminative and how to fuse information cross multi-granularity. In this work, we propose a novel framework for fine-grained visual classification to tackle these problems. In particular, we propose: (i) a progressive training strategy that effectively fuses features from different granularities, and (ii) a random jigsaw patch generator that encourages the network to learn features at specific granularities. We obtain state-of-the-art performances on several standard FGVC benchmark datasets, where the proposed method consistently outperforms existing methods or delivers competitive results. The code will be available at https://github.com/PRIS-CV/PMG-Progressive-Multi-Granularity-Training. ",Fine-Grained Visual Classification via Progressive Multi-Granularity Training of Jigsaw Patches " We prove a universal approximation theorem that allows to approximate continuous functionals of c\`adl\`ag (rough) paths uniformly in time and on compact sets of paths via linear functionals of their time-extended signature. Our main motivation to treat this question comes from signature-based models for finance that allow for the inclusion of jumps. Indeed, as an important application, we define a new class of universal signature models based on an augmented L\'evy process, which we call L\'evy-type signature models. They extend continuous signature models for asset prices as proposed e.g. by Arribas et al.(2020) in several directions, while still preserving universality and tractability properties. To analyze this, we first show that the signature process of a generic multivariate L\'evy process is a polynomial process on the extended tensor algebra and then use this for pricing and hedging approaches within L\'evy-type signature models. ",Universal approximation theorems for continuous functions of c\`adl\`ag paths and L\'evy-type signature models " This paper presents Non-Attentive Tacotron based on the Tacotron 2 text-to-speech model, replacing the attention mechanism with an explicit duration predictor. This improves robustness significantly as measured by unaligned duration ratio and word deletion rate, two metrics introduced in this paper for large-scale robustness evaluation using a pre-trained speech recognition model. With the use of Gaussian upsampling, Non-Attentive Tacotron achieves a 5-scale mean opinion score for naturalness of 4.41, slightly outperforming Tacotron 2. The duration predictor enables both utterance-wide and per-phoneme control of duration at inference time. When accurate target durations are scarce or unavailable in the training data, we propose a method using a fine-grained variational auto-encoder to train the duration predictor in a semi-supervised or unsupervised manner, with results almost as good as supervised training. ",Non-Attentive Tacotron: Robust and Controllable Neural TTS Synthesis Including Unsupervised Duration Modeling Quantized vortex states of weakly interacting Bose-Einstein condensate of atoms with attractive interatomic interaction in an axially symmetric harmonic oscillator trap are investigated using the numerical solution of the time-dependent Gross-Pitaevskii (GP) equation obtained by the semi-implicit Crank-Nicholson method. Collapse of the condensate is studied in the presence of deformed traps with a larger frequency along the radial as well as along the axial directions. The critical number of atoms for collapse is calculated as a function of vortex quantum $L$. The critical number increases with angular momentum $L$ of the vortex state but tends to saturate for large $L$. ,Collapse of attractive Bose-Einstein condensed vortex states in a cylindrical trap " We prove an algebra property under pointwise multiplication for Besov spaces defined on Lie groups of polynomial growth. When the setting is restricted to the case of H-type groups, this algebra property is generalized to paraproduct estimates. ",Besov algebras on Lie groups of polynomial growth " Noncommutative geometry, an offshoot of string theory, replaces point-like objects by smeared objects. The resulting uncertainty may cause a black hole to be observationally indistinguishable from a traversable wormhole, while the latter, in turn, may become observationally indistinguishable from a gravastar. The same noncommutative-geometry background allows the theoretical construction of thin-shell wormholes from gravastars and may even serve as a model for dark energy. ","Seeking connections between wormholes, gravastars, and black holes via noncommutative geometry" The TOTEM experiment at the LHC is dedicated to the measurement of the total pp cross section and to the study of elastic scattering and of diffractive dissociation processes. TOTEM is here presented with a general overview on the main features of its experimental apparatus and of its physics programme. ,The TOTEM Experiment at the LHC " We consider quantum quenches in models of free scalars and fermions with a generic time-dependent mass $m(t)$ that goes from $m_0$ to zero. We prove that, as anticipated in MSS \cite{Mandal:2015jla}, the post-quench dynamics can be described in terms of a state of the generalized Calabrese-Cardy form $|\psi \rangle$= $\exp[-\kappa_2 H -\sum_{n>2}^\infty \kappa_n W_n]| \hbox{Bd} \rangle$. The $W_n$ ($n=2,3,...$, $W_2=H$) here represent the conserved $W_\infty$ charges and $| \hbox{Bd} \rangle$ represents a conformal boundary state. Our result holds irrespective of whether the pre-quench state is a ground state or a squeezed state, and is proved without recourse to perturbation expansion in the $\kappa_n$'s as in MSS. We compute exact time-dependent correlators for some specific quench protocols $m(t)$. The correlators explicitly show thermalization to a generalized Gibbs ensemble (GGE), with inverse temperature $\beta= 4\kappa_2$, and chemical potentials $\mu_n=4\kappa_n$. In case the pre-quench state is a ground state, it is possible to retrieve the exact quench protocol $m(t)$ from the final GGE, by an application of inverse scattering techniques. Another notable result, which we interpret as a UV/IR mixing, is that the long distance and long time (IR) behaviour of some correlators depends crucially on all $\kappa_n$'s, although they are highly irrelevant couplings in the usual RG parlance. This indicates subtleties in RG arguments when applied to non-equilibrium dynamics. ",Thermalization in 2D critical quench and UV/IR mixing We prove a differential Harnack inequality for the solution of the parabolic Allen-Cahn equation $ \frac{\partial f}{\partial t}=\triangle f-(f^3-f)$ on a closed n-dimensional manifold. As a corollary we find a classical Harnack inequality. We also formally compare the standing wave solution to a gradient estimate of Modica from the 1980s for the elliptic equation. ,A Harnack inequality for the parabolic Allen-Cahn equation A basic component in Internet applications is the electronic mail and its various implications. The paper proposes a mechanism for automatically classifying emails and create dynamic groups that belong to these messages. Proposed mechanisms will be based on natural language processing techniques and will be designed to facilitate human-machine interaction in this direction. ,Clasificarea distribuita a mesajelor de e-mail " This paper presents a fast and effective computer algebraic method for analyzing and verifying non-linear integer arithmetic circuits using a novel algebraic spectral model. It introduces a concept of algebraic spectrum, a numerical form of polynomial expression; it uses the distribution of coefficients of the monomials to determine the type of arithmetic function under verification. In contrast to previous works, the proof of functional correctness is achieved by computing an algebraic spectrum combined with a local rewriting of word-level polynomials. The speedup is achieved by propagating coefficients through the circuit using And-Inverter Graph (AIG) datastructure. The effectiveness of the method is demonstrated with experiments including standard and Booth multipliers, and other synthesized non-linear arithmetic circuits up to 1024 bits containing over 12 million gates. ",Spectral Approach to Verifying Non-linear Arithmetic Circuits Perfect cloning of a known set of states with arbitrary prior probabilities is possible if we allow the cloner to sometimes fail completely. In the optimal case the probability of failure is at its minimum allowed by the laws of quantum mechanics. Here we show that it is possible to lower the failure rate below that of the perfect probabilistic cloner but the price to pay is that the clones are not perfect; the global fidelity is less than one. We determine the optimal fidelity of a cloner with a Fixed Failure Rate (FFR cloner) in the case of a pair of known states. Optimality is shown to be attainable by a measure-and-prepare protocol in the limit of infinitely many clones. The optimal protocol consists of discrimination with a fixed rate of inconclusive outcome followed by preparation of the appropriate clones. The convergence shows a symmetry-breaking second-order phase transition in the fidelity of the approximate infinite clones. ,Optimal Cloning of Quantum States with a Fixed Failure Rate " We introduce two applications of polygraphs to categorification problems. We compute first, from a coherent presentation of an $n$-category, a coherent presentation of its Karoubi envelope. For this, we extend the construction of Karoubi envelope to $n$-polygraphs and linear $(n,n-1)$-polygraphs. The second problem treated in this paper is the construction of Grothendieck decategorifications for $(n,n-1)$-polygraphs. This construction yields a rewriting system presenting for example algebras categorified by a linear monoidal category. We finally link quasi-convergence of such rewriting systems to the uniqueness of direct sum decompositions for linear $(n-1,n-1)$-categories. ",Linear polygraphs applied to categorification We determine the metric dimension of the annihilating-ideal graph of a local finite commutative principal ring and a finite commutative principal ring with two maximal ideals. We also find the bounds for the metric dimension of the annihilating-ideal graph of an arbitrary finite commutative principal ring. ,The metric dimension of the annihilating-ideal graph of a finite commutative ring " This paper presents a distributed control architecture for voltage and frequency stabilization in AC islanded microgrids. In the primary control layer, each generation unit is equipped with a local controller acting on the corresponding voltage-source converter. Following the plug-and-play design approach previously proposed by some of the authors, whenever the addition/removal of a distributed generation unit is required, feasibility of the operation is automatically checked by designing local controllers through convex optimization. The update of the voltage-control layer, when units plug -in/-out, is therefore automatized and stability of the microgrid is always preserved. Moreover, local control design is based only on the knowledge of parameters of power lines and it does not require to store a global microgrid model. In this work, we focus on bus-connected microgrid topologies and enhance the primary plug-and-play layer with local virtual impedance loops and secondary coordinated controllers ensuring bus voltage tracking and reactive power sharing. In particular, the secondary control architecture is distributed, hence mirroring the modularity of the primary control layer. We validate primary and secondary controllers by performing experiments with balanced, unbalanced and nonlinear loads, on a setup composed of three bus-connected distributed generation units. Most importantly, the stability of the microgrid after the addition/removal of distributed generation units is assessed. Overall, the experimental results show the feasibility of the proposed modular control design framework, where generation units can be added/removed on the fly, thus enabling the deployment of virtual power plants that can be resized over time. ",Plug-and-play and coordinated control for bus-connected AC islanded microgrids " The impulse response function (IRF) of a localized bolus in cerebral blood flow codes important information on the tissue type. It is indirectly accessible both from MR- and CT-imaging methods, at least in principle. In practice, however, noise and limited signal resolution render standard deconvolution techniques almost useless. Parametric signal descriptions look more promising, and it is the aim of this contribution to develop some improvements along this line. ",Signal analysis of impulse response functions in MR- and CT-measurements of cerebral blood flow " Metamaterials and metasurfaces are at the pinnacle of wave propagation engineering, yet their design has thus far been mainly focused on deep-subwavelength periodicities, practically forming an effective medium. Such an approach overlooks important structural degrees-of-freedom, e.g. the interplay between the corrugation periodicity and depth and how it affects the beam transport. Here, we present Slack Metasurfaces - weakly modulated metal-dielectric interfaces unlocking all structural degrees-of-freedom that affect the wave propagation. We experimentally demonstrate control over the anisotropy of surface waves in such metasurfaces, leading to yet, unexplored, dual stage topological transitions. We further utilize these metasurfaces to show unique backward focusing of surface waves driven by an umklapp process - momentum relaxation empowered by the periodic nature of the structure. Our findings can be applied to any type of guided waves, introducing a simple and diverse method for controlling wave propagation in artificial media. ",Topological transitions and surface umklapp scattering in Slack Metasurfaces " The effect of deposition oxygen pressure (P$_{O}$) on phase separation (PS) induced in epitaxial La$_{0.67}$Ca$_{0.33}$MnO$_{3}$/NdGaO$_{3}$(001) films was investigated. Fully oxygenated films grown at high P$_{O}$ are anisotropically strained. They exhibit PS over a wide temperature range, because of the large orthorhombicity of NdGaO$_{3}$ substrates. The paramagnetic insulator-to-ferromagnetic metal (FM) and FM-to-antiferromagnetic insulator (AFI) transitions gradually shift to lower temperatures with decreasing PO. The AFI state is initially weakened (P$_{O}$ >= 30 Pa), but then becomes more robust against the magnetic field (P$_{O}$ < 30 Pa). The out-of-plane film lattice parameter increases with decreasing P$_{O}$. For films grown at P$_{O}$>= 30 Pa, the slight oxygen deficiency may enlarge the lattice unit cell, reduce the anisotropic strain and suppress the AFI state. Films deposited at P$_{O}$ < 30 Pa instead experience an average compressive strain. The enhanced compressive strain and structural defects in the films may lead to the robust AFI state. These results aid our understanding of PS in manganite films. ",Effect of growth oxygen pressure on anisotropic-strain-induced phase separation in epitaxial La$_{0.67}$Ca$_{0.33}$MnO$_{3}$/NdGaO$_{3}$(001) films " A wide variety of methods have been used to compute percolation thresholds. In lattice percolation, the most powerful of these methods consists of microcanonical simulations using the union-find algorithm to efficiently determine the connected clusters, and (in two dimensions) using exact values from conformal field theory for the probability, at the phase transition, that various kinds of wrapping clusters exist on the torus. We apply this approach to percolation in continuum models, finding overlaps between objects with real-valued positions and orientations. In particular, we find precise values of the percolation transition for disks, squares, rotated squares, and rotated sticks in two dimensions, and confirm that these transitions behave as conformal field theory predicts. The running time and memory use of our algorithm are essentially linear as a function of the number of objects at criticality. ",Continuum Percolation Thresholds in Two Dimensions " Bialgebroids (resp. Hopf algebroids) are bialgebras (Hopf algebras) over noncommutative rings. Drinfeld twist techniques are particularly useful in the (deformation) quantization of Lie algebras as well as underlying module algebras (=quantum spaces). Smash product construction combines these two into the new algebra which, in fact, does not depend on the twist. However, we can turn it into bialgebroid in the twist dependent way. Alternatively, one can use Drinfeld twist techniques in a category of bialgebroids. We show that both techniques indicated in the title: twisting of a bialgebroid or constructing a bialgebroid from the twisted bialgebra give rise to the same result in the case of normalized cocycle twist. This can be useful for better description of a quantum deformed phase space. We argue that within this bialgebroid framework one can justify the use of deformed coordinates (i.e. spacetime noncommutativity) which are frequently postulated in order to explain quantum gravity effects. ",Twisted bialgebroids versus bialgebroids from a Drinfeld twist " The entropy of the Gram matrix of a joint purification of an ensemble of K mixed states yields an upper bound for the Holevo information Chi of the ensemble. In this work we combine geometrical and probabilistic aspects of the ensemble in order to obtain useful bounds for Chi. This is done by constructing various correlation matrices involving fidelities between every pair of states from the ensemble. For K=3 quantum states we design a matrix of root fidelities that is positive and the entropy of which is conjectured to upper bound Chi. Slightly weaker bounds are established for arbitrary ensembles. Finally, we investigate correlation matrices involving multi-state fidelities in relation to the Holevo quantity. ",Matrices of fidelities for ensembles of quantum states and the Holevo quantity " A comparative study of high and zero temperature plasma for the case of damping rate, drag and diffusion coefficients have been presented. In each of these quantities, it is revealed how the magnetic interaction dominates over the electric one at zero temperature unlike what happens at high temperature. ",Drag and Diffusion coefficients in extreme scenarios of temperature and chemical potential " Blue noise error patterns are well suited to human perception, and when applied to stochastic rendering techniques, blue noise masks (blue noise textures) minimize unwanted low-frequency noise in the final image. Current methods of applying blue noise masks at each frame independently produce white noise frequency spectra temporally. This white noise results in slower integration convergence over time and unstable results when filtered temporally. Unfortunately, achieving temporally stable blue noise distributions is non-trivial since 3D blue noise does not exhibit the desired 2D blue noise properties, and alternative approaches degrade the spatial blue noise qualities. We propose novel blue noise patterns that, when animated, produce values at a pixel that are well distributed over time, converge rapidly for Monte Carlo integration, and are more stable under TAA, while still retaining spatial blue noise properties. To do so, we propose an extension to the well-known void and cluster algorithm that reformulates the underlying energy function to produce spatiotemporal blue noise masks. These masks exhibit blue noise frequency spectra in both the spatial and temporal domains, resulting in visually pleasing error patterns, rapid convergence speeds, and increased stability when filtered temporally. We demonstrate these improvements on a variety of applications, including dithering, stochastic transparency, ambient occlusion, and volumetric rendering. By extending spatial blue noise to spatiotemporal blue noise, we overcome the convergence limitations of prior blue noise works, enabling new applications for blue noise distributions. ",Scalar Spatiotemporal Blue Noise Masks " Canonical correlation analysis investigates linear relationships between two sets of variables, but often works poorly on modern data sets due to high-dimensionality and mixed data types such as continuous, binary and zero-inflated. To overcome these challenges, we propose a semiparametric approach for sparse canonical correlation analysis based on Gaussian copula. Our main contribution is a truncated latent Gaussian copula model for data with excess zeros, which allows us to derive a rank-based estimator of the latent correlation matrix for mixed variable types without the estimation of marginal transformation functions. The resulting canonical correlation analysis method works well in high-dimensional settings as demonstrated via numerical studies, as well as in application to the analysis of association between gene expression and micro RNA data of breast cancer patients. ",Sparse semiparametric canonical correlation analysis for data of mixed types " In this expository article we relate the presentation of weighted estimates in the book of Martinez to the Bergman kernel approach of Sj\""ostrand. It is meant as an introduction to the Helffer--Sj\""ostrand theory (designed for the study of quantum resonances) in the simplest setting and to its adaptations to compact manifolds. ",An introduction to microlocal complex deformations " We define the hitting time for a model of continuous-time open quantum walks in terms of quantum jumps. Our starting point is a master equation in Lindblad form, which can be taken as the quantum analogue of the rate equation for a classical continuous-time Markov chain. The quantum jump method is well known in the quantum optics community and has also been applied to simulate open quantum walks in discrete time. This method however, is well-suited to continuous-time problems. It is shown here that a continuous-time hitting problem is amenable to analysis via quantum jumps: The hitting time can be defined as the time of the first jump. Using this fact, we derive the distribution of hitting times and explicit exressions for its statistical moments. Simple examples are considered to illustrate the final results. We then show that the hitting statistics obtained via quantum jumps is consistent with a previous definition for a measured walk in discrete time [Phys.~Rev.~A~{\bf 73}, 032341 (2006)] (when generalised to allow for non-unitary evolution and in the limit of small time steps). A caveat of the quantum-jump approach is that it relies on the final state (the state which we want to hit) to share only incoherent edges with other vertices in the graph. We propose a simple remedy to restore the applicability of quantum jumps when this is not the case and show that the hitting-time statistics will again converge to that obtained from the measured discrete walk in appropriate limits. ",Hitting statistics from quantum jumps In this paper we study some properties of Fibonacci-sum set-graphs. The aforesaid graphs are an extension of the notion of Fibonacci-sum graphs to the notion of set-graphs. The colouring of Fibonacci-sum graphs is also discussed. A number of challenging research problems are posed in the conclusion. ,Some Properties of Fibonacci-Sum Set-Graphs " Supersymmetric gauge theories in four dimensions can display interesting non-perturbative phenomena. Although the superpotential dynamically generated by these phenomena can be highly nontrivial, it can often be exactly determined. We discuss some general techniques for analyzing the Wilsonian superpotential and demonstrate them with simple but non-trivial examples. ",Exact Superpotentials in Four Dimensions " In this paper we study the regularized analytic torsion of finite volume hyperbolic manifolds. We consider sequences of coverings $X_i$ of a fixed hyperbolic orbifold $X_0$. Our main result is that for certain sequences of coverings and strongly acyclic flat bundles, the analytic torsion divided by the index of the covering, converges to the $L^2$-torsion. Our results apply to certain sequences of arithmetic groups, in particular to sequences of principal congruence subgroups of $\SO^0(d,1)(\Z)$ and to sequences of principal congruence subgroups or Hecke subgroups of Bianchi groups. ",The analytic torsion and its asymptotic behaviour for sequences of hyperbolic manifolds of finite volume " Fluctuation measurements are important sources of information on the mechanism of particle production at LHC energies. This article reports the first experimental results on third-order cumulants of the net-proton distributions in Pb$-$Pb collisions at a center-of-mass energy $\sqrt{s_{\rm NN}} = 5.02$ TeV recorded by the ALICE detector. The results on the second-order cumulants of net-proton distributions at $\sqrt{s_{\rm NN}} = 2.76$ and $5.02$ TeV are also discussed in view of effects due to the global and local baryon number conservation. The results demonstrate the presence of long-range rapidity correlations between protons and antiprotons. Such correlations originate from the early phase of the collision. The experimental results are compared with HIJING and EPOS model calculations, and the dependence of the fluctuation measurements on the phase-space coverage is examined in the context of lattice quantum chromodynamics (LQCD) and hadron resonance gas (HRG) model estimations. The measured third-order cumulants are consistent with zero within experimental uncertainties of about 4% and are described well by LQCD and HRG predictions. ",Closing in on critical net-baryon fluctuations at LHC energies: cumulants up to third order in Pb$-$Pb collisions " Several Riemannian metrics and families of Riemannian metrics were defined on the manifold of Symmetric Positive Definite (SPD) matrices. Firstly, we formalize a common general process to define families of metrics: the principle of deformed metrics. We relate the recently introduced family of alpha-Procrustes metrics to the general class of mean kernel metrics by providing a sufficient condition under which elements of the former belongs to the latter. Secondly, we focus on the principle of balanced bilinear forms that we recently introduced. We give a new sufficient condition under which the balanced bilinear form is a metric. It allows us to introduce the Mixed-Euclidean (ME) metrics which generalize the Mixed-Power-Euclidean (MPE) metrics. We unveal their link with the (u, v)-divergences and the ($\alpha$, $\beta$)-divergences of information geometry and we provide an explicit formula of the Riemann curvature tensor. We show that the sectional curvature of all ME metrics can take negative values and we show experimentally that the sectional curvature of all MPE metrics but the log-Euclidean, power-Euclidean and power-affine metrics can take positive values. ",The geometry of mixed-Euclidean metrics on symmetric positive definite matrices " We briefly summarize motivations for testing the weak equivalence principle and then review recent torsion-balance results that compare the differential accelerations of beryllium-aluminum and beryllium-titanium test body pairs with precisions at the part in $10^{13}$ level. We discuss some implications of these results for the gravitational properties of antimatter and dark matter, and speculate about the prospects for further improvements in experimental sensitivity. ",Torsion-balance tests of the weak equivalence principle " The scarce knowledge of the initial stages of quark-gluon plasma before the thermalization is mostly inferred through the low-$p_\perp$ sector. We propose a complementary approach in this report - the use of high-$p_\perp$ probes' energy loss. We study the effects of four commonly assumed initial stages, whose temperature profiles differ only before the thermalization, on high-$p_\perp$ $R_{AA}$ and $v_2$ predictions. The predictions are based on our Dynamical Radiative and Elastic ENergy-loss Approach (DREENA) framework. We report insensitivity of $v_2$ to the initial stages, making it unable to distinguish between different cases. $R_{AA}$ displays sensitivity to the presumed initial stages, but current experimental precision does not allow resolution between these cases. We further revise the commonly accepted procedure of fitting the energy loss parameters, for each individual initial stage, to the measured $R_{AA}$. We show that the sensitivity of $v_2$ to various initial stages obtained through such procedure is mostly a consequence of fitting procedure, which may obscure the physical interpretations. Overall, the simultaneous study of high-$p_\perp$ observables, with unchanged energy loss parametrization and restrained temperature profiles, is crucial for future constraints on initial stages. ",Utilizing high-$p_\perp$ theory and data to constrain the initial stages of quark-gluon plasma " We report observations of nanosecond nanometer scale heterogeneous dynamics in a free flowing colloidal jet revealed by ultrafast x-ray speckle visibility spectroscopy. The nanosecond double-bunch mode of the Linac Coherent Light Source free electron laser enabled the production of pairs of femtosecond coherent hard x-ray pulses. By exploring the anisotropic summed speckle visibility which relates to the correlation functions, we are able to evaluate not only the average particle flow rate in a colloidal nanoparticle jet, but also the heterogeneous flow field within. The reported methodology presented here establishes the foundation for the study of nano- and atomic-scale heterogeneous fluctuations in complex matter using x-ray free electron laser sources. ",Nanoscale heterogeneous dynamics probed by nanosecond x-ray speckle visibility spectroscopy " Astronomers have proposed a number of mechanisms to produce supernova explosions. Although many of these mechanisms are now not considered primary engines behind supernovae, they do produce transients that will be observed by upcoming ground-based surveys and NASA satellites. Here we present the first radiation-hydrodynamics calculations of the spectra and light curves from three of these ""failed"" supernovae: supernovae with considerable fallback, accretion induced collapse of white dwarfs, and energetic helium flashes (also known as type .Ia supernovae). ",Spectra and Light Curves of Failed Supernovae " Differentially private deep learning has recently witnessed advances in computational efficiency and privacy-utility trade-off. We explore whether further improvements along the two axes are possible and provide affirmative answers leveraging two instantiations of \emph{group-wise clipping}. To reduce the compute time overhead of private learning, we show that \emph{per-layer clipping}, where the gradient of each neural network layer is clipped separately, allows clipping to be performed in conjunction with backpropagation in differentially private optimization. This results in private learning that is as memory-efficient and almost as fast per training update as non-private learning for many workflows of interest. While per-layer clipping with constant thresholds tends to underperform standard flat clipping, per-layer clipping with adaptive thresholds matches or outperforms flat clipping under given training epoch constraints, hence attaining similar or better task performance within less wall time. To explore the limits of scaling (pretrained) models in differentially private deep learning, we privately fine-tune the 175 billion-parameter GPT-3. We bypass scaling challenges associated with clipping gradients that are distributed across multiple devices with \emph{per-device clipping} that clips the gradient of each model piece separately on its host device. Privately fine-tuning GPT-3 with per-device clipping achieves a task performance at $\epsilon=1$ better than what is attainable by non-privately fine-tuning the largest GPT-2 on a summarization task. ",Exploring the Limits of Differentially Private Deep Learning with Group-wise Clipping " The different forms of propagation of relativistic electron plasma wavepackets in terms of Airy functions are studied. It is shown that exact solutions can be constructed showing accelerated propagations along coordinates transverse to the thermal speed cone coordinate. Similarly, Airy propagation is a solution for relativistic electron plasma waves in the paraxial approximation. This regime is considered in time-domain, when paraxial approximation is considered for frequency, and in space-domain, when paraxial approximation is considered for wavelength. In both different cases, the wavepackets remains structured in the transverse plane. Using these solutions we are able to define generalized and arbitrary Airy wavepackets for electron plasma waves, depeding on arbitrary spectral functions. Examples of this construction are presented. These electron plasma Airy wavepackets are the most general solutions of this kind. ",Exact and paraxial Airy propagation of relativistic electron plasma wavepackets " The Javalambre Photometric Local Universe Survey (J-PLUS) is an ongoing 12 band photometric optical survey, observing thousands of square degrees of the Northern Hemisphere from the dedicated JAST80 telescope at the Observatorio Astrof\'isico de Javalambre (OAJ). Observational strategy is a critical point in this large survey. To plan the best observations, it is necessary to select pointings depending on object visibility, the pointing priority and status and location and phase of the Moon. In this context, the J-PLUS Tracking Tool, a web application, has been implemented, which includes tools to plan the best observations, as well as tools to create the command files for the telescope; to track the observations; and to know the status of the survey. In this environment, robustness is an important point. To obtain it, a feedback software system has been implemented. This software automatically decides and marks which observations are valid or which must be repeated. It bases its decision on the data obtained from the data management pipeline database using a complex system of pointing and filter statuses. This contribution presents J-PLUS Tracking Tool and all feedback software system. ",J-PLUS Tracking Tool: Scheduler and Tracking software for the Observatorio Astrof\'isico de Javalambre (OAJ) " In 1935, Einstein, Podolsky and Rosen (EPR) questioned the completeness of quantum mechanics by devising a quantum state of two massive particles with maximally correlated space and momentum coordinates. The EPR criterion qualifies such continuous-variable entangled states, where a measurement of one subsystem seemingly allows for a prediction of the second subsystem beyond the Heisenberg uncertainty relation. Up to now, continuous-variable EPR correlations have only been created with photons, while the demonstration of such strongly correlated states with massive particles is still outstanding. Here, we report on the creation of an EPR-correlated two-mode squeezed state in an ultracold atomic ensemble. The state shows an EPR entanglement parameter of 0.18(3), which is 2.4 standard deviations below the threshold 1/4 of the EPR criterion. We also present a full tomographic reconstruction of the underlying many-particle quantum state. The state presents a resource for tests of quantum nonlocality and a wide variety of applications in the field of continuous-variable quantum information and metrology. ",Satisfying the Einstein-Podolsky-Rosen criterion with massive particles " We demonstrate, for the first time, a scheme that generates radially-polarized light using Goos-Hanchen shift of a cylindrically symmetric Total Internal Reflection. It allows ultra-broadband radial polarization conversion for wavelengths differing >1 micron. ",Ultra-Broadband Radial Polarization Conversion based on Goos-Hanchen Shift " We report a structural transition found in Ca10(Ir4As8)(Fe2-xIrxAs2)5, which exhibits superconductivity at 16 K. The c-axis parameter is doubled below a structural transition temperature of approximately 100 K, while the tetragonal symmetry with space group P4/n (No.85) is unchanged at all temperatures measured. Our synchrotron x-ray diffraction study clearly shows iridium ions at a non-coplanar position shift along the z-direction at the structural phase transition. We discuss that the iridium displacements affect superconductivity in Fe2As2 layers. ",Synchrotron X-ray Diffraction Study of Structural Phase Transition in Ca10(Ir4As8)(Fe2-xIrxAs2) A preliminary result of the solar axion search experiment at the University of Tokyo is presented. We searched for axions which could be produced in the solar core by exploiting the axion helioscope. The helioscope consists of a superconducting magnet with field strength of 4 Tesla over 2.3 meters. From the absence of the axion signal we set a 95 % confidence level upper limit on the axion coupling to two photons $g_{a\gamma\gamma} < 6.0 \times 10^{-10} GeV^{-1}$ for the axion mass $m_a < 0.03$ eV. This is the first solar axion search experiment whose sensitivity to $g_{a\gamma\gamma}$ exceeds the limit inferred from the solar age consideration. ,The Tokyo Axion Helioscope Experiment " We report on the development of commercially fabricated multi-chroic antenna coupled Transition Edge Sensor (TES) bolometer arrays for Cosmic Microwave Background (CMB) polarimetry experiments. CMB polarimetry experiments have deployed instruments in stages. Stage-II experiments deployed with O(1,000) detectors and reported successful detection of B-mode (divergent free) polarization pattern in the CMB. Stage-III experiments have recently started observing with O(10,000) detectors with wider frequency coverage. A concept for a Stage-IV experiment, CMB-S4, is emerging to make a definitive measurement of CMB polarization from the ground with O(400,000) detectors. The orders of magnitude increase in detector count for CMB-S4 requires a new approach in detector fabrication to increase fabrication throughput.and reduce cost. We report on collaborative efforts with two commercial micro-fabrication foundries to fabricate antenna coupled TES bolometer detectors. The detector design is based on the sinuous antenna coupled dichroic detector from the POLARBEAR-2 experiment. The TES bolometers showed the expected I-V response and the RF performance agrees with simulation. We will discuss the motivation, design consideration, fabrication processes, test results, and how industrial detector fabrication could be a path to fabricate hundreds of detector wafers for future CMB polarimetry experiments. ",Commercialization of micro-fabrication of antenna-coupled Transition Edge Sensor bolometer detectors for studies of the Cosmic Microwave Background " We produce synthetic images and SEDs from radiation hydrodynamical simulations of radiatively driven implosion. The synthetically imaged bright rimmed clouds (BRCs) are morphologically similar to those observed in star forming regions. Using nebular diagnostic line-ratios, simulated Very Large Array (VLA) radio images, H{\alpha} imaging and SED fitting we compute the neutral cloud and ionized boundary layer gas densities and temperatures and perform a virial stability analysis for each model cloud. We determine that the neutral cloud temperatures derived by SED fitting are hotter than the dominant neutral cloud temperature by 1 - 2 K due to emission from warm dust. This translates into a change in the calculated cloud mass by 8-35 %. Using a constant mass conversion factor (C{\nu}) for BRCs of different class is found to give rise to errors in the cloud mass of up to a factor of 3.6. The ionized boundary layer (IBL) electron temperature calculated using diagnostic line ratios is more accurate than assuming the canonical value adopted for radio diagnostics of 10^4 K. Both radio diagnostics and diagnostic line ratios are found to underestimate the electron density in the IBL. Each system is qualitatively correctly found to be in a state in which the pressure in the ionized boundary layer is greater than the supporting cloud pressure, implying that the objects are being compressed. We find that observationally derived mass loss estimates agree with those on the simulation grid and introduce the concept of using the mass loss flux to give an indication of the relative strength of photo-evaporative flow between clouds. The effect of beam size on these diagnostics in radio observations is found to be a mixing of the bright rim and ambient cloud and HII region fluxes, which leads to an underestimate of the cloud properties relative to a control diagnostic. ",Testing diagnostics of triggered star formation " In this work, we consider weighted anisotropic Hardy inequalities and trace Hardy inequalities involving a general Finsler metric. We follow a unifying approach, by establishing first a sharp interpolation between them, extending the corresponding nonweighted version, being established recently by a different approach. Then, passing to bounded domains, we obtain successive sharp improvements by adding remainder terms involving sharp weights and optimal constants, resulting in an infinite series-type improvement. The results extend, into the Finsler context, the earlier known ones within the Euclidean setting. The generalization of our results to cones is also discussed. ",Series expansion of weighted Finsler-Kato-Hardy inequalities " As learning-based methods make their way from perception systems to planning/control stacks, robot control systems have started to enjoy the benefits that data-driven methods provide. Because control systems directly affect the motion of the robot, data-driven methods, especially black box approaches, need to be used with caution considering aspects such as stability and interpretability. In this paper, we describe a differentiable and hierarchical control architecture. The proposed representation, called \textit{multi-abstractive neural controller}, uses the input image to control the transitions within a novel discrete behavior planner (referred to as the visual automaton generative network, or \textit{vAGN}). The output of a vAGN controls the parameters of a set of dynamic movement primitives which provides the system controls. We train this neural controller with real-world driving data via behavior cloning and show improved explainability, sample efficiency, and similarity to human driving. ",Multi-Abstractive Neural Controller: An Efficient Hierarchical Control Architecture for Interactive Driving " We give a unified description of twisted forms of classical reductive groups schemes. Such group schemes are constructed from algebraic objects of finite rank, excluding some exceptions of small rank. These objects, augmented odd form algebras, consist of $2$-step nilpotent groups with an action of the underlying commutative ring, hence we develop basic descent theory for them. In addition, we describe classical isotropic reductive groups as odd unitary groups up to an isogeny. ",Twisted forms of classical groups " Ontology matching is a core task when creating interoperable and linked open datasets. In this paper, we explore a novel structure-based mapping approach which is based on knowledge graph embeddings: The ontologies to be matched are embedded, and an approach known as absolute orientation is used to align the two embedding spaces. Next to the approach, the paper presents a first, preliminary evaluation using synthetic and real-world datasets. We find in experiments with synthetic data, that the approach works very well on similarly structured graphs; it handles alignment noise better than size and structural differences in the ontologies. ",Ontology Matching Through Absolute Orientation of Embedding Spaces " Our work aims to study tools offered to students and tutors involved in face-to-face or blended project- based learning activities. Project-based learning is often applied in the case of complex learning (i.e. which aims at making learners acquire various linked skills or develop their behaviours). In comparison to traditional learning, this type of learning relies on co-development, collective responsibility and co-operation. Learners are the principal actors of their learning. These trainings rest on rich and complex organizations, particularly for tutors, and it is difficult to apply innovative educational strategies. Our aim, in a bottom-up approach, is (1) to observe, according to Knowledge Management methods, a course characterized by these three criteria. The observed course concerns project management learning. Its observation allows us (2) to highlight and to analyze the problems encountered by the actors (students, tutors, designers) and (3) to propose tools to solve or improve them. We particularly study the relevance and the limits of the existing monitoring and experience sharing tools. We finally propose a result in the form of the tool MEShaT (Monitoring and Experience Sharing Tool) and end on the perspectives offered by these researches. ",Combiner suivi de l'activite? et partage d'exp\'eriences en apprentissage par projet pour les acteurs tuteurs et apprenants " The Automunge open source python library platform for tabular data pre-processing automates feature engineering data transformations of numerical encoding and missing data infill to received tidy data on bases fit to properties of columns in a designated train set for consistent and efficient application to subsequent data pipelines such as for inference, where transformations may be applied to distinct columns in ""family tree"" sets with generations and branches of derivations. Included in the library of transformations are methods to extract structure from bounded categorical string sets by way of automated string parsing, in which comparisons between entries in the set of unique values are parsed to identify character subset overlaps which may be encoded by appended columns of boolean overlap detection activations or by replacing string entries with identified overlap partitions. Further string parsing options, which may also be applied to unbounded categoric sets, include extraction of numeric substring partitions from entries or search functions to identify presence of specified substring partitions. The aggregation of these methods into ""family tree"" sets of transformations are demonstrated for use to automatically extract structure from categoric string compositions in relation to the set of entries in a column, such as may be applied to prepare categoric string set encodings for machine learning without human intervention. ",Parsed Categoric Encodings with Automunge " Optical losses degrade the sensitivity of laser interferometric instruments. They reduce the number of signal photons and introduce technical noise associated with diffuse light. In quantum-enhanced metrology, they break the entanglement between correlated photons. Such decoherence is one of the primary obstacles in achieving high levels of quantum noise reduction in precision metrology. In this work, we compare direct measurements of cavity and mirror losses in the Caltech 40m gravitational-wave detector prototype interferometer with numerical estimates obtained from semi-analytic intra-cavity wavefront simulations using mirror surface profile maps. We show a unified approach to estimating the total loss in optical cavities (such as the LIGO gravitational detectors) that will lead towards the engineering of systems with minimum decoherence for quantum-enhanced precision metrology. ",Scattering Loss in Precision Metrology due to Mirror Roughness " The sampling of the configuration space in diffusion Monte Carlo (DMC) is done using walkers moving randomly. In a previous work on the Hubbard model [\href{https://doi.org/10.1103/PhysRevB.60.2299}{Assaraf et al.~Phys.~Rev.~B \textbf{60}, 2299 (1999)}], it was shown that the probability for a walker to stay a certain amount of time in the same state obeys a Poisson law and that the on-state dynamics can be integrated out exactly, leading to an effective dynamics connecting only different states. Here, we extend this idea to the general case of a walker trapped within domains of arbitrary shape and size. The equations of the resulting effective stochastic dynamics are derived. The larger the average (trapping) time spent by the walker within the domains, the greater the reduction in statistical fluctuations. A numerical application to the Hubbard model is presented. Although this work presents the method for finite linear spaces, it can be generalized without fundamental difficulties to continuous configuration spaces. ",Diffusion Monte Carlo using domains in configuration space " Liquid capillary-bridge formation between solid particles has a critical influence on the rheological properties of granular materials and, in particular, on the efficiency of fluidized bed reactors. The available analytical and semi-analytical methods have inherent limitations, and often do not cover important aspects, like the presence of non-axisymmetric bridges. Here, we conduct numerical simulations of the capillary bridge formation between equally and unequally-sized solid particles using the lattice Boltzmann method, and provide an assessment of the accuracy of different families of analytical models. We find that some of the models taken into account are shown to perform better than others. However, all of them fail to predict the capillary force for contact angles larger than $\pi/2$, where a repulsive capillary force attempts to push the solid particle outwards to minimize the surface energy, especially at a small separation distance. ",Capillary-bridge Forces Between Solid Particles: Insights from Lattice Boltzmann Simulations " The advances in deep neural networks (DNN) have significantly enhanced real-time detection of anomalous data in IoT applications. However, the complexity-accuracy-delay dilemma persists: complex DNN models offer higher accuracy, but typical IoT devices can barely afford the computation load, and the remedy of offloading the load to the cloud incurs long delay. In this paper, we address this challenge by proposing an adaptive anomaly detection scheme with hierarchical edge computing (HEC). Specifically, we first construct multiple anomaly detection DNN models with increasing complexity, and associate each of them to a corresponding HEC layer. Then, we design an adaptive model selection scheme that is formulated as a contextual-bandit problem and solved by using a reinforcement learning policy network. We also incorporate a parallelism policy training method to accelerate the training process by taking advantage of distributed models. We build an HEC testbed using real IoT devices, implement and evaluate our contextual-bandit approach with both univariate and multivariate IoT datasets. In comparison with both baseline and state-of-the-art schemes, our adaptive approach strikes the best accuracy-delay tradeoff on the univariate dataset, and achieves the best accuracy and F1-score on the multivariate dataset with only negligibly longer delay than the best (but inflexible) scheme. ",Adaptive Anomaly Detection for Internet of Things in Hierarchical Edge Computing: A Contextual-Bandit Approach " In this paper, we show that existing recognition and localization deep architectures, that have not been exposed to eye tracking data or any saliency datasets, are capable of predicting the human visual saliency. We term this as implicit saliency in deep neural networks. We calculate this implicit saliency using expectancy-mismatch hypothesis in an unsupervised fashion. Our experiments show that extracting saliency in this fashion provides comparable performance when measured against the state-of-art supervised algorithms. Additionally, the robustness outperforms those algorithms when we add large noise to the input images. Also, we show that semantic features contribute more than low-level features for human visual saliency detection. ",Implicit Saliency in Deep Neural Networks " A Kerr microresonator frequency comb has enabled the generation of low-phase-noise millimeter- and terahertz-waves in conjunction with an ultrafast photodiode. It is intriguing to employ the new light source in wireless communication at above 100 GHz band, where a carrier signal with a high signal-to-noise ratio is desired to achieve higher data rates. In this study, we demonstrate two simple and efficient architectures of wireless links based on a microresonator comb. We show experimentally that simultaneous modulation and detection of multiple comb lines result in >10 times stronger modulation signal strength than two-line detection at a receiver. Successful transmission of complex modulation format up to 64 quadrature amplitude modulation proves that a microresonator comb and the proposed modulation method are effective in modern wireless communication. ",300 GHz wireless link based on an integrated Kerr soliton comb " Dark matter neutralinos in the constrained minimal supersymmetric model (CMSSM) may account for the recent cosmic ray electron and positron observations reported by the PAMELA and ATIC experiments either through self annihilation or via decay. However, to achieve this, both scenarios require new physics beyond the 'standard' CMSSM, and a unified explanation of the two experiments suggests a neutralino mass of order 700 GeV - 2 TeV. A relatively light neutralino with mass around 100 GeV (300 GeV) can accomodate the PAMELA but not the ATIC observations based on a model of annihilating (decaying) neutralinos. We study the implications of these scenarios for Higgs and sparticle spectroscopy in the CMSSM and highlight some benchmark points. An estimate of neutrino flux expected from the annihilating and decaying neutralino scenarios is provided. ",CMSSM Spectroscopy in light of PAMELA and ATIC " We explicitly compute the Dolbeault cohomologies of certain domains in complex space generalizing the classical Hartogs figure. The cohomology groups are non-Hausdorff topological vector spaces, and it is possible to identify the reduced (Hausdorff) and the indiscrete part of the cohomology. ",Some non-pseudoconvex domains with explicitly computable non-Hausdorff Dolbeault cohomology " We have studied a mathematical relationship between holographic Wilsonian renormalization group(HWRG) and stochastic quantization(SQ) of scalar field with arbitrary mass in AdS spacetime. In the stochastic theory, the field is described by an equation with a form of harmonic oscillator with time dependent frequency and its Euclidean action also shows explicit time dependent kernel in it. We have obtained the stochastic 2-point correlation function and demonstrate that it reproduces the radial evolution of the double trace operator correctly via the suggested relation given in arXiv:1209.2242. Moreover, we justify our stochastic procedure with time dependent kernel by showing that it can be mapped to a new stochastic frame with a standard kernel without time dependence. Finally, we consider more general boundary conditions for the stochastic field to reproduce the radial evolution of the holographic boundary effective action when alternative quantization is allowed. We extensively study the Neumann boundary condition case and confirm that even in this case, the relation between HWRG and SQ is precisely hold. ",Stochastic quantization and holographic Wilsonian renormalization group of scalar theories with arbitrary mass " A particular case of initial data for the two-dimensional Euler equations is studied numerically. The results show that the Godunov method does not always converge to the physical solution, at least not on feasible grids. Moreover, they suggest that entropy solutions (in the weak entropy inequality sense) are not well-posed. ",A possible counterexample to wellposedness of entropy solutions and to Godunov scheme convergence " This thesis concerns the study of random walks in random environments (RWRE). Since there are two levels of randomness for random walks in random environments, there are two different distributions for the random walk that can be studied. The quenched distribution is the law of the random walk conditioned on a given environment. The annealed distribution is the quenched law averaged over all environments. The main results of the thesis fall into two categories: quenched limiting distributions for one-dimensional, transient RWRE and annealed large deviations for multidimensional RWRE. The analysis of the quenched distributions for transient, one-dimensional RWRE falls into two separate cases. First, when an annealed central limit theorem holds, we prove that a quenched central limit theorem also holds but with a random (depending on the environment) centering. In contrast, when the annealed limit distribution is not Gaussian, we prove that there is no quenched limiting distribution for the RWRE. Moreover, we show that for almost every environment, there exist two random (depending on the environment) sequences of times, along which random walk has different quenched limiting distributions. While an annealed large deviation principle for multidimensional RWRE was known previously, very little qualitative information was available about the annealed large deviation rate function. We prove that if the law on environments is non-nestling, then the annealed large deviation rate function is analytic in a neighborhood of its unique zero (which is the limiting velocity of the RWRE). ",Limiting distributions and large deviations for random walks in random environments " The electric quadrupole moment and the magnetic moment of the 11Li halo nucleus have been measured with more than an order of magnitude higher precision than before, |Q| = 33.3(5)mb and mu=3.6712(3)mu_N, revealing a 8.8(1.5)% increase of the quadrupole moment relative to that of 9Li. This result is compared to various models that aim at describing the halo properties. In the shell model an increased quadrupole moment points to a significant occupation of the 1d orbits, whereas in a simple halo picture this can be explained by relating the quadrupole moments of the proton distribution to the charge radii. Advanced models so far fail to reproduce simultaneously the trends observed in the radii and quadrupole moments of the lithium isotopes. ",Precision Measurement of 11Li moments: Influence of Halo Neutrons on the 9Li Core " This is an expository article of our work on analogies between knot theory and algebraic number theory. We shall discuss foundational analogies between knots and primes, 3-manifolds and number rings mainly from the group-theoretic point of view. ","Analogies between Knots and Primes, 3-Manifolds and Number Rings" " We present a new diagnostic diagram for local ultraluminous infrared galaxies (ULIRGs) and quasars, analysing particularly the Spitzer Space Telescope's Infrared Spectrograph (IRS) spectra of 102 local ULIRGs and 37 Palomar Green quasars. Our diagram is based on a special non-linear mapping of these data, employing the Kernel Principal Component Analysis method. The novelty of this map lies in the fact that it distributes the galaxies under study on the surface of a well-defined ellipsoid, which, in turn, links basic concepts from geometry to physical properties of the galaxies. Particularly, we have found that the equatorial direction of the ellipsoid corresponds to the evolution of the power source of ULIRGs, starting from the pre-merger phase, moving through the starburst-dominated coalescing stage towards the active galactic nucleus (AGN)-dominated phase, and finally terminating with the post-merger quasar phase. On the other hand, the meridian directions distinguish deeply obscured power sources of the galaxies from unobscured ones. These observations have also been verified by comparison with simulated ULIRGs and quasars using radiative transfer models. The diagram correctly identifies unique galaxies with extreme features that lie distinctly away from the main distribution of the galaxies. Furthermore, special two-dimensional projections of the ellipsoid recover almost monotonic variations of the two main physical properties of the galaxies, the silicate and PAH features. This suggests that our diagram naturally extends the well-known Spoon diagram and it can serve as a diagnostic tool for existing and future infrared spectroscopic data, such as those provided by the James Webb Space Telescope. ",Classification of local ultraluminous infrared galaxies and quasars with kernel principal component analysis " We compute an explicit algebraic deformation quantization for an affine Poisson variety described by an ideal in a polynomial ring, and inheriting its Poisson structure from the ambient space. ",On the deformation quantization of affine algebraic varieties " To understand the essence of the exciton Mott transition in three-dimensional electron-hole systems, the metal-insulator transition is studied for a two-band Hubbard model in infinite dimensions with interactions of electron-electron (hole-hole) repulsion U and electron-hole attraction -U'. By using the dynamical mean-field theory, the phase diagram in the U-U' plane is obtained (which is exact in infinite dimensions) assuming that electron-hole pairs do not condense. When both electron and hole bands are half-filled, two types of insulating states appear: the Mott-Hubbard insulator for U > U' and the biexciton-like insulator for U < U'. Even when away from half-filling, we find the phase transition between the exciton- or biexciton-like insulator and a metallic state. This transition can be assigned to the exciton Mott transition, whereas the Mott-Hubbard transition is absent. ",Phase diagram for the exciton Mott transition in infinite-dimensional electron-hole systems " We study correlations of hydrodynamic fluctuations in shear flow analytically and also by dissipative particle dynamics~(DPD) simulations. The hydrodynamic equations are linearized around the macroscopic velocity field and then solved by a perturbation method in Fourier-transformed space. The autocorrelation functions~(ACFs) from the analytical method are compared with results obtained from DPD simulations under the same shear-flow conditions. Upto a moderate shear rate, various ACFs from the two approaches agree with each other well. At large shear rates, discrepancies between the two methods are observed, hence revealing strong additional coupling between different fluctuating variables, which is not considered in the analytical approach. In addition, the results at low and moderate shear rates can serve as benchmarks for developing multiscale algorithms for coupling of heterogeneous solvers, such as a hybrid simulation of molecular dynamics and fluctuating hydrodynamics solver, where thermal fluctuations are indispensable. ",Analytical and Computational Studies of Correlations of Hydrodynamic Fluctuations in Shear Flow " Green algae of the $Volvocine$ lineage, spanning from unicellular $Chlamydomonas$ to vastly larger $Volvox$, are models for the study of the evolution of multicellularity, flagellar dynamics, and developmental processes. Phototactic steering in these organisms occurs without a central nervous system, driven solely by the response of individual cells. All such algae spin about a body-fixed axis as they swim; directional photosensors on each cell thus receive periodic signals when that axis is not aligned with the light. The flagella of $Chlamydomonas$ and $Volvox$ both exhibit an adaptive response to such signals in a manner that allows for accurate phototaxis, but in the former the two flagella have distinct responses, while the thousands of flagella on the surface of spherical $Volvox$ colonies have essentially identical behaviour. The planar 16-cell species $Gonium~pectorale$ thus presents a conundrum, for its central 4 cells have a $Chlamydomonas$-like beat that provide propulsion normal to the plane, while its 12 peripheral cells generate rotation around the normal through a $Volvox$-like beat. Here, we combine experiment, theory, and computations to reveal how $Gonium$, perhaps the simplest differentiated colonial organism, achieves phototaxis. High-resolution cell tracking, particle image velocimetry of flagellar driven flows, and high-speed imaging of flagella on micropipette-held colonies show how, in the context of a recently introduced model for $Chlamydomonas$ phototaxis, an adaptive response of the peripheral cells alone leads to photo-reorientation of the entire colony. The analysis also highlights the importance of local variations in flagellar beat dynamics within a given colony, which can lead to enhanced reorientation dynamics. ","Motility and Phototaxis of $Gonium$, the Simplest Differentiated Colonial Alga" " We report the detection of the sulfur-bearing species NCS, HCCS, H2CCS, H2CCCS, and C4S for the first time in space. These molecules were found towards TMC-1 through the observation of several lines for each species. We also report the detection of C5S for the first time in a cold cloud through the observation of five lines in the 31-50 GHz range. The derived column densities are N(NCS) = (7.8 +/- 0.6)e11 cm-2, N(HCCS) = (6.8 +/- 0.6)e11 cm-2, N(H2CCS) = (7.8 +/- 0.8)e11 cm-2, N(H2CCCS) = (3.7 +/- 0.4)e11 cm-2, N(C4S) = (3.8 +/- 0.4)e10 cm-2, and N(C5S) = (5.0 +/- 1.0)e10 cm-2. The observed abundance ratio between C3S and C4S is 340, that is to say a factor of approximately one hundred larger than the corresponding value for CCS and C3S. The observational results are compared with a state-of-the-art chemical model, which is only partially successful in reproducing the observed abundances. These detections underline the need to improve chemical networks dealing with S-bearing species. ","TMC-1, the starless core sulfur factory: Discovery of NCS, HCCS, H2CCS, H2CCCS, and C4S and detection of C5S" " In this note we generalize a result from a recent paper of Hajac, Reznikoff and Tobolski (2020). In that paper they give conditions they call admissibility on a pushout diagram in the category of directed graphs implying that the $C^*$-algebras of the graphs form a pullback diagram. We consider a larger category of relative graphs that correspond to relative Toeplitz graph algebras. In this setting we give necessary and sufficient conditions on the pushout to get a pullback of $C^*$-algebras. ",Relative graphs and pullbacks of relative Toeplitz graph algebras " This Signal Processing Grand Challenge (SPGC) targets a difficult automatic prediction problem of societal and medical relevance, namely, the detection of Alzheimer's Dementia (AD). Participants were invited to employ signal processing and machine learning methods to create predictive models based on spontaneous speech data. The Challenge has been designed to assess the extent to which predictive models built based on speech in one language (English) generalise to another language (Greek). To the best of our knowledge no work has investigated acoustic features of the speech signal in multilingual AD detection. Our baseline system used conventional machine learning algorithms with Active Data Representation of acoustic features, achieving accuracy of 73.91% on AD detection, and 4.95 root mean squared error on cognitive score prediction. ",Multilingual Alzheimer's Dementia Recognition through Spontaneous Speech: a Signal Processing Grand Challenge " In this paper, we analyze the polarized muon decay at rest (PMDaR) and elastic neutrino-electron scattering (ENES) admitting the non-standard V+A interaction in addition to standard V-A interaction. Considerations are made for Dirac massive muon neutrino and electron antineutrino. Moreover, muon neutrinos are transversely polarized. It means that the outgoing muon-neutrino beam is a mixture of the left- and right-chirality muon neutrinos and has a fixed direction of transverse spin polarization with respect to production plane. We show that the angle-energy distribution of muon neutrinos contains the interference terms between the standard V-A and exotic V+A couplings, which are proportional to the transverse components of muon neutrino spin polarization. They do not vanish in a limit of massless neutrino and include the relative phases to test the CP violation. In consequence, it allows to calculate a neutrino flux and an expected event number in the ENES (detection process) both for the standard model prediction and the case of neutrino left-right mixture. ",Polarized Muon Decay at Rest with V+A Interaction " By using Girsanov transformation and martingale representation, Talagrand-type transportation cost inequalities, with respect to both the uniform and the $L^2$ distances on the global free path space, are established for the segment process associated to a class of neutral functional stochastic differential equations. Neutral functional stochastic partial differential equations are also investigated. ",Transportation Cost Inequalities for Neutral Functional Stochastic Equations " From a linear stability analysis of the Gross Pitaevskii equation for binary Bose Einstein condensates, it is found that the uniform state becomes unstable to a periodic perturbation of wave number k if k exceeds a critical value kc. However we find that a stationary spatially periodic state does not exist. We show the existence of pulse type solutions, when the pulse structure for one condensate is strongly influenced by the presence of the other condensate. ",Stability and The Existence of Coherent Structure in Demixed State of Binary BEC " It is an old and challenging topic to investigate for which discrete groups G the full group C*-algebra C*(G) is residually finite-dimensional (RFD). In particular not much is known about how the RFD property behaves under fundamental constructions, such as amalgamated free products and HNN-extensions. In [CS19] it was proved that central amalgamated free products of virtually abelian groups are RFD. In this paper we prove that this holds much beyond this case. Our method is based on showing a certain approximation property for characters induced from central subgroups. In particular it allows us to prove that free products of polycyclic-by-finite groups amalgamated over finitely generated central subgroups are RFD. On the other hand we prove that the class of RFD C*-algebras (and groups) is not closed under central amalgamated free products. Namely we give an example of RFD groups (in fact finitely generated amenable RF groups) whose central amalgamated free product is not RFD, moreover it is not even maximally almost periodic. This answers a question of Khan and Morris [KM82]. ",Central amalgamation of groups and the RFD property " Photo-induced processes are fundamental in nature, but accurate simulations are seriously limited by the cost of the underlying quantum chemical calculations, hampering their application for long time scales. Here we introduce a method based on machine learning to overcome this bottleneck and enable accurate photodynamics on nanosecond time scales, which are otherwise out of reach with contemporary approaches. Instead of expensive quantum chemistry during molecular dynamics simulations, we use deep neural networks to learn the relationship between a molecular geometry and its high-dimensional electronic properties. As an example, the time evolution of the methylenimmonium cation for one nanosecond is used to demonstrate that machine learning algorithms can outperform standard excited-state molecular dynamics approaches in their computational efficiency while delivering the same accuracy. ",Machine learning enables long time scale molecular photodynamics simulations " The network studied here is based on a standard model in physics, but it appears in various applications ranging from spintronics to neuroscience. When the network is forced by an external signal common to all its elements, there are shown to be two potential (gradient) functions: One for amplitudes and one for phases. But the phase potential disappears when the forcing is removed. The phase potential describes the distribution of in-phase/anti-phase oscillations in the network, as well as resonances in the form of phase locking. A valley in a potential surface corresponds to memory that may be accessed by associative recall. The two potentials derived here exhibit two different forms of memory: structural memory (time domain memory) that is sustained in the free problem, and evoked memory (frequency domain memory) that is sustained by the phase potential, only appearing when the system is illuminated by common external forcing. The common forcing organizes the network into those elements that are locked to forcing frequencies and other elements that may form secluded sub-networks. The secluded networks may perform independent operations such as pattern recognition and logic computations. Various control methods for shaping the network's outputs are demonstrated. ",A Frequency-Phase Potential for a Forced STNO Network: an Example of Evoked Memory " The standard notion of NS-NS 3-form flux is lifted to Hitchin's generalized geometry. This generalized flux is given in terms of an integral of a modified Nijenhuis operator over a generalized 3-cycle. Explicitly evaluating the generalized flux in a number of familiar examples, we show that it can compute three-form flux, geometric flux and non-geometric Q-flux. Finally, a generalized connection that acts on generalized vectors is described and we show how the flux arises from it. ",NS-NS fluxes in Hitchin's generalized geometry We provide convincing empirical evidence that long range interactions strongly enhance the rectification effect which takes place in mass graded systems. Even more importantly the rectification does not decrease with the increase of the system size. Large rectification is obtained also for the equal mass case and with graded on-site potential. These results allow to overcome current limitations of the rectification mechanism and open the way for a realistic implementation of efficient thermal diodes. ,Ingredients for an efficient thermal diode " We present a new family of exact four-dimensional Taub-NUT spacetimes in Einstein-$\Lambda$ theory supplemented with a conformally coupled scalar field exhibiting a power-counting super-renormalizable potential. The construction proceeds as follows: A solution of a conformally coupled theory with a conformal potential, henceforth the seed $(g_{\mu\nu},\phi)$, is transformed by the action of a specific change of frame in addition with a simultaneous shift of the seed scalar. The new configuration, $(\bar{g}_{\mu\nu},\bar{\phi})$, solves the field equations of a conformally coupled theory with the aforementioned super-renormalizable potential. The solution spectrum of the seed is notoriously enhanced. We highlight the existence of two types of exact black bounces given by de Sitter and anti-de Sitter geometries that transit across three different configurations each. The de Sitter geometries transit from a regular black hole with event and cosmological horizons to a bouncing cosmology connecting two de Sitter Universes with different values of the asymptotic cosmological constant. An intermediate phase represented by a de Sitter wormhole or by a bouncing cosmology that connects two de Sitter Universes is shown, this under the presence of a cosmological horizon. On the other hand, the anti-de Sitter geometries transit from a regular black hole with inner and event horizons to a wormhole that connects two asymptotic boundaries with different constant curvatures. The intermediate phase is given by an anti-de Sitter regular black hole with a single event horizon that appears in two different settings. As a regular anti-de Sitter black hole inside of an anti-de Sitter wormhole or as an anti-de Sitter regular black hole with an internal cosmological bounce. These geometries are smoothly connected by the mass parameter only. Other black holes, bouncing cosmologies and wormholes are also found. ",AdS-Taub-NUT spacetimes and exact black bounces with scalar hair " Assume that abelian categories $A, B$ over a field admit countable direct limits and that these limits are exact. Let $F: D^+_{dg}(A) --> D^+_{dg}(B)$ be a DG quasi-functor such that the functor $Ho(F): D^+(A) \to D^+(B)$ carries $D^{\geq 0}(A)$ to $D^{\geq 0}(B)$ and such that, for every $i>0$, the functor $H^i F: A \to B$ is effaceable. We prove that $F$ is canonically isomorphic to the right derived DG functor $RH^0(F)$. We also prove a similar result for bounded derived DG categories in a more general setting. We give an example showing that the corresponding statements for triangulated functors are false. We prove a formula that expresses Hochschild cohomology of the categories $ D^b_{dg}(A)$, $ D^+_{dg}(A) $ as the $Ext$ groups in the abelian category of left exact functors $A \to Ind B$ . ",On the derived DG functors " We develop the theory of linear algebra over a (Z_2)^n-commutative algebra (n in N), which includes the well-known super linear algebra as a special case (n=1). Examples of such graded-commutative algebras are the Clifford algebras, in particular the quaternion algebra H. Following a cohomological approach, we introduce analogues of the notions of trace and determinant. Our construction reduces in the classical commutative case to the coordinate-free description of the determinant by means of the action of invertible matrices on the top exterior power, and in the supercommutative case it coincides with the well-known cohomological interpretation of the Berezinian. ",Cohomological Approach to the Graded Berezinian " In this paper, we compare the regularities of symbolic and ordinary powers of edge ideals of weighted oriented graphs. For a weighted oriented graph $D$, we give a lower bound for $\reg(I(D)^{k})$, if $V^+$ are sinks. If $D$ has an induced directed path $(x_i,x_j),(x_j,x_r) \in E(D)$ of length $2$ with $w(x_j)\geq 2$, then we show that $\reg(I(D)^{k})\leq \reg(I(D)^{k})$ for all $k\geq 2$. In particular, if $D$ is bipartite, then the above inequality holds for all $k\geq 2$. For any weighted oriented graph $D$, if $V^+$ are sink vertices, then we show that $\reg(I(D)^{k}) \leq \reg(I(D)^{k})$ with $k=2,3$. We further study when these regularities are equal. As a consequence, we give sharp upper bounds for regularity of symbolic powers of certain class of weighted oriented graphs. ",Regularity comparison of symbolic and ordinary powers of weighted oriented graphs " We report on the realization and characterization of a magnetic microtrap for ultra cold atoms near a straight superconducting Nb wire with circular cross section. The trapped atoms are used to probe the magnetic field outside the superconducting wire. The Meissner effect shortens the distance between the trap and the wire, reduces the radial magnetic-field gradients and lowers the trap depth. Measurements of the trap position reveal a complete exclusion of the magnetic field from the superconducting wire for temperatures lower than 6K. As the temperature is further increased, the magnetic field partially penetrates the superconducting wire; hence the microtrap position is shifted towards the position expected for a normal-conducting wire. ",Meissner effect in superconducting microtraps " Let $u$ be a positive harmonic function in the unit ball $B_1 \subset \mathbb{R}^n$ and let $\mu$ be the boundary measure of $u$. Consider a point $x\in \partial B_1$ and let $n(x)$ denote the unit normal vector at $x$. Let $\alpha$ be a number in $(-1,n-1]$ and $A \in [0,+\infty) $. We prove that $u(x+n(x)t)t^{\alpha} \to A$ as $t \to +0$ if and only if $\frac{\mu({B_r(x)})}{r^{n-1}} r^{\alpha} \to C_\alpha A$ as $r\to+0$, where ${C_\alpha= \frac{\pi^{n/2}}{\Gamma(\frac{n-\alpha+1}{2})\Gamma(\frac{\alpha+1}{2})}}$. For $\alpha=0$ it follows from the theorems by Rudin and Loomis which claim that a positive harmonic function has a limit along the normal iff the boundary measure has the derivative at the corresponding point of the boundary. For $\alpha=n-1$ it concerns about the point mass of $\mu$ at $x$ and it follows from the Beurling minimal principle. For the general case of $\alpha \in (-1,n-1)$ we prove it with the help of the Wiener Tauberian theorem in a similar way to Rudin's approach. Unfortunately this approach works for a ball or a half-space only but not for a general kind of domain. In dimension $2$ one can use conformal mappings and generalise the statement above to sufficiently smooth domains, in dimension $n\geq 3$ we showed that this generalisation is possible for $\alpha\in [0,n-1]$ due to harmonic measure estimates. The last method leads to an extension of the theorems by Loomis, Ramey and Ullrich on non-tangential limits of harmonic functions to positive solutions of elliptic differential equations with Holder continuous coefficients. ",On the Boundary Behavior of Positive Solutions of Elliptic Differential Equations " In this paper, we study geometric properties of basins of attraction of monotone systems. Our results are based on a combination of monotone systems theory and spectral operator theory. We exploit the framework of the Koopman operator, which provides a linear infinite-dimensional description of nonlinear dynamical systems and spectral operator-theoretic notions such as eigenvalues and eigenfunctions. The sublevel sets of the dominant eigenfunction form a family of nested forward-invariant sets and the basin of attraction is the largest of these sets. The boundaries of these sets, called isostables, allow studying temporal properties of the system. Our first observation is that the dominant eigenfunction is increasing in every variable in the case of monotone systems. This is a strong geometric property which simplifies the computation of isostables. We also show how variations in basins of attraction can be bounded under parametric uncertainty in the vector field of monotone systems. Finally, we study the properties of the parameter set for which a monotone system is multistable. Our results are illustrated on several systems of two to four dimensions. ",Geometric Properties of Isostables and Basins of Attraction of Monotone Systems " Phase transitions are characterized by a sharp change in the type of dynamics of microparticles, and their description usually requires quantum mechanics. Recently, a peculiar type of conductors was discovered in which two-dimensional (2D) electrons form a viscous fluid. In this work we reveal that such electron fluid in high-quality samples can be formed from ballistic electrons via a phase transition. For this purpose, we theoretically study the evolution of a ballistic flow of 2D weakly interacting electrons with an increase of magnetic field and trace an emergence of a fluid fraction at a certain critical field. Such restructuring of the flow manifests itself in a kink in magnetic-field dependencies of the longitudinal and the Hall resistances. It is remarkable that the studied phase transition has a classical-mechanical origin and is determined by both the ballistic size effects and the electron-electron scattering. Our analysis shows that this effect was apparently observed in the recent transport experiments on 2D electrons in graphene and high-mobility GaAs quantum wells. ",Ballistic-hydrodynamic phase transition in flow of two-dimensional electrons " Oscillations from high energy photons into light pseudoscalar particles in an external magnetic field are expected to occur in some extensions of the standard model. It is usually assumed that those axionlike particles (ALPs) could produce a drop in the energy spectra of gamma ray sources and possibly decrease the opacity of the Universe for TeV gamma rays. We show here that these assumptions are in fact based on an average behavior that cannot happen in real observations of single sources. We propose a new method to search for photon-ALP oscillations, taking advantage of the fact that a single observation would deviate from the average expectation. Our method is based on the search for irregularities in the energy spectra of gamma ray sources. We predict features that are unlikely to be produced by known astrophysical processes and a new signature of ALPs that is easily falsifiable. ",Irregularity in gamma ray source spectra as a signature of axionlike particles " With the hit of new pandemic threats, scientific frameworks are needed to understand the unfolding of the epidemic. The use of mobile apps that are able to trace contacts is of utmost importance in order to control new infected cases and contain further propagation. Here we present a theoretical approach using both percolation and message--passing techniques, to the role of contact tracing, in mitigating an epidemic wave. We show how the increase of the app adoption level raises the value of the epidemic threshold, which is eventually maximized when high-degree nodes are preferentially targeted. Analytical results are compared with extensive Monte Carlo simulations showing good agreement for both homogeneous and heterogeneous networks. These results are important to quantify the level of adoption needed for contact-tracing apps to be effective in mitigating an epidemic. ",A message-passing approach to epidemic tracing and mitigation with apps " Optimal caching of files in a content distribution network (CDN) is a problem of fundamental and growing commercial interest. Although many different caching algorithms are in use today, the fundamental performance limits of network caching algorithms from an online learning point-of-view remain poorly understood to date. In this paper, we resolve this question in the following two settings: (1) a single user connected to a single cache, and (2) a set of users and a set of caches interconnected through a bipartite network. Recently, an online gradient-based coded caching policy was shown to enjoy sub-linear regret. However, due to the lack of known regret lower bounds, the question of the optimality of the proposed policy was left open. In this paper, we settle this question by deriving tight non-asymptotic regret lower bounds in both of the above settings. In addition to that, we propose a new Follow-the-Perturbed-Leader-based uncoded caching policy with near-optimal regret. Technically, the lower-bounds are obtained by relating the online caching problem to the classic probabilistic paradigm of balls-into-bins. Our proofs make extensive use of a new result on the expected load in the most populated half of the bins, which might also be of independent interest. We evaluate the performance of the caching policies by experimenting with the popular MovieLens dataset and conclude the paper with design recommendations and a list of open problems. ",Fundamental Limits of Online Network-Caching " HTTP/2 (h2) is a new standard for Web communications that already delivers a large share of Web traffic. Unlike HTTP/1, h2 uses only one underlying TCP connection. In a cellular network with high loss and sudden spikes in latency, which the TCP stack might interpret as loss, using a single TCP connection can negatively impact Web performance. In this paper, we perform an extensive analysis of real world cellular network traffic and design a testbed to emulate loss characteristics in cellular networks. We use the emulated cellular network to measure h2 performance in comparison to HTTP/1.1, for webpages synthesized from HTTP Archive repository data. Our results show that, in lossy conditions, h2 achieves faster page load times (PLTs) for webpages with small objects. For webpages with large objects, h2 degrades the PLT. We devise a new domain-sharding technique that isolates large and small object downloads on separate connections. Using sharding, we show that under lossy cellular conditions, h2 over multiple connections improves the PLT compared to h2 with one connection and HTTP/1.1 with six connections. Finally, we recommend content providers and content delivery networks to apply h2-aware domain-sharding on webpages currently served over h2 for improved mobile Web performance. ",Domain-Sharding for Faster HTTP/2 in Lossy Cellular Networks Two different constructions of an invariant of an odd dimensional hyperbolic manifold in the K-group $K_{2n-1}(\bar \Bbb Q)\otimes \Bbb Q$ are given. The volume of the manifold is equal to the value of the Borel regulator on that element. The scissor congruence groups in non euclidian geometries are studied and their relationship with algebraic K-theory of the field of complex numbers is discussed. ,Volumes of hyperbolic manifolds and mixed Tate motives " Pulmonary embolus (PE) refers to obstruction of pulmonary arteries by blood clots. PE accounts for approximately 100,000 deaths per year in the United States alone. The clinical presentation of PE is often nonspecific, making the diagnosis challenging. Thus, rapid and accurate risk stratification is of paramount importance. High-risk PE is caused by right ventricular (RV) dysfunction from acute pressure overload, which in return can help identify which patients require more aggressive therapy. Reconstructed four-chamber views of the heart on chest CT can detect right ventricular enlargement. CT pulmonary angiography (CTPA) is the golden standard in the diagnostic workup of suspected PE. Therefore, it can link between diagnosis and risk stratification strategies. We developed a weakly supervised deep learning algorithm, with an emphasis on a novel attention mechanism, to automatically classify RV strain on CTPA. Our method is a 3D DenseNet model with integrated 3D residual attention blocks. We evaluated our model on a dataset of CTPAs of emergency department (ED) PE patients. This model achieved an area under the receiver operating characteristic curve (AUC) of 0.88 for classifying RV strain. The model showed a sensitivity of 87% and specificity of 83.7%. Our solution outperforms state-of-the-art 3D CNN networks. The proposed design allows for a fully automated network that can be trained easily in an end-to-end manner without requiring computationally intensive and time-consuming preprocessing or strenuous labeling of the data.We infer that unmarked CTPAs can be used for effective RV strain classification. This could be used as a second reader, alerting for high-risk PE patients. To the best of our knowledge, there are no previous deep learning-based studies that attempted to solve this problem. ",Weakly Supervised Attention Model for RV StrainClassification from volumetric CTPA Scans " In the present article we propose a new hybrid shape function for wormhole (WH)s in the modified $f(R,T)$ gravity. The proposed shape function satisfied the conditions of WH geometry. Geometrical behavior of WH solutions are discussed in both anisotropic and isotropic cases respectively. Also, the stability of this model is obtained by determining the equilibrium condition. The radial null energy condition and weak energy condition are validated in the proposed shape function indicating the absence of exotic matter in modified $f(R,T)$ gravity. ","Wormhole model with a hybrid shape function in f(R,T) gravity" " Audio-visual multi-modal modeling has been demonstrated to be effective in many speech related tasks, such as speech recognition and speech enhancement. This paper introduces a new time-domain audio-visual architecture for target speaker extraction from monaural mixtures. The architecture generalizes the previous TasNet (time-domain speech separation network) to enable multi-modal learning and at meanwhile it extends the classical audio-visual speech separation from frequency-domain to time-domain. The main components of proposed architecture include an audio encoder, a video encoder that extracts lip embedding from video streams, a multi-modal separation network and an audio decoder. Experiments on simulated mixtures based on recently released LRS2 dataset show that our method can bring 3dB+ and 4dB+ Si-SNR improvements on two- and three-speaker cases respectively, compared to audio-only TasNet and frequency-domain audio-visual networks ",Time Domain Audio Visual Speech Separation Sum of the time component of $\Sigma$ term and the induced pseudo scalar term in axial current is shown to be the t-channel pion pole in Born terms for pion electroproduction near threshold. We also show that this $\Sigma$ term represents the charged pseudo scalar quark density matrix elements in nucleon and manifests itself in the $L_0^+$ amplitude on this reaction. ,$\Sigma$ Like Term in Pion Electroproduction Near Threshold We indicate the tentative source of instability in the two-dimensional black hole background. There are relevant operators among the tachyon and the higher level vertex operators in the conformal field theory. Connection of this instability with Hawking radiation is not obvious. The situation is somewhat analogous to fields in the background of a negative mass Euclidean Schwarzschild solution (in four dimensions). Speculation is made about decay of the Minkowski black hole into finite temperature flat space. ,Instabilities in the gravitational background and string theory " This paper considers the distributed information bottleneck (D-IB) problem for a primitive Gaussian diamond channel with two relays and MIMO Rayleigh fading. The channel state is an independent and identically distributed (i.i.d.) process known at the relays but unknown to the destination. The relays are oblivious, i.e., they are unaware of the codebook and treat the transmitted signal as a random process with known statistics. The bottleneck constraints prevent the relays to communicate the channel state information (CSI) perfectly to the destination. To evaluate the bottleneck rate, we provide an upper bound by assuming that the destination node knows the CSI and the relays can cooperate with each other, and also two achievable schemes with simple symbol-by-symbol relay processing and compression. Numerical results show that the lower bounds obtained by the proposed achievable schemes can come close to the upper bound on a wide range of relevant system parameters. ",Distributed Information Bottleneck for a Primitive Gaussian Diamond MIMO Channel " We consider global monopoles as well as black holes with global monopole hair in Einstein-Goldstone model with a cosmological constant in four spacetime dimensions. Similar to the $\Lambda=0$ case, the mass of these solutions defined in the standard way diverges. We use a boundary counterterm subtraction method to compute the mass and action of $\Lambda \neq 0$ configurations. The mass of the asymptotically de Sitter solutions computed in this way turns out to take positive values in a specific parameter range and, for a relaxed set of asymptotic boundary conditions, yields a counterexample to the maximal mass conjecture. ","Global monopoles, cosmological constant and maximal mass conjecture" " Besov-type and Triebel-Lizorkin-type spaces $\dot B^{s,\tau}_{p,q}$ and $\dot F^{s,\tau}_{p,q}$ on $\mathbb{R}^n$ consist of a general family of function spaces that cover not only the well-known Besov and Triebel-Lizorkin spaces $\dot B^{s}_{p,q}$ and $\dot F^{s}_{p,q}$ (when $\tau=0$) but also several other spaces of interest, such as Morrey spaces and $Q$ spaces. In this memoir, we introduce and study matrix-weighted versions $\dot B^{s,\tau}_{p,q}(W)$ and $\dot F^{s,\tau}_{p,q}(W)$ of these general function spaces on $\mathbb{R}^n$, where $W$ is a matrix-valued Muckenhoupt $A_p$ weight on $\mathbb R^n$. Our contributions include several characterizations of these spaces in terms of both the $\varphi$-transform of Frazier and Jawerth and the related sequence spaces $\dot b^{s,\tau}_{p,q}(W)$ and $\dot f^{s,\tau}_{p,q}(W)$, almost diagonal conditions that imply the boundedness of weakly defined operators on these spaces, and consequences for the boundedness of classical operators like pseudo-differential operators, trace operators, and Calderon-Zygmund operators. Results of this type are completely new on this level of generality, but many of them also improve the known results in the unweighted spaces $\dot B^{s,\tau}_{p,q}$ and $\dot F^{s,\tau}_{p,q}$ or, with $\tau=0$, in the weighted spaces $\dot B^{s}_{p,q}(W)$ and $\dot F^{s}_{p,q}(W)$. Several of our results are conveniently stated in terms of a new concept of the $A_p$-dimension $d\in[0,n)$ of a matrix weight $W\in A_p$ on $\mathbb R^n$ and, in several cases, the obtained estimates are shown to be sharp. In particular, for certain parameter ranges, we are able to characterize the sharp almost diagonal conditions that imply the boundedness of operators on these spaces. ",Matrix-Weighted Besov-Type and Triebel-Lizorkin-Type Spaces " I comment on Ulvi Yurtsever's result, which states that the entropy of a truncated bosonic Fock space is given by a holographic bound when the energy of the Fock states is constrained gravitationally. The derivation given in Yurtsever's paper contains an subtle mistake, which invalidates the result. A more restrictive, non-holographic entropy bound is derived. ",Entropy bound and local quantum field theory " We present multiwavelength (X-ray/optical/near-infrared/millimetre) observations of GRB 051022 between 2.5 hours and ~1.15 yr after the event. It is the most intense gamma-ray burst (~ 10^-4 erg cm^-2) detected by HETE-2, with the exception of the nearby GRB 030329. Optical and near infrared observations did not detect the afterglow despite a strong afterglow at X-ray wavelengths. Millimetre observations at Plateau de Bure (PdB) detected a source and a flare, confirming the association of this event with a moderately bright (R = 21.5) galaxy. Spectroscopic observations of this galaxy show strong [O II], Hbeta and [O III] emission lines at a redshift of 0.809. The spectral energy distribution of the galaxy implies Av (rest frame) = 1.0 and a starburst occuring ~ 25 Myr ago, during which the star-forming-rate reached >= 25 Msun/yr. In conjunction with the spatial extent (~ 1'') it suggests a very luminous (Mv = - 21.8) blue compact galaxy, for which we also find with Z Zsun. The X-ray spectrum shows evidence of considerable absorption by neutral gas with NH, X-ray = 3.47(+0.48/-0.47) x 10^22 cm^-2 (rest frame). Absorption by dust in the host galaxy at z = 0.809 certainly cannot account for the non-detection of the optical afterglow, unless the dust-to-gas ratio is quite different than that seen in our Galaxy (i.e. large dust grains). It is likely that the afterglow of the dark GRB 051022 was extinguished along the line of sight by an obscured, dense star forming region in a molecular cloud within the parent host galaxy. This galaxy is different from most GRB hosts being brighter than L* by a factor of 3. We have also derived a SFR ~ 50 Msun/yr and predict that this host galaxy will be detected at sub-mm wavelengths. ",The dark nature of GRB 051022 and its host galaxy " Over the last few years, we have witnessed the availability of an increasing data generated from non-Euclidean domains, which are usually represented as graphs with complex relationships, and Graph Neural Networks (GNN) have gained a high interest because of their potential in processing graph-structured data. In particular, there is a strong interest in exploring the possibilities in performing convolution on graphs using an extension of the GNN architecture, generally referred to as Graph Convolutional Neural Networks (ConvGNN). Convolution on graphs has been achieved mainly in two forms: spectral and spatial convolutions. Due to the higher flexibility in exploring and exploiting the graph structure of data, there is recently an increasing interest in investigating the possibilities that the spatial approach can offer. The idea of finding a way to adapt the network behaviour to the inputs they process to maximize the total performances has aroused much interest in the neural networks literature over the years. This paper presents a novel method to adapt the behaviour of a ConvGNN to the input proposing a method to perform spatial convolution on graphs using input-specific filters, which are dynamically generated from nodes feature vectors. The experimental assessment confirms the capabilities of the proposed approach, which achieves satisfying results using a low number of filters. ",Adaptive Filters in Graph Convolutional Neural Networks " Terahertz time-domain spectroscopy (THz-TDS) is a non-invasive, non-contact and label-free technique for biological and chemical sensing as THz-spectra is less energetic and lies in the characteristic vibration frequency regime of proteins and DNA molecules. However, THz-TDS is less sensitive for detection of micro-organisms of size equal to or less than $ \lambda/100 $ (where, $ \lambda $ is wavelength of incident THz wave) and, molecules in extremely low concentrated solutions (like, a few femtomolar). After successful high-throughput fabrication of nanostructures, nanoantennas and metamaterials were found to be indispensable in enhancing the sensitivity of conventional THz-TDS. These nanostructures lead to strong THz field enhancement which when in resonance with absorption spectrum of absorptive molecules, causing significant changes in the magnitude of the transmission spectrum, therefore, enhancing the sensitivity and allowing detection of molecules and biomaterials in extremely low concentrated solutions. Hereby, we review the recent developments in ultra-sensitive and selective nanogap biosensors. We have also provided an in-depth review of various high-throughput nanofabrication techniques. We also discussed the physics behind the field enhancements in sub-skin depth as well as sub-nanometer sized nanogaps. We introduce finite-difference time-domain (FDTD) and molecular dynamics (MD) simulations tools to study THz biomolecular interactions. Finally, we provide a comprehensive account of nanoantenna enhanced sensing of viruses (like, H1N1) and biomolecules such as artificial sweeteners which are addictive and carcinogenic. ",Nanoantenna Enhanced Terahertz Interaction of Biomolecules " We investigate the nature of the double color-magnitude sequence observed in the Gaia DR2 HR diagram of stars with high transverse velocities. The stars in the reddest-color sequence are likely dominated by the dynamically-hot tail of the thick disk population. Information from Nissen & Schuster (2010) and from the APOGEE survey suggests that stars in the blue-color sequence have elemental abundance patterns that can be explained by this population having a relatively low star-formation efficiency during its formation. In dynamical and orbital spaces, such as the `Toomre diagram', the two sequences show a significant overlap, but with a tendency for stars on the blue-color sequence to dominate regions with no or retrograde rotation and high total orbital energy. In the plane defined by the maximal vertical excursion of the orbits versus their apocenters, stars of both sequences redistribute into discrete wedges. We conclude that stars which are typically assigned to the halo in the solar vicinity are actually both accreted stars lying along the blue sequence in the HR diagram, and the low velocity tail of the old Galactic disk, possibly dynamically heated by past accretion events. Our results imply that a halo population formed in situ and responsible for the early chemical enrichment prior to the formation of the thick disk is yet to be robustly identified, and that what has been defined as the stars of the in situ stellar halo of the Galaxy may be in fact fossil records of its last significant merger. ",In disguise or out of reach: first clues about in-situ and accreted stars in the stellar halo of the Milky Way from Gaia DR2 " Two major causes of death in the United States and worldwide are stroke and myocardial infarction. The underlying cause of both is thrombi released from ruptured or eroded unstable atherosclerotic plaques that occlude vessels in the heart (myocardial infarction) or the brain (stroke). Clinical studies show that plaque composition plays a more important role than lesion size in plaque rupture or erosion events. To determine the plaque composition, various cell types in 3D cardiovascular immunofluorescent images of plaque lesions are counted. However, counting these cells manually is expensive, time-consuming, and prone to human error. These challenges of manual counting motivate the need for an automated approach to localize and count the cells in images. The purpose of this study is to develop an automatic approach to accurately detect and count cells in 3D immunofluorescent images with minimal annotation effort. In this study, we used a weakly supervised learning approach to train the HoVer-Net segmentation model using point annotations to detect nuclei in fluorescent images. The advantage of using point annotations is that they require less effort as opposed to pixel-wise annotation. To train the HoVer-Net model using point annotations, we adopted a popularly used cluster labeling approach to transform point annotations into accurate binary masks of cell nuclei. Traditionally, these approaches have generated binary masks from point annotations, leaving a region around the object unlabeled (which is typically ignored during model training). However, these areas may contain important information that helps determine the boundary between cells. Therefore, we used the entropy minimization loss function in these areas to encourage the model to output more confident predictions on the unlabeled areas. Our comparison studies indicate that the HoVer-Net model trained using our weakly ... ",Weakly Supervised Deep Instance Nuclei Detection using Points Annotation in 3D Cardiovascular Immunofluorescent Images " Single visual object tracking from an unmanned aerial vehicle (UAV) poses fundamental challenges such as object occlusion, small-scale objects, background clutter, and abrupt camera motion. To tackle these difficulties, we propose to integrate the 3D structure of the observed scene into a detection-by-tracking algorithm. We introduce a pipeline that combines a model-free visual object tracker, a sparse 3D reconstruction, and a state estimator. The 3D reconstruction of the scene is computed with an image-based Structure-from-Motion (SfM) component that enables us to leverage a state estimator in the corresponding 3D scene during tracking. By representing the position of the target in 3D space rather than in image space, we stabilize the tracking during ego-motion and improve the handling of occlusions, background clutter, and small-scale objects. We evaluated our approach on prototypical image sequences, captured from a UAV with low-altitude oblique views. For this purpose, we adapted an existing dataset for visual object tracking and reconstructed the observed scene in 3D. The experimental results demonstrate that the proposed approach outperforms methods using plain visual cues as well as approaches leveraging image-space-based state estimations. We believe that our approach can be beneficial for traffic monitoring, video surveillance, and navigation. ",Integration of the 3D Environment for UAV Onboard Visual Object Tracking Let $A$ be an abelian variety over an algebraically closed field. We show that $A$ is the automorphism group scheme of some smooth projective variety if and only if $A$ has only finitely many automorphisms as an algebraic group. This generalizes a result of Lombardo and Maffei for complex abelian varieties. ,Abelian varieties as automorphism groups of smooth projective varieties in arbitrary characteristics " In the development of modern cockpits, there is a trend towards the use of large displays that combine information about air navigation and the status of aircraft equipment. Flight and equipment performance information generated by multiple flight control systems should be graphically displayed in an easy-to-read form on widescreen multifunction displays. It is usually generated by independent systems whose output must not interfere with each other in accordance with the requirements of the ARINC 653 standard. This paper presents a solution to the problem of displaying ARINC 653 applications, which further improves security and portability, when running multiple applications on a single screen of one physical device. ",Cross-platform graphics subsystem for an ARINC 653-compatible real-time operating system English Translation of Paul Drude's 1902 investigation into the factors affecting the self-resonance of single-layer solenoid coils. The ratio of the self-resonant half-wavelength to the conductor length is is found to be principally dependent on the height-to diameter ratio of the coil and the dielectric constants of any insulating materials involved in the construction. ,On the construction of Tesla transformers. Period of oscillation and self-inductance of the coil. (Zur construction von Teslatransformatoren. Schwingungsdauer und Selbstinduction von Drahtspulen) " We revisit supersymmetric solutions to five dimensional ungauged N=1 supergravity with dynamic hypermultiplets. In particular we focus on a truncation to the axion-dilaton contained in the universal hypermultiplet. The relevant solutions are fibrations over a four-dimensional Kahler base with a holomorphic axion-dilaton. We focus on solutions with additional symmetries and classify Killing vectors which preserve the additional structure imposed by supersymmetry; in particular we extend the existing classification of solutions with a space-like U(1) isometry to the case where the Killing vector is rotational. We elaborate on general geometrical aspects which we illustrate in some simple examples. We especially discuss solutions describing the backreaction of M2-branes, which for example play a role in the black hole deconstruction proposal for microstate geometries. ",Unlocking the Axion-Dilaton in 5D Supergravity " Hierarchical structure formation inevitably leads to the formation of supermassive binary black holes (BBHs) with a sub-parsec separation in galactic nuclei. However, to date there has been no unambiguous detection of such systems. In an effort to search for potential observational signatures of supermassive BBHs, we performed high-resolution smoothed particle hydrodynamics (SPH) simulations of two black holes in a binary of moderate eccentricity surrounded by a circumbinary disk. Building on our previous work, which has shown that gas can periodically transfer from the circumbinary disk to the black holes when the binary is on an eccentric orbit, the current set of simulations focuses on the formation of the individual accretion disks, their evolution and mutual interaction, and the predicted radiative signature. The variation in mass transfer with orbital phase from the circumbinary disk induces periodic variations in the light curve of the two accretion disks at ultraviolet wavelengths, but not in the optical or near-infrared. Searches for this signal offer a promising method to detect supermassive BBHs. ",A supermassive binary black hole with triple disks " Typically, locally repairable codes (LRCs) and regenerating codes have been studied independently of each other, and it has not been clear how the parameters of one relate to those of the other. In this paper, a novel connection between locally repairable codes and exact regenerating codes is established. Via this connection, locally repairable codes are interpreted as exact regenerating codes. Further, some of these codes are shown to perform better than time-sharing codes between minimum bandwidth regenerating and minimum storage regenerating codes. ",A Connection Between Locally Repairable Codes and Exact Regenerating Codes " This paper presents a new hypothesis on a macro law in the universe, the law of increasing complexity, to formulate the assumption that the universe we observe and the biosphere on Earth are getting more diverse and complex with time. This formulation utilizes a quantitative definition of the complexity of organized matters, organized complexity (OC) [6]. We then apply this law to the coincidence (or fine-tuning) problem about the fundamental physical constants. We introduce a new principle, the principle of increasing complexity, on the law of increasing complexity and explain the coincidence with this new principle without using the anthropic principle. The principle implies that an (approximate) reduction of this macro law to fundamental physical laws would lead to a concrete analysis of the coincidence problem of fundamental physical constants. ",On the Arrow of Time and Organized Complexity in the Universe " Path sampling approaches have become invaluable tools to explore the mechanisms and dynamics of so-called rare events that are characterized by transitions between metastable states separated by sizeable free energy barriers. Their practical application, in particular to ever more complex molecular systems, is, however, not entirely trivial. Focusing on replica exchange transition interface sampling (RETIS) and forward flux sampling (FFS), we discuss a range of analysis tools that can be used to assess the quality and convergence of such simulations which is crucial to obtain reliable results. The basic ideas of a step-wise evaluation are exemplified for the study of nucleation in several systems with different complexity, providing a general guide for the critical assessment of RETIS and FFS simulations. ",Practical guide to replica exchange transition interface sampling and forward flux sampling " A ""book with k pages"" consists of a straight line (the ""spine"") and k half-planes (the ""pages""), such that the boundary of each page is the spine. If a graph is drawn on a book with k pages in such a way that the vertices lie on the spine, and each edge is contained in a page, the result is a k-page book drawing (or simply a k-page drawing). The k-page crossing number nu_k(G) of a graph G is the minimum number of crossings in a k-page drawing of G. In this paper we investigate the k-page crossing numbers of complete graphs K_n. We use semidefinite programming techniques to give improved lower bounds on nu_k(K_n) for various values of k. We also use a maximum satisfiability reformulation to calculate the exact value of nu_k(K_n) for several values of k and n. Finally, we investigate the best construction known for drawing K_n in k pages, calculate the resulting number of crossings, and discuss this upper bound in the light of the new results reported in this paper. ",Improved lower bounds on book crossing numbers of complete graphs " The theory of (tight) wavelet frames has been extensively studied in the past twenty years and they are currently widely used for image restoration and other image processing and analysis problems. The success of wavelet frame based models, including balanced approach and analysis based approach, is due to their capability of sparsely approximating piecewise smooth functions like images. Motivated by the balanced approach and analysis based approach, we shall propose a wavelet frame based $\ell_0$ minimization model, where the $\ell_0$ ""norm"" of the frame coefficients is penalized. We adapt the penalty decomposition (PD) method to solve the proposed optimization problem. Numerical results showed that the proposed model solved by the PD method can generate images with better quality than those obtained by either analysis based approach or balanced approach in terms of restoring sharp features as well as maintaining smoothness of the recovered images. Some convergence analysis of the PD method will also be provided. ",$\ell_0$ Minimization for Wavelet Frame Based Image Restoration " Under a nonlinear regression model with univariate response an algorithm for the generation of sequential adaptive designs is studied. At each stage, the current design is augmented by adding $p$ design points where $p$ is the dimension of the parameter of the model. The augmenting $p$ points are such that, at the current parameter estimate, they constitute the locally D-optimal design within the set of all saturated designs. Two relevant subclasses of nonlinear regression models are focused on, which were considered in previous work of the authors on the adaptive Wynn algorithm: firstly, regression models satisfying the `saturated identifiability condition' and, secondly, generalized linear models. Adaptive least squares estimators and adaptive maximum likelihood estimators in the algorithm are shown to be strongly consistent and asymptotically normal, under appropriate assumptions. For both model classes, if a condition of `saturated D-optimality' is satisfied, the almost sure asymptotic D-optimality of the generated design sequence is implied by the strong consistency of the adaptive estimators employed by the algorithm. The condition states that there is a saturated design which is locally D-optimal at the true parameter point (in the class of all designs). ",A $p$-step-ahead sequential adaptive algorithm for D-optimal nonlinear regression design " We extend the symbol calculus and study the limit operator theory for $\sigma$-compact, \'{e}tale and amenable groupoids, in the Hilbert space case. This approach not only unifies various existing results which include the cases of exact groups and discrete metric spaces with Property A, but also establish new limit operator theories for group/groupoid actions and uniform Roe algebras of groupoids. In the process, we extend a monumental result by Exel, Nistor and Prudhon, showing that the invertibility of an element in the groupoid $C^*$-algebra of a $\sigma$-compact amenable groupoid with a Haar system is equivalent to the invertibility of its images under regular representations. ",Limit operator theory for groupoids " Nowadays, people start to use online reservation systems to plan their vacations since they have vast amount of choices available. Selecting when and where to go from this large-scale options is getting harder. In addition, sometimes consumers can miss the better options due to the wealth of information to be found on the online reservation systems. In this sense, personalized services such as recommender systems play a crucial role in decision making. Two traditional recommendation techniques are content-based and collaborative filtering. While both methods have their advantages, they also have certain disadvantages, some of which can be solved by combining both techniques to improve the quality of the recommendation. The resulting system is known as a hybrid recommender system. This paper presents a new hybrid hotel recommendation system that has been developed by combining content-based and collaborative filtering approaches that recommends customer the hotel they need and save them from time loss. ",Hotel Recommendation System Based on User Profiles and Collaborative Filtering " In this work, we introduce the notion of Context-Based Prediction Models. A Context-Based Prediction Model determines the probability of a user's action (such as a click or a conversion) solely by relying on user and contextual features, without considering any specific features of the item itself. We have identified numerous valuable applications for this modeling approach, including training an auxiliary context-based model to estimate click probability and incorporating its prediction as a feature in CTR prediction models. Our experiments indicate that this enhancement brings significant improvements in offline and online business metrics while having minimal impact on the cost of serving. Overall, our work offers a simple and scalable, yet powerful approach for enhancing the performance of large-scale commercial recommender systems, with broad implications for the field of personalized recommendations. ",Unleash the Power of Context: Enhancing Large-Scale Recommender Systems with Context-Based Prediction Models We briefly review the recent progress in the exclusive determination of |V(ub)| using QCD sum rules on the light cone. ,Obtaining |V(ub)| exclusively: a theoretical perspective " The period enforcer algorithm for self-suspending real-time tasks is a technique for suppressing the ""back-to-back"" scheduling penalty associated with deferred execution. Originally proposed in 1991, the algorithm has attracted renewed interest in recent years. This note revisits the algorithm in the light of recent developments in the analysis of self-suspending tasks, carefully re-examines and explains its underlying assumptions and limitations, and points out three observations that have not been made in the literature to date: (i) period enforcement is not strictly superior (compared to the base case without enforcement) as it can cause deadline misses in self-suspending task sets that are schedulable without enforcement; (ii) to match the assumptions underlying the analysis of the period enforcer, a schedulability analysis of self-suspending tasks subject to period enforcement requires a task set transformation for which no solution is known in the general case, and which is subject to exponential time complexity (with current techniques) in the limited case of a single self-suspending task; and (iii) the period enforcer algorithm is incompatible with all existing analyses of suspension-based locking protocols, and can in fact cause ever-increasing suspension times until a deadline is missed. ",A Note on the Period Enforcer Algorithm for Self-Suspending Tasks " Partially Observable Markov Decision Processes (POMDPs) are rich environments often used in machine learning. But the issue of information and causal structures in POMDPs has been relatively little studied. This paper presents the concepts of equivalent and counterfactually equivalent POMDPs, where agents cannot distinguish which environment they are in though any observations and actions. It shows that any POMDP is counterfactually equivalent, for any finite number of turns, to a deterministic POMDP with all uncertainty concentrated into the initial state. This allows a better understanding of POMDP uncertainty, information, and learning. ","Counterfactual equivalence for POMDPs, and underlying deterministic environments" " Machine learning has proved invaluable for a range of different tasks, yet it also proved vulnerable to evasion attacks, i.e., maliciously crafted perturbations of input data designed to force mispredictions. In this paper we propose a novel technique to verify the security of decision tree models against evasion attacks with respect to an expressive threat model, where the attacker can be represented by an arbitrary imperative program. Our approach exploits the interpretability property of decision trees to transform them into imperative programs, which are amenable for traditional program analysis techniques. By leveraging the abstract interpretation framework, we are able to soundly verify the security guarantees of decision tree models trained over publicly available datasets. Our experiments show that our technique is both precise and efficient, yielding only a minimal number of false positives and scaling up to cases which are intractable for a competitor approach. ",Certifying Decision Trees Against Evasion Attacks by Program Analysis " We study the long-distance asymptotic behavior of various correlation functions for the one-dimensional (1D) attractive Hubbard model in a partially polarized phase through the Bethe ansatz and conformal field theory approaches. We particularly find the oscillating behavior of these correlation functions with spatial power-law decay, of which the pair (spin) correlation function oscillates with a frequency $\Delta k_F$ ($2\Delta k_F$). Here $\Delta k_F=\pi(n_\uparrow-n_\downarrow)$ is the mismatch in the Fermi surfaces of spin-up and spin-down particles. Consequently, the pair correlation function in momentum space has peaks at the mismatch $k=\Delta k_F$, which has been observed in recent numerical work on this model. These singular peaks in momentum space together with the spatial oscillation suggest an analog of the Fulde-Ferrell-Larkin-Ovchinnikov (FFLO) state in the 1D Hubbard model. The parameter $\beta$ representing the lattice effect becomes prominent in critical exponents which determine the power-law decay of all correlation functions. We point out that the backscattering of unpaired fermions and bound pairs within their own Fermi points gives a microscopic origin of the FFLO pairing in 1D. ",Asymptotic correlation functions and FFLO signature for the one-dimensional attractive Hubbard model " Tur\'an, Mitrinovi\'c-Adamovi\'c and Wilker type inequalities are deduced for regular Coulomb wave functions. The proofs are based on a Mittag-Leffler expansion for the regular Coulomb wave function, which may be of independent interest. Moreover, some complete monotonicity results concerning the Coulomb zeta functions and some interlacing properties of the zeros of Coulomb wave functions are given. ",Tur\'an type inequalities for regular Coulomb wave functions " We discuss the static axially symmetric regular solutions, obtained recently in Einstein-Yang-Mills and Einstein-Yang-Mills-dilaton theory [1]. These asymptotically flat solutions are characterized by the winding number $n>1$ and the node number $k$ of the purely magnetic gauge field. The well-known spherically symmetric solutions have winding number $n=1$. The axially symmetric solutions satisfy the same relations between the metric and the dilaton field as their spherically symmetric counterparts. Exhibiting a strong peak along the $\rho$-axis, the energy density of the matter fields of the axially symmetric solutions has a torus-like shape. For fixed winding number $n$ with increasing node number $k$ the solutions form sequences. The sequences of magnetically neutral non-abelian axially symmetric regular solutions with winding number $n$ tend to magnetically charged abelian spherically symmetric limiting solutions, corresponding to ``extremal'' Einstein-Maxwell-dilaton solutions for finite values of $\gamma$ and to extremal Reissner-Nordstr\o m solutions for $\gamma=0$, with $n$ units of magnetic charge. ",Static Axially Symmetric Einstein-Yang-Mills-Dilaton Solutions: I.Regular Solutions " We estimate a characteristic timescale for star formation in the spiral arms of disk galaxies, going from atomic hydrogen (HI) to dust-enshrouded massive stars. Drawing on high-resolution HI data from The HI Nearby Galaxy Survey and 24$\mu$m images from the Spitzer Infrared Nearby Galaxies Survey we measure the average angular offset between the HI and 24$\mu$m emissivity peaks as a function of radius, for a sample of 14 nearby disk galaxies. We model these offsets assuming an instantaneous kinematic pattern speed, $\Omega_p$, and a timescale, t(HI-->24$\mu$m), for the characteristic time span between the dense \hi phase and the formation of massive stars that heat the surrounding dust. Fitting for $\Omega_p$ and t(HI-->24$\mu$m), we find that the radial dependence of the observed angular offset (of the \hi and 24$\mu$m emission) is consistent with this simple prescription; the resulting corotation radii of the spiral patterns are typically $R_{cor}\simeq 2.7 R_{s}$, consistent with independent estimates. The resulting values of t(HI-->24$\mu$m) for the sample are in the range 1--4 Myr. We have explored the possible impact of non-circular gas motions on the estimate of t(HI-->24$\mu$m) and have found it to be substantially less than a factor of 2. This implies that a short timescale for the most intense phase of the ensuing star formation in spiral arms, and implies a considerable fraction of molecular clouds exist only for a few Myr before forming stars. However, our analysis does not preclude that some molecular clouds persist considerably longer. If much of the star formation in spiral arms occurs within this short interval t(HI-->24$\mu$m), then star formation must be inefficient, in order to avoid the short-term depletion of the gas reservoir. ",Geometrically Derived Timescales for Star Formation in Spiral Galaxies " A new $\mu$TCA DAQ system was introduced in CANDLES experiment with SpaceWire-to-GigabitEthernet (SpaceWire-GigabitEthernet) network for data readout and Flash Analog-to-Digital Converters (FADCs). With SpaceWire-GigabitEthernet, we can construct a flexible DAQ network with multi-path access to FADCs by using off-the-shelf computers. FADCs are equipped 8 event buffers, which act as de-randomizer to detect sequential decays from the background. SpaceWire-GigabitEthernet has high latency (about 100 $\mu$sec) due to long turnaround time, while GigabitEthernet has high throughput. To reduce dead-time, we developed the DAQ system with 4 ""crate-parallel"" (modules in crates are read in parallel) reading threads. As a result, the readout time is reduced by 4 times: 40 msec down to 10 msec. With improved performance, it is expected to achieve higher background suppression for CANDLES experiment. Moreover, for energy calibration, ""event-parallel"" reading process (events are read in parallel) is also introduced to reduce measurement time. With 2 ""event-parallel"" reading processes, the data rate is increased 2 times. ",$\mu$TCA DAQ system and parallel reading in CANDLES experiment Linear representations for a subclass of boolean symmetric functions selected by a parity condition are shown to constitute a generalization of the linear constraints on probabilities introduced by Boole. These linear constraints are necessary to compute probabilities of events with relations between the. arbitrarily specified with propositional calculus boolean formulas. ,Generalizing Fuzzy Logic Probabilistic Inferences " The relationship between the densities of ground-state wave functions (i.e., the minimizers of the Rayleigh--Ritz (RR) variation principle) and the ground-state densities in density-functional theory (i.e., the minimizers of the Hohenberg--Kohn (HK) variation principle) is studied within the framework of convex conjugation, in a generic setting covering molecular systems, solid-state systems, and more. Having introduced admissible density functionals as functionals that produce the exact ground ground-state energy for a given external potential by minimizing over densities in the HK variation principle, necessary sufficient conditions on such functionals are established to ensure that the RR ground-state densities and the HK ground-state densities are identical. We apply the results to molecular systems in the BO-approximation. For any given potential $v \in L^{3/2}(\mathbb{R}^3) + L^{\infty}(\mathbb{R}^3)$, we establish a one-to-one correspondence between the mixed ground-state densities of the RR variation principle and the mixed ground-state densities of the HK variation principle when the Lieb density-matrix constrained-search universal density functional is taken as the admissible functional. A similar one-to-one correspondence is established between the pure ground-state densities of the RR variation principle and the pure ground-state densities obtained using the HK variation principle with the Levy--Lieb pure-state constrained-search functional. In other words, all physical ground-state densities (pure or mixed) are recovered with these functionals and no false densities (i.e., minimizing densities that are not physical) exist. The importance of topology (i.e., choice of Banach space of densities and potentials) is emphasized and illustrated. The relevance of these results for current-density-functional theory is examined. ",Ground-state densities from the Rayleigh--Ritz variation principle and from density-functional theory " Each of neutrinos has a non - zero mass and regardless of whether it is a Dirac or a Majorana mass, can possess both anapole and electric dipole moments. Between their form factors appears a connection, for example, at the longitudinal neutrinos scattering on spinless nuclei. We discuss a theory, in which a mass consists of vector and axial - vector components responsible for separateness of leptonic current into the vector and axial - vector parts of the same charge or dipole moment. Such a model can explain the absence of truly neutral neutrinos vector interactions and the availability of an axial - vector structure of a Majorana mass. Thereby it relates the two neutrinos of a different nature. We derive an equation which unites the masses to a ratio of the anapole and electric dipole form factors of any lepton and its neutrino as a consequence of their unification in families of doublets and singlets. This testifies in favor of the existence of the left (right) dileptons and paradileptons of the axial - vector currents. Each of them answers to conservation of an axial - vector charge and any lepton flavor. Therefore, an axial - vector mass, anapole and electric dipole moment of the neutrino become proportional respectively to an axial - vector mass, anapole and electric dipole moment of a particle of the same families. ",Family Structure of Leptons and Their Currents of an Axial Vector Nature " We investigate the formation of methane line at 2.3 ${\rm \mu m}$ in Brown Dwarf Gliese 229B. Two sets of model parameters with (a) $T_{\rm eff}=940 $K and $\log (g) =5.0$, (b) $T_{\rm eff}=1030 $K and $ \log(g)=5.5$ are adopted both of which provide excellent fit for the synthetic continuum spectra with the observed flux at a wide range of wavelengths. In the absence of observational data for individual molecular lines, we set the additional parameters that are needed in order to model the individual lines by fitting the calculated flux with the observed flux at the continuum. A significant difference in the amount of flux at the core of the line is found with the two different models although the flux at the ontinuum remains the same. Hence, we show that if spectroscopic observation at $2.3{\rm \mu m}$ with a resolution as high as $R \simeq 200,000$ is possible then a much better constraint on the surface gravity and on the metallicity of the object could be obtained by fitting the theoretical model of individual molecular line with the observed data. ",Line Formation in the Atmosphere of Brown Dwarf Gliese 229B: $CH_4$ at 2.3 micron " The energy losses at collisions of heavy multiply charged ions with light atoms and polarization losses at moving through the matter have been considered under circumstances the ion charge Z>>1 and the relative colliding velocity v>>1, so that Z~v=1 for their emitted X-rays and gamma-rays, but perhaps even in ultraviolet through optical light at higher redshift. So far there are only inconclusive hints of this from z~4 active-galactic nucleii and gamma-ray bursts viewed with Fermi and Hubble Space Telescope. If correct though, that would impose a significant, fundamental resolution limit for galaxies out to z~8 in the era of the James Webb Space Telescope and the next generation of ground-based telescopes using adaptive optics. ",Limits to Seeing High-Redshift Galaxies Due to Planck-Scale-Induced Blurring " We apply a new method based upon thermofield dynamics (TFD) to study entanglement of finite-spin systems with non-competitive external fields for both equilibrium and non-equilibrium cases. For the equilibrium finite-spin systems, the temperature dependence of the extended density matrices is derived using this method, and the effect of non-competitive external field is demonstrated. For the non-equilibrium finite-spin systems, the time dependence of the extended density matrices and the extended entanglement entropies is derived in accordance with von Noumann equation, and the dissipative dynamics of the finite-spin systems is argued. Consequently, the applicability of the TFD-based method to describe entanglement is confirmed in both equilibrium and non-equilibrium cases with the external fieds. ",Dissipative dynamics of entangled finite-spin systems with non-competitive external fields We consider special solution to the 3D compressible Navier-Stokes system with and without the Coriolis force and dry friction and find the respective initial data implying a finite time gradient catastrophe. ,Exact solutions to the compressible Navier-Stokes equations with the Coriolis and friction terms " I study theoretically quadrupolar topological insulators under applied static electric field rotated along the crystal axis. I demonstrate, that the energy spectrum of this structure is a Wannier-Stark ladder that is quantized and directly distinguishes between the topological phase, possessing localized corner states, and the trivial phase, lacking the corner states. These results may find applications in the characterization of rapidly emerging higher-order topological phases of light and matter. ",Distinguishing trivial and topological quadrupolar insulators by Wannier-Stark ladders " Jets, jet-medium interaction and hydrodynamic evolution of fluctuations in initial parton density all lead to the final anisotropic dihadron azimuthal correlations in high-energy heavy-ion collisions. We remove the harmonic flow background and study the net correlations from different sources with different initial conditions within the AMPT model. We also study $\gamma$-hadron correlations which are only influenced by jet-medium interactions. ",Initial fluctuations and dihadron and $\gamma$-hadron correlations in high-energy heavy ion collisions " With the tremendous amount of computing because of the wide usage of internet it is observed that some user(s) are not able to manage their desktop with antivirus software properly installed. It is happening few times, that we allow our friends, students and colleagues to sit on our networked PC. Sometimes the user is unaware of the situation that there workstations are unsecured and so some one else could also be monitoring your flow of information and your most important data could go haywire, resulting into leakage of most confidential data to unwanted or malicious user(s). Example of some such documents could be question papers designed by the faculty member by various universities. Now a day most of the universities are having the biggest threat about the question papers and many other confidential documents designed by their faculty members. We in this paper present the solution to over come such a situation using the concept of Steganography. Steganography is a technique through which one can hide information into some cover object. This technique, if used, in positive direction could be of great help to solve such a problem and even other. ",An approach to secure highly confidential documents of any size in the corporate or institutes having unsecured networks In this paper we examine predictions from different models of nondiagonal parton distributions. This will be achieved by examining whether certain predictions of relationships between diagonal and nondiagonal parton distributions also hold after having evolved the different distributions. ,Study of Nondiagonal Parton Distribution Models " In our previous work [Y. Angelopoulos, S. Aretakis, and D. Gajic, Late-time asymptotics for the wave equation on spherically symmetric stationary backgrounds, in Advances in Mathematics 323 (2018), 529-621] we showed that the coefficient in the precise leading-order late-time asymptotics for solutions to the wave equation with smooth, compactly supported initial data on Schwarzschild backgrounds is proportional to the time-inverted Newman-Penrose constant (TINP), that is the Newman-Penrose constant of the associated time integral. The time integral (and hence the TINP constant) is canonically defined in the domain of dependence of any Cauchy hypersurface along which the stationary Killing field is non-vanishing. As a result, an explicit expression of the late-time polynomial tails was obtained in terms of initial data on Cauchy hypersurfaces intersecting the future event horizon to the future of the bifurcation sphere. In this paper, we extend the above result to Cauchy hypersurfaces intersecting the bifurcation sphere via a novel geometric interpretation of the TINP constant in terms of a modified gradient flux on Cauchy hypersurfaces. We show, without appealing to the time integral construction, that a general conservation law holds for these gradient fluxes. This allows us to express the TINP constant in terms of initial data on Cauchy hypersurfaces for which the time integral construction breaks down. ",Asymptotics for scalar perturbations from a neighborhood of the bifurcation sphere " Many data-fitting applications require the solution of an optimization problem involving a sum of large number of functions of high dimensional parameter. Here, we consider the problem of minimizing a sum of $n$ functions over a convex constraint set $\mathcal{X} \subseteq \mathbb{R}^{p}$ where both $n$ and $p$ are large. In such problems, sub-sampling as a way to reduce $n$ can offer great amount of computational efficiency. Within the context of second order methods, we first give quantitative local convergence results for variants of Newton's method where the Hessian is uniformly sub-sampled. Using random matrix concentration inequalities, one can sub-sample in a way that the curvature information is preserved. Using such sub-sampling strategy, we establish locally Q-linear and Q-superlinear convergence rates. We also give additional convergence results for when the sub-sampled Hessian is regularized by modifying its spectrum or Levenberg-type regularization. Finally, in addition to Hessian sub-sampling, we consider sub-sampling the gradient as way to further reduce the computational complexity per iteration. We use approximate matrix multiplication results from randomized numerical linear algebra (RandNLA) to obtain the proper sampling strategy and we establish locally R-linear convergence rates. In such a setting, we also show that a very aggressive sample size increase results in a R-superlinearly convergent algorithm. While the sample size depends on the condition number of the problem, our convergence rates are problem-independent, i.e., they do not depend on the quantities related to the problem. Hence, our analysis here can be used to complement the results of our basic framework from the companion paper, [38], by exploring algorithmic trade-offs that are important in practice. ",Sub-Sampled Newton Methods II: Local Convergence Rates " The implementation of a proof-of-concept Lattice Quantum Chromodynamics kernel on the Cell processor is described in detail, illustrating issues encountered in the porting process. The resulting code performs up to 45GFlop/s per socket, indicating that the Cell processor is likely to be a good platform for future Lattice QCD calculations. ",Performance of a Lattice Quantum Chromodynamics Kernel on the Cell Processor " Rocky planets and moons experiencing solar wind sputtering are continuously supplying their enveloping exosphere with ejected neutral atoms. To understand the quantity and properties of the ejecta, well established Binary Collision Approximation Monte Carlo codes like TRIM with default settings are used predominantly. Improved models such as SDTrimSP have come forward and together with new experimental data the underlying assumptions have been challenged. We introduce a hybrid model, combining the previous surface binding approach with a new bulk binding model akin to Hofs\""ass & Stegmaier (2023). In addition, we expand the model implementation by distinguishing between free and bound components sourced from mineral compounds such as oxides or sulfides. The use of oxides and sulfides also enables the correct setting of the mass densities of minerals, which was previously limited to the manual setting of individual atomic densities of elements. All of the energies and densities used are thereby based on tabulated data, so that only minimal user input and no fitting of parameters are required. We found unprecedented agreement between the newly implemented hybrid model and previously published sputter yields for incidence angles up to 45{\deg} from surface normal. Good agreement is found for the angular distribution of mass sputtered from enstatite MgSiO$_3$ compared to latest experimental data. Energy distributions recreate trends of experimental data of oxidized metals. Similar trends are to be expected from future mineral experimental data. The model thus serves its purpose of widespread applicability and ease of use for modelers of rocky body exospheres. ",New compound and hybrid binding energy sputter model for modeling purposes in agreement with experimental data " Quantum data is susceptible to decoherence induced by the environment and to errors in the hardware processing it. A future fault-tolerant quantum computer will use quantum error correction (QEC) to actively protect against both. In the smallest QEC codes, the information in one logical qubit is encoded in a two-dimensional subspace of a larger Hilbert space of multiple physical qubits. For each code, a set of non-demolition multi-qubit measurements, termed stabilizers, can discretize and signal physical qubit errors without collapsing the encoded information. Experimental demonstrations of QEC to date, using nuclear magnetic resonance, trapped ions, photons, superconducting qubits, and NV centers in diamond, have circumvented stabilizers at the cost of decoding at the end of a QEC cycle. This decoding leaves the quantum information vulnerable to physical qubit errors until re-encoding, violating a basic requirement for fault tolerance. Using a five-qubit superconducting processor, we realize the two parity measurements comprising the stabilizers of the three-qubit repetition code protecting one logical qubit from physical bit-flip errors. We construct these stabilizers as parallelized indirect measurements using ancillary qubits, and evidence their non-demolition character by generating three-qubit entanglement from superposition states. We demonstrate stabilizer-based quantum error detection (QED) by subjecting a logical qubit to coherent and incoherent bit-flip errors on its constituent physical qubits. While increased physical qubit coherence times and shorter QED blocks are required to actively safeguard quantum information, this demonstration is a critical step toward larger codes based on multiple parity measurements. ",Detecting bit-flip errors in a logical qubit using stabilizer measurements " By an exotic algebraic structure on the affine space ${\bf C}^n$ we mean a smooth affine algebraic variety which is diffeomorphic to ${\bf R}^{2n}$ but not isomorphic to ${\bf C}^n$. This is a survey of the recent developement on the subject, which emphasizes its analytic aspects and points out some open problems. ",On exotic algebraic structures on affine spaces " Some of the advances made in the literature to understand the phase transitions of quark matter in the presence of strong magnetic field and finite temperature (zero quark chemical potential) are reviewed. We start by discussing the physics behind the Magnetic catalysis (MC) at zero/finite temperature and then focus on the lattice predictions for inverse magnetic catalysis (IMC) at high temperature and strong magnetic fields. Possible explanations for the IMC are covered as well. Finally, we discuss recent efforts to modify QCD (quantum chromodynamics) effective models in order to reproduce the IMC observed on the lattice simulations. We emphasize the fact that applying thermo-magnetic effects on the coupling constant of the NJL model significantly improve the effectiveness of the NJL model to obtain a reasonable physical description of hot and magnetized quark matter being in agreement with lattice results. ",Inverse magnetic catalysis -- how much do we know about? " Counterion adsorption on a flexible polyelectrolyte chain in a spherical cavity is considered by taking a ""permuted"" charge distribution on the chain so that the ""adsorbed"" counterions are allowed to move along the backbone. We compute the degree of ionization by using self-consistent field theory (SCFT) and compare with the previously developed variational theory. Analysis of various contributions to the free energy in both theories reveals that the equilibrium degree of ionization is attained mainly as an interplay of the adsorption energy of counterions on the backbone, the translational entropy of the small ions, and their correlated density fluctuations. Degree of ionization computed from SCFT is significantly lower than that from the variational formalism. The difference is entirely due to the density fluctuations of the small ions in the system, which are accounted for in the variational procedure. When these fluctuations are deliberately suppressed in the truncated variational procedure, there emerges a remarkable quantitative agreement in the various contributing factors to the equilibrium degree of ionization, in spite of the fundamental differences in the approximations and computational procedures used in these two schemes. Nevertheless, since the significant effects from density fluctuations of small ions are not captured by the SCFT, and due to the close agreement between SCFT and the other contributing factors in the more transparent variational procedure, the latter is a better computational tool for obtaining the degree of ionization. ",Counterion adsorption on flexible polyelectrolytes: comparison of theories In this work we investigate how a circumstellar disk affects the radiation emitted by an embedded star. We show correlations obtained from broad-band observations of bipolar nebulae indicating that an orientation effect is at play in these systems. The FIR radiation relative to total radiation increases with inclination while the NIR and BVR fractions decrease. This is an expected effect if we consider the system as being made up of a dense dusty disk being irradiated by a hot star. We calculate 2-D models to try and reproduce the observed behavior with different disk and star configurations. ,Orientation effects in bipolar nebulae: Can disks do it? " Multi-target tracking is mainly challenged by the nonlinearity present in the measurement equation, and the difficulty in fast and accurate data association. To overcome these challenges, the present paper introduces a grid-based model in which the state captures target signal strengths on a known spatial grid (TSSG). This model leads to \emph{linear} state and measurement equations, which bypass data association and can afford state estimation via sparsity-aware Kalman filtering (KF). Leveraging the grid-induced sparsity of the novel model, two types of sparsity-cognizant TSSG-KF trackers are developed: one effects sparsity through $\ell_1$-norm regularization, and the other invokes sparsity as an extra measurement. Iterative extended KF and Gauss-Newton algorithms are developed for reduced-complexity tracking, along with accurate error covariance updates for assessing performance of the resultant sparsity-aware state estimators. Based on TSSG state estimates, more informative target position and track estimates can be obtained in a follow-up step, ensuring that track association and position estimation errors do not propagate back into TSSG state estimates. The novel TSSG trackers do not require knowing the number of targets or their signal strengths, and exhibit considerably lower complexity than the benchmark hidden Markov model filter, especially for a large number of targets. Numerical simulations demonstrate that sparsity-cognizant trackers enjoy improved root mean-square error performance at reduced complexity when compared to their sparsity-agnostic counterparts. ",Tracking Target Signal Strengths on a Grid using Sparsity " Weighted Timed Games (WTG for short) are the most widely used model to describe controller synthesis problems involving real-time issues. Unfortunately, they are notoriously difficult, and undecidable in general. As a consequence, one-clock WTGs have attracted a lot of attention, especially because they are known to be decidable when only non-negative weights are allowed. However, when arbitrary weights are considered, despite several recent works, their decidability status was still unknown. In this paper, we solve this problem positively and show that the value function can be computed in exponential time (if weights are encoded in unary). ",Decidability of One-Clock Weighted Timed Games with Arbitrary Weights " Spectroscopic and photometric observations of the peculiar object AM 2049-691 are presented here. Its systemic velocity is V(GSR) = (10956 +-30) km/s, and the derived distance (H(0) = 75 km/s/Mpc) results 146 Mpc. A bridge is observed between two very distinct nuclei whose separation is about 10 kpc, as well as two tails that emerge from the extremes SW and NE of the main body and extend up to 41 and 58 kpc respectively. The spectral characteristics of the all observed zones are typical of H II regions of low excitation. The internal reddening is quit high, particularly in the NE nucleus. All the derived equivalent widths of the H(alpha)+[N II] lines indicate enhanced star formation compared with isolated galaxies, specially in the NE nucleus; the equivalent width corresponding to the integrated spectrum reflects starburst activity in the whole object, and is compatible with a merger of two disk galaxies. All the observed characteristics of AM 2049-691 indicate it is a merger, where a overabundance of nitrogen is detected in one of the nuclei, which has the most evolved population and would be the most massive one. The detected total IR emission is not very high. The integrated total color B - V corresponds to a Sc-Scd galaxy and its average integrated population is about F7 type. Indicative B - V colors of the nuclei, corrected for internal absorption, are in agreement with the spectroscopic results. The central radial velocity dispersions at the nuclei suggest that the most massive galaxy would be the progenitor of the SW component. The observed radial velocity curve shows the presence of two subsystems, each one associated with a different nucleus. ",The Merging System Am 2049-691 " Trained human pilots or operators still stand out through their efficient, robust, and versatile skills in guidance tasks such as driving agile vehicles in spatial environments or performing complex surgeries. This research studies how humans learn a task environment for agile behavior. The hypothesis is that sensory-motor primitives previously described as interaction patterns and proposed as units of behavior for organization and planning of behavior provide elements of memory structure needed to efficiently learn task environments. The paper presents a modeling and analysis framework using the interaction patterns to formulate learning as a graph learning process and apply the framework to investigate and evaluate human learning and decision-making while operating in unknown environments. This approach emphasizes the effects of agent-environment dynamics (e.g., a vehicle controlled by a human operator), which is not emphasized in existing environment learning studies. The framework is applied to study human data collected from simulated first-person guidance experiments in an obstacle field. Subjects were asked to perform multiple trials and find minimum-time routes between prespecified start and goal locations without priori knowledge of the environment. ",Human Learning of Unknown Environments in Agile Guidance Tasks " We report spatially and temporally resolved measurements of magnetic fields generated by petawatt laser-solid interactions with high spatial resolution, using optical polarimetry. The polarimetric measurements map the megagauss magnetic field profiles generated by the fast electron currents at the target rear. The magnetic fields at the rear of a 50 $\mu$m thick aluminum target exhibit distinct and unambiguous signatures of electron beam filamentation. These results are corroborated by hybrid simulations. ",Micron-Scale Mapping of Megagauss Magnetic Fields in Petawatt Laser-Solid Interactions " We discuss a simple extension of the Standard Model (SM) that provides an explicit realization of the dark-matter (DM) neutrino-portal paradigm. The dark sector is composed of a scalar $ \Phi $ and a Dirac fermion $ \Psi $, with the latter assumed to be lighter than the former. These particles interact with the SM through the exchange of a set of heavy Dirac fermion mediators that are neutral under all local SM symmetries, and also under the dark-sector symmetry that stabilizes the $ \Psi $ against decay. We show that this model can accommodate all experimental and observational constraints provided the DM mass is below $\sim 35\, \gev $ or is in a resonant region of the Higgs or $Z$ boson. We also show that if the dark scalar and dark fermion are almost degenerate in mass, heavier DM fermions are not excluded. We note that in this scenario DM annihilation in the cores of astrophysical objects and the galactic halo produces a monochromatic neutrino beam of energy $ \mfe $, which provides a clear signature for this paradigm. Other experimental signatures are also discussed. ",A realistic model for Dark Matter interactions in the neutrino portal paradigm " Reversible logic allows low power dissipating circuit design and founds its application in cryptography, digital signal processing, quantum and optical information processing. This paper presents a novel quantum cost efficient reversible BCD adder for nanotechnology based systems using PFAG gate. It has been demonstrated that the proposed design offers less hardware complexity and requires minimum number of garbage outputs than the existing counterparts. The remarkable property of the proposed designs is that its quantum realization is given in NMR technology. ",Quantum Cost Efficient Reversible BCD Adder for Nanotechnology Based Systems " A limitation of many clustering algorithms is the requirement to tune adjustable parameters for each application or even for each dataset. Some techniques require an \emph{a priori} estimate of the number of clusters while density-based techniques usually require a scale parameter. Other parametric methods, such as mixture modeling, make assumptions about the underlying cluster distributions. Here we introduce a non-parametric clustering method that does not involve tunable parameters and only assumes that clusters are unimodal, in the sense that they have a single point of maximal density when projected onto any line, and that clusters are separated from one another by a separating hyperplane of relatively lower density. The technique uses a non-parametric variant of Hartigan's dip statistic using isotonic regression as the kernel operation repeated at every iteration. We compare the method against k-means++, DBSCAN, and Gaussian mixture methods and show in simulations that it performs better than these standard methods in many situations. The algorithm is suited for low-dimensional datasets with a large number of observations, and was motivated by the problem of ""spike sorting"" in neural electrical recordings. Source code is freely available. ",Unimodal clustering using isotonic regression: ISO-SPLIT " We study how primordial non-Gaussianities affect the clustering of voids at large scales. We derive a formula of the bias of voids induced from the non-Gaussianities by making use of the functional integral method. In a similar way as of haloes, we find that primordial non-Gaussianities can generate scale-dependence in the bias of voids at large scales. In addition, we show that by observing the cross power spectrum of voids and haloes we could check the consistency relation between the non-linearity parameters f_NL and tau_NL. Large voids (high peak objects) would be good targets since the effects of non-Gaussianities are more prominent while the effects of ""void-in-cloud"" are less significant. ",Void bias from primordial non-Gaussianities " We report measurements of charmonia produced in two-photon collisions and decaying to four-meson final states, where the meson is either a charged pion or a charged kaon. The analysis is based on a 395fb^{-1} data sample accumulated with the Belle detector at the KEKB electron-positron collider. We observe signals for the three C-even charmonia eta_c(1S), chi_{c0}(1P) and chi_{c2}(1P) in the pi^+pi^-pi^+pi^-, K^+K^-pi^+pi^- and K^+K^-K^+K^- decay modes. No clear signals for eta_c(2S) production are found in these decay modes. We have also studied resonant structures in charmonium decays to two-body intermediate meson resonances. We report the products of the two-photon decay width and the branching fractions, Gamma_{gamma gamma}B, for each of the charmonium decay modes. ",Study of charmonia in four-meson final states produced in two-photon collisions " Persistent homology, an algebraic method for discerning structure in abstract data, relies on the construction of a sequence of nested topological spaces known as a filtration. Two-parameter persistent homology allows the analysis of data simultaneously filtered by two parameters, but requires a bifiltration -- a sequence of topological spaces simultaneously indexed by two parameters. To apply two-parameter persistence to digital images, we first must consider bifiltrations constructed from digital images, which have scarcely been studied. We introduce the value-offset bifiltration for grayscale digital image data. We present efficient algorithms for computing this bifiltration with respect to the taxicab distance and for approximating it with respect to the Euclidean distance. We analyze the runtime complexity of our algorithms, demonstrate the results on sample images, and contrast the bifiltrations obtained from real images with those obtained from random noise. ",Value-Offset Bifiltrations for Digital Images " Chalcopyrite compounds are extensively explored for their exotic topological phases and associated phenomena in a variety of experiments. Here, we discuss the electrical transport properties of a direct energy gap semiconductor CdGeAs$_{2}$. The observed transverse magnetoresistance (MR) is found to be around 136% at a temperature of 1.8 K and a magnetic field of 14 T, following the semiclassical exponent MR $\sim$ $B^{2.18}$. The MR analysis exhibits a violation of the Kohler rule, suggesting the involvement of multiple carriers in the system. Below 15 K, with decreasing magnetic field, the MR increases, leading to the well known quantum interference phenomenon weak localization (WL). The analysis of the magnetoconductivity data based on the Hikami-Larkin-Nagaoka (HLN) model unveils three dimensional nature of the WL and the weak spin-orbit coupling in CdGeAs$_{2}$. The phase coherence length follows the $L_{\phi}$ $\sim$ $T^{-0.66}$ power law, which exhibits the 3D nature of the observed WL feature. ",A Trivial Geometrical Phase of an Electron Wavefunction in a Direct Band Gap Semiconductor CdGeAs$_{2}$ " We review several aspects of Yang-Mills theory (YMT) in two dimensions, related to its perturbative and topological properties. Consistency between light-front and equal-time formulations is thoroughly discussed. ",Light-front vacuum and instantons in two dimensions " Undulator radiation from synchrotron light sources must be transported down a beamline from the source to the sample. A partially coherent photon beam may be represented in phase space using a Wigner function, and its transport may use some similar techniques as is familiar in particle beam transport. We describe this process in the case that the beamline is composed of linear focusing and defocusing sections as well as apertures. We present a compact representation of the beamline map involving linear transformations and convolutions. We create a 1:1 imaging system (4f system) with a single slit on the image plane and observe the radiation downstream to it. We propagate a Gaussian beam and undulator radiation down this sample beamline, drawing parameters from current and future ultra low emittance light sources. We derive an analytic expression for the partially coherent Gaussian case including passage through a single slit aperture. We benchmark the Wigner function calculation against the analytical expression and a partially coherent calculation in the Synchrotron Radiation Workshop (SRW) code. ",Propagation of partially coherent radiation using Wigner functions " The quantum cascade laser (QCL) has evolved to be a compact, powerful source of coherent mid-infrared (mid-IR) light. However, its fast gain dynamics strongly restricts the formation of ultrashort pulses. As such, the shortest pulses reported so far were limited to a few picoseconds with some hundreds of milliwatts of peak power, strongly narrowing their applicability for time-resolved and nonlinear experiments. Here, we demonstrate an alternative approach capable of producing near-transform-limited sub-picosecond pulses with several watts of peak power. Starting from a frequency modulated phase-locked state, which most efficiently exploits the gain of the active region, ultrashort high peak power pulses are generated via external pulse compression. We assess their temporal nature by means of a novel optical sampling method, coherent beat note interferometry and interferometric autocorrelation. These results open new pathways for nonlinear physics in the mid-infrared. ",Femtosecond pulses from a mid-infrared quantum cascade laser " We bound the excess risk of interpolating deep linear networks trained using gradient flow. In a setting previously used to establish risk bounds for the minimum $\ell_2$-norm interpolant, we show that randomly initialized deep linear networks can closely approximate or even match known bounds for the minimum $\ell_2$-norm interpolant. Our analysis also reveals that interpolating deep linear models have exactly the same conditional variance as the minimum $\ell_2$-norm solution. Since the noise affects the excess risk only through the conditional variance, this implies that depth does not improve the algorithm's ability to ""hide the noise"". Our simulations verify that aspects of our bounds reflect typical behavior for simple data distributions. We also find that similar phenomena are seen in simulations with ReLU networks, although the situation there is more nuanced. ",Deep Linear Networks can Benignly Overfit when Shallow Ones Do " Latest experimental and evaluated $\alpha$-decay half-lives between 82$\leq$Z$\leq$118 have been used to modify two empirical formulas: (i) Horoi scaling law [J. Phys. G \textbf{30}, 945 (2004)], and Sobiczewski formula [Acta Phys. Pol. B \textbf{36}, 3095 (2005)] by adding asymmetry dependent terms ($I$ and $I^2$) and refitting of the coefficients. The results of these modified formulas are found with significant improvement while compared with other 21 formulas, and, therefore, are used to predict $\alpha$-decay half-lives with more precision in the unknown superheavy region. The formula of spontaneous fission (SF) half-life proposed by Bao \textit{et al.} [J. Phys. G \textbf{42}, 085101 (2015)] is further modified by using ground-state shell-plus-pairing correction taken from FRDM-2012 and using latest experimental and evaluated spontaneous fission half-lives between 82$\leq$Z$\leq$118. Using these modified formulas, contest between $\alpha$-decay and SF is probed for the nuclei within the range 112$\leq$Z$\leq$118 and consequently probable half-lives and decay modes are estimated. Potential decay chains of $^{286-302}$Og and $^{287-303}$119 (168$\leq$N$\leq$184: island of stability) are analyzed which are found in excellent agreement with available experimental data. In addition, four different machine learning models: XGBoost, Random Forest (RF), Decision Trees (DTs), and Multilayer Perceptron (MLP) neural network are used to train a predictor for $\alpha$-decay and SF half-lives prediction. The prediction of decay modes using XGBoost and MLP are found in excellent agreement with available experimental decay modes along with our predictions obtained by above mentioned modified formulas. ",Modified empirical formulas and machine learning for $\alpha$-decay systematics " Formulas for longitudinal electric conductivity and dielectric permeability in the quantum non-degenerate collisional plasma with the frequency of collisions depending on momentum in Mermin' approach are received. The kinetic equation in momentum space in relaxation approximation is used. It is shown that when Planck's constant tends to zero, the deduced formula passes to the corresponding formula for classical plasma. It is shown also that when frequency of collisions of particles of plasma tends to zero (plasma passes to collisionless one), the deduced formula passes to the known Lindhard' formula received for collisionless plasmas. It is shown, that when frequency of collisions is a constant, the deduced formula for dielectric permeability passes in known Mermin' formula. ",Longitudinal electric conductivity and dielectric permeability in quantum plasma with variable frequency of collisions in Mermin' approach " We construct a duality between several simple physical systems by showing that they are different aspects of the same quantum theory. Examples include the free relativistic massless particle and the hydrogen atom in any number of dimensions. The key is the gauging of the Sp(2) duality symmetry that treats position and momentum (x,p) as a doublet in phase space. As a consequence of the gauging, the Minkowski space-time vectors (x^\mu, p^\mu) get enlarged by one additional space-like and one additional time-like dimensions to (x^M,p^M). A manifest global symmetry SO(d,2) rotates (x^M,p^M) like d+2 dimensional vectors. The SO(d,2) symmetry of the parent theory may be interpreted as the familiar conformal symmetry of quantum field theory in Minkowski spacetime in one gauge, or as the dynamical symmetry of a totally different physical system in another gauge. Thanks to the gauge symmetry, the theory permits various choices of ``time'' which correspond to different looking Hamiltonians, while avoiding ghosts. Thus we demonstrate that there is a physical role for a spacetime with two times when taken together with a gauged duality symmetry that produces appropriate constraints. ","Gauged Duality, Conformal Symmetry, and Spacetime with Two Times" " In this paper, we derive optimal transmission policies for energy harvesting sensors to maximize the utility obtained over a finite horizon. First, we consider a single energy harvesting sensor, with discrete energy arrival process, and a discrete energy consumption policy. Under this model, we show that the optimal finite horizon policy is a threshold policy, and explicitly characterize the thresholds, and the thresholds can be precomputed using a recursion. Next, we address the case of multiple sensors, with only one of them allowed to transmit at any given time to avoid interference, and derive an explicit optimal policy for this scenario as well. ",Finite-Horizon Optimal Transmission Policies for Energy Harvesting Sensors " We report on the first measurement of exclusive single-photon muoproduction on the proton by COMPASS using 160 GeV/$c$ polarized $\mu^+$ and $\mu^-$ beams of the CERN SPS impinging on a liquid hydrogen target. We determine the dependence of the average of the measured $\mu^+$ and $\mu^-$ cross sections for deeply virtual Compton scattering on the squared four-momentum transfer $t$ from the initial to the final final proton. The slope $B$ of the $t$-dependence is fitted with a single exponential function, which yields $B=(4.3 \ \pm \ 0.6_{\text{stat}}\_{- \ 0.3}^{+ \ 0.1}\big\rvert_{\text{sys}}) (\text{GeV}/c)^{-2}$. This result can be converted into an average transverse extension of partons in the proton, $\sqrt{\langle r_{\perp}^2 \rangle} = (0.58 \ \pm \ 0.04_{\text{stat}}\_{- \ 0.02}^{+ \ 0.01}\big\rvert_{\text{sys}})\text{fm}$. For this measurement, the average virtuality of the photon mediating the interaction is $\langle Q^2 \rangle = 1.8\,(\text{GeV/}c)^2$ and the average value of the Bjorken variable is $\langle x_{\text{Bj}} \rangle = 0.056$. ",Transverse Extension of Partons in the Proton probed by Deeply Virtual Compton Scattering " Thermal history of the universe after big-bang nucleosynthesis (BBN) is well understood both theoretically and observationally, and recent cosmological observations also begin to reveal the inflationary dynamics. However, the epoch between inflation and BBN is scarcely known. In this paper we show that the detection of the stochastic gravitational wave background around 1Hz provides useful information about thermal history well before BBN. In particular, the reheating temperature of the universe may be determined by future space-based laser interferometer experiments such as DECIGO and/or BBO if it is around 10^{6-9} GeV, depending on the tensor-to-scalar ratio $r$ and dilution factor $F$. ",Probing reheating temperature of the universe with gravitational wave background " We study the existence and uniqueness of Lp-bounded mild solutions for a class ofsemilinear stochastic evolutions equations driven by a real L\'evy processes withoutGaussian component not square integrable for instance the stable process through atruncation method by separating the big and small jumps together with the classicaland simple Banach fixed point theorem ; under local Lipschitz, Holder, linear growthconditions on the coefficients. ",On bounded mild solutions for a class of semilinear stochastic evolution equation driven by stable process " We report the first detection of C$^{15}$N in diffuse molecular gas from a detailed examination of CN absorption lines in archival VLT/UVES spectra of stars probing local diffuse clouds. Absorption from the C$^{15}$N isotopologue is confidently detected (at $\gtrsim4\sigma$) in three out of the four directions studied and appears as a very weak feature between the main $^{12}$CN and $^{13}$CN absorption components. Column densities for each CN isotopologue are determined through profile fitting, after accounting for weak additional line-of-sight components of $^{12}$CN, which are seen in the absorption profiles of CH and CH$^+$ as well. The weighted mean value of C$^{14}$N/C$^{15}$N for the three sight lines with detections of C$^{15}$N is $274\pm18$. Since the diffuse molecular clouds toward our target stars have relatively high gas kinetic temperatures and relatively low visual extinctions, their C$^{14}$N/C$^{15}$N ratios should not be affected by chemical fractionation. The mean C$^{14}$N/C$^{15}$N ratio that we obtain should therefore be representative of the ambient $^{14}$N/$^{15}$N ratio in the local interstellar medium. Indeed, our mean value agrees well with that derived from millimeter-wave observations of CN, HCN, and HNC in local molecular clouds. ",The C$^{14}$N/C$^{15}$N Ratio in Diffuse Molecular Clouds " Using numerical simulations, we investigate the gravitational evolution of filamentary molecular cloud structures and their condensation into dense protostellar cores. One possible process is the so called 'edge effect', the pile-up of matter at the end of the filament due to self-gravity. This effect is predicted by theory but only rarely observed. To get a better understanding of the underlying processes we used a simple analytic approach to describe the collapse and the corresponding collapse time. We identify a model of two distinct phases: The first phase is free fall dominated, due to the self-gravity of the filament. In the second phase, after the turning point, the collapse is balanced by the ram pressure, produced by the inside material of the filament, which leads to a constant collapse velocity. This approach reproduces the established collapse time of uniform density filaments and agrees well with our hydrodynamic simulations. In addition, we investigate the influence of different radial density profiles on the collapse. We find that the deviations compared to the uniform filament are less than 10%. Therefore, the analytic collapse model of the uniform density filament is an excellent general approach. ",Filament collapse: a two phase process " The Pomeron, which dominates high energy elastic and diffractive hadronic processes, must be largely gluonic in nature. We use a recent picture of a scalar glueball/sigma system with coupling of the sigma to glue determined from experiment to predict strong peripheral sigma production seen in the p p $\pi^o\pi^o$ final state. ",Peripheral Production of Sigmas in Proton Proton Collisions " We report the structural and magnetic properties of a new system LiRhMnO$_{4}$ (LRMO) through x-ray diffraction, bulk magnetization, heat capacity and $^{7}$Li nuclear magnetic resonance (NMR) measurements. LRMO crystallizes in the cubic space group $\mathit{Fd}$$\bar{3}$$\mathit{m}$. From the DC susceptibility data, we obtained the Curie-Weiss temperature $\mathrm{\theta}_{\mathrm{CW}}$ = -26 K and Curie constant $\mathit{C}$ = 1.79 Kcm$^{3}$/mol suggesting antiferromagnetic correlations among the magnetic Mn$^{4+}$ ions with an effective spin $\mathit{S}$ = $\frac{3}{2}$. At $\mathit{H}$ = 50 Oe, the field cooled and zero-field cooled magnetizations bifurcate at a freezing temperature, $T_{f}$ = 4.45 K, which yields the frustration parameter $\mathit{f=\frac{\mid\theta_{CW}\mid}{T_{f}}}>$5. AC susceptibility, shows a cusp-like peak at around $T_{f}$, with the peak position shifting as a function of the driving frequency, confirming a spin-glass-like transition in LRMO. LRMO also shows typical spin-glass characteristics such as memory effect, aging effect and relaxation. In the heat capacity, there is no sharp anomaly down to 2 K indicative of long-range ordering. The field sweep $^{7}$Li NMR spectra show broadening with decreasing temperature without any spectral line shift. The $^{7}$Li NMR spin-lattice and spin-spin relaxation rates also show anomalies due to spin freezing near $T_{f}$. ",Structural and magnetic properties of a new cubic spinel LiRhMnO$_{4}$ " A conventional approach to train neural ordinary differential equations (ODEs) is to fix an ODE solver and then learn the neural network's weights to optimize a target loss function. However, such an approach is tailored for a specific discretization method and its properties, which may not be optimal for the selected application and yield the overfitting to the given solver. In our paper, we investigate how the variability in solvers' space can improve neural ODEs performance. We consider a family of Runge-Kutta methods that are parameterized by no more than two scalar variables. Based on the solvers' properties, we propose an approach to decrease neural ODEs overfitting to the pre-defined solver, along with a criterion to evaluate such behaviour. Moreover, we show that the right choice of solver parameterization can significantly affect neural ODEs models in terms of robustness to adversarial attacks. Recently it was shown that neural ODEs demonstrate superiority over conventional CNNs in terms of robustness. Our work demonstrates that the model robustness can be further improved by optimizing solver choice for a given task. The source code to reproduce our experiments is available at https://github.com/juliagusak/neural-ode-metasolver. ",Meta-Solver for Neural Ordinary Differential Equations " Temporal data, representing chronological observations of complex systems, has always been a typical data structure that can be widely generated by many domains, such as industry, medicine and finance. Analyzing this type of data is extremely valuable for various applications. Thus, different temporal data analysis tasks, eg, classification, clustering and prediction, have been proposed in the past decades. Among them, causal discovery, learning the causal relations from temporal data, is considered an interesting yet critical task and has attracted much research attention. Existing causal discovery works can be divided into two highly correlated categories according to whether the temporal data is calibrated, ie, multivariate time series causal discovery, and event sequence causal discovery. However, most previous surveys are only focused on the time series causal discovery and ignore the second category. In this paper, we specify the correlation between the two categories and provide a systematical overview of existing solutions. Furthermore, we provide public datasets, evaluation metrics and new perspectives for temporal data causal discovery. ",Causal Discovery from Temporal Data: An Overview and New Perspectives " We investigate the surface width $W$ of solid-on-solid surfaces in the vicinity of the roughening temperature $T_r$. Above $T_r$, $W^2$ is expected to diverge with the system size $L$ like $\ln L$. However, close to $T_r$ a clean $\ln{L}$ behavior can only be seen on extremely large lattices. Starting from the Kosterlitz-Thouless renormalization group, we derive an improved formula that describes the small $L$ behavior on both sides of $T_r$. For the Discrete Gaussian model, we used the valleys-to-mountains-reflections cluster algorithm in order to simulate the fluctuating solid-on-solid surface. The base plane above which the surface is defined is an $L \times L$ square lattice. In the simulation we took $8\leq L\leq 256$. The improved formula fits the numerical results very well. {}From the analysis, we estimate the roughening temperature to be $T_r = 0.755(3)$. ",The Solid-on-Solid Surface Width Around the Roughening Transition " Recent data from heavy ion collisions at RHIC show unexpectedly large near-angle correlations that broaden longitudinally with increasing centrality. The amplitude of this ridge-like correlation rises rapidly, reaches a maximum, and then falls in the most central collisions. In this letter we explain how this behavior can be explained as initial-state coordinate-space anisotropies converted into final-state momentum-space correlations. We propose $v_n^2/\epsilon_{n,\mathrm{part}}^{2}$ as a useful way to study length scales and provide a prediction for the ridge in Pb+Pb collisions at $\sqrt{s_{\mathrm{NN}}}=$ 2.76 TeV. ",The Rise and Fall of the Ridge in Heavy Ion Collisions " Mean field calculations, based on the D1S, D1N and D1M parametrizations of the Gogny energy density functional, have been carried out to obtain the potential energy surfaces relevant to fission in several Ra isotopes with the neutron number 144 $\le$ N $\le$ 176. Inner and outer barrier heights as well as first and second isomer excitation energies are given. The existence of a well developed third minimum along the fission paths of Ra nuclei, is analyzed in terms of the energetics of the ""fragments"" defining such elongated configuration. The masses and charges of the fission fragments are studied as functions of the neutron number in the parent Ra isotope. The comparison between fission and $\alpha$-decay half-lives, reveals that the former becomes faster for increasing neutron numbers. Though there exists a strong variance of the results with respect to the parameters used in the computation of the spontaneous fission rate, a change in tendency is observed at N=164 with a steady increase that makes heavier neutron-rich Ra isotopes stable against fission, diminishing the importance of fission recycling in the r-process. ",Microscopic description of fission in neutron-rich Radium isotopes with the Gogny energy density functional " We study the scattering of the light-flavor pseudoscalar mesons ($\pi, K, \eta$) off the ground-state charmed mesons ($D,D_s$) within chiral effective field theory. The recent lattice simulation results on various scattering lengths and the finite-volume spectra both in the moving and center-of-mass frames, most of which are obtained at unphysical meson masses, are used to constrain the free parameters in our theory. Explicit formulas to include the $S$- and $P$-wave mixing to determine the finite-volume energy levels are provided. After a successful reproduction of the lattice data, we perform a chiral extrapolation to predict the quantities with physical meson masses, including phase shifts, inelasticities, resonance pole positions and the corresponding residues from the scattering of the light pseudoscalar and charmed mesons. ",Towards a precise determination of the scattering amplitudes of the charmed and light-flavor pseudoscalar mesons " Multishot Magnetic Resonance Imaging (MRI) has recently gained popularity as it accelerates the MRI data acquisition process without compromising the quality of final MR image. However, it suffers from motion artifacts caused by patient movements which may lead to misdiagnosis. Modern state-of-the-art motion correction techniques are able to counter small degree motion, however, their adoption is hindered by their time complexity. This paper proposes a Generative Adversarial Network (GAN) for reconstructing motion free high-fidelity images while reducing the image reconstruction time by an impressive two orders of magnitude. ",Automating Motion Correction in Multishot MRI Using Generative Adversarial Networks " We study the current flow paths between two edges in a random resistor network on a $L\times L$ square lattice. Each resistor has resistance $e^{ax}$, where $x$ is a uniformly-distributed random variable and $a$ controls the broadness of the distribution. We find (a) the scaled variable $u\equiv L/a^\nu$, where $\nu$ is the percolation connectedness exponent, fully determines the distribution of the current path length $\ell$ for all values of $u$. For $u\gg 1$, the behavior corresponds to the weak disorder limit and $\ell$ scales as $\ell\sim L$, while for $u\ll 1$, the behavior corresponds to the strong disorder limit with $\ell\sim L^{d_{\scriptsize opt}}$, where $d_{\scriptsize opt} = 1.22\pm0.01$ is the optimal path exponent. (b) In the weak disorder regime, there is a length scale $\xi\sim a^\nu$, below which strong disorder and critical percolation characterize the current path. ",Current Flow in Random Resistor Networks: The Role of Percolation in Weak and Strong Disorder " The probability distribution of percolation thresholds in finite lattices were first believed to follow a normal Gaussian behaviour. With increasing computer power and more efficient simulational techniques, this belief turned to a stretched exponential behaviour, instead. Here, based on a further improvement of Monte Carlo data, we show evidences that this question is not yet answered at all. ",Are the Tails of Percolation Thresholds Gaussians ? " We construct a class of quadratic irrationals having continued fractions of period $n\geq2$ with ""small"" partial quotients for which certain integer multiples have continued fractions of period $1$, $2$ or $4$ with ""large"" partial quotients. We then show that numbers in the period of the new continued fraction are simple functions of the numbers in the periods of the original continued fraction. We give generalizations of some of the continued fractions and show that polynomials arising from the generalizations are related to Chebyshev and Fibonacci polynomials. We then show that some of these polynomials have a hyperbolic root distribution. ",Multiples of long period small element continued fractions to short period large elements continued fractions " We give a probabilistic proof of relative Fatou's theorem for $(-\Delta)^{\alpha/2}$-harmonic functions (equivalently for symmetric $\alpha$-stable processes) in bounded $\kappa$-fat open set where $\alpha \in (0,2)$. That is, if $u$ is positive $(-\Delta)^{\alpha/2}$-harmonic function in a bounded $\kappa$-fat open set $D$ and $h$ is singular positive $(-\Delta)^{\alpha/2}$-harmonic function in $D$, then non-tangential limits of $u/h$ exist almost everywhere with respect to the Martin-representing measure of $h$. It is also shown that, under the gaugeability assumption, relative Fatou's theorem is true for operators obtained from the generator of the killed $\alpha$-stable process in bounded $\kappa$-fat open set $D$ through non-local Feynman-Kac transforms. As an application, relative Fatou's theorem for relativistic stable processes is also true if $D$ is bounded $C^{1,1}$-open set. ",Relative Fatou's Theorem for $(-\Delta)^{\alpha/2}$-harmonic Functions in Bounded $\kappa$-fat Open Set " Dynamic population gratings (DPGs) in rare-earth doped fibers are prevalent devices in fiber lasers for the production of single-longitudinal-mode emission, Q-switched pulses, and wavelength self-sweeping regimes. This study presents a transition from Q-switched state to continuous wave (CW) state, accompanying irregular mode-hopping, in an erbium-doped fiber laser with a heavily-doped DPG centered at 1549.95 nm. Our results demonstrate that the transition between these two states can be achieved by adjusting the pump power. The repetition frequency of the Q-switched pulse increases monotonically with the increasing pump power, while the pulse duration initially narrows and then expands because the reduced peak intensity weakens the nonlinear effect. Additionally, modulation peaks are evident on both the Q-switched pulse train and the CW background, which are induced by the irregular mode-hopping caused by the DPG. Furthermore, we observe that the central wavelength fluctuates within a range of 0.05 nm. These results provide valuable insight into the DPG effect in heavily-doped fibers. ",Observation of Q-switched and continuous wave regimes with mode-hopping in Er-doped fiber lasers incorporating a dynamic population grating " This paper is concerned with the regulation control of a one-dimensional reaction-diffusion equation in the presence of a state-delay in the reaction term. The objective is to achieve the PI regulation of the right Dirichlet trace with a command selected as the left Dirichlet trace. The control design strategy consists of the design of a PI controller on a finite dimensional truncated model obtained by spectral reduction. By an adequate selection of the number of modes of the original infinite-dimensional system, we show that the proposed control design procedure achieves both the exponential stabilization of the original infinite-dimensional system as well as the setpoint regulation of the right Dirichlet trace. ",Integral action for setpoint regulation control of a reaction-diffusion equation in the presence of a state delay " The Berber, or Amazigh language family is a low-resource North African vernacular language spoken by the indigenous Berber ethnic group. It has its own unique alphabet called Tifinagh used across Berber communities in Morocco, Algeria, and others. The Afroasiatic language Berber is spoken by 14 million people, yet lacks adequate representation in education, research, web applications etc. For instance, there is no option of translation to or from Amazigh / Berber on Google Translate, which hosts over 100 languages today. Consequently, we do not find specialized educational apps, L2 (2nd language learner) acquisition, automated language translation, and remote-access facilities enabled in Berber. Motivated by this background, we propose a supervised approach called DaToBS for Detection and Transcription of Berber Signs. The DaToBS approach entails the automatic recognition and transcription of Tifinagh characters from signs in photographs of natural environments. This is achieved by self-creating a corpus of 1862 pre-processed character images; curating the corpus with human-guided annotation; and feeding it into an OCR model via the deployment of CNN for deep learning based on computer vision models. We deploy computer vision modeling (rather than language models) because there are pictorial symbols in this alphabet, this deployment being a novel aspect of our work. The DaToBS experimentation and analyses yield over 92 percent accuracy in our research. To the best of our knowledge, ours is among the first few works in the automated transcription of Berber signs from roadside images with deep learning, yielding high accuracy. This can pave the way for developing pedagogical applications in the Berber language, thereby addressing an important goal of outreach to underrepresented communities via AI in education. ",Optical Character Recognition and Transcription of Berber Signs from Images in a Low-Resource Language Amazigh " We completely characterize the Crawford number attainment set and the numerical radius attainment set of a bounded linear operator on a Hilbert space. We study the intersection properties of the corresponding attainment sets of numerical radius, Crawford number, norm, minimum norm of a bounded linear operator defined on a normed space. Our study illustrates the similarities and the differences of the extremal properties of a bounded linear operator on a Hilbert space and a general normed space. ",On numerical radius and Crawford number attainment sets of a bounded linear operator It is shown that any generalized Kac-Moody Lie algebra g that has no mutually orthogonal imaginary simple roots can be written as the vector space direct sum of a Kac-Moody subalgebra and subalgebras isomorphic to free Lie algebras over certain modules for the Kac-Moody subalgebra. Also included is a detailed discussion of Borcherds' construction of the Monster Lie algebra from a vertex algebra and an elementary proof of Borcherds' theorem relating Lie algebras with `an almost positive definite bilinear form' to generalized Kac-Moody algebras. (Preprint version 1996) ,"Generalized Kac-Moody Lie algebras, free Lie algebras and the structure of the Monster Lie algebra" " In this paper, we study the construction of the supersymmetric extensions of vertex algebras. In particular, for $N = n \in \mathbb{Z}_{+}$, we show the universal enveloping $N = n$ supersymmetric (SUSY) vertex algebra of an $N = n$ SUSY Lie conformal algebra can be extended to an $N = n' > n$ SUSY vertex algebra. ",Supersymmetric extension of universal enveloping vertex algebras " We construct the first explicit (i.e., non-random) examples of Salem sets in $\mathbb{R}^n$ of arbitrary prescribed Hausdorff dimension. This completely resolves a problem proposed by Kahane more than 60 years ago. The construction is based on a form of Diophantine approximation in number fields. ",Explicit Salem sets in $\mathbb{R}^n$ " We rule out a certain $9$-dimensional algebra over an algebraically closed field to be the basic algebra of a block of a finite group, thereby completing the classification of basic algebras of dimension at most $12$ of blocks of finite group algebras. ",A $9$-dimensional algebra which is not a block of a finite group " An elegant breadboard model of the LISA phasemeter is currently under development by a Danish-German consortium. The breadboard is build in the frame of an ESA technology development activity to demonstrate the feasibility and readiness of the LISA metrology baseline architecture. This article gives an overview about the breadboard design and its components, including the distribution of key functionalities. ",Breadboard model of the LISA phasemeter " Information transport and processing by pure magnonic spin currents in insulators is a promising alternative to conventional charge-current driven spintronic devices. The absence of Joule heating as well as the reduced spin wave damping in insulating ferromagnets has been suggested to enable the implementation of efficient logic devices. After the proof of concept for a logic majority gate based on the superposition of spin waves has been successfully demonstrated, further components are required to perform complex logic operations. A key component is a switch that corresponds to a conventional magnetoresistive spin valve. Here, we report on magnetization orientation dependent spin signal detection in collinear magnetic multilayers with spin transport by magnonic spin currents. We find in Y3Fe5O12|CoO|Co tri-layers that the detected spin signal depends on the relative alignment of Y3Fe5O12 and Co. This demonstrates a spin valve behavior with an effect amplitude of 120% in our systems. We demonstrate the reliability of the effect and investigate the origin by both temperature and power dependent measurements, showing that spin rectification effects and a magnetic layer alignment dependent spin transport effect result in the measured signal. ",Ferroic collinear multilayer magnon spin valve " We study a compactification of the moduli space of theta characteristics, giving a modular interpretation of the geometric points and describing the boundary stratification. This space is different from the moduli space of spin curves. The modular description and the boundary stratification of the new compactification are encoded by a tropical moduli space. We show that this tropical moduli space is a refinement of the moduli space of spin tropical curves. We describe explicitly the induced decomposition of its cones. ",The moduli space of quasistable spin curves " Text classifiers are applied at scale in the form of one-size-fits-all solutions. Nevertheless, many studies show that classifiers are biased regarding different languages and dialects. When measuring and discovering these biases, some gaps present themselves and should be addressed. First, ``Does language, dialect, and topical content vary across geographical regions?'' and secondly ``If there are differences across the regions, do they impact model performance?''. We introduce a novel dataset called GeoOLID with more than 14 thousand examples across 15 geographically and demographically diverse cities to address these questions. We perform a comprehensive analysis of geographical-related content and their impact on performance disparities of offensive language detection models. Overall, we find that current models do not generalize across locations. Likewise, we show that while offensive language models produce false positives on African American English, model performance is not correlated with each city's minority population proportions. Warning: This paper contains offensive language. ",Measuring Geographic Performance Disparities of Offensive Language Classifiers " We present a dynamical system that naturally exhibits two unstable attractors that are completely enclosed by each others basin volume. This counter-intuitive phenomenon occurs in networks of pulse-coupled oscillators with delayed interactions. We analytically and numerically investigate this phenomenon and clarify the mechanism underlying it: Upon continuously removing the non-invertibility of the system, the set of two unstable attractors becomes a set of two non-attracting saddle states that are heteroclinically connected to each other. This transition from a network of unstable attractors to a heteroclinic cycle constitutes a new type of bifurcation in dynamical systems. ",Bifurcation From Networks of Unstable Attractors to Heteroclinic Switching We present a new chiral expansion scheme for the nucleon-nucleon scattering amplitude which preserves unitarity exactly. Our effective field theory builds on the power counting rules for 2-nucleon reducible diagrams proposed in \cite{lutz}. We evaluate the leading order terms of the isospin one scattering amplitude and elaborate in detail on the $^1S_0$ phase shift. Our chiral description of the $^1S_0$-phase shift does compete in quality with modern phenomenological nucleon-nucleon potentials. We describe elastic and inelastic scattering quantitatively up to laboratory energies of $E_{\rm lab} \simeq 600$ MeV. ,Effective chiral theory of nucleon-nucleon scattering " Context: Citations are a key measure of scientific performance in most fields, including software engineering. However, there is limited research that studies which characteristics of articles' metadata (title, abstract, keywords, and author list) are driving citations in this field. Objective: In this study, we propose a simple theoretical model for how citations come to be with respect to article metadata, we hypothesize theoretical linkages between metadata characteristics and citations of articles, and we empirically test these hypotheses. Method: We use multiple regression analyses to examine a data set comprising the titles, abstracts, keywords, and authors of 16,131 software engineering articles published between 1990 and 2020 in 20 highly influential software engineering venues. Results: We find that number of authors, number of keywords, number of question marks and dividers in the title, number of acronyms, abstract length, abstract propositional idea density, and corresponding authors in the core Anglosphere are significantly related to citations. Conclusion: Various characteristics of articles' metadata are linked to the frequency with which the corresponding articles are cited. These results partially confirm and partially go counter to prior findings in software engineering and other disciplines. ",Text and Team: What Article Metadata Characteristics Drive Citations in Software Engineering? " We discuss the sensitivity of the CASSINI experiments to gravitational waves emitted by the in-spiral of compact binaries. We show that the maximum distance reachable by the instrument is $\sim 100$ Mpc. In particular, CASSINI can detect massive black hole binaries with chirp mass $\simgt 10^6 \Ms$ in the Virgo Cluster with signal-to-noise ratio between 5 and 30 and possible compact objects of mass $\simgt 30 \Ms$ orbiting the massive black hole that our Galactic Centre is likely to harbour. ",Coalescing binaries and Doppler experiments " There has been a rapid progress in the task of Visual Question Answering with improved model architectures. Unfortunately, these models are usually computationally intensive due to their sheer size which poses a serious challenge for deployment. We aim to tackle this issue for the specific task of Visual Question Answering (VQA). A Convolutional Neural Network (CNN) is an integral part of the visual processing pipeline of a VQA model (assuming the CNN is trained along with entire VQA model). In this project, we propose an efficient and modular neural architecture for the VQA task with focus on the CNN module. Our experiments demonstrate that a sparsely activated CNN based VQA model achieves comparable performance to a standard CNN based VQA model architecture. ",Learning Sparse Mixture of Experts for Visual Question Answering This article is a survey on the local well-posedness problem for the general EPDiff equation. The main contribution concerns recent results on local existence of the geodesics on $\mathrm{Diff}(\mathbb{T}^{d})$ and $\mathrm{Diff}(\mathbb{R}^{d})$ when the inertia operator is a non-local Fourier multiplier. ,Local well-posedness of the EPDiff equation: a survey " The antiferromagnetic Heisenberg model on the icosahedron, which consists of 20 edge-sharing triangles and belongs to the icosahedral $I_h$ symmetry group, presents unconventional properties at the classical and quantum level. These originate in the frustrated nature of the interactions between the spins. For classical spins the magnetization is discontinuous in a magnetic field. Here we examine the importance of the connectivity of the icosahedron for the appearance of the magnetization discontinuity, and also investigate the transition from the classical to the quantum limit. The influence of connectivity on the magnetic properties is revealed by considering the cluster as being made up of a closed strip of a triangular lattice with two additional spins attached. The classical magnetization discontinuity is shown to evolve continuously from the discontinuity effected by these two spins when they are uncoupled to the cluster. In the second part the transition from the classical to the quantum limit is examined by focusing on the low energy spectrum taking fully into account the spatial and the spin symmetry of the model in the characterization of the states. A symmetry analysis of the highly degenerate due to the connectivity lowest energy classical manifold identifies as its direct fingerprint the low energy quantum states for spin magnitude as low as $s=1$, with the latter following a tower of states behavior which relates to the icosahedron having a structure reminiscent of a depleted triangular lattice. The classical character of the AHM for small $s$ is also detected on the ground state energy and correlation functions. On the other hand the classical magnetization discontinuity in a field eventually disappears for small $s$, after a weak reentrant behavior. ",Antiferromagnetic Heisenberg Model on the Icosahedron: Influence of Connectivity and the Transition from the Classical to the Quantum Limit " We have extended the time baseline for observations of the proper motions of radio sources in the Orion BN/KL region from 14.7 to 22.5 years. We present improved determinations for the sources BN and I. In addition, we address the proper motions of the double radio source n, that have been questioned in the literature. We confirm that all three sources are moving away at transverse velocities of tens of km s$^{-1}$ from a region in-between them, where they were located about 500 years ago. Source n exhibits a new component that we interpret as due to a one-sided ejection of free-free emitting plasma that took place after 2006.36. We used the highly accurate relative proper motions between sources BN and I to determine that their closest separation took place in the year 1475$\pm$6, when they were within $\sim$100 AU or less from each other in the plane of the sky. ",The Proper Motions of the Double Radio Source n in the Orion BN/KL Region " We investigate the mechanical behavior of a confined granular packing of irregular polyhedral particles under repeated heating and cooling cycles by means of numerical simulations with the Non-Smooth Contact Dynamics method. Assuming a homogeneous temperature distribution as well as constant temperature rate, we study the effect of the container shape, and coefficients of thermal expansions on the pressure buildup at the confining walls and the density evolution. We observe that small changes in the opening angle of the confinement can lead to a drastic peak pressure reduction. Furthermore, the displacement fields over several thermal cycles are obtained and we discover the formation of convection cells inside the granular material having the shape of a torus. The root mean square of the vorticity is then calculated from the displacement fields and a quadratic dependency on the ratio of thermal expansion coefficients is established. ",Behavior of confined granular beds under cyclic thermal loading " We study deposition dynamics of Na and Na$_2$ on an Ar substrate, both species neutral as well as charged. The system is modeled by a hierarchical approach describing the Na valence electrons by time-dependent density-functional theory while Na core, Ar atoms and their dynamical polarizability are treated by molecular dynamics. We explore effects of Na charge and initial kinetic energy of the impinging Na system. We find that neutral Na is captured into a loosely bound adsorbate state for sufficiently low impact energy. The charged monomers are more efficiently captured and the cation Na$^+$ even penetrates the surface layer. For charged dimers, we come to different final configurations depending on the process, direct deposit of Na$_2^+$ as a whole, or sequential deposit. In any case, charge dramatically amplifies the excitation of the matrix, in particular at the side of the Ar dipoles. The presence of a charge also enhances the binding to the surface and favours accumulation of larger compounds. ",Deposition dynamics of Na monomers and dimers on an Ar(001) substrate " In order to backward integrate the orbits of Milky Way (MW) dwarf galaxies, much effort has been invested in recent years to constrain their initial phase-space coordinates. Yet equally important are the assumptions on the potential that the dwarf galaxies experience over time, especially given the fact that the MW is currently accreting the Large Magellanic Cloud (LMC). In this work, using a dark matter-only zoom-in simulation, we test whether the use of common parametric forms of the potential is adequate to successfully backward integrate the orbits of the subhaloes from their present-day positions. We parametrise the recovered orbits and compare them with those from the simulations. We find that simple symmetric parametric forms of the potential fail to capture the complexities and the inhomogeneities of the true potential experienced by the subhaloes. More specifically, modelling a recent massive accretion like that of the LMC as a sum of two spherical parametric potentials leads to substantial errors in the recovered parameters of the orbits. These errors rival those caused due to a) a 30\% uncertainty in the virial mass of the MW and b) not modelling the potential of the recently accreted massive satellite. Our work suggests that i) the uncertainties in the parameters of the recovered orbits of some MW dwarfs may be under-estimated and that ii) researchers should characterise the uncertainties inherent to their choice of integration techniques and assumptions of the potential against cosmological zoom-in simulations of the MW, which include a recently-accreted LMC. ",Uncertainties associated with the backward integration of dwarf satellites using simple parametric potentials " In recent years, cosmic observational data have reported that our present universe is undergoing an accelerated expansion, which has been termed as mysterious ""dark energy"" phenomena, that is, the origin of dark energy has not been determined yet. According to our previous work that a new equation of state \cite{X.Shan and H.Chen1993} can be employed to explain the dark energy, and we are very interested in investigating the astrophysical scale properties of dark energy based on the new cosmological fluids given that the universe is filled with a dark energy fluid everywhere. Hence, in this paper, we study the exact solutions of spherically-symmetrical Einstein field equations describing wormholes supported by phantom energy from Shan-Chen (SC) fluids by considering an obvious relation between the transversal pressure and energy density which is different from our previous work \cite{111}. We have still investigated the important case $\psi\approx1$ which corresponds to the "" saturation effect "", and this regime corresponds to an effective form of "" asymptotic freedom "", occurring at cosmological rather than subnuclear scales. Then we find out two solutions by making some special choices for the shape function $b(r)$ and discuss the singularities of the solutions and find that the spacetimes are both geodesically incomplete. At the same time, it is worth noting that whether the solutions are geodesically complete or incomplete depends on the appropriate choice of the shape function $b(r)$ through our works. In addition, we match our interior solutions to the exterior Schwarzschild solutions and calculate out the total mass of the wormhole when $r\leq a$. ",Modeling phantom energy wormholes from Shan-Chen fluids " In a previous paper [J. Chem. Phys. 121 4501 (2004)] a unique bipolar decomposition, Psi = Psi1 + Psi2 was presented for stationary bound states Psi of the one-dimensional Schroedinger equation, such that the components Psi1 and Psi2 approach their semiclassical WKB analogs in the large action limit. The corresponding bipolar quantum trajectories, as defined in the usual Bohmian mechanical formulation, are classical-like and well-behaved, even when Psi has many nodes, or is wildly oscillatory. A modification for discontinuous potential stationary stattering states was presented in a second paper [J. Chem. Phys. 124 034115 (2006)], whose generalization for continuous potentials is given here. The result is an exact quantum scattering methodology using classical trajectories. For additional convenience in handling the tunneling case, a constant velocity trajectory version is also developed. ",Reconciling Semiclassical and Bohmian Mechanics: III. Scattering states for continuous potentials " Covering skill (a.k.a., option) discovery has been developed to improve the exploration of reinforcement learning in single-agent scenarios with sparse reward signals, through connecting the most distant states in the embedding space provided by the Fiedler vector of the state transition graph. However, these option discovery methods cannot be directly extended to multi-agent scenarios, since the joint state space grows exponentially with the number of agents in the system. Thus, existing researches on adopting options in multi-agent scenarios still rely on single-agent option discovery and fail to directly discover the joint options that can improve the connectivity of the joint state space of agents. In this paper, we show that it is indeed possible to directly compute multi-agent options with collaborative exploratory behaviors among the agents, while still enjoying the ease of decomposition. Our key idea is to approximate the joint state space as a Kronecker graph -- the Kronecker product of individual agents' state transition graphs, based on which we can directly estimate the Fiedler vector of the joint state space using the Laplacian spectrum of individual agents' transition graphs. This decomposition enables us to efficiently construct multi-agent joint options by encouraging agents to connect the sub-goal joint states which are corresponding to the minimum or maximum values of the estimated joint Fiedler vector. The evaluation based on multi-agent collaborative tasks shows that the proposed algorithm can successfully identify multi-agent options, and significantly outperforms prior works using single-agent options or no options, in terms of both faster exploration and higher cumulative rewards. ",Learning Multi-agent Skills for Tabular Reinforcement Learning using Factor Graphs " Coupled oscillators such as lasers, OPO's and BEC polaritons can rapidly and efficiently dissipate into a stable phase locked state that can be mapped onto the minimal energy (ground state) of classical spin Hamiltonians. However, for degenerate or near degenerate ground state manifolds, statistical fair sampling is required to obtain a complete knowledge of the minimal energy state, which needs many repetitions of simulations under identical conditions. We show that with dissipatively coupled lasers such fair sampling can be achieved rapidly and accurately by exploiting the many longitudinal modes of each laser to form an ensemble of identical but independent simulators, acting in parallel. We fairly sampled the ground state manifold of square, triangular and Kagome lattices by measuring their coherence function identifying manifolds composed of a single, doubly degenerate, and highly degenerate ground states, respectively. ",Rapid fair sampling of XY spin Hamiltonian with a laser simulator " Self-supervised learning has been widely applied to train high-quality vision transformers. Unleashing their excellent performance on memory and compute constraint devices is therefore an important research topic. However, how to distill knowledge from one self-supervised ViT to another has not yet been explored. Moreover, the existing self-supervised knowledge distillation (SSKD) methods focus on ConvNet based architectures are suboptimal for ViT knowledge distillation. In this paper, we study knowledge distillation of self-supervised vision transformers (ViT-SSKD). We show that directly distilling information from the crucial attention mechanism from teacher to student can significantly narrow the performance gap between both. In experiments on ImageNet-Subset and ImageNet-1K, we show that our method AttnDistill outperforms existing self-supervised knowledge distillation (SSKD) methods and achieves state-of-the-art k-NN accuracy compared with self-supervised learning (SSL) methods learning from scratch (with the ViT-S model). We are also the first to apply the tiny ViT-T model on self-supervised learning. Moreover, AttnDistill is independent of self-supervised learning algorithms, it can be adapted to ViT based SSL methods to improve the performance in future research. The code is here: https://github.com/wangkai930418/attndistill ",Attention Distillation: self-supervised vision transformer students need more guidance " Let G be group; a finite p-subgroup S of G is a Sylow p-subgroup if every finite p-subgroup of G is conjugate to a subgroup of S. In this paper, we examine the relations between the fusion system over S which is given by conjugation in G and a certain chamber system C, on which G acts chamber transitively with chamber stabilizer N_G(S). Next, we introduce the notion of a fusion system with a parabolic family and we show that a chamber system can be associated to such a fusion system. We determine some conditions the chamber system has to fulfill in order to assure the saturation of the underlying fusion system. We give an application to fusion systems with parabolic families of classical type. ",Saturated fusion systems with parabolic families " The entanglement content of superpositions of pairs of degenerate eigenstates of a bipartite system are considered in the case that both are also eigenstates of the $z$ component of the total angular momentum. It is shown that the von Neumann entropy of the state that is obtained tracing out one of the parts of the system has a definite convexity (concavity) as a function of the superposition parameter and that its convexity (concavity) can be predicted using a quantity of information that measures the entropy shared by the states at the extremes of the superposition. Several examples of two particle system, whose eigenfunctions and density matrices can be obtained exactly, are analyzed thoroughly. ",Convexity properties of superpositions of degenerate bipartite eigenstates " We present our studies on jet-induced modifications of the characteristic of the bulk nuclear matter. To describe such a matter, we use efficient relativistic hydrodynamic simulations in (3+1) dimensions employing the Graphics Processing Unit (GPU) in the parallel programming framework. We use Cartesian coordinates in the calculations to ensure a high spatial resolution that is constant throughout the evolution of the system. We show our results on how jets modify the hydrodynamics fields and discuss the implications. ",Jet-induced modifications of the characteristic of the bulk nuclear matter In the present paper the asymptotic formulae for the first moment of the Riemann zeta-function on the critical line is proven under assumption of the Riemann Hypothesis. ,On Asymptotic Formula for the First Moment of the Riemann Zeta-Function on the Critical Line " In this paper we propose a unified framework for structured prediction with latent variables which includes hidden conditional random fields and latent structured support vector machines as special cases. We describe a local entropy approximation for this general formulation using duality, and derive an efficient message passing algorithm that is guaranteed to converge. We demonstrate its effectiveness in the tasks of image segmentation as well as 3D indoor scene understanding from single images, showing that our approach is superior to latent structured support vector machines and hidden conditional random fields. ",Efficient Structured Prediction with Latent Variables for General Graphical Models " We have investigated uniaxial and hydrostatic pressure effects on superconductivity in Fe1.07Te0.88S0.12 through magnetic-susceptibility measurements down to 1.8 K. The superconducting transition temperature Tc is enhanced by out-of-plane pressure (uniaxial pressure along the c-axis); the onset temperature of the superconductivity reaches 11.8 K at 0.4 GPa. In contrast, Tc is reduced by in-plane pressure (uniaxial pressure along the ab-plane) and hydrostatic pressure. Taking into account these results, it is inferred that the superconductivity of Fe1+yTe1-xSx is enhanced when the lattice constant c considerably shrinks. This implies that the relationship between Tc and the anion height for Fe1+yTe1-xSx is similar to that applicable to most iron-based superconductors. We consider the reduction of Tc by hydrostatic pressure due to suppression of spin fluctuations because the system moves away from antiferromagnetic ordering, and the enhancement of Tc by out-of-plane pressure due to the anion height effect on Tc. ",Anisotropic pressure effects on superconductivity in Fe1+yTe1-xSx " We study the near horizon limit of a four dimensional extreme rotating black hole. The limiting metric is a completely nonsingular vacuum solution, with an enhanced symmetry group SL(2,R) x U(1). We show that many of the properties of this solution are similar to the AdS_2 x S^2 geometry arising in the near horizon limit of extreme charged black holes. In particular, the boundary at infinity is a timelike surface. This suggests the possibility of a dual quantum mechanical description. A five dimensional generalization is also discussed. ",The Extreme Kerr Throat Geometry: A Vacuum Analog of AdS_2 x S^2 " This paper studies equable parallelograms whose vertices lie on the integer lattice. Using Rosenberger's Theorem on generalised Markov equations, we show that the g.c.d. of the side lengths of such parallelograms can only be 3, 4 or 5, and in each of these cases the set of parallelograms naturally forms an infinite tree all of whose vertices have degree 4, bar the root. The paper then focuses on what we call Pythagorean equable parallelograms. These are lattice equable parallelograms whose complement in a circumscribing rectangle consists of two Pythagorean triangles. We prove that for these parallelograms the shortest side can only be 3, 4, 5, 6 or 10, and there are five infinite families of such parallelograms, given by solutions to corresponding Pell-like equations. ",Lattice Equable Quadrilaterals I -- Parallelograms The dynamics of Rydberg states of atomic hydrogen illuminated by resonant elliptically polarized microwaves is investigated both semiclassically and quantum mechanically in a simplified two-dimensional model of an atom. Semiclassical predictions for quasienergies of the system are found to be in a very good agreement with exact quantum data enabling a classification of possible types of motion and their dynamics with the change of the ellipticity of the microwaves. Particular attention is paid to the dynamics of the nonspreading wave packet states which are found to exist for an arbitrary microwave polarization. ,H atom in elliptically polarized microwaves: Semiclassical versus quantum resonant dynamics " Hydrogen emission lines can provide extensive information about star-forming galaxies in both the local and high-redshift Universe. We present a detailed Lyman continuum (LyC), Lyman-alpha (Ly{\alpha}), and Balmer line (H{\alpha} and H\b{eta}) radiative transfer study of a high-resolution isolated Milky-Way simulation using the Arepo-RT radiation hydrodynamics code with the SMUGGLE galaxy formation model. The realistic framework includes stellar feedback, non-equilibrium thermochemistry, and dust grain evolution in the interstellar medium (ISM). We extend our Cosmic Ly{\alpha} Transfer (COLT) code with photoionization equilibrium Monte Carlo radiative transfer for self-consistent end-to-end (non-)resonant line predictions. Accurate LyC reprocessing to recombination emission requires modelling pre-absorption by dust (27.5%), helium ionization (8.7%), and anisotropic escape fractions (7.9%), as these reduce the available budget for hydrogen line emission (55.9%). We investigate the role of the multiphase dusty ISM, disc geometry, gas kinematics, and star formation activity in governing the physics of emission and escape, focusing on the time variability, gas phase structure, and spatial, spectral, and viewing angle dependence of the emergent photons. Isolated disc simulations are well-suited for comprehensive observational comparisons with local H{\alpha} surveys, but would require a proper cosmological circumgalactic medium (CGM) environment as well as less dust absorption and rotational broadening to serve as analogs for high-redshift Ly{\alpha} emitting galaxies. Future applications of our framework to next-generation cosmological simulations of galaxy formation including radiation-hydrodynamics that resolve <10 pc multiphase ISM and <1 kpc CGM structures will provide crucial insights and predictions for current and upcoming Ly{\alpha} observations. ",The physics of Lyman-alpha escape from disc-like galaxies " In this paper, we study two classes of Kirchhoff type problems set on a double phase framework. That is, the functional space where finding solutions coincides with the Musielak-Orlicz-Sobolev space $W^{1,\mathcal H}_0(\Omega)$, with modular function $\mathcal H$ related to the so called double phase operator. Via a variational approach, we provide existence and multiplicity results. ",Existence and multiplicity results for Kirchhoff type problems on a double phase setting " We study the problem of realizing the full spectrum of bipedal locomotion on a real robot with sim-to-real reinforcement learning (RL). A key challenge of learning legged locomotion is describing different gaits, via reward functions, in a way that is intuitive for the designer and specific enough to reliably learn the gait across different initial random seeds or hyperparameters. A common approach is to use reference motions (e.g. trajectories of joint positions) to guide learning. However, finding high-quality reference motions can be difficult and the trajectories themselves narrowly constrain the space of learned motion. At the other extreme, reference-free reward functions are often underspecified (e.g. move forward) leading to massive variance in policy behavior, or are the product of significant reward-shaping via trial-and-error, making them exclusive to specific gaits. In this work, we propose a reward-specification framework based on composing simple probabilistic periodic costs on basic forces and velocities. We instantiate this framework to define a parametric reward function with intuitive settings for all common bipedal gaits - standing, walking, hopping, running, and skipping. Using this function we demonstrate successful sim-to-real transfer of the learned gaits to the bipedal robot Cassie, as well as a generic policy that can transition between all of the two-beat gaits. ",Sim-to-Real Learning of All Common Bipedal Gaits via Periodic Reward Composition " We study the problem of cooperative localization of a large network of nodes in integer-coordinated unit disk graphs, a simplified but useful version of general random graph. Exploiting the property that the radius $r$ sets clear cut on the connectivity of two nodes, we propose an essential philosophy that ""no connectivity is also useful information just like the information being connected"" in unit disk graphs. Exercising this philosophy, we show that the conventional network localization problem can be re-formulated to significantly reduce the search space, and that global rigidity, a necessary and sufficient condition for the existence of unique solution in general graphs, is no longer necessary. While the problem is still NP-hard, we show that a (depth-first) tree-search algorithm with memory O(N) ($N$ is the network size) can be developed, and for practical setups, the search complexity and speed is very manageable, and is magnitudes less than the conventional problem, especially when the graph is sparse or when only very limited anchor nodes are available. ",Network Localization on Unit Disk Graphs " In this paper, we study the problem of robust sparse mean estimation, where the goal is to estimate a $k$-sparse mean from a collection of partially corrupted samples drawn from a heavy-tailed distribution. Existing estimators face two critical challenges in this setting. First, they are limited by a conjectured computational-statistical tradeoff, implying that any computationally efficient algorithm needs $\tilde\Omega(k^2)$ samples, while its statistically-optimal counterpart only requires $\tilde O(k)$ samples. Second, the existing estimators fall short of practical use as they scale poorly with the ambient dimension. This paper presents a simple mean estimator that overcomes both challenges under moderate conditions: it runs in near-linear time and memory (both with respect to the ambient dimension) while requiring only $\tilde O(k)$ samples to recover the true mean. At the core of our method lies an incremental learning phenomenon: we introduce a simple nonconvex framework that can incrementally learn the top-$k$ nonzero elements of the mean while keeping the zero elements arbitrarily small. Unlike existing estimators, our method does not need any prior knowledge of the sparsity level $k$. We prove the optimality of our estimator by providing a matching information-theoretic lower bound. Finally, we conduct a series of simulations to corroborate our theoretical findings. Our code is available at https://github.com/huihui0902/Robust_mean_estimation. ",Robust Sparse Mean Estimation via Incremental Learning " While artificial intelligence (AI) holds promise for addressing societal challenges, issues of exactly which tasks to automate and to what extent to do so remain understudied. We approach this problem of task delegability from a human-centered perspective by developing a framework on human perception of task delegation to AI. We consider four high-level factors that can contribute to a delegation decision: motivation, difficulty, risk, and trust. To obtain an empirical understanding of human preferences in different tasks, we build a dataset of 100 tasks from academic papers, popular media portrayal of AI, and everyday life, and administer a survey based on our proposed framework. We find little preference for full AI control and a strong preference for machine-in-the-loop designs, in which humans play the leading role. Among the four factors, trust is the most correlated with human preferences of optimal human-machine delegation. This framework represents a first step towards characterizing human preferences of AI automation across tasks. We hope this work encourages future efforts towards understanding such individual attitudes; our goal is to inform the public and the AI research community rather than dictating any direction in technology development. ","Ask Not What AI Can Do, But What AI Should Do: Towards a Framework of Task Delegability" " The emergent mechanism provides a possible way to resolve the big bang singularity problem by assuming that our universe originates from the Einstein static (ES) state. Thus, the existence of a stable ES solution becomes a very crucial prerequisite for the emergent scenario. In this paper, we study the stability of an ES universe in gravity theory with a non-minimal coupling between the kinetic term of a scalar field and the Einstein tensor. We find that the ES solution is stable under both scalar and tensor perturbations when the model parameters satisfy certain conditions, which indicates that the big bang singularity can be avoided successfully by the emergent mechanism in the non-minimally kinetic coupled gravity. ",Stability of Einstein static universe in gravity theory with a non-minimal derivative coupling Collapsing Bose-Einstein condensates are rich and complex quantum systems for which quantitative explanation by simple models has proved elusive. We present new experimental data on the collapse of high density Rb-85 condensates with attractive interactions and find quantitative agreement with the predictions of the Gross-Pitaevskii equation. The collapse data and measurements of the decay of atoms from our condensates allow us to put new limits on the value of the Rb-85 three-body loss coefficient K_3 at small positive and negative scattering lengths. ,Bosenova and three-body loss in a Rb-85 Bose-Einstein condensate " In this paper we derive Ward-Takahashi identities from the path integral of supersymmetric five-dimensional field theories with an $SU(1,3)$ spacetime symmetry in the presence of instantons. We explicitly show how $SU(1,3)$ is enhanced to $SU(1,3)\times U(1)$ where the additional $U(1)$ acts non-perturbatively. Solutions to such Ward-Takahashi identities were previously obtained from correlators of six-dimensional Lorentzian conformal field theories but where the instanton number was replaced by the momentum along a null direction. Here we study the reverse procedure whereby we construct correlation functions out of towers of five-dimensional operators which satisfy the Ward-Takahashi identities of a six-dimensional conformal field theory. This paves the way to computing observables in six dimensions using five-dimensional path integral techniques. We also argue that, once the instanton sector is included into the path integral, the coupling of the five-dimensional Lagrangian must be quantised, leaving no free continuous parameters. ",Five-Dimensional Path Integrals for Six-Dimensional Conformal Field Theories " The robustness of Text-to-SQL parsers against adversarial perturbations plays a crucial role in delivering highly reliable applications. Previous studies along this line primarily focused on perturbations in the natural language question side, neglecting the variability of tables. Motivated by this, we propose the Adversarial Table Perturbation (ATP) as a new attacking paradigm to measure the robustness of Text-to-SQL models. Following this proposition, we curate ADVETA, the first robustness evaluation benchmark featuring natural and realistic ATPs. All tested state-of-the-art models experience dramatic performance drops on ADVETA, revealing models' vulnerability in real-world practices. To defend against ATP, we build a systematic adversarial training example generation framework tailored for better contextualization of tabular data. Experiments show that our approach not only brings the best robustness improvement against table-side perturbations but also substantially empowers models against NL-side perturbations. We release our benchmark and code at: https://github.com/microsoft/ContextualSP. ",Towards Robustness of Text-to-SQL Models Against Natural and Realistic Adversarial Table Perturbation " We report on a member of the spin-disordered honeycomb lattice antiferromagnet in a quasi-one-dimensional cobaltate Ba_3Co_2O_6(CO_3)_0.7. Resistivity exhibits as semimetallic along the face-sharing CoO6 chains. Magnetic susceptibility shows strongly anisotropic Ising-spin character with the easy axis along the chain due to significant spin-orbit coupling and a trigonal crystal field. Nevertheless, ^135Ba NMR detects no indication of the long-range magnetic order down to 0.48 K. Marginally itinerant electrons possess large entropy and low-lying excitations with a Wilson ratio R_W = 116, which highlight interplays of charge, spin, and orbital in the disordered ground state. ",Spin disorder in an Ising honeycomb chain cobaltate " We present the first comparison of observed stellar continuum spectra of high-redshift galaxies and mock galaxy spectra generated from hydrodynamical simulations. The mock spectra are produced from the IllustrisTNG TNG100 simulation combined with stellar population models and take into account dust attenuation and realistic observational effects (aperture effects and noise). We compare the simulated $D_n4000$ and EW(H$\delta$) of galaxies with $10.5 \leq \log(M_\ast/M_\odot) \leq 11.5$ at $0.6 \leq z \leq 1.0$ to the observed distributions from the LEGA-C survey. TNG100 globally reproduces the observed distributions of spectral indices, implying that the age distribution of galaxies in TNG100 is generally realistic. Yet there are small but significant differences. For old galaxies, TNG100 shows small $D_n4000$ when compared to LEGA-C, while LEGA-C galaxies have larger EW(H$\delta$) at fixed $D_n4000$. There are several possible explanations: 1) LEGA-C galaxies have overall older ages combined with small contributions (a few percent in mass) from younger ($<1$~Gyr) stars, while TNG100 galaxies may not have such young sub-populations; 2) the spectral mismatch could be due to systematic uncertainties in the stellar population models used to convert stellar ages and metallicities to observables. In conclusion, the latest cosmological galaxy formation simulations broadly reproduce the global age distribution of galaxies at $z\sim1$ and, at the same time, the high quality of the latest observed and simulated datasets help constrain stellar population synthesis models as well as the physical models underlying the simulations. ",Towards precise galaxy evolution: a comparison between spectral indices of $z\sim1$ galaxies in the IllustrisTNG simulation and the LEGA-C survey " Bundle adjustment (BA) on LiDAR point clouds has been extensively investigated in recent years due to its ability to optimize multiple poses together, resulting in high accuracy and global consistency for point cloud. However, the accuracy and speed of LiDAR bundle adjustment depend on the quality of plane extraction, which provides point association for LiDAR BA. In this study, we propose a novel and efficient voxel-based approach for plane extraction that is specially designed to provide point association for LiDAR bundle adjustment. To begin, we partition the space into multiple voxels of a fixed size and then split these root voxels based on whether the points are on the same plane, using an octree structure. We also design a novel plane determination method based on principle component analysis (PCA), which segments the points into four even quarters and compare their minimum eigenvalues with that of the initial point cloud. Finally, we adopt a plane merging method to prevent too many small planes from being in a single voxel, which can increase the optimization time required for BA. Our experimental results on HILTI demonstrate that our approach achieves the best precision and least time cost compared to other plane extraction methods. ",An Efficient Plane Extraction Approach for Bundle Adjustment on LiDAR Point clouds " We present the COmoving Lagrangian Acceleration (COLA) method: an N-body method for solving for Large Scale Structure (LSS) in a frame that is comoving with observers following trajectories calculated in Lagrangian Perturbation Theory (LPT). Unlike standard N-body methods, the COLA method can straightforwardly trade accuracy at small-scales in order to gain computational speed without sacrificing accuracy at large scales. This is especially useful for cheaply generating large ensembles of accurate mock halo catalogs required to study galaxy clustering and weak lensing, as those catalogs are essential for performing detailed error analysis for ongoing and future surveys of LSS. As an illustration, we ran a COLA-based N-body code on a box of size 100Mpc/h with particles of mass ~5*10^9Msolar/h. Running the code with only 10 timesteps was sufficient to obtain an accurate description of halo statistics down to halo masses of at least 10^11Msolar/h. This is only at a modest speed penalty when compared to mocks obtained with LPT. A standard detailed N-body run is orders of magnitude slower than our COLA-based code. The speed-up we obtain with COLA is due to the fact that we calculate the large-scale dynamics exactly using LPT, while letting the N-body code solve for the small scales, without requiring it to capture exactly the internal dynamics of halos. Achieving a similar level of accuracy in halo statistics without the COLA method requires at least 3 times more timesteps than when COLA is employed. ",Solving Large Scale Structure in Ten Easy Steps with COLA " Pre-trained language models (PrLMs) have demonstrated superior performance due to their strong ability to learn universal language representations from self-supervised pre-training. However, even with the help of the powerful PrLMs, it is still challenging to effectively capture task-related knowledge from dialogue texts which are enriched by correlations among speaker-aware utterances. In this work, we present SPIDER, Structural Pre-traIned DialoguE Reader, to capture dialogue exclusive features. To simulate the dialogue-like features, we propose two training objectives in addition to the original LM objectives: 1) utterance order restoration, which predicts the order of the permuted utterances in dialogue context; 2) sentence backbone regularization, which regularizes the model to improve the factual correctness of summarized subject-verb-object triplets. Experimental results on widely used dialogue benchmarks verify the effectiveness of the newly introduced self-supervised tasks. ",Structural Pre-training for Dialogue Comprehension " We count the connected components in the moduli space of PU(p,q)-representations of the fundamental group for a closed oriented surface. The components are labelled by pairs of integers which arise as topological invariants of the flat bundles associated to the representations. Our results show that for each allowed value of these invariants, which are bounded by a Milnor-Wood type inequality, there is a unique non-empty connected component. Interpreting the moduli space of representations as a moduli space of Higgs bundles, we take a Morse theoretic approach using a certain smooth proper function on the Higgs moduli space. A key step is the identification of the function's local minima as moduli spaces of holomorphic triples. We prove that these moduli spaces of triples are non-empty and irreducible. ","Representations of the fundamental group of a surface in PU(p,q) and holomorphic triples" We analyze the current instability of the steady state with a direct current for an ungated two-dimensional(2D) electron layer for arbitrary current and the carrier scattering strength. We demonstrate the possibility of single-mode generator operating in terahertz frequency range. ,Current instability and single-mode THz generation in ungated two-dimensional electron gas " We theoretically investigate a possible idea to introduce magnetic impurities to a superfluid Fermi gas. In the presence of population imbalance ($N_\uparrow>N_\downarrow$, where $N_\sigma$ is the number of Fermi atoms with pseudospin $\sigma=\uparrow,\downarrow$), we show that nonmagnetic potential scatterers embedded in the system are magnetized in the sense that some of excess $\uparrow$-spin atoms are localized around them. They destroy the superfluid order parameter around them, as in the case of magnetic impurity effect discussed in the superconductivity literature. This pair-breaking effect naturally leads to localized excited states below the superfluid excitation gap. To confirm our idea in a simply manner, we treat an attractive Fermi Hubbard model within the mean-field theory at T=0. We self-consistently determine superfluid properties around a nonmagnetic impurity, such as the superfluid order parameter, local population imbalance, as well as single-particle density of states, in the presence of population imbalance. Since the competition between superconductivity and magnetism is one of the most fundamental problems in condensed matter physics, our results would be useful for the study of this important issue in cold Fermi gases. ",Formation of magnetic impurities and pair-breaking effect in a superfluid Fermi gas " A model independent theoretical analysis of recent experimental data on deuteron photodisintegration with polarized laser beams is presented. We find that it is important to distinguish between the three isovector E1 amplitudes $E1_v^j$ in reaction channels with total angular momentum $j=0,1,2$ and that the isoscalar M1 amplitude $M1_s$ is non-zero in the photon energy range $3.5 MeV < E_\gamma < 10 MeV$ ",Deuteron photodisintegration with polarized lasers " For many machine learning tasks, the input data lie on a low-dimensional manifold embedded in a high dimensional space and, because of this high-dimensional structure, most algorithms are inefficient. The typical solution is to reduce the dimension of the input data using standard dimension reduction algorithms such as ISOMAP, LAPLACIAN EIGENMAPS or LLES. This approach, however, does not always work in practice as these algorithms require that we have somewhat ideal data. Unfortunately, most data sets either have missing entries or unacceptably noisy values. That is, real data are far from ideal and we cannot use these algorithms directly. In this paper, we focus on the case when we have missing data. Some techniques, such as matrix completion, can be used to fill in missing data but these methods do not capture the non-linear structure of the manifold. Here, we present a new algorithm MR-MISSING that extends these previous algorithms and can be used to compute low dimensional representation on data sets with missing entries. We demonstrate the effectiveness of our algorithm by running three different experiments. We visually verify the effectiveness of our algorithm on synthetic manifolds, we numerically compare our projections against those computed by first filling in data using nlPCA and mDRUR on the MNIST data set, and we also show that we can do classification on MNIST with missing data. We also provide a theoretical guarantee for MR-MISSING under some simplifying assumptions. ",Unsupervised Metric Learning in Presence of Missing Data " Chemotaxis can be defined as an innate behavioural response by an organism to a directional stimulus, in which bacteria, and other single-cell or multicellular organisms direct their movements according to certain chemicals in their environment. This is important for bacteria to find food (e.g., glucose) by swimming towards the highest concentration of food molecules, or to flee from poisons. Based on self-organized computational approaches and similar stigmergic concepts we derive a novel swarm intelligent algorithm. What strikes from these observations is that both eusocial insects as ant colonies and bacteria have similar natural mechanisms based on stigmergy in order to emerge coherent and sophisticated patterns of global collective behaviour. Keeping in mind the above characteristics we will present a simple model to tackle the collective adaptation of a social swarm based on real ant colony behaviors (SSA algorithm) for tracking extrema in dynamic environments and highly multimodal complex functions described in the well-know De Jong test suite. Later, for the purpose of comparison, a recent model of artificial bacterial foraging (BFOA algorithm) based on similar stigmergic features is described and analyzed. Final results indicate that the SSA collective intelligence is able to cope and quickly adapt to unforeseen situations even when over the same cooperative foraging period, the community is requested to deal with two different and contradictory purposes, while outperforming BFOA in adaptive speed. Results indicate that the present approach deals well in severe Dynamic Optimization problems. ",Computational Chemotaxis in Ants and Bacteria over Dynamic Environments " Recent research works have focused on generating human models and garments from their 2D images. However, state-of-the-art researches focus either on only a single layer of the garment on a human model or on generating multiple garment layers without any guarantee of the intersection-free geometric relationship between them. In reality, people wear multiple layers of garments in their daily life, where an inner layer of garment could be partially covered by an outer one. In this paper, we try to address this multi-layer modeling problem and propose the Layered-Garment Net (LGN) that is capable of generating intersection-free multiple layers of garments defined by implicit function fields over the body surface, given the person's near front-view image. With a special design of garment indication fields (GIF), we can enforce an implicit covering relationship between the signed distance fields (SDF) of different layers to avoid self-intersections among different garment surfaces and the human body. Experiments demonstrate the strength of our proposed LGN framework in generating multi-layer garments as compared to state-of-the-art methods. To the best of our knowledge, LGN is the first research work to generate intersection-free multiple layers of garments on the human body from a single image. ",Layered-Garment Net: Generating Multiple Implicit Garment Layers from a Single Image " The Fat-Link Irrelevant Clover (FLIC) fermion action provides a new form of nonperturbative O(a) improvement and allows efficient access to the light quark-mass regime. FLIC fermions enable the construction of the nonperturbatively O(a)-improved conserved vector current without the difficulties associated with the fine tuning of the improvement coefficients. The simulations are performed with an O(a^2) mean-field improved plaquette-plus-rectangle gluon action on a 20^3 x 40 lattice with a lattice spacing of 0.128 fm, enabling the first simulation of baryon form factors at light quark masses on a large volume lattice. Magnetic moments, electric charge radii and magnetic radii are extracted from these form factors, and show interesting chiral nonanalytic behavior in the light quark mass regime. ",Electromagnetic Form Factors with FLIC Fermions " In the past few decades polynomial curves with Pythagorean Hodograph (for short PH curves) have received considerable attention due to their usefulness in various CAD/CAM areas, manufacturing, numerical control machining and robotics. This work deals with classes of PH curves built-upon exponential-polynomial spaces (for short EPH curves). In particular, for the two most frequently encountered exponential-polynomial spaces, we first provide necessary and sufficient conditions to be satisfied by the control polygon of the B\'{e}zier-like curve in order to fulfill the PH property. Then, for such EPH curves, fundamental characteristics like parametric speed or cumulative and total arc length are discussed to show the interesting analogies with their well-known polynomial counterparts. Differences and advantages with respect to ordinary PH curves become commendable when discussing the solutions to application problems like the interpolation of first-order Hermite data. Finally, a new evaluation algorithm for EPH curves is proposed and shown to compare favorably with the celebrated de Casteljau-like algorithm and two recently proposed methods: Wo\'zny and Chudy's algorithm and the dynamic evaluation procedure by Yang and Hong. ",Construction and evaluation of PH curves in exponential-polynomial spaces " We propose sequential Monte Carlo based algorithms for maximum likelihood estimation of the static parameters in hidden Markov models with an intractable likelihood using ideas from approximate Bayesian computation. The static parameter estimation algorithms are gradient based and cover both offline and online estimation. We demonstrate their performance by estimating the parameters of three intractable models, namely the alpha-stable distribution, g-and-k distribution, and the stochastic volatility model with alpha-stable returns, using both real and synthetic data. ",Parameter Estimation in Hidden Markov Models with Intractable Likelihoods Using Sequential Monte Carlo " In this work, we study the $D\bar{D}$, $DD$, $D\bar{D}_s$, $DD_s$, $D_s\bar{D}_s$ and $D_sD_s$ tetraquark molecular states with the $J^{PC}=0^{++}$ via the QCD sum rules. The prediction $M_{D_s\bar {D}_s} = 3.98\pm0.10\, \rm{GeV}$ is in very good agreement with the experimental value $M_{X(3960)} = 3956 \pm 5\pm 10 \,\rm{MeV}$ from the LHCb collaboration and supports assigning the $X(3960)$ as the $D_s^+D_s^-$ molecular state with the $J^{PC}=0^{++}$. We take account of our previous works on the four-quark states consisting of two color-neutral clusters, and acquire the mass spectrum of the ground state hidden-charm and doubly-charm tetraquark molecular states. ",Analysis of the $X(3960)$ and related tetraquark molecular states via the QCD sum rules " We introduce a framework that enables efficient sampling from learned probability distributions for MRI reconstruction. Different from conventional deep learning-based MRI reconstruction techniques, samples are drawn from the posterior distribution given the measured k-space using the Markov chain Monte Carlo (MCMC) method. In addition to the maximum a posteriori (MAP) estimate for the image, which can be obtained with conventional methods, the minimum mean square error (MMSE) estimate and uncertainty maps can also be computed. The data-driven Markov chains are constructed from the generative model learned from a given image database and are independent of the forward operator that is used to model the k-space measurement. This provides flexibility because the method can be applied to k-space acquired with different sampling schemes or receive coils using the same pre-trained models. Furthermore, we use a framework based on a reverse diffusion process to be able to utilize advanced generative models. The performance of the method is evaluated on an open dataset using 10-fold undersampling in k-space. ",Bayesian MRI Reconstruction with Joint Uncertainty Estimation using Diffusion Models " Anomaly detection aims to distinguish observations that are rare and different from the majority. While most existing algorithms assume that instances are i.i.d., in many practical scenarios, links describing instance-to-instance dependencies and interactions are available. Such systems are called attributed networks. Anomaly detection in attributed networks has various applications such as monitoring suspicious accounts in social media and financial fraud in transaction networks. However, it remains a challenging task since the definition of anomaly becomes more complicated and topological structures are heterogeneous with nodal attributes. In this paper, we propose a spectral convolution and deconvolution based framework -- SpecAE, to project the attributed network into a tailored space to detect global and community anomalies. SpecAE leverages Laplacian sharpening to amplify the distances between representations of anomalies and the ones of the majority. The learned representations along with reconstruction errors are combined with a density estimation model to perform the detection. They are trained jointly as an end-to-end framework. Experiments on real-world datasets demonstrate the effectiveness of SpecAE. ",SpecAE: Spectral AutoEncoder for Anomaly Detection in Attributed Networks " We consider the production of $Z$ bosons associated with beauty quarks at the LHC using a combined $k_T$ + collinear QCD factorization approach, that interpolates between small $x$ and large $x$ physics. Our consideration is based on the off-shell gluon-gluon fusion subprocess $g^* g^* \to Z Q\bar Q$ at the leading order ${\cal O}(\alpha\alpha_s^2)$ (where the $Z$ boson further decays into a lepton pair), calculated in the $k_T$-factorization approach, and several subleading ${\cal O}(\alpha \alpha_s^2 )$ and ${\cal O}(\alpha \alpha_s^3 )$ subprocesses involving quark-antiquark and quark-gluon interactions, taken into account in conventional (collinear) QCD factorization. The contributions from double parton scattering are discussed as well. The transverse momentum dependent (or unintegrated) gluon densities in a proton are derived from Catani-Ciafaloni-Fiorani-Marchesini (CCFM) evolution equation. We achieve reasonably good agreement with the latest data taken by CMS and ATLAS Collaborations. The comparison of our results with next-to-leading-order pQCD predictions, obtained in the collinear QCD factorization, is presented. We discuss the uncertainties of our calculations and demonstrate the importance of subleading quark involving contributions in describing the LHC data in the whole kinematic region. ",Associated production of $Z$ bosons and $b$-jets at the LHC in the combined $k_T$ + collinear QCD factorization approach " A proof of work (PoW) is an important cryptographic construct enabling a party to convince others that they invested some effort in solving a computational task. Arguably, its main impact has been in the setting of cryptocurrencies such as Bitcoin and its underlying blockchain protocol, which received significant attention in recent years due to its potential for various applications as well as for solving fundamental distributed computing questions in novel threat models. PoWs enable the linking of blocks in the blockchain data structure and thus the problem of interest is the feasibility of obtaining a sequence (chain) of such proofs. In this work, we examine the hardness of finding such chain of PoWs against quantum strategies. We prove that the chain of PoWs problem reduces to a problem we call multi-solution Bernoulli search, for which we establish its quantum query complexity. Effectively, this is an extension of a threshold direct product theorem to an average-case unstructured search problem. Our proof, adding to active recent efforts, simplifies and generalizes the recording technique of Zhandry (Crypto'19). As an application, we revisit the formal treatment of security of the core of the Bitcoin consensus protocol, the Bitcoin backbone (Eurocrypt'15), against quantum adversaries, while honest parties are classical and show that protocol's security holds under a quantum analogue of the classical ``honest majority'' assumption. Our analysis indicates that the security of Bitcoin backbone is guaranteed provided the number of adversarial quantum queries is bounded so that each quantum query is worth $O(p^{-1/2})$ classical ones, where $p$ is the success probability of a single classical query to the protocol's underlying hash function. Somewhat surprisingly, the wait time for safe settlement in the case of quantum adversaries matches the safe settlement time in the classical case. ",Quantum Multi-Solution Bernoulli Search with Applications to Bitcoin's Post-Quantum Security Choosing the regularization parameter for inverse problems is of major importance for the performance of the regularization method. We will introduce a fast version of the Lepskij balancing principle and show that it is a valid parameter choice method for Tikhonov regularization both in a deterministic and a stochastic noise regime as long as minor conditions on the solution are fulfilled. ,Parameter Choice by Fast Balancing " A magnetic model for low/hard state (LHS) of black hole X-ray binaries (BHXBs),H1743-322 and GX 339-4, is proposed based on the transportation of magnetic field from a companion into an accretion disk around a black hole (BH). This model consists of a truncated thin disk with an inner advection-dominated accretion flow (ADAF). The spectral profiles of the sources are fitted in agreement with the data observed at four different dates corresponding to the rising phase of the LHS. In addition, the association of the LHS with quasi-steady jet is modelled based on transportation of magnetic field, where the Blandford-Znajek (BZ) and Blandford-Payne (BP) processes are invoked to drive the jets from BH and inner ADAF. It turns out that the steep radio/X-ray correlations observed in H1743-322 and GX 339-4 can be interpreted based on our model. ",A magnetic model for low/hard state of black hole binaries " We studied the emergence process of 42 active region (ARs) by analyzing the time derivative, R(t), of the total unsigned flux. Line-of-sight magnetograms acquired by the Helioseismic and Magnetic Imager (HMI) onboard the Solar Dynamics Observatory (SDO) were used. A continuous piecewise linear fitting to the R(t)-profile was applied to detect an interval, dt_2, of nearly-constant R(t) covering one or several local maxima. The averaged over dt_2 magnitude of R(t) was accepted as an estimate of the maximal value of the flux growth rate, R_MAX, which varies in a range of (0.5-5)x10^20 Mx hour^-1 for active regions with the maximal total unsigned flux of (0.5-3)x10^22 Mx. The normalized flux growth rate, R_N, was defined under an assumption that the saturated total unsigned flux, F_MAX, equals unity. Out of 42 ARs in our initial list, 36 event were successfully fitted and they form two subsets (with a small overlap of 8 events): the ARs with a short (<13 hours) interval dt_2 and a high (>0.024 hour^-1) normalized flux emergence rate, R_N, form the ""rapid"" emergence event subset. The second subset consists of ""gradual"" emergence events and it is characterized by a long (>13 hours) interval dt_2 and a low R_N (<0.024 hour^-1). In diagrams of R_MAX plotted versus F_MAX, the events from different subsets are not overlapped and each subset displays an individual power law. The power law index derived from the entire ensemble of 36 events is 0.69+-0.10. The ""rapid"" emergence is consistent with a ""two-step"" emergence process of a single twisted flux tube. The ""gradual"" emergence is possibly related to a consecutive rising of several flux tubes emerging at nearly the same location in the photosphere. ",Analysis of the flux growth rate in emerging active regions on the Sun " The aim of this paper is to find the exact solutions of the Schrodinger equation. As is known, the Schrodinger equation can be reduced to the continuum equation. In this paper, using the non-linear Legendre transform the equation of continuity is linearized. Particular solutions of such a linear equation are found in the paper and an inverse Legendre transform is considered for them with subsequent construction of solutions of the Schrodinger equation. Examples of the classical and quantum systems are considered. ",A new class of exact solutions of the Schrodinger equation " Twitter is one of the popular social media platforms where people share news or reactions towards an event or topic using short text messages called ""tweets"". Emotion analysis in these tweets can play a vital role in understanding peoples' feelings towards the underlying event or topic. In this work, we present our visual analytics tool, called TECVis, that focuses on providing comparison views of peoples' emotion feelings in tweets towards an event or topic. The comparison is done based on geolocations or timestamps. TECVis provides several interaction and filtering options for navigation and better exploration of underlying tweet data for emotion feelings comparison. ",TECVis: A Visual Analytics Tool to Compare People's Emotion Feelings " The problem of computing Craig Interpolants has recently received a lot of interest. In this paper, we address the problem of efficient generation of interpolants for some important fragments of first order logic, which are amenable for effective decision procedures, called Satisfiability Modulo Theory solvers. We make the following contributions. First, we provide interpolation procedures for several basic theories of interest: the theories of linear arithmetic over the rationals, difference logic over rationals and integers, and UTVPI over rationals and integers. Second, we define a novel approach to interpolate combinations of theories, that applies to the Delayed Theory Combination approach. Efficiency is ensured by the fact that the proposed interpolation algorithms extend state of the art algorithms for Satisfiability Modulo Theories. Our experimental evaluation shows that the MathSAT SMT solver can produce interpolants with minor overhead in search, and much more efficiently than other competitor solvers. ",Efficient Generation of Craig Interpolants in Satisfiability Modulo Theories " The central levels problem asserts that the subgraph of the $(2m+1)$-dimensional hypercube induced by all bitstrings with at least $m+1-\ell$ many 1s and at most $m+\ell$ many 1s, i.e., the vertices in the middle $2\ell$ levels, has a Hamilton cycle for any $m\geq 1$ and $1\le \ell\le m+1$. This problem was raised independently by Buck and Wiedemann, Savage, by Gregor and \v{S}krekovski, and by Shen and Williams, and it is a common generalization of the well-known middle levels problem, namely the case $\ell=1$, and classical binary Gray codes, namely the case $\ell=m+1$. In this paper we present a general constructive solution of the central levels problem. Our results also imply the existence of optimal cycles through any sequence of $\ell$ consecutive levels in the $n$-dimensional hypercube for any $n\ge 1$ and $1\le \ell \le n+1$. Moreover, extending an earlier construction by Streib and Trotter, we construct a Hamilton cycle through the $n$-dimensional hypercube, $n\geq 2$, that contains the symmetric chain decomposition constructed by Greene and Kleitman in the 1970s, and we provide a loopless algorithm for computing the corresponding Gray code. ",On the central levels problem " It has been proposed that the coherent propagation of long-lived heavy neutrino mass eigenstates can lead to an oscillating rate of lepton number conserving (LNC) and violating (LNV) events, as a function of the distance between the production and displaced decay vertices. We discuss this phenomenon, which we refer to as heavy neutrino-antineutrino oscillations, in the framework of quantum field theory (QFT), using the formalism of external wave packets. General formulae for the oscillation probabilities and the number of expected events are derived and the coherence and localisation conditions that have to be satisfied in order for neutrino-antineutrino oscillations to be observable are discussed. The formulae are then applied to a low scale seesaw scenario, which features two nearly mass degenerate heavy neutrinos that can be sufficiently long lived to produce a displaced vertex when their masses are below the $W$ boson mass. The leading and next-to-leading order oscillation formulae for this scenario are derived. For an example parameter point used in previous studies, the kinematics of the considered LNC/LNV processes are simulated, to check that the coherence and localisation conditions are satisfied. Our results show that the phenomenon of heavy neutrino-antineutrino oscillations can indeed occur in low scale seesaw scenarios and that the previously used leading order formulae, derived with a plane wave approach, provide a good approximation for the considered example parameter point. ",Heavy Neutrino-Antineutrino Oscillations in Quantum Field Theory " Mathematical modelling of unemployment dynamics attempts to predict the probability of a job seeker finding a job as a function of time. This is typically achieved by using information in unemployment records. These records are right censored, making survival analysis a suitable approach for parameter estimation. The proposed model uses a deep artificial neural network (ANN) as a non-linear hazard function. Through embedding, high-cardinality categorical features are analysed efficiently. The posterior distribution of the ANN parameters are estimated using a variational Bayes method. The model is evaluated on a time-to-employment data set spanning from 2011 to 2020 provided by the Slovenian public employment service. It is used to determine the employment probability over time for each individual on the record. Similar models could be applied to other questions with multi-dimensional, high-cardinality categorical data including censored records. Such data is often encountered in personal records, for example in medical records. ",Variational Bayes survival analysis for unemployment modelling " We investigate the method for constructing the invariant mass using the M_T2-assisted on-shell (MAOS) approximation to the invisible particle momenta in the cascade decays of a new particle resonance produced at hadron colliders. We note that the MAOS reconstruction can be defined in several different ways, while keeping the efficiency of approximation at a similar level, and one of them provides a unique solution for each event. It is shown that the invariant mass distribution constructed with the MAOS momenta exhibits a peak at the heavy resonance mass, regardless of the chosen MAOS scheme and the detailed mass spectrum of the particles in the cascade. We stress that the MAOS invariant mass can be used as a clean signal of new particle resonance produced at the LHC, as well as a model-independent method to measure the masses of new particles involved in the event. ",Reconstructing the heavy resonance at hadron colliders " Interface modes have been widely explored in the field of electronics, optics, acoustics and nanophononics. One strategy to generate them is band inversion in one-dimensional superlattices. Most realizations of this type of topological states have so far been explored for a single kind of excitation. Despite its potential in the manipulation and engineering of interactions, platforms for the simultaneous topological confinement of multiple excitations remain an open challenge. GaAs/AlAs heterostructures exhibit enhanced optomechanical interactions due to the intrinsic colocalization of light and sound. In this work, we designed, fabricated, and experimentally studied a multilayered structure based on GaAs/AlAs. Due to the simultaneously inverted band structures for light and phonons, colocalized interface modes for both 1.34 eV photons and 18 GHz phonons appear. We experimentally validated the concept by optical reflectivity and coherent phonon generation and detection. Furthermore, we theoretically analyzed the performance of different topological designs presenting colocalized states in time-domain Brillouin scattering and deduce engineering rules. Potential future applications include the engineering of robust optomechanical resonators, compatible with the incorporation of active media such as quantum wells and quantum dots. ",Topological optical and phononic interface mode by simultaneous band inversion " In the geolocation field where high-level programs and low-level devices coexist, it is often difficult to find a friendly user inter- face to configure all the parameters. The challenge addressed in this paper is to propose intuitive and simple, thus natural lan- guage interfaces to interact with low-level devices. Such inter- faces contain natural language processing and fuzzy represen- tations of words that facilitate the elicitation of business-level objectives in our context. ",Dealing with natural language interfaces in a geolocation context " We use an optical cavity in the regime of intermediate coupling between atom and cavity mode to detect single moving atoms. Degenerate polarization modes allow excitation of the atoms in one mode and collection of spontaneous emission in the other, while keeping separate the two sources of light; we obtain a higher confidence and efficiency of detection by adding cavity-enhanced Faraday rotation. Both methods greatly benefit from coincidence detection of photons, attaining fidelities in excess of 99% in less than 1 microsecond. Detailed studies of the second-order intensity autocorrelation function of light from the signal mode reveal evidence of antibunched photon emissions and the dynamics of single-atom transits. ",Atom detection in a two-mode optical cavity with intermediate coupling: Autocorrelation studies " Many features of spin models can be interpreted in geometrical terms by means of the properties of well defined clusters of spins. In case of spontaneous symmetry breaking, the phase transition of models like the q-state Potts model, O(n), etc., can be equivalently described as a percolation transition of clusters. We study here the behaviour of such clusters when the presence of an external field H breaks explicitly the global symmetry of the Hamiltonian of the theory. We find that these clusters have still some interesting relationships with thermal features of the model. ",Cluster Percolation and Explicit Symmetry Breaking in Spin Models " The 8-\textit{Pmmn} borophene is one kind of new elemental monolayer, which hosts anisotropic and tilted massless Dirac fermions (MDF). The planar \textit{p-n} junction (PNJ) structure as the basic component of various novel devices based on the monolayer material has attracted increasing attention. Here, we analytically study the transport properties of anisotropic and tilted MDF across 8-\textit{Pmmn} borophene PNJ. Similar to the isotropic MDF across graphene junctions, perfect transmission exists but its direction departures the normal direction of borophene PNJ induced by the anisotropy and tilt, i.e., oblique Klein tunneling. The oblique Klein tunneling does not depend on the doping levels in \textit{N} and \textit{P} regions of PNJ as the normal Klein tunneling but depends on the junction direction. Furthermore, we analytically derive the special junction direction for the maximal difference between perfect transmission direction and the normal direction of PNJ and clearly distinguish the respective contribution of anisotropy and tilt underlying the oblique Klein tunneling. In light of the rapid advances of experimental technologies, we expect the oblique Klein tunneling to be observable in the near future. ",Oblique Klein tunneling in 8-Pmmn borophene p-n junctions " It is known that, in string sigma-model metric, the `extreme' fivebrane solution of D=10 supergravity interpolates between D=10 Minkowski spacetime and a supersymmetric $S^3$ compactification to a linear dilaton vacuum. We show here that, in {\it fivebrane} sigma-model metric, the extreme string solution of D=10 supergravity interpolates between Minkowski spacetime and a hitherto unknown supersymmetric $S^7$ compactification of d=10 supergravity to a three-dimensional anti-de Sitter generalization of the linear dilaton vacuum, which may be invariantly characterized in terms of conformal Killing vectors. The dilaton field diverges near the string core but this divergence may be eliminated by re-interpreting the string solution as the extreme membrane solution of 11-dimensional supergravity. We show that the latter has an analytic extension through a regular degenerate event horizon to an interior region containing a curvature singularity. We obtain analogous results for other extended object solutions of supergravity theories. ",Macroscopic superstrings as interpolating solitons " Recent experiments on the Co-based $d^7$ honeycomb materials BaCo$_2$(AsO$_4$)$_2$ and BaCo$_2$(PO$_4$)$_2$ have drawn renewed interest to XXZ models with competing interactions beyond the nearest neighbor exchange. While the case of dominant antiferromagnetic exchange has been extensively studied, the actual materials exhibit a ferromagnetic nearest neighbor exchange. Here we show that such a sign change of the dominant nearest neighbor coupling has substantial consequences on the phase diagram. In particular, the nature of the quantum disordered phase of interest changes from a plaquette valence bond crystal to a long-range entangled spin liquid phase. By means of complementary numerical simulations, based on exact diagonalization and the pseudo-fermion functional renormalization group, we demonstrate the formation of a gapless spin liquid state at the heart of the ferromagnetic phase diagram in the isotropic Heisenberg limit, which gives way to out-of-plane Ising magnetic order upon inclusion of XXZ anisotropy. The magnetically ordered phases surrounding this quantum disordered region are much less sensitive to the sign change of the dominant interaction and can be understood from their classical analogs. We briefly comment on the relevance of our results for Co-based $d^7$ honeycomb materials. ","Frustrated Ferromagnetism of Honeycomb Cobaltates: Incommensurate Spirals, Quantum Disordered Phases, and Out-of-Plane Ising Order" " The formation of planets in compact or highly eccentric binaries and the migration of hot Jupiters are two outstanding problems in planet formation. The hot Jupiter $\tau$ Boo Ab orbits the primary star in the long-period ($P \gtrsim 1000\,$yr), highly eccentric ($e \sim 0.9$) binary $\tau$ Bo\""otis. Due to the long orbital period, the orbit of the stellar binary is poorly constrained. Here we aim to constrain the orbit of the stellar binary $\tau$ Boo AB in order to investigate the formation and migration history of the system. The mutual orbital inclination of the stellar companion and the hot Jupiter has important implications for planet migration. The binary eccentricity and periastron distance are important for understanding the conditions under which $\tau$ Boo Ab formed. We combine more than 150 years of astrometric data with twenty-five years of high precision radial velocities. The combination of sky-projected and line-of-sight measurements places tight constraints on the orbital inclination, eccentricity, and periastron distance of $\tau$ Boo AB. We find an orbital inclination of $47.2^{+2.7}_{-3.7}\,$deg, periastron distance of $28.3^{+2.3}_{-3.0}\,$au and eccentricity of $0.87^{+0.04}_{-0.03}$. We find that the orbital inclinations of $\tau$ Boo Ab and $\tau$ Boo B, as well as the stellar spin-axis of $\tau$ Boo A coincide at $\sim \! 45$ degrees, a result consistent with the assumption of a well-aligned, coplanar system. The likely aligned, coplanar configuration suggests planetary migration within a well-aligned protoplanetary disc. Due to the high eccentricity and small periastron distance of $\tau$ Boo B, the protoplanetary disc was tidally truncated at $\approx \! 6\,$au. We suggest that $\tau$ Boo Ab formed near the edge of the truncated disc and migrated inwards with high eccentricity due to spiral waves generated by the stellar companion. ","Constraining the orbit of the planet-hosting binary $\tau$ Bo\""otis: Clues about planetary formation and migration" " We investigate the most interesting decay processes involving axions, photons and the lightest pseudoscalar mesons, making use of a chiral effective Lagrangian model with $L=3$ light quark flavors, which also includes the flavor-singlet pseudoscalar meson and implements the $U(1)$ axial anomaly of the fundamental theory. In particular, we compute the electromagnetic coupling of the axion to photons and we compare our result with the prediction of the Chiral Effective Lagrangian with $L=2$ light quark flavors. Moreover, we study the decay channels $\eta/\eta'\rightarrow \pi\pi a$ and we estimate the corresponding decay widths, using the existing bounds on the $U(1)_{PQ}$ breaking scale. ",Study of the interactions of the axion with mesons and photons using a chiral effective Lagrangian model " We investigate metric learning in the context of dynamic time warping (DTW), the by far most popular dissimilarity measure used for the comparison and analysis of motion capture data. While metric learning enables a problem-adapted representation of data, the majority of methods has been proposed for vectorial data only. In this contribution, we extend the popular principle offered by the large margin nearest neighbors learner (LMNN) to DTW by treating the resulting component-wise dissimilarity values as features. We demonstrate that this principle greatly enhances the classification accuracy in several benchmarks. Further, we show that recent auxiliary concepts such as metric regularization can be transferred from the vectorial case to component-wise DTW in a similar way. We illustrate that metric regularization constitutes a crucial prerequisite for the interpretation of the resulting relevance profiles. ",Efficient Metric Learning for the Analysis of Motion Data " Persistent homology is one of the most active branches of Computational Algebraic Topology with applications in several contexts such as optical character recognition or analysis of point cloud data. In this paper, we report on the formal development of certified programs to compute persistent Betti numbers, an instrumental tool of persistent homology, using the Coq proof assistant together with the SSReflect extension. To this aim it has been necessary to formalize the underlying mathematical theory of these algorithms. This is another example showing that interactive theorem provers have reached a point where they are mature enough to tackle the formalization of nontrivial mathematical theories. ",Computing Persistent Homology within Coq/SSReflect " We consider the modifications to squark production in the presence of a naturally heavier Dirac gluino. First generation squark production is highly suppressed, providing an interesting but challenging signal find or rule out. No dedicated searches for supersymmetry with a Dirac gluino have been performed, however a reinterpretation of a ""decoupled gluino"" simplified model suggests the bounds on a common first and second generation squark mass is much smaller than in the MSSM: $\lsim 850$ GeV for a massless LSP, and no bound for an LSP heavier than about 300 GeV. We compare and contrast the squark production cross sections between a model with a Dirac gluino and one with a Majorana gluino, updating earlier results in the literature to a $pp$ collider operating at $\sqrt{s} = 14$ and 33 TeV. Associated production of squark+gluino is likely very small at $\sqrt{s} = 14$ TeV, while is a challenging but important signal at even higher energy $pp$ colliders. Several other salient implications of Dirac gauginos are mentioned, with some thought-provoking discussion as it regards the importance of the various experiments planned or proposed. ",Dirac Gauginos in Supersymmetry -- Suppressed Jets + MET Signals: A Snowmass Whitepaper " As in-space exploration increases, autonomous systems will play a vital role in building the necessary facilities to support exploration. To this end, an autonomous system must be able to assign tasks in a scheme that efficiently completes all of the jobs in the desired project. This research proposes a flexible job shop problem (FJSP) representation to characterize an autonomous assembly project and then proposes both a mixed integer programming (MIP) solution formulation and a reinforcement learning (RL) solution formulation. The MIP formulation encodes all of the constraints and interjob dynamics a priori and was able to solve for the optimal solution to minimize the makespan. The RL formulation did not converge to an optimal solution but did successfully learn implicitly interjob dynamics through interaction with the reward function. Future work will include developing a solution formulation that utilizes the strengths of both proposed solution methods to handle scaling in size and complexity. ",A Flexible Job Shop Scheduling Representation of the Autonomous In-Space Assembly Task Assignment Problem " We study the nonequilibrium properties of a nonergodic random quantum chain in which highly excited eigenstates exhibit critical properties usually associated with quantum critical ground states. The ground state and excited states of this system belong to different universality classes, characterized by infinite-randomness quantum critical behavior. Using strong-disorder renormalization group techniques, we show that the crossover between the zero and finite energy density regimes is universal. We analytically derive a flow equation describing the unitary dynamics of this isolated system at finite energy density from which we obtain universal scaling functions along the crossover. ",Universal crossover from ground state to excited-state quantum criticality " Near-term applications of quantum information processors will rely on optimized circuit implementations to minimize gate depth and therefore mitigate the impact of gate errors in noisy intermediate-scale quantum (NISQ) computers. More expressive gate sets can significantly reduce the gate depth of generic circuits. Similarly, structured algorithms can benefit from a gate set that more directly matches the symmetries of the problem. The XY interaction generates a family of gates that provides expressiveness well tailored to quantum chemistry as well as to combinatorial optimization problems, while also offering reductions in circuit depth for more generic circuits. Here we implement the full family of XY entangling gates in a transmon-based superconducting qubit architecture. We use a composite pulse scheme that requires calibration of only a single gate pulse and maintains constant gate time for all members of the family. This allows us to maintain a high fidelity implementation of the gate across all entangling angles. The average fidelity of gates sampled from this family ranges from $95.67 \pm 0.60\%$ to $99.01 \pm 0.15\%$, with a median fidelity of $97.35 \pm 0.17\%$, which approaches the coherence-limited gate fidelity of the qubit pair. We furthermore demonstrate the utility of XY in a quantum approximation optimization algorithm in enabling circuit depth reductions as compared to the CZ only case. ",Implementation of the XY interaction family with calibration of a single pulse " Although the intrinsic conductance of an interacting one-dimensional system is renormalized by the electron-electron correlations, it has been known for some time that this renormalization is washed out by the presence of the (non-interacting) electrodes to which the wire is connected. Here, we study the transient conductance of such a wire: a finite voltage bias is suddenly applied across the wire and we measure the current before it has enough time to reach its stationary value. These calculations allow us to extract the Sharvin (contact) resistance of Luttinger and Fermi liquids. In particular, we find that a perfect junction between a Fermi liquid electrode and a Luttinger liquid electrode is characterized by a contact resistance that consists of half the quantum of conductance in series with half the intrinsic resistance of an infinite Luttinger liquid. These results were obtained using two different methods: a dynamical Hartree-Fock approach and a self-consistent Boltzmann approach. Although these methods are formally approximate we find a perfect match with the exact results of Luttinger/Fermi liquid theory. ",Transient and Sharvin resistances of Luttinger liquids " We prove that the rational function semifield of a tropical curve is finitely generated as a semifield over the tropical semifield $\boldsymbol{T} := ( \boldsymbol{R} \cup \{ - \infty \}, \operatorname{max}, +)$ by giving a specific finite generating set. Also, we show that for a finite harmonic morphism between tropical curves $\varphi : \varGamma \to \varGamma^{\prime}$, the rational function semifield of $\varGamma$ is finitely generated as a $\varphi^{\ast}(\operatorname{Rat}(\varGamma^{\prime}))$-algebra, where $\varphi^{\ast}(\operatorname{Rat}(\varGamma^{\prime}))$ stands for the pull-back of the rational function semifield of $\varGamma^{\prime}$ by $\varphi$. ",Rational function semifields of tropical curves are finitely generated over the tropical semifield " Sound measurements on a sequence of related, similar constructions with slightly different dimensions confirm a simple picture of the air modes of the internal resonator banjo's body. For the purpose of this study, the air modes are decoupled from the soundboard (i.e., [drum] head) modes by replacing the head with 3/4"" plywood. The resulting characteristic features survive the strong coupling of the air modes to the head and are in accord with the qualitative distinctions recognized by banjo players. ",Air modes of the Bacon internal resonator banjo " In this study, we provide an alternatively reformulated interpretation of Gibbons-Hawking radiation as well as inflation. By using a spacetime quantization procedure, proposed recently by L.C. C\'eleri et al., in anti-de Sitter space we show that Gibbons-Hawking radiation is an intrinsic property of the concerned space, that arises due to the existence of a scalar field whose quanta ""carry"" a length $l$ (i.e. the radius of the hyperboloid curvature). Furthermore, within the context of Tsallis q-framework, we propose an inflationary model that depends on the non-extensive parameter $q$. The main source of such an inflation is the same scalar field mentioned before. Being constrained by the observational data, the q-parameter along with the rest of the model's parameters has been used to estimate the time at which inflation ends as well as the reheating temperature. The latter is found to be related to Gibbons-Hawking temperature. Thus, the present model offers an alternative perspective regarding the nature of the cosmic background radiation (CMB). ",Can Gibbons-Hawking Radiation and Inflation Arise Due to Spacetime Quanta? " Examining residuals such as Pearson and deviance residuals, is a standard tool for assessing normal regression. However, for discrete response, these residuals cluster on lines corresponding to distinct response values. Their distributions are far from normality; graphical and quantitative inspection of these residuals provides little information for model diagnosis. Marshall and Spiegelhalter (2003) defined a cross-validatory predictive p-value for identifying outliers. Predictive p-values are uniformly distributed for continuous response but not for discrete response. We propose to use randomized predictive p-values (RPP) for diagnosing models with discrete responses. RPPs can be transformed to ""residuals"" with normal distribution, called NRPPs by us. NRPPs can be used to diagnose all regression models with scalar response using the same way for diagnosing normal regression. The NRPPs are nearly the same as the randomized quantile residuals (RQR), which are previously proposed by Dunn and Smyth (1996) but remain little known by statisticians. This paper provides an exposition of RQR using the RPP perspective. The contributions of this exposition include: (1) we give a rigorous proof of uniformity of RPP and illustrative examples to explain the uniformity under the true model; (2) we conduct extensive simulation studies to demonstrate the normality of NRPPs under the true model; (3) our simulation studies also show that the NRPP method is a versatile diagnostic tool for detecting many kinds of model inadequacies due to lack of complexity. The effectiveness of NRPP is further demonstrated with a health utilization dataset. ",Randomized Predictive P-values: A Versatile Model Diagnostic Tool with Unified Reference Distribution " The fragility of a glassforming liquid characterizes how rapidly its relaxation dynamics slow down with cooling. The viscosity of strong liquids follows an Arrhenius law with a temperature-independent barrier height to rearrangements responsible for relaxation, whereas fragile liquids experience a much faster increase in their dynamics, suggesting a barrier height that increases with decreasing temperature. Strong glassformers are typically network glasses, while fragile glassformers are typically molecular or hard-sphere-like. As a result of these differences at the microscopic level, strong and fragile glassformers are usually treated separately from a theoretical point of view. Silica is the archetypal strong glassformer at low temperatures, but also exhibits a mysterious strong-to-fragile crossover at higher temperatures. Here we show that softness, a structure-based machine learned parameter that has previously been applied to fragile glassformers provides a useful description of model liquid silica in the strong and fragile regimes, and through the strong-to-fragile crossover. Just as for fragile glassformers, the relationship between softness and dynamics is invariant and Arrhenius in all regimes, but the average softness changes with temperature. The strong-to-fragile crossover in silica is not due to a sudden, qualitative change in structure, but can be explained by a simple Arrhenius form with a continuously and linearly changing local structure. Our results unify the study of liquid silica under a single simple conceptual picture. ",Unifying framework for strong and fragile liquids via machine learning: a study of liquid silica " Performance unpredictability is a major roadblock towards cloud adoption, and has performance, cost, and revenue ramifications. Predictable performance is even more critical as cloud services transition from monolithic designs to microservices. Detecting QoS violations after they occur in systems with microservices results in long recovery times, as hotspots propagate and amplify across dependent services. We present Seer, an online cloud performance debugging system that leverages deep learning and the massive amount of tracing data cloud systems collect to learn spatial and temporal patterns that translate to QoS violations. Seer combines lightweight distributed RPC-level tracing, with detailed low-level hardware monitoring to signal an upcoming QoS violation, and diagnose the source of unpredictable performance. Once an imminent QoS violation is detected, Seer notifies the cluster manager to take action to avoid performance degradation altogether. We evaluate Seer both in local clusters, and in large-scale deployments of end-to-end applications built with microservices with hundreds of users. We show that Seer correctly anticipates QoS violations 91% of the time, and avoids the QoS violation to begin with in 84% of cases. Finally, we show that Seer can identify application-level design bugs, and provide insights on how to better architect microservices to achieve predictable performance. ",Leveraging Deep Learning to Improve the Performance Predictability of Cloud Microservices " It is generally accepted that relatively more permanent (i.e., more temporally persistent) traits are more valuable for biometric performance than less permanent traits. Although this finding is intuitive, there is no current work identifying exactly where in the biometric analysis temporal persistence makes a difference. In this paper, we answer this question. In a recent report, we introduced the intraclass correlation coefficient (ICC) as an index of temporal persistence for such features. In that report, we also showed that choosing only the most temporally persistent features yielded superior performance in 12 of 14 datasets. Motivated by those empirical results, we present a novel approach using synthetic features to study which aspects of a biometric identification study are influenced by the temporal persistence of features. What we show is that using more temporally persistent features produces effects on the similarity score distributions that explain why this quality is so key to biometric performance. The results identified with the synthetic data are largely reinforced by an analysis of two datasets, one based on eye-movements and one based on gait. There was one difference between the synthetic and real data: In real data, features are intercorrelated, with the level of intercorrelation increasing with increasing ICC. This increasedhttps://www.overleaf.com/project/5e2b14694c5dc600017292e6 intercorrelation in real data was associated with an increase in the spread of the impostor similarity score distributions. Removing these intercorrelations for real datasets with a decorrelation step produced results which were very similar to that obtained with synthetic features. ",Why Temporal Persistence of Biometric Features is so Valuable for Classification Performance " The spontaneous formation of patterns in dynamical systems is a rich phenomenon that transcends scientific boundaries. Here, we report our observation of coupled optical-atomic pattern formation, which results in the creation of self-organized, multimode structures in free-space laser-driven cold atoms. We show that this process gives rise to spontaneous three-dimensional Sisyphus cooling even at very low light intensities and the emergence of self-organized atomic structures on both sub- and super-wavelength scales. ",Spontaneous emergence of free-space optical and atomic patterns " A plethora of recent research has proposed several automated methods based on machine learning (ML) and deep learning (DL) to detect cybersickness in Virtual reality (VR). However, these detection methods are perceived as computationally intensive and black-box methods. Thus, those techniques are neither trustworthy nor practical for deploying on standalone VR head-mounted displays (HMDs). This work presents an explainable artificial intelligence (XAI)-based framework VR-LENS for developing cybersickness detection ML models, explaining them, reducing their size, and deploying them in a Qualcomm Snapdragon 750G processor-based Samsung A52 device. Specifically, we first develop a novel super learning-based ensemble ML model for cybersickness detection. Next, we employ a post-hoc explanation method, such as SHapley Additive exPlanations (SHAP), Morris Sensitivity Analysis (MSA), Local Interpretable Model-Agnostic Explanations (LIME), and Partial Dependence Plot (PDP) to explain the expected results and identify the most dominant features. The super learner cybersickness model is then retrained using the identified dominant features. Our proposed method identified eye tracking, player position, and galvanic skin/heart rate response as the most dominant features for the integrated sensor, gameplay, and bio-physiological datasets. We also show that the proposed XAI-guided feature reduction significantly reduces the model training and inference time by 1.91X and 2.15X while maintaining baseline accuracy. For instance, using the integrated sensor dataset, our reduced super learner model outperforms the state-of-the-art works by classifying cybersickness into 4 classes (none, low, medium, and high) with an accuracy of 96% and regressing (FMS 1-10) with a Root Mean Square Error (RMSE) of 0.03. ",VR-LENS: Super Learning-based Cybersickness Detection and Explainable AI-Guided Deployment in Virtual Reality " We present a novel method for injecting temporality into entailment graphs to address the problem of spurious entailments, which may arise from similar but temporally distinct events involving the same pair of entities. We focus on the sports domain in which the same pairs of teams play on different occasions, with different outcomes. We present an unsupervised model that aims to learn entailments such as win/lose $\rightarrow$ play, while avoiding the pitfall of learning non-entailments such as win $\not\rightarrow$ lose. We evaluate our model on a manually constructed dataset, showing that incorporating time intervals and applying a temporal window around them, are effective strategies. ",Incorporating Temporal Information in Entailment Graph Mining " An ordered graph $H$ is a simple graph with a linear order on its vertex set. The corresponding Tur\'an problem, first studied by Pach and Tardos, asks for the maximum number $\text{ex}_<(n,H)$ of edges in an ordered graph on $n$ vertices that does not contain $H$ as an ordered subgraph. It is known that $\text{ex}_<(n,H) > n^{1+\varepsilon}$ for some positive $\varepsilon=\varepsilon(H)$ unless $H$ is a forest that has a proper 2-coloring with one color class totally preceding the other one. Making progress towards a conjecture of Pach and Tardos, we prove that $\text{ex}_<(n,H) =n^{1+o(1)}$ holds for all such forests that are ""degenerate"" in a certain sense. This class includes every forest for which an $n^{1+o(1)}$ upper bound was previously known, as well as new examples. Our proof is based on a density-increment argument. ",On the Tur\'an number of ordered forests " Starting with a mathematical boundary value problem for the magnetic vector potential in an axisymmetric cylindrical coordinate system, we derive a general solution for any arbitrary current distribution using the method of Green's functions. We use this to derive an analytic form for an hourglass magnetic field pattern created by electrical currents that are concentrated near (but not confined within) the equatorial plane of a cylindrical coordinate system. Our solution is not characterized by a cusp at the equatorial plane, as in previous solutions based on a current sheet. The pattern we derive provides a very good fit to hourglass magnetic field patterns emerging from three-dimensional numerical simulations of core formation, and can in principle be used for source-fitting of observed magnetic hourglass patterns. ",A Mathematical Model for an Hourglass Magnetic Field " CoDEx is a Mathematica package that calculates the Wilson Coefficients (WCs) corresponding to effective operators up to mass dimension-6. Once the part of the Lagrangian involving single as well as multiple degenerate heavy fields, belonging to some Beyond Standard Model (BSM) theory, is given, the package can then integrate out propagators from the tree as well as 1-loop diagrams of that BSM theory. It then computes the associated WCs up to 1-loop level, for two different bases: ""Warsaw"" and ""SILH"". CoDEx requires only very basic information about the heavy field(s), e.g., Colour, Isospin, Hyper-charge, Mass, and Spin. The package first calculates the WCs at the high scale (mass of the heavy field(s)). We then have an option to perform the renormalisation group evolutions (RGEs) of these operators in ""Warsaw"" basis, a complete one (unlike ""SILH""), using the anomalous dimension matrix. Thus, one can get all effective operators at the electro-weak scale, generated from any such BSM theory, containing heavy fields of spin: 0, 1/2, and 1. We have provided many example models (both here and in the package-documentation) that more or less encompass different choices of heavy fields and interactions. Relying on the status of the present day precision data, we restrict ourselves up to dimension-6 effective operators. This will be generalised for any dimensional operators in a later version. Site: https://effexteam.github.io/CoDEx ",CoDEx: Wilson coefficient calculator connecting SMEFT to UV theory " The $r$-index (Gagie et al., JACM 2020) represented a breakthrough in compressed indexing of repetitive text collections, outperforming its alternatives by orders of magnitude. Its space usage, $\mathcal{O}(r)$ where $r$ is the number of runs in the Burrows-Wheeler Transform of the text, is however larger than Lempel-Ziv and grammar-based indexes, and makes it uninteresting in various real-life scenarios of milder repetitiveness. In this paper we introduce the $sr$-index, a variant that limits the space to $\mathcal{O}(\min(r,n/s))$ for a text of length $n$ and a given parameter $s$, at the expense of multiplying by $s$ the time per occurrence reported. The $sr$-index is obtained by carefully subsampling the text positions indexed by the $r$-index, in a way that we prove is still able to support pattern matching with guaranteed performance. Our experiments demonstrate that the $sr$-index sharply outperforms virtually every other compressed index on repetitive texts, both in time and space, even matching the performance of the $r$-index while using 1.5--3.0 times less space. Only some Lempel-Ziv-based indexes achieve better compression than the $sr$-index, using about half the space, but they are an order of magnitude slower. ",A Fast and Small Subsampled R-index " Cygnus X-3, among the Galactic X-ray binaries, is one of the brightest radio sources. Radio image as well as spectral analysis shows the presence of outflow in the form of jets. The SED in the X-ray band shows a complicated structure and evolution. Here we review some recent results of the correlation of the long-term radio and X-ray emission, in the various states of the source, with emphasis on the low (hard) state. We compare the results with those of other common Galactic microquasars, and attempt to provide a consistent picture of the accretion-ejection mechanism in these types of binaries. ",The Radio : X-ray Correlation of Cygnus X-3 and other Galactic Microquasars " Atomic form factors are widely used for the characterization of targets and specimens, from crystallography to biology. By using recent mathematical results, here we derive an analytical expression for the atomic form factor within the independent particle model constructed from nonrelativistic screened hydrogenic wavefunctions. The range of validity of this analytical expression is checked by comparing the analytically obtained form factors with the ones obtained within the Hartee-Fock method. As an example, we apply our analytical expression for the atomic form factor to evaluate the differential cross section for Rayleigh scattering off neutral atoms. ",Analytical evaluation of atomic form factors: application to Rayleigh scattering " In this paper, we consider the problem of finding a minimum common partition of two strings. The problem has its application in genome comparison. As it is an NP-hard, discrete combinatorial optimization problem, we employ a metaheuristic technique, namely, MAX-MIN ant system to solve this problem. To achieve better efficiency we first map the problem instance into a special kind of graph. Subsequently, we employ a MAX-MIN ant system to achieve high quality solutions for the problem. Experimental results show the superiority of our algorithm in comparison with the state of art algorithm in the literature. The improvement achieved is also justified by standard statistical test. ",Solving the Minimum Common String Partition Problem with the Help of Ants " We propose a class of weighted $L_2$-type tests of fit to the Gamma distribution. Our novel procedure is based on a fixed point property of a new transformation connected to a Steinian characterization of the family of Gamma distributions. We derive the weak limits of the statistic under the null hypothesis and under contiguous alternatives. Further, we establish the global consistency of the tests and apply a parametric bootstrap technique in a Monte Carlo simulation study to show the competitiveness to existing procedures. ",A new characterization of the Gamma distribution and associated goodness of fit tests We show that systems with the structure introduced by Climenhaga and Thompson have dense intermediate pressures and dense intermediate entropies of ergodic measures. The result applies to the Ma\~n\'e diffeomorphisms. ,Denseness of intermediate pressures for systems with the Climenhaga-Thompson structures " Whenever an architect or a team of architects begins an architectural design, there are certain goals set to achieve. There are many factors involved in setting up goals for the architecture design such as type of the project, end user perspective, functional and non-functional requirements and so on. This paper reviews and further elaborates strategy for the usability characteristics of software architecture. Although user centered designs are tremendously gaining popularity, still in many design scenarios, usability is barely even considered as one of the primary goals. This work provides an opportunity to compare different strategies and evaluate their pros and cons. ",Usability as a Dominant Quality Attribute " In this paper, all possible orbits of test particles are investigated by using phase plane method in regular Hayward black hole space-time. Our results show that the time-like orbits are divided into four types: unstable circular orbits, separates stable orbits, stable hyperbolic orbits and elliptical orbits in regular Hayward black hole space-time. We find that the orbital properties vary with the change of $\ell$ (a convenient encoding of the central energy density $3/8\pi\ell^{2}$). If $\ell =\frac{1}{3}$ and $b < 3.45321$, the test particles which moving toward the black hole will definitely be plunging into the black hole. In addition, it is obtained that the innermost stable circular orbit happens at $r_{min}$ = 5.93055 for $b$ = 3.45321. ",Orbital motion of test particles in regular Hayward black hole space-time " We apply the vortex path model of critical currents to a comprehensive analysis of contemporary data on defect-engineered superconductors, showing that it provides a consistent and detailed interpretation of the experimental data for a diverse range of materials. We address the question of whether electron mass anisotropy plays a role of any consequence in determining the form of this data and conclude that it does not. By abandoning this false interpretation of the data, we are able to make significant progress in understanding the real origin of the observed behavior. In particular, we are able to explain a number of common features in the data including shoulders at intermediate angles, a uniform response over a wide angular range and the greater discrimination between individual defect populations at higher fields. We also correct several misconceptions including the idea that a peak in the angular dependence of the critical current is a necessary signature of strong correlated pinning, and conversely that the existence of such a peak implies the existence of correlated pinning aligned to the particular direction. The consistency of the vortex path model with the principle of maximum entropy is introduced. ",The interpretation of the field angle dependence of the critical current in defect-engineered superconductors " A write-once memory (wom) is a storage medium formed by a number of ``write-once'' bit positions (wits), where each wit initially is in a `0' state and can be changed to a `1' state irreversibly. Examples of write-once memories include SLC flash memories and optical disks. This paper presents a low complexity coding scheme for rewriting such write-once memories, which is applicable to general problem configurations. The proposed scheme is called the \emph{position modulation code}, as it uses the positions of the zero symbols to encode some information. The proposed technique can achieve code rates higher than state-of-the-art practical solutions for some configurations. For instance, there is a position modulation code that can write 56 bits 10 times on 278 wits, achieving rate 2.01. In addition, the position modulation code is shown to achieve a rate at least half of the optimal rate. ",Position Modulation Code for Rewriting Write-Once Memories " We focus on developing an effective Direction Of Arrival (DOA) estimation method for wideband sources based on the gridless sparse concept. Previous coherent methods have been designed by dividing wideband frequencies into a few subbands which are transferred to a reference subband using focusing matrices. In this work, as opposed to the previous techniques, we propose a convex optimization problem that leads to an accurate wideband DOA estimation method with no need for any focusing matrix. Moreover, in this method, no initial DOA estimates are required and it can be used for any arbitrary linear arrays. Numerical simulations show that in comparison to some well-known techniques, the proposed method generates outstanding accuracy and better robustness to noise. The effectiveness of the method is also verified in presence of close adjacent sources. ",Super-Resolution DOA Estimation for Wideband Signals using Arbitrary Linear Arrays without Focusing Matrices " The human brain exhibits a relatively stable spatiotemporal organization that supports brain function and can be manipulated via local brain stimulation. Such perturbations to local cortical dynamics are globally integrated by distinct neural systems. However, it remains unclear how and why local changes in neural activity affect large-scale system dynamics. Here, we briefly review empirical and computational studies addressing how localized perturbations affect brain activity. We then systematically analyze a model of large-scale brain dynamics, assessing how localized changes in brain activity at the different sites affect whole-brain dynamics. We find that local stimulation induces changes in brain activity that can be summarized by relatively smooth tuning curves, which relate a region's effectiveness as a stimulation site to its position within the cortical hierarchy. Our results also support the notion that brain hubs, operating in a slower regime, are more resilient to focal perturbations and critically contribute to maintain stability in global brain dynamics. In contrast, perturbations of peripheral regions, characterized by faster activity, have greater impact on functional connectivity. As a parallel with this region-level result, we also find that peripheral systems such as the visual and sensorimotor networks were more affected by local perturbations than high-level systems such as the cingulo-opercular network. Our results highlight the importance of a periphery-to-core hierarchy to determine the effect of local stimulation on the brain network. We also provide novel resources to orient empirical work aiming at manipulating functional connectivity using non-invasive brain stimulation. ",Mapping how local perturbations influence systems-level brain dynamics " I study the time-evolution of a particle prepared in the ground state of an infinite well after the latter is suddenly expanded. It turns out that the probability density $|\Psi(x, t)|^{2}$ shows up quite a surprising behaviour: for definite times, {\it plateaux} appear for which $|\Psi(x, t)|^{2}$ is constant on finite intervals for $x$. Elements of theoretical explanation are given by analyzing the singular component of the second derivative $\partial_{xx}\Psi(x, t)$. Analytical closed expressions are obtained for some specific times, which easily allow to show that, at these times, the density organizes itself into regular patterns provided the size of the box in large enough; more, above some critical time-dependent size, the density patterns are independent of the expansion parameter. It is seen how the density at these times simply results from a construction game with definite rules acting on the pieces of the initial density. ",Surprises in the suddenly-expanded infinite well " (abbreviated) We aim to determine the masses of the envelopes, disks, and central stars of young stellar objects (YSOs) in the Class I stage. We observed the embedded Class I objects IRS 63 and Elias 29 in the rho Ophiuchi star-forming region with the Submillimeter Array (SMA) at 1.1 mm. IRS 63 and Elias 29 are both clearly detected in the continuum, with peak fluxes of 459 resp. 47 mJy/beam. The continuum emission toward Elias 29 is clearly resolved, whereas IRS 63 is consistent with a point source down to a scale of 3 arcsec (400 AU). The SMA data are combined with single-dish data, and disk masses of 0.055 and >= 0.007 MSun and envelope masses of 0.058 and >= 0.058 MSun are empirically determined for IRS 63 and Elias 29, respectively. The disk+envelope systems are modelled with the axisymmetric radiative-transfer code RADMC, yielding disk and envelope masses that differ from the empirical results by factors of a few. HCO+ J = 3-2 is detected toward both sources, HCN J = 3-2 is not. The HCO+ position-velocity diagrams are indicative of Keplerian rotation. For a fiducial inclination of 30 degrees, we find stellar masses of 0.37 +/- 0.13 and 2.5 +/- 0.6 MSun for IRS 63 and Elias 29, respectively. We conclude that the sensitivity and spatial resolution of the SMA at 1.1 mm allow a good separation of the disks around Class I YSOs from their circumstellar envelopes and environments, and the spectral resolution makes it possible to resolve their dynamical structure and estimate the masses of the central stars. The ratios of the envelope and disk masses are found to be 0.2 and 6 for IRS 63 and Elias 29, respectively. This is lower than the values for Class 0 sources, which have Menv/Mdisk >= 10, suggesting that this ratio is a tracer of the evolutionary stage of a YSO. ","SMA observations of young disks: separating envelope, disk, and stellar masses in class I YSOs" " Neuroimaging-driven prediction of brain age, defined as the predicted biological age of a subject using only brain imaging data, is an exciting avenue of research. In this work we seek to build models of brain age based on functional connectivity while prioritizing model interpretability and understanding. This way, the models serve to both provide accurate estimates of brain age as well as allow us to investigate changes in functional connectivity which occur during the ageing process. The methods proposed in this work consist of a two-step procedure: first, linear latent variable models, such as PCA and its extensions, are employed to learn reproducible functional connectivity networks present across a cohort of subjects. The activity within each network is subsequently employed as a feature in a linear regression model to predict brain age. The proposed framework is employed on the data from the CamCAN repository and the inferred brain age models are further demonstrated to generalize using data from two open-access repositories: the Human Connectome Project and the ATR Wide-Age-Range. ",Interpretable brain age prediction using linear latent variable models of functional connectivity " As a means to better understanding manifolds with positive curvature, there has been much recent interest in the study of non-negatively curved manifolds which contain either a point or an open dense set of points at which all 2-planes have positive curvature. We study infinite families of biquotients defined by Eschenburg and Bazaikin from this viewpoint, together with torus quotients of $S^3 \x S^3$. ",On the curvature of biquotients " In this paper, we introduce a generalization of Balancing and Balancing-Lucas numbers. We describe some of their properties also we give the related matrix representation and divisibility properties. ",Generalized Balancing and Balancing-Lucas numbers " This paper presents the design, implementation, and evaluation of the PyTorch distributed data parallel module. PyTorch is a widely-adopted scientific computing package used in deep learning research and applications. Recent advances in deep learning argue for the value of large datasets and large models, which necessitates the ability to scale out model training to more computational resources. Data parallelism has emerged as a popular solution for distributed training thanks to its straightforward principle and broad applicability. In general, the technique of distributed data parallelism replicates the model on every computational resource to generate gradients independently and then communicates those gradients at each iteration to keep model replicas consistent. Despite the conceptual simplicity of the technique, the subtle dependencies between computation and communication make it non-trivial to optimize the distributed training efficiency. As of v1.5, PyTorch natively provides several techniques to accelerate distributed data parallel, including bucketing gradients, overlapping computation with communication, and skipping gradient synchronization. Evaluations show that, when configured appropriately, the PyTorch distributed data parallel module attains near-linear scalability using 256 GPUs. ",PyTorch Distributed: Experiences on Accelerating Data Parallel Training " Machine reading comprehension (MRC) poses new challenges over logical reasoning, which aims to understand the implicit logical relations entailed in the given contexts and perform inference over them. Due to the complexity of logic, logical relations exist at different granularity levels. However, most existing methods of logical reasoning individually focus on either entity-aware or discourse-based information but ignore the hierarchical relations that may even have mutual effects. In this paper, we propose a holistic graph network (HGN) which deals with context at both discourse level and word level, as the basis for logical reasoning, to provide a more fine-grained relation extraction. Specifically, node-level and type-level relations, which can be interpreted as bridges in the reasoning process, are modeled by a hierarchical interaction mechanism to improve the interpretation of MRC systems. Experimental results on logical reasoning QA datasets (ReClor and LogiQA) and natural language inference datasets (SNLI and ANLI) show the effectiveness and generalization of our method, and in-depth analysis verifies its capability to understand complex logical relations. ",Modeling Hierarchical Reasoning Chains by Linking Discourse Units and Key Phrases for Reading Comprehension " Identifying weakly coupled nuclear spins around single electron spins is a key step of implementing quantum information processing using coupled electron-nuclei spin systems or sensing like single spin nuclear magnetic resonance detection using diamond defect spins. Dynamical decoupling control of the center electron spin with periodic pulse sequences [e.g., the Carre-Purcell-Meiboom-Gill (CPMG) sequence] has been successfully used to identify single nuclear spins and to resolve structure of nuclear spin clusters. Here, we design a new type of pulse sequences by replacing the repetition unit (a single $\pi$-pulse) of the CPMG sequence with a group of nonuniformly-spaced $\pi$-pulses. Using nitrogen-vacancy center system in diamond, we show that the designed pulse sequence improves the resolution of nuclear spin noise spectroscopy, and more information about the surrounding nuclear spins is extracted. The principle of dynamical decoupling design proposed in this paper is useful in many systems (e.g., defect spin qubit in solids, trapped ion and superconducting qubit) for high-resolution noise spectroscopy. ",Dynamical decoupling design for identifying weakly coupled nuclear spins in a bath " We extend the Bardakci-Thorn (BT) worldsheet formalism to supersymmetric non-abelian gauge theory. Our method covers the cases of N =1,2,4 extended supersymmetry. This task requires the introduction of spinor valued Grassmann variables on the worldsheet analogous to those of the supersymmetric formulation of superstring theory. As in the pure Yang-Mills case, the worldsheet formalism automatically generates the correct quartic vertices from the cubic vertices. We also discuss coupling renormalization to one loop order. ",BT Worldsheet for Supersymmetric Gauge Theories " Through a combination of monitoring the Raman spectral characteristics of 2D materials grown on copper catalyst layers, and wafer scale automated detection of the fraction of transferred material, we reproducibly achieve transfers with over 97.5% monolayer hexagonal boron nitride and 99.7% monolayer graphene coverage, for up to 300 mm diameter wafers. We find a strong correlation between the transfer coverage obtained for graphene and the emergence of a lower wavenumber 2D- peak component, with the concurrent disappearance of the higher wavenumber 2D+ peak component during oxidation of the catalyst surface. The 2D peak characteristics can therefore act as an unambiguous predictor of the success of the transfer. The combined monitoring and transfer process presented here is highly scalable and amenable for roll-to-roll processing. ",Raman Spectral Indicators of Catalyst Decoupling for Transfer of CVD Grown 2D Materials " The ARGO-YBJ experiment will be installed at YangBaJing Cosmic Ray Laboratory (Tibet, P.R. China), 4300 m a.s.l. . It consists of a full coverage of about 10^4 m^2 realized with RPC chambers. A small carpet of about 50 m^2 has been operated at YangBaJing in order to check the RPC performance in these high altitude conditions. Results concerning efficiency and time resolution are reported. ",Performance of RPCs operated at the YangBaJing Laboratory Lagragian quantization rules for general gauge theories are proposed on a basis of a superfield formulation of the standard BRST symmetry. Independence of the $S$-matrix on a choice of the gauge is proved. The Ward identities in terms of superfields are derived. ,Superfield formulation of the Lagrangian BRST quantization method " Let S=K[x_1,..., x_n], let A,B be finitely generated graded S-modules, and let m=(x_1,...,x_n). We give bounds for the Castelnuovo-Mumford regularity of the local cohomology of Tor_i(A,B) under the assumption that the Krull dimension of Tor_1(A,B) is at most 1. We apply the results to syzygies, Groebner bases, products and powers of ideals, and to the relationship of the Rees and Symmetric algebras. For example we show that any homogeneous linearly presented m-primary ideal has some power equal to a power of m; and if the first (roughly) (n-1)/2 steps of the resolution of I are linear, then I^2 is a power of m. ",The Regularity of Tor and Graded Betti Numbers " Traditional methods for inference in change point detection often rely on a large number of observed data points and can be inaccurate in non-asymptotic settings. With the rise of mobile health and digital phenotyping studies, where patients are monitored through the use of smartphones or other digital devices, change point detection is needed in non-asymptotic settings where it may be important to identify behavioral changes that occur just days before an adverse event such as relapse or suicide. Furthermore, analytical and computationally efficient means of inference are necessary for the monitoring and online analysis of large-scale digital phenotyping cohorts. We extend the result for asymptotic tail probabilities of the likelihood ratio test to the multivariate change point detection setting, and demonstrate through simulation its inaccuracy when the number of observed data points is not large. We propose a non-asymptotic approach for inference on the likelihood ratio test, and compare the efficiency of this estimated p-value to the popular empirical p-value obtained through simulation of the null distribution. The accuracy and power of this approach relative to competing methods is demonstrated through simulation and through the detection of a change point in the behavior of a patient with schizophrenia in the week prior to relapse. ",Non-asymptotic inference for multivariate change point detection " We construct embeddings G of the category of graphs into categories of R-modules over a commutative ring R which are almost full in the sense that the maps induced by the functoriality of G R[Hom_Graphs(X,Y)] --> Hom_R(GX,GY) are isomorphisms. The symbol R[S] above denotes the free R-module with the basis S. This implies that, for any cotorsion-free ring R, the categories of R-modules are not less complicated than the category of graphs. A similar embedding of graphs into the category of vector spaces with four distinguished subspaces (over any field, e.g. F_2={0,1} is obtained). ",An axiomatic construction of an almost full embedding of the category of graphs into the category of R-objects " Asteroids and comets dissipate energy when they rotate about the axis different from the axis of the maximal moment of inertia. We show that the most efficient internal relaxation happens at the double frequency of body's tumbling. Therefore the earlier estimates that ignore double frequency input underestimate the internal relaxation in asteroids and comets. We show that the Earth seismological data may poorly represent acoustic properties of asteroids and comet as internal relaxation increases in the presence of moisture. At the same time owing to non-linearlity of inelastic relaxation small angle nutations can persist for very long time spans, but our ability to detect such precessions is limited by the resolution of the radar-generated images. Wobbling may provide valuable information on the composition and structure of asteroids and on their recent history of external impacts. ",Inelastic Dissipation in Wobbling Asteroids and Comets " We obtain a continuous Wick rotation for Dirac, Majorana and Weyl spinors $\psi \to \exp ({1\over 2} \theta \gamma^4 \gamma^5)\psi$ which interpolates between Minkowski and Euclidean field theories. ",A continuous Wick rotation for spinor fields and supersymmetry in Euclidean space " Tetrahedral frame fields have applications to certain classes of nematic liquid crystals and frustrated media. We consider the problem of constructing a tetrahedral frame field in three dimensional domains in which the boundary normal vector is included in the frame on the boundary. To do this we identify an isomorphism between a given tetrahedral frame and a symmetric, traceless third order tensor under a particular nonlinear constraint. We then define a Ginzburg-Landau-type functional which penalizes the associated nonlinear constraint. Using gradient descent, one retrieves a globally defined limiting tensor outside of a singular set. The tetrahedral frame can then be recovered from this tensor by a determinant maximization method, developed in this work. The resulting numerically generated frame fields are smooth outside of one dimensional filaments that join together at triple junctions. ",Tetrahedral frame fields via constrained third order symmetric tensors " We report results of a spectropolarimetric and photometric monitoring of the weak-line T Tauri star LkCa4 within the MaTYSSE programme, involving ESPaDOnS at the Canada-France-Hawaii Telescope. Despite an age of only 2Myr and a similarity with prototypical classical T Tauri stars, LkCa4 shows no evidence for accretion and probes an interesting transition stage for star and planet formation. Large profile distortions and Zeeman signatures are detected in the unpolarized and circularly-polarized lines of LkCa4 using Least-Squares Deconvolution (LSD), indicating the presence of brightness inhomogeneities and magnetic fields at the surface of LkCa4. Using tomographic imaging, we reconstruct brightness and magnetic maps of LkCa4 from sets of unpolarized and circularly-polarized LSD profiles. The large-scale field is strong and mainly axisymmetric, featuring a ~2kG poloidal component and a ~1kG toroidal component encircling the star at equatorial latitudes - the latter making LkCa4 markedly different from classical TTauri stars of similar mass and age. The brightness map includes a dark spot overlapping the magnetic pole and a bright region at mid latitudes - providing a good match to the contemporaneous photometry. We also find that differential rotation at the surface of LkCa4 is small, typically ~5.5x weaker than that of the Sun, and compatible with solid-body rotation. Using our tomographic modelling, we are able to filter out the activity jitter in the RV curve of LkCa4 (of full amplitude 4.3km/s) down to a rms precision of 0.055km/s. Looking for hot Jupiters around young Sun-like stars thus appears feasible, even though we find no evidence for such planets around LkCa4. ",Modelling the magnetic activity & filtering radial velocity curves of young Suns: the weak-line T Tauri star LkCa 4 " The core velocity dispersion (CVD) is a potentially useful tool for studying the turbulent velocity field of molecular clouds. CVD is based on centroid velocities of dense gas clumps, thus is less prone to density fluctuation and reflects more directly the cloud velocity field. Prior work demonstrated that the Taurus molecular cloud CVD resembles the well-known Larson's linewidth-size relation of molecular clouds. In this work, we studied the dependence of the CVD on the line-of-sight thickness of molecular clouds, a quantity which cannot be measured by direct means. We produced a simple statistical model of cores within clouds and analyzed the CVD of a variety of hydrodynamical simulations. We show that the relation between the CVD and the 2D projected separation of cores ($L_{2D}$) is sensitive to the cloud thickness. When the cloud is thin, the index of CVD-$L_{2D}$ relation ($\gamma$ in the relation CVD$\sim L_{2D}^{\gamma}$) reflects the underlying energy spectrum ($E(k)\sim k^{-\beta}$) in that $\gamma\sim(\beta-1)/2$. The CVD-$L_{2D}$ relation becomes flatter ($\gamma\to 0$) for thicker clouds. We used this result to constrain the thicknesses of Taurus, Perseus, and Ophiuchus. We conclude that Taurus has a ratio of cloud depth to cloud length smaller than about 1/10-1/8, i.e. it is a sheet. A simple geometric model fit to the linewidth-size relation indicates that the Taurus cloud has a $\sim 0.7$ pc line-of-sight dimension. In contrast, Perseus and Ophiuchus are thicker and have ratios of cloud depth to cloud length larger than about 1/10-1/8. ","A New Method for Constraining Molecular Cloud Thickness: A study of Taurus, Perseus and Ophiuchus" " The nonlocal nature of the Majorana zero modes implies an inherent teleportation channel and unique transport signatures for Majorana identification. In this work we make an effort to eliminate some inconsistencies between the Bogoliubov-de Gennes equation based treatment and the method using the associated regular fermion number states of vacancy and occupation within the `second quantization' framework. We first consider a rather simple `quantum dot--Majorana wire--quantum dot' system, then a more experimentally relevant setup by replacing the quantum dots with transport leads. For the latter setup, based on the dynamical evolution of electron-hole excitations, we develop a single-particle-wavefunction approach to quantum transport, which renders both the conventional quantum scattering theory and the steady-state nonequilibrium Green's function formalism as its stationary limit. Further, we revisit the issue of Majorana tunneling spectroscopy and consider in particular the two-lead coupling setup. We present comprehensive discussions with detailed comparisons, and predict a zero-bias-limit conductance of $e^2/h$ (for symmetric coupling to the leads),which is a half of the popular result of the zero-bias-peak, or, the so-called Majorana quantized conductance ($2e^2/h$). The present work may arouse a need to reexamine some existing studies and the proposed treatment is expected to be involved in analyzing future experiments in this fast developing field. ",Revisit the non-locality of Majorana zero modes and teleportation: Bogoliubov-de Gennes equation based treatment " We undertake a general study of the boundary (or edge) modes that arise in gauge and gravitational theories defined on a space with boundary, either asymptotic or at finite distance, focusing on efficient techniques for computing the corresponding boundary action. Such actions capture all the dynamics of the system that are implied by its asymptotic symmetry group, such as correlation functions of the corresponding conserved currents. Working in the covariant phase space formalism, we develop a collection of approaches for isolating the boundary modes and their dynamics, and illustrate with various examples, notably AdS$_3$ gravity (with and without a gravitational Chern-Simons terms) subject to assorted boundary conditions. ",Systematics of Boundary Actions in Gauge Theory and Gravity " This paper analyzes the performance of a multiuser integrated sensing and communications (ISAC) system, where nonorthogonal multiple access (NOMA) is exploited to mitigate inter-user interference. Closed-form expressions are derived to evaluate the outage probability, ergodic communication rate, and sensing rate. Furthermore, asymptotic analyses are carried out to unveil diversity orders and high signal-to-noise ratio (SNR) slopes of the considered NOMA-ISAC system. As the further advance, the achievable sensing-communication rate region of ISAC is characterized. It is proved that ISAC system is capable of achieving a larger rate region than the conventional frequency-division sensing and communications (FDSAC) system. ",NOMA-ISAC: Performance Analysis and Rate Region Characterization " This paper presents a framework to solve the strategic bidding problem of participants in an electricity market cleared by employing the full AC Optimal Power Flow (ACOPF) problem formulation. Traditionally, the independent system operators (ISOs) leveraged DC Optimal Power Flow (DCOPF) problem formulation to settle the electricity market. The main quest of this work is to find what would be the challenges and opportunities if ISOs leverage the full ACOPF as the market-clearing Problem (MCP)? This paper presents tractable mathematical programming with equilibrium constraints for the convexified AC market-clearing problem. Market participants maximize their profit via strategic bidding while considering the reactive power dispatch of generation units. The equilibrium constraints are procured by presenting the dual form of the relaxed ACOPF problem. The strategic bidding problem with ACOPF-based MCP improves the exactness of the location marginal prices (LMPs) and profit of market participants compared to the one with DCOPF. It is shown that the strategic bidding problem with DCOFP-based MCP is unable to model the limitations of reactive power support. The presented results display cases where the proposed strategic bidding method renders $52.3\%$ more profit for the Generation Company (GENCO) than the DCOPF-based MCP model. The proposed strategic bidding framework also addresses the challenges in coupling real and reactive power dispatch of generation constraints, ramping constraints, demand response implications with curtailable and time shiftable loads, and AC line flow constraints. Therefore, the presented method will help market participants leverage the more accurate ACOPF model in the strategic bidding problem. ",Strategic Bidding in Electricity Markets with Convexified AC Market-Clearing Process " Recent simulations have shown that a high-energy proton bunch can excite strong plasma wakefields and accelerate a bunch of electrons to the energy frontier in a single stage of acceleration. It therefore paves the way towards a compact future collider design using the proton beams from existing high-energy proton machines, e.g. Tevatron or the LHC. This paper addresses some key issues in designing a compact electron-positron linear collider and an electron-proton collider based on existing CERN accelerator infrastructure. ",Collider design issues based on proton-driven plasma wakefield acceleration A specially designed and produced edge filter with pronounced nonlinear effects is carefully characterized. The nonlinear effects are estimated at the intensities close to the laser-induced damage. ,Characterization of nonlinear effects in edge filters " We analyze an experimental method for creating interesting nonclassical states by processing the entanglement generated when two large coherent states interact in a cross-Kerr medium. We specifically investigate the effects of loss and noise in every mode of the experiment, as well as the effect of ""binning"" the post-selection outcomes. Even with these imperfections, we find an optimal set of currently-achievable parameters which would allow a proof-of-principle demonstration of number squeezing in states with large mean photon number. We discuss other useful states which can be generated with the same experimental tools, including a class of states which contain coherent superpositions of differing photon numbers, e.g. good approximations to the state $\frac{1}{\sqrt{2}} (|0\rangle+|20\rangle)$. Finally, we suggest one possible application of this state in the field of optomechanics. ",Verifying cross-Kerr induced number squeezing: a case study " We consider the semiclassical limit for the Heisenberg-von Neumann equation with a potential which consists of the sum of a repulsive Coulomb potential, plus a Lipschitz potential whose gradient belongs to $BV$; this assumption on the potential guarantees the well posedness of the Liouville equation in the space of bounded integrable solutions. We find sufficient conditions on the initial data to ensure that the quantum dynamics converges to the classical one. More precisely, we consider the Husimi functions of the solution of the Heisenberg-von Neumann equation, and under suitable assumptions on the initial data we prove that they converge, as $\e \to 0$, to the unique bounded solution of the Liouville equation (locally uniformly in time). ",Semiclassical limit for mixed states with singular and rough potentials " The main result of the paper is motivated by the following two, apparently unrelated graph optimization problems: (A) as an extension of Edmonds' disjoint branchings theorem, characterize digraphs comprising $k$ disjoint branchings $B_i$ each having a specified number $\mu _i$ of arcs, (B) as an extension of Ryser's maximum term rank formula, determine the largest possible matching number of simple bipartite graphs complying with degree-constraints. The solutions to these problems and to their generalizations will be obtained from a new min-max theorem on covering a supermodular function by a simple degree-constrained bipartite graph. A specific feature of the result is that its minimum cost extension is already NP-complete. Therefore classic polyhedral tools themselves definitely cannot be sufficient for solving the problem, even though they make some good service in our approach. ",Supermodularity in Unweighted Graph Optimization I: Branchings and Matchings " Carrier and phonon dynamics in Bi2Se3 crystals are studied by a spatially resolved ultrafast pump-probe technique. Pronounced oscillations in differential reflection are observed with two distinct frequencies, and are attributed to coherent optical and acoustic phonons, respectively. The rising time of the signal indicates that the thermalization and energy relaxation of hot carriers are both sub-ps in this material. We found that the thermalization and relaxation time decreases with the carrier density. The expansion of the differential reflection profile allows us to estimate an ambipolar carrier diffusion coefficient on the order of 500 square centimeters per second. A long-term slow expansion of the profile shows a thermal diffusion coefficient of 1.2 square centimeters per second. ",Spatially resolved femtosecond pump-probe study of topological insulator Bi2Se3 " Liouville-type theorems for the steady incompressible Navier-Stokes system are investigated for solutions in a three-dimensional slab with either no-slip boundary conditions or periodic boundary conditions. When the no-slip boundary conditions are prescribed, we prove that any bounded solution is trivial if it is axisymmetric or $ru^r$ is bounded, and that general three-dimensional solutions must be Poiseuille flows when the velocity is not big in $L^\infty$ space. When the periodic boundary conditions are imposed on the slab boundaries, we prove that the bounded solutions must be constant vectors if either the swirl or radial velocity is independent of the angular variable, or $ru^r$ decays to zero as $r$ tends to infinity. The proofs are based on the fundamental structure of the equations and energy estimates. The key technique is to establish a Saint-Venant type estimate that characterizes the growth of Dirichlet integral of nontrivial solutions. ",Liouville-type theorems for steady solutions to the Navier-Stokes system in a slab " In this work, we present a semi-numerical solution of a fractal telegraphic dual-porosity fluid flow model. It combines Laplace transform and finite difference schemes. The Laplace transform handles the time variable whereas the finite difference method deals with the spatial coordinate. This semi-numerical scheme is not restricted by space discretization and allows the computation of a solution at any time without compromising numerical stability or the mass conservation principle. Our formulation results in a non-analytically-solvable second-order differential equation whose numerical treatment outcomes in a tri-diagonal linear algebraic system. Moreover, we describe comparisons between semi-numerical and semi-analytical solutions for particular cases. Results agree well with those from semi-analytic solutions. Furthermore, we expose a parametric analysis from the coupled model in order to show the effects of relevant parameters on pressure profiles and flow rates for the case where neither analytic nor semi-analytic solutions are available. ",Semi-numerical solution for a fractal telegraphic dual-porosity fluid flow model " We study observational consequences arising from dark matter (DM) of non-thermal origin, produced by dark freeze-out from a hidden sector heat bath. We assume this heat bath was populated by feebly-coupled mediator particles, produced via a Higgs portal interaction with the Standard Model (SM). The dark sector then attained internal equilibrium with a characteristic temperature different from the SM photon temperature. We find that even if the coupling between the DM and the SM sectors is very weak, the scenario allows for indirect observational signals. We show how the expected strength of these signals depends on the temperature of the hidden sector at DM freeze-out. ",Prospects for indirect detection of frozen-in dark matter " We present the first results of applying Gaussian Mixture Models in the stellar kinematic space of normalized angular momentum and binding energy on NIHAO high resolution galaxies to separate the stars into multiple components. We exemplify this method using a simulated Milky Way analogue, whose stellar component hosts: thin and thick discs, classical and pseudo bulges, and a stellar halo. The properties of these stellar structures are in good agreement with observational expectations in terms of sizes, shapes and rotational support. Interestingly, the two kinematic discs show surface mass density profiles more centrally concentrated than exponentials, while the bulges and the stellar halo are purely exponential. We trace back in time the Lagrangian mass of each component separately to study their formation history. Between z~3 and the end of halo virialization, z~1.3, all components lose a fraction of their angular momentum. The classical bulge loses the most (~95%) and the thin disc the least (~60%). Both bulges formed their stars in-situ at high redshift, while the thin disc formed ~98% in-situ, but with a constant SFR~1.5M$_{\rm\odot}$/yr$^{\rm-1}$ over the last ~11 Gyr. Accreted stars (6% of total stellar mass) are mainly incorporated to the thick disc or the stellar halo, which formed ex-situ 8% and 45% of their respective masses. Our analysis pipeline is freely available at https://github.com/aobr/gsf. ",Introducing galactic structure finder: the multiple stellar kinematic structures of a simulated Milky Way mass galaxy " We show that scattering from the boundary of static, higher-order topological insulators (HOTIs) can be used to simulate the behavior of (time-periodic) Floquet topological insulators. We consider D-dimensional HOTIs with gapless corner states which are weakly probed by external waves in a scattering setup. We find that the unitary reflection matrix describing back-scattering from the boundary of the HOTI is topologically equivalent to a (D-1)-dimensional nontrivial Floquet operator. To characterize the topology of the reflection matrix, we introduce the concept of `nested' scattering matrices. Our results provide a route to engineer topological Floquet systems in the lab without the need for external driving. As benefit, the topological system does not to suffer from decoherence and heating. ",Simulating Floquet topological phases in static systems " Using the Fourier-Laplace transform, we describe the isomonodromy equations for meromorphic connections on the Riemann sphere with unramified irregular singularities as those for connections with a (possibly ramified) irregular singularity and a regular singularity. This generalizes some results of Harnad and Woodhouse. ",Fourier-Laplace transform and isomonodromic deformations " ICMP timestamp request and response packets have been standardized for nearly 40 years, but have no modern practical application, having been superseded by NTP. However, ICMP timestamps are not deprecated, suggesting that while hosts must support them, little attention is paid to their implementation and use. In this work, we perform active measurements and find 2.2 million hosts on the Internet responding to ICMP timestamp requests from over 42,500 unique autonomous systems. We develop a methodology to classify timestamp responses, and find 13 distinct classes of behavior. Not only do these behaviors enable a new fingerprinting vector, some behaviors leak important information about the host e.g., OS, kernel version, and local timezone. ",Sundials in the Shade: An Internet-wide Perspective on ICMP Timestamps " We prove that universal quantum computation can be realized---using only linear optics and $\chi^{(2)}$ (three-wave mixing) interactions---in any $(n+1)$-dimensional qudit basis of the $n$-pump-photon subspace. First, we exhibit a strictly universal gate set for the qubit basis in the one-pump-photon subspace. Next, we demonstrate qutrit-basis universality by proving that $\chi^{(2)}$ Hamiltonians and photon-number operators generate the full $\mathfrak{u}(3)$ Lie algebra in the two-pump-photon subspace, and showing how the qutrit controlled-$Z$ gate can be implemented with only linear optics and $\chi^{(2)}$ interactions. We then use proof by induction to obtain our general qudit result. Our induction proof relies on coherent photon injection/subtraction, a technique enabled by $\chi^{(2)}$ interaction between the encoding modes and ancillary modes. Finally, we show that coherent photon injection is more than a conceptual tool in that it offers a route to preparing high-photon-number Fock states from single-photon Fock states. ",Qudit-Basis Universal Quantum Computation using $\chi^{(2)}$ Interactions " Time series has wide applications in the real world and is known to be difficult to forecast. Since its statistical properties change over time, its distribution also changes temporally, which will cause severe distribution shift problem to existing methods. However, it remains unexplored to model the time series in the distribution perspective. In this paper, we term this as Temporal Covariate Shift (TCS). This paper proposes Adaptive RNNs (AdaRNN) to tackle the TCS problem by building an adaptive model that generalizes well on the unseen test data. AdaRNN is sequentially composed of two novel algorithms. First, we propose Temporal Distribution Characterization to better characterize the distribution information in the TS. Second, we propose Temporal Distribution Matching to reduce the distribution mismatch in TS to learn the adaptive TS model. AdaRNN is a general framework with flexible distribution distances integrated. Experiments on human activity recognition, air quality prediction, and financial analysis show that AdaRNN outperforms the latest methods by a classification accuracy of 2.6% and significantly reduces the RMSE by 9.0%. We also show that the temporal distribution matching algorithm can be extended in Transformer structure to boost its performance. ",AdaRNN: Adaptive Learning and Forecasting of Time Series " The entangled behavior of different dimensional systems driven by classical external random field is investigated. The amount of the survival entanglement between the components of each system is quantified. There are different behaviors of entanglement that come into view decay, sudden death, sudden birth and long-lived entanglement. The maximum entangled states which can be generated from any of theses suggested systems are much fragile than the partially entangled ones. The systems of larger dimensions are more robust than those of smaller dimensions systems, where the entanglement decay smoothly, gradually and may vanish for a very short time. For the class of $2\times 3$ dimensional system, the one parameter family is found to be more robust than the two parameters family. Although the entanglement of driven $ 2 \times 3$ dimensional system is very sensitive to the classical external random field, one can use them to generate a long-lived entanglement. ",Sudden death and rebirth of Entanglement for Different Dimensional Systems driven by a Classical Random External Field " The N\'eel temperature, staggered magnetization density, as well as the spinwave velocity of a three-dimensional (3D) quantum Heisenberg model with antiferromagnetic disorder (randomness) are calculated using first principles non-perturbative quantum Monte Carlo simulations. In particular, we examine the validity of universal scaling relations that are related to these three studied physical quantities. These relations are relevant to experimental data and are firmly established for clean (regular) 3D dimerized spin-1/2 Heisenberg models. Remarkably, our numerical results show that the considered scaling relations remain true for the investigated model with the introduced disorder. In addition, while the presence of disorder may change the physical properties of regular dimerized models, hence leading to different critical theories, both the obtained data of N\'eel temperature and staggered magnetization density in our study are fully compatible with the expected critical behaviour for clean dimerized systems. As a result, it is persuasive to conclude that the related quantum phase transitions of the considered disordered model and its clean counterparts are governed by the same critical theory, which is not always the case in general. Finally, we also find smooth scalings even emerging when both the data of the investigated disordered model as well as its associated clean system are taken into account. This in turn implies that, while in a restricted sense, the considered scaling relations for 3D spin-1/2 antiferromagnets are indeed universal. ","Universal scalings of N\'eel temperature, staggered magnetization density, and spinwave velocity of three-dimensional disordered and clean quantum antiferromagnets" " To understand and improve DRAM performance, reliability, security and energy efficiency, prior works study characteristics of commodity DRAM chips. Unfortunately, state-of-the-art open source infrastructures capable of conducting such studies are obsolete, poorly supported, or difficult to use, or their inflexibility limit the types of studies they can conduct. We propose DRAM Bender, a new FPGA-based infrastructure that enables experimental studies on state-of-the-art DRAM chips. DRAM Bender offers three key features at the same time. First, DRAM Bender enables directly interfacing with a DRAM chip through its low-level interface. This allows users to issue DRAM commands in arbitrary order and with finer-grained time intervals compared to other open source infrastructures. Second, DRAM Bender exposes easy-to-use C++ and Python programming interfaces, allowing users to quickly and easily develop different types of DRAM experiments. Third, DRAM Bender is easily extensible. The modular design of DRAM Bender allows extending it to (i) support existing and emerging DRAM interfaces, and (ii) run on new commercial or custom FPGA boards with little effort. To demonstrate that DRAM Bender is a versatile infrastructure, we conduct three case studies, two of which lead to new observations about the DRAM RowHammer vulnerability. In particular, we show that data patterns supported by DRAM Bender uncovers a larger set of bit-flips on a victim row compared to the data patterns commonly used by prior work. We demonstrate the extensibility of DRAM Bender by implementing it on five different FPGAs with DDR4 and DDR3 support. DRAM Bender is freely and openly available at https://github.com/CMU-SAFARI/DRAM-Bender. ",DRAM Bender: An Extensible and Versatile FPGA-based Infrastructure to Easily Test State-of-the-art DRAM Chips " We derive a posteriori error estimates for a fully discrete finite element approximation of the stochastic Cahn-Hilliard equation. The a posteriori bound is obtained by a splitting of the equation into a linear stochastic partial differential equation (SPDE) and a nonlinear random partial differential equation (RPDE). The resulting estimate is robust with respect to the interfacial width parameter and is computable since it involves the discrete principal eigenvalue of a linearized (stochastic) Cahn-Hilliard operator. Furthermore, the estimate is robust with respect to topological changes as well as the intensity of the stochastic noise. We provide numerical simulations to demonstrate the practicability of the proposed adaptive algorithm. ",Robust a posteriori estimates for the stochastic Cahn-Hilliard equation " I highlight several concerns regarding the consistency of Type Ia supernova data in the publicly available Pantheon and JLA compilations. The measured heliocentric redshifts (zhel) of $\sim$150 SNe Ia as reported in the Pantheon catalogue are significantly discrepant from those in JLA - with 58 having differences amounting to between 5 and 137 times the quoted measurement uncertainty. The discrepancy seems to have been introduced in the process of rectifying a previously reported issue. The Pantheon catalogue until very recently had the redshifts of all SNe Ia up to z $\sim$ 0.3 modified under the guise of 'peculiar velocity corrections' - although there is no information on peculiar velocities at such high redshifts. While this has reportedly been rectified on Github by removing peculiar velocity corrections for z > 0.08, the impact of this on the published cosmological analysis of the Pantheon catalogue is not stated. In JLA, the effect of these 'corrections' is to significantly bias the inferred value of $\Omega_{\Lambda}$ towards higher values, while the equivalent effect on Pantheon cannot be ascertained due to the unavailability of the individual components of the covariance matrix in the public domain. I provide Jupyter notebooks and URLs in order to allow the reader to ascertain the veracity of these assertions. ",Concerns about the reliability of publicly available SNe Ia data " Operationally accessible entanglement in bipartite systems of indistinguishable particles could be reduced due to restrictions on the allowed local operations as a result of particle number conservation. In order to quantify this effect, Wiseman and Vaccaro [Phys. Rev. Lett. 91, 097902 (2003)] introduced an operational measure of the von Neumann entanglement entropy. Motivated by advances in measuring R\'enyi entropies in quantum many-body systems subject to conservation laws, we derive a generalization of the operational entanglement that is both computationally and experimentally accessible. Using the Widom theorem, we investigate its scaling with the size of a spatial subregion for free fermions and find a logarithmically violated area law scaling, similar to the spatial entanglement entropy, with at most, a double-log leading-order correction. A modification of the correlation matrix method confirms our findings in systems of up to $10^5$ particles. ",R\'enyi generalization of the operational entanglement entropy " This investigation deals with the effect of annealing temperature on the structural, topographical and optical properties of Zinc Oxide thin films prepared by sol gel method. The structural properties were studied using X-ray diffraction and the recorded patterns indicated that all the films had a preferred orientation along (002) plane and the crystallinity along with the grain size were augmented with annealing temperature. The topographical modification of the films due to heat treatment was probed by atomic force microscopy which revealed that annealing roughened the surface of the film. The optical properties were examined by a UV visible spectrophotometer which exhibited that maximum transmittance reached nearly 90% and it diminished with increasing annealing temperature. ","Influence of annealing temperature on the structural, topographical and optical properties of sol gel derived ZnO thin films" " Noting that the extensive astrometric observations of the double comet Wirtanen (C/1956 F1) made by E. Roemer have never been published, I replicate the contents of a fortuitously discovered copy of her measurement records of the companion's offsets from the main mass in 1957-1959 and use with such data by others to refine the fragmentation solution. The sublimation-driven nongravitational acceleration is shown to essentially control the companion's motion in the orbital plane. The fragmentation parameters derived by the author in 1978 have now been improved and strong disagreement with the independent results by Roemer is noted. The revised model is employed to predict the positions of the companion on the plates exposed by Roemer on 25 September 1960, which she reported to show the principal nucleus but not the companion. At my request, these plates have now been scanned and processed at the Lowell Observatory, and the companion is found to be located at the predicted position. The images of the main mass and the companion on one of the two plates are displayed. ",On Elusive Observations and a Sly Companion of Comet Wirtanen (C/1956 F1) " Local environment of fluorescent dyes could strongly affect emission dynamics of the latter. In particular, both signal intensities and emission lifetimes are highly sensitive to solvent temperatures. Here, temperature-dependent behavior Rhodamine B fluorescence in water and ethanol solutions was experimentally investigated. Phase transition point of liquid water to ice was shown to have a dramatic impact on both in intensity (30-fold drop) and in lifetime (from 2.68ns down to 0.13ns) of the dye luminescence along with the spectral shift of spectral maxima from 590 to 625nm. At the same time, ethanol solvent does not lead any similar behavior. The reported results and approaches enable further investigations of dye-solvent interactions and studies of physical properties of liquids at phase transition points. ",Rhodamine B as a probe for phase transition in liquid solutions " The purpose of this paper is to investigate the connection between the Lotka-Volterra system and combinatorics. We study several context-free grammars associated with the Lotka-Volterra system. Some combinatorial arrays, involving the Stirling numbers of the second kind and Eulerian numbers, are generated by these context-free grammars. In particular, we present grammatical characterization of some statistics on cyclically ordered partitions. ",Some combinatorial arrays related to the Lotka-Volterra system " We present some results on the usefulness of upgraded Tevatron, LHC proton proton collider and linear e+e- colliders in the TeV range to test the idea of a strongly interacting sector as responsible for the electroweak symmetry breaking. The calculations are performed within an effective lagrangian description, called the BESS model, which provides for a rather general frame based on the standing point of custodial symmetry and gauge invariance, without specifying any dynamical scheme. ",Some Results on the BESS Model at Future Colliders " We combine experimental bounds on the electric dipole moments of the neutron and electron with cosmological limits on the relic density of a gaugino-type LSP neutralino to constrain certain CP-violating phases appearing in the MSSM. We find that in the Constrained MSSM, the phase |\theta_\mu | < \pi/10, while the phase \theta_A remains essentially unconstrained. ",CP-Violating Phases in the MSSM " Let $g$ be a holomorphic map of $B$, where $B$ is the unit ball of ${C}^n$. Let $0-1$ and $\alpha>0$. This paper gives some necessary and sufficient conditions for the Extended Ces$\acute{a}$ro Operators induced by $g$ to be bounded or compact between generalized Besov space $B(p,q)$ and $\alpha$- Bloch space ${\mathcal B}^\alpha.$ ",Extended Ces$\acute{a}$RO Operators between Generalized Besov Spaces and Bloch Type Spaces in the Unit Ball " Two-photon imaging of calcium indicators allows simultaneous recording of responses of hundreds of neurons over hours and even days, but provides a relatively indirect measure of their spiking activity. Existing deconvolution algorithms attempt to recover spikes from observed imaging data, which are then commonly subjected to the same analyses that are applied to electrophysiologically recorded spikes (e.g., estimation of average firing rates, or tuning curves). Here we show, however, that in the presence of noise this approach is often heavily biased. We propose an alternative analysis that aims to estimate the underlying rate directly, by integrating over the unobserved spikes instead of committing to a single estimate of the spike train. This approach can be used to estimate average firing rates or tuning curves directly from the imaging data, and is sufficiently flexible to incorporate prior knowledge about tuning structure. We show that directly estimated rates are more accurate than those obtained from averaging of spikes estimated through deconvolution, both on simulated data and on imaging data acquired in mouse visual cortex. ",Direct Estimation of Firing Rates from Calcium Imaging Data " We have studied crystal structure, magnetism and electric transport properties of a europium fulleride Eu6C60 and its Sr-substituted compounds, Eu6-xSrxC60. They have a bcc structure, which is an isostructure of other M6C60 (M represents an alkali atom or an alkaline earth atom). Magnetic measurements revealed that magnetic moment is ascribed to the divalent europium atom with S = 7/2 spin, and a ferromagnetic transition was observed at TC = 10 - 14 K. In Eu6C60, we also confirm the ferromagnetic transition by heat capacity measurement. The striking feature in Eu6-xSrxC60} is very large negative magnetoresistance at low temperature; the resistivity ratio \rho(H = 9 T)/\rho(H = 0 T) reaches almost 10^{-3} at 1 K in Eu6C60. Such large magnetoresistance is the manifestation of a strong pi-f interaction between conduction carriers on C60 and 4f electrons of Eu. ",Ferromagnetism and giant magnetoresistance in the rare earth fullerides Eu6-xSrxC60 " We investigate the scaling properties of the Barkhausen effect, recording the noise in several soft ferromagnetic materials: polycrystals with different grain sizes and amorphous alloys. We measure the Barkhausen avalanche distributions and determine the scaling exponents. In the limit of vanishing external field rate, we can group the samples in two distinct classes, characterized by exponents \tau = 1.50 \pm 0.05 or \tau = 1.27 \pm 0.03, for the avalanche size distributions. We interpret these results in terms of the depinning transition of domain walls and obtain an expression relating the cutoff of the distributions to the demagnetizing factor which is in quantitative agreement with experiments. ",Scaling exponents for Barkhausen avalanches in polycrystals and amorphous ferromagnets " Graphics Processing Units (GPUs) were once used solely for graphical computation tasks but with the increase in the use of machine learning applications, the use of GPUs to perform general-purpose computing has increased in the last few years. GPUs employ a massive amount of threads, that in turn achieve a high amount of parallelism, to perform tasks. Though GPUs have a high amount of computation power, they face the problem of cache contention due to the SIMT model that they use. A solution to this problem is called ""cache bypassing"". This paper presents a predictive model that analyzes the access patterns of various machine learning algorithms and determines whether certain data should be stored in the cache or not. It presents insights on how well each model performs on different datasets and also shows how minimizing the size of each model will affect its performance The performance of most of the models were found to be around 90% with KNN performing the best but not with the smallest size. We further increase the features by splitting the addresses into chunks of 4 bytes. We observe that this increased the performance of the neural network substantially and increased the accuracy to 99.9% with three neurons. ",Cache Bypassing for Machine Learning Algorithms " We present preliminary results of an ongoing survey to probe the physical nature of a sample of open clusters (POCRs, Bica et al. 2001) believed to be in an advanced stage of dynamical evolution. This kind of object is very important to constrain dynamical $N$-body models of clusters evolution and dissolution, and to understand the formation of the Galactic field stars population. Moreover they might represent what remains of much larger old clusters, which are of paramount importance to probe the early stages of the Galactic disk evolution. In this paper we present our results for NGC 1901. ",A survey of Probable Open Cluster Remnants in the Galactic Disk We realize the enveloping algebra of the positive part of a symmetrizable Kac-Moody algebra as a convolution algebra of constructible functions on module varieties of some Iwanaga-Gorenstein algebras of dimension 1. ,Quivers with relations for symmetrizable Cartan matrices II : Convolution algebras " With the attention mechanism, transformers achieve significant empirical successes. Despite the intuitive understanding that transformers perform relational inference over long sequences to produce desirable representations, we lack a rigorous theory on how the attention mechanism achieves it. In particular, several intriguing questions remain open: (a) What makes a desirable representation? (b) How does the attention mechanism infer the desirable representation within the forward pass? (c) How does a pretraining procedure learn to infer the desirable representation through the backward pass? We observe that, as is the case in BERT and ViT, input tokens are often exchangeable since they already include positional encodings. The notion of exchangeability induces a latent variable model that is invariant to input sizes, which enables our theoretical analysis. - To answer (a) on representation, we establish the existence of a sufficient and minimal representation of input tokens. In particular, such a representation instantiates the posterior distribution of the latent variable given input tokens, which plays a central role in predicting output labels and solving downstream tasks. - To answer (b) on inference, we prove that attention with the desired parameter infers the latent posterior up to an approximation error, which is decreasing in input sizes. In detail, we quantify how attention approximates the conditional mean of the value given the key, which characterizes how it performs relational inference over long sequences. - To answer (c) on learning, we prove that both supervised and self-supervised objectives allow empirical risk minimization to learn the desired parameter up to a generalization error, which is independent of input sizes. Particularly, in the self-supervised setting, we identify a condition number that is pivotal to solving downstream tasks. ",An Analysis of Attention via the Lens of Exchangeability and Latent Variable Models " We have investigated the possible existence of a $^7$H resonant state, considered as a five-body system consisting of a $^3$H core with four valence neutrons. To this aim, an effective n-$^3$H potential is constructed in order to reproduce the low energy elastic neutron scattering on $^3$H phase shifts and the $^5$H resonant ground state in terms of $^3$H-n-n system. The variational Gaussian Expansion Method is used to solve the 5-body Schr\""{o}dinger equation, while the resonant state parameters were estimated by means of the stabilization method. We have not found any sign of a narrow low energy resonance in the vicinity of $^3$H+4n threshold. However, we have identified a very broad structure at $E_R\approx 9$ MeV above this threshold, which corresponds to the $^7$H J$^{\pi}$=1/2$^+$ ground state. In the vicinity of this state, we have also identified a broad structure corresponding to the ground state of $^6$H isotope with quantum numbers $J^{\pi}=2^-$. ",$^7$H ground state as a $^3$H+4n resonance " Experimental measurements and numerical simulations of turbulent flows are characterised by a trade-off between accuracy and resolution. In this study, we bridge this gap using Physics Informed Neural Networks (PINNs) constrained by the Reynolds-Averaged Navier-Stokes (RANS) equations and accurate sparse pointwise mean velocity measurements for data assimilation (DA). Firstly, by constraining the PINN with sparse data and the under-determined RANS equations without closure, we show that the mean flow is reconstructed to a higher accuracy than a RANS solver using the Spalart-Allmaras (SA) turbulence model. Secondly, we propose the SA turbulence model augmented PINN (PINN-DA-SA), which outperforms the former approach - up to 73% reduction in mean velocity reconstruction error with coarse measurements. The additional SA physics constraints improve flow reconstructions in regions with high velocity and pressure gradients and separation. Thirdly, we compare the PINN-DA-SA approach to a variational data assimilation using the same sparse velocity measurements and physics constraints. The PINN-DA-SA achieves lower reconstruction error across a range of data resolutions. This is attributed to discretisation errors in the variational methodology that are avoided by PINNs. We demonstrate the method using high fidelity measurements from direct numerical simulation of the turbulent periodic hill at Re=5600. ",Turbulence model augmented physics informed neural networks for mean flow reconstruction The Hanbury-Brown-Twiss (HBT) radii have been calculated from the two particle correlation functions with virtual photons produced in the collisions of two nuclei at ultra-relativistic energies. We show that the variation of the HBT radii with the invariant mass of the virtual photon can be used to characterize and distinguish the hadronic as well as the partonic phase that might have produced initially in the collisions. It has been illustrated that the non-monotonic variation of the HBT radii with invariant mass provides an access to the development of collective flow in the system. ,Characterizing quark gluon plasma by dilepton interferometry " We discuss recent Chandra and XMM-Newton observations of the bright isolated neutron star RX J1856.5-3754 and suggest that the absence of any line features is due to effects of a high magnetic field strength (~10^13 G). Using different models for the temperature distribution across the neutron star surface assuming blackbody emission to fit the optical and X-ray spectrum and we derive a conservative lower limit of the ""apparent"" neutron star radius of 16.5 km x (d/117 pc). This corresponds to the radius for the ""true"" (de-redshifted) radius of 14 km for a 1.4 Msun neutron star, indicating a stiff equation of state at high densities. A comparison of the result with mass-radius diagrams shows that quark stars and neutron stars with quark matter cores can be ruled out with high confidence. ",The Puzzles of RX J1856.5-3754: Neutron Star or Quark Star? " The newly developed ""void expansion method"" allows for an efficient generation of porous packings of spherical particles over a wide range of volume fractions using the discrete element method. Particles are randomly placed under addition of much smaller ""void-particles"". Then, the void-particle radius is increased repeatedly, thereby rearranging the structural particles until formation of a dense particle packing. The structural particles' mean coordination number was used to characterize the evolving microstructures. At some void radius, a transition from an initially low to a higher mean coordination number is found, which was used to characterize the influence of the various simulation parameters. For structural and void-particle stiffnesses of the same order of magnitude, the transition is found at constant total volume fraction slightly below the random close packing limit. For decreasing void-particle stiffness the transition is shifted towards a smaller void-particle radius and becomes smoother. ",Generation of Porous Particle Structures using the Void Expansion Method " We study reversible deterministic dynamics of classical charged particles on a lattice with hard-core interaction. It is rigorously shown that the system exhibits three types of transport phenomena, ranging from ballistic, through diffusive to insulating. By obtaining an exact expressions for the current time-autocorrelation function we are able to calculate the linear response transport coefficients, such as the diffusion constant and the Drude weight. Additionally, we calculate the long- time charge profile after an inhomogeneous quench and obtain diffusive profile with the Green-Kubo diffusion constant. Exact analytical results are corroborated by Monte-Carlo simulations. ",Diffusion in deterministic interacting lattice systems " Though wakefield acceleration in crystal channels has been previously proposed, x-ray wakefield acceleration has only recently become a realistic possibility since the invention of the single-cycled optical laser compression technique. We investigate the acceleration due to a wakefield induced by a coherent, ultrashort x-ray pulse guided by a nanoscale channel inside a solid material. By two-dimensional particle in- cell computer simulations, we show that an acceleration gradient of TeV/cm is attainable. This is about 3 orders of magnitude stronger than that of the conventional plasma-based wakefield accelerations, which implies the possibility of an extremely compact scheme to attain ultrahigh energies. In addition to particle acceleration, this scheme can also induce the emission of high energy photons at ~O(10-100) MeV. Our simulations confirm such high energy photon emissions, which is in contrast with that induced by the optical laser driven wakefield scheme. In addition to this, the significantly improved emittance of the energetic electrons has been discussed. ",Particle-in-cell simulation of x-ray wakefield acceleration and betatron radiation in nanotubes " We present a higher genus generalization of $bc$-Motzkin numbers, which are themselves a generalization of Catalan numbers, and we derive a recursive formula which can be used to calculate them. Further, we show that this leads to a topological recursion which is identical to the topological recursion that had previously been proved by Dumitrescu and Mulase for generalized Catalan numbers, and which is an example of the Eynard-Orantin topological recursion. ",Topological Recursion for Generalized $bc$-Motzkin Numbers " At very high energies, the high parton densities (characterized by a semi-hard saturation scale \Lambda_s) ensure that parton distributions can be described by a classical effective field theory with remarkable properties analogous to those of spin glass systems. This color glass condensate (CGC) of gluons also provides the initial conditions for multi-particle production in high energy nuclear collisions. In this talk, we briefly summarize recent theoretical and phenomenological progress in the CGC approach to small x physics. In particular, we discuss recent numerical work on the real time gluodynamics of partons after a nuclear collision. The implications of this work for the theoretical study of thermalization in nuclear collisions and on the phenomenological interpretation of results of the recent RHIC experiments are also discussed. ",Small x physics and the initial conditions in heavy ion collisions " An exactly solvable model of the sawtooth chain with Ising and Heisenberg bonds and with coupling to lattice distortion for Heisenberg bonds is considered in the magnetic field. Using the direct transfer-matrix formalism an exact description of the thermodynamic functions is obtained. The ground state phase diagrams for all regions of parameters values containing phases corresponding to the magnetization plateaus at $M=0,1/4$ and 1/2 have been obtained. Exact formulas for bond distortions for various ground states are presented. A novel mechanism of magnetization plateau stabilization corresponding to $M=1/4$ state is reported. ",Lattice distortions in a sawtooth chain with Heisenberg and Ising bonds " We numerically evolve spherically symmetric solutions to the linear wave equation on some expanding Friedmann-Lema\^itre-Robertson-Walker (FLRW) spacetimes and study the respective asymptotics for large times. We find a quantitative relation between the expansion rate of the underlying background universe and the decay rate of linear waves, also in the context of spatially-hyperbolic spacetimes, for which rigorous proofs of decay rates are not generally known. A prominent role in the decay mechanism is shown to be played by tails, i.e. scattered waves propagating in the interior of the lightcone. ",Decay of solutions of the wave equation in cosmological spacetimes -- a numerical analysis " We propose that glass-forming liquids are intrinsically under the influences of both fluctuating interactions and random fields well-known in the field of spin systems. This is due to the frustration between the isotropic and anisotropic parts of effective intermolecular interactions. Our model indicates the existence of two key temperatures relevant to glass transition, the density ordering point T^*_m and the Vogel-Fulcher temperature T_0. Between T^*_m and T_0, a system has features similar to the `Griffiths phase', while below $T_0$ it has those peculiar to the `spin-glass phase'. This picture naturally and universally explains vitrification behavior from its strong to fragile limit. ",A simple physical model of liquid-glass transition: intrinsic fluctuating interactions and random fields hidden in glass-forming liquids " W43-MM1 is a young region, very rich in terms of high-mass star formation. We aim to systematically identify the massive cores which contain a hot core and compare their molecular composition. We used ALMA high-spatial resolution (2500 au) data of W43-MM1 to identify line-rich protostellar cores and make a comparative study of their temperature and molecular composition. The identification of hot cores is based on both the spatial distribution of the complex organic molecules and the contribution of molecular lines relative to the continuum intensity. We rely on the analysis of CH3CN and CH3CCH to estimate the temperatures of the selected cores. Finally, we rescale the spectra of the different hot cores based on their CH3OCHO line intensities to directly compare the detections and line intensities of the other species. W43-MM1 turns out to be a region rich in massive hot cores with at least 1 less massive and 7 massive hot cores. The excitation temperature of CH3CN is of the same order for all of them (120-160 K). There is a factor of up to 30 difference in the intensity of the complex organic molecules (COMs) lines. However the molecular emission of the hot cores appears to be the same within a factor 2-3. This points towards both a similar chemical composition and excitation of most of the COMs over these massive cores, which span about an order of magnitude in core mass. In contrast, CH3CCH emission is found to preferentially trace more the envelope, with a temperature ranging from 50 K to 90 K. Core 1, the most massive hot core of W43-MM1, shows a richer line spectrum than the other cores. In core 2, the emission of O-bearing molecules does not peak at the dust continuum core center; the blue and red shifted emission correspond to the outflow lobes, suggesting a formation via the sublimation of the ice mantles through shocks or UV irradiation on the walls of the cavity. ","ALMA-IMF IV -- A comparative study of the main hot cores in W43-MM1: detection, temperature and molecular composition" " Bring Your Own Device (BYOD) is a growing trend among enterprises, aiming to improve workers' mobility and productivity via their smartphones. The threats and dangers posed by the smartphones to the enterprise are also ever-growing. Such dangers can be mitigated by running the enterprise software inside a ""secure container"" on the smartphone. In our work we present a systematic assessment of security critical areas in design and implementation of a secure container for Android using reverse engineering and attacker-inspired methods. We do this through a case-study of Samsung KNOX, a real-world product deployed on millions of devices. Our research shows how KNOX security features work behind the scenes and lets us compare the vendor's public security claims against reality. Along the way we identified several design weaknesses and a few vulnerabilities that were disclosed to Samsung. ",Secure Containers in Android: the Samsung KNOX Case Study " We discuss stability and phase coherence of 1D trapped Bose gases and find that inelastic decay processes, such as 3-body recombination, are suppressed in the strongly interacting (Tonks-Girardeau) and intermediate regimes. This is promising for achieving these regimes with a large number of particles. ""Fermionization"" of the system reduces the phase coherence length, and at T=0 the gas is fully phase coherent only deeply in the weakly interacting (Gross-Pitaevskii) regime. ",Stability and phase coherence of trapped 1D Bose gases We prove that the energy conditions in arXiv:1302.5093v7 are implied by the fractional A2 conditions and testing conditions for the vector of fractional Riesz transforms when one measure is supported on a line. Then we apply the main theorem in arXiv:1302.5093v7 to obtain a T1 theorem for the fractional Riesz transforms when one weight is supported on a line. ,"A geometric condition, necessity of energy, and two weight boundedness of fractional Riesz transforms" We review various connections between condensed matter systems with the Nambu-Jona Lasinio model and nonlinear sigma models. The field theoretical description of interacting systems offers a systematic framework to describe the dynamical generation of condensates. Resent findings of a duality between the Nambu-Jona Lasinio model and the nonlinear sigma model enables us to investigate various properties underlying both theories. In this review we mainly focus on inhomogeneous condensations in static situations. The various methods developed in the Nambu-Jona Lasinio model reveal the inhomogeneous phase structures and also yield new inhomogeneous solutions in the nonlinear sigma model owing to the duality. The recent progress on interacting systems in finite systems is also reviewed. ,Nambu-Jona Lasinio and Nonlinear Sigma Models in Condensed Matter Systems " In this paper, we propose new lower and upper bounds on the linear extension complexity of regular $n$-gons. Our bounds are based on the equivalence between the computation of (i) an extended formulation of size $r$ of a polytope $P$, and (ii) a rank-$r$ nonnegative factorization of a slack matrix of the polytope $P$. The lower bound is based on an improved bound for the rectangle covering number (also known as the boolean rank) of the slack matrix of the $n$-gons. The upper bound is a slight improvement of the result of Fiorini, Rothvoss and Tiwary [Extended Formulations for Polygons, Discrete Comput. Geom. 48(3), pp. 658-668, 2012]. The difference with their result is twofold: (i) our proof uses a purely algebraic argument while Fiorini et al. used a geometric argument, and (ii) we improve the base case allowing us to reduce their upper bound $2 \left\lceil \log_2(n) \right\rceil$ by one when $2^{k-1} < n \leq 2^{k-1}+2^{k-2}$ for some integer $k$. We conjecture that this new upper bound is tight, which is suggested by numerical experiments for small $n$. Moreover, this improved upper bound allows us to close the gap with the best known lower bound for certain regular $n$-gons (namely, $9 \leq n \leq 13$ and $21 \leq n \leq 24$) hence allowing for the first time to determine their extension complexity. ",On the Linear Extension Complexity of Regular n-gons " Following Richardson and using results of Kesten on First-passage percolation, we obtain an upper bound on the expected perimeter in an Eden Growth Process. Using results of the author from a problem in Statistical Mechanics, we show that the average perimeter of the lattice animals resulting from a very natural family of ""growth histories"" does not obey a similar bound. ",The Expected Perimeter in Eden and Related Growth Processes " Shannon quantum information entropies $S_{\rho,\gamma}$, Fisher informations $I_{\rho,\gamma}$, Onicescu energies $O_{\rho,\gamma}$ and complexities $e^SO$ are calculated both in position (subscript $\rho$) and momentum ($\gamma$) spaces for azimuthally symmetric 2D nanoring that is placed into combination of transverse uniform magnetic field $\bf B$ and Aharonov-Bohm (AB) flux $\phi_{AB}$ and whose potential profile is modeled by superposition of quadratic and inverse quadratic dependencies on radius $r$. Increasing intensity $B$ flattens momentum waveforms $\Phi_{nm}({\bf k})$ and in the limit of infinitely large fields they turn to zero, what means that the position wave functions $\Psi_{nm}({\bf r})$, which are their Fourier counterparts, tend in this limit to the $\delta$-functions. Position (momentum) Shannon entropy depends on the field $B$ as a negative (positive) logarithm of $\omega_{eff}\equiv\left(\omega_0^2+\omega_c^2/4\right)^{1/2}$, where $\omega_0$ determines the quadratic steepness of the confining potential and $\omega_c$ is a cyclotron frequency. This makes the sum ${S_\rho}_{nm}+{S_\gamma}_{nm}$ a field-independent quantity that increases with the principal $n$ and azimuthal $m$ quantum numbers and does satisfy entropic uncertainty relation. Position Fisher information does not depend on $m$, linearly increases with $n$ and varies as $\omega_{eff}$ whereas its $n$ and $m$ dependent Onicescu counterpart ${O_\rho}_{nm}$ changes as $\omega_{eff}^{-1}$. The products ${I_\rho}_{nm}{I_\gamma}_{nm}$ and ${O_\rho}_{nm}{O_\gamma}_{nm}$ are $B$-independent quantities. A dependence of the measures on the ring geometry is discussed. It is argued that a variation of the position Shannon entropy or Onicescu energy with the AB field uniquely determines an associated persistent current as a function of $\phi_{AB}$ at $B=0$. An inverse statement is correct too. ",Quantum information measures of the Aharonov-Bohm ring in uniform magnetic fields " We introduce a tracking-by-detection method that integrates a deep object detector with a particle filter tracker under the regularization framework where the tracked object is represented by a sparse dictionary. A novel observation model which establishes consensus between the detector and tracker is formulated that enables us to update the dictionary with the guidance of the deep detector. This yields an efficient representation of the object appearance through the video sequence hence improves robustness to occlusion and pose changes. Moreover we propose a new state vector consisting of translation, rotation, scaling and shearing parameters that allows tracking the deformed object bounding boxes hence significantly increases robustness to scale changes. Numerical results reported on challenging VOT2016 and VOT2018 benchmarking data sets demonstrate that the introduced tracker, L1DPF-M, achieves comparable robustness on both data sets while it outperforms state-of-the-art trackers on both data sets where the improvement achieved in success rate at IoU-th=0.5 is 11% and 9%, respectively. ",Integration of Regularized l1 Tracking and Instance Segmentation for Video Object Tracking " We show that the partially topological twisted N=16, D=2 super Yang-Mill theory gives rise to a N_T=8 Hodge-type cohomological gauge theory with global SU(4) symmetry. ","N_T=8, D=2 Hodge-type cohomological gauge theory with global SU(4) symmetry" " For any real number $x \in [0,1)$, we denote by $q_n(x)$ the denominator of the $n$-th convergent of the continued fraction expansion of $x$ $(n \in \mathbb{N})$. It is well-known that the Lebesgue measure of the set of points $x \in [0,1)$ for which $\log q_n(x)/n$ deviates away from $\pi^2/(12\log2)$ decays to zero as $n$ tends to infinity. In this paper, we study the rate of this decay by giving an upper bound and a lower bound. What is interesting is that the upper bound is closely related to the Hausdorff dimensions of the level sets for $\log q_n(x)/n$. As a consequence, we obtain a large deviation type result for $\log q_n(x)/n$, which indicates that the rate of this decay is exponential. ",The denominators of convergents for continued fractions " Deep learning, and in particular Recurrent Neural Networks (RNN) have shown superior accuracy in a large variety of tasks including machine translation, language understanding, and movie frame generation. However, these deep learning approaches are very expensive in terms of computation. In most cases, Graphic Processing Units (GPUs) are in used for large scale implementations. Meanwhile, energy efficient RNN approaches are proposed for deploying solutions on special purpose hardware including Field Programming Gate Arrays (FPGAs) and mobile platforms. In this paper, we propose an effective quantization approach for Recurrent Neural Networks (RNN) techniques including Long Short Term Memory (LSTM), Gated Recurrent Units (GRU), and Convolutional Long Short Term Memory (ConvLSTM). We have implemented different quantization methods including Binary Connect {-1, 1}, Ternary Connect {-1, 0, 1}, and Quaternary Connect {-1, -0.5, 0.5, 1}. These proposed approaches are evaluated on different datasets for sentiment analysis on IMDB and video frame predictions on the moving MNIST dataset. The experimental results are compared against the full precision versions of the LSTM, GRU, and ConvLSTM. They show promising results for both sentiment analysis and video frame prediction. ",Effective Quantization Approaches for Recurrent Neural Networks " A number of recent works employ bilinear Hamiltonian interactions between Linear Quantum Stochastic Systems (LQSSs). Contrary to naturally occurring Hamiltonian interactions between physical systems, such interactions must be engineered. In this work, we propose a simple model for the implementation of an arbitrary bilinear interaction between two given LQSSs via a feedback interconnection. ",Implementation of Bilinear Hamiltonian Interactions between Linear Quantum Stochastic Systems via Feedback " We present the Spectral Energy Distribution of HR 4049 based on literature data and new continuum measurements at 850 micron. The SED shows variable absorption in the UV, and a large IR excess, both caused by circumstellar dust. The shape of the IR excess from 1 micron all the way down to 850 micron can be nearly perfectly fitted with a single blackbody function at T ~ 1150 K or alternatively with a sum of blackbodies in a narrow temperature range. The energy emitted in this IR continuum radiation is about one-third of the stellar luminosity. We show that this blackbody radiation must be due to the presence of a circumbinary disk with a large height. This disk must also be gas-rich, in agreement with the observations of molecular bands in the ISO-SWS spectrum. We present two possible scenario's for explaining the shape and the intensity of the IR excess. The first scenario involves large grains (a >1 mm) that each radiate like a blackbody. The second scenario argues that the blackbody radiation is due to a very optically thick circumbinary disk. We investigate if such a disk would indeed produce blackbody radiation by presenting results from radiative transfer calculations. We further quantify the properties of such a disk and its stability in the framework of (hydro)dynamics, grain settling, radiation pressure and grain drift. The virtues and shortcomings of both models for the origin of the IR blackbody are discussed by contrasting them with other observations and assessing them in the framework of (binary) (post-)AGB evolution. ",The dust disk of HR4049 " We investigate the stability of a neutrino-dominated accretion flow (NDAF), which is expected to be formed in the gravitational collapse of a massive star or the merger of a neutron star binary, based on the variable-$\alpha$ prescription. Recent magnetohydrodynamic (MHD) simulations shows that the viscosity parameter $\alpha$ is proportional to the power of the magnetic Prandtl number ${\rm Pm}=\nu/\eta$, where $\nu$ and $\eta$ are the kinematic viscosity and electric resistivity of the fluid, respectively. In the inner region of a hyperaccretion flow, the viscosity and resistivity are carried by mildly, relativistically degenerated electrons. We fit the dependence of the magnetic Prandtl number on density and temperature by a simple analytic form, and derive the condition for an NDAF to be dynamically unstable. As demonstrations we perform simple one-dimensional simulations of NDAFs with the variable-$\alpha$ and show that the mass accretion becomes highly time-variable in the unstable branch. This mechanism may account for the rapid variability observed in the prompt emission of gamma-ray bursts (GRBs). The mass ejection from a hyperaccretion flow due to viscous heating, which makes a kilonova/macronova emission in the merger of a neutron star binary, is also briefly discussed. ",Neutrino-Dominated Accretion Flows with Magnetic Prandtl Number-Dependent MRI-driven Turbulence " In this article we study singular subelliptic $p$-Laplace equations and best constants in Sobolev inequalities on nilpotent Lie groups. We prove solvability of these subelliptic $p$-Laplace equations and existence of the minimizer of the corresponding variational problem. It leads to existence of the best constant in the corresponding $(q,p)$-Sobolev inequality, $06 galaxies through ALMA observations of [CII] " We study the complexity of a class of problems involving satisfying constraints which remain the same under translations in one or more spatial directions. In this paper, we show hardness of a classical tiling problem on an N x N 2-dimensional grid and a quantum problem involving finding the ground state energy of a 1-dimensional quantum system of N particles. In both cases, the only input is N, provided in binary. We show that the classical problem is NEXP-complete and the quantum problem is QMA_EXP-complete. Thus, an algorithm for these problems which runs in time polynomial in N (exponential in the input size) would imply that EXP = NEXP or BQEXP = QMA_EXP, respectively. Although tiling in general is already known to be NEXP-complete, to our knowledge, all previous reductions require that either the set of tiles and their constraints or some varying boundary conditions be given as part of the input. In the problem considered here, these are fixed, constant-sized parameters of the problem. Instead, the problem instance is encoded solely in the size of the system. ",The Quantum and Classical Complexity of Translationally Invariant Tiling and Hamiltonian Problems " In this paper we extend to non-compact Riemannian manifolds with boundary the use of two important tools in the geometric analysis of compact spaces, namely, the weak maximum principle for subharmonic functions and the integration by parts. The first one is a new form of the classical Ahlfors maximum principle whereas the second one is a version for manifolds with boundary of the so called Kelvin-Nevanlinna-Royden criterion of parabolicity. In fact, we will show that the validity of non-compact versions of these tools serve as a characterization of the Neumann parabolicity of the space. The motivation underlying this study is to obtain new information on the geometry of graphs with prescribed mean curvature inside a Riemannian product of the type $N\times\mathbb{R}$. In this direction two kind of results will be presented: height estimates for constant mean curvature graphs parametrized over unbounded domains in a complete manifold and slice type results for graphs whose superlevel sets have finite volume. ",Global maximum principles and divergence theorems on complete manifolds with boundary " We develop an approach, by calculating the autocorrelation function of spins, to derive the magnetic field gradient induced transverse ($T_2$) relaxation of spins undergoing restricted diffusion. This approach is an extension to the method adopted by McGregor. McGregor's approach solves the problem only in the fast diffusion limit; however, our approach yields a single analytical solution suitable in all diffusion regimes, including the intermediate regime. This establishes a direct connection between the well-known Torrey's slow diffusion result and the fast diffusion result. We also perform free induction decay measurements on spin-exchange optically polarized $^3$He gas with different diffusion constants. The transverse relaxation profiles are compared with the theory and satisfactory agreement has been found throughout all diffusion regimes. In addition to the transverse relaxation, this approach is also applicable to solving the longitudinal relaxation ($T_1$) regardless of the diffusion limits. It turns out that the longitudinal relaxation in the slow diffusion limit differs by a factor of two, compared with that in the fast diffusion limit. ",General Solution to Gradient Induced Transverse and Longitudinal Relaxation of Spins Undergoing Restricted Diffusion " The present paper provides exact expressions for the probability distributions of linear functionals of the two-parameter Poisson--Dirichlet process $\operatorname {PD}(\alpha,\theta)$. We obtain distributional results yielding exact forms for density functions of these functionals. Moreover, several interesting integral identities are obtained by exploiting a correspondence between the mean of a Poisson--Dirichlet process and the mean of a suitable Dirichlet process. Finally, some distributional characterizations in terms of mixture representations are proved. The usefulness of the results contained in the paper is demonstrated by means of some illustrative examples. Indeed, our formulae are relevant to occupation time phenomena connected with Brownian motion and more general Bessel processes, as well as to models arising in Bayesian nonparametric statistics. ",Distributions of linear functionals of two parameter Poisson--Dirichlet random measures " In this letter we derive a new method for direct calculation of the modal contributions to thermal conductivity, which is termed Green-Kubo modal analysis (GKMA). The GKMA method combines the lattice dynamics formalism with the Green-Kubo formula for thermal conductivity, such that the thermal conductivity becomes a direct summation of modal contributions, where one need not define the phonon velocity. As a result the GKMA method can be applied to any material/group of atoms where the atoms vibrate around stable equilibrium positions, which includes not only crystalline line compounds, but also random alloys, amorphous materials and even molecules. The GKMA method provides new insight into the nature of phonon transport, as it casts the problem in terms of mode-mode correlation instead of scattering, and provides a general unified formalism that can be used to understand phonon-phonon interactions in essentially any class of materials or structures where the atoms vibrate around stable equilibrium sites. ",Direct Calculation of Modal Contributions to Thermal Conductivity via Green-Kubo Modal Analysis: Crystalline and Amorphous Silicon " We consider the Euler--Darboux equation with parameters modulo 1/2 and generalization to the space 3D analogue. Due to the fact that the Cauchy problem in its classical formulation is incorrect for such parameter values, the authors propose formulations and solutions of modified Cauchy-type problems with parameter values: a) $\alpha=\beta=\displaystyle\frac{1}{2}$, b) $\alpha=-\,\displaystyle\frac{1}{2}$, $\beta=+\,\displaystyle\frac{1}{2}$, c) $\alpha=\beta=-\,\displaystyle\frac{1}{2}$. The obtained result is used to formulate an analogue of the $\Delta_1$ problem in the first quadrant with the setting of boundary conditions with displacement on the coordinate axes and non-standard conjugation conditions on the singularity line of the coefficients of the equation $y=x$. The first of these conditions glues the normal derivatives of the desired solution, the~second contains the limit values of the combination of the solution and its normal derivatives. The problem was reduced to a uniquely solvable system of integral equations. ",The modified problems for the equation of Euler--Darboux in the case of parameters on the module equal to 1/2 " Inversion by Direct Iteration (InDI) is a new formulation for supervised image restoration that avoids the so-called ``regression to the mean'' effect and produces more realistic and detailed images than existing regression-based methods. It does this by gradually improving image quality in small steps, similar to generative denoising diffusion models. Image restoration is an ill-posed problem where multiple high-quality images are plausible reconstructions of a given low-quality input. Therefore, the outcome of a single step regression model is typically an aggregate of all possible explanations, therefore lacking details and realism. The main advantage of InDI is that it does not try to predict the clean target image in a single step but instead gradually improves the image in small steps, resulting in better perceptual quality. While generative denoising diffusion models also work in small steps, our formulation is distinct in that it does not require knowledge of any analytic form of the degradation process. Instead, we directly learn an iterative restoration process from low-quality and high-quality paired examples. InDI can be applied to virtually any image degradation, given paired training data. In conditional denoising diffusion image restoration the denoising network generates the restored image by repeatedly denoising an initial image of pure noise, conditioned on the degraded input. Contrary to conditional denoising formulations, InDI directly proceeds by iteratively restoring the input low-quality image, producing high-quality results on a variety of image restoration tasks, including motion and out-of-focus deblurring, super-resolution, compression artifact removal, and denoising. ",Inversion by Direct Iteration: An Alternative to Denoising Diffusion for Image Restoration " A good correlation has been found in star-forming galaxies between the soft X-ray and the far infrared or radio luminosities. In this work we analyze the relation between the soft X-ray and far infrared luminosities as predicted by evolutionary population synthesis models, aiming first to test the validity of the soft X-ray luminosity as a star formation rate estimator, using the already known calibration of the FIR luminosity as a proxy, and second proposing a calibration based on the predictions of evolutionary synthesis models. We have computed the soft X-ray and far infrared luminosities expected for a massive starburst as a function of evolutionary state, efficiency of the conversion of mechanical energy into soft X-ray luminosity, star formation regime (instantaneous or extended) and dust abundance, and have compared the predictions with observational values for a sample of 62 star-forming galaxies taken from the literature. The observational Lsoftx/Lfir ratios are consistent with the predictions by the models under realistic assumptions (young starbursts, efficiency in the re-processing of mechanical energy of few percent), confirming the direct relation of the diffuse soft X-ray emission with the star formation episode. The soft X-ray emission of the diffuse, extended gas surrounding massive star forming regions can be used as a Star Formation Rate estimator. The empirical calibrations presented in the literature are supported by the predictions of evolutionary synthesis models and by the analysis of a larger sample of star-forming galaxies, but it is biased towards galaxies dominated by relatively unevolved starbursts. ",Soft X-ray to Far Infrared luminosities ratio in star-forming galaxies " We have found one of the possible conditions under which the Beltrami equation with degeneration of ellipticity has a continuous solution of the Sobolev class. With some additional requirements, this solution is homeomorphic ",On the existence of a solution of the Beltrami equation with degeneration " As a step towards understanding the in-medium evolution of a hard jet, we consider the interference pattern for the medium induced gluon radiation produced by a color singlet quark-antiquark antenna embedded in a QCD medium with size L. We focus on the typical kinematics for medium-induced gluon radiation in the BDMPS-Z regime, that is, short formation times \tau_f << L and relatively large emission angles \theta >>\theta_c = 2/\sqrt{\hat q L^3}, with \hat q the `jet quenching' parameter. We demonstrate that, for a dipole opening angle \theta_{q\bar q} larger than \theta_c, the interference between the medium-induced gluon emissions by the quark and the antiquark is parametrically suppressed with respect to the corresponding direct emissions. Physically, this is so since the direct emissions can be delocalized anywhere throughout the medium and thus yield contributions proportional to L. On the contrary, the interference occurs only between gluons emitted at very early times, within the characteristic time scales for quantum and color coherence between the two emitters, which in this regime are much smaller than L. This implies that, for \theta_{q\bar q} >> \theta_c, the medium-induced radiation by the dipole is simply the sum of the two BDMPS-Z spectra individually produced by the quark and the antiquark, without coherence effects like angular ordering. For \theta_{q\bar q} << \theta_c, the BDMPS-Z radiation by the dipole vanishes. ",Interference effects in medium-induced gluon radiation " A model for observable effects of electromagnetic vacuum fluctuations is presented. The model involves a probe pulse which traverses a slab of nonlinear optical material with a nonzero second order polarizability. We argue that the pulse interacts with the ambient vacuum fluctuations of other modes of the quantized electric field, and these vacuum fluctuations cause variations in the flight time of the pulse through the material. The geometry of the slab of material defines a sampling function for the quantized electric field, which in turn determines that vacuum modes whose wavelengths are of the order of the thickness of the slab give the dominant contribution. Some numerical estimates are made, which indicate that fractional fluctuations in flight time of the order of $10^{-9}$ are possible in realistic situations. The model presented here is both an illustration of a physical effect of vacuum fluctuations, and an analog model for the lightcone fluctuations predicted by quantum gravity. ",Vacuum Lightcone Fluctuations in a Dielectric " In this paper a fairly complete mathematical model of CP-PLL, which reliable enough to serve as a tool for credible analysis of dynamical properties of these circuits, is studied. We refine relevant mathematical definitions of the hold-in and pull-in ranges related to the local and global stability. Stability analysis of the steady state for the charge-pump phase locked loop is non-trivial: straight-forward linearization of available CP-PLL models may lead to incorrect conclusions, because the system is not smooth near the steady state and may experience overload. In this work necessary details for local stability analysis are presented and the hold-in range is computed. An upper estimate of the pull-in range is obtained via the analysis of limit cycles. The study provided an answer to Gardner's conjecture on the similarity of transient responses of CP-PLL and equivalent classical PLL and to conjectures on the infinite pull-in range of CP-PLL with proportionally-integrating filter. ",Nonlinear analysis of charge-pump phase-locked loop: the hold-in and pull-in ranges " Plasmoid-mediated reconnection is investigated for nonlinear Peeling-Ballooning Edge-Localized Modes (P-B ELMs). The formation of current sheets and the transition to 3-D current sheet instability is demonstrated through fully nonlinear resistive MHD simulations of P-B ELMs in DIII- D discharges. Large-scale axisymmetric current sheets, as well as small-scale poloidally extending current sheets, are formed as the coherent P-B ELM filaments nonlinearly evolve. It is observed that, at high Lundquist numbers, these current sheets break during a reconnection burst, i.e. a secondary exponential growth of intermediate modes followed by relaxation due to the suppression of P-B drive. ",Plasmoid-mediated reconnection during nonlinear peeling-ballooning edge-localized modes " We perform direct numerical simulations of the equations of magnetohydrodynamics with external random forcing and in the presence of gravity. The domain is divided into two parts: a lower layer where the forcing is helical and an upper layer where the helicity of the forcing is zero with a smooth transition in between. At early times, a large-scale helical dynamo develops in the bottom layer. At later times the dynamo saturates, but the vertical magnetic field continues to develop and rises to form dynamic bipolar structures at the top, which later disappear and reappear. Some of the structures look similar to $\delta$ spots observed in the Sun. This is the first example of magnetic flux concentrations, owing to strong density stratification, from self-consistent dynamo simulations that generate bipolar, super-equipartition strength, magnetic structures whose energy density can exceeds the turbulent kinetic energy by even a factor of ten. ",Intense bipolar structures from stratified helical dynamos " For a square-free bivariate polynomial $p$ of degree $n$ we introduce a simple and fast numerical algorithm for the construction of $n\times n$ matrices $A$, $B$, and $C$ such that $\det(A+xB+yC)=p(x,y)$. This is the minimal size needed to represent a bivariate polynomial of degree $n$. Combined with a square-free factorization one can now compute $n \times n$ matrices for any bivariate polynomial of degree $n$. The existence of such symmetric matrices was established by Dixon in 1902, but, up to now, no simple numerical construction has been found, even if the matrices can be nonsymmetric. Such representations may be used to efficiently numerically solve a system of two bivariate polynomials of small degree via the eigenvalues of a two-parameter eigenvalue problem. The new representation speeds up the computation considerably. ",Minimal determinantal representations of bivariate polynomials " We investigate thoroughly the temporal evolution of the universe temperature as a function of the Hubble parameter associated with the Stochastic Gravitational Wave (SGW), that formed at the cosmological QCD phase transition epoch to the current epoch, within the Generalized Uncertainty Principle (GUP) framework. Here we use GUP version which provide constraints on the minimum measurable length and the maximum observable momentum, that characterized by a free parameter $\alpha$. We study the effect of this parameter on the SGW background. We show that the effect can slightly enhance the SGW frequency in the lower frequency regime which might be important in the detection of SGW in the future GW detection facilities. ","Minimal length, maximal momentum and stochastic gravitational waves spectrum generated from cosmological QCD phase transition" " In (2+1)-dimensional general relativity, the path integral for a manifold $M$ can be expressed in terms of a topological invariant, the Ray-Singer torsion of a flat bundle over $M$. For some manifolds, this makes an explicit computation of transition amplitudes possible. In this paper, we evaluate the amplitude for a simple topology-changing process. We show that certain amplitudes for spatial topology change are nonvanishing---in fact, they can be infrared divergent---but that they are infinitely suppressed relative to similar topology-preserving amplitudes. ",Topology Change in (2+1)-Dimensional Gravity " A set of functions is introduced which generalizes the famous Schur polynomials and their connection to Grasmannian manifolds. These functions are shown to provide a new method of constructing solutions to the KP hierarchy of nonlinear partial differential equations. Specifically, just as the Schur polynomials are used to expand tau-functions as a sum, it is shown that it is natural to expand a quotient of tau-functions in terms of these generalized Schur functions. The coefficients in this expansion are found to be constrained by the Pl\""ucker relations of a grassmannian. ","Grassmannians, Nonlinear Wave Equations and Generalized Schur Functions" " Let $G$ be a graph, and let $f_G$ be the sum of $(-1)^{|A|}$, over all stable sets $A$. If $G$ is a cycle with length divisible by three, then $f_G= \pm 2$. Motivated by topological considerations, G. Kalai and R. Meshulam made the conjecture that,if no induced cycle of a graph $G$ has length divisible by three, then $|f_G|\le 1$. We prove this conjecture. ",Proof of the Kalai-Meshulam conjecture " Deep convolutional neural networks (CNNs) obtain outstanding results in tasks that require human-level understanding of data, like image or speech recognition. However, their computational load is significant, motivating the development of CNN-specialized accelerators. This work presents NEURAghe, a flexible and efficient hardware/software solution for the acceleration of CNNs on Zynq SoCs. NEURAghe leverages the synergistic usage of Zynq ARM cores and of a powerful and flexible Convolution-Specific Processor deployed on the reconfigurable logic. The Convolution-Specific Processor embeds both a convolution engine and a programmable soft core, releasing the ARM processors from most of the supervision duties and allowing the accelerator to be controlled by software at an ultra-fine granularity. This methodology opens the way for cooperative heterogeneous computing: while the accelerator takes care of the bulk of the CNN workload, the ARM cores can seamlessly execute hard-to-accelerate parts of the computational graph, taking advantage of the NEON vector engines to further speed up computation. Through the companion NeuDNN SW stack, NEURAghe supports end-to-end CNN-based classification with a peak performance of 169 Gops/s, and an energy efficiency of 17 Gops/W. Thanks to our heterogeneous computing model, our platform improves upon the state-of-the-art, achieving a frame rate of 5.5 fps on the end-to-end execution of VGG-16, and 6.6 fps on ResNet-18. ",NEURAghe: Exploiting CPU-FPGA Synergies for Efficient and Flexible CNN Inference Acceleration on Zynq SoCs These notes aim to give a first introduction to intersection cohomology and perverse sheaves with applications to representation theory or quantum groups in mind. ,An introduction to perverse sheaves " We study the dynamics of self-trapping in Bose-Einstein condensates (BECs) loaded in deep optical lattices with Gaussian initial conditions, when the dynamics is well described by the Discrete Nonlinear Schr\""odinger Equation (DNLS). In the literature an approximate dynamical phase diagram based on a variational approach was introduced to distinguish different dynamical regimes: diffusion, self-trapping and moving breathers. However, we find that the actual DNLS dynamics shows a completely different diagram than the variational prediction. We numerically calculate a detailed dynamical phase diagram accurately describing the different dynamical regimes. It exhibits a complex structure which can readily be tested in current experiments in BECs in optical lattices and in optical waveguide arrays. Moreover, we derive an explicit theoretical estimate for the transition to self-trapping in excellent agreement with our numerical findings, which may be a valuable guide as well for future studies on a quantum dynamical phase diagram based on the Bose-Hubbard Hamiltonian. ",Dynamical phase diagram of Gaussian BEC wave packets in optical lattices " An extra Abelian gauge symmetry is motivated in many new physics models in both supersymmetric and nonsupersymmetric cases. Such a new gauge symmetry may interact with both the observable sector and the hidden sector. We systematically investigate the most general residual discrete symmetries in both sectors from a common Abelian gauge symmetry. Those discrete symmetries can ensure the stability of the proton and the dark matter candidate. A hidden sector dark matter candidate (lightest U-parity particle or LUP) interacts with the standard model fields through the gauge boson Z', which may selectively couple to quarks or leptons only. We make a comment on the implications of the discrete symmetry and the leptonically coupling dark matter candidate, which has been highlighted recently due to the possibility of the simultaneous explanation of the DAMA and the PAMELA results. We also show how to construct the most general U(1) charges for a given discrete symmetry, and discuss the relation between the U(1) gauge symmetry and R-parity. ",Common gauge origin of discrete symmetries in observable sector and hidden sector " The canonical commutation relation is the hallmark of quantum theory and Heisenberg's uncertainty relation is a direct consequence of it. But despite its fundamental role in quantum theory, surprisingly, its genuine direct experimental test has hitherto not been performed. In this article, we present a novel scheme to directly test the canonical commutation relation between two dichotomic observables, by exploiting the notion of weak measurement. The imaginary part of a suitably formulated weak value enables this direct test. The measurement of the weak value of a path-qubit observable in a neutron interferometer experiment is used to verify the approach. The experiment is realized using a newly developed technique in our neutron interferometric setup where the neutron's spin/energy degree of freedom serves as ancilla. ",Direct experimental test of commutation relation via weak value We consider interactions of fermions with the domain wall bubbles produced during a first order phase transition. A new exact solution of the Dirac equations is obtained for a wall profile incorporating a position dependent phase factor. The reflection coefficients are obtained. ,Fermion Scattering in domain walls with a locally dependent phase " Ensemble data assimilation in flood forecasting depends strongly on the density, frequency and statistics of errors associated with the observation network. This work focuses on the assimilation of 2D flood extent data, expressed in terms of wet surface ratio, in addition to the in-situ water level data. The objective is to improve the representation of the flood plain dynamics with a TELEMAC-2D model and an Ensemble Kalman Filter (EnKF). The EnKF control vector is composed of friction coefficients and corrective parameters to the input forcing. It is augmented with the water level state averaged over selected subdomains of the floodplain. This work focuses on the 2019 flood event that occurred over the Garonne Marmandaise catchment. The merits of assimilating SAR-derived flood plain data complementary to in-situ water level observations are shown in the control parameter and observation spaces with 1D and 2D assessment metrics. It was also shown that the assimilation of Wet surface Ratio in the flood plain complementary to in-situ data in the river bed brings significative improvement when a corrective term on flood plain hydraulic state is included in the control vector. Yet, it has barely no impact in the river bed that is sufficiently well described by in-situ data. We highlighted that the correction of the hydraulic state in the flood plain significantly improved the flood dynamics, especially during the recession. This proof-of-concept study paves the way towards near-real-time flood forecast, making the most of remote sensing-derived flood observations. ",Enhancing Flood Forecasting with Dual State-Parameter Estimation and Ensemble-based SAR Data Assimilation " If $\phi$ is an automorphism of a group $G$ and $x,y\in G$, we say that $x$ and $y$ are $\phi$-twisted conjugates if there exists an $z\in G$ such that $y=z.x.\phi(z^{-1})$. This is an equivalence relation. If there are infinitely many $\phi$-twisted conjugacy classes for every automorphism $\phi$ of $G$ we say that $G$ has the $R_\infty$-property. We prove that the generalized Richard Thompson groups $F_n$ and $F(l,A,P)$ have the $R_\infty$-property. ",Twisted conjugacy in generalized Thompson groups of type F " In this note we investigate the structure of the space $\Jj$ of smooth almost complex structures on $S^2\times S^2$ that are compatible with some symplectic form. This space has a natural stratification that changes as the cohomology class of the form changes and whose properties are very closely connected to the topology of the group of symplectomorphisms of $S^2\times S^2$. By globalizing standard gluing constructions in the theory of stable maps, we show that the strata of $\Jj$ are Fr\'echet manifolds of finite codimension, and that the normal link of each stratum is a finite dimensional stratified space. The topology of these links turns out to be surprisingly intricate, and we work out certain cases. Our arguments apply also to other ruled surfaces, though they give complete information only for bundles over $S^2$ and $T^2$. ",Almost Complex Structures on $S^2\times S^2$ " We explore the statistical properties of energy transfer in ensembles of doubly-driven Random- Matrix Floquet Hamiltonians, based on universal symmetry arguments. The energy pumping efficiency distribution P(E) is associated with the Hamiltonian parameter ensemble and the eigenvalue statistics of the Floquet operator. For specific Hamiltonian ensembles, P(E) undergoes a transition that cannot be associated with a symmetry breaking of the instantaneous Hamiltonian. The Floquet eigenvalue spacing distribution indicates the considered ensembles constitute generic nonintegrable Hamiltonian families. As a step towards Hamiltonian engineering, we develop a machine-learning classifier to understand the relative parameter importance in resulting high conversion efficiency. We propose Random Floquet Hamiltonians as a general framework to investigate frequency conversion effects in a new class of generic dynamical processes beyond adiabatic pumps. ",Energy Transfer in Random-Matrix ensembles of Floquet Hamiltonians " We propose a learning-based method for Lyapunov stability analysis of piecewise affine dynamical systems in feedback with piecewise affine neural network controllers. The proposed method consists of an iterative interaction between a learner and a verifier, where in each iteration, the learner uses a collection of samples of the closed-loop system to propose a Lyapunov function candidate as the solution to a convex program. The learner then queries the verifier, which solves a mixed-integer program to either validate the proposed Lyapunov function candidate or reject it with a counterexample, i.e., a state where the stability condition fails. This counterexample is then added to the sample set of the learner to refine the set of Lyapunov function candidates. We design the learner and the verifier based on the analytic center cutting-plane method, in which the verifier acts as the cutting-plane oracle to refine the set of Lyapunov function candidates. We show that when the set of Lyapunov functions is full-dimensional in the parameter space, the overall procedure finds a Lyapunov function in a finite number of iterations. We demonstrate the utility of the proposed method in searching for quadratic and piecewise quadratic Lyapunov functions. ",Learning Lyapunov Functions for Piecewise Affine Systems with Neural Network Controllers " Information-theoretic security--widely accepted as the strictest notion of security--relies on channel coding techniques that exploit the inherent randomness of propagation channels to strengthen the security of communications systems. Within this paradigm, we explore strategies to improve secure connectivity in a wireless network. We first consider the intrinsically secure communications graph (iS-graph), a convenient representation of the links that can be established with information-theoretic security on a large-scale network. We then propose and characterize two techniques--sectorized transmission and eavesdropper neutralization--which are shown to dramatically enhance the connectivity of the iS-graph. ",Techniques for Enhanced Physical-Layer Security " We present a logical calculus for reasoning about information flow in quantum programs. In particular we introduce a dynamic logic that is capable of dealing with quantum measurements, unitary evolutions and entanglements in compound quantum systems. We give a syntax and a relational semantics in which we abstract away from phases and probabilities. We present a sound proof system for this logic, and we show how to characterize by logical means various forms of entanglement (e.g. the Bell states) and various linear operators. As an example we sketch an analysis of the teleportation protocol. ",The Logic of Quantum Programs " We show that a search for $e^+ e^- \to t \bar q$ $(\bar q = \bar c, \bar u)$ events at LEP--2 collider provide a possibility to improve significantly the modern constraints on coupling constants of anomalous $t$--quark interaction via flavor--changing neutral currents. ",Search for Anomalous Top--Quark Interaction at LEP-2 Collider " In this paper, a time-varying distributed convex optimization problem is studied for continuous-time multi-agent systems. Control algorithms are designed for the cases of single-integrator and double-integrator dynamics. Two discontinuous algorithms based on the signum function are proposed to solve the problem in each case. Then in the case of double-integrator dynamics, two continuous algorithms based on, respectively, a time-varying and a fixed boundary layer are proposed as continuous approximations of the signum function. Also, to account for inter-agent collision for physical agents, a distributed convex optimization problem with swarm tracking behavior is introduced for both single-integrator and double-integrator dynamics. ",Distributed Convex Optimization for Continuous-Time Dynamics with Time-Varying Cost Function " In this paper, Hamiltonian and energy preserving reduced-order models are developed for the rotating thermal shallow water equation (RTSWE) in the non-canonical Hamiltonian form with the state-dependent Poisson matrix. The high fidelity full solutions are obtained by discretizing the RTSWE in space with skew-symmetric finite-differences, that preserve the Hamiltonian structure. The resulting skew-gradient system is integrated in time with the energy preserving average vector field (AVF) method. The reduced-order model (ROM) is constructed in the same way as the full order model (FOM), preserving the reduced skew-symmetric structure and integrating in time with the AVF method. Relying on structure-preserving discretizations in space and time and applying proper orthogonal decomposition (POD) with the Galerkin projection, an energy preserving reduced order model (ROM) is constructed. The nonlinearities in the ROM are computed by applying the discrete empirical interpolation (DEIM) method to reduce the computational cost. The computation of the reduced-order solutions is accelerated further by the use of tensor techniques. The overall procedure yields a clear separation of the offline and online computational cost of the reduced solutions. The accuracy and computational efficiency of the ROMs are demonstrated for a numerical test problem. Preservation of the energy (Hamiltonian), and other conserved quantities, i.e. mass, buoyancy, and total vorticity show that the reduced-order solutions ensure the long-term stability of the solutions while exhibiting several orders of magnitude computational speedup over the FOM. ",Energy preserving reduced-order modelling of thermal shallow water equation " A new class UF of problems is introduced, strictly included in the class NP, which arises in the analysis of the time verifying the intermediate results of computations. The implications of the introduction of this class are considered. First of all, we prove that $P\not= NP$ and establish that it needs to consider the problem ""P vs UF"" instead the problem ""P vs NP"". Also, we determine the set-theoretical of properties of a one-way functions that used in cryptology. ",On the structure of the class NP " The Accurate Particle Tracer (APT) code is designed for large-scale particle simulations on dynamical systems. Based on a large variety of advanced geometric algorithms, APT possesses long-term numerical accuracy and stability, which are critical for solving multi-scale and non-linear problems. Under the well-designed integrated and modularized framework, APT serves as a universal platform for researchers from different fields, such as plasma physics, accelerator physics, space science, fusion energy research, computational mathematics, software engineering, and high-performance computation. The APT code consists of seven main modules, including the I/O module, the initialization module, the particle pusher module, the parallelization module, the field configuration module, the external force-field module, and the extendible module. The I/O module, supported by Lua and Hdf5 projects, provides a user-friendly interface for both numerical simulation and data analysis. A series of new geometric numerical methods and key physical problems, such as runaway electrons in tokamaks and energetic particles in Van Allen belt, have been studied using APT. As an important realization, the APT-SW version has been successfully distributed on the world's fastest computer, the Sunway TaihuLight supercomputer, by supporting master-slave architecture of Sunway many-core processors. ",The Accurate Particle Tracer Code " Formulae expressing the trace of the composition of several (up to five) adjoint actions of elements of the Griess algebra of a vertex operator algebra are derived under certain assumptions on the action of the automorphism group. They coincide, when applied to the moonshine module $V^\natural$, with the trace formulae obtained in a different way by S.Norton, and the spectrum of idempotents related to 2A, 2B, 3A and 4A element of the Monster is determined by the representation theory of Virasoro algebra at $c=1/2$, $W_3$ algebra at $c=4/5$ or $W_4$ algebra at $c=1$. The generalization to the trace function on the whole space is also given for the composition of two adjoint actions, which can be used to compute the McKay-Thompson series for a 2A involution of the Monster. ",Norton's Trace Formulae for the Griess Algebra of a Vertex Operator Algebra with Larger Symmetry " A \emph{domatic} (\emph{total domatic}) \emph{$k$-coloring} of a graph $G$ is an assignment of $k$ colors to the vertices of $G$ such that each vertex contains vertices of all $k$ colors in its closed neighborhood (neighborhood). The \emph{domatic} (\emph{total domatic}) \emph{number} of $G$, denoted $d(G)$ ($d_t (G)$), is the maximum $k$ for which $G$ has a domatic (total domatic) $k$-coloring. In this paper, we show that for two non-trivial graphs $G$ and $H$, the domatic and total domatic numbers of their Cartesian product $G \cart H$ is bounded above by $\max\{|V(G)|, |V(H)|\}$ and below by $\max\{d(G), d(H)\}$. Both these bounds are tight for an infinite family of graphs. Further, we show that if $H$ is bipartite, then $d_t(G \cart H)$ is bounded below by $2\min\{d_t(G),d_t(H)\}$ and $d(G \cart H)$ is bounded below by $2\min\{d(G),d_t(H)\}$. These bounds give easy proofs for many of the known bounds on the domatic and total domatic numbers of hypercubes \cite{chen,zel4} and the domination and total domination numbers of hypercubes \cite{har,joh} and also give new bounds for Hamming graphs. We also obtain the domatic (total domatic) number and domination (total domination) number of $n$-dimensional torus $\mathop{\cart}\limits_{i=1}^{n} C_{k_i}$ with some suitable conditions to each $k_i$, which turns out to be a generalization of a result due to Gravier \cite{grav2} %[\emph{Total domination number of grid graphs}, Discrete Appl. Math. 121 (2002) 119-128] and give easy proof of a result due to Klav\v{z}ar and Seifter \cite{sand}. ",On Domatic and Total Domatic Numbers of Product Graphs " To ensure trust in AI models, it is becoming increasingly apparent that evaluation of models must be extended beyond traditional performance metrics, like accuracy, to other dimensions, such as fairness, explainability, adversarial robustness, and distribution shift. We describe an empirical study to evaluate multiple model types on various metrics along these dimensions on several datasets. Our results show that no particular model type performs well on all dimensions, and demonstrate the kinds of trade-offs involved in selecting models evaluated along multiple dimensions. ","An Empirical Study of Accuracy, Fairness, Explainability, Distributional Robustness, and Adversarial Robustness" " In computational materials science, mechanical properties are typically extracted from simulations by means of analysis routines that seek to mimic their experimental counterparts. However, simulated data often exhibit uncertainties that can propagate into final predictions in unexpected ways. Thus, modelers require data analysis tools that (i) address the problems posed by simulated data, and (ii) facilitate uncertainty quantification. In this manuscript, we discuss three case studies in materials modeling where careful data analysis can be leveraged to address specific instances of these issues. As a unifying theme, we highlight the idea that attention to physical and mathematical constraints surrounding the generation of computational data can significantly enhance its analysis. ",The Role of Data Analysis in Uncertainty Quantification: Case Studies for Materials Modeling " We consider the propagation of light in arbitrarily curved step-index optical fibers. Using a multiple-scales approximation scheme, set-up in Fermi normal coordinates, the full vectorial Maxwell equations are solved in a perturbative manner. At leading order, this provides a rigorous derivation of Rytov's law. At next order, we obtain non-trivial dynamics of the electromagnetic field, characterized by two coupling constants, the phase and the polarization curvature moments, which describe the curvature response of the light's phase and its polarization vector, respectively. The latter can be viewed as an inverse spin Hall effect of light, where the direction of propagation is constrained along the optical fiber and the polarization evolves in a frequency-dependent way. ",Polarization transport in optical fibers beyond Rytov's law " The ExoClock project is an inclusive, integrated, and interactive platform that was developed to monitor the ephemerides of the Ariel targets to increase the mission efficiency. The project makes the best use of all available resources, i.e., observations from ground telescopes, mid-time values from the literature and finally, observations from space instruments. Currently, the ExoClock network includes 280 participants with telescopes capable of observing 85\% of the currently known Ariel candidate targets. This work includes the results of $\sim$1600 observations obtained up to the 31st of December 2020 from the ExoClock network. These data in combination with $\sim$2350 mid-time values collected from the literature are used to update the ephemerides of 180 planets. The analysis shows that 40\% of the updated ephemerides will have an impact on future scheduling as either they have a significantly improved precision, or they have revealed biases in the old ephemerides. With the new observations, the observing coverage and rate for half of the planets in the sample has been doubled or more. Finally, from a population perspective, we identify that the differences in the 2028 predictions between the old and the new ephemerides have an STD that is double what is expected from gaussian uncertainties. These findings have implications for planning future observations, where we will need to account for drifts potentially greater than the prediction uncertainties. The updated ephemerides are open and accessible to the wider exoplanet community both from our Open Science Framework (OSF) repository and our website. ",ExoClock project II: A large-scale integrated study with 180 updated exoplanet ephemerides " An effective atomic cross-chain swap protocol is introduced by Herlihy [Herlihy, 2018] as a distributed coordination protocol in order to exchange assets across multiple blockchains among multiple parties. An atomic cross-chain swap protocol guarantees; (1) if all parties conform to the protocol, then all assets are exchanged among parties, (2) even if some parties or coalitions of parties deviate from the protocol, no party conforming to the protocol suffers a loss, and (3) no coalition has an incentive to deviate from the protocol. Herlihy [Herlihy, 2018] invented this protocol by using hashed timelock contracts. A cross-chain swap is modeled as a directed graph D = (V,A). Vertex set V denotes a set of parties and arc set A denotes a set of proposed asset transfers. Herlihy's protocol uses the graph topology and signature information to set appropriate hashed timelock contracts. The space complexity of the protocol (i.e., the total number of bits written in the blockchains in a swap) is O(|A|^2). The local time complexity of the protocol (i.e., the maximum execution time of a contract in a swap to transfer the corresponding asset) is O(|V||L|), where L is a feedback vertex set computed by the protocol. We propose a new atomic cross-chain swap protocol which uses only signature information and improves the space complexity to O(|A||V|) and the local time complexity to O(|V|). ",Atomic Cross-Chain Swaps with Improved Space and Local Time Complexity " Existing face completion solutions are primarily driven by end-to-end models that directly generate 2D completions of 2D masked faces. By having to implicitly account for geometric and photometric variations in facial shape and appearance, such approaches result in unrealistic completions, especially under large variations in pose, shape, illumination and mask sizes. To alleviate these limitations, we introduce 3DFaceFill, an analysis-by-synthesis approach for face completion that explicitly considers the image formation process. It comprises three components, (1) an encoder that disentangles the face into its constituent 3D mesh, 3D pose, illumination and albedo factors, (2) an autoencoder that inpaints the UV representation of facial albedo, and (3) a renderer that resynthesizes the completed face. By operating on the UV representation, 3DFaceFill affords the power of correspondence and allows us to naturally enforce geometrical priors (e.g. facial symmetry) more effectively. Quantitatively, 3DFaceFill improves the state-of-the-art by up to 4dB higher PSNR and 25% better LPIPS for large masks. And, qualitatively, it leads to demonstrably more photorealistic face completions over a range of masks and occlusions while preserving consistency in global and component-wise shape, pose, illumination and eye-gaze. ",3DFaceFill: An Analysis-By-Synthesis Approach to Face Completion " Distinct-$n$ score\cite{Li2016} is a widely used automatic metric for evaluating diversity in language generation tasks. However, we observed that the original approach for calculating distinct scores has evident biases that tend to assign higher penalties to longer sequences. We refine the calculation of distinct scores by scaling the number of distinct tokens based on their expectations. We provide both empirical and theoretical evidence to show that our method effectively removes the biases existing in the original distinct score. Our experiments show that our proposed metric, \textit{Expectation-Adjusted Distinct (EAD)}, correlates better with human judgment in evaluating response diversity. To foster future research, we provide an example implementation at \url{https://github.com/lsy641/Expectation-Adjusted-Distinct}. ",Rethinking and Refining the Distinct Metric " We introduce a new measurable quantity, $R_{\Delta \phi}$, for studies of the rapidity and transverse momentum dependence of dijet azimuthal decorrelations in hadron-hadron collisions. In pQCD, $R_{\Delta \phi}$ is computed as a ratio of three-jet and dijet cross sections in which the parton distribution functions cancel to a large extent. At the leading order, $R_{\Delta \phi}$ is proportional to $\alpha_s$, and the transverse momentum dependence of can therefore be exploited to determine $\alpha_s$. We compute the NLO pQCD theory predictions and non-perturbative corrections for $R_{\Delta \phi}$ at the LHC and the Tevatron and investigate the corresponding uncertainties. From this, we estimate the theory uncertainties for $\alpha_s$ determinations based on $R_{\Delta \phi}$ at both colliders. The potential of $R_{\Delta \phi}$ measurements for tuning Monte Carlo event generators is also demonstrated. ",A new quantity for studies of dijet azimuthal decorrelations The precise knowledge of parton distribution functions (PDFs) is indispensable to the accurate calculation of hadron-initiated QCD hard scattering observables. Much of our information on PDFs is extracted by comparing deep inelastic scattering (DIS) and lepton pair production (LPP) data to convolutions of the PDFs with the partonic cross sections of these processes. It is known that partonic cross sections receive large corrections in regions of phase space near partonic thresholds that can be resummed using threshold resummation techniques. The effect of threshold resummation on DIS and LPP differs because partonic thresholds for the two processes occur in different kinematic regions. Recent global fits for PDFs have included DIS data from the large Bjorken x and moderate Q^2 region where threshold effects have been shown to be large. The present project explores the effects of simultaneously incorporating threshold resummation in both DIS and LPP and to evaluate the effects of such additions on global fits. ,Threshold Resummation and the Determinations of Parton Distribution Functions " Traditional Chinese medicine (TCM) relies on natural medical products to treat symptoms and diseases. While clinical data have demonstrated the effectiveness of selected TCM-based treatments, the mechanistic root of how TCM herbs treat diseases remains largely unknown. More importantly, current approaches focus on single herbs or prescriptions, missing the high-level general principles of TCM. To uncover the mechanistic nature of TCM on a system level, in this work we establish a generic network medicine framework for TCM from the human protein interactome. Applying our framework reveals a network pattern between symptoms (diseases) and herbs in TCM. We first observe that genes associated with a symptom are not distributed randomly in the interactome, but cluster into localized modules; furthermore, a short network distance between two symptom modules is indicative of the symptoms' co-occurrence and similarity. Next, we show that the network proximity of a herb's targets to a symptom module is predictive of the herb's effectiveness in treating the symptom. We validate our framework with real-world hospital patient data by showing that (1) shorter network distance between symptoms of inpatients correlates with higher relative risk (co-occurrence), and (2) herb-symptom network proximity is indicative of patients' symptom recovery rate after herbal treatment. Finally, we identified novel herb-symptom pairs in which the herb's effectiveness in treating the symptom is predicted by network and confirmed in hospital data, but previously unknown to the TCM community. These predictions highlight our framework's potential in creating herb discovery or repurposing opportunities. In conclusion, network medicine offers a powerful novel platform to understand the mechanism of traditional medicine and to predict novel herbal treatment against diseases. ",Network medicine framework reveals generic herb-symptom effectiveness of Traditional Chinese Medicine " In this paper, we study the competition between finite-size effects (i.e. discernibility of particles) and dipole-dipole interactions in few-atom systems coupled to the electromagnetic field in vacuum. We consider two hallmarks of cooperative effects, superradiance and subradiance, and compute for each the rate of energy radiated by the atoms and the coherence of the atomic state during the time evolution. We adopt a statistical approach in order to extract the typical behavior of the atomic dynamics and average over random atomic distributions in spherical containers with prescribed $k_0R$ with $k_0$ the radiation wavenumber and $R$ the average interatomic distance. Our approach allows us to highlight the tradeoff between finite-size effects and dipole-dipole interactions in superradiance/subradiance. In particular, we show the existence of an optimal value of $k_0R$ for which the superradiant intensity and coherence pulses are the less affected by dephasing effects induced by dipole-dipole interactions and finite-size effects. ",Competition between finite-size effects and dipole-dipole interactions in few-atom systems " Gravitational-wave memory refers to the permanent displacement of the test masses in an idealized (freely-falling) gravitational-wave interferometer. Inspiraling binaries produce a particularly interesting form of memory--the Christodoulou memory. Although it originates from nonlinear interactions at 2.5 post-Newtonian order, the Christodoulou memory affects the gravitational-wave amplitude at leading (Newtonian) order. Previous calculations have computed this non-oscillatory amplitude correction during the inspiral phase of binary coalescence. Using an ""effective-one-body"" description calibrated with the results of numerical relativity simulations, the evolution of the memory during the inspiral, merger, and ringdown phases, as well as the memory's final saturation value, are calculated. Using this model for the memory, the prospects for its detection are examined, particularly for supermassive black hole binary coalescences that LISA will detect with high signal-to-noise ratios. Coalescing binary black holes also experience center-of-mass recoil due to the anisotropic emission of gravitational radiation. These recoils can manifest themselves in the gravitational-wave signal in the form of a ""linear"" memory and a Doppler shift of the quasi-normal-mode frequencies. The prospects for observing these effects are also discussed. ",Gravitational-wave memory revisited: memory from the merger and recoil of binary black holes " The source HD172555 is a young A7V star surrounded by a debris disk with a gaseous component. Here, we present the detection of variable absorption features detected simultaneously in the Ca II K and H doublet lines (at 3,933 and 3,968 Angstrom). We identified the presence of these absorption signatures at four different epochs in the 129 HARPS high-resolution spectra gathered between 2004 and 2011. These transient absorption features are most likely due to Falling Evaporating Bodies (FEBs, or exocomets) that produce absorbing gas observed transiting in front of the central star. We also detect a stable Ca II absorption component at the star's radial velocity. With no corresponding detection in the Na I line, the resulting very low upper limit for the NaI/CaII ratio suggests that this absorption is due to circumstellar gas. ",Exocomets in the circumstellar gas disk of HD 172555 " We study the slowly varying, non-autonomous quantum dynamics of a translation invariant spin or fermion system on the lattice $\mathbb Z^d$. This system is assumed to be initially in thermal equilibrium, and we consider realizations of quasi-static processes in the adiabatic limit. By combining the Gibbs variational principle with the notion of quantum weak Gibbs states introduced in [Jak\v{s}i\'c, Pillet, Tauber, arXiv:2204.00440], we establish a number of general structural results regarding such realizations. In particular, we show that such a quasi-static process is incompatible with the property of approach to equilibrium studied in this previous work. ",A note on adiabatic time evolution and quasi-static processes in translation-invariant quantum systems " We consider certain elliptic modular graph functions that arise in the asymptotic expansion around the non--separating node of genus two string invariants that appear in the integrand of the $D^8 R^4$ interaction in the low momentum expansion of the four graviton amplitude in type II superstring theory. These elliptic modular graphs have links given by the Green function, as well its holomorphic and anti--holomorphic derivatives. Using appropriate auxiliary graphs at various intermediate stages of the analysis, we show that each graph can be expressed solely in terms of graphs with links given only by the Green function and not its derivatives. This results in a reduction in the number of basis elements in the space of elliptic modular graphs. ",Relations between elliptic modular graphs " The epoch of galaxy formation provides an important additional test of cosmological theories. Cold-dark-matter (CDM) models with cosmological constant ($\Lambda$) are designed to account for the observed excess power in galaxy distribution, but at the same time suppress the power on small scales pushing galaxy formation to recent epochs. We point out that the recently discovered high redshift galaxy, 53W091, with accurate age measurements (Dunlop et al 1996) provides a particularly important test of these models. In the flat $\Lambda$-dominated Universe, the redshift of formation of 53W091 decreases with decreasing $\Omega$. However, in the modified CDM models decreasing $\Omega$ suppresses the small scale power in the density field and this effect turns out to be dominant. We estimate the mass of the galaxy and show that it represents a very rare and unlikely event in the density field of such models. Similar problems would occur in other modifications of the CDM cosmogonies. ",Modified Cold-Dark-Matter models in light of 53W091 an old galaxy at high z " We show that for a typical coordinate projection of a subgaussian class of functions, the infimum over signs $\inf_{(\epsilon_i)}{\sup_{f\in F}}|{\sum_{i=1}^k\epsilon_i}f(X_i)|$ is asymptotically smaller than the expectation over signs as a function of the dimension $k$, if the canonical Gaussian process indexed by $F$ is continuous. To that end, we establish a bound on the discrepancy of an arbitrary subset of $\mathbb {R}^k$ using properties of the canonical Gaussian process the set indexes, and then obtain quantitative structural information on a typical coordinate projection of a subgaussian class. ","Discrepancy, chaining and subgaussian processes" " Materials with large magnetocrystalline anisotropy and strong electric field effects are highly needed to develop new types of memory devices based on electric field control of spin orientations. Instead of using modified transition metal films, we propose that certain monolayer transition metal dichalcogenides are the ideal candidate materials for this purpose. Using density functional calculations, we show that they exhibit not only a large magnetocrystalline anisotropy (MCA), but also colossal voltage modulation under external field. Notably, in some materials like CrSe_2 and FeSe_2, where spins show a strong preference for in-plane orientation, they can be switched to out-of-plane direction. This effect is attributed to the large band character alteration that the transition metal d-states undergo around the Fermi energy due to the electric field. We further demonstrate that strain can also greatly change MCA, and can help to improve the modulation efficiency while combined with an electric field. ",Voltage-Controllable Colossal Magnetocrystalline Anisotropy in Single Layer Transition Metal Dichalcogenides " Spectroscopic and photometric data for likely member stars of five Galactic globular clusters (M3, M53, M71, M92, and NGC 5053) and three open clusters (M35, NGC 2158, and NGC 6791) are processed by the current version of the SEGUE Stellar Parameter Pipeline (SSPP), in order to determine estimates of metallicities and radial velocities for the clusters. These results are then compared to values from the literature. We find that the mean metallicity (<[Fe/H]>) and mean radial velocity () estimates for each cluster are almost all within 2{\sigma} of the adopted literature values; most are within 1{\sigma}. We also demonstrate that the new version of the SSPP achieves small, but noteworthy, improvements in <[Fe/H]> estimates at the extrema of the cluster metallicity range, as compared to a previous version of the pipeline software. These results provide additional confidence in the application of the SSPP for studies of the abundances and kinematics of stellar populations in the Galaxy. ",The SEGUE Stellar Parameter Pipeline. IV. Validation with an Extended Sample of Galactic Globular and Open Clusters " Identifying potential abuses of human rights through imagery is a novel and challenging task in the field of computer vision, that will enable to expose human rights violations over large-scale data that may otherwise be impossible. While standard databases for object and scene categorisation contain hundreds of different classes, the largest available dataset of human rights violations contains only 4 classes. Here, we introduce the `Human Rights Archive Database' (HRA), a verified-by-experts repository of 3050 human rights violations photographs, labelled with human rights semantic categories, comprising a list of the types of human rights abuses encountered at present. With the HRA dataset and a two-phase transfer learning scheme, we fine-tuned the state-of-the-art deep convolutional neural networks (CNNs) to provide human rights violations classification CNNs (HRA-CNNs). We also present extensive experiments refined to evaluate how well object-centric and scene-centric CNN features can be combined for the task of recognising human rights abuses. With this, we show that HRA database poses a challenge at a higher level for the well studied representation learning methods, and provide a benchmark in the task of human rights violations recognition in visual context. We expect this dataset can help to open up new horizons on creating systems able of recognising rich information about human rights violations. Our dataset, codes and trained models are available online at https://github.com/GKalliatakis/Human-Rights-Archive-CNNs. ",Exploring object-centric and scene-centric CNN features and their complementarity for human rights violations recognition in images " Motivated by a question of A. Rapinchuk concerning general reductive groups, we are investigating the following question: Given a finitely generated integral domain $R$ with field of fractions $F$, is there a \emph{finitely generated subgroup} $\Gamma$ of $SL_2(F)$ containing $SL_2(R)$? We shall show in this paper that the answer to this question is negative for any polynomial ring $R$ of the form $R = R_0[s,t]$, where $R_0$ is a finitely generated integral domain with infinitely many (non--associate) prime elements. The proof applies Bass--Serre theory and reduces to analyzing which elements of $SL_2(R)$ can be generated by elementary matrices with entries in a given finitely generated $R$--subalgbra of $F$. Using Bass--Serre theory, we can also exhibit new classes of rings which do not have the $GE_2$ property introduced by P.M. Cohn. ",On finite and elementary generation of SL_2(R) " Gradient-descent-based algorithms and their stochastic versions have widespread applications in machine learning and statistical inference. In this work we perform an analytic study of the performances of one of them, the Langevin algorithm, in the context of noisy high-dimensional inference. We employ the Langevin algorithm to sample the posterior probability measure for the spiked matrix-tensor model. The typical behaviour of this algorithm is described by a system of integro-differential equations that we call the Langevin state evolution, whose solution is compared with the one of the state evolution of approximate message passing (AMP). Our results show that, remarkably, the algorithmic threshold of the Langevin algorithm is sub-optimal with respect to the one given by AMP. We conjecture this phenomenon to be due to the residual glassiness present in that region of parameters. Finally we show how a landscape-annealing protocol, that uses the Langevin algorithm but violate the Bayes-optimality condition, can approach the performance of AMP. ",Marvels and Pitfalls of the Langevin Algorithm in Noisy High-dimensional Inference " Exponentially stable extended adaptive observer is proposed for a class of linear time-invariant systems with unknown parameters and overparameterization. It allows one to reconstruct unmeasured states and bounded external disturbance produced by a known linear exosystem with unknown initial conditions if a weak requirement of regressor finite excitation is met. In contrast to the existing solutions, the proposed observer reconstructs the original (physical) states of the system rather than the virtual one of its observer canonical form. Simulation results to validate the developed theory are presented. ",Extended Adaptive Observer for Linear Systems with Overparametrization " Using the keyboard LEDs to send data optically was proposed in 2002 by Loughry and Umphress [1] (Appendix A). In this paper we extensively explore this threat in the context of a modern cyber-attack with current hardware and optical equipment. In this type of attack, an advanced persistent threat (APT) uses the keyboard LEDs (Caps-Lock, Num-Lock and Scroll-Lock) to encode information and exfiltrate data from airgapped computers optically. Notably, this exfiltration channel is not monitored by existing data leakage prevention (DLP) systems. We examine this attack and its boundaries for today's keyboards with USB controllers and sensitive optical sensors. We also introduce smartphone and smartwatch cameras as components of malicious insider and 'evil maid' attacks. We provide the necessary scientific background on optical communication and the characteristics of modern USB keyboards at the hardware and software level, and present a transmission protocol and modulation schemes. We implement the exfiltration malware, discuss its design and implementation issues, and evaluate it with different types of keyboards. We also test various receivers, including light sensors, remote cameras, 'extreme' cameras, security cameras, and smartphone cameras. Our experiment shows that data can be leaked from air-gapped computers via the keyboard LEDs at a maximum bit rate of 3000 bit/sec per LED given a light sensor as a receiver, and more than 120 bit/sec if smartphones are used. The attack doesn't require any modification of the keyboard at hardware or firmware levels. ",CTRL-ALT-LED: Leaking Data from Air-Gapped Computers via Keyboard LEDs " The periodic events occurring in Eta Carinae have been widely monitored during the last three 5.5-year cycles. The last one recently occurred in January 2009 and more exhaustive observations have been made at different wavelength ranges. If these events are produced when the binary components approach periastron, the timing and sampling of the photometric features can provide more information about the geometry and physics of the system. Thus, we continued with our ground-based optical photometric campaign started in 2003 to record the behaviour of the 2009.0 event in detail. This time the observation program included a new telescope to obtain information from other photometric bands. The daily monitoring consists of the acquisition of CCD images through standard UBVRI filters and a narrow Halpha passband. The subsequent differential photometry includes the central region of the object and the whole Homunculus nebula. The results of our relative UBVRIHalpha photometry, performed from November 2008 up to the end of March 2009, are presented in this work, which comprises the totality of the event. The initial rising branch, the maximum, the dip to the minimum and the recovering rising phase strongly resemble a kind of eclipse. All these features happened on time - according to that predicted - although there are some photometric differences in comparison with the previous event. We made a new determination of 2022.8 days for the period value using the present and previous ""eclipse-like"" event data. These results strongly support the binarity hypothesis for Eta Car. In this paper, the complete dataset with the photometry of the 2009.0 event is provided to make it readily available for further analysis. ","The Eta Carinae optical 2009.0 event, a new ""eclipse-like"" phenomenon" The main goal of this paper is to prove the following two conjectures for genus up to two: 1. Witten's conjecture on the relations between higher spin curves and Gelfand--Dickey hierarchy. 2. Virasoro conjecture for target manifolds with conformal semisimple Frobenius manifolds. The main technique used in the proof is the invariance of tautological equations under loop group action. ,"Witten's conjecture, Virasoro conjecture, and invariance of tautological equations" " The prediction and selection of lesion features are two important tasks in voxel-based neuroimage analysis. Existing multivariate learning models take two tasks equivalently and optimize simultaneously. However, in addition to lesion features, we observe that there is another type of feature, which is commonly introduced during the procedure of preprocessing steps, which can improve the prediction result. We call such a type of feature as procedural bias. Therefore, in this paper, we propose that the features/voxels in neuroimage data are consist of three orthogonal parts: lesion features, procedural bias, and null features. To stably select lesion features and leverage procedural bias into prediction, we propose an iterative algorithm (termed GSplit LBI) as a discretization of differential inclusion of inverse scale space, which is the combination of Variable Splitting scheme and Linearized Bregman Iteration (LBI). Specifically, with a variable the splitting term, two estimators are introduced and split apart, i.e. one is for feature selection (the sparse estimator) and the other is for prediction (the dense estimator). Implemented with Linearized Bregman Iteration (LBI), the solution path of both estimators can be returned with different sparsity levels on the sparse estimator for the selection of lesion features. Besides, the dense the estimator can additionally leverage procedural bias to further improve prediction results. To test the efficacy of our method, we conduct experiments on the simulated study and Alzheimer's Disease Neuroimaging Initiative (ADNI) database. The validity and the benefit of our model can be shown by the improvement of prediction results and the interpretability of visualized procedural bias and lesion features. ",Leveraging both Lesion Features and Procedural Bias in Neuroimaging: An Dual-Task Split dynamics of inverse scale space " We separate a Bose-Einstein condensate into an array of 2D sheets using a 1D optical lattice, and then excite quantized vibrational motion in the direction normal to the sheets. Collisions between atoms induce vibrational de-excitation, transferring the large excitation energy into back-to-back outgoing atoms, imaged as rings in the 2D plane. The ring diameters correspond to vibrational energy level differences, and edge-on imaging allows identification of the final vibrational states. Time dependence of these data provides a nearly complete characterization of the decay process including the energies, populations, and lifetimes of the lowest two excited vibrational levels. The measured decay rates represent a suppression of collisional de-excitation due to the reduced dimensionality, a matter wave analog to inhibited spontaneous emission. ",Collisional de-excitation in a quasi-2D degenerate Bose gas " In this work we provide a detailed study of the CP violating phase transition (CPPT) which is a new mechanism proposed to produce a baryon asymmetry. This mechanism exploits the Weinberg operator whose coefficient is dynamically realised from the vacuum expectation values (VEVs) of new scalars. In the specific case of the first order phase transition, the scalar VEVs vary in the bubble wall which separates the two phases. This results in a spacetime varying coefficient for the Weinberg operator. The interference of two Weinberg operators at different spacetime points generates a CP asymmetry between lepton and anti-lepton production/annihilation processes, which eventually results in an asymmetry between baryon and anti-baryon number densities in the early Universe. We present the calculation of the lepton asymmetry, based on non-equilibrium quantum field theory methods, in full. We consider the influence of the bubble wall characteristics and the impact of thermal effects on the lepton asymmetry and draw a comparison between the CPPT mechanism and electroweak baryogenesis. ",Leptogenesis via Varying Weinberg Operator: the Closed-Time-Path Approach " We present a unified description of different types of matter-wave solitons that can emerge in quasi one-dimensional spin-orbit coupled (SOC) Bose-Einstein condensates (BECs). This description relies on the reduction of the original two-component Gross-Pitaevskii SOC-BEC model to a single nonlinear Schr\""{o}dinger equation, via a multiscale expansion method. This way, we find approximate bright and dark soliton solutions, for attractive and repulsive interatomic interactions respectively, for different regimes of the SOC interactions. Beyond this, our approach also reveals ""negative mass"" regimes, where corresponding ""negative mass"" bright or dark solitons can exist for repulsive or attractive interactions, respectively. Such a unique opportunity stems from the structure of the excitation spectrum of the SOC-BEC. Numerical results are found to be in excellent agreement with our analytical predictions. ",Positive and negative mass solitons in spin-orbit coupled Bose-Einstein condensates " We report the first search for the Sagittarius tidal stream of axion dark matter around 4.55 $\mu$eV using CAPP-12TB haloscope data acquired in March of 2022. Our result excluded the Sagittarius tidal stream of Dine-Fischler-Srednicki-Zhitnitskii and Kim-Shifman-Vainshtein-Zakharov axion dark matter densities of $\rho_a\gtrsim0.184$ and $\gtrsim0.025$ GeV/cm$^{3}$, respectively, over a mass range from 4.51 to 4.59 $\mu$eV at a 90% confidence level. ",Search for the Sagittarius Tidal Stream of Axion Dark Matter around 4.55 $\mu$eV " We study the question of dividing a collection of indivisible goods amongst a set of agents. The main objective of research in the area is to achieve one of two goals: fairness or efficiency. On the fairness side, envy-freeness is the central fairness criterion in economics, but envy-free allocations typically do not exist when the goods are indivisible. A recent line of research shows that envy-freeness can be achieved if a small quantity of a homogeneous divisible good (money) is introduced into the system, or equivalently, if transfer payments are allowed between the agents. A natural question to explore, then, is whether transfer payments can be used to provide high welfare in addition to envy-freeness, and if so, how much money is needed to be transferred. We show that for general monotone valuations, there always exists an allocation with transfers that is envy-free and whose Nash social welfare (NSW) is at least an $e^{-1/e}$-fraction of the optimal Nash social welfare. Additionally, when the agents have additive valuations, an envy-free allocation with negligible transfers and whose NSW is within a constant factor of optimal can be found in polynomial time. Consequently, we demonstrate that the seemingly incompatible objectives of fairness and high welfare can be achieved simultaneously via transfer payments, even for general valuations, when the welfare objective is NSW. On the other hand, we show that a similar result is impossible for utilitarian social welfare: any envy-freeable allocation that achieves a constant fraction of the optimal welfare requires non-negligible transfers. To complement this result we present algorithms that compute an envy-free allocation with a given target welfare and with bounded transfers. ",Two Birds With One Stone: Fairness and Welfare via Transfers The extraction of the magnetic form factor of the nucleon at zero momentum transfer is usually performed by adopting a parametrization for its momentum dependence and fitting the results obtained at finite momenta. We present position space methods that rely on taking the derivative of relevant correlators to extract directly the magnetic form factor at zero momentum without the need to assume a functional form for its momentum dependence. These methods are explored on one ensemble using $N_f=2+1+1$ Wilson twisted mass fermions. ,Extraction of the isovector magnetic form factor of the nucleon at zero momentum " Consider the Euclidean functional integral representation of any physical process in the electroweak model. Integrating out the fermion degrees of freedom introduces twenty-four fermion determinants. These multiply the Gaussian functional measures of the Maxwell, $Z$, $W$ and Higgs fields to give an effective functional measure. Suppose the functional integral over the Maxwell field is attempted first. This paper is concerned with the large amplitude behavior of the Maxwell effective measure. It is assumed that the large amplitude variation of this measure is insensitive to the presence of the $Z$, $W$ and $H$ fields; they are assumed to be a subdominant perturbation of the large amplitude Maxwell sector. Accordingly, we need only examine the large amplitude variation of a single QED fermion determinant. To facilitate this the Schwinger proper time representation of this determinant is decomposed into a sum of three terms. The advantage of this is that the separate terms can be non-perturbatively estimated for a measurable class of large amplitude random fields in four dimensions. It is found that the QED fermion determinant grows faster than $\exp \left[ce^2\int\mathrm d^4x\, F_{\mu\nu}^2\right]$, $c>0$, in the absence of zero mode supporting random background potentials. This raises doubt on whether the QED fermion determinant is integrable with any Gaussian measure whose support does not include zero mode supporting potentials. Including zero mode supporting background potentials can result in a decaying exponential growth of the fermion determinant. This is \textit{prima facie} evidence that Maxwellian zero modes are necessary for the non-perturbative quantization of QED and, by implication, for the non-perturbative quantization of the electroweak model. ",Non-perturbative quantization of the electroweak model's electrodynamic sector " The density functional theory is used to study the electronic structure of a quantum wire in a magnetic field. In a GaAs quantum wire, a critical density has been found, below which the electron density has a strong spatial inhomogeneity. The effect of the disorder potential on the electron density profile in a two-dimensional electron gas is estimated. ",Electronic structure of a wide quantum wire in a magnetic field at $\nu = 1$ " We present physical properties of radio galaxies (RGs) with $f_{\rm 1.4 GHz} >$ 1 mJy discovered by Subaru Hyper Supreme-Cam (HSC) and VLA Faint Images of the Radio Sky at Twenty-Centimeters (FIRST) survey. For 1056 FIRST RGs at $0 < z \leq 1.7$ with HSC counterparts in about 100 deg$^2$, we compiled multi-wavelength data of optical, near-infrared (IR), mid-IR, far-IR, and radio (150 MHz). We derived their color excess ($E (B-V)_{*}$), stellar mass, star formation rate (SFR), IR luminosity, the ratio of IR and radio luminosity ($q_{\rm IR}$), and radio spectral index ($\alpha_{\rm radio}$) that are derived from the SED fitting with CIGALE. We also estimated Eddington ratio based on stellar mass and integration of the best-fit SEDs of AGN component. We found that $E (B-V)_{*}$, SFR, and IR luminosity clearly depend on redshift while stellar mass, $q_{\rm IR}$, and $\alpha_{\rm radio}$ do not significantly depend on redshift. Since optically-faint ($i_{\rm AB} \geq 21.3$) RGs that are newly discovered by our RG survey tend to be high redshift, they tend to not only have a large dust extinction and low stellar mass but also have high SFR and AGN luminosity, high IR luminosity, and high Eddington ratio compared to optically-bright ones. The physical properties of a fraction of RGs in our sample seem to differ from a classical view of RGs with massive stellar mass, low SFR, and low Eddington ratio, demonstrating that our RG survey with HSC and FIRST provides us curious RGs among entire RG population. ","A Wide and Deep Exploration of Radio Galaxies with Subaru HSC (WERGS). II. Physical Properties derived from the SED Fitting with Optical, Infrared, and Radio Data" " Associated to every closed, embedded submanifold $N$ of a connected Riemannian manifold $M$, there is the distance function $d_N$ which measures the distance of a point in $M$ from $N$. We analyze the square of this function and show that it is Morse-Bott on the complement of the cut locus $\mathrm{Cu}(N)$ of $N$, provided $M$ is complete. Moreover, the gradient flow lines provide a deformation retraction of $M-\mathrm{Cu}(N)$ to $N$. If $M$ is a closed manifold, then we prove that the Thom space of the normal bundle of $N$ is homeomorphic to $M/\mathrm{Cu}(N)$. We also discuss several interesting results which are either applications of these or related observations regarding the theory of cut locus. These results include, but are not limited to, a computation of the local homology of singular matrices, a classification of the homotopy type of the cut locus of a homology sphere inside a sphere, a deformation of the indefinite unitary group $U(p,q)$ to $U(p)\times U(q)$ and a geometric deformation of $GL(n,\mathbb{R} )$ to $O(n,\mathbb{R} )$ which is different from the Gram-Schmidt retraction. \bigskip \noindent If a compact Lie group $G$ acts on a Riemannian manifold $M$ freely then $M/G$ is a manifold. In addition, if the action is isometric, then the metric of $M$ induces a metric on $M/G$. We show that if $N$ is a $G$-invariant submanifold of $M$, then the cut locus $\mathrm{Cu}(N)$ is $G$-invariant, and $\mathrm{Cu}(N)/G = \mathrm{Cu}\left( N/G \right) $ in $M/G$. An application of this result to complex projective hypersurfaces has been provided. ",Cut Locus of Submanifolds: A Geometric and Topological Viewpoint " We prove that any distributional limit of finite planar graphs in which the degree of the root has an exponential tail is almost surely recurrent. As a corollary, we obtain that the uniform infinite planar triangulation and quadrangulation (UIPT and UIPQ) are almost surely recurrent, resolving a conjecture of Angel, Benjamini and Schramm. We also settle another related problem of Benjamini and Schramm. We show that in any bounded degree, finite planar graph the probability that the simple random walk started at a uniform random vertex avoids its initial location for T steps is at most C/log T. ",Recurrence of planar graph limits We study pp and p\bar p collisions which lead to the exclusive production of J/psi or Upsilon from the pomeron-odderon and the pomeron-photon fusion. We calculate scattering amplitudes of these processes in the lowest order approximation and in the framework of k_T factorization. We present estimates of cross sections for the kinematic conditions of the Tevatron and of the LHC. ,Exclusive J/psi and Upsilon hadroproduction and the QCD odderon " Enhancing a recent result of Bayart and Ruzsa we obtain a Birkhoff-type characterization of upper frequently hypercyclic operators and a corresponding Upper Frequent Hypercyclicity Criterion. As an application we characterize upper frequently hypercyclic weighted backward shifts on sequence spaces, which in turn allows us to come up with various counter-examples in linear dynamics that are substantially simpler than those previously obtained in the literature. More generally, we introduce the notion of upper Furstenberg families $\mathcal{A}$ and show that our main results hold for $\mathcal{A}$-hypercyclic operators with respect to such families. ",Upper frequent hypercyclicity and related notions Efficient solution of the lowest eigenmodes is studied for a family of related eigenvalue problems with common $2\times 2$ block structure. It is assumed that the upper diagonal block varies between different versions while the lower diagonal block and the range of the coupling blocks remains unchanged. Such block structure naturally arises when studying the effect of a subsystem to the eigenmodes of the full system. The proposed method is based on interpolation of the resolvent function after some of its singularities have been removed by a spectral projection. Singular value decomposition can be used to further reduce the dimension of the computational problem. Error analysis of the method indicates exponential convergence with respect to the number of interpolation points. Theoretical results are illustrated by two numerical examples related to finite element discretisation of the Laplace operator. ,Efficient solution of symmetric eigenvalue problems from families of coupled systems " Leader election and arbitrary pattern formation are funda- mental tasks for a set of autonomous mobile robots. The former consists in distinguishing a unique robot, called the leader. The latter aims in arranging the robots in the plane to form any given pattern. The solv- ability of both these tasks turns out to be necessary in order to achieve more complex tasks. In this paper, we study the relationship between these two tasks in a model, called CORDA, wherein the robots are weak in several aspects. In particular, they are fully asynchronous and they have no direct means of communication. They cannot remember any previous observation nor computation performed in any previous step. Such robots are said to be oblivious. The robots are also uniform and anonymous, i.e, they all have the same program using no global parameter (such as an identity) allowing to differentiate any of them. Moreover, we assume that none of them share any kind of common coordinate mechanism or common sense of direction and we discuss the influence of a common handedness (i.e., chirality). In such a system, Flochini et al. proved in [11] that it is possible to elect a leader for n \geq 3 robots if it is possible to form any pattern for n \geq 3. In this paper, we show that the converse is true for n \geq 4 when the robots share a common handedness and for n \geq 5 when they do not. Thus, we deduce that with chirality (resp. without chirality) both problems are equivalent for n \geq 4 (resp. n \geq 5) in CORDA. ",Leader Election Problem Versus Pattern Formation Problem The ideals of the Lie algebras of unitriangular polynomial derivations are classified. An isomorphism criterion is given for the Lie factor algebras of the Lie algebras of unitriangular polynomial derivations. ,Lie algebras of triangular polynomial derivations and an isomorphism criterion for their Lie factor algebras " We discuss several steady-state rotation and oscillation modes of the planar parametric rotator and pendulum with damping. We consider a general elliptic trajectory of the suspension point for both rotator and pendulum, for the latter at an arbitrary angle with gravity, with linear and circular trajectories as particular cases. We treat the damped, non-linear equation of motion of the parametric rotator and pendulum perturbatively for small parametric excitation and damping, although our perturbative approach can be extended to other regimes as well. Our treatment involves only ordinary second-order differential equations with constant coefficients, and provides numerically accurate perturbative solutions in terms of elementary functions. Some of the steady-state rotation and oscillation modes studied here have not been discussed in the previous literature. Other well-known ones, such as parametric resonance and the inverted pendulum, are extended to elliptic parametric excitation tilted with respect to gravity. The results presented here should be accessible to advanced undergraduates, and of interest to graduate students and specialists in the field of non-linear mechanics. ",Steady states of the parametric rotator and pendulum " Light-weight time-of-flight (ToF) depth sensors are small, cheap, low-energy and have been massively deployed on mobile devices for the purposes like autofocus, obstacle detection, etc. However, due to their specific measurements (depth distribution in a region instead of the depth value at a certain pixel) and extremely low resolution, they are insufficient for applications requiring high-fidelity depth such as 3D reconstruction. In this paper, we propose DELTAR, a novel method to empower light-weight ToF sensors with the capability of measuring high resolution and accurate depth by cooperating with a color image. As the core of DELTAR, a feature extractor customized for depth distribution and an attention-based neural architecture is proposed to fuse the information from the color and ToF domain efficiently. To evaluate our system in real-world scenarios, we design a data collection device and propose a new approach to calibrate the RGB camera and ToF sensor. Experiments show that our method produces more accurate depth than existing frameworks designed for depth completion and depth super-resolution and achieves on par performance with a commodity-level RGB-D sensor. Code and data are available at https://zju3dv.github.io/deltar/. ",DELTAR: Depth Estimation from a Light-weight ToF Sensor and RGB Image " We explore the nature of the Bose condensation transition in driven open quantum systems, such as exciton-polariton condensates. Using a functional renormalization group approach formulated in the Keldysh framework, we characterize the dynamical critical behavior that governs decoherence and an effective thermalization of the low frequency dynamics. We identify a critical exponent special to the driven system, showing that it defines a new dynamical universality class. Hence critical points in driven systems lie beyond the standard classification of equilibrium dynamical phase transitions. We show how the new critical exponent can be probed in experiments with driven cold atomic systems and exciton-polariton condensates. ",Dynamical Critical Phenomena in Driven-Dissipative Systems " Marc Rieffel had introduced the notion of the quantum Gromov-Hausdorff distance on compact quantum metric spaces and found a sequence of matrix algebras that converges to the space of continuous functions on $2$-sphere in this distance. One finds applications of similar approximations in many places in the theoretical physics literature. In this paper, we have defined a compact quantum metric space structure on the sequence of Toeplitz algebras on generalized Bergman spaces and have proved that the sequence converges to the space of continuous function on odd spheres in the quantum Gromov-Hausdorff distance. ",Sequences of operator algebras converging to odd spheres in the quantum Gromov-Hausdorff distance " We have obtained accurate CCD narrow band H_beta and H_alpha photometry of Giant HII Regions in M 33, NGC 6822 and M 101. Comparison with previous determinations of emission line fluxes show large discrepancies; their probable origins are discussed. Combining our new photometric data with global velocity dispersion (sigma) derived from emission line widths we have reviewed the L(H_beta)-sigma relation. A reanalysis of the properties of the GEHRs included in our sample shows that age spread and the superposition of components in multiple regions introduce a considerable spread in the regression. Combining the information available in the literature regarding ages of the associated clusters, evolutionary footprints on the interstellar medium, and kinematical properties of the knots that build up the multiple GEHRs, we have found that a subsample - which we refer to as young and single GEHRs - do follow a tight relation in the L-sigma plane. ",Narrowband CCD Photometry of Giant HII Regions " We give a new randomized distributed algorithm for $(\Delta+1)$-coloring in the LOCAL model, running in $O(\sqrt{\log \Delta})+ 2^{O(\sqrt{\log \log n})}$ rounds in a graph of maximum degree~$\Delta$. This implies that the $(\Delta+1)$-coloring problem is easier than the maximal independent set problem and the maximal matching problem, due to their lower bounds of $\Omega \left( \min \left( \sqrt{\frac{\log n}{\log \log n}}, \frac{\log \Delta}{\log \log \Delta} \right) \right)$ by Kuhn, Moscibroda, and Wattenhofer [PODC'04]. Our algorithm also extends to list-coloring where the palette of each node contains $\Delta+1$ colors. We extend the set of distributed symmetry-breaking techniques by performing a decomposition of graphs into dense and sparse parts. ",Distributed $(\Delta+1)$-Coloring in Sublogarithmic Rounds " Heusler alloys based magnetic tunnel junctions can potentially provide high magnetoresistance, small damping and fast switching. Here junctions with Co2FeAl as a ferromagnetic electrode are fabricated by room temperature sputtering on Si/SiO2 substrates. The doping of Boron in Co2FeAl is found to have a large positive impact on the structural, magnetic and transport properties of the junctions, with a reduced interfacial roughness and substantial improved tunneling magnetoresistance. A two-level magnetoresistance is also observed in samples annealed at low temperature, which is believed to be related to the memristive effect of the tunnel barrier with impurities. ",Impact of Boron doping to the tunneling magnetoresistance of Heusler alloy Co2FeAl " In this article, we analyze the propagation of Wigner measures of a family of solutions to a system of semi-classical pseudodifferential equations presenting eigenvalues crossings on hypersurfaces. We prove the propagation along classical trajectories under a geometric condition which is satisfied for example as soon as the Hamiltonian vector fields are transverse or tangent at finite order to the crossing set. We derive resolvent estimates for semi-classical Schr\""odinger operator with matrix-valued potential under a geometric condition of the same type on the crossing set and we analyze examples of degenerate situations where one can prove transfers between the modes. ",Degenerated codimension 1 crossings and resolvent estimates " We thoroughly examine the ground state of the triangular lattice of Pb on Si(111) using scanning tunneling microscopy. We detect charge-order, accompanied by a subtle structural reconstruction. Applying the extended variational cluster approach we map out the phase diagram as a function of local and non-local Coulomb interactions. Comparing the experimental data with the theoretical modeling leads us to conclude that electron correlations are the driving force of the charge-ordered state in Pb/Si(111), rather than Fermi surface nesting. These results resolve the discussion about the origin of the well known $3\times 3$ reconstruction forming below $86\,$K. By exploiting the tunability of correlation strength, hopping parameters and bandfilling, this material class represents a promising platform to search for exotic states of matter, in particular, for chiral topological superconductivity. ",Correlation-driven charge order in a frustrated two-dimensional atom lattice " We report on the study of the population of B and Be stars in SMC young clusters, performed with the Wide Field Imager in slitless spectroscopic mode at ESO/T2.2m with a filter centered at Halpha. First, we explain the reduction methods we used and our selection of different types of objects. Second, we present results on the proportion of Be stars in SMC clusters, and we compare this proportion to the one observed in the Milky Way. Finally, we also present results on a statistical study of variability of Be stars with OGLE. ",Be stars in open clusters in the Small Magellanic Cloud " We discuss the issue of parity violation in quantum gravity. In particular, we study the coupling of fermionic degrees of freedom in the presence of torsion and the physical meaning of the Immirzi parameter from the viewpoint of effective field theory. We derive the low-energy effective lagrangian which turns out to involve two parameters, one measuring the non-minimal coupling of fermions in the presence of torsion, the other being the Immirzi parameter. In the case of non-minimal coupling the effective lagrangian contains an axial-vector interaction leading to parity violation. Alternatively, in the case of minimal coupling there is no parity violation and the effective lagrangian contains only the usual axial-axial interaction. In this situation the real values of the Immirzi parameter are not at all constrained. On the other hand, purely imaginary values of the Immirzi parameter lead to violations of unitarity for the case of non-minimal coupling. Finally, the effective lagrangian blows up for the positive and negative unit imaginary values of the Immirzi parameter. ","Quantum Gravity, Torsion, Parity Violation and all that" " An operator deformed quantum algebra is discovered exploiting the quantum Yang-Baxter equation with trigonometric R-matrix. This novel Hopf algebra along with its $q \to 1$ limit appear to be the most general Yang-Baxter algebra underlying quantum integrable systems. Three different directions of application of this algebra in integrable systems depending on different sets of values of deforming operators are identified. Fixed values on the whole lattice yield subalgebras linked to standard quantum integrable models, while the associated Lax operators generate and classify them in an unified way. Variable values construct a new series of quantum integrable inhomogeneous models. Fixed but different values at different lattice sites can produce a novel class of integrable hybrid models including integrable matter-radiation models and quantum field models with defects, in particular, a new quantum integrable sine-Gordon model with defect. ",Yang-Baxter algebra and generation of quantum integrable models " The exchange-correlation potential experienced by an electron in the free space adjacent to a solid surface or to a low-dimensional system defines the fundamental image states and is generally important in surface- and nano-science. Here we determine the potential near the two- and one-dimensional electron gases (EG), doing this analytically at the level of the exact exchange of the density-functional theory (DFT). We find that, at $r_\perp\gg k_F^{-1}$, where $r_\perp$ is the distance from the EG and $k_F$ is the Fermi radius, the potential obeys the already known asymptotic $-e^2/r_\perp$, while at $r_\perp \lesssim k_F^{-1}$, but {\em still in vacuum}, qualitative and quantitative deviations of the exchange potential from the asymptotic law occur. The playground of the excitations to the low-lying image states falls into the latter regime, causing significant departure from the Rydberg series. In general, our analytical exchange potentials establish benchmarks for numerical approaches in the low-dimensional science, where DFT is by far the most common tool. ","Exact ""exact exchange"" potential of two- and one-dimensional electron gases beyond the asymptotic limit" The difference between the CP asymmetries of the $B^0 \to K^+ \pi^-$ and $B^+ \to K^+ \pi^0$ decays has been recently confirmed with an evidence larger than $5\sigma$'s. We discuss it as a possible signal of new physics associated with new (large) CP violation in the electroweak penguin contributions. We propose a supersymmetry breaking scheme where such new sources of CP violation occur in the flavor non-universal trilinear scalar couplings. ,B -> K pi Puzzle and New Sources of CP Violation in Supersymmetry " The near-zone ``Love'' symmetry resolves the naturalness issue of black hole Love number vanishing with $\text{SL}\left(2,\mathbb{R}\right)$ representation theory. Here, we generalize this proposal to $5$-dimensional asymptotically flat and doubly spinning (Myers-Perry) black holes. We consider the scalar response of Myers-Perry black holes and extract its static scalar Love numbers. In agreement with the naturalness arguments, these Love numbers are, in general, non-zero and exhibit logarithmic running unless certain resonant conditions are met; these conditions include new cases with no previously known analogs. We show that there exist two near-zone truncations of the equations of motion that exhibit enhanced $\text{SL}\left(2,\mathbb{R}\right)$ Love symmetries that explain the vanishing of the static scalar Love numbers in the resonant cases. These Love symmetries can be interpreted as local $\text{SL}\left(2,\mathbb{R}\right)\times\text{SL}\left(2,\mathbb{R}\right)$ near-zone symmetries spontaneously broken down to global $\text{SL}\left(2,\mathbb{R}\right)\times U\left(1\right)$ symmetries by the periodic identification of the azimuthal angles. We also discover an infinite-dimensional extension of the Love symmetry into $\text{SL}\left(2,\mathbb{R}\right)\ltimes\hat{U}\left(1\right)_{\mathcal{V}}^2$ that contains both Love symmetries as particular subalgebras, along with a family of $\text{SL}\left(2,\mathbb{R}\right)$ subalgebras that reduce to the exact near-horizon Myers-Perry black hole isometries in the extremal limit. Finally, we show that the Love symmetries acquire a geometric interpretation as isometries of subtracted (effective) black hole geometries that preserve the internal structure of the black hole and interpret these non-extremal $\text{SL}\left(2,\mathbb{R}\right)$ structures as remnants of the enhanced isometry of the near-horizon extremal geometries. ",Scalar Love numbers and Love symmetries of 5-dimensional Myers-Perry black holes " Collector's seals provide important clues about the ownership of a book. They contain much information pertaining to the essential elements of ancient materials and also show the details of possession, its relation to the book, the identity of the collectors and their social status and wealth, amongst others. Asian collectors have typically used artistic ancient characters rather than modern ones to make their seals. In addition to the owner's name, several other words are used to express more profound meanings. A system that automatically recognizes these characters can help enthusiasts and professionals better understand the background information of these seals. However, there is a lack of training data and labelled images, as samples of some seals are scarce and most of them are degraded images. It is necessary to find new ways to make full use of such scarce data. While these data are available online, they do not contain information on the characters'position. The goal of this research is to provide retrieval tools assist in obtaining more information from Asian collector's seals imprints without consuming a lot of computational resources. In this paper, a character segmentation method is proposed to predict the candidate characters'area without any labelled training data that contain character coordinate information. A retrieval-based recognition system that focuses on a single character is also proposed to support seal retrieval and matching. The experimental results demonstrate that the proposed character segmentation method performs well on Asian collector's seals, with 92% of the test data being correctly segmented. ",Character Segmentation in Asian Collector's Seal Imprints: An Attempt to Retrieval Based on Ancient Character Typeface " Semantic parsing of large-scale 3D point clouds is an important research topic in computer vision and remote sensing fields. Most existing approaches utilize hand-crafted features for each modality independently and combine them in a heuristic manner. They often fail to consider the consistency and complementary information among features adequately, which makes them difficult to capture high-level semantic structures. The features learned by most of the current deep learning methods can obtain high-quality image classification results. However, these methods are hard to be applied to recognize 3D point clouds due to unorganized distribution and various point density of data. In this paper, we propose a 3DCNN-DQN-RNN method which fuses the 3D convolutional neural network (CNN), Deep Q-Network (DQN) and Residual recurrent neural network (RNN) for an efficient semantic parsing of large-scale 3D point clouds. In our method, an eye window under control of the 3D CNN and DQN can localize and segment the points of the object class efficiently. The 3D CNN and Residual RNN further extract robust and discriminative features of the points in the eye window, and thus greatly enhance the parsing accuracy of large-scale point clouds. Our method provides an automatic process that maps the raw data to the classification results. It also integrates object localization, segmentation and classification into one framework. Experimental results demonstrate that the proposed method outperforms the state-of-the-art point cloud classification methods. ",3DCNN-DQN-RNN: A Deep Reinforcement Learning Framework for Semantic Parsing of Large-scale 3D Point Clouds " We consider a supersymmetric hybrid inflation scenario in which the U(1) $R$-symmetry is explicitly broken by Planck scale suppressed operators in the superpotential. We provide an example with minimal K\""ahler potential, with the $R$-symmetry breaking term relevant during inflation being $\alpha S^4$, where $S$ denotes the well-known gauge singlet inflaton superfield. The inflationary potential takes into account the radiative and supergravity corrections, as well as the soft supersymmetry breaking terms. For successful inflation, with the scalar spectral index in the currently preferred range, $n_s \approx 0.97 \pm 0.010$, $| \alpha |$ $\lesssim 10^{-7}$. The tensor to scalar ratio $r \lesssim 10^{-4}$, while $| \text{d} n_s / \text{d} \ln k | \sim \mathcal{O}(10^{-3})-\mathcal{O}(10^{-4})$ ",$R$-Symmetry Breaking in Supersymmetric Hybrid Inflation " The development of the Automatic License Plate Recognition (ALPR) system has received much attention for the English license plate. However, despite being the sixth largest population around the world, no significant progress can be tracked in the Bengali language countries or states for the ALPR system addressing their more alarming traffic management with inadequate road-safety measures. This paper reports a computationally efficient and reasonably accurate Automatic License Plate Recognition (ALPR) system for Bengali characters with a new end-to-end DNN model that we call Bengali License Plate Network(BLPnet). The cascaded architecture for detecting vehicle regions prior to vehicle license plate (VLP) in the model is proposed to eliminate false positives resulting in higher detection accuracy of VLP. Besides, a lower set of trainable parameters is considered for reducing the computational cost making the system faster and more compatible for a real-time application. With a Computational Neural Network (CNN)based new Bengali OCR engine and word-mapping process, the model is characters rotation invariant, and can readily extract, detect and output the complete license plate number of a vehicle. The model feeding with17 frames per second (fps) on real-time video footage can detect a vehicle with the Mean Squared Error (MSE) of 0.0152, and the mean license plate character recognition accuracy of 95%. While compared to the other models, an improvement of 5% and 20% were recorded for the BLPnetover the prominent YOLO-based ALPR model and the Tesseract model for the number-plate detection accuracy and time requirement, respectively. ",BLPnet: A new DNN model and Bengali OCR engine for Automatic License Plate Recognition " We propose an enhanced peer-review process where the reviewers are encouraged to truthfully disclose their reviews. We start by modelling that process using a Bayesian model where the uncertainty regarding the quality of the manuscript is taken into account. After that, we introduce a scoring function to evaluate the reported reviews. Under mild assumptions, we show that reviewers strictly maximize their expected scores by telling the truth. We also show how those scores can be used in order to reach consensus. ",On a Reliable Peer-Review Process We present an unprecedented view on the morphology and kinematics of the extended narrow-line region (ENLR) and molecular gas around the prototypical hyper-luminous quasar 3C273 ($L\sim10^{47}$ erg/s at z=0.158) based on VLT-MUSE optical 3D spectroscopy and ALMA observations. We find that: 1) The ENLR size of 12.1$\pm$0.2kpc implies a smooth continuation of the size-luminosity relation out to large radii or a much larger break radius as previously proposed. 2) The kinematically disturbed ionized gas with line splits reaching 1000km/s out to 6.1$\pm$1.5kpc is aligned along the jet axis. 3) The extreme line broadening on kpc scales is caused by spatial and spectral blending of many distinct gas clouds separated on sub-arcsecond scales with different line-of-sight velocities. The ENLR velocity field combined with the known jet orientation rule out a simple scenario of a radiatively-driven radial expansion of the outflow. Instead we propose that a pressurized expanding hot gas cocoon created by the radio jet is impacting on an inclined gas disk leading to transverse and/or backflow motion with respect to our line-of-sight. The molecular gas morphology may either be explained by a density wave at the front of the outflow expanding along the jet direction as predicted by positive feedback scenario or the cold gas may be trapped in a stellar over-density caused by a recent merger event. Using 3C273 as a template for observations of high-redshift hyper-luminous AGN reveals that large-scale ENLRs and kpc scale outflows may often be missed due to the brightness of the nuclei and the limited sensitivity of current near-IR instrumentation. ,Jet-driven galaxy-scale gas outflows in the hyper-luminous quasar 3C273 " An isomorphism between two graphs is a bijection between their vertices that preserves the edges. We consider the problem of determining whether two finite undirected weighted graphs are isomorphic, and finding an isomorphism relating them if the answer is positive. In this paper we introduce effective probabilistic linear programming (LP) heuristics to solve the graph isomorphism problem. We motivate our heuristics by showing guarantees under some conditions, and present numerical experiments that show effectiveness of these heuristics in the general case. ",Linear Programming Heuristics for the Graph Isomorphism Problem " Self-consistency based Kohn-Sham density functional theory (KS-DFT) electronic structure calculations with Gaussian basis sets are reported for a set of 17 protein-like molecules with geometries obtained from the protein data bank. It is found that in many cases such calculations do not converge due to vanishing HOMO-LUMO gaps. A sequence of polyproline I helix molecules is also studied, and it is found that self-consistency calculations using pure functionals fail to converge for helices longer than six proline units. Since the computed gap is strongly correlated to the fraction of Hartree-Fock exchange, test calculations using both pure and hybrid density functionals are reported. The tested methods include the pure functionals BLYP, PBE, and LDA, as well as Hartree-Fock and the hybrid functionals BHandHLYP, B3LYP, and PBE0. The effect of including solvent molecules in the calculations is studied, and it is found that the inclusion of explicit solvent molecules around the protein fragment in many cases gives a larger gap, but that convergence problems due to vanishing gaps still occur in calculations with pure functionals. In order to achieve converged results, some modeling of the charge distribution of solvent water molecules outside the electronic structure calculation is needed. Representing solvent water molecules by a simple point charge distribution is found to give non-vanishing HOMO-LUMO gaps for the tested protein-like systems also for pure functionals. ",Difficulties in applying pure Kohn-Sham density functional theory electronic structure methods to protein molecules The Fourier series method is used to solve the homogeneous equation governing the motion of the harmonic oscillator. It is shown that the general solution to the problem can be found in a surprisingly simple way for the case of the simple harmonic oscillator. It is also shown that the damped harmonic oscillator is susceptible to the analysis. ,Harmonic oscillator: an analysis via Fourier series " Collider searches for energetic particles recoiling against missing transverse energy allow to place strong bounds on the interactions between dark matter (DM) and standard model particles. In this article we update and extend LHC constraints on effective dimension-7 operators involving DM and electroweak gauge bosons. A concise comparison of the sensitivity of the mono-photon, mono-W, mono-Z, mono-W/Z, invisible Higgs-boson decays in the vector boson fusion mode and the mono-jet channel is presented. Depending on the parameter choices, either the mono-photon or the mono-jet data provide the most stringent bounds at the moment. We furthermore explore the potential of improving the current 8 TeV limits at 14 TeV. Future strategies capable of disentangling the effects of the different effective operators involving electroweak gauge bosons are discussed as well. ",LHC constraints on gauge boson couplings to dark matter " Collision cross sections and rate constants between two ground- state oxygen molecules are investigated theoretically at translational energies below $\sim 1$K and in zero magnetic field. We present calculations for elastic and spin- changing inelastic collision rates for different isotopic combinations of oxygen atoms as a prelude to understanding their collisional stability in ultracold magnetic traps. A numerical analysis has been made in the framework of a rigid- rotor model that accounts fully for the singlet, triplet, and quintet potential energy surfaces in this system. The results offer insights into the effectiveness of evaporative cooling and the properties of molecular Bose- Einstein condensates, as well as estimates of collisional lifetimes in magnetic traps. Specifically, $^{17}O_{2}$ looks like a good candidate for ultracold studies, while $^{16}O_{2}$ is unlikely to survive evaporative cooling. Since $^{17}O_{2}$ is representative of a wide class of molecules that are paramagnetic in their ground state we conclude that many molecules can be successfully magnetically trapped at ultralow temperatures. ",Ultracold collisions of oxygen molecules " We characterize the capacity region to within 2 bits/s/Hz and the symmetric capacity to within 1 bit/s/Hz for the two-user Gaussian interference channel (IC) with feedback. We develop achievable schemes and derive a new outer bound to arrive at this conclusion. One consequence of the result is that feedback provides multiplicative gain, i.e., the gain becomes arbitrarily large for certain channel parameters. It is a surprising result because feedback has been so far known to provide no gain in memoryless point-to-point channels and only bounded additive gain in multiple access channels. The gain comes from using feedback to maximize resource utilization, thereby enabling more efficient resource sharing between the interfering users. The result makes use of a deterministic model to provide insights into the Gaussian channel. This deterministic model is a special case of El Gamal-Costa deterministic model and as a side-generalization, we establish the exact feedback capacity region of this general class of deterministic ICs. ",Feedback Capacity of the Gaussian Interference Channel to within 2 Bits " Using the ASAS data, we determine the pulsational frequencies of the high-amplitude $\delta$ Sct star BP Pegasi. The analysis revealed only the two known, independent frequencies. On the basis of multicolour Str\""omgren photometry, we independently find that both frequencies can only be associated with radial modes which, according to the frequency ratio, are fundamental and first overtone modes. The models fitting the two frequencies depend strongly on the opacity data. For low values of the mixing length parameter $\alpha_{\rm MLT}\approx 0.5$, only the OPAL seismic models in the post-main sequence phase of evolution are caught within the observed error box. Seismic OP and OPLIB models can only reach the error box if we increase $\alpha_{\rm MLT}$ to at least 2.0. Then, including the non-adiabatic parameter $f$ into our seismic modelling, we constrain various parameters, employing Monte Carlo-based Bayesian analysis. ",Asteroseismology of the double-radial mode $\delta$ Scuti star BP Pegasi " Numerical calculations for Majorana zero modes on a one-dimensional chain are performed using the technique of block diagonalization for general parameter settings. It is found that Majorana zero modes occur near the ends of the chain and decay exponentially away from the ends. The phase diagrams show that Majorana zero modes of a long-enough chain indeed have a parameter domain of $2t>|\mu|$ as predicted from the bulk property of the chain, but a short chain has a much smaller parameter domain than the prediction. Through a numerical simulation Majorana zero modes are found to be robust under the disturbance of noise. Finally the reversion of the parity of the ground states is studied by applying a bias voltage on the quantum dot at an end of the chain. It is found that for a weak coupling between a chain and a quantum dot the parity of the ground states can be reversed through adiabatically tuning the bias voltage. ",Majorana zero modes on a 1D chain for quantum computation " We present preliminary results of the first and on-going radial velocity survey for circumbinary planets. With a novel radial velocity technique employing an iodine absorption cell we achieve an unprecedented RV precision of up to 2 m/s for double-lined binary stars. The high resolution spectra collected with the Keck I/Hires, TNG/Sarg and Shane/CAT/Hamspec telescopes/spectrographs over the years 2003-2008 allow us to derive RVs and compute planet detection limits for ten double-lined binary stars. For this initial sample of targets, we can rule out planets on dynamically stable orbits with masses as small as ~0.3 to 3 MJup for the orbital periods of up to ~5.3 years. Even though the presented sample of stars is too small to make any strong conclusions, it is clear that the search for circumbinary planets is now technique-wise possible and eventually will provide new constraints for the planet formation theories. ","The Radial Velocity TATOOINE Search for Circumbinary Planets: Planet Detection Limits for a Sample of Double-lined Binary Stars - Initial Results from Keck I/Hires, Shane/CAT/Hamspec and TNG/Sarg Observations" " A subgraph is constructed by using a subset of vertices and edges of a given graph. There exist many graph properties that are hereditary for subgraphs. Hence, researchers from different communities have paid a great deal of attention in studying numerous subgraph problems, on top of the ordinary graph problems. Many algorithms are proposed in studying subgraph problems, where one common approach is by extracting the patterns and structures of a given graph. Due to the complex structures of certain types of graphs and to improve overall performances of the existing frameworks, machine learning techniques have recently been employed in dealing with various subgraph problems. In this article, we present a comprehensive review on five well known subgraph problems that have been tackled by using machine learning methods. They are subgraph isomorphism (both counting and matching), maximum common subgraph, community detection and community search problems. We provide an outline of each proposed method, and examine its designs and performances. We also explore non-learning-based algorithms for each problem and a brief discussion is given. We then suggest some promising research directions in this area, hoping that relevant subgraph problems can be tackled by using a similar strategy. Since there is a huge growth in employing machine learning techniques in recent years, we believe that this survey will serve as a good reference point to relevant research communities. ",A Survey on Machine Learning Solutions for Graph Pattern Extraction " A randomized misfit approach is presented for the efficient solution of large-scale PDE-constrained inverse problems with high-dimensional data. The purpose of this paper is to offer a theory-based framework for random projections in this inverse problem setting. The stochastic approximation to the misfit is analyzed using random projection theory. By expanding beyond mean estimator convergence, a practical characterization of randomized misfit convergence can be achieved. The theoretical results developed hold with any valid random projection in the literature. The class of feasible distributions is broad yet simple to characterize compared to previous stochastic misfit methods. This class includes very sparse random projections which provide additional computational benefit. A different proof for a variant of the Johnson-Lindenstrauss lemma is also provided. This leads to a different intuition for the $O(\epsilon^{-2})$ factor in bounds for Johnson-Lindenstrauss results. The main contribution of this paper is a theoretical result showing the method guarantees a valid solution for small reduced misfit dimensions. The interplay between Johnson-Lindenstrauss theory and Morozov's discrepancy principle is shown to be essential to the result. The computational cost savings for large-scale PDE-constrained problems with high- dimensional data is discussed. Numerical verification of the developed theory is presented for model problems of estimating a distributed parameter in an elliptic partial differential equation. Results with different random projections are presented to demonstrate the viability and accuracy of the proposed approach. ",A Data-Scalable Randomized Misfit Approach for Solving Large-Scale PDE-Constrained Inverse Problems " Both feature selection and hyperparameter tuning are key tasks in machine learning. Hyperparameter tuning is often useful to increase model performance, while feature selection is undertaken to attain sparse models. Sparsity may yield better model interpretability and lower cost of data acquisition, data handling and model inference. While sparsity may have a beneficial or detrimental effect on predictive performance, a small drop in performance may be acceptable in return for a substantial gain in sparseness. We therefore treat feature selection as a multi-objective optimization task. We perform hyperparameter tuning and feature selection simultaneously because the choice of features of a model may influence what hyperparameters perform well. We present, benchmark, and compare two different approaches for multi-objective joint hyperparameter optimization and feature selection: The first uses multi-objective model-based optimization. The second is an evolutionary NSGA-II-based wrapper approach to feature selection which incorporates specialized sampling, mutation and recombination operators. Both methods make use of parameterized filter ensembles. While model-based optimization needs fewer objective evaluations to achieve good performance, it incurs computational overhead compared to the NSGA-II, so the preferred choice depends on the cost of evaluating a model on given data. ",Multi-Objective Hyperparameter Tuning and Feature Selection using Filter Ensembles " The paper deals with $\Sigma-$composition of terms, which allows us to extend the derivation rules in formal deduction of identities. The concept of essential variables and essential positions of terms with respect to a set of identities is a key step in the simplification of the process of formal deduction. $\Sigma-$composition of terms is defined as replacement between $\Sigma$-equal terms. This composition induces $\Sigma R-$deductively closed sets of identities. In analogy to balanced identities we introduce and investigate $\Sigma-$balanced identities for a given set of identities $\Sigma$. ",Essential variables and positions in terms " The thermodynamic uncertainty relation is a universal trade-off relation connecting the precision of a current with the average dissipation at large times. For continuous time Markov chains (also called Markov jump processes) this relation is valid in the time-homogeneous case, while it fails in the time-periodic case. The latter is relevant for the study of several small thermodynamic systems. We consider here a time-periodic Markov chain with continuous time and a broad class of functionals of stochastic trajectories, which are general linear combinations of the empirical flow and the empirical density. Inspired by the analysis done in our previous work [1], we provide general methods to get local quadratic bounds for large deviations, which lead to universal lower bounds on the ratio of the diffusion coefficient to the squared average value in terms of suitable universal rates, independent of the empirical functional. These bounds are called ""generalized thermodynamic uncertainty relations"" (GTUR's), being generalized versions of the thermodynamic uncertainty relation to the time-periodic case and to functionals which are more general than currents. Previously, GTUR's in the time-periodic case have been obtained in [1, 27, 42]. Here we recover the GTUR's in [1, 27] and produce new ones, leading to even stronger bounds and also to new trade-off relations for time-homogeneous systems. Moreover, we generalize to arbitrary protocols the GTUR obtained in [42] for time-symmetric protocols. We also generalize to the time-periodic case the GTUR obtained in [19] for the so called dynamical activity, and provide a new GTUR which, in the time-homogeneous case, is stronger than the one in [19]. The unifying picture is completed with a comprehensive comparison between the different GTUR's. ",A unifying picture of generalized thermodynamic uncertainty relations " Knowledge distillation becomes a de facto standard to improve the performance of small neural networks. Most of the previous works propose to regress the representational features from the teacher to the student in a one-to-one spatial matching fashion. However, people tend to overlook the fact that, due to the architecture differences, the semantic information on the same spatial location usually vary. This greatly undermines the underlying assumption of the one-to-one distillation approach. To this end, we propose a novel one-to-all spatial matching knowledge distillation approach. Specifically, we allow each pixel of the teacher feature to be distilled to all spatial locations of the student features given its similarity, which is generated from a target-aware transformer. Our approach surpasses the state-of-the-art methods by a significant margin on various computer vision benchmarks, such as ImageNet, Pascal VOC and COCOStuff10k. Code will be released soon. ",Knowledge Distillation via the Target-aware Transformer " Robin Forman's highly influential 2002 paper titled A User's Guide to Discrete Morse Theory presents an overview of the subject in a very readable manner. As a proof of concept, the author determines the topology (homotopy type) of the abstract simplicial complex of disconnected graphs of order $n$ (which was previously done by Victor Vassiliev using classical topological methods) using discrete Morse theoretic techniques, which are purely combinatorial in nature. The techniques involve the construction (and verification) of a discrete gradient vector field on the complex. However, the verification part relies on a claim that doesn't seem to hold. In this note, we provide a couple of counterexamples against this specific claim and also an alternative proof of the claim that the constructed discrete vector field is indeed a gradient vector field. ",A note on an application of discrete Morse theoretic techniques on the complex of disconnected graphs " Vector solitary waves are nonlinear waves of coupled polarizations that propagate with constant velocity and shape. In mechanics, they hold the potential to control locomotion, mitigate shocks and transfer information, among other functionalities. Recently, such elastic waves were numerically observed in compressible rubber-like laminates. Here, we conduct numerical experiments to characterize the possible vector solitary waves in these laminates, and expose a new type of waves whose amplitude and velocity oscillate periodically without dispersing in time. This oscillation is a manifestation of a periodic transfer of energy between the two wave polarizations, which we consider as internal mode of the solitary wave. We find that the vector solitary waves propagate faster at higher amplitudes, and determine a lower bound for their velocity. We describe a procedure for identifying which initial strains generate such vector solitary waves. This procedure also enables an additional classification between tensile and compressive solitary waves, according to the way that the axial strain changes as the waves propagate. ",Oscillating vector solitary waves in soft laminates " In this paper we analyze the gravitational field of a global monopole in the context of $f(R)$ gravity. More precisely, we show that the field equations obtained are expressed in terms of $F(R)=\frac{df(R)}{dR}$. Since we are dealing with a spherically symmetric system, we assume that $F(R)$ is a function of the radial coordinate only. Moreover, adopting the weak field approximation, we can provide all components of the metric tensor. A comparison with the corresponding results obtained in General Relativity and in the Brans-Dicke theory is also made. ",Gravitational Field of a Global Monopole in a Modified Gravity " Iron self-diffusion in nano-composite FeZr alloy has been investigated using neutron reflectometry technique as a function of applied compressive stress. A composite target of (Fe+Zr) and (57Fe+Zr) was alternatively sputtered to deposit chemically homogeneous multilayer (CHM) structure, [Fe75Zr25/57Fe75Zr25]10. The multilayers were deposited on to a bent Si wafer using a 3-point bending device. Post-deposition, the bending of the substrate was released which results in an applied compressive stress on to the multilayer. In the as-deposited state, the alloy multilayer forms an amorphous phase, which crystallizes into a nano-composite phase when heated at 373 K. Bragg peaks due to isotopic contrast were observed from CHM, when measured by neutron reflectivity, while x-ray reflectivity showed a pattern corresponding to a single layer. Self-diffusion of iron was measured with the decay of the intensities at the Bragg peaks in the neutron reflectivity pattern after thermal annealing at different temperatures. It was found that the self-diffusion of iron slows down with an increase in the strength of applied compressive stress. ",Iron self-diffusion in FeZr/$^{57}$FeZr multilayers measured by neutron reflectometry: Effect of applied compressive stress " This exposition contains a short and streamlined proof of the recent result of Kwan, Letzter, Sudakov and Tran that every triangle-free graph with minimum degree $d$ contains an induced bipartite subgraph with average degree $\Omega(\ln d/\ln\ln d)$. ",A note on dense bipartite induced subgraphs We present the first example of a poly-time computable Julia set with a recurrent critical point: we prove that the Julia set of the Feigenbaum map is computable in polynomial time. ,Poly-time Computability of the Feigenbaum Julia set " Monitoring propeller failures is vital to maintain the safe and reliable operation of quadrotor UAVs. The simulation-to-reality UAV fault diagnosis technique offer a secure and economical approach to identify faults in propellers. However, classifiers trained with simulated data perform poorly in real flights due to the wind disturbance in outdoor scenarios. In this work, we propose an uncertainty-based fault classifier (UFC) to address the challenge of sim-to-real UAV fault diagnosis in windy scenarios. It uses the ensemble of difference-based deep convolutional neural networks (EDDCNN) to reduce model variance and bias. Moreover, it employs an uncertainty-based decision framework to filter out uncertain predictions. Experimental results demonstrate that the UFC can achieve 100% fault-diagnosis accuracy with a data usage rate of 33.6% in the windy outdoor scenario. ",Simulation-to-reality UAV Fault Diagnosis in windy environments " We present the notion of higher Kirillov brackets on the sections of an even line bundle over a supermanifold. When the line bundle is trivial we shall speak of higher Jacobi brackets. These brackets are understood furnishing the module of sections with an $L_{\infty}$-algebra, which we refer to as a homotopy Kirillov algebra. We are then to higher Kirillov algebroids as higher generalisations of Jacobi algebroids. Furthermore, we show how to associate a higher Kirillov algebroid and a homotopy BV-algebra with every higher Kirillov manifold. In short, we construct homotopy versions of some of the well-known theorems related to Kirillov's local Lie algebras. ",Kirillov structures up to homotopy " I present quenched domain wall fermion and 2+1 flavor improved Kogut-Susskind fermion calculations of the hadronic vacuum polarization which are used to calculate the ${\cal O}(\alpha^2)$ hadronic contribution to the anomalous magnetic moment of the muon. Together with previous quenched calcuations, the new results confirm that in the quenched theory the hadronic contribution is signifcantly smaller ($\sim 30%$) than the value obtained from the total cross section of $e^+e^-$ annhilation to hadrons. The 2+1 flavor results show an increasing contribution to $g-2$ as the quark mass is reduced. ",Lattice calculation of the lowest-order hadronic contribution to the muon anomalous magnetic momen " Starting from the Euler's identity, the author improved Riemann's results, discovered the relationship between the Riemann Zeta function and the prime function, and obtained two new corollaries based on Riemann hypothesis is tenable. From these corollaries, the author found the relationship between the p_m, its subscript m and the t_m, and obtains a complete table of primes (less than 5000). ",New results based on Riemann hypothesis is tenable " New solutions for second-order intertwining relations in two-dimensional SUSY QM are found via the repeated use of the first order supersymmetrical transformations with intermediate constant unitary rotation. Potentials obtained by this method - two-dimensional generalized P\""oschl-Teller potentials - appear to be shape-invariant. The recently proposed method of $SUSY-$separation of variables is implemented to obtain a part of their spectra, including the ground state. Explicit expressions for energy eigenvalues and corresponding normalizable eigenfunctions are given in analytic form. Intertwining relations of higher orders are discussed. ",New Two-Dimensional Quantum Models Partially Solvable by Supersymmetrical Approach " Copper oxide nanostructures with spherical (0D), needle (1D) and hierarchical cauliflower (3D) morphologies are used to demonstrate superhydrophobic, superoleophobic and slippery behavior. These nanostructures are synthesized on galvanized steel substrates using a simple chemical bath deposition method by tuning precursor concentration. Subsequent coating of low surface energy polymer, polydimethylsiloxane, results in superhydrophobicity with water contact angle ~160(2){\deg} and critical sliding angle ~2{\deg}. When functionalized with low-surface energy perfluoroalkyl silane, these surfaces display high repellency for low surface tension oils and hydrocarbons. Among them, the hierarchical cauliflower morphology exhibits better re-entrant structure thus show the best superoleophobicity with 149{\deg} contact angle for dodecane having surface tension 25.3 mNm-1. If these nanostructured substrates are infused with lubricant Silicone oil, they show excellent slippery behavior for water drops. Due to the lubricating nature of Silicone oil, the Silicone oil infused slippery surfaces (SOIS) show low contact angle hysteresis (~2{\deg}) and critical tilt angle (~2{\deg}). The hierarchical cauliflower nanostrcuture exhibit better slippery characteristics and stability compared to the other nanostructured surfaces. ","Uniting Superhydrophobic, Superoleophobic and Lubricating Fluid Infused Slippery Behavior on Copper Oxide Nano-structured Substrates" " We construct infinitely many signed graphs having symmetric spectrum, by using the NEPS and rooted product of signed graphs. We also present a method for constructing large cospectral signed graphs. Although the obtained family contains only a minority of signed graphs, it strengthen the belief that the signed graphs with symmetric spectrum are deeper than bipartite graphs, i.e the unsigned graphs with symmetric spectrum. ",Some non-sign-symmetric signed graphs with symmetric spectrum " We show that the analytic continuations of Helson zeta functions $ \zeta_\chi (s)= \sum_1^{\infty}\chi(n)n^{-s} $ can have essentially arbitrary poles and zeroes in the strip $ 21/40 < \Re s < 1 $ (unconditionally), and in the whole critical strip $ 1/2 < \Re s <1 $ under Riemann Hypothesis. ",On zeroes and poles of Helson zeta functions " Current methods of deploying robots that operate in dynamic, uncertain environments, such as Uncrewed Aerial Systems in search \& rescue missions, require nearly continuous human supervision for vehicle guidance and operation. These methods do not consider high-level mission context resulting in cumbersome manual operation or inefficient exhaustive search patterns. We present a human-centered autonomous framework that infers geospatial mission context through dynamic feature sets, which then guides a probabilistic target search planner. Operators provide a set of diverse inputs, including priority definition, spatial semantic information about ad-hoc geographical areas, and reference waypoints, which are probabilistically fused with geographical database information and condensed into a geospatial distribution representing an operator's preferences over an area. An online, POMDP-based planner, optimized for target searching, is augmented with this reward map to generate an operator-constrained policy. Our results, simulated based on input from five professional rescuers, display effective task mental model alignment, 18\% more victim finds, and 15 times more efficient guidance plans then current operational methods. ",Human-Centered Autonomy for UAS Target Search " On Earth, the development of technology required easy access to open air combustion, which is only possible when oxygen partial pressure, P(O$_2$), is above 18\%. This suggests that only planets with significant atmospheric oxygen concentrations will be capable of developing ``advanced'' technospheres and hence detectable technosignatures. ",The Oxygen Bottleneck for Technospheres " Performance Monitoring Unit (PMU) is a common hardware module in Intel CPUs. It can be used to record various CPU behaviors therefore it is often used for performance analysis and optimization. Of the 65536 event spaces, Intel has officially published only 200 or so. In this paper, we design a hidden PMU event collection method. And we found a large number of undocumented PMU events in CPUs of Skylake, Kabylake, and Alderlake microarchitectures. We further demonstrate the existence of these events by using them for transient execution attack detection and build-side channel attacks. This also implies that these hidden PMU events have huge exploitation potential and security threats. ",Exploration and Exploitation of Hidden PMU Events " We consider an efficiently decodable non-adaptive group testing (NAGT) problem that meets theoretical bounds. The problem is to find a few specific items (at most $d$) satisfying certain characteristics in a colossal number of $N$ items as quickly as possible. Those $d$ specific items are called \textit{defective items}. The idea of NAGT is to pool a group of items, which is called \textit{a test}, then run a test on them. If the test outcome is \textit{positive}, there exists at least one defective item in the test, and if it is \textit{negative}, there exists no defective items. Formally, a binary $t \times N$ measurement matrix $\mathcal{M} = (m_{ij})$ is the representation for $t$ tests where row $i$ stands for test $i$ and $m_{ij} = 1$ if and only if item $j$ belongs to test $i$. There are three main objectives in NAGT: minimize the number of tests $t$, construct matrix $\mathcal{M}$, and identify defective items as quickly as possible. In this paper, we present a strongly explicit construction of $\mathcal{M}$ for when the number of defective items is at most 2, with the number of tests $t \simeq 16 \log{N} = O(\log{N})$. In particular, we need only $K \simeq N \times 16\log{N} = O(N\log{N})$ bits to construct such matrices, which is optimal. Furthermore, given these $K$ bits, any entry in the matrix can be constructed in time $O \left(\ln{N}/ \ln{\ln{N}} \right)$. Moreover, $\mathcal{M}$ can be decoded with high probability in time $O\left( \frac{\ln^2{N}}{\ln^2{\ln{N}}} \right)$. When the number of defective items is greater than 2, we present a scheme that can identify at least $(1-\epsilon)d$ defective items with $t \simeq 32 C(\epsilon) d \log{N} = O(d \log{N})$ in time $O \left( d \frac{\ln^2{N}}{\ln^2{\ln{N}}} \right)$ for any close-to-zero $\epsilon$, where $C(\epsilon)$ is a constant that depends only on $\epsilon$. ",Non-Adaptive Group Testing Framework based on Concatenation Code The geometric realizations of Lusztig's symmetries of symmetrizable quantum groups are given in this paper. This construction is a generalization of that in [19]. ,Geometric realizations of Lusztig's symmetries of symmetrizable quantum groups " We prove a new bound for the number of connected components of a real regular elliptic surface with a real section and we show the sharpness of this bound. Furthermore, all possible values for the Betti numbers of such a surface are realized. ",Nombres de Betti des surfaces elliptiques r\'{e}elles " The growing need to deal with massive instances motivates the design of algorithms balancing the quality of the solution with applicability. For the latter, an important measure is the \emph{adaptive complexity}, capturing the number of sequential rounds of parallel computation needed. In this work we obtain the first \emph{constant factor} approximation algorithm for non-monotone submodular maximization subject to a knapsack constraint with \emph{near-optimal} $O(\log n)$ adaptive complexity. Low adaptivity by itself, however, is not enough: one needs to account for the total number of function evaluations (or value queries) as well. Our algorithm asks $\tilde{O}(n^2)$ value queries, but can be modified to run with only $\tilde{O}(n)$ instead, while retaining a low adaptive complexity of $O(\log^2n)$. Besides the above improvement in adaptivity, this is also the first \emph{combinatorial} approach with sublinear adaptive complexity for the problem and yields algorithms comparable to the state-of-the-art even for the special cases of cardinality constraints or monotone objectives. Finally, we showcase our algorithms' applicability on real-world datasets. ",Submodular Maximization subject to a Knapsack Constraint: Combinatorial Algorithms with Near-optimal Adaptive Complexity " This papers deals with the constrained discounted control of piecewise deterministic Markov process (PDMPs) in general Borel spaces. The control variable acts on the jump rate and transition measure, and the goal is to minimize the total expected discounted cost, composed of positive running and boundary costs, while satisfying some constraints also in this form. The basic idea is, by using the special features of the PDMPs, to re-write the problem via an embedded discrete-time Markov chain associated to the PDMP and re-formulate the problem as an infinite dimensional linear programming (LP) problem, via the occupation measures associated to the discrete-time process. It is important to stress however that our new discrete-time problem is not in the same framework of a general constrained discrete-time Markov Decision Process and, due to that, some conditions are required to get the equivalence between the continuous-time problem and the LP formulation. We provide in the sequel sufficient conditions for the solvability of the associated LP problem, based on a generalization of Theorem 4.1 in [8]. In the Appendix we present the proof of this generalization which, we believe, is of interest on its own. The paper is concluded with some examples to illustrate the obtained results. ",A Linear Programming Formulation for Constrained Discounted Continuous Control for Piecewise Deterministic Markov Processes " We study the $S_3$-symmetric two Higgs doublet model by adding two generations of vector like leptons (VLL) which are odd under a discrete $Z_2$ symmetry. The lightest neutral component of the VLL acts as a dark matter (DM) whereas the full VLL set belongs to a dark sector with no mixings allowed with the standard model fermions. We analyse the model in light of dark matter and collider searches. We show that the DM is compatible with the current relic density data as well as satisfying all direct and indirect dark matter search constraints. We choose some representative points in the model parameter space allowed by all aforementioned dark matter constraints and present a detailed collider analysis of multi-lepton signal viz. the mono-lepton, di-lepton, tri-lepton and four-lepton along with missing transverse energy in the final state using both the cut-based analysis and multivariate analysis respectively at the high luminosity 14 TeV LHC run. ",Dark Matter and Collider Searches in $S_3$-Symmetric 2HDM with Vector Like Leptons " An algorithm is proposed for determining asymptotics of the sum of a perturbative series in the strong coupling limit using given values of the expansion coefficients. Operation of the algorithm is illustrated by test examples, method for estimating errors is developed, and an optimization procedure is described. Application of the algorithm to the $\phi^4$ theory gives a behavior $\beta(g)\approx 7.4 g^{0.96}$ at large $g$ for its Gell-Mann -- Low function. The fact that the exponent is close to unity can be interpreted as a manifestation of the logarithmic branching of the type $\beta(g)\sim g (\ln g)^{-\gamma}$ (with $\gamma\approx 0.14$), which is confirmed by independent evidence. In any case, the $\phi^4$ theory is internally consistent. The procedure of summing perturbartive series with arbitrary values of expansion parameter is discussed. ",Summing Divergent Perturbative Series in a Strong Coupling Limit. The Gell-Mann - Low Function of the \phi^4 Theory " We give elementary proofs of several Stirling's precise bounds. We first improve all the precise bounds from the literature and give new precise bounds. In particular, we show that for all $n\ge 8$ $$\sqrt{2\pi n}\left(\frac{n}{e}\right)^n e^{\frac{1}{12n}-\frac{1}{360n^3+103n}} \ge n!\ge \sqrt{2\pi n}\left(\frac{n}{e}\right)^n e^{\frac{1}{12n}-\frac{1}{360n^3+102n}}$$ and for all $n\ge 3$ $$\sqrt{2\pi n}\left(\frac{n}{e}\right)^n e^{\frac{1}{12n+\frac{2}{5n}-\frac{1.1}{10n^3}}} \ge n!\ge \sqrt{2\pi n}\left(\frac{n}{e}\right)^n e^{\frac{1}{12n+\frac{2}{5n}-\frac{0.9}{10n^3}}}.$$ ",Elementary Proofs of Some Stirling Bounds " Graphene is the two-dimensional building block for carbon allotropes of every other dimensionality. Since its experimental discovery, graphene continues to attract enormous interest, in particular as a new kind of matter, in which electron transport is governed by a Dirac-like wave equation, and as a model system for studying electronic and phonon properties of other, more complex, graphitic materials[1-4]. Here, we uncover the constitutive relation of graphene and probe new physics of its optical phonons, by studying its Raman spectrum as a function of uniaxial strain. We find that the doubly degenerate E2g optical mode splits in two components, one polarized along the strain and the other perpendicular to it. This leads to the splitting of the G peak into two bands, which we call G+ and G-, by analogy with the effect of curvature on the nanotube G peak[5-7]. Both peaks red shift with increasing strain, and their splitting increases, in excellent agreement with first-principles calculations. Their relative intensities are found to depend on light polarization, which provides a useful tool to probe the graphene crystallographic orientation with respect to the strain. The singly degenerate 2D and 2D' bands also red shift, but do not split for small strains. We study the Gruneisen parameters for the phonons responsible for the G, D and D' peaks. These can be used to measure the amount of uniaxial or biaxial strain, providing a fundamental tool for nanoelectronics, where strain monitoring is of paramount importance[8, 9] ","Uniaxial Strain in Graphene by Raman Spectroscopy: G peak splitting, Gruneisen Parameters and Sample Orientation" " Early screening of patients is a critical issue in order to assess immediate and fast responses against the spread of COVID-19. The use of nasopharyngeal swabs has been considered the most viable approach; however, the result is not immediate or, in the case of fast exams, sufficiently accurate. Using Chest X-Ray (CXR) imaging for early screening potentially provides faster and more accurate response; however, diagnosing COVID from CXRs is hard and we should rely on deep learning support, whose decision process is, on the other hand, ""black-boxed"" and, for such reason, untrustworthy. We propose an explainable two-step diagnostic approach, where we first detect known pathologies (anomalies) in the lungs, on top of which we diagnose the illness. Our approach achieves promising performance in COVID detection, compatible with expert human radiologists. All of our experiments have been carried out bearing in mind that, especially for clinical applications, explainability plays a major role for building trust in machine learning algorithms. ",A two-step explainable approach for COVID-19 computer-aided diagnosis from chest x-ray images " We show that the skew-growth function of a dual Artin monoid of finite type P has exactly rank(P) =: l simple real zeros on the interval (0, 1]. The proofs for types A_l and B_l are based on an unexpected fact that the skew-growth functions, up to a trivial factor, are expressed by Jacobi polynomials due to a Rodrigues type formula in the theory of orthogonal polynomials. The skew-growth functions for type D_l also satisfy Rodrigues type formulae, but the relation with Jacobi polynomials is not straightforward, and the proof is intricate. We show that the smallest root converges to zero as the rank l of all the above types tend to infinity. ",Zero loci of skew-growth functions for dual Artin monoids We prove local existence and uniqueness of static spherically symmetric solutions of the Einstein-Yang-Mills equations for an arbitrary compact semisimple gauge group in the so-called regular case. By this we mean the equations obtained when the rotation group acts on the principal bundle on which the Yang-Mills connection takes its values in a particularly simple way (the only one ever considered in the literature). The boundary value problem that results for possible asymptotically flat soliton or black hole solutions is very singular and just establishing that local power series solutions exist at the center and asymptotic solutions at infinity amounts to a nontrivial algebraic problem. We discuss the possible field equations obtained for different group actions and solve the algebraic problem on how the local solutions depend on initial data at the center and at infinity. ,Local existence proofs for the boundary value problem for static spherically symmetric Einstein-Yang-Mills fields with compact gauge groups " We construct examples of surfaces of general type with $p_g=1$, $q=0$ and $K^2=6$. We use as key varieties Fano fourfolds and Calabi-Yau threefolds that are zero section of some special homogeneous vector bundle on Grassmannians. We link as well our construction to a classical Campedelli surface, using the Pfaffian-Grassmannian correspondence. ","Surfaces of general type with $p_g=1$, $q=0$, $K^2=6$ and Grassmannians" " Spin waves are propagating disturbances in magnetically ordered materials, analogous to lattice waves in solid systems and are often described from a quasiparticle point of view as magnons. The attractive advantages of Joule-heat-free transmission of information, utilization of the phase of the wave as an additional degree of freedom and lower footprint area compared to conventional charge-based devices have made spin waves or magnon spintronics a promising candidate for beyond-CMOS wave-based computation. However, any practical realization of an all-magnon based computing system must undergo the essential steps of a careful selection of materials and demonstrate robustness with respect to thermal noise or variability. Here, we aim at identifying suitable materials and theoretically demonstrate the possibility of achieving error-free clocked non-volatile spin wave logic device, even in the presence of thermal noise and clock jitter or clock skew. ",Overcoming thermal noise in non-volatile spin wave logic " Occupational segregation is widely considered as one major reason leading to the gender discrimination in labor market. Using large-scale Chinese resume data of online job seekers, we uncover an interesting phenomenon that occupations with higher proportion of men have smaller gender wage gap measured by the female-male ratio on wage. We further show that the severity of occupational segregation in China is low both overall and regionally, and the inter-occupational discrimination is much smaller than the intra-occupational discrimination. That is to say, Chinese women do not face large barriers when changing their occupations. Accordingly, we suggest Chineses women a new way to narrow the gender wage gap: to join male-dominated occupations. Meanwhile, it is worth noticing that although the gender wage gap is smaller in male-dominated occupations, it does not mean that the gender discrimination is smaller there. ",Jumping to male-dominated occupations: A novel way to reduce gender wage gap for Chinese women " Pion pion scattering is studied in a generalized linear sigma model which contains two scalar nonets (one of quark-antiquark type and the other of diquark-antidiquark type) and two corresponding pseudoscalar nonets. An interesting feature concerns the mixing of the four isosinglet scalar mesons which yield poles in the scattering amplitude. Some realism is introduced by enforcing exact unitarity via the K-matrix method. It is shown that a reasonable agreement with experimental data is obtained up to about 1 GeV. The poles in the unitarized scattering amplitude are studied in some detail. The lowest pole clearly represents the sigma meson (or f0(600)) with a mass and decay width around 500 MeV. The second pole invites comparison with the f0(980) which has a mass around 1 GeV and decay width around 100 MeV. The third and fourth poles, resemble some of the isosinglet state in the complicated 1-2 GeV region. Some comparison is made to the situation in the usual SU(3) linear sigma model with a single scalar nonet. ",Chiral Nonet Mixing in pi pi Scattering " The singular values of products of standard complex Gaussian random matrices, or sub-blocks of Haar distributed unitary matrices, have the property that their probability distribution has an explicit, structured form referred to as a polynomial ensemble. It is furthermore the case that the corresponding bi-orthogonal system can be determined in terms of Meijer G-functions, and the correlation kernel given as an explicit double contour integral. It has recently been shown that the Hermitised product $X_M \cdots X_2 X_1A X_1^T X_2^T \cdots X_M^T$, where each $X_i$ is a standard real complex Gaussian matrix, and $A$ is real anti-symmetric shares exhibits analogous properties. Here we use the theory of spherical functions and transforms to present a theory which, for even dimensions, includes these properties of the latter product as a special case. As an example we show that the theory also allows for a treatment of this class of Hermitised product when the $X_i$ are chosen as sub-blocks of Haar distributed real orthogonal matrices. ",Multiplicative Convolution of Real Asymmetric and Real Antisymmetric Matrices " This paper deals with the comparative use of the chemical shift surfaces to simulate experimental 13C CPMAS data on amorphous solid state disaccharides, paying particular attention to -1-1 linkage of trehalose, to -1,4 linkage between pyranose rings (lactose) and to linkage implying a furanose ring (sucrose). The combination of molecular mechanics with DFT/GIAO ab-initio methods provides reliable structural information on the conformational distribution in the glass. The results are interpreted in terms of an enhanced flexibility that trehalose experiences in amorphous solid state compared to the other sugars. An attempt to relate this property to the balance between intra- and inter-molecular hydrogen bonding network in the glass is presented. ",Exploring conformational energy landscape of glassy disaccharides by CPMAS 13C NMR and DFT/GIAO simulations. II. Enhanced molecular flexibility in amorphous trehalose " Virtual screening, including molecular docking, plays an essential role in drug discovery. Many traditional and machine-learning based methods are available to fulfil the docking task. The traditional docking methods are normally extensively time-consuming, and their performance in blind docking remains to be improved. Although the runtime of docking based on machine learning is significantly decreased, their accuracy is still limited. In this study, we take the advantage of both traditional and machine-learning based methods, and present a method Deep Site and Docking Pose (DSDP) to improve the performance of blind docking. For the traditional blind docking, the entire protein is covered by a cube, and the initial positions of ligands are randomly generated in the cube. In contract, DSDP can predict the binding site of proteins and provide an accurate searching space and initial positions for the further conformational sampling. The docking task of DSDP makes use of the score function and a similar but modified searching strategy of AutoDock Vina, accelerated by implementation in GPUs. We systematically compare its performance with the state-of-the-art methods, including Autodock Vina, GNINA, QuickVina, SMINA, and DiffDock. DSDP reaches a 29.8% top-1 success rate (RMSD < 2 {\AA}) on an unbiased and challenging test dataset with 1.2 s wall-clock computational time per system. Its performances on DUD-E dataset and the time-split PDBBind dataset used in EquiBind, TankBind, and DiffDock are also effective, presenting a 57.2% and 41.8% top-1 success rate with 0.8 s and 1.0 s per system, respectively. ",DSDP: A Blind Docking Strategy Accelerated by GPUs " The analysis of the 21 cm signature of cosmic string wakes is extended in several ways. First we consider the constraints on $G\mu$ from the absorption signal of shock heated wakes laid down much later than matter radiation equality. Secondly we analyze the signal of diffuse wake, that is those wakes in which there is a baryon overdensity but which have not shock heated. Finally we compare the size of these signals compared to the expected thermal noise per pixel which dominates over the background cosmic gas brightness temperature and find that the cosmic string signal will exceed the thermal noise of an individual pixel in the Square Kilometre Array for string tensions $G\mu > 2.5 \times 10^{-8}$. ",The 21 cm Signature of Shock Heated and Diffuse Cosmic String Wakes " The joint modeling of longitudinal and time-to-event data is an active area of statistics research that has received a lot of attention in the recent years. More recently, a new and attractive application of this type of models has been to obtain individualized predictions of survival probabilities and/or of future longitudinal responses. The advantageous feature of these predictions is that they are dynamically updated as extra longitudinal responses are collected for the subjects of interest, providing real time risk assessment using all recorded information. The aim of this paper is two-fold. First, to highlight the importance of modeling the association structure between the longitudinal and event time responses that can greatly influence the derived predictions, and second, to illustrate how we can improve the accuracy of the derived predictions by suitably combining joint models with different association structures. The second goal is achieved using Bayesian model averaging, which, in this setting, has the very intriguing feature that the model weights are not fixed but they are rather subject- and time-dependent, implying that at different follow-up times predictions for the same subject may be based on different models. ",Combining Dynamic Predictions from Joint Models for Longitudinal and Time-to-Event Data using Bayesian Model Averaging " For $k \geq 1$ and $n \geq 2k$, the well known Kneser graph $\operatorname{KG}(n,k)$ has all $k$-element subsets of an $n$-element set as vertices; two such subsets are adjacent if they are disjoint. Schrijver constructed a vertex-critical subgraph $\operatorname{SG}(n,k)$ of $\operatorname{KG}(n,k)$ with the same chromatic number. In this paper, we compute the diameter of the graph $\operatorname{SG}(2k+r,k)$ with $r \geq 1$. We obtain an exact value of the diameter of $\operatorname{SG}(2k+r,k)$ when $r \in \{1,2\}$ or when $r \geq k-3$. For the remained cases, when $3 \leq r \leq k-4$, we obtain that the diameter of $\operatorname{SG}(2k+r,k)$ belongs to the integer interval $[4..k-r-1]$. ",On the diameter of Schrijver graphs " We formulate a class of angular Gaussian distributions that allows different degrees of isotropy for directional random variables of arbitrary dimension. Through a series of novel reparameterization, this distribution family is indexed by parameters with meaningful statistical interpretations that can range over the entire real space of an adequate dimension. The new parameterization greatly simplifies maximum likelihood estimation of all model parameters, which in turn leads to theoretically sound and numerically stable inference procedures to infer key features of the distribution. Byproducts from the likelihood-based inference are used to develop graphical and numerical diagnostic tools for assessing goodness of fit of this distribution in a data application. Simulation study and application to data from a hydrogeology study are used to demonstrate implementation and performance of the inference procedures and diagnostics methods. ",Elliptically symmetric distributions for directional data of arbitrary dimension " We study identification of stochastic Wiener dynamic systems using so-called indirect inference. The main idea is to first fit an auxiliary model to the observed data and then in a second step, often by simulation, fit a more structured model to the estimated auxiliary model. This two-step procedure can be used when the direct maximum-likelihood estimate is difficult or intractable to compute. One such example is the identification of stochastic Wiener systems, i.e.,~linear dynamic systems with process noise where the output is measured using a non-linear sensor with additive measurement noise. It is in principle possible to evaluate the log-likelihood cost function using numerical integration, but the corresponding optimization problem can be quite intricate. This motivates studying consistent, but sub-optimal, identification methods for stochastic Wiener systems. We will consider indirect inference using the best linear approximation as an auxiliary model. We show that the key to obtain a reliable estimate is to use uncertainty weighting when fitting the stochastic Wiener model to the auxiliary model estimate. The main technical contribution of this paper is the corresponding asymptotic variance analysis. A numerical evaluation is presented based on a first-order finite impulse response system with a cubic non-linearity, for which certain illustrative analytic properties are derived. ",Identification of Stochastic Wiener Systems using Indirect Inference " We derive from first principles a one-way radiative transfer equation for the wave intensity resolved over directions (Wigner transform of the wave field) in random media. It is an initial value problem with excitation from a source which emits waves in a preferred, forward direction. The equation is derived in a regime with small random fluctuations of the wave speed but long distances of propagation with respect to the wavelength, so that cumulative scattering is significant. The correlation length of the medium and the scale of the support of the source are slightly larger than the wavelength, and the waves propagate in a wide cone with opening angle less than $180^o$, so that the backward and evanescent waves are negligible. The scattering regime is a bridge between that of radiative transfer, where the waves propagate in all directions and the paraxial regime, where the waves propagate in a narrow angular cone. We connect the one-way radiative transport equation with the equations satisfied by the Wigner transform of the wave field in these regimes. ",Derivation of a one-way radiative transfer equation in random media " We provide a complete spectral analysis of all self-adjoint operators acting on $\ell^{2}(\mathbb{Z})$ which are associated with two doubly infinite Jacobi matrices with entries given by $$ q^{-n+1}\delta_{m,n-1}+q^{-n}\delta_{m,n+1} $$ and $$ \delta_{m,n-1}+\alpha q^{-n}\delta_{m,n}+\delta_{m,n+1}, $$ respectively, where $q\in(0,1)$ and $\alpha\in\mathbb{R}$. As an application, we derive orthogonality relations for the Ramanujan entire function and the third Jackson $q$-Bessel function. ",Spectral analysis of two doubly infinite Jacobi matrices with exponential entries " We consider a banking network represented by a system of stochastic differential equations coupled by their drift. We assume a core-periphery structure, and that the banks in the core hold a bubbly asset. The banks in the periphery have not direct access to the bubble, but can take initially advantage from its increase by investing on the banks in the core. Investments are modeled by the weight of the links, which is a function of the robustness of the banks. In this way, a preferential attachment mechanism towards the core takes place during the growth of the bubble. We then investigate how the bubble distort the shape of the network, both for finite and infinitely large systems, assuming a non vanishing impact of the core on the periphery. Due to the influence of the bubble, the banks are no longer independent, and the law of large numbers cannot be directly applied at the limit. This results in a term in the drift of the diffusions which does not average out, and that increases systemic risk at the moment of the burst. We test this feature of the model by numerical simulations. ",Financial asset bubbles in banking networks " The origin and early evolution of eukaryotes are one of the major transitions in the evolution of life on earth. One of its most interesting aspects is the emergence of cellular organelles, their dynamics, their functions, and their divergence. Cell compartmentalization and architecture in prokaryotes is a less understood complex property. In eukaryotes it is related to cell size, specific genomic architecture, evolution of cell cycles, biogenesis of membranes and endosymbiotic processes. Explaining cell evolution through form and function demands an interdisciplinary approach focused on microbial diversity, phylogenetic and functional cell biology. Two centuries of views on eukaryotic origin have completed the disciplinary tools necessarily to answer these questions. We have moved from Haeckel SCALA NATURAE to the un-rooted tree of life. However, the major relations among cell domains are still elusive and keep the nature of eukaryotic ancestor enigmatic. Here we present a review on state of art views of eukaryogenesis; the background and perspectives of different disciplines involved in this topic ",On the ideas of the origin of eukaryotes: a critical review " We consider topological sigma models with generalized Kahler target spaces. The mirror map is constructed explicitly for a special class of target spaces and the topological A and B model are shown to be mirror pairs in the sense that the observables, the instantons and the anomalies are mapped to each other. We also apply the construction to open topological models and show that A branes are mapped to B branes. Furthermore, we demonstrate a relation between the field strength on the brane and a two-vector on the mirror manifold. ",Mirror symmetry for topological sigma models with generalized Kahler geometry " For a massive scalar field with a general curvature coupling parameter, we investigate the finite temperature contributions to the Hadamard function and to the charge and current densities in the geometry of a magnetic flux carrying generalized cosmic string embedded in $(D+1)$-dimensional locally AdS spacetime with a compactified spatial dimension. For $D=4$, the geometry on the AdS boundary, in the context of the AdS/CFT duality, corresponds to a cosmic string as a linear defect, compactified along its axis. In contrast to the case of the Minkowski bulk, the upper bound on the chemical potential does not depend on the field mass and is completely determined by the length of compact dimension and by the enclosed magnetic flux. The only nonzero components correspond to the charge density, to the azimuthal current and to the current along the compact dimension. They are periodic functions of magnetic fluxes with the period equal to the flux quantum. The charge density is an odd function and the currents are even functions of the chemical potential. At high temperatures the influence of the gravitational field and topology on the charge density is subdominant and the leading term in the corresponding expansion coincides with that for the charge density in the Minkowski spacetime. The current densities are topology-induced quantities and their behavior at high temperatures is completely different with the linear dependence on the temperature. At small temperatures and for chemical potentials smaller than the critical value, the thermal expectation values are exponentially suppressed for both massive and massless fields. ",Finite temperature charge and current densities around a cosmic string in AdS spacetime with compact dimension " We develop a theory of semialgebra Grassmann triples via Hasse-Schmidt derivations, which formally generalizes results such as the Cayley-Hamilton theorem in linear algebra, thereby providing a unified approach to classical linear algebra and tropical algebra. ",Grassman semialgebras and the Cayley-Hamilton theorem " Millions of drivers worldwide have enjoyed financial benefits and work schedule flexibility through a ride-sharing economy, but meanwhile they have suffered from the lack of a sense of identity and career achievement. Equipped with social identity and contest theories, financially incentivized team competitions have been an effective instrument to increase drivers' productivity, job satisfaction, and retention, and to improve revenue over cost for ride-sharing platforms. While these competitions are overall effective, the decisive factors behind the treatment effects and how they affect the outcomes of individual drivers have been largely mysterious. In this study, we analyze data collected from more than 500 large-scale team competitions organized by a leading ride-sharing platform, building machine learning models to predict individual treatment effects. Through a careful investigation of features and predictors, we are able to reduce out-sample prediction error by more than 24%. Through interpreting the best-performing models, we discover many novel and actionable insights regarding how to optimize the design and the execution of team competitions on ride-sharing platforms. A simulated analysis demonstrates that by simply changing a few contest design options, the average treatment effect of a real competition is expected to increase by as much as 26%. Our procedure and findings shed light on how to analyze and optimize large-scale online field experiments in general. ",Predicting Individual Treatment Effects of Large-scale Team Competitions in a Ride-sharing Economy " We examine the clustering and kinematics of young stellar objects (YSOs) in the North America/Pelican Nebulae, as revealed by Gaia astrometry, in relation to the structure and motions of the molecular gas, as indicated in molecular line maps. The Gaia parallaxes and proper motions allow us to significantly refine previously published lists of YSOs, demonstrating that many of the objects previously thought to form a distributed population turn out to be non-members. The members are subdivided into at least 6 spatio-kinematic groups, each of which is associated with its own molecular cloud component or components. Three of the groups are expanding, with velocity gradients of 0.3-0.5 km/s/pc, up to maximum velocities of ~8 km/s away from the groups' centers. The two known O-type stars associated with the region, 2MASS J20555125+4352246 and HD 199579, are rapidly escaping one of these groups, following the same position-velocity relation as the low-mass stars. We calculate that a combination of gas expulsion and tidal forces from the clumpy distribution of molecular gas could impart the observed velocity gradients within the groups. However, on a global scale, the relative motions of the groups do not appear either divergent or convergent. The velocity dispersion of the whole system is consistent with the kinetic energy gained due to gravitational collapse of the complex. Most of the stellar population has ages similar to the free-fall timescales for the natal clouds. Thus, we suggest the nearly free-fall collapse of a turbulent molecular cloud as the most likely scenario for star formation in this complex. ",The Formation of a Stellar Association in the NGC 7000/IC 5070 Complex: Results from Kinematic Analysis of Stars and Gas " Noncritical M-theory in 2+1 dimensions has been defined as a double-scaling limit of a nonrelativistic Fermi liquid on a flat two-dimensional plane. Here we study this noncritical M-theory in the limit of high energies, analogous to the \alpha'\to\infty limit of string theory. In the related case of two-dimensional Type 0A strings, it has been argued that the conformal \alpha'\to\infty limit leads to AdS_2 with a propagating fermion whose mass is set by the value of the RR flux. Here we provide evidence that in the high-energy limit, the natural ground state of noncritical M-theory similarly describes the AdS_2\times S^1 spacetime, with a massless propagating fermion. We argue that the spacetime effective theory in this background is captured by a topological higher-spin extension of conformal Chern-Simons gravity in 2+1 dimensions, consistently coupled to a massless Dirac field. Intriguingly, the two-dimensional plane populated by the original nonrelativistic fermions is essentially the twistor space associated with the symmetry group of the AdS_2\times S^1 spacetime; thus, at least in the high-energy limit, noncritical M-theory can be nonperturbatively described as a ""Fermi liquid on twistor space."" ",Strings on AdS_2 and the High-Energy Limit of Noncritical M-Theory " Real-world generalization, e.g., deciding to approach a never-seen-before animal, relies on contextual information as well as previous experiences. Such a seemingly easy behavioral choice requires the interplay of multiple neural mechanisms, from integrative encoding to category-based inference, weighted differently according to the circumstances. Here, we argue that a comprehensive theory of the neuro-cognitive substrates of real-world generalization will greatly benefit from empirical research with three key elements. First, the ecological validity provided by multimodal, naturalistic paradigms. Second, the model stability afforded by deep sampling. Finally, the statistical rigor granted by predictive modeling and computational controls. ","A roadmap to reverse engineering real-world generalization by combining naturalistic paradigms, deep sampling, and predictive computational models" " We present a new analysis of $A_N$ in $p^\uparrow p\to \pi\, X$ within the collinear twist-3 factorization formalism. We incorporate recently derived Lorentz invariance relations into our calculation and focus on input from the kinematical twist-3 functions, which are weighted integrals of transverse momentum dependent (TMD) functions. In particular, we use the latest extractions of the Sivers and Collins functions with TMD evolution to compute certain terms in $A_N$. Consequently, we are able to constrain the remaining contributions from the lesser known dynamical twist-3 correlators. ","Phenomenological constraints on $A_N$ in $p^\uparrow p\to \pi\, X$ from Lorentz invariance relations" " We review the properties and nature of luminous high-redshift radio galaxies (HzRGs, z > 2) and the environments in which they are located. HzRGs have several distinct constituents which interact with each other - relativistic plasma, gas in various forms, dust, stars and an active galactic nucleus (AGN). These building blocks provide unique diagnostics about conditions in the early Universe. We discuss the properties of each constituent. Evidence is presented that HzRGs are massive forming galaxies and the progenitors of brightest cluster galaxies in the local Universe. HzRGs are located in overdense regions in the early Universe and are frequently surrounded by protoclusters. We review the properties and nature of these radio-selected protoclusters. Finally we consider the potential for future progress in the field during the next few decades. A compendium of known HzRGs is given in an appendix. ",Distant Radio Galaxies and their Environments " A common problem affecting neural network (NN) approximations of model predictive control (MPC) policies is the lack of analytical tools to assess the stability of the closed-loop system under the action of the NN-based controller. We present a general procedure to quantify the performance of such a controller, or to design minimum complexity NNs with rectified linear units (ReLUs) that preserve the desirable properties of a given MPC scheme. By quantifying the approximation error between NN-based and MPC-based state-to-input mappings, we first establish suitable conditions involving two key quantities, the worst-case error and the Lipschitz constant, guaranteeing the stability of the closed-loop system. We then develop an offline, mixed-integer optimization-based method to compute those quantities exactly. Together these techniques provide conditions sufficient to certify the stability and performance of a ReLU-based approximation of an MPC control law. ",Reliably-stabilizing piecewise-affine neural network controllers " Polar codes are a channel coding scheme for the next generation of wireless communications standard (5G). The belief propagation (BP) decoder allows for parallel decoding of polar codes, making it suitable for high throughput applications. However, the error-correction performance of polar codes under BP decoding is far from the requirements of 5G. It has been shown that the error-correction performance of BP can be improved if the decoding is performed on multiple permuted factor graphs of polar codes. However, a different BP decoding scheduling is required for each factor graph permutation which results in the design of a different decoder for each permutation. Moreover, the selection of the different factor graph permutations is at random, which prevents the decoder to achieve a desirable error-correction performance with a small number of permutations. In this paper, we first show that the permutations on the factor graph can be mapped into suitable permutations on the codeword positions. As a result, we can make use of a single decoder for all the permutations. In addition, we introduce a method to construct a set of predetermined permutations which can provide the correct codeword if the decoding fails on the original permutation. We show that for the 5G polar code of length $1024$, the error-correction performance of the proposed decoder is more than $0.25$ dB better than that of the BP decoder with the same number of random permutations at the frame error rate of $10^{-4}$. ",On the Decoding of Polar Codes on Permuted Factor Graphs " Reliable detection of out-of-distribution (OOD) inputs is increasingly understood to be a precondition for deployment of machine learning systems. This paper proposes and investigates the use of contrastive training to boost OOD detection performance. Unlike leading methods for OOD detection, our approach does not require access to examples labeled explicitly as OOD, which can be difficult to collect in practice. We show in extensive experiments that contrastive training significantly helps OOD detection performance on a number of common benchmarks. By introducing and employing the Confusion Log Probability (CLP) score, which quantifies the difficulty of the OOD detection task by capturing the similarity of inlier and outlier datasets, we show that our method especially improves performance in the `near OOD' classes -- a particularly challenging setting for previous methods. ",Contrastive Training for Improved Out-of-Distribution Detection " The study examines the relationship between mobile financial services and individual financial behavior in India wherein a sizeable population is yet to be financially included. Addressing the endogeneity associated with the use of mobile financial services using an instrumental variable method, the study finds that the use of mobile financial services increases the likelihood of investment, having insurance and borrowing from formal financial institutions. Further, the analysis highlights that access to mobile financial services have the potential to bridge the gender divide in financial inclusion. Fastening the pace of access to mobile financial services may partially alter pandemic induced poverty. ",Effect of mobile financial services on financial behavior in developing economies-Evidence from India " Let $T$ be a `properly metrisable' topological space, that is, locally compact, separable and metrisable. Let $\mathcal{M}^T$ be the non-empty set of all proper metrics $d$ on $T$ compatible with its topology, and equip $\mathcal{M}^T$ with the topology of uniform convergence, where the metrics are regarded as functions on $T^2$. We prove that if $T$ is uncountable then the set $\mathcal{A}^T_f$ of metrics $d\in\mathcal{M}^T$ for which the Lipschitz-free space $\mathcal{F}(T,d)$ fails the approximation property is a dense set in $\mathcal{M}^T$. Combining this with a result of Dalet, we conclude that for any properly metrisable space $T$, $\mathcal{A}^T_f$ is either empty or dense in $\mathcal{M}^T$. ",Lipschitz-free spaces over properly metrisable spaces and approximation properties " We study the generation number of massless fermions in compactifications recently found in heterotic supergravity. The internal spaces are products of two-dimensional spaces of constant curvature, and the standard embedding is not assumed. The generation number is constrained by the equations of motion and the Bianchi identity. In the case that the Euler characteristics of the three internal submanifolds are $(2,-2,-2)$, the generation number is three or less. We also show that three-generation solutions exist for arbitrary Euler characteristics of the negatively curved 2-manifolds, and we present some explicit solutions. ",Three-generation solutions of equations of motion in heterotic supergravity " The efficiency of recent star formation (SF) in galaxies increases with increasing projected distance from the centre of a cluster out to several times its virial radius (R_v). Using a complete sample of galaxies in 268 clusters from the SDSS DR4, we investigate how, at a given projected radius from the cluster centre, M* and SF properties of a galaxy depend on its absolute line-of-sight velocity in the cluster rest frame, |v_LOS|. We find that for R<0.5 R_v, the fraction of high mass non-BCG galaxies increases towards the centre for low |v_LOS|. At a given projected radius, the fraction of Galaxies with Ongoing or Recent (<1-3 Gyr) Efficient Star Formation (GORES, with EW(H_delta)>2 ang & D_4000>1.5) is slightly but significantly lower for low |v_LOS| galaxies than for their high velocity counterparts. We study these observational trends with the help of a dark matter (DM) cosmological simulation. We find that the backsplash particles account for at least one-third (half) of all particles at projected radii slightly greater than the virial radius and |v_LOS|