text
stringlengths
138
2.38k
labels
sequencelengths
6
6
Predictions
sequencelengths
1
3
Title: Tangle-tree duality in abstract separation systems, Abstract: We prove a general width duality theorem for combinatorial structures with well-defined notions of cohesion and separation. These might be graphs and matroids, but can be much more general or quite different. The theorem asserts a duality between the existence of high cohesiveness somewhere local and a global overall tree structure. We describe cohesive substructures in a unified way in the format of tangles: as orientations of low-order separations satisfying certain consistency axioms. These axioms can be expressed without reference to the underlying structure, such as a graph or matroid, but just in terms of the poset of the separations themselves. This makes it possible to identify tangles, and apply our tangle-tree duality theorem, in very diverse settings. Our result implies all the classical duality theorems for width parameters in graph minor theory, such as path-width, tree-width, branch-width or rank-width. It yields new, tangle-type, duality theorems for tree-width and path-width. It implies the existence of width parameters dual to cohesive substructures such as $k$-blocks, edge-tangles, or given subsets of tangles, for which no width duality theorems were previously known. Abstract separation systems can be found also in structures quite unlike graphs and matroids. For example, our theorem can be applied to image analysis by capturing the regions of an image as tangles of separations defined as natural partitions of its set of pixels. It can be applied in big data contexts by capturing clusters as tangles. It can be applied in the social sciences, e.g. by capturing as tangles the few typical mindsets of individuals found by a survey. It could also be applied in pure mathematics, e.g. to separations of compact manifolds.
[ 0, 0, 1, 0, 0, 0 ]
[ "Mathematics", "Computer Science" ]
Title: Essentially No Barriers in Neural Network Energy Landscape, Abstract: Training neural networks involves finding minima of a high-dimensional non-convex loss function. Knowledge of the structure of this energy landscape is sparse. Relaxing from linear interpolations, we construct continuous paths between minima of recent neural network architectures on CIFAR10 and CIFAR100. Surprisingly, the paths are essentially flat in both the training and test landscapes. This implies that neural networks have enough capacity for structural changes, or that these changes are small between minima. Also, each minimum has at least one vanishing Hessian eigenvalue in addition to those resulting from trivial invariance.
[ 0, 0, 0, 1, 0, 0 ]
[ "Computer Science", "Mathematics" ]
Title: Approximating the Backbone in the Weighted Maximum Satisfiability Problem, Abstract: The weighted Maximum Satisfiability problem (weighted MAX-SAT) is a NP-hard problem with numerous applications arising in artificial intelligence. As an efficient tool for heuristic design, the backbone has been applied to heuristics design for many NP-hard problems. In this paper, we investigated the computational complexity for retrieving the backbone in weighted MAX-SAT and developed a new algorithm for solving this problem. We showed that it is intractable to retrieve the full backbone under the assumption that . Moreover, it is intractable to retrieve a fixed fraction of the backbone as well. And then we presented a backbone guided local search (BGLS) with Walksat operator for weighted MAX-SAT. BGLS consists of two phases: the first phase samples the backbone information from local optima and the backbone phase conducts local search under the guideline of backbone. Extensive experimental results on the benchmark showed that BGLS outperforms the existing heuristics in both solution quality and runtime.
[ 1, 0, 0, 0, 0, 0 ]
[ "Computer Science", "Mathematics" ]
Title: New type integral inequalities for convex functions with applications II, Abstract: We have recently established some integral inequalities for convex functions via the Hermite-Hadamard's inequalities. In continuation here, we also establish some interesting new integral inequalities for convex functions via the Hermite--Hadamard's inequalities and Jensen's integral inequality. Useful applications involving special means are also included.
[ 0, 0, 1, 0, 0, 0 ]
[ "Mathematics" ]
Title: Physical problem solving: Joint planning with symbolic, geometric, and dynamic constraints, Abstract: In this paper, we present a new task that investigates how people interact with and make judgments about towers of blocks. In Experiment~1, participants in the lab solved a series of problems in which they had to re-configure three blocks from an initial to a final configuration. We recorded whether they used one hand or two hands to do so. In Experiment~2, we asked participants online to judge whether they think the person in the lab used one or two hands. The results revealed a close correspondence between participants' actions in the lab, and the mental simulations of participants online. To explain participants' actions and mental simulations, we develop a model that plans over a symbolic representation of the situation, executes the plan using a geometric solver, and checks the plan's feasibility by taking into account the physical constraints of the scene. Our model explains participants' actions and judgments to a high degree of quantitative accuracy.
[ 1, 0, 0, 1, 0, 0 ]
[ "Computer Science", "Quantitative Biology" ]
Title: A Stochastic Model for File Lifetime and Security in Data Center Networks, Abstract: Data center networks are an important infrastructure in various applications of modern information technologies. Note that each data center always has a finite lifetime, thus once a data center fails, then it will lose all its storage files and useful information. For this, it is necessary to replicate and copy each important file into other data centers such that this file can increase its lifetime of staying in a data center network. In this paper, we describe a large-scale data center network with a file d-threshold policy, which is to replicate each important file into at most d-1 other data centers such that this file can maintain in the data center network under a given level of data security in the long-term. To this end, we develop three relevant Markov processes to propose two effective methods for assessing the file lifetime and data security. By using the RG-factorizations, we show that the two methods are used to be able to more effectively evaluate the file lifetime of large-scale data center networks. We hope the methodology and results given in this paper are applicable in the file lifetime study of more general data center networks with replication mechanism.
[ 1, 0, 0, 0, 0, 0 ]
[ "Computer Science", "Statistics" ]
Title: A Theoretical Perspective of Solving Phaseless Compressed Sensing via Its Nonconvex Relaxation, Abstract: As a natural extension of compressive sensing and the requirement of some practical problems, Phaseless Compressed Sensing (PCS) has been introduced and studied recently. Many theoretical results have been obtained for PCS with the aid of its convex relaxation. Motivated by successful applications of nonconvex relaxed methods for solving compressive sensing, in this paper, we try to investigate PCS via its nonconvex relaxation. Specifically, we relax PCS in the real context by the corresponding $\ell_p$-minimization with $p\in (0,1)$. We show that there exists a constant $p^\ast\in (0,1]$ such that for any fixed $p\in(0, p^\ast)$, every optimal solution to the $\ell_p$-minimization also solves the concerned problem; and derive an expression of such a constant $p^\ast$ by making use of the known data and the sparsity level of the concerned problem. These provide a theoretical basis for solving this class of problems via the corresponding $\ell_p$-minimization.
[ 0, 0, 1, 0, 0, 0 ]
[ "Mathematics", "Computer Science" ]
Title: Affine Rough Models, Abstract: The goal of this survey article is to explain and elucidate the affine structure of recent models appearing in the rough volatility literature, and show how it leads to exponential-affine transform formulas.
[ 0, 0, 0, 0, 0, 1 ]
[ "Quantitative Finance", "Mathematics" ]
Title: When flux standards go wild: white dwarfs in the age of Kepler, Abstract: White dwarf stars have been used as flux standards for decades, thanks to their staid simplicity. We have empirically tested their photometric stability by analyzing the light curves of 398 high-probability candidates and spectroscopically confirmed white dwarfs observed during the original Kepler mission and later with K2 Campaigns 0-8. We find that the vast majority (>97 per cent) of non-pulsating and apparently isolated white dwarfs are stable to better than 1 per cent in the Kepler bandpass on 1-hr to 10-d timescales, confirming that these stellar remnants are useful flux standards. From the cases that do exhibit significant variability, we caution that binarity, magnetism, and pulsations are three important attributes to rule out when establishing white dwarfs as flux standards, especially those hotter than 30,000 K.
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics" ]
Title: Stable representations of posets, Abstract: The purpose of this paper is to study stable representations of partially ordered sets (posets) and compare it to the well known theory for quivers. In particular, we prove that every indecomposable representation of a poset of finite type is stable with respect to some weight and construct that weight explicitly in terms of the dimension vector. We show that if a poset is primitive then Coxeter transformations preserve stable representations. When the base field is the field of complex numbers we establish the connection between the polystable representations and the unitary $\chi$-representations of posets. This connection explains the similarity of the results obtained in the series of papers.
[ 0, 0, 1, 0, 0, 0 ]
[ "Mathematics" ]
Title: A Practical Bandit Method with Advantages in Neural Network Tuning, Abstract: Stochastic bandit algorithms can be used for challenging non-convex optimization problems. Hyperparameter tuning of neural networks is particularly challenging, necessitating new approaches. To this end, we present a method that adaptively partitions the combined space of hyperparameters, context, and training resources (e.g., total number of training iterations). By adaptively partitioning the space, the algorithm is able to focus on the portions of the hyperparameter search space that are most relevant in a practical way. By including the resources in the combined space, the method tends to use fewer training resources overall. Our experiments show that this method can surpass state-of-the-art methods in tuning neural networks on benchmark datasets. In some cases, our implementations can achieve the same levels of accuracy on benchmark datasets as existing state-of-the-art approaches while saving over 50% of our computational resources (e.g. time, training iterations).
[ 1, 0, 0, 1, 0, 0 ]
[ "Computer Science", "Statistics" ]
Title: Dynamic Security Analysis of Power Systems by a Sampling-Based Algorithm, Abstract: Dynamic security analysis is an important problem of power systems on ensuring safe operation and stable power supply even when certain faults occur. No matter such faults are caused by vulnerabilities of system components, physical attacks, or cyber-attacks that are more related to cyber-security, they eventually affect the physical stability of a power system. Examples of the loss of physical stability include the Northeast blackout of 2003 in North America and the 2015 system-wide blackout in Ukraine. The nonlinear hybrid nature, that is, nonlinear continuous dynamics integrated with discrete switching, and the high degree of freedom property of power system dynamics make it challenging to conduct the dynamic security analysis. In this paper, we use the hybrid automaton model to describe the dynamics of a power system and mainly deal with the index-1 differential-algebraic equation models regarding the continuous dynamics in different discrete states. The analysis problem is formulated as a reachability problem of the associated hybrid model. A sampling-based algorithm is then proposed by integrating modeling and randomized simulation of the hybrid dynamics to search for a feasible execution connecting an initial state of the post-fault system and a target set in the desired operation mode. The proposed method enables the use of existing power system simulators for the synthesis of discrete switching and control strategies through randomized simulation. The effectiveness and performance of the proposed approach are demonstrated with an application to the dynamic security analysis of the New England 39-bus benchmark power system exhibiting hybrid dynamics. In addition to evaluating the dynamic security, the proposed method searches for a feasible strategy to ensure the dynamic security of the system in face of disruptions.
[ 1, 0, 0, 0, 0, 0 ]
[ "Computer Science", "Physics" ]
Title: Macro diversity in Cellular Networks with Random Blockages, Abstract: Blocking objects (blockages) between a transmitter and receiver cause wireless communication links to transition from line-of-sight (LOS) to non-line-of-sight (NLOS) propagation, which can greatly reduce the received power, particularly at higher frequencies such as millimeter wave (mmWave). We consider a cellular network in which a mobile user attempts to connect to two or more base stations (BSs) simultaneously, to increase the probability of at least one LOS link, which is a form of macrodiversity. We develop a framework for determining the LOS probability as a function of the number of BSs, when taking into account the correlation between blockages: for example, a single blockage close to the device -- including the user's own body -- could block multiple BSs. We consider the impact of the size of blocking objects on the system reliability probability and show that macrodiversity gains are higher when the blocking objects are small. We also show that the BS density must scale as the square of the blockage density to maintain a given level of reliability.
[ 1, 0, 1, 0, 0, 0 ]
[ "Computer Science", "Physics" ]
Title: HybridNet: Classification and Reconstruction Cooperation for Semi-Supervised Learning, Abstract: In this paper, we introduce a new model for leveraging unlabeled data to improve generalization performances of image classifiers: a two-branch encoder-decoder architecture called HybridNet. The first branch receives supervision signal and is dedicated to the extraction of invariant class-related representations. The second branch is fully unsupervised and dedicated to model information discarded by the first branch to reconstruct input data. To further support the expected behavior of our model, we propose an original training objective. It favors stability in the discriminative branch and complementarity between the learned representations in the two branches. HybridNet is able to outperform state-of-the-art results on CIFAR-10, SVHN and STL-10 in various semi-supervised settings. In addition, visualizations and ablation studies validate our contributions and the behavior of the model on both CIFAR-10 and STL-10 datasets.
[ 0, 0, 0, 1, 0, 0 ]
[ "Computer Science", "Statistics" ]
Title: Multipath Error Correction in Radio Interferometric Positioning Systems, Abstract: The radio interferometric positioning system (RIPS) is an accurate node localization method featuring a novel phase-based ranging process. Multipath is the limiting error source for RIPS in ground-deployed scenarios or indoor applications. There are four distinct channels involved in the ranging process for RIPS. Multipath reflections affect both the phase and amplitude of the ranging signal for each channel. By exploiting untapped amplitude information, we put forward a scheme to estimate each channel's multipath profile, which is then subsequently used to correct corresponding errors in phase measurements. Simulations show that such a scheme is very effective in reducing multipath phase errors, which are essentially brought down to the level of receiver noise under moderate multipath conditions. It is further demonstrated that ranging errors in RIPS are also greatly reduced via the proposed scheme.
[ 1, 0, 0, 0, 0, 0 ]
[ "Computer Science", "Physics" ]
Title: Network of vertically c-oriented prism shaped InN nanowalls grown on c-GaN/sapphire template by chemical vapor deposition technique, Abstract: Networks of vertically c-oriented prism shaped InN nanowalls, are grown on c-GaN/sapphire templates using a CVD technique, where pure indium and ammonia are used as metal and nitrogen precursors. A systematic study of the growth, structural and electronic properties of these samples shows a preferential growth of the islands along [11-20] and [0001] directions leading to the formation of such a network structure, where the vertically [0001] oriented tapered walls are laterally align along one of the three [11-20] directions. Inclined facets of these walls are identified as r-planes [(1-102)-planes] of wurtzite InN. Onset of absorption for these samples is observed to be higher than the band gap of InN suggesting a high background carrier concentration in this material. Study of the valence band edge through XPS indicates the formation of positive depletion regions below the r-plane side facets of the walls. This is in contrast with the observation for c-plane InN epilayers, where electron accumulation is often reported below the top surface.
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics" ]
Title: Convolved subsampling estimation with applications to block bootstrap, Abstract: The block bootstrap approximates sampling distributions from dependent data by resampling data blocks. A fundamental problem is establishing its consistency for the distribution of a sample mean, as a prototypical statistic. We use a structural relationship with subsampling to characterize the bootstrap in a new and general manner. While subsampling and block bootstrap differ, the block bootstrap distribution of a sample mean equals that of a $k$-fold self-convolution of a subsampling distribution. Motivated by this, we provide simple necessary and sufficient conditions for a convolved subsampling estimator to produce a normal limit that matches the target of bootstrap estimation. These conditions may be linked to consistency properties of an original subsampling distribution, which are often obtainable under minimal assumptions. Through several examples, the results are shown to validate the block bootstrap for means under significantly weakened assumptions in many existing (and some new) dependence settings, which also addresses a standing conjecture of Politis, Romano and Wolf(1999). Beyond sample means, the convolved subsampling estimator may not match the block bootstrap, but instead provides a hybrid-resampling estimator of interest in its own right. For general statistics with normal limits, results also establish the consistency of convolved subsampling under minimal dependence conditions, including non-stationarity.
[ 0, 0, 1, 1, 0, 0 ]
[ "Statistics", "Mathematics" ]
Title: Computing Simple Multiple Zeros of Polynomial Systems, Abstract: Given a polynomial system f associated with a simple multiple zero x of multiplicity {\mu}, we give a computable lower bound on the minimal distance between the simple multiple zero x and other zeros of f. If x is only given with limited accuracy, we propose a numerical criterion that f is certified to have {\mu} zeros (counting multiplicities) in a small ball around x. Furthermore, for simple double zeros and simple triple zeros whose Jacobian is of normalized form, we define modified Newton iterations and prove the quantified quadratic convergence when the starting point is close to the exact simple multiple zero. For simple multiple zeros of arbitrary multiplicity whose Jacobian matrix may not have a normalized form, we perform unitary transformations and modified Newton iterations, and prove its non-quantified quadratic convergence and its quantified convergence for simple triple zeros.
[ 0, 0, 1, 0, 0, 0 ]
[ "Mathematics", "Computer Science" ]
Title: Latent Geometry and Memorization in Generative Models, Abstract: It can be difficult to tell whether a trained generative model has learned to generate novel examples or has simply memorized a specific set of outputs. In published work, it is common to attempt to address this visually, for example by displaying a generated example and its nearest neighbor(s) in the training set (in, for example, the L2 metric). As any generative model induces a probability density on its output domain, we propose studying this density directly. We first study the geometry of the latent representation and generator, relate this to the output density, and then develop techniques to compute and inspect the output density. As an application, we demonstrate that "memorization" tends to a density made of delta functions concentrated on the memorized examples. We note that without first understanding the geometry, the measurement would be essentially impossible to make.
[ 1, 0, 0, 1, 0, 0 ]
[ "Computer Science", "Statistics", "Mathematics" ]
Title: Landau Collision Integral Solver with Adaptive Mesh Refinement on Emerging Architectures, Abstract: The Landau collision integral is an accurate model for the small-angle dominated Coulomb collisions in fusion plasmas. We investigate a high order accurate, fully conservative, finite element discretization of the nonlinear multi-species Landau integral with adaptive mesh refinement using the PETSc library (www.mcs.anl.gov/petsc). We develop algorithms and techniques to efficiently utilize emerging architectures with an approach that minimizes memory usage and movement and is suitable for vector processing. The Landau collision integral is vectorized with Intel AVX-512 intrinsics and the solver sustains as much as 22% of the theoretical peak flop rate of the Second Generation Intel Xeon Phi, Knights Landing, processor.
[ 1, 0, 0, 0, 0, 0 ]
[ "Physics", "Computer Science" ]
Title: Finite-time scaling at the Anderson transition for vibrations in solids, Abstract: A model in which a three-dimensional elastic medium is represented by a network of identical masses connected by springs of random strengths and allowed to vibrate only along a selected axis of the reference frame, exhibits an Anderson localization transition. To study this transition, we assume that the dynamical matrix of the network is given by a product of a sparse random matrix with real, independent, Gaussian-distributed non-zero entries and its transpose. A finite-time scaling analysis of system's response to an initial excitation allows us to estimate the critical parameters of the localization transition. The critical exponent is found to be $\nu = 1.57 \pm 0.02$ in agreement with previous studies of Anderson transition belonging to the three-dimensional orthogonal universality class.
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics", "Mathematics" ]
Title: On the Status of the Measurement Problem: Recalling the Relativistic Transactional Interpretation, Abstract: In view of a resurgence of concern about the measurement problem, it is pointed out that the Relativistic Transactional Interpretation (RTI) remedies issues previously considered as drawbacks or refutations of the original TI. Specifically, once one takes into account relativistic processes that are not representable at the non-relativistic level (such as particle creation and annihilation, and virtual propagation), absorption is quantitatively defined in unambiguous physical terms. In addition, specifics of the relativistic transactional model demonstrate that the Maudlin `contingent absorber' challenge to the original TI cannot even be mounted: basic features of established relativistic field theories (in particular, the asymmetry between field sources and the bosonic fields, and the fact that slow-moving bound states, such as atoms, are not offer waves) dictate that the `slow-moving offer wave' required for the challenge scenario cannot exist. It is concluded that issues previously considered obstacles for TI are no longer legitimately viewed as such, and that reconsideration of the transactional picture is warranted in connection with solving the measurement problem.
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics" ]
Title: Mixed Precision Training of Convolutional Neural Networks using Integer Operations, Abstract: The state-of-the-art (SOTA) for mixed precision training is dominated by variants of low precision floating point operations, and in particular, FP16 accumulating into FP32 Micikevicius et al. (2017). On the other hand, while a lot of research has also happened in the domain of low and mixed-precision Integer training, these works either present results for non-SOTA networks (for instance only AlexNet for ImageNet-1K), or relatively small datasets (like CIFAR-10). In this work, we train state-of-the-art visual understanding neural networks on the ImageNet-1K dataset, with Integer operations on General Purpose (GP) hardware. In particular, we focus on Integer Fused-Multiply-and-Accumulate (FMA) operations which take two pairs of INT16 operands and accumulate results into an INT32 output.We propose a shared exponent representation of tensors and develop a Dynamic Fixed Point (DFP) scheme suitable for common neural network operations. The nuances of developing an efficient integer convolution kernel is examined, including methods to handle overflow of the INT32 accumulator. We implement CNN training for ResNet-50, GoogLeNet-v1, VGG-16 and AlexNet; and these networks achieve or exceed SOTA accuracy within the same number of iterations as their FP32 counterparts without any change in hyper-parameters and with a 1.8X improvement in end-to-end training throughput. To the best of our knowledge these results represent the first INT16 training results on GP hardware for ImageNet-1K dataset using SOTA CNNs and achieve highest reported accuracy using half-precision
[ 1, 0, 0, 0, 0, 0 ]
[ "Computer Science" ]
Title: Nanostructured complex oxides as a route towards thermal behavior in artificial spin ice systems, Abstract: We have used soft x-ray photoemission electron microscopy to image the magnetization of single domain La$_{0.7}$Sr$_{0.3}$MnO$_{3}$ nano-islands arranged in geometrically frustrated configurations such as square ice and kagome ice geometries. Upon thermal randomization, ensembles of nano-islands with strong inter-island magnetic coupling relax towards low-energy configurations. Statistical analysis shows that the likelihood of ensembles falling into low-energy configurations depends strongly on the annealing temperature. Annealing to just below the Curie temperature of the ferromagnetic film (T$_{C}$ = 338 K) allows for a much greater probability of achieving low energy configurations as compared to annealing above the Curie temperature. At this thermally active temperature of 325 K, the ensemble of ferromagnetic nano-islands explore their energy landscape over time and eventually transition to lower energy states as compared to the frozen-in configurations obtained upon cooling from above the Curie temperature. Thus, this materials system allows for a facile method to systematically study thermal evolution of artificial spin ice arrays of nano-islands at temperatures modestly above room temperature.
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics" ]
Title: A Simple Convolutional Generative Network for Next Item Recommendation, Abstract: Convolutional Neural Networks (CNNs) have been recently introduced in the domain of session-based next item recommendation. An ordered collection of past items the user has interacted with in a session (or sequence) are embedded into a 2-dimensional latent matrix, and treated as an image. The convolution and pooling operations are then applied to the mapped item embeddings. In this paper, we first examine the typical session-based CNN recommender and show that both the generative model and network architecture are suboptimal when modeling long-range dependencies in the item sequence. To address the issues, we introduce a simple, but very effective generative model that is capable of learning high-level representation from both short- and long-range item dependencies. The network architecture of the proposed model is formed of a stack of \emph{holed} convolutional layers, which can efficiently increase the receptive fields without relying on the pooling operation. Another contribution is the effective use of residual block structure in recommender systems, which can ease the optimization for much deeper networks. The proposed generative model attains state-of-the-art accuracy with less training time in the next item recommendation task. It accordingly can be used as a powerful recommendation baseline to beat in future, especially when there are long sequences of user feedback.
[ 0, 0, 0, 1, 0, 0 ]
[ "Computer Science" ]
Title: A Domain Specific Language for Performance Portable Molecular Dynamics Algorithms, Abstract: Developers of Molecular Dynamics (MD) codes face significant challenges when adapting existing simulation packages to new hardware. In a continuously diversifying hardware landscape it becomes increasingly difficult for scientists to be experts both in their own domain (physics/chemistry/biology) and specialists in the low level parallelisation and optimisation of their codes. To address this challenge, we describe a "Separation of Concerns" approach for the development of parallel and optimised MD codes: the science specialist writes code at a high abstraction level in a domain specific language (DSL), which is then translated into efficient computer code by a scientific programmer. In a related context, an abstraction for the solution of partial differential equations with grid based methods has recently been implemented in the (Py)OP2 library. Inspired by this approach, we develop a Python code generation system for molecular dynamics simulations on different parallel architectures, including massively parallel distributed memory systems and GPUs. We demonstrate the efficiency of the auto-generated code by studying its performance and scalability on different hardware and compare it to other state-of-the-art simulation packages. With growing data volumes the extraction of physically meaningful information from the simulation becomes increasingly challenging and requires equally efficient implementations. A particular advantage of our approach is the easy expression of such analysis algorithms. We consider two popular methods for deducing the crystalline structure of a material from the local environment of each atom, show how they can be expressed in our abstraction and implement them in the code generation framework.
[ 1, 1, 0, 0, 0, 0 ]
[ "Computer Science", "Physics", "Quantitative Biology" ]
Title: CMB anisotropies at all orders: the non-linear Sachs-Wolfe formula, Abstract: We obtain the non-linear generalization of the Sachs-Wolfe + integrated Sachs-Wolfe (ISW) formula describing the CMB temperature anisotropies. Our formula is valid at all orders in perturbation theory, is also valid in all gauges and includes scalar, vector and tensor modes. A direct consequence of our results is that the maps of the logarithmic temperature anisotropies are much cleaner than the usual CMB maps, because they automatically remove many secondary anisotropies. This can for instance, facilitate the search for primordial non-Gaussianity in future works. It also disentangles the non-linear ISW from other effects. Finally, we provide a method which can iteratively be used to obtain the lensing solution at the desired order.
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics" ]
Title: On a binary system of Prendiville: The cubic case, Abstract: We prove sharp decoupling inequalities for a class of two dimensional non-degenerate surfaces in R^5, introduced by Prendiville. As a consequence, we obtain sharp bounds on the number of integer solutions of the Diophantine systems associated with these surfaces.
[ 0, 0, 1, 0, 0, 0 ]
[ "Mathematics" ]
Title: Causal Inference by Stochastic Complexity, Abstract: The algorithmic Markov condition states that the most likely causal direction between two random variables X and Y can be identified as that direction with the lowest Kolmogorov complexity. Due to the halting problem, however, this notion is not computable. We hence propose to do causal inference by stochastic complexity. That is, we propose to approximate Kolmogorov complexity via the Minimum Description Length (MDL) principle, using a score that is mini-max optimal with regard to the model class under consideration. This means that even in an adversarial setting, such as when the true distribution is not in this class, we still obtain the optimal encoding for the data relative to the class. We instantiate this framework, which we call CISC, for pairs of univariate discrete variables, using the class of multinomial distributions. Experiments show that CISC is highly accurate on synthetic, benchmark, as well as real-world data, outperforming the state of the art by a margin, and scales extremely well with regard to sample and domain sizes.
[ 1, 0, 0, 0, 0, 0 ]
[ "Computer Science", "Statistics" ]
Title: Automated Speed and Lane Change Decision Making using Deep Reinforcement Learning, Abstract: This paper introduces a method, based on deep reinforcement learning, for automatically generating a general purpose decision making function. A Deep Q-Network agent was trained in a simulated environment to handle speed and lane change decisions for a truck-trailer combination. In a highway driving case, it is shown that the method produced an agent that matched or surpassed the performance of a commonly used reference model. To demonstrate the generality of the method, the exact same algorithm was also tested by training it for an overtaking case on a road with oncoming traffic. Furthermore, a novel way of applying a convolutional neural network to high level input that represents interchangeable objects is also introduced.
[ 1, 0, 0, 0, 0, 0 ]
[ "Computer Science" ]
Title: Reducing biases on $H_0$ measurements using strong lensing and galaxy dynamics: results from the EAGLE simulation, Abstract: Cosmological parameter constraints from observations of time-delay lenses are becoming increasingly precise. However, there may be significant bias and scatter in these measurements due to, among other things, the so-called mass-sheet degeneracy. To estimate these uncertainties, we analyze strong lenses from the largest EAGLE hydrodynamical simulation. We apply a mass-sheet transformation to the radial density profiles of lenses, and by selecting lenses near isothermality, we find that the bias on H0 can be reduced to 5% with an intrinsic scatter of 10%, confirming previous results performed on a different simulation data set. We further investigate whether combining lensing observables with kinematic constraints helps to minimize this bias. We do not detect any significant dependence of the bias on lens model parameters or observational properties of the galaxy, but depending on the source--lens configuration, a bias may still exist. Cross lenses provide an accurate estimate of the Hubble constant, while fold (double) lenses tend to be biased low (high). With kinematic constraints, double lenses show bias and intrinsic scatter of 6% and 10%, respectively, while quad lenses show bias and intrinsic scatter of 0.5% and 10%, respectively. For lenses with a reduced $\chi^2 > 1$, a power-law dependence of the $\chi^2$ on the lens environment (number of nearby galaxies) is seen. Lastly, we model, in greater detail, the cases of two double lenses that are significantly biased. We are able to remove the bias, suggesting that the remaining biases could also be reduced by carefully taking into account additional sources of systematic uncertainty.
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics", "Astrophysics" ]
Title: Heisenberg Modules over Quantum 2-tori are metrized quantum vector bundles, Abstract: The modular Gromov-Hausdorff propinquity is a distance on classes of modules endowed with quantum metric information, in the form of a metric form of a connection and a left Hilbert module structure. This paper proves that the family of Heisenberg modules over quantum two tori, when endowed with their canonical connections, form a family of metrized quantum vector bundles, as a first step in proving that Heisenberg modules form a continuous family for the modular Gromov-Hausdorff propinquity.
[ 0, 0, 1, 0, 0, 0 ]
[ "Mathematics", "Physics" ]
Title: Toward Controlled Generation of Text, Abstract: Generic generation and manipulation of text is challenging and has limited success compared to recent deep generative modeling in visual domain. This paper aims at generating plausible natural language sentences, whose attributes are dynamically controlled by learning disentangled latent representations with designated semantics. We propose a new neural generative model which combines variational auto-encoders and holistic attribute discriminators for effective imposition of semantic structures. With differentiable approximation to discrete text samples, explicit constraints on independent attribute controls, and efficient collaborative learning of generator and discriminators, our model learns highly interpretable representations from even only word annotations, and produces realistic sentences with desired attributes. Quantitative evaluation validates the accuracy of sentence and attribute generation.
[ 1, 0, 0, 1, 0, 0 ]
[ "Computer Science", "Statistics" ]
Title: Human peripheral blur is optimal for object recognition, Abstract: Our eyes sample a disproportionately large amount of information at the centre of gaze with increasingly sparse sampling into the periphery. This sampling scheme is widely believed to be a wiring constraint whereby high resolution at the centre is achieved by sacrificing spatial acuity in the periphery. Here we propose that this sampling scheme may be optimal for object recognition because the relevant spatial content is dense near an object and sparse in the surrounding vicinity. We tested this hypothesis by training deep convolutional neural networks on full-resolution and foveated images. Our main finding is that networks trained on images with foveated sampling show better object classification compared to networks trained on full resolution images. Importantly, blurring images according to the human blur function yielded the best performance compared to images with shallower or steeper blurring. Taken together our results suggest that, peripheral blurring in our eyes may have evolved for optimal object recognition, rather than merely to satisfy wiring constraints.
[ 0, 0, 0, 0, 1, 0 ]
[ "Computer Science", "Quantitative Biology" ]
Title: Acceleration of Mean Square Distance Calculations with Floating Close Structure in Metadynamics Simulations, Abstract: Molecular dynamics simulates the~movements of atoms. Due to its high cost, many methods have been developed to "push the~simulation forward". One of them, metadynamics, can hasten the~molecular dynamics with the~help of variables describing the~simulated process. However, the~evaluation of these variables can include numerous mean square distance calculations that introduce substantial computational demands, thus jeopardize the~benefit of the~approach. Recently, we proposed an~approximative method that significantly reduces the~number of these distance calculations. Here we evaluate the~performance and the~scalability on two molecular systems. We assess the~maximal theoretical speed-up based on the reduction of distance computations and Ahmdal's law and compare it to the~practical speed-up achieved with our implementation.
[ 1, 0, 0, 0, 0, 0 ]
[ "Physics", "Computer Science" ]
Title: Neutronic Analysis on Potential Accident Tolerant Fuel-Cladding Combination U$_3$Si$_2$-FeCrAl, Abstract: Neutronic performance is investigated for a potential accident tolerant fuel (ATF),which consists of U$_3$Si$_2$ fuel and FeCrAl cladding. In comparison with current UO$_2$-Zr system, FeCrAl has a better oxidation resistance but a larger thermal neutron absorption cross section. U$_3$Si$_2$ has a higher thermal conductivity and a higher uranium density, which can compensate the reactivity suppressed by FeCrAl. Based on neutronic investigations, a possible U$_3$Si$_2$-FeCrAl fuel-cladding systemis taken into consideration. Fundamental properties of the suggested fuel-cladding combination are investigated in a fuel assembly.These properties include moderator and fuel temperature coefficients, control rods worth, radial power distribution (in a fuel rod), and different void reactivity coefficients. The present work proves that the new combination has less reactivity variation during its service lifetime. Although, compared with the current system, it has a little larger deviation on power distribution and a little less negative temperature coefficient and void reactivity coefficient and its control rods worth is less important, variations of these parameters are less important during the service lifetime of fuel. Hence, U$_3$Si$_2$-FeCrAl system is a potential ATF candidate from a neutronic view.
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics" ]
Title: A Macdonald refined topological vertex, Abstract: We consider the refined topological vertex of Iqbal et al, as a function of two parameters (x, y), and deform it by introducing Macdonald parameters (q, t), as in the work of Vuletic on plane partitions, to obtain 'a Macdonald refined topological vertex'. In the limit q -> t, we recover the refined topological vertex of Iqbal et al. In the limit x -> y, we obtain a qt-deformation of the topological vertex of Aganagic et al. Copies of the vertex can be glued to obtain qt-deformed 5D instanton partition functions that have well-defined 4D limits and, for generic values of (q, t), contain infinite-towers of poles for every pole in the limit q -> t.
[ 0, 0, 1, 0, 0, 0 ]
[ "Mathematics", "Physics" ]
Title: Estimating Achievable Range of Ground Robots Operating on Single Battery Discharge for Operational Efficacy Amelioration, Abstract: Mobile robots are increasingly being used to assist with active pursuit and law enforcement. One major limitation for such missions is the resource (battery) allocated to the robot. Factors like nature and agility of evader, terrain over which pursuit is being carried out, plausible traversal velocity and the amount of necessary data to be collected all influence how long the robot can last in the field and how far it can travel. In this paper, we develop an analytical model that analyzes the energy utilization for a variety of components mounted on a robot to estimate the maximum operational range achievable by the robot operating on a single battery discharge. We categorize the major consumers of energy as: 1.) ancillary robotic functions such as computation, communication, sensing etc., and 2.) maneuvering which involves propulsion, steering etc. Both these consumers draw power from the common power source but the achievable range is largely affected by the proportion of power available for maneuvering. For this case study, we performed experiments with real robots on planar and graded surfaces and evaluated the estimation error for each case.
[ 1, 0, 0, 0, 0, 0 ]
[ "Computer Science", "Mathematics" ]
Title: Likelihood ratio test for variance components in nonlinear mixed effects models, Abstract: Mixed effects models are widely used to describe heterogeneity in a population. A crucial issue when adjusting such a model to data consists in identifying fixed and random effects. From a statistical point of view, it remains to test the nullity of the variances of a given subset of random effects. Some authors have proposed to use the likelihood ratio test and have established its asymptotic distribution in some particular cases. Nevertheless, to the best of our knowledge, no general variance components testing procedure has been fully investigated yet. In this paper, we study the likelihood ratio test properties to test that the variances of a general subset of the random effects are equal to zero in both linear and nonlinear mixed effects model, extending the existing results. We prove that the asymptotic distribution of the test is a chi-bar-square distribution, that is to say a mixture of chi-square distributions, and we identify the corresponding weights. We highlight in particular that the limiting distribution depends on the presence of correlations between the random effects but not on the linear or nonlinear structure of the mixed effects model. We illustrate the finite sample size properties of the test procedure through simulation studies and apply the test procedure to two real datasets of dental growth and of coucal growth.
[ 0, 0, 0, 1, 0, 0 ]
[ "Statistics", "Mathematics" ]
Title: Solving $\ell^p\!$-norm regularization with tensor kernels, Abstract: In this paper, we discuss how a suitable family of tensor kernels can be used to efficiently solve nonparametric extensions of $\ell^p$ regularized learning methods. Our main contribution is proposing a fast dual algorithm, and showing that it allows to solve the problem efficiently. Our results contrast recent findings suggesting kernel methods cannot be extended beyond Hilbert setting. Numerical experiments confirm the effectiveness of the method.
[ 0, 0, 1, 1, 0, 0 ]
[ "Computer Science", "Mathematics" ]
Title: Complementary views on electron spectra: From Fluctuation Diagnostics to real space correlations, Abstract: We study the relation between the microscopic properties of a many-body system and the electron spectra, experimentally accessible by photoemission. In a recent paper [Phys. Rev. Lett. 114, 236402 (2015)], we introduced the "fluctuation diagnostics" approach, to extract the dominant wave vector dependent bosonic fluctuations from the electronic self-energy. Here, we first reformulate the theory in terms of fermionic modes, to render its connection with resonance valence bond (RVB) fluctuations more transparent. Secondly, by using a large-U expansion, where U is the Coulomb interaction, we relate the fluctuations to real space correlations. Therefore, it becomes possible to study how electron spectra are related to charge, spin, superconductivity and RVB-like real space correlations, broadening the analysis of an earlier work [Phys. Rev. B 89, 245130 (2014)]. This formalism is applied to the pseudogap physics of the two-dimensional Hubbard model, studied in the dynamical cluster approximation. We perform calculations for embedded clusters with up to 32 sites, having three inequivalent K-points at the Fermi surface. We find that as U is increased, correlation functions gradually attain values consistent with an RVB state. This first happens for correlation functions involving the antinodal point and gradually spreads to the nodal point along the Fermi surface. Simultaneously a pseudogap opens up along the Fermi surface. We relate this to a crossover from a Kondo-like state to an RVB-like localized cluster state and to the presence of RVB and spin fluctuations. These changes are caused by a strong momentum dependence in the cluster bath-couplings along the Fermi surface. We also show, from a more algorithmic perspective, how the time-consuming calculations in fluctuation diagnostics can be drastically simplified.
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics" ]
Title: Extracting Automata from Recurrent Neural Networks Using Queries and Counterexamples, Abstract: We present a novel algorithm that uses exact learning and abstraction to extract a deterministic finite automaton describing the state dynamics of a given trained RNN. We do this using Angluin's L* algorithm as a learner and the trained RNN as an oracle. Our technique efficiently extracts accurate automata from trained RNNs, even when the state vectors are large and require fine differentiation.
[ 1, 0, 0, 0, 0, 0 ]
[ "Computer Science", "Mathematics" ]
Title: GBDT of discrete skew-selfadjoint Dirac systems and explicit solutions of the corresponding non-stationary problems, Abstract: Generalized Bäcklund-Darboux transformations (GBDTs) of discrete skew-selfadjoint Dirac systems have been successfully used for explicit solving of direct and inverse problems of Weyl-Titchmarsh theory. During explicit solving of the direct and inverse problems, we considered GBDTs of the trivial initial systems. However, GBDTs of arbitrary discrete skew-selfadjoint Dirac systems are important as well and we introduce these transformations in the present paper. The obtained results are applied to the construction of explicit solutions of the interesting related non-stationary systems.
[ 0, 0, 1, 0, 0, 0 ]
[ "Mathematics", "Physics" ]
Title: Smooth Neighbors on Teacher Graphs for Semi-supervised Learning, Abstract: The recently proposed self-ensembling methods have achieved promising results in deep semi-supervised learning, which penalize inconsistent predictions of unlabeled data under different perturbations. However, they only consider adding perturbations to each single data point, while ignoring the connections between data samples. In this paper, we propose a novel method, called Smooth Neighbors on Teacher Graphs (SNTG). In SNTG, a graph is constructed based on the predictions of the teacher model, i.e., the implicit self-ensemble of models. Then the graph serves as a similarity measure with respect to which the representations of "similar" neighboring points are learned to be smooth on the low-dimensional manifold. We achieve state-of-the-art results on semi-supervised learning benchmarks. The error rates are 9.89%, 3.99% for CIFAR-10 with 4000 labels, SVHN with 500 labels, respectively. In particular, the improvements are significant when the labels are fewer. For the non-augmented MNIST with only 20 labels, the error rate is reduced from previous 4.81% to 1.36%. Our method also shows robustness to noisy labels.
[ 1, 0, 0, 1, 0, 0 ]
[ "Computer Science", "Statistics" ]
Title: Achieving Spectrum Efficient Communication Under Cross-Technology Interference, Abstract: In wireless communication, heterogeneous technologies such as WiFi, ZigBee and BlueTooth operate in the same ISM band.With the exponential growth in the number of wireless devices, the ISM band becomes more and more crowded. These heterogeneous devices have to compete with each other to access spectrum resources, generating cross-technology interference (CTI). Since CTI may destroy wireless communication, this field is facing an urgent and challenging need to investigate spectrum efficiency under CTI. In this paper, we introduce a novel framework to address this problem from two aspects. On the one hand, from the perspective of each communication technology itself, we propose novel channel/link models to capture the channel/link status under CTI. On the other hand, we investigate spectrum efficiency from the perspective by taking all heterogeneous technologies as a whole and building crosstechnology communication among them. The capability of direct communication among heterogeneous devices brings great opportunities to harmoniously sharing the spectrum with collaboration rather than competition.
[ 1, 0, 0, 0, 0, 0 ]
[ "Computer Science" ]
Title: A Galactic Cosmic Ray Electron Intensity Increase of a factor of up to 100 At Energies between 3 and 50 MeV in the Heliosheath between the Termination Shock and the Heliopause Due to Solar Modulation As Measured by Voyager 1, Abstract: We have derived background corrected intensities of 3-50 MeV galactic electrons observed by Voyager 1 as it passes through the heliosheath from 95 to 122 AU. The overall intensity change of the background corrected data from the inner to the outer boundary of the heliosheath is a maximum of a factor ~100 at 15 MeV. At lower energies this fractional change becomes less and the corrected electron spectra in the heliosheath becomes progressively steeper, reaching values ~ -2.5 for the spectral index just outside of the termination shock. At higher energies the spectra of electrons has an exponent changing from the negative LIS spectral index of -1.3 to values approaching zero in the heliosheath as a result of the solar modulation of the galactic electron component. The large modulation effects observed below ~100 MV are possible evidence for enhanced diffusion as part of the modulation process for electrons in the heliosheath.
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics" ]
Title: On self-affine sets, Abstract: We survey the dimension theory of self-affine sets for general mathematical audience. The article is in Finnish.
[ 0, 0, 1, 0, 0, 0 ]
[ "Mathematics" ]
Title: Multiplicative slices, relativistic Toda and shifted quantum affine algebras, Abstract: We introduce the shifted quantum affine algebras. They map homomorphically into the quantized $K$-theoretic Coulomb branches of $3d\ {\mathcal N}=4$ SUSY quiver gauge theories. In type $A$, they are endowed with a coproduct, and they act on the equivariant $K$-theory of parabolic Laumon spaces. In type $A_1$, they are closely related to the open relativistic quantum Toda lattice of type $A$.
[ 0, 0, 1, 0, 0, 0 ]
[ "Mathematics", "Physics" ]
Title: High temperature thermodynamics of the honeycomb-lattice Kitaev-Heisenberg model: A high temperature series expansion study, Abstract: We develop high temperature series expansions for the thermodynamic properties of the honeycomb-lattice Kitaev-Heisenberg model. Numerical results for uniform susceptibility, heat capacity and entropy as a function of temperature for different values of the Kitaev coupling $K$ and Heisenberg exachange coupling $J$ (with $|J|\le |K|$) are presented. These expansions show good convergence down to a temperature of a fraction of $K$ and in some cases down to $T=K/10$. In the Kitaev exchange dominated regime, the inverse susceptibility has a nearly linear temperature dependence over a wide temperature range. However, we show that already at temperatures $10$-times the Curie-Weiss temperature, the effective Curie-Weiss constant estimated from the data can be off by a factor of 2. We find that the magnitude of the heat capacity maximum at the short-range order peak, is substantially smaller for small $J/K$ than for $J$ of order or larger than $K$. We suggest that this itself represents a simple marker for the relative importance of the Kitaev terms in these systems. Somewhat surprisingly, both heat capacity and susceptibility data on Na$_2$IrO$_3$ are consistent with a dominant {\it antiferromagnetic} Kitaev exchange constant of about $300-400$ $K$.
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics" ]
Title: Inference in Sparse Graphs with Pairwise Measurements and Side Information, Abstract: We consider the statistical problem of recovering a hidden "ground truth" binary labeling for the vertices of a graph up to low Hamming error from noisy edge and vertex measurements. We present new algorithms and a sharp finite-sample analysis for this problem on trees and sparse graphs with poor expansion properties such as hypergrids and ring lattices. Our method generalizes and improves over that of Globerson et al. (2015), who introduced the problem for two-dimensional grid lattices. For trees we provide a simple, efficient, algorithm that infers the ground truth with optimal Hamming error has optimal sample complexity and implies recovery results for all connected graphs. Here, the presence of side information is critical to obtain a non-trivial recovery rate. We then show how to adapt this algorithm to tree decompositions of edge-subgraphs of certain graph families such as lattices, resulting in optimal recovery error rates that can be obtained efficiently The thrust of our analysis is to 1) use the tree decomposition along with edge measurements to produce a small class of viable vertex labelings and 2) apply an analysis influenced by statistical learning theory to show that we can infer the ground truth from this class using vertex measurements. We show the power of our method in several examples including hypergrids, ring lattices, and the Newman-Watts model for small world graphs. For two-dimensional grids, our results improve over Globerson et al. (2015) by obtaining optimal recovery in the constant-height regime.
[ 1, 0, 0, 0, 0, 0 ]
[ "Computer Science", "Statistics", "Mathematics" ]
Title: Guiding Reinforcement Learning Exploration Using Natural Language, Abstract: In this work we present a technique to use natural language to help reinforcement learning generalize to unseen environments. This technique uses neural machine translation, specifically the use of encoder-decoder networks, to learn associations between natural language behavior descriptions and state-action information. We then use this learned model to guide agent exploration using a modified version of policy shaping to make it more effective at learning in unseen environments. We evaluate this technique using the popular arcade game, Frogger, under ideal and non-ideal conditions. This evaluation shows that our modified policy shaping algorithm improves over a Q-learning agent as well as a baseline version of policy shaping.
[ 1, 0, 0, 1, 0, 0 ]
[ "Computer Science", "Statistics" ]
Title: Cell Coverage Extension with Orthogonal Random Precoding for Massive MIMO Systems, Abstract: In this paper, we investigate a coverage extension scheme based on orthogonal random precoding (ORP) for the downlink of massive multiple-input multiple-output (MIMO) systems. In this scheme, a precoding matrix consisting of orthogonal vectors is employed at the transmitter to enhance the maximum signal-to-interference-plus-noise ratio (SINR) of the user. To analyze and optimize the ORP scheme in terms of cell coverage, we derive the analytical expressions of the downlink coverage probability for two receiver structures, namely, the single-antenna (SA) receiver and multiple-antenna receiver with antenna selection (AS). The simulation results show that the analytical expressions accurately capture the coverage behaviors of the systems employing the ORP scheme. It is also shown that the optimal coverage performance is achieved when a single precoding vector is used under the condition that the threshold of the signal-to-noise ratio of the coverage is greater than one. The performance of the ORP scheme is further analyzed when different random precoder groups are utilized over multiple time slots to exploit precoding diversity. The numerical results show that the proposed ORP scheme over multiple time slots provides a substantial coverage gain over the space-time coding scheme despite its low feedback overhead.
[ 1, 0, 1, 0, 0, 0 ]
[ "Computer Science", "Physics" ]
Title: Quasiparticles and charge transfer at the two surfaces of the honeycomb iridate Na$_2$IrO$_3$, Abstract: Direct experimental investigations of the low-energy electronic structure of the Na$_2$IrO$_3$ iridate insulator are sparse and draw two conflicting pictures. One relies on flat bands and a clear gap, the other involves dispersive states approaching the Fermi level, pointing to surface metallicity. Here, by a combination of angle-resolved photoemission, photoemission electron microscopy, and x-ray absorption, we show that the correct picture is more complex and involves an anomalous band, arising from charge transfer from Na atoms to Ir-derived states. Bulk quasiparticles do exist, but in one of the two possible surface terminations the charge transfer is smaller and they remain elusive.
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics" ]
Title: Breaking Bivariate Records, Abstract: We establish a fundamental property of bivariate Pareto records for independent observations uniformly distributed in the unit square. We prove that the asymptotic conditional distribution of the number of records broken by an observation given that the observation sets a record is Geometric with parameter 1/2.
[ 0, 0, 1, 1, 0, 0 ]
[ "Mathematics", "Statistics" ]
Title: A Bag-of-Paths Node Criticality Measure, Abstract: This work compares several node (and network) criticality measures quantifying to which extend each node is critical with respect to the communication flow between nodes of the network, and introduces a new measure based on the Bag-of-Paths (BoP) framework. Network disconnection simulation experiments show that the new BoP measure outperforms all the other measures on a sample of Erdos-Renyi and Albert-Barabasi graphs. Furthermore, a faster (still O(n^3)), approximate, BoP criticality relying on the Sherman-Morrison rank-one update of a matrix is introduced for tackling larger networks. This approximate measure shows similar performances as the original, exact, one.
[ 1, 1, 0, 0, 0, 0 ]
[ "Computer Science", "Mathematics" ]
Title: Selection of quasi-stationary states in the Navier-Stokes equation on the torus, Abstract: The two dimensional incompressible Navier-Stokes equation on $D_\delta := [0, 2\pi\delta] \times [0, 2\pi]$ with $\delta \approx 1$, periodic boundary conditions, and viscosity $0 < \nu \ll 1$ is considered. Bars and dipoles, two explicitly given quasi-stationary states of the system, evolve on the time scale $\mathcal{O}(e^{-\nu t})$ and have been shown to play a key role in its long-time evolution. Of particular interest is the role that $\delta$ plays in selecting which of these two states is observed. Recent numerical studies suggest that, after a transient period of rapid decay of the high Fourier modes, the bar state will be selected if $\delta \neq 1$, while the dipole will be selected if $\delta = 1$. Our results support this claim and seek to mathematically formalize it. We consider the system in Fourier space, project it onto a center manifold consisting of the lowest eight Fourier modes, and use this as a model to study the selection of bars and dipoles. It is shown for this ODE model that the value of $\delta$ controls the behavior of the asymptotic ratio of the low modes, thus determining the likelihood of observing a bar state or dipole after an initial transient period. Moreover, in our model, for all $\delta \approx 1$, there is an initial time period in which the high modes decay at the rapid rate $\mathcal{O}(e^{-t/\nu})$, while the low modes evolve at the slower $\mathcal{O}(e^{-\nu t})$ rate. The results for the ODE model are proven using energy estimates and invariant manifolds and further supported by formal asymptotic expansions and numerics.
[ 0, 0, 1, 0, 0, 0 ]
[ "Mathematics", "Physics" ]
Title: Geometric Enclosing Networks, Abstract: Training model to generate data has increasingly attracted research attention and become important in modern world applications. We propose in this paper a new geometry-based optimization approach to address this problem. Orthogonal to current state-of-the-art density-based approaches, most notably VAE and GAN, we present a fresh new idea that borrows the principle of minimal enclosing ball to train a generator G\left(\bz\right) in such a way that both training and generated data, after being mapped to the feature space, are enclosed in the same sphere. We develop theory to guarantee that the mapping is bijective so that its inverse from feature space to data space results in expressive nonlinear contours to describe the data manifold, hence ensuring data generated are also lying on the data manifold learned from training data. Our model enjoys a nice geometric interpretation, hence termed Geometric Enclosing Networks (GEN), and possesses some key advantages over its rivals, namely simple and easy-to-control optimization formulation, avoidance of mode collapsing and efficiently learn data manifold representation in a completely unsupervised manner. We conducted extensive experiments on synthesis and real-world datasets to illustrate the behaviors, strength and weakness of our proposed GEN, in particular its ability to handle multi-modal data and quality of generated data.
[ 1, 0, 0, 1, 0, 0 ]
[ "Computer Science", "Mathematics" ]
Title: Out-degree reducing partitions of digraphs, Abstract: Let $k$ be a fixed integer. We determine the complexity of finding a $p$-partition $(V_1, \dots, V_p)$ of the vertex set of a given digraph such that the maximum out-degree of each of the digraphs induced by $V_i$, ($1\leq i\leq p$) is at least $k$ smaller than the maximum out-degree of $D$. We show that this problem is polynomial-time solvable when $p\geq 2k$ and ${\cal NP}$-complete otherwise. The result for $k=1$ and $p=2$ answers a question posed in \cite{bangTCS636}. We also determine, for all fixed non-negative integers $k_1,k_2,p$, the complexity of deciding whether a given digraph of maximum out-degree $p$ has a $2$-partition $(V_1,V_2)$ such that the digraph induced by $V_i$ has maximum out-degree at most $k_i$ for $i\in [2]$. It follows from this characterization that the problem of deciding whether a digraph has a 2-partition $(V_1,V_2)$ such that each vertex $v\in V_i$ has at least as many neighbours in the set $V_{3-i}$ as in $V_i$, for $i=1,2$ is ${\cal NP}$-complete. This solves a problem from \cite{kreutzerEJC24} on majority colourings.
[ 1, 0, 0, 0, 0, 0 ]
[ "Computer Science", "Mathematics" ]
Title: Presymplectic convexity and (ir)rational polytopes, Abstract: In this paper, we extend the Atiyah--Guillemin--Sternberg convexity theorem and Delzant's classification of symplectic toric manifolds to presymplectic manifolds. We also define and study the Morita equivalence of presymplectic toric manifolds and of their corresponding framed momentum polytopes, which may be rational or non-rational. Toric orbifolds, quasifolds and non-commutative toric varieties may be viewed as the quotient of our presymplectic toric manifolds by the kernel isotropy foliation of the presymplectic form.
[ 0, 0, 1, 0, 0, 0 ]
[ "Mathematics", "Physics" ]
Title: Unsupervised Learning of Mixture Regression Models for Longitudinal Data, Abstract: This paper is concerned with learning of mixture regression models for individuals that are measured repeatedly. The adjective "unsupervised" implies that the number of mixing components is unknown and has to be determined, ideally by data driven tools. For this purpose, a novel penalized method is proposed to simultaneously select the number of mixing components and to estimate the mixing proportions and unknown parameters in the models. The proposed method is capable of handling both continuous and discrete responses by only requiring the first two moment conditions of the model distribution. It is shown to be consistent in both selecting the number of components and estimating the mixing proportions and unknown regression parameters. Further, a modified EM algorithm is developed to seamlessly integrate model selection and estimation. Simulation studies are conducted to evaluate the finite sample performance of the proposed procedure. And it is further illustrated via an analysis of a primary biliary cirrhosis data set.
[ 0, 0, 0, 1, 0, 0 ]
[ "Statistics", "Mathematics" ]
Title: Anomalous electron states, Abstract: By the certain macroscopic perturbations in condensed matter anomalous electron wells can be formed due to a local reduction of electromagnetic zero point energy. These wells are narrow, of the width $\sim 10^{-11}cm$, and with the depth $\sim 1MeV$. Such anomalous states, from the formal standpoint of quantum mechanics, correspond to a singular solution of a wave equation produced by the non-physical $\delta(\vec R)$ source. The resolution, on the level of the Standard Model, of the tiny region around the formal singularity shows that the state is physical. The creation of those states in an atomic system is of the formal probability $\exp(-1000)$. The probability becomes not small under a perturbation which rapidly varies in space, on the scale $10^{-11}cm$. In condensed matter such perturbation may relate to acoustic shock waves. In this process the short scale is the length of the standing de Broglie wave of a reflected lattice atom. Under electron transitions in the anomalous well (anomalous atom) $keV$ X-rays are expected to be emitted. A macroscopic amount of anomalous atoms, of the size $10^{-11}cm$ each, can be formed in a solid resulting in ${\it collapsed}$ ${\it matter}$ with $10^9$ times enhanced density.
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics" ]
Title: Temporal processing and context dependency in C. elegans mechanosensation, Abstract: A quantitative understanding of how sensory signals are transformed into motor outputs places useful constraints on brain function and helps reveal the brain's underlying computations. We investigate how the nematode C. elegans responds to time-varying mechanosensory signals using a high-throughput optogenetic assay and automated behavior quantification. In the prevailing picture of the touch circuit, the animal's behavior is determined by which neurons are stimulated and by the stimulus amplitude. In contrast, we find that the behavioral response is tuned to temporal properties of mechanosensory signals, like its integral and derivative, that extend over many seconds. Mechanosensory signals, even in the same neurons, can be tailored to elicit different behavioral responses. Moreover, we find that the animal's response also depends on its behavioral context. Most dramatically, the animal ignores all tested mechanosensory stimuli during turns. Finally, we present a linear-nonlinear model that predicts the animal's behavioral response to stimulus.
[ 0, 0, 0, 0, 1, 0 ]
[ "Quantitative Biology" ]
Title: On the putative essential discreteness of q-generalized entropies, Abstract: It has been argued in [EPL {\bf 90} (2010) 50004], entitled {\it Essential discreteness in generalized thermostatistics with non-logarithmic entropy}, that "continuous Hamiltonian systems with long-range interactions and the so-called q-Gaussian momentum distributions are seen to be outside the scope of non-extensive statistical mechanics". The arguments are clever and appealing. We show here that, however, some mathematical subtleties render them unconvincing
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics", "Mathematics" ]
Title: Emergent electronic structure of CaFe2As2, Abstract: CaFe2As2 exhibits collapsed tetragonal (cT) structure and varied exotic behavior under pressure at low temperatures that led to debate on linking the structural changes to its exceptional electronic properties like superconductivity, magnetism, etc. Here, we investigate the electronic structure of CaFe2As2 forming in different structures employing density functional theory. The results indicate better stability of the cT phase with enhancement in hybridization induced effects and shift of the energy bands towards lower energies. The Fermi surface centered around $\Gamma$ point gradually vanishes with the increase in pressure. Consequently, the nesting between the hole and electron Fermi surfaces associated to the spin density wave state disappears indicating a pathway to achieve the proximity to quantum fluctuations. The magnetic moment at the Fe sites diminishes in the cT phase consistent with the magnetic susceptibility results. Notably, the hybridization of Ca 4s states (Ca-layer may be treated as a charge reservoir layer akin to those in cuprate superconductors) is significantly enhanced in the cT phase revealing its relevance in its interesting electronic properties.
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics" ]
Title: Smart "Predict, then Optimize", Abstract: Many real-world analytics problems involve two significant challenges: prediction and optimization. Due to the typically complex nature of each challenge, the standard paradigm is to predict, then optimize. By and large, machine learning tools are intended to minimize prediction error and do not account for how the predictions will be used in a downstream optimization problem. In contrast, we propose a new and very general framework, called Smart "Predict, then Optimize" (SPO), which directly leverages the optimization problem structure, i.e., its objective and constraints, for designing successful analytics tools. A key component of our framework is the SPO loss function, which measures the quality of a prediction by comparing the objective values of the solutions generated using the predicted and observed parameters, respectively. Training a model with respect to the SPO loss is computationally challenging, and therefore we also develop a surrogate loss function, called the SPO+ loss, which upper bounds the SPO loss, has desirable convexity properties, and is statistically consistent under mild conditions. We also propose a stochastic gradient descent algorithm which allows for situations in which the number of training samples is large, model regularization is desired, and/or the optimization problem of interest is nonlinear or integer. Finally, we perform computational experiments to empirically verify the success of our SPO framework in comparison to the standard predict-then-optimize approach.
[ 1, 0, 0, 1, 0, 0 ]
[ "Computer Science", "Mathematics", "Statistics" ]
Title: Density estimation on small datasets, Abstract: How might a smooth probability distribution be estimated, with accurately quantified uncertainty, from a limited amount of sampled data? Here we describe a field-theoretic approach that addresses this problem remarkably well in one dimension, providing an exact nonparametric Bayesian posterior without relying on tunable parameters or large-data approximations. Strong non-Gaussian constraints, which require a non-perturbative treatment, are found to play a major role in reducing distribution uncertainty. A software implementation of this method is provided.
[ 1, 0, 0, 0, 1, 0 ]
[ "Statistics", "Mathematics" ]
Title: Security Trust Zone in 5G Networks, Abstract: Fifth Generation (5G) telecommunication system is going to deliver a flexible radio access network (RAN). Security functions such as authorization, authentication and accounting (AAA) are expected to be distributed from central clouds to edge clouds. We propose a novel architectural security solution that applies to 5G networks. It is called Trust Zone (TZ) that is designed as an enhancement of the 5G AAA in the edge cloud. TZ also provides an autonomous and decentralized security policy for different tenants under variable network conditions. TZ also initiates an ability of disaster cognition and extends the security functionalities to a set of flexible and highly available emergency services in the edge cloud.
[ 1, 0, 0, 0, 0, 0 ]
[ "Computer Science" ]
Title: Few-shot learning of neural networks from scratch by pseudo example optimization, Abstract: In this paper, we propose a simple but effective method for training neural networks with a limited amount of training data. Our approach inherits the idea of knowledge distillation that transfers knowledge from a deep or wide reference model to a shallow or narrow target model. The proposed method employs this idea to mimic predictions of reference estimators that are more robust against overfitting than the network we want to train. Different from almost all the previous work for knowledge distillation that requires a large amount of labeled training data, the proposed method requires only a small amount of training data. Instead, we introduce pseudo training examples that are optimized as a part of model parameters. Experimental results for several benchmark datasets demonstrate that the proposed method outperformed all the other baselines, such as naive training of the target model and standard knowledge distillation.
[ 0, 0, 0, 1, 0, 0 ]
[ "Computer Science", "Statistics" ]
Title: Identities and congruences involving the Fubini polynomials, Abstract: In this paper, we investigate the umbral representation of the Fubini polynomials $F_{x}^{n}:=F_{n}(x)$ to derive some properties involving these polynomials. For any prime number $p$ and any polynomial $f$ with integer coefficients, we show $(f(F_{x}))^{p}\equiv f(F_{x})$ and we give other curious congruences.
[ 0, 0, 1, 0, 0, 0 ]
[ "Mathematics" ]
Title: High-sensitivity Kinetic Inductance Detectors for CALDER, Abstract: Providing a background discrimination tool is crucial for enhancing the sensitivity of next-generation experiments searching for neutrinoless double- beta decay. The development of high-sensitivity (< 20 eV RMS) cryogenic light detectors allows simultaneous read-out of the light and heat signals and enables background suppression through particle identification. The Cryogenic wide- Area Light Detector with Excellent Resolution (CALDER) R&D already proved the potential of this technique using the phonon-mediated Kinetic Inductance Detectors (KIDs) approach. The first array prototype with 4 Aluminum KIDs on a 2 $\times$ 2 cm2 Silicon substrate showed a baseline resolution of 154 $\pm$ 7 eV RMS. Improving the design and the readout of the resonator, the next CALDER prototype featured an energy resolution of 82 $\pm$ 4 eV, by sampling the same substrate with a single Aluminum KID.
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics" ]
Title: An Exploration of Approaches to Integrating Neural Reranking Models in Multi-Stage Ranking Architectures, Abstract: We explore different approaches to integrating a simple convolutional neural network (CNN) with the Lucene search engine in a multi-stage ranking architecture. Our models are trained using the PyTorch deep learning toolkit, which is implemented in C/C++ with a Python frontend. One obvious integration strategy is to expose the neural network directly as a service. For this, we use Apache Thrift, a software framework for building scalable cross-language services. In exploring alternative architectures, we observe that once trained, the feedforward evaluation of neural networks is quite straightforward. Therefore, we can extract the parameters of a trained CNN from PyTorch and import the model into Java, taking advantage of the Java Deeplearning4J library for feedforward evaluation. This has the advantage that the entire end-to-end system can be implemented in Java. As a third approach, we can extract the neural network from PyTorch and "compile" it into a C++ program that exposes a Thrift service. We evaluate these alternatives in terms of performance (latency and throughput) as well as ease of integration. Experiments show that feedforward evaluation of the convolutional neural network is significantly slower in Java, while the performance of the compiled C++ network does not consistently beat the PyTorch implementation.
[ 1, 0, 0, 0, 0, 0 ]
[ "Computer Science" ]
Title: Monotonicity patterns and functional inequalities for classical and generalized Wright functions, Abstract: In this paper our aim is to present the completely monotonicity and convexity properties for the Wright function. As consequences of these results, we present some functional inequalities. Moreover, we derive the monotonicity and log-convexity results for the generalized Wright functions. As applications, we present several new inequalities (like Turán type inequalities) and we prove some geometric properties for four--parametric Mittag--Leffler functions.
[ 0, 0, 1, 0, 0, 0 ]
[ "Mathematics" ]
Title: Galaxy And Mass Assembly: the evolution of the cosmic spectral energy distribution from z = 1 to z = 0, Abstract: We present the evolution of the Cosmic Spectral Energy Distribution (CSED) from $z = 1 - 0$. Our CSEDs originate from stacking individual spectral energy distribution fits based on panchromatic photometry from the Galaxy and Mass Assembly (GAMA) and COSMOS datasets in ten redshift intervals with completeness corrections applied. Below $z = 0.45$, we have credible SED fits from 100 nm to 1 mm. Due to the relatively low sensitivity of the far-infrared data, our far-infrared CSEDs contain a mix of predicted and measured fluxes above $z = 0.45$. Our results include appropriate errors to highlight the impact of these corrections. We show that the bolometric energy output of the Universe has declined by a factor of roughly four -- from $5.1 \pm 1.0$ at $z \sim 1$ to $1.3 \pm 0.3 \times 10^{35}~h_{70}$~W~Mpc$^{-3}$ at the current epoch. We show that this decrease is robust to cosmic variance, SED modelling and other various types of error. Our CSEDs are also consistent with an increase in the mean age of stellar populations. We also show that dust attenuation has decreased over the same period, with the photon escape fraction at 150~nm increasing from $16 \pm 3$ at $z \sim 1$ to $24 \pm 5$ per cent at the current epoch, equivalent to a decrease in $A_\mathrm{FUV}$ of 0.4~mag. Our CSEDs account for $68 \pm 12$ and $61 \pm 13$ per cent of the cosmic optical and infrared backgrounds respectively as defined from integrated galaxy counts and are consistent with previous estimates of the cosmic infrared background with redshift.
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics" ]
Title: Asymptotic formula of the number of Newton polygons, Abstract: In this paper, we enumerate Newton polygons asymptotically. The number of Newton polygons is computable by a simple recurrence equation, but unexpectedly the asymptotic formula of its logarithm contains growing oscillatory terms. As the terms come from non-trivial zeros of the Riemann zeta function, an estimation of the amplitude of the oscillating part is equivalent to the Riemann hypothesis.
[ 0, 0, 1, 0, 0, 0 ]
[ "Mathematics" ]
Title: Stochastic Variance Reduction Methods for Policy Evaluation, Abstract: Policy evaluation is a crucial step in many reinforcement-learning procedures, which estimates a value function that predicts states' long-term value under a given policy. In this paper, we focus on policy evaluation with linear function approximation over a fixed dataset. We first transform the empirical policy evaluation problem into a (quadratic) convex-concave saddle point problem, and then present a primal-dual batch gradient method, as well as two stochastic variance reduction methods for solving the problem. These algorithms scale linearly in both sample size and feature dimension. Moreover, they achieve linear convergence even when the saddle-point problem has only strong concavity in the dual variables but no strong convexity in the primal variables. Numerical experiments on benchmark problems demonstrate the effectiveness of our methods.
[ 1, 0, 1, 1, 0, 0 ]
[ "Computer Science", "Statistics", "Mathematics" ]
Title: Measuring the polarization of electromagnetic fields using Rabi-rate measurements with spatial resolution: experiment and theory, Abstract: When internal states of atoms are manipulated using coherent optical or radio-frequency (RF) radiation, it is essential to know the polarization of the radiation with respect to the quantization axis of the atom. We first present a measurement of the two-dimensional spatial distribution of the electric-field amplitude of a linearly-polarized pulsed RF electric field at $\sim 25.6\,$GHz and its angle with respect to a static electric field. The measurements exploit coherent population transfer between the $35$s and $35$p Rydberg states of helium atoms in a pulsed supersonic beam. Based on this experimental result, we develop a general framework in the form of a set of equations relating the five independent polarization parameters of a coherently oscillating field in a fixed laboratory frame to Rabi rates of transitions between a ground and three excited states of an atom with arbitrary quantization axis. We then explain how these equations can be used to fully characterize the polarization in a minimum of five Rabi rate measurements by rotation of an external bias-field, or, knowing the polarization of the driving field, to determine the orientation of the static field using two measurements. The presented technique is not limited to Rydberg atoms and RF fields but can also be applied to characterize optical fields. The technique has potential for sensing the spatiotemporal properties of electromagnetic fields, e.g., in metrology devices or in hybrid experiments involving atoms close to surfaces.
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics" ]
Title: Software-based Microarchitectural Attacks, Abstract: Modern processors are highly optimized systems where every single cycle of computation time matters. Many optimizations depend on the data that is being processed. Software-based microarchitectural attacks exploit effects of these optimizations. Microarchitectural side-channel attacks leak secrets from cryptographic computations, from general purpose computations, or from the kernel. This leakage even persists across all common isolation boundaries, such as processes, containers, and virtual machines. Microarchitectural fault attacks exploit the physical imperfections of modern computer systems. Shrinking process technology introduces effects between isolated hardware elements that can be exploited by attackers to take control of the entire system. These attacks are especially interesting in scenarios where the attacker is unprivileged or even sandboxed. In this thesis, we focus on microarchitectural attacks and defenses on commodity systems. We investigate known and new side channels and show that microarchitectural attacks can be fully automated. Furthermore, we show that these attacks can be mounted in highly restricted environments such as sandboxed JavaScript code in websites. We show that microarchitectural attacks exist on any modern computer system, including mobile devices (e.g., smartphones), personal computers, and commercial cloud systems. This thesis consists of two parts. In the first part, we provide background on modern processor architectures and discuss state-of-the-art attacks and defenses in the area of microarchitectural side-channel attacks and microarchitectural fault attacks. In the second part, a selection of our papers are provided without modification from their original publications. I have co-authored these papers, which have subsequently been anonymously peer-reviewed, accepted, and presented at renowned international conferences.
[ 1, 0, 0, 0, 0, 0 ]
[ "Computer Science" ]
Title: Boundary problems for the fractional and tempered fractional operators, Abstract: For characterizing the Brownian motion in a bounded domain: $\Omega$, it is well-known that the boundary conditions of the classical diffusion equation just rely on the given information of the solution along the boundary of a domain; on the contrary, for the Lévy flights or tempered Lévy flights in a bounded domain, it involves the information of a solution in the complementary set of $\Omega$, i.e., $\mathbb{R}^n\backslash \Omega$, with the potential reason that paths of the corresponding stochastic process are discontinuous. Guided by probability intuitions and the stochastic perspectives of anomalous diffusion, we show the reasonable ways, ensuring the clear physical meaning and well-posedness of the partial differential equations (PDEs), of specifying `boundary' conditions for space fractional PDEs modeling the anomalous diffusion. Some properties of the operators are discussed, and the well-posednesses of the PDEs with generalized boundary conditions are proved.
[ 0, 0, 1, 0, 0, 0 ]
[ "Mathematics", "Physics" ]
Title: Bayesian Unification of Gradient and Bandit-based Learning for Accelerated Global Optimisation, Abstract: Bandit based optimisation has a remarkable advantage over gradient based approaches due to their global perspective, which eliminates the danger of getting stuck at local optima. However, for continuous optimisation problems or problems with a large number of actions, bandit based approaches can be hindered by slow learning. Gradient based approaches, on the other hand, navigate quickly in high-dimensional continuous spaces through local optimisation, following the gradient in fine grained steps. Yet, apart from being susceptible to local optima, these schemes are less suited for online learning due to their reliance on extensive trial-and-error before the optimum can be identified. In this paper, we propose a Bayesian approach that unifies the above two paradigms in one single framework, with the aim of combining their advantages. At the heart of our approach we find a stochastic linear approximation of the function to be optimised, where both the gradient and values of the function are explicitly captured. This allows us to learn from both noisy function and gradient observations, and predict these properties across the action space to support optimisation. We further propose an accompanying bandit driven exploration scheme that uses Bayesian credible bounds to trade off exploration against exploitation. Our empirical results demonstrate that by unifying bandit and gradient based learning, one obtains consistently improved performance across a wide spectrum of problem environments. Furthermore, even when gradient feedback is unavailable, the flexibility of our model, including gradient prediction, still allows us outperform competing approaches, although with a smaller margin. Due to the pervasiveness of bandit based optimisation, our scheme opens up for improved performance both in meta-optimisation and in applications where gradient related information is readily available.
[ 1, 0, 0, 0, 0, 0 ]
[ "Computer Science", "Statistics", "Mathematics" ]
Title: Detecting Bot Activity in the Ethereum Blockchain Network, Abstract: The Ethereum blockchain network is a decentralized platform enabling smart contract execution and transactions of Ether (ETH) [1], its designated cryptocurrency. Ethereum is the second most popular cryptocurrency with a market cap of more than 100 billion USD, with hundreds of thousands of transactions executed daily by hundreds of thousands of unique wallets. Tens of thousands of those wallets are newly generated each day. The Ethereum platform enables anyone to freely open multiple new wallets [2] free of charge (resulting in a large number of wallets that are controlled by the same entities). This attribute makes the Ethereum network a breeding space for activity by software robots (bots). The existence of bots is widespread in different digital technologies and there are various approaches to detect their activity such as rule-base, clustering, machine learning and more [3,4]. In this work we demonstrate how bot detection can be implemented using a network theory approach.
[ 1, 0, 0, 0, 0, 0 ]
[ "Computer Science", "Quantitative Finance" ]
Title: Near-infrared laser thermal conjunctivoplasty, Abstract: Conjunctivochalasis is a common cause of tear dysfunction due to the conjunctiva becoming loose and wrinkly with age. The current solutions to this disease include either surgical excision in the operating room, or thermoreduction of the loose tissue with hot wire in the clinic. We developed a near-infrared (NIR) laser thermal conjunctivoplasty (LTC) system, which gently shrinks the redundant tissue. The NIR light is mainly absorbed by water, so the heating is even and there is no bleeding. The system utilizes a 1460-nm programmable laser diode system as a light source. A miniaturized handheld probe delivers the laser light and focuses the laser into a 10x1 mm2 line. A foot pedal is used to deliver a preset number of calibrated laser pulses. A fold of loose conjunctiva is grasped by a pair of forceps. The infrared laser light is delivered through an optical fiber and a laser line is focused exactly on the conjunctival fold by a cylindrical lens. Ex vivo experiments using porcine eye were performed with the optimal laser parameters. It was found that up to 50% of conjunctiva shrinkage could be achieved.
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics", "Quantitative Biology" ]
Title: VTA: An Open Hardware-Software Stack for Deep Learning, Abstract: Hardware acceleration is an enabler for ubiquitous and efficient deep learning. With hardware accelerators being introduced in datacenter and edge devices, it is time to acknowledge that hardware specialization is central to the deep learning system stack. This technical report presents the Versatile Tensor Accelerator (VTA), an open, generic, and customizable deep learning accelerator design. VTA is a programmable accelerator that exposes a RISC-like programming abstraction to describe operations at the tensor level. We designed VTA to expose the most salient and common characteristics of mainstream deep learning accelerators, such as tensor operations, DMA load/stores, and explicit compute/memory arbitration. VTA is more than a standalone accelerator design: it's an end-to-end solution that includes drivers, a JIT runtime, and an optimizing compiler stack based on TVM. The current release of VTA includes a behavioral hardware simulator, as well as the infrastructure to deploy VTA on low-cost FPGA development boards for fast prototyping. By extending the TVM stack with a customizable, and open source deep learning hardware accelerator design, we are exposing a transparent end-to-end deep learning stack from the high-level deep learning framework, down to the actual hardware design and implementation. This forms a truly end-to-end, from software-to-hardware open source stack for deep learning systems.
[ 0, 0, 0, 1, 0, 0 ]
[ "Computer Science" ]
Title: Exact partial information decompositions for Gaussian systems based on dependency constraints, Abstract: The Partial Information Decomposition (PID) [arXiv:1004.2515] provides a theoretical framework to characterize and quantify the structure of multivariate information sharing. A new method (Idep) has recently been proposed for computing a two-predictor PID over discrete spaces. [arXiv:1709.06653] A lattice of maximum entropy probability models is constructed based on marginal dependency constraints, and the unique information that a particular predictor has about the target is defined as the minimum increase in joint predictor-target mutual information when that particular predictor-target marginal dependency is constrained. Here, we apply the Idep approach to Gaussian systems, for which the marginally constrained maximum entropy models are Gaussian graphical models. Closed form solutions for the Idep PID are derived for both univariate and multivariate Gaussian systems. Numerical and graphical illustrations are provided, together with practical and theoretical comparisons of the Idep PID with the minimum mutual information PID (Immi). [arXiv:1411.2832] In particular, it is proved that the Immi method generally produces larger estimates of redundancy and synergy than does the Idep method. In discussion of the practical examples, the PIDs are complemented by the use of deviance tests for the comparison of Gaussian graphical models.
[ 0, 0, 0, 1, 1, 0 ]
[ "Statistics", "Mathematics" ]
Title: Finite-time generalization of the thermodynamic uncertainty relation, Abstract: For fluctuating currents in non-equilibrium steady states, the recently discovered thermodynamic uncertainty relation expresses a fundamental relation between their variance and the overall entropic cost associated with the driving. We show that this relation holds not only for the long-time limit of fluctuations, as described by large deviation theory, but also for fluctuations on arbitrary finite time scales. This generalization facilitates applying the thermodynamic uncertainty relation to single molecule experiments, for which infinite timescales are not accessible. Importantly, often this finite-time variant of the relation allows inferring a bound on the entropy production that is even stronger than the one obtained from the long-time limit. We illustrate the relation for the fluctuating work that is performed by a stochastically switching laser tweezer on a trapped colloidal particle.
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics" ]
Title: Composition Properties of Inferential Privacy for Time-Series Data, Abstract: With the proliferation of mobile devices and the internet of things, developing principled solutions for privacy in time series applications has become increasingly important. While differential privacy is the gold standard for database privacy, many time series applications require a different kind of guarantee, and a number of recent works have used some form of inferential privacy to address these situations. However, a major barrier to using inferential privacy in practice is its lack of graceful composition -- even if the same or related sensitive data is used in multiple releases that are safe individually, the combined release may have poor privacy properties. In this paper, we study composition properties of a form of inferential privacy called Pufferfish when applied to time-series data. We show that while general Pufferfish mechanisms may not compose gracefully, a specific Pufferfish mechanism, called the Markov Quilt Mechanism, which was recently introduced, has strong composition properties comparable to that of pure differential privacy when applied to time series data.
[ 1, 0, 0, 1, 0, 0 ]
[ "Computer Science", "Statistics" ]
Title: Well-posedness of a Model for the Growth of Tree Stems and Vines, Abstract: The paper studies a PDE model for the growth of a tree stem or a vine, having the form of a differential inclusion with state constraints. The equations describe the elongation due to cell growth, and the response to gravity and to external obstacles. The main theorem shows that the evolution problem is well posed, until a specific "breakdown configuration" is reached. A formula is proved, characterizing the reaction produced by unilateral constraints. At a.e. time t, this is determined by the minimization of an elastic energy functional under suitable constraints.
[ 0, 0, 1, 0, 0, 0 ]
[ "Mathematics", "Quantitative Biology" ]
Title: Hindsight policy gradients, Abstract: A reinforcement learning agent that needs to pursue different goals across episodes requires a goal-conditional policy. In addition to their potential to generalize desirable behavior to unseen goals, such policies may also enable higher-level planning based on subgoals. In sparse-reward environments, the capacity to exploit information about the degree to which an arbitrary goal has been achieved while another goal was intended appears crucial to enable sample efficient learning. However, reinforcement learning agents have only recently been endowed with such capacity for hindsight. In this paper, we demonstrate how hindsight can be introduced to policy gradient methods, generalizing this idea to a broad class of successful algorithms. Our experiments on a diverse selection of sparse-reward environments show that hindsight leads to a remarkable increase in sample efficiency.
[ 1, 0, 0, 0, 0, 0 ]
[ "Computer Science", "Statistics" ]
Title: Electron-Phonon Interaction in Ternary Rare-Earth Copper Antimonides LaCuSb2 and La(Cu0.8Ag0.2)Sb2 probed by Yanson Point-Contact Spectroscopy, Abstract: Investigation of the electron-phonon interaction (EPI) in LaCuSb2 and La(Cu0.8Ag0.2)Sb2 compounds by Yanson point-contact spectroscopy (PCS) has been carried out. Point-contact spectra display a pronounced broad maximum in the range of 10÷20 mV caused by EPI. Variation of the position of this maximum is likely connected with anisotropic phonon spectrum in these layered compounds. The absence of phonon features after the main maximum allows the assessment of the Debye energy of about 40 meV. The EPI constant for the LaCuSb2 compound was estimated to be {\lambda}=0.2+/-0.03. A zero-bias minimum in differential resistance for the latter compound is observed for some point contacts, which vanishes at about 6 K, pointing to the formation of superconducting phase under point contact, while superconducting critical temperature of the bulk sample is only 1K.
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics" ]
Title: Time-dependent linear-response variational Monte Carlo, Abstract: We present the extension of variational Monte Carlo (VMC) to the calculation of electronic excitation energies and oscillator strengths using time-dependent linear-response theory. By exploiting the analogy existing between the linear method for wave-function optimisation and the generalised eigenvalue equation of linear-response theory, we formulate the equations of linear-response VMC (LR-VMC). This LR-VMC approach involves the first-and second-order derivatives of the wave function with respect to the parameters. We perform first tests of the LR-VMC method within the Tamm-Dancoff approximation using single-determinant Jastrow-Slater wave functions with different Slater basis sets on some singlet and triplet excitations of the beryllium atom. Comparison with reference experimental data and with configuration-interaction-singles (CIS) results shows that LR-VMC generally outperforms CIS for excitation energies and is thus a promising approach for calculating electronic excited-state properties of atoms and molecules.
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics", "Mathematics" ]
Title: Replica analysis of overfitting in regression models for time-to-event data, Abstract: Overfitting, which happens when the number of parameters in a model is too large compared to the number of data points available for determining these parameters, is a serious and growing problem in survival analysis. While modern medicine presents us with data of unprecedented dimensionality, these data cannot yet be used effectively for clinical outcome prediction. Standard error measures in maximum likelihood regression, such as p-values and z-scores, are blind to overfitting, and even for Cox's proportional hazards model (the main tool of medical statisticians), one finds in literature only rules of thumb on the number of samples required to avoid overfitting. In this paper we present a mathematical theory of overfitting in regression models for time-to-event data, which aims to increase our quantitative understanding of the problem and provide practical tools with which to correct regression outcomes for the impact of overfitting. It is based on the replica method, a statistical mechanical technique for the analysis of heterogeneous many-variable systems that has been used successfully for several decades in physics, biology, and computer science, but not yet in medical statistics. We develop the theory initially for arbitrary regression models for time-to-event data, and verify its predictions in detail for the popular Cox model.
[ 0, 1, 0, 1, 0, 0 ]
[ "Statistics", "Mathematics", "Quantitative Biology" ]
Title: Spinor analysis, Abstract: "Let us call the novel quantities which, in addition to the vectors and tensors, have appeared in the quantum mechanics of the spinning electron, and which in the case of the Lorentz group are quite differently transformed from tensors, as spinors for short. Is there no spinor analysis that every physicist can learn, such as tensor analysis, and with the aid of which all the possible spinors can be formed, and secondly, all the invariant equations in which spinors occur?" So Mr Ehrenfest asked me and the answer will be given below.
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics", "Mathematics" ]
Title: Identifiability of phylogenetic parameters from k-mer data under the coalescent, Abstract: Distances between sequences based on their $k$-mer frequency counts can be used to reconstruct phylogenies without first computing a sequence alignment. Past work has shown that effective use of k-mer methods depends on 1) model-based corrections to distances based on $k$-mers and 2) breaking long sequences into blocks to obtain repeated trials from the sequence-generating process. Good performance of such methods is based on having many high-quality blocks with many homologous sites, which can be problematic to guarantee a priori. Nature provides natural blocks of sequences into homologous regions---namely, the genes. However, directly using past work in this setting is problematic because of possible discordance between different gene trees and the underlying species tree. Using the multispecies coalescent model as a basis, we derive model-based moment formulas that involve the divergence times and the coalescent parameters. From this setting, we prove identifiability results for the tree and branch length parameters under the Jukes-Cantor model of sequence mutations.
[ 0, 0, 1, 0, 0, 0 ]
[ "Quantitative Biology", "Statistics" ]
Title: Using Transfer Learning for Image-Based Cassava Disease Detection, Abstract: Cassava is the third largest source of carbohydrates for human food in the world but is vulnerable to virus diseases, which threaten to destabilize food security in sub-Saharan Africa. Novel methods of cassava disease detection are needed to support improved control which will prevent this crisis. Image recognition offers both a cost effective and scalable technology for disease detection. New transfer learning methods offer an avenue for this technology to be easily deployed on mobile devices. Using a dataset of cassava disease images taken in the field in Tanzania, we applied transfer learning to train a deep convolutional neural network to identify three diseases and two types of pest damage (or lack thereof). The best trained model accuracies were 98% for brown leaf spot (BLS), 96% for red mite damage (RMD), 95% for green mite damage (GMD), 98% for cassava brown streak disease (CBSD), and 96% for cassava mosaic disease (CMD). The best model achieved an overall accuracy of 93% for data not used in the training process. Our results show that the transfer learning approach for image recognition of field images offers a fast, affordable, and easily deployable strategy for digital plant disease detection.
[ 1, 0, 0, 0, 0, 0 ]
[ "Computer Science", "Quantitative Biology" ]
Title: Discovery of Extreme [OIII]+H$β$ Emitting Galaxies Tracing an Overdensity at z~3.5 in CDF-South, Abstract: Using deep multi-wavelength photometry of galaxies from ZFOURGE, we group galaxies at $2.5<z<4.0$ by the shape of their spectral energy distributions (SEDs). We identify a population of galaxies with excess emission in the $K_s$-band, which corresponds to [OIII]+H$\beta$ emission at $2.95<z<3.65$. This population includes 78% of the bluest galaxies with UV slopes steeper than $\beta = -2$. We de-redshift and scale this photometry to build two composite SEDs, enabling us to measure equivalent widths of these Extreme [OIII]+H$\beta$ Emission Line Galaxies (EELGs) at $z\sim3.5$. We identify 60 galaxies that comprise a composite SED with [OIII]+H$\beta$ rest-frame equivalent width of $803\pm228$\AA\ and another 218 galaxies in a composite SED with equivalent width of $230\pm90$\AA. These EELGs are analogous to the `green peas' found in the SDSS, and are thought to be undergoing their first burst of star formation due to their blue colors ($\beta < -1.6$), young ages ($\log(\rm{age}/yr)\sim7.2$), and low dust attenuation values. Their strong nebular emission lines and compact sizes (typically $\sim1.4$ kpc) are consistent with the properties of the star-forming galaxies possibly responsible for reionizing the universe at $z>6$. Many of the EELGs also exhibit Lyman-$\alpha$ emission. Additionally, we find that many of these sources are clustered in an overdensity in the Chandra Deep Field South, with five spectroscopically confirmed members at $z=3.474 \pm 0.004$. The spatial distribution and photometric redshifts of the ZFOURGE population further confirm the overdensity highlighted by the EELGs.
[ 0, 1, 0, 0, 0, 0 ]
[ "Astrophysics" ]
Title: Learning to attend in a brain-inspired deep neural network, Abstract: Recent machine learning models have shown that including attention as a component results in improved model accuracy and interpretability, despite the concept of attention in these approaches only loosely approximating the brain's attention mechanism. Here we extend this work by building a more brain-inspired deep network model of the primate ATTention Network (ATTNet) that learns to shift its attention so as to maximize the reward. Using deep reinforcement learning, ATTNet learned to shift its attention to the visual features of a target category in the context of a search task. ATTNet's dorsal layers also learned to prioritize these shifts of attention so as to maximize success of the ventral pathway classification and receive greater reward. Model behavior was tested against the fixations made by subjects searching images for the same cued category. Both subjects and ATTNet showed evidence for attention being preferentially directed to target goals, behaviorally measured as oculomotor guidance to targets. More fundamentally, ATTNet learned to shift its attention to target like objects and spatially route its visual inputs to accomplish the task. This work makes a step toward a better understanding of the role of attention in the brain and other computational systems.
[ 0, 0, 0, 0, 1, 0 ]
[ "Computer Science", "Quantitative Biology" ]
Title: Anisotropic functional Laplace deconvolution, Abstract: In the present paper we consider the problem of estimating a three-dimensional function $f$ based on observations from its noisy Laplace convolution. Our study is motivated by the analysis of Dynamic Contrast Enhanced (DCE) imaging data. We construct an adaptive wavelet-Laguerre estimator of $f$, derive minimax lower bounds for the $L^2$-risk when $f$ belongs to a three-dimensional Laguerre-Sobolev ball and demonstrate that the wavelet-Laguerre estimator is adaptive and asymptotically near-optimal in a wide range of Laguerre-Sobolev spaces. We carry out a limited simulations study and show that the estimator performs well in a finite sample setting. Finally, we use the technique for the solution of the Laplace deconvolution problem on the basis of DCE Computerized Tomography data.
[ 0, 0, 0, 1, 0, 0 ]
[ "Mathematics", "Statistics" ]
Title: Parameter Sharing Deep Deterministic Policy Gradient for Cooperative Multi-agent Reinforcement Learning, Abstract: Deep reinforcement learning for multi-agent cooperation and competition has been a hot topic recently. This paper focuses on cooperative multi-agent problem based on actor-critic methods under local observations settings. Multi agent deep deterministic policy gradient obtained state of art results for some multi-agent games, whereas, it cannot scale well with growing amount of agents. In order to boost scalability, we propose a parameter sharing deterministic policy gradient method with three variants based on neural networks, including actor-critic sharing, actor sharing and actor sharing with partially shared critic. Benchmarks from rllab show that the proposed method has advantages in learning speed and memory efficiency, well scales with growing amount of agents, and moreover, it can make full use of reward sharing and exchangeability if possible.
[ 1, 0, 0, 0, 0, 0 ]
[ "Computer Science", "Statistics" ]
Title: On a Neumann-type series for modified Bessel functions of the first kind, Abstract: In this paper, we are interested in a Neumann-type series for modified Bessel functions of the first kind which arises in the study of Dunkl operators associated with dihedral groups and as an instance of the Laguerre semigroup constructed by Ben Said-Kobayashi-Orsted. We first revisit the particular case corresponding to the group of square-preserving symmetries for which we give two new and different proofs other than the existing ones. The first proof uses the expansion of powers in a Neumann series of Bessel functions while the second one is based on a quadratic transformation for the Gauss hypergeometric function and opens the way to derive further expressions when the orders of the underlying dihedral groups are powers of two. More generally, we give another proof of De Bie \& al formula expressing this series as a $\Phi_2$-Horn confluent hypergeometric function. In the course of proving, we shed the light on the occurrence of multiple angles in their formula through elementary symmetric functions, and get a new representation of Gegenbauer polynomials.
[ 0, 0, 1, 0, 0, 0 ]
[ "Mathematics" ]
Title: A Modern Search for Wolf-Rayet Stars in the Magellanic Clouds. III. A Third Year of Discoveries, Abstract: For the past three years we have been conducting a survey for WR stars in the Large and Small Magellanic Clouds (LMC, SMC). Our previous work has resulted in the discovery of a new type of WR star in the LMC, which we are calling WN3/O3. These stars have the emission-line properties of a WN3 star (strong N V but no N IV), plus the absorption-line properties of an O3 star (Balmer hydrogen plus Pickering He II but no He I). Yet these stars are 15x fainter than an O3 V star would be by itself, ruling out these being WN3+O3 binaries. Here we report the discovery of two more members of this class, bringing the total number of these objects to 10, 6.5% of the LMC's total WR population. The optical spectra of nine of these WN3/O3s are virtually indistinguishable from each other, but one of the newly found stars is significantly different, showing a lower excitation emission and absorption spectrum (WN4/O4-ish). In addition, we have newly classified three unusual Of-type stars, including one with a strong C III 4650 line, and two rapidly rotating "Oef" stars. We also "rediscovered" a low mass x-ray binary, RX J0513.9-6951, and demonstrate its spectral variability. Finally, we discuss the spectra of ten low priority WR candidates that turned out not to have He II emission. These include both a Be star and a B[e] star.
[ 0, 1, 0, 0, 0, 0 ]
[ "Physics" ]
Title: Mathematical modeling of Zika disease in pregnant women and newborns with microcephaly in Brazil, Abstract: We propose a new mathematical model for the spread of Zika virus. Special attention is paid to the transmission of microcephaly. Numerical simulations show the accuracy of the model with respect to the Zika outbreak occurred in Brazil.
[ 0, 0, 1, 0, 0, 0 ]
[ "Mathematics", "Quantitative Biology" ]