@article{7117, abstract = {We propose a novel generic shape optimization method for CAD models based on the eXtended Finite Element Method (XFEM). Our method works directly on the intersection between the model and a regular simulation grid, without the need to mesh or remesh, thus removing a bottleneck of classical shape optimization strategies. This is made possible by a novel hierarchical integration scheme that accurately integrates finite element quantities with sub-element precision. For optimization, we efficiently compute analytical shape derivatives of the entire framework, from model intersection to integration rule generation and XFEM simulation. Moreover, we describe a differentiable projection of shape parameters onto a constraint manifold spanned by user-specified shape preservation, consistency, and manufacturability constraints. We demonstrate the utility of our approach by optimizing mass distribution, strength-to-weight ratio, and inverse elastic shape design objectives directly on parameterized 3D CAD models.}, author = {Hafner, Christian and Schumacher, Christian and Knoop, Espen and Auzinger, Thomas and Bickel, Bernd and Bächer, Moritz}, issn = {0730-0301}, journal = {ACM Transactions on Graphics}, number = {6}, publisher = {ACM}, title = {{X-CAD: Optimizing CAD Models with Extended Finite Elements}}, doi = {10.1145/3355089.3356576}, volume = {38}, year = {2019}, } @article{6189, abstract = {Suspended particles can alter the properties of fluids and in particular also affect the transition fromlaminar to turbulent flow. An earlier study [Mataset al.,Phys. Rev. Lett.90, 014501 (2003)] reported howthe subcritical (i.e., hysteretic) transition to turbulent puffs is affected by the addition of particles. Here weshow that in addition to this known transition, with increasing concentration a supercritical (i.e.,continuous) transition to a globally fluctuating state is found. At the same time the Newtonian-typetransition to puffs is delayed to larger Reynolds numbers. At even higher concentration only the globallyfluctuating state is found. The dynamics of particle laden flows are hence determined by two competinginstabilities that give rise to three flow regimes: Newtonian-type turbulence at low, a particle inducedglobally fluctuating state at high, and a coexistence state at intermediate concentrations.}, author = {Agrawal, Nishchal and Choueiri, George H and Hof, Björn}, issn = {10797114}, journal = {Physical Review Letters}, number = {11}, publisher = {American Physical Society}, title = {{Transition to turbulence in particle laden flows}}, doi = {10.1103/PhysRevLett.122.114502}, volume = {122}, year = {2019}, } @phdthesis{6371, abstract = {Decades of studies have revealed the mechanisms of gene regulation in molecular detail. We make use of such well-described regulatory systems to explore how the molecular mechanisms of protein-protein and protein-DNA interactions shape the dynamics and evolution of gene regulation. i) We uncover how the biophysics of protein-DNA binding determines the potential of regulatory networks to evolve and adapt, which can be captured using a simple mathematical model. ii) The evolution of regulatory connections can lead to a significant amount of crosstalk between binding proteins. We explore the effect of crosstalk on gene expression from a target promoter, which seems to be modulated through binding competition at non-specific DNA sites. iii) We investigate how the very same biophysical characteristics as in i) can generate significant fitness costs for cells through global crosstalk, meaning non-specific DNA binding across the genomic background. iv) Binding competition between proteins at a target promoter is a prevailing regulatory feature due to the prevalence of co-regulation at bacterial promoters. However, the dynamics of these systems are not always straightforward to determine even if the molecular mechanisms of regulation are known. A detailed model of the biophysical interactions reveals that interference between the regulatory proteins can constitute a new, generic form of system memory that records the history of the input signals at the promoter. We demonstrate how the biophysics of protein-DNA binding can be harnessed to investigate the principles that shape and ultimately limit cellular gene regulation. These results provide a basis for studies of higher-level functionality, which arises from the underlying regulation. }, author = {Igler, Claudia}, issn = {2663-337X}, keywords = {gene regulation, biophysics, transcription factor binding, bacteria}, pages = {152}, publisher = {Institute of Science and Technology Austria}, title = {{On the nature of gene regulatory design - The biophysics of transcription factor binding shapes gene regulation}}, doi = {10.15479/AT:ISTA:6371}, year = {2019}, } @article{10286, abstract = {In this paper, we evaluate clock signals generated in ring oscillators and self-timed rings and the way their jitter can be transformed into random numbers. We show that counting the periods of the jittery clock signal produces random numbers of significantly better quality than the methods in which the jittery signal is simply sampled (the case in almost all current methods). Moreover, we use the counter values to characterize and continuously monitor the source of randomness. However, instead of using the widely used statistical variance, we propose to use Allan variance to do so. There are two main advantages: Allan variance is insensitive to low frequency noises such as flicker noise that are known to be autocorrelated and significantly less circuitry is required for its computation than that used to compute commonly used variance. We also show that it is essential to use a differential principle of randomness extraction from the jitter based on the use of two identical oscillators to avoid autocorrelations originating from external and internal global jitter sources and that this fact is valid for both kinds of rings. Last but not least, we propose a method of statistical testing based on high order Markov model to show the reduced dependencies when the proposed randomness extraction is applied.}, author = {Allini, Elie Noumon and Skórski, Maciej and Petura, Oto and Bernard, Florent and Laban, Marek and Fischer, Viktor}, issn = {2569-2925}, journal = {IACR Transactions on Cryptographic Hardware and Embedded Systems}, number = {3}, pages = {214--242}, publisher = {International Association for Cryptologic Research}, title = {{Evaluation and monitoring of free running oscillators serving as source of randomness}}, doi = {10.13154/tches.v2018.i3.214-242}, volume = {2018}, year = {2018}, } @inproceedings{10883, abstract = {Solving parity games, which are equivalent to modal μ-calculus model checking, is a central algorithmic problem in formal methods, with applications in reactive synthesis, program repair, verification of branching-time properties, etc. Besides the standard compu- tation model with the explicit representation of games, another important theoretical model of computation is that of set-based symbolic algorithms. Set-based symbolic algorithms use basic set operations and one-step predecessor operations on the implicit description of games, rather than the explicit representation. The significance of symbolic algorithms is that they provide scalable algorithms for large finite-state systems, as well as for infinite-state systems with finite quotient. Consider parity games on graphs with n vertices and parity conditions with d priorities. While there is a rich literature of explicit algorithms for parity games, the main results for set-based symbolic algorithms are as follows: (a) the basic algorithm that requires O(nd) symbolic operations and O(d) symbolic space; and (b) an improved algorithm that requires O(nd/3+1) symbolic operations and O(n) symbolic space. In this work, our contributions are as follows: (1) We present a black-box set-based symbolic algorithm based on the explicit progress measure algorithm. Two important consequences of our algorithm are as follows: (a) a set-based symbolic algorithm for parity games that requires quasi-polynomially many symbolic operations and O(n) symbolic space; and (b) any future improvement in progress measure based explicit algorithms immediately imply an efficiency improvement in our set-based symbolic algorithm for parity games. (2) We present a set-based symbolic algorithm that requires quasi-polynomially many symbolic operations and O(d · log n) symbolic space. Moreover, for the important special case of d ≤ log n, our algorithm requires only polynomially many symbolic operations and poly-logarithmic symbolic space.}, author = {Chatterjee, Krishnendu and Dvořák, Wolfgang and Henzinger, Monika H and Svozil, Alexander}, booktitle = {22nd International Conference on Logic for Programming, Artificial Intelligence and Reasoning}, issn = {2398-7340}, location = {Awassa, Ethiopia}, pages = {233--253}, publisher = {EasyChair}, title = {{Quasipolynomial set-based symbolic algorithms for parity games}}, doi = {10.29007/5z5k}, volume = {57}, year = {2018}, } @inproceedings{11, abstract = {We report on a novel strategy to derive mean-field limits of quantum mechanical systems in which a large number of particles weakly couple to a second-quantized radiation field. The technique combines the method of counting and the coherent state approach to study the growth of the correlations among the particles and in the radiation field. As an instructional example, we derive the Schrödinger–Klein–Gordon system of equations from the Nelson model with ultraviolet cutoff and possibly massless scalar field. In particular, we prove the convergence of the reduced density matrices (of the nonrelativistic particles and the field bosons) associated with the exact time evolution to the projectors onto the solutions of the Schrödinger–Klein–Gordon equations in trace norm. Furthermore, we derive explicit bounds on the rate of convergence of the one-particle reduced density matrix of the nonrelativistic particles in Sobolev norm.}, author = {Leopold, Nikolai K and Pickl, Peter}, location = {Munich, Germany}, pages = {185 -- 214}, publisher = {Springer}, title = {{Mean-field limits of particles in interaction with quantised radiation fields}}, doi = {10.1007/978-3-030-01602-9_9}, volume = {270}, year = {2018}, } @article{1215, abstract = {Two generalizations of Itô formula to infinite-dimensional spaces are given. The first one, in Hilbert spaces, extends the classical one by taking advantage of cancellations when they occur in examples and it is applied to the case of a group generator. The second one, based on the previous one and a limit procedure, is an Itô formula in a special class of Banach spaces having a product structure with the noise in a Hilbert component; again the key point is the extension due to a cancellation. This extension to Banach spaces and in particular the specific cancellation are motivated by path-dependent Itô calculus.}, author = {Flandoli, Franco and Russo, Francesco and Zanco, Giovanni A}, journal = {Journal of Theoretical Probability}, number = {2}, pages = {789--826}, publisher = {Springer}, title = {{Infinite-dimensional calculus under weak spatial regularity of the processes}}, doi = {10.1007/s10959-016-0724-2}, volume = {31}, year = {2018}, } @inproceedings{185, abstract = {We resolve in the affirmative conjectures of A. Skopenkov and Repovš (1998), and M. Skopenkov (2003) generalizing the classical Hanani-Tutte theorem to the setting of approximating maps of graphs on 2-dimensional surfaces by embeddings. Our proof of this result is constructive and almost immediately implies an efficient algorithm for testing whether a given piecewise linear map of a graph in a surface is approximable by an embedding. More precisely, an instance of this problem consists of (i) a graph G whose vertices are partitioned into clusters and whose inter-cluster edges are partitioned into bundles, and (ii) a region R of a 2-dimensional compact surface M given as the union of a set of pairwise disjoint discs corresponding to the clusters and a set of pairwise disjoint "pipes" corresponding to the bundles, connecting certain pairs of these discs. We are to decide whether G can be embedded inside M so that the vertices in every cluster are drawn in the corresponding disc, the edges in every bundle pass only through its corresponding pipe, and every edge crosses the boundary of each disc at most once.}, author = {Fulek, Radoslav and Kynčl, Jan}, isbn = {978-3-95977-066-8}, location = {Budapest, Hungary}, publisher = {Schloss Dagstuhl - Leibniz-Zentrum für Informatik}, title = {{Hanani-Tutte for approximating maps of graphs}}, doi = {10.4230/LIPIcs.SoCG.2018.39}, volume = {99}, year = {2018}, } @inproceedings{188, abstract = {Smallest enclosing spheres of finite point sets are central to methods in topological data analysis. Focusing on Bregman divergences to measure dissimilarity, we prove bounds on the location of the center of a smallest enclosing sphere. These bounds depend on the range of radii for which Bregman balls are convex.}, author = {Edelsbrunner, Herbert and Virk, Ziga and Wagner, Hubert}, location = {Budapest, Hungary}, pages = {35:1 -- 35:13}, publisher = {Schloss Dagstuhl - Leibniz-Zentrum für Informatik}, title = {{Smallest enclosing spheres and Chernoff points in Bregman geometry}}, doi = {10.4230/LIPIcs.SoCG.2018.35}, volume = {99}, year = {2018}, } @article{306, abstract = {A cornerstone of statistical inference, the maximum entropy framework is being increasingly applied to construct descriptive and predictive models of biological systems, especially complex biological networks, from large experimental data sets. Both its broad applicability and the success it obtained in different contexts hinge upon its conceptual simplicity and mathematical soundness. Here we try to concisely review the basic elements of the maximum entropy principle, starting from the notion of ‘entropy’, and describe its usefulness for the analysis of biological systems. As examples, we focus specifically on the problem of reconstructing gene interaction networks from expression data and on recent work attempting to expand our system-level understanding of bacterial metabolism. Finally, we highlight some extensions and potential limitations of the maximum entropy approach, and point to more recent developments that are likely to play a key role in the upcoming challenges of extracting structures and information from increasingly rich, high-throughput biological data.}, author = {De Martino, Andrea and De Martino, Daniele}, journal = {Heliyon}, number = {4}, publisher = {Elsevier}, title = {{An introduction to the maximum entropy approach and its application to inference problems in biology}}, doi = {10.1016/j.heliyon.2018.e00596}, volume = {4}, year = {2018}, } @book{3300, abstract = {This book first explores the origins of this idea, grounded in theoretical work on temporal logic and automata. The editors and authors are among the world's leading researchers in this domain, and they contributed 32 chapters representing a thorough view of the development and application of the technique. Topics covered include binary decision diagrams, symbolic model checking, satisfiability modulo theories, partial-order reduction, abstraction, interpolation, concurrency, security protocols, games, probabilistic model checking, and process algebra, and chapters on the transfer of theory to industrial practice, property specification languages for hardware, and verification of real-time systems and hybrid systems. The book will be valuable for researchers and graduate students engaged with the development of formal methods and verification tools.}, author = {Clarke, Edmund M. and Henzinger, Thomas A and Veith, Helmut and Bloem, Roderick}, isbn = {978-3-319-10574-1}, pages = {XLVIII, 1212}, publisher = {Springer Nature}, title = {{Handbook of Model Checking}}, doi = {10.1007/978-3-319-10575-8}, year = {2018}, } @inbook{37, abstract = {Developmental processes are inherently dynamic and understanding them requires quantitative measurements of gene and protein expression levels in space and time. While live imaging is a powerful approach for obtaining such data, it is still a challenge to apply it over long periods of time to large tissues, such as the embryonic spinal cord in mouse and chick. Nevertheless, dynamics of gene expression and signaling activity patterns in this organ can be studied by collecting tissue sections at different developmental stages. In combination with immunohistochemistry, this allows for measuring the levels of multiple developmental regulators in a quantitative manner with high spatiotemporal resolution. The mean protein expression levels over time, as well as embryo-to-embryo variability can be analyzed. A key aspect of the approach is the ability to compare protein levels across different samples. This requires a number of considerations in sample preparation, imaging and data analysis. Here we present a protocol for obtaining time course data of dorsoventral expression patterns from mouse and chick neural tube in the first 3 days of neural tube development. The described workflow starts from embryo dissection and ends with a processed dataset. Software scripts for data analysis are included. The protocol is adaptable and instructions that allow the user to modify different steps are provided. Thus, the procedure can be altered for analysis of time-lapse images and applied to systems other than the neural tube.}, author = {Zagórski, Marcin P and Kicheva, Anna}, booktitle = {Morphogen Gradients }, isbn = {978-1-4939-8771-9}, issn = {1064-3745}, pages = {47 -- 63}, publisher = {Springer Nature}, title = {{Measuring dorsoventral pattern and morphogen signaling profiles in the growing neural tube}}, doi = {10.1007/978-1-4939-8772-6_4}, volume = {1863}, year = {2018}, } @article{305, abstract = {The hanging-drop network (HDN) is a technology platform based on a completely open microfluidic network at the bottom of an inverted, surface-patterned substrate. The platform is predominantly used for the formation, culturing, and interaction of self-assembled spherical microtissues (spheroids) under precisely controlled flow conditions. Here, we describe design, fabrication, and operation of microfluidic hanging-drop networks.}, author = {Misun, Patrick and Birchler, Axel and Lang, Moritz and Hierlemann, Andreas and Frey, Olivier}, journal = {Methods in Molecular Biology}, pages = {183 -- 202}, publisher = {Springer}, title = {{Fabrication and operation of microfluidic hanging drop networks}}, doi = {10.1007/978-1-4939-7792-5_15}, volume = {1771}, year = {2018}, } @inproceedings{325, abstract = {Probabilistic programs extend classical imperative programs with real-valued random variables and random branching. The most basic liveness property for such programs is the termination property. The qualitative (aka almost-sure) termination problem asks whether a given program program terminates with probability 1. While ranking functions provide a sound and complete method for non-probabilistic programs, the extension of them to probabilistic programs is achieved via ranking supermartingales (RSMs). Although deep theoretical results have been established about RSMs, their application to probabilistic programs with nondeterminism has been limited only to programs of restricted control-flow structure. For non-probabilistic programs, lexicographic ranking functions provide a compositional and practical approach for termination analysis of real-world programs. In this work we introduce lexicographic RSMs and show that they present a sound method for almost-sure termination of probabilistic programs with nondeterminism. We show that lexicographic RSMs provide a tool for compositional reasoning about almost-sure termination, and for probabilistic programs with linear arithmetic they can be synthesized efficiently (in polynomial time). We also show that with additional restrictions even asymptotic bounds on expected termination time can be obtained through lexicographic RSMs. Finally, we present experimental results on benchmarks adapted from previous work to demonstrate the effectiveness of our approach.}, author = {Agrawal, Sheshansh and Chatterjee, Krishnendu and Novotny, Petr}, location = {Los Angeles, CA, USA}, number = {POPL}, publisher = {ACM}, title = {{Lexicographic ranking supermartingales: an efficient approach to termination of probabilistic programs}}, doi = {10.1145/3158122}, volume = {2}, year = {2018}, } @inbook{408, abstract = {Adventitious roots (AR) are de novo formed roots that emerge from any part of the plant or from callus in tissue culture, except root tissue. The plant tissue origin and the method by which they are induced determine the physiological properties of emerged ARs. Hence, a standard method encompassing all types of AR does not exist. Here we describe a method for the induction and analysis of AR that emerge from the etiolated hypocotyl of dicot plants. The hypocotyl is formed during embryogenesis and shows a determined developmental pattern which usually does not involve AR formation. However, the hypocotyl shows propensity to form de novo roots under specific circumstances such as removal of the root system, high humidity or flooding, or during de-etiolation. The hypocotyl AR emerge from a pericycle-like cell layer surrounding the vascular tissue of the central cylinder, which is reminiscent to the developmental program of lateral roots. Here we propose an easy protocol for in vitro hypocotyl AR induction from etiolated Arabidopsis seedlings.}, author = {Trinh, Hoang and Verstraeten, Inge and Geelen, Danny}, booktitle = {Root Development }, issn = {1064-3745}, pages = {95 -- 102}, publisher = {Springer Nature}, title = {{In vitro assay for induction of adventitious rooting on intact arabidopsis hypocotyls}}, doi = {10.1007/978-1-4939-7747-5_7}, volume = {1761}, year = {2018}, } @inbook{411, abstract = {Immunolocalization is a valuable tool for cell biology research that allows to rapidly determine the localization and expression levels of endogenous proteins. In plants, whole-mount in situ immunolocalization remains a challenging method, especially in tissues protected by waxy layers and complex cell wall carbohydrates. Here, we present a robust method for whole-mount in situ immunolocalization in primary root meristems and lateral root primordia in Arabidopsis thaliana. For good epitope preservation, fixation is done in an alkaline paraformaldehyde/glutaraldehyde mixture. This fixative is suitable for detecting a wide range of proteins, including integral transmembrane proteins and proteins peripherally attached to the plasma membrane. From initiation until emergence from the primary root, lateral root primordia are surrounded by several layers of differentiated tissues with a complex cell wall composition that interferes with the efficient penetration of all buffers. Therefore, immunolocalization in early lateral root primordia requires a modified method, including a strong solvent treatment for removal of hydrophobic barriers and a specific cocktail of cell wall-degrading enzymes. The presented method allows for easy, reliable, and high-quality in situ detection of the subcellular localization of endogenous proteins in primary and lateral root meristems without the need of time-consuming crosses or making translational fusions to fluorescent proteins.}, author = {Karampelias, Michael and Tejos, Ricardo and Friml, Jirí and Vanneste, Steffen}, booktitle = {Root Development. Methods and Protocols}, editor = {Ristova, Daniela and Barbez, Elke}, pages = {131 -- 143}, publisher = {Springer}, title = {{Optimized whole mount in situ immunolocalization for Arabidopsis thaliana root meristems and lateral root primordia}}, doi = {10.1007/978-1-4939-7747-5_10}, volume = {1761}, year = {2018}, } @article{456, abstract = {Inhibition of the endoplasmic reticulum stress pathway may hold the key to Zika virus-associated microcephaly treatment. }, author = {Novarino, Gaia}, journal = {Science Translational Medicine}, number = {423}, publisher = {American Association for the Advancement of Science}, title = {{Zika-associated microcephaly: Reduce the stress and race for the treatment}}, doi = {10.1126/scitranslmed.aar7514}, volume = {10}, year = {2018}, } @article{53, abstract = {In 2013, a publication repository was implemented at IST Austria and 2015 after a thorough preparation phase a data repository was implemented - both based on the Open Source Software EPrints. In this text, designed as field report, we will reflect on our experiences with Open Source Software in general and specifically with EPrints regarding technical aspects but also regarding their characteristics of the user community. The second part is a pleading for including the end users in the process of implementation, adaption and evaluation.}, author = {Petritsch, Barbara and Porsche, Jana}, journal = {VÖB Mitteilungen}, number = {1}, pages = {199 -- 206}, publisher = {Vereinigung Österreichischer Bibliothekarinnen und Bibliothekare}, title = {{IST PubRep and IST DataRep: the institutional repositories at IST Austria}}, doi = {10.31263/voebm.v71i1.1993}, volume = {71}, year = {2018}, } @article{536, abstract = {We consider the problem of consensus in the challenging classic model. In this model, the adversary is adaptive; it can choose which processors crash at any point during the course of the algorithm. Further, communication is via asynchronous message passing: there is no known upper bound on the time to send a message from one processor to another, and all messages and coin flips are seen by the adversary. We describe a new randomized consensus protocol with expected message complexity O(n2log2n) when fewer than n / 2 processes may fail by crashing. This is an almost-linear improvement over the best previously known protocol, and within logarithmic factors of a known Ω(n2) message lower bound. The protocol further ensures that no process sends more than O(nlog3n) messages in expectation, which is again within logarithmic factors of optimal. We also present a generalization of the algorithm to an arbitrary number of failures t, which uses expected O(nt+t2log2t) total messages. Our approach is to build a message-efficient, resilient mechanism for aggregating individual processor votes, implementing the message-passing equivalent of a weak shared coin. Roughly, in our protocol, a processor first announces its votes to small groups, then propagates them to increasingly larger groups as it generates more and more votes. To bound the number of messages that an individual process might have to send or receive, the protocol progressively increases the weight of generated votes. The main technical challenge is bounding the impact of votes that are still “in flight” (generated, but not fully propagated) on the final outcome of the shared coin, especially since such votes might have different weights. We achieve this by leveraging the structure of the algorithm, and a technical argument based on martingale concentration bounds. Overall, we show that it is possible to build an efficient message-passing implementation of a shared coin, and in the process (almost-optimally) solve the classic consensus problem in the asynchronous message-passing model.}, author = {Alistarh, Dan-Adrian and Aspnes, James and King, Valerie and Saia, Jared}, issn = {01782770}, journal = {Distributed Computing}, number = {6}, pages = {489--501}, publisher = {Springer}, title = {{Communication-efficient randomized consensus}}, doi = {10.1007/s00446-017-0315-1}, volume = {31}, year = {2018}, } @article{554, abstract = {We analyse the canonical Bogoliubov free energy functional in three dimensions at low temperatures in the dilute limit. We prove existence of a first-order phase transition and, in the limit (Formula presented.), we determine the critical temperature to be (Formula presented.) to leading order. Here, (Formula presented.) is the critical temperature of the free Bose gas, ρ is the density of the gas and a is the scattering length of the pair-interaction potential V. We also prove asymptotic expansions for the free energy. In particular, we recover the Lee–Huang–Yang formula in the limit (Formula presented.).}, author = {Napiórkowski, Marcin M and Reuvers, Robin and Solovej, Jan}, issn = {00103616}, journal = {Communications in Mathematical Physics}, number = {1}, pages = {347--403}, publisher = {Springer}, title = {{The Bogoliubov free energy functional II: The dilute Limit}}, doi = {10.1007/s00220-017-3064-x}, volume = {360}, year = {2018}, } @inbook{562, abstract = {Primary neuronal cell culture preparations are widely used to investigate synaptic functions. This chapter describes a detailed protocol for the preparation of a neuronal cell culture in which giant calyx-type synaptic terminals are formed. This chapter also presents detailed protocols for utilizing the main technical advantages provided by such a preparation, namely, labeling and imaging of synaptic organelles and electrophysiological recordings directly from presynaptic terminals.}, author = {Dimitrov, Dimitar and Guillaud, Laurent and Eguchi, Kohgaku and Takahashi, Tomoyuki}, booktitle = {Neurotrophic Factors}, editor = {Skaper, Stephen D.}, pages = {201 -- 215}, publisher = {Springer}, title = {{Culture of mouse giant central nervous system synapses and application for imaging and electrophysiological analyses}}, doi = {10.1007/978-1-4939-7571-6_15}, volume = {1727}, year = {2018}, } @inbook{59, abstract = {Graph-based games are an important tool in computer science. They have applications in synthesis, verification, refinement, and far beyond. We review graphbased games with objectives on infinite plays. We give definitions and algorithms to solve the games and to give a winning strategy. The objectives we consider are mostly Boolean, but we also look at quantitative graph-based games and their objectives. Synthesis aims to turn temporal logic specifications into correct reactive systems. We explain the reduction of synthesis to graph-based games (or equivalently tree automata) using synthesis of LTL specifications as an example. We treat the classical approach that uses determinization of parity automata and more modern approaches.}, author = {Bloem, Roderick and Chatterjee, Krishnendu and Jobstmann, Barbara}, booktitle = {Handbook of Model Checking}, editor = {Henzinger, Thomas A and Clarke, Edmund M. and Veith, Helmut and Bloem, Roderick}, isbn = {978-3-319-10574-1}, pages = {921 -- 962}, publisher = {Springer}, title = {{Graph games and reactive synthesis}}, doi = {10.1007/978-3-319-10575-8_27}, year = {2018}, } @inbook{60, abstract = {Model checking is a computer-assisted method for the analysis of dynamical systems that can be modeled by state-transition systems. Drawing from research traditions in mathematical logic, programming languages, hardware design, and theoretical computer science, model checking is now widely used for the verification of hardware and software in industry. This chapter is an introduction and short survey of model checking. The chapter aims to motivate and link the individual chapters of the handbook, and to provide context for readers who are not familiar with model checking.}, author = {Clarke, Edmund and Henzinger, Thomas A and Veith, Helmut}, booktitle = {Handbook of Model Checking}, editor = {Henzinger, Thomas A}, pages = {1 -- 26}, publisher = {Springer}, title = {{Introduction to model checking}}, doi = {10.1007/978-3-319-10575-8_1}, year = {2018}, } @inbook{61, abstract = {We prove that there is no strongly regular graph (SRG) with parameters (460; 153; 32; 60). The proof is based on a recent lower bound on the number of 4-cliques in a SRG and some applications of Euclidean representation of SRGs. }, author = {Bondarenko, Andriy and Mellit, Anton and Prymak, Andriy and Radchenko, Danylo and Viazovska, Maryna}, booktitle = {Contemporary Computational Mathematics}, pages = {131 -- 134}, publisher = {Springer}, title = {{There is no strongly regular graph with parameters (460; 153; 32; 60)}}, doi = {10.1007/978-3-319-72456-0_7}, year = {2018}, } @article{6354, abstract = {Blood platelets are critical for hemostasis and thrombosis, but also play diverse roles during immune responses. We have recently reported that platelets migrate at sites of infection in vitro and in vivo. Importantly, platelets use their ability to migrate to collect and bundle fibrin (ogen)-bound bacteria accomplishing efficient intravascular bacterial trapping. Here, we describe a method that allows analyzing platelet migration in vitro, focusing on their ability to collect bacteria and trap bacteria under flow.}, author = {Fan, Shuxia and Lorenz, Michael and Massberg, Steffen and Gärtner, Florian R}, issn = {2331-8325}, journal = {Bio-Protocol}, keywords = {Platelets, Cell migration, Bacteria, Shear flow, Fibrinogen, E. coli}, number = {18}, publisher = {Bio-Protocol}, title = {{Platelet migration and bacterial trapping assay under flow}}, doi = {10.21769/bioprotoc.3018}, volume = {8}, year = {2018}, } @misc{6459, author = {Petritsch, Barbara}, keywords = {Open Access, Publication Analysis}, location = {Graz, Austria}, publisher = {IST Austria}, title = {{Open Access at IST Austria 2009-2017}}, doi = {10.5281/zenodo.1410279}, year = {2018}, } @inbook{6525, abstract = {This chapter finds an agreement of equivariant indices of semi-classical homomorphisms between pairwise mirror branes in the GL2 Higgs moduli space on a Riemann surface. On one side of the agreement, components of the Lagrangian brane of U(1,1) Higgs bundles, whose mirror was proposed by Hitchin to be certain even exterior powers of the hyperholomorphic Dirac bundle on the SL2 Higgs moduli space, are present. The agreement arises from a mysterious functional equation. This gives strong computational evidence for Hitchin’s proposal.}, author = {Hausel, Tamás and Mellit, Anton and Pei, Du}, booktitle = {Geometry and Physics: Volume I}, isbn = {9780198802013}, pages = {189--218}, publisher = {Oxford University Press}, title = {{Mirror symmetry with branes by equivariant verlinde formulas}}, doi = {10.1093/oso/9780198802013.003.0009}, year = {2018}, } @article{690, abstract = {We consider spectral properties and the edge universality of sparse random matrices, the class of random matrices that includes the adjacency matrices of the Erdős–Rényi graph model G(N, p). We prove a local law for the eigenvalue density up to the spectral edges. Under a suitable condition on the sparsity, we also prove that the rescaled extremal eigenvalues exhibit GOE Tracy–Widom fluctuations if a deterministic shift of the spectral edge due to the sparsity is included. For the adjacency matrix of the Erdős–Rényi graph this establishes the Tracy–Widom fluctuations of the second largest eigenvalue when p is much larger than N−2/3 with a deterministic shift of order (Np)−1.}, author = {Lee, Jii and Schnelli, Kevin}, journal = {Probability Theory and Related Fields}, number = {1-2}, publisher = {Springer}, title = {{Local law and Tracy–Widom limit for sparse random matrices}}, doi = {10.1007/s00440-017-0787-8}, volume = {171}, year = {2018}, } @article{703, abstract = {We consider the NP-hard problem of MAP-inference for undirected discrete graphical models. We propose a polynomial time and practically efficient algorithm for finding a part of its optimal solution. Specifically, our algorithm marks some labels of the considered graphical model either as (i) optimal, meaning that they belong to all optimal solutions of the inference problem; (ii) non-optimal if they provably do not belong to any solution. With access to an exact solver of a linear programming relaxation to the MAP-inference problem, our algorithm marks the maximal possible (in a specified sense) number of labels. We also present a version of the algorithm, which has access to a suboptimal dual solver only and still can ensure the (non-)optimality for the marked labels, although the overall number of the marked labels may decrease. We propose an efficient implementation, which runs in time comparable to a single run of a suboptimal dual solver. Our method is well-scalable and shows state-of-the-art results on computational benchmarks from machine learning and computer vision.}, author = {Shekhovtsov, Alexander and Swoboda, Paul and Savchynskyy, Bogdan}, issn = {01628828}, journal = {IEEE Transactions on Pattern Analysis and Machine Intelligence}, number = {7}, pages = {1668--1682}, publisher = {IEEE}, title = {{Maximum persistency via iterative relaxed inference with graphical models}}, doi = {10.1109/TPAMI.2017.2730884}, volume = {40}, year = {2018}, } @inproceedings{7116, abstract = {Training deep learning models has received tremendous research interest recently. In particular, there has been intensive research on reducing the communication cost of training when using multiple computational devices, through reducing the precision of the underlying data representation. Naturally, such methods induce system trade-offs—lowering communication precision could de-crease communication overheads and improve scalability; but, on the other hand, it can also reduce the accuracy of training. In this paper, we study this trade-off space, and ask:Can low-precision communication consistently improve the end-to-end performance of training modern neural networks, with no accuracy loss?From the performance point of view, the answer to this question may appear deceptively easy: compressing communication through low precision should help when the ratio between communication and computation is high. However, this answer is less straightforward when we try to generalize this principle across various neural network architectures (e.g., AlexNet vs. ResNet),number of GPUs (e.g., 2 vs. 8 GPUs), machine configurations(e.g., EC2 instances vs. NVIDIA DGX-1), communication primitives (e.g., MPI vs. NCCL), and even different GPU architectures(e.g., Kepler vs. Pascal). Currently, it is not clear how a realistic realization of all these factors maps to the speed up provided by low-precision communication. In this paper, we conduct an empirical study to answer this question and report the insights.}, author = {Grubic, Demjan and Tam, Leo and Alistarh, Dan-Adrian and Zhang, Ce}, booktitle = {Proceedings of the 21st International Conference on Extending Database Technology}, isbn = {9783893180783}, issn = {2367-2005}, location = {Vienna, Austria}, pages = {145--156}, publisher = {OpenProceedings}, title = {{Synchronous multi-GPU training for deep learning with low-precision communications: An empirical study}}, doi = {10.5441/002/EDBT.2018.14}, year = {2018}, } @inproceedings{7407, abstract = {Proofs of space (PoS) [Dziembowski et al., CRYPTO'15] are proof systems where a prover can convince a verifier that he "wastes" disk space. PoS were introduced as a more ecological and economical replacement for proofs of work which are currently used to secure blockchains like Bitcoin. In this work we investigate extensions of PoS which allow the prover to embed useful data into the dedicated space, which later can be recovered. Our first contribution is a security proof for the original PoS from CRYPTO'15 in the random oracle model (the original proof only applied to a restricted class of adversaries which can store a subset of the data an honest prover would store). When this PoS is instantiated with recent constructions of maximally depth robust graphs, our proof implies basically optimal security. As a second contribution we show three different extensions of this PoS where useful data can be embedded into the space required by the prover. Our security proof for the PoS extends (non-trivially) to these constructions. We discuss how some of these variants can be used as proofs of catalytic space (PoCS), a notion we put forward in this work, and which basically is a PoS where most of the space required by the prover can be used to backup useful data. Finally we discuss how one of the extensions is a candidate construction for a proof of replication (PoR), a proof system recently suggested in the Filecoin whitepaper. }, author = {Pietrzak, Krzysztof Z}, booktitle = {10th Innovations in Theoretical Computer Science Conference (ITCS 2019)}, isbn = {978-3-95977-095-8}, issn = {1868-8969}, location = {San Diego, CA, United States}, pages = {59:1--59:25}, publisher = {Schloss Dagstuhl - Leibniz-Zentrum für Informatik}, title = {{Proofs of catalytic space}}, doi = {10.4230/LIPICS.ITCS.2019.59}, volume = {124}, year = {2018}, } @article{6001, abstract = {The concurrent memory reclamation problem is that of devising a way for a deallocating thread to verify that no other concurrent threads hold references to a memory block being deallocated. To date, in the absence of automatic garbage collection, there is no satisfactory solution to this problem; existing tracking methods like hazard pointers, reference counters, or epoch-based techniques like RCU are either prohibitively expensive or require significant programming expertise to the extent that implementing them efficiently can be worthy of a publication. None of the existing techniques are automatic or even semi-automated. In this article, we take a new approach to concurrent memory reclamation. Instead of manually tracking access to memory locations as done in techniques like hazard pointers, or restricting shared accesses to specific epoch boundaries as in RCU, our algorithm, called ThreadScan, leverages operating system signaling to automatically detect which memory locations are being accessed by concurrent threads. Initial empirical evidence shows that ThreadScan scales surprisingly well and requires negligible programming effort beyond the standard use of Malloc and Free.}, author = {Alistarh, Dan-Adrian and Leiserson, William and Matveev, Alexander and Shavit, Nir}, issn = {2329-4949}, journal = {ACM Transactions on Parallel Computing}, number = {4}, publisher = {Association for Computing Machinery}, title = {{ThreadScan: Automatic and scalable memory reclamation}}, doi = {10.1145/3201897}, volume = {4}, year = {2018}, } @inproceedings{7812, abstract = {Deep neural networks (DNNs) continue to make significant advances, solving tasks from image classification to translation or reinforcement learning. One aspect of the field receiving considerable attention is efficiently executing deep models in resource-constrained environments, such as mobile or embedded devices. This paper focuses on this problem, and proposes two new compression methods, which jointly leverage weight quantization and distillation of larger teacher networks into smaller student networks. The first method we propose is called quantized distillation and leverages distillation during the training process, by incorporating distillation loss, expressed with respect to the teacher, into the training of a student network whose weights are quantized to a limited set of levels. The second method, differentiable quantization, optimizes the location of quantization points through stochastic gradient descent, to better fit the behavior of the teacher model. We validate both methods through experiments on convolutional and recurrent architectures. We show that quantized shallow students can reach similar accuracy levels to full-precision teacher models, while providing order of magnitude compression, and inference speedup that is linear in the depth reduction. In sum, our results enable DNNs for resource-constrained environments to leverage architecture and accuracy advances developed on more powerful devices.}, author = {Polino, Antonio and Pascanu, Razvan and Alistarh, Dan-Adrian}, booktitle = {6th International Conference on Learning Representations}, location = {Vancouver, Canada}, title = {{Model compression via distillation and quantization}}, year = {2018}, } @unpublished{8547, abstract = {The cerebral cortex contains multiple hierarchically organized areas with distinctive cytoarchitectonical patterns, but the cellular mechanisms underlying the emergence of this diversity remain unclear. Here, we have quantitatively investigated the neuronal output of individual progenitor cells in the ventricular zone of the developing mouse neocortex using a combination of methods that together circumvent the biases and limitations of individual approaches. We found that individual cortical progenitor cells show a high degree of stochasticity and generate pyramidal cell lineages that adopt a wide range of laminar configurations. Mathematical modelling these lineage data suggests that a small number of progenitor cell populations, each generating pyramidal cells following different stochastic developmental programs, suffice to generate the heterogenous complement of pyramidal cell lineages that collectively build the complex cytoarchitecture of the neocortex.}, author = {Llorca, Alfredo and Ciceri, Gabriele and Beattie, Robert J and Wong, Fong K. and Diana, Giovanni and Serafeimidou, Eleni and Fernández-Otero, Marian and Streicher, Carmen and Arnold, Sebastian J. and Meyer, Martin and Hippenmeyer, Simon and Maravall, Miguel and Marín, Oscar}, booktitle = {bioRxiv}, publisher = {Cold Spring Harbor Laboratory}, title = {{Heterogeneous progenitor cell behaviors underlie the assembly of neocortical cytoarchitecture}}, doi = {10.1101/494088}, year = {2018}, } @inbook{86, abstract = {Responsiveness—the requirement that every request to a system be eventually handled—is one of the fundamental liveness properties of a reactive system. Average response time is a quantitative measure for the responsiveness requirement used commonly in performance evaluation. We show how average response time can be computed on state-transition graphs, on Markov chains, and on game graphs. In all three cases, we give polynomial-time algorithms.}, author = {Chatterjee, Krishnendu and Henzinger, Thomas A and Otop, Jan}, booktitle = {Principles of Modeling}, editor = {Lohstroh, Marten and Derler, Patricia and Sirjani, Marjan}, pages = {143 -- 161}, publisher = {Springer}, title = {{Computing average response time}}, doi = {10.1007/978-3-319-95246-8_9}, volume = {10760}, year = {2018}, } @article{9229, author = {Danzl, Johann G}, issn = {2500-2295}, journal = {Opera Medica et Physiologica}, number = {S1}, pages = {11}, publisher = {Lobachevsky State University of Nizhny Novgorod}, title = {{Diffraction-unlimited optical imaging for synaptic physiology}}, doi = {10.20388/omp2018.00s1.001}, volume = {4}, year = {2018}, } @inproceedings{6005, abstract = {Network games are widely used as a model for selfish resource-allocation problems. In the classicalmodel, each player selects a path connecting her source and target vertices. The cost of traversingan edge depends on theload; namely, number of players that traverse it. Thus, it abstracts the factthat different users may use a resource at different times and for different durations, which playsan important role in determining the costs of the users in reality. For example, when transmittingpackets in a communication network, routing traffic in a road network, or processing a task in aproduction system, actual sharing and congestion of resources crucially depends on time.In [13], we introducedtimed network games, which add a time component to network games.Each vertexvin the network is associated with a cost function, mapping the load onvto theprice that a player pays for staying invfor one time unit with this load. Each edge in thenetwork is guarded by the time intervals in which it can be traversed, which forces the players tospend time in the vertices. In this work we significantly extend the way time can be referred toin timed network games. In the model we study, the network is equipped withclocks, and, as intimed automata, edges are guarded by constraints on the values of the clocks, and their traversalmay involve a reset of some clocks. We argue that the stronger model captures many realisticnetworks. The addition of clocks breaks the techniques we developed in [13] and we developnew techniques in order to show that positive results on classic network games carry over to thestronger timed setting.}, author = {Avni, Guy and Guha, Shibashis and Kupferman, Orna}, issn = {1868-8969}, location = {Liverpool, United Kingdom}, publisher = {Schloss Dagstuhl - Leibniz-Zentrum für Informatik}, title = {{Timed network games with clocks}}, doi = {10.4230/LIPICS.MFCS.2018.23}, volume = {117}, year = {2018}, } @article{315, abstract = {More than 100 years after Grigg’s influential analysis of species’ borders, the causes of limits to species’ ranges still represent a puzzle that has never been understood with clarity. The topic has become especially important recently as many scientists have become interested in the potential for species’ ranges to shift in response to climate change—and yet nearly all of those studies fail to recognise or incorporate evolutionary genetics in a way that relates to theoretical developments. I show that range margins can be understood based on just two measurable parameters: (i) the fitness cost of dispersal—a measure of environmental heterogeneity—and (ii) the strength of genetic drift, which reduces genetic diversity. Together, these two parameters define an ‘expansion threshold’: adaptation fails when genetic drift reduces genetic diversity below that required for adaptation to a heterogeneous environment. When the key parameters drop below this expansion threshold locally, a sharp range margin forms. When they drop below this threshold throughout the species’ range, adaptation collapses everywhere, resulting in either extinction or formation of a fragmented metapopulation. Because the effects of dispersal differ fundamentally with dimension, the second parameter—the strength of genetic drift—is qualitatively different compared to a linear habitat. In two-dimensional habitats, genetic drift becomes effectively independent of selection. It decreases with ‘neighbourhood size’—the number of individuals accessible by dispersal within one generation. Moreover, in contrast to earlier predictions, which neglected evolution of genetic variance and/or stochasticity in two dimensions, dispersal into small marginal populations aids adaptation. This is because the reduction of both genetic and demographic stochasticity has a stronger effect than the cost of dispersal through increased maladaptation. The expansion threshold thus provides a novel, theoretically justified, and testable prediction for formation of the range margin and collapse of the species’ range.}, author = {Polechova, Jitka}, issn = {15449173}, journal = {PLoS Biology}, number = {6}, publisher = {Public Library of Science}, title = {{Is the sky the limit? On the expansion threshold of a species’ range}}, doi = {10.1371/journal.pbio.2005372}, volume = {16}, year = {2018}, } @article{9471, abstract = {The DEMETER (DME) DNA glycosylase catalyzes genome-wide DNA demethylation and is required for endosperm genomic imprinting and embryo viability. Targets of DME-mediated DNA demethylation reside in small, euchromatic, AT-rich transposons and at the boundaries of large transposons, but how DME interacts with these diverse chromatin states is unknown. The STRUCTURE SPECIFIC RECOGNITION PROTEIN 1 (SSRP1) subunit of the chromatin remodeler FACT (facilitates chromatin transactions), was previously shown to be involved in the DME-dependent regulation of genomic imprinting in Arabidopsis endosperm. Therefore, to investigate the interaction between DME and chromatin, we focused on the activity of the two FACT subunits, SSRP1 and SUPPRESSOR of TY16 (SPT16), during reproduction in Arabidopsis. We found that FACT colocalizes with nuclear DME in vivo, and that DME has two classes of target sites, the first being euchromatic and accessible to DME, but the second, representing over half of DME targets, requiring the action of FACT for DME-mediated DNA demethylation genome-wide. Our results show that the FACT-dependent DME targets are GC-rich heterochromatin domains with high nucleosome occupancy enriched with H3K9me2 and H3K27me1. Further, we demonstrate that heterochromatin-associated linker histone H1 specifically mediates the requirement for FACT at a subset of DME-target loci. Overall, our results demonstrate that FACT is required for DME targeting by facilitating its access to heterochromatin.}, author = {Frost, Jennifer M. and Kim, M. Yvonne and Park, Guen Tae and Hsieh, Ping-Hung and Nakamura, Miyuki and Lin, Samuel J. H. and Yoo, Hyunjin and Choi, Jaemyung and Ikeda, Yoko and Kinoshita, Tetsu and Choi, Yeonhee and Zilberman, Daniel and Fischer, Robert L.}, issn = {1091-6490}, journal = {Proceedings of the National Academy of Sciences}, keywords = {Multidisciplinary}, number = {20}, pages = {E4720--E4729}, publisher = {National Academy of Sciences}, title = {{FACT complex is required for DNA demethylation at heterochromatin during reproduction in Arabidopsis}}, doi = {10.1073/pnas.1713333115}, volume = {115}, year = {2018}, } @inproceedings{186, abstract = {A drawing of a graph on a surface is independently even if every pair of nonadjacent edges in the drawing crosses an even number of times. The ℤ2-genus of a graph G is the minimum g such that G has an independently even drawing on the orientable surface of genus g. An unpublished result by Robertson and Seymour implies that for every t, every graph of sufficiently large genus contains as a minor a projective t × t grid or one of the following so-called t-Kuratowski graphs: K3, t, or t copies of K5 or K3,3 sharing at most 2 common vertices. We show that the ℤ2-genus of graphs in these families is unbounded in t; in fact, equal to their genus. Together, this implies that the genus of a graph is bounded from above by a function of its ℤ2-genus, solving a problem posed by Schaefer and Štefankovič, and giving an approximate version of the Hanani-Tutte theorem on orientable surfaces.}, author = {Fulek, Radoslav and Kynčl, Jan}, location = {Budapest, Hungary}, pages = {40.1 -- 40.14}, publisher = {Schloss Dagstuhl - Leibniz-Zentrum für Informatik}, title = {{The ℤ2-Genus of Kuratowski minors}}, doi = {10.4230/LIPIcs.SoCG.2018.40}, volume = {99}, year = {2018}, } @inproceedings{433, abstract = {A thrackle is a graph drawn in the plane so that every pair of its edges meet exactly once: either at a common end vertex or in a proper crossing. We prove that any thrackle of n vertices has at most 1.3984n edges. Quasi-thrackles are defined similarly, except that every pair of edges that do not share a vertex are allowed to cross an odd number of times. It is also shown that the maximum number of edges of a quasi-thrackle on n vertices is 3/2(n-1), and that this bound is best possible for infinitely many values of n.}, author = {Fulek, Radoslav and Pach, János}, location = {Boston, MA, United States}, pages = {160 -- 166}, publisher = {Springer}, title = {{Thrackles: An improved upper bound}}, doi = {10.1007/978-3-319-73915-1_14}, volume = {10692}, year = {2018}, } @misc{9837, abstract = {Both classical and recent studies suggest that chromosomal inversion polymorphisms are important in adaptation and speciation. However, biases in discovery and reporting of inversions make it difficult to assess their prevalence and biological importance. Here, we use an approach based on linkage disequilibrium among markers genotyped for samples collected across a transect between contrasting habitats to detect chromosomal rearrangements de novo. We report 17 polymorphic rearrangements in a single locality for the coastal marine snail, Littorina saxatilis. Patterns of diversity in the field and of recombination in controlled crosses provide strong evidence that at least the majority of these rearrangements are inversions. Most show clinal changes in frequency between habitats, suggestive of divergent selection, but only one appears to be fixed for different arrangements in the two habitats. Consistent with widespread evidence for balancing selection on inversion polymorphisms, we argue that a combination of heterosis and divergent selection can explain the observed patterns and should be considered in other systems spanning environmental gradients.}, author = {Faria, Rui and Chaube, Pragya and Morales, Hernán E. and Larsson, Tomas and Lemmon, Alan R. and Lemmon, Emily M. and Rafajlović, Marina and Panova, Marina and Ravinet, Mark and Johannesson, Kerstin and Westram, Anja M and Butlin, Roger K.}, publisher = {Dryad}, title = {{Data from: Multiple chromosomal rearrangements in a hybrid zone between Littorina saxatilis ecotypes}}, doi = {10.5061/dryad.72cg113}, year = {2018}, } @inbook{10864, abstract = {We prove that every congruence distributive variety has directed Jónsson terms, and every congruence modular variety has directed Gumm terms. The directed terms we construct witness every case of absorption witnessed by the original Jónsson or Gumm terms. This result is equivalent to a pair of claims about absorption for admissible preorders in congruence distributive and congruence modular varieties, respectively. For finite algebras, these absorption theorems have already seen significant applications, but until now, it was not clear if the theorems hold for general algebras as well. Our method also yields a novel proof of a result by P. Lipparini about the existence of a chain of terms (which we call Pixley terms) in varieties that are at the same time congruence distributive and k-permutable for some k.}, author = {Kazda, Alexandr and Kozik, Marcin and McKenzie, Ralph and Moore, Matthew}, booktitle = {Don Pigozzi on Abstract Algebraic Logic, Universal Algebra, and Computer Science}, editor = {Czelakowski, J}, isbn = {9783319747712}, issn = {2211-2766}, pages = {203--220}, publisher = {Springer Nature}, title = {{Absorption and directed Jónsson terms}}, doi = {10.1007/978-3-319-74772-9_7}, volume = {16}, year = {2018}, } @inproceedings{184, abstract = {We prove that for every d ≥ 2, deciding if a pure, d-dimensional, simplicial complex is shellable is NP-hard, hence NP-complete. This resolves a question raised, e.g., by Danaraj and Klee in 1978. Our reduction also yields that for every d ≥ 2 and k ≥ 0, deciding if a pure, d-dimensional, simplicial complex is k-decomposable is NP-hard. For d ≥ 3, both problems remain NP-hard when restricted to contractible pure d-dimensional complexes.}, author = {Goaoc, Xavier and Paták, Pavel and Patakova, Zuzana and Tancer, Martin and Wagner, Uli}, location = {Budapest, Hungary}, pages = {41:1 -- 41:16}, publisher = {Schloss Dagstuhl - Leibniz-Zentrum für Informatik}, title = {{Shellability is NP-complete}}, doi = {10.4230/LIPIcs.SoCG.2018.41}, volume = {99}, year = {2018}, } @inproceedings{285, abstract = {In graph theory, as well as in 3-manifold topology, there exist several width-type parameters to describe how "simple" or "thin" a given graph or 3-manifold is. These parameters, such as pathwidth or treewidth for graphs, or the concept of thin position for 3-manifolds, play an important role when studying algorithmic problems; in particular, there is a variety of problems in computational 3-manifold topology - some of them known to be computationally hard in general - that become solvable in polynomial time as soon as the dual graph of the input triangulation has bounded treewidth. In view of these algorithmic results, it is natural to ask whether every 3-manifold admits a triangulation of bounded treewidth. We show that this is not the case, i.e., that there exists an infinite family of closed 3-manifolds not admitting triangulations of bounded pathwidth or treewidth (the latter implies the former, but we present two separate proofs). We derive these results from work of Agol and of Scharlemann and Thompson, by exhibiting explicit connections between the topology of a 3-manifold M on the one hand and width-type parameters of the dual graphs of triangulations of M on the other hand, answering a question that had been raised repeatedly by researchers in computational 3-manifold topology. In particular, we show that if a closed, orientable, irreducible, non-Haken 3-manifold M has a triangulation of treewidth (resp. pathwidth) k then the Heegaard genus of M is at most 48(k+1) (resp. 4(3k+1)).}, author = {Huszár, Kristóf and Spreer, Jonathan and Wagner, Uli}, issn = {18688969}, location = {Budapest, Hungary}, publisher = {Schloss Dagstuhl - Leibniz-Zentrum für Informatik}, title = {{On the treewidth of triangulated 3-manifolds}}, doi = {10.4230/LIPIcs.SoCG.2018.46}, volume = {99}, year = {2018}, } @misc{13059, abstract = {This dataset contains a GitHub repository containing all the data, analysis, Nextflow workflows and Jupyter notebooks to replicate the manuscript titled "Fast and accurate large multiple sequence alignments with a root-to-leaf regressive method". It also contains the Multiple Sequence Alignments (MSAs) generated and well as the main figures and tables from the manuscript. The repository is also available at GitHub (https://github.com/cbcrg/dpa-analysis) release `v1.2`. For details on how to use the regressive alignment algorithm, see the T-Coffee software suite (https://github.com/cbcrg/tcoffee).}, author = {Garriga, Edgar and di Tommaso, Paolo and Magis, Cedrik and Erb, Ionas and Mansouri, Leila and Baltzis, Athanasios and Laayouni, Hafid and Kondrashov, Fyodor and Floden, Evan and Notredame, Cedric}, publisher = {Zenodo}, title = {{Fast and accurate large multiple sequence alignments with a root-to-leaf regressive method}}, doi = {10.5281/ZENODO.2025846}, year = {2018}, } @phdthesis{49, abstract = {Nowadays, quantum computation is receiving more and more attention as an alternative to the classical way of computing. For realizing a quantum computer, different devices are investigated as potential quantum bits. In this thesis, the focus is on Ge hut wires, which turned out to be promising candidates for implementing hole spin quantum bits. The advantages of Ge as a material system are the low hyperfine interaction for holes and the strong spin orbit coupling, as well as the compatibility with the highly developed CMOS processes in industry. In addition, Ge can also be isotopically purified which is expected to boost the spin coherence times. The strong spin orbit interaction for holes in Ge on the one hand enables the full electrical control of the quantum bit and on the other hand should allow short spin manipulation times. Starting with a bare Si wafer, this work covers the entire process reaching from growth over the fabrication and characterization of hut wire devices up to the demonstration of hole spin resonance. From experiments with single quantum dots, a large g-factor anisotropy between the in-plane and the out-of-plane direction was found. A comparison to a theoretical model unveiled the heavy-hole character of the lowest energy states. The second part of the thesis addresses double quantum dot devices, which were realized by adding two gate electrodes to a hut wire. In such devices, Pauli spin blockade was observed, which can serve as a read-out mechanism for spin quantum bits. Applying oscillating electric fields in spin blockade allowed the demonstration of continuous spin rotations and the extraction of a lower bound for the spin dephasing time. Despite the strong spin orbit coupling in Ge, the obtained value for the dephasing time is comparable to what has been recently reported for holes in Si. All in all, the presented results point out the high potential of Ge hut wires as a platform for long-lived, fast and fully electrically tunable hole spin quantum bits.}, author = {Watzinger, Hannes}, issn = {2663-337X}, pages = {77}, publisher = {Institute of Science and Technology Austria}, title = {{Ge hut wires - from growth to hole spin resonance}}, doi = {10.15479/AT:ISTA:th_1033}, year = {2018}, } @phdthesis{201, abstract = {We describe arrangements of three-dimensional spheres from a geometrical and topological point of view. Real data (fitting this setup) often consist of soft spheres which show certain degree of deformation while strongly packing against each other. In this context, we answer the following questions: If we model a soft packing of spheres by hard spheres that are allowed to overlap, can we measure the volume in the overlapped areas? Can we be more specific about the overlap volume, i.e. quantify how much volume is there covered exactly twice, three times, or k times? What would be a good optimization criteria that rule the arrangement of soft spheres while making a good use of the available space? Fixing a particular criterion, what would be the optimal sphere configuration? The first result of this thesis are short formulas for the computation of volumes covered by at least k of the balls. The formulas exploit information contained in the order-k Voronoi diagrams and its closely related Level-k complex. The used complexes lead to a natural generalization into poset diagrams, a theoretical formalism that contains the order-k and degree-k diagrams as special cases. In parallel, we define different criteria to determine what could be considered an optimal arrangement from a geometrical point of view. Fixing a criterion, we find optimal soft packing configurations in 2D and 3D where the ball centers lie on a lattice. As a last step, we use tools from computational topology on real physical data, to show the potentials of higher-order diagrams in the description of melting crystals. The results of the experiments leaves us with an open window to apply the theories developed in this thesis in real applications.}, author = {Iglesias Ham, Mabel}, issn = {2663-337X}, pages = {171}, publisher = {Institute of Science and Technology Austria}, title = {{Multiple covers with balls}}, doi = {10.15479/AT:ISTA:th_1026}, year = {2018}, } @phdthesis{68, abstract = {The most common assumption made in statistical learning theory is the assumption of the independent and identically distributed (i.i.d.) data. While being very convenient mathematically, it is often very clearly violated in practice. This disparity between the machine learning theory and applications underlies a growing demand in the development of algorithms that learn from dependent data and theory that can provide generalization guarantees similar to the independent situations. This thesis is dedicated to two variants of dependencies that can arise in practice. One is a dependence on the level of samples in a single learning task. Another dependency type arises in the multi-task setting when the tasks are dependent on each other even though the data for them can be i.i.d. In both cases we model the data (samples or tasks) as stochastic processes and introduce new algorithms for both settings that take into account and exploit the resulting dependencies. We prove the theoretical guarantees on the performance of the introduced algorithms under different evaluation criteria and, in addition, we compliment the theoretical study by the empirical one, where we evaluate some of the algorithms on two real world datasets to highlight their practical applicability.}, author = {Zimin, Alexander}, issn = {2663-337X}, pages = {92}, publisher = {Institute of Science and Technology Austria}, title = {{Learning from dependent data}}, doi = {10.15479/AT:ISTA:TH1048}, year = {2018}, } @phdthesis{83, abstract = {A proof system is a protocol between a prover and a verifier over a common input in which an honest prover convinces the verifier of the validity of true statements. Motivated by the success of decentralized cryptocurrencies, exemplified by Bitcoin, the focus of this thesis will be on proof systems which found applications in some sustainable alternatives to Bitcoin, such as the Spacemint and Chia cryptocurrencies. In particular, we focus on proofs of space and proofs of sequential work. Proofs of space (PoSpace) were suggested as more ecological, economical, and egalitarian alternative to the energy-wasteful proof-of-work mining of Bitcoin. However, the state-of-the-art constructions of PoSpace are based on sophisticated graph pebbling lower bounds, and are therefore complex. Moreover, when these PoSpace are used in cryptocurrencies like Spacemint, miners can only start mining after ensuring that a commitment to their space is already added in a special transaction to the blockchain. Proofs of sequential work (PoSW) are proof systems in which a prover, upon receiving a statement x and a time parameter T, computes a proof which convinces the verifier that T time units had passed since x was received. Whereas Spacemint assumes synchrony to retain some interesting Bitcoin dynamics, Chia requires PoSW with unique proofs, i.e., PoSW in which it is hard to come up with more than one accepting proof for any true statement. In this thesis we construct simple and practically-efficient PoSpace and PoSW. When using our PoSpace in cryptocurrencies, miners can start mining on the fly, like in Bitcoin, and unlike current constructions of PoSW, which either achieve efficient verification of sequential work, or faster-than-recomputing verification of correctness of proofs, but not both at the same time, ours achieve the best of these two worlds.}, author = {Abusalah, Hamza M}, issn = {2663-337X}, pages = {59}, publisher = {Institute of Science and Technology Austria}, title = {{Proof systems for sustainable decentralized cryptocurrencies}}, doi = {10.15479/AT:ISTA:TH_1046}, year = {2018}, }