Project acronym 2DNanoSpec
Project Nanoscale Vibrational Spectroscopy of Sensitive 2D Molecular Materials
Researcher (PI) Renato ZENOBI
Host Institution (HI) EIDGENOESSISCHE TECHNISCHE HOCHSCHULE ZUERICH
Call Details Advanced Grant (AdG), PE4, ERC-2016-ADG
Summary I propose to investigate the nanometer scale organization of delicate 2-dimensional molecular materials using nanoscale vibrational spectroscopy. 2D structures are of great scientific and technological importance, for example as novel materials (graphene, MoS2, WS2, etc.), and in the form of biological membranes and synthetic 2D-polymers. Powerful methods for their analysis and imaging with molecular selectivity and sufficient spatial resolution, however, are lacking. Tip-enhanced Raman spectroscopy (TERS) allows label-free spectroscopic identification of molecular species, with ≈10 nm spatial resolution, and with single molecule sensitivity for strong Raman scatterers. So far, however, TERS is not being carried out in liquids, which is the natural environment for membranes, and its application to poor Raman scatterers such as components of 2D polymers, lipids, or other membrane compounds (proteins, sugars) is difficult. TERS has the potential to overcome the restrictions of other optical/spectroscopic methods to study 2D materials, namely (i) insufficient spatial resolution of diffraction-limited optical methods; (ii) the need for labelling for all methods relying on fluorescence; and (iii) the inability of some methods to work in liquids. I propose to address a number of scientific questions associated with the spatial organization, and the occurrence of defects in sensitive 2D molecular materials. The success of these studies will also rely critically on technical innovations of TERS that notably address the problem of energy dissipation. This will for the first time allow its application to study of complex, delicate 2D molecular systems without photochemical damage.
Summary
I propose to investigate the nanometer scale organization of delicate 2-dimensional molecular materials using nanoscale vibrational spectroscopy. 2D structures are of great scientific and technological importance, for example as novel materials (graphene, MoS2, WS2, etc.), and in the form of biological membranes and synthetic 2D-polymers. Powerful methods for their analysis and imaging with molecular selectivity and sufficient spatial resolution, however, are lacking. Tip-enhanced Raman spectroscopy (TERS) allows label-free spectroscopic identification of molecular species, with ≈10 nm spatial resolution, and with single molecule sensitivity for strong Raman scatterers. So far, however, TERS is not being carried out in liquids, which is the natural environment for membranes, and its application to poor Raman scatterers such as components of 2D polymers, lipids, or other membrane compounds (proteins, sugars) is difficult. TERS has the potential to overcome the restrictions of other optical/spectroscopic methods to study 2D materials, namely (i) insufficient spatial resolution of diffraction-limited optical methods; (ii) the need for labelling for all methods relying on fluorescence; and (iii) the inability of some methods to work in liquids. I propose to address a number of scientific questions associated with the spatial organization, and the occurrence of defects in sensitive 2D molecular materials. The success of these studies will also rely critically on technical innovations of TERS that notably address the problem of energy dissipation. This will for the first time allow its application to study of complex, delicate 2D molecular systems without photochemical damage.
Max ERC Funding
2 311 696 €
Duration
Start date: 2017-09-01, End date: 2022-08-31
Project acronym ALGILE
Project Foundations of Algebraic and Dynamic Data Management Systems
Researcher (PI) Christoph Koch
Host Institution (HI) ECOLE POLYTECHNIQUE FEDERALE DE LAUSANNE
Call Details Starting Grant (StG), PE6, ERC-2011-StG_20101014
Summary "Contemporary database query languages are ultimately founded on logic and feature an additive operation – usually a form of (multi)set union or disjunction – that is asymmetric in that additions or updates do not always have an inverse. This asymmetry puts a greater part of the machinery of abstract algebra for equation solving outside the reach of databases. However, such equation solving would be a key functionality that problems such as query equivalence testing and data integration could be reduced to: In the current scenario of the presence of an asymmetric additive operation they are undecidable. Moreover, query languages with a symmetric additive operation (i.e., which has an inverse and is thus based on ring theory) would open up databases for a large range of new scientific and mathematical applications.
The goal of the proposed project is to reinvent database management systems with a foundation in abstract algebra and specifically in ring theory. The presence of an additive inverse allows to cleanly define differences between queries. This gives rise to a database analog of differential calculus that leads to radically new incremental and adaptive query evaluation algorithms that substantially outperform the state of the art techniques. These algorithms enable a new class of systems which I call Dynamic Data Management Systems. Such systems can maintain continuously fresh query views at extremely high update rates and have important applications in interactive Large-scale Data Analysis. There is a natural connection between differences and updates, motivating the group theoretic study of updates that will lead to better ways of creating out-of-core data processing algorithms for new storage devices. Basing queries on ring theory leads to a new class of systems, Algebraic Data Management Systems, which herald a convergence of database systems and computer algebra systems."
Summary
"Contemporary database query languages are ultimately founded on logic and feature an additive operation – usually a form of (multi)set union or disjunction – that is asymmetric in that additions or updates do not always have an inverse. This asymmetry puts a greater part of the machinery of abstract algebra for equation solving outside the reach of databases. However, such equation solving would be a key functionality that problems such as query equivalence testing and data integration could be reduced to: In the current scenario of the presence of an asymmetric additive operation they are undecidable. Moreover, query languages with a symmetric additive operation (i.e., which has an inverse and is thus based on ring theory) would open up databases for a large range of new scientific and mathematical applications.
The goal of the proposed project is to reinvent database management systems with a foundation in abstract algebra and specifically in ring theory. The presence of an additive inverse allows to cleanly define differences between queries. This gives rise to a database analog of differential calculus that leads to radically new incremental and adaptive query evaluation algorithms that substantially outperform the state of the art techniques. These algorithms enable a new class of systems which I call Dynamic Data Management Systems. Such systems can maintain continuously fresh query views at extremely high update rates and have important applications in interactive Large-scale Data Analysis. There is a natural connection between differences and updates, motivating the group theoretic study of updates that will lead to better ways of creating out-of-core data processing algorithms for new storage devices. Basing queries on ring theory leads to a new class of systems, Algebraic Data Management Systems, which herald a convergence of database systems and computer algebra systems."
Max ERC Funding
1 480 548 €
Duration
Start date: 2012-01-01, End date: 2016-12-31
Project acronym AlgoRNN
Project Recurrent Neural Networks and Related Machines That Learn Algorithms
Researcher (PI) Juergen Schmidhuber
Host Institution (HI) UNIVERSITA DELLA SVIZZERA ITALIANA
Call Details Advanced Grant (AdG), PE6, ERC-2016-ADG
Summary Recurrent neural networks (RNNs) are general parallel-sequential computers. Some learn their programs or weights. Our supervised Long Short-Term Memory (LSTM) RNNs were the first to win pattern recognition contests, and recently enabled best known results in speech and handwriting recognition, machine translation, etc. They are now available to billions of users through the world's most valuable public companies including Google and Apple. Nevertheless, in lots of real-world tasks RNNs do not yet live up to their full potential. Although universal in theory, in practice they fail to learn important types of algorithms. This ERC project will go far beyond today's best RNNs through novel RNN-like systems that address some of the biggest open RNN problems and hottest RNN research topics: (1) How can RNNs learn to control (through internal spotlights of attention) separate large short-memory structures such as sub-networks with fast weights, to improve performance on many natural short-term memory-intensive tasks which are currently hard to learn by RNNs, such as answering detailed questions on recently observed videos? (2) How can such RNN-like systems metalearn entire learning algorithms that outperform the original learning algorithms? (3) How to achieve efficient transfer learning from one RNN-learned set of problem-solving programs to new RNN programs solving new tasks? In other words, how can one RNN-like system actively learn to exploit algorithmic information contained in the programs running on another? We will test our systems existing benchmarks, and create new, more challenging multi-task benchmarks. This will be supported by a rather cheap, GPU-based mini-brain for implementing large RNNs.
Summary
Recurrent neural networks (RNNs) are general parallel-sequential computers. Some learn their programs or weights. Our supervised Long Short-Term Memory (LSTM) RNNs were the first to win pattern recognition contests, and recently enabled best known results in speech and handwriting recognition, machine translation, etc. They are now available to billions of users through the world's most valuable public companies including Google and Apple. Nevertheless, in lots of real-world tasks RNNs do not yet live up to their full potential. Although universal in theory, in practice they fail to learn important types of algorithms. This ERC project will go far beyond today's best RNNs through novel RNN-like systems that address some of the biggest open RNN problems and hottest RNN research topics: (1) How can RNNs learn to control (through internal spotlights of attention) separate large short-memory structures such as sub-networks with fast weights, to improve performance on many natural short-term memory-intensive tasks which are currently hard to learn by RNNs, such as answering detailed questions on recently observed videos? (2) How can such RNN-like systems metalearn entire learning algorithms that outperform the original learning algorithms? (3) How to achieve efficient transfer learning from one RNN-learned set of problem-solving programs to new RNN programs solving new tasks? In other words, how can one RNN-like system actively learn to exploit algorithmic information contained in the programs running on another? We will test our systems existing benchmarks, and create new, more challenging multi-task benchmarks. This will be supported by a rather cheap, GPU-based mini-brain for implementing large RNNs.
Max ERC Funding
2 500 000 €
Duration
Start date: 2017-10-01, End date: 2022-09-30
Project acronym AMSEL
Project Atomic Force Microscopy for Molecular Structure Elucidation
Researcher (PI) Leo Gross
Host Institution (HI) IBM RESEARCH GMBH
Call Details Consolidator Grant (CoG), PE4, ERC-2015-CoG
Summary Molecular structure elucidation is of great importance in synthetic chemistry, pharmacy, life sciences, energy and environmental sciences, and technology applications. To date structure elucidation by atomic force microscopy (AFM) has been demonstrated for a few, small and mainly planar molecules. In this project high-risk, high-impact scientific questions will be solved using structure elucidation with the AFM employing a novel tool and novel methodologies.
A combined low-temperature scanning tunneling microscope/atomic force microscope (LT-STM/AFM) with high throughput and in situ electrospray deposition method will be developed. Chemical resolution will be achieved by novel measurement techniques, in particular the usage of different and novel tip functionalizations and combination with Kelvin probe force microscopy. Elements will be identified using substructure recognition provided by a database that will be erected and by refined theory and simulations.
The developed tools and techniques will be applied to molecules of increasing fragility, complexity, size, and three-dimensionality. In particular samples that are challenging to characterize with conventional methods will be studied. Complex molecular mixtures will be investigated molecule-by-molecule taking advantage of the single-molecule sensitivity. The absolute stereochemistry of molecules will be determined, resolving molecules with multiple stereocenters. The operation of single molecular machines as nanocars and molecular gears will be investigated. Reactive intermediates generated with atomic manipulation will be characterized and their on-surface reactivity will be studied by AFM.
Summary
Molecular structure elucidation is of great importance in synthetic chemistry, pharmacy, life sciences, energy and environmental sciences, and technology applications. To date structure elucidation by atomic force microscopy (AFM) has been demonstrated for a few, small and mainly planar molecules. In this project high-risk, high-impact scientific questions will be solved using structure elucidation with the AFM employing a novel tool and novel methodologies.
A combined low-temperature scanning tunneling microscope/atomic force microscope (LT-STM/AFM) with high throughput and in situ electrospray deposition method will be developed. Chemical resolution will be achieved by novel measurement techniques, in particular the usage of different and novel tip functionalizations and combination with Kelvin probe force microscopy. Elements will be identified using substructure recognition provided by a database that will be erected and by refined theory and simulations.
The developed tools and techniques will be applied to molecules of increasing fragility, complexity, size, and three-dimensionality. In particular samples that are challenging to characterize with conventional methods will be studied. Complex molecular mixtures will be investigated molecule-by-molecule taking advantage of the single-molecule sensitivity. The absolute stereochemistry of molecules will be determined, resolving molecules with multiple stereocenters. The operation of single molecular machines as nanocars and molecular gears will be investigated. Reactive intermediates generated with atomic manipulation will be characterized and their on-surface reactivity will be studied by AFM.
Max ERC Funding
2 000 000 €
Duration
Start date: 2016-06-01, End date: 2021-05-31
Project acronym AOC
Project Adversary-Oriented Computing
Researcher (PI) Rachid Guerraoui
Host Institution (HI) ECOLE POLYTECHNIQUE FEDERALE DE LAUSANNE
Call Details Advanced Grant (AdG), PE6, ERC-2013-ADG
Summary "Recent technological evolutions, including the cloud, the multicore, the social and the mobiles ones, are turning computing ubiquitously distributed. Yet, building high-assurance distributed programs is notoriously challenging. One of the main reasons is that these systems usually seek to achieve several goals at the same time. In short, they need to be efficient, responding effectively in various average-case conditions, as well as reliable, behaving correctly in severe, worst-case conditions. As a consequence, they typically intermingle different strategies: each to cope with some specific condition, e.g., with or without node failures, message losses, time-outs, contention, cache misses,
over-sizing, malicious attacks, etc. The resulting programs end up hard to design, prove, verify, implement, test and debug. Not surprisingly, there are anecdotal evidences of the fragility of the most celebrated distributed systems.
The goal of this project is to contribute to building high-assurance distributed programs by introducing a new dimension for separating and isolating their concerns, as well as a new scheme for composing and reusing them in a modular manner. In short, the project will explore the inherent power and limitations of a novel paradigm, Adversary-Oriented Computing (AOC). Sub-programs, each implementing a specific strategy to cope with a given adversary, modelling a specific working condition, are designed, proved, verified, implemented, tested and debugged independently. They are then composed, possibly dynamically, as black-boxes within the same global program. The AOC project is ambitious and it seeks to fundamentally revisit the way distributed algorithms are designed and distributed systems are implemented. The gain expected in comparison with today's approaches is substantial, and I believe it will be proportional to the degree of difficulty of the distributed problem at hand."
Summary
"Recent technological evolutions, including the cloud, the multicore, the social and the mobiles ones, are turning computing ubiquitously distributed. Yet, building high-assurance distributed programs is notoriously challenging. One of the main reasons is that these systems usually seek to achieve several goals at the same time. In short, they need to be efficient, responding effectively in various average-case conditions, as well as reliable, behaving correctly in severe, worst-case conditions. As a consequence, they typically intermingle different strategies: each to cope with some specific condition, e.g., with or without node failures, message losses, time-outs, contention, cache misses,
over-sizing, malicious attacks, etc. The resulting programs end up hard to design, prove, verify, implement, test and debug. Not surprisingly, there are anecdotal evidences of the fragility of the most celebrated distributed systems.
The goal of this project is to contribute to building high-assurance distributed programs by introducing a new dimension for separating and isolating their concerns, as well as a new scheme for composing and reusing them in a modular manner. In short, the project will explore the inherent power and limitations of a novel paradigm, Adversary-Oriented Computing (AOC). Sub-programs, each implementing a specific strategy to cope with a given adversary, modelling a specific working condition, are designed, proved, verified, implemented, tested and debugged independently. They are then composed, possibly dynamically, as black-boxes within the same global program. The AOC project is ambitious and it seeks to fundamentally revisit the way distributed algorithms are designed and distributed systems are implemented. The gain expected in comparison with today's approaches is substantial, and I believe it will be proportional to the degree of difficulty of the distributed problem at hand."
Max ERC Funding
2 147 012 €
Duration
Start date: 2014-06-01, End date: 2019-05-31
Project acronym ATMOPACS
Project Atmospheric Organic Particulate Matter, Air Quality and Climate Change Studies
Researcher (PI) Spyridon Pandis
Host Institution (HI) FOUNDATION FOR RESEARCH AND TECHNOLOGY HELLAS
Call Details Advanced Grant (AdG), PE10, ERC-2010-AdG_20100224
Summary Despite its importance for human health and climate change organic aerosol (OA) remains one of the least understood aspects of atmospheric chemistry. We propose to develop an innovative new framework for the description of OA in chemical transport and climate models that will be able to overcome the challenges posed by the chemical complexity of OA while capturing its essential features.
The objectives of ATMOPACS are: (i) The development of a new unified framework for the description of OA based on its two most important parameters: volatility and oxygen content. (ii) The development of measurement techniques for the volatility distribution and oxygen content distribution of OA. This will allow the experimental characterization of OA in this new “coordinate system”. (iii) The study of the major OA processes (partitioning, chemical aging, hygroscopicity, CCN formation, nucleation) in this new framework combining lab and field measurements. (iv) The development and evaluation of the next generation of regional and global CTMs using the above framework. (v) The quantification of the importance of the various sources and formation pathways of OA in Europe and the world, of the sensitivity of OA to emission control strategies, and its role in the direct and indirect effects of aerosols on climate.
The proposed work involves a combination of laboratory measurements, field measurements including novel “atmospheric perturbation experiments”, OA model development, and modelling in urban, regional, and global scales. Therefore, it will span the system scales starting from the nanoscale to the global. The modelling tools that will be developed will be made available to all other research groups.
Summary
Despite its importance for human health and climate change organic aerosol (OA) remains one of the least understood aspects of atmospheric chemistry. We propose to develop an innovative new framework for the description of OA in chemical transport and climate models that will be able to overcome the challenges posed by the chemical complexity of OA while capturing its essential features.
The objectives of ATMOPACS are: (i) The development of a new unified framework for the description of OA based on its two most important parameters: volatility and oxygen content. (ii) The development of measurement techniques for the volatility distribution and oxygen content distribution of OA. This will allow the experimental characterization of OA in this new “coordinate system”. (iii) The study of the major OA processes (partitioning, chemical aging, hygroscopicity, CCN formation, nucleation) in this new framework combining lab and field measurements. (iv) The development and evaluation of the next generation of regional and global CTMs using the above framework. (v) The quantification of the importance of the various sources and formation pathways of OA in Europe and the world, of the sensitivity of OA to emission control strategies, and its role in the direct and indirect effects of aerosols on climate.
The proposed work involves a combination of laboratory measurements, field measurements including novel “atmospheric perturbation experiments”, OA model development, and modelling in urban, regional, and global scales. Therefore, it will span the system scales starting from the nanoscale to the global. The modelling tools that will be developed will be made available to all other research groups.
Max ERC Funding
2 496 000 €
Duration
Start date: 2011-01-01, End date: 2015-12-31
Project acronym ATTOLIQ
Project Attosecond X-ray spectroscopy of liquids
Researcher (PI) Hans Jakob WÖRNER
Host Institution (HI) EIDGENOESSISCHE TECHNISCHE HOCHSCHULE ZUERICH
Call Details Consolidator Grant (CoG), PE4, ERC-2017-COG
Summary Charge and energy transfer are the key steps underlying most chemical reactions and biological transformations. The purely electronic dynamics that control such processes take place on attosecond time scales. A complete understanding of these dynamics on the electronic level therefore calls for new experimental methods with attosecond resolution that are applicable to aqueous environments. We propose to combine the element sensitivity of X-ray spectroscopy with attosecond temporal resolution and ultrathin liquid microjets to study electronic dynamics of relevance to chemical, biological and photovoltaic processes. We will build on our recent achievements in demonstrating femtosecond time-resolved measurements in the water, attosecond pho-toelectron spectroscopy on a liquid microjet and measuring and controlling attosecond charge migration in isolated molecules. We will first concentrate on liquid water to study its electronic dynamics following outer-valence ionization, the formation pathway of the solvated electron and the time scales and intermolecular Coulombic decay following inner-valence or core-level ionization. Second, we will turn to solvated species and measure electronic dynamics and charge migration in solvated molecules, transition-metal complexes and pho-toexcited nanoparticles. These goals will be achieved by developing several innovative experimental tech-niques. We will develop a source of isolated attosecond pulses covering the water window (285-538 eV) and combine it with a flat liquid microjet to realize attosecond transient absorption in liquids. We will complement these measurements with attosecond X-ray emission spectroscopy, Auger spectroscopy and a novel hetero-dyne-detected variant of resonant inelastic Raman scattering, exploiting the large bandwidth that is naturally available from attosecond X-ray sources.
Summary
Charge and energy transfer are the key steps underlying most chemical reactions and biological transformations. The purely electronic dynamics that control such processes take place on attosecond time scales. A complete understanding of these dynamics on the electronic level therefore calls for new experimental methods with attosecond resolution that are applicable to aqueous environments. We propose to combine the element sensitivity of X-ray spectroscopy with attosecond temporal resolution and ultrathin liquid microjets to study electronic dynamics of relevance to chemical, biological and photovoltaic processes. We will build on our recent achievements in demonstrating femtosecond time-resolved measurements in the water, attosecond pho-toelectron spectroscopy on a liquid microjet and measuring and controlling attosecond charge migration in isolated molecules. We will first concentrate on liquid water to study its electronic dynamics following outer-valence ionization, the formation pathway of the solvated electron and the time scales and intermolecular Coulombic decay following inner-valence or core-level ionization. Second, we will turn to solvated species and measure electronic dynamics and charge migration in solvated molecules, transition-metal complexes and pho-toexcited nanoparticles. These goals will be achieved by developing several innovative experimental tech-niques. We will develop a source of isolated attosecond pulses covering the water window (285-538 eV) and combine it with a flat liquid microjet to realize attosecond transient absorption in liquids. We will complement these measurements with attosecond X-ray emission spectroscopy, Auger spectroscopy and a novel hetero-dyne-detected variant of resonant inelastic Raman scattering, exploiting the large bandwidth that is naturally available from attosecond X-ray sources.
Max ERC Funding
2 750 000 €
Duration
Start date: 2018-04-01, End date: 2023-03-31
Project acronym ATTOSCOPE
Project Measuring attosecond electron dynamics in molecules
Researcher (PI) Hans Jakob Wörner
Host Institution (HI) EIDGENOESSISCHE TECHNISCHE HOCHSCHULE ZUERICH
Call Details Starting Grant (StG), PE4, ERC-2012-StG_20111012
Summary "The goal of the present proposal is to realize measurements of electronic dynamics in polyatomic
molecules with attosecond temporal resolution (1 as = 10^-18s). We propose to study electronic
rearrangements following photoexcitation, charge migration in a molecular chain induced by
ionization and non-adiabatic multi-electron dynamics in an intense laser field. The grand question
addressed by this research is the characterization of electron correlations which control the shape, properties and function of molecules. In all three proposed projects, a time-domain approach appears to be the most suitable since it reduces complex molecular dynamics to the purely electronic dynamics by exploiting the hierarchy of motional time scales. Experimentally, we propose to realize an innovative experimental setup. A few-cycle infrared (IR) pulse will be used to generate attosecond pulses in the extreme-ultraviolet (XUV) by high-harmonic generation. The IR pulse will be separated from the XUV by means of an innovative interferometer. Additionally, it will permit the introduction of a controlled attosecond delay between the two pulses. We propose to use the attosecond pulses as a tool to look inside individual IR- or UV-field cycles to better understand light-matter interactions. Time-resolved pump-probe experiments will be carried out on polyatomic molecules by detecting the energy and angular distribution of photoelectrons in a velocity-map imaging spectrometer. These experiments are expected to provide new insights
into the dynamics of multi-electron systems along with new results for the validation and
improvement of theoretical models. Multi-electron dynamics is indeed a very complex subject
on its own and even more so in the presence of strong laser fields. The proposed experiments
directly address theses challenges and are expected to provide new insights that will be beneficial to a wide range of scientific research areas."
Summary
"The goal of the present proposal is to realize measurements of electronic dynamics in polyatomic
molecules with attosecond temporal resolution (1 as = 10^-18s). We propose to study electronic
rearrangements following photoexcitation, charge migration in a molecular chain induced by
ionization and non-adiabatic multi-electron dynamics in an intense laser field. The grand question
addressed by this research is the characterization of electron correlations which control the shape, properties and function of molecules. In all three proposed projects, a time-domain approach appears to be the most suitable since it reduces complex molecular dynamics to the purely electronic dynamics by exploiting the hierarchy of motional time scales. Experimentally, we propose to realize an innovative experimental setup. A few-cycle infrared (IR) pulse will be used to generate attosecond pulses in the extreme-ultraviolet (XUV) by high-harmonic generation. The IR pulse will be separated from the XUV by means of an innovative interferometer. Additionally, it will permit the introduction of a controlled attosecond delay between the two pulses. We propose to use the attosecond pulses as a tool to look inside individual IR- or UV-field cycles to better understand light-matter interactions. Time-resolved pump-probe experiments will be carried out on polyatomic molecules by detecting the energy and angular distribution of photoelectrons in a velocity-map imaging spectrometer. These experiments are expected to provide new insights
into the dynamics of multi-electron systems along with new results for the validation and
improvement of theoretical models. Multi-electron dynamics is indeed a very complex subject
on its own and even more so in the presence of strong laser fields. The proposed experiments
directly address theses challenges and are expected to provide new insights that will be beneficial to a wide range of scientific research areas."
Max ERC Funding
1 999 992 €
Duration
Start date: 2012-09-01, End date: 2017-08-31
Project acronym BALANCE
Project Mapping Dispersion Spectroscopically in Large Gas-Phase Molecular Ions
Researcher (PI) Peter CHEN
Host Institution (HI) EIDGENOESSISCHE TECHNISCHE HOCHSCHULE ZUERICH
Call Details Advanced Grant (AdG), PE4, ERC-2018-ADG
Summary We use IR spectroscopy of trapped ions in a cryogenic FT-ICR spectrometer to probe non-covalent, “dispersion” interactions in large, gas-phase molecular ions. We will measure conformational equilibria by N-H frequency shifts, and correlate gas-phase IR frequency to the N-H-N bond angle in an ionic H-bond. Substituents on “onium” cations can adopt various conformations, whose energies map interaction potentials. Substituents on their proton-bound dimers interact non-covalently through dispersion forces, whose quantitative evaluation in large molecules has remained difficult despite dispersion becoming increasingly cited as a design principle in the construction of catalysts and materials. The non-covalent interactions bend the N-H-N bond, leading to large shifts in the IR frequency. The proton-bound dimer acts like a molecular balance where the non-covalent interaction, is set against the bending potential in an ionic hydrogen bond. Despite encouragingly accurate calculations for small molecules, experimental benchmarks for large molecules in the gas phase remain scarce, and there is evidence that the good results for small molecules may not extrapolate reliably to large molecules. The present proposal introduces a new experimental probe of non-covalent interactions, providing a sensitive test of the diverging results coming from various computational methods and other experiments. The experiment must be done on isolated molecules in the gas phase, as previous work has shown that solvation substantially cancels out the attractive potential. Accordingly, the proposed experimental design, which involves a custom-built spectrometer, newly available tunable IR sources, chemical synthesis of custom substrates, and quantum calculations up to coupled-cluster levels of theory, showcases how an interdisciplinary approach combining physical and organic chemistry can solve a fundamental problem that impacts how we understand steric effects in organic chemistry.
Summary
We use IR spectroscopy of trapped ions in a cryogenic FT-ICR spectrometer to probe non-covalent, “dispersion” interactions in large, gas-phase molecular ions. We will measure conformational equilibria by N-H frequency shifts, and correlate gas-phase IR frequency to the N-H-N bond angle in an ionic H-bond. Substituents on “onium” cations can adopt various conformations, whose energies map interaction potentials. Substituents on their proton-bound dimers interact non-covalently through dispersion forces, whose quantitative evaluation in large molecules has remained difficult despite dispersion becoming increasingly cited as a design principle in the construction of catalysts and materials. The non-covalent interactions bend the N-H-N bond, leading to large shifts in the IR frequency. The proton-bound dimer acts like a molecular balance where the non-covalent interaction, is set against the bending potential in an ionic hydrogen bond. Despite encouragingly accurate calculations for small molecules, experimental benchmarks for large molecules in the gas phase remain scarce, and there is evidence that the good results for small molecules may not extrapolate reliably to large molecules. The present proposal introduces a new experimental probe of non-covalent interactions, providing a sensitive test of the diverging results coming from various computational methods and other experiments. The experiment must be done on isolated molecules in the gas phase, as previous work has shown that solvation substantially cancels out the attractive potential. Accordingly, the proposed experimental design, which involves a custom-built spectrometer, newly available tunable IR sources, chemical synthesis of custom substrates, and quantum calculations up to coupled-cluster levels of theory, showcases how an interdisciplinary approach combining physical and organic chemistry can solve a fundamental problem that impacts how we understand steric effects in organic chemistry.
Max ERC Funding
2 446 125 €
Duration
Start date: 2019-05-01, End date: 2024-04-30
Project acronym BDE
Project Beyond Distance Estimates: A New Theory of Heuristics for State-Space Search
Researcher (PI) Malte HELMERT
Host Institution (HI) UNIVERSITAT BASEL
Call Details Consolidator Grant (CoG), PE6, ERC-2018-COG
Summary "Many problems in computer science can be cast as state-space search, where the
objective is to find a path from an initial state to a goal state in a
directed graph called a ""state space"". State-space search is challenging due
to the state explosion problem a.k.a. ""curse of dimensionality"": interesting
state spaces are often astronomically large, defying brute-force exploration.
State-space search has been a core research problem in Artificial Intelligence
since its early days and is alive as ever. Every year, a substantial fraction
of research published at the ICAPS and SoCS conferences is concerned with
state-space search, and the topic is very active at general AI conferences
such as IJCAI and AAAI.
Algorithms in the A* family, dating back to 1968, are still the go-to approach
for state-space search. A* is a graph search algorithm whose only
""intelligence"" stems from a so-called ""heuristic function"", which estimates
the distance from a state to the nearest goal state. The efficiency of A*
depends on the accuracy of this estimate, and decades of research have pushed
the envelope in devising increasingly accurate estimates.
In this project, we question the ""A* + distance estimator"" paradigm and
explore three new directions that go beyond the classical approach:
1. We propose a new paradigm of declarative heuristics, where heuristic
information is not represented as distance estimates, but as properties of
solutions amenable to introspection and general reasoning.
2. We suggest moving the burden of creativity away from the human expert by
casting heuristic design as a meta-optimization problem that can be solved
automatically.
3. We propose abandoning the idea of exploring sequential paths in state
spaces, instead transforming state-space search into combinatorial
optimization problems with no explicit sequencing aspect. We argue that the
""curse of sequentiality"" is as bad as the curse of dimensionality and must
be addressed head-on."
Summary
"Many problems in computer science can be cast as state-space search, where the
objective is to find a path from an initial state to a goal state in a
directed graph called a ""state space"". State-space search is challenging due
to the state explosion problem a.k.a. ""curse of dimensionality"": interesting
state spaces are often astronomically large, defying brute-force exploration.
State-space search has been a core research problem in Artificial Intelligence
since its early days and is alive as ever. Every year, a substantial fraction
of research published at the ICAPS and SoCS conferences is concerned with
state-space search, and the topic is very active at general AI conferences
such as IJCAI and AAAI.
Algorithms in the A* family, dating back to 1968, are still the go-to approach
for state-space search. A* is a graph search algorithm whose only
""intelligence"" stems from a so-called ""heuristic function"", which estimates
the distance from a state to the nearest goal state. The efficiency of A*
depends on the accuracy of this estimate, and decades of research have pushed
the envelope in devising increasingly accurate estimates.
In this project, we question the ""A* + distance estimator"" paradigm and
explore three new directions that go beyond the classical approach:
1. We propose a new paradigm of declarative heuristics, where heuristic
information is not represented as distance estimates, but as properties of
solutions amenable to introspection and general reasoning.
2. We suggest moving the burden of creativity away from the human expert by
casting heuristic design as a meta-optimization problem that can be solved
automatically.
3. We propose abandoning the idea of exploring sequential paths in state
spaces, instead transforming state-space search into combinatorial
optimization problems with no explicit sequencing aspect. We argue that the
""curse of sequentiality"" is as bad as the curse of dimensionality and must
be addressed head-on."
Max ERC Funding
1 997 510 €
Duration
Start date: 2019-02-01, End date: 2024-01-31
Project acronym BEFINE
Project mechanical BEhavior of Fluid-INduced Earthquakes
Researcher (PI) Marie, Estelle, Solange VIOLAY
Host Institution (HI) ECOLE POLYTECHNIQUE FEDERALE DE LAUSANNE
Call Details Starting Grant (StG), PE10, ERC-2017-STG
Summary Fluids play an important role in fault zone and in earthquakes generation. Fluid pressure reduces the normal effective stress, lowering the frictional strength of the fault, potentially triggering earthquake ruptures. Fluid injection induced earthquakes (FIE) are direct evidence of the effect of fluid pressure on the fault strength. In addition, natural earthquake sequences are often associated with high fluid pressures at seismogenic depths. Although simple in theory, the mechanisms that govern the nucleation, propagation and recurrence of FIEs are poorly constrained, and our ability to assess the seismic hazard that is associated with natural and induced events remains limited. This project aims to enhance our knowledge of FIE mechanisms over entire seismic cycles through multidisciplinary approaches, including the following:
- Set-up and installation of a new and unique rock friction apparatus that is dedicated to the study of FIEs.
- Low strain rate friction experiments (coupled with electrical conductivity measurements) to investigate the influence of fluids on fault creep and earthquake recurrence.
- Intermediate strain rate friction experiments to investigate the effect of fluids on fault stability during earthquake nucleation.
- High strain rate friction experiments to investigate the effect of fluids on fault weakening during earthquake propagation.
- Post-mortem experimental fault analyses with state-of-art microstructural techniques.
- The theoretical friction law will be calibrated with friction experiments and faulted rock microstructural observations.
These steps will produce fundamental discoveries regarding natural earthquakes and tectonic processes and help scientists understand and eventually manage the occurrence of induced seismicity, an increasingly hot topic in geo-engineering. The sustainable exploitation of geo-resources is a key research and technology challenge at the European scale, with a substantial economical and societal impact.
Summary
Fluids play an important role in fault zone and in earthquakes generation. Fluid pressure reduces the normal effective stress, lowering the frictional strength of the fault, potentially triggering earthquake ruptures. Fluid injection induced earthquakes (FIE) are direct evidence of the effect of fluid pressure on the fault strength. In addition, natural earthquake sequences are often associated with high fluid pressures at seismogenic depths. Although simple in theory, the mechanisms that govern the nucleation, propagation and recurrence of FIEs are poorly constrained, and our ability to assess the seismic hazard that is associated with natural and induced events remains limited. This project aims to enhance our knowledge of FIE mechanisms over entire seismic cycles through multidisciplinary approaches, including the following:
- Set-up and installation of a new and unique rock friction apparatus that is dedicated to the study of FIEs.
- Low strain rate friction experiments (coupled with electrical conductivity measurements) to investigate the influence of fluids on fault creep and earthquake recurrence.
- Intermediate strain rate friction experiments to investigate the effect of fluids on fault stability during earthquake nucleation.
- High strain rate friction experiments to investigate the effect of fluids on fault weakening during earthquake propagation.
- Post-mortem experimental fault analyses with state-of-art microstructural techniques.
- The theoretical friction law will be calibrated with friction experiments and faulted rock microstructural observations.
These steps will produce fundamental discoveries regarding natural earthquakes and tectonic processes and help scientists understand and eventually manage the occurrence of induced seismicity, an increasingly hot topic in geo-engineering. The sustainable exploitation of geo-resources is a key research and technology challenge at the European scale, with a substantial economical and societal impact.
Max ERC Funding
1 982 925 €
Duration
Start date: 2018-03-01, End date: 2023-02-28
Project acronym BIGCODE
Project Learning from Big Code: Probabilistic Models, Analysis and Synthesis
Researcher (PI) Martin Vechev
Host Institution (HI) EIDGENOESSISCHE TECHNISCHE HOCHSCHULE ZUERICH
Call Details Starting Grant (StG), PE6, ERC-2015-STG
Summary The goal of this proposal is to fundamentally change the way we build and reason about software. We aim to develop new kinds of statistical programming systems that provide probabilistically likely solutions to tasks that are difficult or impossible to solve with traditional approaches.
These statistical programming systems will be based on probabilistic models of massive codebases (also known as ``Big Code'') built via a combination of advanced programming languages and powerful machine learning and natural language processing techniques. To solve a particular challenge, a statistical programming system will query a probabilistic model, compute the most likely predictions, and present those to the developer.
Based on probabilistic models of ``Big Code'', we propose to investigate new statistical techniques in the context of three fundamental research directions: i) statistical program synthesis where we develop techniques that automatically synthesize and predict new programs, ii) statistical prediction of program properties where we develop new techniques that can predict important facts (e.g., types) about programs, and iii) statistical translation of programs where we investigate new techniques for statistical translation of programs (e.g., from one programming language to another, or to a natural language).
We believe the research direction outlined in this interdisciplinary proposal opens a new and exciting area of computer science. This area will combine sophisticated statistical learning and advanced programming language techniques for building the next-generation statistical programming systems.
We expect the results of this proposal to have an immediate impact upon millions of developers worldwide, triggering a paradigm shift in the way tomorrow's software is built, as well as a long-lasting impact on scientific fields such as machine learning, natural language processing, programming languages and software engineering.
Summary
The goal of this proposal is to fundamentally change the way we build and reason about software. We aim to develop new kinds of statistical programming systems that provide probabilistically likely solutions to tasks that are difficult or impossible to solve with traditional approaches.
These statistical programming systems will be based on probabilistic models of massive codebases (also known as ``Big Code'') built via a combination of advanced programming languages and powerful machine learning and natural language processing techniques. To solve a particular challenge, a statistical programming system will query a probabilistic model, compute the most likely predictions, and present those to the developer.
Based on probabilistic models of ``Big Code'', we propose to investigate new statistical techniques in the context of three fundamental research directions: i) statistical program synthesis where we develop techniques that automatically synthesize and predict new programs, ii) statistical prediction of program properties where we develop new techniques that can predict important facts (e.g., types) about programs, and iii) statistical translation of programs where we investigate new techniques for statistical translation of programs (e.g., from one programming language to another, or to a natural language).
We believe the research direction outlined in this interdisciplinary proposal opens a new and exciting area of computer science. This area will combine sophisticated statistical learning and advanced programming language techniques for building the next-generation statistical programming systems.
We expect the results of this proposal to have an immediate impact upon millions of developers worldwide, triggering a paradigm shift in the way tomorrow's software is built, as well as a long-lasting impact on scientific fields such as machine learning, natural language processing, programming languages and software engineering.
Max ERC Funding
1 500 000 €
Duration
Start date: 2016-04-01, End date: 2021-03-31
Project acronym BIOCARB
Project Carbonate Biomineralization in the Marine Environment: Paleo-climate proxies and the origin of vital effects
Researcher (PI) Anders Meibom
Host Institution (HI) ECOLE POLYTECHNIQUE FEDERALE DE LAUSANNE
Call Details Advanced Grant (AdG), PE10, ERC-2009-AdG
Summary This interdisciplinary proposal has the objective to greatly enhance our understanding of fundamental biomineralization processes involved in the formation of calcium carbonates by marine organisms, such as corals, foraminifera and bivalves, in order to better understand vital effects. This is essential to the application of these carbonates as proxies for global (paleo-) environmental change. The core of the proposal is an experimental capability that I have pioneered during 2008: Dynamic stable isotopic labeling during formation of carbonate skeletons, tests, and shells, combined with NanoSIMS imaging. The NanoSIMS ion microprobe is a state-of-the-art analytical technology that allows precise elemental and isotopic imaging with a spatial resolution of ~100 nanometers. NanoSIMS imaging of the isotopic label(s) in the resulting biocarbonates and in associated cell-structures will be used to uncover cellular-level transport processes, timescales of formation of different biocarbonate components, as well as trace-elemental and isotopic fractionations. This will uncover the origin of vital effects. With this proposal, I establish a new scientific frontier and guarantee European leadership. The technical and scientific developments resulting from this work are broadly applicable and will radically change scientific ideas about marine carbonate biomineralization and compositional vital effects.
Summary
This interdisciplinary proposal has the objective to greatly enhance our understanding of fundamental biomineralization processes involved in the formation of calcium carbonates by marine organisms, such as corals, foraminifera and bivalves, in order to better understand vital effects. This is essential to the application of these carbonates as proxies for global (paleo-) environmental change. The core of the proposal is an experimental capability that I have pioneered during 2008: Dynamic stable isotopic labeling during formation of carbonate skeletons, tests, and shells, combined with NanoSIMS imaging. The NanoSIMS ion microprobe is a state-of-the-art analytical technology that allows precise elemental and isotopic imaging with a spatial resolution of ~100 nanometers. NanoSIMS imaging of the isotopic label(s) in the resulting biocarbonates and in associated cell-structures will be used to uncover cellular-level transport processes, timescales of formation of different biocarbonate components, as well as trace-elemental and isotopic fractionations. This will uncover the origin of vital effects. With this proposal, I establish a new scientific frontier and guarantee European leadership. The technical and scientific developments resulting from this work are broadly applicable and will radically change scientific ideas about marine carbonate biomineralization and compositional vital effects.
Max ERC Funding
2 182 000 €
Duration
Start date: 2010-07-01, End date: 2015-06-30
Project acronym BIOMOL. SIMULATION
Project Development of multi-scale molecular models, force fields and computer software for biomolecular simulation
Researcher (PI) Willem Frederik Van Gunsteren
Host Institution (HI) EIDGENOESSISCHE TECHNISCHE HOCHSCHULE ZUERICH
Call Details Advanced Grant (AdG), PE4, ERC-2008-AdG
Summary During the past decades the PI has helped shape the research field of computer simulation of biomolecular systems at the atomic level. He has carried out one of the first molecular dynamics (MD) simulations of proteins, and has since then contributed many different methodological improvements and developed one of the major atomic-level force fields for simulations of proteins, carbohydrates, nucleotides and lipids. Methodology and force field have been implemented in a set of programs called GROMOS (GROningen MOlecular Simulation package), which is currently used in hundreds of academic and industrial research groups from over 50 countries on all continents. It is proposed to develop a next generation of molecular models, force fields, multi-scaling simulation methodology and software for biomolecular simulations which is at least an order of magnitude more accurate in terms of energetics, and which is 1000 times more efficient through the use of coarse-grained molecular models than the currently available software and models.
Summary
During the past decades the PI has helped shape the research field of computer simulation of biomolecular systems at the atomic level. He has carried out one of the first molecular dynamics (MD) simulations of proteins, and has since then contributed many different methodological improvements and developed one of the major atomic-level force fields for simulations of proteins, carbohydrates, nucleotides and lipids. Methodology and force field have been implemented in a set of programs called GROMOS (GROningen MOlecular Simulation package), which is currently used in hundreds of academic and industrial research groups from over 50 countries on all continents. It is proposed to develop a next generation of molecular models, force fields, multi-scaling simulation methodology and software for biomolecular simulations which is at least an order of magnitude more accurate in terms of energetics, and which is 1000 times more efficient through the use of coarse-grained molecular models than the currently available software and models.
Max ERC Funding
1 320 000 €
Duration
Start date: 2008-11-01, End date: 2014-09-30
Project acronym BLACARAT
Project "Black Carbon in the Atmosphere: Emissions, Aging and Cloud Interactions"
Researcher (PI) Martin Gysel Beer
Host Institution (HI) PAUL SCHERRER INSTITUT
Call Details Consolidator Grant (CoG), PE10, ERC-2013-CoG
Summary "Atmospheric aerosol particles have been shown to impact the earth's climate because they scatter and absorb solar radiation (direct effect) and because they can modify the microphysical properties of clouds by acting as cloud condensation nuclei or ice nuclei (indirect effects). Radiative forcing by anthropogenic aerosols remains poorly quantified, thus leading to considerable uncertainty in our understanding of the earth’s climate response to the radiative forcing by greenhouse gases. Black carbon (BC), mostly emitted by anthropogenic combustion processes and biomass burning, is an important component of atmospheric aerosols. Estimates show that BC may be the second strongest contributor (after CO2) to global warming. Adverse health effects due to particulate air pollution have also been associated with traffic-related BC particles. These climate and health effects brought BC emission reductions into the political focus of possible mitigation strategies with immediate and multiple benefits for human well-being.
Laboratory experiments aim at the physical and chemical characterisation of BC emissions from diesel engines and biomass burning under controlled conditions. A mobile laboratory equipped with state-of-the-art aerosol sensors will be used to determine the contribution of different BC sources to atmospheric BC loadings, and to investigate the evolution of the relevant BC properties with atmospheric aging during transport from sources to remote areas. The interactions of BC particles with clouds as a function of BC properties will be investigated with in-situ measurements by operating quantitative single particle instruments behind a novel sampling inlet, which makes selective sampling of interstitial, cloud droplet residual or ice crystal residual particles possible. Above experimental studies aim at improving our understanding of BC’s atmospheric life cycle and will be used in model simulations for quantitatively assessing the atmospheric impacts of BC."
Summary
"Atmospheric aerosol particles have been shown to impact the earth's climate because they scatter and absorb solar radiation (direct effect) and because they can modify the microphysical properties of clouds by acting as cloud condensation nuclei or ice nuclei (indirect effects). Radiative forcing by anthropogenic aerosols remains poorly quantified, thus leading to considerable uncertainty in our understanding of the earth’s climate response to the radiative forcing by greenhouse gases. Black carbon (BC), mostly emitted by anthropogenic combustion processes and biomass burning, is an important component of atmospheric aerosols. Estimates show that BC may be the second strongest contributor (after CO2) to global warming. Adverse health effects due to particulate air pollution have also been associated with traffic-related BC particles. These climate and health effects brought BC emission reductions into the political focus of possible mitigation strategies with immediate and multiple benefits for human well-being.
Laboratory experiments aim at the physical and chemical characterisation of BC emissions from diesel engines and biomass burning under controlled conditions. A mobile laboratory equipped with state-of-the-art aerosol sensors will be used to determine the contribution of different BC sources to atmospheric BC loadings, and to investigate the evolution of the relevant BC properties with atmospheric aging during transport from sources to remote areas. The interactions of BC particles with clouds as a function of BC properties will be investigated with in-situ measurements by operating quantitative single particle instruments behind a novel sampling inlet, which makes selective sampling of interstitial, cloud droplet residual or ice crystal residual particles possible. Above experimental studies aim at improving our understanding of BC’s atmospheric life cycle and will be used in model simulations for quantitatively assessing the atmospheric impacts of BC."
Max ERC Funding
1 992 015 €
Duration
Start date: 2014-03-01, End date: 2019-02-28
Project acronym CASSANDRA
Project Accelerating mass loss of Greenland: firn and the shifting runoff limit
Researcher (PI) Horst MACHGUTH
Host Institution (HI) UNIVERSITE DE FRIBOURG
Call Details Consolidator Grant (CoG), PE10, ERC-2018-COG
Summary Meltwater running off the flanks of the Greenland ice sheet contributes roughly 60% to its mass loss, the rest being due to calving. Only meltwater originating from below the elevation of the runoff limit leaves the ice sheet, contributing to mass loss; melt at higher elevations refreezes in the porous firn and does not drive mass loss. Therefore any shift in the runoff limit modifies mass loss and subsequent sea level rise. New evidence shows surface runoff at increasingly high elevations, outpacing the rate at which the equilibrium line elevation rises. This research proposal focuses on the runoff limit as a powerful yet poorly understood modulator of Greenland mass balance. We will track the runoff limit over the full satellite era using two of the largest and oldest remote sensing archives, Landsat and the Advanced Very High Resolution Radiometer (AVHRR). We will establish time series of the runoff limit for all regions of Greenland to identify the mechanisms driving fluctuations in the runoff limit. This newly gained process understanding and a wealth of in-situ measurements will then be used to build firn hydrology models capable of simulating runoff and the associated runoff limit over time. Eventually, the firn hydrology models will be applied to reconcile estimates of Greenland past, present and future mass balance. Covering the entire satellite era and all of Greenland, the focus on the runoff limit will constitute a paradigm shift leading to major advance in our understanding of how vulnerable the surface of the ice sheet reacts to climate change and how the changing surface impacts runoff and thus Greenland's role in the global sea level budget.
Summary
Meltwater running off the flanks of the Greenland ice sheet contributes roughly 60% to its mass loss, the rest being due to calving. Only meltwater originating from below the elevation of the runoff limit leaves the ice sheet, contributing to mass loss; melt at higher elevations refreezes in the porous firn and does not drive mass loss. Therefore any shift in the runoff limit modifies mass loss and subsequent sea level rise. New evidence shows surface runoff at increasingly high elevations, outpacing the rate at which the equilibrium line elevation rises. This research proposal focuses on the runoff limit as a powerful yet poorly understood modulator of Greenland mass balance. We will track the runoff limit over the full satellite era using two of the largest and oldest remote sensing archives, Landsat and the Advanced Very High Resolution Radiometer (AVHRR). We will establish time series of the runoff limit for all regions of Greenland to identify the mechanisms driving fluctuations in the runoff limit. This newly gained process understanding and a wealth of in-situ measurements will then be used to build firn hydrology models capable of simulating runoff and the associated runoff limit over time. Eventually, the firn hydrology models will be applied to reconcile estimates of Greenland past, present and future mass balance. Covering the entire satellite era and all of Greenland, the focus on the runoff limit will constitute a paradigm shift leading to major advance in our understanding of how vulnerable the surface of the ice sheet reacts to climate change and how the changing surface impacts runoff and thus Greenland's role in the global sea level budget.
Max ERC Funding
1 989 181 €
Duration
Start date: 2019-05-01, End date: 2024-04-30
Project acronym CAUSALPATH
Project Next Generation Causal Analysis: Inspired by the Induction of Biological Pathways from Cytometry Data
Researcher (PI) Ioannis Tsamardinos
Host Institution (HI) PANEPISTIMIO KRITIS
Call Details Consolidator Grant (CoG), PE6, ERC-2013-CoG
Summary Discovering the causal mechanisms of a complex system of interacting components is necessary in order to control it. Computational Causal Discovery (CD) is a field that offers the potential to discover causal relations under certain conditions from observational data alone or with a limited number of interventions/manipulations.
An important, challenging biological problem that may take decades of experimental work is the induction of biological cellular pathways; pathways are informal causal models indispensable in biological research and drug design. Recent exciting advances in flow/mass cytometry biotechnology allow the generation of large-sample datasets containing measurements on single cells, thus setting the problem of pathway learning suitable for CD methods.
CAUSALPATH builds upon and further advances recent breakthrough developments in CD methods to enable the induction of biological pathways from cytometry and other omics data. As a testbed problem we focus on the differentiation of human T-cells; these are involved in autoimmune and inflammatory diseases, as well as cancer and thus, are targets of new drug development for a range of chronic diseases. The biological problem acts as our campus for general novel formalisms, practical algorithms, and useful tools development, pointing to fundamental CD problems: presence of feedback cycles, presence of latent confounding variables, CD from time-course data, Integrative Causal Analysis (INCA) of heterogeneous datasets and others.
Three features complement CAUSALPATH’s approach: (A) methods development will co-evolve with biological wet-lab experiments periodically testing the algorithmic postulates, (B) Open-source tools will be developed for the non-expert, and (C) Commercial exploitation of the results will be sought out.
CAUSALPATH brings together an interdisciplinary team, committed to this vision. It builds upon the PI’s group recent important results on INCA algorithms.
Summary
Discovering the causal mechanisms of a complex system of interacting components is necessary in order to control it. Computational Causal Discovery (CD) is a field that offers the potential to discover causal relations under certain conditions from observational data alone or with a limited number of interventions/manipulations.
An important, challenging biological problem that may take decades of experimental work is the induction of biological cellular pathways; pathways are informal causal models indispensable in biological research and drug design. Recent exciting advances in flow/mass cytometry biotechnology allow the generation of large-sample datasets containing measurements on single cells, thus setting the problem of pathway learning suitable for CD methods.
CAUSALPATH builds upon and further advances recent breakthrough developments in CD methods to enable the induction of biological pathways from cytometry and other omics data. As a testbed problem we focus on the differentiation of human T-cells; these are involved in autoimmune and inflammatory diseases, as well as cancer and thus, are targets of new drug development for a range of chronic diseases. The biological problem acts as our campus for general novel formalisms, practical algorithms, and useful tools development, pointing to fundamental CD problems: presence of feedback cycles, presence of latent confounding variables, CD from time-course data, Integrative Causal Analysis (INCA) of heterogeneous datasets and others.
Three features complement CAUSALPATH’s approach: (A) methods development will co-evolve with biological wet-lab experiments periodically testing the algorithmic postulates, (B) Open-source tools will be developed for the non-expert, and (C) Commercial exploitation of the results will be sought out.
CAUSALPATH brings together an interdisciplinary team, committed to this vision. It builds upon the PI’s group recent important results on INCA algorithms.
Max ERC Funding
1 724 000 €
Duration
Start date: 2015-01-01, End date: 2019-12-31
Project acronym chromo-SUMMIT
Project Decoding dynamic chromatin signaling by single-molecule multiplex detection
Researcher (PI) Beat FIERZ
Host Institution (HI) ECOLE POLYTECHNIQUE FEDERALE DE LAUSANNE
Call Details Consolidator Grant (CoG), PE4, ERC-2016-COG
Summary Transient multivalent interactions are critical for biological processes such as signaling pathways controlling chromatin function. Chromatin, the nucleoprotein complex organizing the genome, is dynamically regulated by post-translational modifications (PTMs) of the chromatin fiber. Protein effectors interact with combinations of these PTMs through multivalent interactions, deposit novel PTMs, thereby propagate signaling cascades and remodel chromatin structure. To reveal the underlying molecular mechanisms, methods outside classical biochemistry are required, in particular due to the combinational complexity of chromatin PTMs and the transient supramolecular interactions crucial for their recognition. Here, we develop a novel approach, where we synthesize arrays of chemically defined designer chromatin fibers and use dynamic multiplex single-molecule imaging to dissect multivalent signaling processes in chromatin. Our studies target a key pathway, the DNA damage response (DDR), which regulates DNA repair processes central to cell survival and is critically implicated in cancer. Detailed knowledge is of utmost importance to develop targeted therapeutic interventions. We thus employ advanced peptide and protein chemistry to generate libraries of chromatin fibers of a defined PTM state that is encoded in the chromatin DNA. With the library immobilized in a flow cell, we use single-molecule detection to directly observe signaling processes by key DDR effectors in real time. Subsequent in situ polony decoding allows the identification of each chromatin fiber’s modification state, enabling broad sampling of signaling outcomes. Finally, we use dynamic computational models to integrate the effector-chromatin interaction network and test key mechanisms in cancer-based cell culture. Together, these methods will yield fundamental insight into chromatin and DDR signaling and will be of broad use for chemical and biomedical research with applications beyond the chromatin field.
Summary
Transient multivalent interactions are critical for biological processes such as signaling pathways controlling chromatin function. Chromatin, the nucleoprotein complex organizing the genome, is dynamically regulated by post-translational modifications (PTMs) of the chromatin fiber. Protein effectors interact with combinations of these PTMs through multivalent interactions, deposit novel PTMs, thereby propagate signaling cascades and remodel chromatin structure. To reveal the underlying molecular mechanisms, methods outside classical biochemistry are required, in particular due to the combinational complexity of chromatin PTMs and the transient supramolecular interactions crucial for their recognition. Here, we develop a novel approach, where we synthesize arrays of chemically defined designer chromatin fibers and use dynamic multiplex single-molecule imaging to dissect multivalent signaling processes in chromatin. Our studies target a key pathway, the DNA damage response (DDR), which regulates DNA repair processes central to cell survival and is critically implicated in cancer. Detailed knowledge is of utmost importance to develop targeted therapeutic interventions. We thus employ advanced peptide and protein chemistry to generate libraries of chromatin fibers of a defined PTM state that is encoded in the chromatin DNA. With the library immobilized in a flow cell, we use single-molecule detection to directly observe signaling processes by key DDR effectors in real time. Subsequent in situ polony decoding allows the identification of each chromatin fiber’s modification state, enabling broad sampling of signaling outcomes. Finally, we use dynamic computational models to integrate the effector-chromatin interaction network and test key mechanisms in cancer-based cell culture. Together, these methods will yield fundamental insight into chromatin and DDR signaling and will be of broad use for chemical and biomedical research with applications beyond the chromatin field.
Max ERC Funding
1 999 815 €
Duration
Start date: 2017-05-01, End date: 2022-04-30
Project acronym CICERO
Project Cold Ion Chemistry - Experiments within a Rydberg Orbit
Researcher (PI) Frédéric MERKT
Host Institution (HI) EIDGENOESSISCHE TECHNISCHE HOCHSCHULE ZUERICH
Call Details Advanced Grant (AdG), PE4, ERC-2016-ADG
Summary "To date no experiment has investigated ion-molecule reactions at temperatures significantly below about 20 K, for two reasons: (i) Cooling the translational and internal degrees of freedom of ions and molecules is extremely challenging. (ii) Even very weak stray electric fields accelerate the ions. A potential difference of only 1 mV across the reaction volume imparts a kinetic energy of 1 meV to ions, which corresponds to a temperature of about 12 K. Quantum mechanical effects arising from the translational and the frozen or hindered rotational motion of the reactants in the intermolecular potential are only expected to be significant below 20 K and have therefore not been observed yet in ion-molecule reactions, even for reactions involving the lightest ions and molecules. This proposal aims at developing a new experimental method to study ion-molecule reactions at temperatures down to 100 mK and to study ion-molecule reactions involving light species, with particular emphasis placed on the observation and quantification of quantum effects in low-temperature ion-molecule chemistry. To reach this goal, we will study the ion-molecule reactions within the orbit of a highly excited Rydberg electron, which will shield the reaction from stray fields without affecting its outcome. To reach very low collision energies, we will use a merged-beam approach relying on a surface-electrode Rydberg-Stark deflector. In the preparatory phase of this proposal, we have carried out a proof-of-principle measurement of the H2+ + H2 -> H3+ + H reaction below 1 K using a simplified version of the ""ideal"" instrument and demonstrated the feasibility of our method. We now plan to exploit the full potential of our new approach and study important ion-molecule reactions in a temperature range thought until now to be experimentally inaccessible."
Summary
"To date no experiment has investigated ion-molecule reactions at temperatures significantly below about 20 K, for two reasons: (i) Cooling the translational and internal degrees of freedom of ions and molecules is extremely challenging. (ii) Even very weak stray electric fields accelerate the ions. A potential difference of only 1 mV across the reaction volume imparts a kinetic energy of 1 meV to ions, which corresponds to a temperature of about 12 K. Quantum mechanical effects arising from the translational and the frozen or hindered rotational motion of the reactants in the intermolecular potential are only expected to be significant below 20 K and have therefore not been observed yet in ion-molecule reactions, even for reactions involving the lightest ions and molecules. This proposal aims at developing a new experimental method to study ion-molecule reactions at temperatures down to 100 mK and to study ion-molecule reactions involving light species, with particular emphasis placed on the observation and quantification of quantum effects in low-temperature ion-molecule chemistry. To reach this goal, we will study the ion-molecule reactions within the orbit of a highly excited Rydberg electron, which will shield the reaction from stray fields without affecting its outcome. To reach very low collision energies, we will use a merged-beam approach relying on a surface-electrode Rydberg-Stark deflector. In the preparatory phase of this proposal, we have carried out a proof-of-principle measurement of the H2+ + H2 -> H3+ + H reaction below 1 K using a simplified version of the ""ideal"" instrument and demonstrated the feasibility of our method. We now plan to exploit the full potential of our new approach and study important ion-molecule reactions in a temperature range thought until now to be experimentally inaccessible."
Max ERC Funding
2 130 088 €
Duration
Start date: 2017-06-01, End date: 2022-05-31
Project acronym CLUSTER
Project Birth of solids: atomic-scale processes in crystal nucleation
Researcher (PI) Rolf Erni
Host Institution (HI) EIDGENOSSISCHE MATERIALPRUFUNGS- UND FORSCHUNGSANSTALT
Call Details Consolidator Grant (CoG), PE4, ERC-2015-CoG
Summary The goal of this project is to explore the fundamental processes which trigger the nucleation and growth of solids. Condensed matter is formed by clustering of atoms, ions or molecules. This initial step is key for the onset of crystallization, condensation and precipitate formation. Yet, despite of the scientific and technological significance of these phenomena, on an atomistic level we merely have expectations on how atoms should behave rather than experimental evidence about how the growth of solid matter is initiated. The classical nucleation theory is commonly in agreement with experiments, provided the original and the final stages are inspected qualitatively. However, the classical theory does not define what fundamentally constitutes a pre-nucleation state or how a nucleus is formed at all. CLUSTER aims at investigating the very early stages of crystalline matter formation on an unprecedented length scale. It shall explore the atomic mechanisms which prompt the formation of solids. Complemented by density functional theory calculations and molecular dynamics simulations, in-situ high-resolution electron microscopy shall be used to investigate the formation, dynamics, stability and evolution of tiniest atomic clusters which represent the embryos of solid matter. Firstly, we investigate the 3D structure of clusters deposited on suspended graphene. Secondly, we focus on cluster formation, the evolution of sub-critical nuclei and the onset of particle growth by thermal activation. Thirdly, using a novel liquid-cell approach in the transmission electron microscope, we control and monitor in-situ cluster formation and precipitation in supersaturated solutions. The results of CLUSTER, which will advance the understanding of the birth of solid matter, are important for the controlled synthesis of (nano-)materials, for cluster science and catalysis and for the development of novel materials.
Summary
The goal of this project is to explore the fundamental processes which trigger the nucleation and growth of solids. Condensed matter is formed by clustering of atoms, ions or molecules. This initial step is key for the onset of crystallization, condensation and precipitate formation. Yet, despite of the scientific and technological significance of these phenomena, on an atomistic level we merely have expectations on how atoms should behave rather than experimental evidence about how the growth of solid matter is initiated. The classical nucleation theory is commonly in agreement with experiments, provided the original and the final stages are inspected qualitatively. However, the classical theory does not define what fundamentally constitutes a pre-nucleation state or how a nucleus is formed at all. CLUSTER aims at investigating the very early stages of crystalline matter formation on an unprecedented length scale. It shall explore the atomic mechanisms which prompt the formation of solids. Complemented by density functional theory calculations and molecular dynamics simulations, in-situ high-resolution electron microscopy shall be used to investigate the formation, dynamics, stability and evolution of tiniest atomic clusters which represent the embryos of solid matter. Firstly, we investigate the 3D structure of clusters deposited on suspended graphene. Secondly, we focus on cluster formation, the evolution of sub-critical nuclei and the onset of particle growth by thermal activation. Thirdly, using a novel liquid-cell approach in the transmission electron microscope, we control and monitor in-situ cluster formation and precipitation in supersaturated solutions. The results of CLUSTER, which will advance the understanding of the birth of solid matter, are important for the controlled synthesis of (nano-)materials, for cluster science and catalysis and for the development of novel materials.
Max ERC Funding
2 271 250 €
Duration
Start date: 2016-06-01, End date: 2021-05-31