Project acronym ASTROFLOW
Project The influence of stellar outflows on exoplanetary mass loss
Researcher (PI) Aline VIDOTTO
Host Institution (HI) THE PROVOST, FELLOWS, FOUNDATION SCHOLARS & THE OTHER MEMBERS OF BOARD OF THE COLLEGE OF THE HOLY & UNDIVIDED TRINITY OF QUEEN ELIZABETH NEAR DUBLIN
Call Details Consolidator Grant (CoG), PE9, ERC-2018-COG
Summary ASTROFLOW aims to make ground-breaking progress in our physical understanding of exoplanetary mass loss, by quantifying the influence of stellar outflows on atmospheric escape of close-in exoplanets. Escape plays a key role in planetary evolution, population, and potential to develop life. Stellar irradiation and outflows affect planetary mass loss: irradiation heats planetary atmospheres, which inflate and more likely escape; outflows cause pressure confinement around otherwise freely escaping atmospheres. This external pressure can increase, reduce or even suppress escape rates; its effects on exoplanetary mass loss remain largely unexplored due to the complexity of such interactions. I will fill this knowledge gap by developing a novel modelling framework of atmospheric escape that will, for the first time, consider the effects of realistic stellar outflows on exoplanetary mass loss. My expertise in stellar wind theory and 3D magnetohydrodynamic simulations is crucial for producing the next-generation models of planetary escape. My framework will consist of state-of-the-art, time-dependent, 3D simulations of stellar outflows (Method 1), which will be coupled to novel 3D simulations of atmospheric escape (Method 2). My models will account for the major underlying physical processes of mass loss. With this, I will determine the response of planetary mass loss to realistic stellar particle, magnetic and radiation environments and will characterise the physical conditions of the escaping material. I will compute how its extinction varies during transit and compare synthetic line profiles to atmospheric escape observations from, eg, Hubble and our NASA cubesat CUTE. Strong synergy with upcoming observations (JWST, TESS, SPIRou, CARMENES) also exists. Determining the lifetime of planetary atmospheres is essential to understanding populations of exoplanets. ASTROFLOW’s work will be the foundation for future research of how exoplanets evolve under mass-loss processes.
Summary
ASTROFLOW aims to make ground-breaking progress in our physical understanding of exoplanetary mass loss, by quantifying the influence of stellar outflows on atmospheric escape of close-in exoplanets. Escape plays a key role in planetary evolution, population, and potential to develop life. Stellar irradiation and outflows affect planetary mass loss: irradiation heats planetary atmospheres, which inflate and more likely escape; outflows cause pressure confinement around otherwise freely escaping atmospheres. This external pressure can increase, reduce or even suppress escape rates; its effects on exoplanetary mass loss remain largely unexplored due to the complexity of such interactions. I will fill this knowledge gap by developing a novel modelling framework of atmospheric escape that will, for the first time, consider the effects of realistic stellar outflows on exoplanetary mass loss. My expertise in stellar wind theory and 3D magnetohydrodynamic simulations is crucial for producing the next-generation models of planetary escape. My framework will consist of state-of-the-art, time-dependent, 3D simulations of stellar outflows (Method 1), which will be coupled to novel 3D simulations of atmospheric escape (Method 2). My models will account for the major underlying physical processes of mass loss. With this, I will determine the response of planetary mass loss to realistic stellar particle, magnetic and radiation environments and will characterise the physical conditions of the escaping material. I will compute how its extinction varies during transit and compare synthetic line profiles to atmospheric escape observations from, eg, Hubble and our NASA cubesat CUTE. Strong synergy with upcoming observations (JWST, TESS, SPIRou, CARMENES) also exists. Determining the lifetime of planetary atmospheres is essential to understanding populations of exoplanets. ASTROFLOW’s work will be the foundation for future research of how exoplanets evolve under mass-loss processes.
Max ERC Funding
1 999 956 €
Duration
Start date: 2019-09-01, End date: 2024-08-31
Project acronym BeyondOpposition
Project Opposing Sexual and Gender Rights and Equalities: Transforming Everyday Spaces
Researcher (PI) Katherine Browne
Host Institution (HI) NATIONAL UNIVERSITY OF IRELAND MAYNOOTH
Call Details Consolidator Grant (CoG), SH2, ERC-2018-COG
Summary OPPSEXRIGHTS will be the first large-scale, transnational study to consider the effects of recent Sexual and Gender Rights and Equalities (SGRE) on those who oppose them, by exploring opponents’ experiences of the transformation of everyday spaces. It will work beyond contemporary polarisations, creating new possibilities for social transformation. This cutting-edge research engages with the dramatically altered social and political landscapes in the late 20th and early 21st Century created through the development of lesbian, gay, bisexual, and trans, and women’s rights. Recent reactionary politics highlight the pressing need to understand the position of those who experience these new social orders as a loss. The backlash to SGRE has coalesced into various resistances that are tangibly different to the classic vilification of homosexuality, or those that are anti-woman. Some who oppose SGRE have found themselves the subject of public critique; in the workplace, their jobs threatened, while at home, engagements with schools can cause family conflicts. This is particularly visible in the case studies of Ireland, UK and Canada because of SGRE. A largescale transnational systematic database will be created using low risk (media and organisational discourses; participant observation at oppositional events) and higher risk (online data collection and interviews) methods. Experimenting with social transformation, OPPSEXRIGHTS will work to build bridges between ‘enemies’, including families and communities, through innovative discussion and arts-based workshops. This ambitious project has the potential to create tangible solutions that tackle contemporary societal issues, which are founded in polarisations that are seemingly insurmountable.
Summary
OPPSEXRIGHTS will be the first large-scale, transnational study to consider the effects of recent Sexual and Gender Rights and Equalities (SGRE) on those who oppose them, by exploring opponents’ experiences of the transformation of everyday spaces. It will work beyond contemporary polarisations, creating new possibilities for social transformation. This cutting-edge research engages with the dramatically altered social and political landscapes in the late 20th and early 21st Century created through the development of lesbian, gay, bisexual, and trans, and women’s rights. Recent reactionary politics highlight the pressing need to understand the position of those who experience these new social orders as a loss. The backlash to SGRE has coalesced into various resistances that are tangibly different to the classic vilification of homosexuality, or those that are anti-woman. Some who oppose SGRE have found themselves the subject of public critique; in the workplace, their jobs threatened, while at home, engagements with schools can cause family conflicts. This is particularly visible in the case studies of Ireland, UK and Canada because of SGRE. A largescale transnational systematic database will be created using low risk (media and organisational discourses; participant observation at oppositional events) and higher risk (online data collection and interviews) methods. Experimenting with social transformation, OPPSEXRIGHTS will work to build bridges between ‘enemies’, including families and communities, through innovative discussion and arts-based workshops. This ambitious project has the potential to create tangible solutions that tackle contemporary societal issues, which are founded in polarisations that are seemingly insurmountable.
Max ERC Funding
1 988 652 €
Duration
Start date: 2019-10-01, End date: 2024-09-30
Project acronym DELPHI
Project Computing Answers to Complex Questions in Broad Domains
Researcher (PI) Jonathan Berant
Host Institution (HI) TEL AVIV UNIVERSITY
Call Details Starting Grant (StG), PE6, ERC-2018-STG
Summary The explosion of information around us has democratized knowledge and transformed its availability for
people around the world. Still, since information is mediated through automated systems, access is bounded
by their ability to understand language.
Consider an economist asking “What fraction of the top-5 growing countries last year raised their co2 emission?”.
While the required information is available, answering such complex questions automatically is
not possible. Current question answering systems can answer simple questions in broad domains, or complex
questions in narrow domains. However, broad and complex questions are beyond the reach of state-of-the-art.
This is because systems are unable to decompose questions into their parts, and find the relevant information
in multiple sources. Further, as answering such questions is hard for people, collecting large datasets to train
such models is prohibitive.
In this proposal I ask: Can computers answer broad and complex questions that require reasoning over
multiple modalities? I argue that by synthesizing the advantages of symbolic and distributed representations
the answer will be “yes”. My thesis is that symbolic representations are suitable for meaning composition, as
they provide interpretability, coverage, and modularity. Complementarily, distributed representations (learned
by neural nets) excel at capturing the fuzziness of language. I propose a framework where complex questions
are symbolically decomposed into sub-questions, each is answered with a neural network, and the final answer
is computed from all gathered information.
This research tackles foundational questions in language understanding. What is the right representation
for reasoning in language? Can models learn to perform complex actions in the face of paucity of data?
Moreover, my research, if successful, will transform how we interact with machines, and define a role for
them as research assistants in science, education, and our daily life.
Summary
The explosion of information around us has democratized knowledge and transformed its availability for
people around the world. Still, since information is mediated through automated systems, access is bounded
by their ability to understand language.
Consider an economist asking “What fraction of the top-5 growing countries last year raised their co2 emission?”.
While the required information is available, answering such complex questions automatically is
not possible. Current question answering systems can answer simple questions in broad domains, or complex
questions in narrow domains. However, broad and complex questions are beyond the reach of state-of-the-art.
This is because systems are unable to decompose questions into their parts, and find the relevant information
in multiple sources. Further, as answering such questions is hard for people, collecting large datasets to train
such models is prohibitive.
In this proposal I ask: Can computers answer broad and complex questions that require reasoning over
multiple modalities? I argue that by synthesizing the advantages of symbolic and distributed representations
the answer will be “yes”. My thesis is that symbolic representations are suitable for meaning composition, as
they provide interpretability, coverage, and modularity. Complementarily, distributed representations (learned
by neural nets) excel at capturing the fuzziness of language. I propose a framework where complex questions
are symbolically decomposed into sub-questions, each is answered with a neural network, and the final answer
is computed from all gathered information.
This research tackles foundational questions in language understanding. What is the right representation
for reasoning in language? Can models learn to perform complex actions in the face of paucity of data?
Moreover, my research, if successful, will transform how we interact with machines, and define a role for
them as research assistants in science, education, and our daily life.
Max ERC Funding
1 499 375 €
Duration
Start date: 2019-04-01, End date: 2024-03-31
Project acronym DIFFOP
Project Nonlinear Data and Signal Analysis with Diffusion Operators
Researcher (PI) Ronen TALMON
Host Institution (HI) TECHNION - ISRAEL INSTITUTE OF TECHNOLOGY
Call Details Starting Grant (StG), PE6, ERC-2018-STG
Summary Nowadays, extensive collection and storage of massive data sets have become a routine in multiple disciplines and in everyday life. These large amounts of intricate data often make data samples arithmetic and basic comparisons problematic, raising new challenges to traditional data analysis objectives such as filtering and prediction. Furthermore, the availability of such data constantly pushes the boundaries of data analysis to new emerging domains, ranging from neuronal and social network analysis to multimodal sensor fusion. The combination of evolved data and new domains drives a fundamental change in the field of data analysis. Indeed, many classical model-based techniques have become obsolete since their models do not embody the richness of the collected data. Today, one notable avenue of research is the development of nonlinear techniques that transition from data to creating representations, without deriving models in closed-form. The vast majority of such existing data-driven methods operate directly on the data, a hard task by itself when the data are large and elaborated. The goal of this research is to develop a fundamentally new methodology for high dimensional data analysis with diffusion operators, making use of recent transformative results in manifold and geometry learning. More concretely, shifting the focus from processing the data samples themselves and considering instead structured data through the lens of diffusion operators will introduce new powerful “handles” to data, capturing their complexity efficiently. We will study the basic theory behind this nonlinear analysis, develop new operators for this purpose, and devise efficient data-driven algorithms. In addition, we will explore how our approach can be leveraged for devising efficient solutions to a broad range of open real-world data analysis problems, involving intrinsic representations, sensor fusion, time-series analysis, network connectivity inference, and domain adaptation.
Summary
Nowadays, extensive collection and storage of massive data sets have become a routine in multiple disciplines and in everyday life. These large amounts of intricate data often make data samples arithmetic and basic comparisons problematic, raising new challenges to traditional data analysis objectives such as filtering and prediction. Furthermore, the availability of such data constantly pushes the boundaries of data analysis to new emerging domains, ranging from neuronal and social network analysis to multimodal sensor fusion. The combination of evolved data and new domains drives a fundamental change in the field of data analysis. Indeed, many classical model-based techniques have become obsolete since their models do not embody the richness of the collected data. Today, one notable avenue of research is the development of nonlinear techniques that transition from data to creating representations, without deriving models in closed-form. The vast majority of such existing data-driven methods operate directly on the data, a hard task by itself when the data are large and elaborated. The goal of this research is to develop a fundamentally new methodology for high dimensional data analysis with diffusion operators, making use of recent transformative results in manifold and geometry learning. More concretely, shifting the focus from processing the data samples themselves and considering instead structured data through the lens of diffusion operators will introduce new powerful “handles” to data, capturing their complexity efficiently. We will study the basic theory behind this nonlinear analysis, develop new operators for this purpose, and devise efficient data-driven algorithms. In addition, we will explore how our approach can be leveraged for devising efficient solutions to a broad range of open real-world data analysis problems, involving intrinsic representations, sensor fusion, time-series analysis, network connectivity inference, and domain adaptation.
Max ERC Funding
1 260 000 €
Duration
Start date: 2019-02-01, End date: 2024-01-31
Project acronym EMERGE
Project Reconstructing the emergence of the Milky Way’s stellar population with Gaia, SDSS-V and JWST
Researcher (PI) Dan Maoz
Host Institution (HI) TEL AVIV UNIVERSITY
Call Details Advanced Grant (AdG), PE9, ERC-2018-ADG
Summary Understanding how the Milky Way arrived at its present state requires a large volume of precision measurements of our Galaxy’s current makeup, as well as an empirically based understanding of the main processes involved in the Galaxy’s evolution. Such data are now about to arrive in the flood of quality information from Gaia and SDSS-V. The demography of the stars and of the compact stellar remnants in our Galaxy, in terms of phase-space location, mass, age, metallicity, and multiplicity are data products that will come directly from these surveys. I propose to integrate this information into a comprehensive picture of the Milky Way’s present state. In parallel, I will build a Galactic chemical evolution model, with input parameters that are as empirically based as possible, that will reproduce and explain the observations. To get those input parameters, I will measure the rates of supernovae (SNe) in nearby galaxies (using data from past and ongoing surveys) and in high-redshift proto-clusters (by conducting a SN search with JWST), to bring into sharp focus the element yields of SNe and the distribution of delay times (the DTD) between star formation and SN explosion. These empirically determined SN metal-production parameters will be used to find the observationally based reconstruction of the Galaxy’s stellar formation history and chemical evolution that reproduces the observed present-day Milky Way stellar population. The population census of stellar multiplicity with Gaia+SDSS-V, and particularly of short-orbit compact-object binaries, will hark back to the rates and the element yields of the various types of SNe, revealing the connections between various progenitor systems, their explosions, and their rates. The plan, while ambitious, is feasible, thanks to the data from these truly game-changing observational projects. My team will perform all steps of the analysis and will combine the results to obtain the clearest picture of how our Galaxy came to be.
Summary
Understanding how the Milky Way arrived at its present state requires a large volume of precision measurements of our Galaxy’s current makeup, as well as an empirically based understanding of the main processes involved in the Galaxy’s evolution. Such data are now about to arrive in the flood of quality information from Gaia and SDSS-V. The demography of the stars and of the compact stellar remnants in our Galaxy, in terms of phase-space location, mass, age, metallicity, and multiplicity are data products that will come directly from these surveys. I propose to integrate this information into a comprehensive picture of the Milky Way’s present state. In parallel, I will build a Galactic chemical evolution model, with input parameters that are as empirically based as possible, that will reproduce and explain the observations. To get those input parameters, I will measure the rates of supernovae (SNe) in nearby galaxies (using data from past and ongoing surveys) and in high-redshift proto-clusters (by conducting a SN search with JWST), to bring into sharp focus the element yields of SNe and the distribution of delay times (the DTD) between star formation and SN explosion. These empirically determined SN metal-production parameters will be used to find the observationally based reconstruction of the Galaxy’s stellar formation history and chemical evolution that reproduces the observed present-day Milky Way stellar population. The population census of stellar multiplicity with Gaia+SDSS-V, and particularly of short-orbit compact-object binaries, will hark back to the rates and the element yields of the various types of SNe, revealing the connections between various progenitor systems, their explosions, and their rates. The plan, while ambitious, is feasible, thanks to the data from these truly game-changing observational projects. My team will perform all steps of the analysis and will combine the results to obtain the clearest picture of how our Galaxy came to be.
Max ERC Funding
1 859 375 €
Duration
Start date: 2019-10-01, End date: 2024-09-30
Project acronym FIAT
Project The Foundations of Institutional AuThority: a multi-dimensional model of the separation of powers
Researcher (PI) Eoin CAROLAN
Host Institution (HI) UNIVERSITY COLLEGE DUBLIN, NATIONAL UNIVERSITY OF IRELAND, DUBLIN
Call Details Consolidator Grant (CoG), SH2, ERC-2018-COG
Summary ‘Almost three centuries later, it is past time to rethink Montesquieu’s holy trinity’ (Ackerman, 2010).
As Ackerman (and many others) have observed, political reality has long left the traditional model of the separation of powers behind. The problems posed by this gap between constitutional theory and political practice have recently acquired fresh urgency as developments in Hungary, Poland, Turkey, Russia, the UK, US, Bolivia and elsewhere place the separation of powers under strain. These include the emergence of authoritarian leaders; personalisation of political authority; recourse to non-legal plebiscites; and the capture or de-legitimisation of other constitutional bodies.
This project argues that these difficulties are rooted in a deeper problem with constitutional thinking about institutional power: a constitution-as-law approach that equates the conferral of legal power with the authority to exercise it. This makes it possible for a gap to emerge between legal accounts of authority and its diverse –and potentially conflicting (Cotterrell)– sociological foundations. Where that gap exists, the practical authority of an institution (or constitution) may be vulnerable to challenge from rival and more socially-resonant claims (Scheppele (2017)).
It is this gap between legal norms and social facts that the project aims to investigate – and ultimately bridge.
How is authority established? How is it maintained? How might it fail? And how does the constitution (as rule? representation (Saward)? mission statement (King)?) shape, re-shape and come to be shaped by those processes? By investigating these questions across six case studies, the project will produce a multi-dimensional account of institutional authority that takes seriously the sociological influence of both law and culture.
The results from these cases provide the evidential foundation for the project’s final outputs: a new model and new evaluative measures of the separation of powers.
Summary
‘Almost three centuries later, it is past time to rethink Montesquieu’s holy trinity’ (Ackerman, 2010).
As Ackerman (and many others) have observed, political reality has long left the traditional model of the separation of powers behind. The problems posed by this gap between constitutional theory and political practice have recently acquired fresh urgency as developments in Hungary, Poland, Turkey, Russia, the UK, US, Bolivia and elsewhere place the separation of powers under strain. These include the emergence of authoritarian leaders; personalisation of political authority; recourse to non-legal plebiscites; and the capture or de-legitimisation of other constitutional bodies.
This project argues that these difficulties are rooted in a deeper problem with constitutional thinking about institutional power: a constitution-as-law approach that equates the conferral of legal power with the authority to exercise it. This makes it possible for a gap to emerge between legal accounts of authority and its diverse –and potentially conflicting (Cotterrell)– sociological foundations. Where that gap exists, the practical authority of an institution (or constitution) may be vulnerable to challenge from rival and more socially-resonant claims (Scheppele (2017)).
It is this gap between legal norms and social facts that the project aims to investigate – and ultimately bridge.
How is authority established? How is it maintained? How might it fail? And how does the constitution (as rule? representation (Saward)? mission statement (King)?) shape, re-shape and come to be shaped by those processes? By investigating these questions across six case studies, the project will produce a multi-dimensional account of institutional authority that takes seriously the sociological influence of both law and culture.
The results from these cases provide the evidential foundation for the project’s final outputs: a new model and new evaluative measures of the separation of powers.
Max ERC Funding
1 997 628 €
Duration
Start date: 2020-01-01, End date: 2024-12-31
Project acronym FTHPC
Project Fault Tolerant High Performance Computing
Researcher (PI) Oded Schwartz
Host Institution (HI) THE HEBREW UNIVERSITY OF JERUSALEM
Call Details Consolidator Grant (CoG), PE6, ERC-2018-COG
Summary Supercomputers are strategically crucial for facilitating advances in science and technology: in climate change research, accelerated genome sequencing towards cancer treatments, cutting edge physics, devising engineering innovative solutions, and many other compute intensive problems. However, the future of super-computing depends on our ability to cope with the ever increasing rate of faults (bit flips and component failure), resulting from the steadily increasing machine size and decreasing operating voltage. Indeed, hardware trends predict at least two faults per minute for next generation (exascale) supercomputers.
The challenge of ascertaining fault tolerance for high-performance computing is not new, and has been the focus of extensive research for over two decades. However, most solutions are either (i) general purpose, requiring little to no algorithmic effort, but severely degrading performance (e.g., checkpoint-restart), or (ii) tailored to specific applications and very efficient, but requiring high expertise and significantly increasing programmers' workload. We seek the best of both worlds: high performance and general purpose fault resilience.
Efficient general purpose solutions (e.g., via error correcting codes) have revolutionized memory and communication devices over two decades ago, enabling programmers to effectively disregard the very
likely memory and communication errors. The time has come for a similar paradigm shift in the computing regimen. I argue that exciting recent advances in error correcting codes, and in short probabilistically checkable proofs, make this goal feasible. Success along these lines will eliminate the bottleneck of required fault-tolerance expertise, and open exascale computing to all algorithm designers and programmers, for the benefit of the scientific, engineering, and industrial communities.
Summary
Supercomputers are strategically crucial for facilitating advances in science and technology: in climate change research, accelerated genome sequencing towards cancer treatments, cutting edge physics, devising engineering innovative solutions, and many other compute intensive problems. However, the future of super-computing depends on our ability to cope with the ever increasing rate of faults (bit flips and component failure), resulting from the steadily increasing machine size and decreasing operating voltage. Indeed, hardware trends predict at least two faults per minute for next generation (exascale) supercomputers.
The challenge of ascertaining fault tolerance for high-performance computing is not new, and has been the focus of extensive research for over two decades. However, most solutions are either (i) general purpose, requiring little to no algorithmic effort, but severely degrading performance (e.g., checkpoint-restart), or (ii) tailored to specific applications and very efficient, but requiring high expertise and significantly increasing programmers' workload. We seek the best of both worlds: high performance and general purpose fault resilience.
Efficient general purpose solutions (e.g., via error correcting codes) have revolutionized memory and communication devices over two decades ago, enabling programmers to effectively disregard the very
likely memory and communication errors. The time has come for a similar paradigm shift in the computing regimen. I argue that exciting recent advances in error correcting codes, and in short probabilistically checkable proofs, make this goal feasible. Success along these lines will eliminate the bottleneck of required fault-tolerance expertise, and open exascale computing to all algorithm designers and programmers, for the benefit of the scientific, engineering, and industrial communities.
Max ERC Funding
1 824 467 €
Duration
Start date: 2019-06-01, End date: 2024-05-31
Project acronym HARMONIC
Project Discrete harmonic analysis for computer science
Researcher (PI) Yuval FILMUS
Host Institution (HI) TECHNION - ISRAEL INSTITUTE OF TECHNOLOGY
Call Details Starting Grant (StG), PE6, ERC-2018-STG
Summary Boolean function analysis is a topic of research at the heart of theoretical computer science. It studies functions on n input bits (for example, functions computed by Boolean circuits) from a spectral perspective, by treating them as real-valued functions on the group Z_2^n, and using techniques from Fourier and functional analysis. Boolean function analysis has been applied to a wide variety of areas within theoretical computer science, including hardness of approximation, learning theory, coding theory, and quantum complexity theory.
Despite its immense usefulness, Boolean function analysis has limited scope, since it is only appropriate for studying functions on {0,1}^n (a domain known as the Boolean hypercube). Discrete harmonic analysis is the study of functions on domains possessing richer algebraic structure such as the symmetric group (the group of all permutations), using techniques from representation theory and Sperner theory. The considerable success of Boolean function analysis suggests that discrete harmonic analysis could likewise play a central role in theoretical computer science.
The goal of this proposal is to systematically develop discrete harmonic analysis on a broad variety of domains, with an eye toward applications in several areas of theoretical computer science. We will generalize classical results of Boolean function analysis beyond the Boolean hypercube, to domains such as finite groups, association schemes (a generalization of finite groups), the quantum analog of the Boolean hypercube, and high-dimensional expanders (high-dimensional analogs of expander graphs). Potential applications include a quantum PCP theorem and two outstanding open questions in hardness of approximation: the Unique Games Conjecture and the Sliding Scale Conjecture. Beyond these concrete applications, we expect that the fundamental results we prove will have many other applications that are hard to predict in advance.
Summary
Boolean function analysis is a topic of research at the heart of theoretical computer science. It studies functions on n input bits (for example, functions computed by Boolean circuits) from a spectral perspective, by treating them as real-valued functions on the group Z_2^n, and using techniques from Fourier and functional analysis. Boolean function analysis has been applied to a wide variety of areas within theoretical computer science, including hardness of approximation, learning theory, coding theory, and quantum complexity theory.
Despite its immense usefulness, Boolean function analysis has limited scope, since it is only appropriate for studying functions on {0,1}^n (a domain known as the Boolean hypercube). Discrete harmonic analysis is the study of functions on domains possessing richer algebraic structure such as the symmetric group (the group of all permutations), using techniques from representation theory and Sperner theory. The considerable success of Boolean function analysis suggests that discrete harmonic analysis could likewise play a central role in theoretical computer science.
The goal of this proposal is to systematically develop discrete harmonic analysis on a broad variety of domains, with an eye toward applications in several areas of theoretical computer science. We will generalize classical results of Boolean function analysis beyond the Boolean hypercube, to domains such as finite groups, association schemes (a generalization of finite groups), the quantum analog of the Boolean hypercube, and high-dimensional expanders (high-dimensional analogs of expander graphs). Potential applications include a quantum PCP theorem and two outstanding open questions in hardness of approximation: the Unique Games Conjecture and the Sliding Scale Conjecture. Beyond these concrete applications, we expect that the fundamental results we prove will have many other applications that are hard to predict in advance.
Max ERC Funding
1 473 750 €
Duration
Start date: 2019-03-01, End date: 2024-02-29
Project acronym HOLI
Project Deep Learning for Holistic Inference
Researcher (PI) Amir Globerson
Host Institution (HI) TEL AVIV UNIVERSITY
Call Details Consolidator Grant (CoG), PE6, ERC-2018-COG
Summary Machine learning has rapidly evolved in the last decade, significantly improving accuracy on tasks such as image classification. Much of this success can be attributed to the re-emergence of neural nets. However, learning algorithms are still far from achieving the capabilities of human cognition. In particular, humans can rapidly organize an input stream (e.g., textual or visual) into a set of entities, and understand the complex relations between those. In this project I aim to create a general methodology for semantic interpretation of input streams. Such problems fall under the structured-prediction framework, to which I have made numerous contributions. The proposal identifies and addresses three key components required for a comprehensive and empirically effective approach to the problem.
First, we consider the holistic nature of semantic interpretations, where a top-down process chooses a coherent interpretation among the vast number of options. We argue that deep-learning architectures are ideally suited for modeling such coherence scores, and propose to develop the corresponding theory and algorithms. Second, we address the complexity of the semantic representation, where a stream is mapped into a variable number of entities, each having multiple attributes and relations to other entities. We characterize the properties a model should satisfy in order to produce such interpretations, and propose novel models that achieve this. Third, we develop a theory for understanding when such models can be learned efficiently, and how well they can generalize. To achieve this, we address key questions of non-convex optimization, inductive bias and generalization. We expect these contributions to have a dramatic impact on AI systems, from machine reading of text to image analysis. More broadly, they will help bridge the gap between machine learning as an engineering field, and the study of human cognition.
Summary
Machine learning has rapidly evolved in the last decade, significantly improving accuracy on tasks such as image classification. Much of this success can be attributed to the re-emergence of neural nets. However, learning algorithms are still far from achieving the capabilities of human cognition. In particular, humans can rapidly organize an input stream (e.g., textual or visual) into a set of entities, and understand the complex relations between those. In this project I aim to create a general methodology for semantic interpretation of input streams. Such problems fall under the structured-prediction framework, to which I have made numerous contributions. The proposal identifies and addresses three key components required for a comprehensive and empirically effective approach to the problem.
First, we consider the holistic nature of semantic interpretations, where a top-down process chooses a coherent interpretation among the vast number of options. We argue that deep-learning architectures are ideally suited for modeling such coherence scores, and propose to develop the corresponding theory and algorithms. Second, we address the complexity of the semantic representation, where a stream is mapped into a variable number of entities, each having multiple attributes and relations to other entities. We characterize the properties a model should satisfy in order to produce such interpretations, and propose novel models that achieve this. Third, we develop a theory for understanding when such models can be learned efficiently, and how well they can generalize. To achieve this, we address key questions of non-convex optimization, inductive bias and generalization. We expect these contributions to have a dramatic impact on AI systems, from machine reading of text to image analysis. More broadly, they will help bridge the gap between machine learning as an engineering field, and the study of human cognition.
Max ERC Funding
1 932 500 €
Duration
Start date: 2019-02-01, End date: 2024-01-31
Project acronym iEXTRACT
Project Information Extraction for Everyone
Researcher (PI) Yoav Goldberg
Host Institution (HI) BAR ILAN UNIVERSITY
Call Details Starting Grant (StG), PE6, ERC-2018-STG
Summary Staggering amounts of information are stored in natural language documents, rendering them unavailable to data-science techniques. Information Extraction (IE), a subfield of Natural Language Processing (NLP), aims to automate the extraction of structured information from text, yielding datasets that can be queried, analyzed and combined to provide new insights and drive research forward.
Despite tremendous progress in NLP, IE systems remain mostly inaccessible to non-NLP-experts who can greatly benefit from them. This stems from the current methods for creating IE systems: the dominant machine-learning (ML) approach requires technical expertise and large amounts of annotated data, and does not provide the user control over the extraction process. The previously dominant rule-based approach unrealistically requires the user to anticipate and deal with the nuances of natural language.
I aim to remedy this situation by revisiting rule-based IE in light of advances in NLP and ML. The key idea is to cast IE as a collaborative human-computer effort, in which the user provides domain-specific knowledge, and the system is in charge of solving various domain-independent linguistic complexities, ultimately allowing the user to query
unstructured texts via easily structured forms.
More specifically, I aim develop:
(a) a novel structured representation that abstracts much of the complexity of natural language;
(b) algorithms that derive these representations from texts;
(c) an accessible rule language to query this representation;
(d) AI components that infer the user extraction intents, and based on them promote relevant examples and highlight extraction cases that require special attention.
The ultimate goal of this project is to democratize NLP and bring advanced IE capabilities directly to the hands of
domain-experts: doctors, lawyers, researchers and scientists, empowering them to process large volumes of data and
advance their profession.
Summary
Staggering amounts of information are stored in natural language documents, rendering them unavailable to data-science techniques. Information Extraction (IE), a subfield of Natural Language Processing (NLP), aims to automate the extraction of structured information from text, yielding datasets that can be queried, analyzed and combined to provide new insights and drive research forward.
Despite tremendous progress in NLP, IE systems remain mostly inaccessible to non-NLP-experts who can greatly benefit from them. This stems from the current methods for creating IE systems: the dominant machine-learning (ML) approach requires technical expertise and large amounts of annotated data, and does not provide the user control over the extraction process. The previously dominant rule-based approach unrealistically requires the user to anticipate and deal with the nuances of natural language.
I aim to remedy this situation by revisiting rule-based IE in light of advances in NLP and ML. The key idea is to cast IE as a collaborative human-computer effort, in which the user provides domain-specific knowledge, and the system is in charge of solving various domain-independent linguistic complexities, ultimately allowing the user to query
unstructured texts via easily structured forms.
More specifically, I aim develop:
(a) a novel structured representation that abstracts much of the complexity of natural language;
(b) algorithms that derive these representations from texts;
(c) an accessible rule language to query this representation;
(d) AI components that infer the user extraction intents, and based on them promote relevant examples and highlight extraction cases that require special attention.
The ultimate goal of this project is to democratize NLP and bring advanced IE capabilities directly to the hands of
domain-experts: doctors, lawyers, researchers and scientists, empowering them to process large volumes of data and
advance their profession.
Max ERC Funding
1 499 354 €
Duration
Start date: 2019-05-01, End date: 2024-04-30