Catalogue Search | MBRL
Search Results Heading
Explore the vast range of titles available.
MBRLSearchResults
-
DisciplineDiscipline
-
Is Peer ReviewedIs Peer Reviewed
-
Item TypeItem Type
-
SubjectSubject
-
YearFrom:-To:
-
More FiltersMore FiltersSourceLanguage
Done
Filters
Reset
1,007
result(s) for
"Colloquium Papers"
Sort by:
On instabilities of deep learning in image reconstruction and the potential costs of AI
by
Antun, Vegard
,
Renna, Francesco
,
Hansen, Anders C.
in
Algorithms
,
Applied Mathematics
,
COLLOQUIUM PAPERS
2020
Deep learning, due to its unprecedented success in tasks such as image classification, has emerged as a new tool in image reconstruction with potential to change the field. In this paper, we demonstrate a crucial phenomenon: Deep learning typically yields unstable methods for image reconstruction. The instabilities usually occur in several forms: 1) Certain tiny, almost undetectable perturbations, both in the image and sampling domain, may result in severe artefacts in the reconstruction; 2) a small structural change, for example, a tumor, may not be captured in the reconstructed image; and 3) (a counterintuitive type of instability) more samples may yield poorer performance. Our stability test with algorithms and easy-to-use software detects the instability phenomena. The test is aimed at researchers, to test their networks for instabilities, and for government agencies, such as the Food and Drug Administration (FDA), to secure safe use of deep learning methods.
Journal Article
Benign overfitting in linear regression
by
Lugosi, Gábor
,
Long, Philip M.
,
Tsigler, Alexander
in
Accuracy
,
Artificial neural networks
,
Benign
2020
The phenomenon of benign overfitting is one of the key mysteries uncovered by deep learning methodology: deep neural networks seem to predict well, even with a perfect fit to noisy training data. Motivated by this phenomenon, we consider when a perfect fit to training data in linear regression is compatible with accurate prediction. We give a characterization of linear regression problems for which the minimum norm interpolating prediction rule has near-optimal prediction accuracy. The characterization is in terms of two notions of the effective rank of the data covariance. It shows that overparameterization is essential for benign overfitting in this setting: the number of directions in parameter space that are unimportant for prediction must significantly exceed the sample size. By studying examples of data covariance properties that this characterization shows are required for benign overfitting, we find an important role for finite-dimensional data: the accuracy of the minimum norm interpolating prediction rule approaches the best possible accuracy for a much narrower range of properties of the data distribution when the data lie in an infinite-dimensional space vs. when the data lie in a finite-dimensional space with dimension that grows faster than the sample size.
Journal Article
The frontier of simulation-based inference
by
Cranmer, Kyle
,
Louppe, Gilles
,
Brehmer, Johann
in
Approximate Bayesian Computation
,
COLLOQUIUM PAPERS
,
Computer science
2020
Many domains of science have developed complex simulations to describe phenomena of interest. While these simulations provide high-fidelity models, they are poorly suited for inference and lead to challenging inverse problems. We review the rapidly developing field of simulation-based inference and identify the forces giving additional momentum to the field. Finally, we describe how the frontier is expanding so that a broad audience can appreciate the profound influence these developments may have on science.
Journal Article
Science audiences, misinformation, and fake news
by
Krause, Nicole M.
,
Scheufele, Dietram A.
in
Arthur M. Sackler on the Science of Science Communication III
,
COLLOQUIUM PAPERS
,
Communication
2019
Concerns about public misinformation in the United States—ranging from politics to science—are growing. Here, we provide an overview of how and why citizens become (and sometimes remain) misinformed about science. Our discussion focuses specifically on misinformation among individual citizens. However, it is impossible to understand individual information processing and acceptance without taking into account social networks, information ecologies, and other macro-level variables that provide important social context. Specifically, we show how being misinformed is a function of a person’s ability and motivation to spot falsehoods, but also of other group-level and societal factors that increase the chances of citizens to be exposed to correct(ive) information. We conclude by discussing a number of research areas—some of which echo themes of the 2017 National Academies of Sciences, Engineering, and Medicine’s Communicating Science Effectively report—that will be particularly important for our future understanding of misinformation, specifically a systems approach to the problem of misinformation, the need for more systematic analyses of science communication in new media environments, and a (re)focusing on traditionally underserved audiences.
Journal Article
Understanding the role of individual units in a deep neural network
by
Strobelt, Hendrik
,
Bau, David
,
Lapedriza, Agata
in
Artificial neural networks
,
Classification
,
COLLOQUIUM PAPERS
2020
Deep neural networks excel at finding hierarchical representations that solve complex tasks over large datasets. How can we humans understand these learned representations? In this work, we present network dissection, an analytic framework to systematically identify the semantics of individual hidden units within image classification and image generation networks. First, we analyze a convolutional neural network (CNN) trained on scene classification and discover units that match a diverse set of object concepts. We find evidence that the network has learned many object classes that play crucial roles in classifying scene classes. Second, we use a similar analytic method to analyze a generative adversarial network (GAN) model trained to generate scenes. By analyzing changes made when small sets of units are activated or deactivated, we find that objects can be added and removed from the output scenes while adapting to the context. Finally, we apply our analytic framework to understanding adversarial attacks and to semantic image editing.
Journal Article
The unreasonable effectiveness of deep learning in artificial intelligence
2020
Deep learning networks have been trained to recognize speech, caption photographs, and translate text between languages at high levels of performance. Although applications of deep learning networks to real-world problems have become ubiquitous, our understanding of why they are so effective is lacking. These empirical results should not be possible according to sample complexity in statistics and nonconvex optimization theory. However, paradoxes in the training and effectiveness of deep learning networks are being investigated and insights are being found in the geometry of high-dimensional spaces. A mathematical theory of deep learning would illuminate how they function, allow us to assess the strengths and weaknesses of different network architectures, and lead to major improvements. Deep learning has provided natural ways for humans to communicate with digital devices and is foundational for building artificial general intelligence. Deep learning was inspired by the architecture of the cerebral cortex and insights into autonomy and general intelligence may be found in other brain regions that are essential for planning and survival, but major breakthroughs will be needed to achieve these goals.
Journal Article
Emergent linguistic structure in artificial neural networks trained by self-supervision
by
Hewitt, John
,
Khandelwal, Urvashi
,
Levy, Omer
in
Artificial neural networks
,
COLLOQUIUM PAPERS
,
Computer Sciences
2020
This paper explores the knowledge of linguistic structure learned by large artificial neural networks, trained via self-supervision, whereby the model simply tries to predict a masked word in a given context. Human language communication is via sequences of words, but language understanding requires constructing rich hierarchical structures that are never observed explicitly. The mechanisms for this have been a prime mystery of human language acquisition, while engineering work has mainly proceeded by supervised learning on treebanks of sentences hand labeled for this latent structure. However, we demonstrate that modern deep contextual language models learn major aspects of this structure, without any explicit supervision. We develop methods for identifying linguistic hierarchical structure emergent in artificial neural networks and demonstrate that components in these models focus on syntactic grammatical relationships and anaphoric coreference. Indeed, we show that a linear transformation of learned embeddings in these models captures parse tree distances to a surprising degree, allowing approximate reconstruction of the sentence tree structures normally assumed by linguists. These results help explain why these models have brought such large improvements across many language-understanding tasks.
Journal Article
Misinformation in and about science
by
West, Jevin D.
,
Bergstrom, Carl T.
in
Arthur M. Sackler on Advancing the Science and Practice of Science Communication: Misinformation about Science in the Public Sphere
,
Biomedical Research - ethics
,
Climate change
2021
Humans learn about the world by collectively acquiring information, filtering it, and sharing what we know. Misinformation undermines this process. The repercussions are extensive. Without reliable and accurate sources of information, we cannot hope to halt climate change, make reasoned democratic decisions, or control a global pandemic. Most analyses of misinformation focus on popular and social media, but the scientific enterprise faces a parallel set of problems—from hype and hyperbole to publication bias and citation misdirection, predatory publishing, and filter bubbles. In this perspective, we highlight these parallels and discuss future research directions and interventions.
Journal Article
Fast reinforcement learning with generalized policy updates
by
Barreto, André
,
Precup, Doina
,
Hou, Shaobo
in
COLLOQUIUM PAPERS
,
Computer Sciences
,
Decision making
2020
The combination of reinforcement learning with deep learning is a promising approach to tackle important sequential decision-making problems that are currently intractable. One obstacle to overcome is the amount of data needed by learning systems of this type. In this article, we propose to address this issue through a divide-and-conquer approach. We argue that complex decision problems can be naturally decomposed into multiple tasks that unfold in sequence or in parallel. By associating each task with a reward function, this problem decomposition can be seamlessly accommodated within the standard reinforcement-learning formalism. The specific way we do so is through a generalization of two fundamental operations in reinforcement learning: policy improvement and policy evaluation. The generalized version of these operations allow one to leverage the solution of some tasks to speed up the solution of others. If the reward function of a task can be well approximated as a linear combination of the reward functions of tasks previously solved, we can reduce a reinforcement-learning problem to a simpler linear regression. When this is not the case, the agent can still exploit the task solutions by using them to interact with and learn about the environment. Both strategies considerably reduce the amount of data needed to solve a reinforcement-learning problem.
Journal Article
Renewable electricity storage using electrolysis
by
Turner, John A.
,
Mallouk, Thomas E.
,
Hitt, Jeremy L.
in
Arthur M. Sackler on the Status and Challenges in Decarbonizing our Energy Landscape
,
Carbon dioxide
,
Chemical bonds
2020
Electrolysis converts electrical energy into chemical energy by storing electrons in the form of stable chemical bonds. The chemical energy can be used as a fuel or converted back to electricity when needed. Water electrolysis to hydrogen and oxygen is a well-established technology, whereas fundamental advances in CO₂ electrolysis are still needed to enable short-term and seasonal energy storage in the form of liquid fuels. This paper discusses the electrolytic reactions that can potentially enable renewable energy storage, including water, CO₂ and N₂ electrolysis. Recent progress and major obstacles associated with electrocatalysis and mass transfer management at a system level are reviewed. We conclude that knowledge and strategies are transferable between these different electrochemical technologies, although there are also unique complications that arise from the specifics of the reactions involved.
Journal Article