Annual Reviews home
0
Skip to content
  • For Librarians & Agents
  • For Authors
  • Knowable Magazine
  • Institutional Login
  • Login
  • Register
  • Activate
  • 0 Cart
  • Help
Annual Reviews home
  • JOURNALS A-Z
    • Analytical Chemistry
    • Animal Biosciences
    • Anthropology
    • Astronomy and Astrophysics
    • Biochemistry
    • Biomedical Data Science
    • Biomedical Engineering
    • Biophysics
    • Cancer Biology
    • Cell and Developmental Biology
    • Chemical and Biomolecular Engineering
    • Clinical Psychology
    • Computer Science
    • Condensed Matter Physics
    • Control, Robotics, and Autonomous Systems
    • Criminology
    • Developmental Psychology
    • Earth and Planetary Sciences
    • Ecology, Evolution, and Systematics
    • Economics
    • Entomology
    • Environment and Resources
    • Financial Economics
    • Fluid Mechanics
    • Food Science and Technology
    • Genetics
    • Genomics and Human Genetics
    • Immunology
    • Law and Social Science
    • Linguistics
    • Marine Science
    • Materials Research
    • Medicine
    • Microbiology
    • Neuroscience
    • Nuclear and Particle Science
    • Nutrition
    • Organizational Psychology and Organizational Behavior
    • Pathology: Mechanisms of Disease
    • Pharmacology and Toxicology
    • Physical Chemistry
    • Physiology
    • Phytopathology
    • Plant Biology
    • Political Science
    • Psychology
    • Public Health
    • Resource Economics
    • Sociology
    • Statistics and Its Application
    • Virology
    • Vision Science
    • Article Collections
    • Events
    • Shot of Science
  • JOURNAL INFO
    • Copyright & Permissions
    • Add To Your Course Reader
    • Expected Publication Dates
    • Impact Factor Rankings
    • Access Metadata
    • RSS Feeds
  • PRICING & SUBSCRIPTIONS
    • General Ordering Info
    • Online Activation Instructions
    • Personal Pricing
    • Institutional Pricing
    • Society Partnerships
  •     S2O    
  •     GIVE    
  • ABOUT
    • What We Do
    • Founder & History
    • Our Team
    • Careers
    • Press Center
    • Events
    • News
    • Global Access
    • DEI
    • Directory
    • Help/FAQs
    • Contact Us
  • Home >
  • Annual Review of Fluid Mechanics >
  • Volume 52, 2020 >
  • Brunton, pp 477-508
  • Save
  • Email
  • Share

Machine Learning for Fluid Mechanics

  • Home
  • Annual Review of Fluid Mechanics
  • Volume 52, 2020
  • Brunton, pp 477-508
  • Facebook
  • Twitter
  • LinkedIn
Download PDF

Machine Learning for Fluid Mechanics

Annual Review of Fluid Mechanics

Vol. 52:477-508 (Volume publication date January 2020)
First published as a Review in Advance on September 12, 2019
https://doi.org/10.1146/annurev-fluid-010719-060214

Steven L. Brunton,1 Bernd R. Noack,2,3 and Petros Koumoutsakos4

1Department of Mechanical Engineering, University of Washington, Seattle, Washington 98195, USA

2LIMSI (Laboratoire d'Informatique pour la Mécanique et les Sciences de l'Ingénieur), CNRS UPR 3251, Université Paris-Saclay, F-91403 Orsay, France

3Institut für Strömungsmechanik und Technische Akustik, Technische Universität Berlin, D-10634 Berlin, Germany

4Computational Science and Engineering Laboratory, ETH Zurich, CH-8092 Zurich, Switzerland; email: [email protected]

Download PDF Article Metrics
  • Permissions
  • Reprints

  • Download Citation
  • Citation Alerts
Sections
  • Abstract
  • Keywords
  • INTRODUCTION
  • MACHINE LEARNING FUNDAMENTALS
  • FLOW MODELING WITH MACHINE LEARNING
  • FLOW OPTIMIZATION AND CONTROL USING MACHINE LEARNING
  • DISCUSSION AND OUTLOOK
  • SUMMARY POINTS
  • FUTURE ISSUES
  • disclosure statement
  • acknowledgments
  • literature cited

Abstract

The field of fluid mechanics is rapidly advancing, driven by unprecedented volumes of data from experiments, field measurements, and large-scale simulations at multiple spatiotemporal scales. Machine learning (ML) offers a wealth of techniques to extract information from data that can be translated into knowledge about the underlying fluid mechanics. Moreover, ML algorithms can augment domain knowledge and automate tasks related to flow control and optimization. This article presents an overview of past history, current developments, and emerging opportunities of ML for fluid mechanics. We outline fundamental ML methodologies and discuss their uses for understanding, modeling, optimizing, and controlling fluid flows. The strengths and limitations of these methods are addressed from the perspective of scientific inquiry that considers data as an inherent part of modeling, experiments, and simulations. ML provides a powerful information-processing framework that can augment, and possibly even transform, current lines of fluid mechanics research and industrial applications.

Keywords

machine learning, data-driven modeling, optimization, control

1. INTRODUCTION

Fluid mechanics has traditionally dealt with massive amounts of data from experiments, field measurements, and large-scale numerical simulations. Indeed, in the past few decades, big data have been a reality in fluid mechanics research (Pollard et al. 2016) due to high-performance computing architectures and advances in experimental measurement capabilities. Over the past 50 years, many techniques were developed to handle such data, ranging from advanced algorithms for data processing and compression to fluid mechanics databases (Perlman et al. 2007, Wu & Moin 2008). However, the analysis of fluid mechanics data has relied, to a large extent, on domain expertise, statistical analysis, and heuristic algorithms.

The growth of data today is widespread across scientific disciplines, and gaining insight and actionable information from data has become a new mode of scientific inquiry as well as a commercial opportunity. Our generation is experiencing an unprecedented confluence of (a) vast and increasing volumes of data; (b) advances in computational hardware and reduced costs for computation, data storage, and transfer; (c) sophisticated algorithms; (d) an abundance of open source software and benchmark problems; and (e) significant and ongoing investment by industry on data-driven problem solving. These advances have, in turn, fueled renewed interest and progress in the field of machine learning (ML) to extract information from these data. ML is now rapidly making inroads in fluid mechanics. These learning algorithms may be categorized into supervised, semisupervised, and unsupervised learning (see Figure 1), depending on the information available about the data to the learning machine (LM).

figure
Figure 1 

ML provides a modular and agile modeling framework that can be tailored to address many challenges in fluid mechanics, such as reduced-order modeling, experimental data processing, shape optimization, turbulence closure modeling, and control. As scientific inquiry shifts from first principles to data-driven approaches, we may draw a parallel with the development of numerical methods in the 1940s and 1950s to solve the equations of fluid dynamics. Fluid mechanics stands to benefit from learning algorithms and in return presents challenges that may further advance these algorithms to complement human understanding and engineering intuition.

In addition to outlining successes, we must note the importance of understanding how learning algorithms work and when these methods succeed or fail. It is important to balance excitement about the capabilities of ML with the reality that its application to fluid mechanics is an open and challenging field. In this context, we also highlight the benefit of incorporating domain knowledge about fluid mechanics into learning algorithms. We envision that the fluid mechanics community can contribute to advances in ML reminiscent of advances in numerical methods in the last century.

1.1. Historical Overview

The interface between ML and fluid dynamics has a long and possibly surprising history. In the early 1940s, Kolmogorov, a founder of statistical learning theory, considered turbulence as one of its prime application domains (Kolmogorov 1941). Advances in ML in the 1950s and 1960s were characterized by two distinct developments. On one side, we may distinguish cybernetics (Wiener 1965) and expert systems designed to emulate the thinking process of the human brain, and on the other side, machines such as the perceptron (Rosenblatt 1958) aimed to automate processes such as classification and regression. The use of perceptrons for classification created significant excitement. However, this excitement was quenched by findings that their capabilities had severe limitations (Minsky & Papert 1969): Single-layer perceptrons were only able to learn linearly separable functions and were not capable of learning the XOR function. It was known that multilayer perceptrons could learn the XOR function, but perhaps their advancement was limited given the computational resources of the times (a recurring theme in ML research). The reduced interest in perceptrons was soon accompanied by a reduced interest in artificial intelligence (AI) in general.

Another branch of ML, closely related to budding ideas of cybernetics in the early 1960s, was pioneered by two graduate students: Ingo Rechenberg and Hans-Paul Schwefel at the Technical University of Berlin. They performed experiments in a wind tunnel on a corrugated structure composed of five linked plates with the goal of finding their optimal angles to reduce drag (see Figure 2). Their breakthrough involved adding random variations to these angles, where the randomness was generated using a Galton board (an analog random number generator). Most importantly, the size of the variance was learned (increased/decreased) based on the success rate (positive/negative) of the experiments. Despite its brilliance, the work of Rechenberg and Schwefel has received little recognition in the fluid mechanics community, even though a significant number of applications in fluid mechanics and aerodynamics use ideas that can be traced back to their work. Renewed interest in the potential of AI for aerodynamics applications materialized almost simultaneously with the early developments in computational fluid dynamics in the early 1980s. Attention was given to expert systems to assist in aerodynamic design and development processes (Mehta & Kutler 1984).

figure
Figure 2 

An indirect link between fluid mechanics and ML was the so-called Lighthill report in 1974 that criticized AI programs in the United Kingdom as not delivering on their grand claims. This report played a major role in the reduced funding and interest in AI in the United Kingdom and subsequently in the United States that is known as the AI winter. Lighthill's main argument was based on his perception that AI would never be able to address the challenge of the combinatorial explosion between possible configurations in the parameter space. He used the limitations of language processing systems of that time as a key demonstration of the failures for AI. In Lighthill's defense, 40 years ago the powers of modern computers as we know them today may have been difficult to fathom. Indeed, today one may watch Lighthill's speech on the internet while an ML algorithm automatically provides the captions.

The reawakening of interest in ML, and in neural networks (NNs) in particular, came in the late 1980s with the development of the backpropagation algorithm (Rumelhart et al. 1986). This enabled the training of NNs with multiple layers, even though in the early days at most two layers were the norm. Other sources of stimulus were the works by Hopfield (1982), Gardner (1988), and Hinton & Sejnowski (1986), who developed links between ML algorithms and statistical mechanics. However, these developments did not attract many researchers from fluid mechanics. In the early 1990s a number of applications of NNs in flow-related problems were developed in the context of trajectory analysis and classification for particle tracking velocimetry (PTV) and particle image velocimetry (PIV) (Teo et al. 1991, Grant & Pan 1995) as well as for identifying the phase configurations in multiphase flows (Bishop & James 1993). The link between proper orthogonal decomposition (POD) and linear NNs (Baldi & Hornik 1989) was exploited in order to reconstruct turbulence flow fields and the flow in the near-wall region of a channel flow using wall-only information (Milano & Koumoutsakos 2002). This application was one of the first to also use multiple layers of neurons to improve compression results, marking perhaps the first use of deep learning, as it is known today, in the field of fluid mechanics.

In the past few years, we have experienced a renewed blossoming of ML applications in fluid mechanics. Much of this interest is attributed to the remarkable performance of deep learning architectures, which hierarchically extract informative features from data. This has led to several advances in data-rich and model-limited fields such as the social sciences and in companies for which prediction is a key financial factor. Fluid mechanics is not a model-limited field, and it is rapidly becoming data rich. We believe that this confluence of first principles and data-driven approaches is unique and has the potential to transform both fluid mechanics and ML.

1.2. Challenges and Opportunities for Machine Learning in Fluid Dynamics

Fluid dynamics presents challenges that differ from those tackled in many applications of ML, such as image recognition and advertising. In fluid flows it is often important to precisely quantify the underlying physical mechanisms in order to analyze them. Furthermore, fluid flows exhibit complex, multiscale phenomena the understanding and control of which remain largely unresolved. Unsteady flow fields require algorithms capable of addressing nonlinearities and multiple spatiotemporal scales that may not be present in popular ML algorithms. In addition, many prominent applications of ML, such as playing the game Go, rely on inexpensive system evaluations and an exhaustive categorization of the process that must be learned. This is not the case in fluids, where experiments may be difficult to repeat or automate and where simulations may require large-scale supercomputers operating for extended periods of time.

ML has also become instrumental in robotics, and algorithms such as reinforcement learning (RL) are used routinely in autonomous driving and flight. While many robots operate in fluids, it appears that the subtleties of fluid dynamics are not presently a major concern in their design. Reminiscent of the pioneering days of flight, solutions imitating natural forms and processes are often the norm (see the sidebar titled Learning Fluid Mechanics: From Living Organisms to Machines). We believe that deeper understanding and exploitation of fluid mechanics will become critical in the design of robotic devices when their energy consumption and reliability in complex flow environments become a concern.

In the context of flow control, actively or passively manipulating flow dynamics for an engineering objective may change the nature of the system, making predictions based on data of uncontrolled systems impossible. Although flow data are vast in some dimensions, such as spatial resolution, they may be sparse in others; for example, it may be expensive to perform parametric studies. Furthermore, flow data can be highly heterogeneous, requiring special care when choosing the type of LM. In addition, many fluid systems are nonstationary, and even for stationary flows it may be prohibitively expensive to obtain statistically converged results.

Fluid dynamics is central to transportation, health, and defense systems, and it is therefore essential that ML solutions are interpretable, explainable, and generalizable. Moreover, it is often necessary to provide guarantees on performance, which are presently rare. Indeed, there is a poignant lack of convergence results, analysis, and guarantees in many ML algorithms. It is also important to consider whether the model will be used for interpolation within a parameter regime or for extrapolation, which is considerably more challenging. Finally, we emphasize the importance of cross-validation on withheld data sets to prevent overfitting in ML.

LEARNING FLUID MECHANICS: FROM LIVING ORGANISMS TO MACHINES

Birds, bats, insects, fish, whales, and other aquatic and aerial life-forms perform remarkable feats of fluid manipulation, optimizing and controlling their shape and motion to harness unsteady fluid forces for agile propulsion, efficient migration, and other exquisite maneuvers. The range of fluid flow optimization and control observed in biology is breathtaking and has inspired humans for millennia. How do these organisms learn to manipulate the flow environment?

To date, we know of only one species that manipulates fluids through knowledge of the Navier–Stokes equations. Humans have been innovating and engineering devices to harness fluids since before the dawn of recorded history, from dams and irrigation to mills and sailing. Early efforts were achieved through intuitive design, although recent quantitative analysis and physics-based design have enabled a revolution in performance over the past hundred years. Indeed, physics-based engineering of fluid systems is a high-water mark of human achievement. However, there are serious challenges associated with equation-based analysis of fluids, including high dimensionality and nonlinearity, which defy closed-form solutions and limit real-time optimization and control efforts. At the beginning of a new millennium, with increasingly powerful tools in machine learning and data-driven optimization, we are again learning how to learn from experience.

We suggest that this nonexhaustive list of challenges need not be a barrier; to the contrary, it should provide a strong motivation for the development of more effective ML techniques. These techniques will likely impact several disciplines if they are able to solve fluid mechanics problems. The application of ML to systems with known physics, such as fluid mechanics, may provide deeper theoretical insights into algorithms. We also believe that hybrid methods that combine ML and first principles models will be a fertile ground for development.

This review is structured as follows: Section 2 outlines the fundamental algorithms of ML, followed by discussions of their applications to flow modeling (Section 3) and optimization and control (Section 4). We provide a summary and outlook of this field in Section 5.

2. MACHINE LEARNING FUNDAMENTALS

The learning problem can be formulated as the process of estimating associations between inputs, outputs, and parameters of a system using a limited number of observations (Cherkassky & Mulier 2007). We distinguish between a generator of samples, the system in question, and an LM, as in Figure 3. We emphasize that the approximations by LMs are fundamentally stochastic, and their learning process can be summarized as the minimization of a risk functional:

1.
equation 1
where the data (inputs) and (outputs) are samples from a probability distribution p, defines the structure and the parameters of the LM, and the loss function L balances the various learning objectives (e.g., accuracy, simplicity, smoothness, etc.). We emphasize that the risk functional is weighted by a probability distribution that also constrains the predictive capabilities of the LM. The various types of learning algorithms can be grouped into three major categories: supervised, unsupervised, and semisupervised, as in Figure 1. These distinctions signify the degree to which external supervisory information from an expert is available to the LM.
figure
Figure 3 

2.1. Supervised Learning

Supervised learning implies the availability of corrective information to the LM. In its simplest and most common form, this implies labeled training data, with labels corresponding to the output of the LM. Minimization of the cost function, which depends on the training data, will determine the unknown parameters of the LM. In this context, supervised learning dates back to the regression and interpolation methods proposed centuries ago by Gauss (Meijering 2002). A commonly employed loss function is

2.
equation 2
Alternative loss functions may reflect different constraints on the LM such as sparsity (Hastie et al. 2009, Brunton & Kutz 2019). The choice of the approximation function reflects prior knowledge about the data, and the choice between linear and nonlinear methods directly bears on the computational cost associated with the learning methods.

2.1.1. Neural networks.

NNs are arguably the most well-known methods in supervised learning. They are fundamental nonlinear function approximators, and in recent years several efforts have been dedicated to understanding their effectiveness. The universal approximation theorem (Hornik et al. 1989) states that any function may be approximated by a sufficiently large and deep network. Recent work has shown that sparsely connected, deep NNs are information theoretic–optimal nonlinear approximators for a wide range of functions and systems (Bölcskei et al. 2019).

The power and flexibility of NNs emanate from their modular structure based on the neuron as a central building element, a caricature of neurons in the human brain. Each neuron receives an input, processes it through an activation function, and produces an output. Multiple neurons can be combined into different structures that reflect knowledge about the problem and the type of data. Feedforward networks are among the most common structures, and they are composed of layers of neurons, where a weighted output from one layer is the input to the next layer. NN architectures have an input layer that receives the data and an output layer that produces a prediction. Nonlinear optimization methods, such as backpropagation (Rumelhart et al. 1986), are used to identify the network weights to minimize the error between the prediction and labeled training data. Deep NNs involve multiple layers and various types of nonlinear activation functions. When the activation functions are expressed in terms of convolutional kernels, a powerful class of networks emerges, namely convolutional neural networks (CNNs), with great success in image and pattern recognition (Krizhevsky et al. 2012, Goodfellow et al. 2016), Grossberg et al. 1988).

Recurrent neural networks (RNNs), depicted in Figure 4, are of particular interest to fluid mechanics. They operate on sequences of data (e.g., images from a video, time series, etc.), and their weights are obtained by backpropagation through time. RNNs have been quite successful for natural language processing and speech recognition. Their architecture takes into account the inherent order of the data, thus augmenting some of the pioneering applications of classical NNs on signal processing (Rico-Martinez et al. 1992). However, their effectiveness has been hindered by diminishing or exploding gradients that emerge during their training. The renewed interest in RNNs is largely attributed to the development of the long short-term memory (LSTM) (Hochreiter & Schmidhuber 1997) algorithms that deploy cell states and gating mechanisms to store and forget information about past inputs, thus alleviating the problems with gradients and the transmission of long-term information from which standard RNNs suffer. An extended architecture called the multidimensional LSTM network (Graves et al. 2007) was proposed to efficiently handle high-dimensional spatiotemporal data. Several potent alternatives to RNNs have appeared over the years; the echo state network has been used with success in predicting the output of several dynamical systems (Pathak et al. 2018).

figure
Figure 4 

2.1.2. Classification: support vector machines and random forests.

Classification is a supervised learning task that can determine the label or category of a set of measurements from a priori labeled training data. It is perhaps the oldest method for learning, starting with the perceptron (Rosenblatt 1958), which could classify between two types of linearly separable data. Two fundamental classification algorithms are support vector machines (SVMs) (Schölkopf & Smola 2002) and random forests (Breiman 2001), which have been widely adopted in industry. The problem can be specified by the following loss functional, which is expressed here for two classes:

3.
equation 3
The output of the LM is an indicator of the class to which the data belong. The risk functional quantifies the probability of misclassification, and the task is to minimize the risk based on the training data by suitable choice of . Random forests are based on an ensemble of decision trees that hierarchically split the data using simple conditional statements; these decisions are interpretable and fast to evaluate at scale. In the context of classification, an SVM maps the data into a high-dimensional feature space on which a linear classification is possible.

2.2. Unsupervised Learning

This learning task implies the extraction of features from the data by specifying certain global criteria, without the need for supervision or a ground-truth label for the results. The types of problems involved here include dimensionality reduction, quantization, and clustering.

2.2.1. Dimensionality reduction I: proper orthogonal decomposition, principal component analysis, and autoencoders.

The extraction of flow features from experimental data and large-scale simulations is a cornerstone of flow modeling. Moreover, identifying lower-dimensional representations for high-dimensional data can be used as preprocessing for all tasks in supervised learning algorithms. Dimensionality reduction can also be viewed as an information-filtering bottleneck where the data are processed through a lower-dimensional representation before being mapped back to the ambient dimension. The classical POD algorithm belongs to this category of learning and is discussed more in Section 3. The POD, or linear principal component analysis (PCA) as it is more widely known, can be formulated as a two-layer NN (an autoencoder) with a linear activation function for its linearly weighted input, which can be trained by stochastic gradient descent (see Figure 5). This formulation is an algorithmic alternative to linear eigenvalue/eigenvector problems in terms of NNs, and it offers a direct route to the nonlinear regime and deep learning by adding more layers and a nonlinear activation function on the network. Unsupervised learning algorithms have seen limited use in the fluid mechanics community, and we believe that they deserve further exploration. In recent years, the ML community has developed numerous autoencoders that, when properly matched with the possible features of the flow field, can lead to significant insight for reduced-order modeling of stationary and time-dependent data.

figure
Figure 5 

2.2.2. Dimensionality reduction II: discrete principal curves and self-organizing maps.

The mapping between high-dimensional data and a low-dimensional representation can be structured through an explicit shaping of the lower-dimensional space, possibly reflecting an a priori knowledge about this subspace. These techniques can be seen as extensions of the linear autoencoders, where the encoder and decoder can be nonlinear functions. This nonlinearity may, however, come at the expense of losing the inverse relationship between the encoder and decoder functions that is one of the strengths of linear PCA. An alternative is to define the decoder as an approximation of the inverse of the encoder, leading to the method of principal curves. Principal curves are structures on which the data are projected during the encoding step of the learning algorithm. In turn, the decoding step amounts to an approximation of the inverse of this mapping by adding, for example, some smoothing onto the principal curves. An important version of this process is the self-organizing map (SOM) introduced by Grossberg (1976) and Kohonen (1995). In SOMs the projection subspace is described into a finite set of values with specified connectivity architecture and distance metrics. The encoder step amounts to identifying for each data point the closest node point on the SOM, and the decoder step is a weighted regression estimate using, for example, kernel functions that take advantage of the specified distance metric between the map nodes. This modifies the node centers, and the process can be iterated until the empirical risk of the autoencoder has been minimized. The SOM capabilities can be exemplified by comparing it to linear PCA for a two-dimensional set of points. The linear PCA will provide as an approximation the least squares straight line between the points, whereas the SOM will map the points onto a curved line that better approximates the data. We note that SOMs can be extended to areas beyond floating point data, and they offer an interesting way for creating databases based on features of flow fields.

2.2.3. Clustering and vector quantization.

Clustering is an unsupervised learning technique that identifies similar groups in the data. The most common algorithm is k-means clustering, which partitions data into k clusters; an observation belongs to the cluster with the nearest centroid, resulting in a partition of data space into Voronoi cells.

Vector quantizers identify representative points for data that can be partitioned into a predetermined number of clusters. These points can then be used instead of the full data set so that future samples can be approximated by them. The vector quantizer provides a mapping between the data and the coordinates of the cluster centers. The loss function is usually the squared distortion of the data from the cluster centers, which must be minimized to identify the parameters of the quantizer,

4.
equation 4
We note that vector quantization is a data reduction method not necessarily employed for dimensionality reduction. In the latter, the learning problem seeks to identify low-dimensional features in high-dimensional data, whereas quantization amounts to finding representative clusters of the data. Vector quantization must also be distinguished from clustering, as in the former the number of desired centers is determined a priori, whereas clustering aims to identify meaningful groupings in the data. When these groupings are represented by some prototypes, then clustering and quantization have strong similarities.

2.3. Semisupervised Learning

Semisupervised learning algorithms operate under partial supervision, either with limited labeled training data or with other corrective information from the environment. Two algorithms in this category are generative adversarial networks (GANs) and RL. In both cases, the LM is (self-)trained through a game-like process discussed below.

2.3.1. Generative adversarial networks.

GANs are learning algorithms that result in a generative model, i.e., a model that produces data according to a probability distribution that mimics that of the data used for its training. The LM is composed of two networks that compete with each other in a zero-sum game (Goodfellow et al. 2014). The generative network produces candidate data examples that are evaluated by the discriminative, or critic, network to optimize a certain task. The generative network's training objective is to synthesize novel examples of data to fool the discriminative network into misclassifying them as belonging to the true data distribution. The weights of these networks are obtained through a process, inspired by game theory, called adversarial learning. The final objective of the GAN training process is to identify the generative model that produces an output that reflects the underlying system. Labeled data are provided by the discriminator network, and the function to be minimized is the Kullback–Liebler divergence between the two distributions. In the ensuing game, the discriminator aims to maximize the probability of discriminating between true data and data produced by the generator, while the generator aims to minimize the same probability. Because the generative and discriminative networks essentially train themselves, after initialization with labeled training data, this procedure is often called self-supervised. This self-training process adds to the appeal of GANs, but at the same time one must be cautious about whether an equilibrium will ever be reached in the above-mentioned game. As with other training algorithms, large amounts of data help the process, but at the moment, there is no guarantee of convergence.

2.3.2. Reinforcement learning.

RL is a mathematical framework for problem solving (Sutton & Barto 2018) that implies goal-directed interactions of an agent with its environment. In RL the agent has a repertoire of actions and perceives states. Unlike in supervised learning, the agent does not have labeled information about the correct actions but instead learns from its own experiences in the form of rewards that may be infrequent and partial; thus, this is termed semisupervised learning. Moreover, the agent is concerned not only with uncovering patterns in its actions or in the environment but also with maximizing its long-term rewards. RL is closely linked to dynamic programming (Bellman 1952), as it also models interactions with the environment as a Markov decision process. Unlike dynamic programming, RL does not require a model of the dynamics, such as a Markov transition model, but proceeds by repeated interaction with the environment through trial and error. We believe that it is precisely this approximation that makes it highly suitable for complex problems in fluid dynamics. The two central elements of RL are the agent's policy, a mapping between the state s of the system and the optimal action a, and the value function that represents the utility of reaching the state s for maximizing the agent's long-term rewards.

Games are one of the key applications of RL that exemplify its strengths and limitations. One of the early successes of RL is the backgammon learner of Tesauro (1992). The program started out from scratch as a novice player, trained by playing a couple of million times against itself, won the computer backgammon olympiad, and eventually became comparable to the three best human players in the world. In recent years, advances in high-performance computing and deep NN architectures have produced agents that are capable of performing at or above human performance at video games and strategy games much more complicated than backgammon, such as Go (Mnih et al. 2015) and the AI gym (Mnih et al. 2015, Silver et al. 2016). It is important to emphasize that RL requires significant computational resources due to the large numbers of episodes required to properly account for the interaction of the agent and the environment. This cost may be trivial for games, but it may be prohibitive in experiments and flow simulations, a situation that is rapidly changing (Verma et al. 2018).

A core remaining challenge for RL is the long-term credit assignment (LTCA) problem, especially when rewards are sparse or very delayed in time (for example, consider the case of a perching bird or robot). LTCA implies inference, from a long sequence of states and actions, of causal relations between individual decisions and rewards. Several efforts address these issues by augmenting the original sparsely rewarded objective with densely rewarded subgoals (Schaul et al. 2015). A related issue is the proper accounting of past experience by the agent as it actively forms a new policy (Novati et al. 2019).

2.4. Stochastic Optimization: A Learning Algorithms Perspective

Optimization is an inherent part of learning, as a risk functional is minimized in order to identify the parameters of the LM. There is, however, one more link that we wish to highlight in this review: that optimization (and search) algorithms can be cast in the context of learning algorithms and more specifically as the process of learning a probability distribution that contains the design points that maximize a certain objective. This connection was pioneered by Rechenberg (1973) and Schwefel (1977), who introduced evolution strategies (ES) and adapted the variance of their search space based on the success rate of their experiments. This process is also reminiscent of the operations of selection and mutation that are key ingredients of genetic algorithms (GAs) (Holland 1975) and genetic programming (Koza 1992). ES and GAs can be considered as hybrids between gradient search strategies, which may effectively march downhill toward a minimum, and Latin hypercube or Monte Carlo sampling methods, which maximally explore the search space. Genetic programming was developed in the late 1980s by J.R. Koza, a PhD student of John Holland. Genetic programming generalized parameter optimization to function optimization, initially coded as a tree of operations (Koza 1992). A critical aspect of these algorithms is that they rely on an iterative construction of the probability distribution, based on data values of the objective function. This iterative construction can be lengthy and practically impossible for problems with expensive objective function evaluations.

Over the past 20 years, ES and GAs have begun to converge into estimation of distribution algorithms (EDAs). The covariance matrix adaptation ES (CMA-ES) algorithm (Ostermeier et al. 1994, Hansen et al. 2003) is a prominent example of ES using an adaptive estimation of the covariance matrix of a Gaussian probability distribution to guide the search for optimal parameters. This covariance matrix is adapted iteratively using the best points in each iteration. The CMA-ES is closely related to several other algorithms, such as mixed Bayesian optimization algorithms (Pelikan et al. 2004), and the reader is referred to Kern et al. (2004) for a comparative review. In recent years, this line of work has evolved into the more generalized information-geometric optimization (IGO) framework (Ollivier et al. 2017). IGO algorithms allow for families of probability distributions whose parameters are learned during the optimization process and maintain the cost function invariance as a major design principle. The resulting algorithm makes no assumption on the objective function to be optimized, and its flow is equivalent to a stochastic gradient descent. These techniques have proven to be effective on several simplified benchmark problems; however, their scaling remains unclear, and there are few guarantees for convergence in cost function landscapes such as those encountered in complex fluid dynamics problems. We note also that there is an interest in deploying these optimization methods to minimize the cost functions often associated with classical ML tasks (Salimans et al. 2017).

2.5. Important Topics We Have Not Covered: Bayesian Inference and Gaussian Processes

There are several learning algorithms that this review does not address but that demand particular attention from the fluid mechanics community. First and foremost, we wish to mention Bayesian inference, which aims to inform the model structure and its parameters from data in a probabilistic framework. Bayesian inference is fundamental for uncertainty quantification, and it is also fundamentally a learning method, as data are used to adapt the models. In fact, the alternative view is also possible, where every ML framework can be cast in a Bayesian framework (Barber 2012, Theodoridis 2015). The optimization algorithms outlined in this review provide a direct link. Whereas optimization algorithms aim to provide the best parameters of a model for given data in a stochastic manner, Bayesian inference aims to provide the full probability distribution. It may be argued that Bayesian inference may be even more powerful than ML, as it provides probability distributions for all parameters, leading to robust predictions, rather than single values, as is usually the case with classical ML algorithms. However, a key drawback for Bayesian inference is its computational cost, as it involves sampling and integration in high-dimensional spaces, which can be prohibitive for expensive function evaluations (e.g., wind tunnel experiments or large-scale direct numerical simulation). Along the same lines, one must mention Gaussian processes (GPs), which resemble kernel-based methods for regression. However, GPs develop these kernels adaptively based on the available data. They also provide probability distributions for the respective model parameters. GPs have been used extensively in problems related to time-dependent problems, and they may be considered competitors, albeit more costly, to RNNs. Finally, we note the use of GPs as surrogates for expensive cost functions in optimization problems using ES and GAs.

3. FLOW MODELING WITH MACHINE LEARNING

First principles, such as conservation laws, have been the dominant building blocks for flow modeling over the past centuries. However, for high Reynolds numbers, scale-resolving simulations using the most prominent model in fluid mechanics, the Navier–Stokes equations, are beyond our current computational resources. An alternative is to perform simulations based on approximations of these equations (as often practiced in turbulence modeling) or laboratory experiments for a specific configuration. However, simulations and experiments are expensive for iterative optimization, and simulations are often too slow for real-time control (Brunton & Noack 2015). Consequently, considerable effort has been placed on obtaining accurate and efficient reduced-order models that capture essential flow mechanisms at a fraction of the cost (Rowley & Dawson 2016). ML provides new avenues for dimensionality reduction and reduced-order modeling in fluid mechanics by providing a concise framework that complements and extends existing methodologies.

We distinguish here two complementary efforts: dimensionality reduction and reduced-order modeling. Dimensionality reduction involves extracting key features and dominant patterns that may be used as reduced coordinates where the fluid is compactly and efficiently described (Taira et al. 2017). Reduced-order modeling describes the spatiotemporal evolution of the flow as a parametrized dynamical system, although it may also involve developing a statistical map from parameters to averaged quantities, such as drag.

There have been significant efforts to identify coordinate transformations and reductions that simplify dynamics and capture essential flow physics; the POD is a notable example (Lumley 1970). Model reduction, such as Galerkin projection of the Navier–Stokes equations onto an orthogonal basis of POD modes, benefits from a close connection to the governing equations; however, it is intrusive, requiring human expertise to develop models from a working simulation. ML provides modular algorithms that may be used for data-driven system identification and modeling. Unique aspects of data-driven modeling of fluid flows include the availability of partial prior knowledge of the governing equations, constraints, and symmetries. With advances in simulation capabilities and experimental techniques, fluid dynamics is becoming a data-rich field, thus becoming amenable to ML algorithms.

In this review, we distinguish ML algorithms to model flow (a) kinematics through the extraction flow features and (b) dynamics through the adoption of various learning architectures.

3.1. Flow Feature Extraction

Pattern recognition and data mining are core strengths of ML. Many techniques have been developed by the ML community that are readily applicable to spatiotemporal fluid data. We discuss linear and nonlinear dimensionality reduction techniques, followed by clustering and classification. We also consider accelerated measurement and computation strategies, as well as methods to process experimental flow field data.

3.1.1. Dimensionality reduction: linear and nonlinear embeddings.

A common approach in fluid dynamics simulation and modeling is to define an orthogonal linear transformation from physical coordinates into a modal basis. The POD provides such an orthogonal basis for complex geometries based on empirical measurements. Sirovich (1987) introduced the snapshot POD, which reduces the computation to a simple data-driven procedure involving a singular value decomposition. Interestingly, in the same year, Sirovich used POD to generate a low-dimensional feature space for the classification of human faces, which is a foundation for much of modern computer vision (Sirovich & Kirby 1987).

POD is closely related to the algorithm of PCA, one of the fundamental algorithms of applied statistics and ML, to describe correlations in high-dimensional data. We recall that the PCA can be expressed as a two-layer neural network, called an autoencoder, to compress high-dimensional data for a compact representation, as shown in Figure 5. This network embeds high-dimensional data into a low-dimensional latent space and then decodes from the latent space back to the original high-dimensional space. When the network nodes are linear and the encoder and decoder are constrained to be transposes of one another, the autoencoder is closely related to the standard POD/PCA decomposition (Baldi & Hornik 1989) (see also Figure 6). However, the structure of the NN autoencoder is modular, and by using nonlinear activation units for the nodes, it is possible to develop nonlinear embeddings, potentially providing more compact coordinates. This observation led to the development of one of the first applications of deep NNs to reconstruct the near-wall velocity field in a turbulent channel flow using wall pressure and shear (Milano & Koumoutsakos 2002). More powerful autoencoders are available today in the ML community, and this link deserves further exploration.

figure
Figure 6 

On the basis of the universal approximation theorem (Hornik et al. 1989), which states that a sufficiently large NN can represent an arbitrarily complex input–output function, deep NNs are increasingly used to obtain more effective nonlinear coordinates for complex flows. However, deep learning often implies the availability of large volumes of training data that far exceed the parameters of the network. The resulting models are usually good for interpolation but may not be suitable for extrapolation when the new input data have different probability distributions than the training data (see Equation 1). In many modern ML applications, such as image classification, the training data are so vast that it is natural to expect that most future classification tasks will fall within an interpolation of the training data. For example, the ImageNet data set in 2012 (Krizhevsky et al. 2012) contained over 15 million labeled images, which sparked the current movement in deep learning (LeCun et al. 2015). Despite the abundance of data from experiments and simulations, the fluid mechanics community is still distanced from this working paradigm. However, it may be possible in the coming years to curate large, labeled, and complete-enough fluid databases to facilitate the deployment of such deep learning algorithms.

3.1.2. Clustering and classification.

Clustering and classification are cornerstones of ML. There are dozens of mature algorithms to choose from, depending on the size of the data and the desired number of categories. The k-means algorithm has been successfully employed by Kaiser et al. (2014) to develop a data-driven discretization of a high-dimensional phase space for the fluid mixing layer. This low-dimensional representation, in terms of a small number of clusters, enabled tractable Markov transition models of how the flow evolves in time from one state to another. Because the cluster centroids exist in the data space, it is possible to associate each cluster centroid with a physical flow field, lending additional interpretability. Amsallem et al. (2012) used k-means clustering to partition phase space into separate regions, in which local reduced-order bases were constructed, resulting in improved stability and robustness to parameter variations.

Classification is also widely used in fluid dynamics to distinguish between various canonical behaviors and dynamic regimes. Classification is a supervised learning approach where labeled data are used to develop a model to sort new data into one of several categories. Recently, Colvert et al. (2018) investigated the classification of wake topology (e.g., 2S, 2P + 2S, 2P + 4S) behind a pitching airfoil from local vorticity measurements using NNs; extensions have compared performance for various types of sensors (Alsalman et al. 2018). Wang & Hemati (2017) used the k-nearest-neighbors algorithm to detect exotic wakes. Similarly, NNs have been combined with dynamical systems models to detect flow disturbances and estimate their parameters (Hou et al. 2019). Related graph and network approaches in fluids by Nair & Taira (2015) have been used for community detection in wake flows (Meena et al. 2018). Finally, one of the earliest examples of ML classification in fluid dynamics by Bright et al. (2013) was based on sparse representation (Wright et al. 2009).

3.1.3. Sparse and randomized methods.

In parallel to ML, there have been great strides in sparse optimization and randomized linear algebra. ML and sparse algorithms are synergistic in that underlying low-dimensional representations facilitate sparse measurements (Manohar et al. 2018) and fast randomized computations (Halko et al. 2011). Decreasing the amount of data to train and execute a model is important when a fast decision is required, as in control. In this context, algorithms for the efficient acquisition and reconstruction of sparse signals, such as compressed sensing (Donoho 2006), have already been leveraged for compact representations of wall-bounded turbulence (Bourguignon et al. 2014) and for POD-based flow reconstruction (Bai et al. 2014).

Low-dimensional structure in data also facilitates accelerated computations via randomized linear algebra (Halko et al. 2011, Mahoney 2011). If a matrix has low-rank structure, then there are efficient matrix decomposition algorithms based on random sampling; this is closely related to the idea of sparsity and the high-dimensional geometry of sparse vectors. The basic idea is that if a large matrix has low-dimensional structure, then with high probability this structure will be preserved after projecting the columns or rows onto a random low-dimensional subspace, facilitating efficient downstream computations. These so-called randomized numerical methods have the potential to transform computational linear algebra, providing accurate matrix decompositions at a fraction of the cost of deterministic methods. For example, randomized linear algebra may be used to efficiently compute the singular value decomposition, which is used to compute PCA (Rokhlin et al. 2009, Halko et al. 2011).

3.1.4. Superresolution and flow cleansing.

Much of ML is focused on imaging science, providing robust approaches to improve resolution and remove noise and corruption based on statistical inference. These superresolution and denoising algorithms have the potential to improve the quality of both simulations and experiments in fluids.

Superresolution involves the inference of a high-resolution image from low-resolution measurements, leveraging the statistical structure of high-resolution training data. Several approaches have been developed for superresolution, for example, based on a library of examples (Freeman et al. 2002), sparse representation in a library (Yang et al. 2010), and most recently CNNs (Dong et al. 2014). Experimental flow field measurements from PIV (Adrian 1991, Willert & Gharib 1991) provide a compelling application where there is a tension between local flow resolution and the size of the imaging domain. Superresolution could leverage expensive and high-resolution data on smaller domains to improve the resolution on a larger imaging domain. Large-eddy simulations (LES) (Germano et al. 1991, Meneveau & Katz 2000) may also benefit from superresolution to infer the high-resolution structure inside a low-resolution cell that is required to compute boundary conditions. Recently, Fukami et al. (2018) developed a CNN-based superresolution algorithm and demonstrated its effectiveness on turbulent flow reconstruction, showing that the energy spectrum is accurately preserved. One drawback of superresolution is that it is often extremely costly computationally, making it useful for applications where high-resolution imaging may be prohibitively expensive; however, improved NN-based approaches may drive the cost down significantly. We note also that Xie et al. (2018) recently employed GANs for superresolution.

The processing of experimental PIV and particle tracking has also been one of the first applications of ML. NNs have been used for fast PIV (Knaak et al. 1997) and PTV (Labonté 1999), with impressive demonstrations for three-dimensional Lagrangian particle tracking (Ouellette et al. 2006). More recently, deep CNNs have been used to construct velocity fields from PIV image pairs (Lee et al. 2017). Related approaches have also been used to detect spurious vectors in PIV data (Liang et al. 2003) to remove outliers and fill in corrupt pixels.

3.2. Modeling Flow Dynamics

A central goal of modeling is to balance efficiency and accuracy. When modeling physical systems, interpretability and generalizability are also critical considerations.

3.2.1. Linear models through nonlinear embeddings: dynamic mode decomposition and Koopman analysis.

Many classical techniques in system identification may be considered ML, as they are data-driven models that generalize beyond the training data. Dynamic mode decomposition (DMD) (Schmid 2010, Kutz et al. 2016) is a modern approach to extract spatiotemporal coherent structures from time series data of fluid flows, resulting in a low-dimensional linear model for the evolution of these dominant coherent structures. DMD is based on data-driven regression and is equally valid for time-resolved experimental and numerical data. DMD is closely related to the Koopman operator (Rowley et al. 2009, Mezic 2013), which is an infinite-dimensional linear operator that describes how all measurement functions of the system evolve in time. Because the DMD algorithm is based on linear flow field measurements (i.e., direct measurements of the fluid velocity or vorticity field), the resulting models may not be able to capture nonlinear transients.

Recently, there has been a concerted effort to identify a coordinate system where the nonlinear dynamics appears linear. The extended DMD (Williams et al. 2015) and variational approach of conformation dynamics (Noé & Nüske 2013, Nüske et al. 2016) enrich the model with nonlinear measurements, leveraging kernel methods (Williams et al. 2015) and dictionary learning (Li et al. 2017). These special nonlinear measurements are generally challenging to represent, and deep learning architectures are now used to identify nonlinear Koopman coordinate systems where the dynamics appear linear (Takeishi et al. 2017, Lusch et al. 2018, Mardt et al. 2018, Wehmeyer & Noé 2018). The VAMPnet architecture (Mardt et al. 2018, Wehmeyer & Noé 2018) uses a time-lagged autoencoder and a custom variational score to identify Koopman coordinates on an impressive protein folding example. Based on the performance of VAMPnet, fluid dynamics may benefit from neighboring fields, such as molecular dynamics, which have similar modeling issues, including stochasticity, coarse-grained dynamics, and separation of timescales.

3.2.2. Neural network modeling.

Over the last three decades, NNs have been used to model dynamical systems and fluid mechanics problems. Early examples include the use of NNs to learn the solutions of ordinary and partial differential equations (Dissanayake & Phan-Thien 1994, Gonzalez-Garcia et al. 1998, Lagaris et al. 1998). We note that the potential of these works has not been fully explored, and in recent years there have been further advances (Chen et al. 2018, Raissi & Karniadakis 2018), including discrete and continuous-in-time networks. We note also the possibility of using these methods to uncover latent variables and reduce the number of parametric studies often associated with partial differential equations (Raissi et al. 2019). NNs are also frequently employed in nonlinear system identification techniques such as NARMAX, which are often used to model fluid systems (Glaz et al. 2010, Semeraro et al. 2016). In fluid mechanics, NNs were widely used to model heat transfer (Jambunathan et al. 1996), turbomachinery (Pierret & Van den Braembussche 1999), turbulent flows (Milano & Koumoutsakos 2002), and other problems in aeronautics (Faller & Schreck 1996).

RNNs with LSTMs (Hochreiter & Schmidhuber 1997) have been revolutionary for speech recognition, and they are considered one of the landmark successes of AI. They are currently being used to model dynamical systems and for data-driven predictions of extreme events (Vlachas et al. 2018, Wan et al. 2018). An interesting finding of these studies is that combining data-driven and reduced-order models is a potent method that outperforms each of its components on several studies. GANs (Goodfellow et al. 2014) are also being used to capture physics (Wu et al. 2018). GANs have potential to aid in the modeling and simulation of turbulence (Kim et al. 2018), although this field is nascent.

Despite the promise and widespread use of NNs in dynamical systems, several challenges remain. NNs are fundamentally interpolative, and so the function is well approximated only in the span (or under the probability distribution) of the sampled data used to train them. Thus, caution should be exercised when using NN models for an extrapolation task. In many computer vision and speech recognition examples, the training data are so vast that nearly all future tasks may be viewed as an interpolation on the training data, although this scale of training has not been achieved to date in fluid mechanics. Similarly, NN models are prone to overfitting, and care must be taken to cross-validate models on a sufficiently chosen test set; best practices are discussed by Goodfellow et al. (2016). Finally, it is important to explicitly incorporate partially known physics, such as symmetries, constraints, and conserved quantities.

3.2.3. Parsimonious nonlinear models.

Parsimony is a recurring theme in mathematical physics, from Hamilton's principle of least action to the apparent simplicity of many governing equations. In contrast to the raw representational power of NNs, ML algorithms are also being employed to identify minimal models that balance predictive accuracy with model complexity, preventing overfitting and promoting interpretability and generalizability. Genetic programming was recently used to discover conservation laws and governing equations (Schmidt & Lipson 2009). Sparse regression in a library of candidate models has also been proposed to identify dynamical systems (Brunton et al. 2016) and partial differential equations (Rudy et al. 2017, Schaeffer 2017). Loiseau & Brunton (2018) identified sparse reduced-order models of several flow systems, enforcing energy conservation as a constraint. In both genetic programming and sparse identification, a Pareto analysis is used to identify models that have the best trade-off between model complexity, measured in number of terms, and predictive accuracy. In cases where the physics is known, this approach typically discovers the correct governing equations, providing exceptional generalizability compared with other leading algorithms in ML.

3.2.4. Closure models with machine learning.

The use of ML to develop turbulence closures is an active area of research (Duraisamy et al. 2019). The extreme separation of spatiotemporal scales in turbulent flows makes it exceedingly costly to resolve all scales in simulation, and even with Moore's law, we are decades away from resolving all scales in relevant configurations (e.g., aircraft, submarines, etc.). It is common to truncate small scales and model their effect on the large scales with a closure model. Common approaches include Reynolds-averaged Navier–Stokes (RANS) and LES. However, these models may require careful tuning to match data from fully resolved simulations or experiments.

ML has been used to identify and model discrepancies in the Reynolds stress tensor between a RANS model and high-fidelity simulations (Ling & Templeton 2015, Parish & Duraisamy 2016, Ling et al. 2016b, Xiao et al. 2016, Singh et al. 2017, Wang et al. 2017). Ling & Templeton (2015) compared SVMs, Adaboost decision trees, and random forests to classify and predict regions of high uncertainty in the Reynolds stress tensor. Wang et al. (2017) used random forests to build a supervised model for the discrepancy in the Reynolds stress tensor. Xiao et al. (2016) leveraged sparse online velocity measurements in a Bayesian framework to infer these discrepancies. In related work, Parish & Duraisamy (2016) developed the field inversion and ML modeling framework that builds corrective models based on inverse modeling. This framework was later used by Singh et al. (2017) to develop an NN enhanced correction to the Spalart–Allmaras RANS model, with excellent performance. A key result by Ling et al. (2016b) employed the first deep network architecture with many hidden layers to model the anisotropic Reynolds stress tensor, as shown in Figure 7. Their novel architecture incorporates a multiplicative layer to embed Galilean invariance into the tensor predictions. This provides an innovative and simple approach to embed known physical symmetries and invariances into the learning architecture (Ling et al. 2016a), which we believe will be essential in future efforts that combine learning for physics. For LES closures, Maulik et al. (2019) have employed artificial NNs to predict the turbulence source term from coarsely resolved quantities.

figure
Figure 7 

3.2.5. Challenges of machine learning for dynamical systems.

Applying ML to model physical dynamical systems poses several unique challenges and opportunities. Model interpretability and generalizability are essential cornerstones in physics. A well-crafted model will yield hypotheses for phenomena that have not been observed before. This principle is for example exhibited in the parsimonious formulation of classical mechanics in Newton's second law.

High-dimensional systems, such as those encountered in unsteady fluid dynamics, have the challenges of multiscale dynamics, sensitivity to noise and disturbances, latent variables, and transients, all of which require careful attention when applying ML techniques. In ML for dynamics, we distinguish two tasks: discovering unknown physics and improving models by incorporating known physics. Many learning architectures cannot readily incorporate physical constraints in the form of symmetries, boundary conditions, and global conservation laws. This is a critical area for continued development, and several recent works have presented generalizable physics models (Battaglia et al. 2018).

4. FLOW OPTIMIZATION AND CONTROL USING MACHINE LEARNING

Learning algorithms are well suited to flow optimization and control problems involving black-box or multimodal cost functions. These algorithms are iterative and often require several orders of magnitude more cost function evaluations than gradient-based algorithms (Bewley et al. 2001). Moreover, they do not offer guarantees of convergence, and we suggest that they be avoided when techniques such as adjoint methods are applicable. At the same time, techniques such as RL have been shown to outperform even optimal flow control strategies (Novati et al. 2019). Indeed, there are several classes of flow control and optimization problems where learning algorithms may be the method of choice, as described below.

OPTIMIZATION AND CONTROL: BOUNDARIES ERASED BY FAST COMPUTERS

Optimization and control are intimately related, and the boundaries are becoming even less distinct with increasingly fast computers, as summarized by Tsiotras & Mesbahi (2017, p. 195):

Interestingly, the distinction between optimization and control is largely semantic and (alas!) implementation-dependent. If one has the capability of solving optimization problems fast enough on the fly to close the loop, then one has (in principle) a feedback control law... Not surprisingly then, the same algorithm can be viewed as solving an optimization or a control problem, based solely on the capabilities of the available hardware. With the continued advent of faster and more capable computer hardware architectures, the boundary between optimization and control will become even more blurred. However, when optimization is embedded in the implementation of feedback control, the classical problems of control such as robustness to model uncertainty, time delays, and process and measurement noise become of paramount importance, particularly for high-performance aerospace systems.

In contrast to flow modeling, learning algorithms for optimization and control interact with the data sampling process in several ways. First, in line with the modeling efforts described in earlier sections, ML can be applied to develop explicit surrogate models that relate the cost function and the control/optimization parameters. Surrogate models such as NNs can then be amenable to even gradient-based methods, although they often get stuck in local minima. Multifidelity algorithms (Perdikaris et al. 2016) can also be employed to combine surrogates with the cost function of the complete problem. As the learning progresses, new data are requested as guided by the results of the optimization. Alternatively, the optimization or control problem may be described in terms of learning probability distributions of parameters that minimize the cost function. These probability distributions are constructed from cost function samples obtained during the optimization process. Furthermore, the high-dimensional and nonconvex optimization procedures that are currently employed to train nonlinear LMs are well suited to the high-dimensional, nonlinear optimization problems in flow control.

We remark that the lines between optimization and control are becoming blurred by the availability of powerful computers (see the sidebar titled Optimization and Control: Boundaries Erased by Fast Computers). However, the range of critical spatiotemporal scales and the nonlinearity of the underlying processes will likely render real-time optimization for flow control a challenge for decades to come.

4.1. Stochastic Flow Optimization: Learning Probability Distributions

Stochastic optimization includes ES and GAs, which were originally developed based on bio-inspired principles. However, in recent years these algorithms have been placed in a learning framework (Kern et al. 2004).

Stochastic optimization has found widespread use in engineering design, in particular as many engineering problems involve black-box-type cost functions. A much-abbreviated list of applications includes aerodynamic shape optimization (Giannakoglou et al. 2006), uninhabited aerial vehicles (UAVs) (Hamdaoui et al. 2010), shape and motion optimization in artificial swimmers (Gazzola et al. 2012, Van Rees et al. 2015), and improved power extraction in crossflow turbines (Strom et al. 2017). We refer readers to the review article by Skinner & Zare-Behtash (2018) for an extensive comparison of gradient-based and stochastic optimization algorithms for aerodynamics.

These algorithms involve large numbers of iterations, and they can benefit from massively parallel computer architectures. Advances in automation have also facilitated their application in experimental (Strom et al. 2017, Martin & Gharib 2018) and industrial settings (Bueche et al. 2002). We note that stochastic optimization algorithms are well suited to address the experimental and industrial challenges associated with uncertainty, such as unexpected system behavior, partial descriptions of the system and its environment, and exogenous disturbances. Hansen et al. (2009) proposed an approach to enhance the capabilities of evolutionary algorithms for online optimization of a combustor test rig.

Stochastic flow optimization will continue to benefit from advances in computer hardware and experimental techniques. At the same time, convergence proofs, explainability, and reliability are outstanding issues that need to be taken into consideration when deploying such algorithms in fluid mechanics problems. Hybrid algorithms that combine in a problem-specific manner stochastic techniques and gradient-based methods may offer the best strategy for flow control problems.

4.2. Flow Control with Machine Learning

Feedback flow control modifies the behavior of a fluid dynamic system through actuation that is informed by sensor measurements. Feedback is necessary to stabilize an unstable system, attenuate sensor noise, and compensate for external disturbances and model uncertainty. Challenges of flow control include a high-dimensional state, nonlinearity, latent variables, and time delays. ML algorithms have been used extensively in control, system identification, and sensor placement.

4.2.1. Neural networks for control.

NNs have received significant attention for system identification (see Section 3) and control, including applications in aerodynamics (Phan et al. 1995). The application of NNs to turbulence flow control was pioneered by Lee et al. (1997). The skin-friction drag of a turbulent boundary layer was reduced using local wall-normal blowing and suction based on few skin-friction sensors. A sensor-based control law was learned from a known optimal full-information controller, with little loss in overall performance. Furthermore, a single-layer network was optimized for skin-friction drag reduction without incorporating any prior knowledge of the actuation commands. Both strategies led to a conceptually simple local opposition control. Several other studies employ NNs, e.g., for phasor control (Rabault et al. 2019) or even frequency cross-talk. The need to optimize many parameters is the price for the theoretical advantage of approximating arbitrary nonlinear control laws. NN control may require exorbitant computational or experimental resources for configurations with complex high-dimensional nonlinearities and many sensors and actuators. At the same time, the training time of NNs has been improved by several orders of magnitude since these early applications, which warrant further investigation into their potential for flow control.

4.2.2. Genetic algorithms for control.

GAs have been deployed to solve several flow control problems. They require that the structure of the control law be prespecified and contain only a few adjustable parameters. An example of the use of GA for control design in fluids was used for experimental mixing optimization of the backward-facing step (Benard et al. 2016). As with NN control, the learning time increases with the number of parameters, making it challenging or even prohibitive for controllers with nonlinearities (e.g., a constant-linear-quadratic law), signal history (e.g., a Kalman filter), or multiple sensors and actuators.

Genetic programming has been used extensively in active control for engineering applications (Dracopoulos 1997, Fleming & Purshouse 2002) and in recent years in several flow control plants. This includes the learning of multifrequency open-loop actuation, multi-input sensor feedback, and distributed control. We refer readers to Duriez et al. (2016) for an in-depth description of the method and to Noack (2018) for an overview of the plants. We remark that most control laws have been obtained within 1,000 test evaluations, each requiring only a few seconds in a wind tunnel.

4.3. Flow Control via Reinforcement Learning

In recent years, RL has advanced beyond the realm of games and has become a fundamental mode of problem solving in a growing number of domains, including to reproduce the dynamics of hydrological systems (Loucks et al. 2005), actively control the oscillatory laminar flow around bluff bodies (Guéniat et al. 2016), study the individual (Gazzola et al. 2014) or collective motion of fish (Gazzola et al. 2016, Novati et al. 2017, Verma et al. 2018), maximize the range of simulated (Reddy et al. 2016) and robotic (Reddy et al. 2018) gliders, optimize the kinematic motion of UAVs (Kim et al. 2004, Tedrake et al. 2009), and optimize the motion of microswimmers (Colabrese et al. 2017, 2018). Figure 8 provides a schematic of RL with compelling examples related to fluid mechanics.

figure
Figure 8 

Fluid mechanics knowledge is essential for applications of RL, as success or failure hinges on properly selecting states, actions, and rewards that reflect the governing mechanisms of the flow problem. Natural organisms and their sensors, such as the visual system in a bird or the lateral line in a fish, can guide the choice of states. As sensor technologies progress at a rapid pace, the algorithmic challenge may be that of optimal sensor placement (Papadimitriou & Papadimitriou 2015, Manohar et al. 2018). The actions reflect the flow actuation device and may involve body deformation or wing flapping. Rewards may include energetic factors, such as the cost of transport or proximity to the center of a fish school to avoid predation. The computational cost of RL remains a challenge to its widespread adoption, but we believe this deficiency can be mediated by the parallelism inherent in RL. There is growing interest in methods designed to be transferable from low-accuracy (e.g., two-dimensional) to high-accuracy (e.g., three-dimensional) simulations (Verma et al. 2018) or from simulations to related real-world applications (Richter et al. 2016, Bousmalis et al. 2017).

5. DISCUSSION AND OUTLOOK

This review presents ML algorithms for the perspective of fluid mechanics. The interface of the two fields has a long history and has attracted a renewed interest in the last few years. This review addresses applications of ML in problems of flow modeling, optimization, and control in experiments and simulations. It highlights some successes of ML in critical fluid mechanics tasks, such as dimensionality reduction, feature extraction, PIV processing, superresolution, reduced-order modeling, turbulence closure, shape optimization, and flow control. It discusses lessons learned from these efforts and justifies the current interest in light of the technological advances of our times. Our goal is to provide a deeper understanding of ML and its context in fluid mechanics. ML comprises data-driven optimization and applied regression techniques that are well suited for high-dimensional, nonlinear problems, such as those encountered in fluid dynamics; fluid mechanics expertise will be necessary to formulate these optimization and regression problems.

ML algorithms present an arsenal of tools, largely unexplored in fluid mechanics research, that can augment existing modes of inquiry. Fluid mechanics knowledge and centuries-old conservation laws remain relevant in the era of big data. Such knowledge can help frame more precise questions and assist in reducing the large computational cost often associated with the application of ML algorithms in flow control and optimization. The exploration and visualization of high-dimensional search spaces can be simplified by ML and increasingly abundant high-performance computing resources.

In the near future, experience with ML algorithms will help frame new questions in fluid mechanics, extending decades-old linearized models and linear approaches to the nonlinear regime. The transition to the nonlinear realm of ML is facilitated by the abundance of open source software and methods and the prevalent openness of the ML community. In the long term, ML will undoubtedly offer a fresh look into old problems of fluid mechanics under the light of data. Interpreting the ML solutions, and refining the problem statement, will again require fluid mechanics expertise.

A word of caution is necessary to balance the current excitement about data-driven research and the (almost magical) powers of ML. After all, an ML algorithm will always provide some kind of answer to any question that is based on its training data—data that may not even be relevant to the question at hand. Properly formulating the question and selecting the data, the LM, and its training are all critical components of the learning process. Applying ML algorithms to fluid mechanics faces numerous outstanding challenges (and opportunities!). Although many fields of ML are concerned with raw predictive performance, applications in fluid mechanics often require models that are explainable and generalizable and have guarantees.

Although deep learning will undoubtedly become a critical tool in several aspects of flow modeling, not all ML is deep learning. It is important to consider several factors when choosing methods, including the quality and quantity of data, the desired inputs and outputs, the cost function to be optimized, whether the task involves interpolation or extrapolation, and how important it is for models to be explainable. It is also important to cross-validate ML models; otherwise, results may be prone to overfitting. It is also important to develop and adapt ML algorithms that are not only physics informed but also physics consistent, a major outstanding challenge in AI. This review concludes with a call for action in the fluid mechanics community to further embrace open and reproducible research products and standards. Reproducibility is a cornerstone of science, and several frameworks are currently developed to render this into a systematic scientific process (Barber 2015). It is increasingly possible to document procedures, archive code, and host data so that others can reproduce results. Data are essential for ML; thus, creating and curating benchmark data sets and software will spur interest among researchers in related fields, driving progress. These fluid benchmarks are more challenging than the traditional image data sets encountered in ML: Fluid data are multimodal and multifidelity, they have high resolution in some dimensions and are sparse in others, many tasks balance multiple objectives, and foremost, data come from dynamical systems, where many tasks do not admit postmortem analysis.

We are entering a new and exciting era in fluid mechanics research. Centuries of theoretical developments based on first principles are now merging with data-driven analysis. This fusion will provide solutions to many long-sought problems in fluid dynamics, first and foremost the enhanced understanding of its governing mechanisms.

SUMMARY POINTS

1. 

Machine learning (ML) entails powerful information-processing algorithms that are relevant for modeling, optimization, and control of fluids. Effective problem solvers will have expertise in ML and in-depth knowledge of fluid mechanics.

2. 

Fluid mechanics has been traditionally concerned with big data. For decades it has used ML to understand, predict, optimize, and control flows. Currently, ML capabilities are advancing at an incredible rate, and fluid mechanics is beginning to tap into the full potential of these powerful methods.

3. 

Many tasks in fluid mechanics, such as reduced-order modeling, shape optimization, and feedback control, may be posed as optimization and regression tasks. ML can improve optimization performance and reduce convergence time. ML is also used for dimensionality reduction and identifying low-dimensional manifolds and discrete flow regimes, which benefit understanding.

4. 

Flow control strategies have been traditionally based on the precise sequence from understanding to modeling and then to control. The ML paradigm suggests more flexibility and iterates between data-driven and first principle approaches.

FUTURE ISSUES

1. 

ML algorithms often come without guarantees for performance, robustness, or convergence, even for well-defined tasks. How can interpretability, generalizability, and explainability of the results be achieved?

2. 

Incorporating and enforcing known flow physics is a challenge and opportunity for ML algorithms. Can we hybridize data-driven and first principle approaches in fluid mechanics?

3. 

There are many possibilities to discover new physical mechanisms, symmetries, constraints, and invariances from fluids data.

4. 

Data-driven modeling may be a potent alternative in revisiting existing empirical laws in fluid mechanics.

5. 

ML encourages open sharing of data and software. Can this assist the development of frameworks for reproducible and open science in fluid mechanics?

6. 

Fluids researchers will benefit from interfacing with the ML community, where the latest advances are reported at peer-reviewed conferences.

disclosure statement

The authors are not aware of any biases that might be perceived as affecting the objectivity of this review.

acknowledgments

S.L.B. acknowledges funding from the Army Research Office (ARO W911NF-17-1-0306, W911NF-17-1-0422) and the Air Force Office of Scientific Research (AFOSR FA9550-18-1-0200). B.R.N. acknowledges funding from LIMSI-CNRS, Université Paris Sud (SMEMaG), the French National Research Agency (ANR-11-IDEX-0003-02, ANR-17-ASTR-0022), and the German Research Foundation (CRC880, SE 2504/2-1, SE 2504/3-1). P.K. acknowledges funding from an ERC Advanced Investigator Award (FMCoBe, No. 34117), the Swiss National Science Foundation, and the Swiss National Supercomputing Centre. The authors are grateful for discussions with Nathan Kutz (University of Washington), Jean-Christophe Loiseau (ENSAM ParisTech, Paris), François Lusseyran (LIMSI-CNRS, Paris), Guido Novati (ETH Zurich), Luc Pastur (ENSTA ParisTech, Paris), and Pantelis Vlachas (ETH Zurich).

literature cited

  • 1.
    Adrian RJ. 1991. Particle-imaging techniques for experimental fluid mechanics. Annu. Rev. Fluid Mech. 23:261–304
    • Link
    • Web of Science ®
    • ADS
    • Google Scholar
  • 2.
    Alsalman M, Colvert B, Kanso E. 2018. Training bioinspired sensors to classify flows. Bioinspiration Biomim. 14:016009
    • Crossref
    • Medline
    • Web of Science ®
    • Google Scholar
    Article Location
  • 3.
    Amsallem D, Zahr MJ, Farhat C. 2012. Nonlinear model order reduction based on local reduced-order bases. Int. J. Numer. Meth. Eng. 92:891–916
    • Crossref
    • Web of Science ®
    • Google Scholar
    Article Location
  • 4.
    Bai Z, Wimalajeewa T, Berger Z, Wang G, Glauser M, Varshney PK. 2014. Low-dimensional approach for reconstruction of airfoil data via compressive sensing. AIAA J. 53:920–33
    • Crossref
    • ADS
    • Google Scholar
    Article Location
  • 5.
    Baldi P, Hornik K. 1989. Neural networks and principal component analysis: learning from examples without local minima. Neural Netw. 2:53–58
    • Crossref
    • Web of Science ®
    • Google Scholar
    Article Locations:
    • Article Location
    • Article Location
    More AR articles citing this reference

    • Statistical Mechanics of Deep Learning

      Yasaman Bahri,1 Jonathan Kadmon,2 Jeffrey Pennington,1 Sam S. Schoenholz,1 Jascha Sohl-Dickstein,1 and Surya Ganguli1,21Google Brain, Google Inc., Mountain View, California 94043, USA2Department of Applied Physics, Stanford University, Stanford, California 94035, USA; email: [email protected]
      Annual Review of Condensed Matter Physics Vol. 11: 501 - 528
      • ...Reference 39 proved that the error landscape of linear neural networks with one hidden layer has no local minima that are not also global minima; all higher-error critical points are saddle points....

  • 6.
    Barber D. 2012. Bayesian Reasoning and Machine Learning. Cambridge, UK: Cambridge Univ. Press
    • Google Scholar
    Article Location
  • 7.
    Barber RF, Candes EJ. 2015. Controlling the false discovery rate via knock-offs. Ann. Stat. 43:2055–85 Reproducible science: a framework.
    • Crossref
    • Web of Science ®
    • Google Scholar
    Article Location
    More AR articles citing this reference

    • Large-Scale Global and Simultaneous Inference: Estimation and Testing in Very High Dimensions

      T. Tony Cai1 and Wenguang Sun21Department of Statistics, The Wharton School, University of Pennsylvania, Philadelphia, Pennsylvania 19104; email: [email protected]2Department of Data Sciences and Operations, Marshall School of Business, University of Southern California, Los Angeles, California 90089; email: [email protected]
      Annual Review of Economics Vol. 9: 411 - 439
      • ...Javanmard & Montanari 2014, Liu & Luo 2014, Lockhart et al. 2014, Van de Geer et al. 2014, Zhang & Zhang 2014, Barber & Candès 2015, Xia et al. 2017, Cai & Guo 2017)....

  • 8.
    Battaglia PW, Hamrick JB, Bapst V, Sanchez-Gonzalez A, Zambaldi V, 2018. Relational inductive biases, deep learning, and graph networks. arXiv:1806.01261 [cs.LG]
    • ADS
    • Google Scholar
    Article Location
  • 9.
    Bellman R. 1952. On the theory of dynamic programming. PNAS 38:716–19
    • Crossref
    • Medline
    • Web of Science ®
    • ADS
    • Google Scholar
    Article Location
  • 10.
    Benard N, Pons-Prats J, Periaux J, Bugeda G, Braud P, 2016. Turbulent separated shear flow control by surface plasma actuator: experimental optimization by genetic algorithm approach. Exp. Fluids 57:22
    • Crossref
    • Web of Science ®
    • Google Scholar
    Article Location
  • 11.
    Bewley TR, Moin P, Temam R. 2001. DNS-based predictive control of turbulence: an optimal benchmark for feedback algorithms. J. Fluid Mech. 447:179–225
    • Crossref
    • Web of Science ®
    • ADS
    • Google Scholar
    Article Location
    More AR articles citing this reference

    • Nonlinear Nonmodal Stability Theory

      R.R. Kerswell1,21School of Mathematics, Bristol University, Bristol BS8 1TW, United Kingdom; email: [email protected]2Current affiliation: Department of Applied Mathematics and Theoretical Physics, Centre for Mathematical Sciences, Cambridge University, Cambridge CB3 0WA, United Kingdom
      Annual Review of Fluid Mechanics Vol. 50: 319 - 345
      • ...so the use of fully nonlinear methods in control is still impractical (e.g., Joslin et al. 1995, 1997; Gunzburger 2000; Bewley et al. 2001...
    • A Linear Systems Approach to Flow Control

      John Kim1 and Thomas R. Bewley21Department of Mechanical and Aerospace Engineering, University of California, Los Angeles, Los Angeles, California 90095-1597; email: [email protected]2Department of Mechanical and Aerospace Engineering, University of California, San Diego, La Jolla, California 92093-0411; email: [email protected]
      Annual Review of Fluid Mechanics Vol. 39: 383 - 417
      • ...the benchmark problem of relaminarizing fully developed channel-flow turbulence (in numerical simulations) via a distribution of blowing/suction on the wall as the control was solved for the first time this way (Bewley et al. 2001)....
      • ...as found in the application of this method to the control of channel-flow turbulence, as reported in Bewley et al. (2001)....

  • 12.
    Bishop CM, James GD. 1993. Analysis of multiphase flows using dual-energy gamma densitometry and neural networks. Nucl. Instrum. Methods Phys. Res. 327:580–93
    • Crossref
    • Web of Science ®
    • ADS
    • Google Scholar
    Article Locations:
    • Article Location
    • Article Location
  • 13.
    Bölcskei H, Grohs P, Kutyniok G, Petersen P. 2019. Optimal approximation with sparsely connected deep neural networks. SIAM J. Math. Data Sci. 1:8–45 Theoretical analysis of the approximation properties of deep neural networks.
    • Crossref
    • Google Scholar
    Article Location
  • 14.
    Bourguignon JL, Tropp JA, Sharma AS, McKeon BJ. 2014. Compact representation of wall-bounded turbulence using compressive sampling. Phys. Fluids 26:015109
    • Crossref
    • Web of Science ®
    • Google Scholar
    Article Location
  • 15.
    Bousmalis K, Irpan A, Wohlhart P, Bai Y, Kelcey M, 2017. Using simulation and domain adaptation to improve efficiency of deep robotic grasping. arXiv:1709.07857 [cs.LG]
    • ADS
    • Google Scholar
    Article Location
  • 16.
    Breiman L. 2001. Random forests. Mach. Learn. 45:5–32
    • Crossref
    • Web of Science ®
    • Google Scholar
    Article Location
    More AR articles citing this reference

    • DNA Methylation Profiling: An Emerging Paradigm for Cancer Diagnosis

      Antonios Papanicolau-Sengos and Kenneth AldapeLaboratory of Pathology, National Cancer Institute, Bethesda, Maryland 20892, USA; email: [email protected], [email protected]
      Annual Review of Pathology: Mechanisms of Disease Vol. 17: 295 - 321
      • ...and the final class is defined as the most frequently voted output (38)....
    • Machine Learning for the Study of Plankton and Marine Snow from Images

      Jean-Olivier Irisson,1 Sakina-Dorothée Ayata,1 Dhugal J. Lindsay,2 Lee Karp-Boss,3 and Lars Stemmann11Laboratoire d'Océanographie de Villefranche, Sorbonne Université, CNRS, F-06230 Villefranche-sur-Mer, France; email: [email protected], [email protected], [email protected]2Advanced Science-Technology Research (ASTER) Program, Institute for Extra-Cutting-Edge Science and Technology Avant-Garde Research (X-STAR), Japan Agency for Marine-Earth Science and Technology, Yokosuka, Kanagawa 237-0021, Japan; email: [email protected]3School of Marine Sciences, University of Maine, Orono, Maine 04469, USA; email: [email protected]
      Annual Review of Marine Science Vol. 14: 277 - 301
      • ...such as a support vector machine (Cortes & Vapnik 1995) or a random forest (RF) (Breiman 2001), ...
    • Small Steps with Big Data: Using Machine Learning in Energy and Environmental Economics

      Matthew C. Harding1 and Carlos Lamarche21Department of Economics and Department of Statistics, University of California, Irvine, California 92697; email: [email protected]2Department of Economics, Gatton College of Business and Economics, University of Kentucky, Lexington, Kentucky 40506
      Annual Review of Resource Economics Vol. 13: 469 - 488
      • ...regularization approaches are needed to avoid issues. Breiman (1996, 2001) develops a series of approaches that lead to improvements in the performance of tree methods....
      • ...The random forest introduced by Breiman (2001) builds on a large number of trees that are formed based on a bootstrap sample of the original sample (bagging, ...
    • Generating and Using Transcriptomically Based Retinal Cell Atlases

      Karthik Shekhar1 and Joshua R. Sanes21Department of Chemical and Biomolecular Engineering; Helen Wills Neuroscience Institute; and California Institute for Quantitative Biosciences, QB3, University of California, Berkeley, California 94720, USA; email: [email protected]2Center for Brain Science and Department of Molecular and Cell Biology, Harvard University, Cambridge, Massachusetts 02138, USA; email: [email protected]
      Annual Review of Vision Science Vol. 7: 43 - 72
      • ...This comparison is frequently performed using multiclass supervised classification approaches such as Random Forest and XGBoost (Breiman 2001, Chen & Guestrin 2016)....
    • Artificial Intelligence, Predictive Policing, and Risk Assessment for Law Enforcement

      Richard A. BerkDepartments of Statistics and Criminology, University of Pennsylvania, Philadelphia, Pennsylvania 19104, USA; email: [email protected]
      Annual Review of Criminology Vol. 4: 209 - 237
      • ...Perhaps the two most important and effective ML methods used for criminal justice risk assessment are random forests (Breiman 2001a)...
      • ...One simple approach computes the decline in accuracy when each predictor in turn is precluded from affecting the fitted values (Breiman 2001a)....
    • Machine Learning in Materials Discovery: Confirmed Predictions and Their Underlying Approaches

      James E. Saal,1 Anton O. Oliynyk,2 and Bryce Meredig11Citrine Informatics, Redwood City, California 94063, USA; email: [email protected]2Department of Chemistry and Biochemistry, Manhattan College, Riverdale, New York 10471, USA
      Annual Review of Materials Research Vol. 50: 49 - 69
      • ...The popular random forest (RF) (59, 60) and (deep) neural network (NN) (61, 62)...
    • Sparse Data–Driven Learning for Effective and Efficient Biomedical Image Segmentation

      John A. Onofrey,1,2 Lawrence H. Staib,1,3 Xiaojie Huang,1,6 Fan Zhang,1 Xenophon Papademetris,1,3 Dimitris Metaxas,4 Daniel Rueckert,5 and James S. Duncan1,31Department of Radiology and Biomedical Imaging, Yale School of Medicine, New Haven, Connecticut 06520, USA; email: [email protected]2Department of Urology, Yale School of Medicine, New Haven, Connecticut 06520, USA3Department of Biomedical Engineering, Yale University, New Haven, Connecticut 06520, USA; email: [email protected]4Department of Computer Science, Rutgers University, Piscataway, New Jersey 08854, USA5Department of Computing, Imperial College London, London SW7 2AZ, United Kingdom6Citadel Securities, Chicago, Illinois 60603, USA
      Annual Review of Biomedical Engineering Vol. 22: 127 - 153
      • ...segmentation methods use one or more of these appearance features in conjunction with classification methods such as support vector machines (26), random forests (27), ...
    • Machine Learning in Epidemiology and Health Outcomes Research

      Timothy L. Wiemken1 and Robert R. Kelley21Center for Health Outcomes Research, Saint Louis University, Saint Louis, Missouri 63104, USA; email: [email protected]2Department of Computer Science, Bellarmine University, Louisville, Kentucky 40205, USA; email: [email protected]
      Annual Review of Public Health Vol. 41: 21 - 36
      • ... as compared with a moderate size (several hundred cases) necessary for a behavioral/cognition outcome evaluating functional magnetic resonance imaging (fMRI) (11)....
      • ...Examples of machine learning missing data imputers are ripe in the literature, largely basing the models on random forest approaches (11, 55, 60)....
    • Robust Small Area Estimation: An Overview

      Jiming Jiang1 and J. Sunil Rao21Department of Statistics, University of California, Davis, California 95616, USA; email: [email protected]2Department of Public Health Sciences, University of Miami, Miami, Florida 33136, USA
      Annual Review of Statistics and Its Application Vol. 7: 337 - 360
      • ...They then went a step further and proposed random forests (RFs) (Brieman 2001) for SAE, ...
    • Big Data in Industrial-Organizational Psychology and Human Resource Management: Forward Progress for Organizational Research and Practice

      Frederick L. Oswald,1 Tara S. Behrend,2 Dan J. Putka,3 and Evan Sinar41Department of Psychological Sciences, Rice University, Houston, Texas 77005, USA; email: [email protected]2Department of Organizational Sciences and Communication, George Washington University, Washington, DC 20052, USA3Human Resources Research Organization, Alexandria, Virginia 22314, USA4BetterUp, Pittsburgh, Pennsylvania 15243, USA
      Annual Review of Organizational Psychology and Organizational Behavior Vol. 7: 505 - 533
      • ...The next step up in complexity for exploring both nonlinearities and interactions involves methods that extend traditional classification and regression tree (CART) models (Breiman et al. 1984), such as random forests (Breiman 2001a)...
    • Big Data and Artificial Intelligence Modeling for Drug Discovery

      Hao ZhuDepartment of Chemistry and Center for Computational and Integrative Biology, Rutgers University, Camden, New Jersey 08102, USA; email: [email protected]
      Annual Review of Pharmacology and Toxicology Vol. 60: 573 - 589
      • ...which were developed based on nonlinear modeling algorithms such as k-nearest neighbors (77), support vector machines (78), and random forest (79, 80), ...
    • Machine Learning Methods That Economists Should Know About

      Susan Athey1,2,3 and Guido W. Imbens1,2,3,41Graduate School of Business, Stanford University, Stanford, California 94305, USA; email: [email protected], [email protected]2Stanford Institute for Economic Policy Research, Stanford University, Stanford, California 94305, USA3National Bureau of Economic Research, Cambridge, Massachusetts 02138, USA4Department of Economics, Stanford University, Stanford, California 94305, USA
      Annual Review of Economics Vol. 11: 685 - 725
      • ...Regression trees (Breiman et al. 1984) and their extension, random forests (Breiman 2001a), ...
      • ...For better estimates of , random forests (Breiman 2001a) build on the regression tree algorithm....
    • Machine Learning for Sociology

      Mario Molina and Filiz GaripDepartment of Sociology, Cornell University, Ithaca, New York 14853, USA; email: [email protected], [email protected]
      Annual Review of Sociology Vol. 45: 27 - 45
      • ...A version called random forests averages over multiple trees (Breiman 2001a), ...
      • ...and specify a parametric (typically linear) model to relate the inputs to an output (Breiman 2001a, Donoho 2017)....
    • Discovering Pathway and Cell Type Signatures in Transcriptomic Compendia with Machine Learning

      Gregory P. Way1,2 and Casey S. Greene21Genomics and Computational Biology Graduate Group, Perelman School of Medicine, University of Pennsylvania, Philadelphia, Pennsylvania 19104, USA2Department of Systems Pharmacology and Translational Therapeutics, University of Pennsylvania, Philadelphia, Pennsylvania 19104, USA; email: [email protected]
      Annual Review of Biomedical Data Science Vol. 2: 1 - 17
      • ...and RFs will determine over many iterations features used to split samples based on information content (21, 22)....
    • Using Statistics to Assess Lethal Violence in Civil and Inter-State War

      Patrick Ball and Megan PriceHuman Rights Data Analysis Group, San Francisco, California 94110, USA; email: [email protected]
      Annual Review of Statistics and Its Application Vol. 6: 63 - 84
      • ...we chose binary classification algorithms that perform better than alternating decision trees [e.g., gradient boosting (Chen & Guestrin 2016), random forests (Breiman 2001), ...
    • Personalized Cancer Genomics

      Richard M. SimonR Simon Consulting, Potomac, Maryland 20854, USA; email: [email protected]
      Annual Review of Statistics and Its Application Vol. 5: 169 - 182
      • ...partial least squares discriminant analysis (Nguyen & Rocke 2002, Boulesteix & Strimmer 2007), random forests (Breiman 2001), ...
    • Well-Being Dynamics and Poverty Traps

      Christopher B. Barrett,1 Teevrat Garg,2,3 and Linden McBride11Charles H. Dyson School of Applied Economics and Management, Cornell University, Ithaca, New York 14853; email: [email protected], [email protected]2Grantham Research Institute on Climate Change and Environment, London School of Economics, London, WC2A 2AE, United Kingdom3School of Global Policy and Strategy, University of California, San Diego, La Jolla, California 92093; email: [email protected]
      Annual Review of Resource Economics Vol. 8: 303 - 327
      • ...One alternative nonparametric approach involves using machine learning methods such as classification and regression trees (Breiman 2001, Loh 2002), ...
    • Advances and Challenges in Genomic Selection for Disease Resistance

      Jesse Poland1 and Jessica Rutkoski2,31Wheat Genetics Resource Center, Department of Plant Pathology and Department of Agronomy, Kansas State University, Manhattan, Kansas, 66506; email: [email protected]2Plant Breeding and Genetics Section, Cornell University, Ithaca, New York, 14853; email: [email protected]3International Maize and Wheat Research Center (CIMMYT), Texcoco, Estado de México, 56237 Mexico
      Annual Review of Phytopathology Vol. 54: 79 - 98
      • ...Two such models are Reproducing Kernel Hilbert Space (RKHS) (21) and Random Forest (RF) (9)....
    • League Tables for Hospital Comparisons

      Sharon-Lise T. Normand,1 Arlene S. Ash,2 Stephen E. Fienberg,3 Thérèse A. Stukel,4 Jessica Utts,5 and Thomas A. Louis61Department of Health Care Policy, Harvard Medical School, and Department of Biostatistics, Harvard T.H. Chan School of Public Health, Boston, Massachusetts 02115; email: [email protected]2Department of Quantitative Health Sciences, University of Massachusetts Medical School, Worcester, Massachusetts 01605; email: [email protected]3Department of Statistics, Carnegie Mellon University, Pittsburgh, Pennsylvania 15213; email: [email protected]4Institute for Clinical Evaluative Sciences, Toronto, Ontario M4N 3M5, Canada, and the Institute of Health Policy, Management & Evaluation, University of Toronto, Toronto, Ontario M5T 3M6, Canada, and Dartmouth Institute for Health Policy and Clinical Practice, Hanover, New Hampshire 03766; email: [email protected]5Department of Statistics, University of California, Irvine, California 92697; email: [email protected]6Department of Biostatistics, Johns Hopkins Bloomberg School of Public Health, Baltimore, Maryland 21205; email: [email protected]
      Annual Review of Statistics and Its Application Vol. 3: 21 - 50
      • ...and boosting (Berk 2008, Breiman 2001, Hastie et al. 2009, McCaffrey et al. 2004), ...
    • Computerized Adaptive Diagnosis and Testing of Mental Health Disorders

      Robert D. Gibbons,1 David J. Weiss,2 Ellen Frank,3 and David Kupfer31Center for Health Statistics and Departments of Medicine and Public Health Sciences, University of Chicago, Chicago, Illinois 60612; email: [email protected]2Department of Psychology, University of Minnesota, Minneapolis, Minnesota 554553Department of Psychiatry and Western Psychiatric Institute and Clinic, University of Pittsburgh, Pittsburgh, Pennsylvania 15213
      Annual Review of Clinical Psychology Vol. 12: 83 - 104
      • ...a Diagnostic and Statistical Manual of Mental Disorders (DSM) diagnosis of major depressive disorder (MDD)]. Decision trees (Brieman 2001, Brieman et al. 1984, Quinlan 1993) represent an attractive framework for designing adaptive predictive tests because their corresponding models can be represented as a sequence of binary decisions....
      • ...Decision trees (Brieman 2001, Brieman et al. 1984, Quinlan 1993) represent a model in terms of a flow chart....
      • ...ensemble models constructed of averages of hundreds of decision trees have received considerable attention in statistics and machine learning (Brieman 1996, 2001...
      • ...Random forests require minimal human intervention and have historically exhibited good performance across a wide range of domains (Brieman 2001, Hastie et al. 2009)....
    • Diboson Production at Colliders

      Mark S. NeubauerDepartment of Physics, University of Illinois at Urbana-Champaign, Urbana, Illinois 61801; email: [email protected]
      Annual Review of Nuclear and Particle Science Vol. 61: 223 - 250
      • ...dijet mass) that can sensitively distinguish between signal and background as input to a random forest (75) multivariate event classifier. Figure 9 shows the dijet mass distribution obtained from the results of the random forest output fit....

  • 17.
    Bright I, Lin G, Kutz JN. 2013. Compressive sensing based machine learning strategies for characterizing the flow around a cylinder with limited pressure measurements. Phys. Fluids 25:127102
    • Crossref
    • Web of Science ®
    • ADS
    • Google Scholar
    Article Location
  • 18.
    Brunton SL, Kutz JN. 2019. Data-Driven Science and Engineering: Machine Learning, Dynamical Systems, and Control. New York: Cambridge Univ. Press
    • Crossref
    • Google Scholar
    Article Location
  • 19.
    Brunton SL, Noack BR. 2015. Closed-loop turbulence control: progress and challenges. Appl. Mech. Rev. 67:050801
    • Crossref
    • Web of Science ®
    • ADS
    • Google Scholar
    Article Location
    More AR articles citing this reference

    • Model Reduction for Flow Analysis and Control

      Clarence W. Rowley and Scott T.M. DawsonDepartment of Mechanical and Aerospace Engineering, Princeton University, Princeton, New Jersey 08544; email: [email protected]
      Annual Review of Fluid Mechanics Vol. 49: 387 - 417
      • ... for a recent review of model reduction methods for linear systems and to Brunton & Noack (2015) for a comprehensive overview of methods used for the control of turbulence....

  • 20.
    Brunton SL, Proctor JL, Kutz JN. 2016. Discovering governing equations from data by sparse identification of nonlinear dynamical systems. PNAS 113:3932–37
    • Crossref
    • Medline
    • Web of Science ®
    • ADS
    • Google Scholar
    Article Location
    More AR articles citing this reference

    • Turbulence Modeling in the Age of Data

      Karthik Duraisamy,1, Gianluca Iaccarino,2, and Heng Xiao3,1Department of Aerospace Engineering, University of Michigan, Ann Arbor, Michigan 48109, USA; email: [email protected]2Department of Mechanical Engineering, Stanford University, Stanford, California 94305, USA; email: [email protected]3Kevin T. Crofton Department of Aerospace and Ocean Engineering, Virginia Tech, Blacksburg, Virginia 24060, USA; email: [email protected]
      Annual Review of Fluid Mechanics Vol. 51: 357 - 377
      • ...What is the right balance between data and models? Recent works (Schmidt & Lipson 2009, Brunton et al. 2016, Raissi et al. 2017) have explored the possibility of extracting models purely from data....
    • Improving Lateral Flow Assay Performance Using Computational Modeling

      David Gasperino,1 Ted Baughman,1 Helen V. Hsieh,1 David Bell,1 and Bernhard H. Weigl1,21Intellectual Ventures Laboratory, Bellevue, Washington 98007, USA2Department of Bioengineering, University of Washington, Seattle, Washington 98195, USA
      Annual Review of Analytical Chemistry Vol. 11: 219 - 244
      • ...the approach could be automated to incrementally add reaction complexity to an initially simple reaction framework (105, 106)....

  • 21.
    Bueche D, Stoll P, Dornberger R, Koumoutsakos P. 2002. Multi-objective evolutionary algorithm for the optimization of noisy combustion problems. IEEE Trans. Syst. Man Cybern. C 32:460–73
    • Crossref
    • Google Scholar
    Article Location
  • 22.
    Chen TQ, Rubanova Y, Bettencourt J, Duvenaud DK. 2018. Neural ordinary differential equations. In Advances in Neural Information Processing Systems 31, ed. S Bengio, H Wallach, H Larochelle, K Grauman, N Cesa-Bianchi, R Garnett, pp. 6571–83. Red Hook, NY: Curran Assoc.
    • Google Scholar
    Article Location
    More AR articles citing this reference

    • Koopman Operators for Estimation and Control of Dynamical Systems

      Samuel E. Otto and Clarence W. RowleyDepartment of Mechanical and Aerospace Engineering, Princeton University, Princeton, New Jersey 08544, USA; email: [email protected]
      Annual Review of Control, Robotics, and Autonomous Systems Vol. 4: 59 - 87
      • ...we think that nonlinear evolution of the observables could be easily incorporated into existing recurrent autoencoder architectures using the “neural ODE” framework developed by Chen et al. (44)....

  • 23.
    Cherkassky V, Mulier FM. 2007. Learning from Data: Concepts, Theory, and Methods. Hoboken, NJ: John Wiley & Sons
    • Crossref
    • Google Scholar
    Article Locations:
    • Article Location
    • Article Location
    More AR articles citing this reference

    • Quantitative Molecular Positron Emission Tomography Imaging Using Advanced Deep Learning Techniques

      Habib Zaidi1,2,3,4 and Issam El Naqa5,6,71Division of Nuclear Medicine and Molecular Imaging, Geneva University Hospital, 1211 Geneva, Switzerland; email: [email protected]2Geneva Neuroscience Centre, University of Geneva, 1205 Geneva, Switzerland3Department of Nuclear Medicine and Molecular Imaging, University of Groningen, 9700 RB Groningen, Netherlands4Department of Nuclear Medicine, University of Southern Denmark, DK-5000 Odense, Denmark5Department of Machine Learning, Moffitt Cancer Center, Tampa, Florida 33612, USA6Department of Radiation Oncology, University of Michigan, Ann Arbor, Michigan 48109, USA7Department of Oncology, McGill University, Montreal, Quebec H3A 1G5, Canada
      Annual Review of Biomedical Engineering Vol. 23: 249 - 276
      • ...or analytically by using complexity measures such as the Vapnik–Chervonenkis dimension (138)....

  • 24.
    Colabrese S, Gustavsson K, Celani A, Biferale L. 2017. Flow navigation by smart microswimmers via reinforcement learning. Phys. Rev. Lett. 118:158004
    • Crossref
    • Medline
    • Web of Science ®
    • ADS
    • Google Scholar
    Article Location
    More AR articles citing this reference

    • Advances in Bioconvection

      Martin A. BeesDepartment of Mathematics, University of York, York YO19 5DD, United Kingdom; email: [email protected]
      Annual Review of Fluid Mechanics Vol. 52: 449 - 476
      • ...Colabrese et al. (2017) demonstrated how smart gyrotactic swimmers learn effective strategies (by shifting their preferred swimming direction from the vertical) to escape from situations where they would normally be trapped by flow structures....

  • 25.
    Colabrese S, Gustavsson K, Celani A, Biferale L. 2018. Smart inertial particles. Phys. Rev. Fluids 3:084301
    • Crossref
    • Web of Science ®
    • Google Scholar
    Article Location
  • 26.
    Colvert B, Alsalman M, Kanso E. 2018. Classifying vortex wakes using neural networks. Bioinspiration Biomim. 13:025003
    • Crossref
    • Medline
    • Web of Science ®
    • Google Scholar
    Article Location
  • 27.
    Dissanayake M, Phan-Thien N. 1994. Neural-network-based approximations for solving partial differential equations. Comm. Numer. Meth. Eng. 10:195–201
    • Crossref
    • Google Scholar
    Article Location
  • 28.
    Dong C, Loy CC, He K, Tang X. 2014. Learning a deep convolutional network for image super-resolution. In Computer Vision—ECCV 2014, ed. D Fleet, T Pajdla, B Schiele, T Tuytelaars, pp. 184–99. Cham, Switz.: Springer
    • Crossref
    • Google Scholar
    Article Location
  • 29.
    Donoho DL. 2006. Compressed sensing. IEEE Trans. Inf. Theory 52:1289–306
    • Crossref
    • Web of Science ®
    • Google Scholar
    Article Location
    More AR articles citing this reference

    • System Identification: A Machine Learning Perspective

      A. Chiuso and G. PillonettoDepartment of Information Engineering, University of Padova, 35131 Padova, Italy; email: [email protected]
      Annual Review of Control, Robotics, and Autonomous Systems Vol. 2: 281 - 304
      • ...The use of regularization in inverse problems to find sparse solutions has received significant attention in diverse areas of machine learning and signal processing, including dictionary learning and matrix factorization problems (95), compressive sensing (96), ...
    • Electrophysiological Source Imaging: A Noninvasive Window to Brain Dynamics

      Bin He,1,2 Abbas Sohrabpour,2 Emery Brown,3 and Zhongming Liu41Department of Biomedical Engineering, Carnegie Mellon University, Pittsburgh, Pennsylvania 15213, USA; email: [email protected]2Department of Biomedical Engineering, University of Minnesota, Minneapolis, Minnesota 55455, USA3Institute for Medical Engineering and Science, Massachusetts Institute of Technology, Cambridge, Massachusetts 02139, USA4Weldon School of Biomedical Engineering, School of Electrical and Computer Engineering, and Purdue Institute of Integrative Neuroscience, Purdue University, West Lafayette, Indiana 47906, USA
      Annual Review of Biomedical Engineering Vol. 20: 171 - 196
      • ...sources may be piecewise continuous; although the source distribution is not sparse, its edge or spatial gradient is sparse (56)....
      • ...sparsity is imposed in other domains, such as the gradient domain (56), ...
    • Coded Apertures in Mass Spectrometry

      Jason J. Amsden,1 Michael E. Gehm,1 Zachary E. Russell,2 Evan X. Chen,1 Shane T. Di Dona,1 Scott D. Wolter,3 Ryan M. Danell,4 Gottfried Kibelka,5 Charles B. Parker,1 Brian R. Stoner,1,6 David J. Brady,1 and Jeffrey T. Glass11Department of Electrical and Computer Engineering, Duke University, Durham, North Carolina 27708; email: [email protected]2Ion Innovations, Roswell, Georgia 300753Department of Physics, Elon University, Elon, North Carolina 272784Danell Consulting, Winterville, North Carolina 285905CMS Field Products, OI Analytical, Pelham, Alabama 351246Discovery Science and Technology, RTI International, Research Triangle Park, North Carolina 27709
      Annual Review of Analytical Chemistry Vol. 10: 141 - 156
      • ...For example, compressive sensing theory (5–9, 10) is the realization that natural signals contain correlations that render full sampling redundant and allow signal recovery from undersampled measurement; adaptive sensing theory (11...
    • Magnetic Resonance Imaging and Velocity Mapping in Chemical Engineering Applications

      Lynn F. Gladden and Andrew J. SedermanDepartment of Chemical Engineering and Biotechnology, University of Cambridge, Cambridge CB2 3RA, United Kingdom; email: [email protected]ac.uk, [email protected]
      Annual Review of Chemical and Biomolecular Engineering Vol. 8: 227 - 247
      • ...Two such methods that are now becoming more widely used in MR are compressed sensing (32...
    • Structured Regularizers for High-Dimensional Problems: Statistical and Computational Issues

      Martin J. WainwrightDepartment of Statistics and Department of Electrical Engineering and Computer Sciences, University of California, Berkeley, California 94704; email: [email protected]
      Annual Review of Statistics and Its Application Vol. 1: 233 - 253
      • ...Bickel et al. 2009, Candes & Tao 2007, Donoho 2006, Donoho & Tanner 2008, Zhang & Huang 2008), ...
    • Electron Tomography in the (S)TEM: From Nanoscale Morphological Analysis to 3D Atomic Imaging

      Zineb Saghi1 and Paul A. Midgley21BIONAND (Andalusian Center for Nanomedicine and Biotechnology), 29590, Málaga, Spain; email: [email protected]2Department of Materials Science and Metallurgy, University of Cambridge, Cambridge CB2 3QZ, United Kingdom; email: [email protected]
      Annual Review of Materials Research Vol. 42: 59 - 79
      • ...Compressed sensing-electron tomography (CS-ET) takes advantage of the sparsity of the tomogram in a chosen transform domain and employs a convex optimization algorithm to find the sparsest solution in the domain, subject to consistency with the acquired data (79, 80)....
    • Compressed Sensing, Sparsity, and Dimensionality in Neuronal Information Processing and Data Analysis

      Surya Ganguli1 and Haim Sompolinsky2,31Department of Applied Physics, Stanford University, Stanford, California 94305; email: [email protected]2Edmond and Lily Safra Center for Brain Sciences, Interdisciplinary Center for Neural Computation, Hebrew University, Jerusalem 91904, Israel; email: [email protected]3Center for Brain Science, Harvard University, Cambridge, Massachusetts 02138
      Annual Review of Neuroscience Vol. 35: 485 - 508
      • ...the field of compressed sensing (CS) (Candes et al. 2006, Candes & Tao 2006, Donoho 2006...
      • ...investigators have further shown that no measurement matrices and no reconstruction algorithm can yield sparse signal recovery with substantially fewer measurements (Candes & Tao 2006, Donoho 2006) than that shown in Equation 1....

  • 30.
    Dracopoulos DC. 1997. Evolutionary Learning Algorithms for Neural Adaptive Control. London: Springer-Verlag
    • Crossref
    • Google Scholar
    Article Location
  • 31.
    Duraisamy K, Iaccarino G, Xiao H. 2019. Turbulence modeling in the age of data. Annu. Rev. Fluid Mech. 51:357–77
    • Link
    • Web of Science ®
    • ADS
    • Google Scholar
  • 32.
    Duriez T, Brunton SL, Noack BR. 2016. Machine Learning Control: Taming Nonlinear Dynamics and Turbulence. Cham, Switz.: Springer
    • Google Scholar
    Article Location
  • 33.
    Faller WE, Schreck SJ. 1996. Neural networks: applications and opportunities in aeronautics. Prog. Aerosp. Sci. 32:433–56
    • Crossref
    • Web of Science ®
    • Google Scholar
    Article Location
  • 34.
    Fleming PJ, Purshouse RC. 2002. Evolutionary algorithms in control systems engineering: a survey. Control Eng. Pract. 10:1223–41
    • Crossref
    • Web of Science ®
    • Google Scholar
    Article Location
  • 35.
    Freeman WT, Jones TR, Pasztor EC. 2002. Example-based super-resolution. IEEE Comput. Graph. 22:56–65
    • Crossref
    • Web of Science ®
    • Google Scholar
    Article Location
  • 36.
    Fukami K, Fukagata K, Taira K. 2018. Super-resolution reconstruction of turbulent flows with machine learning. arXiv:1811.11328 [physics.flu-dyn]
    • ADS
    • Google Scholar
    Article Location
  • 37.
    Gardner E. 1988. The space of interactions in neural network models. J. Phys. A 21:257
    • Crossref
    • Web of Science ®
    • ADS
    • Google Scholar
    More AR articles citing this reference

    • Statistical Mechanics of Deep Learning

      Yasaman Bahri,1 Jonathan Kadmon,2 Jeffrey Pennington,1 Sam S. Schoenholz,1 Jascha Sohl-Dickstein,1 and Surya Ganguli1,21Google Brain, Google Inc., Mountain View, California 94043, USA2Department of Applied Physics, Stanford University, Stanford, California 94035, USA; email: [email protected]
      Annual Review of Condensed Matter Physics Vol. 11: 501 - 528
      • ...the focus in physics has been on asymptotically exact calculations of training and test errors in a thermodynamic limit in which both N and P become large but the measurement density α = P/N remains O(1) (126...
    • Compressed Sensing, Sparsity, and Dimensionality in Neuronal Information Processing and Data Analysis

      Surya Ganguli1 and Haim Sompolinsky2,31Department of Applied Physics, Stanford University, Stanford, California 94305; email: [email protected]2Edmond and Lily Safra Center for Brain Sciences, Interdisciplinary Center for Neural Computation, Hebrew University, Jerusalem 91904, Israel; email: [email protected]3Center for Brain Science, Harvard University, Cambridge, Massachusetts 02138
      Annual Review of Neuroscience Vol. 35: 485 - 508
      • ...This question has been studied exhaustively (Gardner 1988, Seung et al. 1992)...

  • 38.
    Gazzola M, Hejazialhosseini B, Koumoutsakos P. 2014. Reinforcement learning and wavelet adapted vortex methods for simulations of self-propelled swimmers. SIAM J. Sci. Comput. 36:B622–39
    • Crossref
    • Web of Science ®
    • Google Scholar
    Article Location
  • 39.
    Gazzola M, Tchieu A, Alexeev D, De Brauer A, Koumoutsakos P. 2016. Learning to school in the presence of hydrodynamic interactions. J. Fluid Mech. 789:726–49
    • Crossref
    • Web of Science ®
    • ADS
    • Google Scholar
    Article Location
  • 40.
    Gazzola M, Van Rees WM, Koumoutsakos P. 2012. C-start: optimal start of larval fish. J. Fluid Mech. 698:5–18
    • Crossref
    • Web of Science ®
    • ADS
    • Google Scholar
    Article Location
    More AR articles citing this reference

    • Immersed Methods for Fluid–Structure Interaction

      Boyce E. Griffith1 and Neelesh A. Patankar21Departments of Mathematics, Applied Physical Sciences, and Biomedical Engineering, University of North Carolina, Chapel Hill, North Carolina 27599, USA; email: [email protected]2Department of Mechanical Engineering, Northwestern University, Evanston, Illinois 60208, USA; email: [email protected]
      Annual Review of Fluid Mechanics Vol. 52: 421 - 448
      • ...; Borazjani & Sotiropoulos 2008, 2009; Shirgaonkar et al. 2008; Gazzola et al. 2012...
    • Biomimetic Survival Hydrodynamics and Flow Sensing

      Michael S. Triantafyllou,1 Gabriel D. Weymouth,2 and Jianmin Miao31Department of Mechanical Engineering, Massachusetts Institute of Technology, Cambridge, Massachusetts 02139; email: [email protected]2Southampton Marine and Maritime Institute, University of Southampton, SO16 7QF Southampton, United Kingdom3School of Mechanical and Aerospace Engineering, Nanyang Technological University, 639798 Singapore
      Annual Review of Fluid Mechanics Vol. 48: 1 - 24
      • ...Gazzola et al. (2012) optimized through simulation a single type of maneuver (C-shaped) of a simply shaped fish (zebrafish larvae) that involves only its body and caudal fin....
      • ...Simulation shed further light on the actuation mechanics. Gazzola et al. (2012) used tracer particles to connect the flow features observed during the first phase of the maneuver, ...
      • ...a principal conclusion of Gazzola et al. (2012) is that the strikingly large initial curvature of the C-start maneuver serves to engage as much fluid mass as possible: The body motion conveys significant kinetic (added mass–related) energy, ...
      • ...Figure adapted with permission from Gazzola et al. (2012)....

  • 41.
    Germano M, Piomelli U, Moin P, Cabot WH. 1991. A dynamic subgrid-scale eddy viscosity model. Phys. Fluids 3:1760–65
    • Crossref
    • Web of Science ®
    • ADS
    • Google Scholar
    Article Location
    More AR articles citing this reference

    • Statistical Properties of Subgrid-Scale Turbulence Models

      Robert D. Moser,1,2 Sigfried W. Haering,3 and Gopal R. Yalla11Oden Institute for Computational Engineering and Sciences, University of Texas at Austin, Austin, Texas 78712, USA; email: [email protected]2Walker Department of Mechanical Engineering, University of Texas at Austin, Austin, Texas 78712, USA3Argonne National Laboratory, Lemont, Illinois 60439, USA
      Annual Review of Fluid Mechanics Vol. 53: 255 - 286
      • ...The dynamic procedure (Germano et al. 1991) has been widely used to determine constants in subgrid models....
      • ...this relation is overdetermined since C is just a scalar in a tensor equation. Germano et al. (1991) originally proposed contracting this equation with to obtain a scalar equation, ...
      • ...a model of this type that conserves momentum and is formulated for a priori consistency of dissipation outperforms dynamic Smagorinsky (Germano et al. 1991)...
      • ... is to determine the optimal estimate dynamically based on information from a coarser filtering of the LES solution, as in the dynamic model (Germano et al. 1991)....
    • Toward Constitutive Models for Momentum, Species, and Energy Transport in Gas–Particle Flows

      Sankaran Sundaresan, Ali Ozel, and Jari KolehmainenDepartment of Chemical and Biological Engineering, Princeton University, Princeton, New Jersey 08544, USA; email: [email protected]
      Annual Review of Chemical and Biomolecular Engineering Vol. 9: 61 - 81
      • ...Dynamic adjustment of the Smagorinsky constant using a scale similarity approach (78) is common in single-phase turbulence modeling....
    • Wall-Modeled Large-Eddy Simulation for Complex Turbulent Flows

      Sanjeeb T. Bose1,2 and George Ilhwan Park3,41Cascade Technologies, Inc., Palo Alto, California 94303; email: [email protected]2Institute for Computational and Mathematical Engineering, Stanford University, Stanford, California 943053Center for Turbulence Research, Stanford University, Stanford, California 943054Department of Mechanical Engineering and Applied Mechanics, University of Pennsylvania, Philadelphia, Pennsylvania 19104
      Annual Review of Fluid Mechanics Vol. 50: 535 - 561
      • ...dynamic Smagorinsky model (Germano et al. 1991, Lilly 1992)] can be adjusted to prevent the log-layer mismatch by a proper consideration of the turbulent kinetic energy budget....
    • Aerodynamics of Heavy Vehicles

      Haecheon Choi,1,2 Jungil Lee,2,3 and Hyungmin Park11School of Mechanical and Aerospace Engineering and2Institute of Advanced Machinery and Design, Seoul National University, Seoul 151-744, Korea; email: [email protected], [email protected]3Department of Mechanical Engineering, Ajou University, Suwon 443-749, Korea; email: [email protected]
      Annual Review of Fluid Mechanics Vol. 46: 441 - 468
      • ...The well-known dynamic Smagorinsky model (Germano et al. 1991, Lilly 1992) overcame this weakness, ...
    • Physics and Computation of Aero-Optics

      Meng Wang,1 Ali Mani,2 and Stanislav Gordeyev11Department of Aerospace and Mechanical Engineering, University of Notre Dame, Notre Dame, Indiana 46556; email: [email protected], [email protected], [email protected]2Department of Mechanical Engineering, Stanford University, Stanford, California 94305; email: [email protected]
      Annual Review of Fluid Mechanics Vol. 44: 299 - 321
      • ... for which the dynamic procedure (Germano et al. 1991, Moin et al. 1991, Lilly 1992) provides a robust way of computing the model coefficient from the resolved scales, ...
    • Wavelet Methods in Computational Fluid Dynamics

      Kai Schneider1,2 and Oleg V. Vasilyev31Laboratoire de Modélisation en Mećanique et Procédés Propres, CNRS et Universités d'Aix-Marseille, 13451 Marseille cedex 20, France; email: [email protected]2Centre de Mathématiques et d'Informatique, Université d'Aix-Marseille I, 13453 Marseille cedex 13, France3Department of Mechanical Engineering, University of Colorado, Boulder, Colorado 80309; email: [email protected]olorado.edu
      Annual Review of Fluid Mechanics Vol. 42: 473 - 503
      • ...see Germano et al. 1991, Lesieur & Métais 1996, Meneveau & Katz 2000)....
    • Magnetohydrodynamic Turbulence at Low Magnetic Reynolds Number

      Bernard Knaepen1 and René Moreau21Université Libre de Bruxelles, Boulevard du Triomphe, Campus Plaine CP231, B-1050 Ixelles Belgium; email: [email protected]2Laboratoire EPM, ENSHMG, BP 95, 38402 Saint-Martin d’Hères France; email: [email protected]
      Annual Review of Fluid Mechanics Vol. 40: 25 - 45
      • ...that is automatically calibrated during the simulation (Germano et al. 1991)....
    • COMPUTATIONAL PREDICTION OF FLOW-GENERATED SOUND

      Meng Wang,1 Jonathan B. Freund,2 and Sanjiva K. Lele31Center for Turbulence Research, Stanford University/NASA Ames Research Center, Stanford, California 94305; email: [email protected]2Department of Theoretical and Applied Mechanics, University of Illinois at Urbana-Champaign, Urbana, Illinois 61801; email: [email protected]3Departments of Mechanical Engineering and Aeronautics & Astronautics, Stanford University, Stanford, California 94305; email: [email protected]
      Annual Review of Fluid Mechanics Vol. 38: 483 - 512
      • ...which requires no adjustable constant and near-wall damping function (Germano et al. 1991, Lilly 1992), ...
    • LARGE-EDDY SIMULATION OF TURBULENT COMBUSTION

      Heinz PitschDepartment of Mechanical Engineering, Stanford University, Stanford, California 94305; email: [email protected]
      Annual Review of Fluid Mechanics Vol. 38: 453 - 482
      • ...knowledge of the large-scale dynamics and the assumption that an applied model should be valid independently of the filter size leads to the formulation of the so-called dynamic models (Germano et al. 1991, Moin et al. 1991), ...
    • TURBULENT MIXING

      Paul E. DimotakisGraduate Aeronautical Laboratories, California Institute of Technology, Pasadena, California 91125; email: [email protected]
      Annual Review of Fluid Mechanics Vol. 37: 329 - 356
      • ...; scale-dependent dynamic models that observe the Germano et al. (1991)...
    • WALL-LAYER MODELS FOR LARGE-EDDY SIMULATIONS

      Ugo Piomelli and Elias BalarasDepartment of Mechanical Engineering, University of Maryland, College Park, Maryland 20742; e-mail: [email protected] , [email protected]
      Annual Review of Fluid Mechanics Vol. 34: 349 - 374
      • ...the explicit filtering operations commonly used to perform a dynamic evaluation of the model coefficients (Germano et al. 1991) are ill defined near the wall....
    • Scale-Invariance and Turbulence Models for Large-Eddy Simulation

      Charles Meneveau1 and Joseph Katz11Department of Mechanical Engineering, Center for Environmental and Applied Fluid Mechanics, The Johns Hopkins University, Baltimore, Maryland 21218; email: [email protected]
      Annual Review of Fluid Mechanics Vol. 32: 1 - 32
      • ...the dynamic Smagorinsky model (Germano et al 1991) is reviewed in Section 4....
      • ...The dynamic model (Germano et al 1991) consists of using the resolved scales to measure the model coefficient during the simulation, ...
      • ...Originally, Germano et al (1991) have contracted Equation 10 with ....
      • ...The procedure of averaging over directions of statistical homogeneity used by Germano et al (1991) circumvents these problems....

  • 42.
    Giannakoglou K, Papadimitriou D, Kampolis I. 2006. Aerodynamic shape design using evolutionary algorithms and new gradient-assisted metamodels. Comput. Methods Appl. Mech. Eng. 195:6312–29
    • Crossref
    • Web of Science ®
    • ADS
    • Google Scholar
    Article Location
  • 43.
    Glaz B, Liu L, Friedmann PP. 2010. Reduced-order nonlinear unsteady aerodynamic modeling using a surrogate-based recurrence framework. AIAA J. 48:2418–29
    • Crossref
    • ADS
    • Google Scholar
    Article Location
  • 44.
    Gonzalez-Garcia R, Rico-Martinez R, Kevrekidis I. 1998. Identification of distributed parameter systems: a neural net based approach. Comput. Chem. Eng. 22:S965–68
    • Crossref
    • Web of Science ®
    • Google Scholar
    Article Location
  • 45.
    Goodfellow I, Bengio Y, Courville A. 2016. Deep Learning. Cambridge, MA: MIT Press
    • Google Scholar
    Article Locations:
    • Article Location
    • Article Location
    More AR articles citing this reference

    • Machine Learning for Sustainable Energy Systems

      Priya L. Donti1,2 and J. Zico Kolter1,31Computer Science Department, Carnegie Mellon University, Pittsburgh, Pennsylvania 15213, USA; email: [email protected]2Department of Engineering and Public Policy, Carnegie Mellon University, Pittsburgh, Pennsylvania 15213, USA3Bosch Center for Artificial Intelligence, Pittsburgh, Pennsylvania 15222, USA
      Annual Review of Environment and Resources Vol. 46:
      • ...Deep learning (25, 26), currently one of the more prominent approaches to ML, ...
    • Lightness Perception in Complex Scenes

      Richard F. MurrayDepartment of Psychology and Centre for Vision Research, York University, Toronto M3J 1P3, Canada; email: [email protected]
      Annual Review of Vision Science Vol. 7: 417 - 436
      • ...Two examples of this are Markov random fields (Koller & Friedman 2009) and artificial neural networks (Goodfellow et al. 2016), ...
      • ...We have computational tools that can turn hypotheses about the natural scene statistics that guide lightness perception into algorithms for inference (Goodfellow et al. 2016, Koller & Friedman 2009)....
    • Estimating DSGE Models: Recent Advances and Future Challenges

      Jesús Fernández-Villaverde1 and Pablo A. Guerrón-Quintana21Department of Economics, University of Pennsylvania, Philadelphia, Pennsylvania 19104, USA; email: [email protected]2Department of Economics, Boston College, Chestnut Hill, Massachusetts 02467, USA; email: [email protected]
      Annual Review of Economics Vol. 13: 229 - 252
      • ...Related to this point is the third challenge for DSGE models: the incorporation of machine learning methods (Goodfellow et al. 2016)....
    • Quantitative Molecular Positron Emission Tomography Imaging Using Advanced Deep Learning Techniques

      Habib Zaidi1,2,3,4 and Issam El Naqa5,6,71Division of Nuclear Medicine and Molecular Imaging, Geneva University Hospital, 1211 Geneva, Switzerland; email: [email protected]2Geneva Neuroscience Centre, University of Geneva, 1205 Geneva, Switzerland3Department of Nuclear Medicine and Molecular Imaging, University of Groningen, 9700 RB Groningen, Netherlands4Department of Nuclear Medicine, University of Southern Denmark, DK-5000 Odense, Denmark5Department of Machine Learning, Moffitt Cancer Center, Tampa, Florida 33612, USA6Department of Radiation Oncology, University of Michigan, Ann Arbor, Michigan 48109, USA7Department of Oncology, McGill University, Montreal, Quebec H3A 1G5, Canada
      Annual Review of Biomedical Engineering Vol. 23: 249 - 276
      • ...A typical CNN (Figure 2) consists of several specialized layers, including convolutional and pooling layers (45)....
    • Inferring Macroscale Brain Dynamics via Fusion of Simultaneous EEG-fMRI

      Marios G. Philiastides,1 Tao Tu,2 and Paul Sajda31Institute of Neuroscience and Psychology, University of Glasgow, Glasgow G12 8AD, Scotland; email: [email protected]2Department of Biomedical Engineering, Columbia University, New York, NY 10027, USA3Departments of Biomedical Engineering, Electrical Engineering, and Radiology and the Data Science Institute, Columbia University, New York, NY 10027, USA; email: [email protected]
      Annual Review of Neuroscience Vol. 44: 315 - 334
      • ...deep learning (Goodfellow et al. 2016) offers an exciting approach to consider when fusing EEG-fMRI....
    • Applications of Machine and Deep Learning in Adaptive Immunity

      Margarita Pertseva,1,2 Beichen Gao,1 Daniel Neumeier,1 Alexander Yermanos,1,3,4 and Sai T. Reddy11Department of Biosystems Science and Engineering, ETH Zurich, 4058 Basel, Switzerland; email: [email protected]2Life Science Zurich Graduate School, ETH Zurich and University of Zurich, 8006 Zurich, Switzerland3Department of Pathology and Immunology, University of Geneva, 1205 Geneva, Switzerland4Department of Biology, Institute of Microbiology and Immunology, ETH Zurich, 8093 Zurich, Switzerland
      Annual Review of Chemical and Biomolecular Engineering Vol. 12: 39 - 62
      • ...especially those that utilize neural networks and DL, is the lack of interpretability (51)....
      • ...Full coverage of DL models is outside the scope of this review; the interested reader could refer to several additional resources (50, 51, 55)....
      • ...The structure of training data also dictates which type of learning can be performed: supervised or unsupervised (51)....
      • ...Typical unsupervised learning tasks are data clustering and dimensionality reduction performed with methods such as principal component analysis or k-means clustering (51)....
    • Machine Learning for Social Science: An Agnostic Approach

      Justin Grimmer,1 Margaret E. Roberts,2 and Brandon M. Stewart31Department of Political Science and Hoover Institution, Stanford University, Stanford, California 94305, USA; email: [email protected]2Department of Political Science and Halıcıoğlu Data Science Institute, University of California San Diego, La Jolla, California 92093, USA; email: [email protected]3Department of Sociology and Office of Population Research, Princeton University, Princeton, New Jersey 08540, USA; email: [email protected]
      Annual Review of Political Science Vol. 24: 395 - 419
      • ...we point the reader to accessible textbooks (Bishop 2006, Murphy 2012, Goodfellow et al. 2016)...
    • Koopman Operators for Estimation and Control of Dynamical Systems

      Samuel E. Otto and Clarence W. RowleyDepartment of Mechanical and Aerospace Engineering, Princeton University, Princeton, New Jersey 08544, USA; email: [email protected]
      Annual Review of Control, Robotics, and Autonomous Systems Vol. 4: 59 - 87
      • ...This idea has led to approximation techniques for the Koopman operator involving learned dictionaries (35) or deep learning autoencoders (36) suitably adapted for this task....
    • Artificial Intelligence, Predictive Policing, and Risk Assessment for Law Enforcement

      Richard A. BerkDepartments of Statistics and Criminology, University of Pennsylvania, Philadelphia, Pennsylvania 19104, USA; email: [email protected]
      Annual Review of Criminology Vol. 4: 209 - 237
      • ...A related conception works backward from particular tasks whose execution exemplifies artificial intelligence (Goodfellow et al. 2016, ...
      • ...The details are too lengthy to specify here and can vary depending on the kind of deep learning in play (Goodfellow et al. 2016)....
      • ...although there are several other regularization tools that are commonly used (Goodfellow et al. 2016)....
      • ...The convolution processes can be seen as part of data preprocessing that extracts the essential features of the image to be passed along as predictors to a neural network (Goodfellow et al. 2016)....
      • ...Recurrent neural networks (RNNs) have been developed for pooled cross-section data (Goodfellow et al. 2016) so that the temporal structure can be exploited....
      • ...Autoencoder (Goodfellow et al. 2016) is a new tool that can be seen as a nonlinear version of principal component analysis....
    • Identifying Regulatory Elements via Deep Learning

      Mira Barshai,1, Eitamar Tripto,2, and Yaron Orenstein11School of Electrical and Computer Engineering, Ben-Gurion University of the Negev, Beer-Sheva 8410501, Israel; email: [email protected]2Department of Biomedical Engineering, Ben-Gurion University of the Negev, Beer-Sheva 8410501, Israel
      Annual Review of Biomedical Data Science Vol. 3: 315 - 338
      • ...with the aim of forcing the learned representations of the input to assume useful properties (59)....
    • Toward Community-Driven Big Open Brain Science: Open Big Data and Tools for Structure, Function, and Genetics

      Adam S. Charles,1,2 Benjamin Falk,1 Nicholas Turner,3 Talmo D. Pereira,4 Daniel Tward,1 Benjamin D. Pedigo,1 Jaewon Chung,1 Randal Burns,1 Satrajit S. Ghosh,5,6 Justus M. Kebschull,1,7 William Silversmith,4 and Joshua T. Vogelstein1,21Department of Biomedical Engineering, Johns Hopkins University, Baltimore, Maryland 21218, USA; email: [email protected]2Institute for Computational Medicine, Kavli Neuroscience Discovery Institute, and Center for Imaging Science, Johns Hopkins University, Baltimore, Maryland 21218, USA3Department of Computer Science, Princeton University, Princeton, New Jersey 08540, USA4Princeton Neuroscience Institute, Princeton University, Princeton, New Jersey 08540, USA5McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, Massachusetts 02139, USA6Department of Otolaryngology–Head and Neck Surgery, Harvard Medical School, Boston, Massachusetts 02115, USA7Stanford University, Palo Alto, California 94305, USA
      Annual Review of Neuroscience Vol. 43: 441 - 464
      • ...The leading image analysis tools are CNNs (Goodfellow et al. 2016), ...
      • ...such as Long Short-Term Memory and other recurrent neural networks (RNNs) (Goodfellow et al. 2016), ...
    • Opportunities and Challenges for Machine Learning in Materials Science

      Dane Morgan and Ryan JacobsDepartment of Materials Science and Engineering, University of Wisconsin–Madison, Madison, Wisconsin 53706, USA; email: [email protected], [email protected]
      Annual Review of Materials Research Vol. 50: 71 - 103
      • ...more detailed information on these general ML methods is covered in References 41–44....
    • Machine Learning for Molecular Simulation

      Frank Noé,1,2,3 Alexandre Tkatchenko,4 Klaus-Robert Müller,5,6,7 and Cecilia Clementi1,3,81Department of Mathematics and Computer Science, Freie Universität Berlin, 14195 Berlin, Germany; email: [email protected]2Department of Physics, Freie Universität Berlin, 14195 Berlin, Germany3Department of Chemistry and Center for Theoretical Biological Physics, Rice University, Houston, Texas 77005, USA; email: [email protected]4Physics and Materials Science Research Unit, University of Luxembourg, 1511 Luxembourg, Luxembourg; email: [email protected]5Department of Computer Science, Technical University Berlin, 10587 Berlin, Germany; email: [email protected]6Max-Planck-Institut für Informatik, 66123 Saarbrücken, Germany7Department of Brain and Cognitive Engineering, Korea University, Seoul 136-713, South Korea8Department of Physics, Rice University, Houston, Texas 77005, USA
      Annual Review of Physical Chemistry Vol. 71: 361 - 390
      • ...and we refer to the literature for an introduction to statistical learning theory (3, 4) and deep learning (5, 6)....
    • Big Data in Industrial-Organizational Psychology and Human Resource Management: Forward Progress for Organizational Research and Practice

      Frederick L. Oswald,1 Tara S. Behrend,2 Dan J. Putka,3 and Evan Sinar41Department of Psychological Sciences, Rice University, Houston, Texas 77005, USA; email: [email protected]2Department of Organizational Sciences and Communication, George Washington University, Washington, DC 20052, USA3Human Resources Research Organization, Alexandria, Virginia 22314, USA4BetterUp, Pittsburgh, Pennsylvania 15243, USA
      Annual Review of Organizational Psychology and Organizational Behavior Vol. 7: 505 - 533
      • ... and modern variants on neural networks, namely deep learning (Goodfellow et al. 2016)....
    • Massively Parallel Assays and Quantitative Sequence–Function Relationships

      Justin B. Kinney and David M. McCandlishSimons Center for Quantitative Biology, Cold Spring Harbor Laboratory, Cold Spring Harbor, New York 11724, USA; email: [email protected], [email protected]
      Annual Review of Genomics and Human Genetics Vol. 20: 99 - 127
      • ...One exciting and increasingly popular strategy that might help to answer some of these lingering questions is the use of deep learning techniques (48) for modeling protein–DNA and protein–RNA specificity....
    • Machine Learning for Sociology

      Mario Molina and Filiz GaripDepartment of Sociology, Cornell University, Ithaca, New York 14853, USA; email: [email protected], [email protected]
      Annual Review of Sociology Vol. 45: 27 - 45
      • ...We can also categorize ML tools by the kind of experience they are allowed to have during the learning process (Goodfellow et al. 2016), ...
      • ...but the best method for the particular question at hand (Goodfellow et al. 2016, ...
      • ...UML searches for a representation of the inputs that is more useful than X itself (Goodfellow et al. 2016)....
      • ...1Supervised and unsupervised learning are not formally defined terms (Goodfellow et al. 2016)....
    • Sentiment Analysis

      Robert A. StineDepartment of Statistics, University of Pennsylvania, Philadelphia, Pennsylvania 19104, USA; email: [email protected]
      Annual Review of Statistics and Its Application Vol. 6: 287 - 308
      • ...problems in sentiment analysis remain a popular benchmark. Goodfellow et al. (2016) provide an overview of deep learning in general, ...
    • Deep Learning and Its Application to LHC Physics

      Dan Guest,1 Kyle Cranmer,2 and Daniel Whiteson11Department of Physics and Astronomy, University of California, Irvine, California 92697, USA2Physics Department, New York University, New York, NY 10003, USA
      Annual Review of Nuclear and Particle Science Vol. 68: 161 - 181
      • ...Modern deep learning is characterized by the composition of modular, differentiable components (20)....

  • 46.
    Goodfellow I, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, 2014. Generative adversarial nets. In Advances in Neural Information Processing Systems 27, ed. Z Ghahramani, M Welling, C Cortes, ND Lawrence, KQ Weinberger, pp. 2672–80. Red Hook, NY: Curran Assoc. Powerful deep learning architecture that learns through a game between a network that can generate new data and a network that is an expert classifier.
    • Google Scholar
    Article Locations:
    • Article Location
    • Article Location
    More AR articles citing this reference

    • Probabilistic Machine Learning for Healthcare

      Irene Y. Chen,1, Shalmali Joshi,2, Marzyeh Ghassemi,2,3 and Rajesh Ranganath4,5,61Department of Electrical Engineering and Computer Science, Massachusetts Institute of Technology, Cambridge, Massachusetts 02139, USA; email: [email protected]2Vector Institute, Toronto, Ontario M5G 1M1, Canada; email: [email protected]3Institute for Medical and Evaluative Sciences, Massachusetts Institute of Technology, Cambridge, Massachusetts 02139, USA4Department of Computer Science, Courant Institute, New York University, New York, NY 10012, USA5Center for Data Science, New York University, New York, NY 10012, USA6Department of Population Health, New York University Grossman School of Medicine, New York, NY 10016, USA
      Annual Review of Biomedical Data Science Vol. 4: 393 - 415
      • ...Generative adversarial networks (GANs) (104) use two neural networks to first create artificial imitations of the training data and then separately decide whether a given sample was genuine or counterfeit....
    • Representation Learning: A Statistical Perspective

      Jianwen Xie,1 Ruiqi Gao,2 Erik Nijkamp,2 Song-Chun Zhu,2 and Ying Nian Wu21Hikvision Research Institute, Santa Clara, California 95054, USA2Department of Statistics, University of California, Los Angeles, California 90095, USA; email: [email protected]
      Annual Review of Statistics and Its Application Vol. 7: 303 - 335
      • ...then the resulting model is commonly called generator network (Goodfellow et al. 2014, Kingma & Welling 2014)....
      • ...as in generative adversarial networks (GAN) (Goodfellow et al. 2014, Radford et al. 2015), ...
      • ...Sharp synthesis can be achieved by GAN (Goodfellow et al. 2014, Radford et al. 2015), ...
      • ...It is different from GAN (Goodfellow et al. 2014), in which the discriminator eventually becomes confused because the generated data become similar to the real data....
    • Deep Learning: The Good, the Bad, and the Ugly

      Thomas SerreDepartment of Cognitive Linguistic and Psychological Sciences, Carney Institute for Brain Science, Brown University, Providence, Rhode Island 02818, USA; email: [email protected]
      Annual Review of Vision Science Vol. 5: 399 - 426
      • ...Another recent method for image synthesis that is gaining in popularity is the generative adversarial network (GAN) (Goodfellow et al. 2014)....

  • 47.
    Grant I, Pan X. 1995. An investigation of the performance of multi layer, neural networks applied to the analysis of PIV images. Exp. Fluids 19:159–66
    • Crossref
    • Web of Science ®
    • Google Scholar
    Article Location
  • 48.
    Graves A, Fernández S, Schmidhuber J. 2007. Multi-dimensional recurrent neural networks. In Artificial Neural Networks—ICANN 2007, ed. JM de Sa, LA Alexandre, W Duch, D Mandic, pp. 549–58. Berlin: Springer
    • Crossref
    • Google Scholar
    Article Location
  • 49.
    Grossberg S. 1976. Adaptive pattern classification and universal recoding, I: parallel development and coding of neural feature detectors. Biol. Cybernet. 23:121–34
    • Crossref
    • Google Scholar
    Article Locations:
    • Article Location
    • Article Location
    More AR articles citing this reference

    • Long-Range Neural Synchrony in Behavior

      Alexander Z. Harris1,2 and Joshua A. Gordon1,21Department of Psychiatry, Columbia University, New York, New York 10032; email: [email protected], [email protected]2Division of Integrative Neuroscience, New York State Psychiatric Institute, New York, New York 10032
      Annual Review of Neuroscience Vol. 38: 171 - 194
      • ...The possibility that gamma synchrony could solve the so-called binding problem is grounded in previous theoretical models (Grossberg 1976, Milner 1974, von der Malsburg 1981)....
    • PSYCHOBIOLOGICAL MODELS OF HIPPOCAMPAL FUNCTION IN LEARNING AND MEMORY

      Mark A. Gluck and Catherine E. MyersCenter for Molecular and Behavioral Neuroscience, Rutgers University, 197 University Avenue, Newark, New Jersey; e-mail: [email protected]
      Annual Review of Psychology Vol. 48: 481 - 514
      • ...Node j becomes active if the sum of inputs exceeds some firing threshold (cf Grossberg 1976, Kohonen 1984, McCulloch & Pitts 1943, Rosenblatt 1962): 1 In Equation 1, ...
      • ...such as nonspecific modulatory influences that determine the storage rates in CA3 (Grossberg 1976, Hasselmo et al 1995, Murre 1996, Treves & Rolls 1992), ...
      • ...The resulting network is similar to the unsupervised, competitive-learning systems developed by Kohonen (1984), Rumelhart & Zipser (1985), Grossberg (1976), ...

  • 50.
    Grossberg S. 1988. Nonlinear neural networks: Principles, mechanisms, and architectures. Neural Netw. 1:17–61
    • Crossref
    • Google Scholar
    Article Locations:
    • Article Location
    • Article Location
    • Article Location
    • Article Location
  • 51.
    Guéniat F, Mathelin L, Hussaini MY. 2016. A statistical learning strategy for closed-loop control of fluid flows. Theor. Comput. Fluid Dyn. 30:497–510
    • Crossref
    • Web of Science ®
    • Google Scholar
  • 52.
    Halko N, Martinsson PG, Tropp JA. 2011. Finding structure with randomness: probabilistic algorithms for constructing approximate matrix decompositions. SIAM Rev. 53:217–88
    • Crossref
    • Web of Science ®
    • Google Scholar
    More AR articles citing this reference

    • Dynamic Mode Decomposition and Its Variants

      Peter J. SchmidDepartment of Mathematics, Imperial College London, London, United Kingdom; email: [email protected]
      Annual Review of Fluid Mechanics Vol. 54: 225 - 254
      • ...This near rank-deficiency can be exploited effectively by randomized methods (Halko et al. 2011)....
    • From Bypass Transition to Flow Control and Data-Driven Turbulence Modeling: An Input–Output Viewpoint

      Mihailo R. JovanovićMing Hsieh Department of Electrical and Computer Engineering, University of Southern California, Los Angeles, California 90089, USA; email: [email protected]
      Annual Review of Fluid Mechanics Vol. 53: 311 - 345
      • ...Dominant singular values of the state-transition and frequency response operators can be computed iteratively (Schmid 2007) or via randomized techniques (Halko et al. 2011)....

  • 53.
    Hamdaoui M, Chaskalovic J, Doncieux S, Sagaut P. 2010. Using multiobjective evolutionary algorithms and data-mining methods to optimize ornithopters’ kinematics. J. Aircraft 47:1504–16
    • Crossref
    • Web of Science ®
    • Google Scholar
    Article Location
  • 54.
    Hansen N, Müller SD, Koumoutsakos P. 2003. Reducing the time complexity of the derandomized evolution strategy with covariance matrix adaptation (CMA-ES). Evol. Comput. 11:1–18
    • Crossref
    • Medline
    • Web of Science ®
    • Google Scholar
    Article Location
  • 55.
    Hansen N, Niederberger AS, Guzzella L, Koumoutsakos P. 2009. A method for handling uncertainty in evolutionary optimization with an application to feedback control of combustion. IEEE Trans. Evol. Comput. 13:180–97
    • Crossref
    • Web of Science ®
    • Google Scholar
    Article Location
  • 56.
    Hastie T, Tibshirani R, Friedman J, Hastie T, Friedman J, Tibshirani R. 2009. The Elements of Statistical Learning, Vol. 2. New York: Springer
    • Crossref
    • Google Scholar
    Article Location
    More AR articles citing this reference

    • Small Steps with Big Data: Using Machine Learning in Energy and Environmental Economics

      Matthew C. Harding1 and Carlos Lamarche21Department of Economics and Department of Statistics, University of California, Irvine, California 92697; email: [email protected]2Department of Economics, Gatton College of Business and Economics, University of Kentucky, Lexington, Kentucky 40506
      Annual Review of Resource Economics Vol. 13: 469 - 488
      • ...cross-validation) for the selection of the tuning parameter λ (see, e.g., Hastie et al. 2009)....
    • Emerging Applications of Machine Learning in Food Safety

      Xiangyu Deng,1 Shuhao Cao,2 and Abigail L. Horn31Center for Food Safety, University of Georgia, Griffin, Georgia 30223, USA; email: [email protected]2Department of Mathematics and Statistics, Washington University, St. Louis, Missouri 63105, USA; email: [email protected]3Department of Preventive Medicine, University of Southern California, Los Angeles, California 90032, USA; email: [email protected]
      Annual Review of Food Science and Technology Vol. 12: 513 - 538
      • ...although at the cost of losing unbiasedness (Hastie et al. 2009) and parameter-prediction mechanics that are comprehensible by humans....
    • Artificial Intelligence, Predictive Policing, and Risk Assessment for Law Enforcement

      Richard A. BerkDepartments of Statistics and Criminology, University of Pennsylvania, Philadelphia, Pennsylvania 19104, USA; email: [email protected]
      Annual Review of Criminology Vol. 4: 209 - 237
      • ...and Xj is the jth predictor from a set of P predictors (Hastie et al. 2009)....
      • ...Figure 4 A simple deep-learning neural network with one hidden layer composed of M nodes [adapted from Berk (2020b) and Hastie et al. (2009)]....
      • ...ML methods are not constrained in this manner (Hastie et al. 2009)....
      • ...Another approach displays how each predictor is related to the response holding all other predictors constant without the use of the standard covariance adjustments employed by linear regression (Hastie et al. 2009)....
    • Machine Learning in Epidemiology and Health Outcomes Research

      Timothy L. Wiemken1 and Robert R. Kelley21Center for Health Outcomes Research, Saint Louis University, Saint Louis, Missouri 63104, USA; email: [email protected]2Department of Computer Science, Bellarmine University, Louisville, Kentucky 40205, USA; email: [email protected]
      Annual Review of Public Health Vol. 41: 21 - 36
      • ...and Cox proportional hazards often used in epidemiology and biostatistics (28)....
    • A Survey of Tuning Parameter Selection for High-Dimensional Regression

      Yunan Wu and Lan WangSchool of Statistics, University of Minnesota, Minneapolis, Minnesota 55455, USA; email: [email protected]
      Annual Review of Statistics and Its Application Vol. 7: 209 - 226
      • ...Statistical methods for analyzing high-dimensional data have been the focus of an enormous amount of research in the past decade or so; readers are directed to the books of Hastie et al. (2009), Bühlmann & Van de Geer (2011), Hastie et al. (2015), ...
    • Big Data in Industrial-Organizational Psychology and Human Resource Management: Forward Progress for Organizational Research and Practice

      Frederick L. Oswald,1 Tara S. Behrend,2 Dan J. Putka,3 and Evan Sinar41Department of Psychological Sciences, Rice University, Houston, Texas 77005, USA; email: [email protected]2Department of Organizational Sciences and Communication, George Washington University, Washington, DC 20052, USA3Human Resources Research Organization, Alexandria, Virginia 22314, USA4BetterUp, Pittsburgh, Pennsylvania 15243, USA
      Annual Review of Organizational Psychology and Organizational Behavior Vol. 7: 505 - 533
      • ...principal components regression; Hastie et al. 2009) or the composition and number of those components (e.g., ...
      • ...which provides a useful blend of conceptual and theoretical description with practical implementation; and Hastie et al. (2009), ...
    • Machine Learning Methods That Economists Should Know About

      Susan Athey1,2,3 and Guido W. Imbens1,2,3,41Graduate School of Business, Stanford University, Stanford, California 94305, USA; email: [email protected], [email protected]2Stanford Institute for Economic Policy Research, Stanford University, Stanford, California 94305, USA3National Bureau of Economic Research, Cambridge, Massachusetts 02138, USA4Department of Economics, Stanford University, Stanford, California 94305, USA
      Annual Review of Economics Vol. 11: 685 - 725
      • ...and many textbooks discuss ML methods alongside more traditional statistical methods (e.g., Hastie et al. 2009, Efron & Hastie 2016)....
      • ...including the work of Efron & Hastie (2016); Hastie et al. (2009), ...
      • ...Hastie et al. (2009, 2015) discuss what they call the sparsity principle: ...
    • Machine Learning for Sociology

      Mario Molina and Filiz GaripDepartment of Sociology, Cornell University, Ithaca, New York 14853, USA; email: [email protected], [email protected]
      Annual Review of Sociology Vol. 45: 27 - 45
      • ...This error comprises two components: bias and variance (Hastie et al. 2009)....
      • ...estimating its generalization (prediction) error on new data (Hastie et al. 2009)....
      • ...and a quarter each for validation and testing (Hastie et al. 2009)....
      • ...Boosting involves giving more weight to misclassified observations over repeated estimation (Hastie et al. 2009)....
      • ...4Hastie et al. (2009, table 10.1) compare different methods on several criteria (e.g., ...
    • Precision Medicine

      Michael R. Kosorok1 and Eric B. Laber2 1Department of Biostatistics and Department of Statistics and Operations Research, University of North Carolina at Chapel Hill, Chapel Hill, North Carolina 27599, USA; email: [email protected]2Department of Statistics, North Carolina State University, Raleigh, North Carolina 27695, USA; email: [email protected]
      Annual Review of Statistics and Its Application Vol. 6: 263 - 286
      • ... uses this framework with support vector machines (see, e.g., chapter 12 of Hastie et al. 2009)...
    • Forecasting Methods in Finance

      Allan TimmermannRady School of Management, University of California, San Diego, La Jolla, California 92093, USA; email: [email protected]
      Annual Review of Financial Economics Vol. 10: 449 - 479
      • ...have been developed in recent years (for an excellent introduction, see Hastie, Tibshirani & Friedman 2009)....
    • Big Data Approaches for Modeling Response and Resistance to Cancer Drugs

      Peng Jiang,1 William R. Sellers,2 and X. Shirley Liu11Dana–Farber Cancer Institute and Harvard T.H. Chan School of Public Health, Boston, Massachusetts 02215, USA; email: [email protected]2Broad Institute of MIT and Harvard, Cambridge, Massachusetts 02142, USA; email: [email protected]
      Annual Review of Biomedical Data Science Vol. 1: 1 - 27
      • ...These penalties help find coefficients of the optimal solution in high-dimensional settings while preventing the regression procedure from overfitting the training data (86)....
    • Machine Learning Approaches for Clinical Psychology and Psychiatry

      Dominic B. Dwyer, Peter Falkai, and Nikolaos KoutsoulerisDepartment of Psychiatry and Psychotherapy, Section for Neurodiagnostic Applications, Ludwig-Maximilian University, Munich 80638, Germany; email: [email protected], [email protected], [email protected]
      Annual Review of Clinical Psychology Vol. 14: 91 - 118
      • ...there are several well-established and highly regarded comprehensive methodological guides to machine learning that include formal statistical nomenclature (Bishop 2006, Hastie et al. 2009, James et al. 2015)....
      • ...and the long computational time (Hastie et al. 2009, Varoquaux et al. 2017)....
      • ...This process is then repeated for a prespecified number of k folds and results in more stable estimates of generalizability outside the sample because the training groups are more variable and there are more individuals in the left-out test sets (Hastie et al. 2009)....
      • ...and while authors recommend 5- or 10-fold CV (Breiman & Spector 1992) or statistical criteria (Hastie et al. 2009, James et al. 2015), ...
      • ...Aspects of the SVM algorithm have developed over time to include the ability to characterize nonlinear hyperplanes by using a data transformation implemented by a kernel function (Hastie et al. 2009, James et al. 2015)....
    • Treatment Selection in Depression

      Zachary D. Cohen and Robert J. DeRubeisDepartment of Psychology, University of Pennsylvania, Philadelphia, Pennsylvania 19104, USA; email: [email protected]
      Annual Review of Clinical Psychology Vol. 14: 209 - 236
      • ...researchers must weigh the increased flexibility and predictive power of such approaches against the interpretability (Hastie et al. 2009, James et al. 2013)...
      • ...data from the to-be-predicted patient cannot be included in the course of development of the algorithm (Hastie et al. 2009)....
    • Forecasting in Economics and Finance

      Graham Elliott1 and Allan Timmermann1,21Department of Economics, University of California, San Diego, La Jolla, California 92093; email: [email protected]2Center for Research in Econometric Analysis of Time Series, Aarhus University, DK-8210 Aarhus, Denmark
      Annual Review of Economics Vol. 8: 81 - 110
      • ...these methods will receive further consideration in future work. Hastie et al. (2009) provide a terrific introduction to statistical learning methods, ...
    • League Tables for Hospital Comparisons

      Sharon-Lise T. Normand,1 Arlene S. Ash,2 Stephen E. Fienberg,3 Thérèse A. Stukel,4 Jessica Utts,5 and Thomas A. Louis61Department of Health Care Policy, Harvard Medical School, and Department of Biostatistics, Harvard T.H. Chan School of Public Health, Boston, Massachusetts 02115; email: [email protected]2Department of Quantitative Health Sciences, University of Massachusetts Medical School, Worcester, Massachusetts 01605; email: [email protected]3Department of Statistics, Carnegie Mellon University, Pittsburgh, Pennsylvania 15213; email: [email protected]4Institute for Clinical Evaluative Sciences, Toronto, Ontario M4N 3M5, Canada, and the Institute of Health Policy, Management & Evaluation, University of Toronto, Toronto, Ontario M5T 3M6, Canada, and Dartmouth Institute for Health Policy and Clinical Practice, Hanover, New Hampshire 03766; email: [email protected]5Department of Statistics, University of California, Irvine, California 92697; email: [email protected]6Department of Biostatistics, Johns Hopkins Bloomberg School of Public Health, Baltimore, Maryland 21205; email: [email protected]
      Annual Review of Statistics and Its Application Vol. 3: 21 - 50
      • ...and boosting (Berk 2008, Breiman 2001, Hastie et al. 2009, McCaffrey et al. 2004), ...
    • Far Right Parties in Europe

      Matt GolderDepartment of Political Science, Pennsylvania State University, University Park, Pennsylvania 16802; email: [email protected]
      Annual Review of Political Science Vol. 19: 477 - 497
      • ...They have yet to exploit recent developments in data mining, particularly with respect to cluster analysis (Hastie et al. 2009)....
    • Computerized Adaptive Diagnosis and Testing of Mental Health Disorders

      Robert D. Gibbons,1 David J. Weiss,2 Ellen Frank,3 and David Kupfer31Center for Health Statistics and Departments of Medicine and Public Health Sciences, University of Chicago, Chicago, Illinois 60612; email: [email protected]2Department of Psychology, University of Minnesota, Minneapolis, Minnesota 554553Department of Psychiatry and Western Psychiatric Institute and Clinic, University of Pittsburgh, Pittsburgh, Pennsylvania 15213
      Annual Review of Clinical Psychology Vol. 12: 83 - 104
      • ...have received considerable attention in statistics and machine learning (Brieman 1996, Hastie et al. 2009)....
      • ...decision trees have frequently suffered from poor performance (Hastie et al. 2009) because algorithms used to build trees from data can exhibit sensitivity to small changes in the data sets that are provided....
      • ...Random forests require minimal human intervention and have historically exhibited good performance across a wide range of domains (Brieman 2001, Hastie et al. 2009)....
    • Modular Brain Networks

      Olaf Sporns1,2 and Richard F. Betzel11Department of Psychological and Brain Sciences, Indiana University, Bloomington, Indiana 47405; email: [email protected]du2Indiana University Network Science Institute, Indiana University, Bloomington, Indiana 47405
      Annual Review of Psychology Vol. 67: 613 - 640
      • ...Distance-based modules.One of the simplest methods for detecting modules in complex networks is to extend distance-based clustering techniques to be compatible with network data (Hastie et al. 2009)....
    • Analytics of Insurance Markets

      Edward W. FreesWisconsin School of Business, University of Wisconsin–Madison, Madison, Wisconsin 53706; email: [email protected]
      Annual Review of Financial Economics Vol. 7: 253 - 277
      • ...predictive analytics means advanced data-mining tools such as described in The Elements of Statistical Learning: Data Mining, Inference, and Prediction (Hastie, Tibshirani & Friedman 2009)....
    • Empirical Comparative Law

      Holger SpamannHarvard Law School, Harvard University, Cambridge, Massachusetts 02138; email: [email protected]
      Annual Review of Law and Social Science Vol. 11: 131 - 153
      • ...25This is the “bet on sparsity” principle coined by Hastie et al. (2009, ...
    • Valid Post-Selection and Post-Regularization Inference: An Elementary, General Approach

      Victor Chernozhukov,1 Christian Hansen,2 and Martin Spindler3 1Department of Economics, Massachusetts Institute of Technology, Cambridge, Massachusetts 02142; email: [email protected] 2University of Chicago Booth School of Business, Chicago, Illinois 60637; email: [email protected] 3Munich Center for the Economics of Aging, 80799 Munich, Germany; email: [email protected]
      Annual Review of Economics Vol. 7: 649 - 688
      • ...Many other interesting procedures beyond those mentioned in this review have been developed for estimating high-dimensional models (see, e.g., Hastie et al. 2009 for a textbook review)....
    • Statistical Foundations for Model-Based Adjustments

      Sander Greenland1 and Neil Pearce2,31Department of Epidemiology and Department of Statistics, University of California, Los Angeles, California 90095-1772; email: [email protected]2Departments of Medical Statistics and Non-Communicable Disease Epidemiology, Faculty of Epidemiology and Population Health, London School of Hygiene and Tropical Medicine, London WC1E 7HT, United Kingdom; email: [email protected]3Centre for Public Health Research, Massey University, Wellington 6140, New Zealand
      Annual Review of Public Health Vol. 36: 89 - 108
      • ...We focus entirely on methods for observational studies of causal effects, there being some excellent texts on purely predictive modeling (79, 80, 94, 122, 132)....
      • ...but we must leave many difficult issues about model specification and diagnostics to more detailed discussions (3, 60, 61, 67, 79, 80, 94, 122, 132, 140)....
      • ... [even though all modern software allows use of better criteria (60, 79, 80, 132)]....
      • ...the resulting model often yields much poorer predictions than can be obtained with modern techniques (79, 80, 132)....
      • ...Defects can be corrected by using advanced resampling and cross-validation methods (80, 122, 132, 140), ...
      • ...such as the Akaike information criterion (AIC) or Bayesian information criterion (BIC) (79, 80, 132); nonetheless, ...
      • ...Again, there are ways to account for this problem (79, 80, 132, 140), but we know of none that are easy to implement with popular software....
      • ...such as cross-validation or bootstrapping, to a level that rivals much more sophisticated algorithms (80)....
      • ... and can be applied similarly to outcome modeling (80) as long as exposure is forced into the model....
    • High-Dimensional Statistics with a View Toward Applications in Biology

      Peter Bühlmann, Markus Kalisch, and Lukas MeierSeminar for Statistics, ETH Zürich, CH-8092 Zürich, Switzerland; email: [email protected], [email protected], [email protected]
      Annual Review of Statistics and Its Application Vol. 1: 255 - 278
      • ...Assessing the accuracy of prediction is relatively straightforward using the tool of cross-validation (cf. Hastie et al. 2009)....
    • Sensors and Decoding for Intracortical Brain Computer Interfaces

      Mark L. Homer,1 Arto V. Nurmikko,2 John P. Donoghue,4,3 and Leigh R. Hochberg4,2,51Biomedical Engineering,2School of Engineering,3Department of Neuroscience, Brown University, Providence, Rhode Island 02912; email: [email protected]4Center for Neurorestoration and Neurotechnology, Veterans Affairs Medical Center, Providence, Rhode Island 029085Neurology, Massachusetts General Hospital, Harvard Medical School, Boston, Massachusetts 02114
      Annual Review of Biomedical Engineering Vol. 15: 383 - 405
      • ...akin to forward stepwise regression, can find promising, though typically not optimal, subsets (83)....
    • Sparse High-Dimensional Models in Economics

      Jianqing Fan,1,2 Jinchi Lv,3 and Lei Qi1,21Bendheim Center for Finance and 2Department of Operations Research and Financial Engineering, Princeton University, Princeton, New Jersey 08544; email: [email protected], [email protected]3Information and Operations Management Department, Marshall School of Business, University of Southern California, Los Angeles, California 90089; email: [email protected]
      Annual Review of Economics Vol. 3: 291 - 317
      • ...such as text and document classification and computer vision (see Hastie et al. 2009 for more examples)....
    • Species Distribution Models: Ecological Explanation and Prediction Across Space and Time

      Jane Elith1 and John R. Leathwick21School of Botany, The University of Melbourne, Victoria 3010, Australia; email: [email protected]2National Institute of Water and Atmospheric Research, Hamilton, New Zealand; email: [email protected]
      Annual Review of Ecology, Evolution, and Systematics Vol. 40: 677 - 697
      • ...both within the model-fitting process, and for model evaluation (Hastie et al. 2009)....
      • ...Other model averaging techniques from computer science use a range of approaches to concurrently develop a set of models that together predict well (Hastie et al. 2009)....
      • ...The different information criteria provide a range of trade-offs between model complexity and predictive performance and can be used within cross-validation to select a model (Hastie et al. 2009)....

  • 57.
    Hinton GE, Sejnowski TJ. 1986. Learning and relearning in Boltzmann machines. In Parallel Distributed Processing: Explorations in the Microstructure of Cognition, Vol. 1, ed. DE Rumelhart, J McClelland, pp. 282–317. Cambridge, MA: MIT Press
    • Google Scholar
    Article Location
  • 58.
    Hochreiter S, Schmidhuber J. 1997. Long short-term memory. Neural Comput. 9:1735–80 Regularization of recurrent neural networks and major contributor to the success of Google Translate.
    • Crossref
    • Medline
    • Web of Science ®
    • Google Scholar
    Article Locations:
    • Article Location
    • Article Location
    • Article Location
    More AR articles citing this reference

    • Methods for Robot Behavior Adaptation for Cognitive Neurorehabilitation

      Alyssa Kubota and Laurel D. RiekDepartment of Computer Science and Engineering, University of California, San Diego, La Jolla, California, USA; email: [email protected], [email protected]
      Annual Review of Control, Robotics, and Autonomous Systems Vol. 5: 109 - 135
      • ...are able to learn long-term dependencies throughout the data stream by implementing an input gate and output gate to protect stored memory from irrelevant input (123)....
    • Current Advances in Neural Networks

      Víctor Gallego1 and David Ríos Insua1,21Institute of Mathematical Sciences (ICMAT-CSIC), 28049 Madrid, Spain; email: [email protected]2Department of Statistics and Operations Research, Universidad Complutense de Madrid, 28003 Madrid, Spain
      Annual Review of Statistics and Its Application Vol. 9: 197 - 222
      • ...as in natural language processing (NLP) (e.g., Hochreiter & Schmidhuber 1997, Chung et al. 2014)....
      • ...gating architectures to improve the stability have been proposed and successfully applied in real-life tasks, such as long short-term memory (LSTM) (Hochreiter & Schmidhuber 1997)...
    • Neurocomputational Models of Language Processing

      John T. Hale,1 Luca Campanelli,1,2,3 Jixing Li,4 Shohini Bhattasali,5 Christophe Pallier,6 and Jonathan R. Brennan71Department of Linguistics, University of Georgia, Athens, Georgia, USA; email: [email protected]2Haskins Laboratories, New Haven, Connecticut, USA3Department of Communicative Disorders, University of Alabama, Tuscaloosa, Alabama, USA4Neuroscience of Language Lab, NYU Abu Dhabi, Abu Dhabi, United Arab Emirates5Institute for Advanced Computer Studies, Department of Linguistics, and Program in Neuroscience and Cognitive Science, University of Maryland, College Park, Maryland, USA6Cognitive Neuroimaging Unit, INSERM U992, Gif-sur-Yvette, France7Department of Linguistics, University of Michigan, Ann Arbor, Michigan, USA
      Annual Review of Linguistics Vol. 8: 427 - 446
      • ...Recurrent nets based on gated connections such as the Long Short-Term Memory of Hochreiter & Schmidhuber (1997)...
    • Balancing Flexibility and Interference in Working Memory

      Timothy J. BuschmanPrinceton Neuroscience Institute and Department of Psychology, Princeton University, Princeton, New Jersey 08544, USA; email: [email protected]
      Annual Review of Vision Science Vol. 7: 367 - 388
      • ...working memory is tightly controlled to maximize the use of this limited resource (Gazzaley & Nobre 2012, Hochreiter & Schmidhuber 1997, Vogel et al. 2005)....
    • Quantitative Molecular Positron Emission Tomography Imaging Using Advanced Deep Learning Techniques

      Habib Zaidi1,2,3,4 and Issam El Naqa5,6,71Division of Nuclear Medicine and Molecular Imaging, Geneva University Hospital, 1211 Geneva, Switzerland; email: [email protected]2Geneva Neuroscience Centre, University of Geneva, 1205 Geneva, Switzerland3Department of Nuclear Medicine and Molecular Imaging, University of Groningen, 9700 RB Groningen, Netherlands4Department of Nuclear Medicine, University of Southern Denmark, DK-5000 Odense, Denmark5Department of Machine Learning, Moffitt Cancer Center, Tampa, Florida 33612, USA6Department of Radiation Oncology, University of Michigan, Ann Arbor, Michigan 48109, USA7Department of Oncology, McGill University, Montreal, Quebec H3A 1G5, Canada
      Annual Review of Biomedical Engineering Vol. 23: 249 - 276
      • ...which can learn to store and forget internal memory information and create long-term dependencies through a data sequence (47)....
    • Extension of Plant Phenotypes by the Foliar Microbiome

      Christine V. Hawkes,1 Rasmus Kjøller,2 Jos M. Raaijmakers,3 Leise Riber,4 Svend Christensen,4 Simon Rasmussen,5 Jan H. Christensen,4 Anders Bjorholm Dahl,6 Jesper Cairo Westergaard,4 Mads Nielsen,7 Gina Brown-Guedira,8 and Lars Hestbjerg Hansen41Department of Plant and Microbial Biology, North Carolina State University, Raleigh, North Carolina 27695, USA; email: [email protected]2Department of Biology, University of Copenhagen, 2100 Copenhagen Ø, Denmark; email: [email protected]3Department of Microbial Ecology, Netherlands Institute of Ecology, 6708 PB Wageningen, The Netherlands; email: [email protected]4Department of Plant and Environmental Sciences, University of Copenhagen, 1871 Frederiksberg C, Denmark; email: [email protected], [email protected], [email protected], [email protected], [email protected]5Novo Nordisk Foundation Center for Protein Research, Faculty of Health and Medical Sciences, University of Copenhagen, 2200 Copenhagen N, Denmark; email: [email protected]6Department of Applied Mathematics and Computer Science, Technical University of Denmark, 2800 Lyngby, Denmark; email: [email protected]7Department of Computer Science, University of Copenhagen, 2100 Copenhagen Ø, Denmark; email: [email protected]8Plant Science Research Unit, USDA Agricultural Research Service and Department of Crop and Soil Sciences, North Carolina State University, Raleigh, North Carolina 27695, USA; email: [email protected]
      Annual Review of Plant Biology Vol. 72: 823 - 846
      • ...temporal aspects can be modelled using time-resolved networks such as long short-term memory networks (50) to predict cluster or trajectory memberships, ...
    • Emerging Applications of Machine Learning in Food Safety

      Xiangyu Deng,1 Shuhao Cao,2 and Abigail L. Horn31Center for Food Safety, University of Georgia, Griffin, Georgia 30223, USA; email: [email protected]2Department of Mathematics and Statistics, Washington University, St. Louis, Missouri 63105, USA; email: [email protected]3Department of Preventive Medicine, University of Southern California, Los Angeles, California 90032, USA; email: [email protected]
      Annual Review of Food Science and Technology Vol. 12: 513 - 538
      • ... and speech (Hochreiter & Schmidhuber 1997); classification of objects such as cats, ...
    • Syntactic Structure from Deep Learning

      Tal Linzen1 and Marco Baroni2,3,41Department of Linguistics and Center for Data Science, New York University, New York, NY 10003, USA; email: [email protected]2Facebook AI Research, Paris 75002, France; email: [email protected]3Catalan Institute for Research and Advanced Studies, Barcelona 08010, Spain4Departament de Traducció i Ciències del Llenguatge, Universitat Pompeu Fabra, Barcelona 08018, Spain
      Annual Review of Linguistics Vol. 7: 195 - 212
      • ...such as long short-term memory networks (LSTMs) (Hochreiter & Schmidhuber 1997)...
    • Identifying Regulatory Elements via Deep Learning

      Mira Barshai,1, Eitamar Tripto,2, and Yaron Orenstein11School of Electrical and Computer Engineering, Ben-Gurion University of the Negev, Beer-Sheva 8410501, Israel; email: [email protected]2Department of Biomedical Engineering, Ben-Gurion University of the Negev, Beer-Sheva 8410501, Israel
      Annual Review of Biomedical Data Science Vol. 3: 315 - 338
      • ...Hochreiter & Schmidhuber (56) substituted the hidden units in RNNs with long short-term memory (LSTM) units....
    • Robots That Use Language

      Stefanie Tellex,1 Nakul Gopalan,2 Hadas Kress-Gazit,3 and Cynthia Matuszek41Department of Computer Science, Brown University, Providence, Rhode Island 02912, USA; email: [email protected]2School of Interactive Computing, Georgia Institute of Technology, Atlanta, Georgia 30332, USA; email: [email protected]3Sibley School of Mechanical and Aerospace Engineering, Cornell University, Ithaca, New York 14853, USA; email: [email protected]4Department of Computer Science and Electrical Engineering, University of Maryland, Baltimore County, Baltimore, Maryland 21250, USA; email: [email protected]
      Annual Review of Control, Robotics, and Autonomous Systems Vol. 3: 25 - 55
      • ...Modern approaches use word vectors to capture or learn structure, such as long short-term memory units (LSTMs) (42)...
    • Representation Learning: A Statistical Perspective

      Jianwen Xie,1 Ruiqi Gao,2 Erik Nijkamp,2 Song-Chun Zhu,2 and Ying Nian Wu21Hikvision Research Institute, Santa Clara, California 95054, USA2Department of Statistics, University of California, Los Angeles, California 90095, USA; email: [email protected]
      Annual Review of Statistics and Its Application Vol. 7: 303 - 335
      • ...The other class consists of recurrent neural networks (Hochreiter & Schmidhuber 1997), ...
    • Deep Learning: The Good, the Bad, and the Ugly

      Thomas SerreDepartment of Cognitive Linguistic and Psychological Sciences, Carney Institute for Brain Science, Brown University, Providence, Rhode Island 02818, USA; email: [email protected]
      Annual Review of Vision Science Vol. 5: 399 - 426
      • ...Popular recurrent neural networks such as Long Short Term Memory (LSTM) units (Hochreiter & Schmidhuber 1997)...
    • Deep Learning and Its Application to LHC Physics

      Dan Guest,1 Kyle Cranmer,2 and Daniel Whiteson11Department of Physics and Astronomy, University of California, Irvine, California 92697, USA2Physics Department, New York University, New York, NY 10003, USA
      Annual Review of Nuclear and Particle Science Vol. 68: 161 - 181
      • ...These alleviate the exploding and vanishing gradient problem at the expense of a more complicated recurrent unit; examples are long-short-term-memory units (LSTMs) (25)...
    • Deep Learning in Biomedical Data Science

      Pierre BaldiDepartment of Computer Science, Institute for Genomics and Bioinformatics, and Center for Machine Learning and Intelligent Systems, University of California, Irvine, California 92697, USA; email: [email protected]
      Annual Review of Biomedical Data Science Vol. 1: 181 - 205
      • ...A long short-term memory (LSTM) is a kind of recurrent neural network building block capable of learning or storing contextual information over different temporal or spatial length scales (6, 7)....
    • Computational Neuroscience: Mathematical and Statistical Perspectives

      Robert E. Kass,1 Shun-Ichi Amari,2 Kensuke Arai,3 Emery N. Brown,4,5 Casey O. Diekman,6 Markus Diesmann,7,8 Brent Doiron,9 Uri T. Eden,3 Adrienne L. Fairhall,10 Grant M. Fiddyment,3 Tomoki Fukai,2 Sonja Grün,7,8 Matthew T. Harrison,11 Moritz Helias,7,8 Hiroyuki Nakahara,2 Jun-nosuke Teramae,12 Peter J. Thomas,13 Mark Reimers,14 Jordan Rodu,15 Horacio G. Rotstein,16,17 Eric Shea-Brown,10 Hideaki Shimazaki,18,19 Shigeru Shinomoto,19 Byron M. Yu,20 and Mark A. Kramer31Department of Statistics, Machine Learning Department, and Center for the Neural Basis of Cognition, Carnegie Mellon University, Pittsburgh, Pennsylvania 15213, USA; email: [email protected]2Mathematical Neuroscience Laboratory, RIKEN Brain Science Institute, Wako, Saitama Prefecture 351-0198, Japan3Department of Mathematics and Statistics, Boston University, Boston, Massachusetts 02215, USA4Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, Massachusetts 02139, USA5Department of Anesthesia, Harvard Medical School, Boston, Massachusetts 02115, USA6Department of Mathematical Sciences, New Jersey Institute of Technology, Newark, New Jersey 07102, USA7Institute of Neuroscience and Medicine, Jülich Research Centre, 52428 Jülich, Germany8Department of Theoretical Systems Neurobiology, Institute of Biology, RWTH Aachen University, 52062 Aachen, Germany9Department of Mathematics, University of Pittsburgh, Pittsburgh, Pennsylvania 15260, USA10Department of Physiology and Biophysics, University of Washington, Seattle, Washington 98105, USA11Division of Applied Mathematics, Brown University, Providence, Rhode Island 02912, USA12Department of Integrated Theoretical Neuroscience, Osaka University, Suita, Osaka Prefecture 565-0871, Japan13Department of Mathematics, Applied Mathematics, and Statistics, Case Western Reserve University, Cleveland, Ohio 44106, USA14Department of Neuroscience, Michigan State University, East Lansing, Michigan 48824, USA15Department of Statistics, University of Virginia, Charlottesville, Virginia 22904, USA16Federated Department of Biological Sciences, Rutgers University/New Jersey Institute of Technology, Newark, New Jersey 07102, USA17Institute for Brain and Neuroscience Research, New Jersey Institute of Technology, Newark, New Jersey 07102, USA18Honda Research Institute Japan, Wako, Saitama Prefecture 351-0188, Japan19Department of Physics, Kyoto University, Kyoto, Kyoto Prefecture 606-8502, Japan20Department of Electrical and Computer Engineering and Department of Biomedical Engineering, Carnegie Mellon University, Pittsburgh, Pennsylvania 15213, USA
      Annual Review of Statistics and Its Application Vol. 5: 183 - 214
      • ...LSTM (Hochreiter & Schmidhuber 1997) enables neural networks to take as input sequential data of arbitrary length and learn long-term dependencies by incorporating a memory module where information can be added or forgotten according to functions of the current input and state of the system....
    • Deep Neural Networks: A New Framework for Modeling Biological Vision and Brain Information Processing

      Nikolaus KriegeskorteMedical Research Council Cognition and Brain Sciences Unit, University of Cambridge, Cambridge CB2 7EF, United Kingdom; email: [email protected]
      Annual Review of Vision Science Vol. 1: 417 - 446
      • ...One solution to this problem is offered by the long short-term memory (LSTM) architecture (Hochreiter & Schmidhuber 1997), ...

  • 59.
    Holland JH. 1975. Adaptation in Natural and Artificial Systems: An Introductory Analysis with Applications to Biology, Control, and Artificial Intelligence. Ann Arbor: Univ. Mich. Press
    • Google Scholar
    Article Location
  • 60.
    Hopfield JJ. 1982. Neural networks and physical systems with emergent collective computational abilities. PNAS 79:2554–58
    • Crossref
    • Medline
    • Web of Science ®
    • ADS
    • Google Scholar
    Article Location
    More AR articles citing this reference

    • Neural Algorithms and Circuits for Motor Planning

      Hidehiko K. Inagaki1, Susu Chen2, Kayvon Daie2,3, Arseny Finkelstein2,4, Lorenzo Fontolan2, Sandro Romani2, and Karel Svoboda2,31Max Planck Florida Institute for Neuroscience, Jupiter, Florida, USA; email: [email protected]2Janelia Research Campus, Howard Hughes Medical Institute, Ashburn, Virginia, USA3Allen Institute for Neural Dynamics, Seattle, Washington, USA; email: [email protected]4Department of Physiology and Pharmacology, Sackler Faculty of Medicine, Tel Aviv University, Tel Aviv-Yafo, Israel
      Annual Review of Neuroscience Vol. 45: 249 - 271
      • ...Dynamical systems theory provides a mechanistic framework for explaining population neuronal activity (Amari 1972, Wilson & Cowan 1972, Hopfield 1982, Sompolinsky et al. 1988, Kleinfeld et al. 1990, Amit & Brunel 1997, Laurent 2002, Brody et al. 2003b, Stopfer et al. 2003, Sussillo & Abbott 2009, Druckmann & Chklovskii 2012, Gallego et al. 2017, Vyas et al. 2020, Ebitz & Hayden 2021)....
      • ...attractors have been hypothesized as a mechanism of memory (Amari 1972, Hopfield 1982, Chaudhuri & Fiete 2016)....
      • ...Point attractors can store discrete memories (Amari 1972, Amit & Brunel 1997, Hopfield 1982), ...
      • ...What happens in tasks with multiple choices? Attractor networks can accommodate a large number of fixed points (Hopfield 1982, Amit et al. 1985)....
    • Neurophysiology of Remembering

      György Buzsáki,1,2 Sam McKenzie,3 and Lila Davachi4,51Neuroscience Institute and Department of Neurology, NYU Grossman School of Medicine, New York University, New York, NY 10016, USA; email: [email protected]2Center for Neural Science, New York University, New York, NY 10003, USA3Department of Neurosciences, University of New Mexico, Albuquerque, New Mexico 87131, USA4Department of Psychology, Columbia University, New York, NY 10027, USA5Center for Biomedical Imaging and Neuromodulation, Nathan S. Kline Institute for Psychiatric Research, Orangeburg, New York 10962, USA
      Annual Review of Psychology Vol. 73: 187 - 215
      • ...or schema (Hopfield 1982, McKenzie et al. 2014, Samsonovich & McNaughton 1997)—and to provide a stable and robust balance against competing needs, ...
    • Interneuron Types as Attractors and Controllers

      Gord Fishell1,2,3 and Adam Kepecs4,51Department of Neurobiology, Blavatnik Institute, Harvard Medical School, Boston, Massachusetts 02115, USA; email: [email protected]2Stanley Center for Psychiatric Research, Broad Institute, Cambridge, Massachusetts 02142, USA3Center for Genomics and Systems Biology, New York University Abu Dhabi, Abu Dhabi, United Arab Emirates4Cold Spring Harbor Laboratory, Cold Spring Harbor, New York 11724, USA5Department of Neuroscience, Washington University in St. Louis, St. Louis, Missouri 63130, USA; email: [email protected]
      Annual Review of Neuroscience Vol. 43: 1 - 30
      • ...An interesting side point for a neuroscientist is that these networks have a very similar mathematical formalism to Hopfield neural networks in which attractors represent neural activity patterns, each encoding a memory (Hopfield 1982)....
    • Synaptic Plasticity Forms and Functions

      Jeffrey C. Magee and Christine GrienbergerDepartment of Neuroscience and Howard Hughes Medical Institute, Baylor College of Medicine, Houston, Texas 77030, USA; email: [email protected]
      Annual Review of Neuroscience Vol. 43: 95 - 117
      • ...forming a simple association that allows even fragments of the associated input pattern to evoke the correct output activity (Andersen 1972, Kohonen 1972, Hopfield 1982)....
    • Nonequilibrium Thermodynamics in Cell Biology: Extending Equilibrium Formalism to Cover Living Systems

      Xiaona Fang1 and Jin Wang1,21Department of Chemistry, Stony Brook University, Stony Brook, New York 11794, USA; email: [email protected]2Department of Physics and Astronomy, Stony Brook University, Stony Brook, New York 11794, USA
      Annual Review of Biophysics Vol. 49: 227 - 246
      • ...one can study the dynamics of the neural networks that are crucial for cognitive functions such as decision making (1, 45, 53, 91)....
    • Machine-Learning Quantum States in the NISQ Era

      Giacomo Torlai1 and Roger G. Melko2,31Center for Computational Quantum Physics, Flatiron Institute, New York, NY 10010, USA; email: [email protected]2Department of Physics and Astronomy, University of Waterloo, Waterloo, Ontario N2L 3G1, Canada; email: [email protected]3Perimeter Institute for Theoretical Physics, Waterloo, Ontario N2L 2Y5, Canada
      Annual Review of Condensed Matter Physics Vol. 11: 325 - 344
      • ...established by the works of condensed matter physicists William Little (15, 16) and John Hopfield (17)....
      • ...The Hopfield network, introduced in 1982 as a model for associative memories (17), ...
      • ...there exists a learning rule to modify the interactions in such a way that these states become local minima in the energy landscape (17)....
    • Big Data and Artificial Intelligence Modeling for Drug Discovery

      Hao ZhuDepartment of Chemistry and Center for Computational and Integrative Biology, Rutgers University, Camden, New Jersey 08102, USA; email: [email protected]
      Annual Review of Pharmacology and Toxicology Vol. 60: 573 - 589
      • ...which was designed as a computational tool in the 1980s (92), ...
    • Computational Models of Memory Search

      Michael J. KahanaDepartment of Psychology, University of Pennsylvania, Philadelphia, Pennsylvania 19104, USA; email: [email protected]
      Annual Review of Psychology Vol. 71: 107 - 138
      • ...For example, in a Hopfield network (Hopfield 1982), we would apply the same storage equation to binary vectors....
      • ...The symmetric nature of these linear models also applies to their nonlinear variants (Hopfield 1982)....
    • Antagonistic Phenomena in Network Dynamics

      Adilson E. Motter1 and Marc Timme2,31Department of Physics and Astronomy and Northwestern Institute on Complex Systems, Northwestern University, Evanston, Illinois 60208, USA; email: [email protected]2Chair of Network Dynamics, Institute for Theoretical Physics and Center for Advancing Electronics (cfaed), Technical University of Dresden, 01062 Dresden, Germany; email: [email protected]3Network Dynamics, Max Planck Institute for Dynamics and Self-Organization, 37077 Göttingen, Germany
      Annual Review of Condensed Matter Physics Vol. 9: 463 - 484
      • ...ranging from neural and biochemical circuits to self-organized communication networks (86...
    • Computational Neuroscience: Mathematical and Statistical Perspectives

      Robert E. Kass,1 Shun-Ichi Amari,2 Kensuke Arai,3 Emery N. Brown,4,5 Casey O. Diekman,6 Markus Diesmann,7,8 Brent Doiron,9 Uri T. Eden,3 Adrienne L. Fairhall,10 Grant M. Fiddyment,3 Tomoki Fukai,2 Sonja Grün,7,8 Matthew T. Harrison,11 Moritz Helias,7,8 Hiroyuki Nakahara,2 Jun-nosuke Teramae,12 Peter J. Thomas,13 Mark Reimers,14 Jordan Rodu,15 Horacio G. Rotstein,16,17 Eric Shea-Brown,10 Hideaki Shimazaki,18,19 Shigeru Shinomoto,19 Byron M. Yu,20 and Mark A. Kramer31Department of Statistics, Machine Learning Department, and Center for the Neural Basis of Cognition, Carnegie Mellon University, Pittsburgh, Pennsylvania 15213, USA; email: [email protected]2Mathematical Neuroscience Laboratory, RIKEN Brain Science Institute, Wako, Saitama Prefecture 351-0198, Japan3Department of Mathematics and Statistics, Boston University, Boston, Massachusetts 02215, USA4Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, Massachusetts 02139, USA5Department of Anesthesia, Harvard Medical School, Boston, Massachusetts 02115, USA6Department of Mathematical Sciences, New Jersey Institute of Technology, Newark, New Jersey 07102, USA7Institute of Neuroscience and Medicine, Jülich Research Centre, 52428 Jülich, Germany8Department of Theoretical Systems Neurobiology, Institute of Biology, RWTH Aachen University, 52062 Aachen, Germany9Department of Mathematics, University of Pittsburgh, Pittsburgh, Pennsylvania 15260, USA10Department of Physiology and Biophysics, University of Washington, Seattle, Washington 98105, USA11Division of Applied Mathematics, Brown University, Providence, Rhode Island 02912, USA12Department of Integrated Theoretical Neuroscience, Osaka University, Suita, Osaka Prefecture 565-0871, Japan13Department of Mathematics, Applied Mathematics, and Statistics, Case Western Reserve University, Cleveland, Ohio 44106, USA14Department of Neuroscience, Michigan State University, East Lansing, Michigan 48824, USA15Department of Statistics, University of Virginia, Charlottesville, Virginia 22904, USA16Federated Department of Biological Sciences, Rutgers University/New Jersey Institute of Technology, Newark, New Jersey 07102, USA17Institute for Brain and Neuroscience Research, New Jersey Institute of Technology, Newark, New Jersey 07102, USA18Honda Research Institute Japan, Wako, Saitama Prefecture 351-0188, Japan19Department of Physics, Kyoto University, Kyoto, Kyoto Prefecture 606-8502, Japan20Department of Electrical and Computer Engineering and Department of Biomedical Engineering, Carnegie Mellon University, Pittsburgh, Pennsylvania 15213, USA
      Annual Review of Statistics and Its Application Vol. 5: 183 - 214
      • ...Hopfield (1982) applied statistical physics tools to introduce an energy function and showed that a simple update rule would decrease the energy so that the network would settle to a pattern-matching attractor state....
    • Replay Comes of Age

      David J. FosterDepartment of Psychology and Helen Wills Neuroscience Institute, University of California, Berkeley, California 94720; email: [email protected]
      Annual Review of Neuroscience Vol. 40: 581 - 602
      • ...the problem of creating a simple chain between place cells becomes analogous to that of storing sequences of distributed patterns in a recurrent neural network such as a Hopfield network (Hopfield 1982)....
    • Inhibitory Plasticity: Balance, Control, and Codependence

      Guillaume Hennequin,1, Everton J. Agnes,2, and Tim P. Vogels21Computational and Biological Learning Lab, Department of Engineering, University of Cambridge, Cambridge CB2 3EJ, United Kingdom2Centre for Neural Circuits and Behaviour, Department of Physiology, Anatomy and Genetics, University of Oxford, Oxford OX1 3SR, United Kingdom; email: [email protected]
      Annual Review of Neuroscience Vol. 40: 557 - 579
      • ...Hopfield (1982) mathematically described such memories as patterns of active and silent neurons (Figure 5a) and constructed simplified neuronal networks in which such patterns become fixed points of the collective dynamics....
    • Mechanisms of Persistent Activity in Cortical Circuits: Possible Neural Substrates for Working Memory

      Joel Zylberberg1,2,3 and Ben W. Strowbridge4,51Department of Physiology and Biophysics, Center for Neuroscience, and Computational Bioscience Program, University of Colorado School of Medicine, Aurora, Colorado 800452Department of Applied Mathematics, University of Colorado, Boulder, Colorado 803093Learning in Machines and Brains Program, Canadian Institute for Advanced Research, Toronto, Ontario M5G 1Z8, Canada4Department of Neurosciences, Case Western Reserve University School of Medicine, Cleveland, Ohio 44106; email: [email protected]5Department of Physiology and Biophysics, Case Western Reserve University School of Medicine, Cleveland, Ohio 44106
      Annual Review of Neuroscience Vol. 40: 603 - 627
      • ...does the persistence rely on the connectivity within the neural circuit to extend the representation beyond the timescale of activity in isolated neurons (Compte et al. 2000, Druckmann & Chklovskii 2012, Goldman 2009, Hopfield 1982, Lim & Goldman 2013, Seung 1996), ...
      • ...The classic way to achieve persistent representations is for the neural population to have specific patterns of activity that are reinforced by the synaptic connectivity in the network (Hopfield 1982); we refer to these as privileged activity patterns....
      • ...been studied with spiking (or binary) neuron models (see the sidebar titled Different Classes of Computational Neuron Models for descriptions of the different types of neuron models) (Amit & Brunel 1997, Hopfield 1982)....
      • ...This type of neuron model is used in the discrete attractor network of Hopfield (1982)....
    • Imaging and Optically Manipulating Neuronal Ensembles

      Luis Carrillo-Reid,1,2 Weijian Yang,1,2 Jae-eun Kang Miller,1,2 Darcy S. Peterka,1,2 and Rafael Yuste1,2,31NeuroTechnology Center, Columbia University, New York, NY 100272Department of Biological Sciences, Columbia University, New York, NY 100273Department of Neuroscience, Columbia University, New York, NY 10027; email: [email protected]
      Annual Review of Biophysics Vol. 46: 271 - 293
      • ...These ideas were later developed into formal models of networks of interconnected neurons where coactive neurons form functional states called attractors (33), ...
    • Information Processing in Living Systems

      Gašper Tkačik1 and William Bialek21Institute of Science and Technology Austria, A-3400 Klosterneuburg, Austria; email: [email protected]2Joseph Henry Laboratories of Physics and Lewis-Sigler Institute for Integrative Genomics, Princeton University, Princeton, New Jersey 08544; email: [email protected]
      Annual Review of Condensed Matter Physics Vol. 7: 89 - 117
      • ...Perhaps the best developed set of ideas in this direction concerns the dynamics of neural networks (39...
      • ...the redundant strategy with information-maximizing connectivity yields a network with strong attractor states to which different stimuli map uniquely, recovering Hopfield-like associative memory (39) from an optimization principle....
    • Whatever Happened to Solid State Physics?

      John J.HopfieldDepartment of Molecular Biology, Princeton University, Princeton, New Jersey 08544; email: [email protected] Institute for Advanced Study, Princeton, New Jersey 08540

      Annual Review of Condensed Matter Physics Vol. 5: 1 - 13
      • ...It was to provide an entryway to working on neuroscience for many physicists and is the most cited paper I have ever written (11)....
    • Electrical Compartmentalization in Dendritic Spines

      Rafael YusteHHMI, Departments of Biological Sciences and Neuroscience, and Kavli Institute for Brain Science, Columbia University, New York, NY 10027; email: [email protected]
      Annual Review of Neuroscience Vol. 36: 429 - 449
      • ... and perform relatively sophisticated computations such as associative memory (Hopfield 1982), ...
    • The Impact of Adult Neurogenesis on Olfactory Bulb Circuits and Computations

      Gabriel Lepousez,1,2, Matthew T. Valley,1,2, and Pierre-Marie Lledo1,21Laboratory of Perception and Memory, Institut Pasteur, F-75015 Paris, France; email: [email protected]2Unité de Recherche Associée (URA2182), Centre National de la Recherche Scientifique (CNRS), F-75015 Paris, France
      Annual Review of Physiology Vol. 75: 339 - 363
      • ...the activity-dependent integration of newborn GABAergic neurons into the M/T cell network may provide new inhibitory connections that stabilize specific attractor states in M/T cell population coding (146, 147, 148)....
    • Attractor Dynamics of Spatially Correlated Neural Activity in the Limbic System

      James J. Knierim1 and Kechen Zhang21Krieger Mind/Brain Institute and Department of Neuroscience, Johns Hopkins University, Baltimore, Maryland 21218; email: [email protected]2Department of Biomedical Engineering, Johns Hopkins University, Baltimore, Maryland 21205; email: [email protected]
      Annual Review of Neuroscience Vol. 35: 267 - 285
      • ...Neural networks with symmetric connections (equal reciprocal connections between neurons) always allow an energy function (Lyapunov function) (Hopfield 1982, Cohen & Grossberg 1983, Hopfield 1984). (b) Schematic examples of various attractors in neural network models, ...
      • ...which allows multiple point attractors, each corresponding to a stored memory pattern (Hopfield 1982)....
    • Compressed Sensing, Sparsity, and Dimensionality in Neuronal Information Processing and Data Analysis

      Surya Ganguli1 and Haim Sompolinsky2,31Department of Applied Physics, Stanford University, Stanford, California 94305; email: [email protected]2Edmond and Lily Safra Center for Brain Sciences, Interdisciplinary Center for Neural Computation, Hebrew University, Jerusalem 91904, Israel; email: [email protected]3Center for Brain Science, Harvard University, Cambridge, Massachusetts 02138
      Annual Review of Neuroscience Vol. 35: 485 - 508
      • ...The influential idea of attractor dynamics (Hopfield 1982) suggests how single stimuli can be stored as stable patterns of activity, ...
    • Emotion, Cognition, and Mental State Representation in Amygdala and Prefrontal Cortex

      C. Daniel Salzman1,2,3,4,5,6 and Stefano Fusi11Department of Neuroscience, Columbia University, New York, NY 10032;2Department of Psychiatry, Columbia University, New York, NY 10032;3W.M. Keck Center on Brain Plasticity and Cognition, Columbia University, New York, NY 10032;4Kavli Institute for Brain Sciences, Columbia University, New York, NY 10032;5Mahoney Center for Brain and Behavior, Columbia University, New York, NY 10032; email: [email protected], [email protected]6New York State Psychiatric Institute, New York, NY 10032
      Annual Review of Neuroscience Vol. 33: 173 - 202
      • ...Attractor networks have been proposed as models for associative and working memory (Amit 1989, Hopfield 1982), ...
    • Place Cells, Grid Cells, and the Brain's Spatial Representation System

      Edvard I. Moser,1 Emilio Kropff,1,2 and May-Britt Moser11Kavli Institute for Systems Neuroscience and Centre for the Biology of Memory, Norwegian University of Science and Technology, 7489 Trondheim, Norway2Cognitive Neuroscience Sector, International School for Advanced Studies, Trieste, Italy; email: [email protected]
      Annual Review of Neuroscience Vol. 31: 69 - 89
      • ...a stable firing state sustained by recurrent connections with robust performance in the presence of noise (Hopfield 1982, Amit 1989, Rolls & Treves 1998)....
      • ...How could memories be stored in the place cell system? On the basis of the extensive intrinsic connectivity and modifiability of the CA3 network, theoretical work has indicated attractor dynamics (Hopfield 1982, Amit 1989)...
    • The Mind and Brain of Short-Term Memory

      John Jonides, Richard L. Lewis, Derek Evan Nee, Cindy A. Lustig, Marc G. Berman, and Katherine Sledge MooreDepartment of Psychology, University of Michigan, Ann Arbor, Michigan 48109; email: [email protected]
      Annual Review of Psychology Vol. 59: 193 - 224
      • ...self-sustaining patterns observed in certain classes of recurrent networks (Hopfield 1982, Rougier et al. 2005, Polk et al. 2002)....
      • ...A plausible neural mechanism for the recovery of this activity pattern at retrieval is the emergent pattern-completion property of attractor networks (Hopfield 1982)....
    • Physics of Proteins

      Jayanth R. Banavar1 and Amos Maritan21Department of Physics, The Pennsylvania State University, University Park, Pennsylvania 16802; email: [email protected]2INFM and Dipartimento di Fisica ‘G. Galilei’, Università di Padova, 35131 Padova, Italy; email: [email protected]
      Annual Review of Biophysics and Biomolecular Structure Vol. 36: 261 - 280
      • ...Spin glasses are thus characterized by stability (each ground state is a local minimum) and diversity (2, 47) (there are many local minima), ...
    • INFERENCE AND COMPUTATION WITH POPULATION CODES

      Alexandre Pouget,1 Peter Dayan,2 and Richard S. Zemel31Department of Brain and Cognitive Sciences, Meliora Hall, University of Rochester, Rochester, New York, 14627; email: [email protected] 2Gatsby Computational Neuroscience Unit, Alexandra House, 17 Queen Square, London WC1N 3AR, United Kingdom; email: [email protected] 3Department of Computer Science, University of Toronto, Toronto, Ontario M5S 1A4 Canada; email: [email protected]
      Annual Review of Neuroscience Vol. 26: 381 - 410
      • ...Both of these problems can be addressed by utilizing recurrent connections within the population to make it behave like an autoassociative memory (Hopfield 1982)....
    • From Factors to Actors: Computational Sociology and Agent-Based Modeling

      Michael W. Macy and Robert WillerDepartment of Sociology, Cornell University, Ithaca, New York 84153; e-mail: [email protected] [email protected]
      Annual Review of Sociology Vol. 28: 143 - 166
      • ...Structural differentiation based on positive and negative influence has been studied using attractor neural networks, a cognitive modeling technique developed by Hopfield (1982)...
    • Odor Encoding as an Active, Dynamical Process: Experiments, Computation, and Theory

      Gilles Laurent,1 Mark Stopfer,1 Rainer W Friedrich,1 Misha I Rabinovich,2 Alexander Volkovskii,2 and Henry DI Abarbanel,2,3Division of Biology 139-74, 1California Institute of Technology, Pasadena, California 91125; e-mail: [email protected] [email protected] [email protected] Institute for Nonlinear Science, 2University of California San Diego, La Jolla, California 92093-0402; e-mail: [email protected] [email protected] [email protected] 3Department of Physics and Marine Physical Laboratory, Scripps Institution of Oceanography, University of California San Diego, La Jolla, California 92093
      Annual Review of Neuroscience Vol. 24: 263 - 297
      • ...The most common nonlinear dynamical models of early olfactory processing (representation and recognition) lead to the idea of “coding with attractors” (“Hopfield nets” [Cohen & Grossberg 1982, Hopfield 1982])....
    • An Integrative Theory of Prefrontal Cortex Function

      Earl K. MillerCenter for Learning and Memory, RIKEN-MIT Neuroscience Research Center and Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, Massachusetts 02139; e-mail: [email protected] Jonathan D. CohenCenter for the Study of Brain, Mind, and Behavior and Department of Psychology, Princeton University, Princeton, New Jersey 08544; e-mail: [email protected]
      Annual Review of Neuroscience Vol. 24: 167 - 202
      • ...propose that the recirculation of activity through closed (or “recurrent”) loops of interconnected neurons, or attractor networks (Hopfield 1982), ...
    • Morphological Changes in Dendritic Spines Associated with Long-Term Synaptic Plasticity

      Rafael Yuste1 and Tobias Bonhoeffer21Department of Biological Sciences, Columbia University, New York, NY 10027; e-mail: [email protected] 2Max Planck Institut für Neurobiologie, München-Martinsried, 82152 Germany; e-mail: [email protected]
      Annual Review of Neuroscience Vol. 24: 1071 - 1089
      • ...LTP furthermore nicely relates to neural network theories of brain function because it implements a local learning rule, an essential element for associative neuronal networks (Hopfield 1982), ...
    • Neuroengineering Models of Brain Disease

      Leif H. FinkelDepartment of Bioengineering, University of Pennsylvania, Philadelphia, Pennsylvania 19104; e-mail: [email protected]
      Annual Review of Biomedical Engineering Vol. 2: 577 - 606
      • ...Hopfield (19) derived a formula for setting the synaptic weights such that any desired activity pattern can be made an attractor state of the network, ...
    • Evolutionary Computation: An Overview

      Melanie Mitchell* and Charles E. Taylor***Santa Fe Institute, 1399 Hyde Park Road, Santa Fe, New Mexico 87501; e-mail: [email protected] ,**Department of Organismic Biology, Ecology, and Evolution, University of California, Los Angeles, California 90095; e-mail: [email protected]
      Annual Review of Ecology and Systematics Vol. 30: 593 - 616
      • ...the number of memories reliably learned and stored is approximately 0.15 times the number of nodes in a completely connected system (32)....
    • THEORY OF PROTEIN FOLDING: The Energy Landscape Perspective

      José Nelson OnuchicDepartment of Physics, University of California at San Diego, La Jolla, California 92093-0319 Zaida Luthey-Schulten and Peter G. WolynesSchool of Chemical Sciences, University of Illinois, Urbana, Illinois 61801
      Annual Review of Physical Chemistry Vol. 48: 545 - 600
      • ...Indeed the energy landscape theory of structure prediction has many mathematical parallels with the theories of learning used in connectionist artificial intelligence (204, 205)....
    • THE SIGNIFICANCE OF NEURAL ENSEMBLE CODES DURING BEHAVIOR AND COGNITION

      Sam A. Deadwyler and Robert E. HampsonDepartment of Physiology and Pharmacology Neuroscience Program, Bowman Gray School of Medicine, Wake Forest University, Winston-Salem, North Carolina 27157
      Annual Review of Neuroscience Vol. 20: 217 - 244
      • ...providing a parallel spread of excitatory and inhibitory signals to cells whose firing may appear quite divergent from the original input sources (Hopfield 1982, Rumelhart & McClelland 1986, Hopfield & Tank 1986)....
      • ...A second major influence was the explosive development in neural computation and the modeling of neural networks made possible by the meteoric changes in accessibility and power of laboratory computers (Hopfield 1982, Churchland & Sejnowski 1989)....
    • PSYCHOBIOLOGICAL MODELS OF HIPPOCAMPAL FUNCTION IN LEARNING AND MEMORY

      Mark A. Gluck and Catherine E. MyersCenter for Molecular and Behavioral Neuroscience, Rutgers University, 197 University Avenue, Newark, New Jersey; e-mail: [email protected]
      Annual Review of Psychology Vol. 48: 481 - 514
      • ...A network of n nodes is able to store only about 0.15n random patterns before they begin to interfere with one another (Hopfield 1982)....

  • 61.
    Hornik K, Stinchcombe M, White H. 1989. Multilayer feedforward networks are universal approximators. Neural Netw. 2:359–66
    • Crossref
    • Web of Science ®
    • Google Scholar
    Article Locations:
    • Article Location
    • Article Location
    More AR articles citing this reference

    • Recent Challenges in Actuarial Science

      Paul Embrechts and Mario V. WüthrichRiskLab, Department of Mathematics, ETH Zurich, Zurich, Switzerland, CH-8092; email: [email protected], [email protected]
      Annual Review of Statistics and Its Application Vol. 9: 119 - 140
      • ...Based on well-known universality theorems (see, e.g., Cybenko 1989, Hornik et al. 1989, Isenbeck & Rüschendorf 1992), ...
    • Small Steps with Big Data: Using Machine Learning in Energy and Environmental Economics

      Matthew C. Harding1 and Carlos Lamarche21Department of Economics and Department of Statistics, University of California, Irvine, California 92697; email: [email protected]2Department of Economics, Gatton College of Business and Economics, University of Kentucky, Lexington, Kentucky 40506
      Annual Review of Resource Economics Vol. 13: 469 - 488
      • ...Early theoretical results showing how well networks approximate unknown functions were established by Hornik et al. (1989)...
    • Face Recognition by Humans and Machines: Three Fundamental Advances from Deep Learning

      Alice J. O'Toole1 and Carlos D. Castillo21School of Behavioral and Brain Sciences, The University of Texas at Dallas, Richardson, Texas 75080, USA; email: [email protected]2Department of Electrical and Computer Engineering, Johns Hopkins University, Baltimore, Maryland 21218, USA; email: [email protected]
      Annual Review of Vision Science Vol. 7: 543 - 570
      • ...the universal approximation theorem (Hornik et al. 1989) ensures that both types of networks can approximate any complex continuous function relating the input (visual image) to the output (face identity)....
    • Data Science in Chemical Engineering: Applications to Molecular Science

      Chowdhury Ashraf,1 Nisarg Joshi,1 David A.C. Beck,1,2 and Jim Pfaendtner11Department of Chemical Engineering, University of Washington, Seattle, Washington 98195, USA; email: [email protected], [email protected]2eScience Institute, University of Washington, Seattle, Washington 98195, USA
      Annual Review of Chemical and Biomolecular Engineering Vol. 12: 15 - 37
      • ...much of the recent focus in the field is focused on developing QSAR models free from the mathematical limitations of linear modeling as facilitated by NN, which are effective as universal function approximators (90)....
    • Statistical Mechanics of Deep Learning

      Yasaman Bahri,1 Jonathan Kadmon,2 Jeffrey Pennington,1 Sam S. Schoenholz,1 Jascha Sohl-Dickstein,1 and Surya Ganguli1,21Google Brain, Google Inc., Mountain View, California 94043, USA2Department of Applied Physics, Stanford University, Stanford, California 94035, USA; email: [email protected]
      Annual Review of Condensed Matter Physics Vol. 11: 501 - 528
      • ...Seminal results (19, 20) demonstrate that shallow networks, with only one hidden layer of neurons, ...
      • ...Importantly, the early results on function approximation in References 19 and 20...
    • Machine Learning Methods That Economists Should Know About

      Susan Athey1,2,3 and Guido W. Imbens1,2,3,41Graduate School of Business, Stanford University, Stanford, California 94305, USA; email: [email protected], [email protected]2Stanford Institute for Economic Policy Research, Stanford University, Stanford, California 94305, USA3National Bureau of Economic Research, Cambridge, Massachusetts 02138, USA4Department of Economics, Stanford University, Stanford, California 94305, USA
      Annual Review of Economics Vol. 11: 685 - 725
      • ...which were the focus of a small econometrics literature in the 1990s (Hornik et al. 1989, White 1992) but more recently have become a very prominent part of the literature on ML in various subtle reincarnations....
      • ...Neural networks were studied in the econometric literature in the 1990s but did not catch on at the time (see Hornik et al. 1989, White 1992)....
    • Deep Learning and Its Application to LHC Physics

      Dan Guest,1 Kyle Cranmer,2 and Daniel Whiteson11Department of Physics and Astronomy, University of California, Irvine, California 92697, USA2Physics Department, New York University, New York, NY 10003, USA
      Annual Review of Nuclear and Particle Science Vol. 68: 161 - 181
      • ...because of the theoretical analysis that demonstrated that any function can be approximated by a shallow network (14)....
    • Forecasting in Economics and Finance

      Graham Elliott1 and Allan Timmermann1,21Department of Economics, University of California, San Diego, La Jolla, California 92093; email: [email protected]2Center for Research in Econometric Analysis of Time Series, Aarhus University, DK-8210 Aarhus, Denmark
      Annual Review of Economics Vol. 8: 81 - 110
      • ...these models can approximate very general sets of functions arbitrarily well (see Hornik et al. 1989)....
    • COMPLETE FUNCTIONAL CHARACTERIZATION OF SENSORY NEURONS BY SYSTEM IDENTIFICATION

      Michael C.-K. Wu,1 Stephen V. David,2 and Jack L. Gallant3,41Biophysics Graduate Group, 3Department of Psychology, 4Program in Neuroscience,University of California, Berkeley, California 94720; email: [email protected]2Institute for Systems Research, University of Maryland, College Park, Maryland 20742
      Annual Review of Neuroscience Vol. 29: 477 - 505
      • ...they are universal approximators (Hammer & Gersmann 2003, Hornik et al. 1989)], ...

  • 62.
    Hou W, Darakananda D, Eldredge J. 2019. Machine learning based detection of flow disturbances using surface pressure measurements. Paper presented at AIAA Atmospheric Flight Mechanics Conference 2019, San Diego, Calif., AIAA Pap. 2019-1148
    • Crossref
    • Google Scholar
    Article Location
  • 63.
    Jambunathan K, Hartle S, Ashforth-Frost S, Fontama V. 1996. Evaluating convective heat transfer coefficients using neural networks. Int. J. Heat Mass Transf. 39:2329–32
    • Crossref
    • Web of Science ®
    • Google Scholar
    Article Location
  • 64.
    Kaiser E, Noack BR, Cordier L, Spohn A, Segond M, 2014. Cluster-based reduced-order modelling of a mixing layer. J. Fluid Mech. 754:365–414
    • Crossref
    • Web of Science ®
    • ADS
    • Google Scholar
    Article Location
    More AR articles citing this reference

    • Dynamic Mode Decomposition and Its Variants

      Peter J. SchmidDepartment of Mathematics, Imperial College London, London, United Kingdom; email: [email protected]
      Annual Review of Fluid Mechanics Vol. 54: 225 - 254
      • ...the Frobenius–Perron approach has been exploited in a data-driven cluster-based analysis of shear flow instabilities (Kaiser et al. 2014)...

  • 65.
    Kern S, Müller SD, Hansen N, Büche D, Ocenasek J, Koumoutsakos P. 2004. Learning probability distributions in continuous evolutionary algorithms—a comparative review. Nat. Comput. 3:77–112
    • Crossref
    • Google Scholar
    Article Locations:
    • Article Location
    • Article Location
  • 66.
    Kim B, Azevedo VC, Thuerey N, Kim T, Gross M, Solenthaler B. 2018. Deep fluids: a generative network for parameterized fluid simulations. arXiv:1806.02071 [cs.LG]
    • ADS
    • Google Scholar
    Article Location
  • 67.
    Kim HJ, Jordan MI, Sastry S, Ng AY. 2004. Autonomous helicopter flight via reinforcement learning. In Advances in Neural Information Processing Systems 17, ed. B Scholkopf, Y Vis, JC Platt, pp. 799–806. Cambridge, MA: MIT Press
    • Google Scholar
    Article Location
  • 68.
    Knaak M, Rothlubbers C, Orglmeister R. 1997. A Hopfield neural network for flow field computation based on particle image velocimetry/particle tracking velocimetry image sequences. IEEE Int. Conf. Neural Netw. 1:48–52
    • Google Scholar
    Article Location
  • 69.
    Kohonen T. 1995. Self-Organizing Maps. Berlin: Springer-Verlag
    • Crossref
    • Google Scholar
    Article Location
    More AR articles citing this reference

    • PERCEPTUAL LEARNING

      Robert L. GoldstonePsychology Building, Indiana University, Bloomington, Indiana 47405; e-mail: [email protected]
      Annual Review of Psychology Vol. 49: 585 - 612
      • .... Kohonen (1995) has developed a framework for describing neural networks that develop topological structures with learning....

  • 70.
    Kolmogorov A. 1941. The local structure of turbulence in incompressible viscous fluid for very large Reynolds number. Cr. Acad. Sci. USSR 30:301–5
    • Google Scholar
    Article Location
  • 71.
    Koza JR. 1992. Genetic Programming: On the Programming of Computers by Means of Natural Selection. Boston: MIT Press
    • Google Scholar
    Article Locations:
    • Article Location
    • Article Location
  • 72.
    Krizhevsky A, Sutskever I, Hinton GE. 2012. Imagenet classification with deep convolutional neural networks. In Advances in Neural Information Processing Systems 25, ed. F Pereira, CJC Burges, L Bottou, KQ Weinberger, pp. 1097–105. Red Hook, NY: Curran Assoc.
    • Google Scholar
    Article Locations:
    • Article Location
    • Article Location
    More AR articles citing this reference

    • Advances in Inference and Representation for Simultaneous Localization and Mapping

      David M. Rosen,1 Kevin J. Doherty,2 Antonio Terán Espinoza,2 and John J. Leonard21Laboratory for Information and Decision Systems, Massachusetts Institute of Technology, Cambridge, Massachusetts 02139, USA; email: [email protected]2Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology, Cambridge, Massachusetts 02139, USA; email: [email protected], [email protected], [email protected]
      Annual Review of Control, Robotics, and Autonomous Systems Vol. 4: 215 - 242
      • ...the expectation of SLAM methods to perform certain scene understanding tasks coupled with classical geometric estimation coincided with the recent successes (and accessibility) of machine learning methods for perception tasks [especially in computer vision (76, 77)], ...
    • Machine Learning for Structural Materials

      Taylor D. Sparks,1 Steven K. Kauwe,1 Marcus E. Parry,1 Aria Mansouri Tehrani,2 and Jakoah Brgoch21Department of Materials Science and Engineering, University of Utah, Salt Lake City, Utah 84112, USA; email: [email protected]2Department of Chemistry, University of Houston, Texas 77204, USA
      Annual Review of Materials Research Vol. 50: 27 - 48
      • ...they passed the micrographs into a pretrained caffe model (51) and used the output from the fully connected sixth layer as features for learning (illustrated in Figure 3)....
    • Robots That Use Language

      Stefanie Tellex,1 Nakul Gopalan,2 Hadas Kress-Gazit,3 and Cynthia Matuszek41Department of Computer Science, Brown University, Providence, Rhode Island 02912, USA; email: [email protected]2School of Interactive Computing, Georgia Institute of Technology, Atlanta, Georgia 30332, USA; email: [email protected]3Sibley School of Mechanical and Aerospace Engineering, Cornell University, Ithaca, New York 14853, USA; email: [email protected]4Department of Computer Science and Electrical Engineering, University of Maryland, Baltimore County, Baltimore, Maryland 21250, USA; email: [email protected]
      Annual Review of Control, Robotics, and Autonomous Systems Vol. 3: 25 - 55
      • ...this work used classifiers to recognize an object class within an image (107)....
    • Representation Learning: A Statistical Perspective

      Jianwen Xie,1 Ruiqi Gao,2 Erik Nijkamp,2 Song-Chun Zhu,2 and Ying Nian Wu21Hikvision Research Institute, Santa Clara, California 95054, USA2Department of Statistics, University of California, Los Angeles, California 90095, USA; email: [email protected]
      Annual Review of Statistics and Its Application Vol. 7: 303 - 335
      • ...If we generalize the linear mapping from h to x to a nonlinear mapping parameterized by a deep network (LeCun et al. 1998, Krizhevsky et al. 2012), ...
      • ...such as deep neural networks (LeCun et al. 1998, Krizhevsky et al. 2012), ...
      • ...One consists of convolutional neural networks (LeCun et al. 1998, Krizhevsky et al. 2012), ...
    • Deep Learning: The Good, the Bad, and the Ugly

      Thomas SerreDepartment of Cognitive Linguistic and Psychological Sciences, Carney Institute for Brain Science, Brown University, Providence, Rhode Island 02818, USA; email: [email protected]
      Annual Review of Vision Science Vol. 5: 399 - 426
      • ...That year, a network (Krizhevsky et al. 2012), since dubbed AlexNet (after its lead developer, ...
      • ... and their computer vision relatives (LeCun et al. 1998, Krizhevsky et al. 2012) is the degree to which visual representations are constrained by task demand....
      • ...earlier networks included eight layers or fewer (Fukushima 1980, Krizhevsky et al. 2012, LeCun et al. 1998, Riesenhuber & Poggio 1999)....
      • ...Early computational neuroscience work started with AlexNet (Krizhevsky et al. 2012)...
    • Material Perception

      Roland W. FlemingDepartment of Experimental Psychology, University of Giessen, 35394 Giessen, Germany; email: [email protected]
      Annual Review of Vision Science Vol. 3: 365 - 388
      • ...Although state-of-the art algorithms can now outperform humans at certain object recognition tasks (e.g., Krizhevsky et al. 2012, He et al. 2015), ...

  • 73.
    Kutz JN, Brunton SL, Brunton BW, Proctor JL. 2016. Dynamic Mode Decomposition: Data-Driven Modeling of Complex Systems. Philadelphia: SIAM
    • Crossref
    • Google Scholar
    Article Location
  • 74.
    Labonté G. 1999. A new neural network for particle-tracking velocimetry. Exp. Fluids 26:340–46
    • Crossref
    • Web of Science ®
    • Google Scholar
    Article Location
  • 75.
    Lagaris IE, Likas A, Fotiadis DI. 1998. Artificial neural networks for solving ordinary and partial differential equations. IEEE Trans. Neural Netw. 9:987–1000
    • Crossref
    • Medline
    • Google Scholar
    Article Location
  • 76.
    LeCun Y, Bengio Y, Hinton G. 2015. Deep learning. Nature 521:436–44
    • Crossref
    • Medline
    • Web of Science ®
    • ADS
    • Google Scholar
    Article Location
    More AR articles citing this reference

    • Detection and Monitoring of Viral Infections via Wearable Devices and Biometric Data

      Craig J. Goergen,1 MacKenzie J. Tweardy,2 Steven R. Steinhubl,2,3 Stephan W. Wegerich,2 Karnika Singh,4 Rebecca J. Mieloszyk,5 and Jessilyn Dunn41Weldon School of Biomedical Engineering, Purdue University, West Lafayette, Indiana, USA; email: [email protected]2physIQ Inc., Chicago, Illinois, USA3Scripps Research Translational Institute, La Jolla, California, USA4Department of Biomedical Engineering, Duke University, Durham, North Carolina, USA5Microsoft Research, Redmond, Washington, USA
      Annual Review of Biomedical Engineering Vol. 24: 1 - 27
      • ...and random forests, with a greater current interest in neural networks (78) (Figure 3)....
    • Real-Time Functional MRI in the Treatment of Mental Health Disorders

      Vincent Taschereau-Dumouchel,1,2 Cody A. Cushing,3 and Hakwan Lau41Department of Psychiatry and Addictology, Université de Montréal, Montréal, Québec, Canada; email: [email protected]2Centre de Recherche de l'Institut Universitaire en Santé Mentale de Montréal, Montréal, Québec, Canada3Department of Psychology, University of California, Los Angeles, California, USA4RIKEN Center for Brain Science, Wakoshi, Saitama, Japan
      Annual Review of Clinical Psychology Vol. 18: 125 - 154
      • ...Modern machine learning algorithms have recently achieved remarkable successes in the field of pattern recognition and computer vision (LeCun et al. 2015, Goodfellow et al. 2016)....
      • ...Thanks to the recent explosion of research in artificial intelligence and machine learning (LeCun et al. 2015)...
    • Semantic Structure in Deep Learning

      Ellie PavlickDepartment of Computer Science, Brown University, Providence, Rhode Island, USA; email: [email protected]
      Annual Review of Linguistics Vol. 8: 447 - 471
      • ...traditional DSMs have fallen to the wayside in favor of linguistic representations derived from deep learning (LeCun et al. 2015)....
    • Machine Learning for the Study of Plankton and Marine Snow from Images

      Jean-Olivier Irisson,1 Sakina-Dorothée Ayata,1 Dhugal J. Lindsay,2 Lee Karp-Boss,3 and Lars Stemmann11Laboratoire d'Océanographie de Villefranche, Sorbonne Université, CNRS, F-06230 Villefranche-sur-Mer, France; email: [email protected], [email protected], [email protected]2Advanced Science-Technology Research (ASTER) Program, Institute for Extra-Cutting-Edge Science and Technology Avant-Garde Research (X-STAR), Japan Agency for Marine-Earth Science and Technology, Yokosuka, Kanagawa 237-0021, Japan; email: [email protected]3School of Marine Sciences, University of Maine, Orono, Maine 04469, USA; email: [email protected]
      Annual Review of Marine Science Vol. 14: 277 - 301
      • ...We now tend to separate classic machine learning from deep learning (LeCun et al. 2015)....
    • Small Steps with Big Data: Using Machine Learning in Energy and Environmental Economics

      Matthew C. Harding1 and Carlos Lamarche21Department of Economics and Department of Statistics, University of California, Irvine, California 92697; email: [email protected]2Department of Economics, Gatton College of Business and Economics, University of Kentucky, Lexington, Kentucky 40506
      Annual Review of Resource Economics Vol. 13: 469 - 488
      • ...the literature has concentrated its attention on multilayer networks and generalizations to Deep Learning (LeCun et al. 2015, Farrell et al. 2018)....
    • Face Recognition by Humans and Machines: Three Fundamental Advances from Deep Learning

      Alice J. O'Toole1 and Carlos D. Castillo21School of Behavioral and Brain Sciences, The University of Texas at Dallas, Richardson, Texas 75080, USA; email: [email protected]2Department of Electrical and Computer Engineering, Johns Hopkins University, Baltimore, Maryland 21218, USA; email: [email protected]
      Annual Review of Vision Science Vol. 7: 543 - 570
      • ...DCNNs also emulate computational aspects of the ventral visual system (Fukushima 1988, Krizhevsky et al. 2012, LeCun et al. 2015) and support surprisingly direct, ...
    • Spatial Integration in Normal Face Processing and Its Breakdown in Congenital Prosopagnosia

      Galia Avidan1 and Marlene Behrmann21Department of Psychology and Department of Cognitive and Brain Sciences, Ben-Gurion University of the Negev, Beer-Sheva 8410501, Israel; email: [email protected]2Department of Psychology and Neuroscience Institute, Carnegie Mellon University, Pittsburgh, Pennsylvania 15213, USA
      Annual Review of Vision Science Vol. 7: 301 - 321
      • ...focusing on a certain location of the stimulus enables the processing of information at that location but also the generation of a prediction of the next location to be processed (Ji et al. 2013, LeCun et al. 2015) (for an illustration of a model dCNN as a coarse analogy to ventral pathway function, ...
    • Optical Coherence Tomography and Glaucoma

      Alexi Geevarghese,1 Gadi Wollstein,1,2,3 Hiroshi Ishikawa,1,2 and Joel S. Schuman1,2,3,41Department of Ophthalmology, NYU Langone Health, NYU Grossman School of Medicine, New York, NY 10016, USA; email: [email protected]2Department of Biomedical Engineering, NYU Tandon School of Engineering, Brooklyn, New York 11201, USA3Center for Neural Science, NYU College of Arts and Sciences, New York, NY 10003, USA4Department of Physiology and Neuroscience, NYU Langone Health, NYU Grossman School of Medicine, New York, NY 10016, USA
      Annual Review of Vision Science Vol. 7: 693 - 726
      • ...Importance is applied to each node on the basis of an iterative training process that determines the optimal weights that yield the smallest classification error (LeCun et al. 2015, Zheng et al. 2019). ...
    • Quantitative Molecular Positron Emission Tomography Imaging Using Advanced Deep Learning Techniques

      Habib Zaidi1,2,3,4 and Issam El Naqa5,6,71Division of Nuclear Medicine and Molecular Imaging, Geneva University Hospital, 1211 Geneva, Switzerland; email: [email protected]2Geneva Neuroscience Centre, University of Geneva, 1205 Geneva, Switzerland3Department of Nuclear Medicine and Molecular Imaging, University of Groningen, 9700 RB Groningen, Netherlands4Department of Nuclear Medicine, University of Southern Denmark, DK-5000 Odense, Denmark5Department of Machine Learning, Moffitt Cancer Center, Tampa, Florida 33612, USA6Department of Radiation Oncology, University of Michigan, Ann Arbor, Michigan 48109, USA7Department of Oncology, McGill University, Montreal, Quebec H3A 1G5, Canada
      Annual Review of Biomedical Engineering Vol. 23: 249 - 276
      • ...but recent studies have shown that it is most effective with deep neural network (DNN) methods due to their universal approximation nature (42, 43)....
    • Extension of Plant Phenotypes by the Foliar Microbiome

      Christine V. Hawkes,1 Rasmus Kjøller,2 Jos M. Raaijmakers,3 Leise Riber,4 Svend Christensen,4 Simon Rasmussen,5 Jan H. Christensen,4 Anders Bjorholm Dahl,6 Jesper Cairo Westergaard,4 Mads Nielsen,7 Gina Brown-Guedira,8 and Lars Hestbjerg Hansen41Department of Plant and Microbial Biology, North Carolina State University, Raleigh, North Carolina 27695, USA; email: [email protected]2Department of Biology, University of Copenhagen, 2100 Copenhagen Ø, Denmark; email: [email protected]3Department of Microbial Ecology, Netherlands Institute of Ecology, 6708 PB Wageningen, The Netherlands; email: [email protected]4Department of Plant and Environmental Sciences, University of Copenhagen, 1871 Frederiksberg C, Denmark; email: [email protected], [email protected], [email protected], [email protected], [email protected]5Novo Nordisk Foundation Center for Protein Research, Faculty of Health and Medical Sciences, University of Copenhagen, 2200 Copenhagen N, Denmark; email: [email protected]6Department of Applied Mathematics and Computer Science, Technical University of Denmark, 2800 Lyngby, Denmark; email: [email protected]7Department of Computer Science, University of Copenhagen, 2100 Copenhagen Ø, Denmark; email: [email protected]8Plant Science Research Unit, USDA Agricultural Research Service and Department of Crop and Soil Sciences, North Carolina State University, Raleigh, North Carolina 27695, USA; email: [email protected]
      Annual Review of Plant Biology Vol. 72: 823 - 846
      • ...and neural networks have been applied in a wide array of fields within the last 50 years (for an overview, see 75)....
    • Applications of Machine and Deep Learning in Adaptive Immunity

      Margarita Pertseva,1,2 Beichen Gao,1 Daniel Neumeier,1 Alexander Yermanos,1,3,4 and Sai T. Reddy11Department of Biosystems Science and Engineering, ETH Zurich, 4058 Basel, Switzerland; email: [email protected]2Life Science Zurich Graduate School, ETH Zurich and University of Zurich, 8006 Zurich, Switzerland3Department of Pathology and Immunology, University of Geneva, 1205 Geneva, Switzerland4Department of Biology, Institute of Microbiology and Immunology, ETH Zurich, 8093 Zurich, Switzerland
      Annual Review of Chemical and Biomolecular Engineering Vol. 12: 39 - 62
      • ...one of its major limitations is that the feature extraction step can be tedious and often requires domain-specific knowledge (50)....
      • ...DL uses a class of algorithms that find a relevant set of features required to perform a particular task in a more automated manner (50)....
      • ...the computed result of one layer acts as an input to the next layer, resulting in an increasingly abstract data representation (50)....
      • ...Full coverage of DL models is outside the scope of this review; the interested reader could refer to several additional resources (50, 51, 55)....
    • Syntactic Structure from Deep Learning

      Tal Linzen1 and Marco Baroni2,3,41Department of Linguistics and Center for Data Science, New York University, New York, NY 10003, USA; email: [email protected]2Facebook AI Research, Paris 75002, France; email: [email protected]3Catalan Institute for Research and Advanced Studies, Barcelona 08010, Spain4Departament de Traducció i Ciències del Llenguatge, Universitat Pompeu Fabra, Barcelona 08018, Spain
      Annual Review of Linguistics Vol. 7: 195 - 212
      • ...which have been rebranded as deep learning (LeCun et al. 2015), ...
    • Toward Realizing the Promise of Educational Neuroscience: Improving Experimental Design in Developmental Cognitive Neuroscience Studies

      Usha GoswamiCentre for Neuroscience in Education, University of Cambridge, Cambridge CB2 3EB, United Kingdom; email: [email protected]
      Annual Review of Developmental Psychology Vol. 2: 133 - 155
      • ...clusters of real medical symptoms) and then acquire expertise that can exceed that of human operators (for example, in medical diagnosis; see LeCun et al. 2015)....
      • ...“a giraffe is standing in the forest with trees in the background”; LeCun et al. 2015)....
    • Spatial Metabolomics and Imaging Mass Spectrometry in the Age of Artificial Intelligence

      Theodore Alexandrov1,21Structural and Computational Biology Unit, European Molecular Biology Laboratory, 69117 Heidelberg, Germany; email: [email protected]2Skaggs School of Pharmacy and Pharmaceutical Sciences, University of California, San Diego, La Jolla, California 92093, USA
      Annual Review of Biomedical Data Science Vol. 3: 61 - 87
      • ...a method that has transformed machine learning by outperforming other methods, first for computer vision and later for other problems (49)....
    • Identifying Regulatory Elements via Deep Learning

      Mira Barshai,1, Eitamar Tripto,2, and Yaron Orenstein11School of Electrical and Computer Engineering, Ben-Gurion University of the Negev, Beer-Sheva 8410501, Israel; email: [email protected]2Department of Biomedical Engineering, Ben-Gurion University of the Negev, Beer-Sheva 8410501, Israel
      Annual Review of Biomedical Data Science Vol. 3: 315 - 338
      • ...termed deep learning, has been revolutionizing the data science world (45)....
      • ...Prediction accuracy has been improving tremendously for image and text processing tasks (45)....
    • Computational Approaches for Unraveling the Effects of Variation in the Human Genome and Microbiome

      Chengsheng Zhu,1 Maximilian Miller,1 Zishuo Zeng,1 Yanran Wang,1 Yannick Mahlich,1 Ariel Aptekmann,1 and Yana Bromberg1,21Department of Biochemistry and Microbiology, Rutgers University, New Brunswick, New Jersey 08873, USA; email: [email protected], [email protected]2Department of Genetics, Rutgers University, Piscataway, New Jersey 08854, USA
      Annual Review of Biomedical Data Science Vol. 3: 411 - 432
      • ...a class of machine learning algorithms well suited to processing high-dimensional data, provide new means for this type of analysis (185)....
    • Synaptic Plasticity Forms and Functions

      Jeffrey C. Magee and Christine GrienbergerDepartment of Neuroscience and Howard Hughes Medical Institute, Baylor College of Medicine, Houston, Texas 77030, USA; email: [email protected]
      Annual Review of Neuroscience Vol. 43: 95 - 117
      • ...the learning rules used are essentially the same (Woodrow & Hoff 1960, Rumelhart et al. 1986, LeCun et al. 2015) (Figure 2c–e)....
      • ...While there are relatively straightforward methods to accomplish this in ANNs (Rumelhart et al. 1986, LeCun et al. 2015), ...
    • Opportunities and Challenges for Machine Learning in Materials Science

      Dane Morgan and Ryan JacobsDepartment of Materials Science and Engineering, University of Wisconsin–Madison, Madison, Wisconsin 53706, USA; email: [email protected], [email protected]
      Annual Review of Materials Research Vol. 50: 71 - 103
      • ...The large number of ML models and their many technical details are well covered in many texts and reviews (41–43, 151), ...
    • Machine Learning in Materials Discovery: Confirmed Predictions and Their Underlying Approaches

      James E. Saal,1 Anton O. Oliynyk,2 and Bryce Meredig11Citrine Informatics, Redwood City, California 94063, USA; email: [email protected]2Department of Chemistry and Biochemistry, Manhattan College, Riverdale, New York 10471, USA
      Annual Review of Materials Research Vol. 50: 49 - 69
      • ... and (deep) neural network (NN) (61, 62) algorithms are illustrated conceptually in Figure 5. ...
    • Machine Learning for Molecular Simulation

      Frank Noé,1,2,3 Alexandre Tkatchenko,4 Klaus-Robert Müller,5,6,7 and Cecilia Clementi1,3,81Department of Mathematics and Computer Science, Freie Universität Berlin, 14195 Berlin, Germany; email: [email protected]2Department of Physics, Freie Universität Berlin, 14195 Berlin, Germany3Department of Chemistry and Center for Theoretical Biological Physics, Rice University, Houston, Texas 77005, USA; email: [email protected]4Physics and Materials Science Research Unit, University of Luxembourg, 1511 Luxembourg, Luxembourg; email: [email protected]5Department of Computer Science, Technical University Berlin, 10587 Berlin, Germany; email: [email protected]6Max-Planck-Institut für Informatik, 66123 Saarbrücken, Germany7Department of Brain and Cognitive Engineering, Korea University, Seoul 136-713, South Korea8Department of Physics, Rice University, Houston, Texas 77005, USA
      Annual Review of Physical Chemistry Vol. 71: 361 - 390
      • ...and we refer to the literature for an introduction to statistical learning theory (3, 4) and deep learning (5, 6)....
    • Machine-Learning Quantum States in the NISQ Era

      Giacomo Torlai1 and Roger G. Melko2,31Center for Computational Quantum Physics, Flatiron Institute, New York, NY 10010, USA; email: [email protected]2Department of Physics and Astronomy, University of Waterloo, Waterloo, Ontario N2L 3G1, Canada; email: [email protected]3Perimeter Institute for Theoretical Physics, Waterloo, Ontario N2L 2Y5, Canada
      Annual Review of Condensed Matter Physics Vol. 11: 325 - 344
      • ...Artificial neural networks, the bedrock of modern machine learning and artificial intelligence (8), ...
    • Statistical Mechanics of Deep Learning

      Yasaman Bahri,1 Jonathan Kadmon,2 Jeffrey Pennington,1 Sam S. Schoenholz,1 Jascha Sohl-Dickstein,1 and Surya Ganguli1,21Google Brain, Google Inc., Mountain View, California 94043, USA2Department of Applied Physics, Stanford University, Stanford, California 94035, USA; email: [email protected]
      Annual Review of Condensed Matter Physics Vol. 11: 501 - 528
      • ...Deep neural networks, with multiple hidden layers (1), have achieved remarkable success across many fields, ...
    • Use of Mechanistic Nutrition Models to Identify Sustainable Food Animal Production

      Mark D. Hanigan1 and Veridiana L. Daley1,21Department of Dairy Science, Virginia Polytechnic Institute and State University, Blacksburg, Virginia 24061, USA; email: [email protected], [email protected]2National Animal Nutrition Program (NANP), Department of Animal & Food Sciences, University of Kentucky, Lexington, Kentucky 40546, USA
      Annual Review of Animal Biosciences Vol. 8: 355 - 376
      • ...outputs, and detection of patterns in the input variables (121); thus, ...
    • Distributional Semantics and Linguistic Theory

      Gemma Boleda1,21Department of Translation and Language Sciences, Universitat Pompeu Fabra, Barcelona 08018, Spain; email: [email protected]2Catalan Institution for Research and Advanced Studies (ICREA), Barcelona 08010, Spain
      Annual Review of Linguistics Vol. 6: 213 - 234
      • ...Neural networks are a type of machine learning algorithm, recently revamped as deep learning (LeCun et al. 2015), ...
    • Big Data and Artificial Intelligence Modeling for Drug Discovery

      Hao ZhuDepartment of Chemistry and Center for Computational and Integrative Biology, Rutgers University, Camden, New Jersey 08102, USA; email: [email protected]
      Annual Review of Pharmacology and Toxicology Vol. 60: 573 - 589
      • ...The milestone paper of deep learning was published at almost the same time (103), ...
    • Concepts and Compositionality: In Search of the Brain's Language of Thought

      Steven M. Frankland1 and Joshua D. Greene21Princeton Neuroscience Institute, Princeton University, Princeton, New Jersey 08544, USA; email: [email protected]2Department of Psychology and Center for Brain Science, Harvard University, Cambridge, Massachusetts 02138, USA; email: [email protected]
      Annual Review of Psychology Vol. 71: 273 - 303
      • ...proponents of the LoT hypothesis suspect that human comprehension depends on complex semantic representations with internal representations that are far more structurally constrained. LeCun et al. (2015, ...
    • Data-Driven Approaches to Understanding Visual Neuron Activity

      Daniel A. ButtsDepartment of Biology and Program in Neuroscience and Cognitive Science, University of Maryland, College Park, Maryland 20742, USA; email: [email protected]
      Annual Review of Vision Science Vol. 5: 451 - 477
      • ...the recent machine learning–driven successes in computer vision (Kriegeskorte 2015, Krizhevsky et al. 2012, LeCun et al. 2015, Serre 2019) suggest a new range of possible approaches, ...
      • ...making such methods broadly accessible for using DNNs to fit a larger variety of data (LeCun et al. 2015)....
      • ...refer to solving tasks such as object and face recognition and have played a crucial role in driving the development of DNNs (LeCun et al. 2015, Schmidhuber 2015, Serre 2019)....
    • Machine Learning Methods That Economists Should Know About

      Susan Athey1,2,3 and Guido W. Imbens1,2,3,41Graduate School of Business, Stanford University, Stanford, California 94305, USA; email: [email protected], [email protected]2Stanford Institute for Economic Policy Research, Stanford University, Stanford, California 94305, USA3National Bureau of Economic Research, Cambridge, Massachusetts 02138, USA4Department of Economics, Stanford University, Stanford, California 94305, USA
      Annual Review of Economics Vol. 11: 685 - 725
      • ...This suggests that using a deep model expresses a useful preference over the space of functions the model can learn. (LeCun et al. 2015, ...
    • Computational and Informatic Advances for Reproducible Data Analysis in Neuroimaging

      Russell A. Poldrack,1 Krzysztof J. Gorgolewski,1 and Gaël Varoquaux21Department of Psychology, Stanford University, Stanford, California 94305, USA; email: [email protected]2Parietal Team, Inria and NeuroSpin/CEA (Atomic Energy Commission), 91191 Gif/-sur-Yvette, France
      Annual Review of Biomedical Data Science Vol. 2: 119 - 138
      • ...Machine learning has opened new alleys in extracting information from texts, images, genomes, etc. (42), ...
    • Scientific Discovery Games for Biomedical Research

      Rhiju Das,1 Benjamin Keep,2Peter Washington,3 and Ingmar H. Riedel-Kruse31Department of Biochemistry and Department of Physics, Stanford University, Stanford, California 94305, USA; email: [email protected]2Department of Learning Sciences, Stanford University, Stanford, California 94305, USA3Department of Bioengineering, Stanford University, Stanford, California 94305, USA; email: [email protected]
      Annual Review of Biomedical Data Science Vol. 2: 253 - 279
      • ...it will be important to compare results to more recent algorithmic methods for the same visual tasks, which have been improving at an impressive pace (85)....
    • The Challenge of Big Data and Data Science

      Henry E. BradyDepartment of Political Science and Goldman School of Public Policy, University of California, Berkeley, California 94720, USA; email: [email protected]
      Annual Review of Political Science Vol. 22: 297 - 323
      • ...has succeeded at difficult pattern recognition tasks such as speech and image recognition, natural language processing, and bioinformatics (LeCun et al. 2015)....
    • System Identification: A Machine Learning Perspective

      A. Chiuso and G. PillonettoDepartment of Information Engineering, University of Padova, 35131 Padova, Italy; email: [email protected]
      Annual Review of Control, Robotics, and Autonomous Systems Vol. 2: 281 - 304
      • ...which have recently garnered renewed interest thanks to deep networks’ success in classification and pattern recognition (7)....
    • Deep Learning and Its Application to LHC Physics

      Dan Guest,1 Kyle Cranmer,2 and Daniel Whiteson11Department of Physics and Astronomy, University of California, Irvine, California 92697, USA2Physics Department, New York University, New York, NY 10003, USA
      Annual Review of Nuclear and Particle Science Vol. 68: 161 - 181
      • ...when a convergence of techniques enabled training of very large neural networks that greatly outperformed the previous state of the art (2...
    • Invariant Recognition Shapes Neural Representations of Visual Input

      Andrea Tacchetti, Leyla Isik, and Tomaso A. PoggioCenter for Brains, Minds and Machines, MIT, Cambridge, Massachusetts 02139, USA; email: [email protected], [email protected], [email protected]
      Annual Review of Vision Science Vol. 4: 403 - 422
      • ... and the availability of powerful computational models (Serre et al. 2007a, Kriegeskorte 2015, LeCun et al. 2015), ...
      • ...specific instances of this class of models achieved human-level performance on a number of perceptual tasks (Kriegeskorte 2015, LeCun et al. 2015), ...
      • ...and one model with convolutional templates learned by optimizing performance on an action recognition task (LeCun et al. 2015)....
    • Hyperspectral Sensors and Imaging Technologies in Phytopathology: State of the Art

      A.-K. Mahlein,1 M.T. Kuska,2 J. Behmann,2 G. Polder,3 and A. Walter41Institute of Sugar Beet Research (IfZ), 37079 Göttingen, Germany; email: [email protected]2Institute of Crop Science and Resource Conservation (INRES)–Plant Diseases and Plant Protection, University of Bonn, 53115 Bonn, Germany3Greenhouse Horticulture, Wageningen University and Research, 6708PB Wageningen, Netherlands4Institute of Agricultural Sciences, ETH Zürich, 8092 Zürich, Switzerland
      Annual Review of Phytopathology Vol. 56: 535 - 558
      • ...Recently, deep learning arose for data analysis from machine learning (86)....
      • ...the general trend to use deep learning approaches has changed the way of data interpretation in many application fields (86)....
    • Computational Methods for Understanding Mass Spectrometry–Based Shotgun Proteomics Data

      Pavel Sinitcyn, Jan Daniel Rudolph, and Jürgen CoxComputational Systems Biochemistry Research Group, Max Planck Institute of Biochemistry, 82152 Martinsried, Germany; email: [email protected]
      Annual Review of Biomedical Data Science Vol. 1: 207 - 234
      • ...Deep learning (145, 146) is gaining traction in proteomics (75) and will likely find more applications in the future....
    • Defining Phenotypes from Clinical Data to Drive Genomic Research

      Jamie R. Robinson,1,2 Wei-Qi Wei,1 Dan M. Roden,1,3,4 and Joshua C. Denny1,31Department of Biomedical Informatics, Vanderbilt University Medical Center, Nashville, Tennessee 37232, USA; email: [email protected]2Department of General Surgery, Vanderbilt University Medical Center, Nashville, Tennessee 37232, USA3Department of Medicine, Vanderbilt University Medical Center, Nashville, Tennessee 37232, USA4Department of Pharmacology, Vanderbilt University Medical Center, Nashville, Tennessee 37232, USA
      Annual Review of Biomedical Data Science Vol. 1: 69 - 92
      • ...The key aspect of deep learning is that these layers of features are learned from the data rather than designed by domain experts (86)....
    • Toward an Integrative Theory of Thalamic Function

      Rajeev V. Rikhye,1,2 Ralf D. Wimmer,1,3 and Michael M. Halassa1,2,31Department of Brain and Cognitive Science, Massachusetts Institute of Technology, Cambridge, Massachusetts 02139, USA; email: [email protected]2McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, Massachusetts 02139, USA3Stanley Center for Psychiatric Genetics, Broad Institute, Cambridge, Massachusetts 02139, USA
      Annual Review of Neuroscience Vol. 41: 163 - 183
      • ...and recent advances in coupling artificial HCNNs with more efficient learning algorithms have given rise to the revolution of machines that are almost on par with humans in their ability to recognize objects (Hassabis et al. 2017, LeCun et al. 2015). ...
    • Computational Principles of Supervised Learning in the Cerebellum

      Jennifer L. Raymond1 and Javier F. Medina21Department of Neurobiology, Stanford University School of Medicine, Stanford, California 94305, USA; email: [email protected]2Department of Neuroscience, Baylor College of Medicine, Houston, Texas 77030, USA; email: [email protected]
      Annual Review of Neuroscience Vol. 41: 233 - 253
      • ...the process of finding a suitable representation of the input data is called feature engineering and is a critical step that often determines whether the algorithm will succeed or fail (Bengio et al. 2013, LeCun et al. 2015). (c) Instructive signals compose the third element....
    • Machine Learning Approaches for Clinical Psychology and Psychiatry

      Dominic B. Dwyer, Peter Falkai, and Nikolaos KoutsoulerisDepartment of Psychiatry and Psychotherapy, Section for Neurodiagnostic Applications, Ludwig-Maximilian University, Munich 80638, Germany; email: [email protected], [email protected], [email protected]
      Annual Review of Clinical Psychology Vol. 14: 91 - 118
      • ...The idea of meta-learning is an important concept in fields such as deep learning (LeCun et al. 2015), ...
    • Big Data in Public Health: Terminology, Machine Learning, and Privacy

      Stephen J. Mooney1 and Vikas Pejaver21Harborview Injury Prevention and Research Center, University of Washington, Seattle, Washington 98122, USA; email: [email protected]2Department of Biomedical Informatics and Medical Education and the eScience Institute, University of Washington, Seattle, Washington 98109, USA; email: [email protected]
      Annual Review of Public Health Vol. 39: 95 - 112
      • ...have been used extensively in image classification and natural language processing (68)....
    • Computational Neuroscience: Mathematical and Statistical Perspectives

      Robert E. Kass,1 Shun-Ichi Amari,2 Kensuke Arai,3 Emery N. Brown,4,5 Casey O. Diekman,6 Markus Diesmann,7,8 Brent Doiron,9 Uri T. Eden,3 Adrienne L. Fairhall,10 Grant M. Fiddyment,3 Tomoki Fukai,2 Sonja Grün,7,8 Matthew T. Harrison,11 Moritz Helias,7,8 Hiroyuki Nakahara,2 Jun-nosuke Teramae,12 Peter J. Thomas,13 Mark Reimers,14 Jordan Rodu,15 Horacio G. Rotstein,16,17 Eric Shea-Brown,10 Hideaki Shimazaki,18,19 Shigeru Shinomoto,19 Byron M. Yu,20 and Mark A. Kramer31Department of Statistics, Machine Learning Department, and Center for the Neural Basis of Cognition, Carnegie Mellon University, Pittsburgh, Pennsylvania 15213, USA; email: [email protected]2Mathematical Neuroscience Laboratory, RIKEN Brain Science Institute, Wako, Saitama Prefecture 351-0198, Japan3Department of Mathematics and Statistics, Boston University, Boston, Massachusetts 02215, USA4Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, Massachusetts 02139, USA5Department of Anesthesia, Harvard Medical School, Boston, Massachusetts 02115, USA6Department of Mathematical Sciences, New Jersey Institute of Technology, Newark, New Jersey 07102, USA7Institute of Neuroscience and Medicine, Jülich Research Centre, 52428 Jülich, Germany8Department of Theoretical Systems Neurobiology, Institute of Biology, RWTH Aachen University, 52062 Aachen, Germany9Department of Mathematics, University of Pittsburgh, Pittsburgh, Pennsylvania 15260, USA10Department of Physiology and Biophysics, University of Washington, Seattle, Washington 98105, USA11Division of Applied Mathematics, Brown University, Providence, Rhode Island 02912, USA12Department of Integrated Theoretical Neuroscience, Osaka University, Suita, Osaka Prefecture 565-0871, Japan13Department of Mathematics, Applied Mathematics, and Statistics, Case Western Reserve University, Cleveland, Ohio 44106, USA14Department of Neuroscience, Michigan State University, East Lansing, Michigan 48824, USA15Department of Statistics, University of Virginia, Charlottesville, Virginia 22904, USA16Federated Department of Biological Sciences, Rutgers University/New Jersey Institute of Technology, Newark, New Jersey 07102, USA17Institute for Brain and Neuroscience Research, New Jersey Institute of Technology, Newark, New Jersey 07102, USA18Honda Research Institute Japan, Wako, Saitama Prefecture 351-0188, Japan19Department of Physics, Kyoto University, Kyoto, Kyoto Prefecture 606-8502, Japan20Department of Electrical and Computer Engineering and Department of Biomedical Engineering, Carnegie Mellon University, Pittsburgh, Pennsylvania 15213, USA
      Annual Review of Statistics and Its Application Vol. 5: 183 - 214
      • ...3.4.5. Deep learning.Deep learning (le Cun et al. 2015) is an outgrowth of PDP modeling (see Section 1.4)....
      • ...receptive fields (le Cun et al. 2015) identify a very specific input pattern, ...
    • Neural Circuitry of Reward Prediction Error

      Mitsuko Watabe-Uchida,1, Neir Eshel,1,2, and Naoshige Uchida11Department of Molecular and Cellular Biology, Center for Brain Science, Harvard University, Cambridge, Massachusetts 02138; email: [email protected], [email protected]2Department of Psychiatry and Behavioral Sciences, Stanford University School of Medicine, Stanford, California 94305; email: [email protected]
      Annual Review of Neuroscience Vol. 40: 373 - 394
      • ...versus simple box-and-arrow computations? As is the case in modern artificial neural networks (LeCun et al. 2015), ...
    • Toward a Rational and Mechanistic Account of Mental Effort

      Amitai Shenhav,1,2 Sebastian Musslick,3 Falk Lieder,4 Wouter Kool,5 Thomas L. Griffiths,6 Jonathan D. Cohen,3,7 and Matthew M. Botvinick8,91Department of Cognitive, Linguistic and Psychological Sciences, Brown University, Providence, Rhode Island 02912; email: [email protected]2Brown Institute for Brain Science, Brown University, Providence, Rhode Island 029123Princeton Neuroscience Institute, Princeton University, Princeton, New Jersey 085444Helen Wills Neuroscience Institute, University of California, Berkeley, California 947205Department of Psychology, Harvard University, Cambridge, Massachusetts 021386Department of Psychology, University of California, Berkeley, California 947207Department of Psychology, Princeton University, Princeton, New Jersey 085408Google DeepMind, London M1C 4AG, United Kingdom9Gatsby Computational Neuroscience Unit, University College London, London W1T 4JG, United Kingdom
      Annual Review of Neuroscience Vol. 40: 99 - 124
      • ... and is driving the current explosion of interest in deep learning networks within the machine learning community (Bengio et al. 2013, Caruana 1998, LeCun et al. 2015)....
    • The Role of Variability in Motor Learning

      Ashesh K. Dhawale,1,2 Maurice A. Smith,2,3 and Bence P. Ölveczky1,21Department of Organismic and Evolutionary Biology, Harvard University, Cambridge, Massachusetts 02138; email: [email protected]2Center for Brain Science, Harvard University, Cambridge, Massachusetts 021383John A. Paulson School of Engineering and Applied Sciences, Harvard University, Cambridge, Massachusetts 02138
      Annual Review of Neuroscience Vol. 40: 479 - 498
      • ...been due to the use of convolutional network architectures that reduce dramatically the dimensionality of the solution space by enforcing highly symmetric patterns in the weights to be learned (LeCun et al. 1998, 2015...
      • ...Another key to the success of deep learning networks has been the use of unsupervised methods to pretrain networks based on the statistics of the input data (Hinton et al. 2006, LeCun et al. 2015, Lee et al. 2009)....
    • Deep Learning in Medical Image Analysis

      Dinggang Shen,1,2 Guorong Wu,1 and Heung-Il Suk21Department of Radiology, University of North Carolina, Chapel Hill, North Carolina 27599; email: [email protected]2Department of Brain and Cognitive Engineering, Korea University, Seoul 02841, Republic of Korea; email: [email protected]
      Annual Review of Biomedical Engineering Vol. 19: 221 - 248
      • ...and then discovers the informative representations in a self-taught manner (8, 9)....
      • ...Deep neural networks can discover hierarchical feature representations such that higher-level features can be derived from lower-level features (9)....
    • Personal Sensing: Understanding Mental Health Using Ubiquitous Sensors and Machine Learning

      David C. Mohr,1 Mi Zhang,2 and Stephen M. Schueller11Center for Behavioral Intervention Technologies and Department of Preventive Medicine, Northwestern University, Chicago, Illinois 60611; email: [email protected], [email protected]2Department of Electrical and Computer Engineering, Michigan State University, East Lansing, Michigan 48824; email: [email protected]
      Annual Review of Clinical Psychology Vol. 13: 23 - 47
      • ...they do not generalize well to challenging problems involving large-scale datasets (LeCun et al. 2015)....
    • Visual Object Recognition: Do We (Finally) Know More Now Than We Did?

      Isabel Gauthier1 and Michael J. Tarr21Department of Psychology, Vanderbilt University, Nashville, Tennessee 37240-7817; email: [email protected]2Department of Psychology, Center for the Neural Basis of Cognition, Carnegie Mellon University, Pittsburgh, Pennsylvania 15213
      Annual Review of Vision Science Vol. 2: 377 - 396
      • ...most typically embodied—as illustrated in Figure 3—in convolutional neural networks (CNNs) (LeCun et al. 2015). (By one estimate, ...
      • ...Figure and caption adapted, with permission, from LeCun et al. (2015)....
    • Early Visual Cortex as a Multiscale Cognitive Blackboard

      Pieter R. Roelfsema1,2,3 and Floris P. de Lange41Netherlands Institute for Neuroscience, 1105 BA Amsterdam, The Netherlands; email: [email protected]2Department of Integrative Neurophysiology, VU University Amsterdam, 1081 HV Amsterdam, The Netherlands3Psychiatry Department, Academic Medical Center, 1105 AZ Amsterdam, The Netherlands4Donders Institute for Brain, Cognition and Behavior, Radboud University, 6525 EN Nijmegen, The Netherlands
      Annual Review of Vision Science Vol. 2: 131 - 151
      • ...Recent progress in deep learning has been made in the recognition of semantic categories in photographs by using neural networks consisting of many layers (LeCun et al. 2015)....
    • Deep Neural Networks: A New Framework for Modeling Biological Vision and Brain Information Processing

      Nikolaus KriegeskorteMedical Research Council Cognition and Brain Sciences Unit, University of Cambridge, Cambridge CB2 7EF, United Kingdom; email: [email protected]
      Annual Review of Vision Science Vol. 1: 417 - 446
      • ...I argue that recent advances in neural network models (LeCun et al. 2015) will usher in a new era of computational neuroscience, ...

  • 77.
    Lee C, Kim J, Babcock D, Goodman R. 1997. Application of neural networks to turbulence control for drag reduction. Phys. Fluids 9:1740–47
    • Crossref
    • Web of Science ®
    • ADS
    • Google Scholar
    Article Location
    More AR articles citing this reference

    • Microelectromechanical Systems–Based Feedback Control of Turbulence for Skin Friction Reduction

      Nobuhide Kasagi,1 Yuji Suzuki,1 and Koji Fukagata21Department of Mechanical Engineering, The University of Tokyo, Bunkyo-ku, Tokyo 113-8656, Japan; email: [email protected]; [email protected]2Department of Mechanical Engineering, Keio University, Kohoku-ku, Yokohama 223-8522, Japan; email: [email protected]
      Annual Review of Fluid Mechanics Vol. 41: 231 - 251
      • ..., seem to achieve the best performance. Lee et al. (1997)...
      • ...It is well-known that spanwise wall shear stress is useful for the state estimation of wall turbulence and thus for feedback control (e.g., Lee et al. 1997)....
    • MECHANISMS ON TRANSVERSE MOTIONS IN TURBULENT WALL FLOWS

      G.E. KarniadakisDivision of Applied Mathematics, Brown University, Providence, Rhode Island 02912; Department of Ocean Engineering, Massachusetts Institute of Technology, Cambridge, Massachusetts 02139; email: [email protected] Kwing-So ChoiSchool of Mechanical, Materials, and Manufacturing Engineering and Management, University of Nottingham, University Park, Nottingham NG7 2RD, United Kingdom; email: [email protected]
      Annual Review of Fluid Mechanics Vol. 35: 45 - 62
      • ...the focus in modern drag-reduction techniques has been on controlling locally individual streamwise vortices based on sophisticated but often rather complex closed-loop control strategies (e.g., Lee et al. 1997, Rathnasingham & Breuer 1997)....
    • DIRECT NUMERICAL SIMULATION: A Tool in Turbulence Research

      Parviz Moin and Krishnan MaheshCenter for Turbulence Research, Stanford University, Stanford, CA 94305 NASA Ames Research Center, Moffett Field, California 94035; e-mail: [email protected] ; e-mail: [email protected]
      Annual Review of Fluid Mechanics Vol. 30: 539 - 578
      • ...A novel approach was developed by Lee et al (1997) who used DNS to train a neural network to approximate this influence and then provided the optimal wall-actuations that would minimize the drag....
    • MICRO-ELECTRO-MECHANICAL-SYSTEMS (MEMS) AND FLUID FLOWS

      Chih-Ming HoMechanical and Aerospace Engineering Department, University of California at Los Angeles, Los Angeles, California 90095; e-mail: [email protected] Yu-Chong TaiElectrical Engineering Department, California Institute of Technology, Pasadena, California 91125; e-mail: [email protected]
      Annual Review of Fluid Mechanics Vol. 30: 579 - 612
      • ...and numerical (Lee et al 1997) experiments show successful examples of using simple neural network–based processing for controlling transitional or turbulent flows....
      • ...Neural networks are a viable approach (Jacobson & Reynolds 1995, Lee et al 1997)....

  • 78.
    Lee Y, Yang H, Yin Z. 2017. PIV-DCNN: cascaded deep convolutional neural networks for particle image velocimetry. Exp. Fluids 58:171
    • Crossref
    • Web of Science ®
    • Google Scholar
    Article Location
  • 79.
    Li Q, Dietrich F, Bollt EM, Kevrekidis IG. 2017. Extended dynamic mode decomposition with dictionary learning: a data-driven adaptive spectral decomposition of the Koopman operator. Chaos 27:103111
    • Crossref
    • Medline
    • Web of Science ®
    • ADS
    • Google Scholar
    Article Location
    More AR articles citing this reference

    • Dynamic Mode Decomposition and Its Variants

      Peter J. SchmidDepartment of Mathematics, Imperial College London, London, United Kingdom; email: [email protected]
      Annual Review of Fluid Mechanics Vol. 54: 225 - 254
      • ...or discontinuous spectral elements, have been suggested (Li et al. 2017), ...
      • ...both the dictionary and the operator are the control variables for rendering the residual a minimum (Li et al. 2017)....
      • ...a three-layer feed-forward neural network of the form and for k = 0, 1, 2 (see Li et al. 2017)....
    • Koopman Operators for Estimation and Control of Dynamical Systems

      Samuel E. Otto and Clarence W. RowleyDepartment of Mechanical and Aerospace Engineering, Princeton University, Princeton, New Jersey 08544, USA; email: [email protected]
      Annual Review of Control, Robotics, and Autonomous Systems Vol. 4: 59 - 87
      • ...This idea has led to approximation techniques for the Koopman operator involving learned dictionaries (35)...
    • Statistics of Extreme Events in Fluid Flows and Waves

      Themistoklis P. SapsisDepartment of Mechanical Engineering, Massachusetts Institute of Technology, Cambridge, Massachusetts 01239, USA; email: [email protected]
      Annual Review of Fluid Mechanics Vol. 53: 85 - 111
      • ...Several works have focused on extracting dynamical information from data (Kevrekidis et al. 2015, Li et al. 2017, Takeishi et al. 2017, Wan & Sapsis 2017, Yeung et al. 2017, Lusch et al. 2018, Vlachas et al. 2018, Otto & Rowley 2019, Brunton et al. 2020)....
    • Machine Learning for Molecular Simulation

      Frank Noé,1,2,3 Alexandre Tkatchenko,4 Klaus-Robert Müller,5,6,7 and Cecilia Clementi1,3,81Department of Mathematics and Computer Science, Freie Universität Berlin, 14195 Berlin, Germany; email: [email protected]2Department of Physics, Freie Universität Berlin, 14195 Berlin, Germany3Department of Chemistry and Center for Theoretical Biological Physics, Rice University, Houston, Texas 77005, USA; email: [email protected]4Physics and Materials Science Research Unit, University of Luxembourg, 1511 Luxembourg, Luxembourg; email: [email protected]5Department of Computer Science, Technical University Berlin, 10587 Berlin, Germany; email: [email protected]6Max-Planck-Institut für Informatik, 66123 Saarbrücken, Germany7Department of Brain and Cognitive Engineering, Korea University, Seoul 136-713, South Korea8Department of Physics, Rice University, Houston, Texas 77005, USA
      Annual Review of Physical Chemistry Vol. 71: 361 - 390
      • ...Extended dynamic mode decomposition with dictionary learning (121) uses an architecture similar to that of VAMPnets but is optimized by minimizing the regression error in latent space....
      • ...such as constant functions (see Section 2.4), a suitable regularization must be employed (121). ...

  • 80.
    Liang D, Jiang C, Li Y. 2003. Cellular neural network to detect spurious vectors in PIV data. Exp. Fluids 34:52–62
    • Crossref
    • Web of Science ®
    • Google Scholar
    Article Location
  • 81.
    Ling J, Jones R, Templeton J. 2016a. Machine learning strategies for systems with invariance properties. J. Comput. Phys. 318:22–35
    • Crossref
    • Web of Science ®
    • ADS
    • Google Scholar
    Article Location
    More AR articles citing this reference

    • Turbulence Modeling in the Age of Data

      Karthik Duraisamy,1, Gianluca Iaccarino,2, and Heng Xiao3,1Department of Aerospace Engineering, University of Michigan, Ann Arbor, Michigan 48109, USA; email: [email protected]2Department of Mechanical Engineering, Stanford University, Stanford, California 94305, USA; email: [email protected]3Kevin T. Crofton Department of Aerospace and Ocean Engineering, Virginia Tech, Blacksburg, Virginia 24060, USA; email: [email protected]
      Annual Review of Fluid Mechanics Vol. 51: 357 - 377
      • ... illustrated a scheme for crafting features based on flow physics and normalizing them using local quantities; later work has expanded this approach by using invariants of a tensorial set (Ling et al. 2016a)...

  • 82.
    Ling J, Kurzawski A, Templeton J. 2016b. Reynolds averaged turbulence modelling using deep neural networks with embedded invariance. J. Fluid Mech. 807:155–66
    • Crossref
    • Web of Science ®
    • ADS
    • Google Scholar
    Article Locations:
    • Article Location
    • Article Location
    • Article Location
    More AR articles citing this reference

    • Fluid Dynamics of Axial Turbomachinery: Blade- and Stage-Level Simulations and Models

      Richard D. Sandberg1 and Vittorio Michelassi21Department of Mechanical Engineering, University of Melbourne, Parkville, Australia; email: [email protected]2Turbomachinery & Process Solutions, Baker Hughes, Florence, Italy
      Annual Review of Fluid Mechanics Vol. 54: 255 - 285
      • ...Ling et al. 2016, Weatheritt & Sandberg 2016) to ensure that physical constraints are inherently satisfied....
    • Modeling Turbulent Flows in Porous Media

      Brian D. Wood,1 Xiaoliang He,2 and Sourabh V. Apte21School of Chemical, Biological, and Environmental Engineering, Oregon State University, Corvallis, Oregon 97330, USA; email: [email protected]2School of Mechanical, Industrial, and Manufacturing Engineering, Oregon State University, Corvallis, Oregon 97330, USA
      Annual Review of Fluid Mechanics Vol. 52: 171 - 203
      • ...Duraisamy et al. 2019, Ling et al. 2016) have made this approach possible....
    • Turbulence Modeling in the Age of Data

      Karthik Duraisamy,1, Gianluca Iaccarino,2, and Heng Xiao3,1Department of Aerospace Engineering, University of Michigan, Ann Arbor, Michigan 48109, USA; email: [email protected]2Department of Mechanical Engineering, Stanford University, Stanford, California 94305, USA; email: [email protected]3Kevin T. Crofton Department of Aerospace and Ocean Engineering, Virginia Tech, Blacksburg, Virginia 24060, USA; email: [email protected]
      Annual Review of Fluid Mechanics Vol. 51: 357 - 377
      • ...Ling et al. (2016b) proposed a neural network architecture with embedded invariance properties to learn the coefficients c(θ, ...
    • Toward Constitutive Models for Momentum, Species, and Energy Transport in Gas–Particle Flows

      Sankaran Sundaresan, Ali Ozel, and Jari KolehmainenDepartment of Chemical and Biological Engineering, Princeton University, Princeton, New Jersey 08544, USA; email: [email protected]
      Annual Review of Chemical and Biomolecular Engineering Vol. 9: 61 - 81
      • ...user-friendly code (which could conceivably embed deep learning methods; see 100, 101) to propose constitutive models for microscopic stress, ...

  • 83.
    Ling J, Templeton J. 2015. Evaluation of machine learning algorithms for prediction of regions of high Reynolds averaged Navier Stokes uncertainty. Phys. Fluids 27:085103
    • Crossref
    • Web of Science ®
    • Google Scholar
    Article Locations:
    • Article Location
    • Article Location
    More AR articles citing this reference

    • Turbulence Modeling in the Age of Data

      Karthik Duraisamy,1, Gianluca Iaccarino,2, and Heng Xiao3,1Department of Aerospace Engineering, University of Michigan, Ann Arbor, Michigan 48109, USA; email: [email protected]2Department of Mechanical Engineering, Stanford University, Stanford, California 94305, USA; email: [email protected]3Kevin T. Crofton Department of Aerospace and Ocean Engineering, Virginia Tech, Blacksburg, Virginia 24060, USA; email: [email protected]
      Annual Review of Fluid Mechanics Vol. 51: 357 - 377
      • ...This marker was found to correlate well with regions where the prediction of the Reynolds stress divergence was inaccurate. Ling & Templeton (2015) employed databases of DNS and RANS solutions and formulated the evaluation of the potential adequacy of the RANS model as a classification problem in ML....
      • ...Figure adapted with permission from Ling & Templeton (2015)....
      • ...The work of Ling & Templeton (2015) illustrated a scheme for crafting features based on flow physics and normalizing them using local quantities; later work has expanded this approach by using invariants of a tensorial set (Ling et al. 2016a)...

  • 84.
    Loiseau JC, Brunton SL. 2018. Constrained sparse Galerkin regression. J. Fluid Mech. 838:42–67
    • Crossref
    • Web of Science ®
    • ADS
    • Google Scholar
    Article Location
  • 85.
    Loucks D, van Beek E, Stedinger J, Dijkman J, Villars M. 2005. Water Resources Systems Planning and Management: An Introduction to Methods, Vol. 2. Cham, Switz.: Springer
    • Google Scholar
    Article Location
  • 86.
    Lumley J. 1970. Stochastic Tools in Turbulence. New York: Academic
    • Google Scholar
    Article Location
    More AR articles citing this reference

    • John Leask Lumley: Whither Turbulence?

      Sidney Leibovich1 and Zellman Warhaft11Sibley School of Mechanical and Aerospace Engineering, Cornell University, Ithaca, New York 14853; email: [email protected], [email protected]

      Annual Review of Fluid Mechanics Vol. 50: 1 - 23
      • ...He wrote six books: Structure of Atmospheric Turbulence (Lumley & Panofsky 1964), Stochastic Tools in Turbulence (Lumley 1970a), ...
      • ...His book Stochastic Tools in Turbulence (Lumley 1970a) contains (in a very terse style) virtually everything a student needs to know to approach problems in turbulence....
    • Model Reduction for Flow Analysis and Control

      Clarence W. Rowley and Scott T.M. DawsonDepartment of Mechanical and Aerospace Engineering, Princeton University, Princeton, New Jersey 08544; email: [email protected]
      Annual Review of Fluid Mechanics Vol. 49: 387 - 417
      • ...Our aim is to review some well-developed techniques, such as proper orthogonal decomposition (POD; Lumley 1970), ...
      • ...Karhunen-Loève expansion) but was first introduced to the fluid mechanics community by Lumley (1970)....
    • DYNAMICS AND CONTROL OF HIGH-REYNOLDS-NUMBER FLOW OVER OPEN CAVITIES

      Clarence W. Rowley1 and David R. Williams21Department of Mechanical and Aerospace Engineering, Princeton University, Princeton, New Jersey 08544; email: [email protected]2Mechanical, Materials, and Aerospace Engineering Department, Illinois Institute of Technology, Chicago, Illinois 60616; email: [email protected]
      Annual Review of Fluid Mechanics Vol. 38: 251 - 276
      • ...This method was pioneered by Lumley (1970), and used to model the dynamics of a turbulent boundary layer by Aubry et al. (1988)...

  • 87.
    Lusch B, Kutz JN, Brunton SL. 2018. Deep learning for universal linear embeddings of nonlinear dynamics. Nat. Commun. 9:4950
    • Crossref
    • Medline
    • Web of Science ®
    • ADS
    • Google Scholar
    Article Location
    More AR articles citing this reference

    • Dynamic Mode Decomposition and Its Variants

      Peter J. SchmidDepartment of Mathematics, Imperial College London, London, United Kingdom; email: [email protected]
      Annual Review of Fluid Mechanics Vol. 54: 225 - 254
      • ...This type of machine learning technique for finding effective observables for Koopman analysis has been applied to systems with discrete spectra (Otto & Rowley 2019) and with discrete and continuous spectra (Lusch et al. 2018), ...
    • Koopman Operators for Estimation and Control of Dynamical Systems

      Samuel E. Otto and Clarence W. RowleyDepartment of Mechanical and Aerospace Engineering, Princeton University, Princeton, New Jersey 08544, USA; email: [email protected]
      Annual Review of Control, Robotics, and Autonomous Systems Vol. 4: 59 - 87
      • ...This point of view is also consistent with numerical methods such as those described by Lusch et al. (6), ...
      • ...The closely related approaches of References 6 and 37–42 utilize an autoencoder neural network together with recurrent linear dynamics in the latent space (e.g., ...
    • Statistics of Extreme Events in Fluid Flows and Waves

      Themistoklis P. SapsisDepartment of Mechanical Engineering, Massachusetts Institute of Technology, Cambridge, Massachusetts 01239, USA; email: [email protected]
      Annual Review of Fluid Mechanics Vol. 53: 85 - 111
      • ...Several works have focused on extracting dynamical information from data (Kevrekidis et al. 2015, Li et al. 2017, Takeishi et al. 2017, Wan & Sapsis 2017, Yeung et al. 2017, Lusch et al. 2018, Vlachas et al. 2018, Otto & Rowley 2019, Brunton et al. 2020)....

  • 88.
    Mahoney MW. 2011. Randomized algorithms for matrices and data. Found. Trends Mach. Learn. 3:123–224
    • Google Scholar
    Article Location
  • 89.
    Manohar K, Brunton BW, Kutz JN, Brunton SL. 2018. Data-driven sparse sensor placement for reconstruction: demonstrating the benefits of exploiting known patterns. IEEE Control Syst. Mag. 38:63–86
    • Crossref
    • Web of Science ®
    • Google Scholar
    Article Locations:
    • Article Location
    • Article Location
  • 90.
    Mardt A, Pasquali L, Wu H, Noé F. 2018. VAMPnets for deep learning of molecular kinetics. Nat. Commun. 9:5
    • Crossref
    • Medline
    • Web of Science ®
    • ADS
    • Google Scholar
    Article Locations:
    • Article Location
    • Article Location
    More AR articles citing this reference

    • Machine Learning for Molecular Simulation

      Frank Noé,1,2,3 Alexandre Tkatchenko,4 Klaus-Robert Müller,5,6,7 and Cecilia Clementi1,3,81Department of Mathematics and Computer Science, Freie Universität Berlin, 14195 Berlin, Germany; email: [email protected]2Department of Physics, Freie Universität Berlin, 14195 Berlin, Germany3Department of Chemistry and Center for Theoretical Biological Physics, Rice University, Houston, Texas 77005, USA; email: [email protected]4Physics and Materials Science Research Unit, University of Luxembourg, 1511 Luxembourg, Luxembourg; email: [email protected]5Department of Computer Science, Technical University Berlin, 10587 Berlin, Germany; email: [email protected]6Max-Planck-Institut für Informatik, 66123 Saarbrücken, Germany7Department of Brain and Cognitive Engineering, Korea University, Seoul 136-713, South Korea8Department of Physics, Rice University, Houston, Texas 77005, USA
      Annual Review of Physical Chemistry Vol. 71: 361 - 390
      • ...Analysis and simulation of MD trajectories have also been affected by ML—for instance, for the definition of optimal reaction coordinates (22...
      • ...the estimation of free energy surfaces (25, 28–30), the construction of Markov state models (24, 26, 31)...
      • ...SoftMax can be exploited in VAMPnets to simultaneously learn an embedding from configurations to metastable states and a Markov transition matrix (Section 4.4) (24)....
      • ...VAMPnets (24) were introduced to replace the complicated and error-prone approach of constructing Markov state models by (a) searching for optimal features; (b) combining them into a low-dimensional representation , ...
      • ...Figure 6 VAMPnet and application to alanine dipeptide. (a) A VAMPnet (24)....
      • ...Figure adapted from Reference 24....
      • ...In Reference 24, parameters were shared between the VAMPnet nodes, and a unique embedding was thereby learned....
      • ...the embedding encodes the space of the dominant Markov operator eigenfunctions (24)....
      • ...Mardt et al. (24) chose to use a SoftMax layer as an output layer, ...
      • ...The results described in Reference 24 (see, e.g., Figure 6) were competitive with and sometimes surpassed the state-of-the-art handcrafted Markov state model analysis pipeline....
      • ...Combined with networks that learn a representation, such as DTNN/SchNet (14, 100) and VAMPnets (24), ...

  • 91.
    Martin N, Gharib M. 2018. Experimental trajectory optimization of a flapping fin propulsor using an evolutionary strategy. Bioinspiration Biomim. 14:016010
    • Crossref
    • Medline
    • Web of Science ®
    • Google Scholar
    Article Location
  • 92.
    Maulik R, San O, Rasheed A, Vedula P. 2019. Subgrid modelling for two-dimensional turbulence using neural networks. J. Fluid Mech. 858:122–44
    • Crossref
    • Web of Science ®
    • ADS
    • Google Scholar
    Article Location
  • 93.
    Meena MG, Nair AG, Taira K. 2018. Network community-based model reduction for vortical flows. Phys. Rev. E 97:063103
    • Crossref
    • Medline
    • Web of Science ®
    • Google Scholar
    Article Location
    More AR articles citing this reference

    • Dynamic Mode Decomposition and Its Variants

      Peter J. SchmidDepartment of Mathematics, Imperial College London, London, United Kingdom; email: [email protected]
      Annual Review of Fluid Mechanics Vol. 54: 225 - 254
      • ...graph-theoretical methods (Gopalakrishnan Meena et al. 2018) have emerged in an effort to describe complex fluid behavior....

  • 94.
    Mehta UB, Kutler P. 1984. Computational aerodynamics and artificial intelligence. NASA Tech. Rep. NASA-TM-85994, Ames Res. Cent., Moffett Field, CA
    • Google Scholar
    Article Location
  • 95.
    Meijering E. 2002. A chronology of interpolation: from ancient astronomy to modern signal and image processing. Proc. IEEE 90:319–42
    • Crossref
    • Web of Science ®
    • Google Scholar
    Article Location
  • 96.
    Meneveau C, Katz J. 2000. Scale-invariance and turbulence models for large-eddy simulation. Annu. Rev. Fluid Mech. 32:1–32
    • Link
    • Web of Science ®
    • ADS
    • Google Scholar
  • 97.
    Mezic I. 2013. Analysis of fluid flows via spectral properties of the Koopman operator. Annu. Rev. Fluid Mech. 45:357–78
    • Link
    • Web of Science ®
    • ADS
    • Google Scholar
  • 98.
    Milano M, Koumoutsakos P. 2002. Neural network modeling for near wall turbulent flow. J. Comput. Phys. 182:1–26
    • Crossref
    • Web of Science ®
    • ADS
    • Google Scholar
    Article Locations:
    • Article Location
    • Article Location
    • Article Location
    More AR articles citing this reference

    • Microelectromechanical Systems–Based Feedback Control of Turbulence for Skin Friction Reduction

      Nobuhide Kasagi,1 Yuji Suzuki,1 and Koji Fukagata21Department of Mechanical Engineering, The University of Tokyo, Bunkyo-ku, Tokyo 113-8656, Japan; email: [email protected]; [email protected]2Department of Mechanical Engineering, Keio University, Kohoku-ku, Yokohama 223-8522, Japan; email: [email protected]
      Annual Review of Fluid Mechanics Vol. 41: 231 - 251
      • ...A similar trend can be found in the flow estimation using a neural network (Milano & Koumoutsakos 2002)....

  • 99.
    Minsky M, Papert SA. 1969. Perceptrons: An Introduction to Computational Geometry. Cambridge, MA: MIT Press
    • Google Scholar
    Article Location
    More AR articles citing this reference

    • Machine-Learning Quantum States in the NISQ Era

      Giacomo Torlai1 and Roger G. Melko2,31Center for Computational Quantum Physics, Flatiron Institute, New York, NY 10010, USA; email: [email protected]2Department of Physics and Astronomy, University of Waterloo, Waterloo, Ontario N2L 3G1, Canada; email: [email protected]3Perimeter Institute for Theoretical Physics, Waterloo, Ontario N2L 2Y5, Canada
      Annual Review of Condensed Matter Physics Vol. 11: 325 - 344
      • ...It was later shown that a single-layer perceptron is only capable of learning linearly separable functions (12), ...

  • 100.
    Mnih V, Kavukcuoglu K, Silver D, Rusu AA, Veness J, 2015. Human-level control through deep reinforcement learning. Nature 518:529–33
    • Crossref
    • Medline
    • Web of Science ®
    • ADS
    • Google Scholar
    Article Locations:
    • Article Location
    • Article Location
    More AR articles citing this reference

    • Partially Observable Markov Decision Processes and Robotics

      Hanna KurniawatiSchool of Computing, Australian National University, Canberra, Australian Capital Territory, Australia; email: [email protected]
      Annual Review of Control, Robotics, and Autonomous Systems Vol. 5: 253 - 277
      • ...Some of the early work (86–88) was model free, directly learning the policy or value function without learning the POMDP model....
    • Safe Learning in Robotics: From Learning-Based Control to Safe Reinforcement Learning

      Lukas Brunke,1,2,3, Melissa Greeff,1,2,3, Adam W. Hall,1,2,3, Zhaocong Yuan,1,2,3, Siqi Zhou,1,2,3, Jacopo Panerati,1,2,3 and Angela P. Schoellig1,2,31Institute for Aerospace Studies, University of Toronto, Toronto, Ontario, Canada; email: [email protected], [email protected], [email protected], [email protected], [email protected], [email protected], [email protected]2University of Toronto Robotics Institute, Toronto, Ontario, Canada3Vector Institute for Artificial Intelligence, Toronto, Ontario, Canada
      Annual Review of Control, Robotics, and Autonomous Systems Vol. 5: 411 - 444
      • ...This method learns an ensemble of deep Q-networks (102) and defines the risk of an action based on the variance of its value predictions....
      • ... proposed certified lower bounds for the value predictions from a deep Q-network (102), ...
    • Olfactory Sensing and Navigation in Turbulent Environments

      Gautam Reddy,1 Venkatesh N. Murthy,2,3 and Massimo Vergassola4,51NSF–Simons Center for Mathematical & Statistical Analysis of Biology, Harvard University, Cambridge, Massachusetts, USA2Department of Molecular & Cellular Biology, Harvard University, Cambridge, Massachusetts, USA3Center for Brain Science, Harvard University, Cambridge, Massachusetts, USA4Laboratoire de physique de l'École Normale Supérieure, CNRS, PSL Research University, Sorbonne Université, Paris, France; email: [email protected]5Department of Physics, University of California, San Diego, La Jolla, California, USA
      Annual Review of Condensed Matter Physics Vol. 13: 191 - 213
      • ...neural network-based approaches have reached astounding success in well-controlled simulated environments (170...
    • Learning-Based Model Predictive Control: Toward Safe Learning in Control

      Lukas Hewing, Kim P. Wabersich, Marcel Menner, and Melanie N. ZeilingerInstitute for Dynamic Systems and Control, ETH Zurich, Zurich 8092, Switzerland; email: [email protected], [email protected], [email protected], [email protected]
      Annual Review of Control, Robotics, and Autonomous Systems Vol. 3: 269 - 296
      • ...has shown great success in solving complex and high-dimensional control tasks (see, e.g., 125, 126), ...
    • Q-Learning: Theory and Applications

      Jesse Clifton and Eric LaberDepartment of Statistics, North Carolina State University, Raleigh, North Carolina 27695, USA; email: [email protected]
      Annual Review of Statistics and Its Application Vol. 7: 279 - 301
      • ...and poker (Bowling et al. 2015, Mnih et al. 2015, Silver et al. 2017)....
      • ...or the points earned in a game (Mnih et al. 2015)....
      • ...which achieved performance comparable to that of human experts in a number of Atari video games using only raw pixels as a state representation (Mnih et al. 2015)....
    • Big Data and Artificial Intelligence Modeling for Drug Discovery

      Hao ZhuDepartment of Chemistry and Center for Computational and Integrative Biology, Rutgers University, Camden, New Jersey 08102, USA; email: [email protected]
      Annual Review of Pharmacology and Toxicology Vol. 60: 573 - 589
      • ...and it is now the base of image/speech recognition, video analysis, language understanding, and other relevant applications (131)....
    • Deep Learning: The Good, the Bad, and the Ugly

      Thomas SerreDepartment of Cognitive Linguistic and Psychological Sciences, Carney Institute for Brain Science, Brown University, Providence, Rhode Island 02818, USA; email: [email protected]
      Annual Review of Vision Science Vol. 5: 399 - 426
      • ...challenging our superiority complex over machines: AI has now beaten the best human players at Atari games (Mnih et al. 2015), ...
    • Has Dynamic Programming Improved Decision Making?

      John RustDepartment of Economics, Georgetown University, Washington, DC 20057, USA; email: [email protected]
      Annual Review of Economics Vol. 11: 833 - 858
      • ...For example, Mnih et al. (2015, p. 529) showed that by combining Q-learning with multilayer or deep neural networks, ...
    • Scientific Discovery Games for Biomedical Research

      Rhiju Das,1 Benjamin Keep,2Peter Washington,3 and Ingmar H. Riedel-Kruse31Department of Biochemistry and Department of Physics, Stanford University, Stanford, California 94305, USA; email: [email protected]2Department of Learning Sciences, Stanford University, Stanford, California 94305, USA3Department of Bioengineering, Stanford University, Stanford, California 94305, USA; email: [email protected]
      Annual Review of Biomedical Data Science Vol. 2: 253 - 279
      • ...AI agents outperform human players in several instances, e.g., classic Atari video games (24)...
    • A Tour of Reinforcement Learning: The View from Continuous Control

      Benjamin RechtDepartment of Electrical Engineering and Computer Sciences, University of California, Berkeley, California 94720, USA; email: [email protected]
      Annual Review of Control, Robotics, and Autonomous Systems Vol. 2: 253 - 279
      • ...The goal now is to analyze the features x and then subsequently choose a policy that emits u so that r is large.1 There are an endless number of problems where this formulation is applied (3, 9, 10), from online decision-making in games (1, 11...
      • ...some early results in RL have shown promise in training optimal controllers directly from pixels (12, 58)....
    • A Separation Principle for Control in the Age of Deep Learning

      Alessandro Achille and Stefano SoattoDepartment of Computer Science, University of California, Los Angeles, California 90095, USA; email: [email protected], [email protected]
      Annual Review of Control, Robotics, and Autonomous Systems Vol. 1: 287 - 307
      • ...While in some tasks, such as stochastic optimal control (reinforcement learning) (39), ...
    • Computational Neuroscience: Mathematical and Statistical Perspectives

      Robert E. Kass,1 Shun-Ichi Amari,2 Kensuke Arai,3 Emery N. Brown,4,5 Casey O. Diekman,6 Markus Diesmann,7,8 Brent Doiron,9 Uri T. Eden,3 Adrienne L. Fairhall,10 Grant M. Fiddyment,3 Tomoki Fukai,2 Sonja Grün,7,8 Matthew T. Harrison,11 Moritz Helias,7,8 Hiroyuki Nakahara,2 Jun-nosuke Teramae,12 Peter J. Thomas,13 Mark Reimers,14 Jordan Rodu,15 Horacio G. Rotstein,16,17 Eric Shea-Brown,10 Hideaki Shimazaki,18,19 Shigeru Shinomoto,19 Byron M. Yu,20 and Mark A. Kramer31Department of Statistics, Machine Learning Department, and Center for the Neural Basis of Cognition, Carnegie Mellon University, Pittsburgh, Pennsylvania 15213, USA; email: [email protected]2Mathematical Neuroscience Laboratory, RIKEN Brain Science Institute, Wako, Saitama Prefecture 351-0198, Japan3Department of Mathematics and Statistics, Boston University, Boston, Massachusetts 02215, USA4Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, Massachusetts 02139, USA5Department of Anesthesia, Harvard Medical School, Boston, Massachusetts 02115, USA6Department of Mathematical Sciences, New Jersey Institute of Technology, Newark, New Jersey 07102, USA7Institute of Neuroscience and Medicine, Jülich Research Centre, 52428 Jülich, Germany8Department of Theoretical Systems Neurobiology, Institute of Biology, RWTH Aachen University, 52062 Aachen, Germany9Department of Mathematics, University of Pittsburgh, Pittsburgh, Pennsylvania 15260, USA10Department of Physiology and Biophysics, University of Washington, Seattle, Washington 98105, USA11Division of Applied Mathematics, Brown University, Providence, Rhode Island 02912, USA12Department of Integrated Theoretical Neuroscience, Osaka University, Suita, Osaka Prefecture 565-0871, Japan13Department of Mathematics, Applied Mathematics, and Statistics, Case Western Reserve University, Cleveland, Ohio 44106, USA14Department of Neuroscience, Michigan State University, East Lansing, Michigan 48824, USA15Department of Statistics, University of Virginia, Charlottesville, Virginia 22904, USA16Federated Department of Biological Sciences, Rutgers University/New Jersey Institute of Technology, Newark, New Jersey 07102, USA17Institute for Brain and Neuroscience Research, New Jersey Institute of Technology, Newark, New Jersey 07102, USA18Honda Research Institute Japan, Wako, Saitama Prefecture 351-0188, Japan19Department of Physics, Kyoto University, Kyoto, Kyoto Prefecture 606-8502, Japan20Department of Electrical and Computer Engineering and Department of Biomedical Engineering, Carnegie Mellon University, Pittsburgh, Pennsylvania 15213, USA
      Annual Review of Statistics and Its Application Vol. 5: 183 - 214
      • ...in part through the incorporation of RL (see Section 3.4.3) into deep learning architectures (Mnih et al. 2015)....
    • Zebrafish Behavior: Opportunities and Challenges

      Michael B. Orger and Gonzalo G. de PolaviejaChampalimaud Research, Champalimaud Foundation, 1400-038 Lisbon, Portugal; email: [email protected], [email protected]
      Annual Review of Neuroscience Vol. 40: 125 - 147
      • ...Combining existing technology and new results in artificial intelligence (Mnih et al. 2015) should soon allow for quantitative detailed measurements of the behavior of individual zebrafish in groups in 3-D, ...
    • Reinforcement Learning and Episodic Memory in Humans and Animals: An Integrative Framework

      Samuel J. Gershman1 and Nathaniel D. Daw21Department of Psychology and Center for Brain Science, Harvard University, Cambridge, Massachusetts 02138; email: [email protected]2Princeton Neuroscience Institute and Department of Psychology, Princeton University, Princeton, New Jersey 08544
      Annual Review of Psychology Vol. 68: 101 - 128
      • ...] was a major advance in machine learning that continues to provide the foundation for modern applications (e.g., Mnih et al. 2015)....
      • ...which can learn to discover good parametric representations from a large amount of training data (Mnih et al. 2015)....
    • Deep Neural Networks: A New Framework for Modeling Biological Vision and Brain Information Processing

      Nikolaus KriegeskorteMedical Research Council Cognition and Brain Sciences Unit, University of Cambridge, Cambridge CB2 7EF, United Kingdom; email: [email protected]
      Annual Review of Vision Science Vol. 1: 417 - 446
      • ...we will also need to employ unsupervised and reinforcement learning techniques (Sutton & Barto 1998, Mnih et al. 2015)....

  • 101.
    Nair AG, Taira K. 2015. Network-theoretic approach to sparsified discrete vortex dynamics. J. Fluid Mech. 768:549–71
    • Crossref
    • Web of Science ®
    • ADS
    • Google Scholar
    Article Location
  • 102.
    Noack BR. 2018. Closed-loop turbulence control—from human to machine learning (and retour). In Fluid-Structure-Sound Interactions and Control: Proceedings of the 4th Symposium on Fluid-Structure-Sound Interactions and Control, ed. Y Zhou, M Kimura, G Peng, AD Lucey, L Hung, pp. 23–32. Singapore: Springer
    • Google Scholar
    Article Location
  • 103.
    Noé F, Nüske F. 2013. A variational approach to modeling slow processes in stochastic dynamical systems. SIAM Multiscale Model Simul. 11:635–55
    • Crossref
    • Web of Science ®
    • Google Scholar
    Article Location
    More AR articles citing this reference

    • Machine Learning for Molecular Simulation

      Frank Noé,1,2,3 Alexandre Tkatchenko,4 Klaus-Robert Müller,5,6,7 and Cecilia Clementi1,3,81Department of Mathematics and Computer Science, Freie Universität Berlin, 14195 Berlin, Germany; email: [email protected]2Department of Physics, Freie Universität Berlin, 14195 Berlin, Germany3Department of Chemistry and Center for Theoretical Biological Physics, Rice University, Houston, Texas 77005, USA; email: [email protected]4Physics and Materials Science Research Unit, University of Luxembourg, 1511 Luxembourg, Luxembourg; email: [email protected]5Department of Computer Science, Technical University Berlin, 10587 Berlin, Germany; email: [email protected]6Max-Planck-Institut für Informatik, 66123 Saarbrücken, Germany7Department of Brain and Cognitive Engineering, Korea University, Seoul 136-713, South Korea8Department of Physics, Rice University, Houston, Texas 77005, USA
      Annual Review of Physical Chemistry Vol. 71: 361 - 390
      • ...then the minimum regression error of , the variational approach of Markov processes (VAMP) (77, 89), ...
      • ...This problem can be avoided by following a variational approach to approximating the leading terms of the spectral decomposition (Equation 10) (77, 89)....
      • ...The variational approach for conformation dynamics (89) states that for dynamics obeying detailed balance (21), ...
      • ...This is possible because with the variational approach for conformational dynamics and VAMP principles (Section 2.4) (77, 89), ...
      • ...the loss function automatically becomes a variational approach for conformational dynamics score (89)....

  • 104.
    Novati G, Mahadevan L, Koumoutsakos P. 2019. Controlled gliding and perching through deep-reinforcement-learning. Phys. Rev. Fluids 4:093902
    • Crossref
    • Web of Science ®
    • Google Scholar
    Article Locations:
    • Article Location
    • Article Location
  • 105.
    Novati G, Verma S, Alexeev D, Rossinelli D, Van Rees WM, Koumoutsakos P. 2017. Synchronisation through learning for two self-propelled swimmers. Bioinspiration Biomim. 12: