Collaborative Research: Scalable Linear Algebra and Neural Network Theory

Information

  • NSF Award
  • 2134247
Owner
  • Award Id
    2134247
  • Award Effective Date
    9/1/2021 - 4 years ago
  • Award Expiration Date
    8/31/2024 - a year ago
  • Award Amount
    $ 439,618.00
  • Award Instrument
    Continuing Grant

Collaborative Research: Scalable Linear Algebra and Neural Network Theory

These projects will use randomized numerical linear algebra building blocks to develop improved methods in stochastic optimization theory and statistical/machine learning theory. The motivation is that, while machine learning and deep learning methodology has transformed certain applications, such as computer vision and natural language processing, its promised impact on many other areas has yet to be seen. The reason for this is the flip side of why it has been successful where it has. In the applications where it has had the most remarkable successes, people have adopted the following strategy: get large quantities of data; train a neural network model using stochastic first order methods; and implement and apply the model in a user-facing industrial application. There are many well-known limitations with this general approach, ranging from the need for large quantities of data and daunting compute resources to interpretability and robustness issues. These limitations are particularly apparent when using neural networks for problems such as high-performance computing, fluid mechanics/dynamics, temporal supply chain forecasting problems, biotechnology, etc., where interpretability is paramount. This work aims to address central technical issues underlying this approach, namely: while linear algebraic techniques are central to the design and use of modern neural network models, current methodology uses linear algebra in relatively superficial ways. If we have stronger control over the linear algebraic methods, the community will have a more practical theory to guide neural network use in a broad range of applications beyond computer vision and natural language processing. These methods will enable qualitatively more refined scalable implementations and applications of neural network models in a range of scientific and engineering domains. Broader impacts of these projects include mentoring of grant-supported graduate students and postdoctoral researchers.<br/><br/>Technically, the work will focus on three general directions: optimization theory, including convex optimization based neural network and going beyond optimization; scalable linear algebra theory, including randomized linear algebra for neural networks, and sparse randomized linear algebra; and statistics and machine learning theory, including implicit regularization, and learning with limited non-iid data. More broadly, the goal is to provide a basis for practical theory that can guide practice, in a manner analogous to how linear algebraic and functional analytic methods underlie practical and useful theory in a broad range of scientific/engineering applications. We expect that such a challenging task is possible since many of the recent developments in machine learning theory and neural network practice have parallels in scientific computing, where there is a long history of what may be called scalable linear algebra for physical/engineering theory. Many of the methods to be developed may be viewed as bridging the interdisciplinary gap between these old ideas and the new challenges we face; and principal investigators have a history of developing interdisciplinary classes, summer schools, workshops related to the topics of the proposed work, and they will continue to do so.<br/><br/>This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.

  • Program Officer
    Christopher Starkcstark@nsf.gov7032924869
  • Min Amd Letter Date
    8/12/2021 - 4 years ago
  • Max Amd Letter Date
    8/12/2021 - 4 years ago
  • ARRA Amount

Institutions

  • Name
    International Computer Science Institute
  • City
    Berkeley
  • State
    CA
  • Country
    United States
  • Address
    2150 Shattuck Ave, Suite 1100
  • Postal Code
    947041345
  • Phone Number
    5106662900

Investigators

  • First Name
    Amir
  • Last Name
    Gholaminejad
  • Email Address
    amirgh@berkeley.edu
  • Start Date
    8/12/2021 12:00:00 AM
  • First Name
    Michael
  • Last Name
    Mahoney
  • Email Address
    mmahoney@icsi.berkeley.edu
  • Start Date
    8/12/2021 12:00:00 AM

Program Element

  • Text
    OFFICE OF MULTIDISCIPLINARY AC
  • Code
    1253
  • Text
    Special Projects - CCF
  • Code
    2878
  • Text
    IIS Special Projects
  • Code
    7484
  • Text
    CDS&E-MSS
  • Code
    8069

Program Reference

  • Text
    Artificial Intelligence (AI)
  • Text
    Machine Learning Theory