Repository logo
Log In(current)
  1. Home
  2. Colleges & Schools
  3. Graduate School
  4. Doctoral Dissertations
  5. Nonparametric Bayesian Deep Learning for Scientific Data Analysis
Details

Nonparametric Bayesian Deep Learning for Scientific Data Analysis

Date Issued
December 1, 2020
Author(s)
Agrawal, Devanshu
Advisor(s)
Jacob D. Hinkle
Additional Advisor(s)
Jacob Hinkle, Georgia Tourassi, Vasileios Maroulas, Hairong Qi
Abstract

Deep learning (DL) has emerged as the leading paradigm for predictive modeling in a variety of domains, especially those involving large volumes of high-dimensional spatio-temporal data such as images and text. With the rise of big data in scientific and engineering problems, there is now considerable interest in the research and development of DL for scientific applications. The scientific domain, however, poses unique challenges for DL, including special emphasis on interpretability and robustness. In particular, a priority of the Department of Energy (DOE) is the research and development of probabilistic ML methods that are robust to overfitting and offer reliable uncertainty quantification (UQ) on high-dimensional noisy data that is limited in size relative to its complexity. Gaussian processes (GPs) are nonparametric Bayesian models that are naturally robust to overfitting and offer UQ out-of-the-box. Unfortunately, traditional GP methods lack the balance of expressivity and domain-specific inductive bias that is key to the success of DL. Recently, however, a number of approaches have emerged to incorporate the DL paradigm into GP methods, including deep kernel learning (DKL), deep Gaussian processes (DGPs), and neural network Gaussian processes (NNGPs). In this work, we investigate DKL, DGPs, and NNGPs as paradigms for developing robust models for scientific applications. First, we develop DKL for text classification, and apply both DKL and Bayesian neural networks (BNNs) to the problem of classifying cancer pathology reports, with BNNs attaining new state-of-the-art results. Next, we introduce the deep ensemble kernel learning (DEKL) method, which is just as powerful as DKL while admitting easier model parallelism. Finally, we derive a new model called a ``bottleneck NNGP'' by unifying the DGP and NNGP paradigms, thus laying the groundwork for a new class of methods for future applications.

Subjects

Gaussian process

Neural network

Probability

Ensemble

Natural language proc...

Cancer

Disciplines
Applied Mathematics
Artificial Intelligence and Robotics
Computer Sciences
Data Science
Probability
Statistics and Probability
Degree
Doctor of Philosophy
File(s)
Thumbnail Image
Name

agrawal_dissertation_revision2.pdf

Size

4.15 MB

Format

Adobe PDF

Checksum (MD5)

ffbb64ab57b8fc7c00e47900d175e61d

Learn more about how TRACE supports reserach impact and open access here.

Built with DSpace-CRIS software - Extension maintained and optimized by 4Science

  • Privacy policy
  • End User Agreement
  • Send Feedback
  • Contact
  • Libraries at University of Tennessee, Knoxville
Repository logo COAR Notify