Deep Generative Models that Solve PDEs: Distributed Computing for Training Large Data-Free Models

Thumbnail Image
Date
2020-07-24
Authors
Botelho, Sergio
Joshi, Ameya
Khara, Biswajit
Sarkar, Soumik
Hegde, Chinmay
Adavani, Santi
Ganapathysubramanian, Baskar
Major Professor
Advisor
Committee Member
Journal Title
Journal ISSN
Volume Title
Publisher
Authors
Person
Research Projects
Organizational Units
Organizational Unit
Mechanical Engineering
The Department of Mechanical Engineering at Iowa State University is where innovation thrives and the impossible is made possible. This is where your passion for problem-solving and hands-on learning can make a real difference in our world. Whether you’re helping improve the environment, creating safer automobiles, or advancing medical technologies, and athletic performance, the Department of Mechanical Engineering gives you the tools and talent to blaze your own trail to an amazing career.
Organizational Unit
Journal Issue
Is Version Of
Versions
Series
Department
Mechanical EngineeringElectrical and Computer EngineeringPlant Sciences Institute
Abstract

Recent progress in scientific machine learning (SciML) has opened up the possibility of training novel neural network architectures that solve complex partial differential equations (PDEs). Several (nearly data free) approaches have been recently reported that successfully solve PDEs, with examples including deep feed forward networks, generative networks, and deep encoder-decoder networks. However, practical adoption of these approaches is limited by the difficulty in training these models, especially to make predictions at large output resolutions (≥1024×1024). Here we report on a software framework for data parallel distributed deep learning that resolves the twin challenges of training these large SciML models - training in reasonable time as well as distributing the storage requirements. Our framework provides several out of the box functionality including (a) loss integrity independent of number of processes, (b) synchronized batch normalization, and (c) distributed higher-order optimization methods. We show excellent scalability of this framework on both cloud as well as HPC clusters, and report on the interplay between bandwidth, network topology and bare metal vs cloud. We deploy this approach to train generative models of sizes hitherto not possible, showing that neural PDE solvers can be viably trained for practical applications. We also demonstrate that distributed higher-order optimization methods are 2−3× faster than stochastic gradient-based methods and provide minimal convergence drift with higher batch-size.

Comments

This is a pre-print of the article Botelho, Sergio, Ameya Joshi, Biswajit Khara, Soumik Sarkar, Chinmay Hegde, Santi Adavani, and Baskar Ganapathysubramanian. "Deep Generative Models that Solve PDEs: Distributed Computing for Training Large Data-Free Models." arXiv preprint arXiv:2007.12792 (2020). Posted with permission.

Description
Keywords
Citation
DOI
Source
Copyright
Wed Jan 01 00:00:00 UTC 2020
Collections