Researchers from Oak Ridge National Laboratory developed a novel design and training strategy for reversible ResNets that reduces the dimensionality of high-dimensional machine learning models for complex physical systems.
Developing reduced-order models of complex physical systems is computationally expensive. ORNL researchers have developed a neural network-based approach that reduces the number of inputs necessary to develop these models and, by extension, the complexity of HPC applications. The team’s method:
- reduced a 20-dimensional model to 1-dimension.
- reduced the error rate (compared to a standard NN) from 35.1% to 1.6%.
Input reduction is achieved by employing residual neural networks, or ResNets, which utilize shortcuts to bypass layers. The ORNL team’s approach can be used for a wide range of applications (and even experimental data), such as the team’s acceleration of the design process of multi-layer composite shells (which are used in pressure vessels, reservoirs and tanks, and rocket and spacecraft parts) by determining optimum ply angles.
The researchers are currently working on scaling the algorithm up to ORNL’s Summit supercomputer, currently the world’s most powerful.
Research team: Guannan Zhang and Jacob Hinkle, both of ORNL
Funding: This material was based upon work supported by the Artificial Intelligence Initiative at the Oak Ridge National Laboratory (ORNL). ORNL is operated by UT-Battelle, LLC., for the U.S. Department of Energy under Contract DE-AC05-00OR22725.
Contact: Guannan Zhang, firstname.lastname@example.org