AI

Transfer Learning in Scalable Graph Neural Network for Improved Physical Simulation

1 Mins read

In recent years, graph neural network (GNN) based models showed promising results in simulating complex physical systems. However, training dedicated graph network simulator can be costly, as most models are confined to fully supervised training. Extensive data generated from traditional simulators is required to train the model. It remained unexplored how transfer learning could be applied to improve the model performance and training efficiency. In this work, we introduce a pretraining and transfer learning paradigm for graph network simulator.
First, We proposed the scalable graph U-net (SGUNet). By incorporating an innovative depth-first search (DFS) pooling, the SGUNet is configurable to adaptable different mesh size and resolutions for different simulation tasks. To enable the transfer learning between different configured SGUNet, we propose a set of mapping functions to align the parameters between pretrained model and target model. An extra normalization term is also added into loss to constrain the similarity between the pretrained weights and target model weights for better generalization performance. Then we created a dataset for pretraining the simulators. It includes 20,000 physical simulations with 3D shapes randomly selected from the open source A Big CAD (ABC) datasets. We demonstrate that with our proposed transfer learning methods, model fine-tuned with a small portion of the training data could reach even better performance compared with the one trained from scratch. On 2D Deformable Plate, our pretrained model fine-tuned on 1/16 of the training data could achieve 11.05% improvement compared to model trained from scratch.


Source link

Related posts
AI

Meta AI Introduces MLGym: A New AI Framework and Benchmark for Advancing AI Research Agents

3 Mins read
The ambition to accelerate scientific discovery through AI has been longstanding, with early efforts such as the Oak Ridge Applied AI Project…
AI

Microsoft Researchers Introduces BioEmu-1: A Deep Learning Model that can Generate Thousands of Protein Structures Per Hour on a Single GPU

3 Mins read
Proteins are the essential component behind nearly all biological processes, from catalyzing reactions to transmitting signals within cells. While advances like AlphaFold…
AI

Building a Legal AI Chatbot: A Step-by-Step Guide Using bigscience/T0pp LLM, Open-Source NLP Models, Streamlit, PyTorch, and Hugging Face Transformers

3 Mins read
In this tutorial, we will build an efficient Legal AI CHatbot using open-source tools. It provides a step-by-step guide to creating a…

 

 

Leave a Reply

Your email address will not be published. Required fields are marked *