Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2021 Jun 23;7(26):eabf3658.
doi: 10.1126/sciadv.abf3658. Print 2021 Jun.

One for all: Universal material model based on minimal state-space neural networks

Affiliations

One for all: Universal material model based on minimal state-space neural networks

Colin Bonatti et al. Sci Adv. .

Abstract

Computational models describing the mechanical behavior of materials are indispensable when optimizing the stiffness and strength of structures. The use of state-of-the-art models is often limited in engineering practice due to their mathematical complexity, with each material class requiring its own distinct formulation. Here, we develop a recurrent neural network framework for material modeling by introducing "Minimal State Cells." The framework is successfully applied to datasets representing four distinct classes of materials. It reproduces the three-dimensional stress-strain responses for arbitrary loading paths accurately and replicates the state space of conventional models. The final result is a universal model that is flexible enough to capture the mechanical behavior of any engineering material while providing an interpretable representation of their state.

PubMed Disclaimer

Figures

Fig. 1
Fig. 1. Mechanical material modeling workflow.
(A) Examples of classes of engineering materials (middle) each requiring a different model (top) and implementation when taking the conventional physics-based modeling approach. A single universal model (bottom) replaces the different models when adopting a machine learning–based systematic solution. (B) Information flow in an MSC with three inputs and three outputs, ns = 2, w = 5, dt = 2, and do = 1. Each weighted connection (thin line) represents a parameter of the model. Bias terms are not represented. Symbols ψ and φ, respectively, represent the hyperbolic tangent and logistic sigmoid activation functions.
Fig. 2
Fig. 2. Effect of model architecture on the performance of MSCs.
(A to D) Effect of the number of state variables ns of MSCs, with comparison (A and B) to single-layer GRUs and LSTMs. (A) Isotropic-hardening model (7 state variables). (B) Mixed-hardening model (12 state variables). (C) Foam model (7 state variables). (D) Rubber model (7 state variables). (E and F) Effect of the number of parameters on the mean squared error on the validation datasets for optimal state-space MSCs compared with stacked LSTMs and GRUs for the (E) isotropic-hardening model and (F) mixed-hardening model.
Fig. 3
Fig. 3. Predictive capabilities of MSCs and learned features for different materials.
Results for compact MSC models (with dt = 4 and w = 25) after training with random walk data compared to Finite Element Model (FEM) results. (A to C) First stress component prediction on the worst predicted example for the mixed-hardening material: (A) validation dataset, (B) 500-long test dataset, and (C) 100,000-long test dataset. (D) Bauschinger effect for cyclic loading of the mixed-hardening material. (E) Mullins effect during unloading-reloading of the rubber material. (F and G) Spearman correlation coefficients between state variables of MSCs and state variables of the physics-based models evaluated on the validation dataset of (F) the isotropic-hardening model and (G) the crushable foam model. (H and I) Internal state variable of an MSC model as a function of the equivalent plastic strain of the corresponding physics-based model for the 500 validation examples for the (H) isotropic-hardening model and (I) crushable foam model. (J and K) Extraction of yield surface and demonstration of their self-similar evolution by plotting iso-contours of the MSC-inferred internal state variable. Gray lines indicate the isolines of the theoretical yield surface for comparison. (J) Iso-levels (0.0001, 0.01, 0.02, 0.03, 0.04, and 0.05) obtained from 3000 isochoric proportional loading path simulations for the isotropic-hardening model. (K) Iso-levels (0.003, 0.03, 0.06, 0.09, 0.12, and 0.15) obtained from 1000 proportional loading path simulations for the crushable foam model.

References

    1. Si X., Hu C., Zhang J., A review of recurrent neural networks: Lstm cells and network architectures. Neural Comput. 31, 1235–1270 (2019). - PubMed
    1. Sutskever I., Vinyals O., Le Q. V., Sequence to sequence learning with neural networks. Adv. Neural Inf. Process. Syst. 4, 3104–3112 (2014).
    1. R. Pascanu, C. Gulcehre, K. Cho, Y. Bengio, How to construct deep recurrent neural networks, in Proceedings of the 2nd International Conference on Learning Representations (ICLR 2014) (2014), pp. 1–13.
    1. Bengio Y., Simard P., Frasconi P., Learning long-term dependencies with gradient descent is difficult. IEEE Trans. Neural Netw. 5, 157–166 (1994). - PubMed
    1. Hochreiter S., Urgen Schmidhuber J., Long short-term memory. Neural Comput. 9, 1735–1780 (1997). - PubMed