Abstract
We present an approach for the data-driven modeling of nonlinear viscoelastic materials at small strains which is based on physics-augmented neural networks (NNs) and requires only stress and strain paths for training. The model is built on the concept of generalized standard materials and is therefore thermodynamically consistent by construction. It consists of a free energy and a dissipation potential, which can be either expressed by the components of their tensor arguments or by a suitable set of invariants. The two potentials are described by fully/partially input convex neural networks. For training of the NN model by paths of stress and strain, an efficient and flexible training method based on a long short-term memory cell is developed to automatically generate the internal variable(s) during the training process. The proposed method is benchmarked and thoroughly compared with existing approaches. Different databases with either ideal or noisy stress data are generated for training by using a conventional nonlinear viscoelastic reference model. The coordinate-based and the invariant-based formulation are compared and the advantages of the latter are demonstrated. Afterwards, the invariant-based model is calibrated by applying the three training methods using ideal or noisy stress data. All methods yield good results, but differ in computation time and usability for large data sets. The presented training method based on a recurrent cell turns out to be particularly robust and widely applicable. We show that the presented model together with the recurrent cell for training yield complete and accurate 3D constitutive models even for sparse bi- or uniaxial training data.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
Despite a large number of existing classical constitutive models for nonlinear elastic and inelastic materials [1, 2], the description of novel materials with complex constitutive behavior remains a challenging task. The choice of a suitable model and the identification of the corresponding material parameters is time-consuming and does not necessarily lead to results with the desired accuracy, so that the development of new specialized models may be necessary. For this reason, a new class of approaches has emerged in recent years, which are often referred to as data-driven or data-based methods [3, 4].
1.1 Constitutive modeling with neural networks
Historically, the pioneering work of Ghaboussi et al. [5] from the beginning of the 1990s is particularly noteworthy. In this work, NNs, specifically feedforward neural networks (FNNs), are used for the first time to predict hysteresis in uniaxial and multiaxial stress states. Therein, the FNN is fed with information from several previous time steps to enable it to learn history-dependent behavior. In Furukawa and Yagawa [6], a constitutive model based on an FNN is presented that can be used to train uniaxial viscoplastic behavior. However, this requires internal variables in the training process, whose experimental identification is also described. Despite the rather simple nature of these models from today’s point of view, approaches of this kind indicate the potential of data-driven constitutive modeling. Without having to decide on a specific model, it is possible to learn complex material behavior. With the recent rise in popularity of NNs and the associated rapid progress in efficiency and accessibility, many different methods have emerged in an extremely short time, extending and improving these approaches. For example, the works [7,8,9,10,11] propose advanced techniques using FNNs, while [10,11,12,13,14] make use of recurrent neural networks (RNNs), which are capable of making predictions based on past events due to their internal structure [15].
Thus, recurrent architectures are an appealing way to model inelastic behavior, since the provision of history variables or internal variables can be avoided [17]. In particular, the development of advanced RNN cells such as long short-term memory (LSTM) [18] or gated recurrent units (GRUs) [19], which provide increased memory capacity and enable an efficient training, lead to great popularity and rapid progress in this field.
What remains a challenge for NN-based approaches is the lack of a fundamental inclusion of physics, especially the second law of thermodynamics. The networks map input quantities directly to the variable of interest, for example the stress, and thus approximate only the phenomenological relationship between input and output [5, 6]. Compared to most classical constitutive models, the physical motivation is completely missing. This has some significant downsides. Such models, often denoted as black box models, cannot guarantee robustness of the predictions beyond the training range covered by the used training data. Exceeding this range may not only lead to wrong but also to physically implausible predictions [20, 21]. Furthermore, the training process is entirely driven by the training data and is not supported by existing knowledge from classical constitutive modeling. This may lead to more complex optimization problems that exhibit gradient conflicts and require more training data [21]. For this reason, it seems natural to integrate the existing knowledge from continuum mechanics and constitutive modeling into the NN architecture to combine the advantages of both concepts. This kind of NN-based constitutive models or scientific machine learning approaches in general are labeled as physics-informed [22, 23], mechanics-informed [24], physics-augmented [25], physics-constrained [26], thermodynamics-informed [27] or thermodynamics-based [21]. These approaches have been intensively pursued for a few years with great success in constitutive modeling for elastic [20, 23, 25, 26, 28,29,30] elastoplastic [21, 27, 31,32,33,34,35,36] or viscoelastic behavior [36,37,38,39,40]. Thereby, a distinction must be made between methods with weak fulfillment of the principles by an additional term in the loss function and strong fulfillment with a priori compliance with the respective principle by constraining the architecture of the network [41]. According to the comparative study presented in Rosenkranz et al. [17], the second approach is more promising since it is more efficient in terms of required data, robust and can extrapolate very well due to the high degree of incorporated physics, but involves some difficulties. The challenge here is to efficiently restrict the network without loosing too much flexibility.
For elastic and especially hyperelastic materials, many works have been published on that topic, e.g., [20, 24, 26, 28, 29, 42,43,44], among others, in which different requirements for a constitutive model are incorporated in a strong sense. Thereby, an elastic potential is approximated by using an FNN with the deformation or strain invariants as input. To allow the calibration of the network directly by tuples of stress and strain, the derivative of the potential with respect to the deformation, i.e., the network’s input, is included into the loss, which is also called Sobolev training [20, 45]. With easily accessible implementations for automatic differentiation [46] in popular libraries like TensorFlow, PyTorch or JAX, this is no longer a major difficulty and is used in a wide variety of research areas [22, 47]. Furthermore, polyconvex NN models [20, 29, 48] have been formulated by using fully input convex neural networks (FICNNs) as introduced by Amos et al. [49]. Also parametrized polyconvex models [50] have been formulated with partially input convex neural networks (PICNNs) [49].
Regarding NN-based constitutive modeling of inelastic behavior with a strong physical background, a variety of works have been published meanwhile. Thereby, approaches applying the concept of internal variables have shown to be particularily well suited. Remarkable works on the NN-based modeling of plasticity in recent years are, for example [21, 27, 31,32,33, 51,52,53], among others. In many of these works, however, the thermodynamic consistency is only fulfilled in a weak sense by adding a penalty term to the loss or internal variables must be known prior to training. When using the models in multiscale problems, internal variables can be determined, e.g., by autoencoders [54, 55], but in real experimental setups the determination might not be practical. Thus, there are still some challenges to overcome, in particular with regard to the fulfillment of physical conditions by construction or the provision of internal variables during training. Only the elasto-plastic NN models [32, 33] a priori fulfill the second law of thermodynamics and do not require internal variables in the training data set at the same time. A pioneering NN-based approach to model viscoelastic behavior is presented in [56]. Thereby, thermodynamic consistency is ensured by using a convex dissipation potential but internal variables are required for training. Several approaches based on a similar modeling strategy can be found [17, 39, 44, 57]. In contrast to [56], however, no prior knowledge of the internal variable(s) is needed for training there.
1.2 Objectives and contributions of this work
As outlined above, NN-based constitutive models for viscoelastic problems using internal variables and accounting for fundamental physics as the second law of thermodynamics have so far received comparatively little attention. In addition, the techniques for providing internal variables during training are not satisfactory in every case, as either a high computational effort is required or flexibility and accuracy are not sufficient. Thus, within this contribution, we present a physics-augmented NN model for viscoelastic materials and an efficient training method which only requires stress and strain paths for training. The model is built on the concept of generalized standard materials and is therefore thermodynamically consistent by construction. It consists of a free energy and a dissipation potential, which can be either expressed by the coordinates of their tensor arguments or by a suitable set of invariants. The two potentials are expressed by FICNNs/PICNNs [49]. For training of the NN model by paths of stress and strain, an efficient and flexible training method based on an LSTM cell is developed to automatically provide the internal variable(s) during the training process. The focus of this work is on a comprehensive benchmark test of the proposed method based on existing approaches. These include a method that obtains the internal variable by integrating the evolution equation over the entire sequence [37], while the other method uses an auxiliary FNN for the internal variable(s) [24]. Databases for training are generated by using a conventional nonlinear viscoelastic reference model. These training data sets include either multiaxial, biaxial, uniaxial stress states with either ideal or noisy data points. The coordinate-based and the invariant-based formulation are compared and the three training methods are applied. We show that the presented framework yields complete and accurate 3D constitutive models with all of these datasets, using only a single short training sequence. An overview about this work and a classification in a larger context is schematically illustrated in Fig. 1.
The article is organized as follows: After a short summary of the concept of generalized standard materials in Sect. 2, the NN-based model is described in Sect. 3. In Sect. 4, three methods to calibrate the NN model are explained. Subsequently, the presented model and the training methods are tested and compared within numerical examples in Sect. 5 using different data sets with stress states of decreasing complexity and either ideal or noisy stress data. After a discussion of the results, some closing remarks are given in Sect. 6.
Notation
Throughout this work, several important quantities are tensors of different ranks. The space of tensors of rank \(n\ge 1\) is denoted as \(\mathcal {L}_{n}\). Especially, the cases of \(n=2\) and \(n=4\), i.e., tensors of rank two and four, are of importance herein. A tensor of rank two is denoted with upright bold symbols as \({\textbf {T}} = T_{ij} \varvec{e}_i \otimes \varvec{e}_j \in \mathcal {L}_{2}\) and a tensor of rank four with blackboard bold symbols as \({\mathbb {C}} = C_{ijkl} \varvec{e}_i \otimes \varvec{e}_j \otimes \varvec{e}_k \otimes \varvec{e}_l \in \mathcal {L}_{4}\). Therein, the Einstein summation convention is used, \(\otimes \) is the dyadic product and \(\varvec{e}_i \in \mathcal {L}_{1}\) is the i-th Cartesian basis vector. For tensors of rank two, only symmetric second order tensors \({\textbf {S}}\in \mathcal {S}\mathcalligra {ym}_2\subset \mathcal {L}_{2}\) are relevant in the scope of this work, with \(\mathcal {S}\mathcalligra {ym}_2= \left\{ \,{\textbf {S}}\in \mathcal {L}_{2} \mid {\textbf {S}}={\textbf {S}}^\top \, \right\} \) and \({\textbf {S}}^\top =S_{ji}\varvec{e}_i \otimes \varvec{e}_j\) denoting the transpose of \({\textbf {S}}\). For tensors of rank four, the subset \(\mathcal {S}\mathcalligra {ym}_4\subset \mathcal {L}_{4}\) of tensors with major and minor symmetries is defined as \(\mathcal {S}\mathcalligra {ym}_4= \left\{ \,{\mathbb {C}}\in \mathcal {L}_{4} \mid C_{ijkl}=C_{ijlk}=C_{jikl}=C_{klij} \, \right\} \). Some special fourth order tensors required here are the fully symmetric fourth order identity tensor \({\mathbb {I}}^{\text {S}}=\frac{1}{2}\left( \delta _{ik}\delta _{jl} + \delta _{il}\delta _{jk} \right) \varvec{e}_i \otimes \varvec{e}_j \otimes \varvec{e}_k \otimes \varvec{e}_l \in \mathcal {S}\mathcalligra {ym}_4\), the spherical projector \({\mathbb {I}}^{\text {K}}=\frac{1}{3}\varvec{1}\otimes \varvec{1} \in \mathcal {S}\mathcalligra {ym}_4\) with \(\varvec{1}=\delta _{ij}\varvec{e}_i\otimes \varvec{e}_j\) and the deviatoric projector given by \({\mathbb {I}}^{\text {D}}={\mathbb {I}}^{\text {S}}-{\mathbb {I}}^{\text {K}}\). In the above definitions, denotes double contraction of adjacent indices. Furthermore, \({{\,\textrm{tr}\,}}({\textbf {T}})= T_{kk}\) is the trace of \({\textbf {T}}\in \mathcal {L}_{2}\), the square of a tensor is \({\textbf {T}}^2=T_{ik}T_{kj}\varvec{e}_i\otimes \varvec{e}_j\) and the Frobenius norm is denoted by \(\Vert {\textbf {T}} \Vert \) in this work and is defined as .
2 Generalized standard materials
The concept of generalized standard materials (GSMs) [58,59,60,61,62,63,64,65] allows the formulation of thermodynamically consistent constitutive models that are entirely described by two scalar valued functions. In the context of this work, thermodynamic consistency is equivalent to satisfying the Clausius-Duhem inequality
where \(\mathcal {D}\) is the dissipation rate, \(\varvec{\upsigma }\in \mathcal {S}\mathcalligra {ym}_2\) is the stress, \({{\dot{\varvec{\upvarepsilon }}}} \in \mathcal {S}\mathcalligra {ym}_2\) is the strain rate, and \(\psi \) is the Helmholtz free energy density, or free energy for short. This free energy \(\psi (\varvec{\upvarepsilon }, {\textbf {q}}_\alpha )\) depends on the current strain \(\varvec{\upvarepsilon }\in \mathcal {S}\mathcalligra {ym}_2\) and a set of internal variables \({\textbf {q}}_\alpha \in \mathcal {S}\mathcalligra {ym}_2\). The stress \(\varvec{\upsigma }\) and the internal forces \(\varvec{\uptau }_\alpha \in \mathcal {S}\mathcalligra {ym}_2\) are obtained by differentiating the free energy with respect to the strain or the internal variables, respectively, i.e.,
In the context of GSMs, there exists another potential besides the free energy, the so-called dissipation potential \(\phi ({\dot{{\textbf {q}}}}_\alpha , {\textbf {q}}_\alpha , \varvec{\upvarepsilon })\), which depends on the rates of the internal variables \({\dot{{\textbf {q}}}}_\alpha \) and possibly on the strain and the internal variables themselves. Differentiating the dissipation potential with respect to the rate of the internal variables again yields the internal forces
which are denoted with an additional \(\hat{(\cdot )}\) to distinguish them from the internal forces \(\varvec{\uptau }_\alpha \) obtained with the free energy. To construct a constitutive model that complies with in Eq. (1), the dissipation potential must satisfy all of the following three requirements:
-
(i)
\(\phi ({\dot{{\textbf {q}}}}_\alpha , {\textbf {q}}_\alpha , \varvec{\upvarepsilon })\) must be convex in all \({\dot{{\textbf {q}}}}_\alpha \),
-
(ii)
\(\phi ({\dot{{\textbf {q}}}}_\alpha =\varvec{0}, {\textbf {q}}_\alpha , \varvec{\upvarepsilon }) = 0\) and
-
(iii)
\(\phi ({\dot{{\textbf {q}}}}_\alpha , {\textbf {q}}_\alpha , \varvec{\upvarepsilon }) \ge 0\).
The material law is formulated with the Biot equation [66], that relates both potentials via
Evaluating Eq. (4) gives rise to the evolution laws [67] and is thus the evolution equation for the internal variables given in an implicit form.Footnote 1 The stated conditions on \(\phi \) automatically construct these evolution laws such that inequality Eq. (1) is satisfied. This framework of GSMs contains various classical constitutive models, including the viscoelastic reference model in Sect. 5.1.1 that is used to generate the data for the numerical experiments.
3 Formulation of the physics-augmented neural network-based model
Based on the theoretical background from Sect. 2, an NN-based constitutive model for viscoelastic materials is presented. It uses a single internal variable \({\textbf {q}}\in \mathcal {S}\mathcalligra {ym}_2\) of strain type to model the inelastic behavior. The model adapts the concept of generalized standard materials, where the two potentials are expressed as FNNs with incorporated convexity constraints [20, 25, 37, 39, 43, 49, 50], see App. A. First, the modeling approaches for free energy and dissipation potential are described, followed by the prediction process using the adapted free energy and dissipation potential. The training methods to find the weights and biases of the FNNs without requiring the internal variable in the training data set are explained in Sect. 4.
3.1 Modeling of the potentials
The potentials \(\psi \) and \(\phi \) can be either expressed in terms of tensor coordinates (\(\psi (\varvec{\upvarepsilon }, {\textbf {q}})\) and \(\phi ({\dot{{\textbf {q}}}}, {\textbf {q}}, \varvec{\upvarepsilon })\)) or with a set of invariants of those tensors (\(\psi (\mathcal {I}^\psi )\) and \(\phi (\mathcal {I}^\phi )\)) [20, 57, 68]. Since some details must be taken into account in the invariant formulation, this formulation is explained in more detail below. The coordinate formulation is obtained analogously with the corresponding simplifications.
3.1.1 Free energy
The free energy is additively decomposed into an equilibrium part \(\psi ^{\text {eq}}(\varvec{\upvarepsilon })\) and a non-equilibrium part \(\psi ^{\text {ov}}(\varvec{\upvarepsilon }, {\textbf {q}})\), i.e., \(\psi (\varvec{\upvarepsilon }, {\textbf {q}}) =\ \psi ^{\text {eq}}(\varvec{\upvarepsilon })\ + \psi ^{\text {ov}}(\varvec{\upvarepsilon }, {\textbf {q}})\). For the non-equilibrium part it is assumed, that the dependency on \((\varvec{\upvarepsilon }, {\textbf {q}})\) can be expressed as \(\psi ^{\text {ov}}(\varvec{\upvarepsilon }, {\textbf {q}}) =\ \psi ^{\text {ov}}({\textbf {p}}(\varvec{\upvarepsilon }, {\textbf {q}}))\), where \({\textbf {p}}= \varvec{\upvarepsilon }- {\textbf {q}}\) [39]. The free energy is formulated with the following assumptions:
-
(i)
\(\psi \) is an isotropic tensor function.
-
(ii)
In the initial state \(\varvec{\upvarepsilon }={\textbf {q}}=\varvec{0}\), the free energy vanishes, i.e., \(\psi (\varvec{0}, \varvec{0})=0\).
-
(iii)
In the initial state, the stress vanishes, i.e., \(\partial _{\varvec{\upvarepsilon }}\psi (\varvec{0}, \varvec{0})=\varvec{0}\).
-
(iv)
In the initial state, the internal force vanishes, i.e.,
\(-\partial _{{\textbf {q}}}\psi (\varvec{0}, \varvec{0})=\varvec{0}\).
-
(v)
\(\psi (\varvec{\upvarepsilon }, {\textbf {q}})\) is assumed to be convex in \(\varvec{\upvarepsilon }\) and \({\textbf {q}}\).
Condition (v) is used for several reasons. On the one hand, the convexity of \(\psi \) results in the positive definiteness of the material tangent \(\partial _{\varvec{\upvarepsilon }\varvec{\upvarepsilon }}\psi \), which offers numerical advantages in the application [69]. On the other hand, the resulting restrictions simplify the training of the model. Furthermore, (v) is fulfilled for the used reference model from Sect. 5.1.1. Furterhmore, it should be noted that other works assume the convexity of \(\psi \) as well [39, 70].
These constraints have to be incorporated into the FNN representation of \(\psi \), where \(\psi ^{\text {eq}}\) and \(\psi ^{\text {ov}}\) are each described by an FNN. These FNNs are denoted as \({\tilde{\psi }}^{\text {eq}}\) and \({\tilde{\psi }}^{\text {ov}}\), respectively. Requirements (i)–(v) can be satisfied as follows:
-
(i)
In order to ensure an isotropic free energy, \(\psi ^{\text {eq}}\) and \(\psi ^{\text {ov}}\) are expressed in terms of invariants of their tensor arguments [20], summarized in the invariant set \(\mathcal {I}^{\psi ^{\text {eq}}} =\ ( I^{\psi ^{\text {eq}}}_\alpha )_{\alpha =1}^3\) and \(\mathcal {I}^{\psi ^{\text {ov}}}=\ ( I^{\psi ^{\text {ov}}}_\alpha )_{\alpha =1}^3\), resp. That is, \(\psi ^{\text {eq}}(\varvec{\upvarepsilon })\) becomes \(\psi ^{\text {eq}}(\mathcal {I}^{\psi ^{\text {eq}}})\) and \(\psi ^{\text {ov}}({\textbf {p}})\) becomes \(\psi ^{\text {ov}}(\mathcal {I}^{\psi ^{\text {ov}}})\) with
$$\begin{aligned} \mathcal {I}^{\psi ^{\text {eq}}}(\varvec{\upvarepsilon })&= (\,{{\,\textrm{tr}\,}}{\varvec{\upvarepsilon }},\, {{\,\textrm{tr}\,}}{\varvec{\upvarepsilon }^2},\, {{\,\textrm{tr}\,}}{\varvec{\upvarepsilon }^4}\,) \end{aligned}$$(5)$$\begin{aligned} \mathcal {I}^{\psi ^{\text {ov}}}({\textbf {p}})&= (\,{{\,\textrm{tr}\,}}{{\textbf {p}}},\, {{\,\textrm{tr}\,}}{{{\textbf {p}}}^2},\, {{\,\textrm{tr}\,}}{{{\textbf {p}}}^4}\,) \quad . \end{aligned}$$(6)The specific choice of the invariant sets is explained in (v), where the convexity properties are discussed.
-
(ii)
In order to ensure \(\psi (\varvec{0},\varvec{0})=0\), each part \(\psi ^{\text {eq}}(\mathcal {I}^{\psi ^{\text {eq}}}(\varvec{\upvarepsilon }))\) and \(\psi ^{\text {ov}}(\mathcal {I}^{\psi ^{\text {ov}}}({\textbf {p}}))\) is set to zero in the initial state, i.e., \(\psi ^{\text {eq}}(\mathcal {I}^{\psi ^{\text {eq}}}(\varvec{0}))=0\) and \(\psi ^{\text {ov}}(\mathcal {I}^{\psi ^{\text {ov}}}(\varvec{0}))=0\). These requirements are not incorporated into the network functions themselves. Instead, a correction term is appended to the definitions of \(\psi ^{\text {eq}}(\mathcal {I}^{\psi ^{\text {eq}}})\) and \(\psi ^{\text {ov}}(\mathcal {I}^{\psi ^{\text {ov}}})\), that is, \(\psi ^{\text {eq}}(\mathcal {I}^{\psi ^{\text {eq}}})=\ {\tilde{\psi }}^{\text {eq}}(\mathcal {I}^{\psi ^{\text {eq}}}) -\ \psi ^{\text {eq}}_0\) and \(\psi ^{\text {ov}}(\mathcal {I}^{\psi ^{\text {ov}}})=\ {\tilde{\psi }}^{\text {ov}}(\mathcal {I}^{\psi ^{\text {ov}}}) - \psi ^{\text {ov}}_0\), where \(\psi ^{\text {eq}}_0=\ {\tilde{\psi }}^{\text {eq}}(\mathcal {I}^{\psi ^{\text {eq}}}(\varvec{\upvarepsilon }=\varvec{0}))\) and \(\psi ^{\text {eq}}_0=\ {\tilde{\psi }}^{\text {eq}}(\mathcal {I}^{\psi ^{\text {ov}}}({\textbf {p}}=\varvec{0}))\).
-
(iii)
For the stress to vanish in the initial state, the equilibrium stress \(\partial _{\varvec{\upvarepsilon }}\psi ^{\text {eq}}\) and the overstress \(\partial _{\varvec{\upvarepsilon }}\psi ^{\text {ov}}\) must vanish. This is enforced with another set of correction terms \(\psi ^{\text {eq}}_{\varvec{\upsigma }}\) and \(\psi ^{\text {ov}}_{\varvec{\upsigma }}\), such that \(\psi ^{\text {eq}}(\mathcal {I}^{\psi ^{\text {eq}}})=\ {\tilde{\psi }}^{\text {eq}}(\mathcal {I}^{\psi ^{\text {eq}}}) -\ \psi ^{\text {eq}}_0-\ \psi ^{\text {eq}}_{\varvec{\upsigma }}(\mathcal {I}^{\psi ^{\text {eq}}})\) and \(\psi ^{\text {ov}}(\mathcal {I}^{\psi ^{\text {ov}}})=\ {\tilde{\psi }}^{\text {ov}}(\mathcal {I}^{\psi ^{\text {ov}}}) -\ \psi ^{\text {ov}}_0-\ \psi ^{\text {ov}}_{\varvec{\upsigma }}(\mathcal {I}^{\psi ^{\text {ov}}})\) [20, 24, 56]. These correction terms are defined such that, when differentiated with respect to \(\varvec{\upvarepsilon }\), they compensate the error in the initial stress prediction by \({\tilde{\psi }}^{\text {eq}}(\mathcal {I}^{\psi ^{\text {eq}}})\) or \({\tilde{\psi }}^{\text {ov}}(\mathcal {I}^{\psi ^{\text {ov}}})\), respectively. A possible, but inconvenient way to achieve this is to set
(7)and analogous for \(\psi ^{\text {ov}}\). However, this leads to a free energy that depends on not only the invariants, but also on the strain tensor itself, which may violate (i). Therefore, \(\psi ^{\text {eq}}_{\varvec{\upsigma }}(\mathcal {I}^{\psi ^{\text {eq}}})\) and \(\psi ^{\text {ov}}_{\varvec{\upsigma }}(\mathcal {I}^{\psi ^{\text {ov}}})\) may only depend on the invariants. This is achieved with
$$\begin{aligned} \psi ^{\text {eq}}_{\varvec{\upsigma }}(\mathcal {I}^{\psi ^{\text {eq}}})= & {} \frac{\partial {\tilde{\psi }}^{\text {eq}}}{\partial I^{\psi ^{\text {eq}}}_1} \Bigg |_{\varvec{\upvarepsilon }=\varvec{0}} I^{\psi ^{\text {eq}}}_{1} \quad \text {and} \end{aligned}$$(8)$$\begin{aligned} \psi ^{\text {ov}}_{\varvec{\upsigma }}(\mathcal {I}^{\psi ^{\text {ov}}})= & {} \frac{\partial {\tilde{\psi }}^{\text {ov}}}{\partial I^{\psi ^{\text {ov}}}_1} \Bigg |_{{\textbf {p}}=\varvec{0}} I^{\psi ^{\text {ov}}}_{1}\;. \end{aligned}$$(9)Since \(I^{\psi ^{\text {eq}}}_1={{\,\textrm{tr}\,}}{\varvec{\upvarepsilon }}\) and \(I^{\psi ^{\text {ov}}}_1={{\,\textrm{tr}\,}}{{\textbf {p}}}\) are linear functions in \(\varvec{\upvarepsilon }\) and \({\textbf {p}}\), this does not effect convexity of \(\psi \).
-
(iv)
The internal force vanishes, if \(-\partial _{{\textbf {q}}}\psi ^{\text {ov}} \Bigg |_{{\textbf {p}}=\varvec{0}} = \varvec{0}\). Since
\(-\partial _{{\textbf {q}}}\psi ^{\text {ov}}=\partial _{\varvec{\upvarepsilon }}\psi ^{\text {ov}}\) holds and \(\partial _{\varvec{\upvarepsilon }}\psi ^{\text {ov}} \Bigg |_{{\textbf {p}}=\varvec{0}}=\varvec{0}\) is already assured with (iii), this requirement is already secured.
-
(v)
Convexity in \(\varvec{\upvarepsilon }\) and \({\textbf {q}}\) can be achieved if two conditions are met: (a) The chosen invariants are convex with respect to \(\varvec{\upvarepsilon }\) and \({\textbf {q}}\) and (b) \({\tilde{\psi }}^{\text {eq}}\) and \({\tilde{\psi }}^{\text {ov}}\) are convex and non-decreasing in those invariants [29]. Since both \(\psi ^{\text {eq}}(\varvec{\upvarepsilon })\) and \(\psi ^{\text {ov}}({\textbf {p}})\) depend on a single symmetric tensor of rank two, a complete set of invariants composes, e.g., the invariants \(({{\,\textrm{tr}\,}}{{\textbf {S}}},\, {{\,\textrm{tr}\,}}{{\textbf {S}}^2}, \, {{\,\textrm{tr}\,}}{{\textbf {S}}^3})\), where \({\textbf {S}}\in \mathcal {S}\mathcalligra {ym}_2\). The third invariant \({{\,\textrm{tr}\,}}{{\textbf {S}}^3}\), however, is not convex in \({\textbf {S}}\), as shown in App. B. Therefore, the functional basis is adjusted suitably by replacing \({{\,\textrm{tr}\,}}{{\textbf {S}}^3}\) by the convex invariant \({{\,\textrm{tr}\,}}{{\textbf {S}}^4}\). This is allowed, since \({{\,\textrm{tr}\,}}{{\textbf {S}}^3}\) can be expressed by \({{\,\textrm{tr}\,}}{{\textbf {S}}}\), \({{\,\textrm{tr}\,}}{{\textbf {S}}^2}\) and \({{\,\textrm{tr}\,}}{{\textbf {S}}^4}\) using the Cayley-Hamilton theorem and the original basis can be recovered. A comment on the convexity of the used invariants can be found in Appendix B. Consequently, the invariant bases in Eqs. (5) and (6) allow for the construction of convex functions using non-decreasing fully input convex neural networks(FICNNs)Footnote 2 [20, 29, 49, 71] as described in App. A.1.1.
Summarizing, the free energy is modeled with an additive decomposition, where each part is a non-decreasing FICNN with additional correction terms and is expressed in a set of three convex invariants, such that
With that, the free energy representation is physically meaningful, even in the untrained state of the NN model. Details on the hyperparameters of the FICNNs can be found in Sec. A.3 To simplify notation, the weights and biases of the two FICNNs are summarized in a single parameter set \(\varvec{{\mathcalligra {w}}}^\psi \).
The dissipation potential is constructed using similar techniques.
3.1.2 Dissipation potential
The NN-based dissipation potential depends on \(({\dot{{\textbf {q}}}}, \varvec{\upvarepsilon }, {\textbf {q}})\) [17]. Similar to the non-equilibrium part of the free energy, the dependency of \(\varvec{\upvarepsilon }\) and \({\textbf {q}}\) is expressed in terms of \({\textbf {p}}=\varvec{\upvarepsilon }-{\textbf {q}}\). Consequently, \(\phi = \phi ({\dot{{\textbf {q}}}}, {\textbf {p}}(\varvec{\upvarepsilon }, {\textbf {q}}))\). As discussed in Sect. 2, the dissipation potential must obey some constraints:
-
(a)
\(\phi \) is an isotropic tensor function.
-
(b)
\(\phi \) is convex with respect to its first argument, the rate of the internal variable \({\dot{{\textbf {q}}}}\).
-
(c)
If the internal variable does not evolve, the dissipation potential vanishes, i.e., \(\phi (\varvec{0},{\textbf {p}})=0\).
-
(d)
If the inelastic strain does not evolve, the dissipation potential is in a minimum, i.e., \(\partial _{{\dot{{\textbf {q}}}}}\phi (\varvec{0},{\textbf {p}}) = \varvec{0}\).
The respective network to model this function is denoted as \({\tilde{\phi }}\). The stated requirements are enforced with techniques similar to those used to construct the free energy.
-
(a)
Formulating \(\phi \) in invariants enforces isotropy of the resulting function. The network \({\tilde{\phi }}(\mathcal {I}^{\phi }({\dot{{\textbf {q}}}}, {\textbf {p}}))\) depends on the six invariants
$$\begin{aligned} \mathcal {I}^{\phi }({\dot{{\textbf {q}}}}, {\textbf {p}}) = (\,{{\,\textrm{tr}\,}}{\dot{{\textbf {q}}}},\, {{{\,\textrm{tr}\,}}{\dot{{\textbf {q}}}}}^2,\, {{{\,\textrm{tr}\,}}{\dot{{\textbf {q}}}}}^4,\, {{\,\textrm{tr}\,}}{\textbf {p}},\, {{{\,\textrm{tr}\,}}{\textbf {p}}}^2,\, {{{\,\textrm{tr}\,}}{\textbf {p}}}^3\,). \end{aligned}$$(13)Here, no mixed invariants between \({\dot{{\textbf {q}}}}\) and \({\textbf {p}}\) are used.
-
(b)
To enforce convexity with respect to the rate of the internal variable, a non-decreasing partially input convex neural network (PICNN) [49, 50] is used, which is convex with respect to the three invariants \((\,{{\,\textrm{tr}\,}}{\dot{{\textbf {q}}}},\, {{{\,\textrm{tr}\,}}{\dot{{\textbf {q}}}}}^2,\, {{{\,\textrm{tr}\,}}{\dot{{\textbf {q}}}}}^4\,)\) and not necessarily convex with respect to \((\,{{\,\textrm{tr}\,}}{\textbf {p}},\, {{{\,\textrm{tr}\,}}{\textbf {p}}}^2,\, {{{\,\textrm{tr}\,}}{\textbf {p}}}^3\,)\).
-
(c)
Analogous to the free energy, a correction term \(\phi _{0}\) is appended to the definition of \(\phi \), such that \(\phi ={\tilde{\phi }}- \phi _{0}\). \(\phi _{0}\) compensates the offset in \({\tilde{\phi }}\) and is defined as \(\phi _{0}={\tilde{\phi }}(\mathcal {I}^{\phi }({\dot{{\textbf {q}}}}=\varvec{0}, {\textbf {p}}))\).
-
(d)
By adapting the stress correction term for the free energy, a correction term \(\phi _{\varvec{\uptau }}\) is appended, such that \(\phi = {\tilde{\phi }}- \phi _{0}- \phi _{\varvec{\uptau }}\), where \(\phi _{\varvec{\uptau }}\) is
$$\begin{aligned} \phi _{\varvec{\uptau }}= \frac{\partial {\tilde{\phi }}}{\partial I^\phi _1} \Bigg |_{{\dot{{\textbf {q}}}}=\varvec{0}} I_1^{\phi }. \end{aligned}$$(14)
Summarizing, the dissipation potential is modeled with a non-decreasing PICNN and additional correction terms. The dependency on \({\dot{{\textbf {q}}}}\) is expressed with convex invariants of \({\dot{{\textbf {q}}}}\). That is,
The NN model is thus thermodynamically consistent and isotropic by construction. All weights and biases of the PICNN with the hyperparamters from Sec. A.3 are summarized in the parameter set \(\varvec{{\mathcalligra {w}}}^\phi \).
3.1.3 Coordinate formulation
The previous explanations refer to the invariant formulation. For the formulation with coordinates [39], requirements (i) and (a), i.e., the isotropy of the potentials, cannot be fulfilled. In principle, the same techniques are used to comply with the remaining requirements, but instead of convex and non-decreasing NNs, only convex NNs are used, as the tensors themselves are inputs of the networks. The coordinate formulations of the additional terms for zero energy and zero stress in the initial state are briefly explained using the example of \(\psi ^{\text {eq}}\): The term \(\psi ^{\text {eq}}_0\) for zero energy in the initial state is calculated as
and the term \(\psi ^{\text {eq}}_{\varvec{\upsigma }}\) for the stress free initial state becomes
The other additional terms for \(\psi ^{\text {ov}}\) and \(\phi \) are calculated analogously.
3.2 Incremental predictions with the trained models
Once the weights and biases of the networks for free energy and dissipation potential have been adapted, the constitutive model can be used to predict the constitutive response to a prescribed strain path consisting of multiple discrete time steps. In order to solve the evolution equation numerically, an implicit Euler scheme is applied for the temporal discretization, in which the rate of the internal variable is approximated as \({}^{n}{{\dot{{\textbf {q}}}}}\approx ({}^{n}{{\textbf {q}}}-{}^{n-1}{{\textbf {q}}})/{}^{n}{\Delta t}\), where the superscript n corresponds to the n-th time step of a sequence and \({}^{n}{\Delta t} = {}^{n}{t} - {}^{n-1}{t}\) is the time increment. This ensures stability of the evolution of \({\textbf {q}}\). In each time step, the rate of the internal variable is adapted iteratively, such that Eq. (4) is fulfilled up to a prescribed tolerance. That is, the resulting internal force, which is calculated from the free energy, i.e., \(\varvec{\uptau }=-\partial _{{\textbf {q}}}\psi \), and the internal force calculated from the dissipation potential, i.e., \(\hat{\varvec{\uptau }}=\partial _{{\dot{{\textbf {q}}}}}\phi \), are supposed to be identical within a given tolerance. This iterative solution of Eq. (4) is done using a Newton–Raphson scheme, as depicted in Fig. 2.
4 Training methods
Since the internal variable and its rate are arguments of the free energy and the dissipation potential, they must be provided in the training process to render training possible. In the following, however, it is assumed, that they are not present in the training data set. For data obtained through homogenization, a few approaches have been developed to enable the determination of internal variables using autoencoders [54, 55]. The training data set \(\mathscr {D}=(\varvec{\upvarepsilon }(t), \varvec{\upsigma }(t), {\textbf {q}}(t))\) then comprises sequences of stresses, strains and internal variable(s). In real experiments, however, the determination of internal variables is in general not possible, shrinking the data set \(\mathscr {D}=(\varvec{\upvarepsilon }(t), \varvec{\upsigma }(t))\) to sequences of only stresses and strains. This section introduces three methods to address this issue. Two methods from the literature determine the internal variable by integrating the evolution equation over the entire sequence or by means of an additional auxiliary network in the form of an FNN. In the newly developed third method, the auxiliary network is replaced by an RNN.
4.1 Integration
The training approach denoted as integration in the following is probably the most intuitive among the three presented and has been applied in [37, 57]. In every training epoch, the stress response to individual sequences is determined following the scheme provided in Figs. 2 and 3. Starting from the initial state \({}^{0}\varvec{\upvarepsilon }={}^{0}\varvec{\upsigma }={}^{0}{\textbf {q}}=\varvec{0}\), the stress \({}^{1}\varvec{\upsigma }\) for the new strain \({}^{1}\varvec{\upvarepsilon }\) and the corresponding time step \({}^{1}\Delta t\) is calculated. The obtained internal variable is then passed on to the next time step and so forth until the last time step of the sequence is reached. The stresses \(\varvec{\upsigma }\) determined this way are compared to the expected stresses \({{\bar{\varvec{\upsigma }}}}\) from the training data set using the loss function
where \({{\,\textrm{MAE}\,}}(\varvec{\upsigma }, {{\bar{\varvec{\upsigma }}}})\) is the mean absolute error between predicted stress \(\varvec{\upsigma }\) and expected stress \({{\bar{\varvec{\upsigma }}}}\). Compared to the mean squared error, the mean absolute error performed best for the problems in this article and is therefore used within the scope of this work. Furthermore, \(s_{\varvec{\upsigma }}\) is the normalization factor for the stress, that normalizes the loss to magnitude 1. The calculation of \(s_{\varvec{\upsigma }}\) is explained in A.2. To adapt the weights and biases of the networks, the loss function Eq. (18) is minimized in the optimization problem
where the parameters \(\varvec{{\mathcalligra {w}}}^\psi \) and \(\varvec{{\mathcalligra {w}}}^\phi \) correspond to the weights and biases of the two FICNNs of the free energy and the PICNN of the dissipation potential. They are constrained by the sets \(\mathcal C^\psi \) and \({\mathcal {C}}^\phi \) containing restrictions for weights and biases of these ICNNs as described in App. A. Due to the iterative solution scheme, the evaluation of the loss function is very expensive [37]. Hence, an optimizer with fast convergence is favorable. The optimizer Sequential Least Squares Programming (SLSQP) has shown to perform well for this application [17, 20, 26, 72] and is therefore used in the numerical examples in Sect. 5.
4.2 Auxiliary feedforward network
To avoid the challenges of the integration method, another method has been proposed by As’ad and Farhat [39] and is adopted with slight modifications herein [17]. This method introduces an additional FNN, that is supposed to learn the temporal course of the internal variable. Let \({{\tilde{{\textbf {q}}}}}(t)\) denote this additional FNN. It depends on only a single input, the time t, and has six outputs corresponding to the six independent entries of the symmetric tensor \({\textbf {q}}\), as shown in Fig. 4. For the training of the networks for free energy and dissipation potential, values for the internal variable and its rate are taken from this network, where the rate is approximated as \({}^{n}{\dot{{\textbf {q}}}}\approx ({}^{n}{\textbf {q}}-{}^{n-1}{\textbf {q}}) / {}^{n}\Delta t\) to be consistent with the prediction process. The weights and biases of \({{\tilde{{\textbf {q}}}}}\) are adapted such that the predicted temporal course of \({{\tilde{{\textbf {q}}}}}(t)\) allows the loss function
to become as small as possible. The loss function consists of a term \(\mathcal {L}^{\varvec{\upsigma }}\) for the accurate stress prediction and another \(\mathcal {L}^{\text {Biot}}\) term to comply with the Biot relation Eq. (4) [17]. Thus, the reformulated optimization problem reads
where the parameter set \(\varvec{{\mathcalligra {w}}}^q\) contains weights and biases of the auxiliary FNN as specified in Sec. A.3. This leads to a reasonable representation of the internal variable without explicitly specifying its value in advance or having to integrate every sequence in every iteration of training. However, we have found that starting the optimization with randomly initialized weights for \({{\tilde{{\textbf {q}}}}}\) makes the problem too complex for common optimizers and does not lead to the desired results. In order to facilitate the training, the auxiliary network is pre-trained using the strain \(\varvec{\upvarepsilon }(t)\) as an initial guess for \({{\tilde{{\textbf {q}}}}}(t)\). Once the pre-training and subsequent actual training is finished, the auxiliary network is no longer necessary and the prediction process can be carried out using only free energy and dissipation potential.
Remark 1
With this architecture, the temporal course of the internal variable for a single sequence can be modeled via a single FNN. However, if data from several sequences are available in the data set, another FNN must be added for each additional sequence, making the optimization problem increasingly difficult.
4.3 Auxiliary recurrent network
RNNs have proven to be particularly suitable for describing the behavior of path-dependent systems [10, 12, 14, 51]. Therefore, the auxiliary FNN is now to be replaced by an RNN. Using an RNN cell to generate the internal variable allows to use multiple sequences for training without requiring a new auxiliary network for every training sequence. We use an LSTM cell [18] as RNN cell. This LSTM cell has two different state vectors, that allow the cell to store information from past time steps. These state vectors are the hidden state \(\varvec{\mathcalligra {h}}\) and the cell state \(\varvec{\mathcalligra {c}}\), where \(\varvec{\mathcalligra {h}}\) is the output of the cell for every time step. The training with the auxiliary RNN works as shown in Fig. 5: In each time step, the RNN cell receives the current strain \({}^{n}\varvec{\upvarepsilon }\), the associated stress \({}^{n}\varvec{\upsigma }\) from the data set as well as the time step \({}^{n}\Delta t\). An FNN reduces the output of the RNN cell, the hidden state \({}^{n}\varvec{\mathcalligra {h}}\), to 6 entries corresponding to the 6 independent entries of \({}^{n}{\textbf {q}}\). The rate of the internal variable \({}^{n}{\dot{{\textbf {q}}}}\) is calculated using \({}^{n}\Delta t\) and \({}^{n-1}{\textbf {q}}\) from the previous time step. Through differentiation of \(\psi ({}^{n}\varvec{\upvarepsilon }, {}^{n}{\textbf {q}})\) and \(\phi ({}^{n}{\dot{{\textbf {q}}}}, {}^{n}\varvec{\upvarepsilon }, {}^{n}{\textbf {q}})\), the stress \({}^{n}\varvec{\upsigma }\) and the internal forces \({}^{n}\varvec{\uptau }\) and \({}^{n}\hat{\varvec{\uptau }}\) are obtained. The status of the LSTM cell, i.e., \({}^{n}\varvec{\mathcalligra {h}}\) and \({}^{n}\varvec{\mathcalligra {c}}\), is passed on to the next time step, to obtain the next material state and so on. Using the calculated stresses and internal forces, the loss function
is evaluated. This loss function again contains a term for the error in the stress prediction and a term for compliance with the Biot equation. The optimization problem thus reads
where the weights and biases of the auxiliary LSTM and the connected FNN are summarized in the parameter set \(\varvec{\mathcalligra {w}}^q\). The concrete architecture of the LSTM and FNN can be found in Sec. A.3. Finally, after training, the RNN and the attached FNN can be discarded, leaving \(\psi \) and \(\phi \) as constitutive model.
5 Numerical examples
In the following, the presented framework is examined comprehensively. First, the reference material is introduced and several data sets are generated using this reference model. These data sets differ in terms of the material states they contain, for example multiaxial stress state or uniaxial stress states, and the presence of noise in the stress data. The data sets are then used as the basis for calibrating the model in different scenarios. An overview of the conducted experiments is given in Table 1.
5.1 Data base
5.1.1 Reference material
To generate the training data, a reference model consisting of a spring and a parallel Maxwell element as depicted in Fig. 6 is used. The free energy and the dissipation potential for such a model are defined as
where \({\dot{\varvec{\upvarepsilon }}}^{\text {in}}\) is defined to be the internal variable, \(\varvec{\upvarepsilon }^{\text {el}} = \varvec{\upvarepsilon }- \varvec{\upvarepsilon }^{\text {in}}\),
are the equilibrium and non-equilibrium stiffness tensors, respectively, and \({\mathbb {V}}(\varvec{\upvarepsilon }^{\text {in}}, \varvec{\upvarepsilon })\) is a positive semidefinite fourth order tensor describing the viscous properties of the material. This tensor is not constant, but depends on the overstress via
This ansatz is motivated from [73]. The specific material parameters are given in Table 2.
5.1.2 Generation of the data base
The data base \(\mathscr {D}=\{\,\varvec{\upvarepsilon }(t), \varvec{\upsigma }(t)\,\}\) contains information about the temporal course of strain and stress for a single or multiple sequences. The generation of such sequences is explained in the subsequent paragraphs, where a distinction is made between training data with multi-, bi- and uniaxial stress states and test data to evaluate the prediction quality of trained models. To indicate the different data sets, a superscript and a subscript is appended to the symbol for the data set that describe the type of data and the number of sequences and timesteps. For example, the data set \(\mathscr {D}^{\text {multiax}}_{1\times 200}\) contains ideal multiaxial data in a single sequence of 200 timesteps.
Multiaxial training data Multiaxial training data sequences are obtained by prescribing a randomized strain path \(\varvec{\upvarepsilon }(t)\) and evaluating the respective stress response of the reference constitutive model from Sect. 5.1.1. A possible randomized strain path \(\varvec{\upvarepsilon }(t)\) is shown in Fig. 7. This strain path is created with cubic splines that connect a set of randomly sampled knots [17, 39]. Starting from \(\varepsilon _{ij}^{\text {knot}}=0\) and \(t^{\text {knot}}=0\), a time increment \(\Delta t^{\text {knot}}\) is sampled from a uniform distribution between \(\Delta t_{\text {min}}^{\text {knot}}= {0.2} \textrm{s}\) and \(\Delta t_{\text {max}}^{\text {knot}}={1} \textrm{s}\). The increments of the six independent coordinates of the strain tensor are sampled from a normal distribution with standard deviation \(s_{\Delta \varepsilon }^{\text {knot}}={0.5} \textrm{s}\) around mean 0. If the resulting absolute value of the strain \(|\varepsilon _{ij}^{\text {knot}}+\Delta \varepsilon _{ij}^{\text {knot}}|\) exceeds \({2} \textrm{s}\), the strain increment \(\Delta \varepsilon ^{\text {knot}}\) is sampled again. Once this strain path \(\varvec{\upvarepsilon }(t)\) is generated, it is applied to the reference material with random time increments \(\Delta t\) from a uniform distribution between \(\Delta t_{\text {min}}={0.03}\textrm{s}\) and \(\Delta t_{\text {max}}={0.07} \textrm{s}\) to obtain the corresponding stresses \(\varvec{\upsigma }(t)\). The symbol \(\mathscr {D}^{\text {multiax}}_{1\times 200}\) indicates such a data set with ideal multiaxial data.
Multiaxial training data with noise To investigate the performance of the training methods with non-ideal data, another training data set with noisy stress data is generated. This data set reuses the ideal multiaxial training data and modifies the stress data therein. To receive the noisy stress data, Gaussian noise is added to the ideal stress, such that \({{{\tilde{\sigma }}}}_{ij}=\sigma _{ij}+\Delta \sigma _{ij}\), where \(\Delta \sigma _{ij}\) is sampled from a normal distribution with standard deviation \(s^{\Delta \sigma }={1.5}\textrm{MPa}\) and mean 0. The ideal and noisy data are shown in Fig. 7. To indicate the noise in the stress data, the symbol \({{\tilde{\mathscr {D}}}}^{\text {multiax}}_{1\times 200}\) for this data set has an additional tilde.
Plane strain training data To generate the plane strain data set, the strain paths that were used to generate the multiaxial data set are reused. These strain paths are modified so that the corresponding coordinates are set to zero, i.e. \(\varepsilon _{13}(t)=\varepsilon _{23}(t)=\varepsilon _{33}(t)=0\). The stress response to these strain paths is calculated and the data set is assembled. Plane strain data sets are denoted as \(\mathscr {D}^{\text {plane}}_{1\times 200}\), for example.
Biaxial training data In order to analyze the applicability of the training methods with potential experimental data, a data set with data from a simulated stress driven biaxial tension test is generated. In this biaxial tension test, the temporal course of the stress coordinates \(\sigma _{11}(t)\ne \sigma _{22}(t)\) is prescribed and in general non-zero, while all other coordinates are \(\sigma _{ij}(t)=0\). The path of the non-zero stress coordinates is generated randomly in the same way as for the strain components in the multiaxial training data set. That is, cubic splines are used that connect randomly sampled knots with positions \((t, \sigma ^{\text {knot}}_{ij})\). These random knots are generated with the time increments from a uniform distribution with \(\Delta t_{\text {min}}^{\text {knot}}={0.2} \textrm{s}\) and \(\Delta t_{\text {max}}^{\text {knot}}={1} \textrm{s}\), a standard deviation of the stress increments of \(s_{\Delta \sigma }={12.5}\textrm{MPa}\) and a maximum absolute stress value for the knots of \({25}\textrm{MPa}\). The stress path described by this cubic spline is evaluated with time increments \(\Delta t\) from a uniform distribution within \(\Delta t_{\text {min}}={0.03} \textrm{s}\) and \(\Delta t_{\text {max}}={0.07} \textrm{s}\) and the corresponding strain \(\varvec{\upvarepsilon }(t)\) is determined iteratively, such that the strain \(\varvec{\upvarepsilon }(t)\) results in the prescribed stress path \(\varvec{\upsigma }(t)\). A biaxial data set is indicated with \(\mathscr {D}^{\text {biax}}_{1\times 200}\).
Uniaxial training data To simulate an uniaxial tension test, the course of the stress coordinate \(\sigma _{11}(t)\) is prescribed and non-zero, while all other coordinates are \(\sigma _{ij}(t)=0\). The same technique as for the biaxial data set is used, i.e., cubic splines are generated for \(\sigma _{11}(t)\). Here, the parameters \(\Delta t_{\text {min}}^{\text {knot}}={0.2} \textrm{s}\) and \(\Delta t_{\text {max}}^{\text {knot}}={1} \textrm{s}\) for the limits of \(\Delta t^{\text {knot}}\), the standard deviation \(s_{\Delta \sigma }={25}\textrm{MPa}\) and the maximum absolute value \({50}\textrm{MPa}\) for the stress are used. Again, the spline is scanned with \(\Delta t_{\text {min}}={0.03} \textrm{s}\) and \(\Delta t_{\text {max}}={0.07} \textrm{s}\) to obtain the strain \(\varvec{\upvarepsilon }(t)\). Uniaxial training data is denoted with \(\mathscr {D}^{\text {uniax}}_{1\times 200}\).
Test data Following the training, the models are tested with an unseen path, that is not part of the training data set. For this purpose, a strain path \(\varvec{\upvarepsilon }(t)\) is generated using the method for the generation of multiaxial training data. The corresponding stress response of the reference material is calculated. That is, the trained models are tested on an arbitrary multiaxial test path where all stress and strain coordinates are generally non-zero, i.e., \(\varepsilon _{ij}(t)\ne 0\) and \(\sigma _{ij}(t)\ne 0\). This test sequence is used as a benchmark for all trained models, regardless of the data set used.
5.2 Training with multiaxial data
The presented NN-based constitutive model and training methods are now to be tested in different scenarios. In the first part of the subsection, it is exploited that the used data is generated synthetically and it is assumed that the internal variable is given in the data set to compare the invariant formulation and the coordinate formulation of the potentials in a simple test case, i.e., \(\mathscr {D}=(\varvec{\upvarepsilon }(t), \varvec{\upsigma }(t), {\textbf {q}}(t))\) additionally contains information about the internal variable. Afterwards, the internal variable is removed from the data set and the three training methods are compared for the invariant formulation, using both ideal and noisy stress data.
5.2.1 Comparison of invariant vs. coordinate formulation
To compare the model formulation using invariants and the formulation using tensor coordinates as inputs of the potentials, a data set \(\mathscr {D}_{5\times 200}^{\text {multiax}}\) with 5 sequences à 200 time steps and the corresponding plane strain data set \(\mathscr {D}_{5\times 200}^{\text {plane}}\) are used. These data sets are sufficiently large to show the full potential of both models and contain the internal variable. The training is carried out using the Adam optimizer with 20, 000 epochs, an initial learning rate of 0.01 and an exponential learning rate decay such that the learning rate is multiplied with 0.1 every 4, 000 epochs. To reduce the effect of the random weights initialization, both models are trained 25 times for each data set. The models with the lowest final value of the loss function are then tested on the unseen test strain path. The results are shown as scatter diagrams in Fig. 8a.
For the coordinate formulation, it can be seen that the model is able to make fairly accurate predictions for the test path based on data set \(\mathscr {D}_{5\times 200}^{\text {multiax}}\). Using data set \(\mathscr {D}_{5\times 200}^{\text {plane}}\), however, the model fails to extrapolate from the plane strain data to the full strain states. To illustrate this behavior, Fig. 8b shows the actual course of predicted stresses \(\sigma _{11}\), \(\sigma _{12}\) and \(\sigma _{13}\). It becomes evident that the large deviations from the expected values mainly concern the out-of-plane coordinates for which the training data set lacks sufficient data.
However, looking at the invariant formulation, it can be seen that it produces very accurate results regardless of the used data set, even exceeding the accuracy of the coordinate formulation with data set \(\mathscr {D}_{5\times 200}^{\text {uniax}}\). Thus, the invariant formulation enables extrapolation from plane strain data to full strain states without notable loss of accuracy. Such a well-developed extrapolation behavior also occurs with elastic NN models based on invariants [20, 26]. This benefit arises from the choice of a specific set of invariants that provides additional information about the anisotropy class of the underlying material law to the model. In fact, the change from plane strain to full strain states may not correspond to an extrapolation in the invariant space at all [28].
Moreover, the invariant formulation requires significantly less training data. The decision to use 5 sequences of 200 time steps was made due to the larger amount of data needed for the coordinate formulation, whereas the invariant formulation can operate on a single sequence of 200 steps with similar prediction accuracy.
For this reason, only the invariant formulation is used in the following studies and the data set is reduced to one sequence of length 200.
5.2.2 Comparison of the three training methods
The internal variable is now removed from the data set and has to be generated during the training process. To do so, the methods described in Sect. 4 are applied, where the data sets \(\mathscr {D}_{1\times 200}^{\text {multiax}}\) and its noisy equivalent \({{\tilde{\mathscr {D}}}}_{1\times 200}^{\text {multiax}}\) are used to compare the methods. Again, all models are trained 25 times and the weigths of the training run with the lowest final value of the loss function are chosen. The result graphs show scatter diagrams of the stress predictions for the unseen test path (Fig. 9), the stress predictions over time for the data set \({{\tilde{\mathscr {D}}}}_{1\times 200}^{\text {multiax}}\) (Fig. 10) and the learned course of the internal variable for the training sequence as used in the last epoch of training (Fig. 11). The history of the losses over iterations for the different methods are shown in Appendix C, exemplarily on the example of the data set \(\mathscr {D}_{1\times 200}^{\text {multiax}}\).
Integration
The integration training method yields highly accurate results for the ideal data set and outperforms the other methods in terms of accuracy. However, it should be noted that this method uses the SLSQP optimizer for computational reasons, which has been shown to provide better results for smaller networks than Adam [72]. Since no additional network is required for this method, SLSQP can be used in a computationally efficient manner with fewer function evaluations compared to Adam. Nevertheless, the evaluation of the loss function consumes the majority of computational time for training, since an iterative solution is required at each time step, regardless of the optimizer, see Fig. 3. Adding another sequence to the data set or doubling its length roughly doubles the cost for the evaluation of the loss function, making it more and more inefficient. However, very good results can be achieved with both the ideal training data as well as the noisy training data set as Figs. 9 and 10 show.
Auxiliary feedforward network
Although this training method shows the least accurate predictions for the training with ideal data, the results are still sufficient for both ideal as well as noisy data. In addition to that, the computational time per iteration is very low. However, before the actual training starts, the auxiliary FNN must be pre-trained to a reasonable initial guess for \({\textbf {q}}(t)\) to have good starting values for the weights and biases. Using randomly initialized parameters at the beginning of the actual training did not yield useful results in the numerical experiments shown here. We used \({\textbf {q}}(t)=\varvec{\upvarepsilon }(t)\) as initial guess to pre-train the FNN. Another major drawback of this method is the increasing demand of trainable parameters for a longer sequence or additional sequences in the training data set. Since a single FNN only predicts the temporal course of one sequence, every new sequence requires a new FNN, making the optimization more and more complex. For a single sequence of moderate length, however, the method can yield useful results, see Figs. 9 and 10.
Auxiliary recurrent network
The RNN as auxiliary network for the generation of the internal variable also provides precise results with both data sets. Compared to the integration method, the training is very fast and in contrast to the FNN it does not need a pre-training and the final value of the loss function is smaller, see Fig. 18 in the appendix.
Moreover, using a training data set with several sequences instead of only one sequence does not increase the number of trainable parameters, as the RNN implicitly learns how the internal variable evolves instead of memorizing its temporal course. Thus, we consider a test case which is more complex compared to the works [37, 39, 57], where only one path has been used for training. To do so, several sequences with multiaxial states are added to the training data set to obtain \(\mathscr {D}_{100\times 100}^{\text {multiax}}\) with 100 sequences. The prediction results for the architecture trained with the set \(\mathscr {D}_{100\times 100}^{\text {multiax}}\) are shown in Fig. 12. As can be seen, the stress response for the unseen strain path is even more precise, while the required time per training iteration remains short. Although the data set contains 50-times the number of time steps, the time per iteration increases from \({0.059} \textrm{s}\) to only \({0.112}\textrm{s}\). This shows that the RNN method allows to use multiple sequences efficiently without increasing amount of trainable parameters and within reasonable training time.
Summary
Each of the analyzed techniques exhibits certain advantages compared to the others. In order to provide a comprehensive summary of this comparison, the advantages and disadvantages of each approach are listed in Table 4. The integration training is very accurate. However, the biggest disadvantage of this method is the training time, which also affects the applicability for large data sets with many sequences, such that this combination is not practically manageable due to cost reasons. Using an FNN as an auxiliary network is extremely fast, but requires a pre-training of the FNN. It is also limited to one or very few sequences. In contrast, the RNN is fast, does not require pre-training, and can be applied to large data sets with multiple sequences without further adjustments. This feature is particularly relevant for real-world applications with experimental or homogenization data.
5.3 Training with biaxial data
So far, data sets were used that are difficult to obtain experimentally. In the following, data sets will be used for which experimental approaches are available. The models are trained with the RNN method. First, the data set \(\mathscr {D}^{\text {biax}}_{1\times 200}\) is used, which contains data from a virtual biaxial test in a single sequence with 200 time steps with ideal stress data. The model is again calibrated 25 times with the RNN method and the model with the lowest final loss value is kept and tested with the multiaxial test path. The results of the predictions for this test path are shown in Fig. 13. These show that the model is able to predict arbitrary multiaxial stress states despite the training data set being limited to biaxial data. The slight loss of accuracy compared to the multiaxial data set is due to the lacking coverage of the full invariant range in the training. However, with a suitable choice of load path for the training data, this problem can be minimized and the model can also be trained and applied with biaxial data using a suitable training method—in this case the RNN.
5.4 Training with uniaxial data
The previous paragraph shows that good prediction results can also be achieved with biaxial stress training data. We will proceed by assuming that only data from uniaxial stress testing is available. For this purpose, the data set \(\mathscr {D}^{\text {uniax}}_{1\times 200}\) is used, which contains synthetic data from a uniaxial test. This data set again contains 1 sequence with 200 time steps and ideal stress data. Compared to the data sets used so far, however, the maximum achievable stress and the rate of the stress are increased for this data set as described in Sect. 5.1.2 in order to compensate the unideal coverage of the invariant space. The model is calibrated 25 times using the RNN method. The prediction results of the best model are shown in the right plot in Fig. 13. This shows that the model can predict arbitrary multiaxial states without suffering from substantial limitations in accuracy, despite being restricted to uniaxial training data. However, it can also be seen that there are systematic deviations from the reference for large \(\sigma _{ij}\), which indicate that the model extrapolates strongly in the invariant space within this region.Footnote 3 Nevertheless, it can be stated that by restricting the anisotropy class and incorporating basic physical principles, both the number of data points and the complexity of stress states in the training data can be greatly reduced, which allows to use experimental methods to generate training data for the presented model.
6 Conclusions
This paper proposes a fast and widely applicable method to calibrate inelastic constitutive models without prior knowledge of the internal variables. This method is compared comprehensively with two existing methods. For this comparison, we propose and use a physics-augmented NN-based model for viscoelastic materials based on the concept of GSMs.
In the beginning of this work, the concept of GSMs is briefly described. Subsequently, the NN-based model is presented, which consists of two potentials: the free energy and the dissipation potential. These potentials are constrained in a physically meaningful way so that thermodynamic consistency is ensured in advance. The potentials can be expressed using either tensor coordinates or invariants of the tensor arguments. Following the presentation of the NN-based constitutive model, three training methods are introduced. These comprise a method that integrates entire sequences and two methods that use an additional NN, either an FNN or an RNN, to provide the internal variable. Based on these techniques, numerical experiments are conducted using synthetically generated data from classical constitutive models. First, it is assumed that the internal variable is present in the multiaxial and plane strain training data sets. Using these data sets, the invariant formulation is compared to the coordinate formulation. It is shown, that the invariant formulation requires less data, yields more accurate results and exhibits far better extrapolation capabilities when predicting stresses for arbitrary strain paths with plane strain training data. Henceforth, only the invariant formulation is used in the subsequent studies. In order to compare the three training methods, the internal variable is erased from the multiaxial data set and the performance of the methods is examined using only stress and strain data. It shows, that all three methods are able to provide models that yield accurate predictions for unseen data, even for a small data set with noisy stress data. However, only the proposed RNN model allows to achieve fast and accurate results with large data sets comprising many sequences and is therefore the method with the broadest applicability. Finally, the applicability of the framework with an invariant-based NN model and the RNN training method for biaxial and uniaxial stress training data is examined. We show, that a complete 3D model can be calibrated, that is capable of predicting arbitrary multiaxial stress states, although only bi- or uniaxial stress data is used for training. Thus, this paper proposes a flexible constitutive model for viscoelastic materials and an efficient method for calibrating such models without prior knowledge of the internal variable. The presented training method exceeds the application possibilities of existing approaches and can form the basis for future extensions in the context of NN-based data driven constitutive modeling.
However, this study also has limitations. Specifically, the presented algorithms assume that constitutive behavior can be modeled with a single internal variable expressed by a symmetric second-order tensor. While this assumption is valid for the synthetically generated data used herein, it may not be sufficient to generalize for more complex behavior. In this case, the number of internal variables can be increased stepwise until the desired accuracy is achieved. Therefore, possible future studies include the implementation and testing of such algorithms, the application to elastoplastic materials [27, 32], extension to viscoelasticity at finite strains [39] as well as the use of experimental [38] or homogenization data [26].
Notes
In Haupt [1], it is postulated that the internal variables’ evolution is described by a system of ordinary differential equations (ODEs) of the form \({\dot{{\textbf {q}}}}_\alpha = \varvec{f}_\alpha (\varvec{\upvarepsilon },{\textbf {q}}_1,{\textbf {q}}_2,\ldots ,{\textbf {q}}_n)\), referred to as evolution equations. We refer to Eq. (4) as an implicitly given evolution equation, since the given form of the ODEs cannot be explicitly determined for every permissible potential \(\phi \). The Biot equation is nevertheless equivalent.
Since the invariants \({{\,\textrm{tr}\,}}{\varvec{\upvarepsilon }}\) and \({{\,\textrm{tr}\,}}{{\textbf {p}}}\) are linear functions of \(\varvec{\upvarepsilon }\) and \({\textbf {q}}\), the constraints on the corresponding weights in the passthrough layers of the non-decreasing FICNN as described in Sec. A are in general too restrictive. However, for the here shown NN-based model and the used reference material, this does not effect the prediction quality and is henceforth ignored.
With the uniaxial data set, only a small area in the invariant space can be scanned. In the case of isotropic elasticity, it is even only a single path, since \(\varepsilon _{11}\) and the transverse strains \(\varepsilon _{22}=\varepsilon _{33}\) can be uniquely assigned to each uniaxial stress state [28, 42].
References
Peter H (2000) Continuum mechanics and theory of materials. Springer, Berlin. https://doi.org/10.1007/978-3-662-04775-0
de Souza NEA, Peri D, Owen DRJ (2008) Computational methods for plasticity. Wiley, Chichester. https://doi.org/10.1002/9780470694626
Bock Frederic E et al (2019) A review of the application of machine learning and data mining approaches in continuum materials mechanics. Front Mater 6:110. https://doi.org/10.3389/fmats.2019.00110
Dornheim J et al. (2023) Neural networks for constitutive modeling: from universal function approximators to advanced models and the integration of physics. In: Archives of computational methods in engineering. https://doi.org/10.1007/s11831-023-10009-y
Ghaboussi J, Garrett JH, Wu X (1991) Knowledge-based modeling of material behavior with neural networks. J Eng Mech 117(1):132–153. https://doi.org/10.1061/(ASCE)0733-9399(1991)117:1(132)
Furukawa T, Yagawa G (1998) Implicit constitutive modelling for viscoplasticity using neural networks. Int J Numer Methods Eng 43(2):195–219
Hambli R, Katerchi H, Benhamou C-L (2011) Multiscale methodology for bone remodelling simulation using coupled finite element and neural network computation. Biomech Model Mechanobiol 10(1):133–145. https://doi.org/10.1007/s10237-010-0222-x
Yao CG et al (2014) Artificial neural network modelling to predict hot deformation behaviour of as HIPed FGH4169 superalloy. Mater Sci Technol 30(10):1170–1176. https://doi.org/10.1179/1743284713Y.0000000411
Xiaoxin L et al (2019) A data-driven computational homogenization method based on neural networks for the nonlinear anisotropic electrical response of graphene/polymer nanocomposites. Comput Mech 64(2):307–321. https://doi.org/10.1007/s00466-018-1643-0
Heider Y, Wang K, Sun WC (2020) SO(3)- invariance of informed-graph-based deepneuralnetwork for anisotropic elastoplastic materials. Comput Methods Appl Mech Eng 363:112875. https://doi.org/10.1016/j.cma.2020.112875
Fuchs A et al (2021) DNN2: a hyper-parameter reinforcement learning game for self-design of neural network based elasto-plastic constitutive descriptions. Comput Struct 249:106505. https://doi.org/10.1016/j.compstruc.2021.106505
Ghavamian F, Simone A (2019) Accelerating multiscale finite element simulations of history-dependent materials using a recurrent neural network. Comput Methods Appl Mech Eng 357:112594. https://doi.org/10.1016/j.cma.2019.112594
Mei H et al (2020) Study on constitutive relation of nickel-base superalloy inconel 718 based on long short term memory recurrent neural network. Metals 10(12):1588. https://doi.org/10.3390/met10121588
Bonatti C, Mohr D (2022) On the importance of self-consistency in recurrent neural network models representing elasto-plastic solids. J Mech Phys Solids 158:104697. https://doi.org/10.1016/j.jmps.2021.104697
Rumelhart DE, Hinton GE, Williams RJ (1988) Learning internal representations by error propagation. In: Readings in cognitive science. Elsevier, pp 399–421. https://doi.org/10.1016/B978-1-4832-1446-7.50035-2
Leygue A et al (2018) Data-based derivation of material response. Comput Methods Appl Mech Eng 331:184–196. https://doi.org/10.1016/j.cma.2017.11.013
Rosenkranz M et al (2023) A comparative study on different neural network architectures to model inelasticity. Int J Numer Meth Eng 124(21):4802–4840. https://doi.org/10.1002/nme.7319
Hochreiter S, Schmidhuber J (1997) Long short- term memory. Neural Comput 9(8):1735–1780. https://doi.org/10.1162/neco.1997.9.8.1735
Cho K et al. (2014) Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv: 1406.1078 [cs.CL]
Linden L et al (2023) Neural networks meet hyperelasticity: a guide to enforcing physics. J Mech Phys Solids 179:105363. https://doi.org/10.1016/j.jmps.2023.105363
Masi F et al (2021) Thermodynamics-based artificial neural networks for constitutive modeling. J Mech Phys Solids 147:104277. https://doi.org/10.1016/j.jmps.2020.104277
Raissi M, Perdikaris P, Karniadakis GE (2019) Physicsinformed neural networks: adeep learning framework for solving forward and inverse problems involving nonlinear partial differential equations. J Comput Phys 378:686–707. https://doi.org/10.1016/j.jcp.2018.10.045
Henkes A, Wessels H, Mahnken R (2022) Physics informed neural networks for continuum micromechanics. Comput Methods Appl Mech Eng 393:114790. https://doi.org/10.1016/j.cma.2022.114790
As’ad F, Avery P, Farhat C (2022) A mechanics-informed artificial neural network approach in data-driven constitutive modeling. Int J Numer Meth Eng 123(12):2738–2759. https://doi.org/10.1002/nme.6957
Klein DK et al (2022) Finite electro-elasticity with physics-augmented neural networks. Comput Methods Appl Mech Eng 400:115501. https://doi.org/10.1016/j.cma.2022.115501
Kalina KA et al (2023) FEANN: an efficient data- driven multiscale approach based on physics-constrained neural networks and automated data mining. Comput Mech 71:827–851. https://doi.org/10.1016/j.cma.2022.115501
Vlassis NN, Sun WC (2021) Sobolev training of thermodynamic-informed neural networks for interpretable elasto-plasticity models with level set hardening. Comput Methods Appl Mech Eng 377:113695. https://doi.org/10.1016/j.cma.2021.113695
Kalina KA et al (2022) Automated constitutive modeling of isotropic hyperelasticity based on artificial neural networks. Comput Mech 69(1):213–232. https://doi.org/10.1007/s00466-021-02090-6
Klein DK et al (2022) Polyconvex anisotropic hyperelasticity with neural networks. J Mech Phys Solids 159:104703. https://doi.org/10.1016/j.jmps.2021.104703
Eivazi H et al (2023) FE2 computations with deep neural networks: algorithmic structure, data generation, and implementation. Mathematical and Computational Applications 28(4):91. https://doi.org/10.3390/mca28040091
Vlassis NN, Sun W (2021) Component- based machine learning paradigm for discovering rate-dependent and pressure-sensitive level-set plasticity models. J Appl Mech 10(1115/1):4052684
Meyer KA, Ekre F (2023) Thermodynamically consistent neural network plasticity modeling and discovery of evolution laws. J Mech Phys Solids 180:105416. https://doi.org/10.1016/j.jmps.2023.105416
Fuhg JN et al (2023) Modular machine learning- based elastoplasticity: generalization in the context of limited data. Comput Methods Appl Mech Eng 407:115930. https://doi.org/10.1016/j.cma.2023.115930
Rezaei S, Moeineddin A, Harandi A (2024) Learning solutions of thermodynamics-based nonlinear constitutive material models using physics-informed neural networks. Comput Mech. https://doi.org/10.1007/s00466-023-02435-3
Fuhg JN, Jones RE, Bouklas N (2023) Extreme sparsification of physics-augmented neural networks for interpretable model discovery in mechanics. arXiv: 2310.03652 [cs.CE]
Benady A, Baranger E, Chamoin L (2024) Unsupervised learning of history-dependent constitutive material laws with thermodynamically consistent neural networks in the modified constitutive relation error framework. Working paper or preprint. https://hal.science/hal-04368755
Taç V et al (2023) Data-driven anisotropic finite viscoelasticity using neural ordinary differential equations. Comput Methods Appl Mech Eng 411:116046. https://doi.org/10.1016/j.cma.2023.116046
Abdolazizi KP, Linka K, Cyron CJ (2023) Viscoelastic constitutive artificial neural networks (vCANNs)—a framework for data-driven an isotropic nonlinear finite viscoelasticity. J Comput Phys 499:112704. https://doi.org/10.1016/j.jcp.2023.112704
As’ad F, Farhat C (2023) A mechanics-informed neural network framework for data-driven nonlinear viscoelasticity. In: AIAA SCITECH 2023 Forum. https://doi.org/10.2514/6.2023-0949
Upadhyay K et al. (2023) Physics-informed data-driven discovery of constitutive models with application to strain-rate-sensitive soft materials. arXiv: 2304.13897 [cs.CE]
Cai C et al (2023) Equivariant geometric learning for digital rock physics: estimating formation factor and effective permeability tensors from Morse graph. Int J Multiscale Comput Eng 21(5):1–24. https://doi.org/10.1615/IntJMultCompEng.2022042266
Linka K et al (2021) Constitutive artificial neural networks: a fast and general approach to predictive data driven constitutive modeling by deep learning. J Comput Phys 429:110010. https://doi.org/10.1016/j.jcp.2020.110010
Fuhg JN, Bouklas N, Jones RE (2022) Learning hyperelastic anisotropy from data via a tensor basis neural network. J Mech Phys Solids 168:105022. https://doi.org/10.1016/j.jmps.2022.105022
Tac V, Costabal FS, Tepole AB (2022) Data-driven tissue mechanics with polyconvex neural ordinary differential equations. Comput Methods Appl Mech Eng 398:115248. https://doi.org/10.1016/j.cma.2022.115248
Vlassis NN, Ma R, Sun WC (2020) Geometric deep learning for computational mechanics Part I: anisotropic hyperelasticity. Comput Methods Appl Mech Eng 371:113299. https://doi.org/10.1016/j.cma.2020.113299
Baydin AG et al. (2015) Automatic differentiation in machine learning: a survey. arXiv: 1502.05767 [cs.SC]
Daw A et al. (2021) Physics-guided neural networks (PGNN): an application in lake temperature modeling. arXiv: 1710.11431 [cs.LG]
Thakolkaran P et al (2022) NN-EUCLID: deep-learning hyperelasticity without stress data. J Mech Phys Solids 169:105076. https://doi.org/10.1016/j.jmps.2022.105076
Amos B, Xu L, Zico KJ (2016) Input convex neural networks. arXiv: 1609.07152 [cs.LG]
Klein Dominik K et al (2023) Parametrized polyconvex hyperelasticity with physics-augmented neural networks. Data-Centric Eng 4:e25. https://doi.org/10.1017/dce.2023.21
He X, Chen J-S (2022) Thermodynamically consistent machine-learned internal state variable approach for data-driven modeling of path-dependent materials. IComput Methods Appl Mech Eng 402:115348. https://doi.org/10.1016/j.cma.2022.115348
Weber P, Wagner W, Freitag S (2023) Physically enhanced training for modeling rate- independent plasticity with feedforward neural networks. Comput Mech. https://doi.org/10.1007/s00466-023-02316-9
Malik A et al (2021) A hybrid approach employing neural networks to simulate the elasto-plastic deformation behavior of 3d-foam structures. Adv Eng Mater 24(2):2100641. https://doi.org/10.1002/adem.202100641
Masi F, Stefanou I (2022) Multiscale modeling of inelastic materials with thermodynamicsbased artificial neural networks (TANN). Comput Methods Appl Mech Eng 398:115190. https://doi.org/10.1016/j.cma.2022.115190
Vlassis NN, Sun WC (2023) Geometric deep learning for computational mechanics part II: graph embedding for interpretable multiscale plasticity. Comput Methods Appl Mech Eng 404:115768. https://doi.org/10.1016/j.cma.2022.115768
Huang S et al (2022) Variational onsager neural networks (VONNs): a thermodynamics-based variational learning strategy for non-equilibrium PDEs. J Mech Phys Solids 163:104856. https://doi.org/10.1016/j.jmps.2022.104856
Holthusen H et al. (2023) Theory and implementation of inelastic constitutive artificial neural networks. arXiv: 2311.06380 [cs.LG]
Miehe C, Kiefer B, Rosato D (2011) An incremental variational formulation of dissipative magnetostriction at the macroscopic continuum level. Int J Solids Struct 48(13):1846–1866. https://doi.org/10.1016/j.ijsolstr.2011.02.011
Miehe C, Schotte J, Lambrecht M (2002) Homogenization of inelastic solid materials at ÿnite strains based on incremental minimization principles. Application to the texture analysis of polycrystals. J Mech Phys Solids 50(10):2123–2167. https://doi.org/10.1016/S0022-5096(02)00016-9
Miehe C (2002) Strain-driven homogenization of inelastic microstructures and composites based on an incremental variational formulation. Int J Numer Meth Eng 55(11):1285–1322. https://doi.org/10.1002/nme.515
Mielke A (2006) A mathematical framework for generalized standard materials in the rate-independent case. In: Multifield problems in solid and fluid mechanics, vol 28. Springer, Berlin, pp 399–428. https://doi.org/10.1007/978-3-540-34961-7_12
Ziegler H, Wehrli C (1987) The derivation of constitutive relations from the free energy and the dissipation function. In: Advances in applied mechanics, vol 25. Elsevier, pp 183–238.https://doi.org/10.1016/S0065-2156(08)70278-3
Rice JR (1971) Inelastic constitutive relations for solids: an internal-variable theory and its application to metal plasticity. J Mech Phys Solids 19(6):433–455. https://doi.org/10.1016/0022-5096(71)90010-X
Moreau JJ (2011) On unilateral constraints, friction and plasticity. In: Capriz G, Stampacchia G (eds) New variational techniques in mathematical physics. Springer, Berlin, pp 171–322. https://doi.org/10.1007/978-3-642-10960-7_7
Halphen B, Nguyen QS (1975) Sur les matériaux standard généralisés. Journal deMécanique 14:39–63
Biot MA (1965) Mechanics of incremental deformations. https://hal.science/hal-01352219
Kumar A, Lopez-Pamies O (2016) On the twopotential constitutive modeling of rubber viscoelastic materials. Comptes Rendus Mécanique 344(2):102–112. https://doi.org/10.1016/j.crme.2015.11.004
Fuhg JN, Bouklas N, Jones RE (2023) Stress representations for tensor basis neural networks: alternative formulations to Finger–Rivlin–Ericksen. arXiv:2308.11080 [cond-mat.soft]
Liu M, Liang L, Sun W (2020) A generic physics-informed neural network-based constitutive model for soft biological tissues. Comput Methods Appl Mech Eng 372:113402. https://doi.org/10.1016/j.cma.2020.113402
Flaschel M, Kumar S, De Lorenzis L (2023) Automated discovery of generalized standard material models with EUCLID. Comput Methods Appl Mech Eng 405:115867. https://doi.org/10.1016/j.cma.2022.115867
Bahador B, WaiChing S (2023) Physics-constrained symbolic model discovery for polyconvex incompressible incompressible hyperelastic materials. arXiv: 2310.04286 [cs.CE]
Kalina KA et al (2024) Neural network-based multiscale modeling of finite strain magneto-elasticity with relaxed convexity criteria. Comput Methods Appl Mech Eng 421:116739. https://doi.org/10.1016/j.cma.2023.116739
Kästner M et al (2012) Inelastic material behavior of polymers - experimental characterization, formulation and implementation of a material model. Mech Mater 52:40–57. https://doi.org/10.1016/j.mechmat.2012.04.011
Acknowledgements
All presented computations were performed on a HPC-Cluster at the Center for Information Services and High Performance Computing (ZIH) at TU Dresden. The authors thus thank the ZIH for generous allocations of computer time. MR and MK thank the German Research Foundation (DFG) for the support within the Research Training Group GRK2868 D\(^3\)—Project Number 493401063. The work of KK was supported by a postdoc fellowship of the German Academic Exchange Service (DAAD) and by the Graduate Academy (GA) of TU Dresden. All supports are gratefully acknowledged.
Funding
Open Access funding enabled and organized by Projekt DEAL.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendices
A neural networks
This section describes the concept of input convex feedforward neural networks, the normalization of inputs and outputs and gives an overview about the chosen hyperparameters of the networks, that are used in the numerical examples. In order to provide a compact explanation, some symbols are introduced: The network function is denoted as \(\mathscr {N}(\mathfrak {i})\) and depends on an input vector \(\mathfrak {i}\). The output \(\varvec{u}_l\) of the layer l is calculated using the activation function \(\mathscr {A}_l\), the biases \(\varvec{b}_l\) and the weights \(\varvec{W}_l\) connecting this layer to the previous layer \(l-1\). The following concepts are mainly based on Amos, Xu, and Kolter [49], with extensions to functions that are not only convex with respect to the inputs \(\mathfrak {i}\), but also with respect to some x in a network \(\mathscr {N}(\mathfrak {i}(x))\) by making the network non-decreasing in \(\mathfrak {i}\) [20, 50].
1.1 A.1 Input convex neural networks
1.1.1 A.1.1 Fully input convex neural networks
A fully input convex neural network (FICNN) is an FNN, that is convex in all of its arguments. There are several ways to construct such an FNN, but not all variants are equivalent. Consider for example a normal FNN with non-negative weights across all layers as well as convex and non-decreasing activation functions. Such an FNN is convex, but is also extremely restrictive. Therefore, the more flexible approach of Amos, Xu, and Kolter [49], shown in Fig. 14, is adopted. It is characterized by two different sets of weights: the weights in \(\varvec{W}_l^{\varvec{u}}\) connect the layer \(l-1\) with the layer l and the weights \(\varvec{W}_l^{\mathfrak {i}}\) connect the input \(\mathfrak {i}\) with the layer l. The outputs of the layers \(l\in (1,\ldots ,L)\) is now calculated as
The network \(\mathscr {N}(\mathfrak {i})=\varvec{u}_{L}\) is convex if the following three conditions are fulfilled:
-
(i)
all weights in \(\varvec{W}_l^{\varvec{u}}\) are non-negative,
-
(ii)
all \(\mathscr {A}_l\) are convex, and
-
(iii)
all \(\mathscr {A}_l\) are non-decreasing.
This approach offers greater flexibility since the weights \(\varvec{W}_l^{\mathfrak {i}}\) remain unconstrained as shown in Fig. 14a. A common choice for a convex and non-decreasing activation function is the softplus activation function, which is defined by
So far, the network has been constructed in such a way that it is convex with respect to its direct inputs. However, this must not always be sufficient. For example, in the context of this work, networks are to be implemented, that are convex with respect to a tensor, although the network inputs are invariants. The arbitrariness of the \(\varvec{W}^{\mathfrak {i}}_l\) follows from the fact that the second derivatives of \(\varvec{W}_l^{\mathfrak {i}} \mathfrak {i}\) with respect to \(\mathfrak {i}\) vanish and consequently do not have any influence on the convexity in \(\mathfrak {i}\) of the network function \(\mathscr {N}(\mathfrak {i})\). This changes if \(\mathscr {N}\) is not supposed to be convex in its inputs, but in another variable x, which the inputs depend on. That is, \(\mathscr {N}(\mathfrak {i}(x))\) is supposed to be convex in x. Then, the second derivatives of \(\varvec{W}^{\mathfrak {i}}_l \mathfrak {i}(x)\) with respect to x do not vanish if \(\mathfrak {i}(x)\) is a nonlinear function of x. Consequently, another restriction on the weights arises, see Fig. 14b: besides the conditions (i)–(iii), it must also hold that
-
(iv)
all entries in \(\mathfrak {i}(x)\) are convex functions of x, and
-
(v)
all weights in \(\varvec{W}^{\mathfrak {i}}_l\) are non-negative [29].
1.1.2 A.1.2 Partially input convex neural networks
A partially input convex neural network (PICNN) is, in contrast to an FICNN, only convex in some of its inputs. Let \(\mathscr {N}(\mathfrak {i}^{\cup }, \mathfrak {i}^{\backsim })\) denote the network, that is convex in \(\mathfrak {i}^{\cup }\), but not necessarily convex in \(\mathfrak {i}^{\backsim }\). The architecture of such a network is shown in Fig. 15. The figure shows the two different paths corresponding to the convex (bottom) and non-convex part (top). The non-convex part with the layer outputs \(\varvec{v}_l\) is a regular FNN, i.e., each layer is connected to only the previous layer of the non-convex path. A layer in the convex path with layer outputs \(\varvec{u}_l\), however, depends on not only the previous layer of the convex path, but also on the output of previous layer in the non-convex path and the convex inputs. The outputs of the layers in the non-convex path are defined as
and for the convex path as
The final output of the network is the last layer of the convex path, i.e., \(\mathscr {N}(\mathfrak {i}^{\cup }, \mathfrak {i}^{\backsim })=\varvec{u}_L\). Instead of two sets of weights, a PICNN comprises six different sets of weights (\(\varvec{W}^{\varvec{v}\varvec{v}}_l\), \(\varvec{W}^{\varvec{u}\varvec{u}}_l\), \(\varvec{W}^{\varvec{u}\mathfrak {i}^{\cup }}_l\), \(\varvec{W}^{\varvec{u}\varvec{v}}_l\), \({{\tilde{\varvec{W}}}}^{\varvec{v}\varvec{v}}_l\), \({{\tilde{\varvec{W}}}}^{\mathfrak {i}^{\cup }\varvec{v}}_l\)) to take into account possible products between the two paths and the convex input. The network is convex in \(\mathfrak {i}^{\cup }\) if the following four conditions are fulfilled:
-
(i)
all weights in \(\varvec{W}_l^{\varvec{u}\varvec{u}}\) are non-negative,
-
(ii)
all activations \(\mathscr {A}_l^{\cup }\) are convex,
-
(iii)
all activations \(\mathscr {A}_l^{\cup }\) are non-decreasing, and
-
(iv)
all activations \(\mathscr {A}_l^{\varvec{u}\varvec{v}}\) map to non-negative values.
All other weights may as well take negative values and the activations \(\mathscr {A}_l^{\backsim }\) and \(\mathscr {A}_l^{\mathfrak {i}^{\cup }\varvec{v}}\) can be chosen arbitrarily. A valid choice for \(\mathscr {A}_l^{\cup }\) is, like for the FICNN, the softplus function \({{\,\mathrm{\mathscr{S}\mathscr{P}}\,}}\), which additionally is non-negative and thus can be used as \(\mathscr {A}_l^{\varvec{u}\varvec{v}}\) as well. The mentioned conditions yield a network \(\mathscr {N}(\mathfrak {i}^{\cup }, \mathfrak {i}^{\backsim })\), that is convex in all entries of \(\mathfrak {i}^{\cup }\). As already discussed for the FICNN, this is not always sufficient. If a network \(\mathscr {N}(\mathfrak {i}^{\cup }(x), \mathfrak {i}^{\backsim })\) is supposed to be convex in x, it is not the first function acting on x, which leads to further restrictions on the weights and activations. \(\mathscr {N}(\mathfrak {i}^{\cup }(x), \mathfrak {i}^{\backsim })\) is convex in x, if in addition to (i)–(iv) it holds:
-
(v)
all entries of \(\mathfrak {i}^{\cup }\) are convex functions of x,
-
(vi)
all weights in \(\varvec{W}_l^{\varvec{u}\mathfrak {i}^{\cup }}\) are non-negative, and
-
(vii)
all \(\mathscr {A}_l^{\mathfrak {i}^{\cup }\varvec{v}}\) map to non-negative values.
These additional restrictions are only necessary if the entries of \(\mathfrak {i}^{\cup }\) are non-linear function in x, i.e., the second derivatives do not vanish. Otherwise, the standard architecture is also valid.
1.2 A.2 Normalization
The training process of neural networks is typically more stable and efficient if the inputs and outputs of the network are values of magnitude 1. In particular, since the free energy and the dissipation potential take on large numerical values in the range \(>10^6\), normalization is essential for a successful training. Normalization of a known quantity \((\circ )\) to the range \((-1, 1)\ni {{\tilde{(\circ )}}}\) can be carried out with
where \({\tilde{(\circ )}}\) represents the normalized quantity and \((\circ )_{\text {max}}\) and \((\circ )_{\text {min}}\) are the maximum and minimum value of \((\circ )\) across the whole training data set. For the problem described here, the inputs \(\varvec{\upvarepsilon }\), \({\textbf {q}}\), \({\dot{{\textbf {q}}}}\), t and \(\Delta t\) or the respective invariants \(\mathcal {I}^{\psi ^{\text {eq}}}, \mathcal {I}^{\psi ^{\text {ov}}}\) and \(\mathcal {I}^{\phi }\) have to be normalized, as well as the outputs \(\psi ^{\text {eq}}\), \(\psi ^{\text {ov}}\) and \(\phi \). The difficulty for the proposed model arises from the fact that neither \({\textbf {q}}\) and \({\dot{{\textbf {q}}}}\) as inputs nor \(\psi ^{\text {eq}}\), \(\psi ^{\text {ov}}\) or \(\phi \) as outputs are known before the training, but only \(\varvec{\upvarepsilon }(t)\), \(\varvec{\upsigma }(t)\) and t are known in advance. For this reason, the order of magnitude of the expected values for the unknown variables is estimated based on the available variables. Therefore, the normalization parameters s and m are determined on the basis of the following assumptions: the internal variable and the variable \({\textbf {p}}=\varvec{\upvarepsilon }-{\textbf {q}}\) are normalized with the values of \(\varvec{\upvarepsilon }\), i.e., \(s_{{\textbf {q}}}=s_{{\textbf {p}}}=s_{\varvec{\upvarepsilon }}\), while the rate \({\dot{{\textbf {q}}}}\) is normalized with \({{\dot{\varvec{\upvarepsilon }}}}\). The normalization parameters for the potentials are defined to be \(s_{\psi ^{\text {eq}}}=s_{\psi ^{\text {ov}}}=m_{\psi ^{\text {eq}}}=m_{\psi ^{\text {ov}}}=s_{\varvec{\upvarepsilon }}\cdot s_{\varvec{\upsigma }}\), where the choice of the ms exploits the fact that \(\psi ^{\text {eq}},\psi ^{\text {ov}},\phi \ge 0\). Using these normalization parameters, the network itself maps only from values of magnitudes around 1 to values of magnitude 1, which enables an efficient training. Note, that due to the linear nature of the transformation Eq. (37), this type of normalization does not effect the convexity properties of the resulting network function.
1.3 A.3 Architecture details
1.3.1 A.3.1 Potentials
The free energy and the dissipation potential are implemented as two FICNNs and a PICNN. The details about the size of the networks, i.e., number of hidden layers, number of neurons or used activation functions can be found in Table 5.
1.3.2 A.3.2 Auxiliary networks
Integration No auxiliary network is necessary for this training method.
FNN The auxiliary network for the internal variable is an ordinary FNN with a single input (the time t), 2 hidden layers with 50 neurons each, \(\tanh \) activations in the hidden layers and 6 output neurons with linear activation.
RNN The auxiliary network comprises an RNN cell and a subsequent FNN. The RNN cell is an LSTM cell with 13 inputs (6 for the strain tensor \(\varvec{\upvarepsilon }\), 6 for the stress tensor \(\varvec{\upsigma }\) and one for the time increment \(\Delta t\)), 50 entries in the hidden state and \(\tanh \) activations. The FNN takes the hidden state as input, has no hidden layers, i.e., it consists of only the input and output layer, and 6 has output neurons with linear activations.
B Convexity of the invariant basis
To enforce convexity of a network with respect to a symmetric second order tensor \({\textbf {S}}\in \mathcal {S}\mathcalligra {ym}_2\), the invariant basis \(\mathcal {I}= (\,{{\,\textrm{tr}\,}}{\textbf {S}}, {{\,\textrm{tr}\,}}{\textbf {S}}^2, {{\,\textrm{tr}\,}}{\textbf {S}}^4 \, )\) is used throughout this work. To motivate this choice, the convexity of
-
(i)
\(f({\textbf {S}})={{\,\textrm{tr}\,}}{\textbf {S}}\),
-
(ii)
\(f({\textbf {S}})={{\,\textrm{tr}\,}}{\textbf {S}}^2\),
-
(iii)
\(f({\textbf {S}})={{\,\textrm{tr}\,}}{\textbf {S}}^3\) and
-
(iv)
\(f({\textbf {S}})={{\,\textrm{tr}\,}}{\textbf {S}}^4\)
is examined. Two equivalent definitions of convexity are used here. A function \(f({\textbf {S}})\) is convex, if
or, under the assumption, that f is twice continuously differentiable,
Note that for twice differentiable f, Eq. (41) follows from Eq. (40) and vice versa.
-
(i)
From Eq. (40) follows directly
$$\begin{aligned} {{\,\textrm{tr}\,}}({\textbf {A}}+ \lambda \left[ {\textbf {B}}-{\textbf {A}}\right] ) = {{\,\textrm{tr}\,}}({\textbf {A}}) + \lambda \left[ {{\,\textrm{tr}\,}}({\textbf {B}})-{{\,\textrm{tr}\,}}({\textbf {A}})\right] . \end{aligned}$$(42)As a linear function, the invariant \({{\,\textrm{tr}\,}}({\textbf {S}})\) is thus (weakly) convex.
-
(ii)
Applying Eq. (40) to \(f({\textbf {S}})={{\,\textrm{tr}\,}}{\textbf {S}}^2\) yields
$$\begin{aligned} {{\,\textrm{tr}\,}}\left( \left( {\textbf {A}}+ \lambda \left[ {\textbf {B}}-{\textbf {A}}\right] \right) ^2 \right) =&{{\,\textrm{tr}\,}}\left( {\textbf {A}}^2\right. + \lambda {\textbf {A}}\cdot \left[ {\textbf {B}}-{\textbf {A}}\right] + \nonumber \\&\lambda \left[ {\textbf {B}}-{\textbf {A}}\right] \cdot {\textbf {A}}+ \nonumber \\&\lambda ^2\left. \left[ {\textbf {B}}^2-{\textbf {A}}\cdot {\textbf {B}}-{\textbf {B}}\cdot {\textbf {A}}+{\textbf {A}}^2\right] \right) \, . \end{aligned}$$(43)With \({{\,\textrm{tr}\,}}({\textbf {A}}\cdot {\textbf {B}})={{\,\textrm{tr}\,}}({\textbf {B}}\cdot {\textbf {A}})\), this expression simplifies to
$$\begin{aligned} {{\,\textrm{tr}\,}}({\textbf {A}}^2) + 2\lambda {{\,\textrm{tr}\,}}({\textbf {A}}\cdot {\textbf {B}}-{\textbf {A}}^2)&+ \nonumber \\ \lambda ^2{{\,\textrm{tr}\,}}({\textbf {B}}-2{\textbf {A}}\cdot {\textbf {B}}+{\textbf {A}}^2)&\le {{\,\textrm{tr}\,}}({\textbf {A}}^2) + \nonumber \\&\quad \lambda \left[ {{\,\textrm{tr}\,}}({\textbf {B}}^2) - {{\,\textrm{tr}\,}}({\textbf {A}}^2)\right] \nonumber \\ \lambda ^2{{\,\textrm{tr}\,}}({\textbf {B}}^2-2{\textbf {A}}\cdot {\textbf {B}}+{\textbf {A}}^2)&\le \lambda {{\,\textrm{tr}\,}}({\textbf {B}}^2 -2{\textbf {A}}\cdot {\textbf {B}}+{\textbf {A}}^2) \nonumber \\ \lambda ^2{{\,\textrm{tr}\,}}(({\textbf {B}}-{\textbf {A}})^2)&\le \lambda {{\,\textrm{tr}\,}}(({\textbf {B}}-{\textbf {A}})^2) \nonumber \\ \lambda&\le 1 \quad . \end{aligned}$$(44) -
(iii)
For \(f({\textbf {S}})={{\,\textrm{tr}\,}}{\textbf {S}}^3\), the tensors \({\textbf {A}}=\varvec{0}\) and \({\textbf {B}}=-\varvec{1}\) are considered. With Eq. (40) it follows
$$\begin{aligned} {{\,\textrm{tr}\,}}\left( \left( \varvec{0}+ \lambda \left[ -\varvec{1}-\varvec{0}\right] \right) ^3\right) = -3\lambda ^3&\le {{\,\textrm{tr}\,}}(\varvec{0}^3) + \nonumber \\&\quad \lambda \left[ {{\,\textrm{tr}\,}}((-\varvec{1})^3) - {{\,\textrm{tr}\,}}(\varvec{0}^3)\right] \nonumber \\ -3\lambda ^3&\le -3\lambda \nonumber \\ \lambda ^3&\ge \lambda , \end{aligned}$$(45)which is a contradiction for \(\lambda \in [0,1]\). The invariant \({{\,\textrm{tr}\,}}{\textbf {S}}^3\) is thus not convex in \({\textbf {S}}\).
-
(iv)
To show the convexity of \(f({\textbf {S}})={{\,\textrm{tr}\,}}{\textbf {S}}^4\), Eq. (41) is used. Evaluating Eq. (41) and introducing the abbreviation \(A_{ik}S_{kj}=r_{ij}\) yields
$$\begin{aligned} \frac{\partial f}{\partial S_{ij} \partial S_{kl}} A_{ij} A_{kl} = 4 \left[ 2r_{mn}r_{mn} + r_{mn}r_{nm} \right] . \end{aligned}$$(46)Decomposing \(\varvec{r}\) into the symmetric part \(s_{ij} = s_{ji}\) and antisymmetric part \(a_{ij}=-a_{ji}\), such that \(r_{mn} = s_{mn} + a_{mn}\) and using \(s_{mn}a_{mn} = 0\) finally leads to
$$\begin{aligned} \frac{\partial f}{\partial S_{ij} \partial S_{kl}} A_{ij} A_{kl} = 4 \left[ 3 s_{mn}s_{mn} + a_{kn}a_{kn} \right] \ge 0. \end{aligned}$$(47)
Consequently, \(\mathcal {I}= (\,{{\,\textrm{tr}\,}}{\textbf {S}}, {{\,\textrm{tr}\,}}{\textbf {S}}^2, {{\,\textrm{tr}\,}}{\textbf {S}}^4 \, )\) forms a complete and convex set of invariants.
C Loss over iterations
In 5.2.2, the three presented training methods are compared. Therefor, the data sets \(\mathscr {D}^{\text {multiax}}_{1\times 200}\) and \({{\tilde{\mathscr {D}}}}^{\text {multiax}}_{1\times 200}\) are used for training. The value of the respective loss functions for each method over the number of training iterations using the data set with ideal stress data \(\mathscr {D}^{\text {multiax}}_{1\times 200}\) is shown in this section. Recall, that every training method was used 25 times. To show how reliable each method is, all 25 curves with different random initializations are shown and the best run is highlighted with a thicker curve.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Rosenkranz, M., Kalina, K.A., Brummund, J. et al. Viscoelasticty with physics-augmented neural networks: model formulation and training methods without prescribed internal variables. Comput Mech (2024). https://doi.org/10.1007/s00466-024-02477-1
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s00466-024-02477-1