Chapter 5
Stress
Note: We have encountered vectors in the first chapters on statics of rigid bodies and have intuitively worked with them. To recap, a vector is an object that represents a direction and a magnitude. Geometrically, they are often represented as arrows. In a Cartesian coordinate system (or basis), a vector can be represented by a set of numbers. For example in two dimensions, we denote the components of the vector \(\v {v}\) as the \(x\) and \(y\) components and typically write the vector in the column-form
It is important to emphasize that this is a representation of the vector and the components \(v_i\in \mathbb {R}\) depend on the specific basis \(\hat {x}\) and \(\hat {y}\). This representation is called a tensor of order \(1\). The order of a tensor is sometimes also called degree or rank. Any component-wise representation, such as the one on the right hand side of Eq. \eqref{eq:cartvec}, implies a basis. The basis is written explicitly as \(\hat {x}\) and \(\hat {y}\) in the middle expression of Eq. \eqref{eq:cartvec}. Note that in component notation, the basis vectors are
Formally, a vector is an element of a vector space. A vector space \(V\) (often also called a linear space) is a set of objects (for example the set containing our basis vectors \(\hat {x}\) and \(\hat {y}\) and linear combinations thereof) along with two operations: Addition (of two vectors) and multiplication (of a vector) with a scalar. These operations again yield a vector, i.e. an element of \(V\). This can be expressed as
- \(\v {u},\;\v {v}\in V\), then \(\v {u}+\v {v}\in V\)
- \(a\in \mathbb {R}\) and \(\v {u}\in V\), then \(a\cdot \v {u}\in V\)
and hence Eq. \eqref{eq:cartvec} yields a vector. In general, we may multiply the vectors by an element of a algebraic number field \(\mathbb {F}\) rather than \(\mathbb {R}\). Then we say \(V\) is a vector space over the field \(\mathbb {F}\). In these notes (and our lectures) we will always deal with either real (\(\mathbb {F}=\mathbb {R}\)) or complex (\(\mathbb {F}=\mathbb {C}\)) numbers.
Recall the concept of an algebraic number field in mathematics. A field is an algebraic structure that is a set along with two operations “\(+\)” and “\(\cdot \)” associating an element with two elements of the set. The operations are required to satisfy the field axioms:
- Associativity of addition and multiplication: \(a+(b+c)=(a+b)+c\) and \(a\cdot (b\cdot c)=(a\cdot b)\cdot c\).
- Commutativity of addition and multiplication: \(a+b=b+a\) and \(a\cdot b=b\cdot a\).
- Additive and multiplicative identity: \(0\in \mathbb {F}\) with \(a+0=a\) and \(1\in \mathbb {F}\) with \(1\cdot a=a\).
- Additive inverses: \(\forall a\in \mathbb {F}\) we have an inverse element \(i=-a\) with \(a+i=a+(-a)=0\)
- Multiplicative inverses: \(\forall a\ne 0\) we have an element \(a^{-1}\) with \(a\cdot a^{-1}=1\).
- Distributivity of multiplication over addition: \(a\cdot (b+c)=a\cdot b+a\cdot c\)
Note that in physics a field is typically a quantity that depends on position and not an algebraic number field. A scalar field \(\phi (\v {r})\) would be scalar quantity \(\phi \in \mathbb {R}\) that depends on a vector (the position) \(\v {r}\in V\). While these two meanings of the term field exist, we will always refer to this latter physical meaning in the following. Additionally, we have here used the dot \(\cdot \) to indicate a scalar multiplication, but we will reserve this dot in the rest of these notes for contractions, i.e. operations that implicitly include a sum.
The simplest (and also most important) operation on a vector is a linear transformation. The simplest linear transformation is the multiplication with a scalar \(a\in \mathbb {R}\). In terms of the geometric interpretation of a vector, this multiplication would change the length of the vector by \(a\) but not its direction.
A general linear transformation can in addition change the direction of the vector and therefore represent for example rotations. Given a vector \(\v {u}\) and a scalar \(\alpha \), a linear transformation \(\mathcal {L}\) to a vector \(\v {u}'=\mathcal {L}\v {u}\) has the properties:
We will only deal with linear operations that map between the same vector space \(V\), \(\mathcal {L}: V\mapsto V\). (It may map between quantities with different physical units.) Given a component-wise (Cartesian) representation of a vector, Eq. \eqref{eq:cartvec}, a linear transformation can be expressed as a multiplication by a matrix. This is easily by applying Eqs. \eqref{eq:linscalar} and \eqref{eq:linsum} to
Application of the linear operation to an arbitrary vector, Eq. \eqref{eq:linvec}, can therefore be expressed as
In component-wise notation we can express Eq. \eqref{eq:linmat} as
It is straighforward to “convert” from dyadic notation to component-wise or index notation. Imagine the \(i,j\) component of the resultant matrix of the product
Note that there is also a double contraction, for example
5.1 Rotating the stress tensor
5.1.1 Rotating vectors
A rotation \(\mathcal {R}\) is a linear operation that does not affect the length (or norm) of the vector, \(|\mathcal {R}\v {x}|=|\v {x}|\). Note that if \(\mathcal {R}\) is represented by a matrix \(\t {R}\), then
which is is equal to \(|\v {x}|\) only if \(\t {R}^T\cdot \t {R}=\t {1}\), i.e. if \(\t {R}\) is an orthogonal matrix. Here the superscript \(^T\) indicates the transpose operation, \(R^T_{ij}=R_{ji}\), and \(\t {1}\) is the unit matrix. Note that this relationship implies \(\t {R}^{-1}=\t {R}^T\), i.e. the inverse of an orthogonal matrix is simply its transpose.
Let us assume we want to rotate from a coordinate system (basis vectors) \(\hat {x}\) and \(\hat {y}\) to the primed coordinate system \(\hat {x}'\) and \(\hat {y}'\). Then the rotation should transform the basis vectors of the unprimed into the primed system,
From Eq. \eqref{eq:matcomp} we obtain the components of the rotation matrix as
If all vectors are expressed in the coordinate system \(\hat {x}\) and \(\hat {y}\), then those vectors have the simple representation given by Eq. \eqref{eq:cartvec} and the rotation matrix can be written as
i.e. we just stack the basis vectors as columns together to obtain \(\t {R}\). This is a simple prescription to construct the rotation matrix given the basis vectors of the rotated coordinate system (in terms of the original coordinate system). It is straightforward to see that \(\t {R}\) is an orthogonal matrix,
if \(\hat {x}'\) and \(\hat {y}'\) are orthogonal vectors of unit length.
The basis vectors of a coordinate system rotated counterclockwise by an angle \(\theta \) are given by
and hence the rotation matrix is
It is straightforward to verify that indeed \(\t {R}^T\cdot \t {R}=\t {1}\).
5.1.2 Rotating tensors
A tensor is a representation of a linear transformation: A tensor of order 2 (a matrix) transforms a tensor of order 1 (a vector) into another tensor of order 1. We have already encountered the Cauchy stress tensor \(\t {\sigma }\) of solid mechanics. It transforms an area vector \(\v {A}\) into a force vector \(\v {F}\),
We describe \(\v {F}\), \(\v {A}\) and \(\t {\sigma }\) in terms of their components, i.e. we describe a realization of the linear transformation (and the area and force vectors).
A rotation does not affect the effect of the linear transformation. Hence, if we change the coordinate system from \(\hat {x}\), \(\hat {y}\) to \(\hat {x}'\), \(\hat {y}'\), then the component of our vector \(\v {F}\) change to
Comparing with Eq. \eqref{eq:rotmat} yields the compact notation
Note that the transpose shows up because \(\t {R}\) describes the rotation of the basis and we are here rotating a vector expressed within this basis, which is the inverse operation.
Since we now understand how to rotate vectors, we can ask the question of how to rotate a tensor of order 2. Starting from Eq. \eqref{eq:tensex} and using the inverse of Eq. \eqref{eq:vecrot}, we write
and multiply by \(\t {R}^T\) from the left to yield
In the rotated coordinate system, the tensor attains the representation
since this leave the expression for the linear transformation
invariant. The trace and determinant of the rotated tensor are
and hence invariant under rotation. Note that in general for an \(n\times n\) tensor, there are \(n\) invariants; more on this will be discussed below when talking about eigenvalues.
Since we now understand how to rotate tensors of order 2, we can ask the question how to rotate a tensor of order 4. As an example, we use the stiffness tensor \(\tt {C}\), that we will encounter in the next chapter. This tensor transforms a strain tensor \(\t {\varepsilon }\) into a stress tensor \(\t {\sigma }\),
We can rewrite this using the inverse of Eq. \eqref{eq:tens2rot} as
which we now multiply from the left with \(\t {R}^T\) and from the right with \(\t {R}\). This gives
and hence the transformation rule
Quantities that transform as Eqs. \eqref{eq:vecrot}, \eqref{eq:tens2rot} and \eqref{eq:tens4rot} are called tensors.
5.2 Principal stresses
Let us discuss in more detail what happens if we rotate a symmetric tensor of order 2, i.e. a tensor that fulfills \(\t {\sigma }^T = \t {\sigma }\). From Eq. \eqref{eq:tens2rot} it is straightforward to see, that \(\t {\sigma }'^T=\t {\sigma }'\), i.e. the transformed tensor is also symmetric.
We now explicitly write the rotation for the tensor
using the rotation matrix Eq. \eqref{eq:rotmat}. This gives symmetric \(\t {\sigma }'=\t {R}^T\cdot \t {\sigma }\cdot \t {R}\) with the components
There is a specific rotation angle \(\theta _0\) where the diagonal elements \(\sigma '_{xx}\) and \(\sigma '_{yy}\) become extremal. It is determined from \(\sigma '_{xx,\theta }=\sigma '_{yy,\theta }=0\),
At this rotation angle we find that the off-diagonal components vanish, \(\sigma '_{xy}(\theta _0)=0\), and the rotated matrix is diagonal,
This is the simplest example of the diagonalization of a matrix. The elements of the diagonalized Cauchy stress tensor, \(\sigma _1\) and \(\sigma _2\), are called the principal stresses.
The diagonalization of a symmetric matrix always corresponds to the rotation into a new coordinate system. We have explictly shown this for the two-dimensional case here and will now show it in more generality for the three-dimensional case. In this process, we will encounter the concept of a stress invariant.
5.3 Stress invariants
Equation \eqref{eq:diag} fulfills the eigenvalue equations \(\t {\sigma }\cdot \hat {x}=\sigma _1\hat {x}\) and \(\t {\sigma }\cdot \hat {y}=\sigma _2\hat {y}\). Rather than explicitly computing a rotation, we can ask the question whether we can find a scalar \(\lambda \) and a vector \(\v {v}\) that fulfills the eigenvalue equation
This equation of course has the trivial solution \(\v {v}=0\). It can only have a nontrivial solution if
For a \(n\times n\) matrix, Eq. \eqref{eq:ev} leads to a polynomial of order \(n\) in \(\lambda \) with \(n\) (possibly complex valued) solutions.
For the case of a symmetric \(3\times 3\) matrix, we can write this down explicitly as
The quantities \(I_1\) to \(I_3\) are called invariants. We have shown above explicitly that the trace and the determinant are invariant under rotation. The same holds true for all coefficients of the characteristic polynomial (two of which are actually trace and determinant). This is because
The 3-dimensional tensor therefore has three invariants. These invariants have important physical interpretations. For the stress tensor, \(I_1\) is related to the hydrostatic stress and \(I_2\) to the shear stress.
Note that for a diagonal matrix,
By equating Eqs. \eqref{eq:genI1} to \eqref{eq:genI3} with Eqs. \eqref{eq:diagI1} to \eqref{eq:diagI3} we can calculate the eigenvalues \(\sigma _1\), \(\sigma _2\) and \(\sigma _3\).
Once we have computed the eigenvalues, we can obtain the corresponding eigenvectors by solving
Note that the expressions only determine the direction of \(\v {v}_i\), not its length, and we are free to require \(|\v {v}_i|=1\). Furthermore, let us regard scalar products \(\v {v}_1\cdot \v {v}_2\), then
and if \(\sigma _1\not =\sigma _2\) we must have \(\v {v}_1\cdot \v {v}_2=0\). Hence the eigenvectors of a symmetric matrix are orthonormal, or in other words, they form the basis of a coordinate system.
We can write Eq. \eqref{eq:eigenvectors} in the more compact notation
with \(\t {R}=(\v {v}_1, \v {v}_2, \v {v}_3)\). Multiplying from the left with \(\t {R}^{-1}=\t {R}^T\) (this holds because the eigenvectors are orthonormal) we get
This is nothing else than a coordinate transformation (rotation) of the tensor \(\t {\sigma }\). Since the diagonalization of a symmetric matrix leads to orthonormal eigenvectors, the diagonalization is a rotation of the coordinate system.
5.4 Hydrostatic and deviatoric stress
We call \(\sigma _h = \tr \t {\sigma }/3 = I_1/3\) the hydrostatic stress. It is the stress measure that tells us about volume expansion and contraction. The pressure is the negative of the hydrostatic stress, \(p=-\sigma _h\). Our sign convention is such that positive stresses (negative pressures) mean tension and negative stresses (positive pressures) compression.
Using the hydrostatic stress, we can construct yet another stress tensor that quantifies the deviation from a pure hydrostatic condition with stress \(\sigma _h \t {1}\). We define
the deviatoric stress. Note that this tensor is constructed such that \(\tr \t {s}=0\).
The invariants of the deviatoric stress are commonly denoted by the symbol \(J\). We already know that \(J_1=0\) by construction. The second invariant is given by
From this invariant we can derive the von-Mises stress \(\sigma _{vM}=\sqrt {3 J_2}\). The von-Mises stress characterizes the pure shear contribution to the stress. The second invariant of the deviatoric stress plays an important role in plasticity models, where it is often assumed that a material flows when the von-Moses stress exceeds a certain threshold.