Search Continuum Mechanics Website

\[ {\partial y \over \partial x} = 0 \qquad {\partial y \over \partial y} = 1 \qquad {\partial y \over \partial z} = 0 \]

And similarly, the derivatives of \(z\) with respect to the three variables are

\[ {\partial z \over \partial x} = 0 \qquad {\partial z \over \partial y} = 0 \qquad {\partial z \over \partial z} = 1 \]

At this point, the pattern should be obvious. It can all be summarized in tensor notation as

\[ {\partial x_i \over \partial x_j} = x_{i,j} = \delta_{ij} \]

But nothing has changed the \(y\) value, so it is still equal to 5. (Note that the coordinates are independent, so \(y\) is not required to change just because \(x\) did.) So \(y_{new} - y_{old} = 5 - 5 = 0\) and \(dy/dx = 0\).

In contrast, \(dx/dx = (3-2)/(3-2) = 1\). Clearly \({dx_i \over dx_j} = x_{i,j} = \delta_{ij}\).

\[ v_i \delta_{ij} = v_1 \delta_{1j} + v_2 \delta_{2j} + v_3 \delta_{3j} \]

where the \(i\) subscript has been automatically summed from 1 to 3 because it appeared twice in the term, once in \(v_i\) and once in \(\delta_{ij}\).

Now arbitrarily select a value for \(j\), say \(j = 3\). This gives

\[ v_i \delta_{i3} = v_1 \delta_{13} + v_2 \delta_{23} + v_3 \delta_{33} = v_3 \]

The result is simply \(v_3\) (the value chosen for \(j\)) because \(\delta_{33} = 1\) while \(\delta_{13} = \delta_{23} = 0\). In general, \(v_i \delta_{ij}\) will always equal \(v_j\) for whatever value is selected for \(j\). Therefore, the general rule is

\[ v_i \delta_{ij} = v_j \]

hence the nickname of

The rule applies regardless of the complexity of the term. Recall the vector cross-product term \(\epsilon_{ijk} \omega_j r_k\). Multiply it by \(\delta_{im}\) for example.

\[ \delta_{im} \epsilon_{ijk} \omega_j r_k \quad = \quad \epsilon_{mjk} \omega_j r_k \]

It has the sole effect of substituting \(m\) for \(i\) in the alternating tensor.

\[ W = {1 \over 2} \sigma_{ij} \epsilon_{ij} \]

Although there are several subscripts in the equation, they all expand out (all sum from 1 to 3) because each letter occurs exactly twice. This is amplified by the fact that on the left hand side (LHS) of the equation, \(W\) is undeniably a scalar because it has no subscript(s) at all.

\[ v_i = \epsilon_{ijk} \omega_j r_k \]

This very small tensor equation represents so much. For starters, it represents three equations, not one, because the \(v_i\) term shows clearly that there is an equation for each \(i =\) 1, 2, and 3. Second, each of the three equations contains nine terms because \(j\) and \(k\) both sum to three. However, we've already seen here that \(\epsilon_{ijk}\) is zero in most terms, leaving

\[ v_1 = \omega_2 r_3 - \omega_3 r_2 \] \[ v_2 = \omega_3 r_1 - \omega_1 r_3 \] \[ v_3 = \omega_1 r_2 - \omega_2 r_1 \]

That is a lot packed into one small tensor equation. Of course, \({\bf v} = \boldsymbol{\omega} \times {\bf r}\) actually represents the same equations as well, except that the subscripts in the tensor equation explicitly lead one to the three component equations.

It is important to recognize that tensor notation often provides the freedom to write any tensor term in many forms. In this case, alternative examples include

\[ v_i = \omega_j \epsilon_{ijk} r_k \qquad \qquad v_i = r_k \epsilon_{ijk} \omega_j \qquad \qquad v_i = r_k \omega_j \epsilon_{ijk} \qquad \qquad v_k = \epsilon_{ijk} \omega_i r_j \]

All are equivalent, and technically correct, because the multiplication details are dictated by the subscripts, not the order of the factors. However, it is customary for readability considerations to write \(\epsilon_{ijk}\) first in a term, and write the remaining factors in approximately the same order they would appear in vector or matrix notation. This leads to the original form of \(v_i = \epsilon_{ijk} \omega_j r_k\) as being preferred.

It is also important to recognize forms that are incorrect, and why. One incorrect form is \(v_m = \epsilon_{ijk} \omega_j r_k\). This is wrong because the free indices on the two sides of the equation are different: \(m\) on the LHS and \(i\) on the RHS. While one might guess the author's intent in such a case, such a mismatch of free indices should be avoided.

A similar bad form is the addition of \(a_i + b_j\). This is simply meaningless.

A second incorrect form is \(v_i = \epsilon_{ijj} \omega_j r_j\). This is wrong because \(j\) appears more than twice in the RHS term.

\[ \boldsymbol{\epsilon} = {1 \over E} \left[ (1 + \nu) \boldsymbol{\sigma} - \nu \; {\bf I} \; \text{tr}(\boldsymbol{\sigma}) \right] \]

where:

\(\boldsymbol{\sigma} \; \) is the stress tensor,\(\boldsymbol{\epsilon} \; \; \) is the strain tensor,

\(E\) is the elastic modulus,

\(\nu \; \; \) is Poisson's ratio,

\({\bf I} \; \; \) is the identity matrix,

\(\text{tr(...)}\) is the trace of a tensor, in this case, \(\sigma_{11}+\sigma_{22}+\sigma_{33}\)

Hooke's Law is written in tensor notation as

\[ \epsilon_{ij} = {1 \over E} \left[ (1 + \nu) \sigma_{ij} - \nu \; \delta_{ij} \; \sigma_{kk} \right] \]

This shows clearly that each term is 2nd rank because of the \(i\) and \(j\) indices. Note that \(\sigma_{kk}\) is just a scalar quantity because \(k\) is summed from 1 to 3 since it appears twice. So \(\sigma_{kk}\) is the trace of \(\boldsymbol{\sigma}\). Another key insight is that the entire \(\nu \; \delta_{ij} \; \sigma_{kk}\) term is nonzero only when \(i = j\). This corresponds to the normal stresses and strains and is the classic

Writing all the matrices out gives

\[ \left[ \matrix { \epsilon_{11} & \epsilon_{12} & \epsilon_{13} \\ \epsilon_{21} & \epsilon_{22} & \epsilon_{23} \\ \epsilon_{31} & \epsilon_{32} & \epsilon_{33} } \right] = {1 \over E} \left\{ (1 + \nu) \left[ \matrix { \sigma_{11} & \sigma_{12} & \sigma_{13} \\ \sigma_{21} & \sigma_{22} & \sigma_{23} \\ \sigma_{31} & \sigma_{32} & \sigma_{33} } \right] - \nu \, ( \sigma_{11} + \sigma_{22} + \sigma_{33} ) \left[ \matrix { 1 & 0 & 0 \\ 0 & 1 & 0 \\ 0 & 0 & 1} \right] \right\} \]

So what are the individual component equations represented by the tensor notation equation, and how many are there? Since there are two

\[ \epsilon_{11} = {1 \over E} \left[ (1 + \nu) \sigma_{11} - \nu \; (\sigma_{11} + \sigma_{22} + \sigma_{33}) \right] \qquad \qquad \epsilon_{12} = {1 \over E} \left[ (1 + \nu) \sigma_{12} \right] \] \[ \epsilon_{22} = {1 \over E} \left[ (1 + \nu) \sigma_{22} - \nu \; (\sigma_{11} + \sigma_{22} + \sigma_{33}) \right] \qquad \qquad \epsilon_{13} = {1 \over E} \left[ (1 + \nu) \sigma_{13} \right] \] \[ \epsilon_{33} = {1 \over E} \left[ (1 + \nu) \sigma_{33} - \nu \; (\sigma_{11} + \sigma_{22} + \sigma_{33}) \right] \qquad \qquad \epsilon_{23} = {1 \over E} \left[ (1 + \nu) \sigma_{23} \right] \]

but then clean up into more concise forms

\[ \epsilon_{11} = {1 \over E} \left[ \sigma_{11} - \nu \; (\sigma_{22} + \sigma_{33}) \right] \qquad \qquad \epsilon_{12} = {(1 + \nu) \over E} \sigma_{12} \] \[ \epsilon_{22} = {1 \over E} \left[ \sigma_{22} - \nu \; (\sigma_{11} + \sigma_{33}) \right] \qquad \qquad \epsilon_{13} = {(1 + \nu) \over E} \sigma_{13} \] \[ \epsilon_{33} = {1 \over E} \left[ \sigma_{33} - \nu \; (\sigma_{11} + \sigma_{22}) \right] \qquad \qquad \epsilon_{23} = {(1 + \nu) \over E} \sigma_{23} \]

Again, this is an amazing amount of information packed into a single tensor equation.

For anyone diving directly into this page here, note that \(\epsilon_{12}\), \(\epsilon_{13}\), and \(\epsilon_{23}\) are one-half of the normal shear quantities, i.e., \(\epsilon_{12} = \gamma_{12} / 2\).

Returning to the fundamental equation...

\[ \epsilon_{ij} = {1 \over E} \left[ (1 + \nu) \sigma_{ij} - \nu \; \delta_{ij} \; \sigma_{kk} \right] \]

It is easy to solve for \(\sigma_{ij}\) to get

\[ \sigma_{ij} = {1 \over (1 + \nu)} \left[ E \; \epsilon_{ij} + \nu \; \delta_{ij} \; \sigma_{kk} \right] \]

But there is a problem, a major problem. There is still a stress term, \(\sigma_{kk}\), on the RHS that must be resolved. It cannot be combined with \(\sigma_{ij}\) on the LHS because they are different animals. \(\sigma_{ij}\) is an individual component of the stress tensor, \(\sigma_{kk}\) is the trace of the stress tensor, \( (\sigma_{11} + \sigma_{22} + \sigma_{33}) \).

This hurdle is overcome in a couple steps by first multiplying the equation through (both sides of course) by \(\delta_{ij}\) as follows.

\[ \delta_{ij} \sigma_{ij} = {1 \over (1 + \nu)} \left[ E \; \delta_{ij} \; \epsilon_{ij} + \nu \; \delta_{ij} \; \delta_{ij} \; \sigma_{kk} \right] \]

At this point, several advanced properties of tensor notation kick in. First, note that the entire equation has been transformed from a 2nd rank tensor equation to a scalar equation because there are no longer any free indices. Each term now has \(i\) and \(j\) occurring twice in it, so both are automatically summed from 1 to 3.

Second, the terms containing \( \delta_{ij} \sigma_{ij} \) and \( \delta_{ij} \epsilon_{ij} \) can be simplified by recalling the substitution property of the Kronecker Delta. Since \(\delta_{ij} = 1\) only when \(i = j\), then

\[ \delta_{ij} \sigma_{ij} = \sigma_{ii} = \sigma_{jj} = \sigma_{11} + \sigma_{22} + \sigma_{33} \] and

\[ \delta_{ij} \epsilon_{ij} = \epsilon_{ii} = \epsilon_{jj} = \epsilon_{11} + \epsilon_{22} + \epsilon_{33} \]

So choose, arbitrarily, to write each term with \(j\) subscripts as \(\sigma_{jj}\) and \(\epsilon_{jj}\).

The final interesting term contains \(\delta_{ij} \delta_{ij}\). This equals 3 because

\[ \delta_{ij} \delta_{ij} = \delta_{ii} = \delta_{11} + \delta_{22} + \delta_{33} = 1 + 1 + 1 = 3 \]

So the now-scalar equation reduces to

\[ \sigma_{jj} = {1 \over (1 + \nu)} \left[ E \; \epsilon_{jj} + 3 \; \nu \; \sigma_{kk} \right] \]

It's now time for another round of insight into tensor notation. Note that \(\sigma_{jj}\) and \(\sigma_{kk}\) are in fact equal because both expand to \( \sigma_{11} + \sigma_{22} + \sigma_{33} \). So they can be combined. The \(\sigma_{jj}\) is simply rewritten as \(\sigma_{kk}\), producing

\[ \sigma_{kk} = {1 \over (1 + \nu)} \left[ E \; \epsilon_{jj} + 3 \; \nu \; \sigma_{kk} \right] \]

Now the \(\sigma_{kk}\) terms on the LHS and RHS can be combined to obtain

\[ \sigma_{kk} = {E \; \epsilon_{jj} \over (1 - 2 \nu)} \]

which is the simple clean equation we need to insert back into

\[ \sigma_{ij} = {1 \over (1 + \nu)} \left[ E \; \epsilon_{ij} + \nu \; \delta_{ij} \; \sigma_{kk} \right] \]

to obtain

\[ \sigma_{ij} = {E \over (1 + \nu)} \left[ \epsilon_{ij} + { \nu \over (1 - 2 \nu)} \delta_{ij} \epsilon_{jj} \right] \]

Except this is incorrect because \(j\) now appears three times in the last term. But the solution is simple. Simply rewrite \(\epsilon_{jj}\) as \(\epsilon_{kk}\). This is completely acceptable because both forms expand out to give the same trace of the strain tensor.

\[ \sigma_{ij} = {E \over (1 + \nu)} \left[ \epsilon_{ij} + { \nu \over (1 - 2 \nu)} \delta_{ij} \epsilon_{kk} \right] \]

And we are finally done.

The second key aspect of the example was the demonstrated freedom to change indices when they occur twice. This is because any indices occurring twice are automatically expanded out so it does not matter what letter they are. This was demonstrated by changing \(\sigma_{jj}\) to \(\sigma_{kk}\) and \(\epsilon_{kk}\) to \(\epsilon_{jj}\) during the example.

\[ \begin{eqnarray} \epsilon_{ijk} \epsilon_{lmn} & = & \left| \matrix{ \delta_{il} & \delta_{im} & \delta_{in} \\ \delta_{jl} & \delta_{jm} & \delta_{jn} \\ \delta_{kl} & \delta_{km} & \delta_{kn} } \right| \\ \\ & = & \delta_{il} (\delta_{jm} \delta_{kn} - \delta_{jn} \delta_{km}) + \delta_{im} (\delta_{jn} \delta_{kl} - \delta_{jl} \delta_{kn}) + \delta_{in} (\delta_{jl} \delta_{km} - \delta_{jm} \delta_{kl}) \end{eqnarray} \]

which is too complex to be of much use to anyone! However, multiplying through by \(\delta_{il}\) reduces the above equation to something much more manageable and useful. (Proving this yourself is an excellent homework problem.)

\[ \epsilon_{ijk} \epsilon_{imn} = \delta_{jm} \delta_{kn} - \delta_{jn} \delta_{km} \]

This is a 4th rank tensor equation because there are four free indices, \(j, k, m,\) and \(n\). The \(i\) index is repeated twice on the LHS, so it is summed from 1 to 3. So the equation could be expanded to give

\[ \epsilon_{ijk} \epsilon_{imn} = \epsilon_{1jk} \epsilon_{1mn} + \epsilon_{2jk} \epsilon_{2mn} + \epsilon_{3jk} \epsilon_{3mn} = \delta_{jm} \delta_{kn} - \delta_{jn} \delta_{km} \]

The terms can be evaluated once values are chosen for each of the four free indices.

The identity is used when two alternating tensors are present in a term, which usually arises when the term involves cross products. The benefit of employing it is that once the epsilons are transformed into the deltas, then the substitution property of the Kronecker Deltas can be used to simplify the equation. The following example demonstrates the usefulness of this identity.

\[ Area = {1 \over 2} | \; {\bf a} \times {\bf b}| \]

which can be written in tensor notation as

\[ Area = {1 \over 2} \sqrt{ \epsilon_{ijk} a_j b_k \epsilon_{imn} a_m b_n } \]

At this point, the epsilon-delta identity can be employed to obtain

\[ Area = {1 \over 2} \sqrt{ (\delta_{jm} \delta_{kn} - \delta_{jn} \delta_{km}) a_j b_k a_m b_n } \]

Multiplying through gives

\[ Area = {1 \over 2} \sqrt{ \delta_{jm} \delta_{kn} a_j b_k a_m b_n - \delta_{jn} \delta_{km} a_j b_k a_m b_n } \]

And the substitution property of the delta operator can now be exploited to obtain

\[ Area = {1 \over 2} \sqrt{ a_m a_m b_n b_n - b_m a_m a_n b_n } \]

But \(a_m a_m\) is simply the dot product of the \({\bf a}\) vector with itself, as is \(b_n b_n\). And \(a_m b_m\) and \(a_n b_n\) are both the same dot product of \({\bf a}\) with \({\bf b}\).

\[ Area = {1 \over 2} \sqrt{ ({\bf a} \cdot {\bf a}) ({\bf b} \cdot {\bf b}) - ({\bf a} \cdot {\bf b})^2 } \]

Which offers an alternative calculation to the cross product approach. For what it's worth, the equation can also be written as a determinant.

\[ Area = {1 \over 2} \sqrt{ \left| \matrix{ ({\bf a} \cdot {\bf a}) & ({\bf a} \cdot {\bf b}) \\ ({\bf a} \cdot {\bf b}) & ({\bf b} \cdot {\bf b}) } \right| } \]

\[ \nabla \times ( \nabla \times {\bf v} ) = \nabla ( \nabla \cdot {\bf v} ) - \nabla^2 {\bf v} \]

\( ( \nabla \times {\bf v} ) \) is \( \epsilon_{ijk} v_{k,j} \) and \( \nabla \times ( \nabla \times {\bf v} ) \) is \( \epsilon_{mni} \epsilon_{ijk} v_{k,jn} \). Apply the identity to transform the equation to

\[ \epsilon_{mni} \epsilon_{ijk} v_{k,jn} = ( \delta_{mj} \delta_{nk} - \delta_{mk} \delta_{jn} ) v_{k,jn} \]

Multiply through to get

\[ \epsilon_{mni} \epsilon_{ijk} v_{k,jn} = \delta_{mj} \delta_{nk} v_{k,jn} - \delta_{mk} \delta_{jn} v_{k,jn} \]

As before, apply the substitution properties to obtain

\[ \epsilon_{mni} \epsilon_{ijk} v_{k,jn} = v_{k,mk} - v_{m,nn} \]

Recognize that \( v_{k,mk} \) is the gradient of the divergence of \({\bf v}\). This can be made clearer if the term is written more explicitly as \( (v_{k,k}),_j \). So \( v_{k,jk} = \nabla (\nabla \cdot {\bf v}) \).

The second term is just the Laplacian of \({\bf v}\), i.e., \(\nabla^2 {\bf v}\). This proves

\[ \nabla \times ( \nabla \times {\bf v} ) = \nabla ( \nabla \cdot {\bf v} ) - \nabla^2 {\bf v} \]

Also, please consider visiting an advertiser on this page. Doing so helps generate revenue to support this website.

Bob McGinty

bmcginty@gmail.com

Click here to see a sample page in each of the two formats.

Copyright © 2012 by Bob McGinty