Multilinear Transformations
We introduce now multilinear transformations, which are functions that act on multiple vectors, each in a linear way. Recall from Section 1.2.3 that a linear transformation is a function that sends vectors from one vector space to another in a linear way, and that a multilinear form is a function that sends a collection of vectors to a scalar in a manner that treats each input vector linearly. Multilinear transformations provide the natural generalization of both of these concepts—they can be thought of as the sweet spot in between linear transformations and multilinear
forms where we have lots of input spaces and potentially have a non-trivial output space as well.
Multilinear transformations
provide the natural generalization of both of these concepts.
Definition 7.4.1 (Multilinear Transformations) Suppose V1, V2 , ..., Vp and W are vector spaces over the same field. A multilinear transformation is a function T : V1 × V2 × ··· × Vp → W with the property that, if we fix 1 ≤ j ≤ p and any p − 1 vectors vi ∈ Vi (1 ≤ i ≠ j ≤ p), then the function S : V j → W defined by
S(v) = T (v1 , ... ,vj−1, v, vj+1, ..., vp) for all v ∈ Vj
is a linear transformation. □
The idea is simply that a multilinear transformation is a function that looks like a linear transformation on each of its inputs individually. When there are just p = 2 input spaces we refer to these functions as bilinear transformations, and we note that bilinear forms, are the special case that arises when the output space is W = 𝔽. Similarly, we sometimes call a multilinear transformation with p input spaces a p-linear transformation (much like we sometimes called multilinear forms p-linear forms).
In other words a map f: V x W ⟶ U is said to be a bilinear map from V and W to W if for all v1,v2 ∈ V and w1, w2 in W we have
It is linear in its first argument:
f(v1 + v2, w) = f (v1, w) + f(v2, w) and
f (cv1, w) = cf (v1, w) for all c ∈ 𝕂, v1, v2 ∈ V, and w ∈ W.
It is linear in its second argument:
f (v, w1 + w2) = f (v, w1) + f(v, w2) and
f (v, cw1) = cf(v, w1) for all c ∈ 𝕂, v ∈ V, and w1, w2 ∈ W. □
Definition of cross product in ℝ3
In physics you may have encountered laws such as F = mv x B, that is a bilinear map which assigns to two vectors a third which is orthogonal to either. This map × : ℝ3 × ℝ3 → ℝ3 is called the cross product or vector product and is written as (v, w) ⟼ v × w. Apart from linearity in each of its two arguments we would like the cross product to be anti-symmetric, that is, v × w = −w × v, so that the orthogonal vector points into the opposite direction when the order of the arguments is changed. Finally, given that the standard unit vectors are mutually orthogonal it makes sense to demand that the cross product of two standard unit vectors gives the third. Altogether this motivates the following definition.
Definition 7.4.2 (Cross product) A map × : ℝ2 × ℝ3 → ℝ3 is called a cross product if is satisfied the following conditions for all vectors v, w, u ∈ ℝ3 and all scalars α, β ∈ ℝ.
v × w = −w × v (anti-symmetry)
v × (w + u) = v × w + v × u (linearity in the 2nd argument)
v × (αw) = α(v × w)
e1 × e2 = e3, e2 × e3 = e1, e3 × e1 = e2 (orthogonality)
While (ii) demands linearity in the second argument it is clear, by combining (ii) with (i), that the cross product is also linear in the first argument and, hence, that it is bi-linear. This means that ℝ3 with the cross product forms an algebra, in the sense of Def. 7.4.3. Anti-symmetry implies that v × v = −v × v, so the cross product of any vector with itself vanishes, that is,
v × v = 0 for all v ∈ ℝ3.
The cross product can also be expressed as the formal determinant:
Using cofactor expansion along the first row instead, it expands to
Problem (Cross product) Work out the cross product of the vectors v = (1, 3, 4) and w = (2, 7, −5). Solution: The cross product is
v x w = −43 e1 + 13 e2 + e3.
Definition 7.4.3 An algebra (V, 𝔽, +, ·, ∗) is a vector space (V, 𝔽, +, ·) with a multiplication ∗ : V × V → V which satisfies the following properties, for all v1, v2, w ∈ V and all α1, α2 ∈ 𝔽.
(α1 v1 + α2 v2) ∗ w = α1 (v1 ∗ w) + α2 (v2 ∗ w) (linear in first argument)
w ∗ (α1 v1 + α2 v2) = α1 (w ∗ v1) + α2 (w ∗ v2) (linear in second argument)
If there is a e ∈ V with e ∗ v = v ∗ e = v for all v ∈ V the algebra is called an algebra with unit. If the product ∗ is associative, the algebra is called an associative algebra. In short, an algebra is a vector space with a multiplication which is bi-linear. We will not investigate algebras systematically but, occasionally, it will be useful to point to the above definition when we come across examples of algebras. One simple such example is the vector space ℝ3 with the cross product as multiplication;
Example 7.4.4 Matrix multiplication is a bilinear transformation. That is, if we define the function T : Mm,n × Mn,p → Mm,p that multiplies two matrices together via
T(A, B) = AB for all A ∈ Mm,n, B ∈ Mn,p,
then T× is bilinear. To verify this claim, we just have to recall that matrix multiplication is both left- and right-distributive, so for any matrices A, B, and C of appropriate size and any scalar c, we have
T (A + cB,C) = (A + cB)C = AC + cBC = T (A,C) + cT (B,C) (linear in 1st argument)
T(A, B + cC) = A(B + cC) = AB + cAC = T (A, B) + cT (A,C) (linear in 2nd argument)
Matrix multiplication also become multilinear transformations if we extend them to three or more inputs in the natural way. For example, the function T defined on triples of matrices via T (A,B,C) = ABC is a multilinear (trilinear) transformation. ■