Library / Advanced Mathematics
What Is Tensor Contraction?
Tensor contraction is the process of summing over one or more paired indices of a tensor expression.
It generalizes familiar operations such as inner products, matrix multiplication, and trace.
Definition
Summation Over Matched Indices
In index notation, contraction occurs when one index from one factor is matched with another index
and summed over. Matrix multiplication is a familiar example:
(AB)_{ij} = sum_k A_{ik} B_{kj}. The index k is contracted.
Tensor contraction is therefore not a mysterious special case. It is the natural higher-rank
extension of operations people already know from linear algebra.
Why It Matters
From Physics To AI
Tensor contraction appears in scientific computing, continuum mechanics, relativity, multilinear
algebra, and machine learning. Any time a system is built from matrix and tensor primitives, some
form of contraction is likely nearby.
It also matters computationally because there are often many ways to order a sequence of contractions.
Choosing that order well can dramatically change runtime and memory usage.
Example
Matrix Multiplication As Contraction
People often learn matrix multiplication without being told that it is a tensor contraction. But that
is exactly what it is. The product (AB)_{ij} sums over the repeated index
k, which means one axis from A and one axis from B are tied
together and eliminated by summation.
Seeing matrix multiplication this way helps unify familiar linear algebra with more general tensor
operations. The notation scales up naturally to higher-rank objects.
Optimization Question
Order Can Matter A Lot
In large tensor expressions, different contraction orders can produce the same mathematical result
with very different intermediate tensor sizes. That means contraction is not just a matter of
correctness. It is also a matter of computational cost, which is why symbolic reasoning and
cost-guided search are so relevant here.
This is one reason tensor contraction belongs in the same library as symbolic computation. Once a
tensor expression is represented explicitly, rewrite rules, equivalence reasoning, and extraction
strategies become relevant to performance-sensitive mathematical software.
Notation
Repeated Indices Encode Real Structure
Index notation is powerful because it makes the contraction pattern explicit. A repeated index marks
exactly which dimensions interact and which dimensions remain free. That clarity is one reason tensor
notation scales well from familiar matrix formulas to much larger multilinear expressions.
For software, that same clarity can become metadata for optimization. Once the index structure is
known, the system can reason about valid reorderings and better contraction schedules.
AI Relevance
Why This Topic Shows Up In Model Workloads
Modern AI systems are full of tensor operations even when users never see explicit index notation.
Attention mechanisms, batched matrix products, tensor reshaping, and fused kernels all sit near the
same underlying idea: structured operations on higher-rank arrays.
That makes tensor contraction more than an advanced-math curiosity. It is part of the conceptual
language behind real computational pipelines.