In linear algebra, an alternant matrix is a matrix formed by applying a finite list of functions pointwise to a fixed column of inputs. An alternant determinant is the determinant of a square alternant matrix.
Generally, if
are functions from a set
to a field
, and
, then the alternant matrix has size
and is defined by

or, more compactly,
. (Some authors use the transpose of the above matrix.) Examples of alternant matrices include Vandermonde matrices, for which
, and Moore matrices, for which
.
- The alternant can be used to check the linear independence of the functions
in function space. For example, let
,
and choose
. Then the alternant is the matrix
and the alternant determinant is
. Therefore M is invertible and the vectors
form a basis for their spanning set: in particular,
and
are linearly independent.
- Linear dependence of the columns of an alternant does not imply that the functions are linearly dependent in function space. For example, let
,
and choose
. Then the alternant is
and the alternant determinant is 0, but we have already seen that
and
are linearly independent.
- Despite this, the alternant can be used to find a linear dependence if it is already known that one exists. For example, we know from the theory of partial fractions that there are real numbers A and B for which
. Choosing
,
,
and
, we obtain the alternant
. Therefore,
is in the nullspace of the matrix: that is,
. Moving
to the other side of the equation gives the partial fraction decomposition
.
- If
and
for any
, then the alternant determinant is zero (as a row is repeated).
- If
and the functions
are all polynomials, then
divides the alternant determinant for all
. In particular, if V is a Vandermonde matrix, then
divides such polynomial alternant determinants. The ratio
is therefore a polynomial in
called the bialternant. The Schur polynomial
is classically defined as the bialternant of the polynomials
.