Abstract
A seminal result of Gerstenhaber gives the maximal dimension of a linear space of nilpotent matrices. It also exhibits the structure of such a space when the maximal dimension is attained. Extensions of this result in the direction of linear spaces of matrices with a bounded number of eigenvalues have been studied. In this paper, we answer what is perhaps the most general problem of the kind as proposed by Loewy and Radwan, by solving their conjecture in the positive. We give the maximal dimension of a vector space of matrices with no more than k<n eigenvalues. We also exhibit the structure of the spaces for which this dimension is attained.
COMMUNICATED BY:
2020 Mathematics Subject Classifications:
1. Introduction
This paper presents the positive solution to the Loewy–Radwan conjecture, which has been open for more than twenty years (Theorem 1.1). It belongs to a theory that started over 60 years ago by the famous Gerstenhaber [Citation1] result on linear spaces of nilpotent matrices of maximal dimension. An interested reader may also wish to consider some recent results in the area, such as [Citation2] by Kokol Bukovšek and Omladič, and [Citation3–5] by de Seguins Pazzis. It seems that spaces of matrices satisfying more general conditions on eigenvalues have been studied for the first time by Omladič and Šemrl in [Citation6]. We will first give a brief history of this theme.
Throughout the paper, we fix positive integers n and k<n and suppose V is a linear subspace of the space of complex matrices with the property that each member of V has at most k distinct eigenvalues. Here can be replaced by any algebraically closed field of characteristic zero. We are interested in how large the dimension of such a space can be. The case when all the matrices are assumed nilpotent dates back to Gerstenhaber [Citation1], who proved that the dimension of such a space is at most . Actually, Gerstenhaber proved the result for all fields with at least n elements, and this assumption was later removed (cf. Serežkin [Citation7], and Mathes, Omladič, Radjavi [Citation8]). Moreover, Gerstenhaber showed that when the maximal dimension is attained the space is simultaneously similar to the space of all strictly upper triangular matrices. Consequently, any space of matrices with only one eigenvalue has dimension at most , and in case of equality such a space is simultaneously similar to the space of upper triangular matrices with equal diagonal entries, see [Citation6] and also [Citation9] for a generalization of these results to other fields. The article [Citation6] also contains the maximal possible dimension for a vector space of matrices with at most k distinct eigenvalues when k = 2 and n is odd and when k = n−1 under some additional assumptions. In these two cases the spaces of maximal dimension were also classified in [Citation6].
Later, Loewy and Radwan [Citation10] removed the assumptions needed in [Citation6] and showed that the dimension of a vector space of matrices with most 2, respectively , distinct eigenvalues is at most , respectively . They also showed that for k = 3 the corresponding upper bound for the dimension is and conjectured that the upper bound is for every k<n. On the other hand, de Seguins Pazzis [Citation11] classified -dimensional spaces of matrices having at most 2 distinct eigenvalues and extended the results to other fields. We also note that Jordan algebras of matrices with few eigenvalues were studied in [Citation12] and that Gerstenhaber's theorem was generalized to semisimple Lie algebras [Citation13, Citation14], which was further used to give another proof of the Erdös-Ko-Rado theorem in combinatorics [Citation15].
The aim of this paper is to prove the Loewy–Radwan conjecture and to classify spaces of matrices with at most k distinct eigenvalues and which have the maximal possible dimension among such spaces. More precisely, we are going to show the following.
Theorem 1.1
Let n and k<n be positive integers and let V be a linear subspace of with the property that each member of V has at most k distinct eigenvalues. Then Moreover, if the equality holds and , then there exists such that V is simultaneously similar to the space of all matrices of the form (1) (1) where , and are arbitrary and is an arbitrary upper triangular matrix with equal diagonal entries.
The proof of this theorem is inspired by the proof of the generalization of Gerstenhaber's theorem to semisimple Lie algebras given in [Citation13]. Although our proof is technically more challenging, it is possible to adapt some of the main ideas from [Citation13] to our situation. Let us briefly explain these main ideas. To show that , we first observe that V belongs to some (projective) subvariety of a Grassmannian variety, which is invariant for the action by conjugation of the (solvable) group of invertible upper triangular matrices. The Borel Fixed Point Theorem then enables us to reduce the problem to linear subspaces that are invariant under conjugation by invertible upper triangular matrices. Such subspaces are spanned by diagonal matrices and matrix units, therefore their dimensions and the number of eigenvalues of their members can be estimated in a straightforward way.
To classify the subspaces of maximal dimension, we use an induction on k. We first show that our space must contain a nonderogatory (or cyclic) matrix with k−1 simple eigenvalues. With no loss of generality we assume that this matrix is in the Jordan canonical form. The rest of the proof is based on the following idea. We define a group homomorphism and consider the spaces and . These spaces are invariant for the action of the group , they have the same dimension as V and their members have at most k distinct eigenvalues. We first consider the structure of such spaces. Using the structure of -modules and the condition on the number of eigenvalues we can compute that the lower-right corner of such W has dimension exactly and its members have at most k−1 eigenvalues. For we use the inductive assumption, while for k = 3 we use the main result of [Citation11] together with the existence of a nonderogatory matrix with k−1 simple eigenvalues in W to obtain the structure of its lower-right corner. After that, we show that each such W is of an appropriate form. We apply this result to and , and finally we show that , which implies that V is equal to these two spaces and concludes the proof of the theorem. The above argument uses some results from representation theory, but we make it accessible to the reader by defining and in an equivalent way and then using only methods from linear algebra.
Section 2 consists of the proof of the first part of Theorem 1.1 and the rest of the paper is devoted to the second part. After showing some preliminary results in Section 3, in Section 4 we prove the second part of Theorem 1.1 under an additional assumption that for all . The general case is proved in Section 5.
2. The upper bound on the dimension
In this section we prove the first part of Theorem 1.1, i.e. we show that a linear subspace of whose members have at most k distinct eigenvalues has dimension at most . We will prove this with the help of algebraic geometry, therefore we first show that certain conditions on matrices are open in the Zariski topology. We call an matrix regular or cyclic or nonderogatory whenever its centralizer is n-dimensional. It is well known (see e.g. [Citation16, Section 3.2.4]) that this condition is equivalent to the condition that the characteristic and minimal polynomial of the matrix coincide or that the Jordan canonical form of the matrix has only one Jordan block for each eigenvalue.
Lemma 2.1
Let A be an complex matrix and k be a nonnegative integer. The following conditions are open in the Zariski topology.
(a) | A is regular. | ||||
(b) | A has more than k distinct eigenvalues. | ||||
(c) | A has more than k simple eigenvalues (i.e. they have algebraic multiplicity 1). |
Proof.
It is well known that , the centralizer of A, has dimension at least n (see e.g. [Citation16, Section 3.2.4]), so we may modify the defining condition of regularity into , meaning that , or equivalently , where is defined in the usual way by letting . This amounts to the same as requiring that at least one of the -minors of the matrix of the transformation in a fixed basis of is different from zero – clearly an open condition.
Let be the characteristic polynomial of A. Then it is well known and not hard to see that condition (b) is fulfilled if and only if the degree of gcd is smaller than n−k. It is also a classical result (see e.g. [Citation17, Section 2.1]) that this degree equals rank, where is the Sylvester matrix of the polynomials and . Let us recall that (in block partition made of n−1 respectively n rows) Here, the 's denote the coefficients of and the 's the coefficients of its derivative. So, the matrix A has more than k distinct eigenvalues if and only if the minors of the matrix are not all zero, which is clearly an open condition.
We will show that the condition that has at most k simple roots is closed. This condition is equivalent to the condition that the matrix has at most k nonzero eigenvalues (counted with algebraic multiplicities), which is further equivalent to , a closed condition.
Here we follow some concepts used in [Citation18, Section 3], so that we will give only the general ideas and omit some of the details. Fix an integer k, and define We will denote by m the maximal possible dimension of a vector space V such that . By Lemma 2.1(b) the set X is closed in the Zariski topology and it is clearly homogeneous (i.e. if and , then ), so we may view it as a projective variety. Following [Citation19, Example 6.19] we introduce the Fano variety which is a closed subset of the Grassmannian variety of all m-dimensional subspaces in , considered as a projective variety via the Plücker embedding that sends a vector space V with a basis into The variety is non-empty by the definition of the number m. Let be the solvable algebraic group of all invertible upper triangular matrices and define an action of on by It is obvious that X is invariant under conjugation, and consequently is invariant under the above action, which is given by regular (rational) maps: So, we can apply the following theorem to :
Borel fixed point theorem [Citation20, Theorem 10.4]: Let Z be a non-empty projective variety and G be a connected solvable algebraic group acting on it via regular maps. Then this action has a fixed point in Z.
Using this result we conclude.
Lemma 2.2
There exists a linear subspace such that for all and every invertible upper triangular matrix P.
We now investigate properties of a space satisfying the previous lemma. In the following lemma we denote the set by . The proof of this lemma may be found in [Citation18, Lemma 9]. We present all of it for the sake of completeness, some of it will be used in Lemma 2.4, the rest of it may be of independent interest.
Lemma 2.3
Let be a vector space such that for all and all invertible upper triangular . Then:
(a) | If for some and with i<j we have , then . | ||||
(b) | If for some with i<j we have , then for all with and . If we have for some with i>j, then the commutator belongs to V for all with p<q. | ||||
(c) | If V contains a matrix of the form in the block partition determined by dimensions , then it also contains the matrices and . | ||||
(d) | Claim (c) remains valid if we replace the first column and row with the ith column and row for any . | ||||
(e) | If for some and with we have , then . | ||||
(f) | If for some and with i<j we have , then . |
Using Lemma 2.3 we will now show that a subspace of which satisfies Lemma 2.2 has a very special form.
Lemma 2.4
Let V be a subspace of which satisfies the conditions of Lemma 2.2. Then there exists a sequence of standard subspaces of of dimension at least 2 with trivial intersections such that for each the span is invariant under all members of V, and:
(a) | For any index and any standard basis vectors we have . | ||||
(b) | All belong to V for . |
Proof.
Recall that V is of maximal possible dimension among spaces of matrices having at most k distinct eigenvalues. Let be a maximal chain of subspaces of which are invariant under all members of V and spanned by some of the first vectors among . If , let be spanned by standard basis vectors such that . Note that the inequality follows immediately.
Let be arbitrary index such that and and suppose that for all and all indices j<i. Lemma 2.3(e) and the second part of Lemma 2.3(b) then imply that for all and all p>i and q<i. (Indeed, if for some q<i<p, then by Lemma 2.3(e) and then by Lemma 2.3(b).) However, then the linear span of is a V-invariant subspace of which is strictly contained in (since it does not contain ) and strictly contains (since ), which contradicts maximality of the chain .
It follows that there is an index j<i (which necessarily satisfies ) such that for some . But then, using the second part of Lemma 2.3(b), we conclude in particular that the diagonal matrix is a member of V. Since i was arbitrary such that and , it follows that V contains all those diagonal matrices with trace zero whose only possible nonzero entries correspond to the subspace . General members of the corresponding diagonal block have the maximal number of eigenvalues equal to . Let be arbitrary and let be the diagonal blocks of A corresponding to the chain . We will show that the matrix has at most distinct eigenvalues. Assume the contrary. We have shown above that there exists a diagonal matrix such that its t-th diagonal block has trace zero and nonzero distinct eigenvalues, while all the other blocks are zero. General linear combination of A and D then has at least k + 1 distinct eigenvalues contradicting our assumption. So, the matrix has at most distinct eigenvalues. Then it is clear that any linear combination of A and , where , has at most k distinct eigenvalues. By maximality of the dimension of V it follows that concluding property (a).
Property (b) now follows easily by maximality of the dimension of V.
In the situation of Lemma 2.4 let for t such that , and let otherwise. Furthermore, let l be the dimension of the space of diagonal members of V that correspond to the standard basis vectors which do not belong to any of the subspaces . Note also that V is invariant under projection on the diagonal by Lemma 2.4.
Corollary 2.5
Then .
Proof.
The maximal number of distinct eigenvalues of a member of V is clearly a sum of all and the maximal number of distinct eigenvalues of a matrix from . The last number cannot be smaller than l, as is l-dimensional.
The proof of the main result of this section will be based on the above corollary and the following lemma.
Lemma 2.6
Let k, l, r be positive integers and nonnegative integers satisfying . Then The equality holds if and only if for some t and either l = 1 or k = l = 2.
Proof.
The second entry on the left hand side, multiplied by 2, can be estimated and it is clear that the difference between the right hand side and the left hand side of the first inequality above is equal to which equals zero only if no more than one number is nonzero. Moreover, the equality in the second inequality holds only if . It follows that where we used the fact that . We have equality in the last two inequalities only if l = 1 or k = l = 2.
In the following theorem, we prove that which proves the first part of Theorem 1.1 and solves [Citation10, Conjecture 1.2]. In the theorem we also characterize subspaces where equality holds in Theorem 1.1 and which additionally satisfy conditions of Lemma 2.2 when . We will use this characterization in Lemma 3.2 which is one of the key steps in the characterization of all subspaces where the upper bound in Theorem 1.1 is achieved.
Theorem 2.7
Let and let m be the maximal possible dimension of a subspace of whose all elements have at most k distinct eigenvalues. Then Moreover, if and is a subspace satisfying conditions of Lemma 2.2, then there exists such that V consists of all matrices where , and are arbitrary and is an arbitrary upper triangular matrix with equal diagonal entries.
Proof.
The spaces of matrices described in the theorem are clearly invariant under conjugation by invertible upper triangular matrices, they have dimension and for each they consist of matrices with at most k distinct eigenvalues, so . To prove the converse, by the Borel fixed point theorem it suffices to show that spaces satisfying conditions of Lemma 2.2 have dimension at most . Let V be such a space and let l and be defined as before Corollary 2.5. Note that l is positive, as scalar matrices are in V by the maximality of the dimension of V. We compute where the first entry on the right hand side counts the diagonal elements of V corresponding to standard basis vectors that do not belong to any of the subspaces , the second one counts the entries strictly above the diagonal, and the terms of the third one count the entries below the diagonal and on it corresponding to each of the subspaces . The first part of the theorem now immeadiately follows from Lemma 2.6. Moreover, the equality in the inequality of the lemma holds only if r = 1, for some t and either l = 1 or k = l = 2. In particular, if , then l = 1, r = 1 and , which gives us the possibilities for V described in the theorem.
3. Preliminaries for the structure result
In our considerations, we will often refer to properties that hold generically. As usual in the algebraic geometry, this will mean that the property holds on an open dense subset (in the Zariski topology). Most often generic conditions will be considered on some line. In this case, a property will hold generically if it will hold for all but a finite number of points of the line.
In the proof of the second part of Theorem 1.1 we will need the following condition more than once.
Two zeros condition: Let n be a positive integer, let s be a polynomial of degree no more than n, and let be fixed. For generic the polynomial has no more than two distinct zeros.
Lemma 3.1
The two zeros condition implies that s = 0.
Proof.
Let . By the condition under consideration, has at most two distinct zeros for all but a finite number of scalars μ. Besides, this polynomial has a simple zero at . According to Lemma 2.1(c) (applied e.g. to the companion matrix of a polynomial) the condition that a monic polynomial of given degree has a simple zero is open, so that has a simple zero generically, i.e. for all but a finite number of . In the rest of the proof, we consider such μ that has a simple zero and at most two distinct zeros. For all these values of μ we can write , where α and β may depend on μ. The first Vieta formula determines α as a linear function of β, i.e. Insert this expression into the second and the third Vieta formula to get two polynomial conditions (2) (2) (3) (3) in β and μ. For a fixed as chosen above the polynomial equations (Equation2(2) (2) ) and (Equation3(3) (3) ) have a common solution β, so the resultant (i.e. the determinant of the Sylvester matrix) of these two polynomials has to be zero. (Note that this is a special case of the result from [Citation17] used in Lemma 2.1, or see some standard textbook on algebraic geometry such as [Citation21, Section 3.5].) Since this condition is satisfied for general μ as considered above, all the coefficients at powers of μ in the obtained resultant must be zero. Since μ appears only in the last three columns of the above determinant, the degree of the resultant is at most 3, and it has zero constant term, since all entries of the last column of the determinant are multiples of μ. Now it is clear that the coefficient at μ equals and the coefficient at equals It follows that and . Using (Equation2(2) (2) ) and (Equation3(3) (3) ) one then concludes that and independently of the chosen μ and hence s = 0.
In the proof of Theorem 1.1 we will also need the fact that for a space V of maximal dimension contains a regular matrix that has exactly k−1 simple eigenvalues, i.e. is similar to (4) (4) where all the empty entries of this matrix are zeros and are pairwise distinct.
Lemma 3.2
Let and let V be a subspace of of dimension whose members have at most k distinct eigenvalues. Then V contains a regular element that has k−1 simple eigenvalues.
Proof.
Assume the contrary. Let Y be the set of all matrices that are not regular, and let Z be the set of all matrices with at most k−2 simple eigenvalues. By Lemma 2.1 the sets Y and Z are both closed in the Zariski topology, so is their union. Moreover, the union is clearly invariant under the action of the group of all invertible upper triangular matrices by conjugation. Furthermore, the Zariski closed set is clearly homogeneous, so we may view it as a projective variety. Hence, we can introduce the Fano variety of the union . The assumption that V does not contain a regular element that has k−1 simple eigenvalues implies that the intersection is not empty. This intersection is invariant under , so by the Borel fixed point theorem it has a fixed point . However, since , is then one of the spaces described in Theorem 2.7 and it contains a regular element with k−1 simple eigenvalues, which is similar (with a similarity that swaps the first two block rows and columns of (Equation1(1) (1) )) to a matrix of the form (Equation4(4) (4) ) with pairwise distinct, contradicting the starting assumption on .
4. Structure of some special spaces
Throughout the rest of the paper let . We will prove the main result by induction on k. Let V be a space of maximal dimension satisfying the conditions of Theorem 1.1. Note that by maximality we may assume that V contains all scalar matrices. First we make a reduction that is based on Lemma 3.2. Each regular matrix with k−1 simple eigenvalues which has at most k distinct eigenvalues is similar to a matrix of the form (Equation4(4) (4) ), so we now conjugate the space V by an appropriate invertible matrix to assume that V contains a matrix of the form (Equation4(4) (4) ) for some distinct .
In this section, we describe the structure of the space considered under the following additional assumption.
AA: If a matrix with blocks of respective sizes 1 and n−1 belongs to V, then the matrices , , and belong to V.
This additional assumption is motivated by Representation Theory. Denote by the multiplicative group . Let be a group homomorphism defined by . For each the space is m-dimensional and with elements having at most k distinct eigenvalues, so it belongs to .
Lemma 4.1
Condition AA is equivalent to for all .
Proof.
Condition AA clearly implies . Conversely, a space satisfying this equality is a -module for the action . Now we use the fact that every -module is a direct sum of weight spaces (see e.g. [Citation22, Proposition 22.5.2(iii)]) to see that V can be written as Write a matrix as where , and . It follows easily that members of are nonzero only in the case that j = 0, j = 1, or j = −1. Elements of are of the form , elements of are of the form , and elements of are of the form , after a straightforward computation.
Our next step will be to estimate the dimensions of when . Here is an additional notation we need to introduce. Let be the set of all matrices of the form such that b has first k−2 entries equal to zero. We define similarly .
Lemma 4.2
Let be the smallest index with the property that some row with equals the upper-right corner of a member of (with the convention l = n + 1 if b is always zero). Furthermore, let be an arbitrary lower-left corner of a member of . Then for all .
Proof.
Clearly we may assume . Choose an arbitrary with and . As explained at the beginning of this section V contains a matrix of the form (Equation4(4) (4) ) for some distinct so that it contains for some distinct and arbitrary . Our assumptions imply that this matrix has at most k distinct eigenvalues so that its characteristic polynomial , which is computed in Lemma 4.3 below, has at most k distinct zeros for arbitrary . As shown in Lemma 4.3 the polynomial is the product of and a polynomial of the form for some polynomial . By the assumption the numbers are not zeros of the polynomial , hence they are not zeros of for generic . Consequently, the polynomial has at most two distinct zeros for generic μ. Now we replace t by and divide by to get the following. For generic the polynomial satisfies the Two zeros condition. Using Lemma 3.1 we conclude that (5) (5) for all . Recalling that l is the smallest index with , the equalities (Equation5(5) (5) ) imply that for all as desired.
Lemma 4.3
Proof.
First observe that the columns indexed by contain only one nonzero entry which equals . Let us perform the usual column expansions along all of these columns consecutively to conclude that where We compute by expanding it first along the first row and then along the first column. The final minor is possibly nonzero only in the case of the (p−k + 2)-th column and (q−k + 2)-th row, for , such that , in which case it equals . So, which gives the desired result after a small computation.
Corollary 4.4
.
Proof.
The conclusion of Lemma 4.2 implies easily the desired estimates.
Corollary 4.5
.
Proof.
This follows immediately from Corollary 4.4. Indeed, and the desired inequality follows.
We now recall that . It follows by Corollary 4.5 that Recall that is a linear space of matrices of the form , whose lower-right corners form a space, which we denote by W, of dimension no smaller than . On the other hand, members of W have no more than k−1 distinct eigenvalues. Indeed, if had k distinct eigenvalues, then some linear combination of the corresponding matrix and some matrix of the form (Equation4(4) (4) ) would lie in V and have at least k + 1 distinct eigenvalues, a contradiction. Using Theorem 2.7 we can therefore conclude that the dimension of W is exactly . This fact implies that in all the above inequalities up to and including Corollary 4.4 we have equalities, more precisely:
Corollary 4.6
(a) | |||||
(b) | |||||
(c) |
This proves that the space of lower-left corners of equals the span of . Using this result we now show a version of Lemma 4.2 in which the roles of b and c are interchanged.
Lemma 4.7
An arbitrary upper-right corner of a member of is of the form .
Proof.
If l>k and is arbitrary, then a matrix where are nonzero and pairwise distinct and μ appears in the -st row, belongs to V for all . The characteristic polynomial of this matrix equals where Since the matrix defined above belongs to V, it has at most k distinct eigenvalues, and as in Lemma 4.2 we conclude that the polynomial has at most two distinct zeros for generic . Lemma 3.1 (applied to ) then implies that , so for , as desired.
Recall that is a subspace of dimension whose members have at most k−1 distinct eigenvalues. If , it now follows from the inductive hypothesis that there exists such that the members of W are simultaneously similar to matrices of the form (Equation1(1) (1) ), i.e. , with blocks of respective sizes p, k−2, n−p−k + 1, where is upper triangular with constant diagonal, but other than that the nonzero blocks are arbitrary. If k = 3, then there are more similarity classes of -dimensional spaces of matrices having at most 2 eigenvalues, see [Citation11, Theorem 1.8]. However, the space W contains the lower-right corner of the matrix given by (Equation4(4) (4) ), which has a simple eigenvalue. This additional information together with [Citation11, Theorem 1.8] implies that the members of W are simultaneously similar to matrices of the form (Equation1(1) (1) ) even if k = 3.
The next step is to prove that W is actually equal to the space of matrices obtained from (Equation1(1) (1) ) by interchanging the first two block rows and columns.
Lemma 4.8
The space W of all lower-right corners of is equal to the space of all matrices of the form with blocks of respective sizes k−2, p and n−p−k + 1 for some where is the sum of a scalar matrix and a strictly upper triangular matrix, and all the other nonzero blocks are arbitrary.
Proof.
Recall that the space W is simultaneously similar to the space of matrices described in the lemma. Let P be an invertible matrix that provides this similarity. The space W contains the lower-right corner of a regular matrix of the form (Equation4(4) (4) ) such that Here, A and F are upper triangular with the same constant, say λ, on the diagonal. Denote this matrix of the form (Equation4(4) (4) ) by L and the block matrix on the right by M. Since the two matrices are similar and is the only multiple eigenvalue of L and λ is a multiple eigenvalue of M, we have that . Consequently, the eigenvalues of D are . Write P with blocks of respective sizes k−2, p, n−p−k + 1 as to get (6) (6) where and are nilpotent Jordan blocks of appropriate sizes. The -block of Equation (Equation6(6) (6) ) equals . Since the intersection of the spectra of and F is empty, we conclude that X = 0 [Citation23, Section VIII.1]. We rewrite the - and -block of Equation (Equation6(6) (6) ) into where J is a nilpotent Jordan block of appropriate size. It follows inductively on l that So, the matrix maps into which is included in for all positive integers l. This readily yields that Y = 0 and that Z is upper triangular.
Next, we consider the -block of equation (Equation6(6) (6) ) to get . Since the spectrum of D does not contain λ, we determine that R = 0. Block equation now implies that . As above we deduce that U is upper triangular. Finally, we conclude that W is the space of all matrices of the form where and are strictly upper triangular. Indeed, we know that space W is simultaneously similar to the space of matrices of this form, while we proved here that a similarity matrix is also of this block form with and -blocks upper triangular.
Let us now write matrices with respect to the block partition of respective sizes 1, k−2, p and n−p−k + 1. Then it follows from the above lemma and the equality that consists of all matrices of the form (7) (7) where are arbitrary and are strictly upper triangular. Also, respectively consists of some matrices of the form We now determine the structure of the spaces and .
Lemma 4.9
For any matrix in V the blocks and are zero.
Proof.
Here is a simplified notation for the first row and column that will be useful As in Lemma 4.2 let be the smallest index such that some row with equals the upper-right corner of a member of , with the convention l = n + 1 if is trivial. By that lemma the entries of any lower-left corner of a member of satisfy for all . First, we want to show that Towards a contradiction, we assume that l<k + p. In particular, we have p>0 and . Choose a matrix of the form (Equation7(7) (7) ) with , , where 's are nonzero and distinct, for some s, , and all the other blocks are zero. We add c and described above to this matrix and compute the resulting characteristic polynomial observe that the matrix under this determinant belongs to V. The isolated entry 1 was set in the l-th row and the s-th column. We expand the determinant at all rows and columns that contain only one nonzero entry: where we introduce So, if , then has k distinct zeros. Consequently, by Lemma 2.1(b) the polynomial has k distinct zeros for generic μ. If for such μ, then the polynomial Δ has k + 1 zeros, a contradiction with the standing assumption on V. Therefore, generically the polynomial has k distinct roots, one of which is zero. This implies that , and since we have . Now, in this consideration s is chosen arbitrary from the set , so are all equal to zero.
We have shown that any first column of a member of is of the form , so that , contradicting Corollary 4.6. This shows that Condition holds.
Next, we want to show that l = k + p. Choose c with 1 in the -th position (i.e. ) and zeros elsewhere. Assume towards a contradiction that l>k + p (and hence and l>k) and repeat the above arguments with the roles of and interchanged. Consider a member of V of the form (Equation7(7) (7) ) with and where are nonzero and pairwise distinct, and with a 1 in the -st column and s-th row for , and with zeros everywhere else. Add to this matrix c and where is an arbitrary upper-right corner of and c is as above. Recall that by Lemma 4.7. Computations as above reveal that the characteristic polynomial of this matrix is equal to where the last determinant on the right hand side equals . As before we conclude that for all possible , which implies . The contradiction so obtained brings us to the fact that l = k + p, which proves the lemma.
The above lemma concludes the proof that a space containing a matrix of type (Equation4(4) (4) ) and satisfying condition AA consists of all matrices of the form with blocks of respective sizes k−1, p and n−p−k + 1, where is an upper triangular matrix with equal diagonal entries and all the other nonzero blocks are arbitrary.
5. Structure of the spaces of maximal dimension
In this section, we will prove the second part of Theorem 1.1, i.e. the following theorem.
Theorem 5.1
Let and let V be a subspace of of dimension such that each member of V has at most k distinct eigenvalues. Then there exists such that V is simultaneously similar to the space of all matrices of the form (Equation1(1) (1) ) where , and are arbitrary and is an arbitrary upper triangular matrix with equal diagonal entries.
Let V be a space satisfying the conditions of the theorem. Recall from the beginning of the previous section that we may assume that V contains a matrix of the form (Equation4(4) (4) ) for some distinct . Consider a block partition of V with respect to dimensions 1 and n−1. We define the projection from V to the lower-left corner as . Clearly, consists of all members of V of the form . Next we define the projection from to the diagonal blocks as . Now, consists of all members of V of the form . Let It is clear that .
Lemma 5.2
Elements of have at most k distinct eigenvalues.
Proof.
Let be an arbitrary matrix in . Then and . So there exists such that . Finally, , so there exist , and such that . Members of V have at most k distinct eigenvalues, therefore the matrix has at most k distinct eigenvalues for each . Consequently, the starting matrix has at most k distinct eigenvalues by Lemma 2.1(b).
Next, we define the projection from V to the upper-right corner as . Clearly, consists of all members of V of the form . We define the projection from to the diagonal blocks as . Now, consists of all members of V of the form . Let so that again . Similar arguments as in the proof of Lemma 5.2 show that all members of have at most k distinct eigenvalues. Let us point out that the so-defined spaces and satisfy condition AA from the beginning of Section 4.
Remark 5.3
Motivation for the definition of spaces and comes from Algebraic Geometry and Representation Theory. Recall the group homomorphism defined by . For each the space is m-dimensional and with elements having at most k distinct eigenvalues, so it belongs to the Fano variety . However, is a projective variety, hence there exist limits in . A short computation reveals that is -stable; indeed, The same considerations apply to . Therefore, both spaces are -modules for the action . Observe that this is equivalent to Condition AA.
Remark 5.4
To show that the two definitions of are equivalent choose the following basis of the space V : Here, r and s are chosen consecutively as the maximal possible so that are linearly independent and that are linearly independent. We denote the basis elements by . Recall that V is represented in by the class , which is independent of the choice of the basis (cf. [Citation19, Chapter 6]). In order to get the basis of space , we compute the limits of classes within the Grassmanian determined by exterior products of basis elements:
Note that the elements of the above exterior product are indeed linearly independent, so a basis of is given by Note that the first r elements form a basis of , the next s elements form a basis of , and the rest of the elements form a basis of . So, the two definitions of are equivalent. The same considerations apply to .
Recall that the spaces and satisfy Condition AA, so we may apply the results of Section 4. In the block partition with respect to blocks of sizes 1, k−2, p and n−p−k + 1 the upper-right corner of linear space respectively is made of vectors of the form respectively . Also, the lower-left corner of the linear space respectively is made of all vectors of the form respectively . So, the linear space consists of all matrices of the form , where and are strictly upper triangular. The case of the space goes in the same way. However, the block division there may be based on a different index denoted by q instead of p. We now want to show that these indices are equal and that consequently .
Proposition 5.5
.
Proof.
Recall the definitions of the projections and . Then and Note that and that is injective on . Consequently, , or equivalently . It was shown before the proposition that so . We want to show that the equality holds.
Write elements of V, and with respect to block partition of respective sizes 1, k−2, q, p−q and n−k−p + 1 (where some of these numbers may be zero). Then, sets and consist of matrices of the form where and are strictly upper triangular. So, members of are of the form (8) (8) where and are strictly upper triangular. An easy computation reveals that (9) (9) Let be the projection defined by , where ; the blocks of the first matrix are of sizes 1 and n−1 and the blocks of are of the sizes and n−p−k + 1. It is obvious that . So, φ induces a linear map . Let be arbitrary. Then there exists b such that . Write . Then . Recall that is a projection from V to the upper-right corner. The structure of implies that all (1,3)-blocks of matrices from V are zero. In particular, . If we subtract from the matrix (10) (10) the difference lies in . It follows that for each the matrix (Equation10(10) (10) ) lies in V. In particular, . If then , so . Since is identity on , it follows that . Consequently, and therefore Combining this inequality with (Equation9(9) (9) ) we get .
If , it now immediately follows that p = q and hence . In particular, and , which implies that V contains all upper-right and lower-left corners of its elements with respect to block partition . Therefore .
It remains to get a contradiction in the case k = 3 when p>q. In this case, all the above dimension inequalities become equalities. This means that is surjective and its kernel is , so the induced map is an isomorphism. It follows that for each the matrix belongs to V.
To obtain a contradiction we now adjust the ideas of Lemmas 4.2 and 4.9. Assume first that . Fix distinct nonzero numbers and . In the above observation take and write . Note that and . It follows that for an arbitrary the matrix belongs to V for each . Hence it has at most three distinct eigenvalues. Its characteristic polynomial is equal to By the assumption on V the quartic polynomial in the parentheses has a multiple zero for each , so for each its discriminant is zero. The above discriminant is a polynomial of degree 4 in μ with constant term zero. The coefficient on μ is equal to and the coefficient on is equal to As both coefficients have to be zero, we get . Since i was an arbitrary element from , we get x = 0, a contradiction with the fact that and is an isomorphism.
On the other hand, if p−q = 1, then we consider a similar matrix as above, with the only difference that the -block is taken to be zero. The characteristic polynomial of this matrix equals The same argument as above shows that , i.e. x = 0, which yields again a contradiction. This concludes the proof of the proposition.
Finally, we can finish the proof of Theorem 1.1. By Proposition 5.5 the space V is equal to . Consequently, by the argument just before Proposition 5.5 there exists such that with respect to the block partition of respective sizes k−1, p, n−k−p + 1 the space V consists of all matrices of the form , where A and F are strictly upper triangular and all other nonzero blocks are arbitrary. A similarity that exchanges the first two block rows and columns now brings the space V into the form (Equation1(1) (1) ).
Acknowledgments
The authors are indebted to an anonymous referee for numerous suggestions that helped us improve substantially the presentation and organization of this manuscript.
Disclosure statement
No potential conflict of interest was reported by the author(s).
Additional information
Funding
References
- Gerstenhaber M. On nilalgebras and linear varieties of nilpotent matrices I. Am J Math. 1958;80:614–622. doi: 10.2307/2372773
- Kokol Bukovšek D, Omladič M. Linear spaces of symmetric nilpotent matrices. Linear Algebra Appl. 2017;530:384–404. doi: 10.1016/j.laa.2017.05.030
- de Seguins Pazzis C. The structured Gerstenhaber problem (I). Linear Algebra Appl. 2019;567:263–298. doi: 10.1016/j.laa.2018.08.015
- de Seguins Pazzis C. The structured Gerstenhaber problem (II). Linear Algebra Appl. 2019;569:113–145. doi: 10.1016/j.laa.2018.11.033
- de Seguins Pazzis C. The structured Gerstenhaber problem (III). Linear Algebra Appl. 2020;601:134–169. doi: 10.1016/j.laa.2020.04.021
- Omladič M, Šemrl P. Matrix spaces with bounded number of eigenvalues. Linear Algebra Appl. 1996;249:29–46. doi: 10.1016/0024-3795(95)00253-7
- Serežkin VN. Linear transformations preserving nilpotency. Vesti Akad Navuk BSSR Ser Fiz Mat Navuk. 1985;5:46–50.
- Mathes B, Omladič M, Radjavi H. Linear spaces of nilpotent matrices. Linear Algebra Appl. 1991;149:215–225. doi: 10.1016/0024-3795(91)90335-T
- de Seguins Pazzis C. Spaces of matrices with a sole eigenvalue. Linear Multilinear Algebra. 2012;60:1165–1190. doi: 10.1080/03081087.2011.654118
- Loewy R, Radwan N. On spaces of matrices with a bounded number of eigenvalues. Electron J Linear Algebra. 1998;3:142–152. doi: 10.13001/1081-3810.1020
- de Seguins Pazzis C. Spaces of matrices with few eigenvalues. Linear Algebra Appl. 2014;449:210–311. doi: 10.1016/j.laa.2014.02.015
- Grunenfelder L, Košir T, Omladič M, et al. Maximal Jordan algebras of matrices with bounded number of eigenvalues. Isr J Math. 2002;128:53–75. doi: 10.1007/BF02785418
- Draisma J, Kraft H, Kuttler J. Nilpotent subspaces of maximal dimension in semi-simple lie algebras. Compos Math. 2006;142:464–476. doi: 10.1112/S0010437X05001855
- Meshulam R, Radwan N. On linear subspaces of nilpotent elements in a lie algebra. Linear Algebra Appl. 1998;279:195–199. doi: 10.1016/S0024-3795(98)10010-1
- Woodroofe R. An algebraic groups perspective on Erdös-Ko-Rado. Linear Multilinear Algebra. 2022;70:7825–7835. 10.1080/03081087.2021.2013428.
- Horn RA, Johnson CR. Matrix analysis. 2nd ed. New York: Cambridge University Press; 2013.
- Krein MG, Naimark MA. The method of symmetric and Hermitian forms in the theory of the separation of the roots of algebraic equations. Linear Multilinear Algebra. 1981;10:265–308. doi: 10.1080/03081088108817420. English translation from the Russian of the paper originally published in Kharkov (1936).
- Omladič M, Radjavi H, Šivic K. On approximate commutativity of spaces of matrices. Linear Algebra Appl. 2023;676:251–266. doi: 10.1016/j.laa.2023.06.026
- Harris J. Algebraic geometry: a first course. New York: Springer-Verlag; 1992. (Graduate texts in mathematics; Vol. 133).
- Borel A. Linear algebraic groups. New York: Springer-Verlag; 1991. (Graduate Texts in Mathematics; Vol. 126).
- Cox D, Little J, O'Shea D. Ideals, varieties, and algorithms: an introduction to computational algebraic geometry and commutative algebra. 3rd ed. New York: Springer; 2007. (Undergraduate texts in mathematics).
- Tauvel P, Yu RWT. Lie algebras and algebraic groups. Berlin, Heidelberg: Springer-Verlag; 2005. (Springer monographs in mathematics).
- Gantmacher FR. The theory of matrices. Rhode Island: Providence; 1959. (AMS Chelsea; Vol. 1).