Rodney Canfield Department of Computer Science University of Georgia Athens, GA 30602, USA erc@cs.uga.edu Submitted: August 30, 1996; Accepted: November 27, 1996 Abstract The weight of a
Trang 1of Minimal Weight Vectors is Asymptotically Poisson
Edward A Bender Center for Communications Research
4320 Westerra Court San Diego, CA 92121, USA ed@ccrwest.org
E Rodney Canfield Department of Computer Science University of Georgia Athens, GA 30602, USA erc@cs.uga.edu Submitted: August 30, 1996; Accepted: November 27, 1996
Abstract
The weight of a vector in the finite vector space GF(q)n is the number of nonzero components it contains We show that for a certain range of parameters (n, j, k, w) the number of k-dimensional subspaces having j(q− 1) vectors of minimum weight
w has asymptotically a Poisson distribution with parameter λ =¡n
w
¢ (q−1)w −1qk −n.
As the Poisson parameter grows, the distribution becomes normal
AMS-MOS Subject Classification (1990) Primary: 05A16 Secondary: 05A15, 11T99
Trang 21 Introduction
Almost all the familiar concepts of linear algebra, such as dimension and linear independence, are valid without regard to the characteristic of the underlying field
An example of a characteristic-dependent result is that a nonzero vector cannot be orthogonal to itself; researchers accustomed to real vector spaces must modify their
“intuition” on this point when entering the realm of finite fields
Let q be a prime power, fixed for the remainder of the paper, and GF(q) be the finite field with q elements Because the underlying field is finite, there are many counting problems associated with fundamental concepts of linear algebra; for example, how many subspaces of dimension k are there in the vector space GF(q)n ? The answer is often denoted £n
k
¤
q, and we have
· n k
¸
q
= (1− qn)(1− qn −1)· · · (1 − qn −k+1)
(1− qk)(1− qk −1)· · · (1 − q) ,
the Gaussian polynomial The reader may consult [2] for an introduction to the subject
Define the weight of a vector v in GF(q)n to be the number of nonzero coordi-nates in v The interaction of weight with familiar concepts of linear algebra yields more and harder counting problems Consider the n vectors of weight 1 in GF(2)n; how many vector spaces of dimension k do they span ? The easy answer is the well known binomial coefficient¡n
k
¢ Now consider the¡n
2
¢ vectors of weight 2 in GF(2)n; how many vector spaces of dimension k do they span ? More thought is needed this time, but again the answer is a classical array from combinatorics, the Stirling numbers of the second kind S(n, n− k) If we ask the same question for weight 3
or higher, no simple answer is known and the numbers cannot be computed easily However, that familiar properties of ¡n
k
¢ and S(n, k) persist in the higher weight version is part of a sweeping conjecture that the Whitney numbers of the second kind for any geometric lattice are log-concave [1, p.141]
Extend the notion of weight to subspaces by saying that a subspace V ⊆ GF(q)n
has weight w if w is the minimum weight of all nonzero vectors in V A natural problem is to describe how the £n
k
¤
q subspaces of dimension k are distributed by weight We cannot give a definitive solution to this question, but using asymptotic methods, we can gain some insight into the problem The number of weight w vectors in a vector space over GF(q) is a multiple of q− 1 since multiplication by
a nonzero scalar preserves weight Let p(j; n, k, w) be the fraction of k-dimensional subspaces of GF(q)n containing j(q− 1) vectors of weight w, and no nonzero vector
of weight less than w Masol [3] showed that
p(0; n, k, 1)− e−λ → 0 uniformly as n → ∞ where λ = nqk −n.
We extend this result as follows:
Trang 3Theorem 1 Fix a prime power q, positive constants b≤ 1
2 and B < 1− b, and a function µ(n) = o(1) Then, uniformly for j, k, w satisfying
1 ≤ w ≤ µ(n)nb/ logqn and λ def≡
µ n w
¶ (q− 1)w−1qk−n ≤ B logqn, (1)
we have
p(j; n, k, w)− λje−λ/j!→ 0 as n → ∞
Theorem 2 Fix a prime power q, positive constants b≤ 1
2 and B < 1− b, and a function µ(n) = o((log n)−1/4) When (1) holds,
√
2πλ p(j; n, k, w)− e−(j−λ) 2 /2λ→ 0 uniformly as λ → ∞
We remind the reader of the meaning of uniformity A function f : N` → N goes
to 0 as n→ ∞, uniformly over A ⊆ N`, provided
sup
a∈A n
|f(n, a)| → 0 as n → ∞, where An = {a ∈ N` −1 : (n, a)∈ A}
Since (q − 1)λ/qk is the the probability that a randomly chosen vector has weight w, the distribution of (q− 1)-tuples of weight w vectors in k-dimensional subspaces is asymptotically the same as the distribution of weight w vectors in random samples of qqk−1−1 (nonzero) vectors: They are asymptotically Poisson with parameter λ A point in projective space is the (q− 1)-tuple of scalar multiples of
a nonzero point in GF(q)n Thus, in a projective n-space over GF(q), the previous observation states that the distribution of weight w points is asymptotically the same among random sets of qqk−1−1 points and among random (k− 1)-dimensional projective subspaces, namely Poisson with parameter λ
2 Proof of the theorems
We will find it convenient to work with the somewhat larger
λ def≡ (q− 1)w−1nw
rather than the definition in (1) Since the ratio of the two versions of λ is
¡n
w
¢
w!
nw = (1− O(w/n))w = exp(O(w2/n)) = exp(o(1/(log n)2)),
it is easily seen that the ratio of the two versions of λ tends to 1 and also that the theorems for either version of λ imply the theorems for the other version Since the
Trang 4ratio of the two versions of λ approach 1, the inequality in (2) follows from that in (1) by replacing B with the average of B and 1− b
Let d = n− k, ² = (1 − b − B)/2, and J = (B + ²) logqn Making µ(n) larger
if necessary, we may assume
µ(n) ≥ max(n−b, e−²2logq n/4
Our proof consists of five parts:
(a) Eliminating small k
(b) Some easy estimates
(c) Conversion of the problem to the study of d× n matrices over GF(q) (d) An estimate for j ≤ J
(e) Completion of the proof
Although we will not explicitly state it, all estimates in o(· · ·) and O(· · ·) as well as all estimates of the form · · · → · · · are uniform
(a) Eliminating small k Suppose Theorem 1 holds for k = k0
def
≡ dn/2e We will deduce that it holds for k < k0 From the bound on w in (1), it follows that
λ → 0 whenever k ≤ k0 Since Theorem 1 is equivalent to p(0; n, k, w) = 1− o(1) when λ→ 0, it suffices to prove that p(0; n, k, w) ≥ p(0; n, k0, w) when k < k0
To this end we recall a property of downsets in regular ranked posets A ranked poset P is regular provided that every element of rank k is comparable to the same number of elements of rank k + 1, and likewise k − 1 (This is the requirement that each of the bipartite graphs formed by restricting the covering relation of P
to two adjacent ranks be regular in the graph theoretic sense.) For example, both the subsets of a set and the subspaces of a vector space, ordered by inclusion, are regular A downset in a partially ordered set is a set S such that x ∈ S and y < x imply y ∈ S We claim that if S is a downset then the fraction |S ∩ Pk|/|Pk| decreases with k, where Pk is the set of elements of rank k To see this, let α and
β be the common degrees of elements in the bipartite graph Pk× Pk+1 Clearly
Since S is a downset, every element of S∩ Pk+1 is related to β elements of S∩ Pk Hence
|S ∩ Pk|α ≥ β|S ∩ Pk+1|
Dividing the left side by the left side of (4) and the right side by the right side of (4) proves the claim
Let Ij be the set of subspaces of GF(q)n that contain at most j(q− 1) vectors
of weight w and no nonzero vectors of less weight Since this is a downset in the poset of subspaces of GF(q)n, the fraction of k-dimensional subspaces that lie in Ij
is a decreasing function of k and hence p(0; n, k, w)≥ p(0; n, k0, w) when k < k0
From now on, we assume that k≥ k0= dn/2e (5)
Trang 5(b) Some easy estimates By (1)
λ2w2/n < J2w2/n < µ(n)2 → 0 (6) and
q−d(ej)w = λ(q− 1)
µ ejw (q− 1)n
¶w
≤ λ(q − 1)
µ ejw (q− 1)n
¶
= O(J2w/n) (7) when j ≤ J Taking logarithms in (2) and using Stirling’s formula, we have
d− Jw = w©logqn− logqw− J + O(1)ª− logqλ
= w© (1− B − ²) logqn− logqw + O(1)ª
− O(log log n)
Fix K Since 1− B − ² = (1 + b − B)/2 > b, it follows easily that
w© (1− B − ²) logqn− logqw + Kª
is an increasing function of w for sufficiently large n Hence, for sufficiently large n,
d− j(w − 1) ≥ b logqn when j ≤ J (8)
(c) Conversion to d× n Matrices For V a subspace of GF(q)n let V⊥ be its orthogonal complement, the set of vectors orthogonal to every element of V As noted in the introduction, for nonzero characteristic the intersection V ∩ V⊥ may
have positive dimension; nevertheless, it is easily checked that the map V 7→ V⊥
is a bijection from k-dimensional to d-dimensional subspaces So it suffices to work with V⊥
Let H be a d× n matrix whose rows form a basis for V⊥ (in coding theory
terminology, a checksum matrix for V ) Denote the columns of H by hi Note that P
vihi = 0 if and only if v ∈ V If a set S of vectors is linearly dependent and
no proper subset is, then call the set minimally dependent If w is the minimal weight in V , the previous discussion shows that there is a bijection between sets
of w minimally dependent vectors among the columns of H and (q− 1)-tuples of vectors of weight w in V , where a tuple consists of all nonzero multiples of a weight
w vector
Since every d-dimensional subspace of GF(q)n has the same number of ordered bases, the fraction of ordered bases with a desired property will be the same as the fraction of d-dimensional subspaces with the property We will look at ordered bases
The rows of H are required to be independent; however, the fraction of all d×n matrices with this property is
q−nd
dY−1 i=0
(qn− qi
) =
n
Y
t=k+1
(1− q−t) = expµ
−
n
X
t=k+1
q−t+ O(q−2t)
¶
= exp¡
O(q−k)¢
= 1 + O(q−n/2),
(9)
Trang 6by (5) We can, in effect, ignore the requirement that the rows of H be independent (d) An estimate for j ≤ J In this part of the proof, we will show that
p(j; n, k, w) ≥ ¡λje−λ/j!¢©
1 + O(µ(n)2)ª
+ O(q−n/2) when j ≤ J (10)
It is instructive, and useful, to treat part of the w = 1 case separately A 1-set
of minimally dependent vectors must be {0} Hence the fraction of H containing exactly j such sets is
q−nd
µ
n
j
¶ (qd− 1)n −j = nj
qjdj!exp
³ O(j2/n)− (n − j)q−d(1 + O(q−d)´
uniformly since j = o(n1/2) Using, (9), the theorem now follows easily
For w > 1, we generalize this argument There is a complication: It is now possible for w-sets of minimally dependent vectors to overlap We count some of the subspaces with nonoverlapping cases and show that this consists of almost all subspaces Here is how we choose the columns:
(a) Repeat j times: Choose w columns of minimal dependency that are indepen-dent of the columns already chosen
(b) Choose the remaining n− jw columns to avoid introducing more dependent sets of size w or less
(c) Choose how to order the vectors
Let Na and Nc be the number of ways to carry out the choices in (a) and (c) The number of ways to carry out (b) depends on the choices in (a) Let Nb be a lower bound on the number of ways to carry out (b) We seek a lower bound for NaNbNc Suppose that i sets have been chosen in (a) Since the vectors already chosen span a space of dimension i(w− 1), the next set can be chosen in
(qd− qi(w −1))(qd − qi(w −1)+1)· · · (qd − qi(w −1)+w−2)× (q − 1)w −1
ways and so
Na = ¡
qd(q− 1)¢j(w−1)
j(w−1)−1Y
t=0
(1− qt −d)
= ¡
qd(q− 1)¢j(w−1)exp¡
O(qj(w−1)−d)¢
= ¡
qd(q− 1)¢j(w−1)¡1 + O(n−b)¢
by (8)
An upper bound on the number of vectors that can be expressed as a linear combination of at most w−1 of the vectors in an i-set is h(i, w) =Pl≤w−1(q−1)l¡i
l
¢ Thus
Nb ≥
n−1Y
i=jw
³
qd− h(i, w)´ = qd(n−jw)
n−1Y
i=jw
exp
³
−q−dh(i, w) + O¡
q−2dh(i, w)2¢´
Trang 7provided the expression inside the O( ) is bounded Now
q−dh(i, w) < q−d(q− 1)w −1µ
n
w− 1
¶ O(1) = O(wλ/n), and so
Nb ≥ qd(n −jw)expµ
−q−d X
l ≤w−1
(q− 1)l
n −1
X
i=jw
µ i l
¶ + O(w2λ2/n)
¶ (11)
by (6) Note that for w ≥ 1, using Stirling’s formula,
X
l ≤w−1
µ jw
l + 1
¶
= O(1)
µ jw w
¶
= O((ej)w)
Since
X
l ≤w−1
(q− 1)l
nX−1 i=jw
µ i l
¶
l ≤w−1
(q− 1)l
µµ n
l + 1
¶
−
µ jw
l + 1
¶¶
= (q− 1)w −1µ
n w
¶©
1 + O(w/n)ª
+ O((ejq)w)
= (q− 1)w −1nw
w!
©
1 + O(w2/n)ª
+ O((ejq)w),
we find, using (11), (7), and the definition of λ,
Nb ≥ qd(n −jw)exp³
−λ¡1 + O(w2/n)¢
+ O(J2w/n) + O(w2λ2/n)
´
= qd(n−jw)e−λ¡
1 + O(J2w2/n)¢
Finally, the number of ways to arrange the vectors, taking into account the fact that there is already some ordering among them is
j!(w!)j(n− wj)! =
nwj j!(w!)j
¡
1 + O(w2j2/n)¢
Putting all these results together, we obtain the lower bound
NaNbNc ≥ qnd¡
λje−λ/j!¢©
1 + O(J2w2/n) + O(n−b)ª
Equation (10) follows from (9), (12), and (3)
(e) Completion of the Proof Summing (10) over j ≤ J gives
j ≤J
p(j; n, k, w) ≥ (1 + O(µ(n)2))e−λX
j ≤J
λj j! + O(J q
−n/2)
= 1 + O(µ(n)2)−X
j>J
λj
eλj!,
(13)
Trang 8where the last equality follows from (3), (5), and the definition of J Since the ratio
of consecutive terms in the last summation is less than
λJ+1
(J + 1)!
Á
λJ
J! < λ/J ≤ B
B + ² = 1− ²
B + ²,
we have
X
j>J
λj
eλj! <
B + ²
²
λJ
eλJ! <
B + ²
²
µ λ J
¶J
eJ−λ Since this is an increasing function of λ, it is bounded above by O(Dlogq n) where
D =
µ B
B + ²
¶B+²
e²
For a fixed ², ¡ B
B+²
¢B+²
increases with B; letting B equal 1, and using 0 < ² < 1/2 and 1 + ²≥ e² −² 2 /2, we see
D ≤ (1 + ²)−1−²e² ≤ e−² 2 /2+²3/2 ≤ e−² 2 /4 Thus (13) becomes
j ≤J
p(j; n, k, w) ≥ 1 + O¡µ(n)2¢
+ O(Dlogq n) = 1 + O¡
µ(n)2¢
, (14) where the last equality follows from (3) It follows that
p(j; n, k, w) = λ
je−λ j! + O(µ(n)
2
and p(j; n, k, w) = O(µ(n)2) for j > J This completes the proof of Theorem 1
We now turn to Theorem 2 The standard approximation of the Poisson dis-tribution by the normal says
√
2πλ
µ
λje−λ j!
¶
− e−(j−λ) 2 /2λ → 0 uniformly in j as λ → ∞
(This can be obtained directly from Stirling’s formula.) Since λ1/2µ(n)2 → 0, Theorem 2 follows from (15)
References
1 M Aigner, Whitney numbers In N White (ed.), Combinatorial Geometries, volume II, Cambridge University Press (1987) 139–160
2 J R Goldman and G.-C Rota, On the foundations of combinatorics, IV: Finite vector spaces and Eulerian generating functions, Studies Appl Math 49 (1970) 239–258
3 V I Masol, The asymptotics of the number of k-dimensional subspaces of minimal weight over a finite field, Random Oper and Stoch Eqs 1 (1993) 287–292