ebook img

A Survey of Results on Random Random Walks on Finite Groups PDF

0.32 MB·English
Save to my drive
Quick download
Download
Most books are stored in the elastic cloud where traffic is expensive. For this reason, we have a limit on daily download.

Preview A Survey of Results on Random Random Walks on Finite Groups

4 A Survey of Results on Random Random 0 0 Walks on Finite Groups 2 n a J Martin Hildebrand 5 Department of Mathematics and Statistics 1 University at Albany ] R State University of New York P . Albany, NY 12222 h t a February 1, 2008 m [ 1 v Abstract 9 8 A number of papers have examined various aspects of “random 1 random walks” on finite groups; the purpose of this article is to pro- 1 0 vide a survey of this work and to show, bring together, and discuss 4 some of the arguments and results in this work. This article also pro- 0 vides a number of exercises. Some exercises involve straightforward / h computations; others involve provingdetails in proofsor extendingre- t a sults proved in the article. This article also describes some problems m for further study. : v i X 1 Introduction r a Random walks on the integers Z are familiar to many students of probability. (See, for example, Ch. XIV of Feller, volume 1 [8], Ch. XII of Feller, volume 2 [9], or Ross [21]). Such random walks are of the form X ,X ,X ,... where 0 1 2 X = 0 and X = Z +...+Z where Z ,Z ,...are independent, identically 0 m 1 m 1 2 distributed random variables on Z. A commonly studied random walk on Z has P(Z = 1) = P(Z = 1) = 1/2. Various questions involving such i i − random walks have been well studied. For example, one may ask what is the probability that there exists an m > 0 such that X = 0. In the example m 1 with P(Z = 1) = P(Z = 1) = 1/2, it can be shown that this probability is i i − 1. (See [8], p. 360.) For another example, one can use the DeMoivre-Laplace Limit Theorem to get a good approximation of the distribution of X for m large m. One can examine random walks on sets other then Z. For instance, there are random walks on Z2 or Z3. A symmetric random walk on Z2 has X = 0 (0,0) and P(Z = (1,0)) = P(Z = ( 1,0)) = P(Z = (0,1)) = P(Z = i i i i − (0, 1)) = 1/4 while a symmetric random walk on Z3 has X = (0,0,0) 0 − and P(Z = (1,0,0)) = P(Z = ( 1,0,0)) = P(Z = (0,1,0)) = P(Z = i i i i − (0, 1,0)) = P(Z = (0,0,1)) = P(Z = (0,0, 1)) = 1/6. It can be shown i i − − for this random walk on Z2, the probability that there exists an m > 0 such that X = (0,0) is 1 while for this random walk on Z3, the probability that m there exists an m > 0 such that X = (0,0,0) is less than 1. (See pp. m 360-361 of [8] for a description and proof. Feller attributes these results to Polya [19] and computation of the probability in the walk on Z3 to McCrea and Whipple [17].) One can similarly look at random walks on Z , the integers modulo n, n where n is a positive integer. Like the walks onthe integers, the random walk is of the form X ,X ,X ,... where X = 0 and X = Z +...+Z where 0 1 2 0 m 1 m Z ,Z ,... are i.i.d. random variables on Z . One example of such a random 1 2 n walk has P(Z = 1) = P(Z = 1) = 1/2 and n being an odd positive i i − integer. This random walk on Z corresponds to a finite Markov chain which n is irreducible, aperiodic, and doubly stochastic. (For more details on this notation, see Ross [21].) Thus the stationary probability for this Markov chain will be uniformly distributed on Z . (If P is a probability uniformly n distributed on Z , then P(a) = 1/n for each a Z .) Furthermore, after a n n ∈ large enough number of steps, the position of the random walk will be close to uniformly distributed on Z . n One may consider other probabilities on Z for Z in the random walk. n i For example, on Z , we might have P(Z = 0) = P(Z = 1) = P(Z = 1000 i i i 10) = P(Z = 100) = 1/4. Again Markov chain arguments can often show i that the stationary distribution for the corresponding Markov chain will be uniformly distributed on Z and that after a large enough number of steps, n the position of the random walk will be close to uniformly distributed on Z . A reasonable question to ask is how large should m be to ensure that 1000 X is close to uniformly distributed. m One may generalize the notion of a random walk to an arbitrary finite group G. We shall suppose that the group’s operation is denoted by mul- 2 tiplication. The random walk will be of the form X ,X ,X ,... where X 0 1 2 0 is the identity element of G, X = Z Z ...Z Z , and Z ,Z ,... are m m m−1 2 1 1 2 i.i.d. random variables on G. (A random variable X on G is such that Pr(X = g) 0 for each g G and Pr(X = g) = 1.) An alternate ≥ ∈ g∈G definition of a random walk on G has X = Z Z ...Z Z instead of m 1 2 m−1 m P X = Z Z ...Z Z . If G is not abelian, then the different definitions m m m−1 2 1 may correspond to different Markov chains. However, probabilities involving X alone do not depend on which definition we are using. m An example of a random walk on S , the group of all permutations n on 1,...,n , has P(Z = e) = 1/n where e is the identity element and i { } P(Z = τ) = 2/n2 for each transposition τ. Yet again, Markov chain ar- i guments can show that after a large enough number of steps, this random walk will be close to uniformly distributed over all n! permutations in S . n Again a reasonable question to ask is how large should m be to ensure that X is close to uniformly distributed over all the permutations in S . This m n problem is examined in Diaconis and Shahshahani [4] and also is discussed in Diaconis [3]. A number of works discuss various random walks on finite groups. A cou- ple of overviews are Diaconis’ monograph [3] and Saloff-Coste’s survey [22]. In this article, we shall focus on “random random walks” on finite groups. To do so, we shall pick a probability for Z at random from a set of probabil- i ities for Z . Then, given this probability for Z , we shall examine how close i i X is to uniformly distributed on G. To measure the distance a probability m is from the uniform distribution, we shall use the variation distance. Often, we shall look at the average variation distance of the probability of X from m the uniform distribution; this average is over the choices for the probabilities for Z . The next section will make these ideas more precise. i 2 Notation If P is a probability on G, we define the variation distance of P from the uniform distribution U on G by 1 1 P U = P(s) . k − k 2 − G s∈G(cid:12) | |(cid:12) X(cid:12) (cid:12) (cid:12) (cid:12) Note that U(s) = 1/ G for all s G. (cid:12) (cid:12) | | ∈ EXERCISE. Show that P U 1. k − k ≤ 3 EXERCISE. Show that P U = max P(A) U(A) k − k A⊆G | − | where A ranges over all subsets of G. (Note that A does not have to be a subgroup of G.) ArandomvariableX onGissaidtohaveprobabilityP ifP(s) = Pr(X = s) for each s G. ∈ If P and Q are probabilities on G, we define the convolution of P and Q by P Q(s) = P(t)Q(t−1s). ∗ t∈G X Note that if X and Y are independent random variables on G with proba- bilities P and Q, respectively, then P Q is the probability of the random ∗ variable XY on G. If m is a positive integer and P is a probability on G, we define P∗m = P P∗(m−1) ∗ where 1 if s = e P∗0(s) = 0 otherwise n with e being the identity element of G. Thus if Z ,...,Z are i.i.d. random 1 m variables on G each with probability P, then P∗m is the probability of the random variable Z Z ...Z Z on G. m m−1 2 1 For example, onZ , suppose P(0) = P(1) = P(2) = 1/3. Then P∗2(0) = 10 1/9, P∗2(1) = 2/9, P∗2(2) = 3/9, P∗2(3) = 2/9, P∗2(4) = 1/9, and P∗2(s) = 0 for the remaining elements s Z . Furthermore 10 ∈ 1 1 1 2 1 3 1 2 1 1 1 P∗2 U = + + + + k − k 2 9 − 10 9 − 10 9 − 10 9 − 10 9 − 10 (cid:18)(cid:12) (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) 1 (cid:12) (cid:12) 1 (cid:12) (cid:12) 1 (cid:12) (cid:12) 1 (cid:12) (cid:12) 1 (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) + 0 + 0 + 0 + 0 + 0 (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) − 10 − 10 − 10 − 10 − 10 (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) (cid:12)(cid:19) 1 (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) = (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) 2 (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) (cid:12) EXERCISE. Let Q be a probability on Z . Suppose Q(0) = Q(1) = 10 Q(4) = 1/3. Compute Q∗2 and Q∗2 U . k − k 4 EXERCISE. Consider the probability P in the previous example and the probability Q in the previous exercise. Compute P∗4, Q∗4, P∗4 U , and k − k Q∗4 U . k − k Let a ,a ,...a G. Suppose p ,...,p are positive numbers which sum 1 2 k 1 k ∈ to 1. Let k P (s) = p δ a1,...,ak b s,ab b=1 X where 1 if s = a δ = b s,ab 0 otherwise. n The random walk X ,X ,X ,... where Z ,Z ,... are i.i.d. random variables 0 1 2 1 2 on G with probability P is said to be supported on (a ,...,a ). a1,...,ak 1 k If we know k and p ,...,p , then in theory we can find 1 k P∗m U k a1,...,ak − k for any k-tuple (a ,...,a ) and positive integer m. Thus if we have some 1 k probability distribution for (a ,...,a ), we can in theory find 1 k E( P∗m U ) k a1,...,ak − k for each positive integer m since this variation distance is a function of the random k-tuple (a ,...,a ). 1 k In a random random walk, a typical probability for Z will be P i a1,...,ak where p = p = ... = p = 1/k and (a ,a ,...,a ) chosen uniformly over 1 2 k 1 2 k all k-tuples with distinct elements of G. However, other probabilities for Z i sometimes may be considered instead. For example, on Z , suppose we let p = p = p = 1/3 and we choose 5 1 2 3 (a ,a ,a ) at random over all 3-tuples with distinct elements of Z . Then 1 2 3 5 1 E( P∗m U ) = P∗m U + P∗m U + P∗m U k a1,a2,a3 − k 10 k 0,1,2 − k k 0,1,3 − k k 0,1,4 − k + (cid:0)P∗m U + P∗m U + P∗m U k 0,2,3 − k k 0,2,4 − k k 0,3,4 − k + P∗m U + P∗m U + P∗m U k 1,2,3 − k k 1,2,4 − k k 1,3,4 − k + P∗m U . k 2,3,4 − k (cid:1) Note that we are using facts such as P = P = P since p = p = 0,1,2 2,1,0 1,2,0 1 2 p = 1/3; thus we averaged over 10 terms instead of over 60 terms. 3 5 We shall often write E( P∗m U ) instead of E( P∗m U ). k − k k a1,...,ak − k Often times we shall seek upper bounds on E( P∗m U ). Note that k − k by Markov’s inequality, if E( P∗m U ) u, then Pr( P∗m U cu) k − k ≤ k − k ≥ ≤ 1/c for c > 1 where the probability is over the same choice of the k-tuple (a ,...,a ) as used in determining E( P∗m U ). 1 k k − k Lower bounds tend to be found on P∗m U for all a ,...,a , and we 1 k k − k turn our attention to some lower bounds in the next section. EXERCISE. Compute E( P∗4 U ) if G = Z , k = 3, p = p = 11 1 2 k − k p = 1/3, and (a ,a ,a ) are chosen uniformly from all 3-tuples with distinct 3 1 2 3 elements of G. It may be very helpful to use a computer to deal with the very tedious calculations; however, it may be instructive to be aware of the different values of P∗4 U for the different choices of a , a , and a . In 1 2 3 k − k particular, what can you say about P∗4 U and P∗4 U ? k a1,a2,a3 − k k 0,a2−a1,a3−a1 − k Throughout this article, there are a number of references to logarithms. Unless a base is specified, the expression log refers to the natural logarithm. 3 Lower bounds 3.1 Lower bounds on Z with k = 2 n In examining our lower bounds, we shall first look at an elementary case. Suppose G = Z . We shall consider random walks supported on 2 points. n The lower bound is given by the following: Theorem 1 Suppose p = p = 1/2. Let ǫ > 0 be given. There exists values 1 2 c > 0 and N > 0 such that if n > N and m < cn2, then P∗m U > 1 ǫ k a1,a2 − k − for all a and a in G with a = a . 1 2 1 2 6 Proof: Let F = i : 1 i m,Z = a , and let S = i : 1 m i 1 m |{ ≤ ≤ }| |{ ≤ i m,Z = a . In other words, if we perform m steps of the random walk i 2 ≤ }| supported on (a ,a ) on Z , we add a exactly F times, and we add a 1 2 n 1 m 2 exactly S times. Note that F + S = m and that X = F a + S a . m m m m m 1 m 2 Observe that E(F ) = m/2. Furthermore, observe that by the DeMoivre- m Laplace Limit Theorem (see Ch. VII of [8], for example), there are constants z , z , and N such that if n > N , then Pr(m/2 z m/4 < F < 1 2 1 1 1 m − p 6 m/2+z m/4) > 1 ǫ/2. Furthermore, there exists a constant c such that 2 − if m < cn2 and n > N , then (ǫ/6)n m/4 and Pr(m/2 (ǫ/6)n < F < p 1 ≥ − m m/2+(ǫ/6)n) > 1 ǫ/2. Let A = F a +(m F )a : m/2 (ǫ/6)n < − m {pm 1 − m 2 − F < m/2+(ǫ/6)n . ThusP∗m(A ) > 1 (ǫ/2). However, A (ǫ/3)n+1. m m m } − | | ≤ Thus if n > N and m < cn2, then 1 ǫ ǫ 1 P∗m U > 1 + k − k − 2 − 3 n (cid:18) (cid:19) 5ǫ 1 = 1 − 6 − n Let N = max(N ,6/ǫ). If n > N and m < cn2, then P∗m U > 1 ǫ. 2 1 k − k − EXERCISE. Modify the above theorem and its proof to consider the case where 0 < p < 1 is given and p = 1 p . 1 2 1 − 3.2 Lower bounds on abelian groups with k fixed Now let’s turn our attention to a somewhat more general result. This result is a minor modification of a result of Greenhalgh [10]. Here the probability will be supported on elements a ,a ,...,a . One similarity to the proof of 1 2 k the previous theorem is that we use the number of times in the random walk we choose Z to be a , the number of times we choose Z to be a , etc. i 1 i 2 Theorem 2 Let p = p = ... = p = 1/k. Suppose G is an abelian group 1 2 k of order n. Assume n k 2. Let δ be a given value under 1/2. Then there ≥ ≥ exists a value γ > 0 such that if m < γn2/(k−1), then P∗m U > δ for k a1,...,ak − k any k distinct values a ,a ,...a G. 1 2 k ∈ Proof: This proof slightly modifies an argument in Greenhalgh [10]. That proof, which only considered the case where G = Z , had δ = 1/4 but n considered a broader range of p ,p ,...,p . 1 2 k Since G is abelian, X is completely determined by the number of times m r we pick each element a . Observe that r + ... + r = m. Thus if g = i i 1 k r a +...+r a , we get 1 1 k k m! P∗m(g) Q (~r) := ≥ m ki=1(ri!kri) where ~r = (r ,...,r ). (Note that we mayQuse P∗m instead of P∗m .) 1 k a1,...,ak 7 Suppose r m/k α √m for i = 1,...,k. Then i i | − | ≤ k m(1/k)(1 1/k) Pr( r m/k α √m,i = 1,...,k) 1 − | i− | ≤ i ≥ − α2m i=1 i X k (1/k)(1 1/k) = 1 − − α2 i=1 i X > 2δ where we assume α are large enough constants to make the last inequality i hold. Let R~ = ~r : r +...+r = m and r m/k α √m for i = 1,...,k , 1 k i i { | − | ≤ } and let A = r a +...+r a : (r ,...,r ) R~ . Thus P∗m(A ) > 2δ. m 1 1 k k 1 k m { ∈ } ~ Observe that, with ǫ (~r) 0 uniformly over ~r R as m , we have, m → ∈ → ∞ by Stirling’s formula, m(k−1)/2e−mmm√2πm(1+ǫ (~r)) Q (~r)m(k−1)/2 = m m ki=1krie−ririri√2πri e−m mm 1 1+ǫ (~r) = Q m exp(− ki=1ri) · ki=1(kri)ri · ki=1 mri · (2π)(k−1)/2 P1 Q 1 q 1+ǫ (~r) Q (cid:0)m(cid:1) = 1 · ki=1(kri/m)ri · ki=1 ri/m · (2π)(k−1)/2 provided that all of theQvalues r1,...,rkQare ppositive integers. It can be shown that k r /m (1/k)k/2 uniformly over ~r R~ as i=1 i → ∈ m . → ∞ Q p Now let x = r m/k; thus kr = kx +m. Note that x +...+x = 0 i i i i 1 k − since r +...+r = m. Observe that 1 k k kr ri k kx m/k+xi i i = 1+ m m i=1(cid:18) (cid:19) i=1(cid:18) (cid:19) Y Y k m kx i = exp +x log 1+ . i k m ! Xi=1 (cid:16) (cid:17) (cid:18) (cid:19) Now observe k m kx i +x log 1+ i k m Xi=1 (cid:16) (cid:17) (cid:18) (cid:19) 8 k m kx k2x2 = +x i i +f(m,x ,...,x ) k i m − 2m2 1 k Xi=1 (cid:16) (cid:17)(cid:18) (cid:19) k k k2 k x3 = x2 i=1 i +f(m,x ,...,x ) 2m i − 2m2 1 k i=1 P X where for some constant C > 0, f(m,x ,...,x ) C /√m for all ~r R~. 1 1 k 1 | | ≤ ∈ So for some constant C > 0, 2 k2 k x3 C f(m,x ,...,x ) i=1 i 2 (cid:12) 1 k − 2m2 (cid:12) ≤ √m P (cid:12) (cid:12) (cid:12) (cid:12) for all ~r R~. (cid:12)(cid:12) (cid:12)(cid:12) ∈ Thus for some constant α > 0 and some integer M > 0, we have Q (~r)m(k−1)/2 α m ≥ for all ~r R~ and m M. (We may also assume that M is large enough ∈ ≥ that m/k > α √m if m M. Thus if m M, we have r > 0 for all ~r R~.) i i ≥ ≥ ∈ Now suppose that α 2 m(k−1)/2 ≥ n Thusifg A ,P∗m(g) α/m(k−1)/2 2/nandP∗m(g) (1/n) P∗m(g)/2. m ∈ ≥ ≥ − ≥ Thus P∗m(A ) U(A ) 0.5P∗m(A ) > δ, and so m m m − ≥ P∗m U > δ k a1,...,ak − k if m (α/2)2/(k−1)n2/(k−1) and m M. The following exercise shows ≤ ≥ that the above reasoning suffices if M (α/2)2/(k−1)n2/(k−1), i.e. n ≤ ≥ (2/α)M(k−1)/2. EXERCISE. If P and Q are probabilities on G, show that P Q U k ∗ − k ≤ P U . Use this result to show that if m m , then P∗m2 U 2 1 k − k ≥ k − k ≤ P∗m1 U . k − k For smaller values of n k, observe that P∗0 U = 1 1/n > δ, and ≥ k − k − we may choose γ such that γ N2/(k−1) < 1 where N = (2/α)M(k−1)/2. Let 1 1 γ = min(γ ,(α/2)2/(k−1)). Note that this value γ does not depend on which 1 abelian group G we chose. 2 EXERCISE. Extend the previous theorem andproof to deal with the case where p ,...,p are positive numbers which sum to 1. 1 k 9 EXERCISE. Extend the previous theorem andproof to deal with the case where δ is a given value under 1. (Hint: If δ < 1/b where b > 1, replace “2δ” by “bδ” and replace “2/n” by an appropriate multiple of 1/n.) 3.3 Some lower bounds with k varying with the order of the group Now let’s look at some lower bounds when k varies with n. The following theorem is a rephrasing of a result of Hildebrand [14]. Theorem 3 Let G be an abelian group of order n. If k = (logn)a where ⌊ ⌋ a < 1 is a constant, then for each fixed positive value b, there exists a function f(n) such that f(n) 1 as n and P∗m U f(n) for each → → ∞ k a1,...,ak − k ≥ (a ,...,a ) Gk where m = (logn)b . 1 k ∈ ⌊ ⌋ Proof: The proof also is based on material in [14]. In the first m steps of the random walk, each value a can be picked either 0 times, 1 time, ..., i or (logn)b times. Since the group is abelian, the value X depends only m ⌊ ⌋ on the number of times each a is picked in the first m steps of the random i walk. So after m steps, there are at most (1+ (logn)b )k different possible ⌊ ⌋ values for X . The following exercise implies the theorem. m EXERCISE. Prove the following. If k = (logn)a and a < 1 is a con- ⌊ ⌋ stant, then (1+ (logn)b )k ⌊ ⌋ 0 n → as n . → ∞ Notethatthefunctionf maydependonbbutdoesnotnecessarilydepend on which abelian group G we chose. 2 The following lower bound, mentioned in [14], applies for any group G. Theorem 4 Let G be any group of order n. Suppose k is a function of n. Let ǫ > 0 be given. If logn m = (1 ǫ), ⌊logk⌋ − then P∗m U f(n) for each (a ,...,a ) Gk and some f(n) such k a1,...,ak − k ≥ 1 k ∈ that f(n) 1 as n . → → ∞ Proof: NotethatX hasat mostkm possible values andthatkm n1−ǫ. m Thus P∗m U 1 (n1−ǫ/n). Let f(n) = 1 (n1−ǫ/n). ≤ 2 k a1,...,ak − k ≥ − − 10

See more

The list of books you might like

Most books are stored in the elastic cloud where traffic is expensive. For this reason, we have a limit on daily download.