ebook img

Finite field matrix channels for network coding PDF

0.3 MB·
Save to my drive
Quick download
Download
Most books are stored in the elastic cloud where traffic is expensive. For this reason, we have a limit on daily download.

Preview Finite field matrix channels for network coding

Finite Field Matrix Channels for Network Coding Simon R. Blackburn and Jessica Claridge 6 Department of Mathematics 1 Royal Holloway University of London 0 2 Egham, Surrey TW20 0EX, United Kingdom c e D Abstract 9 In2010,Silva,KschischangandK¨otterstudiedcertainclassesoffinitefieldmatrix ] channels in order to model random linear network coding where exactly t random T errors are introduced. I . In this paper we introduce a generalisation of these matrix channels that al- s c lows the modelling of channels with a varying number of errors. We show that a [ capacity-achieving input distribution can always be taken to have a very restricted form (the distribution should be uniform given the rank of the input matrix). This 2 v result complements, and is inspired by, a paper of Nobrega, Silva and Uchoa-Filho, 7 that establishes a similar result for a class of matrix channels that model network 3 coding with link erasures. Our result shows that the capacity of our channels can 0 be expressed as a maximisation over probability distributions on the set of possible 6 ranks of input matrices: a set of linear rather than exponential size. 0 . 1 0 1 Introduction 6 1 Network coding, first defined in [1], allows intermediate nodes of a network to com- : v pute with and modify data, as opposed to the traditional view of nodes as ‘on/off’ i X switches. This can increase the rate of information flow through a network. It is shown in [8] that linear network coding is sufficient to maximise information flow in r a multicast problems, that is when there is one source node and information is to be transmitted to a set of sink nodes. Moreover,in [6] it is shownthat for generalmul- tisource multicast problems, random linear network coding achieves capacity with probability exponentially approaching 1 with the code length. In random linear network coding, the source injects packets into the network; thesepacketscanbethoughtofasvectorsoflengthmwithentriesinafinitefieldFq (where q is a fixed power of a prime). The packets flow through a network of un- known topology to a sink node. Each intermediate node forwards packets that are random linear combinations of the packets it has received. A sink node attempts to reconstruct the message from these packets. In this context, Silva, Kschischang and K¨otter[13]studied achanneldefinedas follows. We write Fn×m to denote the setof q all n×m matrices over Fq, and write GL(n,q) for the set of all invertible matrices in Fn×n. q Definition 1. The Multiplicative Matrix Channel (MMC) has input set X and out- put set Y, where X =Y =Fn×m. The channel law is q Y =AX where A∈GL(n,q) is chosen uniformly at random. 1 Here the rows of X correspond to the packets transmitted by the source, the rowsofY arethe packetsreceivedbythe sink,andthe matrixAcorrespondsto the linear combinations of packets computed by the intermediate nodes. Inspired by Montanari and Urbanke [9], Silva et al modelled the introduction of random errors into the network by considering the following generalisation of the MMC. We write Fn×m,r for the set of all n×m matrices of rank r. q Definition 2. The Additive Multiplicative Matrix Channel with t errors (AMMC) has input set X and output set Y, where X =Y =Fn×m. The channel law is q Y =A(X +W) where A ∈ GL(n,q) and W ∈ Fn×m,t are chosen uniformly and independently at q random. So the matrix W corresponds to the assumption that exactly t linearly indepen- dent random errors have been introduced. The MMC is exactly the AMMC with zero errors. We note that the AMMC is very different from the error model studied in the well-known paper by K¨otter and Kschischang [7], where the errors are assumed to be adversarial(so the worst case is studied). In [13] the authors give upper and lower bounds on the capacity of the AMMC, which are shown to converge in certain interesting limiting cases. The exact ca- pacity of the AMMC for fixed parameter choices is hard to determine due to the many degrees of freedom involved: the naive formula maximises over a probability distributiononthe setofpossible input matrices,andthis setis exponentially large. In this paper we introduce a generalisation of these matrix channels that allows the modelling of channels were the number of errors is not necessarily fixed. (For example,itenablesthe modellingofsituationswhenatmostterrorsareintroduced, or when the errors are not necessarily linearly independent, or both.) To define our generalisation, we need the following notation which is due to Nobrega, Silva and Uchoa-Filho [10]. Definition 3. Let R be a probability distribution on the set {0,1,...,min{m,n}} of possible ranks of matrices M ∈ Fn×m. We define a distribution on the set Fn×m q q ofmatrices bychoosingr accordingto R,andthen oncer is fixedchoosinga matrix M ∈ Fn×m,r uniformly at random. We say that this distribution is Uniform Given q Rank (UGR) with rank distribution R. We say a distribution on Fn×m is Uniform q Given Rank (UGR) if it is UGR with rank distribution R for some distribution R. We write R(r) for the probability of rank r under the distribution R. So a distributiononFn×m isUGRwithrankdistributionRifandonlyifeachM ∈Fn×m q q of rank r is chosen with probability R(r)/|Fn×m,r|. q Definition 4. Let R be a probability distribution on the set {0,1,...,min{m,n}} of possible ranks of matrices M ∈ Fn×m. The Generalised Additive Multiplicative q MAtrixChannelwithrankerrordistributionR(theGammachannelΓ(R))hasinput set X and output set Y, where X =Y =Fn×m. The channel law is q Y =A(X +W) where A ∈ GL(n,q) is chosen uniformly, where W ∈ Fn×m is UGR with rank q distribution R, and where A and W are chosen independently. We see that the AMMC is the special case of Γ(R) when R is the distribution choosing rank t with probability 1. But the model covers several other very natural situations. For example, we may drop the assumption that the t errors are linearly independent by defining R(r) to be the probability that t vectors span a subspace of dimension r, when the vectors are chosen uniformly and independently. We can also extend this to model situations when the number t of vectors varies according 2 to natural distributions such as the binomial distribution (which arises when, for example, a packet is corrupted with some fixed non-zero probability). In practice, givena particular network,one may runtests on the network to see the actualerror patterns produced and define an empirical distribution on ranks. One could also defineanappropriatedistributionbyconsideringsomecombinationofthe situations described. Throughoutthispaper,weassumethatq,n,mandRarefixedbytheapplication. We will refer to these values as the channel parameters. We note that the Gamma channel assumes that the transfer matrix A is always invertible. Thisisarealisticassumptioninrandomlinearnetworkcodinginstandard situations: thefieldsizeqisnormallylarge,whichmeanslineardependenciesbetween random vectors are less likely. In both [10]and[12]the authorsconsider (different) generalisationsofthe MMC channelthatdonotnecessarilyhaveasquarefullranktransfermatrix. Suchchannels allow modelling of network coding when no erroneous packets are injected into the network, but there may be link erasures. In [10], Nobrega, Silva and Uchoa-Filho define the transfer matrix to be picked from a UGR distribution. One result of [10] is that a capacity-achieving input distribution for their class of MMC channels can always be taken to be UGR. A main result of this paper (Theorem 16) is that a capacity-achieving input distribution for Gamma channels can always be taken to be UGR. Theorem 16 is a significant extension of the result of [10] to a new class of channels; the extension requires new technical ideas. Corollary 18 to the main result of the paper provides an explicit method for computing the capacity of Gamma channels which maximises over a probability distributionoverthesetofpossibleranksofinputmatrices,ratherthanthesetofall input matrices itself. Thus we have reduced the problem of computing the capacity of a Gamma channel to an explicit maximisation over a set of variables of linear rather than exponential size. The remainder of the paper is organised as follows. Section 2 proves some pre- liminary results needed in what follows. In Section 3 we state results from matrix theory that we use. Section 4 establishes a relationship between the distributions of the ranks of input and output matrices for a Gamma channel. Section 5 proves Theorem 16, and Section 6 proves Corollary 18, giving an exact expression for the capacity of the Gamma channels. Finally, in Section 7 we prove the results from matrix theory that we use in earlier sections. 2 Preliminaries on finite-dimensional vector spaces Inthissectionwediscussfinite-dimensionalvectorspacesandconsiderseveralcount- ing problems involving subspaces and quotient spaces. Thefollowingtwolemmasarestandardresultsinthetheoryoffinite-dimensional vector spaces. Lemma 1. Let V be a vector space of dimension d , and let U be a d -dimensional V U subspace of V. The quotient space V/U has dimension d −d . V U Proof. See [5, §22]. Lemma 2. Let V be a vector space and let U be a fixed subspace of V. For any subspace W of V dim([W])=dim(W)−dim(W ∩U), where [W] denotes the image of W in the quotient space V/U. Proof. This is a standard exercise. 3 The following lemma gives the number of subspaces U of V, when, given some subspace V of V, the intersection of U and V is fixed and the image of U in the 1 1 quotient space V/V is fixed. 1 Lemma 3. Let V be a d -dimensional vector space. Let V , V be subspaces of V 1 2 V, of dimensions d and d respectively, such that V ⊆ V . The number of d - V1 V2 2 1 U dimensional subspaces U ⊆ V such that U ∩V = V and the image of U in the 1 2 quotient space V/V is the fixed d −d dimensional space U′, is given by 1 U V2 q(dU−dV2)(dV1−dV2). Proof. Fix abasisfor V ,say{b ,...,b }. Letπ :V →V/V be the mapwhich 2 1,1 1,dV2 1 takes vectors in V to their image in V/V1. For dU′ = dU −dV2, let {y1,...ydU′} be a basis for U′, and let {b ,...,b } be some vectors in V such that π(b )=y , 2,1 2,dU′ 2,i i for i=1,...,dU′. The set {b ,...,b ,b ,...,b } is a linearly independent set that forms 1,1 1,dV2 2,1 2,dU′ the basis ofa space U ofthe requiredform. Moreover,everyspace U ofthe required formhasabasisthatcanbe constructedinthis way. The generalconstructiontakes the basis for V2 and extends this to a basis for a space U by adding any set of dU′ vectors in V whose image under π is {y ,...,y }. 1 dU′ Given the set {b2,1,...,b2,dU′}, all other sets of dU′ vectors in V whose image under π is {y ,...y } can be written in the form {v +b ,...,v +b } for 1 dU′ 1 2,1 dU′ 2,dU′ some v ,...,v ∈ V . Therefore, any space U of the required form has a basis 1 dU′ 1 that can be written as B = {b ,...,b ,v +b ,...,v + b } for some 1,1 1,dV2 1 2,1 dU′ 2,dU′ v ,...,v ∈V . 1 dU′ 1 Given some set v′,...,v′ ∈ V , let B′ = {b ,...,b ,v′ +b ,...,v′ + 1 dU′ 1 1,1 1,dV2 1 2,1 dU′ b2,dU′}. Now Span(B) = Span(B′) if and only if vi −vi′ ∈ V2 for i = 1,...,dU′. That is, the sets B and B′ give rise to the same space if and only if [v ] = [v′] for i i i=1,...,dU′,where[v]denotesthe imageofavectorv inthequotientspaceV1/V2. Therefore there is a bijection betweenspaces U of the requiredformand ordered sets {[v ],...,[v ]} of elements in the quotient space V /V . 1 dU′ 1 2 For i=1,...dU′, there are qdV1−dV2 choices for [vi]∈V1/V2, thus there are qdU′(dV1−dV2) =q(dU−dV2)(dV1−dV2). (1) choices for the ordered set {[v ],...,[v ]}. The result follows. 1 dU′ Given a vector space V and a subspace V ⊆ V, Lemma 3 can be used to count 1 subspaces U of V when either U ∩V is fixed, or the image of U in V/V is fixed, or 1 1 when only the dimensions of these spaces are fixed. These results are given in the following three corollaries. Corollary 4. Let V be a d -dimensional vector space. Let V , V be subspaces V 1 2 of V, of dimensions d and d respectively, such that V ⊆ V . The number of V1 V2 2 1 d -dimensional subspaces U ⊆V such that U ∩V =V , is given by U 1 2 d −d q(dU−dV2)(dV1−dV2) V V1 . d −d (cid:20) U V2(cid:21)q Proof. Consider the quotient space V/V , this is a space of dimension d −d . Let 1 V V1 U′ be a (d −d )-dimensional subspace of V/V . There are U V2 1 d −d V V1 d −d (cid:20) U V2(cid:21)q possible choices for U′, fix one such space. ByLemma3thereareq(dU−dV2)(dV1−dV2)possibilitiesforthespaceU whoseimage inthequotientV/V isthefixedspaceU′. Multiplyingbythenumberofpossibilities 1 for U′, yields the result. 4 Corollary 5. Let V be a d -dimensional vector space. Let V be a d -dimensional V 1 V1 subspace of V. The number of d -dimensional subspaces U ⊆V such that the image U of U in the quotient space V/V1 is some fixed dU′ dimensional space U′, is given by d qdU′(dV1−(dU−dU′)) V1 . (2) (cid:20)dU −dU′(cid:21)q Proof. Let V2 be a (dU −dU′)-dimensional subspace of V1, there are d V1 (cid:20)dU −dU′(cid:21)q possible choices for V , fix one. 2 By Lemma 3 there are q(dU−(dU−dU′))(dV1−(dU−dU′)) = qdU′(dV1−(dU−dU′)) possi- bilities for the space U whose intersectionwith V is the fixed space V . Multiplying 1 2 by the number of possibilities for V , yields the result. 2 Corollary 6. Let V be a d -dimensional vector space. Let V be a d -dimensional V 1 V1 subspace of V. For a subspace U of V, let [U] denote the image of U in the quotient spaceV/V . Thenumberofd -dimensionalsubspacesU ⊆V suchthatdim(U∩V )= 1 U 1 d is equal to the number of d -dimensional subspaces U such that dim([U]) = UV1 U d −d . This number is equal to U UV1 d −d d q(dU−dUV1)(dV1−dUV1) V V1 V1 . d −d d (cid:20) U UV1(cid:21)q(cid:20) UV1(cid:21)q Proof. Note that dim(U∩V )=d if and only if dim([U])=d −d hence the 1 UV1 U UV1 equivalence in the statement of the lemma holds. Let V be a (d )-dimensional 2 UV1 subspace of V , there are 1 d V1 d (cid:20) UV1(cid:21)q possible choices for V , fix one. Next let U′ be a (d −d )-dimensional subspace 2 U UV1 of V/V . There are 1 d −d V V1 d −d (cid:20) U UV1(cid:21)q possible choices for U′, fix one. By Lemma 3 there are q(dU−dUV1)(dV1−dUV1) possibilities for the space U whose intersection with V is the fixed space V , and image in the quotient V/V is the 1 2 1 fixed space U′. Multiplying by the number of possibilities for V and U′, yields the 2 result. 3 Matrices over finite fields This short section describes the notation and results we use from the theory of matrices over finite fields. Let q be a non-trivial prime power, and let Fq be the finite field of order q. In the introduction we defined the notation Fn×m for the set of n×m matrices with q entries in Fq, Fqn×m,r for the subset of Fqn×m consisting of those matrices of rank r and GL(n,q) for the set of invertible matrices in Fn×n. q For a matrix M, we write rk(M) for the rank of M and we write Row(M) for the row space of M. The following results will be proved in Section 7. Lemma 7. Let U be a subspace of Fm of dimension u. The number of matrices q M ∈ Fn×m such that Row(M) = U can be efficiently computed; it depends only on q q, n, m and u. We write f (u) for the number of matrices M of this form. 0 5 Lemma 8. Let U and V be subspaces of Fm of dimensions u and v respectively. q Let h = dim(U ∩V). Let M ∈ Fn×m be a fixed matrix such that Row(M) = U. q Let r be a non-negative integer. The number of matrices W ∈ Fn×m,r such that q Row(W +M) = V can be efficiently computed; it depends only on q, n, m, r, u, v and h. We write f (u,v,h;r) for the number of matrices W of this form. 1 Lemma 9. Let r, r and r be non-negative integers. Let X be a fixed matrix such W X that rk(X)= rX. The number of matrices W ∈Fqn×m,rW such that rk(X +W)=r can be efficiently computed; it depends only on q, n, m, r, r and r . We write W X f (r,r ,r ) for the number of matrices W of this form. 2 X W By an efficient computation, we mean a polynomial (in max{n,m}) number of arithmetic operations. In Section 7 we will prove Lemmas 7, 8 and 9 and give exact expressions for the functions f ,f and f in terms of their inputs and the values q, 0 1 2 n and m. 4 Input and output rank distributions A distribution PX on the input set X of the Gamma channel induces a distribution (the input rank distribution) RX on the set of possible ranks of input matrices. Let RY be the corresponding output rank distribution, induced from the distribution on theoutputsetofthe Gammachannel. Akeyresult(Lemma11)isthatRY depends on only the channel parameters and RX (rather than on PX itself). This section aims to prove this result: it will play a vital role in the proof of Theorem 16 below. Definition 5. Let r,r ,r ∈{0,...,min{n,m}}. Define X B f (r,r ,r ) 2 X B ρ(r;r ,r )= , X B |Fqn×m,rB| where f2 is as defined in Lemma 9. For any fixed matrix X ∈Fqn×m,rX, ρ(r;rX,rB) gives the proportion of matrices B ∈ Fn×m,rB with rk(X +B) = r. Let R be a q probability distribution on the set {0,1,...,min{n,m}} of possible ranks of n×m matrices. Define min{n,m} ρ(r;r )= R(r )ρ(r;r ,r ), X B X B rXB=0 sothatρ(r;r )givesthe weightedaverageofthis proportionoverthepossibleranks X of matrices B. Lemma10. LetX beann×mmatrixsampledfromsomedistributionPX onFqn×m. Let B be an n×m matrix sampled from a UGR distribution with rank distribution R, where X and B are chosen independently. Let r,r ,r ∈ {0,...,min{n,m}}. X B Then ρ(r;r ,r )=Pr(rk(X+B)=r|rk(X)=r and rk(B)=r ), (3) X B X B and ρ(r;r )=Pr(rk(X +B)=r|rk(X)=r ). (4) X X Proof. Let X be a fixed n × m matrix of rank r . Then, since B has a UGR X distribution, Pr(rk(X +B)=r|rk(B)=r ) B |{B ∈Fn×m,rB :rk(X +B)=r}| q = |Fqn×m,rB| f (r,r ,r ) 2 X B = |Fqn×m,rB| =ρ(r;r ,r ). (5) X B 6 Note that (5) only depends on rk(X), not X itself. Hence Pr(rk(X +B)=r|rk(X)=r ,rk(B)=r ) X B = Pr(X =X)Pr(rk(X +B)=r|rk(B)=r ) B X X = Pr(X =X)ρ(r;r ,r ) X B X X =ρ(r;r ,r ), X B where the sums are over matrices X ∈Fn×m,rX. Thus (3) holds. Also q Pr(rk(X +B)=r|rk(X)=r ) X min{n,m} = R(r )ρ(r;r ,r ) (by (3)) B X B rXB=0 =ρ(r;r ). X Thus (4) holds, and so the lemma follows. Lemma 11. For the Gamma channel Γ(R) with input rank distribution RX, the output rank distribution is given by min{n,m} f (r,r ,r ) 2 X B RY(r)= rXX,rB=0 RX(rX)R(rB) |Fqn×m,rB| for r = 1,...,min{n,m}. In particular, RY depends only on the input rank distri- bution (and the channel parameters), not on the input distribution itself. Proof. We have that Pr(rk(X) = rX) = RX(rX) and Pr(rk(B) = rB) = R(rB). Hence, by (3), RY(r)=Pr(rk(Y)=r) min{n,m} = RX(rX)R(rB)ρ(rY;rX,rB) rXX,rB=0 min{n,m} f (r,r ,r ) 2 X B = rXX,rB=0 RX(rX)R(rB) |Fqn×m,rB| . 5 A UGR input distribution achieves capacity This section shows (Theorem 16) that there exists a UGR input distribution to the Gamma channel that achieves capacity. Lemma 12. Let M and M′ be fixed n×m matrices of the same rank. Let B be an n×m matrix picked from a UGR distribution, and let A be an n×n matrix picked uniformly from GL(n,q), with B and A picked independently. Let Y =A(M +B) and let Y′ =A(M′+B). Then H(Y)=H(Y′). Proof. LetAbeafixedn×ninvertiblematrix. SincethematricesAM andAM′have the same rank, there exist invertible matrices R and C such that AM′ = RAMC. Consider the linear transformation ϕ: Fn×m → Fn×m defined by ϕ(Y)= RYC. It q q is simple to check that ϕ is well defined and a bijection. Note that ϕ(A(M +B))=RAMC+RABC =A(M′+A−1RABC). 7 Since B is picked uniformly once its rank is determined, pre- and post-multiplying B by fixed invertible matrices gives a uniform matrix of the same rank, therefore B and A−1RABC have the same distribution. Now Pr(Y =Y|A=A) =Pr(A(M +B)=Y) =Pr(ϕ(A(M +B))=ϕ(Y)) =Pr A(M′+A−1RABC)=ϕ(Y) =Pr(A(M′+B)=ϕ(Y)) (6) (cid:0) (cid:1) =Pr(Y′ =ϕ(Y)|A=A), (7) where(6) holdssince the distributions ofB and A−1RABC are the same. Since (7) is true for any fixed matrix A, it follows that Pr(Y =Y)= Pr(A=A)Pr(Y =Y|A=A) A∈GXL(n,q) = Pr(A=A)Pr(Y′ =ϕ(Y)|A=A) A∈GXL(n,q) =Pr(Y′ =ϕ(Y)). (8) Thus Y and Y′ have the same distribution, up to relabeling by ϕ. In particular, we find that H(Y)=H(Y′). Definition 6. Let M be any n×m matrix of rank r. Let A be an n×n invertible matrix chosen uniformly from GL(n,q). Let B be an n×m matrix chosen from a UGRdistributionwithrankdistributionR,whereAandBarepickedindependently. We define h =H(A(M +B)). r Lemma 12 implies that the value h does not depend on M, only on the rank r r and the channel parameters q,n,m and R. The exact value of h will be calculated r later, in Theorem 17. Lemma13. Consider theGammachannelΓ(R)withinputdistributionPX. Letthe input matrix X be sampled from PX, and let Y be the corresponding output matrix. Then min{n,m} H(Y|X)= RX(r)hr. r=0 X In particular, H(Y|X) depends only on the associated input rank distribution RX and the channel parameters. Proof. Choosing A and B as in the definition of the Gamma channel, we see that H(Y|X)= P(X =X)H(A(X +B)) X∈X X = P(X =X)h rk(X) X∈X X min{n,m} = RX(r)hr, r=0 X whichestablishesthefirstassertionofthe lemma. The secondassertionfollowssince h depends only on r and the channel parameters. r 8 Lemma 14. Let Y1 and Y2 be two random n×m matrices, sampled from distri- butions with the same associated rank distribution RY. If the distribution of Y2 is UGR then H(Y2)≥H(Y1). Proof. For i = 1,2, we see that rk(Yi) is fully determined by Yi and so it follows that H(Yi,rk(Yi)) = H(Yi). Therefore by the chain rule for entropy (e.g. [4, Thm. 2.2.1]), H(Yi)=H(Yi,rk(Yi)) =H(Yi|rk(Yi))+H(rk(Yi)). (9) Since Y2 is distributed uniformly once its rank is determined, H(Y2|rk(Y2) = r) is maximal (e.g. [4, Thm. 2.6.4]), hence H(Y2|rk(Y2)=r)≥H(Y1|rk(Y1)=r). (10) Thus, using (9) H(Y2)=H(Y2|rk(Y2))+H(rk(Y2)) min{n,m} = (RY(r)H(Y2|rk(Y2)=r))+H(rk(Y2)) r=0 X min{n,m} ≥ (RY(r)H(Y1|rk(Y1)=r))+H(rk(Y2)) (11) r=0 X =H(Y1|rk(Y1))+H(rk(Y2)) =H(Y1|rk(Y1))+H(rk(Y1)) (12) =H(Y1), where (11) follows from(10), and(12) follows since the rank distributions of Y1 and Y2 are the same. Lemma 15. Consider the Gamma channel Γ(R). If the input distribution PX is UGR then the induced output distribution PY is also UGR. Proof. SupposetheinputdistributionisUGR,withrankdistributionRX. We start by showing that the distribution of X +B is UGR. Let D be any n×m matrix. Then Pr(X+B =D) = Pr(X =X)Pr(X +B =D|X =X) X∈XFqn×m = RX(rk(X)) Pr(X +B =D), X∈XFnq×m |Fnq×m,rk(X)| since X is sampled from a UGR distribution. Hence Pr(X +B =D) min{n,m} = RX(r) Pr(B =D−X) Xr=0 |Fqn×m,r|X∈XFnq×m,r min{n,m} RX(r) R(rk(D−X)) = , Xr=0 |Fqn×m,r|X∈XFnq×m,r |Fqn×m,rk(D−X)| 9 since X and B are independent, and since B has a UGR distribution with rank distribution R. Now R(rk(D−X)) X∈XFnq×m,r |Fqn×m,rk(D−X)| min{n,m} R(r ) = rXB=0 |{X ∈Fqn×m,r :rk(D−X)=rB}||Fqn×mB,rB| min{n,m} R(r ) B = f (r ,rk(D),r) rXB=0 2 B |Fqn×m,rB| and so Pr(X +B =D) min{n,m} min{n,m} RX(r) R(rB) = f (r ,rk(D),r) . Xr=0 |Fqn×m,r| rXB=0 2 B |Fqn×m,rB| So Pr(X+B =D) does not depend onthe specific matrix D, only its rank. There- fore, given any two n×m matrices D ,D of the same rank, 1 2 Pr(X +B =D )=Pr(X +B =D ). 1 2 Hence X +B has a UGR distribution. Let A be a fixed n×n invertible matrix. Since X +B is picked uniformly once its rank is determined, multiplying X +B by the invertible matrix A will give a uniform matrix of the same rank, therefore A(X +B) has a UGR distribution. So, defining Y = A(X +B) to be the output of the Gamma channel, we see that for any matrix Y Pr(Y =Y|A=A)=Pr(A(X +B)=Y) Pr(rk(A(X +B))=rk(Y)) = |Fnq×m,rk(Y)| Pr(rk(Y)=rk(Y)|A=A) = , |Fnq×m,rk(Y)| where the second equality follows since A(X +B) has a UGR distribution. Thus Pr(Y =Y)= Pr(A=A)Pr(Y =Y|A=A) A∈GL(n,q) X Pr(rk(Y)=rk(Y)|A=A) = Pr(A=A) n×m,rk(Y) A∈GXL(n,q) |Fq | 1 = Pr(rk(Y)=rk(Y)). (13) |Fnq×m,rk(Y)| Since (13) holds for all Y ∈Fn×n it follows that Y has a UGR distribution. q Theorem 16. For the Gamma channel Γ(R), there exists a UGR input distribu- tion that achieves channel capacity. Moreover, given any input distribution PX with associated rank distribution RX, if PX achieves capacity then the UGR distribution with rank distribution RX achieves capacity. Proof. Let X1 be a channel input, with output Y1 such that PX1 is a capacity achieving input distribution. That is maxPX{I(X,Y)} = I(X1,Y1). Then define 10

See more

The list of books you might like

Most books are stored in the elastic cloud where traffic is expensive. For this reason, we have a limit on daily download.