ebook img

Channel Resolvability Theorems for General Sources and Channels PDF

0.22 MB·
Save to my drive
Quick download
Download
Most books are stored in the elastic cloud where traffic is expensive. For this reason, we have a limit on daily download.

Preview Channel Resolvability Theorems for General Sources and Channels

Channel Resolvability Theorems for General Sources and Channels Hideki Yagi Dept. of Computer and Network Engineering University of Electro-Communications Tokyo, Japan Email: [email protected] 7 Abstract—In the problem of channel resolvability, where a [4], [10]. The established general formula provides a new 1 given output probability distribution via a channel is approxi- expressionfortheδ-spectralsup-entropyrate,whichisawell- 0 matedbytransformingtheuniformrandomnumbers,character- 2 known informationquantity in informationspectrum methods izingtheasymptoticallyminimumrateofthesizeoftherandom [3].Theanalysisisalso extendedto thesecond-orderchannel n numbers, called the channel resolvability, has been open. This a paper derives formulas for the channel resolvability for a given resolvability,whichis definedas the asymptoticallyminimum J general sourceandchannelpair.Wealsoinvestigate thechannel second-orderrateofthesizeofuniformrandomnumberswith 9 resolvability in an optimistic sense. It is demonstrated that the respect to a fixed first-order resolvability rate. derivedgeneralformulasrecaptureasingle-letterformulaforthe 2 stationary memoryless source and channel. When the channel ] is the identity mapping, the established formulas reduce to an II. PROBLEMFORMULATION: CHANNEL RESOLVABILITY T alternative form of the spectral sup-entropy rates, which play a I key role in information spectrum methods. The analysis is also Let and be finite or countably infinite alphabets. s. extended to the second-order channel resolvability. Let XnXdenoteYa sequence of n random variables taking c valuesin n with probabilitydistribution P . In this paper, [ I. INTRODUCTION X Xn we identify P with Xn, and both expressions are used Xn 1 Finding the asymptotically minimum rate of the size of the interchangeably. We call X = Xn ∞ a general source. v uniform random numbers (channel resolvability) which can Also, let Wn : n n denot{e a s}tno=ch1astic mapping, and 62 approximatea giventargetoutputdistributionviaa channelis wecallW ={WXn}→∞n=1Yageneralchannel.Wedonotimpose 3 called the problem of channel resolvability. When the varia- anyassumptionssuchasstationarityorergodicityoneitherX 8 tional distance between the target output distribution and the or W. We denote by Y = Yn ∞ the output process via 0 approximateddistributionis requiredto be asymptoticallynot W due to input process X.{ }n=1 . 1 greaterthanδ [0,1),theproblemiscalledtheproblemofδ- ∈ We review the problem of channel resolvability [3] using 0 channelresolvability.Thoughtheseproblemswereintroduced 7 byHanandVerdu´ [4]morethantwo decadesago,thegeneral thevariationaldistanceasanapproximationmeasure.LetUMn 1 formula for the channel resolvability has not been known denote the uniform random number of size Mn, which is : a random variable uniformly distributed over 1,...,M . v in general. A few cases where the channel resolvability has { n} Xi been characterized are the worst input case with δ = 0 by CUonsidveiar aapdperotexrimmiantiisntgicthmeatpaprignegt dϕistr:ibu1t,io.n..P,MYn by usinng Hayashi [5] and the case of the stationary memoryless source Mn n { n} → X r and Wn. We denote by P the approximated output distri- a and channel by Watanabe and Hayashi [11]. Recently, much Y˜n bution via Wn due to the input X˜n :=ϕ (U ) (cf. Fig. 1). attention has been paid to the channel resolvability because n Mn Precisionofthe approximationis measuredbythe variational this technique can be used to guarantee the strong secrecy in distance between P and P . physical-layer security systems [1], [5]. Thus, it is desirable Yn Y˜n Definition 1 (Variational Distance): Letting P and P tocharacterizethechannelresolvabilityforagivenpairofthe Z Z˜ be probability distributions on a countably infinite set , input distribution and the general channel. Z Inthispaper,wecharacterizetheδ-channelresolvabilityfor 1 a general source and a general channel with any δ ∈ [0,1). d(PZ,PZ˜):= 2 |PZ(z)−PZ˜(z)| (1) By taking the maximum over all possible general sources, z∈Z X we can naturally obtain the general formula for the worst is called the variational distance between P and P . ✷ input case. We also investigate the δ-channel resolvability Z Z˜ in an optimistic sense. When we restrict ourselves to the It is easily seen that 0 ≤ d(PZ,PZ˜) ≤ 1, where the left noiseless channel (identity mapping), the problem of channel inequality becomes equality if and only if PZ =PZ˜. resolvability reduces to the problem of source resolvability For any given sequence of random variables {Zn}∞n=1, we introduce quantities which play an important role in informa- ThisresearchissupportedbyJSPSKAKENHIGrantNumberJP16K06340. tion spectrum methods [3]. target distribution We define S∗(δ X,W) | Approximation is measured :=inf R: R is optimistically δ-achievable at X , { } by referredto as the optimistic δ-channelresolvability (at X). ✷ The following channel resolvability theorem is implicitly proved by Hayashi [5] for general sources and channels. approximated distribution Theorem1(Hayashi[5]):Letδ [0,1)befixedarbitrarily. ∈ For any general source X = Xn ∞ and any general Fig.1. Channel Resolvability System channel W = Wn ∞ , { }n=1 { }n=1 S(δ X,W) I (X;Y), (9) δ | ≤ Definition 2 (ε-Limit Superior in Probability): For ε [0,1], ∈ S∗(δ|X,W)≤I∗δ(X;Y), (10) where we define εp-limsupZn:=inf α:limsupPr Zn>α ε , (2) 1 Wn(Yn Xn) n→∞ n→∞ { }≤ I (X;Y):=δp-limsup log | , (11) (cid:26) (cid:27) δ n→∞ n PYn(Yn) εp∗-limsupZ :=inf α:liminfPr Z >α ε . (3) n→∞ n n→∞ { n }≤ I∗(X;Y):=δp∗-limsup 1 logWn(Yn|Xn). (12) n o δ n→∞ n PYn(Yn) For ε = 0, the right-hand sides of (2) and (3) are simply ✷ denoted by p-limsupZ and p∗-limsupZ , respectively. ✷ n n Unfortunately, Theorem 1 does not provide a lower bound n→∞ n→∞ on the δ-channel resolvability. For the worst input case, in The problem of channel resolvability has been introduced contrast, a lower bound has also been given by Hayashi [5]. by Han and Verdu´ [4]. Theorem 2 (Hayashi [5]): For any general channel W = fixDedefianribtiitornari3ly.(δA-CrheasnonlvealbRileitsyolrvaatbeilRity): L0eitsδsa∈id[t0o,1b)ebδe- {Wn}∞n=1, ≥ achievable at X if there exists a deterministic mapping ϕn : supI2δ(X;Y)≤supS(δ|X,W)≤supIδ(X;Y), (13) 1,...,M n satisfying X X X n { }→X supI∗ (X;Y) supS∗(δ X,W) supI∗(X;Y). (14) 1 2δ ≤ | ≤ δ limsup logM R, (4) X X X n n→∞ n ≤ In particular, linm→s∞upd(PYn,PY˜n)≤δ, (5) supS(0|X,W)=supI(X;Y), (15) X X where Y˜n denotes the output via Wn due to the input X˜n = supS∗(0X,W)=supI∗(X;Y), (16) ϕ (U ). We define X | X n Mn where we define S(δ|X,W):=inf{R: R is δ-achievable at X}, (6) 1 Wn(Yn Xn) I(X;Y):=p-limsup log | , (17) which is called the δ-channel resolvability (at X). ✷ n→∞ n PYn(Yn) Equation (5) requires d(PYn,PY˜n) ≤ δ +γ for all large I∗(X;Y):=p∗-limsup 1 logWn(Yn|Xn). (18) n, where γ > 0 is an arbitrary constant. We may consider n→∞ n PYn(Yn) a slightly weaker constraint, which requires d(PYn,PY˜n) ≤ ✷ δ + γ for infinitely many n The following problem is the weaker version of the δ-channel resolvability, introduced by III. MAIN THEOREMS:δ-CHANNELRESOLVABILITY [9] in the context of partial resolvability. Now, we give the general formulas for the δ-channel re- Definition 4 (Optimistic δ-Channel Resolvability): Let δ solvability at a specific input X and its optimistic version. ∈ [0,1) be fixed arbitrarily. A resolvability rate R 0 is ≥ said to be optimistically δ-achievable at X if there exists a Theorem3:Letδ [0,1)be fixedarbitrarily.Foranyinput ∈ deterministic mapping ϕ : 1,...,M n satisfying process X and any general channel W, n n { }→X 1 S(δ X,W)= inf I(Xˆ;Yˆ), (19) linm→s∞upnlogMn ≤R, (7) | Xˆ∈Bδ(X,W) S∗(δ X,W)= inf I(Xˆ;Yˆ), (20) linm→i∞nfd(PYn,PY˜n)≤δ. (8) | Xˆ∈Bδ∗(X,W) where Yˆ = Yˆn ∞ denotes the output process via W due in [6, Theorem 6]. Comparing the two characterizations, the to the input{proce}sns=X1ˆ = Xˆn ∞ , and we define following inequality is obvious for all δ [0,1): { }n=1 ∈ Bδ(X,W):= Xˆ ={Xˆn}∞n=1 :linm→s∞upd(PYn,PYˆn)≤δ , Xˆ∈Biδn(Xf ,W)I(Xˆ;Yˆ)≤Xˆ∈Biδn(Xf ,W) εs≥u0p, Dε(W||Z|Xˆ). (cid:26) (cid:27) Z∈B˜ε(Yˆ) B∗(X,W):= Xˆ = Xˆn ∞ :liminfd(P ,P ) δ . (25) δ n=1 n→∞ Yn Yˆn ≤ (Proof) The nproof i(cid:8)s giv(cid:9)en in Sec. IV. o✷ because D0(W||Yˆ|Xˆ) = I(Xˆ;Yˆ). Also, we have for all δ [0,1): Remark 1: The right-hand sides of (19) and (20) are ∈ nonincreasing functions of δ. Furthermore, these are right- inf I(Xˆ;Yˆ) inf sup D (W Z Xˆ). ε continuous in δ [0,1). ✷ Xˆ∈B∗(X,W) ≤Xˆ∈B∗(X,W) ε≥0, || | Remark 2: A∈s is mentioned in Theorem 1, Hayashi [5, δ δ Z∈B˜ε(Yˆ) (26) Theorem4]hasimplicitlyshownthatanyrateR>I (X;Y) δ isδ-achievableataspecificinputX.Therefore,weobtainthe These relationships are of use to prove Theorems 3 and 4. ✷ following relation between the right-hand side of (19) and δ- spectral sup-mutual information rate I (X;Y): IV. PROOF OF THEOREMS 3AND 4 δ inf I(Xˆ;Yˆ) I (X;Y) (δ [0,1)) (21) A. Finite-Length Bounds δ Xˆ∈Bδ(X,W) ≤ ∈ As we take an information spectrum approach to prove the and analogously general formulas in Theorems 3 and 4, we will use finite- length upper and lower bounds on the variational distance, inf I(Xˆ;Yˆ) I∗(X;Y) (δ [0,1)). (22) Xˆ∈B∗(X,W) ≤ δ ∈ which hold for each blocklength n. δ Intheproofofthedirectpart,we usethe followinglemma. We canfindexamplesofX andW forwhichtheinequalities in (21) and (22) are strict. This statement is also true even in Lemma 1 (Finite-Length Upper Bound [5]): Let Vn be an the case δ =0. ✷ arbitrary input random variable, and its corresponding output viaWn isdenotedbyZn.Then,foranygivenpositiveinteger Although the formulas established in Theorem 3 are suffi- cienttocharacterizeS(δ X,W)andS∗(δ X,W),itrequires Mn, there exists a mapping ϕn : 1,2,...,Mn n such { }→X | | that atedioustasktoderiveasingle-letterformulaforthestationary memoryless source and channel pair. We give alternative d(P ,P ) formulas in the following theorem: Zn Y˜n Theorem4: Letδ [0,1)be fixedarbitrarily.For anyinput 1 Wn(Zn Vn) 1 enc ∈ Pr log | >c + , (27) process X and any general channel W, ≤ (cid:26)n PZn(Zn) (cid:27) 2rMn S(δ X,W)= inf sup Dε(W Z Xˆ), (23) wherec 0isanarbitraryconstantandY˜ndenotestheoutput | Xˆ∈Bδ(X,W)Z∈εB≥˜ε0(,Yˆ) || | via Wn≥due to input X˜n =ϕn(UMn). ✷ In the proof of the converse part, we use the following S∗(δ X,W)= inf sup Dε(W Z Xˆ), (24) lemma. | Xˆ∈Bδ∗(X,W)Z∈εB≥˜ε0(,Yˆ) || | arbLietrmarmyapr2ob(aFbiinliittye-dLiesntrgibthutiLoonwoenr Bno.uTnhde):n,LfoertaPnZynunbifeoramn where Yˆ = Yˆn ∞ denotes the output process via W due randomnumberU ofsizeM anYdadeterministicmapping to input proc{ess X}ˆn==1 Xˆn ∞ , and we define ϕ : 1,2,...,MMn n wenhave { }n=1 n { n}→X Dε(W||Z|Xˆ):=εp-linm→s∞upn1 logWPnZ(Ynˆ(nYˆ|Xnˆ)n), d(PZn,PY˜n)≥Pr(n1 logWPnZ(Yn˜(nY˜|Xn˜)n) ≥c)− Mennc, (28) B˜ε(Y):=(cid:26)Z ={Zn}∞n=1 : linm→s∞upd(PYn,PZn)≤ε(cid:27). twohXe˜ren,X˜annd=c iϕsna(nUMarnb)it,raY˜ryncdoennsotatenst sthaetisofyuitnpgutMvina Wenncd.ue ≤ (Proof) The proof is given in Sec. IV. ✷ (Proof) First, we define Remark 3: Theorems 3 and 4 provide two formulas for the enc δ-channel resolvability S(δ X,W). Although the characteri- T := y n :P (y) P (y) . (29) | n ∈Y Y˜n ≥ M Zn zationin(23)ismorecomplicated,thisexpressioncanbeseen (cid:26) n (cid:27) as a counterpart of the alternative formula for the channel Then, by the definition of the variational distance, it is easily capacity given by Hayashi and Nagaoka [7, Theorem 1] verified that established for quantumchannels. The correspondingformula d(P ,P ) P (T ) P (T ), (30) fortheδ-channelcapacityoverclassicalchannelscanbefound Zn Y˜n ≥ Y˜n n − Zn n wherethesecondtermontheright-handsidecanbeevaluated Let V = Vn ∞ be a general source satisfying V { }n=1 ∈ as B (X,W) and δ M M PZn(Tn)=yX∈TnPZn(y)≤ ennc yX∈TnPY˜n(y)≤ ennc. (31) I(V;Z)≤Xˆ∈Biδn(Xf ,W)I(Xˆ;Yˆ)+γ, (40) To evaluate the first term on the right-hand side of (30), we whereZ = Zn ∞ denotestheoutputprocessviaW dueto { }n=1 borrow an idea given in [11]. Since the input process V. Setting M =en(I(V;Z)+2γ), it follows n from (39) and (40) that Mn 1 PY˜n(y)=Xi=1 MnWn(y|ϕn(i)) (y ∈Yn), (32) linm→s∞upn1 logMn =I(V;Z)+2γ ≤R. (41) denoting Wn (y)=Wn(y ϕ (i)), we have ϕn(i) | n Lemma 1 with c = I(V;Z)+γ guarantees the existence P (T ) of a deterministic mapping ϕ : 1,2,...,M n with Y˜n n n { n} → X the uniform random number U satisfying Mn 1 enc Mn = i=1 MnWϕnn(i)(cid:26)PY˜n(Y˜n)≥ MnPZn(Y˜n)(cid:27) linm→s∞upd(PZn,PY˜n) X 1 Wn(Zn Vn) Mn 1 Mn 1 enc limsupPr log | >I(V;Z)+γ = M Wϕnn(i) M Wϕnn(j)(Y˜n)≥ M PZn(Y˜n). ≤ n→∞ (cid:26)n PZn(Zn) (cid:27) n n n Xi=1 Xj=1  =0, (42) Here, noticing that   where Y˜n denotes the output via Wn due to the input X˜n = 1 M Wϕnn(i)(y)≥encPZn(y) ϕn(UMn). Then, the triangle inequality leads to n =⇒ M1 Wϕnn(j)(y)≥encPZn(y), (33) linm→s∞upd(PYn,PY˜n) n j we obtain the folloXwing lower bound: ≤linm→s∞upd(PYn,PZn)+nl→im∞d(PZn,PY˜n)≤δ, (43) Mn 1 where the last inequality is due to the fact V ∈ Bδ(X,W) P (T ) Wn Wn (Y˜n) encP (Y˜n) . and (42). Combining (41) and (43) concludes that R is δ- Y˜n n ≥ M ϕn(i) ϕn(i) ≥ Zn Xi=1 n n o achievable, and hence (35) holds. (34) To prove (36), for any given γ >0 setting Thus, plugging (31) and (34) into (30), we obtain (28). ✷ R= inf I(Xˆ;Yˆ)+3γ, (44) B. Proof of Theorems 3 and 4 Xˆ∈Bδ∗(X,W) The relations shown in (25) and (26) imply that to prove we show that R is optimistically δ-achievable. Let V = Theorems 3 and 4, it suffices to show Vn ∞ bea generalsourcesatisfyingV B∗(X,W)and { }n=1 ∈ δ S(δ X,W) inf I(Xˆ;Yˆ), (35) I(V;Z) inf I(Xˆ;Yˆ)+γ, (45) | ≤Xˆ∈Bδ(X,W) ≤Xˆ∈Bδ∗(X,W) S∗(δ X,W) inf I(Xˆ;Yˆ) (36) where Z = Zn ∞ denotes the output process via W | ≤Xˆ∈Bδ∗(X,W) due to input{V.}An=lo1ng the same line to prove (35), it in the direct (achievability) part and is easily verified that there exists a deterministic mapping ϕ : 1,2,...,M n satisfying (41) and (42). Then, S(δ X,W) inf sup D (W Z Xˆ), (37) n { n} → X | ≥Xˆ∈Bδ(X,W) ε≥0, ε || | the triangle inequality leads to Z∈B˜ε(Yˆ) liminfd(P ,P ) S∗(δ X,W) inf sup D (W Z Xˆ) (38) n→∞ Yn Y˜n ε | ≥Xˆ∈B∗(X,W) ε≥0, || | liminfd(P ,P )+ lim d(P ,P ) δ, (46) δ Z∈B˜ε(Yˆ) ≤ n→∞ Yn Zn n→∞ Zn Y˜n ≤ in the converse part. where the last inequality is due to the fact V B∗(X,W). ∈ δ 1) Direct part: First, fix γ >0 arbitrarily. Setting Combining (41) and (46) concludes that R is optimistically δ-achievable, and hence (36) holds. ✷ R= inf I(Xˆ;Yˆ)+3γ, (39) Xˆ∈Bδ(X,W) 2)Conversepart: Weshallprove(37)and(38)toestablish we show that R is δ-achievable, which means (35). the converse part of Theorems 3 and 4. Let R be δ-achievable. Then, there exists a mapping ϕ : The following relation can be obtained from Theorems 3 and n 1,2,...,M n satisfying (4) and (5). Let γ > 0 be 5, which gives a new characterization for H (X) and n δ { } → X fixed arbitrarily. From (4), we have H∗(X):=δp∗-limsup 1 log 1 . (55) 1 logM R+γ (47) δ n→∞ n PXn(Xn) n n ≤ for all sufficiently large n. Fixing an ε [0,1) arbitrarily, we Theorem 6: For any general source X, ochuotpoustevainayWZ d∈ueB˜tεo(Y˜in)p,uwt Xh˜ere=Y˜{X˜=n∈{=Y˜ϕnn}∞n(U=M1 nd)e}n∞no=te1s. tBhye Hδ(X)=Xˆ∈iB˜nδf(X)H(Xˆ), (56) using Lemma 2 with c= 1 logM +γ and (47), we have n n H∗(X)= inf H(Xˆ) (57) δ d(PZn,PY˜n) Xˆ∈B˜δ∗(X) for all δ [0,1), where 1 Wn(Y˜n X˜n) ∈ Pr log | >R+2γ e−nγ (48) ≥ (n PZn(Y˜n) )− B˜δ(X):= Xˆ = Xˆn ∞n=1 : linm→s∞upd(PXn,PXˆn)≤δ , (cid:26) (cid:27) for all sufficiently large n. Since Z B˜ (Y˜), we obtain (cid:8) (cid:9) ∈ ε B˜∗(X):= Xˆ = Xˆn ∞ : liminfd(P ,P ) δ . 1 Wn(Y˜n X˜n) δ n=1 n→∞ Xn Xˆn ≤ linm→s∞upPr(nlog PZn(Y˜|n) >R+2γ)≤ε. (49) Equations(5n6)and(cid:8)(57)i(cid:9)ndicatethatH (X)andH∗(X)oca✷n δ δ Since ε [0,1) and Z B˜ (Y˜) have been fixed arbitrarily, be viewed as “smoothed” 0-spectral sup-entropy rates. These ε (49) imp∈lies ∈ equationscanalso beprovendirectlyfromthepropertyof the ∗ δ-spectralsup-entropyratesH (X)andH (X),respectively. R+2γ sup D (W Z X˜). (50) δ δ ε ≥ ε≥0, || | VI. APPLICATIONOF GENERAL FORMULAS TO Z∈B˜ε(Y˜) MEMORYLESSSOURCE AND CHANNEL Since γ > 0 is arbitrary and X˜ B (X,W) follows from Now, let us consider a special case, where and are δ ∈ X Y (8), we obtain finite sets and for each n = 1,2, , both Xn and Wn are ··· memoryless with joint probability R inf sup D (W Z Xˆ), (51) ε Xwˆhe=re YXˆˆn=.≥{TYˆhXˆnu∈}s∞nB,=wδ(1Xed,oWebnt)oaZtien∈sε(B≥˜t3hε07(e,Y)ˆ.o)utput vi|a|W| due to input fPoXrnx(x=)W(nx(1y,.|x..),=xn(cid:26)) QQ∈nini==X11nPPXXan21d((xxyii))WW=21(((yyyii1||xx,.ii))..,ffyoonrr)eovd∈ednYnnn, Bδ∗T(hXe{,pWro})o,fcoofm(p3le8t)inigstahneaplorgoooufsofbtyheuscionngvetrhsee pfaacrtts.X˜ ✷∈ rWwehspe=erect{XivWejlnya}.n∞ndT=hW1eajrse(ojucro=cme1p,Xle2t)el=dyecnh{oaXteranac}tes∞nor=iuz1recdaenbadyndPthXae1cWchha1anninfneenll, V. SOURCE RESOLVABILITY: REVISITED is odd and by PX2W2 if n is even and are known as one of WhenthechannelWnisanidentitymapping,theaddressed the simplest examplesfor which S(δ|X,W) and S∗(δ|X,W) do not coincide in general [3]. Let Y denote the output via problem reduces to the problem of source resolvability [3], j where the target distribution is the general source Xn itself. Wj due to input Xj for j = 1,2. The alternative formulas (23) and (24) are of use to prove the converse parts. In this case, we denote S(δ X,W) simply by S(δ X). | | Theorem 7: For any δ [0,1), For this problem, Steinberg and Verdu´ [10] have shown the ∈ followingtheorem,whichgeneralizestheresolvabilitytheorem S(δ X,W)= max inf I(Xˆ ;Yˆ ), (58) j j established by Han and Verdu´ [4] for δ =0: | j=1,2Xˆj∈B0(Xj,Wj) Theorem 5 (HanandVerdu´ [4], SteinbergandVerdu´ [10]): S∗(δ X,W)= min inf I(Xˆ ;Yˆ ), (59) For any target general source X, | j=1,2Xˆj∈B0(Xj,Wj) j j S(δ|X)=Hδ(X) (δ ∈[0,1)), (52) where Yˆj denotes the output via Wj due to the input Xˆj, where I(Xˆj;Yˆj) denotes the mutual information between Xˆj and Yˆ , and we define B (X ,W ):= Xˆ :P =P . Hδ(X):=δp-linm→s∞upn1 logPXn1(Xn) (53) (PjrIotosfh)ouTlhdebperonooftiicse0dgivthejantitnhjeAc.on(cid:8)stanjt δ dYojes noYˆtja(cid:9)ppear i✷n is the δ-spectral sup-entropy rate for X. ✷ formulas (58) and (59). This result indicates that the strong When the channel Wn is an identity mapping, we have converse holds for the memoryless source and channel pair. I(X;Y)=H(X) because Precisely, for any 1 logWn(Yn|Xn) = 1 log 1 a.s. (54) R< min inf I(Xˆj;Yˆj), (60) n PYn(Yn) n PXn(Xn) j=1,2Xˆj∈B0(Xj,Wj) anymappingϕ : 1,...,M n satisfying(7)produces Hence, only the case R = S(δ X,W) is of our interest. n n { }→X | tY˜hen vdaernioattieosntahledoisuttapnuctevdia(PWYnn,dPuY˜ent)o→inp1ut(Xn˜n→=∞ϕ),(Uwher)e. sSoimlvialbairlliyt,y,wthheencadseiscRus=sinSg∗(tδheXo,pWtim)isisticou(rδp,Rrim)-acrhyaninnteelrerset-. For an i.i.d. sourceX with X =X =X and a sntatioMnanry | ✷ 1 2 memoryless channel W with W = W = W , we obtain the Now, we establish the general formulas for the second- 1 2 following corollary from Theorem 7, which has been proved orderresolvability.Thefollowingtwotheoremscanbeproven by Watanabe and Hayashi [11]. analogously to Theorems 3 and 4 in the first-order case. Corollary 1 (Watanabe and Hayashi [11]): For any i.i.d. Theorem 8: Let δ [0,1) and R 0 be fixed arbitrarily. ∈ ≥ input source X and any stationary memoryless channel W, For any input process X and any general channel W, S(δ X,W)=S∗(δ X,W)= inf I(Xˆ;Yˆ) (61) T(δ,RX,W)= inf I(RXˆ;Yˆ), (67) | | Xˆ∈B0(X,W) | Xˆ∈Bδ(X,W) | for every δ [0,1), where Yˆ denotes the output via W T∗(δ,RX,W)= inf I(RXˆ;Yˆ), (68) induced by in∈put Xˆ. ✷ | Xˆ∈Bδ∗(X,W) | where Yˆ = Yˆn ∞ denotes the output process via W due VII. SECOND-ORDER CHANNEL RESOLVABILITY to the input{proce}sns=X1ˆ = Xˆn ∞ , and we define { }n=1 We turn to considering the second-order resolution rates 1 Wn(Yn Xn) [11]. First, we define the second-order achievability. I(RX;Y):=p-limsup log | nR . Definition 5 ((δ,R)-Channel Resolvability): Let δ [0,1) | n→∞ √n(cid:18) PYn(Yn) − (cid:19) ∈ ✷ and R 0 be fixed arbitrarily. A resolvability rate L is said ≥ We give alternative formulas in the following theorem, to be (δ,R)-achievable at X if there exists a deterministic whichcorrespondtoTheorem4onthefirst-orderresolvability mapping ϕ : 1,...,M n satisfying n { n}→X rates: 1 Theorem 9: Let δ [0,1) and R > 0 be fixed arbitrarily. linm→s∞up√n(logMn−nR)≤L, (62) For any input process∈X and any general channel W, linm→s∞upd(PYn,PY˜n)≤δ, (63) T(δ,R|X,W)=Xˆ∈Biδn(Xf ,W) εs≥u0p, Jε(R|W,Z,Xˆ), where Y˜n denotes the output via Wn due to the input X˜n = Z∈B˜ε(Yˆ) (69) ϕ (U ). We define n Mn T∗(δ,RX,W)= inf sup J (RW,Z,Xˆ), ε T(δ,RX,W):=inf L: L is (δ,R)-achievable at X , | Xˆ∈B∗(X,W) ε≥0, | | { } δ Z∈B˜ε(Yˆ) which is called the (δ,R)-channel resolvability (at X). ✷ (70) As in the first-order case, we addressthe relaxed constraint where Yˆ = Yˆn ∞ denotes the output process via W due on the variational distance. to input proc{ess X}ˆn==1 Xˆn ∞ , and we define Definition 6 (Optimistic (δ,R)-Channel Resolvability): Let { }n=1 δ [0,1) and R 0 be fixed arbitrarily. A resolvability J (RW,Z,Xˆ) ∈ ≥ ε | rate L is said to be optimistically (δ,R)-achievable at X if there existsa deterministicmappingϕ : 1,...,M n 1 Wn(Yˆn Xˆn) n { n}→X :=εp-limsup log | nR . satisfying n→∞ √n PZn(Yˆn) − ! 1 ✷ limsup (logM nR) L, (64) n→∞ √n n− ≤ When the channel is an identity mapping, the problem addressed here reduces to finding the second-order δ-source liminfd(P ,P ) δ, (65) n→∞ Yn Y˜n ≤ resolvability[8].Inthiscase,wedenoteT(δ,RX,W)simply | by T(δ,RX). Nomura and Han [8] have established the fol- where Y˜n denotes the output via Wn due to the input X˜n = | lowing fundamental theorem, which generalizes the theorem ϕ (U ). We define n Mn on the first-order δ-source resolvability given by [4], [10]: T∗(δ,RX,W) Theorem 10 (Nomura and Han [8]): For any target general | source X, :=inf L: L is optimistically (δ,R)-achievable at X , { } S(δ,RX)=Hδ(RX) (δ [0,1)), (71) | | ∈ called the optimistic (δ,R)-channel resolvability (at X). ✷ where Remark 4: By definition, it is easily verified that 1 1 H (RX):=δp-limsup log nR . T(δ,RX,W)= +∞ for R<S(δ|X,W) (66) δ | n→∞ √n(cid:18) PXn(Xn) − (cid:19) | for R>S(δ X,W). ✷ (cid:26) −∞ | Since the channel Wn is the identity mapping, we have whereV = Vn ∞ andZ = Zn ∞ . On the otherhand, { }n=1 { }n=1 I(RX;Y) = H(RX). The following relation can be ob- because it obviously holds that V B (X,W), we have δ | | ∈ tained from Theorems 8 and 10, which gives a new represen- inf I(Xˆ;Yˆ) I(V;Z). (79) tation for Hδ(R|X) and Xˆ∈Bδ(X,W) ≤ H∗(RX):=δp∗-limsup 1 log 1 nR . Sinceγ >0isanarbitraryconstant,(75),(78)and(79)imply δ | n→∞ √n(cid:18) PXn(Xn) − (cid:19) inf I(Xˆ;Yˆ) inf I(Xˆ ;Yˆ ). (80) 1 1 Xˆ∈Bδ(X,W) ≤Xˆ1∈B(X1,W1) Theorem 11: For any general source X, n (ii) For an arbitrary fixed γ > 0, let X be n i.i.d. samples Hδ(RX)= inf H(RXˆ), (72) from source P satisfying X B(X2,W ) and (75) with | Xˆ∈B˜δ(X) | X2 n 2 ∈ 2 2 j = 2. Also, let X be n i.i.d. samples from source P H∗(RX)= inf H(RXˆ) (73) satisfying I(X ;Y )1= 0, where Y denotes the output vXi1a for all δ [0,1)δ an|d R Xˆ0∈,B˜wδ∗h(Xer)e we d|efine H(RXˆ) = Wan1d (dVuen,toZnin)p=1ut(XX11n2.,YSen2t)(fVorn,eZvenn)1n=. T(hXenn1,,Ywen1)obfotarinodd n H0(R|Xˆ)∈and H∗(R|Xˆ)≥=H∗0(R|Xˆ). | ✷ E 1 logWn(Zn|Vn) I(X ;Y ) for all n. Equation (73) as well as (72) can be provendirectly from the (cid:26)n PZn(Zn) (cid:27)≤ 2 2 definitionofthequantitiesonbothsides.Aswasshownin(56) Again, by the weak law of large numbers, we have and(57)inthefirst-ordercase,so-calledsmoothingoperations ∗ appear here; both H (RX) and H (RX) are characterized 1 Wn(Zn Vn) bB˜y∗(HX(R) |cXeˆn)teorefdaagteXδne,rar|lesspoeucrtcieveXlˆy.δin t|he δ-ball B˜δ(X) and nl→im∞Pr(cid:26)nlog PZn(Z|n) >I(X2;Y2)+γ(cid:27)=0(,81) δ APPENDIX A indicating that PROOF OF THEOREM7 I(V;Z) I(X ;Y )+γ. (82) 2 2 ≤ 1) Direct part: On the other hand, it holds that V B∗(X,W) because Without loss of generality, we assume that ∈ δ inf I(Xˆ1;Yˆ1) inf I(Xˆ2;Yˆ2). (74) linm→i∞nfd(PYn,PZn)≤likm→i∞nfd(PY2k,PZ2k)=0. (83) Xˆ1∈B(X1,W1) ≥Xˆ2∈B(X2,W2) Then, we have n (i) First, fixγ >0 arbitrarily.Forj =1,2,letX beni.i.d. j inf I(Xˆ;Yˆ) I(V;Z). (84) samples from source PXj satisfying Xj ∈B(Xj,Wj) and Xˆ∈Bδ∗(X,W) ≤ I(X ;Y ) inf I(Xˆ ;Yˆ )+γ, (75) Since γ > 0 is an arbitrary constant, (75) with j = 2, (82) j j j j ≤Xˆj∈B(Xj,Wj) and (84) imply (wVhner,eZYn)j=de(nXotn1e,sYthn1e) foourtpoudtdvniaaWndj(dVune,Zton)in=put(XXn2j,.YSn2e)t Xˆ∈Biδ∗n(Xf ,W)I(Xˆ;Yˆ)≤Xˆ2∈Bi(nXf2,W2)I(Xˆ2;Yˆ2). (85) for even n. Since the random variable 2) Converse part: 1 logWn(Zn|Vn) = 1 k logW(Zi|Vi) (76) As was argued in [11], we shall use the method of types n PZn(Zn) n i=1 PZi(Zi) [2]. The following notation is introduced. is a sum of independent randomXvariables, where Vn = • Let Px denote the type of x ∈ Xn, i.e., Px(a) denotes the number of occurrence of symbol a in x. v(Val1u,eV2sa,t.i.s.fi,eVsn) and Zn = (Z1,Z2,...,Zn), its expected • Let Pxy denote the joint type of (x,y)∈∈XXn×Yn. E(cid:26)n1 logWPnZ(nZ(nZ|Vn)n)(cid:27)≤I(X1;Y1)+γ for all n. •• dLDieesttfirnPibexuWtthioe(nbs)oetn:s=Yof.Pε-atyPpxic(aal)Wseq(ub|ean)cedsenasote the marginal The weak law of large numbers guarantees TYn,ε:={y∈Yn :|Py(b)−PY(b)|≤ε, ∀b∈Y}, (86) 1 Wn(Zn Vn) nl→im∞Pr(cid:26)nlog PZn(Z|n) >I(X1;Y1)+γ(cid:27)=0, TWn,ε(x):={y∈Yn :|Pxy(a,b)−Px(a)W(b|a)|≤ε, (77) ∀(a,b)∈X ×Y}, (87) which indicates that AY(ε):={P ∈P(X):|PW(b)−PY(b)|≤2|X|ε, I(V;Z) I(X1;Y1)+γ, (78) ∀b∈Y}. (88) ≤ Now, we are in a position to prove the converse part of We invoke the method of squeezing a subsequence of good Theorem 7. We again assume (74) without loss of generality. types in the information spectrum approach as in [12]. Equa- In view of Theorems 3 and 4, we shall show tion (95) implies that there exists some d >0:d >d > n 1 2 { δ satisfying inf sup D (W Z Xˆ) ···→ } ε Xˆ∈Bδ(X,W) ε≥0, || | 1 Wn(Yˆn Xˆn) inZf∈B˜ε(YˆI)(Xˆ ;Yˆ ), (89) dn ≥Pr(nlog PYn(Yˆ|n) >R+2γ) (97) 1 1 ≥Xˆ1∈B(X1,W1) for all n=1,2, . Since ··· and 1 Wn(Yˆn Xˆn) Pr log | >R+2γ inf sup Dε(W Z Xˆ) (n PYn(Yˆn) ) Xˆ∈B∗(X,W) ε≥0, || | δ Z∈B˜ε(Yˆ) = P (x)Wn 1 logWn(Yˆn|x) >R+2γ , inf I(Xˆ2;Yˆ2). (90) x∈Xn Xˆn x (n PYn(Yˆn) ) ≥Xˆ2∈B(X2,W2) X where we use Wn to denote Wn( x) for simplicity, (97) (i) To show (89), we first fix an arbitrary indicates that therexexists some x ·| n satisfying n ∈X R>Xˆ∈Biδn(Xf ,W)Z∈εsB≥˜uε0p(,Yˆ)Dε(W||Z|Xˆ), (91) dn ≥Wxnn(n1 logWPnY(nYˆ(nYˆ|xn)n) >R+2γ). (98) and we shall show that R is not smaller than the right-hand It is important to use the fact following from (98) that there side of (89). For simplicity, we define existsa sequenceofoddnumbers n <n < and 1 2 { ···→∞} P ( ) such that 1 if n is odd γ ∈P X j(n)= (92) 2 if n is even. limsupd δ, lim P =P , (99) (cid:26) i→∞ ni ≤ i→∞ xni γ Then, we can write Xn = Xn and Wn = Wn and the j(n) j(n) where P denotes the type of x n (cf. [12]). The corresponding output is Yn = Yn . Letting γ > 0 be arbi- xn n ∈ X j(n) existenceofsuchaconvergentpointP ( )followsfrom atrnadrilsyetfitxheed,fowlleowdeinfignepγro′b:a=bi|lXity|γd,iτstnri:b=utiPorn{Yonn ∈nT:Ynj(n),γ′} thefactthatP(X)isacompactsetforγfi∈nitPeXX.Fornotational Y simplicity, we use k to denote (odd number) k = n1,n2, ··· P (y)1 y Tn so that (98) and (99) can be rewritten as PYn(y):= Yn {τn∈ Yj(n),γ′} (y ∈Yn), (93) 1 Wk(Yˆk x ) d Wk log | k >R+2γ (100) where 1{E} is the indicator function for the event E. Then, k ≥ xk(k PYk(Yˆk) ) from the property of the set of γ′-typical sequences Tn , Yj(n),γ′ and we have τ 1 as n and hence n → →∞ limsupd δ, lim P =P , (101) nl→im∞d(PYn,PYn)=0. (94) k→∞ k ≤ k→∞ xk γ respectively. The following lemma is of use. Now, we can see that by (91) there exists an Xˆ Bδ(X,W) satisfying ∈ (10L0e)mamnad3(1:0A1s)swumithe tshoamtexkδ∈X[0,k1()ka=ndnP1,n2,··(·))s,awtishfieeres γ ∈ ∈P X R> sup Dε(W Z Xˆ) γ k = n1,n2,··· denotes either odd or even numbers. If k ε≥0, || | − denotes odd numbers, then Z∈B˜ε(Yˆ) P A (2γ), (102) D (W Y Xˆ) γ, (95) γ ∈ Y1 δ ≥ || | − whereas if k denotes even numbers, then where Yˆ = Yˆn ∞ denotes the output via W due to input XYˆ =B{˜Xˆ(nYˆ}∞n){=w1h,i}acnnhd=f1toolldoewrisvefr(o9m5)XwˆehaBve(uXse,dWth)atafnadct(t9h4a)t Pγ ∈AY2(2γ). (103) with∈theδtriangle inequality: ∈ δ (Proof) Let k = n1,n2,··· denote odd numbers. Suppose that P A (2γ). From the right inequality in (101) we γ 6∈ Y1 linm→s∞upd(PYˆn,PYn) o[1b1ta,iLnePmxmka6∈2]AhYa1v(eγ)shfoowrnaltlhalatrigfeyk∈. WTWakt1a,nγa(bxek)a,ntdheHnayahi ≤nl→im∞d(PYn,PYn)+linm→s∞upd(PYn,PYˆn)≤δ. (96) y 6∈TYk1,γ′. (104) Further, if y Tk , then P (y) = 0 by definition, and the weak law of large numbers and under the conditional thus 6∈ Y1,γ′ Yk probability distribution Wk , yielding xk k1 logWPkY(ky(|yx)k) >R+2γ, (105) likm→s∞upWxkk(k1 logWPkY(kYˆ(kYˆ|xk)k) >R+3γ) Therefore, for all y k we have 0 if D(W P P )<R+3γ, ∈Y = 1|| Y1| γ (111) 1 if D(W P P )>R+3γ 1 Wk(y x ) (cid:26) 1|| Y1| γ 1 y Tk (x ) 1 log | k >R+2γ . ∈ W1,γ k ≤ (k PYk(y) ) and from the left inequality in (101) and (108), we obtain (cid:8) (cid:9) R+3γ D(W P P ). (112) Since the set of γ-typical sequences TWk1,γ(xk) satisfies ≥ 1|| Y1| γ Since γ >0 is arbitrary, taking the limit γ 0 for both sides, Wxkk Yˆk ∈TWk1,γ(xk) →1 (k →∞), (106) we obtain ↓ n o this inequality and (100) leads to R limD(W P P ) ≥ γ↓0 1|| Y1| γ kl→im∞dk =1, (107) =D(W1||PYˆ1|PXˆ1)=I(Xˆ1,Yˆ1) (113) which is a contradiction, and hence (102) holds. with some Xˆ B(X ,W ), where Yˆ denotes the output 1 1 1 1 In the case of even numbers k = n ,n , , (103) can be via W due to ∈input Xˆ . Here, we have used the fact that 1 2 1 1 proven analogously. ··· ✷ P A (γ) by Lemma 3 and A (γ) B(X ,W ) as γ ∈ Y1 Y1 → 1 1 Since P (y) P (y)/τ for all y k, we can bound γ 0. Thus, we have Yk ≤ Yk k ∈Y ↓ the right-hand side of (100) from below as R inf I(Xˆ ;Yˆ ), (114) 1 1 1 Wk(Yˆk x ) 1 1 ≥Xˆ1∈B(X1,W1) d Wk log | k >R+2γ+ log k ≥ xk(k PYk(Yˆk) k τn) completing the proof of (89). (ii) To show (90), we first fix an arbitrary 1 Wk(Yˆk x ) ≥Wxkk(k log PYk(Yˆ|k)k >R+3γ) (k ≥k0), R>Xˆ∈Bi∗n(Xf ,W) εs≥u0p, Dε(W||Z|Xˆ). (115) (108) δ Z∈B˜ε(Yˆ) where the second inequality holds for all large odd numbers Recall that we can write Xn = Xn and Wn = Wn j(n) j(n) k. Since the random variable and the corresponding output is Yn = Yn with definition j(n) (92). Let γ > 0 be arbitrarily fixed. We define γ′ := γ, 1 logWk(Yˆk|xk) = 1 k logW1(Yˆi|xk,i) (109) τn :=Pr{Yn ∈TYnj(n),γ′}andsetPYn againasin(93).T|Xhe|n, k PYk(Yˆk) k i=1 PY1(Yˆi) from the property of the set of γ′-typical sequences TYnj(n),γ′, X we have (94). is a sum of conditionallyindependentrandomvariables given Now, for any general source Xˆ = Xˆn ∞ it is easily Xˆk = x = (x ,x ,...,x ), its expected value can be verified that { }n=1 k k,1 k,2 k,k evaluated as sup D (W Z Xˆ) sup D∗(W Z Xˆ), (116) 1 Wk(Yˆk x ) ε≥0, ε || | ≥ ε≥0, ε || | E(k log PYk(Yˆ|k)k Xˆk =xk) Z∈B˜ε(Yˆ) Z∈B˜ε∗(Yˆ) (cid:12) (cid:12) where Yˆ = Yˆn ∞ denotes the output via W due to input 1 k (cid:12) W1(bxk,i) Xˆ and we d{efine}n=1 = W (bx )log | 1 k,i k | P (b) Xi=1bX∈Y Y1 D∗(W Z Xˆ):=εp∗-limsup 1 logWn(Yˆn|Xˆn), 1 W1(bxk,i) ε || | n→∞ n PZn(Yˆn) = kP (a) W (bx )log | k xk 1 | k,i P (b) aX∈X bX∈Y Y1 B˜ε∗(Y):= Z ={Zn}∞n=1 : linm→i∞nfd(PYn,PZn)≤ε =:D(W1||PY1|Pxk), (110) since for all εn>0 and Z it holds that o whereD(W P P )istheconditionaldivergencebetween 1|| Y1| xk D (W Z Xˆ) D∗(W Z Xˆ). (117) W1 and PY1 given Pxk ∈ P(X). Then, we can invoke ε || | ≥ ε || | We can see that by (115) and (116) there exists an Xˆ REFERENCES ∈ B∗(X,W) satisfying δ [1] M.R.BlochandJ.N.Laneman,“Strongsecrecyfromchannel resolv- R> sup D∗(W Z Xˆ) γ ability,” IEEETrans.Inf.Theory, vol.59,no.12,pp.8077–8098,Dec. ε || | − 2013. ε≥0, Z∈B˜∗(Yˆ) [2] I. Csisza´r and J. Ko¨rner, Information Theory: Coding Theorems for ε Discrete Memoryless Systems, 2nd ed., Cambridge University Press, D∗(W Y Xˆ) γ (118) Cambridge, U.K.,2011. ≥ δ || | − [3] T. S. Han, Information-Spectrum Methods in Information Theory, Springer, 2003. wwhheicrhetofodlleorwivsef(r1o1m8)wXˆehaveBu∗s(eXdt,hWat)facatntdha(t9Y4)∈wBi˜thδ∗(Ytˆhe) [4] ITE.ESE. THraannsa.nIdnf.ST.hVeeorrdyu´,,v“oAl.p3p9ro,xniom.a3ti,opnp.th7e5o2r–y77o1f,oMutapyut19st9a3ti.stics,” ∈ δ [5] M.Hayashi,“Generalnonasymptoticandasymptoticformulasinchan- triangle inequality: nelresolvability andidentification capacity andtheirapplication tothe wiretapchannel,” IEEETrans.Inf.Theory,vol.52,no.4,Apr.2006. linm→i∞nfd(PYˆn,PYn) [6] M. Hayashi, “Information spectrum approach to second-order coding rateinchannel coding,”IEEETrans.Inf.Theory,vol.55,no.11,Nov. ≤nl→im∞d(PYn,PYn)+linm→i∞nfd(PYn,PYˆn)≤δ. (119) [7] 2M0.0H9.ayashiandH.Nagaoka,“Generalformulasforcapacityofclassical- Equation (118) implies that there exists some d > 0 ∞ quantum channels,” IEEETrans.Inf.Theory, vol.49,no.7,pp.1753– { n }n=1 1768,July2003. satisfying [8] R. Nomura and T. S. Han, “Second-order resolvability, intrinsic ran- domness,andfixed-lengthsourcecodingformixedsources:information liminfdn δ (120) spectrum approach,” IEEETrans.Inf.Theory,vol.59,no.1,pp.1–16, n→∞ ≤ Jan.2013. and [9] Y. Steinberg, “New converses in the theory of identification via chan- nels,”IEEETrans.Inf.Theory,vol.44,no.3,pp.984–997,May1998. 1 Wn(Yˆn Xˆn) [10] Y. Steinberg and S. Verdu´, “Simulation of random processes and rate- dn ≥Pr(nlog PYn(Yˆ|n) >R+2γ) (121) Jdaisnt.or1t9io9n6.theory,” IEEE Trans. Inf. Theory, vol. 42, no. 1, pp. 63–86, [11] S. Watanabe and M. Hayashi, “Strong converse and second-order for all n=1,2, . Also, (120) indicates that at least one of asymptotics of channel resolvability,” Proc. IEEE Int. Symp. on Inf. ··· the following inequalities holds: Theory,Jun.2014. [12] H. Yagi, T. S. Han, and R. Nomura, “First- and second-order coding liminfd δ or liminfd δ. (122) theorems formixedmemoryless channels with general mixture,” IEEE 2k+1 2k k→∞ ≤ k→∞ ≤ Trans.Inf.Theory,vol.68,no.8,pp.4395–4412, Aug.2016 [13] H. Yagi and T. S. Han, “Variable-length resolvability for general First, we assume that sources,”submittedtoIEEEInt.Symp.onInf.Theory,Jan.2017. liminfd δ (123) k˜→∞ k˜ ≤ for odd k˜ = 1,3, . Similarly to the derivation of (100) ··· and (101), (118) indicates that there exists some x n, a n ∈X sequence k =n ,n , , where n ,n , are odd numbers, 1 2 1 2 ··· ··· and P ( ) such that γ ∈P X 1 Wk(Y˜k x ) d Wk log | k >R+2γ . (124) k ≥ xk(k PYk(Y˜k) ) and lim d δ, lim P =P . (125) k→∞ k ≤ k→∞ xk γ From Lemma 3, we have P A (2γ). (126) γ ∈ Y1 Then, we can invoke the weak law of large numbersas in the derivation of (114) to yield R inf I(Xˆ ;Yˆ ) 1 1 ≥Xˆ1∈B(X1,W1) inf I(Xˆ ;Yˆ ), (127) 2 2 ≥Xˆ2∈B(X2,W2) where we have used (74) for the last inequality. Thus, we obtain (90). In the case where (123) holds for even k˜ = 2,4, , we can show (127) in the analogous way. ··· ✷

See more

The list of books you might like

Most books are stored in the elastic cloud where traffic is expensive. For this reason, we have a limit on daily download.