ebook img

Adaptive Replication in Distributed Content Delivery Networks PDF

0.34 MB·
Save to my drive
Quick download
Download
Most books are stored in the elastic cloud where traffic is expensive. For this reason, we have a limit on daily download.

Preview Adaptive Replication in Distributed Content Delivery Networks

Adaptive Replication in Distributed Content Delivery Networks M. Leconte M. Lelarge L. Massoulie´ Technicolor - INRIA INRIA - E´cole Normale Supe´rieure Microsoft Research - INRIA Joint Center [email protected] [email protected] [email protected] Abstract—We address the problem of content replication in the data center is much more costly than by a small server, large distributed content delivery networks, composed of a data thus requests are served from the small servers in priority. In 4 center assisted by many small servers with limited capabilities fact, we only send requeststo the data center if no idle server 1 and located at the edge of the network. The objective is to storing the corresponding content is present. We assume the 0 optimize the placement of contents on the servers to offload 2 as much as possible the data center. We model the system data center is dimensionedto absorbany such remainingflow constituted by the small servers as a loss network, each loss of requests. We restrict our attention here to contents which n corresponding to a request to the data center. Based on large require immediate service, which is why we do not allow a system / storage behavior, we obtain an asymptotic formula for J requests to be delayed or queued. Stringent delay constraints theoptimalreplicationofcontentsandproposeadaptiveschemes are indeed often the case for video contents –for example 8 relatedtothoseencounteredincachenetworksbutreactinghere to loss events, and faster algorithms generating virtual events in the context of a Video-on-Demand (VoD) service–, which ] at higher rate while keeping the same target replication. We represent already more than half of the Internet traffic. I N showthroughsimulationsthatouradaptiveschemesoutperform In that context, we address the problem of determining significantlystandard replication strategies both in terms of loss s. rates and adaptation speed. whichcontentstostoreinthesmallservers.Thegoalistofind c contentplacementstrategieswhichoffloadasmuchaspossible [ I. INTRODUCTION thedatacenter.Inthisoptic,itisobviousthatpopularcontents 1 shouldbemorereplicated,butthereisstillalotoffreedomin The amount of multimedia traffic transfered over the In- v thereplicationstrategy.Inorderto comparedifferentpolicies, 0 ternet is steadily growing, mainly driven by video traffic. A we model the system constituted by the small servers alone 7 large fraction of this traffic is already carried by dedicated as a loss network, where each loss corresponds to a request 7 content delivery networks (CDNs) and this trend is expected to the data center. The relevant performance metric is then 1 to continue[1]. The largestCDNs are distributed,with a herd . simplythe expectednumberof losses inducedby eachpolicy. 1 ofstoragepointsanddatacentersspreadovermanynetworks. 0 With the cooperationof the InternetService Providers, CDNs We make three main contributions:first, using a mean-field 4 heuristic for our system with a large number of servers with can deploy storage points anywhere in the network. How- 1 limited coordination, we compute accurately the loss rate of ever, these cooperations require a lot of negociations and are : v therefore only slowly getting established, and it also seems eachcontentbasedonitsreplication;buildingonthisanalysis, i we derive an optimized static replication policy and show X necessary to consider alternatives to a global cache-network that it outperforms significantly standard replications in the planning. An easily deployable solution is to make use of r a the ressources (storage, upload bandwidth) already present at literature; finally, we propose adaptive algorithms attaining the optimal replication. Contrary to most caching schemes, the edge of the network. Such an architecture offers a lot of ourevictionrulesarebasedonloss statistics insteadofaccess potential for cost reduction and quality of service upgrade, statistics; in particular our algorithms do not know or learn but it also comes with new design issues [2] as the operation the popularities of the contents and will adapt automatically. of a very distributed overlay of small caches with limited We also propose a simple mechanism relying heavily on our capabilities is harder than that of a more centralized system mean-fieldanalysisthatallowsustospeedupsignificantlythe composed of only a few large data centers. rate of adaptation of our algorithms. At each step, extensive Inthispaper,weconsiderahybridCDNconsistingofbotha simulations support and illustrate our main points. data center and an edge-assistance in the form of many small servers located at the edge of the network; we call such a The rest of the paper is organized as follows: we review system an edge-assisted CDN. The small servers are here to relatedworkinSectionIIanddescribemoreindetailoursys- reducethecostofoperationoftheCDN,whilethedatacenter temmodelinSectionIII.InSectionIV,weanalyseourmodel helpsprovidereliabilityand qualityofserviceguarantees.We under a mean-field approximation and obtain an asymptotic stressthefactthatthesmallservershavealimitedstorageand expression for the optimal replication. Then, in Section V, service capacity as well as limited coordination capabilities, leveraging the insights from the analysis we propose new which makes cheap devices eligible for that role. As their adaptive schemes minimizing the average loss rate, as well respective roles hint, we assume the service of a request by as ways to further enhance their adaptation speed. 2 II. RELATEDWORK trafficwithvarioustypesofcontents.Basedonthedifferences in their respective popularity distributions, [7] advocates that Ouredge-assistedCDNmodelisrelatedtoafewothermod- VoDcontentsarecachedveryclosetothenetworkedgewhile els. A major difference though is our focus on modeling the user-generated contents, web and file sharing are only stored basic constraints of a realistic system, in particular regarding in thenetworkcore.Ina loosesense, we also studysuch type the limited capacity of servers (in terms of storage, service of two-layer systems, with the small servers at the edge of and coordination) and the way requests are handled, i.e. the the network and the data center in the core, and address the matching policy, which role is to determine which server an question of determining whether to store contents at the edge incomingrequestshouldbedirectedto.Weindeedconsiderthe (and in what proportion)or to rely on the core to serve them. most basic matchingpolicy:an idle server storing the content Concerning P2P VoD models, content pre-fetching strate- is chosen uniformly at random; and no queueing of requests, gies have been investigated in many papers (e.g. [3], [8], nore-directionduringserviceandnosplittingareallowed.The [9],[10]),whereoptimalreplicationstrategiesand/oradaptive reason for modeling such stringent constraints and refusing schemeswerederived.Theyworkunderamaximummatching simplifyingassumptionsinthisdomainisthatthisshouldlead assumptionandmostofthemassumeinfinitedivisibilityofthe to fundamental intuitions and qualitative results that should requests,i.e. thata requestcan be servedin parallelby allthe apply to most practical systems. serversstoringthecorrespondingcontent;wewanttotakeinto Our edge-assisted CDN model is very similar to the dis- account more realistic constraints in that respect. tributed server system model of [3], [4]; we merely use a Onecanalsoexplorethedirectionofoptimizingdistributed different name to emphasize the presence in the core of systems with a special care for the geographical aspects as the network of a data center, which makes it clear what in [11], [12], [13]. These papers solve content placement and performance metric to consider, while otherwise availability matchingproblemsbetweenmanyregions,whilenotmodeling of contents for example may become a relevant metric. The in a detailed way what happens inside a region. edge-assistedCDNmodelisalsorelatedtopeer-to-peer(P2P) Finally, most of the work in these related domains fo- VoD systems, which however have to deal with the behavior cuses on different performance metrics: in hierarchical cache of peers, and to cache networks, where caches are deployed networks, [14] addresses the problem of joint dimensioning in a potentially complex overlay structure in which the flow of caches and content placement in order to reduce overall ofrequestsenteringacacheisthe“miss”flowofanotherone. bandwidth consumed; [15] optimizes replication in order to Previous work on P2P VoD typically does not model all the minimize search time and devises elegant adaptive strategies fundamentalconstraintswementioned,andthecachenetworks towards that end. Surprisingly, for various alternative objec- literaturetendstocompletelyobliteratethefactserversbecome tives and network models, the proportional (to popularity) unavailable while they are serving requests and to focus on replicationexhibitsgoodproperties:itminimizesthedownload alternativeperformancemetricssuchassearchtimeornetwork time in a fully connected network with infinite divisibility distance to a copy of requested contents. of requests and convex server efficiency [16], as well as Indistributedservernetworks,amaximummatchingpolicy the average distance to the closest cache holding a copy of (with re-packing of the requests at any time but no splitting the requested content in a random network with exponential or queueing) has been considered in [3], [4]. In this setting, expansion [17]; and recall that [3] also advocated using this in [3] it was shown that replicating contents proportionallyto replication policy. Therefore, due to its ubiquity in the litera- their popularity leads to full efficiency in large system and ture, proportional replication is the natural and most relevant largestoragelimits, andin [4] itwas furtherprovedthatthere schemeto whichwe shouldcompareanyproposedpolicy;we isactuallyawiderangeofpolicieshavingsuchanasymptotic will obtain substantial improvementsover it in our model. property and an exact characterisation of the speed at which efficiencydecreasesforagivenpolicywasobtained.However, III. EDGE-ASSISTEDCDN MODEL thatformulafortheefficiencyistoocomplicatedtoderiveany Inthissection,wedescribeindetailouredge-assistedCDN practical scheme from it, and anyway maximum matching at model. any time is probably unrealistic for most practical systems. As already mentioned, the basic components of an edge- Finding efficient content replications in cache networks is assistedCDNareadatacenterandalargenumbermofsmall a very active area of research. The optimal replication is not servers. The CDN offers access to a catalog of n contents, yet understood, however adaptive strategies are studied; they which all have the same size for simplicity (otherwise, they createreplicasforcontentswhenincomingrequestsarriveand could be fragmented into constant-size segments). The data evict contents based on either random, least frequently used center stores the entire catalog of contents and can serve all (LFU) or least recently used (LRU) rules, or a variation of the requests directed towards it, whereas each small server those. There is an extensive literature on the subject and we can only store a fixed number d of contents and can provide only mention the most relevant papers to the present work. serviceforatmostonerequestatatime.We canrepresentthe An approximationfor LRU cache performanceis proposed in informationofwhichserverstoreswhichcontentinabipartite [5] and further supported in [6]; it allows [7] to study cache graph G = (S ∪C,E), where S is the set of servers, C the networkswithatwo-levelshierarchyofcachesunderamixof set of contents, and there is an edge in E between a content 3 c and a server s if s stores a copy of c; an edge therefore seems reasonable to pay a particular attention to the asymp- indicatesthat a server is able to serve requestsfor the content totics of the performance of the system as n,m → ∞. To with which it is linked. Given numbers of replicas (D ) keepthingssimple,we can assumethatthe numberofservers c c∈C for the contents, we assume that the graph G is a uniform m andthe numberof contentsn grow to infinity together,i.e. randombipartitegraphwithfixeddegreesdontheserversside n = βm for some fixed β. In addition, as storage is cheap andfixeddegreesequence(D ) onthecontentsside. This nowadayscomparedto access bandwidth, it also makes sense c c∈C models the lack of coordination between servers and would to focus on a regime with d large (but still small compared result from them doing independent caching choices. We do to n). In these regimes, the inefficiency will tend to 0 as the notallowrequeststobequeued,delayedorre-allocatedduring system size increases under most reasonable replications (as service,andasaresultaserverwhichbeginsservingarequest was shown in [4] under the maximum matching assumption). will then remain unavailable until it completes its service. However,asthecostofoperationisonlytiedtolosses,further Splitting of requestsis not considered either, in the sense that optimizing an already small inefficency is still important and only one server is providing service for a particular request. can lead to order improvements on the overall system cost. With these constraints, at any time, the subset of edges of G overwhichsomeserviceisbeingprovidedforma generalized IV. ANAPPROXIMATION FORTHELOSS RATEOF A matchingofthegraphG(theanalysisin[4],[18]reliesheavily CONTENT on this). In this section, we propose an approximationto understand Thecontentsofthecatalogmayhavedifferentpopularities, in a precise mannerthe relation between any fixed replication leading to different requests arrival rates λ . We let λ be ofcontentsandthelossratesinthesystem.Thisanalyticalstep c the average request rate, i.e. λ = 1 λ . According to hasmanyadvantages:itallowsustoformallydemonstratethat n c c the independent reference model (IRM), the requests for the to optimize the system one needsto make the loss rates equal P variouscontentsarriveas independentPoisson processes with forallthecontents;asaconsequenceweobtainanexplicitex- ratesλ . Theservice times ofall the requestsare independent pressionfortheoptimalreplication(SubsectionIV-B);finally, c Exponentialrandomvariableswithmean1,whichisconsistent in Subsection V-B, we will leverageour analytical expression with all the servers having the same service capacity (upload for the full distribution of the number of available replicas bandwidth) and all the contents having the same size. We let of a content to propose a mechanism enhancing the speed of ρ denote the expected load of the system, i.e. ρ = nλ. The adaptivealgorithms.We validateour approximationand show m distribution of popularities of the contents is left arbitrary, that our optimized replication strategy largely outperforms although in practice Zipf distributions are often encountered proportionalreplication through extensive simulations. (see [7] for a page-long survey of many studies) and it A. One-Dimensional Markov Chain Approximation for the seemsthereforemandatorytoevaluatethereplicationschemes Number of Available Replicas proposed under a Zipf popularity distribution. When a request arrives, we first look for an idle server For a given fixed constitution of the caches (i.e. a fixed storing the corresponding content; if none is found then the graph G), the system as described in the previous section is requestis sentto the data centerto be served ata highercost. Markovian,the minimumstate space indicatingwhichservers As in the end all the requests are served without delay, be are busy (it is not necessarily to remember which content it by the data center or by a small server, the performance they serve). We want to understand how the loss rate γ for c of the system is mainly described by the cost associated a particular content c relates to the graph G, but using too with the service. This cost is mostly tied to the number of detailed information about G, such as the exact constitution requestsserved by the data center, thereforethe most relevant of the caches containing c, would have the drawback that performancemetrichereisthefractionoftheloadleft-overto it would not lead to a simple analysis when considering the data center. Then,it makessense to view the requeststhat adaptive replication policies (as the graph G would then keep the small servers cannot handle as “lost”. In fact, the system changing). Therefore, we need to obtain a simple enough but consisting of the small servers alone with fixed caches is a still accurate approximate model tying γ to G. c lossnetworkinthesense of[19].Thisimpliesthatthesystem Theexpectedlossrateγ ofcontentcisequaltoitsrequests c corresponds to a reversible stochastic process, with a product arrival rate λ multiplied by the steady state probability that c form stationary distribution π. However, the exact expression c has no replicas available. Let D be the total number of c of that stationary distribution is too complex to be exploited replicas of content c and Z be its number of available c in our case (as opposed to [3], [4], [18] where the maximum replicas, i.e. those stored on a currently idle server. We thus matching assumption yields a much simpler expression). We want to compute π(Z = 0) to get access to γ . However, c c call γ the rate at which requests for content c are lost, and γ the Markovchain describingthe system is too complicatedto c the average loss rate among contents, i.e. γ = 1 γ . The be able to say much on its steady state. In order to simplify n c c main goal is to make γ as small as possible. We refer to the the system, one can remark that in a large such system the P fractionoflost requestsγ/λasthe inefficiencyof the system. state of any fixed number of servers (i.e. their current caches Finally, as in many real cases CDNs are quite large, with andwhethertheyarebusyoridle)areonlyweaklydependent, often thousandsof serversand similar numbersof contents, it andsimilarly the numberof availablereplicasZ of anyfixed c 4 numberofcontentsareonlyweaklydependent.Therefore,itis This yields the following steady-state probability: naturalto approximatethe system by decouplingthe different D +λ /θ contents and the different servers (similar phenomenon are π(Zc =z)=π(Zc =Dc)θeDffc−z cD −cz eff , (1) explained rigorously in [20]). In other words, this amounts to (cid:18) c (cid:19) where the binomial coefficient does not really have a combi- forgettingthe exact constitution of the caches; as a result, the natorialinterpretationasoneofitsargumentsisnotaninteger, correlationbetweencontentswhich are storedtogetherislost. butshouldonlybeunderstoodas k+x = 1 k (i+x), Then,theevolutionofZc becomesaonedimensionalMarkov for k,l∈N and x∈R . l l! i=k−l+1 chain,independentofthevaluesofZc′ forothercontents,and + (cid:0) (cid:1) Q We now have an approximation for the full distribution of we can try to compute its steady-state distribution. thenumberZ ofavailablereplicasofcontentc, whichyields Foranyz <D ,therateoftransitionfromZ =ztoz+1is c c c an approximationforthe lossrate γ =λ π(Z =0). We can alwaysD −z:itistherateatwhichoneoftheD −zoccupied c c c c c also compute the mode z of this distribution: z ≈ Dc−λc, serversstoringccompletesitscurrentservice;wedonotneed c c 1+θeff which can be used as an approximation for the expectation todistinguishwhetheraserverisactuallyservingarequestfor E[Z ] (in fact, simulations show the two are nearly indistin- c or foranothercontentc′ as we assume the service times are c b b guishable). We further simplify the expression obtained by independent and identically distributed across contents. For providing a good approximation for the normalization factor any z > 0, the transitions from Z = z to z −1 are more c π(Z =D ) in Equation (1): complicated. They happen in the two following cases: c c • either a request arrives for content c and as Zc = z > 0 Dc D +λ /θ −1 it is assigned an idle servers storing c; π(Zc =Dc)= θexff c xc eff . • orarequestarrivesforanothercontentc′ anditisserved Xx=0 (cid:18) (cid:19)! by a server which also stores content c. Tothatend,weusethefactthatweaimatsmalllossrates,and Thefirsteventoccursatrateλ andthesecondoneatexpected thus most of the time at small probabilities of unavailability. c rate Therefore, the bulk of the mass of the distribution of Zc |{s∈S : s is idle and c,c′ ∈s}| should be away from 0, which means that the terms for x λc′E |{s∈S : s is idle and c′ ∈s}| , close to Dc should be fairly small in the previousexpression. c′6=c (cid:20) (cid:21) We thus extend artificially the range of the summation in this X wherec∈sindicatesthatthecontentc isstoredontheserver expression and approximate π(Z =D ) as follows: c c s. At any time and for anyc′, |{s∈S : s is idle and c′ ∈s}| is equal to Zc′. The term |{s ∈ S : s is idle and c,c′ ∈ s}| π(Z =D )−1 ≈ Dc+⌊λc/θeff⌋θx Dc+λc/θeff is equal in expectation to the number of idle servers storing c c eff x x=0 (cid:18) (cid:19) c′ (i.e. Zc′) times the probabilitythat such a server also stores ≈ (1+Xθeff)Dc+λc/θeff. c. As we forgetaboutthe correlationsin the caching,this last probability is approximated as the probability to pick one of We obtain the following approximation for π(Zc =0): tch′ewdhe−n1wreemdiasipnaintcghmatemraonrdyomslotthseinZacnavidalielabselervreerpslisctoasrinogf π(Zc =0)≈ 1+θefθf Dc(1+θeff)−θλecff Dc+Dλc/θeff . contentcbetweenalltheremainingslotsinidleservers.Thus, (cid:18) eff(cid:19) (cid:18) c (cid:19) Finally, as we are interested in large systems and large Z (d−1) E[|{s∈S : s is idle and c,c′ ∈s}|]≈ c′ Z , storage/replication,wecanleveragethefollowingasymptotic c (1−ρ )md eff behavior: where ρeff = ρ(1 − γ/λ) is the average load effectively k+x ≈ ex(Hk−CEuler), absorbed by the system, so that the total number of memory k Γ(x+1) (cid:18) (cid:19) slotsontheidleserversis(1−ρeff)md.We alsoneglectedthe for k ∈ N large enough and where Γ is the gamma function: Zc′ memory slots occupied by c′ in these idle servers when Γ(1+x) = 0+∞txe−tdt; Hk is the k-th harmonic number: computing the total number of idle slots (1 − ρeff)md. We Hk = ki=1R1i; and CEuler is the Euler-Mascheroni constant: obtain the following approximation for the expected rate at C =lim (H −lnk)≈0.57721.For large number of Euler k→∞ k which the second event occurs: replicasPD , we thus obtain the approximation: c d−1 ρ d−1 Zcc′6=cλc′(1−ρeff)md ≈Zc1−efρfeff d =Zcθeff, π(Zc =0)≈C(λc) 1+ θ1 −DcDcθλecff, X (cid:18) eff(cid:19) cwohnetreentwcecnoemglpeacrteeddttohethreataegλgrcegaattwehaircrihvarleqrauteestosfarreriqvueesftosr, whereweletC(λc)= e−CEλulcerλc/θeff andthetermDcθλecff is awnidthwreealseotnθaebffly=hi1g−ρheρffeefffdfe−dc1t.ivNeoloteadthρatefft,hecoirnrteesrpeostnidnsgtroegliamrgee, an approximationof eθλ(e1cff+HθDeffc).θeTffhΓe(1+apθλpecffr)oximationfor π(Zc = valuesofθeff,asρeff →1impliesθeff →∞.TheMarkovchain 0) immediately yields an approximation for the loss rate γc: obtained satisfies the local balance equations: for z <Dc, 1 −Dc λc γ =λ π(Z =0)≈λ C(λ ) 1+ Dθeff. (2) (Dc−z)π(Zc =z)=(λc+(z+1)θeff)π(Zc =z+1), c c c c c (cid:18) θeff(cid:19) c 5 Properties class1 class2 class3 Note that the expression of θ involves the average effective eff numberofcontents ni 200 400 400 load ρeff, which itself depends on the average loss rate γ. We popularity λi 9 3 1 thushaveafixedpointequationinγ (justasin[5],[6]),which numberofreplicas Di 200 67 23 wecaneasilysolvenumerically.Indeed,theoutputvalueγ of simulationdata/approximation class1 class2 class3 ourapproximationisadecreasingfunctionoftheinputvalueγ nb.ofavailable replicas E[Zi] 21.7/21.6 7.28/7.25 2.51/2.50 103×lossrates γi 0/10−5 3.31/2.36 79.4/76.3 used in ρ , which implies simple iterative methods converge eff simulationdata/approximation Zipf,α=0.8 Zipf,α=1.2 exponentially fast to a fixed-point value for γ. proportional: 103×inefficiency 5.69/5.46 11.8/11.6 B. Approximation-Based Optimized Replication TABLE I: Properties of the content classes, and numerical results on the accuracy of the approximation. In this subsection, we exploit the approximation obtained in Equation (2) to understand which replication strategy min- imizes the total loss rate in the system. In other words, we approximately solve the optimization problem: C. Accuracy of the Approximation and Performance of the Optimized Replication D ≤md min γ s.t. c c (3) PDc ∈N, ∀c Intheprocessofapproximatingthelossrateγc ofacontent c, we performed many approximations based on asymptotic Note that the approximation from Equation (2) is consistent behaviors for large systems and large storage / replication. with the intuition that γ is a decreasing, convex function of Dc. Indeed, letting x =c 1+θeθffeffeθeff(Dλcc+1) ≤ 1 since we have IEtquisatitohner(e2fo)riesnnoetcteososafrayrotoff,cwhehcickhwwheedthoerintthheisfsoercmtiuolnavoiaf Dc+1≥ λc in the regime of interest, we compute simulation. The systems we simulate are of quite reasonable θeffln(cid:16)1+θ1eff(cid:17) size (with a few hundred contents and servers). However, γ (D +1)−γ (D )=∆γ (D )=−γ (D )(1−x)≤0, the accuracy of the approximation should only improve as c c c c c c c c ∆γ (D −1)=−γ (D ) 1 −1 ≤0. the system grows. We use two scenarios for the popularity c c c c x distribution of the contents: a class model (similar to that The loss rate γ is a convex function of D as(cid:0)shown(cid:1)by c c of [4]), which allows us to compare between many contents 1 with similar characteristics, and a Zipf popularity model, ∆γ (D )−∆γ (D −1)≥γ (D ) x+ −2 ≥0. c c c c c c which is deemed more realistic. We evaluate the accuracy x (cid:18) (cid:19) of the approximation using proportional replication (other As a consequence, the optimization problem (3) is approxi- replications were also simulated and yield similar results). To mately convex, and we thus obtain an approximate solution compute the approximate expressions, we solve numerically by making the first derivatives of the loss rates ∆γ equal. c the fixed point equation in γ. We simulate systems under a Keeping only the dominant orders in D , we have c reasonably high load of 0.9. As often, it is mainly interesting γ λ to know how the system behaves when the load is high, as ∆γc(Dc)=− c 1− c . (4) otherwiseitsperformanceisalmostalwaysgood.However,as 1+θ D eff (cid:18) c(cid:19) requestsarriveandareservedstochastically,iftheaverageload In the first order, equalizing the first derivatives in (4) using weretoocloseto1thentheinstantaneousloadwouldexceed1 theexpressionin(2)leadstoequalizingthenumberofreplicas quite often, which would automatically induce massive losses foreverycontent,i.e.settingDc =D+o(D))whereD = mnd and mask the influence of the replication. In fact, it is easy is the average number of replicas of contents. Going after the to see that we need to work with systems with a number of second order term, we get servers m large compared to ρ in order to mitigate this (1−ρ)2 effect. lnD D =D+(λ −λ) +o(lnD). (5) The setup with the class model is the following: there c c θeffln 1+ θ1eff are n = 1000 contents divided into K = 3 classes; the (cid:16) (cid:17) characteristics of the classes are given in the first part of Wethereforeobtainthattheasymptoticallyoptimalreplication Table I. The popularities in Table I together with n = 1000 is uniform with an adjustment due to popularity of the order andρ=0.9resultinm=3800servers.Eachservercanstore of the logof the averagenumberofreplicas. Thisagreeswith d=20 contents, which is 5% of the catalog of contents. We the results in [4] and goesbeyond.Finally,inserting backthis let the system run for 104 units of time, i.e. contents of class expression for D into Equation (2) yields c 3 should have received close to 104 requests each. −D Figure 1 and the second part of Table I show the results 1 λ (1+o(1)) γc = 1+ Dθeff , (6) for the class model: the left part of Figure 1 shows the θ (cid:18) eff(cid:19) distributions of the numbers of available replicas for all the which shows that the average loss rate γ under optimized contentsagainsttheapproximationfromEquation(1)foreach replication behaves as the loss rate of an imaginary average class; the right part shows the loss rates for all the contents content (one with popularity λ and replication D). against the approximation from Equation (2) for each class. 6 stationary probability000000...012...123555 sacccilllpmaaapsssursssol a123xtiimona tdioanta stationary loss rate00000....0000.12468 sacccilllpmaaapsssursssol a123xtiimona tdioanta of nb. of available replicas123456 naZZbpiipp.pff ar01ov..xa82imilaabtlieo nreplicas stationary loss rates00000...012..12555 laZZopiispppsff r 01ora..x82tiemsation g 00 nu5mb1e0r o1f 5ava20ilab2l5e r3e0plic3a5s 00 200inde4x0 0of c6o0n0ten8t00 1000 lo00 1 2 3 4 5 6 00 50 100 150 200 log of the rank of the content popularity rank of the content Fig. 1: Class model: distribution of number of available Fig. 2:Zipf model:expectednumberof available replicasand replicas and loss rates Vs approximation. loss rates Vs approximation. x 10−3 Although it is not apparent at first sight, the left part of 8 Figure1actuallydisplaysaplotforeachofthe1000contents, s but the graphs for contents of the same class overlap almost ate6 perfectly, which supports our approximation hypothesis that s r s o thestationarydistributionofthenumberofavailablereplicasis y l4 ar notverydependenton the specific serverson which a content n o iscachedorontheothercontentswithwhichitiscached.This stati2 Zipf 0.8 behavior is also apparent on the right part of Figure 1, as the Zipf 1.2 loss rates for the contents of the same class are quite similar. 0 0 50 100 150 200 From Figure 1 and Table I, it appearsthat the approximations popularity rank of the content fromEquations(1)and(2)arequiteaccurate,withforexample Fig. 3: Zipf model: loss rates under optimized replication. a relative errorof around5%in the inefficiencyofthe system (9.68×10−3 Vs 9.20×10−3). We consider such an accuracy is quite good given that some of the approximationsdone are Equation (1); the right part shows the loss rates for all the based on a large storage / replication asymptotic, while the contents against the approximationfrom Equation (2). Again, simulationsetupis with a storagecapacityofd=20contents theresultsfrombothFigure2andTableIconfirmtheaccuracy onlyandcontentsofclass3(responsibleforthebulkoflosses) of the approximations. have only 23 replicas each. Wenowturntoevaluatingtheperformanceoftheoptimized We now turn to the Zipf popularity model. In this model, replication from Equation (5). Figure 3 shows the resulting thecontentsarerankedfromthe mostpopulartotheleast;for loss rates under the optimized replication, with both Zipf a given exponent parameter α, the popularity of the content exponents. This figure is to be compared with the right part of rank i is given by λi = i−αj−αλ. We use two different of Figure 2, which showed the same results for proportional valuesfortheZipfexponentPα,j≤0n.8and1.2,asproposedin[7]. replication.Itisclearthattheoptimizedreplicationsucceedsat Theexponent0.8ismeanttorepresentcorrectlythepopularity reducingtheoverallinefficiencyγ/λcomparedtoproportional distribution for web, file sharing and user generated contents, replication (from 5.7×10−3 to 5.1×10−4 for α=0.8, and whiletheexponent1.2ismorefitforvideo-on-demand,which from 1.2×10−2 to 1.6×10−6 for α=1.2). Note that in the has a more accentuated popularity distribution. We simulate case of Zipf exponent α = 1.2 the popularity distribution is networks of n = 200 contents and m ≈ 2000 servers of more“cacheable”asitismoreaccentuated,andtheoptimized storage capacity d = 10 under a load ρ = 0.9. This yields replication achieves extremely small loss rates. However, the an average content popularity λ = ρm ≈ 9. Note that under loss rates for all the contents are not trully equalized, as n proportional replication the numbers of replicas of the most popularcontentsarestilltoomuchfavored(ascanbeseenfor popularcontentsareactuallylargerthanthenumberofservers Zipf exponent α=0.8). This may be because the expression mforα=1.2;wethusenforcethatnocontentisreplicatedin for optimized replication is asymptotic in the system size and more than 95% of the servers. As expected and confirmed by storagecapacity.Hence,thereisstillroomforadditionalfinite- the simulations, this is anyway benefical to the system. Each size corrections to the optimized replication. setup is simulated for at least 104 units of time. As an outcome of this section, we conclude that the ap- We show the results for the Zipf modelwith both exponent proximationsproposedareaccurate,evenatreasonablesystem values in Figure 2 and in the third part of Table I: the left size. In addition, the optimized scheme largely outperforms part of Figure 2 shows the expected number of available proportionalreplication.Inthenextsection,wederiveadaptive replicas for all the contents against the approximation from schemes equalizing the loss rates of all the contents and 7 achieving similar performances as the optimized replication. equalizethe loss rates. Thiswouldbe achievedfor exampleif we picked a content for eviction uniformly at random among V. DESIGNING ADAPTIVEREPLICATION SCHEMES all contents. However, the contents eligible for eviction are Inapracticalsystem,wewouldwanttoadaptthereplication only those which are available (although some systems may in an online fashion as much as possible, as it provides allowmodifyingthecachesofbusyservers,asservingrequests more reactivity to variations in the popularity of contents. consumes upload bandwidth while updating caches requires Such variationsare naturally expected for reasons such as the only download bandwidth). Therefore, the most natural and introductionofnewcontentsinthecatalogoralossofinterest practicalevictionruleistopickanavailablecontentuniformly for old contents. Thus, blindly enforcing the replication of at random (hereafter, we refer to this policy simply as the Equation (5) is not always desirable in practice. Instead, we RANDOM policy). Then, the eviction rate for content c is can extract general principles from the analysis performed in given by η ∝ π(Z > 0) = 1− γc. So, at equilibrium, we Section IV to guide the design of adaptive algorithms. can expectcto have c γc = Cste, ∀λcc ∈ C. In a large system, 1−γc Inthis section,we firstshow howbasic adaptiverulesfrom with large storage / repλlcication and loss rates tending to zero, cachenetworkscanbetranslatedintoourcontextbasedonthe thedifferencewithareplicationtrullyequalizingthelossrates insightsfromSectionIVtoyieldadaptivealgorithmsminimiz- is negligeable. If we are confronted to a system with a large ingtheoveralllossrate.Then,weshowhowthemoredetailed numberofveryunpopularcontentsthough,wecancompensate information contained in Equation (1) allows the design of for this effect at the cost of maintaining additional counters fasterschemesattainingthesametargetreplication.Finally,we for the number of evictions of each content. validate the path followed in this paper by evaluating through Once the rule for creatingreplicasif fixed,we immediately simulations the adaptive schemes proposed. obtain a family of adaptive algorithms by modifying the evictionrulesfromthecachenetworkcontextaswedidabove A. Simple Adaptive Algorithms for Minimizing Losses for the RANDOM policy. Instead of focusing on “usage” and The analysis in Section IV-B shows that the problem of the incoming requests process as in cache networks with the minimizing the average loss rate γ is approximatelya convex LFU and LRU policies, we react here to the loss process. problem, and that therefore one should aim at equalizing This yields the LFL (least frequently lost) and LRL (least the derivatives ∆γc(Dc) of the stationary loss rates of all recently lost) policies. RANDOM, LRL, and LFL are only the contents. In addition, Equation (4) points out that these three variantsof a generic adaptive algorithmwhich performs derivatives are proportional to −γc in the limit of large the following actions at every loss for content c: replication / storage, and thus equalizingthe loss rates should 1) create an empty slot on an available server, using the provideanapproximatesolutionfortheoptimizationproblem. eviction rule (RANDOM / LRL / LFL); An immediate remark at this point is that it is unnecessary to 2) add a copy of the content c into the empty slot. store contents with very low popularity λ if the loss rate of c the other contents is already larger than λ . Thethreeevictionrulesconsideredhererequireaveryelemen- c An adaptive replication mechanism is characterized by two tary centralized coordinating entity. For the RANDOM rule, rules: a rule for creating new replicas and another one for this coordinator simply checks which contents are available, evicting contentsto make space for the new replicas. In order picks one uniformly at random, say c′, and then chooses a to figure out how to set these rules, we analyze the system random idle server s storing c′. The server s then picks a in the fluid limit regime, with a separation of timescales such random memory slot and clears it, to store instead a copy of that the dynamics of the system with fixed replication have c. In the case of LRL, the coordinator needs in addition to enough time to converge to their steady-state between every maintain an ordered list of the least recently lost content (we modification of the replication. Achieving this separation of call such a list an LRL list). Whenever a loss occurs for c, timescales in practice would require slowing down enough the coordinator picks the least recently lost available content the adaptation mechanism, which reduces the capacity of the c′ based on the LRL list (possibly restricting to contents less system to react to changes. Therefore, we keep in mind such recently lost than c) and then updates the position of c in a separation of timescales as a justification for our theoretical the LRL list. It then picks a random idle server s storing c′, analysis but we do not slow down our adaptive algorithms. which proceeds as for RANDOM. Finally, for LFL, the coor- When trying to equalize the loss rates, it is natural to use dinator would need to maintain estimates of the loss rates of the loss events to trigger the creation of new replicas. Then, eachcontent(bywhatevermeans,e.g.exponentiallyweighted new replicas for content c are created at rate γ , and we movingaverages);when a losshappens,the coordinatorpicks c let η be the rate at which replicas of c are deleted from the available content c′ with the smallest loss rate estimate c the system. In the fluid limit regime under the separation of and then proceeds as the other two rules. This last rule is timescale assumption, the number of replicas of c evolves more complicated as it involves a timescale adjustement for according to D˙ = γ − η , where all the quantities refer theestimation ofthe lossrates, thereforewe willfocusonthe c c c to expectations in the steady-state of the system with fixed first two options in this paper. Note that the LFL policy does replication. At equilibrium, we have D˙ = 0 and γ = η not suffer from the drawback that the eviction rate is biased c c c for all the contents c, thus we need η = Cste for all c to towardspopularcontentsduetotheirsmallunavailability,and c 8 this effect is also attenuated under the LRL policy. We point get fast adaptation). outthatitispossibletooperatethesysteminafullydistributed As a first step towards setting the probability p (Z ), we c c way (finding a random idle server first and then picking a write Equation (1) as follows: content on it by whatever rule), but this approach is biased z λ +iθ into selecting for eviction contents with a large number of π(Z =0)=π(Z =z) c eff . c c D −i+1 available replicas (i.e. popular contents), which will lead to a c i=1 Y replication with reduced efficency. It is of interest for future This shows that, for any fixed value z with π(Z = z) ≥ c work to find a way to unbias such a distributed mechanism. π(Z =0), we can generate events at rate γ by subsampling c c B. Adapting Faster than Losses at the time of a request arrival with Zc = z with a first probability In the algorithms proposed in the previous subsection, we z λ +iθ usethelossesofthesystemastheonlytriggerforcreatingnew q (z)= c eff . c D −i+1 replicas.Itisconvenientastheseeventsaredirectlytiedtothe i=1 c Y quantity we wish to control (the loss rates), however it also Ifonthecontraryz issuchthatπ(Z =z)<π(Z =0),then c c has drawbacks. Firstly, it implies that we want to generate a the value of q (z) given above is larger than 1 as we cannot c newreplicaforacontentpreciselywhenwehavenoavailable generate events at rate γ by subsampling even more unlikely c serverforuploadingit,andthuseitherwesendtworequestsat events. If we generated virtual losses at rate γ as above for c atimetothedatacenterinsteadofone(onefortheuserwhich eachvalueofz,thenthetotalrateofvirtuallossesforcontentc wecouldnotserveandonetogenerateanewcopy)orwemust wouldbeλ Dc min{π(Z =z),π(Z =0)}, whichclearly delaythecreationofthenewreplicaandtolerateaninedaquate c z=1 c c still depends on c. We thus proceed in two additional steps replication in-between, which also hinders the reactivity of R towardssettingp (Z ):wefirstrestricttherangeofadmissible c c the system. Secondly, unless in practice we intend to slow values of Z , for which we may generate virtual losses, by c down the adaptation enough,there will always be oscillations excluding the values z such that π(Z =z)<π(Z =0). In c c inthereplication.IflosseshappenalmostasaPoissonprocess, theregimeθ ≥1,thiscanbedoneinacoarsewaybyletting eff then only a bit of slow-down is necessary to moderate the z∗ = Dc−λc and rejecting all the values z >z∗. Indeed, oscillations, but if they happen in batches (as it may very c θeff c well be for the most popular contents) then we will create z∗ c D −iθ many replicas in a row for the same contents. If in addition qc(zc∗)= Dc −ieff ≤1, we use the LRL or LFL rules for evicting replicas, then the i=0 c Y samecontentswillsuffermanyevictionssuccessively,fuelling and the distribution of Z is unimodal with the mode at c againtheoscillationsinthereplication.Finally,itisverylikely a smaller value Dc−λc ≈ ρz∗. Now, subsampling with thatpopularitiesareconstantlychanging.Ifthesystemisslow probability q (Z ) wθheffen a requecst arrives for content c and c c to react, then the replication may never be in phase with the Z ≤ z∗ would generate events at a total rate z∗γ . Thus, c c c c current popularities but always lagging behind. For all these it suffices to subsample again with probability minc′∈Czc∗′ to z∗ reasons,itisimportanttofindawaytodecoupletheadaptation obtain the same rate of virtual losses for all the ccontents mechanismsfromthelossesinthesystemtosomeextent,and (another approach would be to restrict again the range of at least to find other ways to trigger adaptation. admissible values of z, e.g. to values around the mode z ). c Weproposeasolution,relyingontheanalysisinSectionIV. To sum up, our approach is to generate a virtual loss for Alossforcontentcoccurswhen,uponarrivalofarequestfor content c at each arrival of a request for c with probability b c, its number of available replicas is equal to 0. In the same min z∗ way, whenevera requestarrives, we can use the currentvalue p (Z )= c′∈C c′1(Z ≤z∗)q (Z ). c c z∗ c c c c of Zc to estimate the loss rate of c. Indeed, Equation (1) tells c us how to relate the probability of Zc =z to the loss rate γc, The rate γc at whichvirtuallosses are generatedfor contentc for any z ∈N. Of course, successive samples of Zc are very isthengivenbyγc×minc′∈Czc∗′,whichisindependentofcas correlated(notethatlossesmayalsobecorrelatedthough)and planned.Whenevera virtuallossoccurs,we canuse whatever e wemustbecarefulnottobetooconfidentintheestimatethey algorithmwe wanted to use in the first place with real losses; provide. A simple way to use those estimates to improve the there is no need to distinguish between the real losses and adaptation scheme is to generate virtual loss events, to which the virtual ones. For example, if we use the LRL policy, we any standard adaptive scheme such as those introduced in the update the position of c in the LRL list and create a new previous subsection may then react. To that end, whenever replica for c by evicting a least recently lost available content a request for c arrives, we generate a virtual loss with a (from a server which does not already store c). If we choose certain probability pc(Zc) depending on the current number to test for virtual losses at ends of service for c (which yields of available replicas Z . The objective is to define p (Z ) so the same outcomein distribution, as the system is reversible), c c c that the rates (γ ) of generation of virtual losses satisfy the new replica can simply be uploaded by the server which c c∈C γ = Cste×γ for all c ∈ C (so that the target replication just completed a service for c. Furthermore, in practice, we c c still equalizes leoss rates) and Cste is as high as possible (to advocate estimating the values of zc∗ and pc(Zc) on the fly e g g 9 rather than learning these values for each content: θ can eff 0.015 8 be computed from ρ , which is naturally approximated by fixed optimized fixed proportional eff RANDOM fixed optimized aatncchvnauuenermrreeracbsangotetitemirmonoopauvofutmeeftserbetfethrzohrvec∗reeorcmaffseun;rwirdenrsqeciplum′ne∈ceat(iCsslZntatszrucplc∗)fmyo′o;,,pbriwutewclrieahbsroieccofciahnnonbglncyutasasepennynrpetvsrcsbeoeeeudxrspv.sifmdaeForrarraystoteemettdoxoλawtmhctmhheaepbesinelynet,eotvtawtahieanneerl stationary loss rates00.0.0015 LRLRRALLN DviOrtuMa lvirtual log of nb. of replicas4567 RLRLRRAALLNN DDviOOrtuMMa lvirtual a service ends for one of them. 0 3 0 50 100 150 200 0 1 2 3 4 5 6 In the next subsection, we evaluate the performance of the popularity rank of the content log of the rank of the content adaptative schemes proposed and the virtual losses mecha- nism. Fig. 4: Adaptive schemes: loss rates and replication. C. Evaluation of the Performance through Simulations 400 Before getting to the simulation results, let us just mention 10% most popular contents 350 10% least popular contents that the complexity of simulating the adaptive algorithms RANDOM Ignrdoweesd,veitryisfaesatsywittoh stehee tshyasttesmimsuilzaetin(gwitthhens,ysmtemanfdord)a. eplicas235000 LRLRRALLN DviOrtuMa lvirtual fixed duration t requires Ω(mdλt) operations. Furthermore, er of r200 the time needed for the RANDOM algorithm to converge, mb150 u whenstartedatproportionalreplication,isroughlyoftheorder n100 of max D /γ, where γ is the average loss rate for the 50 c∈C c limit replication, which decreases exponentially fast in d as 0 0 1 2 3 4 5 6 7 8 9 10 seenin Equation(6).Therefore,ifwe wanttocompareallthe time x 104 adaptive schemes, we cannot simulate very large networks. Fig. 5: Adaptive schemes: time evolution. Anyway,our intentionis to show thatour schemeswork even for networks of reasonable size. As in Section IV-C, we used Zipf popularity distributions and secondly that this replication equalizes the loss rates, as with exponents 0.8 and 1.2 and a class model to evaluate the intended.Inaddition,theadaptiveschemesperformevenbetter performance of the various schemes. The results are qualita- thanthe optimizedstatic replication,which suffersfromfinite tively identical under all these models, so we only show the network / finite storage effects, as they manage to find the resultsforZipfpopularitywithexponentα=0.8.Wecompare right balance between popular and unpopular contents. thevariousschemesintermsofthereplicationachievedandits We compare the adaptation speed of the various schemes associated loss rates, as well as the speed at which the target on Figure 5, where we plot both the evolution of the average replication is reached. We do not slow down the dynamics number of replicas of the 10% most popular contents and of the adaptive schemes even though this necessarily induces that of the 10% least popular ones, starting from proportional someoscillationsinthereplicationobtained.Nonetheless,this replication. As expected, the LRL schemes are faster than setup is already sufficient to demonstrate the performance of the RANDOM ones, but more importantly this plot clearly the adaptive schemes. It would be interesting to quantify the demonstrates the speed enhancement offered by the virtual potential improvement if one reduces the oscillations of the loss method of Section V-B. Regarding the benefits of such replication obtained (e.g. by quantifying the variance of the anenhancedreactioncapability,thereisaninterestingproperty stationary distribution for the number of replicas for each whichwe didnotpointoutnorillustratewith thesimulations: content); we leave this out for future work. Also, we did not thevirtuallossschemehasthepotentialtofollowaconstantly accountfortheloadputonthedatacentertocreatenewcopies evolvingpopularityprofileatnocost,astherequiredcreations of the contents; one can simply double the loss rates for the of replicas to adapt to the changing popularities can be done adaptiveschemestocapturethiseffect.Notethatifadaptation without requesting copies of the contents to the data center. speedisnotapriority,onecantradeitofftoalmostcancelthis factorof2.Finally,concerningthevirtuallossmechanism,we VI. CONCLUSION estimateallthequantitiesinvolvedonthefly,asrecommended We addressed the problem of content replication in edge- in the previous section. assisted CDNs, with a special attention to capturing the In Figure 4, we show results for the various adaptive mostimportantconstraintsonservercapabilitiesandmatching schemes.Ontheleftpartofthefigure,weshowthestationary policy. Based on large system and large storage asymptotics, loss rates of all the contents; on the right part we show we derived an accurate approximationfor the performance of in log-log scale the stationary expectation of the numbers anygivenreplication,therebyallowingofflineoptimizationof of replicas for each content. This plot shows firstly that the replication. In addition, levaraging the insights gathered all the adaptive schemes converge to the same replication in the analysis, we designed adaptive schemes converging to 10 the optimal replication. Our basic adaptive algorithms react to losses, but we proposedfurther mechanisms to move away from losses and adapt even faster than they occur. REFERENCES [1] Cisco White Paper, “Cisco visual networking index: Forecast and methodology, 2012-2017,” May 2013. [Online]. Available: http://www.cisco.com/en/US/solutions/collateral/ns341/ns525/ns537/ns705/ns827/white paper c11-481360.pdf [2] J.Dilley, B.Maggs,J.Parikh,H.Prokop,R.Sitaraman, andB.Weihl, “Globallydistributedcontentdelivery,”IEEEInternetComputing,vol.6, no.5,pp.50–58,2002. [3] B. Tanand L.Massoulie´, “Optimal content placement forpeer-to-peer video-on-demandsystems,”inProceedingsIEEEINFOCOM,2011,pp. 694–702. [4] M. Leconte, M.Lelarge, and L.Massoulie´, “Bipartite graph structures for efficient balancing of heterogeneous loads,” ACM SIGMETRICS PerformanceEvaluation Review,vol.40,no.1,pp.41–52,2012. [5] H. Che, Y. Tung, and Z. Wang, “Hierarchical web caching systems: Modeling, design and experimental results,” IEEEJournal onSelected AreasinCommunications, vol.20,no.7,pp.1305–1314, 2002. [6] C. Fricker, P. Robert, and J. Roberts, “A versatile and accurate ap- proximation for lru cache performance,” in Proceedings of the 24th International Teletraffic Congress. [7] C.Fricker,P.Robert,J.Roberts,andN.Sbihi,“Impactoftrafficmixon cachingperformanceinacontent-centricnetwork,”inIEEEConference onComputerCommunicationsWorkshops(INFOCOMWKSHPS),2012, pp.310–315. [8] Y. Zhou, T. Z. Fu, and D. M. Chiu, “On replication algorithm in p2p vod,”IEEE/ACMTransactions onNetworking, pp.233–243,2013. [9] ——, “A unifying model and analysis of p2p vod replication and scheduling,” inProceedings IEEEINFOCOM,2012,pp.1530–1538. [10] D. Ciullo, V. Martina, M. Garetto, E. Leonardi, and G. L. Torrisi, “Stochasticanalysisofself-sustainability inpeer-assistedvodsystems,” inProceedings IEEEINFOCOM,2012,pp.1539–1547. [11] W. Jiang, S. Ioannidis, L. Massoulie´, and F. Picconi, “Orchestrating massivelydistributedcdns,”inProceedingsofthe8thinternationalcon- ferenceonEmergingnetworkingexperimentsandtechnologies. ACM, 2012,pp.133–144. [12] Y.Rochman, H.Levy,andE.Brosh,“Resource placement andassign- ment in distributed network topologies,” in Proceedings IEEE INFO- COM,2013. [13] ——, “Max percentile replication for optimal performance in multi- regionalp2pvodsystems,”inNinthInternationalConferenceonQuan- titative Evaluation ofSystems(QEST). IEEE,2012,pp.238–248. [14] N. Laoutaris, V. Zissimopoulos, and I. Stavrakakis, “On the optimiza- tion of storage capacity allocation for content distribution,” Computer Networks, vol.47,no.3,pp.409–428, 2005. [15] E.CohenandS.Shenker,“Replicationstrategiesinunstructuredpeer-to- peernetworks,”inACMSIGCOMMComputerCommunicationReview, vol.32,no.4,2002,pp.177–190. [16] S.TewariandL.Kleinrock,“Onfairness,optimaldownloadperformance and proportional replication in peer-to-peer networks,” in NETWORK- ING 2005. Networking Technologies, Services, and Protocols; Perfor- manceofComputerandCommunicationNetworks;MobileandWireless Communications Systems. Springer, 2005,pp.709–717. [17] ——, “Proportional replication in peer-to-peer networks.” in Proceed- ingsIEEEINFOCOM,2006. [18] M.Leconte,M.Lelarge,andL.Massoulie´,“Convergence ofmultivari- ate belief propagation, with applications to cuckoo hashing and load balancing,” in Proceedings of the Twenty-Fourth Annual ACM-SIAM Symposium on Discrete Algorithms (SODA), S. Khanna, Ed. SIAM, 2013,pp.35–46. [19] F.P.Kelly,“Lossnetworks,”Theannalsofappliedprobability,pp.319– 378,1991. [20] A.-S. Sznitman, “Topics in propagation of chaos,” in Ecole d’Ete´ de Probabilite´s deSaint-Flour XIX—1989. Springer, 1991,pp.165–251.

See more

The list of books you might like

Most books are stored in the elastic cloud where traffic is expensive. For this reason, we have a limit on daily download.