ebook img

Vision Steered Beam-forming and Transaural Rendering for the Arti cial Life Interactive Video ... PDF

28 Pages·2001·0.47 MB·English
by  
Save to my drive
Quick download
Download
Most books are stored in the elastic cloud where traffic is expensive. For this reason, we have a limit on daily download.

Preview Vision Steered Beam-forming and Transaural Rendering for the Arti cial Life Interactive Video ...

Vision Steered Beam-forming and Transaural Rendering for the Arti(cid:12)cial Life Interactive Video Environment, (ALIVE) Michael A. Casey, William G. Gardner, Sumit Basu MIT Media Laboratory, Cambridge, USA mkc,billg,[email protected] Abstract This paper describes the audio component of a virtual reality system thatusesremotesensingtofreetheuserfrombody-mountedtrackingequip- ment. Position informationis obtained from a camera and used to con- straina beam-formingmicrophonearray, forfar-(cid:12)eldspeechinput,anda two-speakertransauralaudiosystemfor rendering3D audio. 1 Vision Steered Beam-Forming 1.1 Introduction The MediaLab’s ALIVE project, Arti(cid:12)cialLife Interactive Video Environment is a testbed application platform for research into remote-sensing full-body in- teractiveinterfacesforvirtualenvironments. AplanviewoftheALIVEspace is showninFigure1. Acameraontopofthelargevideoprojectionscreencaptures images of the user in the active zone, these images are fed to a visual recogni- tion system where various features are tracked. A mirror image of the user is projected ontothevideoscreen atthefrontofthe spaceandcomputer-rendered scenes are overlayed onto the users imageto produce a combined real/arti(cid:12)cial compositeimage. OneoftheapplicationsintheALIVEspaceisanautonomous agentmodelofadog,called\Silas",whichposseses itsownsystemofbehaviors and motor-control schemes. The user is able to give the dog visual commands bygesturing. Thevisualrecognitionsystemmakesdecisionsonwhichcommand has been issued out of the twelve gestures that it can recognize, see Figure 2. This section describes the design and implementation of an audio component for the ALIVE system that allows the user to give the arti(cid:12)cial dog spoken commandsfromthe free (cid:12)eld without the use of body-mounted microphones. 1 1.2 Free-(cid:12)eld speech recognition Speech recognition applications typically require near-(cid:12)eld, i.e. <1:5m, micro- phone placement for acceptable performance. Beyond this distance the signal to noise ratio ofthe incomingspeech a(cid:11)ects the performace signi(cid:12)cantly. Com- mercialspeech-recognition packages typicallybreakdownover a4 6dB range. (cid:0) The ALIVEspace requires the user tobe free oftheconstraints ofnear-(cid:12)eld microphone placement and the user must be able to move around the active zone of the space with no noticable degradation in performance. As a result there are several potential solutions. One of these is to have a highly directional microphone that can be panned using a motorized control unit,totracktheuser’slocation. Thisrequiresasigni(cid:12)cantamountofmounting and control hardware, and is limited by the speed and accuracy of the drive motors. In addition, it can only track one user at a time. It is preferable to havea directionalresponse that can be steered electronically. This can be done with the well-known technique of beamforming with an array of microphone elements. Though several microphones need to be used for this method, they need not be very directional and they can be permanently mounted in the environment. In addition, the signals from the microphones in the array can be combined in as manyways as the available computationalpower is capable of, allowing for the tracking of multiple moving sound sources from a single microphonearray. 1.3 Adaptive Beamforming Adaptive beamformingstrategies account for movementin source direction by continuously updating the spatial characteristics of the array for an optimal signaltonoiseratio. Thebehaviorofthearrayisthusdeterminedbythenature of the source signal(s). The problem with adaptive strategies for the ALIVE space is that there are typicallymanyobservers around the space and the level of ambient speech-like sound is very high as a result. Adaptive algorithms do not perform well when multiple sources arrive simultaneously from di(cid:11)erent directions [2]. 1.4 Fixed Beamforming Fixedarraystrategiesoptimizethemicrophone(cid:12)lteringforaparticulardirection and don’t change with varying incident source direction. Thus the directional response ofthe arrayis (cid:12)xed toaparticular azimuthand elevation. However,if the target source is non-stationary, the signal enhancement performance is re- duced asthesource movesawayfromthe steering direction. Spatialbeamwidth constraints maybe added to the (cid:12)xed array such that the directionality of the response is traded for beam width to compensate for small movements in the source. As the beam width increases, so too does the level of ambient noise 2 pickup. The active user zone inthe ALIVE space allowsmovementover alarge azimuthrange thus (cid:12)xed arrayformulationsneed to be modi(cid:12)edinreal-timein order to beamformin the direction of the user. 1.5 A Visually Constrained Beamformer The ALIVE space utilizes a visual recognition system called P(cid:12)nder, short for person (cid:12)nder,developedattheMediaLabfortrackingaperson’s hands,faceor any other color-discriminable feature [4]. P(cid:12)nder uses an intensity-normalized color representation of each pixel in the camera imageand multi-wayGaussian classi(cid:12)ers to decide which of several classes each pixel belongs to. Examples of classes are background, left hand, right hand and head. The background class canbe madeupofarbitraryscenery aslongasthecolorvalueofeachpixeldoes notvarybeyondthedecisionboundaryforinclusioninanotherclass. Themean of each cluster gives the coordinates of the class, and the eigenvalues give the orientation. P(cid:12)nder provides updates on each class roughly 6 times a second. Further details on the visual recognition system can be found in [4]. The informationfrom the visual recognition system is used to steer a (cid:12)xed beamformingalgorithm. Azimuth calculations are performed from the 3-space coordinate data provided by the mean of the head class. The re-calculation of weightsfor each new azimuthisa relativelylow-costoperation since the weight update rate is 5 Hz. The use of visual recognition techniques makes it possible to achieve both theoptimalsignal-enhancementperformanceofa(cid:12)xedbeamformerwithnarrow beamwidth and to get the spatial (cid:13)exibilityof an adaptive beamformer. 1.6 Fixed Beamformer Algorithms Inthissectionwedescribea(cid:12)xedbeamformeralgorithmandthedi(cid:11)erentmicro- phone arrangementsthat can be used with it. The geometryofthe microphone arrayis represented bythe set ofvectors rn which describe the position ofeach microphonenrelativetosomereference point(e.g.,the center ofthe array),see Figure 3. The arrayis steered to maximizethe response to planewaves coming fromthe direction rs of frequency fo. Then, for a plane waveincident fromthe direction^ri, at angle (cid:18), the gain is: jkro(cid:1)^ri F((cid:18))e jkr1(cid:1)^ri F((cid:18))e G((cid:18))= ao a1 a2 a3 2 jkr2(cid:1)^ri 3 (1) F((cid:18))e (cid:2) (cid:3)66 F((cid:18))ejkr3(cid:1)^ri 77 6 7 (cid:0)jko^rn(cid:1)^rs 4 5 where an = an e , F((cid:18)) is the gain pattern of each individual mi- j j crophone, and k (2(cid:25)f=c) is the wave number of the incident plane wave. ko is the wave number corresponding to the frequency fo ofthe incident plane wave. 3 Notethatthere isalsoa(cid:30)dependence forF andG,but since weareonlyinter- ested insteering inone dimension,wehave omittedthis factor. Thisexpression can be written morecompactly as: T G((cid:18))=W H (2) where W represents the microphone weights and H is the set of transfer functions between each microphoneand the reference point. In the formulation above, a maxima is created in the gain pattern at the steering angle for the expected frequency, since^ri =^rs and the phase termsin W and H cancel each other. Note, however, that this is not the only set of weights that can be used forW. Forexample,Stadler andRabinowitzpresent amethodofobtainingthe weights with a parameter (cid:12) that arbitrates high directivity and uncorrelated noise gain [7]. This method, when used to obtain maximumdirectivity, yields gainpatterns thatareslightlymoredirectionalthanthebasicweightsdescribed above. The standard performance metric for the directionality of a (cid:12)xed array is the directivity index which is shown in Equation 3, [7]. The directivity index is the ratio of the array output power due to sound arriving from the far (cid:12)eld in the target direction, ((cid:30)0;(cid:18)0), to the output power due to sound arriving from allother directions in a spherically isotropic noise (cid:12)eld, 2 G((cid:30)0;(cid:18)0) D = (cid:25) j2(cid:25) j 2 ; (3) (1=4(cid:25)) (cid:18)=0 (cid:30)=0 G((cid:30);(cid:18)) sin(cid:18)d(cid:30)d(cid:18) j j The directivityindexthus fRormuRlatedisanarrow-bandperformance metric; it is dependent on frequency but the frequency terms are omitted from Equa- tion 3 for simplicity of notation. In order to assess an array for use in speech enhancement a broad-band performance metric must be used. One such metric is the intelligibility-weighted directivity index [7] in which the directivity index is weighted by a set of frequency-dependent coe(cid:14)cients providedbytheANSIstandardforthespeecharticulationindex[1]. Thismetric weights the directivity index in fourteen one-third-octave bands spanning 180 to 4500 Hz [7]. 1.7 Designing the Array An important (cid:12)rst consideration is the choice of array geometry. Two possible architectures are considered; end(cid:12)re arrangement, Figure 3, and broadside ar- rangement, Figure 4. A second factor is the choice of microphone gain pattern for the individualmicrophone elements, F((cid:18)). Since the gain pattern F((cid:18)) can be pulled out of the H vector as a constant multiplier, the gain pattern for the array can be viewed as the product of the microphonegain pattern and an omnidirectional response where F((cid:18)) = 1. This is the well-known principle of 4 pattern multiplication [3] [7]. For omnidirectional microphones, the gain pat- terns for the two layouts are identical but for a rotation. The gain patterns for anend(cid:12)re arraycentered alongthe(cid:18) =0axiswithomnidirectionalmicrophones steered at 15, 45, and 75 degrees is shown in Figure 5. The use of a cardioid response greatly reduces the largelobes appearingin the rear of the array. The corresponding responses for cardioid microphones are shown in Figure 6. Cardioid elements for the broadside array are not as useful, since the null of the cardioid does not eliminate as much of the undesirable lobes of the gain pattern, Figure 7. Note the symmetry of the response about the (cid:18) = 0 axis; the line containingthe microphonearray. This symmetrycan be eliminatedby nulling out one half of the array response using an acoustic re(cid:13)ector or ba(cid:15)e alongoneside ofthe microphonearray. The re(cid:13)ector wille(cid:11)ectively doubleone side of the gain pattern and eliminate the other, while the ba(cid:15)e will eliminate one side and not a(cid:11)ect the other. Thus a good directional response can be achieved between 0 and 90 degrees using both cardioid elements and a ba(cid:15)e for the end(cid:12)re con(cid:12)guration. The incorporation of a second array,on the other side of the ba(cid:15)e, gives the angles zero to -90 degrees. A layout of the ALIVE space with such an array/ba(cid:15)e combinationis shown in Figure 8. The response of each of the above arrays as measured by the directivity index of Equation 3 is given in Table 1. The integration was performed for a (cid:12)xed (cid:30) across all (cid:18). Table 1: Directivity Index for Di(cid:11)erent Array Architectures Array Architecture 15 degrees 45 degrees 75 degrees OmniBroadside 0.0029 0.0022 0.0022 OmniEnd(cid:12)re 0.0022 0.0022 0.0029 Cardioid Broadside 0.0057 0.0038 0.0045 Cardioid End(cid:12)re 0.0036 0.0037 0.0045 1.8 Conclusion In this section we have described a vision-steered microphone array for use in a full-body interaction virtual environment without body-mounted sensing equipment. A preliminary implementation for the ALIVE space has shown that the system works well for constrained grammarsof 10-20commands. The advantages of cross-modal integration of sensory input are paramount in this system since the desirable properties of (cid:12)xed arrays are combined with the steerability of an adaptive system. 5 2 Visually Steered 3-D Audio 2.1 Introduction This section discusses the 3-D audio system that has been developed for the ALIVE project at the Media Lab [4]. The audio system can position sounds at arbitrary azimuthsand elevations around a listener’s head. The system uses stereo loudspeakers arranged conventionally (at 30 degrees with respect to (cid:6) the listener). The system works by reconstructing the acoustic pressures at the listener’searsthatwouldoccurwithafree-(cid:12)eldsoundsourceatthedesiredloca- tion. Thisisaccomplishedbycombiningabinaural spatializerwithatransaural audio system. The spatializer convolves the source sound with the direction dependent (cid:12)lters that simulate the transmission of sound from free-(cid:12)eld to the twoears. Theresultingbinauraloutputofthespatializerissuitableforlistening over headphones. In order to present the audio over loudspeakers, the output of the spatializer is fed to a transaural audio system which delivers binaural signals to the ears using stereo speakers. The transaural system (cid:12)lters the bin- auralsignalssothatthe crosstalk leakagefromeach speaker tothe oppositeear is canceled. Transauraltechnologyis applicabletosituationswhere asinglelistener isin a reasonably constrained position facing stereo speakers. If the listener moves awayfromthe ideallisteningposition,the crosstalkcancellationnolongerfunc- tions,andthe 3-Daudioillusionvanishes. As discussed earlier,the ALIVE sys- temusesvideocamerastotrackthepositionofthelistener’sheadandhands. A goalofthis workis toinvestigate whether the tracking informationcan be used to dynamically adapt the transaural 3-D audio system so that the 3-D audio illusionis maintainedas the listener moves. We will (cid:12)rst brie(cid:13)y review the principles behind binaural spatial synthesis and transaural audio. Then we will discuss the 3-D audio system that has been constructed for the ALIVE project. Finally we will discuss how the head tracking informationcan be used, and give preliminaryresults. 2.2 Principles of binaural spatial synthesis A binaural spatializer simulates the auditory experience of one or more sound sources arbitrarily located around a listener [9]. The basic idea is to reproduce the acoustical signals at the two ears that would occur in a normal listening situation. This is accomplished by convolving each source signal with the pair 1 of head-related transfer functions (HRTFs) that correspond to the direction of the source, and the resulting binaural signalis presented to the listener over headphones. Usually,theHRTFsareequalizedtocompensatefortheheadphone 1 ThetimedomainequivalentofanHRTFiscalledahead-relatedimpulseresponse(HRIR) andisobtainedviatheinverseFouriertransformofanHRTF. Inthispaper,we willusethe termHRTF torefertoboththetimeandfrequencydomainrepresentation. 6 toearfrequencyresponse[26,17]. Aschematicdiagramofasinglesourcesystem is shown in (cid:12)gure 9. The direction of the source ((cid:18) = azimuth, (cid:30) = elevation) determines which pair of HRTFs to use, and the distance (r) determines the gain. Figure 10 shows a multiple source spatializer that adds a constant level of reverberation to enhance distance perception. The simplest implementation of a binaural spatializer uses the measured HRTFs directly as (cid:12)nite impulse response (FIR) (cid:12)lters. Because the head re- sponse persists for several milliseconds, HRTFs can be more than 100 samples long at typical audio sampling rates. The interaural delay can be included in the(cid:12)lterresponses directlyasleadingzerocoe(cid:14)cients, orcanbefactored outin ane(cid:11)ort to shorten the (cid:12)lter lengths. It is also possible to use mimimumphase (cid:12)lters derived fromthe HRTFs [15], since these will in general be shorter than the original HRTFs. This is somewhat risky because the resulting interaural phase may be completely distorted. It would appear, however, that interaural amplitudesasafunctionoffrequencyencodemoreusefuldirectionalinformation than interaural phase [16]. There are several problemscommonto headphone spatializers: The HRTFs used for sythesis are often a generic set and not the speci(cid:12)c (cid:15) HRTFs of the listener. This can cause localization performance to su(cid:11)er [24, 25], particularly in regards to front-back discrimination, elevation perception, and externalization. When the listener’s own head responses are used, theirlocalizationperformanceiscomparabletonaturallistening [26]. The auditory scene created moves with the head. This can be (cid:12)xed by (cid:15) dynamicallytrackingtheorientationoftheheadandupdatingtheHRTFs appropriately. Localizationperformanceandrealismshouldbothimprove when dynamiccues are added [24]. The auditory images created are not perceived as being external to the (cid:15) head, butrather arelocalizedatthe headorinsidethe head. Externaliza- tion can be improved by using the listener’s own head responses, adding reverberation, and adding dynamiccues [12]. Frontalsoundsarelocalizedbetween theearsorontopotthehead,rather (cid:15) than in front ofthe listener. Because we are used to seeing sound sources that are in front of the head, itis di(cid:14)cult to convince the perceptual sys- temthat asound iscomingfromthe front withoutacorresponding visual cue. However,whenusingthelistener’sownHRTF’s,frontalimagingwith headphones can be excellent. 2.3 Implementation of binaural spatializer Our implementation of the binaural spatializer is quite straightforward. The HRTFs were measured using a KEMAR (Knowles Electronics Mannequin for 7 Acoustics Research), which is a high quality dummy-head microphone. The HRTFs were measured in 10 degree elevation increments from -40 to +90 de- grees [13]. In the horizontal plane (0 degrees elevation), measurements were made every 5 degrees of azimuth. In total, 710 directions were measured. The samplingdensity was chosen to be roughly in accordance with the localization resolution of humans. The HRTFs were measured at a 44.1 kHz samplingrate. The raw HRTF measurements contained not only the desired acoustical re- sponse of the dummyhead, but also the response of the measurement system, includingthespeaker, microphones,andassociatedelectronics. Inaddition,the measured HRTFs contained the response of the KEMAR ear canals. This is undesirable,because the (cid:12)nalpresentation ofspatializedaudiotoa listener will involvethe listener’s ownear canals, andthus adouble ear canalresonance will be heard. One way to eliminate all factors which do not vary as a function of direction is to equalize the HRTFs to a di(cid:11)use-(cid:12)eld reference [15]. This is accomplishedby (cid:12)rst formingthe di(cid:11)use-(cid:12)eld average of all the HRTFs: 2 1 2 HDF = H(cid:18)i;(cid:30)k (4) j j N j j i;k X 2 whereH(cid:18)i;(cid:30)k isthemeasuredHRTFforazimuth(cid:18)i andelevation(cid:30)k. HDF j j is therefore the power spectrum which would result from a spatially di(cid:11)use sound(cid:12)eld of white noise excitation. This formulation assumes uniform spa- tial sampling around the head. The HRTFs are equalized using a minimum phase (cid:12)lter whose magnitude is the inverse of HDF . Thus, the di(cid:11)use-(cid:12)eld j j average of the equalized HRTFs is (cid:13)at. Figure 11 shows the di(cid:11)use-(cid:12)eld av- erage of the HRTFs. It is dominated by the ear canal resonance at 2-3 kHz. The low-frequency dropo(cid:11) is a result of the poor low-frequency response of the measurementspeaker. The inverse equalizing (cid:12)lter was gain limitedto prevent excessive noise ampli(cid:12)cationat extreme frequencies. In additionto the di(cid:11)use- (cid:12)eld equalization, the HRTFs were sample rate converted to 32 kHz. This was done in order to reduce the computationalrequirements ofthe spatializer. The (cid:12)nalHRTFswerecroppedto128points(4msec)whichwasmorethansu(cid:14)cient to capture the entire head response including interaural delays. The spatializer convolves a monophonic input signal with a pair of HRTFs to produce a stereophonic (binaural) output. The HRTFs that are closest to the desired azimuth and elevation are used. For e(cid:14)ciency, the convolution is accomplished using an overlap-save block convolver [20] based on the fast Fourier transform (FFT). Because the impulseresponse is 128 points long, the convolution is performed in 128-point blocks, using a 256-point real FFT. The forwardtransformsofallHRTFsarepre-computed. Foreach128-pointblockof inputsamples(every4msec),theforwardtransformofthesamplesiscalculated, and then two spectral multiplies and two inverse FFTs are calculated to form the two 128-point blocks of output samples. In addition to the convolution, a gainmultiplicationis performed to control apparent distance. 8 It isessential that the positionof the source can be changed smoothlywith- out introducing clicks into the output. This is easily accomplished as follows. Every 12 blocks (48 msec) the new source position is sampled and a new set of HRTFs is selected. The input block is convolved with both the previous HRTFs and the new HRTFs, and the two results are crossfaded using a linear crossfade. ThisassumesreasonablecorrelationbetweenthetwopairsofHRTFs. Subsequentblocksareprocessed usingthenewHRTFsuntilthenextpositionis sampled. The samplingrate of position updates is about 20 Hz, which is quite adequate for slow movingsources. 2.4 Performance of binaural spatializer The binaural spatializer (single source, 32 kHz samplingrate) runs in realtime on an SGI Indigo workstation. Source position can be controlled using a MIDI (Musical Instrument DigitalInterface) controller that has a set of sliders which are assigned to control azimuth, elevation, and distance (gain). A constant amount of reverberation can be mixed into the (cid:12)nal output using an external reverberator as shown in (cid:12)gure 10. The spatializer was evaluated using headphones (AKG-K240, which are di(cid:11)use-(cid:12)eld equalized [23, 18]). The input sound, usually music or sound ef- fects, was taken from one channel of a compact disc player. The spatializer worked quite well for lateral and rear directions for all listeners. As expected, somelisteners had problemswithfront-backreversals. Elevationcontrolo(cid:11) the medial plane was also good, though this varied considerably among listeners. Alllisteners experienced poorexternalizationoffrontalsounds. Atzero degrees elevation,as the source was panned across the front, the perception was always ofthe source movingthrough the head between the ears, or sometimesover the top of the head. Externalization was far better at lateral and rear azimuths. Adding reverberation did improve the realism of the distance control, but did not (cid:12)x the problem of frontal externalization. Clearly a problem of using non- individualizedHRTFs with headphones is the di(cid:14)culty of externalizing frontal sources. In order to increase the number ofsources, or to addintegral reverberation, the performance of the spatializer would need to be improved. Several things could be done: Reduce the (cid:12)lter size by simple cropping (rectangular windowing). (cid:15) Reduce the(cid:12)ltersizebyfactoringouttheinterauraldelayandimplement- (cid:15) ing this separately fromthe convolution. Reduce the (cid:12)lter size by using minimumphase (cid:12)lters. (cid:15) Model the HRTFs using in(cid:12)nite impulse response (IIR) (cid:12)lters. (cid:15) 9 Many of these strategies are discussed in [15]. To obtain the best price to performanceratio,commercialspatializersattempttobease(cid:14)cientaspossible, and usually run on dedicated DSPs. Consequently, the (cid:12)lters are modeled as e(cid:14)cientlyaspossible andthe algorithmsarehand-coded. In addition,there are usually serious memory constraints which prevent having a large database of HRTFs,andthusparameterizationandinterpolationofHRTFsisanimportant issue. Lack of memoryis not a problemin our implementation. 2.5 Principles of transaural audio Transaural audio is a method used to deliver binaural signals to the ears of a listener using stereo loudspeakers. The basic idea is to (cid:12)lter the binauralsignal suchthatthesubsequent stereopresentation producesthebinauralsignalatthe ears of the listener. The technique was (cid:12)rst put into practice by Schroeder and Atal [22, 21] and later re(cid:12)ned by Cooper and Bauck [10], who referred to it as \transaural audio". The stereo listening situation is shown in (cid:12)gure 12, where x^L and x^R are the signals sent to the speakers, and yL and yR are the signals atthelistener’s ears. Thesystemcan be fullydescribed bythevector equation: y=Hx^ (5) where: yL HLL HRL x^L y = ;H= ;^x= (6) yR HLR HRR x^R (cid:20) (cid:21) (cid:20) (cid:21) (cid:20) (cid:21) and HXY is the transfer function from speaker X to ear Y. The frequency variablehas been omitted. Ifx isthebinauralsignalwewishtodelivertothe ears, thenwe mustinvert (cid:0) 1 the system transfer matrixH such that x^=H x. The inverse matrixis: (cid:0)1 1 HRR HRL H = (cid:0) (7) HLLHRR HLRHRL HLR HLL (cid:0) (cid:20) (cid:0) (cid:21) This leads to the general transaural (cid:12)lter shown in (cid:12)gure 13. This is often calledacrosstalk cancellation(cid:12)lter, because iteliminatesthe crosstalk between channels. When the listening situation is symmetric, the inverse (cid:12)lter can be speci(cid:12)ed intermsoftheipsilateral(Hi =HLL =HRR)andcontralateral(Hc = HLR =HRL) responses: (cid:0)1 1 Hi Hc H = Hi2 Hc2 Hc (cid:0)Hi (8) (cid:0) (cid:20) (cid:0) (cid:21) CooperandBauckproposedusinga\shu(cid:15)er"implementationofthetransaural (cid:12)lter[10],whichinvolvesformingthe sumanddi(cid:11)erence ofxL and xR,(cid:12)ltering these signals,andthenundoingthe sumanddi(cid:11)erence operation. Thesumand 10

Description:
strain a beam-forming microphone array, for far-field speech input, and a two- speaker transaural audio system for rendering 3D audio. 1 Vision Steered Beam-
See more

The list of books you might like

Most books are stored in the elastic cloud where traffic is expensive. For this reason, we have a limit on daily download.