ebook img

DTIC ADA505907: NAVO MSRC Navigator. Spring 2006 PDF

3.8 MB·English
Save to my drive
Quick download
Download
Most books are stored in the elastic cloud where traffic is expensive. For this reason, we have a limit on daily download.

Preview DTIC ADA505907: NAVO MSRC Navigator. Spring 2006

Report Documentation Page Form Approved OMB No. 0704-0188 Public reporting burden for the collection of information is estimated to average 1 hour per response, including the time for reviewing instructions, searching existing data sources, gathering and maintaining the data needed, and completing and reviewing the collection of information. Send comments regarding this burden estimate or any other aspect of this collection of information, including suggestions for reducing this burden, to Washington Headquarters Services, Directorate for Information Operations and Reports, 1215 Jefferson Davis Highway, Suite 1204, Arlington VA 22202-4302. Respondents should be aware that notwithstanding any other provision of law, no person shall be subject to a penalty for failing to comply with a collection of information if it does not display a currently valid OMB control number. 1. REPORT DATE 3. DATES COVERED 2006 2. REPORT TYPE 00-00-2006 to 00-00-2006 4. TITLE AND SUBTITLE 5a. CONTRACT NUMBER NAVO MSRC Navigator. Spring 2006 5b. GRANT NUMBER 5c. PROGRAM ELEMENT NUMBER 6. AUTHOR(S) 5d. PROJECT NUMBER 5e. TASK NUMBER 5f. WORK UNIT NUMBER 7. PERFORMING ORGANIZATION NAME(S) AND ADDRESS(ES) 8. PERFORMING ORGANIZATION Naval Oceanographic Office (NAVO),Major Shared Resource Center REPORT NUMBER (MSRC),1002 Balch Boulevard,Stennis Space Center,MS,39522 9. SPONSORING/MONITORING AGENCY NAME(S) AND ADDRESS(ES) 10. SPONSOR/MONITOR’S ACRONYM(S) 11. SPONSOR/MONITOR’S REPORT NUMBER(S) 12. DISTRIBUTION/AVAILABILITY STATEMENT Approved for public release; distribution unlimited 13. SUPPLEMENTARY NOTES 14. ABSTRACT 15. SUBJECT TERMS 16. SECURITY CLASSIFICATION OF: 17. LIMITATION OF 18. NUMBER 19a. NAME OF ABSTRACT OF PAGES RESPONSIBLE PERSON a. REPORT b. ABSTRACT c. THIS PAGE Same as 28 unclassified unclassified unclassified Report (SAR) Standard Form 298 (Rev. 8-98) Prescribed by ANSI Std Z39-18 The NAVO MSRC is undergoing a carefully be configured with two gigabytes of memory planned series of enhancements which, when per processor, IBM's “Federation” inter- completed by Fall 2006, will provide one of the processor switch fabric, and IBM's Global most capable, productive, and balanced HPC Parallel File System (GPFS), all of which will environments ever fielded at this MSRC. These facilitate the execution of tremendously large enhancements substantially boost the computational applications and also diverse mixes of large capabilities and resilience of the MSRC across applications — applications which typically run both the classified and unclassified High as DoD Challenge Projects. To supplement this Performance Computing (HPC) environments enormous computational capability, we continue we support for the Department of Defense (DoD) HPC Modernization Program (HPCMP). The most significant enhancements will be the Enhancing the MSRC addition of two very large IBM HPC systems, both of which will be based upon IBM's to Serve You Better POWER5 processor technology: (cid:88)KRAKEN (the 3000-processor unclassified IBM POWER4+), one of the most successful and requested HPC systems within the NAVO MSRC, will be joined by an unclassified to enhance and optimize the internal mass 3000-processor IBM POWER5+ system storage and networking capabilities of the (named BABBAGE). MSRC for both performance and resilience. (cid:88)ROMULUS (the existing 500-processor IBM Finally, most of you are aware that the HPCMP POWER4+ system) will be transitioned to may be sustaining significant operating budget the classified environment and will be joined cuts beginning in FY07. The six HPC centers by a 1900-processor IBM POWER5+ system within the program are slated to receive the (named PASCAL) which will replace the bulk of these cuts. Please be assured that our existing classified IBM POWER4 system primary goal throughout this budget adjustment (MARCELLUS) when it is retired in early process will be the maintenance of a premier Fiscal Year 2007 (FY07). HPC environment with the support you have When all of these upgrades are complete, the come to expect from all of the centers. We effective computing power of the NAVO MSRC invite you, the DoD user community, to let will be essentially tripled, as measured by us continue to assist you in bringing this sustainable performance on the HPCMP cutting-edge capability to bear in support of benchmark suite. All four of these systems will your HPC needs. 2 SPRING 2006 NAVO MSRC NAVIGATOR Contents The Naval Oceanographic Office (NAVO) Major Shared Resource Center (MSRC): Delivering Science to the Warfighter The NAVO MSRC provides Department of Defense (DoD) scientists and engineers with high The Director’s Corner performance computing (HPC) resources, including leading edge computational systems, 2 Enhancing the MSRC to Serve You Better large-scale data storage and archiving, scientific visualization resources and training, and expertise Feature Articles in specific computational technology areas (CTAs). These CTAs include Computational Fluid Dynamics (CFD), Climate/Weather/Ocean 4 Five Dogs, a Hurricane, and Two IBM POWER5+s: Modeling and Simulation (CWO), Environmental A Personal and Professional Katrina Experience Quality Modeling and Simulation (EQM), Computational Electromagnetic and Acoustics 5 Drag Reduction by Microbubbles in a Spatially- (CEA), and Signal/Image Processing (SIP). Developing Turbulent Boundary Layer: Reynolds Number Effect (HPCMP/CAP) NAVO MSRC Code N7 9 Practical Parallel Graphics and Remote Visualization 1002 Balch Boulevard 12 Mapping the Seabed on an Absolute Reference Frame Stennis Space Center, MS 39522 System Using the Real-Time GIPSY (RTG) Global 1-800-993-7677 or [email protected] Differential GPS and RTK Positioning The Porthole 22 Visitors to the Naval Oceanographic Office Major Shared Resource Center 24 Boy Scout Jamboree NAVO MSRC Navigator Navigator Tools and Tips www.navo.hpc.mil/Navigator 25 Hints for Choosing “Consumable CPUs” and NAVO MSRC Navigator is a biannual technical “Consumable Memory” Resource Requests on KRAKEN publication designed to inform users of the news, and ROMULUS events, people, accomplishments, and activities of the Center. For a free subscription or to make Upcoming Events address changes, contact NAVO MSRC at the above address. 27 Coming Events EDITOR: Gioia Furness Petro, [email protected] DESIGNERS: Kerry Townson, [email protected] Lynn Yott, [email protected] Any opinions, conclusions, or recommendations in this publication are those of the author(s) and do not necessarily reflect those of the Navy or NAVO MSRC. All brand names and product names are trademarks or registered trademarks of their respective holders. These names are for information purposes only and do not imply endorsement by the Navy or NAVO MSRC. Approved for Public Release Distribution Unlimited NAVO MSRC NAVIGATOR SPRING 2006 3 Five Dogs, a Hurricane, and Two IBM POWER5+s: A Personal and Professional Katrina Experience Dave Cole, Government NAVO MSRC User Support Lead Articles in the Fall 2005 edition of the Navigator reported country. This brings our pet count to five small, frisky the unprecedented effects of Hurricane Katrina on the dogs! Naval Oceanographic Office Major Shared Resource (cid:88)Sun., 28 Aug.: Katrina now a Category (CAT) 5 Center (NAVO MSRC) and employees. The log that headed for New Orleans/West Mississippi (MS) Coast. follows, which records my evacuation from the Gulf Coast Evacuated to Minden, LA (where my parents live). and eventual return, provides a more personal snapshot of Boarded the frisky dogs. events that is in a small way representative of the ordeals (cid:88)Mon., 29 Aug.: Held second evaluation conference call that they experienced and overcame. at 1330 Central Time as scheduled. Welcome surprise But first, a little background. In January 2005 I became the that my dad, who avoids technology, has a speaker NAVO MSRC Technical Insertion for Fiscal Year 2006 (TI- phone in the kitchen! Used the kitchen table as my 06) Usability Team Chair. As such, I was charged with command post. Northeast quadrant of Katrina plows keeping the NAVO MSRC Usability Team process on into the west MS Coast with a storm surge greater than schedule as part of the Center's acquisition of two large 20 feet (my house is at a 12-foot elevation). Bought a IBM POWER5+clusters through the Department of Defense Universal Serial Bus (USB) compatible keyboard at (DoD) High Performance Computing Modernization Program Wal-Mart for my laptop and began development of the (HPCMP) TI-06 acquisition effort. Usability Brief. As the NAVO MSRC TI-06 Usability Team Chair, my (cid:88)Tue., 30 Aug.: Moved into Best Western for Digital participation began with a meeting in January 2005 with Subscriber Line (DSL) access. Completed the draft the TI-06 Performance Team and essentially culminated Usability Presentation and passed the ball to Tom with a presentation to the TI-06 Collective Acquisition Team Crimmins of the Army Research Laboratory (ARL) at (CAT) in November 2005. Keeping the Usability Team 22:24 so I could begin planning the trek home to pick process on schedule was especially challenging as Hurricane Katrina struck during the last week of the Phase I analysis. (cid:88)Fri., 26 Aug.: Held first Usability evaluation conferencecall Continued Page 21 as scheduled. Departed work under normal hurricane condition status—no preparation needed. (cid:88)Sat., 27 Aug.: Katrina now a threat—spent the day cutting plywood shutters for my house and preparing for evacuation. Liberated two beagles from local kennel for a friend who was out of the Five (usually) frisky dogs. 4 Drag Reduction by Microbubbles in a Spatially- Developing Turbulent Boundary Layer: Reynolds Number Effect (HPCMP/CAP) Antonino Ferrante and Said Elghobashi, Department of Mechanical and Aerospace Engineering, University of California, Irvine INTRODUCTION framework. The bubble acceleration Time integration in this simulation was equation, on the other hand, is solved performed using the second-order Experimental evidence during the past for each bubble to track its trajectory Adams-Bashforth scheme. The three decades indicates that the injection in time.4, 5 discretized Poisson equation for of gaseous microbubbles (diameter pressure was solved using a cosine ranging from 1 to 1000 microns, and The governing equations of the fluid transform in the streamwise direction, at a relatively large volume fraction motion account for the instantaneous a Fast Fourier Transform (FFT) in the (up to F = 0.7)) into a liquid turbulent n local volume fraction of the bubbles. boundary layer over a flat plate1, 2or spanwise direction, and Gauss The bubble equation of motion over axisymmetrical bodies3can reduce elimination in the wall-normal direction. includes terms representing the added the skin friction by as much as 80 The discrete cosine and Fourier mass, carrier fluid inertia, Stokes drag, percent from its value without bubble transforms were computed using the buoyancy, and lift force. The governing injection. However, the basic physical Fastest Fourier Transform in the West mechanisms responsible for that equations4, 5were discretized in space (FFTW) C subroutine library.6 reductionwere not yet fully understood. using a second-order finite difference scheme—except for the mean advection This article discusses the physical mechanisms responsible for the terms, which were evaluated via a reduction of skin friction in a fifth-order upwind differencing scheme. Continued Next Page... microbubble-laden, Spatially-Developing Turbulent Boundary Layer (SDTBL) over a flat plate4, and the effects of increasing Reynolds number on drag Fluid + Bubbles reduction.5This discussion is based on the results of the author's Direct Numerical Simulations (DNS) of a microbubble-laden SDTBL. These simulations were performed on High Z Performance Computing (HPC) highly scalable supercomputers (CRAY T3E <U > (x, z) and IBM Power4+ (KRAKEN)) at the 1 Naval Oceanographic Office Y (NAVOCEANO) Major Shared Resource Center (MSRC). g MATHEMATICAL DESCRIPTION Figure 1 shows a schematic of the X SDTBL flow where the gravitational acceleration vector is perpendicular to Wall the wall and pointing downward. The DNS used in this simulation employs the Eulerian-Lagrangian approach to solve the fluid continuity and momentum equations in an Eulerian Figure 1.Schematic of microbubble-laden turbulent boundary layer flow over a flat wall. NAVO MSRC NAVIGATOR SPRING 2006 5 CAP PARALLEL DNSBLB is more than ten times faster scalability of the code. Figure 2 shows PERFORMANCE TESTS than its T3E version. The scalability the CPU time in seconds per processor, runs of the DNS code were performed per time step needed for the integration During the past four years the authors using up to 1024 processors for two of the governing equations of both the have used their newly-developed different computational meshes: a fluid and bubbles, versus the number parallel code (DNSBLB, written in coarse mesh of 256x256x96 = 6x106 of processors used on KRAKEN. FORTRAN 90/MPI), which performs a grid points, and a fine mesh of Different lines correspond to different DNS of a microbubble-laden SDTBL. 4, 5, 7, 8, 9DNSBLB is parallelized with 1024x1024x96 = 101x106grid computationalmeshes and numbers a One-Dimensional (1D) domain points; and for four different numbers of bubbles. The CPU time monotonically decreases as the number of processors decomposition in the spanwise of bubbles 100 thousand, 16 million, increase (see Figure 2) for all the tests y-direction (j) of the three-dimensional 100 million, and 200 million. performed. computational domain.*DNSBLB was The scalability tests were performed in Figure 2 also shows that for the coarse originally written for the CRAY T3E. a one-way coupling regime, i.e., no mesh with 100 thousand bubbles, the During the Capability Applications effects of bubble on turbulence were slope of the line (CPU time versus Project (CAP) 2004, the CRAY T3E accounted for. The two-way coupling number of processors) is close to version of DNSBLB was converted to increases the Central Processing Unit -0.1, whereas for the fine mesh with run on KRAKEN. The IBM version of (CPU) time but does not worsen the 200 million bubbles, the slope is close to -1. The slopes for the other tests are Coarse mesh: 256x256x96=6.2x106 between -0.1 and -1. This means that Fine mesh: 1024x1024x96=101x106 the performance of the DNS code improves as the number of grid points of the computational mesh and the p e number of bubbles simulated increase. t e S 101 slope = 1 Furthermore, for the fine mesh the m code shows a nearly ideal scalability Ti since the line in Figure 2 has a slope r e close to -1. P & r DRAG REDUCTION BY o s MICROBUBBLES: REYNOLDS s ce NUMBER EFFECT o Pr The DNS results4for the microbubble- er 100 laden SDTBL for RrV= 1430 with P ) volume fraction ranging from Fn= c e 0.001 to 0.02 show that the presence s ( slope = -0.1 of bubbles results in a local positive me divergence of the fluid velocity, DD.U > U ti 0. This creates a positive mean P 10-1 velocity, <U3>, normal to (and away C from) the wall which, in turn, reduces 100 101 102 103 the mean streamwise velocity and Number of Processors displacesthe quasi-streamwise longitudinal vortical structures away from the wall as in Figure 3. This Coarse mesh & 100K Bubbles Fine mesh & 200M Bubbles Coarse mesh & 16M Bubbles slope = -1 displacement has two main effects: Coarse mesh & 100M Bubbles slope = -0.1 Fine mesh & 100M Bubbles (cid:88)It increases the spanwise gaps between the wall streaks associatedwith the sweep events Figure 2.Scalability on IBM P4+ (KRAKEN) of DNS code. and reduces the streamwise *If np is the number of processors and N, even multiple of np, is the number of grid points in the jdirection, then each processor makes calculation on N=npplanes (x-zplanes) of the computational domain. 6 SPRING 2006 NAVO MSRC NAVIGATOR velocity in these streaks, thus per processor and 124 gigabytes of the same amount of reduction in skin reducing the skin friction. maximum total memory. friction at higher Reynolds number. (cid:88)It moves the location of peak The DNS results5show that increasing Reynolds stress production away Reynolds numbers from 1430 to 2900 SUMMARY from the wall to a zone of a smaller decreases the percentage of drag This article has briefly discussed the transverse gradient of the mean reduction from 22 to 19 percent. physical mechanisms of drag reduction streamwise velocity (i.e., smaller Increasing RrV“squeezes” the quasi- by microbubbles and the Reynolds mean shear), thus reducingthe streamwise vortical structures toward number effect.4, 5 production rate of turbulence the wall, whereas microbubbles “push Furthermore, it reports some details kinetic energy and enstrophy. them away” from the wall. on the simulations of parallel DNS During CAP 2004 the authors were The net result of these two opposing code and its scalability on the able to simulate the microbubble- effects determines the amount of skin NAVOCEANO MSRC KRAKEN system. laden SDTBL for RrV= 2900 with friction reduction by the microbubbles. volume fraction F = 0.01 and bubble The displacement action by the In conclusion, the CAP program has n diameter of 40um using a computational microbubbles is a result of the local considerably helped the author's drag mesh of 1024x512x128 = 67x106 positive velocity divergence, DD .U, reduction by microbubbles research grid points and 29 million bubbles. created by their concentration gradients. to explain the effect of Reynolds Thus, the volume fraction of bubbles number on drag reduction by allowing The computations were performed on 512 processors of the KRAKEN that is responsible for the reduction of the use of a large number of processors system. The bubble-laden flow skin friction in a low Reynolds number and extended CPU hours which were simulations required 9.6 CPU hours SDTBL is not sufficient to produce not allowed in the “standard” queues. Single-Phase Flow Bubble-Laden Flow Quasi-Streamwise Vortex Quasi-Streamwise Vortex Displaced Away from Wall ‘sweep’ ‘ejection’ ‘sweep’ ‘ejection’ Induced Vertical Fluid Velocity g g by Velocity DivergenceEffect Wall Wall High Speed Low Speed Reduced Increased Streak Streak High Speed Low Speed Streak Streak Reduced Skin Friction Figure 3.Schematic of the drag reduction mechanism in a microbubble-laden SDTBL. NAVO MSRC NAVIGATOR SPRING 2006 7 Acknowledgements This work was supported by Office of Naval Research (ONR) Grant No. N00014-05-1-0059 under the direction of Dr. Patrick Purtell. The computations were performed on the IBM Power4+ (KRAKEN) located at the NAVO MSRC at the John C. Stennis Space Center (Mississippi), and on CRAY T3E (MD) located at the U.S. Army High Performing Computing Research Center (AHPCRC, Minnesota). The authors thank the outreach group at the NAVOCEANO MSRC and the support group at AHPCRC for allowing jobs with CPU time limits larger than those allowed in the “standard” queues. References 1. Madavan, N.K., S. Deutsch, and C.L. Merkle, “Reduction of Turbulent Skin Friction By Microbubbles,” Physics of Fluids, 27: 356-363, 1984. 2. Pal, S., C. L. Merkle, and S. Deutsch, “Bubble Characteristics and Trajectories in a Microbubble Boundary Layer,” Physics of Fluids, 31(4): 744-751, 1988. 3. Deutsch, S., and S. Pal, “Local Shear Stress Measurements on an Axisymmetrical Body in a Microbubble-Modified Flow Field,” Physics of Fluids, 2: 2140-2146, 1990. 4. Ferrante, A., and S. Elghobashi, “On the Physical Mechanisms of Drag Reduction in a Spatially Developing Turbulent Boundary Layer Laden with Microbubbles,” Journal of Fluid Mechanics, 503: 345-355, 2004. 5. Ferrante, A., and S. Elghobashi, “Reynolds Number Effect on Drag Reduction in a Microbubble- Laden Spatially Developing Turbulent Boundary Layer,” Journal of Fluid Mechanics, 543: 93-106, 2005. 6. Frigo, M., and S.G. Johnson, “The Design and Implementation of FFTW3,” Proceedings of the IEEE, 93(2): 216-231, 2005. 7. Ferrante, A., and S. Elghobashi, “A Robust Method for Generating Inflow Conditions for Direct Simulations of Spatially-Developing Turbulent Boundary Layers,” Journal of Computational Physics, 198:372-387, 2004. 8. Ferrante, A., S. Elghobashi, P. Adams, M. Valenciano, and D. Longmire, “Evolution of Quasi- Streamwise Vortex Tubes and Wall Streaks in a Bubble-Laden Turbulent Boundary Layer Over A Flat Plate,” Physics of Fluids, 16(9): S2, 2004. 9. A. Ferrante, “Reduction of Skin-Friction in a Microbubble-Laden Spatially Developing Turbulent Boundary Layer Over a Flat Plate,” PhD thesis, University of California, Irvine, 2004. 8 SPRING 2006 NAVO MSRC NAVIGATOR Practical Parallel Graphics and Remote Visualization Sean Ziegeler, Visualization Software Engineer, NAVO MSRC VADIC The purpose of scientific visualization disks before visualization or transferring improved performance. This is, quite often, to analyze the output of it “on-the-fly” during visualization improvement is because a parallel a computational model. As computational using a remote file system. file system consists of multiple disks power increases, model output grows The next technique, parallel graphics, spread out over many or all nodes of in size, which introduces new obstacles uses large graphics systems (such as the cluster. An example at the Naval that prevent users from analyzing their clusters) to render in parallel fashion Oceanographic Office (NAVOCEANO) data—especially if they are not collocated by distributing the load across multiple Major Shared Resource Center (MSRC) with the computing resources. Graphics Processing Units (GPU's). is the /scrfile system on the visualization The traditional approach to scientific Finally, remote visualization can deliver cluster, SEAHORSE. visualization from remote sites has the rendered images, reduced now If the data will not fit on /scr, will not been to transfer the output data from merely to pixels (and much smaller be visualized repeatedly, or won't the computational system to the than the original data), from the require the high-speed access of a user's local site as shown in Figure 1. graphics system to the user's local parallel file system, then the Secure This is less feasible with larger data system. This new approach is shown Shell File System (SSHFS) is an because it may be too large to (1) in Figure 2. excellent tool for accessing data on a transfer over the network between separate system as if it were local. sites in a reasonable amount of time, REMOTE DATA ACCESS SSHFS uses the Secure Shell (SSH) (2) fit on the user's system disk program that is already installed on all Staging the data on a visualization space, or (3) render on the user's MSRC systems to facilitate authenticated system is very similar to transferring local graphics computer. and encrypted file access. the data from the computational To overcome all three problems, a system to the user's local system. The difference between SSHFS and a combination of techniques must be This is the preferred method if the traditional file transfer is that the remote utilized: computational data access, data set is to be visualized repeatedly, files will appear to be like any other parallel graphics, and remote visualization. is not too large to fit on the file on the local system. Command Example 1 shows how to “mount” Computational data access involves visualization system, and requires or begin, briefly utilize, and end an the transfer of the data between a high performance access. SSHFS session. computational and a visualization The result of staging the data in system, either by staging it on large parallel on a graphics cluster is Continued Next Page... Figure 1.Traditional scientific visualization by transferring data to user's site. NAVO MSRC NAVIGATOR SPRING 2006 9

See more

The list of books you might like

Most books are stored in the elastic cloud where traffic is expensive. For this reason, we have a limit on daily download.