TLC2 Infrastructure Efforts and Grid Related Research
Transcription
TLC2 Infrastructure Efforts and Grid Related Research
TLC2 Infrastructure Efforts and Grid Related Research Lennart Johnsson Director TLC2 Department of Computer Science UHCC 2nd floor Fleming 232 SR-1, 232 10 TB SR-3, 402G 10 TB rackmount Layer-3 GigE MM 5 TB rackmount Layer-2 GigE PC Picture Layer-2 GigE Layer-2 GigE MM SM SM Fleming 117 SM SM SM SR-1 1xx “PGH 10” RFO863609 Labels A. Switch in “PGH-10” B. Switches at Baylor/Rice PGH 2XX C. D. Switches at UHCC/PGH576 Alternate to B MM MM Layer-3 1/10 GigE 1/10 GigE Dark fiber Baylor 10 GigE 1/10 GigE MM MM Layer-3 GigE Rice MM PGH 6XX 15 TB PGH 576 SR-3 1xx 10 GigE rackmount PGH 216 SM MM Layer-2 GigE PGH 660 E-Science: Data Gathering, Analysis, Simulation, and Collaboration Simulated Higgs Decay CMS LHC Optical Communication costs Larry Roberts, Caspian Networks 1000 8KB Average Price of Storage DRAM 32KB 100 Price/MByte, Dollars In 2010, $1 will buy enough disk space to store Seagate ST500 Wren II 10 1 Range 0.1 0.01 HDD 64KB DRAM oemprc2000aa.prz Paper/Film 256KB Flash 512KB Flash 2MB 1MB 4MB Flash 1MB Flash 4MB IBM0615 64MB Flash 16MB Flash Maxt170 96 MB Flash Camera 128MB Flash Mem. 128MB Flash 64MB IBM0663 IBM 340 MB Microdrive Flash 64MB IBM 9.1GB Ultrastar IBM 1 GB Microdrive Seagate B'cuda4 IBM 8.1GB Flash Projection IBM Deskstar3 of Paper/Film Travelstar DataQuest 2000 IBM Deskstar4 Toshiba Quant 4.5GB 1" HDD Projection IBM 25GB IBM 18.2GB Ultrastar 6.4GB DataQuest 2000 IBM 16.8GB Deskstar Travelstar Seagate 8.6GB IBM 9.1GB Ultra 2XP IBM Deskstar 25GB 1 " HDD IBM Deskstar 37GB IBM Deskstar 75GXP 3.5 " HDD 0.001 1980 Flash 128KB IBM6150 128KB Flash 512KB Seagate ST125 1985 1990 1995 2000 2.5 " HDD 2005 2010 Year Ed Grochowski at Almaden 10,000 Books 35 hrs of CD Quality audio 2 min of DVD Quality Video Functions per chip, Mtransistors 100000 Cost of Computing 80000 DRAM, at Production 70000 DRAM, at Introduction 60000 50000 40000 Today’s most powerful computers (the power of 10,000 PCs at a cost of $100M) will cost a few hundred thousand dollars 30000 20000 MPU, Cost-Performance, at Production MPU, Cost-Performance, at Introduction MPU, High-Performance, at Production ASIC, at Production 10000 0 1999 180 nm 2000 2001 2002 130 nm . 2003 . 2004 2005 90 nm . 2008 2011 2014 60 nm 40 nm 30 nm Cost/Mtransistor $/Mtransistor In 2010, the compute power of today’s top-of-the-line PC can be found in $1 consumer electronics SIA Roadmap 90000 45 40 35 30 25 20 15 10 5 0 SIA Roadmap DRAM, cost x 100, at Introduction, $/Mtransistors DRAM, cost x 100, at Production, $/Mtransistors 1999 2001 2003 2005 2011 . 100 40 nm 130 180 nm nm nm MPU, Cost-Performance, at Introduction, $/Mtransistors MPU, Cost-Performance, at Production, $/Mtransistors MPU, High Performance, at Production, $/Mtransistors GUSTO tesbed for Grid applications demonstrated at Supercomputing97 exhibition Tele-Microscopy Osaka, Japan Mark Ellisman, UCSD www.givenimaging.com Medicine On-line Ê New Sensors—Israeli Video Pill – Battery, Light, & Video Camera – Images Stored on Hip Mounted Device Ê Next Step—Putting You On-Line! – Wireless Internet Transmission – Key Metabolic and Physical Variables Ê Post-Genomic Individualized Medicine – Combine • Genetic Code • Body Data Flow – Use Powerful AI Data Mining Techniques www.bodymedia.com BIOMEDICAL INFORMATICS RESEARCH NETWORK (BIRN) The BIRN is an NCRR initiative aimed at creating a testbed to address biomedical researchers' need to access and analyze data at a variety of levels of aggregation located at diverse sites throughout the country. The BIRN testbed will bring together hardware and develop software necessary for a scalable network of databases and computational resources. Issues of user authentication, data integrity, security, and data ownership will also be addressed. $20M. A NATIONAL DIGITAL MAMMOGRAPHY ARCHIVE Digital Mammography 32.5 million mammograms/yr in the US. Resolution 50 micron 5000 x 4000 = 20,000,000 pixels Dynamic range: 12 bits per pixel Uncompressed image: 40 Mbyte (16 bpp) Four images per patient: 160 Mbytes Total: 5 Pbytes/yr, uncompressed Lossless compression: reduction by a factor of 2 – 3 Lossy compression: reduction by up to a factor of 80 National Ecological Observatory Network $ 100M FY01-06 NSF support requested Ê 10 observatories nationwide; sponsored by NSF, Archbold Biological Station and SDSC Ê 10 geographically distributed observatories nationwide to serve as national research platforms for integrated, cutting-edge research in field biology Ê To enable scientists to conduct experiments on ecological systems at all levels of biological organization – from molecular genetics to whole ecosystems, and across scales – from seconds to geological time, and from microns to regions and continents. Ê Observatories will have scalable computation capabilities and will be networked via satellite and landlines – to each other and to specialized facilities, such as supercomputer centers. Ê By creating one virtual installation via a cutting-edge computational network, all http://www.sdsc.edu/NEON/ members of the field biology research community will be able to access NEON remotely. Large Hadron Collider Large Hadron Collider (LHC) $ 80.9M FY99-04 NSF support requested Ê CERN (Switzerland) and international Consortium Ê Construction of two detectors of the LHC: ATLAS (A Toroidal Large Angle Spectrometer) and CMS (Compact Muon Solenoid) Ê The research, design, and prototyping of Petascale Virtual Data Grids, which will support the LHC as well as the SDSS (Sloan Digital Sky Survey) and LIGO (Laser Interferometer Gravitational-wave Observatory), is being carried out by GriPhyN, a multi- institutional team that received the largest NSF ITR grant in FY00. http://lhc.web.cern.ch/lhc/ National Data Observatory The National Virtual Observatory (NVO) will unite astronomical databases of many earthbound and orbital observatories, taking advantage of the latest computer technology and data storage and analysis techniques. The goal is to maximize the potential for new scientific insights from the data by making them available in an accessible, seamlessly unified form to professional researchers, amateur astronomers and students. $10 M. http://us-vo.org Alliance98 Interactive Collaborative Virtual Environment http://www.pdc.kth.se/projects/alliance98/ The Grid: Computational Steering GEMSviz at iGRID 2000 Paralle ldatorc ent rum KTH Stockholm INET NORDUnet STAR TAP APAN Universityof Hous ton http://www.pdc.kth.se/projects/gemsviz NORDUnet 2000 Erik Engquist 28 Sep 00 - #17 SimDB Simulation Data Base Grid Application Development Software (GrADS) Grids – Resource Allocation Grids – Overhead Cactus – Migration example Processor-DRAM Memory Gap (latency) CPU “Moore’s Law” 10 DRAM 1 µProc 60%/yr. (2X/1.5yr) Processor-Memory Performance Gap: (grows 50% / year) 100 1980 1981 1982 1983 1984 1985 1986 1987 1988 1989 1990 1991 1992 1993 1994 1995 1996 1997 1998 1999 2000 2001_ _ Performance 1000 Time DRAM 9%/yr. (2X/10 yrs) UHFFT Architecture UHFFT Library Unparser Library of FFT Modules Initialization Routines FFT Code Generator Mixed-Radix (Cooly-Tukey) Execution Routines Prime Factor Algorithm Utilities Split-Radix Algorithm Scheduler Key: Optimizer Initializer (Algorithm Abstraction) Fixed library code Generated code Code generator Rader's Algorithm Performance Tuning Methodology Input Parameters System specifics, User options UHFFT Code generator Input Parameters Size, dim., … Library of FFT modules Initialization Select best plan Performance database Execution Calculate one or more FFTs Installation Run-time
Similar documents
Reflections on a Aberrant Mirror
Elect. Commerce New Mkts Mergers&Acq. Int. Ops Other
More information