SlideShare une entreprise Scribd logo
1  sur  23
Télécharger pour lire hors ligne
1


                 August 2, 2010

Blue Gene
 Active Storage

HEC FSIO 2010
Blake G. Fitch         Alex Rayshubskiy
IBM Research           T.J. Chris Ward
                       Mike Pitman
                       Bernard Metzler
bgf@us.ibm.com         Heiko J. Schick
                       Benjamin Krill
                       Peter Morjan
                       Robert S. Germain




                 © 2010 IBM Corporation    © 2002 IBM Corporation
‘Exascale’ Architectures Motivate Active Storage
    Exascale: Three orders of magnitude more of everything?
     –Often taken to mean just ‘exaflop’ – but that’s the easy part!
     –What about the rest of the Exascale ecosystem: storage, network bandwidth, usability?


    Exascale observations:
     –The wetware will not be three orders of magnitude more capable
       • Essential to think about usability. Massive concurrency must be made more usable.
       • Non-heroic programmers and end-users must be able to utilize exascale architectures
     –Data objects will be too large to move -- between or even within datacenters
       • Today, terabyte objects have people carrying around portable disk arrays – this is not scalable.
       • Bandwidth costs not likely to drop 103 either inside out outside the data center
     –Single program (MPI) use-case alone will be insufficient for to exploit exascale capability
       • Anticipate applications formed of composed components rather than just heroic monoliths
       • Some capability created by exascale will be spent enabling interfaces to support compos ability
     –“Tape is Dead, Disk is Tape, Flash is Disk, RAM locality is King” – Jim Gray
       • A rack of disks can deliver ~10GB/s. 1TB/s requires 100 racks or ~$100M
       • Mixing archival storage and online storage will be less efficient in the future
       • Solid state storage and storage class memory – but how will these technologies be used?



    Scalable Bisectional bandwidth will be THE challenge both to create and exploit
2                 Blue Gene Active Storage   HEC FSIO 2010                                      © 2010 IBM Corporation
From K. Borne, Data Science Challenges from Distributed Petabyte Astronomical Data Collections:
    Preparing for the Data Avalanche through Persistence, Parallelization, and Provenance,
    Computing with Massive and Persistent Data (CMPD 08)
    https://computation.llnl.gov/home/workshops/cmpd08/Presentations/Borne.pdf

3          Blue Gene Active Storage                 HEC FSIO 2010                                     © 2010 IBM Corporation
Data-centric and Network-centric Application Classes
    Embarrassingly Parallel                             Network Dependent
     Table scan                                          Join
     Batch serial                                        Sort
      –Floating point intensive                           –“order by” queries
      –Job farm/cloud                                    “group by” queries
     Map-Reduce                                          Map-Reduce
     Other applications without                          Aggregation operations
     internal global data
                                                          –count(), sum(), min(), max(), avg(), …
     dependencies
                                                         Data analysis/OLAP
                                                          –Aggregation with “group by”…
                                                          –Real-time analytics
                                                         HPC applications


4            Blue Gene Active Storage   HEC FSIO 2010                             © 2010 IBM Corporation
Blue Gene Architecture in Review
Blue Gene is not just FLOPs…




                                                  … it’s also torus network,
                                                  power efficiency, and dense
                                                  packaging.

                                                  Focus on scalability rather
                                                  than on configurability
                                                  gives the Blue Gene family’s
                                                  System-on-a-Chip
                                                  architecture unprecedented
                                                  scalability and reliability.
5            Blue Gene Active Storage   HEC FSIO 2010                            © 2010 IBM Corporation
Blue Gene Architectural Advantages
           Main Memory Capacity per Rack                                          Peak Memory Bandwidth per node (byte/flop)
    4500
                                                                                         SGI ICE
    4000
    3500                                                                               Sun TACC
    3000
                                                                                        Itanium 2
    2500
    2000                                                                          Cray XT5 4 core
    1500
    1000                                                                          Cray XT3 2 core

     500                                                                             BG/P 4 core
       0
            LRZ        Cray     ASC        BG/P        Sun        SGI ICE
                                                                                                    0       0.5          1          1.5
            IA64       XT4     Purple                 TACC


     Relative power, space and cooling efficiencies                                                     Low Failure Rates
                                                                                    http://acts.nersc.gov/events/Workshop2006/slides/Simon.pdf

400%

300%

200%

100%                              IBM BG/P

    0%
            Racks/TF          kW/TF          Sq Ft/TF             Tons/TF

                       Sun/Constellation   Cray/XT4     SGI/ICE



6                        Blue Gene Active Storage                 HEC FSIO 2010                                               © 2010 IBM Corporation
Blue Gene Torus: A cost effective all-to-all network
     10GbE network costs from: Data Center Switch Architecture In The Age Of Merchant Silicon
                                     (Farrington, et al, 2009 http://cseweb.ucsd.edu/~vahdat/papers/hoti09.pdf)




           Theoretical 3D and 4D Torus All-to-all throughput per node for 1GB/s Links (~1/2 10GbE)
                            3


                          2 .5
     A2A BW/Node (GB/s)




                            2


                          1 .5


                            1


                          0 .5
                                        3 D -T o ru s
                                        4 D -T o ru s
                            0
                                 0                 5000            10000             15000           20000   25000         30000
                                                                                      N ode C ount

 7                                      Blue Gene Active Storage     HEC FSIO 2010                                   © 2010 IBM Corporation
Blue Gene Data-centric Micro-benchmark Results
                                  Random Loads Per Second                      100

    Blue Gene/L micro benchmarks show                                           10             16,384 Nodes
    order of magnitude improvement over
    largest SMP and clusters




                                                         Giga-Loads/Second
                                                                                 1


                                                                                0.1


                                                                               0.01

    Scalable random pointer chasing                                           0.001
    through terabytes of memory                                                                                                          Migrate Context
                                                                                                                                            Memory Get
                                                                             0.0001

     –BG/L 1.0 GigaLoads / Rack                                                       1       10         100          1000
                                                                                                               Total Read Contexts
                                                                                                                                     10000       100000    1e+006




       • 14 GigaLoads/s for 16 racks
     –P690 0.3 Gigaloads                                                                       Strong Scaling vs. Node Count
                                                                             120
                                                                                          1 TeraByte
    Scalable “Indy” sort of terabytes in                                                  2 TeraByte
                                                                                          3 TeraByte
                                                                             100
    memory                                                                                4 TeraByte
                                                                                          5 TeraByte


                                                        Sort Time (sec.)
                                                                              80          6 TeraByte
     –BG/L 1.0 TB sort in 10 secs                                                         7 TeraByte

                                                                              60
     –BG/L 7.5 TB sort in 100 secs
                                                                              40
    Streams benchmark
                                                                              20
     –BG/L      2.5TB / Sec / Rack
                                                                                0
     –P5 595 0.2TB / Sec / Rack                                                 4000                          8000                           16000
                                                                                                               Node Count
8            Blue Gene Active Storage   HEC FSIO 2010                                                                                   © 2010 IBM Corporation
Blue Gene/P for I/O and Active Storage Explorations
Currently a research platform for scalable data-centric
    computing and BG/Q I/O system prototyping
      Linux 2.6.29 kernel, 32-bit powerpc
      4GB RAM, diskless, 4-way SMP, 1024
      nodes/rack
      4096 nodes available on site
      IP-over-torus to other compute nodes
      IO nodes as IP routers for off-fabric traffic
      MPICH/TCP, OpenMPI/TCP, SLURM, TCP
      sockets
      Running SoftiWARP for OFED RDMA
      (http://gitorious.org/softiwarp)
      Native OFED verbs provider for Blue Gene
      Torus has achieved initial function
      Experimental GPFS version for on-board Parallel
      File System explorations at very high node count
      Testing MVAPICH and OpenMPI


Future – anticipate moving ASF over to and
    integrating with Argonne National Lab’s
    ZeptoOS (http://www.zeptoos.com)



  9                    Blue Gene Active Storage       HEC FSIO 2010   © 2010 IBM Corporation
Blue Gene Active Storage Opportunity – Current Conditions
IBM Enterprise Server
   (Analysis Here)




                                                           IBM System Storage
                                                                 DS8000
  IBM BG/P HPC Solution
 (Scalable Compute Here)
                                          widening gap
                                                                            IBM 3494 Tape Library




                                                               Persistent
                                                                Storage

10             Blue Gene Active Storage    HEC FSIO 2010                            © 2010 IBM Corporation
Blue Gene Active Storage adds parallel processing into the storage hierarchy.
IBM Enterprise Server




                             10
                                            Blue Gene/AS
                               0s              Option
                                    GB
                                       /s



                                                                 IBM System Storage
                                                                       DS8000
      Application
       Solution
        Driver

                                                                                  IBM 3494 Tape Library
                                         Managed as
                                          Scalable
                                     Solid State Storage
                                             with
                                     Embedded Parallel               Persistent
                                         Processing                   Storage

 11                 Blue Gene Active Storage     HEC FSIO 2010                            © 2010 IBM Corporation
Thought Experiment: A Blue Gene Active Storage Machine
      Integrate significant storage class memory (SCM) at each node
         –   For now, Flash memory, maybe similar function to Fusion-io ioDrive Duo
         –   Future systems may deploy Phase Change Memory (PCM), Memristor, or…?
         –   Assume node density will drops 50% -- 512 Nodes/Rack for embedded apps                         ioDrive Duo         One Board   512 Node
         –   Objective: balance Flash bandwidth to network all-to-all throughput                            SLC NAND Cap.       320 GB      160 TB

                                                                                                            Read BW (64K)       1450 MB/s   725 GB/s

      Resulting System Attributes:                                                                          Write BW (64K)      1400 MB/s   700 GB/s

         –   Rack: 0.5 petabyte, 512 Blue Gene processors, and embedded torus network                       Read IOPS (4K)      270,000     138 Mega

         –   700 TB/s I/O bandwidth to Flash – competitive with ~70 large disk controllers                  Write IOPS (4K)     257,000     131 Maga
                •   Order of magnitude less space and power than equivalent perf via disk solution          Mixed R/W
                                                                                                                                207,000     105 Mega
                •   Can configure fewer disk controllers and optimize them for archival use                 IOPs(75/25@4K)




                                                                                                     51
         –   With network all-to-all throughput at 1GB/s per node, anticipate:




                                                                                                       2x
                •   1TB sort from/to persistent storage in order 10 secs.
                •   130 Million IOPs per rack, 700 GB/s I/O bandwidth
         –   Inherit Blue Gene attributes: scalability, reliability, power efficiency,


      Research Challenges (list not exhaustive):
         –   Packaging – can the integration succeed?
         –   Resilience – storage, network, system management, middleware
         –   Data management – need clear split between on-line and archival data
         –   Data structures and algorithms can take specific advantage of the BGAS
             architecture – no one cares it’s not x86 since software is embedded in storage



      Related Work:
         –   Gordon (UCSD) http://nvsl.ucsd.edu/papers/Asplos2009Gordon.pdf
         –   FAWN (CMU) http://www.cs.cmu.edu/~fawnproj/papers/fawn-sosp2009.pdf
         –   RamCloud (Stanford) http://www.stanford.edu/~ouster/cgi-bin/papers/ramcloud.pdf

 12
 12                    Blue Gene Active Storage               HEC FSIO 2010                                                   © 2010 IBM Corporation
Blue Gene Active Storage for Server Acceleration
     Manage Blue Gene hardware as a scalable, solid state                          P Servers          Z servers
     storage device with embedded processing capability


     Integrate this “Active Storage Fabric” with middleware                Data Sets   Blue Gene/ASF
     such as DB2, MySql, GPFS, etc, at the data/storage
     interface using a parallel, key-value in-memory data
     store
                                                                                         Disk Array
     Transparently accelerate server Jobs with ASF:
      –   Create libraries of reusable, low overhead
          Embedded Parallel Modules (EPMs) (scan, sort, join, sum, etc)
      –   EPMs directly access middleware data (tables, files, etc) in                 Robot Tape
          key/value DB based on middleware data model
      –   Middleware makes ASF datasets available via legacy interfaces
          allowing incremental parallelization and interoperability with
          legacy middleware function and host programs

                                                                                   Shelf Tape
     Integrate BG/ASF with information life-cycle and
     other standard IT data management products
13                 Blue Gene Active Storage   HEC FSIO 2010                                    © 2010 IBM Corporation
Blue Gene Active Storage for UNIX Utility Acceleration
                                                                            GPFS servers access file system data
                                                                            blocks in a single ‘disk’ in Active Store
      Huge
                                                 Blue Gene
                                                                            Parallel In-Memory Database accesses
                               CNTL         Active Storage Fabric
     Analytics                                                              block with key of [file_num,block_num]
      Server                  Active
                              GPFS                                          Embedded Parallel Modules (EPMs)
                              RDMA
                                                                            access file data directly with PIMD cleint
                              Router                                        asf_generate_indy_data an EPM
                              RDMA                                          creates Indy.Input synthetic data
                              Router
                                               BG/AS Nodes
                              Active          Linux IP/RDMA                 asf_grep EPM access Indy.Input file
                              GPFS                                          and greps all text records creating
                              RDMA                                          Indy.Temp
                                               PIMD Server
                              Router                                        asf_sort EPM sorts Indy.Temp file and
                              RDMA                                          creates Indy.Output
                              Router          EPMs, ex: grep
                                                                            Tail -1 is a normal unix command which
                           Non-heroic                                       runs on the workstation and gets the
                           End-User                                         last line in the last block of the
                                                                            Indy.Output file
            Active
            GPFS                                         asf_user@asf_host-> cd /active_gpfs
            Client                                       asf_user@asf_host-> asf_generate_indy_data 1TB > Indy.Input
                                                         asf_user@asf_host-> asf_grep “aaaaaa” Indy.Input > Indy.Temp
                                                         asf_user@asf_host-> asf_sort < Indy.Temp >Indy.Output
                                                         asf_user@asf_host-> tail -1 Indy.Output
                                                         9876543210aaaaaa………………………………. … …
                                                         asf_user@asf_host-> rm Indy.Output Indy.Input

14               Blue Gene Active Storage    HEC FSIO 2010                                           © 2010 IBM Corporation
Questions?

     Note: If you are NDA’d with IBM for Blue Gene/Q or are a
     US Government employee and you are interested in the
     confidential version of the Blue Gene/Q Active Storage
     presentation, please contact me.


     Blake Fitch
     bgf@us.ibm.com




15    Blue Gene Active Storage    HEC FSIO 2010                 © 2010 IBM Corporation
Backup




16   Blue Gene Active Storage    HEC FSIO 2010   © 2010 IBM Corporation
Blue Gene/L ASF: Early Performance Comparison with Host Commands
     Benchmark process*:
     • Generate a 22 GB synthetic “Indy Benchmark” data file
     • grep for records with “AAAAAA” (outputs 1/3’d of the data which is used as input to sort)
     • sort the output of grep

      * NOTE: All ASF data taken on untuned system, non-optimized and with full debug printfs, all p55 data taken with unix time on command line)


                                          Function                                                           BG/ASF                   pSeries p55
       (Measured at Host shell and break out of EPM components)                                             512 nodes                   (1 thread)
                                   Total host command time: create Indy File:                              (22GB) 22.26 s                           686s
                          Total host command time: unix grep of Indy File:                                               25.1s                  3180s
                               File Blocks to PIMD Records (Embedded grep) :                                              7.0 s
                                             Grep core function (Embedded grep) :                                          2.5s
                               PIMD Records to File Blocks (Embedded grep) :                                               7.6s
                      Total time on Blue Gene/ASF fabric (Embedded grep) :                                               18.8s
                             Total host command time: sort output of grep:                                             60.41s                   2220s
                                File Blocks to PIMD Records (Embedded sort) :                                              2.9s
                                                 Sort core function (Embedded sort) :                                    12.2s
                                PIMD Records to File Blocks (Embedded sort) :                                            15.8s
                       Total time on Blue Gene/ASF fabric (Embedded sort) :                                              54.7s

17                    Blue Gene Active Storage          HEC FSIO 2010                                                           © 2010 IBM Corporation
Blue Gene/L ASF: Early Scalability Results
     Benchmark process*:
     • Generate a 22 GB synthetic “Indy Benchmark” data file
     • grep for records with “AAAAAA” (outputs 1/3’d of the data which is used as input to sort)
     • sort the output of grep

      * NOTE: All ASF data taken on untuned system, non-optimized and with full debug printfs, all p55 data taken with unix time on command line)


                                          Function                                                           BG/ASF                      BG/ASF
                                                                                                            512 nodes                   8K nodes
                                   Total host command time: create Indy File:                              (22GB) 22.26 s                (1TB) 197s
                          Total host command time: unix grep of Indy File:                                               25.1s                      107s
                               File Blocks to PIMD Records (Embedded grep) :                                              7.0 s                     18.1s
                                             Grep core function (Embedded grep) :                                          2.5s                      5.1s
                               PIMD Records to File Blocks (Embedded grep) :                                               7.6s                     19.9s
                      Total time on Blue Gene/ASF fabric (Embedded grep) :                                               18.8s                      50.2s
                             Total host command time: sort output of grep:                                             60.41s                       120s
                                File Blocks to PIMD Records (Embedded sort) :                                              2.9s                      6.8s
                                                 Sort core function (Embedded sort) :                                    12.2s                      14.8s
                                PIMD Records to File Blocks (Embedded sort) :                                            15.8s                      22.4s
                       Total time on Blue Gene/ASF fabric (Embedded sort) :                                              54.7s                 66.55s

18                    Blue Gene Active Storage          HEC FSIO 2010                                                           © 2010 IBM Corporation
Parallel In-Memory Database (PIMD) 512 Nodes




19       Blue Gene Active Storage   HEC FSIO 2010   © 2010 IBM Corporation
Applications with potentially strong data dependencies
Application                                  Application Class/Trends                 Algorithms

National Security/Intelligence, Social       Graph-related                            Breadth first search, ….
Networks
Molecular simulation                         N-body                                   Spectral methods (FFT)
Financial risk, pricing, market              Real time control, shift from forensic   Graph traversal, Monte Carlo, …
making                                       to real time
Oil exploration                              Inverse problems                         Reverse time migration, finite
                                                                                      difference
Climate modeling, weather                    Fluid dynamics                           Finite difference, spectral methods,
prediction                                                                            …
Recommendation systems                       Interactive query                        Minimization (least squares,
                                                                                      conjugate gradient, …)
Network simulations                                                                   Discrete event
Traffic management                           Real time control                        Real time optimization
Business intelligence/analytics              Data mining/association discovery        K-means clustering, aggregation,
                                                                                      scans, ….
Data deduplication                           Storage management                       Hashing
Load balancing/meshing (cross-                                                        Graph partitioning
domain)
Multi-media/unstructured search              Search                                   MapReduce, hashing, clustering


20                Blue Gene Active Storage       HEC FSIO 2010                                             © 2010 IBM Corporation
BG/Q Active Storage for Expanding Flash Capacity
       1.00E+13




       1.00E+12
                       Next two years                                                 DRAM raw density (bits/cm2)

                                                                                      FLASH raw density (bits/cm2)

                                                                                      PCM raw density (bits/cm2)
       1.00E+11

                                                                                      possible DRAM limit

                                                                                      Pessimistic PCM (2 bits, 2
                                                                                      layers only)
       1.00E+10




       1.00E+09
              2007     2009     2011     2013   2015   2017      2019   2021   2023

ITRS 2008 data
  21                 Blue Gene Active Storage    HEC FSIO 2010                                      © 2010 IBM Corporation
Related Data Intensive Supercomputing Efforts
•Hyperion (LLNL) http://www.supercomputingonline.com/latest/appro-deploys-a-world-class-linux-cluster-testbed-solution-to-llnl-in-support-of-the-hyperion-project   l



•The Linux cluster solution consists of 80 Appro 1U 1423X, dual socket customized server based on Intel Xeon processor 5600
series with 24 GB of memory and 4 PCIe2 x8 slots. The cluster will be integrated with 2x ioSAN carrier cards with 320 GB of
FLASH memory and one 10Gb/s Ethernet and one IBA QDR (40Gb/s) link. The I/O nodes also have 2x 1Gb/s Ethernet links for
low speed cloud file system I/O and two SATA disks for speed comparisons. In addition to these IO nodes, 28 Lustre OSS 1U
Appro 1U 1423X dual socket nodes with 24 GB of memory, one Mellanox Connect-X card, 10Gb/s Ethernet and QDR IBA
(40Gb/s) links and 2 SATA drives will be added. The entire solution will provide 80 Appro server nodes powered by
Fusion-io technology delivering an aggregate 4KB IOPS rate of over 40 million IOPS and 320 GB/s of IO bandwidth and
102 TB of FLASH memory capabilities.


•Flash Gordon (SDSC) http://storageconference.org/2010/Presentations/MSST/10.Norman.pdf




 22                        Blue Gene Active Storage              HEC FSIO 2010                                                                   © 2010 IBM Corporation
Failures per Month per TF
     From:http://acts.nersc.gov/events/Workshop2006/slides/Simon.pdf




23      Blue Gene Active Storage   HEC FSIO 2010             © 2010 IBM Corporation

Contenu connexe

Tendances

A64fx and Fugaku - A Game Changing, HPC / AI Optimized Arm CPU to enable Exas...
A64fx and Fugaku - A Game Changing, HPC / AI Optimized Arm CPU to enable Exas...A64fx and Fugaku - A Game Changing, HPC / AI Optimized Arm CPU to enable Exas...
A64fx and Fugaku - A Game Changing, HPC / AI Optimized Arm CPU to enable Exas...inside-BigData.com
 
01 From K to Fugaku
01 From K to Fugaku01 From K to Fugaku
01 From K to FugakuRCCSRENKEI
 
Intro to Cell Broadband Engine for HPC
Intro to Cell Broadband Engine for HPCIntro to Cell Broadband Engine for HPC
Intro to Cell Broadband Engine for HPCSlide_N
 
08 Supercomputer Fugaku
08 Supercomputer Fugaku08 Supercomputer Fugaku
08 Supercomputer FugakuRCCSRENKEI
 
AI is Impacting HPC Everywhere
AI is Impacting HPC EverywhereAI is Impacting HPC Everywhere
AI is Impacting HPC Everywhereinside-BigData.com
 
The Tofu Interconnect D for the Post K Supercomputer
The Tofu Interconnect D for the Post K SupercomputerThe Tofu Interconnect D for the Post K Supercomputer
The Tofu Interconnect D for the Post K Supercomputerinside-BigData.com
 
Hardware architecture of Summit Supercomputer
 Hardware architecture of Summit Supercomputer Hardware architecture of Summit Supercomputer
Hardware architecture of Summit SupercomputerVigneshwarRamaswamy
 
Petascale Analytics - The World of Big Data Requires Big Analytics
Petascale Analytics - The World of Big Data Requires Big AnalyticsPetascale Analytics - The World of Big Data Requires Big Analytics
Petascale Analytics - The World of Big Data Requires Big AnalyticsHeiko Joerg Schick
 
Experiences in Application Specific Supercomputer Design - Reasons, Challenge...
Experiences in Application Specific Supercomputer Design - Reasons, Challenge...Experiences in Application Specific Supercomputer Design - Reasons, Challenge...
Experiences in Application Specific Supercomputer Design - Reasons, Challenge...Heiko Joerg Schick
 
04 New opportunities in photon science with high-speed X-ray imaging detecto...
04 New opportunities in photon science with high-speed X-ray imaging  detecto...04 New opportunities in photon science with high-speed X-ray imaging  detecto...
04 New opportunities in photon science with high-speed X-ray imaging detecto...RCCSRENKEI
 
High Performance Computing - Challenges on the Road to Exascale Computing
High Performance Computing - Challenges on the Road to Exascale ComputingHigh Performance Computing - Challenges on the Road to Exascale Computing
High Performance Computing - Challenges on the Road to Exascale ComputingHeiko Joerg Schick
 
NNSA Explorations: ARM for Supercomputing
NNSA Explorations: ARM for SupercomputingNNSA Explorations: ARM for Supercomputing
NNSA Explorations: ARM for Supercomputinginside-BigData.com
 
Exceeding the Limits of Air Cooling to Unlock Greater Potential in HPC
Exceeding the Limits of Air Cooling to Unlock Greater Potential in HPCExceeding the Limits of Air Cooling to Unlock Greater Potential in HPC
Exceeding the Limits of Air Cooling to Unlock Greater Potential in HPCinside-BigData.com
 
Multiple Cores, Multiple Pipes, Multiple Threads – Do we have more Parallelis...
Multiple Cores, Multiple Pipes, Multiple Threads – Do we have more Parallelis...Multiple Cores, Multiple Pipes, Multiple Threads – Do we have more Parallelis...
Multiple Cores, Multiple Pipes, Multiple Threads – Do we have more Parallelis...Slide_N
 
Hardware and Software Architectures for the CELL BROADBAND ENGINE processor
Hardware and Software Architectures for the CELL BROADBAND ENGINE processorHardware and Software Architectures for the CELL BROADBAND ENGINE processor
Hardware and Software Architectures for the CELL BROADBAND ENGINE processorSlide_N
 

Tendances (20)

Blue gene
Blue geneBlue gene
Blue gene
 
BLUE GENE/L
BLUE GENE/LBLUE GENE/L
BLUE GENE/L
 
Bluegene
BluegeneBluegene
Bluegene
 
Bluegene
BluegeneBluegene
Bluegene
 
A64fx and Fugaku - A Game Changing, HPC / AI Optimized Arm CPU to enable Exas...
A64fx and Fugaku - A Game Changing, HPC / AI Optimized Arm CPU to enable Exas...A64fx and Fugaku - A Game Changing, HPC / AI Optimized Arm CPU to enable Exas...
A64fx and Fugaku - A Game Changing, HPC / AI Optimized Arm CPU to enable Exas...
 
01 From K to Fugaku
01 From K to Fugaku01 From K to Fugaku
01 From K to Fugaku
 
Intro to Cell Broadband Engine for HPC
Intro to Cell Broadband Engine for HPCIntro to Cell Broadband Engine for HPC
Intro to Cell Broadband Engine for HPC
 
08 Supercomputer Fugaku
08 Supercomputer Fugaku08 Supercomputer Fugaku
08 Supercomputer Fugaku
 
AI is Impacting HPC Everywhere
AI is Impacting HPC EverywhereAI is Impacting HPC Everywhere
AI is Impacting HPC Everywhere
 
The Tofu Interconnect D for the Post K Supercomputer
The Tofu Interconnect D for the Post K SupercomputerThe Tofu Interconnect D for the Post K Supercomputer
The Tofu Interconnect D for the Post K Supercomputer
 
POWER10 innovations for HPC
POWER10 innovations for HPCPOWER10 innovations for HPC
POWER10 innovations for HPC
 
Hardware architecture of Summit Supercomputer
 Hardware architecture of Summit Supercomputer Hardware architecture of Summit Supercomputer
Hardware architecture of Summit Supercomputer
 
Petascale Analytics - The World of Big Data Requires Big Analytics
Petascale Analytics - The World of Big Data Requires Big AnalyticsPetascale Analytics - The World of Big Data Requires Big Analytics
Petascale Analytics - The World of Big Data Requires Big Analytics
 
Experiences in Application Specific Supercomputer Design - Reasons, Challenge...
Experiences in Application Specific Supercomputer Design - Reasons, Challenge...Experiences in Application Specific Supercomputer Design - Reasons, Challenge...
Experiences in Application Specific Supercomputer Design - Reasons, Challenge...
 
04 New opportunities in photon science with high-speed X-ray imaging detecto...
04 New opportunities in photon science with high-speed X-ray imaging  detecto...04 New opportunities in photon science with high-speed X-ray imaging  detecto...
04 New opportunities in photon science with high-speed X-ray imaging detecto...
 
High Performance Computing - Challenges on the Road to Exascale Computing
High Performance Computing - Challenges on the Road to Exascale ComputingHigh Performance Computing - Challenges on the Road to Exascale Computing
High Performance Computing - Challenges on the Road to Exascale Computing
 
NNSA Explorations: ARM for Supercomputing
NNSA Explorations: ARM for SupercomputingNNSA Explorations: ARM for Supercomputing
NNSA Explorations: ARM for Supercomputing
 
Exceeding the Limits of Air Cooling to Unlock Greater Potential in HPC
Exceeding the Limits of Air Cooling to Unlock Greater Potential in HPCExceeding the Limits of Air Cooling to Unlock Greater Potential in HPC
Exceeding the Limits of Air Cooling to Unlock Greater Potential in HPC
 
Multiple Cores, Multiple Pipes, Multiple Threads – Do we have more Parallelis...
Multiple Cores, Multiple Pipes, Multiple Threads – Do we have more Parallelis...Multiple Cores, Multiple Pipes, Multiple Threads – Do we have more Parallelis...
Multiple Cores, Multiple Pipes, Multiple Threads – Do we have more Parallelis...
 
Hardware and Software Architectures for the CELL BROADBAND ENGINE processor
Hardware and Software Architectures for the CELL BROADBAND ENGINE processorHardware and Software Architectures for the CELL BROADBAND ENGINE processor
Hardware and Software Architectures for the CELL BROADBAND ENGINE processor
 

Similaire à Blue Gene Active Storage

Emc isilon technical deep dive workshop
Emc isilon technical deep dive workshopEmc isilon technical deep dive workshop
Emc isilon technical deep dive workshopsolarisyougood
 
Lug best practice_hpc_workflow
Lug best practice_hpc_workflowLug best practice_hpc_workflow
Lug best practice_hpc_workflowrjmurphyslideshare
 
OMI - The Missing Piece of a Modular, Flexible and Composable Computing World
OMI - The Missing Piece of a Modular, Flexible and Composable Computing WorldOMI - The Missing Piece of a Modular, Flexible and Composable Computing World
OMI - The Missing Piece of a Modular, Flexible and Composable Computing WorldAllan Cantle
 
Infoboom future-storage-aug2011-v3
Infoboom future-storage-aug2011-v3Infoboom future-storage-aug2011-v3
Infoboom future-storage-aug2011-v3Tony Pearson
 
Infoboom future-storage-aug2011-v3
Infoboom future-storage-aug2011-v3Infoboom future-storage-aug2011-v3
Infoboom future-storage-aug2011-v3Tony Pearson
 
IoT関連技術の動向@IETF87
IoT関連技術の動向@IETF87IoT関連技術の動向@IETF87
IoT関連技術の動向@IETF87Shoichi Sakane
 
How to Terminate the GLIF by Building a Campus Big Data Freeway System
How to Terminate the GLIF by Building a Campus Big Data Freeway SystemHow to Terminate the GLIF by Building a Campus Big Data Freeway System
How to Terminate the GLIF by Building a Campus Big Data Freeway SystemLarry Smarr
 
Future Cloud Infrastructure
Future Cloud InfrastructureFuture Cloud Infrastructure
Future Cloud Infrastructureexponential-inc
 
Enterprise power systems transition to power7 technology
Enterprise power systems transition to power7 technologyEnterprise power systems transition to power7 technology
Enterprise power systems transition to power7 technologysolarisyougood
 
High Performance Cyberinfrastructure Enabling Data-Driven Science in the Biom...
High Performance Cyberinfrastructure Enabling Data-Driven Science in the Biom...High Performance Cyberinfrastructure Enabling Data-Driven Science in the Biom...
High Performance Cyberinfrastructure Enabling Data-Driven Science in the Biom...Larry Smarr
 
High Performance Cloud Computing
High Performance Cloud ComputingHigh Performance Cloud Computing
High Performance Cloud ComputingAmazon Web Services
 
High Performance Cloud Computing
High Performance Cloud ComputingHigh Performance Cloud Computing
High Performance Cloud ComputingAmazon Web Services
 
Open Storage Sun Intel European Business Technology Tour
Open Storage Sun Intel European Business Technology TourOpen Storage Sun Intel European Business Technology Tour
Open Storage Sun Intel European Business Technology TourWalter Moriconi
 
CTI Group- Blue power technology storwize technical training for customer - p...
CTI Group- Blue power technology storwize technical training for customer - p...CTI Group- Blue power technology storwize technical training for customer - p...
CTI Group- Blue power technology storwize technical training for customer - p...Tri Susilo
 
Report to the NAC
Report to the NACReport to the NAC
Report to the NACLarry Smarr
 
IBM Cloud Object Storage: How it works and typical use cases
IBM Cloud Object Storage: How it works and typical use casesIBM Cloud Object Storage: How it works and typical use cases
IBM Cloud Object Storage: How it works and typical use casesTony Pearson
 
Architecting Next Generation Enterprise Network Storage
Architecting Next Generation Enterprise Network StorageArchitecting Next Generation Enterprise Network Storage
Architecting Next Generation Enterprise Network StorageIMEX Research
 
Oracle Cloud PaaS & IaaS:2019年12月度サービス情報アップデート
Oracle Cloud PaaS & IaaS:2019年12月度サービス情報アップデートOracle Cloud PaaS & IaaS:2019年12月度サービス情報アップデート
Oracle Cloud PaaS & IaaS:2019年12月度サービス情報アップデートオラクルエンジニア通信
 

Similaire à Blue Gene Active Storage (20)

IBM System Blue Gene/P Data Sheet
IBM System Blue Gene/P Data SheetIBM System Blue Gene/P Data Sheet
IBM System Blue Gene/P Data Sheet
 
Emc isilon technical deep dive workshop
Emc isilon technical deep dive workshopEmc isilon technical deep dive workshop
Emc isilon technical deep dive workshop
 
Lug best practice_hpc_workflow
Lug best practice_hpc_workflowLug best practice_hpc_workflow
Lug best practice_hpc_workflow
 
Emc isilon overview
Emc isilon overview Emc isilon overview
Emc isilon overview
 
OMI - The Missing Piece of a Modular, Flexible and Composable Computing World
OMI - The Missing Piece of a Modular, Flexible and Composable Computing WorldOMI - The Missing Piece of a Modular, Flexible and Composable Computing World
OMI - The Missing Piece of a Modular, Flexible and Composable Computing World
 
Infoboom future-storage-aug2011-v3
Infoboom future-storage-aug2011-v3Infoboom future-storage-aug2011-v3
Infoboom future-storage-aug2011-v3
 
Infoboom future-storage-aug2011-v3
Infoboom future-storage-aug2011-v3Infoboom future-storage-aug2011-v3
Infoboom future-storage-aug2011-v3
 
IoT関連技術の動向@IETF87
IoT関連技術の動向@IETF87IoT関連技術の動向@IETF87
IoT関連技術の動向@IETF87
 
How to Terminate the GLIF by Building a Campus Big Data Freeway System
How to Terminate the GLIF by Building a Campus Big Data Freeway SystemHow to Terminate the GLIF by Building a Campus Big Data Freeway System
How to Terminate the GLIF by Building a Campus Big Data Freeway System
 
Future Cloud Infrastructure
Future Cloud InfrastructureFuture Cloud Infrastructure
Future Cloud Infrastructure
 
Enterprise power systems transition to power7 technology
Enterprise power systems transition to power7 technologyEnterprise power systems transition to power7 technology
Enterprise power systems transition to power7 technology
 
High Performance Cyberinfrastructure Enabling Data-Driven Science in the Biom...
High Performance Cyberinfrastructure Enabling Data-Driven Science in the Biom...High Performance Cyberinfrastructure Enabling Data-Driven Science in the Biom...
High Performance Cyberinfrastructure Enabling Data-Driven Science in the Biom...
 
High Performance Cloud Computing
High Performance Cloud ComputingHigh Performance Cloud Computing
High Performance Cloud Computing
 
High Performance Cloud Computing
High Performance Cloud ComputingHigh Performance Cloud Computing
High Performance Cloud Computing
 
Open Storage Sun Intel European Business Technology Tour
Open Storage Sun Intel European Business Technology TourOpen Storage Sun Intel European Business Technology Tour
Open Storage Sun Intel European Business Technology Tour
 
CTI Group- Blue power technology storwize technical training for customer - p...
CTI Group- Blue power technology storwize technical training for customer - p...CTI Group- Blue power technology storwize technical training for customer - p...
CTI Group- Blue power technology storwize technical training for customer - p...
 
Report to the NAC
Report to the NACReport to the NAC
Report to the NAC
 
IBM Cloud Object Storage: How it works and typical use cases
IBM Cloud Object Storage: How it works and typical use casesIBM Cloud Object Storage: How it works and typical use cases
IBM Cloud Object Storage: How it works and typical use cases
 
Architecting Next Generation Enterprise Network Storage
Architecting Next Generation Enterprise Network StorageArchitecting Next Generation Enterprise Network Storage
Architecting Next Generation Enterprise Network Storage
 
Oracle Cloud PaaS & IaaS:2019年12月度サービス情報アップデート
Oracle Cloud PaaS & IaaS:2019年12月度サービス情報アップデートOracle Cloud PaaS & IaaS:2019年12月度サービス情報アップデート
Oracle Cloud PaaS & IaaS:2019年12月度サービス情報アップデート
 

Plus de Heiko Joerg Schick

Da Vinci - A scaleable architecture for neural network computing (updated v4)
Da Vinci - A scaleable architecture for neural network computing (updated v4)Da Vinci - A scaleable architecture for neural network computing (updated v4)
Da Vinci - A scaleable architecture for neural network computing (updated v4)Heiko Joerg Schick
 
Huawei empowers healthcare industry with AI technology
Huawei empowers healthcare industry with AI technologyHuawei empowers healthcare industry with AI technology
Huawei empowers healthcare industry with AI technologyHeiko Joerg Schick
 
The 2025 Huawei trend forecast gives you the lowdown on data centre facilitie...
The 2025 Huawei trend forecast gives you the lowdown on data centre facilitie...The 2025 Huawei trend forecast gives you the lowdown on data centre facilitie...
The 2025 Huawei trend forecast gives you the lowdown on data centre facilitie...Heiko Joerg Schick
 
The Smarter Car for Autonomous Driving
 The Smarter Car for Autonomous Driving The Smarter Car for Autonomous Driving
The Smarter Car for Autonomous DrivingHeiko Joerg Schick
 
From edge computing to in-car computing
From edge computing to in-car computingFrom edge computing to in-car computing
From edge computing to in-car computingHeiko Joerg Schick
 
Need and value for various levels of autonomous driving
Need and value for various levels of autonomous drivingNeed and value for various levels of autonomous driving
Need and value for various levels of autonomous drivingHeiko Joerg Schick
 
Run-Time Reconfiguration for HyperTransport coupled FPGAs using ACCFS
Run-Time Reconfiguration for HyperTransport coupled FPGAs using ACCFSRun-Time Reconfiguration for HyperTransport coupled FPGAs using ACCFS
Run-Time Reconfiguration for HyperTransport coupled FPGAs using ACCFSHeiko Joerg Schick
 
Browser and Management App for Google's Person Finder
Browser and Management App for Google's Person FinderBrowser and Management App for Google's Person Finder
Browser and Management App for Google's Person FinderHeiko Joerg Schick
 
IBM Corporate Service Corps - Helping Create Interactive Flood Maps
IBM Corporate Service Corps - Helping Create Interactive Flood MapsIBM Corporate Service Corps - Helping Create Interactive Flood Maps
IBM Corporate Service Corps - Helping Create Interactive Flood MapsHeiko Joerg Schick
 
Real time Flood Simulation for Metro Manila and the Philippines
Real time Flood Simulation for Metro Manila and the PhilippinesReal time Flood Simulation for Metro Manila and the Philippines
Real time Flood Simulation for Metro Manila and the PhilippinesHeiko Joerg Schick
 
directCell - Cell/B.E. tightly coupled via PCI Express
directCell - Cell/B.E. tightly coupled via PCI ExpressdirectCell - Cell/B.E. tightly coupled via PCI Express
directCell - Cell/B.E. tightly coupled via PCI ExpressHeiko Joerg Schick
 
QPACE QCD Parallel Computing on the Cell Broadband Engine™ (Cell/B.E.)
QPACE QCD Parallel Computing on the Cell Broadband Engine™ (Cell/B.E.)QPACE QCD Parallel Computing on the Cell Broadband Engine™ (Cell/B.E.)
QPACE QCD Parallel Computing on the Cell Broadband Engine™ (Cell/B.E.)Heiko Joerg Schick
 
QPACE - QCD Parallel Computing on the Cell Broadband Engine™ (Cell/B.E.)
QPACE - QCD Parallel Computing on the Cell Broadband Engine™ (Cell/B.E.)QPACE - QCD Parallel Computing on the Cell Broadband Engine™ (Cell/B.E.)
QPACE - QCD Parallel Computing on the Cell Broadband Engine™ (Cell/B.E.)Heiko Joerg Schick
 

Plus de Heiko Joerg Schick (16)

Da Vinci - A scaleable architecture for neural network computing (updated v4)
Da Vinci - A scaleable architecture for neural network computing (updated v4)Da Vinci - A scaleable architecture for neural network computing (updated v4)
Da Vinci - A scaleable architecture for neural network computing (updated v4)
 
Huawei empowers healthcare industry with AI technology
Huawei empowers healthcare industry with AI technologyHuawei empowers healthcare industry with AI technology
Huawei empowers healthcare industry with AI technology
 
The 2025 Huawei trend forecast gives you the lowdown on data centre facilitie...
The 2025 Huawei trend forecast gives you the lowdown on data centre facilitie...The 2025 Huawei trend forecast gives you the lowdown on data centre facilitie...
The 2025 Huawei trend forecast gives you the lowdown on data centre facilitie...
 
The Smarter Car for Autonomous Driving
 The Smarter Car for Autonomous Driving The Smarter Car for Autonomous Driving
The Smarter Car for Autonomous Driving
 
From edge computing to in-car computing
From edge computing to in-car computingFrom edge computing to in-car computing
From edge computing to in-car computing
 
Need and value for various levels of autonomous driving
Need and value for various levels of autonomous drivingNeed and value for various levels of autonomous driving
Need and value for various levels of autonomous driving
 
Run-Time Reconfiguration for HyperTransport coupled FPGAs using ACCFS
Run-Time Reconfiguration for HyperTransport coupled FPGAs using ACCFSRun-Time Reconfiguration for HyperTransport coupled FPGAs using ACCFS
Run-Time Reconfiguration for HyperTransport coupled FPGAs using ACCFS
 
Browser and Management App for Google's Person Finder
Browser and Management App for Google's Person FinderBrowser and Management App for Google's Person Finder
Browser and Management App for Google's Person Finder
 
IBM Corporate Service Corps - Helping Create Interactive Flood Maps
IBM Corporate Service Corps - Helping Create Interactive Flood MapsIBM Corporate Service Corps - Helping Create Interactive Flood Maps
IBM Corporate Service Corps - Helping Create Interactive Flood Maps
 
Real time Flood Simulation for Metro Manila and the Philippines
Real time Flood Simulation for Metro Manila and the PhilippinesReal time Flood Simulation for Metro Manila and the Philippines
Real time Flood Simulation for Metro Manila and the Philippines
 
Slimline Open Firmware
Slimline Open FirmwareSlimline Open Firmware
Slimline Open Firmware
 
Agnostic Device Drivers
Agnostic Device DriversAgnostic Device Drivers
Agnostic Device Drivers
 
The Cell Processor
The Cell ProcessorThe Cell Processor
The Cell Processor
 
directCell - Cell/B.E. tightly coupled via PCI Express
directCell - Cell/B.E. tightly coupled via PCI ExpressdirectCell - Cell/B.E. tightly coupled via PCI Express
directCell - Cell/B.E. tightly coupled via PCI Express
 
QPACE QCD Parallel Computing on the Cell Broadband Engine™ (Cell/B.E.)
QPACE QCD Parallel Computing on the Cell Broadband Engine™ (Cell/B.E.)QPACE QCD Parallel Computing on the Cell Broadband Engine™ (Cell/B.E.)
QPACE QCD Parallel Computing on the Cell Broadband Engine™ (Cell/B.E.)
 
QPACE - QCD Parallel Computing on the Cell Broadband Engine™ (Cell/B.E.)
QPACE - QCD Parallel Computing on the Cell Broadband Engine™ (Cell/B.E.)QPACE - QCD Parallel Computing on the Cell Broadband Engine™ (Cell/B.E.)
QPACE - QCD Parallel Computing on the Cell Broadband Engine™ (Cell/B.E.)
 

Dernier

Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024
Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024
Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024Victor Rentea
 
Six Myths about Ontologies: The Basics of Formal Ontology
Six Myths about Ontologies: The Basics of Formal OntologySix Myths about Ontologies: The Basics of Formal Ontology
Six Myths about Ontologies: The Basics of Formal Ontologyjohnbeverley2021
 
Finding Java's Hidden Performance Traps @ DevoxxUK 2024
Finding Java's Hidden Performance Traps @ DevoxxUK 2024Finding Java's Hidden Performance Traps @ DevoxxUK 2024
Finding Java's Hidden Performance Traps @ DevoxxUK 2024Victor Rentea
 
Polkadot JAM Slides - Token2049 - By Dr. Gavin Wood
Polkadot JAM Slides - Token2049 - By Dr. Gavin WoodPolkadot JAM Slides - Token2049 - By Dr. Gavin Wood
Polkadot JAM Slides - Token2049 - By Dr. Gavin WoodJuan lago vázquez
 
AWS Community Day CPH - Three problems of Terraform
AWS Community Day CPH - Three problems of TerraformAWS Community Day CPH - Three problems of Terraform
AWS Community Day CPH - Three problems of TerraformAndrey Devyatkin
 
Strategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a FresherStrategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a FresherRemote DBA Services
 
MS Copilot expands with MS Graph connectors
MS Copilot expands with MS Graph connectorsMS Copilot expands with MS Graph connectors
MS Copilot expands with MS Graph connectorsNanddeep Nachan
 
"I see eyes in my soup": How Delivery Hero implemented the safety system for ...
"I see eyes in my soup": How Delivery Hero implemented the safety system for ..."I see eyes in my soup": How Delivery Hero implemented the safety system for ...
"I see eyes in my soup": How Delivery Hero implemented the safety system for ...Zilliz
 
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost SavingRepurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost SavingEdi Saputra
 
Architecting Cloud Native Applications
Architecting Cloud Native ApplicationsArchitecting Cloud Native Applications
Architecting Cloud Native ApplicationsWSO2
 
DEV meet-up UiPath Document Understanding May 7 2024 Amsterdam
DEV meet-up UiPath Document Understanding May 7 2024 AmsterdamDEV meet-up UiPath Document Understanding May 7 2024 Amsterdam
DEV meet-up UiPath Document Understanding May 7 2024 AmsterdamUiPathCommunity
 
Boost Fertility New Invention Ups Success Rates.pdf
Boost Fertility New Invention Ups Success Rates.pdfBoost Fertility New Invention Ups Success Rates.pdf
Boost Fertility New Invention Ups Success Rates.pdfsudhanshuwaghmare1
 
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...DianaGray10
 
Platformless Horizons for Digital Adaptability
Platformless Horizons for Digital AdaptabilityPlatformless Horizons for Digital Adaptability
Platformless Horizons for Digital AdaptabilityWSO2
 
Rising Above_ Dubai Floods and the Fortitude of Dubai International Airport.pdf
Rising Above_ Dubai Floods and the Fortitude of Dubai International Airport.pdfRising Above_ Dubai Floods and the Fortitude of Dubai International Airport.pdf
Rising Above_ Dubai Floods and the Fortitude of Dubai International Airport.pdfOrbitshub
 
Apidays New York 2024 - Passkeys: Developing APIs to enable passwordless auth...
Apidays New York 2024 - Passkeys: Developing APIs to enable passwordless auth...Apidays New York 2024 - Passkeys: Developing APIs to enable passwordless auth...
Apidays New York 2024 - Passkeys: Developing APIs to enable passwordless auth...apidays
 
Navigating the Deluge_ Dubai Floods and the Resilience of Dubai International...
Navigating the Deluge_ Dubai Floods and the Resilience of Dubai International...Navigating the Deluge_ Dubai Floods and the Resilience of Dubai International...
Navigating the Deluge_ Dubai Floods and the Resilience of Dubai International...Orbitshub
 
Artificial Intelligence Chap.5 : Uncertainty
Artificial Intelligence Chap.5 : UncertaintyArtificial Intelligence Chap.5 : Uncertainty
Artificial Intelligence Chap.5 : UncertaintyKhushali Kathiriya
 
ProductAnonymous-April2024-WinProductDiscovery-MelissaKlemke
ProductAnonymous-April2024-WinProductDiscovery-MelissaKlemkeProductAnonymous-April2024-WinProductDiscovery-MelissaKlemke
ProductAnonymous-April2024-WinProductDiscovery-MelissaKlemkeProduct Anonymous
 

Dernier (20)

Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024
Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024
Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024
 
Understanding the FAA Part 107 License ..
Understanding the FAA Part 107 License ..Understanding the FAA Part 107 License ..
Understanding the FAA Part 107 License ..
 
Six Myths about Ontologies: The Basics of Formal Ontology
Six Myths about Ontologies: The Basics of Formal OntologySix Myths about Ontologies: The Basics of Formal Ontology
Six Myths about Ontologies: The Basics of Formal Ontology
 
Finding Java's Hidden Performance Traps @ DevoxxUK 2024
Finding Java's Hidden Performance Traps @ DevoxxUK 2024Finding Java's Hidden Performance Traps @ DevoxxUK 2024
Finding Java's Hidden Performance Traps @ DevoxxUK 2024
 
Polkadot JAM Slides - Token2049 - By Dr. Gavin Wood
Polkadot JAM Slides - Token2049 - By Dr. Gavin WoodPolkadot JAM Slides - Token2049 - By Dr. Gavin Wood
Polkadot JAM Slides - Token2049 - By Dr. Gavin Wood
 
AWS Community Day CPH - Three problems of Terraform
AWS Community Day CPH - Three problems of TerraformAWS Community Day CPH - Three problems of Terraform
AWS Community Day CPH - Three problems of Terraform
 
Strategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a FresherStrategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a Fresher
 
MS Copilot expands with MS Graph connectors
MS Copilot expands with MS Graph connectorsMS Copilot expands with MS Graph connectors
MS Copilot expands with MS Graph connectors
 
"I see eyes in my soup": How Delivery Hero implemented the safety system for ...
"I see eyes in my soup": How Delivery Hero implemented the safety system for ..."I see eyes in my soup": How Delivery Hero implemented the safety system for ...
"I see eyes in my soup": How Delivery Hero implemented the safety system for ...
 
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost SavingRepurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
 
Architecting Cloud Native Applications
Architecting Cloud Native ApplicationsArchitecting Cloud Native Applications
Architecting Cloud Native Applications
 
DEV meet-up UiPath Document Understanding May 7 2024 Amsterdam
DEV meet-up UiPath Document Understanding May 7 2024 AmsterdamDEV meet-up UiPath Document Understanding May 7 2024 Amsterdam
DEV meet-up UiPath Document Understanding May 7 2024 Amsterdam
 
Boost Fertility New Invention Ups Success Rates.pdf
Boost Fertility New Invention Ups Success Rates.pdfBoost Fertility New Invention Ups Success Rates.pdf
Boost Fertility New Invention Ups Success Rates.pdf
 
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
 
Platformless Horizons for Digital Adaptability
Platformless Horizons for Digital AdaptabilityPlatformless Horizons for Digital Adaptability
Platformless Horizons for Digital Adaptability
 
Rising Above_ Dubai Floods and the Fortitude of Dubai International Airport.pdf
Rising Above_ Dubai Floods and the Fortitude of Dubai International Airport.pdfRising Above_ Dubai Floods and the Fortitude of Dubai International Airport.pdf
Rising Above_ Dubai Floods and the Fortitude of Dubai International Airport.pdf
 
Apidays New York 2024 - Passkeys: Developing APIs to enable passwordless auth...
Apidays New York 2024 - Passkeys: Developing APIs to enable passwordless auth...Apidays New York 2024 - Passkeys: Developing APIs to enable passwordless auth...
Apidays New York 2024 - Passkeys: Developing APIs to enable passwordless auth...
 
Navigating the Deluge_ Dubai Floods and the Resilience of Dubai International...
Navigating the Deluge_ Dubai Floods and the Resilience of Dubai International...Navigating the Deluge_ Dubai Floods and the Resilience of Dubai International...
Navigating the Deluge_ Dubai Floods and the Resilience of Dubai International...
 
Artificial Intelligence Chap.5 : Uncertainty
Artificial Intelligence Chap.5 : UncertaintyArtificial Intelligence Chap.5 : Uncertainty
Artificial Intelligence Chap.5 : Uncertainty
 
ProductAnonymous-April2024-WinProductDiscovery-MelissaKlemke
ProductAnonymous-April2024-WinProductDiscovery-MelissaKlemkeProductAnonymous-April2024-WinProductDiscovery-MelissaKlemke
ProductAnonymous-April2024-WinProductDiscovery-MelissaKlemke
 

Blue Gene Active Storage

  • 1. 1 August 2, 2010 Blue Gene Active Storage HEC FSIO 2010 Blake G. Fitch Alex Rayshubskiy IBM Research T.J. Chris Ward Mike Pitman Bernard Metzler bgf@us.ibm.com Heiko J. Schick Benjamin Krill Peter Morjan Robert S. Germain © 2010 IBM Corporation © 2002 IBM Corporation
  • 2. ‘Exascale’ Architectures Motivate Active Storage Exascale: Three orders of magnitude more of everything? –Often taken to mean just ‘exaflop’ – but that’s the easy part! –What about the rest of the Exascale ecosystem: storage, network bandwidth, usability? Exascale observations: –The wetware will not be three orders of magnitude more capable • Essential to think about usability. Massive concurrency must be made more usable. • Non-heroic programmers and end-users must be able to utilize exascale architectures –Data objects will be too large to move -- between or even within datacenters • Today, terabyte objects have people carrying around portable disk arrays – this is not scalable. • Bandwidth costs not likely to drop 103 either inside out outside the data center –Single program (MPI) use-case alone will be insufficient for to exploit exascale capability • Anticipate applications formed of composed components rather than just heroic monoliths • Some capability created by exascale will be spent enabling interfaces to support compos ability –“Tape is Dead, Disk is Tape, Flash is Disk, RAM locality is King” – Jim Gray • A rack of disks can deliver ~10GB/s. 1TB/s requires 100 racks or ~$100M • Mixing archival storage and online storage will be less efficient in the future • Solid state storage and storage class memory – but how will these technologies be used? Scalable Bisectional bandwidth will be THE challenge both to create and exploit 2 Blue Gene Active Storage HEC FSIO 2010 © 2010 IBM Corporation
  • 3. From K. Borne, Data Science Challenges from Distributed Petabyte Astronomical Data Collections: Preparing for the Data Avalanche through Persistence, Parallelization, and Provenance, Computing with Massive and Persistent Data (CMPD 08) https://computation.llnl.gov/home/workshops/cmpd08/Presentations/Borne.pdf 3 Blue Gene Active Storage HEC FSIO 2010 © 2010 IBM Corporation
  • 4. Data-centric and Network-centric Application Classes Embarrassingly Parallel Network Dependent Table scan Join Batch serial Sort –Floating point intensive –“order by” queries –Job farm/cloud “group by” queries Map-Reduce Map-Reduce Other applications without Aggregation operations internal global data –count(), sum(), min(), max(), avg(), … dependencies Data analysis/OLAP –Aggregation with “group by”… –Real-time analytics HPC applications 4 Blue Gene Active Storage HEC FSIO 2010 © 2010 IBM Corporation
  • 5. Blue Gene Architecture in Review Blue Gene is not just FLOPs… … it’s also torus network, power efficiency, and dense packaging. Focus on scalability rather than on configurability gives the Blue Gene family’s System-on-a-Chip architecture unprecedented scalability and reliability. 5 Blue Gene Active Storage HEC FSIO 2010 © 2010 IBM Corporation
  • 6. Blue Gene Architectural Advantages Main Memory Capacity per Rack Peak Memory Bandwidth per node (byte/flop) 4500 SGI ICE 4000 3500 Sun TACC 3000 Itanium 2 2500 2000 Cray XT5 4 core 1500 1000 Cray XT3 2 core 500 BG/P 4 core 0 LRZ Cray ASC BG/P Sun SGI ICE 0 0.5 1 1.5 IA64 XT4 Purple TACC Relative power, space and cooling efficiencies Low Failure Rates http://acts.nersc.gov/events/Workshop2006/slides/Simon.pdf 400% 300% 200% 100% IBM BG/P 0% Racks/TF kW/TF Sq Ft/TF Tons/TF Sun/Constellation Cray/XT4 SGI/ICE 6 Blue Gene Active Storage HEC FSIO 2010 © 2010 IBM Corporation
  • 7. Blue Gene Torus: A cost effective all-to-all network 10GbE network costs from: Data Center Switch Architecture In The Age Of Merchant Silicon (Farrington, et al, 2009 http://cseweb.ucsd.edu/~vahdat/papers/hoti09.pdf) Theoretical 3D and 4D Torus All-to-all throughput per node for 1GB/s Links (~1/2 10GbE) 3 2 .5 A2A BW/Node (GB/s) 2 1 .5 1 0 .5 3 D -T o ru s 4 D -T o ru s 0 0 5000 10000 15000 20000 25000 30000 N ode C ount 7 Blue Gene Active Storage HEC FSIO 2010 © 2010 IBM Corporation
  • 8. Blue Gene Data-centric Micro-benchmark Results Random Loads Per Second 100 Blue Gene/L micro benchmarks show 10 16,384 Nodes order of magnitude improvement over largest SMP and clusters Giga-Loads/Second 1 0.1 0.01 Scalable random pointer chasing 0.001 through terabytes of memory Migrate Context Memory Get 0.0001 –BG/L 1.0 GigaLoads / Rack 1 10 100 1000 Total Read Contexts 10000 100000 1e+006 • 14 GigaLoads/s for 16 racks –P690 0.3 Gigaloads Strong Scaling vs. Node Count 120 1 TeraByte Scalable “Indy” sort of terabytes in 2 TeraByte 3 TeraByte 100 memory 4 TeraByte 5 TeraByte Sort Time (sec.) 80 6 TeraByte –BG/L 1.0 TB sort in 10 secs 7 TeraByte 60 –BG/L 7.5 TB sort in 100 secs 40 Streams benchmark 20 –BG/L 2.5TB / Sec / Rack 0 –P5 595 0.2TB / Sec / Rack 4000 8000 16000 Node Count 8 Blue Gene Active Storage HEC FSIO 2010 © 2010 IBM Corporation
  • 9. Blue Gene/P for I/O and Active Storage Explorations Currently a research platform for scalable data-centric computing and BG/Q I/O system prototyping Linux 2.6.29 kernel, 32-bit powerpc 4GB RAM, diskless, 4-way SMP, 1024 nodes/rack 4096 nodes available on site IP-over-torus to other compute nodes IO nodes as IP routers for off-fabric traffic MPICH/TCP, OpenMPI/TCP, SLURM, TCP sockets Running SoftiWARP for OFED RDMA (http://gitorious.org/softiwarp) Native OFED verbs provider for Blue Gene Torus has achieved initial function Experimental GPFS version for on-board Parallel File System explorations at very high node count Testing MVAPICH and OpenMPI Future – anticipate moving ASF over to and integrating with Argonne National Lab’s ZeptoOS (http://www.zeptoos.com) 9 Blue Gene Active Storage HEC FSIO 2010 © 2010 IBM Corporation
  • 10. Blue Gene Active Storage Opportunity – Current Conditions IBM Enterprise Server (Analysis Here) IBM System Storage DS8000 IBM BG/P HPC Solution (Scalable Compute Here) widening gap IBM 3494 Tape Library Persistent Storage 10 Blue Gene Active Storage HEC FSIO 2010 © 2010 IBM Corporation
  • 11. Blue Gene Active Storage adds parallel processing into the storage hierarchy. IBM Enterprise Server 10 Blue Gene/AS 0s Option GB /s IBM System Storage DS8000 Application Solution Driver IBM 3494 Tape Library Managed as Scalable Solid State Storage with Embedded Parallel Persistent Processing Storage 11 Blue Gene Active Storage HEC FSIO 2010 © 2010 IBM Corporation
  • 12. Thought Experiment: A Blue Gene Active Storage Machine Integrate significant storage class memory (SCM) at each node – For now, Flash memory, maybe similar function to Fusion-io ioDrive Duo – Future systems may deploy Phase Change Memory (PCM), Memristor, or…? – Assume node density will drops 50% -- 512 Nodes/Rack for embedded apps ioDrive Duo One Board 512 Node – Objective: balance Flash bandwidth to network all-to-all throughput SLC NAND Cap. 320 GB 160 TB Read BW (64K) 1450 MB/s 725 GB/s Resulting System Attributes: Write BW (64K) 1400 MB/s 700 GB/s – Rack: 0.5 petabyte, 512 Blue Gene processors, and embedded torus network Read IOPS (4K) 270,000 138 Mega – 700 TB/s I/O bandwidth to Flash – competitive with ~70 large disk controllers Write IOPS (4K) 257,000 131 Maga • Order of magnitude less space and power than equivalent perf via disk solution Mixed R/W 207,000 105 Mega • Can configure fewer disk controllers and optimize them for archival use IOPs(75/25@4K) 51 – With network all-to-all throughput at 1GB/s per node, anticipate: 2x • 1TB sort from/to persistent storage in order 10 secs. • 130 Million IOPs per rack, 700 GB/s I/O bandwidth – Inherit Blue Gene attributes: scalability, reliability, power efficiency, Research Challenges (list not exhaustive): – Packaging – can the integration succeed? – Resilience – storage, network, system management, middleware – Data management – need clear split between on-line and archival data – Data structures and algorithms can take specific advantage of the BGAS architecture – no one cares it’s not x86 since software is embedded in storage Related Work: – Gordon (UCSD) http://nvsl.ucsd.edu/papers/Asplos2009Gordon.pdf – FAWN (CMU) http://www.cs.cmu.edu/~fawnproj/papers/fawn-sosp2009.pdf – RamCloud (Stanford) http://www.stanford.edu/~ouster/cgi-bin/papers/ramcloud.pdf 12 12 Blue Gene Active Storage HEC FSIO 2010 © 2010 IBM Corporation
  • 13. Blue Gene Active Storage for Server Acceleration Manage Blue Gene hardware as a scalable, solid state P Servers Z servers storage device with embedded processing capability Integrate this “Active Storage Fabric” with middleware Data Sets Blue Gene/ASF such as DB2, MySql, GPFS, etc, at the data/storage interface using a parallel, key-value in-memory data store Disk Array Transparently accelerate server Jobs with ASF: – Create libraries of reusable, low overhead Embedded Parallel Modules (EPMs) (scan, sort, join, sum, etc) – EPMs directly access middleware data (tables, files, etc) in Robot Tape key/value DB based on middleware data model – Middleware makes ASF datasets available via legacy interfaces allowing incremental parallelization and interoperability with legacy middleware function and host programs Shelf Tape Integrate BG/ASF with information life-cycle and other standard IT data management products 13 Blue Gene Active Storage HEC FSIO 2010 © 2010 IBM Corporation
  • 14. Blue Gene Active Storage for UNIX Utility Acceleration GPFS servers access file system data blocks in a single ‘disk’ in Active Store Huge Blue Gene Parallel In-Memory Database accesses CNTL Active Storage Fabric Analytics block with key of [file_num,block_num] Server Active GPFS Embedded Parallel Modules (EPMs) RDMA access file data directly with PIMD cleint Router asf_generate_indy_data an EPM RDMA creates Indy.Input synthetic data Router BG/AS Nodes Active Linux IP/RDMA asf_grep EPM access Indy.Input file GPFS and greps all text records creating RDMA Indy.Temp PIMD Server Router asf_sort EPM sorts Indy.Temp file and RDMA creates Indy.Output Router EPMs, ex: grep Tail -1 is a normal unix command which Non-heroic runs on the workstation and gets the End-User last line in the last block of the Indy.Output file Active GPFS asf_user@asf_host-> cd /active_gpfs Client asf_user@asf_host-> asf_generate_indy_data 1TB > Indy.Input asf_user@asf_host-> asf_grep “aaaaaa” Indy.Input > Indy.Temp asf_user@asf_host-> asf_sort < Indy.Temp >Indy.Output asf_user@asf_host-> tail -1 Indy.Output 9876543210aaaaaa………………………………. … … asf_user@asf_host-> rm Indy.Output Indy.Input 14 Blue Gene Active Storage HEC FSIO 2010 © 2010 IBM Corporation
  • 15. Questions? Note: If you are NDA’d with IBM for Blue Gene/Q or are a US Government employee and you are interested in the confidential version of the Blue Gene/Q Active Storage presentation, please contact me. Blake Fitch bgf@us.ibm.com 15 Blue Gene Active Storage HEC FSIO 2010 © 2010 IBM Corporation
  • 16. Backup 16 Blue Gene Active Storage HEC FSIO 2010 © 2010 IBM Corporation
  • 17. Blue Gene/L ASF: Early Performance Comparison with Host Commands Benchmark process*: • Generate a 22 GB synthetic “Indy Benchmark” data file • grep for records with “AAAAAA” (outputs 1/3’d of the data which is used as input to sort) • sort the output of grep * NOTE: All ASF data taken on untuned system, non-optimized and with full debug printfs, all p55 data taken with unix time on command line) Function BG/ASF pSeries p55 (Measured at Host shell and break out of EPM components) 512 nodes (1 thread) Total host command time: create Indy File: (22GB) 22.26 s 686s Total host command time: unix grep of Indy File: 25.1s 3180s File Blocks to PIMD Records (Embedded grep) : 7.0 s Grep core function (Embedded grep) : 2.5s PIMD Records to File Blocks (Embedded grep) : 7.6s Total time on Blue Gene/ASF fabric (Embedded grep) : 18.8s Total host command time: sort output of grep: 60.41s 2220s File Blocks to PIMD Records (Embedded sort) : 2.9s Sort core function (Embedded sort) : 12.2s PIMD Records to File Blocks (Embedded sort) : 15.8s Total time on Blue Gene/ASF fabric (Embedded sort) : 54.7s 17 Blue Gene Active Storage HEC FSIO 2010 © 2010 IBM Corporation
  • 18. Blue Gene/L ASF: Early Scalability Results Benchmark process*: • Generate a 22 GB synthetic “Indy Benchmark” data file • grep for records with “AAAAAA” (outputs 1/3’d of the data which is used as input to sort) • sort the output of grep * NOTE: All ASF data taken on untuned system, non-optimized and with full debug printfs, all p55 data taken with unix time on command line) Function BG/ASF BG/ASF 512 nodes 8K nodes Total host command time: create Indy File: (22GB) 22.26 s (1TB) 197s Total host command time: unix grep of Indy File: 25.1s 107s File Blocks to PIMD Records (Embedded grep) : 7.0 s 18.1s Grep core function (Embedded grep) : 2.5s 5.1s PIMD Records to File Blocks (Embedded grep) : 7.6s 19.9s Total time on Blue Gene/ASF fabric (Embedded grep) : 18.8s 50.2s Total host command time: sort output of grep: 60.41s 120s File Blocks to PIMD Records (Embedded sort) : 2.9s 6.8s Sort core function (Embedded sort) : 12.2s 14.8s PIMD Records to File Blocks (Embedded sort) : 15.8s 22.4s Total time on Blue Gene/ASF fabric (Embedded sort) : 54.7s 66.55s 18 Blue Gene Active Storage HEC FSIO 2010 © 2010 IBM Corporation
  • 19. Parallel In-Memory Database (PIMD) 512 Nodes 19 Blue Gene Active Storage HEC FSIO 2010 © 2010 IBM Corporation
  • 20. Applications with potentially strong data dependencies Application Application Class/Trends Algorithms National Security/Intelligence, Social Graph-related Breadth first search, …. Networks Molecular simulation N-body Spectral methods (FFT) Financial risk, pricing, market Real time control, shift from forensic Graph traversal, Monte Carlo, … making to real time Oil exploration Inverse problems Reverse time migration, finite difference Climate modeling, weather Fluid dynamics Finite difference, spectral methods, prediction … Recommendation systems Interactive query Minimization (least squares, conjugate gradient, …) Network simulations Discrete event Traffic management Real time control Real time optimization Business intelligence/analytics Data mining/association discovery K-means clustering, aggregation, scans, …. Data deduplication Storage management Hashing Load balancing/meshing (cross- Graph partitioning domain) Multi-media/unstructured search Search MapReduce, hashing, clustering 20 Blue Gene Active Storage HEC FSIO 2010 © 2010 IBM Corporation
  • 21. BG/Q Active Storage for Expanding Flash Capacity 1.00E+13 1.00E+12 Next two years DRAM raw density (bits/cm2) FLASH raw density (bits/cm2) PCM raw density (bits/cm2) 1.00E+11 possible DRAM limit Pessimistic PCM (2 bits, 2 layers only) 1.00E+10 1.00E+09 2007 2009 2011 2013 2015 2017 2019 2021 2023 ITRS 2008 data 21 Blue Gene Active Storage HEC FSIO 2010 © 2010 IBM Corporation
  • 22. Related Data Intensive Supercomputing Efforts •Hyperion (LLNL) http://www.supercomputingonline.com/latest/appro-deploys-a-world-class-linux-cluster-testbed-solution-to-llnl-in-support-of-the-hyperion-project l •The Linux cluster solution consists of 80 Appro 1U 1423X, dual socket customized server based on Intel Xeon processor 5600 series with 24 GB of memory and 4 PCIe2 x8 slots. The cluster will be integrated with 2x ioSAN carrier cards with 320 GB of FLASH memory and one 10Gb/s Ethernet and one IBA QDR (40Gb/s) link. The I/O nodes also have 2x 1Gb/s Ethernet links for low speed cloud file system I/O and two SATA disks for speed comparisons. In addition to these IO nodes, 28 Lustre OSS 1U Appro 1U 1423X dual socket nodes with 24 GB of memory, one Mellanox Connect-X card, 10Gb/s Ethernet and QDR IBA (40Gb/s) links and 2 SATA drives will be added. The entire solution will provide 80 Appro server nodes powered by Fusion-io technology delivering an aggregate 4KB IOPS rate of over 40 million IOPS and 320 GB/s of IO bandwidth and 102 TB of FLASH memory capabilities. •Flash Gordon (SDSC) http://storageconference.org/2010/Presentations/MSST/10.Norman.pdf 22 Blue Gene Active Storage HEC FSIO 2010 © 2010 IBM Corporation
  • 23. Failures per Month per TF From:http://acts.nersc.gov/events/Workshop2006/slides/Simon.pdf 23 Blue Gene Active Storage HEC FSIO 2010 © 2010 IBM Corporation