This document discusses a large-scale GPU-based cloud burst simulation run by the IceCube collaboration to calibrate simulations of natural ice. The simulation was data-intensive, producing over 130 TB of data and exceeding 10 Gbps of egress bandwidth. Internet2 Cloud Connect service was used to provision over 20 dedicated network links between collaborators' institutions and cloud providers to enable high-throughput data transfer at a lower cost than commercial routes. Careful planning was required to smoothly ramp up the burst and avoid overloading individual network links.
Handwritten Text Recognition for manuscripts and early printed texts
Data-intensive IceCube Cloud Burst
1. Data-intensive IceCube Cloud Burst
leveraging Internet2 Cloud Connect Service
By Igor Sfiligoi – UCSD
For the rest of the IceCube Cloud Burst team (UCSD+UW Madison)
NRP-Pilot Weekly Meeting, Nov 12th 2020
2. IceCube GPU-based Cloud bursts
• Large amount of photon
propagation simulation needed
to properly calibrate natural Ice
• Simulation compute intensive,
and ideal for GPU compute
Nov’20
https://icecube.wisc.edu
Previous: https://doi.org/10.1145/3311790.3396625
Integral: 225 PFLOP hoursfp32
3. Egress data intensive Cloud run
• This IceCube simulation was
relatively heavy in egress data
• 2 GB per job
• Job length ~= 0.5 hour
• And very spiky
• The whole file
is transferred
after compute completed
• Input sizes small-ish
• 0.25 GB
• Cloud burst exceeded 10 GBps
• To make good use of a large fraction
of available Cloud GPUs
https://www.linkedin.com/pulse/cloudy-100-pflops-gbps-icecube-igor-sfiligoi
4. Storage backends
• UW Madison is IceCube’s
home institution
• Large Lustre-based filesystem
• 5x dedicated GridFTP servers,
each with 25 Gbps NIC
• At UCSD we used SDSC Qumulo
• Available as NFS mounts
inside the UCSD network
• Deployed GridFTP pods in Nautilus
• 3 x pods on 3 nodes at UCSD
• Each with 100 Gbps NIC
• Each pod had 5x NFS mountpoints
5. Using Internet2 Cloud Connect Service
• Egress costs are notoriously high
• Using dedicated links cheaper
• If provisioned on demand
• Internet2 acts as provider for
the research community
• For AWS, Azure and GCP
• No 100Gbps links available
• Had to stitch together 20+ links,
each 10Gbps, 5Gbps and 2 Gbps
Each color band belongs
to one network link
https://internet2.edu/services/cloud-connect/
Simplified list price comparison
6. Using Internet2 Cloud Connect Service
• Egress costs are notoriously high
• Using dedicated links cheaper
• If provisioned on demand
• Internet2 acts as provider for
the research community
• For AWS, Azure and GCP
• No 100Gbps links available
• Had to stitch together 20+ links,
each 10Gbps, 5Gbps and 2 Gbps
Each color band belongs
to one network link
https://internet2.edu/services/cloud-connect/
Produced 130 TB of data
• List price for commercial path: $11k
• We paid: $6k
Compute: $26k
Simplified list price comparison
7. The need for many links
• Internet2 has mostly 2x 10 Gbps links with Cloud providers
• The only bright exception is the California link to Azure at 2x 100 Gbps
• The links are shared, so one can never get the whole link for itself
• 5 Gbps limit in AWS and GCP
• 10 Gbps limit in Azure
• The link speeds are rigidly defined
• 1, 2, 5, 10 Gbps
• To fill an (almost) empty 10 Gbps link,
one needs three links: 5 + 2 + 2
9. Very different provisioning in the 3 Clouds
• AWS the most complex
• And requires initiation by
on-prem network engineer
• Many steps after initial request
• Create VPC and subnets
• Accept connection request
• Create VPG
• Associate VPG with VPC
• Create DCG
• Create VIF
• Relay back to on-prem the BGP key
• Establish VPC -> VPG routing
• Associate DCG -> VPG
• And don’t forget the Internet routers
• GCP the simplest
• Create VPC and subnets
• Create Cloud Router
• Create Interconnect
• Provide key to on-prem
• Azure not much harder
• Create VN and subnets
• Make sure the VN has Gateway subnet
• Create ExpressRoute (ER)
• Provide key to on-prem
• Create VNG
• Create connection between ER and VNG
• Note: Azure comes with many more options
to choose from
10. Additional on-prem networking setup needed
• Quote from Michael Hare, UW Madison Network engineer:
In addition to network configuration [at] UW Madison (AS59), we
provisioned BGP based Layer 3 MPLS VPNs (L3VPNs) towards Internet2
via our regional aggregator, BTAA OmniPop.
This work involved reaching out to the BTAA NOC to coordinate on VLAN
numbers and paths and to [the] Internet2 NOC to make sure the newly
provisioned VLANs were configurable inside OESS.
Due to limitations in programmability or knowledge at the time regarding
duplicate IP address towards the cloud (GCP, Azure, AWS) endpoints, we
built several discrete L3VPNs inside the Internet2 network to accomplish
the desired topology.
• Tom Hutton did the UCSD part
11. Spiky nature of workload tricky for networking
• We could not actually do
a “burst” this time
• Results in too many
spikes and valleys
• We tried it at smaller scale
• Noticed that links
to different
providers behave
differently
• Some capped,
some flexible
• Long upload
times when
congested
Capped
Flexible
Aggregated UW Madison Storage Network – Smaller scale bursty test
12. Much more careful during big “burst”
• Ramp up for over 2 hours • Still not perfect
• But much smoother
GBps–Averagedover10mins
GBps–Averagedover10mins
fp32PFLOPS
Ramp up Stable Final push
13. Summary
•Using dedicated links made this Cloud run
a little more challenging
• But cost savings worth it
•Showed that data-intensive high-throughput
Cloud computing is doable
• With plenty of science data generated to show for it
14. Acknowledgments
• I would like to thank NSF for their support of this endeavor as part of
the OAC-1941481, MPS-1148698, OAC-1841530 , OAC-
1826967 and OPP-1600823.
• And all of this would of course not be possible without the hard work
of Michael Hare, David Schultz, Benedikt Riedel, Vladimir Brik,
Steve Barnet, Frank Wuerthwein, Tom Hutton, Matt Zekauskas and
John Hicks.
16. Application logs only provide dt+MBs for egress
• Different averaging techniques give sightly different insights
GBps–Averagedover1min
fp32PFLOPS
Ramp up Stable Final push GBps–Averagedover1min
17. Internet2 Cloud Connect Explained
• Each Cloud provider has its own
“dedicated link” mechanism
• Similar in spirit,
but technically different
• AWS has Direct Connect
https://aws.amazon.com/directconnect/
• Azure has Express Route
https://azure.microsoft.com/en-us/services/expressroute/
• GCP has Cloud Interconnect
https://cloud.google.com/network-connectivity/docs/interconnect
• Internet2 acts as a
service provider for
all three major Cloud providers
• Providing
• The physical network infrastructure
• A portal for on-prem operators
Azure
example