The document discusses how networking in data centers must fundamentally change to adapt to virtualization and growth in storage. Legacy physical networks are challenging to manage at scale and cannot keep up with modern demands, while next generation virtualized networks provide scalability, flexibility, and automation. The document argues that instead of replacing legacy networks, data centers need unified infrastructure management across computing, storage, and networking to efficiently orchestrate flexible infrastructure and realize performance and economic benefits.
2. The Journey to Efficiency
Builds On Virtual Foundation
STRATEGIC
AGILITY
QUALITY OF
DATA CENTER
SERVICE EFFICIENCY
Policy Driven Automation
ECONOMIC
Dynamic Resource Optimization
SAVINGS
Rapid Provisioning
Disaster Recovery
High Availability
Server & Storage Consolidation
3. Why DC Management Must Change
Legacy Networks Next Gen DC Networks
(Physical) (Virtual)
MANAGEMENT CHALLENGES MANAGEMENT SOLUTIONS
• Growth means complexity • High performance and scalability built-
• Does not scale with virtualization in
demands VS • Virtualization-aware components
• Cannot keep up with storage growth • Unified Fabric enabling new levels of
• Made up of discrete devices storage flexibility
• Multitude of management tools • Network building blocks are interactive
and scalable
TECHNOLOGY CHALLENGES • Data center orchestration
• Single Server with Single Application TECHNOLOGY SOLUTIONS
• Discrete Communications and Storage •Dynamic Server with Virtual Applications
Fabrics VS •Unified Fabrics for Communications and
Low Bandwidth in the rack Storage
•Lossy QoS •More bandwidth per port
•Lots of Ports •“Near” Lossless
•Fewer Ports – Higher Bandwidth
Fabric convergence, port proliferation, management sprawl & virtualization
forcing network management changes. Does that mean your legacy network
needs to be thrown out? How do we manage this change?
4. The Next Step in Efficiency
Flexible infrastructure orchestrated through unified
infrastructure management
Unified
Infrastructure
Management
Compute Storage Networking
4
5. Performance: iSCSI, FCoE and FC
10GE, fully offloaded iSCSI stacks up well against FC and FCoE
Throughput (Mbps) Efficiency (Mbps/%CPU)
500 700
450
600
400
350 500
300
400
250
300
200
150 200
100
100
50
0 0
4K 8K 64K 512K 4K 8K 64K 512K 4K 8K 64K 512K 4K 8K 64K 512K
Read Write Read Write
iSCSI Offload FCoE FC iSCSI Offload FCoE FC
IOMeter, 4 Gb/s targets
5
[Source: iSCSI/FC Performance Analysis in Dell CTO Storage Architecture Lab