Data Centers of Today "ho Blade Servers and Switches Super High Density Uptime is a must Speed, cooling and power become the major concerns Centralized applications, web hosting, SAN,
Trang 2When devices speak different languages a
| pee? _
Trang 3Building Control Systems a
HVAC, Power and
Lighting
Trang 4Common Ground - Packets are just like
Preamble| Destination} Source | Type Data CRC/ FCS
8 bytes 6 bytes 6 bytes| 2 bytes 46-1500 bytes 5 Bytes
Trang 5Carrier meets Enterprise
Trang 6
When devices speak a common language a
Trang 8
Early History of Networking — Multiple Users Shared
Computing Resources: User to Device Ratio 1:5 or more
Recent History — All Users had a computer: User to Device Ratio 1:1
Emerging Network — All Users have a computer, IP
Telephone, add to this IP Security Cameras, WiFi Access
Points, Intelligent Building Networks: User to Device Ratio begins to exceed 3:1 10:1
— In excess of three networked devices for each user and growing!
What are the implications of this ratio on our Networks?
Trang 9Total |P Convergence Drives Data ye Center Growth
More Networked Devices
drive additional Equipment
in Data Center
Trang 11and Electrical performance
of the highest importance
Trang 12Enterprise Computer Room
Traditionally housed stand
alone servers in the
Enterprise
Large amount of space
needed
Not designed for Uptime
Density was low and heat
was not a great concern
Mainly used for Email and
data storage
12
Trang 13Carrier Meets Enterprise
Telephone Exchanges/ Central
offices are evolving into Data
Trang 14Data Centers emerge ue
server numbers increased
More Space required
Availability and Power
Density increased and heat
started be become a
concern
Centralized applications and
storage start to become a
critical part of doing
business
Trang 15Data Centers of Today "ho
Blade Servers and Switches
Super High Density
Uptime is a must
Speed, cooling and power
become the major concerns
Centralized applications,
web hosting, SAN,
Redundancy and Reliability
all become table stakes for
doing business
Trang 16Centralized to Distributed and Back? ue
As the number of devices in the network increase there
becomes an increasing need to control what is done locally
Regulations are driving high levels of data storage
The largest risk to the network are the devices at the
periphery
16
Trang 17Thin Clients (dumb terminals) ue
Blade Servers and Switches =< = LỊ |
are common place
Low bandwidth required at
the client end
Centralized, controlled data
storage
17
Trang 18Density, Security and Cost ue
A Complete Business Computing Solution
CearC ube celutions dotrecr Ably v23 92a), Intel Cased PC corputrg to hw ord woos donk trem a ceou'a, contratian location
" "
Better ( lient | tim e The warttonal, Petey PD tant de taken of Ove anc une's Cash ancl rerteced & altho small Lene Pvt shout tha sion of
AWG tame Al the COVE Rite lot Mier) “SCR Toutes derse PC Cates Gativer A Cf?-C©@TAQWM corning 9ò %C (ers avy ee mets 1 Oe crderproe The oetee nchuten & UIÍ-121LWy (3461429 2 lý er Cone Cue Mamaia ( Sate
Super High Density Lượt
Common Platform
PC Blade
Thee @ eae? User) & 3M Corman 4
Cor fgqurabls, intel bated FC Biace thet
Gebvers tI teretonaity to Bc decking
Fast replacement time
Ihe Che đe Cape 6 a conta ized Shaeets Dat tO es ub
way PC Blacus
Easy to upgrade
Str ou Monage nent Suite
Ta Tee ClearC ube Manage nest S2ta (CMS) arpowen T adn nerstoe
a ts Qowe'ls mmoOmMy 13C0CẰ©139tÌ0 tuste MOlLIGes 5 Versriie set of oo
t1FtS SIC 25 AC Spare SWIC eR, Move Ta woemert, 109 13602091
18
Trang 19
Cooling Requirements
Higher density means more
heat
Equipment selection and
airflow become critical as
Space becomes a premium
19
Trang 20Design Stage
Thermal site mapping and
airflow simulators are being
used to design high density
space
Consultants in the Data
Center arena are becoming
more specialized as a result
Again, designs must be
followed to the letter
20
Trang 21New Products to suit the needs
Equipment racks with
dedicated cooling solutions
are starting be emerge
Density is being addressed
Airflow access becomes
critical
21
Trang 22Proper Cooling Techniques "hc
Proper Data Center Design deployment of Hot Aisle/Cold Aisle Cooling
— Good Airflow/Proper Cooling = Optimal Performance of Servers and Switches
22
Trang 23Proper Cooling Techniques
i
Proper Data Center Design deployment of Hot Aisle/Cold Aisle Cooling
— Good Airflow/Proper Cooling = Optimal Performance of Servers and Switches
Telecom
Cable Trays
Perforated Tiles Power Cables
Perforated Tiles Power Cables
23
Trang 24Good Product, Bad Practices
But what happens if poor cable management blocks airflow?
Trang 25The Data Center Design Dilemma:
Density vs Manageability and Reliability
Fundamentally, you cannot use the same
structured cabling products originally
designed for low density LANs and expect
them to perform to the level required ina
Carrier Data Center
25
Trang 26Reliability of Data Center Equipment is ye Directly Tied to Proper Cooling
Cables blocking air inlets and exits will raise the temperature of switches and servers lowering their reliability!
Trang 27Data Center Design Observations
Data Centers are fundamentally different from typical LANs in that they require
Different Architecture Different Level of Reliability
Different Standards
Different Cable Management Needs Different Structured Cabling Solutions
27
Trang 28Data Center Defined „
¢ A Data Center is dedicated area within a building for
— Reliability and nan —
Trang 2910 Gigabit Speeds and Higher (copper and fiber)
Storage Area Networks (Massive Hard Drive Arrays
for the long term storage of information)
Key Point: All of the above are very positive trends for our product lines — Why?
29
Trang 30Blades increase the density of cabling ys
In the data center
¢ Blade Servers, Switches and Clients:
— Increase density of cabling as there are physically more devices per cabinet
— Increase heat generated in cabinet (Airflow becomes more important)
IBM BladeCentere
Trang 31Bringing the Two Together
Typical Enterprise Data Network
— 99.5% Availability
¢ Typical Carrier Voice Network
—~ 99.999% Availability
¢ Voice, video and storage are real-time applications
- An IP converged network demands that the underlying
network infrastructure must be:
— Delay intolerant, reliable and robust
— Provide for at least 99.999% Availability
— The “five nines” !
Trang 32
Tier | - Data Center (Basic)
¢ Single path for power and cooling distribution, no redundant components, 99.671%
availability 28.8 Hours of downtime annually
Tier Il - Data Center (Redundant Components)
« Single path for power and cooling distribution, redundant components, 99.741%
availability
22 Hours of downtime annually
Tier Ill - Data Center (Concurrently Maintainable) Carrier Grade entry level
e Multiple power and cooling distribution paths, but only one path active Redundant components, concurrently maintainable, 99.982% availability
1.6 Hours of downtime annually
Tier |1V - Data Center (Fault Tolerant)
« Multiple active power and cooling distribution paths, redundant components, fault
tolerant, 99.995% availability
0.4 Hours of downtime annually 32
Trang 33Data Center Reliability Tiers
This chart illustrates Tier similarities and differences
TIER I TIER Il T1ER 1H TIER IV
Eduezaher cí Only 1 Only 1 1 active 2 active
delivery paths 1 passive
Support space to mised floor ratio 20% 30% 80-90% 100%
Enitial waktsAt* 20-30 40-50 40-60 50-80
kJ|tkrryxte wwatksJt' 20-30 40-50 100-150 150+
Raised floor height 12” 18” 30-36” 30-36”
Floof loadindl boiwtcls/fF 85 100 150 150+
k#iEty voltage 208, 480 208, 480 12-15kV 12-15kV
Eđont lụa to: implement 3 3to6 15 to 20 15 to 20
Year first depioyed 1965 1970 1985 1995
Tonsiruction Sift’ raised floor $450 $600 $900 $1,100+
Anewal TF dosnntime clue to site 28.8 hrs 22.0 hrs 1.6 hrs 0.4 hrs
Trang 34¢ TIA-568 - Commercial Building Wiring Standard
Covers the design topology and performance requirements for structured wiring inside a
Commercial Building
Inadequate information with regards to Data Centers
Data Center Design Topology is different from
Commercial Buildings
Standards body recognized this and created a Data
Center specific standard, namely TIA-942
34
Trang 35
TIA-942 - Telecommunications Infrastructure
Standard for Data Centers
site selection and sizing Cabling infrastructure administration Architectural and structural considerations Security and fire protection
Electrical, grounding and mechanical systems Application distance limitations
Access-provider coordination and demarcation
35
Trang 36TIA-942 Data Center Network Topology
Carriers ———
Offices, Operations Center,
(Office & Operations
Center LAN Switches)
Trang 37TIA-942 Data Center Functional Areas ue
Entrance Room
Backbone Main Dist Area Carrier Equip &
Cabling (Routers Backbone Vỹ quip
Horizontal LAN/SAN Switches, emarcation)
Cabling
Horiz Dist Area PBX, M13 Muxes) if se
Equip Dist Area
(Rack/Cabinet)
37
Trang 38¢ Gateway to the outside network, this is where the SONET fiber drop will connect into the Data Center
¢ Active Equipment — OCx Fiber Add/Drop
Multiplexer, SHINs box (Self Healing Network
system), M13 Mux
¢ Passive Equipment — Fiber Demarcation Box with
DC Leading Product FL2000)
38
Trang 39Fiber intensive area of Data Center, combines data
from all sources within data center and combines
them onto a small number of high capacity fibers for
transmission to the outside network
Active Equipment — Core Routers (Cisco 12000 or
7000 series), Core Switches (Cisco 6513, 6509), SAN Switch (Cisco 9513)
Passive Equipment — Fiber Distribution Frames (NGF
preferred, FiberGuide and OM3 Cabling)
39
Trang 40Horizontal Distribution Area "oc
° Aggregates the data from servers to several fiber
connects that connect back to the Core Router
and/ or Switches in the MDA
- Active Equipment — Aggregation Switches (Cisco
6513, copper and fiber connections)
¢ Passive Equipment — TrueNet Fiber Panel (for fiber fed servers), CopperTen (for copper fed servers’
Trang 42Another Key Element of the Data Center
SAN Disk Arrays (EMC, IBM, others)
42
Trang 43Storage Area Networks (SANs) require ye
a high density of fiber
¢ SAN’s connect a high capacity hard disk
arrays to the data center via Fiberchannel (not Ethernet, yet)
¢ Number of connections demand high density
fiber cable management solutions (think 24”
FiberGuide and NGF)
g | Sa _
: IIWWWÍ |-
Trang 44Designed to comply with TIA-942 Standard Emphasis on higher levels of Uptime (Tier 3 and 4) Shift from Cat 5e to Cat 6, Augmented Cat 6, and Fiber Higher concentration of cabling (High Density Servers) Designed to support 10 Gigabit Ethernet and Beyond
44
Trang 45Backbone Main Dist Area Carrier Equip &
Cabling (Routers Backbone Vỹ quip
Horizontal LAN/SAN Switches emarcation)
Trang 46Network Engineer Concerns within the m Data Center
Scalable
— Density of equipment, cabinets, frames
— Fast and Accurate Moves, Adds, and Changes Thermal
— Problematic in most data centers (designed before Blade Servers)
— Poor Air Flow is the largest concern
Reliability/ Uptime
— Can't afford any downtime even during expansion
How can ADC positively impact each area of concern?
46
Trang 47Foundation of Structured Cabling ue
¢ Four Key Elements
— Bend radius protection
— Cable and connector access
— Intuitive cable routing paths
— Physical protection
Trang 48Entrance Room Detail „
— Front accessible fiber panel with integrated termination and splice access
— Manages fiber terminations, cable, and splices in a front access solution that can be wall or cabinet/rack mounted
— Integrated cable management and patented angled fiber adapters promote the Four Fundamentals of fiber cable management
Trang 49Fiber Management "oe
Trang 50Main Distribution Area Detail „
¢ IrueNet Fiber Panel
— 5 RU Fiber Patch Panel, Angle Left/Right with integrated and
removable Vertical Cable Guides, with rear accessible termination and slack management
— Aggregates and manages jumpers from servers and SANs Also
accommodates backbone cabling for field termination or splice scenarios
— Supports the Four Fundamentals of fiber cable management, with
Patented Angled Adapters that remove strain from fiber cables and compatibility with Glide Cable Management
Trang 51Main Distribution Area Detall
¢ MTP Solutions
— 12 and 24 fiber MTP-LC Cassettes
— Compatible with three standard TFP Chassis’
— ldeally paired with 12 Fiber Microcable
51
Trang 52
High Density Fibre Distribution Frames
— Ideal for Equipment Rooms and Data Centers
— Superior Cable Management
Trang 53
¢ TrueNet Laser Optimized Fiber
— Laser Optimized 50 micron Multimode Glass enhanced for 10 Gigabit transmission over distances up to 300 or 550 meters
— Greater bandwidth allows 10 Gigabit Ethernet signals to travel
greater distances than Multimode glass designed only for LED
transmission
— Proprietary glass design uses process optimized for Multimode
transmission using VCSELs
53
Trang 54Backbone Cabling Detail "hc
¢ FiberGuide
— The most widely deployed fiber optic raceway system for critical networks
— Provides bend radius protection, physical protection, and segregation of fiber optic jumpers
— Excels at all the key attributes of a good Fiber Raceway solution:
Trang 55
Equipment Distribution Area Detall ue
¢ TracerLight Fiber Jumpers
— Fiber Patch Cord with integrated visual identification system
— Flashing LED can be activated on both ends of jumper via handheld power source
— 100% Accurate identification of fiber circuits resulting in no
downtime due to incorrect circuit identification
- Ses =
= aes Es
55